Article ID Journal Published Year Pages File Type
516372 International Journal of Medical Informatics 2009 8 Pages PDF
Abstract

PurposeWe assessed the current state of commercial natural language processing (NLP) engines for their ability to extract medication information from textual clinical documents.MethodsTwo thousand de-identified discharge summaries and family practice notes were submitted to four commercial NLP engines with the request to extract all medication information. The four sets of returned results were combined to create a comparison standard which was validated against a manual, physician-derived gold standard created from a subset of 100 reports. Once validated, the individual vendor results for medication names, strengths, route, and frequency were compared against this automated standard with precision, recall, and F measures calculated.ResultsCompared with the manual, physician-derived gold standard, the automated standard was successful at accurately capturing medication names (F measure = 93.2%), but performed less well with strength (85.3%) and route (80.3%), and relatively poorly with dosing frequency (48.3%). Moderate variability was seen in the strengths of the four vendors. The vendors performed better with the structured discharge summaries than with the clinic notes in an analysis comparing the two document types.ConclusionAlthough automated extraction may serve as the foundation for a manual review process, it is not ready to automate medication lists without human intervention.

Related Topics
Physical Sciences and Engineering Computer Science Computer Science Applications
Authors
, , , , , , ,