automatic abstraction of imaging observations with their ...automatic abstraction of imaging...

9
Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt, 1 JaA Lipson, 2 Utku Senol, 3 Daniel L Rubin 4,5 1 Department of Biostatistics and Medical Informatics, Akdeniz University Faculty of Medicine, Antalya, Turkey 2 Department of Radiology, Stanford University, Stanford, California, USA 3 Department of Radiology, Akdeniz University Faculty of Medicine, Antalya, Turkey 4 Department of Radiology, Stanford University, Stanford, California, USA 5 Department of Medicine (Biomedical Informatics Research), Stanford University, Stanford, California, USA Correspondence to Dr Daniel L Rubin, Department of Radiology, Stanford University, Richard M. Lucas Center, 1201 Welch Road, Ofce P285, Stanford, CA 94305-5488, USA; [email protected] Received 26 May 2014 Revised 11 August 2014 Accepted 4 September 2014 To cite: Bozkurt S, Lipson JA, Senol U, et al. J Am Med Inform Assoc Published Online First: [ please include Day Month Year] doi:10.1136/amiajnl- 2014-003009 ABSTRACT Background Radiology reports are usually narrative, unstructured text, a format which hinders the ability to input report contents into decision support systems. In addition, reports often describe multiple lesions, and it is challenging to automatically extract information on each lesion and its relationships to characteristics, anatomic locations, and other information that describes it. The goal of our work is to develop natural language processing (NLP) methods to recognize each lesion in free-text mammography reports and to extract its corresponding relationships, producing a complete information frame for each lesion. Materials and methods We built an NLP information extraction pipeline in the General Architecture for Text Engineering (GATE) NLP toolkit. Sequential processing modules are executed, producing an output information frame required for a mammography decision support system. Each lesion described in the report is identied by linking it with its anatomic location in the breast. In order to evaluate our system, we selected 300 mammography reports from a hospital report database. Results The gold standard contained 797 lesions, and our system detected 815 lesions (780 true positives, 35 false positives, and 17 false negatives). The precision of detecting all the imaging observations with their modiers was 94.9, recall was 90.9, and the F measure was 92.8. Conclusions Our NLP system extracts each imaging observation and its characteristics from mammography reports. Although our application focuses on the domain of mammography, we believe our approach can generalize to other domains and may narrow the gap between unstructured clinical report text and structured information extraction needed for data mining and decision support. INTRODUCTION The interpretation of mammography images is challenged by variability among radiologists in their assessment of the likelihood of malignancy given the abnormalities seen in mammography reports, and methods to improve radiologist performance are needed. 13 One approach to reducing radiolo- gist variation is to standardize the vocabulary used in mammography reports. The American College of Radiology (ACR) developed the Breast Imaging-Reporting and Data System (BI-RADS). Terms called BI-RADS descriptorscan be used by radiologists to describe breast density, lesion fea- tures, impression, and recommendations in mam- mography reports. 46 While adoption of BI-RADS can help to stand- ardize the vocabulary of reports and reduce the variation in mammography reporting, it is not a decision support system (DSS) for mammography, although others have used it as a key ingredient of DSS. 78 For example, several works have demon- strated that statistical machine learning models such as neural networks and Bayesian networks can be built using BI-RADS descriptors and clinical data as inputs. 79 Although preliminary work to develop DSS for mammography using standardized vocabulary to describe the imaging features is promising, few DSS for mammography have been adopted in clinical prac- tice, likely due to the challenge of interfacing DSS with the clinical workow. DSS for mammography can disrupt the workow, 7 1012 since these systems require the radiologist to enter their observations in a separate interface, which duplicates the activity of generating the radiology report. Even if DSS are inte- grated into a structured reporting system, radiologists generally nd it most efcient to produce reports as narrative texts rather than composing reports using structured reporting interfaces. On the other hand, recording the imaging features in unstructured narra- tive text prevents the use of reports directly as input to DSS, since the report content is not structured and machine-understandable. Natural language processing (NLP) techniques for information extraction could acquire the struc- tured information from reports needed to provide the inputs to DSS, 13 assuming all pertinent imaging features are recorded in the report. In addition to providing extracted data as inputs for DSS, such NLP methods could improve interoperability among different medical information systems by providing consistent, structured data, enabling epi- demiological research and clinical studies. The goal of this work is to develop and evaluate NLP methods to extract information on lesions and their corresponding imaging features (imaging observations with their modiers) from free-text mammography reports, with the ultimate goal of providing inputs to DSS to guide radiology practice and to reduce variability in mammography interpretations. BACKGROUND Information extraction from free text is a key task for our work. A number of NLP systems have been developed for information extraction from clinical records. 1421 In general, these systems annotate syn- tactic structures (e.g., sections, sentences, phrases (chunks), tokens (words), and their parts-of-speech), perform named entity recognition, map spans of text to concepts (such as drugs, pathologies, treat- ments, etc.) from a controlled vocabulary or ontol- ogy, and identify the negation context of named entities. 2229 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:19. doi:10.1136/amiajnl-2014-003009 1 Research and applications Journal of the American Medical Informatics Association Advance Access published November 13, 2014

Upload: others

Post on 23-Feb-2020

5 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

Automatic abstraction of imaging observations withtheir characteristics from mammography reportsSelen Bozkurt,1 Jafi A Lipson,2 Utku Senol,3 Daniel L Rubin4,5

1Department of Biostatisticsand Medical Informatics,Akdeniz University Faculty ofMedicine, Antalya, Turkey2Department of Radiology,Stanford University, Stanford,California, USA3Department of Radiology,Akdeniz University Faculty ofMedicine, Antalya, Turkey4Department of Radiology,Stanford University, Stanford,California, USA5Department of Medicine(Biomedical InformaticsResearch), Stanford University,Stanford, California, USA

Correspondence toDr Daniel L Rubin,Department of Radiology,Stanford University, RichardM. Lucas Center, 1201 WelchRoad, Office P285, Stanford,CA 94305-5488, USA;[email protected]

Received 26 May 2014Revised 11 August 2014Accepted 4 September 2014

To cite: Bozkurt S,Lipson JA, Senol U, et al. JAm Med Inform AssocPublished Online First:[please include Day MonthYear] doi:10.1136/amiajnl-2014-003009

ABSTRACTBackground Radiology reports are usually narrative,unstructured text, a format which hinders the ability toinput report contents into decision support systems. Inaddition, reports often describe multiple lesions, and it ischallenging to automatically extract information on eachlesion and its relationships to characteristics, anatomiclocations, and other information that describes it. Thegoal of our work is to develop natural languageprocessing (NLP) methods to recognize each lesion infree-text mammography reports and to extract itscorresponding relationships, producing a completeinformation frame for each lesion.Materials and methods We built an NLP informationextraction pipeline in the General Architecture for TextEngineering (GATE) NLP toolkit. Sequential processingmodules are executed, producing an output informationframe required for a mammography decision supportsystem. Each lesion described in the report is identifiedby linking it with its anatomic location in the breast. Inorder to evaluate our system, we selected 300mammography reports from a hospital report database.Results The gold standard contained 797 lesions, andour system detected 815 lesions (780 true positives, 35false positives, and 17 false negatives). The precision ofdetecting all the imaging observations with theirmodifiers was 94.9, recall was 90.9, and the F measurewas 92.8.Conclusions Our NLP system extracts each imagingobservation and its characteristics from mammographyreports. Although our application focuses on the domainof mammography, we believe our approach cangeneralize to other domains and may narrow the gapbetween unstructured clinical report text and structuredinformation extraction needed for data mining anddecision support.

INTRODUCTIONThe interpretation of mammography images ischallenged by variability among radiologists in theirassessment of the likelihood of malignancy giventhe abnormalities seen in mammography reports,and methods to improve radiologist performanceare needed.1–3 One approach to reducing radiolo-gist variation is to standardize the vocabulary usedin mammography reports. The American Collegeof Radiology (ACR) developed the BreastImaging-Reporting and Data System (BI-RADS).Terms called ‘BI-RADS descriptors’ can be used byradiologists to describe breast density, lesion fea-tures, impression, and recommendations in mam-mography reports.4–6

While adoption of BI-RADS can help to stand-ardize the vocabulary of reports and reduce thevariation in mammography reporting, it is not a

decision support system (DSS) for mammography,although others have used it as a key ingredient ofDSS.7 8 For example, several works have demon-strated that statistical machine learning models suchas neural networks and Bayesian networks can bebuilt using BI-RADS descriptors and clinical data asinputs.7–9

Although preliminary work to develop DSS formammography using standardized vocabulary todescribe the imaging features is promising, few DSSfor mammography have been adopted in clinical prac-tice, likely due to the challenge of interfacing DSSwith the clinical workflow. DSS for mammographycan disrupt the workflow,7 10–12 since these systemsrequire the radiologist to enter their observations in aseparate interface, which duplicates the activity ofgenerating the radiology report. Even if DSS are inte-grated into a structured reporting system, radiologistsgenerally find it most efficient to produce reports asnarrative texts rather than composing reports usingstructured reporting interfaces. On the other hand,recording the imaging features in unstructured narra-tive text prevents the use of reports directly as inputto DSS, since the report content is not structured andmachine-understandable.Natural language processing (NLP) techniques

for information extraction could acquire the struc-tured information from reports needed to providethe inputs to DSS,13 assuming all pertinent imagingfeatures are recorded in the report. In addition toproviding extracted data as inputs for DSS, suchNLP methods could improve interoperabilityamong different medical information systems byproviding consistent, structured data, enabling epi-demiological research and clinical studies.The goal of this work is to develop and evaluate

NLP methods to extract information on lesions andtheir corresponding imaging features (imagingobservations with their modifiers) from free-textmammography reports, with the ultimate goal ofproviding inputs to DSS to guide radiology practiceand to reduce variability in mammographyinterpretations.

BACKGROUNDInformation extraction from free text is a key taskfor our work. A number of NLP systems have beendeveloped for information extraction from clinicalrecords.14–21 In general, these systems annotate syn-tactic structures (e.g., sections, sentences, phrases(chunks), tokens (words), and their parts-of-speech),perform named entity recognition, map spans oftext to concepts (such as drugs, pathologies, treat-ments, etc.) from a controlled vocabulary or ontol-ogy, and identify the negation context of namedentities.22–29

Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009 1

Research and applications Journal of the American Medical Informatics Association Advance Access published November 13, 2014

Page 2: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

Few previous works have addressed information extractionfrom mammography reports.30–34 Mammography reports areunique in several respects compared to other types of clinicalnarrative texts, since they contain specialized types of namedentities (BI-RADS descriptors) and unique information extrac-tion goals. Specifically, it is necessary to identify all the imagingobservations and their relationships (e.g., modifiers such asimaging observation characteristics, and contexts such as neg-ation and breast location). It is particularly important to associ-ate the descriptions of imaging observations with the lesions towhich they refer, since there are often multiple lesions in amammogram. In this regard, mammography reports are gener-ally not directly amenable for use in currently available medicalinformation extraction systems, which were not developed spe-cifically for these unique information extraction tasks. Nassifet al32 developed a system to detect BI-RADS concepts in mam-mography reports, but did not consider relationships betweenconcepts. Likewise, Jain et al used MedLEE to encode the clin-ical information in chest and mammogram reports to identifysuspected tuberculosis and breast cancer.31 35 Since MedLEEwas not optimized for the task of correlating body locations andvisual abnormalities, Sevenster et al36 developed a system basedon MedLEE, which extracts clinical findings and body locationsfrom radiology reports and correlates them. They found thatMedLEE occasionally failed to extract breast locations with mul-tiple modifiers, such as upper inner quadrant of the left breast,although they added the BI-RADS body location phrases toMedLEE’s vocabulary. They concluded that MedLEE’s gram-matical rules need to be updated to accommodate this type ofbody location; however, as MedLEE is not open source, onecannot directly modify MedLEE grammar.36

A particularly important limitation of the NLP systems devel-oped to date is that they do not generally recognize and disam-biguate multiple lesions that are described in the same breast,and associate all of the characteristics of each lesion described inthe report with the particular lesion to which they refer.37 Thisis a very important problem in mammography reports, since aconcept (e.g., ‘mass’) is often used multiple times, referring toeither the same or different lesions. Hence, it is important toidentify whether mentions of concepts actually refer to thesame lesion or to different lesions. For instance, figure 1 showsan example report describing two lesions (two different masses)

having different characteristics such as margin, stability, andshape. To provide accurate input for a DSS from this kind ofreport, it is important for the NLP system to recognize the dif-ferent masses and each of their characteristics so as to produce acomplete information frame suitable for input into a DSS, asshown in the output section of figure 1.

Our work tackles the challenge of recognizing mentions ofbreast lesions in mammography reports, extracting their relation-ships, and associating the extracted information with the respect-ive lesions. Our methods recognize named entities and relationspertinent to mammography, and also recognize and link descrip-tions of lesions to the particular lesions to which they refer. Ourapproach produces a succinct, complete information frame foreach lesion, its observations, characteristics, and location in amammogram, as shown in table 1 and figure 1. Such structuredsummaries could be directly input into a DSS to guide physicianinterpretation of the imaging findings.

MATERIALS AND METHODSA common strategy in developing NLP systems is to decomposethe processing strategy into several subtasks, with each subtaskbeing executed sequentially before subsequent higher-level tasksare executed.37 We decomposed our NLP problem into a set ofsuch subtasks, and we created a sequential processing pipelinecomprising a set of sequentially executed processing modules toautomatically recognize the pertinent named entities in mam-mography reports and the relationships among them needed togenerate inputs for DSS. Our pipeline is shown in figure 2.

Since a mammography report can describe multiple lesions,associating each lesion with its respective characteristics can bevery challenging. We developed a strategy to tackle this chal-lenge by using the anatomic locations of lesions to disambiguatelesions and their associated characteristics, since it is veryunlikely that two abnormalities would occur in precisely thesame anatomic location.

Knowledge representationOur information extraction task focuses on recognizing threesemantic types of named entities: imaging observations, theircharacteristics (modifiers of imaging observations), and anatomiclocations of imaging observations. Imaging observations areterms that refer to abnormalities in the breast, such as masses and

Figure 1 Example mammography report describing two different masses and the ideal output from a natural language processing system toextract the information suitable for input to a decision support system.

2 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009

Research and applications

Page 3: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

calcifications. Imaging observation characteristics (also calledmodifiers) are terms that modify imaging observations to describetheir features, such as ‘spiculation’ to describe the margins of amass in the breast. Anatomic locations are regions in whichabnormalities are located, such as the ‘upper outer quadrant ofthe right breast.’ We thus pre-defined the following semantictypes of named entities and their relations: Imaging Observation,Imaging Observation Modifiers, Location, and LocationModifiers. The Location and Location Modifiers are used toassociate observations with an anatomic location in the breast,represented as extracted location relationships (table 1).

We used the BI-RADS controlled terminology4 and a subsetRadLex,38 which contains terms for anatomic locations (theseare not contained in BI-RADS), to provide the preferred namesof all named entities. Since BI-RADS was not available in astructured format, we created a simple ontology structure of thisterminology for our system (‘BI-RADS ontology’). The BI-RADS

ontology comprises an is-a hierarchy of entities with the follow-ing attributes: the preferred name of each term, synonyms, andacronyms (figure 3). We developed the list of synonyms andabbreviations by inspecting a set of 100 reports in addition to apredefined list compiled by experts.

Processing pipelineWe built our processing pipeline in the GATE (GeneralArchitecture for Text Engineering) NLP toolkit.39 GATE is anopen source framework and also provides several configurableprocessing resources to parse words, numbers, and punctuation(Tokenizer), and to annotate documents based on keyword lists(Gazetteer lists). In addition, GATE provides a regular expres-sion pattern matching engine called the Java AnnotationPatterns Engine ( JAPE), which is a multi-pass regular expressionparser tightly integrated with GATE’s document annotation rep-resentation.39 In our pipeline (figure 2), sequential processing

Table 1 Entity types and modifiers (relationships) to be extracted from mammography reports

Entity type Description Example

Imaging Observation Conditions, problems, etc. The breast, overall, demonstrates no focal dominant mass, architecturaldistortion, or suspicious calcifications.

Location Anatomical structure or location of an Imaging Observation The breast, overall, demonstrates no focal dominant mass,architectural distortion, or suspicious calcifications.

Location ModifiersLateralityClock face locationDepthQuadrant locationRegionDistance

Relates an anatomic entity to its sidedness: right, left, bilateralright breast12 o’clock positionanterior depthupper breastcentral region1 cm from the nipple

Imaging ObservationModifiersStabilityMargin signalShape signalDensity signalSize signal

Relates an Imaging Observation to other information about theImaging Observation stable focal asymmetric density

spiculated massirregularly shaped massbreast tissue is largely fatty10×6×10 mm mass

Figure 2 Natural language processing information extraction processing pipeline.

Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009 3

Research and applications

Page 4: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

modules are executed, producing an output of structured infor-mation required for a mammography DSS. The various process-ing modules that accomplish each of the various subtaskscomprising our system are described below, and correspond tothe modules shown in figure 1.

Pre-processing (Tokenizer, POS tagger, numbers/measurements/date)This module pre-processes input reports using standard GATEmodules for tokenization of words, numbers and punctuation,part of speech tagging, stemming, annotation of sentences, nounphrases, verb groups, and measurements.

BI-RADS onto-gazetteerThis processing module processes the input text to annotate thetext using terms in the BI-RADS ontology. We used theOnto-Gazetteer GATE component to access the BI-RADS ontol-ogy and to incorporate it into our pipeline. This processingmodule can associate the entities from a specific gazetteer listwith a class in the BI-RADS ontology and is aware of mappingsbetween lists and class IDs. We created Gazetteer lists for eachclass in the BI-RADS ontology that has synonyms and acronymsso that this processing module can match input text to BI-RADSterms in the ontology by both exact match and synonym match.

Section segmentationThis processing module segments input radiology report textinto sections. Although radiology reports are free text, they areactually loosely structured in terms of report sections. Eachsection has its own heading in most cases (e.g., ‘History,’‘Findings,’ ‘Impression,’ etc.). Section segmentation was imple-mented as a JAPE grammar consisting of a set of phases toannotate section headers and the subtext that belongs to thosesection headers. Sometimes a report may not have a sectionheader to indicate, for example, the Findings section. To tacklethat challenge, this processing component implements an add-itional JAPE grammar that implements logic necessary to iden-tify the section. For example, the Findings section ofmammography reports commonly starts with a sentence whichdescribes the breast density. If there is no section header, theJAPE grammar recognizes a sentence that contains the firstmention of ‘breast density’ in the report as the starting point ofthe Findings section. In addition, sentences, noun phrases, andverb groups in each section are associated with the relatedsection heading using JAPE grammar.

Finding and linking Imaging Observation entities, modifiers, andanatomic locations for each lesionThis processing module scans the noun phrases in the Findingssection (since imaging observations are typically recorded in theFindings sections of radiology reports), detects BI-RADS termsto which they correspond, and produces Imaging Observation,Modifier, and Location annotations to produce informationextraction frames (in examples 1 and 2 below, text representinga named entity is shown in square brackets and its semantic typeis shown as a subscript).

Example 1: Information extraction of an imaging observationand its location.

[The breast tissue]Location is [heterogeneouslydense]ImagingObservation

Example 2: Information extraction of an imaging observation,its modifier, location, and location modifier.

There is [a 1 cm oval nodular density]ImagingObservation with[an obscured margin]ImagingObservation_modifier in [the rightbreast]Location in [the anterior depth]Location_modifier

This processing module implements several rules (using regularexpressions in JAPE) to link modifiers with their entities (table 2).

Determining the context of entities: negation, experiencer,and temporal statusClinical conditions can be modified by several contextual prop-erties that are relevant for our information extraction task;ConText40 identifies three contextual values in addition toNegEx’s negation: hypothetical, historical, and experiencer. Forinstance, a query for patients with a diagnosis of pneumoniamay return false positive records as pneumonia is mentionedbut is negated (e.g., ‘ruled out pneumonia’), is experienced by afamily member (e.g., ‘family history of pneumonia’), oroccurred in the past (e.g., ‘past history of pneumonia’).40

We implemented ConText as a processing resource within ourGATE NLP system. We used ConText to determine whether anImaging Observation entity is negated and to determine its tem-poral status. For example, for input text: ‘No significant massesor calcifications or other findings are seen in the right breast,’the output of this module returns abnormalities as negated, andif it is a ‘a previously described nodular density,’ the outputreturns the abnormality as historical.

Figure 3 Breast Imaging-Reporting and Data System (BI-RADS)ontology. The is-a hierarchy is shown and the entity names are theterm preferred names (synonyms are not shown).

4 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009

Research and applications

Page 5: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

Imaging Observation–Location relationship extractionTo associate the Location information with the ImagingObservations, we wrote JAPE grammar rules (table 4) forPatterns 1 and 2 below and added Location information as afeature for every Imaging Observation. The default value ofLocation was assigned as ‘not identified.’

Pattern 1: First, any Location modifier inside the ImagingObservation annotation (it is annotated as a noun phrase) isadded as a Location feature of the annotation (example 3).

Example 3:[a [right]Location breast enhancing focus tissue]ImagingObservation:

class = Mass, location = breast (Laterality=right)

Pattern 2: If the Location modifier is not within the ImagingObservation entity text, to associate an Imaging Observationwith Location information and to add it as a feature to anImaging Observation entity, we allow a maximum of four stopwords (SW) and one verb group (VG) between Location andImaging Observation concepts for expressions below (table 3).

Co-reference resolutionThe same Imaging Observations can be referenced in differentsentences, or information related to one Imaging Observationcan be divided among different sentences. It is important toidentify whether several Imaging Observation mentions actuallyrefer to the same Imaging Observation. For example, in the sen-tences in example 4, ‘this mass’ and ‘5 mm oval mass’ refer tothe same Imaging Observation. In other words, these areco-referential, and resolution of such co-references serves thecritical role of linking related information.41

Example 4:In the central right breast, 5 cm posterior to the nipple, thereis [a 5 mm oval mass]Imaging Observation. [This mass]Co-referent iswell circumscribed.

We developed a simple co-reference resolution module foronly Imaging Observation entities. First we tag demonstrativepronouns: this, that, these, and those as Pronoun and theImaging Observation which contains any of these pronouns asco-referent while keeping the Imaging Observation features(class, modifier, etc.). Then, for a three sentence span, we findthe Imaging Observations and their co-referents by using JAPEgrammar to define the expressions below:

Expression 1: {Sentence contains {ImagingObservation.class=A}& {Co-referent.class=A}}

Expression 2: {Sentence contains {ImagingObservation.class=A}}{Sentence contains {Co-referent.class=A}}? (optional)

{Sentence contains {Co-referent.class=A}}Where these patterns matched in the Findings section of a

report, the features of the co-referent were cloned to itsImaging Observation pair, and co-referents were removed. Thisprocess helps to avoid duplications in information extractionfrom the report.42

Extraction of abnormalities for each reportTo extract Imaging Observations that describe multiple abnor-malities, it is essential to relate Imaging Observations to the par-ticular abnormality. Our solution approach to this problem is touse the Location of lesions to identify them uniquely. Goochand Roudsari42 considered the set of all mentions, createdsubsets according to mention class, and within each subset, com-pared pairs of mentions in document order. Similarly, taking theset of all Imaging Observations, we create subsets of ImagingObservation entities according to their class (determined fromthe BI-RADS ontology, e.g., mass, calcification, associated find-ings, other findings, special cases), and within each subset,compare pairs of Imaging Observation entities in documentorder. In addition to the class of Imaging Observations, our

Table 2 Sample rules for linking modifiers with Imaging Observation and Location entities

Rules Example sentences Features of the entities

Rule 1: Any modifier inside the text of an ImagingObservation or Location entity is added as a feature ofthe entity and the modifier is removed.

There is [a 1 cmsize ovalshape nodular density]ImagingObservation ImagingObservation: class=mass,size=1 cm (value=1, unit=cm),shape=oval

Rule 2: If there are only stop words (e.g., in, at, of, with,etc.) or verb group(s) between an entity and its modifier,the modifier is added as a feature to the entity and themodifier is removed.

There is [a 1 cmsize ovalshape nodular density]ImagingObservation with an[obscured marginmargin]ImagingObservation_modifier in [the rightlateralitybreast]Location in [the anterior depthdepth]Location_modifier

ImagingObservation: class=mass,size=1 cm (value=1, unit=cm),shape=oval, margin=obscuredLocation: class=breast, laterality=right,depth=anterior

Rule 3: After rule 1 and 2 are applied, if a Findingssentence contains only one entity and its modifier(s), themodifier(s) are added as the feature(s) to the entity(example 6) and the modifier is removed.

[The mass]ImagingObservation in the central region of the left breast is[stablestability]ImagingObservation_modifier

ImagingObservation: class=mass,stability=stable

Table 3 Grammars to detect Location and Imaging Observation relations: stop word (SW), verb group (VG), zero or one (?)

Grammar Example sentences Features of the entities

{Location}{SW}?{SW}?{VG}?{SW}?{SW}?{Imaging Observation}

…right breast demonstrates fibroglandular tissue.In the central right breast approximately 3 o’clock there isa 5 mm oval mass.

ImagingObservation: class=breast_density, LocatedIn=breast: laterality=rightImagingObservation: class=mass, LocatedIn=breast: laterality=right,clock_face_location=3 o’clock, region=central

{Imaging Observation}{SW}?{SW}?{VG}?{SW}?{SW}?{Location}

There is a right breast enhancing focus in the 8 o’clockposition posteriorly.

ImagingObservation: class=mass, LocatedIn=breast: laterality=right,clock_face_location=8 o’clock, depth=central

Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009 5

Research and applications

Page 6: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

comparison criterion for the Imaging Observation is ‘Location.’For example, the first ImagingObservations.class=Mass is testedagainst all succeeding occurrence of ImagingObservations.class=Mass entities in the text, the second against the third,fourth, etc. Comparisons are made based on Location informa-tion for each Imaging Observation. At the end of this processwe identified only the Imaging Observations which differ fromeach other (example 5 in table 5).

In some reports (example 6 in table 6), the Location informa-tion of two observations is exactly the same, but the sizes of themasses are different. Therefore as a second criterion, if Locationinformation is the same and both observations have size fea-tures, their sizes are compared (i.e., we use size to disambiguatethe lesions in the same location).

DatasetIn order to develop and evaluate our system, we selected a setof 500 mammography reports from a hospital report database,which reports were created using a structured reporting applica-tion (PenRad, Buffalo, Minnesota, USA). We selected thereports in this collection so that it had a balanced representationof the different BI-RADS classification codes (varying frombenign to highly suspicious for malignancy). The reportingsystem produced reports with both structured data fields andfree text; thus, it is a suitable gold standard, since the free-textreport has associated radiologist-vetted structured data fields.However, the correspondence between structured data fields inthe database and the free-text report content may not alwaysagree, because the structured database is populated by the initialdraft report created by the radiologist, while corrections or editsto the report during report review had been done only in thetext report. The corresponding structured database was thus notupdated if edits to the text report were made. We dealt with thislimitation by having an expert mammographer review thereports and the corresponding database entries in those reports

that we used for evaluating our system, and recording thecorrect structured data entries for each report.

We used 200 of the 500 reports for iterative developmentand refinement of the NLP system; the 300 not used for devel-oping the system were selected and held out as an independenttest set for the final evaluation of the NLP system. A mammo-grapher reviewed these 300 reports and database entries toestablish the correct set of database entries for use in comparingagainst the study results. Thus, this 300-report dataset com-prised the gold standard for our evaluation.

EvaluationWe used the 300 mammography report held-out dataset to evalu-ate our NLP system. We compared the information extractionframes produced by our system with the structured report infor-mation determined by the radiologist who reviewed the reports.We evaluated the completeness of our information extraction bycalculating the precision, recall, and the F measure for extractingthe correct information frames (each lesion and its character-istics). In doing these assessments, we counted complete matches(all data in the information frame is correct), partial matches (atleast one modifier of an imaging observation describing a lesionis incorrect or absent), and non-matches (description of a lesionis completely missed). We calculated precision as the number ofcorrectly identified items divided by the total number of itemsidentified by our system, and recall as the number of correctlyidentified items divided by the total number of correct items).43

We calculated the F measure as F=(2×precision×recall)/

Table 5 Example 5

Text Imaging Observations

[A right breast enhancing focus] in [the8 o’clock position posteriorly] whichmeasures [5×4 mm (AP×ML)]. In [the 12o’clock position of the right breast atmid depth] there is [an oval 9×3 mm(AP×ML) mass].

1. Imaging Observation:class=mass, size=5×4 mm

2. Location: laterality=right, clockface location=8, depth=posterior

3. Imaging Observation:class=mass, size=9×3 mm

4. Location: laterality=right, clockface location=12, depth=middle

Table 6 Example 6

Text Imaging observations

There is [a 2.5 cm round mass with acircumscribed margin] in [the rightbreast at 12 o’clock in the anteriordepth]. There also is [a 1.5 cm ovalmass with a circumscribed margin] in[the right breast at 12 o’clock in theanterior depth]. There is [a 1.8 cmround mass with a circumscribedmargin] in the [left breast in theanterior depth central to the nipple].Compared to previous films this massis increased in size. There also is [a1.4 cm oval mass with an obscuredmargin] in the [left breast in theanterior depth of the inferior region].Compared to previous films this massis increased in size.

1. Imaging Observation: class=mass,size=2.5 cm, location=(laterality=right, clock facelocation=12, depth= anterior)

2. Imaging Observation: class=mass,size=1.5 cm, location=(laterality=right, clock facelocation=12, depth=anterior)

3. Imaging Observation: class=mass,size=1.5 cm, location=(laterality=left, depth=anterior,region=central), stability=increase

4. Imaging Observation: class=mass,size=1.4 cm, location=(laterality=left, depth=anterior,region=inferior), stability=increase

Table 4 Accuracy in extracting complete descriptions of breast lesions (Imaging Observation, characteristic, and location)

Results (counts using our system) Lesions (without characteristics) Lesions (with characteristics) Size Stability Shape Margin Density

Full match 780 723 184 452 232 165 104Partial match – 57 –

Non-match 17 FN35 FP

– 1 FP5 FN

6 FP18 FN

2 FN 4 FN 1 FN

Precision rate (%) 95.7 88.7 99.4 98.6 – – –

Recall rate (%) 97.8 90.7 97.3 96.1 99.1 97.6 99Total (gold standard set) 797 797 190 476 234 169 105

Results reported in terms of full, partial, and non-matches to the actual information frames determined in the gold standard of 797 lesions.FN, false negative; FP, false positive.

6 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009

Research and applications

Page 7: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

(precision+recall).43 We computed these three metrics at thelevel of exact match. In addition, true positive and false positiverates were recorded.

RESULTSThe gold standard set of 300 reports contained a total of 797reported lesions. In 102 of the reports, the expert mammogra-pher noted a deficiency in the structured data entries in thereporting application database (cases having duplicated, missed,or deficient reporting of lesions and their characteristics due toedits that were made to the text reports but were not reflectedin the reporting application’s database). Only the structureddata entries determined by the expert mammographer wereused in assessing the performance of our NLP system.

Our system detected 815 lesions; 780 were true positives, 35were false positives, and 17 were false negatives (lesionsreported in the gold standard set that were not detected by thesystem). There were 57 partially matched cases among thedetected lesions. In addition, there were 12 cases where calcifi-cation type was not detected correctly and 8 cases where breastdensity was not detected.

For the goal of perfect match information frame extraction,the precision of extracting the mentions of ImagingObservations with their modifiers by our system was 94.9, recallwas 90.9, and the F measure was 92.8. Table 4 shows the finalresults for full, partial, and non-match cases with precision andrecall rates based on number of lesions in the300-mammography report gold standard.

DISCUSSIONIn this paper, we describe an NLP system to extract structuredimaging observation information from mammography reports toproduce an output comprising each imaging observation with itsassociated characteristics and anatomic locations. This informa-tion is necessary to drive a DSS, such as a Bayesian networkmodel, to predict the probability of malignancy of a given suspi-cious finding in a mammogram.7 8 The recall and precision ofour system for achieving perfect information extraction for thistask were 92.9% and 89.4%, respectively, which are reasonablygood results, and if generalizable to other reports, could makeour approach useful for information extraction for driving deci-sion support in mammography with narrative reporting. We arecurrently conducting an evaluation study to assess the impact ofimperfect recall and precision of information extraction on theoutput of a DSS.

A few studies have addressed extracting information frommammography reports, including string matching methods andrule-based methods.30–34 36 Among them, the study mostsimilar to ours was conducted by Nassif et al in 2009.32 Theresults with Nassif ’s system are inferior to ours, likely since itdoes not identify whether multiple mentions of an imagingobservation refer to the same lesion or to a different lesion.Therefore, it overestimates the number of lesions present andcannot link the characteristics of a lesion to the particular lesionbeing described; the system extracts imaging observations basedon how many times the imaging observations were mentionedin the text. Our system recognizes the context of the mentionsof lesion characteristics by identifying the different lesions inthe same (or different) breast and connecting the characteristicsto them. In addition, Nassif ’s system only detects negation foreach imaging observation but does not check the temporality ofan imaging observation. That system only extracts size informa-tion for the last imaging observation in the report. Althoughrelationships were not extracted, their system gave reasonably

accurate results in extracting the characteristics of the imagingobservations.32

It is difficult to establish a large, robust gold standard corpusof reports for our work because most radiology reports areunstructured, and it is costly and time-consuming to readreports and record the structured data by hand. We thus lever-aged a large sample of data coded by a structured reportingapplication, in which the data entry by the radiologist is struc-tured as part of the reporting workflow, and the final output ofthe report is free text. This enabled us to create a large datasetfor building and refining our system, although there were somelimitations to this approach. First, since the bulk of the text isautomatically generated by the reporting system, the variabilityin narrative reporting style was not represented, which couldmake our results seem better than when applied to reports notgenerated using this system. On the other hand, the radiologistsedited this system-produced text in finalizing the report inapproximately a third of cases, which likely added a substantialamount of variety to the language of the reports. Another limi-tation is that if the radiologist edits the text report, correspond-ing changes are not made in the structured database(a functional limitation of the particular reporting system).Upon review of the structured reporting application’s databaseof structured entries and its text reports, we found a number ofcases in which the database lacked some findings. Therefore, toestablish the dataset for use in our final evaluation, we asked anexpert mammographer to review all 300 reports in our testingreport corpus and confirm the accuracy of the extracted findingsrecorded in the report database. Although a larger test set mightbe desirable, we believe this was sufficiently large for this initialevaluation of our system. Future studies with larger and morediverse reports (specifically not generated by a structured report-ing application) would be helpful to confirm the accuracy andgeneralizability of our approach.

Another limitation in our work is that although the regularexpressions we used in our system seemed to work well fordetecting BI-RADS descriptors, this approach is specific toBI-RADS and might not work as well if radiologists use othertypes of terminologies in reporting mammograms. BI-RADS is astandardized terminology and recommended for reporting allmammography, so this is not likely a substantial limitation. Inaddition, our system is extensible, and deficiencies in entity rec-ognition could likely be overcome by adding the appropriateregular expressions to our processing modules. Our system wasdeveloped in the modular GATE platform, and we encapsulatedthe domain-specific processing in JAPE processing resources,whose regular expressions are compact and human-readable. Wethus believe that such extensions are practical in the future.

Although we included the BI-RADS ontology in our system,in our system we used it simply for term lookup, and we didnot leverage the subsumption relations in the ontological ter-minology structure. We did use the ontological structure as avisual resource in creating our system; displaying the class hier-archy of the ontology was helpful in defining how to assignclasses to terms in the text as we developed our JAPE grammars.We are considering extensions to our system in order to usethe ontological knowledge structure in the future to enable thesystem to extract much more meaningful information about thetext.

In radiology reports, some sentences do not describe findings,and eliminating them can improve system performance. Oursystem does section segmentation to eliminate non-relevant sec-tions of the report, except for the Findings section which con-tains the key sentences that are generally relevant to our

Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009 7

Research and applications

Page 8: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

information extraction task. Of course, even within Findingssection of reports, there may be irrelevant sentences. However,the information extraction logic we developed is designed toidentify and extract specific information about imaging findings.Thus, sentences about past examinations, relations between find-ings, or other non-relevant information in text will not beextracted, since it will not meet the criteria for our informationextraction rule logic.

There were some cases in which our system failed to extractthe correct information from mammography reports. To detectnegation and temporality, we used the ConText plug-in, whichin a recent study was reported to have a 97% recall and preci-sion for negation, and 67.4–82.5% recall and 74.2–94.3% pre-cision for temporality (when assigning historical or hypotheticalvalues).44 Although ConText was generally effective in oursystem, we still experienced some problems detecting negationand temporality of concepts. For example, in the sentence,‘A coarse calcification noted in this region was removed,’ oursystem extracted the calcification as being present, because ‘wasremoved’ was not accepted as a negation term. In anotherexample sentence, ‘The previously noted dense spiculated massat the 8 o’clock position appears considerably less apparentbeing decreased in both density and size,’ our system failed toextract the mass because it is mentioned in a historical context.In addition, our system does not detect historical mentions ofmasses or description of their stability over time, which accountsfor failing to extract this example correctly. We could improvethe system in the future by adding a module to recognize histor-ical lesions and description of their stability. Another limitationof our system is related to the span of sentences consideredwhen searching for imaging observations and their modifiers.For co-reference resolution, we assume that a span of three con-secutive sentences is sufficient to link all co-referents and theirmodifiers to their related imaging observations. However, in afew reports we saw that the co-referents could be found in morethan a three-sentence span. In addition, our system does notextract the ‘count’ or ‘plurality’ information for abnormalities.Therefore for the text, ‘several nodular densities’ or ‘two 8 mmintramammary lymph node were reported,’ our system extractedonly one density and one lymph node. An additional limitationis that some of the assumptions we make about report structurein building our system may not generalize to reports created atother institutions; for example, it is possible that sectionheaders may be omitted. To overcome this in our system, ifthere is no section header, a sentence that contains the firstmention of ‘breast density’ is accepted and considered to be thestarting point of the Findings section in the report. We makethis assumption because there is a convention in mammographyreporting that the first sentence in the Findings section describesthe breast density; however, it is possible that in some practicesthis may not be the case, and in such cases, our system mightfail to recognize Findings section if the section header isomitted. Our system is extensible, and such variances could beaccommodated in future versions if we encounter them inpractice.

Our approach is unique in a few respects. A commonapproach in information extraction systems is to create rulesthat find related concepts based on a 5- or 10-word span.32 Onthe other hand, we developed rules based on noun and verbphrases to link the characteristics to the concepts. Our systemalso recognizes and extracts the contexts and relationships ofimaging observations to enable us to extract information abouteach lesion in mammography reports, and to associate thecharacteristics of each lesion to the respective lesions described

in the reports. This is critical for DSS, since such systemsprovide decision support on a per lesion basis. In addition, therules we developed to build our system focus on concepts andrelationships (imaging observations, locations, size, etc.) that areused in radiology domains other than breast imaging. Thus, webelieve that our system may be extended to other domains, andwe plan to expand our work to other types of radiologyreports.

Our information extraction system is based on rule-basedmethods. It is possible we could have created similar functional-ity using statistical methods. Some clinical NLP work is shiftingfrom rule-based or expert-based approaches to statisticalmachine-learning methods, and investigating various useful fea-tures and appropriate algorithms.28 37 Both approaches havetheir advantages and disadvantages.28 33 45 46 Rule-basedmethods are easier to interpret, whereas statistical methods maybe more robust to noise in the unstructured data (assuming theyare built from large, representative corpora). In general, rule-based systems are more useful in narrow domains where humaninvolvement is both essential and available.47 Since mammog-raphy reporting is a narrow domain, we developed our systemusing rule-based methodology. Combining both rule-based andstatistical methods in the future could provide even betterresults and generalizability.

Finally, since our ultimate goal is to extract informationneeded for decision support, we are planning to use the outputof our system as an input for a mammography DSS and toevaluate its effectiveness. We will also study the impact ofimperfect precision and recall of our system on the quality ofdecision support delivered by a DSS.

CONCLUSIONThere is a tremendous amount of data in unstructured free-textmammography reports, and extracting the structured contentabout the imaging observations and characteristics of each lesionreported could be useful to enable decision support. We devel-oped an NLP system to extract information in mammographyreports needed for input into DSS. The main contribution ofthis paper is the text processing methods that extract imagingobservations and their attributes from mammography reports inthe setting of multiple breast lesions, relating the observationsand characteristics to each lesion. We believe our approach willhelp to narrow the gap between unstructured clinical report textand structured data capture needed for decision support andother applications. In future work we will incorporate our NLPsystem into a DSS.

Collaborators Hakan Bulu, Department of Radiology, Stanford University,Stanford, California, USA.

Contributors DLR conceived and directed the project. SB and DLR designed andevaluated the NLP system, analyzed and evaluated the data, wrote the paper, andhad full access to all of the data in the study and take responsibility for the integrityof the data and the accuracy of the data analysis. JAL and US helped to create thegold standard dataset and the annotation schema. JAL and US contributed equally.All authors reviewed and approved the manuscript.

Funding This work was supported by a Scientific and Technological ResearchCouncil of Turkey grant (number 2214-A).

Competing interests None.

Provenance and peer review Not commissioned; externally peer reviewed.

REFERENCES1 Elmore JG, Wells CK, Lee CH, et al. Variability in radiologists’ interpretations of

mammograms. N Engl J Med 1994;331:1493–9.

8 Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009

Research and applications

Page 9: Automatic abstraction of imaging observations with their ...Automatic abstraction of imaging observations with their characteristics from mammography reports Selen Bozkurt,1 Jafi

2 Jiang Y, Nishikawa RM, Schmidt RA, et al. Potential of computer-aided diagnosis toreduce variability in radiologists’ interpretations of mammograms depictingmicrocalcifications. Radiology 2001;220:787–94.

3 Kerlikowske K, Grady D, Barclay J, et al. Variability and accuracy in mammographicinterpretation using the American College of Radiology Breast Imaging Reportingand Data System. J Natl Cancer Inst 1998;90:1801–9.

4 Liberman L, Menell JH. Breast imaging reporting and data system (BI-RADS). RadiolClin North Am 2002;40:409–30, v.

5 Park CS, Lee JH, Yim HW, et al. Observer agreement using the ACR Breast ImagingReporting and Data System (BI-RADS)-ultrasound, First Edition (2003). Korean JRadiol 2007;8:397–402.

6 Burnside ES, Sickles EA, Bassett LW, et al. The ACR BI-RADS experience: learningfrom history. J Am Coll Radiol 2009;6:851–60.

7 Burnside ES, Rubin DL, Fine JP, et al. Bayesian network to predict breast cancer riskof mammographic microcalcifications and reduce number of benign biopsy results:initial experience. Radiology 2006;240:666–73.

8 Burnside ES, Davis J, Chhatwal J, et al. Probabilistic computer model developedfrom clinical data in national mammography database format to classifymammographic findings. Radiology 2009;251:663–72.

9 Ayer T, Chen Q, Burnside ES. Artificial neural networks in mammography interpretationand diagnostic decision making. Comput Math Methods Med 2013;2013:832509.

10 Wu Y, Giger ML, Doi K, et al. Artificial neural networks in mammography:application to decision making in the diagnosis of breast cancer. Radiology1993;187:81–7.

11 Taylor P. Decision support for image interpretation: a mammography workstation.Inf Process Med Imaging 1995;3:227–38.

12 Stivaros SM, Gledson A, Nenadic G, et al. Decision support systems for clinicalradiological practice—towards the next generation. Br J Radiol 2010;83:904–14.

13 Demner-Fushman D, Chapman WW, McDonald CJ. What can natural languageprocessing do for clinical decision support? J Biomed Inform 2009;42:760–72.

14 Friedman C, Alderson PO, Austin JH, et al. A general natural-language textprocessor for clinical radiology. J Am Med Inform Assoc 1994;1:161–74.

15 Chapman WW, Haug PJ. Comparing expert systems for identifying chest x-rayreports that support pneumonia. Proc AMIA Symp 1999:216–20.

16 Savova GK, Masanz JJ, Ogren PV, et al. Mayo clinical Text Analysis and KnowledgeExtraction System (cTAKES): architecture, component evaluation and applications.J Am Med Inform Assoc 2010;17:507–13.

17 Zeng QT, Goryachev S, Weiss S, et al. Extracting principal diagnosis, co-morbidityand smoking status for asthma research: evaluation of a natural languageprocessing system. BMC Med Inform Decis Mak 2006;6:30.

18 Crowley RS, Castine M, Mitchell K, et al. caTIES: a grid based system for codingand retrieval of surgical pathology reports and tissue specimens in support oftranslational research. J Am Med Inform Assoc 2010;17:253–64.

19 Denny JC, Irani PR, Wehbe FH 3rd, et al. The KnowledgeMap project: developmentof a concept-based medical school curriculum database. AMIA Annu Symp Proc2003:195–9.

20 Hahn U, Romacker M, Schulz S. MEDSYNDIKATE—a natural language system forthe extraction of medical information from findings reports. Int J Med Inform2002;67:63–74.

21 Aronson AR, Lang FM. An overview of MetaMap: historical perspective and recentadvances. J Am Med Inform Assoc 2010;17:229–36.

22 Lacson R, Sugarbaker N, Prevedello LM, et al. Retrieval of radiology reports citing criticalfindings with disease-specific customization. Open Med Inform J 2012;6:28–35.

23 Rubin D, Wang D, Chambers DA, et al. Natural language processing for lines anddevices in portable chest x-rays. AMIA Annu Symp Proc 2010;2010:692–6.

24 Do BH, Wu AS, Maley J, et al. Automatic retrieval of bone fracture knowledge usingnatural language processing. J Digit Imaging 2013;26:709–13.

25 Cheng LT, Zheng J, Savova GK, et al. Discerning tumor status from unstructuredMRI reports—completeness of information in existing reports and utility ofautomated natural language processing. J Digit Imaging 2010;23:119–32.

26 Yetisgen-Yildiz M, Gunn ML, Xia F, et al. A text processing pipeline toextract recommendations from radiology reports. J Biomed Inform 2013;46:354–62.

27 Yetisgen-Yildiz M, Gunn ML, Xia F, et al. Automatic identification of criticalfollow-up recommendation sentences in radiology reports. AMIA Annu Symp Proc2011;2011:1593–602.

28 Meystre SM, Savova GK, Kipper-Schuler KC, et al. Extracting information fromtextual documents in the electronic health record: a review of recent research.Yearb Med Inform 2008;35:128–44.

29 Bashyam V, Taira RK. Indexing anatomical phrases in neuro-radiology reports to theUMLS 2005AA. AMIA Annu Symp Proc 2005:26–30.

30 Burnside E, Rubin D, Strasberg H. Automated indexing of mammography reportsusing linear least squares fit. International congress; 14th Computer assistedradiology and surgery; Cars 2000: Comput Assist Radiol Surg2000;1214:449–54.

31 Jain NL, Friedman C. Identification of findings suspicious for breast cancer based onnatural language processing of mammogram reports. Proc AMIA Annu Fall Symp1997:829–33.

32 Nassif H, Woods R, Burnside E, et al. Information Extraction for Clinical DataMining: A Mammography Case Study. 2009 IEEE International Conference on DataMining Workshops (ICDMW 2009); 2009:37–42.

33 Mykowiecka A, Marciniak M, Kupsc A. Rule-based information extraction frompatients’ clinical data. J Biomed Inform 2009;42:923–36.

34 Esuli A, Marcheggiani D, Sebastiani F. An enhanced CRFs-based system forinformation extraction from radiology reports. J Biomed Inform 2013;46:425–35.

35 Jain NL, Knirsch CA, Friedman C, et al. Identification of suspected tuberculosispatients based on natural language processing of chest radiograph reports. ProcAMIA Annu Fall Symp 1996:542–6.

36 Sevenster M, van Ommering R, Qian Y. Automatically correlating clinical findingsand body locations in radiology reports using MedLEE. J Digit Imaging2012;25:240–9.

37 Nadkarni PM, Ohno-Machado L, Chapman WW. Natural language processing: anintroduction. J Am Med Inform Assoc 2011;18:544–51.

38 Langlotz CP. RadLex: a new method for indexing online educational materials.Radiographics 2006;26:1595–7.

39 Cunningham H, Tablan V, Roberts A, et al. Getting more out of biomedicaldocuments with GATE’s full lifecycle open source text analytics. PLoS Comput Biol2013;9:e1002854.

40 Harkema H, Dowling JN, Thornblade T, et al. ConText: an algorithm for determiningnegation, experiencer, and temporal status from clinical reports. J Biomed Inform2009;42:839–51.

41 Zheng J, Chapman WW, Crowley RS, et al. Coreference resolution: a review ofgeneral methodologies and applications in the clinical domain. J Biomed Inform2011;44:1113–22.

42 Gooch P, Roudsari A. Lexical patterns, features and knowledge resources forcoreference resolution in clinical notes. J Biomed Inform 2012;45:901–12.

43 Maynard D, Peters W, Li Y. Metrics for evaluation of ontology-based informationextraction. International World Wide Web Conference; Edinburgh, UK 2006.

44 Meystre S, Haug PJ. Natural language processing to extract medical problems fromelectronic clinical documents: performance evaluation. J Biomed Inform2006;39:589–99.

45 Liu F, Weng C, Yu H. Natural language processing, electronic health records, andclinical research. In: Richesson RL, Andrews JE. eds. Clinical research informatics.London: Springer, 2012:293–310.

46 Friedman C, Rindflesch TC, Corn M. Natural language processing: state of the artand prospects for significant progress, a workshop sponsored by the NationalLibrary of Medicine. J Biomed Inform 2013;46:765–73.

47 Sarawagi S. Information extraction. Foundations and trends in databases2008;1:261–377.

Bozkurt S, et al. J Am Med Inform Assoc 2014;0:1–9. doi:10.1136/amiajnl-2014-003009 9

Research and applications