Skip to main content
Advertisement

Main menu

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • AJNR Case Collection
    • Case of the Week Archive
    • Classic Case Archive
    • Case of the Month Archive
  • Special Collections
    • Spinal CSF Leak Articles (Jan 2020-June 2024)
    • 2024 AJNR Journal Awards
    • Most Impactful AJNR Articles
  • Multimedia
    • AJNR Podcast
    • AJNR Scantastics
    • Video Articles
  • For Authors
    • Submit a Manuscript
    • Author Policies
    • Fast publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Manuscript Submission Guidelines
    • Imaging Protocol Submission
    • Submit a Case for the Case Collection
  • About Us
    • About AJNR
    • Editorial Board
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home
  • Other Publications
    • ajnr

User menu

  • Alerts
  • Log in

Search

  • Advanced search
American Journal of Neuroradiology
American Journal of Neuroradiology

American Journal of Neuroradiology

ASHNR American Society of Functional Neuroradiology ASHNR American Society of Pediatric Neuroradiology ASSR
  • Alerts
  • Log in

Advanced Search

  • Home
  • Content
    • Current Issue
    • Accepted Manuscripts
    • Article Preview
    • Past Issue Archive
    • AJNR Case Collection
    • Case of the Week Archive
    • Classic Case Archive
    • Case of the Month Archive
  • Special Collections
    • Spinal CSF Leak Articles (Jan 2020-June 2024)
    • 2024 AJNR Journal Awards
    • Most Impactful AJNR Articles
  • Multimedia
    • AJNR Podcast
    • AJNR Scantastics
    • Video Articles
  • For Authors
    • Submit a Manuscript
    • Author Policies
    • Fast publishing of Accepted Manuscripts
    • Graphical Abstract Preparation
    • Manuscript Submission Guidelines
    • Imaging Protocol Submission
    • Submit a Case for the Case Collection
  • About Us
    • About AJNR
    • Editorial Board
  • More
    • Become a Reviewer/Academy of Reviewers
    • Subscribers
    • Permissions
    • Alerts
    • Feedback
    • Advertisers
    • ASNR Home
  • Follow AJNR on Twitter
  • Visit AJNR on Facebook
  • Follow AJNR on Instagram
  • Join AJNR on LinkedIn
  • RSS Feeds

Welcome to the new AJNR, Updated Hall of Fame, and more. Read the full announcements.


AJNR is seeking candidates for the position of Associate Section Editor, AJNR Case Collection. Read the full announcement.

 

Research ArticleAdult Brain
Open Access

Analysis of Stroke Detection during the COVID-19 Pandemic Using Natural Language Processing of Radiology Reports

M.D. Li, M. Lang, F. Deng, K. Chang, K. Buch, S. Rincon, W.A. Mehan, T.M. Leslie-Mazwi and J. Kalpathy-Cramer
American Journal of Neuroradiology March 2021, 42 (3) 429-434; DOI: https://doi.org/10.3174/ajnr.A6961
M.D. Li
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for M.D. Li
M. Lang
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for M. Lang
F. Deng
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for F. Deng
K. Chang
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for K. Chang
K. Buch
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for K. Buch
S. Rincon
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for S. Rincon
W.A. Mehan
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for W.A. Mehan
T.M. Leslie-Mazwi
bNeurology and Neurosurgery (T.M.L.-M.), Massachusetts General Hospital, Harvard Medical School, Boston, Massachusetts
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for T.M. Leslie-Mazwi
J. Kalpathy-Cramer
aFrom the Departments of Radiology (M.D.L., M.L., F.D., K.C., K.B., S.R., W.A.M., J.K.-C.)
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for J. Kalpathy-Cramer
  • Article
  • Figures & Data
  • Supplemental
  • Info & Metrics
  • Responses
  • References
  • PDF
Loading

Abstract

BACKGROUND AND PURPOSE: The coronavirus disease 2019 (COVID-19) pandemic has led to decreases in neuroimaging volume. Our aim was to quantify the change in acute or subacute ischemic strokes detected on CT or MR imaging during the pandemic using natural language processing of radiology reports.

MATERIALS AND METHODS: We retrospectively analyzed 32,555 radiology reports from brain CTs and MRIs from a comprehensive stroke center, performed from March 1 to April 30 each year from 2017 to 2020, involving 20,414 unique patients. To detect acute or subacute ischemic stroke in free-text reports, we trained a random forest natural language processing classifier using 1987 randomly sampled radiology reports with manual annotation. Natural language processing classifier generalizability was evaluated using 1974 imaging reports from an external dataset.

RESULTS: The natural language processing classifier achieved a 5-fold cross-validation classification accuracy of 0.97 and an F1 score of 0.74, with a slight underestimation (−5%) of actual numbers of acute or subacute ischemic strokes in cross-validation. Importantly, cross-validation performance stratified by year was similar. Applying the classifier to the complete study cohort, we found an estimated 24% decrease in patients with acute or subacute ischemic strokes reported on CT or MR imaging from March to April 2020 compared with the average from those months in 2017–2019. Among patients with stroke-related order indications, the estimated proportion who underwent neuroimaging with acute or subacute ischemic stroke detection significantly increased from 16% during 2017–2019 to 21% in 2020 (P = .01). The natural language processing classifier performed worse on external data.

CONCLUSIONS: Acute or subacute ischemic stroke cases detected by neuroimaging decreased during the COVID-19 pandemic, though a higher proportion of studies ordered for stroke were positive for acute or subacute ischemic strokes. Natural language processing approaches can help automatically track acute or subacute ischemic stroke numbers for epidemiologic studies, though local classifier training is important due to radiologist reporting style differences.

ABBREVIATIONS:

ASIS
acute or subacute ischemic stroke
COVID-19
coronavirus disease 2019
NLP
natural language processing

There is much concern regarding the impact of the coronavirus disease 2019 (COVID-19) pandemic on the quality of stroke care, including issues with hospital capacity, clinical resource re-allocation, and the safety of patients and clinicians.1,2 Previous reports have shown that there have been substantial decreases in stroke neuroimaging volume during the pandemic.3,4 In addition, acute ischemic infarcts have been found on neuroimaging studies in many hospitalized patients with COVID-19, though the causal relationship is unclear.5,6 Studies like these and other epidemiologic analyses usually rely on the creation of manually curated databases, in which identification of cases can be time-consuming and difficult to update in real-time. One way to facilitate such research is to use natural language processing (NLP), which has shown utility for automated analysis of radiology report data.7 NLP algorithms have been developed previously for the classification of neuroradiology reports for the presence of ischemic stroke findings and acute ischemic stroke subtypes.8,9 Thus, NLP has the potential to facilitate COVID-19 research.

In this study, we developed an NLP machine learning model that classifies radiology reports for the presence or absence of acute or subacute ischemic stroke (ASIS), as opposed to chronic stroke. We used this model to quantify the change in ASIS detected on all CT or MR imaging studies performed at a large comprehensive stroke center during the COVID-19 pandemic in the United States. We also evaluated NLP model generalizability and different training strategies using a sample of radiology reports from a second stroke center.

MATERIALS AND METHODS

This retrospective study was exempted with waiver of informed consent by the institutional review board of Mass General Brigham (Boston, Massachusetts), the integrated health system that includes both Massachusetts General Hospital and Brigham and Women's Hospital.

Radiology Report Extraction

We used a custom hospital-based radiology report search tool to extract head CT and brain MR imaging study reports performed at Massachusetts General Hospital (hospital 1) and its affiliated imaging centers (a comprehensive stroke center) from March 1 to April 30 in each year from 2017 to 2020. At this hospital, head CT and brain MR imaging studies are routinely performed for patients with stroke. Head CTs included noncontrast and contrast-enhanced head CT and CT angiography studies. Brain MRIs included noncontrast and contrast-enhanced brain MRIs and MR angiography studies. After we removed outside imaging studies also stored in the data base, there were 15,627 head CT and 17,151 brain MR imaging reports (a total of 32,778 studies). Of these studies, 15,590 head CT and 16,965 brain MR imaging reports had study “Impressions”, which restricted the analysis to 98.9% and 99.8% of the dataset, respectively. These studies formed the aggregate study cohort, which included a total of 32,555 brain MR imaging and head CT reports on 20,414 unique patients.

Of the original 32,778 study reports extracted, 1000 head CT and 1000 brain MR imaging studies were randomly sampled for manual annotation to serve as training and testing data for an NLP machine learning model. Of these studies, 1987 contained study Impressions (99.4%). The studies without study Impressions were predominantly pediatric brain MR imaging studies that involved a different structure for reporting.

Using a commercial radiology information system (Primordial/Nuance Communications), we also extracted an additional dataset of radiology reports from Brigham and Women’s Hospital (hospital 2) and its affiliated imaging centers (also a comprehensive stroke center). We analyzed the overlap in radiologists and trainees involved in the dictation of these reports between hospitals 1 and 2. The first 500 consecutive head CTs and the first 500 brain MRIs performed in both April 2019 and April 2020 were obtained (a total of 1000 head CTs and 1000 brain MR imaging study reports), with the same inclusion criteria for noncontrast and contrast-enhanced studies, as well as angiographic studies. All of these reports had study Impressions. After removing duplicate study entries in this dataset (26, 1.3%), 1974 head CT and brain MR imaging reports remained for further analysis.

NLP Training Dataset Annotation

For NLP model training, the 1987 study reports sampled from hospital 1 and the 1974 study reports available from hospital 2 were manually annotated, each by a diagnostic radiology resident (F.D. for CT and M.L. for MR imaging from hospital 1 and M.D.L. for CT and MR imaging from hospital 2). The annotators classified each report for the presence of ASIS using the study “Impression.” This finding could be explicitly or implicitly stated in the report, and reports that stated or suggested chronicity of an infarct were not considered to have this finding. For example, “old” or “chronic” infarct suggests chronicity, though more ambiguous terms like “age-indeterminate” or “unclear timeframe” were sometimes found. Reports with ambiguous terms were not considered to have ASIS, unless an expression of newness was conveyed in the report (eg, “new age-indeterminate infarct”).

NLP Machine Learning Model Training and Testing

We trained a random forest machine learning model that takes the radiology report free-text Impression as input and classifies the report for the presence or absence of an ASIS. To train a machine learning model to automatically parse the radiology report text, we used the re (Version 2.2.1), sklearn (Version 0.20.3), and nltk (Version 3.4) packages in Python (Version 3.7.1). Before model training, we used regular expressions to extract sentences with words containing the stems “infarct” or “ischem” from each study Impression. This step helped to focus the algorithm on sentences containing content relevant to the classification task. The words in the extracted sentences were stemmed using the snowball.EnglishStemmer from the nltk package. The extracted and stemmed sentences were then represented as vectors using bag-of-words vectorization with N-grams (n = 2–3; minimum term frequency, 1%), an approach previously used for radiology report natural language processing.10 Negation was dealt with using the nltk mark_negation function, which appends a “_NEG” suffix between words that appear between a negation term and a punctuation mark. These vector representations of the radiology report Impression served as inputs to the random forest NLP classifier.

The random forest NLP classifier was trained using default hyperparameters in sklearn, Version 0.20.3, including 100 trees in the forest using the Gini Impurity for measuring the quality of the data split. Using the manually annotated datasets from hospitals 1 and 2, we evaluated 2 training strategies. First, we trained a classifier using the hospital 1 annotated dataset and tested performance using 5-fold cross-validation, stratified on outcome (ASIS), given the class imbalance. We also tested this classifier on the external hospital 2 annotated dataset. Second, we trained a classifier using the combined hospital 1 and 2 annotated datasets and tested performance using 5-fold cross-validation, also stratified on outcome, but only including hospital 1 reports in the cross-validation to assess performance on hospital 1 data specifically. We repeated this testing using only hospital 2 reports in the cross-validation to assess the performance on the hospital 2 data specifically. When the combined hospital 1 and 2 datasets were used for classifier training, the N-gram minimum term frequency was halved to 0.5%; empirically, the number of N-gram terms was then similar between this classifier and the classifier trained on hospital 1 data only. The Python code for training these random forest classifiers is available at github.com/QTIM-Lab/asis_nlp.

The metrics used to assess model performance included accuracy, precision, recall, and F1 score. Performance was evaluated for CT and MR imaging reports combined, CT reports alone, and MR imaging reports alone, with 5-fold cross-validation when appropriate.

NLP-Based Epidemiologic Analysis

On the basis of the results of the NLP model testing, an NLP classifier was then applied to the complete cohort of 32,555 brain MR imaging and head CT reports from hospital 1 to estimate the number of patients with ASIS. Patients with at least 1 neuroimaging study (CT or MR imaging) with an ASIS during the time period in question were considered to have had an infarct. Demographic information associated with these patients was extracted along with the radiology report text.

Statistics

Statistical testing was performed using the scipy Version 1.1.0 package in Python. The Pearson χ2 test of independence and 1-way ANOVA were used when appropriate. Statistical significance was determined a priori to be P < .05. Performance metrics were reported as the bootstrap median estimate with 95% confidence intervals.11

RESULTS

Manually Annotated Radiology Report Dataset Characteristics

Among the randomly sampled 1987 neuroimaging reports from hospital 1 used for NLP model development, 67 head CT and 68 brain MR imaging reports were manually classified as positive for ASIS (positive in 129 patients from 1904 total unique patients). Among the 1974 neuroimaging reports from hospital 2, 84 head CT and 91 brain MR imaging reports were manually classified as positive for ASIS (positive in 101 patients from 1514 total unique patients). The remainder of studies were negative for ASIS. In the hospital 1 annotated report dataset, 126 unique radiologists and trainees (residents and fellows) were involved in the dictation of these reports. In the hospital 2 annotated report dataset, there were 94 unique radiologists and trainees involved. There was an overlap of 3 radiologists and trainees between these 2 datasets due to radiologists/trainees moving between institutions. The hospital 1 and hospital 2 reports were all free-text without a standardized structure. The manual annotators who read the report Impressions found that they differed stylistically between the hospitals.

NLP Model Performance

Random forest NLP classifier testing performance is summarized in the Online Supplemental Data. The stratified 5-fold cross-validation performance of the NLP classifier trained on the hospital 1 annotated dataset showed an average accuracy of 0.97 (95% CI, 0.96–0.97) and an F1 score of 0.74 (95% CI, 0.72–0.76). When this NLP classifier was tested on the hospital 2 annotated dataset, the performance was lower, with an accuracy of 0.95 (95% CI, 0.94–0.96) and an F1 score of 0.66 (95% CI, 0.59–0.72). In both tests, when the performance results for CT and MR imaging were separately analyzed, we found that the model performed better for MR imaging reports compared with CT reports.

We also trained a random forest NLP classifier using the combined annotated reports from hospitals 1 and 2. In the stratified 5-fold cross-validation performance with testing of only hospital 1 data in the validation folds, the average accuracy was 0.96 (95% CI, 0.96-0.96) and the average F1 score was 0.74 (95% CI, 0.72–0.76). This performance on hospital 1 data was similar compared with the NLP classifier trained using only hospital 1 data. In the stratified 5-fold cross-validation performance with testing of only hospital 2 data in the validation folds, the average accuracy was 0.96 (95% CI, 0.96-0.97) and the average F1 score was 0.79 (95% CI, 0.77–0.80). This performance on hospital 2 data was substantially improved compared with the NLP classifier trained using only hospital 1 data. Because the performance on hospital 1 data was similar between the NLP classifier trained on hospital 1 reports and the NLP classifier trained on hospitals 1 and 2 reports, we used the former classifier for further analysis of the complete hospital 1 dataset.

For the NLP classifier trained on hospital 1 reports, in the 5 cross-validation folds for the combined CT and MR imaging analysis, there was an average of 19.4 (95% CI, 18.6–20.2) true-positive, 6.2 (95% CI, 5.6–6.8) false-positive, 7.6 (95% CI, 6.8–8.4) false-negative, and 364.2 (95% CI, 18.6–20.2) true-negative classifications. In misclassified cases, the reports typically contained uncertainty regarding the chronicity of the infarct (eg, age-indeterminate or not otherwise specified in the study Impression). For each of the 5 cross-validation folds, there was an average of 25.6 positive results predicted (95% CI, 24.6–26.8), compared with 27.0 actual positive results in each validation fold, due to the stratification on outcome. The NLP predicted that the number of cases slightly underestimated the actual number of studies positive for ASIS in the validation folds (average difference, −1.4; 95% CI, 0.2–2.4; expressed as percentages, −5.1%; 95% CI, 0–8.8%).

To ensure that variations in reporting styles within the hospital 1 reports did not systematically differ by year (because our epidemiologic analysis would compare reports from different years), we performed leave-one-year-out cross-validation on the hospital 1 dataset, in which NLP classifiers were trained on data from all years except the year of the excluded validation set (eg, trained on reports from 2018, 2019, and 2020, and then tested on reports from 2020). We found that there was no substantial difference in model performance in each of those validation folds (with overlap of 95% CI), which shows that the model performed similarly across time periods at hospital 1 (Online Supplemental Data). The F1 score was 0.72 in 2020 versus between 0.68 and 0.73 from 2017 to 2019.

While the NLP model systematically slightly underestimated ASIS case numbers, because the model performed similarly from year-to-year, we used this random forest classifier to estimate changes in the numbers of ASISs detected in the complete hospital 1 study cohort of 32,555 head CT and brain MR imaging reports.

ASIS during the COVID-19 Pandemic

Using this random forest NLP classifier, we estimated the number of neuroimaging studies performed and the number of patients with detected ASIS (Table 1). Patients with at least 1 neuroimaging study (CT or MR imaging) with an ASIS during the time period in question were considered to have had ASIS. There was an estimated 24% decrease in patients with ASIS reported on CT or MR imaging from March to April 2020 compared with the average of the same months from 2017 to 2019, after previous year-on-year growth from 2017 to 2019 (Figure). There was a concomitant decrease in the total number of neuroimaging studies performed and patients undergoing neuroimaging in March and April 2020 compared with 2019 (−39% and −41%, respectively).

FIGURE.
  • Download figure
  • Open in new tab
  • Download powerpoint
FIGURE.

Estimated numbers of patients with acute or subacute ischemic strokes detected on CT or MR imaging in March and April from 2017 to 2020 at hospital 1.

View this table:
  • View inline
  • View popup
Table 1:

Natural language processing-based analysis of all radiology reports from hospital 1

In the complete cohort of 32,555 study reports, 32,358 of the reports (99.4%) included structured and/or unstructured text in the study indication field, entered at the time of order entry. Of those cases, we filtered for indications including “stroke,” “neuro deficit,” and “TIA,” which resulted in 5204 study reports (Table 2). In these patients, we found an estimated 21% decrease in ASIS reported from March to April 2020 compared with March to April 2019 (Figure). In the subset of patients who underwent imaging with stroke-related indications, the estimated proportion of patients with ASIS detected increased from 16% during 2017–2019 to 21% in 2020 (P = .01) (Table 2). The estimated proportion of neuroimaging studies with ASIS detected increased from 20% during 2017–2019 and 24% in 2020 (P = .01).

View this table:
  • View inline
  • View popup
Table 2:

NLP-based analysis of radiology reports from hospital 1 containing “stroke,” “neuro deficit,” or “TIA” in the order indication

The average age of patients with ASIS detected was 66  [SD, 17] years, and there was no significant difference in age among any of the years (P = .9). There was also no significant difference in the sex ratio of March to April 2020 compared with the March to April 2017–2019 time periods (P = .8). In aggregate, 56% of patients with ASIS were men. See Online Supplemental Data for data by year.

Neuroimaging Studies Performed per Patient

If the number of neuroimaging studies performed per patient differed between the prepandemic and pandemic time periods, the number of opportunities to detect ASIS in a patient could vary. However, this variance did not appear to be a confounding factor in our analysis because we found no significant difference in the number of neuroimaging studies performed per patient between the March and April 2020 time period and each of the March to April 2017, 2018, or 2019 time periods (P > .2).

DISCUSSION

In this study, we developed a random forest NLP algorithm for automated classification of ASIS in radiology report Impressions and applied this algorithm to reports during and before the COVID-19 pandemic. We found a substantial decrease in the number of patients with ASIS detected on all CT and MR imaging studies performed at a comprehensive stroke center during the pandemic in the United States. This decrease could be related to avoidance of the hospital due to fear of contracting COVID-19, as previously speculated.12,13 Previous studies have shown a 39% decrease in neuroimaging studies performed primarily for stroke thrombectomy evaluation using commercial software in the United States and a 59.7% decrease in stroke code CT-specific cases in New York.3,4 Our study differs because we sought to quantify the decrease in actual ASISs detected on such studies and the rate of detection. Among patients with stroke-related image -order indication, we found a significant increase in the proportion of neuroimaging studies positive for ASIS. This finding could suggest that during the COVID-19 pandemic, imaged patients had, on average, more severe or clear-cut stroke syndromes (with a higher pretest probability of stroke), implying that patients with mild or equivocal symptoms presented to the hospital less often than in previous years.

The NLP machine learning approach that we used in this study can also be applied to additional data relatively easily, which will allow us to continue to monitor the detection of ASIS on neuroimaging at our institution in the future. NLP algorithms have been used to analyze neuroradiology reports for stroke findings, specifically for the presence of any ischemic stroke findings or ischemic stroke subtypes.8,9 The task in our model, however, is relatively challenging in that we sought to identify acute or subacute strokes specifically and deliberately excluded chronic infarcts. There is often uncertainty or ambiguity in radiology reports related to the timeframe for strokes, which can make this task challenging for the NLP algorithm. Thus, it is not surprising that our NLP model performed better on MR imaging reports compared with CT reports, given the superiority of MR imaging for characterizing the age of an infarct.

The NLP classifier trained on only hospital 1 reports showed lower performance when tested on radiology reports from an external site, which was likely due to systematic differences in linguistic reporting styles between the radiology departments in hospitals 1 and 2. While combining training data from hospitals 1 and 2 helped to create a more generalizable classifier with improved performance on hospital 2 data, the test performance of this classifier on data from hospital 1 was not substantially different from the classifier trained on only hospital 1 data. These findings highlight the importance of localized testing of NLP algorithms before clinical deployment. Nevertheless, a locally trained and deployed model can still be useful, as long as its specific use case and limitations are understood.14

Instead of using the radiology report NLP approach presented in our study, we could have used the International Classification of Diseases codes from hospital administrative and billing data. However, the International Classification of Diseases coding is known to have variable sensitivity and specificity for acute stroke in the literature15 and may be particularly problematic for reliably differentiating stroke chronicity. Comparison of NLP analyses of radiology reports versus administrative data base International Classification of Diseases coding could be an avenue of future research.

There are important limitations to this study. First, we used an automated NLP approach for analysis, which systematically slightly underestimates the number of ASISs but may be scaled to analyze large numbers of reports. In the future, newer NLP technologies including deep learning–based algorithms may help improve the ability to perform studies like this one.16 Second, the radiology report is an imperfect reference standard for assessment of ASIS, particularly for CT in which early infarcts may not be seen. In our epidemiologic analysis, patients with at least 1 neuroimaging study with ASIS during the time period of interest were counted as having ASIS. Thus, patients with early infarcts not reported on CT would still be counted as having ASIS if reported on the subsequent MR imaging, reducing the impact of false-negative CTs. However, false-positive head CTs would falsely elevate the count of ASIS. Third, identification of studies with stroke-related indications likely underestimates the total number of studies performed for suspicion of stroke because nonspecific indications like “altered mental status” were not included. This bias should be consistent across each year though; thus, it should not impact our comparison of the positive case rate between the time periods in question.

CONCLUSIONS

We developed an NLP machine learning model to characterize trends in stroke imaging at a comprehensive stroke center before and during the COVID-19 pandemic. The sequelae of decreased detection of strokes remains to be seen, but this algorithm and the shared code can help facilitate future research of these trends.

Footnotes

  • This research was supported by a training grant from the National Institute of Biomedical Imaging and Bioengineering the National Institutes of Health under award No. 5T32EB1680 and the National Cancer Institute of the National Institutes of Health under Award No. F30CA239407 to K. Chang.

  • Disclosures: Ken Chang—RELATED: Grant: National Institutes of Health, Comments: Research reported in this publication was supported by a training grant from the National Institute of Biomedical Imaging and Bioengineering the National Institutes of Health under award No. 5T32EB1680 and the National Cancer Institute of the National Institutes of Health under Award No. F30CA239407 to K. Chang.* Karen Buch—UNRELATED: Employment: Massachusetts General Hospital. William A. Mehan, Jr—UNRELATED: Consultancy: Kura Oncology, Comments: independent image reviewer for head and neck cancer trial; Expert Testimony: CRICO and other medical insurance companies, Comments: expert opinion for medicolegal cases involving neuroimaging studies. Jayashree Kalpathy-Cramer—UNRELATED: Grants/Grants Pending: GE Healthcare, Genentech Patient Foundation*; Travel/Accommodations/Meeting Expenses Unrelated to Activities Listed: IBM. *Money paid to the institution.

Indicates open access to non-subscribers at www.ajnr.org

References

  1. 1.↵
    1. Leira EC,
    2. Russman AN,
    3. Biller J, et al
    . Preserving stroke care during the COVID-19 pandemic: potential issues and solutions. Neurology 2020;95:124–33 doi:10.1212/WNL.0000000000009713 pmid:32385186
    Abstract/FREE Full Text
  2. 2.↵
    1. Markus HS,
    2. Brainin M
    . COVID-19 and stroke: a global World Stroke Organization perspective. Int J Stroke 2020;15:361–64 doi:10.1177/1747493020923472 pmid:32310017
    CrossRefPubMed
  3. 3.↵
    1. Phillips CD,
    2. Shatzkes D,
    3. Moonis G, et al
    . From the eye of the storm: multi-institutional practical perspectives on neuroradiology from the COVID-19 outbreak in New York City. AJNR Am J Neuroradiol 2020;41:960–65 doi:10.3174/ajnr.A6565 pmid:32354706
    CrossRefPubMed
  4. 4.↵
    1. Kansagra AP,
    2. Goyal MS,
    3. Hamilton S, et al
    . Collateral effect of Covid-19 on stroke evaluation in the United States. N Engl J Med 2020;383:400–01 doi:10.1056/NEJMc2014816 pmid:32383831
    CrossRefPubMed
  5. 5.↵
    1. Mahammedi A,
    2. Saba L,
    3. Vagal A, et al
    . Imaging in neurological disease of hospitalized COVID-19 patients: an Italian multicenter retrospective observational study. Radiology 2020;297:E270–73 doi:10.1148/radiol.2020201933 pmid:32437313
    CrossRefPubMed
  6. 6.↵
    1. Jain R,
    2. Young M,
    3. Dogra S, et al
    . COVID-19 related neuroimaging findings: a signal of thromboembolic complications and a strong prognostic marker of poor patient outcome. J Neurol Sci 2020;414:116923 doi:10.1016/j.jns.2020.116923 pmid:32447193
    CrossRefPubMed
  7. 7.↵
    1. Pons E,
    2. Braun LM,
    3. Hunink MG, et al
    . Natural language processing in radiology: a systematic review. Radiology 2016;279:329–43 doi:10.1148/radiol.16142770 pmid:27089187
    CrossRefPubMed
  8. 8.↵
    1. Wheater E,
    2. Mair G,
    3. Sudlow C, et al
    . A validated natural language processing algorithm for brain imaging phenotypes from radiology reports in UK electronic health records. BMC Med Inform Decis Mak 2019;19:184 doi:10.1186/s12911-019-0908-7 pmid:31500613
    CrossRefPubMed
  9. 9.↵
    1. Garg R,
    2. Oh E,
    3. Naidech A, et al
    . Automating ischemic stroke subtype classification using machine learning and natural language processing. J Stroke Cerebrovasc Dis 2019;28:2045–51 doi:10.1016/j.jstrokecerebrovasdis.2019.02.004 pmid:31103549
    CrossRefPubMed
  10. 10.↵
    1. Hassanpour S,
    2. Langlotz CP,
    3. Amrhein TJ, et al
    . Performance of a machine learning classifier of knee MRI reports in two large academic radiology practices: a tool to estimate diagnostic yield. AJR Am J Roentgenol 2017;208:750–73 doi:10.2214/AJR.16.16128 pmid:28140627
    CrossRefPubMed
  11. 11.↵
    1. Efron B,
    2. Tibshirani RJ
    . An Introduction to the Bootstrap. Chapman & Hall/CRC; 1994
  12. 12.↵
    1. Oxley TJ,
    2. Mocco J,
    3. Majidi S, et al
    . Large-vessel stroke as a presenting feature of Covid-19 in the young. N Engl J Med 2020;382:e60 doi:10.1056/NEJMc2009787 pmid:32343504
    CrossRefPubMed
  13. 13.↵
    1. Zhao J,
    2. Rudd A,
    3. Liu R
    . Challenges and potential solutions of stroke care during the coronavirus disease 2019 (COVID-19) outbreak. Stroke 2020;51:1356–57 doi:10.1161/STROKEAHA.120.029701 pmid:32228369
    CrossRefPubMed
  14. 14.↵
    1. Futoma J,
    2. Simons M,
    3. Panch T, et al
    . The myth of generalisability in clinical research and machine learning in health care. Lancet Digit Heal 2020;2:e489–92 doi:10.1016/S2589-7500(20)30186-2 pmid:32864600
    CrossRefPubMed
  15. 15.↵
    1. McCormick N,
    2. Bhole V,
    3. Lacaille D, et al
    . Validity of diagnostic codes for acute stroke in administrative databases: a systematic review. PLoS One 2015;10:e0135834 doi:10.1371/journal.pone.0135834 pmid:6292280
    CrossRefPubMed
  16. 16.↵
    1. Wolf T,
    2. Debut L,
    3. Sanh V, et al
    . HuggingFace’s Transformers: State-of-the-art Natural Language Processing. Oct 9, 2019. https://arxiv.org/abs/1910.03771. Accessed June 15, 2020.
  • Received June 20, 2020.
  • Accepted after revision October 26, 2020.
  • © 2021 by American Journal of Neuroradiology
PreviousNext
Back to top

In this issue

American Journal of Neuroradiology: 42 (3)
American Journal of Neuroradiology
Vol. 42, Issue 3
1 Mar 2021
  • Table of Contents
  • Index by author
  • Complete Issue (PDF)
Advertisement
Print
Download PDF
Email Article

Thank you for your interest in spreading the word on American Journal of Neuroradiology.

NOTE: We only request your email address so that the person you are recommending the page to knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Analysis of Stroke Detection during the COVID-19 Pandemic Using Natural Language Processing of Radiology Reports
(Your Name) has sent you a message from American Journal of Neuroradiology
(Your Name) thought you would like to see the American Journal of Neuroradiology web site.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Cite this article
M.D. Li, M. Lang, F. Deng, K. Chang, K. Buch, S. Rincon, W.A. Mehan, T.M. Leslie-Mazwi, J. Kalpathy-Cramer
Analysis of Stroke Detection during the COVID-19 Pandemic Using Natural Language Processing of Radiology Reports
American Journal of Neuroradiology Mar 2021, 42 (3) 429-434; DOI: 10.3174/ajnr.A6961

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
0 Responses
Respond to this article
Share
Bookmark this article
Analysis of Stroke Detection during the COVID-19 Pandemic Using Natural Language Processing of Radiology Reports
M.D. Li, M. Lang, F. Deng, K. Chang, K. Buch, S. Rincon, W.A. Mehan, T.M. Leslie-Mazwi, J. Kalpathy-Cramer
American Journal of Neuroradiology Mar 2021, 42 (3) 429-434; DOI: 10.3174/ajnr.A6961
del.icio.us logo Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One
Purchase

Jump to section

  • Article
    • Abstract
    • ABBREVIATIONS:
    • MATERIALS AND METHODS
    • RESULTS
    • DISCUSSION
    • CONCLUSIONS
    • Footnotes
    • References
  • Figures & Data
  • Supplemental
  • Info & Metrics
  • Responses
  • References
  • PDF

Related Articles

  • PubMed
  • Google Scholar

Cited By...

  • No citing articles found.
  • Crossref (26)
  • Google Scholar

This article has been cited by the following articles in journals that are participating in Crossref Cited-by Linking.

  • Construction of nanomaterials as contrast agents or probes for glioma imaging
    Wei Zhao, Xiangrong Yu, Shaojun Peng, Yu Luo, Jingchao Li, Ligong Lu
    Journal of Nanobiotechnology 2021 19 1
  • Natural language processing in clinical neuroscience and psychiatry: A review
    Claudio Crema, Giuseppe Attardi, Daniele Sartiano, Alberto Redolfi
    Frontiers in Psychiatry 2022 13
  • Data Extraction from Free-Text Reports on Mechanical Thrombectomy in Acute Ischemic Stroke Using ChatGPT: A Retrospective Analysis
    Nils C. Lehnen, Franziska Dorn, Isabella C. Wiest, Hanna Zimmermann, Alexander Radbruch, Jakob Nikolas Kather, Daniel Paech
    Radiology 2024 311 1
  • Natural Language Processing in Radiology: Update on Clinical Applications
    Pilar López-Úbeda, Teodoro Martín-Noguerol, Krishna Juluru, Antonio Luna
    Journal of the American College of Radiology 2022 19 11
  • Applications of natural language processing in radiology: A systematic review
    Nathaniel Linna, Charles E. Kahn
    International Journal of Medical Informatics 2022 163
  • The Role of Natural Language Processing during the COVID-19 Pandemic: Health Applications, Opportunities, and Challenges
    Mohammed Ali Al-Garadi, Yuan-Chi Yang, Abeed Sarker
    Healthcare 2022 10 11
  • Complication and Sequelae of COVID-19: What Should We Pay Attention to in the Post-Epidemic Era
    Keda Yang, Guangfu Wen, Jinpeng Wang, Siming Zhou, Wacili Da, Yan Meng, Yuchuan Xue, Lin Tao
    Frontiers in Immunology 2021 12
  • Natural Language Processing of Radiology Reports to Detect Complications of Ischemic Stroke
    Matthew I. Miller, Agni Orfanoudaki, Michael Cronin, Hanife Saglam, Ivy So Yeon Kim, Oluwafemi Balogun, Maria Tzalidi, Kyriakos Vasilopoulos, Georgia Fanaropoulou, Nina M. Fanaropoulou, Jack Kalin, Meghan Hutch, Brenton R. Prescott, Benjamin Brush, Emelia J. Benjamin, Min Shin, Asim Mian, David M. Greer, Stelios M. Smirnakis, Charlene J. Ong
    Neurocritical Care 2022 37 S2
  • Transformer versus traditional natural language processing: how much data is enough for automated radiology report classification?
    Eric Yang, Matthew D Li, Shruti Raghavan, Francis Deng, Min Lang, Marc D Succi, Ambrose J Huang, Jayashree Kalpathy-Cramer
    The British Journal of Radiology 2023 96 1149
  • Development and Validation of a Model to Identify Critical Brain Injuries Using Natural Language Processing of Text Computed Tomography Reports
    Victor M. Torres-Lopez, Grace E. Rovenolt, Angelo J. Olcese, Gabriella E. Garcia, Sarah M. Chacko, Amber Robinson, Edward Gaiser, Julian Acosta, Alison L. Herman, Lindsey R. Kuohn, Megan Leary, Alexandria L. Soto, Qiang Zhang, Safoora Fatima, Guido J. Falcone, Seyedmehdi Payabvash, Richa Sharma, Aaron F. Struck, Kevin N. Sheth, M. Brandon Westover, Jennifer A. Kim
    JAMA Network Open 2022 5 8

More in this TOC Section

  • Diagnostic Neuroradiology of Monoclonal Antibodies
  • ML for Glioma Molecular Subtype Prediction
  • Segmentation of Brain Metastases with BLAST
Show more Adult Brain

Similar Articles

Advertisement

Indexed Content

  • Current Issue
  • Accepted Manuscripts
  • Article Preview
  • Past Issues
  • Editorials
  • Editors Choice
  • Fellow Journal Club
  • Letters to the Editor

Cases

  • Case Collection
  • Archive - Case of the Week
  • Archive - Case of the Month
  • Archive - Classic Case

Special Collections

  • Special Collections

Resources

  • News and Updates
  • Turn around Times
  • Submit a Manuscript
  • Author Policies
  • Manuscript Submission Guidelines
  • Evidence-Based Medicine Level Guide
  • Publishing Checklists
  • Graphical Abstract Preparation
  • Imaging Protocol Submission
  • Submit a Case
  • Become a Reviewer/Academy of Reviewers
  • Get Peer Review Credit from Publons

Multimedia

  • AJNR Podcast
  • AJNR SCANtastic
  • Video Articles

About Us

  • About AJNR
  • Editorial Board
  • Not an AJNR Subscriber? Join Now
  • Alerts
  • Feedback
  • Advertise with us
  • Librarian Resources
  • Permissions
  • Terms and Conditions

American Society of Neuroradiology

  • Not an ASNR Member? Join Now

© 2025 by the American Society of Neuroradiology All rights, including for text and data mining, AI training, and similar technologies, are reserved.
Print ISSN: 0195-6108 Online ISSN: 1936-959X

Powered by HighWire