Autonomous Chest Radiograph Reporting Using AI: Estimation of Clinical Impact

Research output: Contribution to journalJournal articleResearchpeer-review

Standard

Autonomous Chest Radiograph Reporting Using AI : Estimation of Clinical Impact. / Plesner, Louis L.; Müller, Felix C.; Nybing, Janus D.; Laustrup, Lene C.; Rasmussen, Finn; Nielsen, Olav W.; Boesen, Mikael; Andersen, Michael B.

In: Radiology, Vol. 307, No. 3, e222268, 2023.

Research output: Contribution to journalJournal articleResearchpeer-review

Harvard

Plesner, LL, Müller, FC, Nybing, JD, Laustrup, LC, Rasmussen, F, Nielsen, OW, Boesen, M & Andersen, MB 2023, 'Autonomous Chest Radiograph Reporting Using AI: Estimation of Clinical Impact', Radiology, vol. 307, no. 3, e222268. https://doi.org/10.1148/RADIOL.222268

APA

Plesner, L. L., Müller, F. C., Nybing, J. D., Laustrup, L. C., Rasmussen, F., Nielsen, O. W., Boesen, M., & Andersen, M. B. (2023). Autonomous Chest Radiograph Reporting Using AI: Estimation of Clinical Impact. Radiology, 307(3), [e222268]. https://doi.org/10.1148/RADIOL.222268

Vancouver

Plesner LL, Müller FC, Nybing JD, Laustrup LC, Rasmussen F, Nielsen OW et al. Autonomous Chest Radiograph Reporting Using AI: Estimation of Clinical Impact. Radiology. 2023;307(3). e222268. https://doi.org/10.1148/RADIOL.222268

Author

Plesner, Louis L. ; Müller, Felix C. ; Nybing, Janus D. ; Laustrup, Lene C. ; Rasmussen, Finn ; Nielsen, Olav W. ; Boesen, Mikael ; Andersen, Michael B. / Autonomous Chest Radiograph Reporting Using AI : Estimation of Clinical Impact. In: Radiology. 2023 ; Vol. 307, No. 3.

Bibtex

@article{ea8cd57cd9a340c783441e285f8f7215,
title = "Autonomous Chest Radiograph Reporting Using AI: Estimation of Clinical Impact",
abstract = "Background: Automated interpretation of normal chest radiographs could alleviate the workload of radiologists. However, the performance of such an artificial intelligence (AI) tool compared with clinical radiology reports has not been established. Purpose: To perform an external evaluation of a commercially available AI tool for (a) the number of chest radiographs autonomously reported, (b) the sensitivity for AI detection of abnormal chest radiographs, and (c) the performance of AI compared with that of the clinical radiology reports. Materials and Methods: In this retrospective study, consecutive posteroanterior chest radiographs from adult patients in four hospitals in the capital region of Denmark were obtained in January 2020, including images from emergency department patients, in-hospital patients, and outpatients. Three thoracic radiologists labeled chest radiographs in a reference standard based on chest radiograph findings into the following categories: critical, other remarkable, unremarkable, or normal (no abnormalities). AI classified chest radiographs as high confidence normal (normal) or not high confidence normal (abnormal). Results: A total of 1529 patients were included for analysis (median age, 69 years [IQR, 55-69 years]; 776 women), with 1100 (72%) classified by the reference standard as having abnormal radiographs, 617 (40%) as having critical abnormal radiographs, and 429 (28%) as having normal radiographs. For comparison, clinical radiology reports were classified based on the text and insufficient reports excluded (n = 22). The sensitivity of AI was 99.1% (95% CI: 98.3, 99.6; 1090 of 1100 patients) for abnormal radiographs and 99.8% (95% CI: 99.1, 99.9; 616 of 617 patients) for critical radiographs. Corresponding sensitivities for radiologist reports were 72.3% (95% CI: 69.5, 74.9; 779 of 1078 patients) and 93.5% (95% CI: 91.2, 95.3; 558 of 597 patients), respectively. Specificity of AI, and hence the potential autonomous reporting rate, was 28.0% of all normal posteroanterior chest radiographs (95% CI: 23.8, 32.5; 120 of 429 patients), or 7.8% (120 of 1529 patients) of all posteroanterior chest radiographs. Conclusion: Of all normal posteroanterior chest radiographs, 28% were autonomously reported by AI with a sensitivity for any abnormalities higher than 99%. This corresponded to 7.8% of the entire posteroanterior chest radiograph production.",
author = "Plesner, {Louis L.} and M{\"u}ller, {Felix C.} and Nybing, {Janus D.} and Laustrup, {Lene C.} and Finn Rasmussen and Nielsen, {Olav W.} and Mikael Boesen and Andersen, {Michael B.}",
note = "Publisher Copyright: {\textcopyright} RSNA, 2023.",
year = "2023",
doi = "10.1148/RADIOL.222268",
language = "English",
volume = "307",
journal = "Radiology",
issn = "0033-8419",
publisher = "Radiological Society of North America, Inc.",
number = "3",

}

RIS

TY - JOUR

T1 - Autonomous Chest Radiograph Reporting Using AI

T2 - Estimation of Clinical Impact

AU - Plesner, Louis L.

AU - Müller, Felix C.

AU - Nybing, Janus D.

AU - Laustrup, Lene C.

AU - Rasmussen, Finn

AU - Nielsen, Olav W.

AU - Boesen, Mikael

AU - Andersen, Michael B.

N1 - Publisher Copyright: © RSNA, 2023.

PY - 2023

Y1 - 2023

N2 - Background: Automated interpretation of normal chest radiographs could alleviate the workload of radiologists. However, the performance of such an artificial intelligence (AI) tool compared with clinical radiology reports has not been established. Purpose: To perform an external evaluation of a commercially available AI tool for (a) the number of chest radiographs autonomously reported, (b) the sensitivity for AI detection of abnormal chest radiographs, and (c) the performance of AI compared with that of the clinical radiology reports. Materials and Methods: In this retrospective study, consecutive posteroanterior chest radiographs from adult patients in four hospitals in the capital region of Denmark were obtained in January 2020, including images from emergency department patients, in-hospital patients, and outpatients. Three thoracic radiologists labeled chest radiographs in a reference standard based on chest radiograph findings into the following categories: critical, other remarkable, unremarkable, or normal (no abnormalities). AI classified chest radiographs as high confidence normal (normal) or not high confidence normal (abnormal). Results: A total of 1529 patients were included for analysis (median age, 69 years [IQR, 55-69 years]; 776 women), with 1100 (72%) classified by the reference standard as having abnormal radiographs, 617 (40%) as having critical abnormal radiographs, and 429 (28%) as having normal radiographs. For comparison, clinical radiology reports were classified based on the text and insufficient reports excluded (n = 22). The sensitivity of AI was 99.1% (95% CI: 98.3, 99.6; 1090 of 1100 patients) for abnormal radiographs and 99.8% (95% CI: 99.1, 99.9; 616 of 617 patients) for critical radiographs. Corresponding sensitivities for radiologist reports were 72.3% (95% CI: 69.5, 74.9; 779 of 1078 patients) and 93.5% (95% CI: 91.2, 95.3; 558 of 597 patients), respectively. Specificity of AI, and hence the potential autonomous reporting rate, was 28.0% of all normal posteroanterior chest radiographs (95% CI: 23.8, 32.5; 120 of 429 patients), or 7.8% (120 of 1529 patients) of all posteroanterior chest radiographs. Conclusion: Of all normal posteroanterior chest radiographs, 28% were autonomously reported by AI with a sensitivity for any abnormalities higher than 99%. This corresponded to 7.8% of the entire posteroanterior chest radiograph production.

AB - Background: Automated interpretation of normal chest radiographs could alleviate the workload of radiologists. However, the performance of such an artificial intelligence (AI) tool compared with clinical radiology reports has not been established. Purpose: To perform an external evaluation of a commercially available AI tool for (a) the number of chest radiographs autonomously reported, (b) the sensitivity for AI detection of abnormal chest radiographs, and (c) the performance of AI compared with that of the clinical radiology reports. Materials and Methods: In this retrospective study, consecutive posteroanterior chest radiographs from adult patients in four hospitals in the capital region of Denmark were obtained in January 2020, including images from emergency department patients, in-hospital patients, and outpatients. Three thoracic radiologists labeled chest radiographs in a reference standard based on chest radiograph findings into the following categories: critical, other remarkable, unremarkable, or normal (no abnormalities). AI classified chest radiographs as high confidence normal (normal) or not high confidence normal (abnormal). Results: A total of 1529 patients were included for analysis (median age, 69 years [IQR, 55-69 years]; 776 women), with 1100 (72%) classified by the reference standard as having abnormal radiographs, 617 (40%) as having critical abnormal radiographs, and 429 (28%) as having normal radiographs. For comparison, clinical radiology reports were classified based on the text and insufficient reports excluded (n = 22). The sensitivity of AI was 99.1% (95% CI: 98.3, 99.6; 1090 of 1100 patients) for abnormal radiographs and 99.8% (95% CI: 99.1, 99.9; 616 of 617 patients) for critical radiographs. Corresponding sensitivities for radiologist reports were 72.3% (95% CI: 69.5, 74.9; 779 of 1078 patients) and 93.5% (95% CI: 91.2, 95.3; 558 of 597 patients), respectively. Specificity of AI, and hence the potential autonomous reporting rate, was 28.0% of all normal posteroanterior chest radiographs (95% CI: 23.8, 32.5; 120 of 429 patients), or 7.8% (120 of 1529 patients) of all posteroanterior chest radiographs. Conclusion: Of all normal posteroanterior chest radiographs, 28% were autonomously reported by AI with a sensitivity for any abnormalities higher than 99%. This corresponded to 7.8% of the entire posteroanterior chest radiograph production.

U2 - 10.1148/RADIOL.222268

DO - 10.1148/RADIOL.222268

M3 - Journal article

C2 - 36880947

AN - SCOPUS:85153803452

VL - 307

JO - Radiology

JF - Radiology

SN - 0033-8419

IS - 3

M1 - e222268

ER -

ID: 363365165