Skip to content
H
Unlocking infinite medical capacity.
resource_desktop_tab_pic
Latest announcements

How to evaluate a diagnostic radiology AI paper?

As artificial intelligence enters practical and daily use, radiologists will increasingly engage with radiology AI algorithms as end-users or stakeholders in deciding whether to implement algorithms in clinical practice. As radiologists, we need to address the most important question: Can we use this AI solution in clinical practice safely?

Most artificial intelligence algorithms will come with a variety of evidence, either a white paper provided by the manufacturer or a peer-reviewed academic paper published in the literature. Since the most common types of AI algorithms will likely be performing computer-aided diagnosis, most of this evidence will take the form of a diagnostic accuracy study for classification.

Evaluating diagnostic accuracy studies is a well-known topic that predates artificial intelligence or machine learning. Well-established guidelines such as the Standards for Reporting Diagnostic Accuracy Studies (STARD)[1] are suited for this task and should be included as part of the assessment of any diagnostic radiology AI paper. 

However, when evaluating a diagnostic radiology AI paper, several additional facets should be considered, particularly around the dataset used to train and test the AI algorithm, the AI training procedure, and the results presented in the paper.

Not all data are equal

The first questions to consider are based on the data referenced in a paper.

1.Training data vs testing data

Is the source of the data used to test the AI algorithm the same source as the data used to train it? If so, how was the testing data kept separate from the training data? If the AI algorithm was trained on the same patients on which it was evaluated; it could memorise patients’ personal characteristics with particular diseases, falsely improving its accuracy on the testing data.

2. Multi-sourced data

The source of the testing data is also important when considering whether the results of the paper are applicable in your particular setting. The testing data’s inclusion and exclusion criteria will indicate if the patients are similar to those in your practice.

Example: The testing data may use positive tuberculosis patients from India but negative controls from Australia; a practice known as a Frankenstein dataset [2]. This is unlikely to mirror your patient population since it’s unlikely that your practice sees patients from both India and Australia at the same time.

3. Hidden stratification

Subgroups within the patient population that are of particular interest to you should be evaluated separately. Good overall performance of the AI algorithm may mask underlying deficiencies.

Example: The AI algorithm may detect pneumothoraces on chest radiographs very well, but may not detect pneumothorax in patients without an intercostal chest drain, otherwise known as hidden stratification[3]. Since intercostal drains are obvious on a chest radiograph, AI algorithms often pick up patients with a pneumothorax who have been treated with a drain quite easily. But it may fail in cases where the patient has yet to be treated, which are more concerning for clinicians.

4. Plausibility and assumption

The training procedure should also be assessed for its plausibility. Staying on the topic of pneumothorax, many papers describe AI algorithms that use 256×256 images to detect pneumothorax on chest radiographs[4]. This is implausible from a medical perspective as we know that the radiological signs of pneumothorax are often very subtle and require high resolution images to detect. This suggests that the AI algorithm may be using other features, such as the presence of an intercostal drain, to identify the pneumothorax, instead of the salient features themselves.

5. Baseline

Establish a robust baseline for the AI algorithm to compare itself to, and a clinically valid metric that such a comparison could be made. The common metric used in diagnostic radiology is the area under the receiver operating curve (AUC), which measures the ability of the AI algorithm to discriminate between positive and negative cases. An AUC of 0.5 indicates that the algorithm is no better than random guessing, and an AUC of 1.0 indicates perfect discrimination. It is less obvious however what a “good” or “acceptable” AUC is, as different modalities and diseases will have different diagnostic performance.

A common baseline used in many studies is that the AI algorithm should be non-inferior to the current standard of care – which is typically a radiologist or clinician evaluating the study without AI. Such a baseline may be established retrospectively or prospectively. Retrospective baselines may not reflect true clinical performance as readers may be less motivated than in clinical practice, or conversely, may overestimate clinical performance if readers are primed to search for the pathology being evaluated.

Conclusion

Many of these principles are discussed in existing checklists such as CLAIM [5], and this article only touches on the basics of evaluating such papers. As the field progresses and matures, the evidence base will grow in size and complexity, but radiologists and clinicians will always hold ultimate responsibility of using AI algorithms in clinical practice safely.


References

1.  Cohen JF, Korevaar DA, Altman DG, et al (2016) STARD 2015 guidelines for reporting diagnostic accuracy studies: Explanation and elaboration. BMJ Open 6:e012799. https://doi.org/10.1136/bmjopen-2016-012799 

2.  Roberts M, Driggs D, Thorpe M, et al (2021) Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans. Nat Mach Intell 3:199–217. https://doi.org/10.1038/s42256-021-00307-0 

3.  Oakden-Rayner L, Dunnmon J, Carneiro G, Re C (2020) Hidden stratification causes clinically meaningful failures in machine learning for medical imaging. In: ACM CHIL 2020 – Proceedings of the 2020 ACM Conference on Health, Inference, and Learning. Association for Computing Machinery, Inc, pp 151–159 

4.  Rajpurkar P, Irvin J, Zhu K, et al (2017) CheXNet: Radiologist-Level Pneumonia Detection on Chest X-Rays with Deep Learning 

5.  Mongan J, Moy L, Kahn CE (2020) Checklist for Artificial Intelligence in Medical Imaging (CLAIM): A Guide for Authors and Reviewers. Radiol Artif Intell 2:e200029. https://doi.org/10.1148/ryai.2020200029 

About Harrison.ai

Harrison.ai is a global healthcare technology company on a mission to urgently scale healthcare capacity through AI-powered medical imaging diagnostic support and workflow solutions. Its radiology (Annalise.ai) and pathology (Franklin.ai) solutions help clinicians deliver faster, more accurate diagnoses, aiding in the early detection of cancer and other medical conditions. Harrison.ai partners with hospitals and private clinics across APAC, EMEA, the United Kingdom and United States. For more information, visit Harrison.ai.

More news from Harrison.ai

AI ‘co-pilot’ used to speed up cancer diagnosis

AI software has been introduced by a West Yorkshire NHS Trust in a bid to speed up diagnosis of conditions such as lung cancer and infections.

Hospital uses AI technology to detect lung cancer

A North Yorkshire hospital has introduced AI technology that could speed up the detection of lung cancer and other serious illnesses.

Med-tech Harrison.ai’s new genAI tool outperforms OpenAI, Gemini

Australian medical technology company Harrison.ai has launched a radiology-specific generative AI model – and it’s already scoring higher than OpenAI and Gemini.

Ready to unlock infinite medical possibility?