Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Trustworthy clinical AI solutions: a unified review of uncertainty quantification in deep learning models for medical image analysis (2210.03736v1)

Published 5 Oct 2022 in eess.IV, cs.AI, cs.CV, and cs.LG

Abstract: The full acceptance of Deep Learning (DL) models in the clinical field is rather low with respect to the quantity of high-performing solutions reported in the literature. Particularly, end users are reluctant to rely on the rough predictions of DL models. Uncertainty quantification methods have been proposed in the literature as a potential response to reduce the rough decision provided by the DL black box and thus increase the interpretability and the acceptability of the result by the final user. In this review, we propose an overview of the existing methods to quantify uncertainty associated to DL predictions. We focus on applications to medical image analysis, which present specific challenges due to the high dimensionality of images and their quality variability, as well as constraints associated to real-life clinical routine. We then discuss the evaluation protocols to validate the relevance of uncertainty estimates. Finally, we highlight the open challenges of uncertainty quantification in the medical field.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Benjamin Lambert (8 papers)
  2. Florence Forbes (35 papers)
  3. Alan Tucholka (4 papers)
  4. Senan Doyle (9 papers)
  5. Harmonie Dehaene (1 paper)
  6. Michel Dojat (18 papers)
Citations (46)