Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
184 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SonoNet: Real-Time Detection and Localisation of Fetal Standard Scan Planes in Freehand Ultrasound (1612.05601v2)

Published 16 Dec 2016 in cs.CV

Abstract: Identifying and interpreting fetal standard scan planes during 2D ultrasound mid-pregnancy examinations are highly complex tasks which require years of training. Apart from guiding the probe to the correct location, it can be equally difficult for a non-expert to identify relevant structures within the image. Automatic image processing can provide tools to help experienced as well as inexperienced operators with these tasks. In this paper, we propose a novel method based on convolutional neural networks which can automatically detect 13 fetal standard views in freehand 2D ultrasound data as well as provide a localisation of the fetal structures via a bounding box. An important contribution is that the network learns to localise the target anatomy using weak supervision based on image-level labels only. The network architecture is designed to operate in real-time while providing optimal output for the localisation task. We present results for real-time annotation, retrospective frame retrieval from saved videos, and localisation on a very large and challenging dataset consisting of images and video recordings of full clinical anomaly screenings. We found that the proposed method achieved an average F1-score of 0.798 in a realistic classification experiment modelling real-time detection, and obtained a 90.09% accuracy for retrospective frame retrieval. Moreover, an accuracy of 77.8% was achieved on the localisation task.

Citations (296)

Summary

  • The paper proposes a CNN-based framework, SonoNet, for real-time detection and localization of fetal ultrasound scan planes.
  • It demonstrates an optimal trade-off between computational efficiency and accuracy, with SonoNet-32 achieving 90.09% retrieval accuracy and an F1-score of 0.798.
  • The work paves the way for automated diagnostics and enhanced training for novice sonographers in resource-limited settings.

Overview of SonoNet: A Deep Dive into Automated Fetal Ultrasound Scan Plane Detection

The research presented in the paper "SonoNet: Real-Time Detection and Localisation of Fetal Standard Scan Planes in Freehand Ultrasound" addresses a critical challenge in the domain of obstetric ultrasonography: the real-time identification and localization of standard fetal scan planes in freehand ultrasound data. This paper contributes to a significant advancement in the field by leveraging convolutional neural networks (CNNs) to facilitate automated scan plane detection and localization, a task traditionally reliant on highly trained sonographers.

Methodology and Network Design

The authors introduce SonoNet, a CNN architecture inspired by the VGG16 model, designed to operate in real-time and achieve optimal performance in detecting and localizing fetal structures. The paper explores several iterations of the SonoNet architecture (SonoNet-64, SonoNet-32, and SonoNet-16), each varying in complexity and capability to handle real-time constraints.

Key design choices include:

  • Transitioning from fully connected layers to a fully convolutional design, thus enabling efficient processing on arbitrary input sizes.
  • Incorporating batch normalization layers to facilitate faster training convergence and enhance performance across different network architectures.
  • Adopting a mean pooling strategy over max pooling to capture full image context, vital for distinguishing between very similar classes, particularly in ultrasound images where variations can be subtle.

These architectural decisions are underpinned by rigorous experiments showcasing the trade-offs between computational efficiency and detection accuracy, making the case for SonoNet-32 as the optimal balance suitable for real-time clinical deployment.

Results

The network's performance is tested on a dataset of over 2,600 ultrasound examinations, highlighting two usage scenarios: real-time frame annotation and retrospective scan plane retrieval. Impressively, SonoNet-32 achieves an average retrieval accuracy of 90.09% for standard planes in retrospective scenarios, underscoring its applicability in automated archival data analysis. Furthermore, the network differentiates between 13 fetal views and background frames with an F1-score of 0.798 in a real-time detection setting, marking a substantial leap in automated ultrasound interpretation.

The paper also reports the weakly supervised localisation accuracy, achieving a mean IOU of 0.62 with bounding boxes correctly identified in 77.8% of cases, a promising result given the variability and noise inherent in ultrasound imaging. The authors employ a saliency-driven backward pass to generate category-specific saliency maps, offering a finer granularity of localization—a task challenging even for human experts.

Implications and Future Directions

The implications of this research extend beyond automated detection, offering promising avenues for improving the training processes for novice sonographers through real-time feedback systems, potentially mitigating the shortage of skilled personnel in this domain. The integration of SonoNet with ultrasound devices could revolutionize access to diagnostics in under-resourced regions, where skilled expertise is scarce.

From a theoretical perspective, the work stimulates further research into improving CNN-based weakly supervised learning applications, challenging the status quo of requiring extensive annotation efforts. Future investigations could explore integrating temporal data to enhance sequence-based decision making, which might improve the detection of dynamic structures such as those observed in cardiac views.

In summary, SonoNet represents a significant stride toward fully autonomous and accessible fetal ultrasound diagnostics. The proposed framework sets the stage for future developments that can expand upon its capabilities, paving the way for more comprehensive automated analysis systems in medical imaging.

Youtube Logo Streamline Icon: https://streamlinehq.com