Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Skin Lesion Analysis Toward Melanoma Detection: A Challenge at the 2017 International Symposium on Biomedical Imaging (ISBI), Hosted by the International Skin Imaging Collaboration (ISIC) (1710.05006v3)

Published 13 Oct 2017 in cs.CV

Abstract: This article describes the design, implementation, and results of the latest instaLLMent of the dermoscopic image analysis benchmark challenge. The goal is to support research and development of algorithms for automated diagnosis of melanoma, the most lethal skin cancer. The challenge was divided into 3 tasks: lesion segmentation, feature detection, and disease classification. Participation involved 593 registrations, 81 pre-submissions, 46 finalized submissions (including a 4-page manuscript), and approximately 50 attendees, making this the largest standardized and comparative study in this field to date. While the official challenge duration and ranking of participants has concluded, the dataset snapshots remain available for further research and development.

Insightful Analysis of the ISBI 2017 ISIC Challenge for Melanoma Detection

This essay provides an in-depth analysis of the paper titled “Skin Lesion Analysis Toward Melanoma Detection: A Challenge at the 2017 International Symposium on Biomedical Imaging (ISBI), hosted by the International Skin Imaging Collaboration (ISIC)”. The paper details the design, execution, and outcomes of a benchmark challenge aimed at advancing algorithms for the automated diagnosis of melanoma, a highly lethal form of skin cancer.

Challenge Overview

Organized by ISIC, the 2017 challenge was the second of its kind, following an initial effort in 2016. The challenge saw significant participation, including 593 registrations, 81 pre-submissions, 46 finalized entries, and approximately 50 attendees at the associated ISBI workshop. The competition was structured into three tasks: lesion segmentation, dermoscopic feature detection, and disease classification.

Dataset and Task Specifications

The challenge utilized an extensive dataset including images and corresponding ground truth annotations, separated into training (2000 images), validation (150 images), and test (600 images) datasets. This segmentation allowed participants to refine their models using validation data and assess final performance on a holdout test set.

  1. Lesion Segmentation: Tasked participants with generating binary masks to predict lesion boundaries in dermoscopic images.
  2. Dermoscopic Feature Classification: Focused on detecting specific dermoscopic features like pigment networks and milia-like cysts within superpixels derived from lesion images.
  3. Disease Classification: Required participants to classify lesions as melanoma, seborrheic keratosis, or benign nevi, with a confidence score reflecting the disease probability.

Evaluation Metrics

Evaluation of the submissions was based on several metrics:

  • Segmentation: Jaccard Index, Dice coefficient, and pixel-wise accuracy.
  • Classification: Area Under Curve (AUC) of the ROC curve, with additional specificities measured at sensitivity levels correlated with clinical management performance (82%, 89%, and 95%).

Results and Insights

Lesion Segmentation

The top performing method achieved an average Jaccard Index of 0.765, an accuracy of 93.4%, and a Dice coefficient of 0.849 using a fully convolutional network ensemble. Despite these strong numerical results, about 26% of images had a Jaccard Index falling below 0.7, indicating room for improvement in the segmentation task, particularly for outlier images.

Dermoscopic Feature Classification

This task saw limited submissions but provided meaningful insights. The highest AUC achieved was 0.895, demonstrating the feasibility of robust feature detection using automated methods. Techniques here utilized deep learning frameworks tailored to clinical dermoscopic feature recognition.

Disease Classification

A larger set of 23 test submissions was analyzed. Ensembles of deep learning models using additional external data sources showed superior performance, with the leading method attaining an average AUC of 0.911. One key takeaway was that classifying seborrheic keratosis appeared easier than melanoma, potentially due to inherent dataset biases.

Implications and Future Directions

The results have several practical and theoretical implications. For segmentation, revisiting evaluation metrics to better align them with clinical relevance (e.g., binary failure rates) could enhance the assessment framework. Additionally, the low participation in dermoscopic feature detection suggests a need to adapt task formats to be more congruent with prevailing image detection benchmarks.

From a broader perspective, it is clear that deep learning, especially when augmented with large and diverse training datasets, holds substantial promise for advancing automated skin lesion analysis. The integrative approach of collaborative fusion, where predictions from multiple models are amalgamated, proved beneficial, reaffirming findings from the 2016 challenge.

Conclusion

The ISBI 2017 ISIC Challenge represented a critical milestone in leveraging machine learning for melanoma detection from dermoscopic images. While substantial progress was made, particularly in segmentation and classification tasks, addressing dataset biases, enhancing participation in feature detection, and refining evaluation metrics are essential future steps. The ongoing availability of these datasets ensures that this field will continue to evolve, fostering the development of improved diagnostic algorithms and ultimately contributing to better clinical outcomes in dermatology.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (11)
  1. Noel C. F. Codella (12 papers)
  2. David Gutman (7 papers)
  3. M. Emre Celebi (25 papers)
  4. Brian Helba (5 papers)
  5. Michael A. Marchetti (1 paper)
  6. Stephen W. Dusza (1 paper)
  7. Aadi Kalloo (2 papers)
  8. Konstantinos Liopyris (2 papers)
  9. Nabin Mishra (2 papers)
  10. Harald Kittler (10 papers)
  11. Allan Halpern (6 papers)
Citations (1,884)