This repo provides a quick start for the ISIC2024 challenge. It offers baseline implementations for reference. We evaluate representative baselines on a sample of the ISIC2024 dataset and explore practical strategies. For details, see the "ISIC2024_demo.csv Statistics" and "Evaluating Baselines on ISIC2024_demo.csv" sections.
Create the environment and install packages
conda create -n env_name python=3.9 -y
conda activate env_name
pip install torch==2.2.2 torchvision==0.17.2 torchaudio==2.2.2 --index-url https://download.pytorch.org/whl/cu118
pip install -r requirements.txt
ISIC2024_demo: download ISIC2024 images from here (SLICE-3D, it should contains 401,059 JPEG images)
ISIC2024_demo_GT: ISIC2024_demo.csv can be found in this repo.
cd CNN_baselines
# This command will train and evaluate weighted ResNet50 with 3 different seeds.
bash run_resnet50.sh
cd ViT_baselines
# This command will train and evaluate weighted ViT-base with 3 different seeds.
bash run_vit.sh
# This command will train and evaluate weighted DINOv2-base with 3 different seeds.
bash run_dinov2.sh
In this repo, we evaluate various models on a subset of ISIC2024, comprising 49,025 images out of the full 401,058 image dataset. The test set includes all images from ACEMID MIA, while the train and validation sets contain images sampled from the remaining attributions.
Split | Count |
---|---|
Total | 49,025 |
Test | 28,665 |
Train | 16,288 |
Validation | 4,072 |
Split | Benign | Malignant |
---|---|---|
Test | 28,632 | 33 |
Train | 16,000 | 288 |
Validation | 4,000 | 72 |
- Test: 0.12%
- Train: 1.77%
- Validation: 1.77%
Attribution | Count |
---|---|
ACEMID MIA | 28,665 |
Memorial Sloan Kettering Cancer Center | 6,979 |
Department of Dermatology, Hospital Clínic de Barcelona | 5,789 |
University Hospital of Basel | 3,633 |
Frazer Institute, The University of Queensland, Dermatology Research Centre | 2,838 |
ViDIR Group, Department of Dermatology, Medical University of Vienna | 682 |
Department of Dermatology, University of Athens | 439 |
Attribution | Benign | Malignant |
---|---|---|
ACEMID MIA | 28,632 | 33 |
Department of Dermatology, Hospital Clínic de Barcelona | 5,717 | 72 |
Department of Dermatology, University of Athens | 433 | 6 |
Frazer Institute, The University of Queensland | 2,757 | 81 |
Memorial Sloan Kettering Cancer Center | 6,805 | 174 |
University Hospital of Basel | 3,620 | 13 |
ViDIR Group, Department of Dermatology, Medical University of Vienna | 668 | 14 |
Model performance may vary depending on hardware. We conduct all experiments on single RTX6000Ada GPU.
Model | ROC AUC | Specificity | Sensitivity |
---|---|---|---|
ResNet50 (non-weighted) | 0.7300 (±0.13) | 0.9795 (±0.03) | 0.0200 (±0.03) |
ResNet50 (weighted) | 0.8278 (±0.03) | 0.8648 (±0.08) | 0.6061 (±0.09) |
EfficientNet-B1 (weighted) | 0.8069 (±0.01) | 0.9366 (±0.001) | 0.4848 (±0.08) |
EfficientNet-B7 (weighted) | 0.8538 (±0.04) | 0.9459 (±0.02) | 0.4646 (±0.03) |
ViT_base (weighted) | 0.8342 (±0.05) | 0.8612 (±0.11) | 0.6465 (±0.12) |
DINOv2-base (weighted) | 0.7437 (±0.24) | 0.6077 (±0.49) | 0.7475 (±0.20) |
- ISIC2024 presents a highly imbalanced machine learning task. Training a model using cross-entropy loss alone results in very low sensitivity.
- We found that simple random weighted sampling is very effective in alleviating this issue.
- Adjusting the prediction threshold should also be useful for improving sensitivity.
- The non-weighted ResNet50 model achieves high specificity but fails to detect any malignant cases (0% sensitivity).
- Weighted sampling significantly improves sensitivity for various models.
- The weighted ResNet50 and weighted ViT-base model shows the best balance between specificity and sensitivity.
- Experiment with different sampling techniques and loss functions designed for imbalanced datasets.
- Fine-tune the prediction threshold to optimize the trade-off between specificity and sensitivity.
- Explore ensemble methods combining multiple models to potentially improve overall performance.
The baselines are provided by the AIM lab and MMAI, led by A/Prof. Zongyuan Ge and Prof. Victoria Mar from Monash University.