This repository contains the implementation of the following paper:
Customized Segment Anything Model for Medical Image Segmentation
Kaidong Zhang, and Dong Liu
Technical report
[Paper]
- Thanks to the high investment from my supevisor, I can finetune the
vit_h
version of SAM for more accurate medical image segmentation. Now, we release thevit_h
version of SAMed (We denote this version as SAMed_h), and the comparison between SAMed and SAMed_h is shown in the table below.
Model | DSC | HD | Aorta | Gallbladder | Kidney (L) | Kidney (R) | Liver | Pancreas | Spleen | Stomach |
---|---|---|---|---|---|---|---|---|---|---|
SAMed | 81.88 | 20.64 | 87.77 | 69.11 | 80.45 | 79.95 | 94.80 | 72.17 | 88.72 | 82.06 |
SAMed_h | 84.30 | 16.02 | 87.81 | 74.72 | 85.76 | 81.52 | 95.76 | 70.63 | 90.46 | 87.77 |
Without bells and whistles, SAMed_h achieves much higher performance than SAMed. Although the model size of vit_h
version is much larger (above 2G) than vit_b
version (~350M), the LoRA checkpoint of SAMed_h does not increase a lot (from 18M to 21M). Therefore, the deployment and storage cost of SAMed_h is nearly on par with SAMed. Since industry prefers to deploy larger and better performing models, we believe SAMed_h is more promising for computer-assisted diagnosis and preoperative planning in practice. For more details about SAMed_h, please visit SAMed_h directory.
- Make a demo.
- Finetune on more datasets
-
Make SAMed based onvit_l
orvit_h
mode of SAM
- Linux (We tested our codes on Ubuntu 18.04)
- Anaconda
- Python 3.7.11
- Pytorch 1.9.1
To get started, first please clone the repo
git clone https://github.com/hitachinsk/SAMed.git
Then, please run the following commands:
conda create -n SAMed python=3.7.11
conda activate SAMed
pip install -r requirements.txt
If you have the raw Synapse dataset, we provide the preprocess script to process and normalize the data for training. Please refer this folder for more details.
We strongly recommand you to try our online demo .
Currently, we provide the SAMed and the SAMed_s models for reproducing our results quickly. The LoRA checkpoints and their corresponding configurations are shown in the table below.
Model | Checkpoint | Configuration | DSC | HD |
---|---|---|---|---|
SAMed | Link | Cfg | 81.88 | 20.64 |
SAMed_s | Link | Cfg | 77.78 | 31.72 |
Here are the instructions:
- Change the directory to the rootdir of this repository.
- Please download the pretrained SAM model (provided by the original repository of SAM) and the LoRA checkpoint of SAMed. Put them in the
./checkpoints
folder. - Please download the testset and put it in the ./testset folder. Then, unzip and delete this file.
- Run this commend to test the performance of SAMed.
python test.py --is_savenii --output_dir <Your output directory>
If everything works, you can find the average DSC is 0.8188 (81.88) and HD is 20.64, which correspond to the Tab.1 of the paper. And check the test results in <Your output directory>
.
What's more, we also provide the SAMed_s model, which utilizes LoRA to finetune the transformer blocks in image encoder and mask decoder. Compared with SAMed, SAMed_s has smaller model size but the performance also drops slightly. If you want to use this model, download and put it in the ./checkpoints_s
folder and run the below command to test its performance.
python test.py --is_savenii --output_dir <Your output directory> --lora_ckpt checkpoints_s/epoch_159.pth --module sam_lora_image_encoder_mask_decoder
The average DSC is 0.7778 (77.78) and HD is 31.72 for SAMed_s, which corresponds to the Tab.3 of the paper.
We use 2 RTX 3090 GPUs for training.
- Please download the processed training set, whose resolution is
224x224
, and put it in<Your folder>
. Then, unzip and delete this file. We also prepare the training set with resolution512x512
for reference, the224x224
version of training set is downsampled from the512x512
version. - Run this command to train SAMed.
python train.py --root_path <Your folder> --output <Your output path> --warmup --AdamW
Check the results in <Your output path>
.
This work is licensed under MIT license. See the LICENSE for details.
If our work inspires your research or some part of the codes are useful for your work, please cite our paper:
@article{samed,
title={Customized Segment Anything Model for Medical Image Segmentation},
author={Kaidong Zhang, and Dong Liu},
journal={arXiv preprint arXiv:2304.13785},
year={2023}
}
If you have any questions, please contact us via
We appreciate the developers of Segment Anything Model and the provider of the Synapse multi-organ segmentation dataset. The code of SAMed is built upon TransUnet and SAM LoRA, and we express our gratitude to these awesome projects.