UltraLight VM-UNet

Parallel Vision Mamba Significantly Reduces Parameters for Skin Lesion Segmentation

Renkai Wu1, Yinghao Liu2, Pengchen Liang1*, Qing Chang1*

1 Shanghai University, 2 University of Shanghai for Science and Technology

ArXiv Preprint (arXiv:2403.20035)

🔥🔥Highlights🔥🔥

1.The UltraLight VM-UNet has only 0.049M parameters, 0.060 GFLOPs, and a model weight file of only 229.1 KB.

2.Parallel Vision Mamba (or Mamba) is a winner for lightweight models.

News🚀

(2024.04.09) The second version of our paper has been uploaded to arXiv with adjustments to the description in the methods section.

(2024.04.04) Added preprocessing step for private datasets.

(2024.04.01) The project code has been uploaded.

(2024.03.29) The first edition of our paper has been uploaded to arXiv. 📃

Abstract

Traditionally for improving the segmentation performance of models, most approaches prefer to use adding more complex modules. And this is not suitable for the medical field, especially for mobile medical devices, where computationally loaded models are not suitable for real clinical environments due to computational resource constraints. Recently, state-space models (SSMs), represented by Mamba, have become a strong competitor to traditional CNNs and Transformers. In this paper, we deeply explore the key elements of parameter influence in Mamba and propose an UltraLight Vision Mamba UNet (UltraLight VM-UNet) based on this. Specifically, we propose a method for processing features in parallel Vision Mamba, named PVM Layer, which achieves excellent performance with the lowest computational load while keeping the overall number of processing channels constant. We conducted comparisons and ablation experiments with several state-of-the-art lightweight models on three skin lesion public datasets and demonstrated that the UltraLight VM-UNet exhibits the same strong performance competitiveness with parameters of only 0.049M and GFLOPs of 0.060. In addition, this study deeply explores the key elements of parameter influence in Mamba, which will lay a theoretical foundation for Mamba to possibly become a new mainstream module for lightweighting in the future.

Different Parallel Vision Mamba (PVM Layer) settings:

Setting Briefly Params GFLOPs DSC
1 No paralleling ( Channel number C) 0.136M 0.060 0.9069
2 Double parallel ( Channel number (C/2)+(C/2)) 0.070M 0.060 0.9073
3 Quadruple parallel ( Channel number (C/4)+(C/4)+(C/4)+(C/4)) 0.049M 0.060 0.9091

0. Main Environments.
The environment installation procedure can be followed by VM-UNet, or by following the steps below (python=3.8):

conda create -n vmunet python=3.8
conda activate vmunet
pip install torch==1.13.0 torchvision==0.14.0 torchaudio==0.13.0 --extra-index-url https://download.pytorch.org/whl/cu117
pip install packaging
pip install timm==0.4.12
pip install pytest chardet yacs termcolor
pip install submitit tensorboardX
pip install triton==2.0.0
pip install causal_conv1d==1.0.0  # causal_conv1d-1.0.0+cu118torch1.13cxx11abiFALSE-cp38-cp38-linux_x86_64.whl
pip install mamba_ssm==1.0.1  # mmamba_ssm-1.0.1+cu118torch1.13cxx11abiFALSE-cp38-cp38-linux_x86_64.whl
pip install scikit-learn matplotlib thop h5py SimpleITK scikit-image medpy yacs

1. Datasets.
Data preprocessing environment installation (python=3.7):

conda create -n tool python=3.7
conda activate tool
pip install h5py
conda install scipy==1.2.1  # scipy1.2.1 only supports python 3.7 and below.
pip install pillow

A. ISIC2017

  1. Download the ISIC 2017 train dataset from this link and extract both training dataset and ground truth folders inside the /data/dataset_isic17/.
  2. Run Prepare_ISIC2017.py for data preparation and dividing data to train, validation and test sets.

B. ISIC2018

  1. Download the ISIC 2018 train dataset from this link and extract both training dataset and ground truth folders inside the /data/dataset_isic18/.
  2. Run Prepare_ISIC2018.py for data preparation and dividing data to train, validation and test sets.

C. PH2

  1. Download the PH2 dataset from this link and extract both training dataset and ground truth folders inside the /data/PH2/.
  2. Run Prepare_PH2.py to preprocess the data and form test sets for external validation.

D. Prepare your own dataset

  1. The file format reference is as follows. (The image is a 24-bit png image. The mask is an 8-bit png image. (0 pixel dots for background, 255 pixel dots for target))
  • './your_dataset/'
    • images
      • 0000.png
      • 0001.png
    • masks
      • 0000.png
      • 0001.png
    • Prepare_your_dataset.py
  1. In the 'Prepare_your_dataset.py' file, change the number of training sets, validation sets and test sets you want.
  2. Run 'Prepare_your_dataset.py'.

2. Train the UltraLight VM-UNet.

python train.py
  • After trianing, you could obtain the outputs in './results/'

3. Test the UltraLight VM-UNet.
First, in the test.py file, you should change the address of the checkpoint in 'resume_model'.

python test.py
  • After testing, you could obtain the outputs in './results/'

4. Additional information.

  • PVM Layer can be very simply embedded into any model to reduce the overall parameters of the model. Please refer to issue 7 for the methodology of calculating model parameters and GFLOPs.

Citation

If you find this repository helpful, please consider citing:

@article{wu2024ultralight,
  title={UltraLight VM-UNet: Parallel Vision Mamba Significantly Reduces Parameters for Skin Lesion Segmentation},
  author={Wu, Renkai and Liu, Yinghao and Liang, Pengchen and Chang, Qing},
  journal={arXiv preprint arXiv:2403.20035},
  year={2024}
}

Acknowledgement

Thanks to Vim, VMamba, VM-UNet and LightM-UNet for their outstanding work.