paper paper
demo
Laboratory of Biomechanical and Image Guided Surgical Systems, Johns Hopkins University
What are SAM, SAMM and SAMME?
- SAM is the vision foundation model developed by Meta, Segment Anything.
- SAMM is an engineering integration of SAM to 3D Slicer, intended for medical image segmentation. The name is the abbreviation of Segment Any Medical Model.
- SAMME is an extended version of SAMM supporting not only the vanilla SAM, but new variants from the community.
Why SAMM and SAMME? - SAMM was a side project for fun initially (you can tell from the name). Later we got some interests from people because hey it's a new model and it's cool. Accurate image segmentation is crucial for medical image analysis as it enables clinicians to extract meaningful information from the image. It also allows for the detection of subtle changes in the tissue or organ of interest, which is essential for monitoring disease progression and treatment response.
- This later became aiming to develop an integration for Segment Anything Model (SAM) and 3D Slicer for future development and validation of the potentials of transferring Foundation Model to the medical image analysis field. More and more variants of SAM emerged, so we thought it's probably worth it to have a platform supporting the addition of new SAMs.
- YichiZhang98 has compiled a nice history of SAM for medical images. \
Make sure you have more than 8GB of VRAM so that it doesn't crash. \
Also see the Installation Guide.
Watch this video. Here are some key points demonstrated in the video.
- 00:00 Start Server
- 00:20 Pick a Model
- 00:22 Calculate Embeddings
- 00:37 3D Bounding Box
- 01:20 Switch Models
- 01:41 Fix Bad Results by Adjust Brightness and Contrast
- 02:33 2D Bounding Box
- 02:45 Prompt Propagation
- 03:05 Change Prompting View
- 03:27 Point Prompts
- 3 View Inference
- Embedding saving
- Data type
- Volume
- 2D Image
- RGB Image (WIP)
- models
- interactions
- positive and negative points
- 2D bounding box
- 3D bounding box
- combination
- automatic segmentation
- training (WIP)
The inference can run in real time (on a 3090) once the embeddings of the images are calculated and loaded. If you'd like to add your own SAM variant, make sure the implementation keeps the Predictor class intact or uses the same interface to call. Make sure the architecture follows the similar component make ups as in the vanilla SAM.
Works both on Linux and Windows. Has Mac support, but not tested yet.
This assumes Cuda, cv2 and pytorch are in your environment.
git clone git@github.com:bingogome/samm.git
conda create --name samm
conda activate samm
pip install git+https://github.com/facebookresearch/segment-anything.git
pip install opencv-python pycocotools matplotlib onnxruntime onnx timm
git clone https://github.com/bowang-lab/MedSAM
pip install -e MedSAM
pip install git+https://github.com/ChaoningZhang/MobileSAM.git
If you are using Windows, it's okay if you don't install pycocotools.
cd samm/samm-python-terminal
mkdir samm-workspace
Then, move the check point files in the samm-workspace
folder.
Start 3D Slicer, in the Python Console:
slicer.util.pip_install("pyyaml")
slicer.util.pip_install("pyzmq")
slicer.util.pip_install("tqdm")
SD Slicer
→ Developer Tools
→ Extension Wizard
.
Extension Tools
→ Select Extension
→ import the samm/samm folder.
Back to terminal, cd to root folder samm
python ./samm-python-terminal/sam_server.py
If it throws an error missing "sam_vit_h_4b8939.pth", move segment-anything/notebooks/sam_vit_h_4b8939.pth to samm/samm-python-terminal/samm-workspace
Follow the demo and Segment Any Medical Model away!
If you use SAMM in your research, please consider use the following BibTeX entry.
@article{liu2024segment,
title={Segment Any Medical Model Extended},
author={Liu, Yihao and Zhang, Jiaming and Diaz-Pinto, Andres and Li, Haowei and Martin-Gomez, Alejandro and Kheradmand, Amir and Armand, Mehran},
journal={arXiv preprint arXiv:2403.18114},
year={2024}
}
@article{liu2023samm,
title={SAMM (Segment Any Medical Model): A 3D Slicer Integration to SAM},
author={Liu, Yihao and Zhang, Jiaming and She, Zhangcong and Kheradmand, Amir and Armand, Mehran},
journal={arXiv preprint arXiv:2304.05622},
year={2023}
}