The official baseline model for "MICCAI 2021 FLARE Challenge: Fast and Low GPU memory Abdominal oRgan sEgmentation", derived from nnUNet [1].
Download models from
- Baidu Net Disk (PW: 2021).
- Google Drive
git clone https://github.com/YaoZhang93/FLARE21nnUNetBaseline.git
cp ./FLARE21nnUNetBaseline/nnUNet.zip ./FLARE21/datasets
cd ./FLARE21nnUNetBaseline/FLARE21/datasets/
unzip nnUNet.zip
cd ../..
python inference/predict_simple.py -i INPUT_FOLDER -o OUTPUT_FOLDER -t Task000_FLARE21Baseline -m CONFIGURATION
INPUT_FOLDER
is the folder path that contains nii files for testingOUTPUT_FOLDER
is the folder path for the predictions of the baseline modelCONFIGUREATION
should be2d
or3d_fullres
referring to 2D or 3D models, respectively.
For more usage, please refer to the repositry of nnUNet.
python get_params.py -m CONFIGURATION
CONFIGURATION
should be2d
or3d_fullres
referring to 2D or 3D models, respectively.
The output for 2d
should be:
Total params: 41,268,192
Trainable params: 41,268,192
Non-trainable params: 0
The output for 3d_fullres
should be:
Total params: 30,787,584
Trainable params: 30,787,584
Non-trainable params: 0
We use torchsummary
to get the summary of the model. A simple usage is
from torchsummary import summary
# The input_size of the baseline model is 1*80*192*160
summary(model, input_size)
Please refer to pytorch-summary document for more details.
We encourage the participants to use it for the analysis of the models. get_params.py
could be an example to adapt it to your own model.
python get_flops.py -m CONFIGURATION
CONFIGURATION
should be2d
or3d_fullres
referring to 2D or 3D models, respectively.
The output for 2d
should be:
Total FLOPs: 61307143168
The output for 3d_fullres
should be:
Total FLOPs: 590861472000
We use fvcore
to get the FLOPs of the model. A simple usage is
from fvcore import FlopCountAnalysis
# The input_size of the baseline model is 1*1*80*192*160
inputs = (torch.randn(input_size),)
flops = FlopCountAnalysis(model, inputs)
Please refer to fvcore document for more details.
We encourage the participants to use it for the analysis of the models. get_flops.py
could be an example to adapt it to your own model.
Please refer to FLARE21 Evaluation Code.
First, compress the segmentation results by
zip -r TeamNameVal1.zip OUTPUT_FOLDER
FLARE21/BaselineVal1.zip
is an example generated by the baseline model.
Then, submit the results on FLARE21 Submission Page.
Build a docker image of the model by
docker build -t docker_image_name .
The configuration of the docker image is in Dokcerfile
. It will call predict.sh
when starting the docker image.
Please refer to the dockerhub and the video tutorial for more details.
[1] Isensee, Fabian, Paul F Jaeger, Simon A A Kohl, Jens Petersen, and Klaus H Maier-Hein. 2021. “nnU-Net: A Self-Configuring Method for Deep Learning-Based Biomedical Image Segmentation.” Nature Methods 18 (2): 203–11.