Introduction of CTPelvic1K Dataset

To build a comprehensive pelvic CT dataset that can replicate practical appearance variations, we curate a large dataset of pelvic CT images (CTPelvic1K with 1,184 3D volumes, 320K CT slices) using the following seven sources.

  • CLINIC and CLINIC-metal. These two sub-datasets are related to pelvic fractures collected from an orthopedic hospital we collaborate with. CLINIC is collected from preoperative images without metal artifact, and CLINIC-metal ismainly collected from postoperative images with metal artifacts.
  • KITS19. This sub-dataset is from the Kits19 challenge 13 which is related to kidney and kidney tumor segmentation.
  • CERVIX and ABDOMEN. These two sub-datasets are from the Multi-Atlas Labeling Beyond the Cranial Vault - Workshop and Challenge 3. They are all multi-organ segmentation datasets for different body regions originally.
  • MSD T10. This sub-dataset comes from the 10th sub-dataset of Medical Segmentation Decathlon 31 and features colon tumor segmentation.
  • COLONOG. This sub-dataset comes from the CT COLONOGRAPHY 16 dataset related to a CT colonography trial. It has prone and supine DICOM images for each patient. We randomly select one of two positions, which have the similar information, of each patient to our large dataset.

KITS19, CERVIX, ABDOMEN, MSD T10, COLONOG, CLINIC, and CLINICmetal are curated separately from different sites and sources and hence have a diverse range of spacing and FOV. The overview of our large-scale CT Pelvic dataset (CTPelvic1K) and some pelvic CT image examples with various conditions are shown in Table 1 and Fig.1. At the same time, chyme, vascular sclerosis, coprolith, and other situations often encountered in the clinic appear in these sub-datasets. Among them, the data of COLONOG, CLINIC, and CLINIC-metal are stored in a DICOM format, so we can access the information about scanner manufacturer of these sub-datasets.


For more information about CTPelvic1K dataset, please read the following paper. Please also cite this paper if you are using CTPelvic1K dataset for your research!

Pengbo Liu, Hu Han, Yuanqi Du, Heqin Zhu, Yinhao Li, Feng Gu, Honghu Xiao, Jun Li, Chunpeng Zhao, Li Xiao, Xinbao Wu, 
S.Kevin Zhou. Deep Learning to Segment Pelvic Bones: Large-scale CT Datasets and Baseline Models. arXiv: 2012.08721 (2020).

We believe this large-scale dataset will promote the development of the whole community and plan to open source the images, annotations, and trained baseline models at this link.

Link:https://pan.baidu.com/s/1iFk21EWAN1nWaJTuYgSUwQ 
Passwd:mdv7 

This repository is still work in progress. Annotations will continue to be enriched with other bone structures and other further applications. If you encounter any problems while using CTPelvic1K dataset, please let us know.

Getting Started for Code

All the commands in this section assume that you are in a terminal and your working directory is the CTPelvic1K folder (the one that has all the subfolders like dataset_conversion, evaluation, ...)

And all commands are already writen in CTPelvic1K/nnunet/runs.py.

Table of Contents

1 Set paths

Have a look at the file CTPelvic1K/nnunet/paths.py and adapt it to your system by setting the paths where you will store raw data, preprocessed data and trained models.

2 Preparing Datasets & Experiment Planning & Preprocessing

Following commands in runs.py are used for dataset preparing, experiment planning and preprocessing.

home_dir = os.environ['HOME']
train_dir = os.path.join(home_dir,'all_data/nnUNet/rawdata/Task11_CTPelvic1K')
output_dir = os.path.join(home_dir, 'all_data/nnUNet/nnUNet_raw/Task11_CTPelvic1K')
$ command_1 = f'python dataset_conversion/JstPelvisSegmentation_5label.py --train_dir {train_dir} --output_dir {output_dir}'
$ command_2 = 'python experiment_planning/plan_and_preprocess_task.py -t Task11_CTPelvic1K -pl 20 -pf 20'

processed_path = os.path.join(home_dir, 'all_data/nnUNet/nnUNet_processed/Task11_CTPelvic1K')
check_save_path = os.path.join(home_dir, 'all_data/nnUNet/nnUNet_processed/Task11_CTPelvic1K/Task11_check')
$ command_3 = f'python preprocessing/lumbosacral_joint_sampling.py --processed_path {processed_path} --check_save_path {check_save_path}'

os.system(command)

train_dir is where the raw downloaded dataset is stored. command_1 will organize these images and labels to output_dir according to the format of Medical Segmentation Decathlon 1 datasets.

command_2 will analyze our dataset and determine how to train it bset with nnU-Net models 2. Task name (Task11_CTPelvic1K) can be set arbitrarily. -pl/-pf determines how many processes will be used for datatset analysis and preprocessing. Generally you want this number to be as high as you have CPU cores, unless you run into memory problems.

processed_path corresponds to the path storing training/validation data after preprocessing. Due to the serious imbalance between simple areas (anatomy of the bones) and difficult areas (sacroiliac joint and lumbosacral joint), we add oversampling operation for joints region in our experiments. command_3 can extract the coordinates of the target area and save them to the .pkl file (key: "Lumbosacral_Region") of each patient for sampling constraints during training. This imbalance problem mainly appears in 3D fullres scene, because of the high resolution of the CT images, so we only introduce this oversampling operation in 3D fullres (and 3D_cascade), i.e. stage1, experiments.

3 Experiments

The training set, validation set and testing set divisions of our experiments are shown in Table 1 and stored in CTPelvic1K/splits_final.pkl. We set up 22 folds in our experiments, each of which corresponds to an experiment setting (different sub-datasets are included in training phase). Experiments setting of different folds are shown below. For example, we train models on dataset1~6 in fold 0 setting and dataset2-6 (ex dataset1) in fold 7 setting.

Datasets Name: ABDOMEN COLONOG MSD_T10 KITS19 CERVIX CLINIC CLINIC-metal
Datasets index: dataset 1 dataset 2 dataset 3 dataset 4 dataset 5 dataset 6 dataset 7
fold: fold 0 fold 1 fold 2 fold 3 fold 4 fold 5 fold 6
Datasets: D1~6 D1 D2 D3 D4 D5 D6
fold: fold 7 fold 8 fold 9 fold 10 fold 11 fold 12
Datasets: ex D1 ex D2 ex D3 ex D4 ex D5 ex D6
Manufacturers Name: SIEMENS GE Philips TOSHIBA
Manufacturers index: Manu 1 Manu 2 Manu 3 Manu 4
fold: fold 21 fold 13 fold 14 fold 15 fold 16
Manufacturers: M1-4 M1 M2 M3 M4
fold: fold 17 fold 18 fold 19 fold 20
Manufacturers: ex M1 ex M2 ex M3 ex M4

3.1 Training Models

parameter settings:
    TASK = 'Task11_CTPelvic1K'
    FOLD = 0
    GPU = 0
$ command_4 = f'python run/run_training.py 2d nnUNetTrainer {TASK} {FOLD} --gpu {GPU}'
$ command_5 = f'python run/run_training.py 3d_fullres nnUNetTrainer {TASK} {FOLD} --gpu {GPU}'
$ command_6 = f'python run/run_training.py 3d_lowres nnUNetTrainer {TASK} {FOLD} --gpu {GPU}'
$ command_7 = f'python run/run_training.py 3d_cascade_fullres nnUNetTrainerCascadeFullRes {TASK} {FOLD} --gpu {GPU}'

nnU-Net 2 uses three different U-Net models and can automatically choose which of them to use. We can try them all on our CTPelvic1K dataset.

Trained models are stored in network_training_output_dir (specified in paths.py).

2D U-Net

command_4

3D U-Net (full resolution)

command_5

3D U-Net Cascade

The 3D U-Net cascade only applies to datasets where the patch size possible in the 'fullres' setting is too small relative to the size of the image data. If the cascade was configured you can run it as follows, otherwise this step can be skipped. command_6

After validation these models will automatically also predict the segmentations for the next stage of the cascade and save them in the correct spacing.

Then run: command_7

3.2 Validation

Just add --validation_only to the corresponding training command. --valbest means validating on the best model of validation set during training phase.

$ command_8  = f'python run/run_training.py 2d nnUNetTrainer {TASK} {FOLD} --gpu {GPU} --validation_only --valbest'
$ command_9  = f'python run/run_training.py 3d_fullres nnUNetTrainer {TASK} {FOLD} --gpu {GPU} --validation_only --valbest'
$ command_10 = f'python run/run_training.py 3d_lowres nnUNetTrainer {TASK} {FOLD} --gpu {GPU} --validation_only --valbest'
$ command_11 = f'python run/run_training.py 3d_cascade_fullres nnUNetTrainerCascadeFullRes {TASK} {FOLD} --gpu {GPU} --validation_only --valbest'

3.3 Testing

You can use trained models to predict test data. In order to be able to do so the test data must be provided in the same format as the training data. If you want to use the trained models in the Link above, you should modify the path in the value of key, 'init', in the dict saved as 'model_best.model.pkl'.

test_data_path = os.path.join(home_dir, 'all_data/nnUNet/rawdata/ipcai2021_ALL_Test')

$ command_12 = f'python inference/predict_simple.py ' \
             f'-i {test_data_path} ' \
             f'-o {test_data_path}/{TASK}__{my_output_identifier}__fold{FOLD}_2d_pred ' \
             f'-t {TASK} ' \
             f'-tr nnUNetTrainer ' \
             f'-m 2d ' \
             f'-f {FOLD} ' \
             f'--gpu {GPU}'

$ command_13 = f'python inference/predict_simple.py ' \
              f'-i {test_data_path} ' \
              f'-o {test_data_path}/{TASK}__{my_output_identifier}__fold{FOLD}_3dfullres_pred ' \
              f'-t {TASK} ' \
              f'-tr nnUNetTrainer ' \
              f'-m 3d_fullres ' \
              f'-f {FOLD} ' \
              f'--gpu {GPU}'


$ command_14 = f'python inference/predict_simple.py ' \
              f'-i {test_data_path} ' \
              f'-o {test_data_path}/{TASK}__{my_output_identifier}__fold{FOLD}_3dlowres_pred ' \
              f'-t {TASK} ' \
              f'-tr nnUNetTrainer ' \
              f'-m 3d_lowres ' \
              f'-f {FOLD} ' \
              f'--gpu {GPU} ' \
              f'--overwrite_existing 0'

my_task_lowres = TASK
my_output_identifier_lowres = 'CTPelvic1K' #your low_res experiment\'s "my_output_identifier" in path
$ command_15 = f'python inference/predict_simple.py ' \
              f'-i {test_data_path} ' \
              f'-o {test_data_path}/{TASK}__{my_output_identifier_lowres}__{my_output_identifier}__fold{FOLD}_3dcascadefullres_pred ' \
              f'-t {TASK} ' \
              f'-tr nnUNetTrainerCascadeFullRes ' \
              f'-m 3d_cascade_fullres ' \
              f'-f {FOLD} ' \
              f'-l {test_data_path}/{my_task_lowres}__{my_output_identifier_lowres}__fold{FOLD}_3dlowres_pred ' \
              f'--gpu {GPU} ' \
              f'--overwrite_existing 0'
...

To run inference for 3D U-Net model, use the command_13. If you wish to use the 2D U-Nets, you can set -m 2d instead of 3d_fullres.

To run inference with the cascade, run the following two commands: command_14, command_15. Here we first predict the low resolution segmentations and then use them for the second stage of the cascade.

3.4 Evaluation

$ command_16 = 'python ../evaluation.py'
$ command_17 = 'python ../save_evaluation_results2csv.py'
$ command_18 = 'python ../save_evaluation_results2csv_Manu.py'

command_16 can calculate the metrics (DC/HD) between the prediction results and the ground truth. Then we will get an evaluation*.pkl file that stores all metrics of each patient.

command_17 and command_18 can convert evaluation*.pkl to .csv files for each dataset and manufacturer.

4 Results

4.1 visualization of segmentation results

vis

4.2 visualization of SDF post-processing results compared with MCR method

post

Large fragments near the anatomical structure are kept with SDF post-processing but are removed by the MCR method.

5 Acknowledgement

Our code is mainly rewritten based on nnU-Net's 2, 4 code. Thanks to Febian, et al.'s excellent work, which is a big contribution to the community.

6 References

1 http://medicaldecathlon.com/
2 https://github.com/MIC-DKFZ/nnUNet
3 https://www.synapse.org/#!Synapse:syn3193805/wiki/89480
4 Fabian Isensee, Paul F. Jäger, Simon A. A. Kohl, Jens Petersen, Klaus H. Maier-Hein "Automated Design of Deep Learning Methods for Biomedical Image Segmentation" arXiv preprint arXiv:1904.08128 (2020).
13 Heller, N., Sathianathen, N., et al.: The kits19 challenge data: 300 kidney tumor cases with clinical context, CT semantic segmentations, and surgical outcomes. arXiv:1904.00445 (2019).
16 Johnson, C.D., Chen, M.H., et al.: Accuracy of CT colonography for detection of large adenomas and cancers. New England Journal of Medicine 359(12), 1207{1217 (2008).
31 Simpson, A.L., Antonelli, M., et al.: A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv:1902.09063 (2019).

7 Citation

If you use our CTPelvic1K dataset, please cite our paper:

Pengbo Liu, Hu Han, Yuanqi Du, Heqin Zhu, Yinhao Li, Feng Gu, Honghu Xiao, Jun Li, Chunpeng Zhao, Li Xiao, Xinbao Wu, 
S.Kevin Zhou. Deep Learning to Segment Pelvic Bones: Large-scale CT Datasets and Baseline Models. arXiv: 2012.08721 (2020).