/Multiple-Sclerosis-Lesion-Segmentation

Implementation of my paper titled "A Light Weighted Deep Learning Framework for Multiple Sclerosis Lesion Segmentation" published in 2019 Fifth International Conference on Image Information Processing (ICIIP). https://ieeexplore.ieee.org/document/8985674

Primary LanguagePythonMIT LicenseMIT

Multiple-Sclerosis-Lesion-Segmentation

This is the implementation of my paper titled "A Light Weighted Deep Learning Framework for Multiple Sclerosis Lesion Segmentation" (https://ieeexplore.ieee.org/document/8985674) published in 2019 Fifth International Conference on Image Information Processing (ICIIP).

What is Multiple Sclerosis ?

MS is a long-term, or persistent, condition that affects the central nervous system (CNS) which often leads to disability. The CNS is made up of optic nerve, brain and spinal cord. Hardening of tissue in the body is defined as sclerosis. In MS, scar tissue appears in the CNS. This leads to travelling of messages between the brain and rest of the body in an uneven way.

Why MRI Scan ?

Magnetic resonance imaging (MRI) can reveal significant scars, also called lesions, on the brain or spinal cord.Lesions show up as white or dark spots, depending on the type of scan.

Dataset :

MICCAI 2016 Dataset : MRI Scans of 15 patients acquired in different image domains with 5 different modalities T1-w, MPRAGE, FLAIR, T1-w gadolinium enhanced and T2-w/DP contrast enhanced images. The scans were already provided with necessary pre-processing like image de-noising, intensity corrected and skull-stripped.

Preprocessing

  1. Each slice in each modality are concatenated along its 3rd dimension and used as an input.
  2. Resizing the input image is (256 * 256 * 5).
  3. Removing Null Samples.
  4. Normalization and Standardization:
    Normalisation(X) = X / max(X) + epsilon Standardization(X) = (X - mean(X))/(std(X) + epsilon) where epsilon = 0.00001 (a very small value) and X represents one channel of any modalityof size (h * w).

Network Architecture :

Schematic diagram of the proposed light weighted U-Net architecture

Training Phase :

  1. Total number of samples : 2834
  2. Input size : (256 * 256 * 5) a.
  3. Weights Initialization : Xavier initialization.
  4. Mini batch size of 32
  5. Optimizer : Adam
  6. Learning rate : 0.001.
  7. Loss function : Binary Cross Entropy
  8. Epochs : 60

Results :

  1. Accuracy : 96.79%
  2. DSC : 0.76
  3. Sensitivity : 0.65
  4. Specificity : 0.86

Plots :

Graphs Loss vs epoch-num Accuracy vs epoch-num Dice_Coefficient vs epoch-num

Thank you