Skip to content

Hetero-Modal Variational Encoder-Decoder for Joint Modality Completion and Segmentation

License

Notifications You must be signed in to change notification settings

ReubenDo/U-HVED

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

8 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Overview

This repository provides source code and pre-trained models for brain tumour segmentation in presence of missing modalities. The models have been trained and tested on BraTS dataset. The method is detailed in [1]. Our method allows to deal with missing modalities.

This implementation is based on NiftyNet and Tensorflow.

If you use any resources in this repository, please cite the following papers:

  • [1] Reuben Dorent, Samuel Joutard, Marc Modat, Sebastien Ourselin, Tom Vercauteren. "AHetero-Modal Variational Encoder-Decoder for Joint Modality Completion and Segmentation." In MICCAI 2019. https://arxiv.org/abs/1907.11150
  • [2] Eli Gibson*, Wenqi Li*, Carole Sudre, Lucas Fidon, Dzhoshkun I. Shakir, Guotai Wang, Zach Eaton-Rosen, Robert Gray, Tom Doel, Yipeng Hu, Tom Whyntie, Parashkev Nachev, Marc Modat, Dean C. Barratt, Sébastien Ourselin, M. Jorge Cardoso^, Tom Vercauteren^. "NiftyNet: a deep-learning platform for medical imaging." Computer Methods and Programs in Biomedicine, 158 (2018): 113-122. https://arxiv.org/pdf/1709.03485 Modality reconstruction and hetero-modal segmentation

Example of FLAIR and T1 completion and tumour segmentation given a subsetof modalities as input. Green: edema; Red: non-enhancing core; Blue: enhancing core.

Requirements

  • A CUDA compatable GPU with memory not less than 12GB is recommended for training. For testing only, a smaller GPU should be suitable.

  • For this work we will first create a python environement

virtualenv -p python3 ./uhved_python
source ./uhved_python/bin/activate
pip install -r requirements.txt
  • A small change in niftynet is required for our multi-modal application.
cp ./utils/user_parameters_custom.py ./uhved_python/lib/python3.6/site-packages/niftynet/utilities
export PYTHONPATH=$PATH:$(pwd)/extensions/:$PYTHONPATH

How to use

1, Prepare data

  • Download BraTS dataset, and uncompress the training and tesing zip files. For example, the training set will be in ./data/BRATS18Training. As preprocessing step, background is cropped. The rest of the preprocessing is in the configuration files.
python ./utils/cropping.py --input ./data/BRATS18Training --output  ./data/BRATS2018

2, Use pre-trained models to segment images

  • To segment multimodal scans using U-HeMIS, run:
python inference.py -n u_hemis -t1 ./data/BRATS2018/HGG3_T1.nii.gz -t2 ./data/BRATS2018/HGG3_T2.nii.gz  -t1c ./data/BRATS2018/HGG3_T1c.nii.gz -fl ./data/BRATS2018/HGG3_Flair.nii.gz
  • To segment multimodal scans using U-HVED, run:
python inference.py -n u_hved -t1 ./data/BRATS2018/HGG3_T1.nii.gz -t2 ./data/BRATS2018/HGG3_T2.nii.gz  -t1c ./data/BRATS2018/HGG3_T1c.nii.gz -fl ./data/BRATS2018/HGG3_Flair.nii.gz
  • The networks allows for missing modalities, consequently the number of modalities as input can vary. For example:
python inference.py -n u_hved -t2 ./data/BRATS2018/HGG3_T2.nii.gz  -t1c ./data/BRATS2018/HGG3_T1c.nii.gz

3, How to train

The following commands show how to train our models using BRATS18. Look at the config files in ./extensions/u_hemis and ./extensions/u_hved for changing the config settings. Documentation is available here: https://niftynet.readthedocs.io/en/dev/config_spec.html.

  • Train models using U_HeMIS. Run:
net_run train -a  u_hemis.application.U_HeMISApplication  -c ./extensions/u_hemis/config.ini
  • Train models using U_HVED. Run:
net_run train -a  u_hved.application.U_HVEDApplication  -c ./extensions/u_hved/config.ini

Copyright

  • Copyright (c) 2019, King's College London. All rights reserved.

About

Hetero-Modal Variational Encoder-Decoder for Joint Modality Completion and Segmentation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages