Generalizable Cross-modality Medical Image Segmentation via Style Augmentation and Dual Normalization

July 25, 2024 · View on GitHub

by Ziqi Zhou, Lei Qi, Xin Yang, Dong Ni, Yinghuan Shi.

Introduction

This repository is for our CVPR 2022 paper 'Generalizable Cross-modality Medical Image Segmentation via Style Augmentation and Dual Normalization'.

Data Preparation

Dataset

BraTS 2018 | MMWHS | Abdominal-MRI | Abdominal-CT

File Organization

T2 as source domain

├── [Your BraTS2018 Path]
    ├── nii_data
        ├── test
            ├── Brats18_2013_4_1
                ├── Brats18_2013_4_1_flair.nii
                ├── Brats18_2013_4_1_t1.nii
                ├── Brats18_2013_4_1_t1ce.ni
                ├── Brats18_2013_4_1_t2.nii
                └── Brats18_2013_4_1_seg.nii
            ...
        └── train
            ...
    ├── npz_data
        ├── train
            ├── t2_ss
                ├── sample0.npz, sample1.npz, xxx
            └── t2_sd
        └── test
            ├── t1
                ├── test_sample0.npz, test_sample1.npz, xxx
            ├── flair
            ├── t2
            └── t1ce

For Brats dataset, we combined the HGG and LGG (totally 285 cases) then random selected 80% cases as training set (228 cases) and rest 20% cases as testing set (57 cases). Due to the access limit to my former university's GPU server, I can't provide the original dataset spliting. I provide a new random dataset spliting (in "Brats_trian.list" and "Brats_test.list") following our original setting.

Training and Testing

Train on source domain T2.

python -W ignore train_dn_unet.py \
  --train_domain_list_1 t2_ss --train_domain_list_2 t2_sd --n_classes 2 \
  --batch_size 128 --n_epochs 50 --save_step 10 --lr 0.004 --gpu_ids 0,1 \
  --result_dir ./results/unet_dn_t2 --data_dir [Your BraTS2018 Npz Training Data Folder]

Test on target domains (T1, T1ce and Flair).

python -W ignore test_dn_unet.py \
  --data_dir [Your BraTS2018 Npz Test Data Folfer] --n_classes 2 \
  --test_domain_list flair t1 t1ce --model_dir ./results/unet_dn_t2/model \
  --batch_size 64 --gpu_ids 0 \

Acknowledgement

The U-Net model is borrowed from Fed-DG. The Style Augmentation (SA) module is based on the nonlinear transformation in Models Genesis. The Dual-Normalizaiton is borrow from DSBN. We thank all of them for their great contributions.

Citation

If you find this project useful for your research, please consider citing:

@inproceedings{zhou2022dn,
  title={Generalizable Cross-modality Medical Image Segmentation via Style Augmentation and Dual Normalization},
  author={Zhou, Ziqi and Qi, Lei and Yang, Xin and Ni, Dong and Shi, Yinghuan},
  booktitle={CVPR},
  year={2022}
}