| # Multi-organ segmentation in abdominal CT |
|
|
| ### **Authors** |
|
|
| Chen Shen<sup>1</sup>, Holger R. Roth<sup>2</sup>, Kazunari Misawa<sup>3</sup>, Kensaku Mori<sup>1</sup> |
|
|
| 1. Nagoya University, Japan |
|
|
| 2. NVIDIA Corporation, USA |
|
|
| 3. Aichi Cancer Center, Japan |
|
|
| ### **Tags** |
|
|
| Segmentation, Multi-organ, Abdominal |
|
|
| ## **Model Description** |
|
|
| This model uses the DiNTS model architecture searched on [Medical Segmentation Decathlon](http://medicaldecathlon.com/) Pancreas [1] and re-trained for multi-organ segmentation from abdominal CT images [2,3]. |
|
|
| ## **Data** |
|
|
| This model was trained on an abdominal CT dataset in portal venous phase collected from Aichi Cancer Center in Japan. Since this is a private dataset, similar models can be trained using other public multi-organ datasets like [BTCV](https://www.synapse.org/#!Synapse:syn3193805/wiki/89480). |
|
|
| For this bundle, we split the 420 cases into training, validation and testing with 300, 60 and 60 cases, respectively. |
|
|
| ## **Output** |
| 8 channels |
|
|
| - 0: Background |
| - 1: Artery |
| - 2: Portal vein |
| - 3: Liver |
| - 4: Spleen |
| - 5: Stomach |
| - 6: Gallbladder |
| - 7: Pancreas |
|
|
| Here is an example of output. |
|
|
|  |
|
|
| ## **Scores** |
|
|
| This model achieves the following Dice score on the validation data (our own split from the whole dataset): |
|
|
| Mean Dice = 88.6% |
|
|
| ## MONAI Bundle Commands |
| In addition to the Pythonic APIs, a few command line interfaces (CLI) are provided to interact with the bundle. The CLI supports flexible use cases, such as overriding configs at runtime and predefining arguments in a file. |
|
|
| For more details usage instructions, visit the [MONAI Bundle Configuration Page](https://docs.monai.io/en/latest/config_syntax.html). |
|
|
| #### Execute model searching: |
|
|
| ``` |
| python -m scripts.search run --config_file configs/search.yaml |
| ``` |
|
|
| #### Execute multi-GPU model searching (recommended): |
|
|
| ``` |
| torchrun --nnodes=1 --nproc_per_node=8 -m scripts.search run --config_file configs/search.yaml |
| ``` |
|
|
| #### Execute training: |
|
|
| ``` |
| python -m monai.bundle run --config_file configs/train.yaml |
| ``` |
|
|
| Please note that if the default dataset path is not modified with the actual path in the bundle config files, you can also override it by using `--dataset_dir`: |
|
|
| ``` |
| python -m monai.bundle run --config_file configs/train.yaml |
| ``` |
|
|
| #### Override the `train` config to execute multi-GPU training: |
|
|
| ``` |
| torchrun --nnodes=1 --nproc_per_node=8 \ |
| -m scripts.search run \ |
| --config_file configs/search.yaml |
| ``` |
|
|
| #### Override the `train` config to execute evaluation with the trained model: |
|
|
| ``` |
| python -m monai.bundle run \ |
| --config_file "['configs/train.yaml','configs/evaluate.yaml']" |
| ``` |
|
|
| #### Execute inference: |
| ``` |
| python -m monai.bundle run --config_file configs/inference.yaml |
| ``` |
|
|
|
|
| #### Export checkpoint for TorchScript: |
|
|
| ``` |
| python -m monai.bundle ckpt_export network_def --filepath models/model.ts --ckpt_file models/model.pt --meta_file configs/metadata.json --config_file configs/inference.yaml |
| ``` |
|
|
| #### Execute inference with the TensorRT model: |
|
|
| ``` |
| python -m monai.bundle run --config_file "['configs/inference.yaml', 'configs/inference_trt.yaml']" |
| ``` |
|
|
|
|
| ## **References** |
|
|
| [1] He, Y., Yang, D., Roth, H., Zhao, C. and Xu, D., 2021. Dints: Differentiable neural network topology search for 3d medical image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (pp. 5841-5850). |
|
|
| |
| [2] Roth, Holger R., et al. "A multi-scale pyramid of 3D fully convolutional networks for abdominal multi-organ segmentation." International conference on medical image computing and computer-assisted intervention. Springer, Cham, 2018. |
| |
|
|
| [3] Shen, Chen, et al. "Effective hyperparameter optimization with proxy data for multi-organ segmentation." Medical Imaging 2022: Image Processing. Vol. 12032. SPIE, 2022. |
|
|
|
|
|
|
| ## **License** |
| The Licensee is not allowed to distribute or make the model to any third party, either for free or for a fee. Reverse engineering of the model is not allowed. This includes, but is not limited to, providing the model as part of a commercial offering, sharing the model on a public or private network, or making the model available for download on the Internet. |
|
|