Official repository for "C-DARL: Contrastive diffusion adversarial representation learning for label-free blood vessel segmentation" published in Medical Image Analysis.
[arXiv][Medical Image Analysis]
- OS : Ubuntu
- Python >= 3.9
- PyTorch >= 1.12.1
In our experiments, we used the publicly available XCAD dataset. Please refer to our main paper.
To train our model, run this command:
python3 main.py -p train -c config/train.json
To test the trained our model, run:
python3 main.py -p test -c config/test.json
You can download our pre-trained model of the XCAD dataset here. Then, you can test the model by saving the pre-trained weights in the directory ./experiments/pretrained_model. To briefly test our method given the pre-trained model, we provided the toy example in the directory './data/'.
@article{kim2024cdarl,
title = {C-DARL: Contrastive diffusion adversarial representation learning for label-free blood vessel segmentation},
journal = {Medical Image Analysis},
volume = {91},
pages = {103022},
year = {2024},
issn = {1361-8415},
doi = {https://doi.org/10.1016/j.media.2023.103022},
author = {Boah Kim and Yujin Oh and Bradford J. Wood and Ronald M. Summers and Jong Chul Ye}
}