Skip to content

Official repository for "C-DARL: Contrastive diffusion adversarial representation learning for label-free blood vessel segmentation"

Notifications You must be signed in to change notification settings

boahK/MEDIA_CDARL

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

MEDIA_CDARL

Official repository for "C-DARL: Contrastive diffusion adversarial representation learning for label-free blood vessel segmentation" published in Medical Image Analysis.

[arXiv][Medical Image Analysis]

Image of The Proposed method

Requirements

  • OS : Ubuntu
  • Python >= 3.9
  • PyTorch >= 1.12.1

Data

In our experiments, we used the publicly available XCAD dataset. Please refer to our main paper.

Training

To train our model, run this command:

python3 main.py -p train -c config/train.json

Test

To test the trained our model, run:

python3 main.py -p test -c config/test.json

Pre-trained Models

You can download our pre-trained model of the XCAD dataset here. Then, you can test the model by saving the pre-trained weights in the directory ./experiments/pretrained_model. To briefly test our method given the pre-trained model, we provided the toy example in the directory './data/'.

Citations

@article{kim2024cdarl,
title = {C-DARL: Contrastive diffusion adversarial representation learning for label-free blood vessel segmentation},
journal = {Medical Image Analysis},
volume = {91},
pages = {103022},
year = {2024},
issn = {1361-8415},
doi = {https://doi.org/10.1016/j.media.2023.103022},
author = {Boah Kim and Yujin Oh and Bradford J. Wood and Ronald M. Summers and Jong Chul Ye}
}

About

Official repository for "C-DARL: Contrastive diffusion adversarial representation learning for label-free blood vessel segmentation"

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages