This is the implementation of our CVPR 2018 work 'Domain Adaptive Faster R-CNN for Object Detection in the Wild'. The aim is to improve the cross-domain robustness of object detection, in the screnario where training and test data are drawn from different distributions. The original paper can be found here.
If you find it helpful for your research, please consider citing:
@inproceedings{chen2018domain,
title={Domain Adaptive Faster R-CNN for Object Detection in the Wild},
author={Chen, Yuhua and Li, Wen and Sakaridis, Christos and Dai, Dengxin and Van Gool, Luc},
booktitle = {Computer Vision and Pattern Recognition (CVPR)},
year={2018}
}
If you encounter any problems with the code, please contact me at yuhua[dot]chen[at]vision[dot]ee[dot]ethz[dot]ch
The implementation is built on the python implementation of Faster RCNN rbgirshick/py-faster-rcnn
-
Build Caffe and pycaffe (see: Caffe installation instructions)
-
Build the Cython modules
cd $FRCN_ROOT/lib make
-
Follow the instrutions of rbgirshick/py-faster-rcnn to download related data.
-
Prepare the dataset, source domain data should start with the filename 'source_', and target domain data with 'target_'.
-
To train the Domain Adaptive Faster R-CNN:
cd $FRCN_ROOT ./tools/train_net.py --gpu {GPU_ID} --solver models/da_faster_rcnn/solver.prototxt --weights data/imagenet_models/VGG16.v2.caffemodel --imdb voc_2007_trainval --iters {NUM_ITER} --cfg {CONFIGURATION_FILE}
An example of adapting from Cityscapes dataset to Foggy Cityscapes dataset is provided:
-
Download the datasets from here. Specifically, we will use gtFine_trainvaltest.zip, leftImg8bit_trainvaltest.zip and leftImg8bit_trainvaltest_foggy.zip.
-
Prepare the data using the scripts in 'prepare_data/prepare_data.m'.
-
Train the Domain Adaptive Faster R-CNN:
cd $FRCN_ROOT ./tools/train_net.py --gpu {GPU_ID} --solver models/da_faster_rcnn/solver.prototxt --weights data/imagenet_models/VGG16.v2.caffemodel --imdb voc_2007_trainval --iters 70000 --cfg models/da_faster_rcnn/faster_rcnn_end2end.yml
-
Test the trained model:
cd $FRCN_ROOT ./tools/test_net.py --gpu {GPU_ID} --def models/da_faster_rcnn/test.prototxt --net output/faster_rcnn_end2end/voc_2007_trainval/vgg16_da_faster_rcnn_iter_70000.caffemodel --imdb voc_2007_test --cfg models/da_faster_rcnn/faster_rcnn_end2end.yml
Detectron-DA-Faster-RCNN in Caffe2(Detectron)
Domain-Adaptive-Faster-RCNN-PyTorch in pytorch