The code provides functionality to train Fast R-CNN and MultiPath Networks in Torch-7.
Corresponding paper: A MultiPath Network for Object Detection http://arxiv.org/abs/1604.02135
If you use MultiPathNet in your research, please cite the relevant papers:
@INPROCEEDINGS{Zagoruyko2016Multipath,
author = {S. Zagoruyko and A. Lerer and T.-Y. Lin and P. O. Pinheiro and S. Gross and S. Chintala and P. Doll{\'{a}}r},
title = {A MultiPath Network for Object Detection},
booktitle = {BMVC}
year = {2016}
}
- Linux
- NVIDIA GPU with compute capability 3.5+
The code depends on Torch-7, fb.python and several other easy-to-install torch packages.
To install Torch, follow http://torch.ch/docs/getting-started.html
Then install additional packages:
luarocks install inn
luarocks install torchnet
luarocks install fbpython
luarocks install class
Evaluation relies on COCO API calls via python interface, because lua interface doesn't support it. Lua API is used to load annotation files in *json to COCO API data structures. This doesn't work for proposal files as they're too big, so we provide converted proposals for sharpmask and selective search in torch format.
First, clone https://github.com/pdollar/coco:
git clone https://github.com/pdollar/coco
Then install LuaAPI:
cd coco
luarocks make LuaAPI/rocks/coco-scm-1.rockspec
And PythonAPI:
cd coco/PythonAPI
make
You will have to add the path to PythonAPI to PYTHONPATH
. Note that this won't work with anaconda as it ships
with it's own libraries which conflict with torch.
The root folder should have a folder data
with the following subfolders:
models/
annotations/
proposals/
models
folder should contain AlexNet and VGG pretrained imagenet files downloaded from here . ResNets can resident in other places specified by resnet_path
env variable.
annotations
should contain *json files downloaded from https://mscoco.org/external. There are *json annotation files for
PASCAL VOC, MSCOCO, ImageNet and other datasets.
proposals
should contain *t7 files downloaded from here
We provide selective search VOC 2007 and VOC 2012 proposals converted from https://github.com/rbgirshick/fast-rcnn and SharpMask proposals for COCO 2015 converted from https://github.com/facebookresearch/deepmask, which can be used to compute proposals for new images as well.
Here is an example structure:
data
|-- annotations
| |-- instances_train2014.json
| |-- instances_val2014.json
| |-- pascal_test2007.json
| |-- pascal_train2007.json
| |-- pascal_train2012.json
| |-- pascal_val2007.json
| `-- pascal_val2012.json
|-- models
| |-- caffenet_fast_rcnn_iter_40000.t7
| |-- imagenet_pretrained_alexnet.t7
| |-- imagenet_pretrained_vgg.t7
| `-- vgg16_fast_rcnn_iter_40000.t7
`-- proposals
|-- VOC2007
| `-- selective_search
| |-- test.t7
| |-- train.t7
| |-- trainval.t7
| `-- val.t7
`-- coco
`-- sharpmask
|-- train.t7
`-- val.t7
Download selective_search proposals for VOC2007:
wget https://s3.amazonaws.com/multipathnet/proposals/VOC2007/selective_search/train.t7
wget https://s3.amazonaws.com/multipathnet/proposals/VOC2007/selective_search/val.t7
wget https://s3.amazonaws.com/multipathnet/proposals/VOC2007/selective_search/trainval.t7
wget https://s3.amazonaws.com/multipathnet/proposals/VOC2007/selective_search/test.t7
Download sharpmask proposals for COCO:
wget https://s3.amazonaws.com/multipathnet/proposals/coco/sharpmask/train.t7
wget https://s3.amazonaws.com/multipathnet/proposals/coco/sharpmask/val.t7
As for the images themselves, provide paths to VOCDevkit and COCO in config.lua
We provide an example of how to extract DeepMask or SharpMask proposals from an image and run recognition MultiPathNet to classify them, then do non-maximum suppression and draw the found objects.
- Clone DeepMask project into the root directory:
git clone https://github.com/facebookresearch/deepmask
- Download DeepMask or SharpMask network:
cd data/models
# download SharpMask based on ResNet-50
wget https://s3.amazonaws.com/deepmask/models/sharpmask/model.t7 -O sharpmask.t7
- Download recognition network:
cd data/models
# download ResNet-18-based model trained on COCO with integral loss
wget https://s3.amazonaws.com/multipathnet/models/resnet18_integral_coco.t7
-
Make sure you have COCO validation .json files in
data/annotations/instances_val2014.json
-
Pick some image and run the script:
th demo.lua -img ./deepmask/data/testImage.jpg
And you should see this image:
See file demo.lua for details.
The repository supports training Fast-RCNN and MultiPath networks with data and model multi-GPU paralellism. Supported base models are the following:
- AlexNet trained in caffe by Ross Girshick, imagenet_pretrained_alexnet.t7
- VGG trained in caffe by Ross Girshick, imagenet_pretrained_vgg.t7
- ResNets trained in torch with fb.resnet.torch by Sam Gross
- inception-v3 trained in tensorflow by Google
- Network-In-Network trained in torch with imagenet-multiGPU.torch by Sergey Zagoruyko
To train Fast-RCNN on VOC2007 trainval with VGG base model and selective search proposals do:
test_nsamples=1000 model=vgg ./scripts/train_fastrcnn_voc2007.sh
The resulting mAP is slightly (~2 mAP) higher than original Fast-RCNN number. We should mention that the code is not exactly the same as we improved ROIPooling by fixing a few bugs, see szagoruyko/imagine-nn#17
To train MultiPathNet with VGG-16 base model on 4 GPUs run:
train_nGPU=4 test_nGPU=1 ./scripts/train_multipathnet_coco.sh
Here is a graph visualization of the network (click to enlarge):
To train ResNet-18 on COCO do:
train_nGPU=4 test_nGPU=1 model=resnet resnet_path=./data/models/resnet/resnet-18.t7 ./scripts/train_coco.sh
We provide original models from Fast-RCNN paper converted to torch format here:
To evaluate these models run:
model=data/models/caffenet_fast_rcnn_iter_40000.t7 ./scripts/eval_fastrcnn_voc2007.sh
model=data/models/vgg_fast_rcnn_iter_40000.t7 ./scripts/eval_fastrcnn_voc2007.sh
Evaluate fast ResNet-18-based network trained with integral loss on COCO val5k split (resnet18_integral_coco.t7 89MB):
test_nGPU=4 test_nsamples=5000 ./scripts/eval_coco.sh
It achieves 24.4 mAP using 400 SharpMask proposals per image:
Average Precision (AP) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.244
Average Precision (AP) @[ IoU=0.50 | area= all | maxDets=100 ] = 0.402
Average Precision (AP) @[ IoU=0.75 | area= all | maxDets=100 ] = 0.268
Average Precision (AP) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.078
Average Precision (AP) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.266
Average Precision (AP) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.394
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 1 ] = 0.249
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets= 10 ] = 0.368
Average Recall (AR) @[ IoU=0.50:0.95 | area= all | maxDets=100 ] = 0.377
Average Recall (AR) @[ IoU=0.50:0.95 | area= small | maxDets=100 ] = 0.135
Average Recall (AR) @[ IoU=0.50:0.95 | area=medium | maxDets=100 ] = 0.444
Average Recall (AR) @[ IoU=0.50:0.95 | area= large | maxDets=100 ] = 0.561