Skip to content

nuggy875/py-MDNet

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

26 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

py-MDNet

by Hyeonseob Nam and Bohyung Han at POSTECH

Update (April, 2019)

  • Migration to python 3.6 & pyTorch 1.0
  • Improving tracking efficiency (~5fps)
  • Code refactoring

Introduction

PyTorch implementation of MDNet, which runs at ~5fps with a single CPU core and a single GPU (GTX 1080 Ti).

If you're using this code for your research, please cite:

@InProceedings{nam2016mdnet,
author = {Nam, Hyeonseob and Han, Bohyung},
title = {Learning Multi-Domain Convolutional Neural Networks for Visual Tracking},
booktitle = {The IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
month = {June},
year = {2016}
}

Prerequisites

  • python 3.6+
  • opencv 3.0+
  • PyTorch 1.0+ and its dependencies
  • for GPU support: a GPU with ~3G memory

Usage

Tracking

 python tracking/run_tracker.py -s DragonBaby [-d (display fig)] [-f (save fig)]
  • You can provide a sequence configuration in two ways (see tracking/gen_config.py):
    • python tracking/run_tracker.py -s [seq name]
    • python tracking/run_tracker.py -j [json path]

Pretraining

  • Download VGG-M (matconvnet model) and save as "models/imagenet-vgg-m.mat"
  • Download VOT datasets into "datasets/VOT/vot201x"
 python pretrain/prepro_vot.py
 python pretrain/train_mdnet.py

Third-party re-implementations

  • Tensorflow, by @AlexQie: code
  • Tensorflow, by @zhyj3038: code
  • MatconvNet (optimized), by @ZjjConan: code

About

MDNet PyTorch implementation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%