Skip to content

KingSpencer/DualNet

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

18 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

DualNet: Continual Learning, Fast and Slow

This project contains the implementation of the following NeurIPS 2021 paper:

Title: DualNet: Continual Learning, Fast and Slow (NeurIPS 2021), [openreview], [pdf].

Authors: Quang Pham, Chenghao Liu, and Steven Hoi

DualNet proposes a novel continual learning framework for (online) continual learning where a slow network gradually accumulates knowledge to improves its general representation over time and a fast network utilizes the slow representation to quickly learn new information.

DualNet

Requirements

  • python 3.7.3
  • pytorch >= 1.8
  • torchvision >= 0.9
  • Kornia >= 0.5.0

Benchmarks

1. Prepare data

The data/ folders contains the train and test splits for the miniImageNet and CORE50 benchmarks. Download the raw data and modify the path in the csv files to point to the raw data folder.

2. Run experiments

To replicate our results on the task aware and task free settings, run

chmod +x scripts/*
./scripts/task_aware.sh
./scripts/task_free.sh

The results will be put in the resuts/ folders.

Semi-supervised learning setting

For the semi-supervised continual learning experiments, run

./scripts/semi.sh --rho 0.1
./scripts/semi.sh --rho 0.25

The parameter rho indicates the percentage of data that are labeled (e.g. 10% or 25%).

Citation

If you found DualNet useful for your research, please consider citing.

@article{pham2021dualnet,
  title={DualNet: Continual Learning, Fast and Slow},
  author={Pham, Quang and Liu, Chenghao and Hoi, Steven},
  journal={Advances in Neural Information Processing Systems},
  volume={34},
  year={2021}
}

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 98.2%
  • Shell 1.8%