Baseline
- CONFIG (from left to right):
- configs/cao_cifar/baseline/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
29.64 |
25.19 |
61.68 |
56.15 |
CE_CE
- Introduction:
- The most commonly used re-weighting method, you can see Eq. (2) in our paper for more details.
- CONFIG:
- configs/cao_cifar/re_weighting/csce/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
31.70 |
23.20 |
67.73 |
63.49 |
Square CS_CE
- Introduction:
- This is a smooth version of CE_CE (smooth CS_CE), which add a hyper-parameter $ \gamma$ to vanilla CS_CE. In smooth CS_CE, the loss weight of class i is defined as: $(\frac{N_{min}}{N_i})^\gamma$, where $\gamma \in [0, 1]$, $N_i$ is the number of images in class i. We set $\gamma = 0.5$ to get a square-root version of CS_CE (Square CE_CE).
- CONFIG:
- configs/cao_cifar/re_weighting/csce/{cifar10_im100_square.yaml, cifar10_im50_square.yaml, cifar100_im100_square.yaml, cifar100_im50_square.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
31.70 |
22.22 |
61.64 |
57.23 |
Focal loss
- Introduction:
- Focal loss makes the model focus training on difficult samples, and you can see Eq. (4) in our paper for more details.
- The Focal loss paper link: Lin et al., ICCV 2017.
- CONFIG:
- configs/cao_cifar/re_weighting/focal/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
28.44 |
22.09 |
62.78 |
58.21 |
ClassBalanceFocal
- Introduction:
- The modified version of Focal loss, which is based on the theory of effective numbers, and you can see Eq. (5) in our paper for more details.
- The ClassBalanceFocal paper link: Cui et al., CVPR 2019.
- CONFIG:
- configs/cao_cifar/re_weighting/cbfocal/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
24.80 |
21.01 |
61.44 |
57.63 |
ClassBalanceCE
- Introduction:
- The modified version of cross-entropy loss, which is based on the theory of effective numbers, and you can see Eq. (6) in our paper for more details.
- The ClassBalanceCE paper link: Cui et al., CVPR 2019.
- CONFIG:
- configs/cao_cifar/re_weighting/cbce/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
29.52 |
22.52 |
61.03 |
56.22 |
CrossEntropyLabelSmooth
- Introduction:
- The commonly used regularization trick, label smoothing, based on cross-entropy loss.
- The CrossEntropyLabelSmooth paper link: Szegedy et al., CVPR 2016.
- CONFIG:
- configs/cao_cifar/re_weighting/cels/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
27.19 |
23.43 |
61.56 |
57.66 |
CrossEntropyLabelAwareSmooth
- Introduction:
- The modified regularization trick, label-aware smoothing, which is based on label smoothing. It assigns different smoothing factors for each class according to the number of training images it contains.
- The CrossEntropyLabelAwareSmooth paper link: Zhong et al., CVPR 2021.
- CONFIG:
- configs/cao_cifar/re_weighting/celas/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
27.49 |
22.04 |
62.32 |
56.22 |
LDAM loss
- Introduction:
- LDAM loss is one of metric learning methods, which aims to assign different margins to different class.
- The LDAM loss paper link: Cao et al., NeurIPS 2019.
- CONFIG:
- configs/cao_cifar/re_weighting/ldam/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
26.34 |
20.99 |
61.12 |
56.41 |
SEQL
- Introduction:
- The softmax equalization loss (SEQL) aims to reduce the gradients of tail classes' negative samples. The author argues that the imbalance of gradients in tail classes' positive and negtive samples causes bad influences.
- The SEQL paper link: Tan et al., CVPR 2020.
- CONFIG:
- configs/cao_cifar/re_weighting/seql/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
-- |
-- |
59.51 |
55.19 |
CDT
- Introduction:
- The authors find that a model significantly over-fits the tail classes, and they argue that feature deviation between the training and test samples causes this problem. So they propose class-dependent temperatures (CDT).
- The CDT paper link: Ye et al., arXiv 2020.
- CONFIG:
- configs/cao_cifar/re_weighting/cdt/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
22.90 |
18.19 |
60.41 |
55.17 |
BalancedSoftmaxCE
- Introduction:
- A simple and effective re-weighting method, and you can see Eq. (4) in the author paper.
- The BalancedSoftmaxCE paper link: Ren et al., NeurIPS 2020.
- CONFIG:
- configs/cao_cifar/re_weighting/bsce/{cifar10_im100.yaml, cifar10_im50.yaml, cifar100_im100.yaml, cifar100_im50.yaml}
- Running commands:
- bash data_parallel_train.sh CONFIG GPU
|
22.46 |
18.89 |
57.01 |
53.45 |