CVPR 2020.
The code and experimental results have be released now 😄.
- 2020/5/10: Some meaningful changes.
- Some methods have been tried to improve the performance of the model on specific operating environments and devices. The specific experimental results can be found in the section "More Experiments" of
./code/readme.md
. - Use black as the repository formatting tool, and set the configuration file
./code/pyproject.toml
.
- Some methods have been tried to improve the performance of the model on specific operating environments and devices. The specific experimental results can be found in the section "More Experiments" of
- 2020/5/6: Add some new attempts and improvements.
- Modified the method of importing model from the directly setting in config.py, and change it to the automatic selection and instantiation of the specific model class according to the model alias (
exp_name
in config.py). - Added a lighter setting for AIM and MInet.
- Added an attempt to
checkpoint
features of PyTorch. - Added learning rate decay strategy with learning rate warm-up. However, the parameter setting is not flexible at present, and needs to be improved.
- A new optimization strategy has been added to imitate the settings from F3Net.
- Upload the MIT LICENSE file.
- Modified the method of importing model from the directly setting in config.py, and change it to the automatic selection and instantiation of the specific model class according to the model alias (
- 2020/4/16: Modify some misleading descriptions in the
readme.md
file. - 2020/4/7: Simplify the structure of the repository.
- 2020/3/29: Update the method of evaluating results. (See the
readme.md
for more details.) - 2020/3/28: Update our code, results, pretrained parameters and some documents.
code
: Complete training and testing code about our method. Thereadme.md
file describes how to use the code.docs
: Github page about out paper. Here are some paper details.
- Paper:
- Baidu Pan: https://pan.baidu.com/s/1zN7m4aeDhRvTOeF2naATRg (baidu: 48au)
- Google Drive: https://drive.google.com/file/d/1gUYu0hO_8Xc5jgpzetuOVFDrqeSOiKZN/view?usp=sharing
- Results & Pretrained Parameters:
Deep-learning based salient object detection methods achieve great progress. However, the variable scale and unknown category of salient objects are great challenges all the time. These are closely related to the utilization of multi-level and multi-scale features. In this paper, we propose the aggregate interaction modules to integrate the features from adjacent levels, in which less noise is introduced because of only using small up-/down-sampling rates. To obtain more efficient multi-scale features from the integrated features, the self-interaction modules are embedded in each decoder unit. Besides, the class imbalance issue caused by the scale variation weakens the effect of the binary cross entropy loss and results in the spatial inconsistency of the predictions. Therefore, we exploit the consistency-enhanced loss to highlight the fore-/back-ground difference and preserve the intra-class consistency. Experimental results on five benchmark datasets demonstrate that the proposed method without any post-processing performs favorably against 23 state-of-the-art approaches. The source code will be publicly available at https://github.com/lartpang/MINet.
@inproceedings{MINet-CVPR2020,
author = {Youwei Pang and Xiaoqi Zhao and Lihe Zhang and Huchuan Lu},
title = {Multi-scale Interactive Network for Salient Object Detection},
booktitle = CVPR,
year = {2020}
}