Skip to content

This repository includes the official implementation our paper "Can CNNs Be More Robust Than Transformers?"

License

Notifications You must be signed in to change notification settings

wongxin/RobustCNN

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Can CNNs Be More Robust Than Transformers?

This is a PyTorch/PyTorch-XLA implementation of the paper Can CNNs Be More Robust Than Transformers?. It can run on both GPU or TPU devices (see TRAIN/TEST instructions). Our implementation is heavily based on the bits_and_tpu branch of TIMM.

Introduction

The recent success of Vision Transformers is shaking the long dominance of Convolutional Neural Networks (CNNs) in image recognition for a decade. Specifically, in terms of robustness on out-of-distribution samples, recent research finds that Transformers are inherently more robust than CNNs, regardless of different training setups. Moreover, it is believed that such superiority of Transformers should largely be credited to their self-attention-like architectures per se. In this work, we question that belief by closely examining the design of Transformers. Our findings lead to three highly effective architecture designs for boosting robustness, yet simple enough to be implemented in several lines of code, namely a) patchifying input images, b) enlarging kernel size, and c) reducing activation layers and normalization layers. Bringing these components together, we are able to build pure CNN architectures without any attention-like operations that is as robust as, or even more robust than, Transformers. We hope this work can help the community better understand the design of robust neural architectures.

Installation

This repo is a modification on the bits_and_tpu branch of TIMM repo. Installation and preparation follow that repo.

Note that this repo is compatible with both GPU and TPU. If you want to run the code on Google Cloud TPU, here are some documents you may find helpful: Google Cloud User's Guide and TIMM bits README

Data preparation

Clean Dataset

Download and extract ImageNet train and val images from http://image-net.org/. The directory structure is the standard layout for the torchvision, and the training and validation data is expected to be in the train folder and val folder respectively:

/path/to/imagenet/
  train/
    class1/
      img1.jpeg
    class2/
      img2.jpeg
  val/
    class1/
      img3.jpeg
    class/2
      img4.jpeg

Out-of-distribution Dataset

Download and extract Stylized-ImageNet, ImageNet-C, ImageNet-R, ImageNet-Sketch val images:

/path/to/datasets/
    class1/
      img1.jpeg
    class/2
      img2.jpeg

Usage

Training Instructions

The training instructions are in TRAIN.md.

Testing Instructions

The testing instructions are in TEST.md.

Pretrained models

Here are ImageNet-1k trained weights for four model instantiations. All models are trained for 300 epochs with DeiT recipe. Please follow the testing script to evaluate them on out-of-distribution dataset.

Pretrained Model ImageNet↑ Stylized-ImageNet↑ ImageNet-C↓ ImageNet-R↑ ImageNet-Sketch↑
DeiT-S (Official Checkpoint) download link 79.8 16.2 42.8 41.9 29.1
Robust-ResNet-DW-Small download link 79.4 18.6 42.3 45.9 33.0
Robust-ResNet-Inverted-DW-Small download link 79.0 19.5 42.1 45.9 32.8
Robust-ResNet-Up-Inverted-DW-Small download link 79.2 20.2 40.9 48.7 35.2
Robust-ResNet-Down-Inverted-DW-Small download link 79.9 19.3 41.6 46.0 32.8

Acknowledgment

This repo is built on timm. And this work is supported by a gift from Open Philanthropy, TPU Research Cloud (TRC) program, and Google Cloud Research Credits program.

Citation

@article{wang2022robustcnn,
  title   = {Can CNNs Be More Robust Than Transformers?}, 
  author  = {Zeyu Wang and Yutong Bai and Yuyin Zhou and Cihang Xie}},
  journal = {arXiv preprint arXiv:2206.03452},
  year    = {2022},
}

About

This repository includes the official implementation our paper "Can CNNs Be More Robust Than Transformers?"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%