From a676ec154fe7ed3f704e35149e0e98864818c15d Mon Sep 17 00:00:00 2001 From: YifanXu <41295724+YifanXu74@users.noreply.github.com> Date: Tue, 14 Sep 2021 16:33:15 +0800 Subject: [PATCH] Update README.md --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index e61a458..544aa18 100644 --- a/README.md +++ b/README.md @@ -3,7 +3,7 @@ This repository contains the PyTorch code for Evo-ViT. This work proposes a slow-fast token evolution approach to accelerate vanilla vision transformers of both flat and deep-narrow structures without additional pre-training and fine-tuning procedures. For details please see [Evo-ViT: Slow-Fast Token Evolution for Dynamic Vision Transformer](https://arxiv.org/abs/2108.01390) by Yifan Xu*, Zhijie Zhang*, Mengdan Zhang, Kekai Sheng, Ke Li, Weiming Dong, Liqing Zhang, Changsheng Xu, and Xing Sun. -![intro](/imgs/method.png) +![intro](imgs/method.png) Our code is based on [pytorch-image-models](https://github.com/rwightman/pytorch-image-models), [DeiT](https://github.com/facebookresearch/deit), and [LeViT](https://github.com/facebookresearch/LeViT). @@ -124,7 +124,7 @@ Add parameter '--layer-wise-prune' if the visualized model is not trained with l The visualization results of Evo-ViT-S are as follows: -![result](/imgs/results.png) +![result](imgs/results.png) # Citation