Shuffle Transformer: Rethinking Spatial Shuffle for Vision Transformer
Zilong Huang, Youcheng Ben, Guozhong Luo, Pei Cheng, Gang Yu, Bin Fu
Code Available — Be the first to reproduce this paper.
ReproduceCode
Abstract
Very recently, Window-based Transformers, which computed self-attention within non-overlapping local windows, demonstrated promising results on image classification, semantic segmentation, and object detection. However, less study has been devoted to the cross-window connection which is the key element to improve the representation ability. In this work, we revisit the spatial shuffle as an efficient way to build connections among windows. As a result, we propose a new vision transformer, named Shuffle Transformer, which is highly efficient and easy to implement by modifying two lines of code. Furthermore, the depth-wise convolution is introduced to complement the spatial shuffle for enhancing neighbor-window connections. The proposed architectures achieve excellent performance on a wide range of visual tasks including image-level classification, object detection, and semantic segmentation. Code will be released for reproduction.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| ADE20K | UperNet Shuffle-B | Validation mIoU | 50.5 | — | Unverified |
| ADE20K | UperNet Shuffle-T | Validation mIoU | 47.6 | — | Unverified |
| ADE20K val | UperNet Shuffle-B | mIoU | 50.5 | — | Unverified |
| ADE20K val | UperNet Shuffle-S | mIoU | 49.6 | — | Unverified |
| ADE20K val | UperNet Shuffle-T | mIoU | 47.6 | — | Unverified |