SOTAVerified

Relational Self-Attention: What's Missing in Attention for Video Understanding

2021-11-02NeurIPS 2021Code Available1· sign in to hype

Manjin Kim, Heeseung Kwon, Chunyu Wang, Suha Kwak, Minsu Cho

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Convolution has been arguably the most important feature transform for modern neural networks, leading to the advance of deep learning. Recent emergence of Transformer networks, which replace convolution layers with self-attention blocks, has revealed the limitation of stationary convolution kernels and opened the door to the era of dynamic feature transforms. The existing dynamic transforms, including self-attention, however, are all limited for video understanding where correspondence relations in space and time, i.e., motion information, are crucial for effective representation. In this work, we introduce a relational feature transform, dubbed the relational self-attention (RSA), that leverages rich structures of spatio-temporal relations in videos by dynamically generating relational kernels and aggregating relational contexts. Our experiments and ablation studies show that the RSA network substantially outperforms convolution and self-attention counterparts, achieving the state of the art on the standard motion-centric benchmarks for video action recognition, such as Something-Something-V1 & V2, Diving48, and FineGym.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
Diving-48RSANet-R50 (16 frames, ImageNet pretrained, a single clip)Accuracy84.2Unverified
Something-Something V1RSANet-R50 (8+16 frames, ImageNet pretrained, 2 clips)Top 1 Accuracy56.1Unverified
Something-Something V1RSANet-R50 (8+16 frames, ImageNet pretrained, a single clip)Top 1 Accuracy55.5Unverified
Something-Something V1RSANet-R50 (16 frames, ImageNet pretrained, a single clip)Top 1 Accuracy54Unverified
Something-Something V1RSANet-R50 (8 frames, ImageNet pretrained, a single clip)Top 1 Accuracy51.9Unverified
Something-Something V2RSANet-R50 (8+16 frames, ImageNet pretrained, 2 clipsTop-1 Accuracy67.7Unverified
Something-Something V2RSANet-R50 (8+16 frames, ImageNet pretrained, a single clip)Top-1 Accuracy67.3Unverified
Something-Something V2RSANet-R50 (16 frames, ImageNet pretrained, a single clip)Top-1 Accuracy66Unverified
Something-Something V2RSANet-R50 (8 frames, ImageNet pretrained, a single clip)Top-1 Accuracy64.8Unverified
Something-Something V2RSANet-R50 (8+16 frames, ImageNet pretrained, 2 clips)Top-5 Accuracy91.1Unverified

Reproductions