Learn to cycle: Time-consistent feature discovery for action recognition
Alexandros Stergiou, Ronald Poppe
Code Available — Be the first to reproduce this paper.
ReproduceCode
Abstract
Generalizing over temporal variations is a prerequisite for effective action recognition in videos. Despite significant advances in deep neural networks, it remains a challenge to focus on short-term discriminative motions in relation to the overall performance of an action. We address this challenge by allowing some flexibility in discovering relevant spatio-temporal features. We introduce Squeeze and Recursion Temporal Gates (SRTG), an approach that favors inputs with similar activations with potential temporal variations. We implement this idea with a novel CNN block that uses an LSTM to encapsulate feature dynamics, in conjunction with a temporal gate that is responsible for evaluating the consistency of the discovered dynamics and the modeled features. We show consistent improvement when using SRTG blocks, with only a minimal increase in the number of GFLOPs. On Kinetics-700, we perform on par with current state-of-the-art models, and outperform these on HACS, Moments in Time, UCF-101 and HMDB-51.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| HACS | SRTG r(2+1)d-101 | Top 1 Accuracy | 84.33 | — | Unverified |
| HACS | SRTG r(2+1)d-50 | Top 1 Accuracy | 83.77 | — | Unverified |
| HACS | SRTG r3d-101 | Top 1 Accuracy | 81.66 | — | Unverified |
| HACS | SRTG r(2+1)d-34 | Top 1 Accuracy | 80.39 | — | Unverified |
| HACS | SRTG r3d-50 | Top 1 Accuracy | 80.36 | — | Unverified |
| HACS | SRTG r3d-34 | Top 1 Accuracy | 78.6 | — | Unverified |