SOTAVerified

Adversarial Background-Aware Loss for Weakly-supervised Temporal Activity Localization

2020-07-13ECCV 2020Code Available1· sign in to hype

Kyle Min, Jason J. Corso

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Temporally localizing activities within untrimmed videos has been extensively studied in recent years. Despite recent advances, existing methods for weakly-supervised temporal activity localization struggle to recognize when an activity is not occurring. To address this issue, we propose a novel method named A2CL-PT. Two triplets of the feature space are considered in our approach: one triplet is used to learn discriminative features for each activity class, and the other one is used to distinguish the features where no activity occurs (i.e. background features) from activity-related features for each video. To further improve the performance, we build our network using two parallel branches which operate in an adversarial way: the first branch localizes the most salient activities of a video and the second one finds other supplementary activities from non-localized parts of the video. Extensive experiments performed on THUMOS14 and ActivityNet datasets demonstrate that our proposed method is effective. Specifically, the average mAP of IoU thresholds from 0.1 to 0.9 on the THUMOS14 dataset is significantly improved from 27.9% to 30.0%.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
ActivityNet-1.3A2CL-PTmAP@0.5:0.9522.5Unverified
THUMOS14A2CL-PTavg-mAP (0.3-0.7)30.6Unverified
THUMOS14A2CL-PTmAP@0.530.1Unverified
THUMOS 2014A2CL-PTmAP@0.1:0.737.8Unverified

Reproductions