Omni-sourced Webly-supervised Learning for Video Recognition
Haodong Duan, Yue Zhao, Yuanjun Xiong, Wentao Liu, Dahua Lin
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/open-mmlab/mmactionOfficialIn paperpytorch★ 1,875
- github.com/open-mmlab/mmaction2pytorch★ 4,959
- github.com/MCG-NJU/CPD-Videopytorch★ 60
Abstract
We introduce OmniSource, a novel framework for leveraging web data to train video recognition models. OmniSource overcomes the barriers between data formats, such as images, short videos, and long untrimmed videos for webly-supervised learning. First, data samples with multiple formats, curated by task-specific data collection and automatically filtered by a teacher model, are transformed into a unified form. Then a joint-training strategy is proposed to deal with the domain gaps between multiple data sources and formats in webly-supervised learning. Several good practices, including data balancing, resampling, and cross-dataset mixup are adopted in joint training. Experiments show that by utilizing data from multiple sources and formats, OmniSource is more data-efficient in training. With only 3.5M images and 800K minutes videos crawled from the internet without human labeling (less than 2% of prior works), our models learned with OmniSource improve Top-1 accuracy of 2D- and 3D-ConvNet baseline models by 3.0% and 3.9%, respectively, on the Kinetics-400 benchmark. With OmniSource, we establish new records with different pretraining strategies for video recognition. Our best models achieve 80.4%, 80.5%, and 83.6 Top-1 accuracies on the Kinetics-400 benchmark respectively for training-from-scratch, ImageNet pre-training and IG-65M pre-training.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| HMDB-51 | OmniSource (SlowOnly-8x8-R101-RGB + I3D Flow) | Average accuracy of 3 splits | 83.8 | — | Unverified |
| UCF101 | OmniSource (SlowOnly-8x8-R101-RGB + I3D-Flow) | 3-fold Accuracy | 98.6 | — | Unverified |