RAFT: Recurrent All-Pairs Field Transforms for Optical Flow
Zachary Teed, Jia Deng
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/princeton-vl/RAFTOfficialIn paperpytorch★ 3,986
- github.com/pytorch/visionpytorch★ 17,584
- github.com/open-mmlab/mmflowpytorch★ 1,050
- github.com/ykasten/layered-neural-atlasespytorch★ 613
- github.com/neu-vi/ezflowpytorch★ 137
- github.com/neu-vig/ezflowpytorch★ 137
- github.com/Visual-Behavior/aloceptionpytorch★ 93
- github.com/plusgood-steven/id-blaupytorch★ 58
- github.com/duke-vision/optical-flow-active-learning-releasepytorch★ 18
- github.com/ToyotaResearchInstitute/att-awarepytorch★ 14
Abstract
We introduce Recurrent All-Pairs Field Transforms (RAFT), a new deep network architecture for optical flow. RAFT extracts per-pixel features, builds multi-scale 4D correlation volumes for all pairs of pixels, and iteratively updates a flow field through a recurrent unit that performs lookups on the correlation volumes. RAFT achieves state-of-the-art performance. On KITTI, RAFT achieves an F1-all error of 5.10%, a 16% error reduction from the best published result (6.10%). On Sintel (final pass), RAFT obtains an end-point-error of 2.855 pixels, a 30% error reduction from the best published result (4.098 pixels). In addition, RAFT has strong cross-dataset generalization as well as high efficiency in inference time, training speed, and parameter count. Code is available at https://github.com/princeton-vl/RAFT.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| KITTI 2015 (train) | RAFT | F1-all | 17.4 | — | Unverified |
| Sintel-clean | RAFT (warm-start) | Average End-Point Error | 1.61 | — | Unverified |
| Sintel-final | RAFT (warm-start) | Average End-Point Error | 2.86 | — | Unverified |
| Spring | RAFT | 1px total | 6.79 | — | Unverified |