PETR: Position Embedding Transformation for Multi-View 3D Object Detection
2022-03-10Code Available3· sign in to hype
Yingfei Liu, Tiancai Wang, Xiangyu Zhang, Jian Sun
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/megvii-research/petrOfficialIn paperpytorch★ 1,044
Abstract
In this paper, we develop position embedding transformation (PETR) for multi-view 3D object detection. PETR encodes the position information of 3D coordinates into image features, producing the 3D position-aware features. Object query can perceive the 3D position-aware features and perform end-to-end object detection. PETR achieves state-of-the-art performance (50.4% NDS and 44.1% mAP) on standard nuScenes dataset and ranks 1st place on the benchmark. It can serve as a simple yet strong baseline for future research. Code is available at https://github.com/megvii-research/PETR.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 3D Object Detection on Argoverse2 Camera Only | PETR | Average mAP | 17.6 | — | Unverified |
| TruckScenes | PETR | NDS | 12.1 | — | Unverified |