SOTAVerified

Integrally Migrating Pre-trained Transformer Encoder-decoders for Visual Object Detection

2022-05-19ICCV 2023Code Available1· sign in to hype

Feng Liu, Xiaosong Zhang, Zhiliang Peng, Zonghao Guo, Fang Wan, Xiangyang Ji, Qixiang Ye

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Modern object detectors have taken the advantages of backbone networks pre-trained on large scale datasets. Except for the backbone networks, however, other components such as the detector head and the feature pyramid network (FPN) remain trained from scratch, which hinders fully tapping the potential of representation models. In this study, we propose to integrally migrate pre-trained transformer encoder-decoders (imTED) to a detector, constructing a feature extraction path which is ``fully pre-trained" so that detectors' generalization capacity is maximized. The essential differences between imTED with the baseline detector are twofold: (1) migrating the pre-trained transformer decoder to the detector head while removing the randomly initialized FPN from the feature extraction path; and (2) defining a multi-scale feature modulator (MFM) to enhance scale adaptability. Such designs not only reduce randomly initialized parameters significantly but also unify detector training with representation learning intendedly. Experiments on the MS COCO object detection dataset show that imTED consistently outperforms its counterparts by 2.4 AP. Without bells and whistles, imTED improves the state-of-the-art of few-shot object detection by up to 7.6 AP. Code is available at https://github.com/LiewFeng/imTED.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
MS-COCO (10-shot)imTED+ViT-SAP15Unverified
MS-COCO (10-shot)imTED+ViT-BAP22.5Unverified
MS-COCO (30-shot)imTED+ViT-BAP30.2Unverified
MS-COCO (30-shot)imTED+ViT-SAP21Unverified

Reproductions