SOTAVerified

Pix2seq: A Language Modeling Framework for Object Detection

2021-09-22ICLR 2022Code Available1· sign in to hype

Ting Chen, Saurabh Saxena, Lala Li, David J. Fleet, Geoffrey Hinton

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We present Pix2Seq, a simple and generic framework for object detection. Unlike existing approaches that explicitly integrate prior knowledge about the task, we cast object detection as a language modeling task conditioned on the observed pixel inputs. Object descriptions (e.g., bounding boxes and class labels) are expressed as sequences of discrete tokens, and we train a neural network to perceive the image and generate the desired sequence. Our approach is based mainly on the intuition that if a neural network knows about where and what the objects are, we just need to teach it how to read them out. Beyond the use of task-specific data augmentations, our approach makes minimal assumptions about the task, yet it achieves competitive results on the challenging COCO dataset, compared to highly specialized and well optimized detection algorithms.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
COCO minivalPix2seq (ViT-L)box AP50Unverified
COCO minivalPix2seq (R50-C4)box AP47.3Unverified
COCO minivalPix2seq (ViT-B)box AP47.1Unverified
COCO minivalPix2seq (R101-DC5)box AP45Unverified
COCO minivalPix2seq (R50-DC5 )box AP43.2Unverified
COCO minivalPix2seq (R50)box AP42.6Unverified

Reproductions