SOTAVerified

A Deep Learning-based Radar and Camera Sensor Fusion Architecture for Object Detection

2020-05-152019 Sensor Data Fusion: Trends, Solutions, Applications (SDF) 2019Code Available1· sign in to hype

Felix Nobis, Maximilian Geisslinger, Markus Weber, Johannes Betz, Markus Lienkamp

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Object detection in camera images, using deep learning has been proven successfully in recent years. Rising detection rates and computationally efficient network structures are pushing this technique towards application in production vehicles. Nevertheless, the sensor quality of the camera is limited in severe weather conditions and through increased sensor noise in sparsely lit areas and at night. Our approach enhances current 2D object detection networks by fusing camera data and projected sparse radar data in the network layers. The proposed CameraRadarFusionNet (CRF-Net) automatically learns at which level the fusion of the sensor data is most beneficial for the detection result. Additionally, we introduce BlackIn, a training strategy inspired by Dropout, which focuses the learning on a specific sensor type. We show that the fusion network is able to outperform a state-of-the-art image-only network for two different datasets. The code for this research will be made available to the public at: https://github.com/TUMFTM/CameraRadarFusionNet.

Tasks

Reproductions