@inproceedings{59854000a4de47e2856ac8dbe247f1e8,
title = "A Deep Learning-based Radar and Camera Sensor Fusion Architecture for Object Detection",
abstract = "Object detection in camera images, using deep learning has been proven successfully in recent years. Rising detection rates and computationally efficient network structures are pushing this technique towards application in production vehicles. Nevertheless, the sensor quality of the camera is limited in severe weather conditions and through increased sensor noise in sparsely lit areas and at night. Our approach enhances current 2D object detection networks by fusing camera data and projected sparse radar data in the network layers. The proposed CameraRadarFusion Net (CRF-Net) automatically learns at which level the fusion of the sensor data is most beneficial for the detection result. Additionally, we introduce BlackIn, a training strategy inspired by Dropout, which focuses the learning on a specific sensor type. We show that the fusion network is able to outperform a state-of-the-art image-only network for two different datasets. The code for this research will be made available to the public at: https://github.com/TUMFTM/CameraRadarFusionNet.",
keywords = "Autonomous Driving, Deep Learning, Low Level Fusion, Multi-modal Sensor Fusion, Neural Fusion, Neural Networks, Object Detection, Radar Processing, Raw Data Fusion, Sensor Fusion",
author = "Felix Nobis and Maximilian Geisslinger and Markus Weber and Johannes Betz and Markus Lienkamp",
note = "Publisher Copyright: {\textcopyright} 2019 IEEE.; 2019 Symposium on Sensor Data Fusion: Trends, Solutions, Applications, SDF 2019 ; Conference date: 15-10-2019 Through 17-10-2019",
year = "2019",
month = oct,
doi = "10.1109/SDF.2019.8916629",
language = "English",
series = "2019 Symposium on Sensor Data Fusion: Trends, Solutions, Applications, SDF 2019",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
booktitle = "2019 Symposium on Sensor Data Fusion",
}