Skip to the content.

Fusing Event-based and RGB camera for Robust Object Detection in Adverse Conditions

The ability to detect objects, under image corruptions and different weather conditions is vital for deep learning models especially when applied to real-world applications such as autonomous driving. Traditional RGB-based detection fails under these conditions and it is thus important to design a sensor suite that is redundant to failures of the primary frame-based detection. Event-based cameras can complement frame-based cameras in low-light conditions and high dynamic range scenarios that an autonomous vehicle can encounter during navigation. Accordingly, we propose a redundant sensor fusion model of event-based and frame-based cameras that is robust to common image corruptions. The method utilizes a voxel grid representation for events as input and proposes a two-parallel feature extractor network for frames and events. Our sensor fusion approach is more robust to corruptions by over 30% compared to only frame-based detections and outperforms the only event-based detection. The model is trained and evaluated on the publicly released DSEC dataset.

Datset and training

The DSEC dataset contains data from two event and frame-based cameras in stereo setup recorded from a moving vehicle in cities across Switzerland and data from both sensor modalities are publicly released. However, they lack The object detection labels which we generated using YOLOv5 on the left RGB images and then transferred to the event frame using homographic transformation and refinement. dsec

To evaluate robustness and efficacy of event-based camera to act as a redundant sensor modality, the RGB images were corrupted by adding noise, blur and weather conditions. The performance is evaluated over 5 severity levels for each of the 15 corruption types. The corruption method is adapted from this work by D. Hendrycks et al. All the evaluated models are trained only clean dataset, only during testing, the RGB images are corrupted as shown in the figure. For sensor fusion models, image dropouts were used with a probaility of 0.15 during training.

15_corruptImages

Sensor Fusion Models

In this work, we propose a Feature Pyramid Net fusion (FPN) which uses ResNet-50 used as backbone for feature extraction as shown in figure below. Events and RGB features are fused at multiple scales before feeding it to the FPN layers. Network_Architecture

Apart from this, we also evaluate the efficacy of combining events and RGB channels early. In this work we have evaluated 2 type of event representations:

Results

Model Input mAP rPC
RetinaNet-50 Event Voxel 0.12 -
RetinaNet-50 Event Gray 0.12 -
RetinaNet-50 RGB 0.25 38.6
Early Fusion Event Gray+RGB 0.26 40.4
FPN-Fusion Event Gray+RGB 0.25 60.8
Early Fusion Event Voxel+RGB 0.24 66.2
FPN-Fusion Event Voxel+RGB 0.24 68.7