Exploring Cross-fusion and Curriculum Learning for Multi-modal Human Detection on Drones

  • Ali Safa
  • , Ilja Ocket
  • , Francky Catthoor
  • , Georges Gielen

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

Abstract

In a number of applications ranging from warehouse management to people search and rescue, drones will need to evolve in the vicinity of human agents. In those situations, robust and fail-safe human detection by drones must be provided. However, human detection systems used on drones are currently based on single imaging cameras, beside a growing number of works investigating more robust detection schemes via sensor fusion. In the drone context, the fusion of standard RGB and event-based cameras has emerged, while in the automotive context, the fusion of RGB with radars has been proposed for up-most safety towards environmental conditions. In this paper, our aim is to debut the investigation of RGB, event-based camera and radar fusion. First, we acquire a novel dataset for the task of people detection in an indoor, industrial setting, by mounting the sensor fusion suite on a drone. Then, we propose a baseline convolutional neural network (CNN) architecture augmented with cross-fusion highways for sensor fusion and people detection. To train the network, we propose a novel multimodal curriculum learning procedure and demonstrate that our method (termed SAUL) greatly enhances the robustness of the system towards hard RGB failures ( on the peak F1 score) and provides a significant gain in detection performance ( on the peak F1 score) compared to the BlackIn procedure, previously proposed for cross-fusion network training. Finally, we report the performance of our system through precision-recall curve analysis and perform additional ablation studies to shed light on the key aspect of our system.

Original languageEnglish
Title of host publicationProceedings of System Engineering for Constrained Embedded Systems - DroneSE
Subtitle of host publicationDrone Systems Engineering - RAPIDO: Rapid Simulation and Performance Evaluation: Methods and Tools, HiPEAC Conference
PublisherAssociation for Computing Machinery
Pages1-7
Number of pages7
ISBN (Electronic)9781450395663
DOIs
Publication statusPublished - 17 Jan 2022
Externally publishedYes
Event2022 Workshop on System Engineering for Constrained Embedded Systems - Drone Systems Engineering and Rapid Simulation and Performance Evaluation: Methods and Tools, DroneSE and RAPIDO 2022 - Presented at HiPEAC 2022 Conference - Budapest, Hungary
Duration: 20 Jun 202222 Jun 2022

Publication series

NameACM International Conference Proceeding Series

Conference

Conference2022 Workshop on System Engineering for Constrained Embedded Systems - Drone Systems Engineering and Rapid Simulation and Performance Evaluation: Methods and Tools, DroneSE and RAPIDO 2022 - Presented at HiPEAC 2022 Conference
Country/TerritoryHungary
CityBudapest
Period20/06/2222/06/22

Keywords

  • curriculum learning
  • deep learning
  • drones
  • people detection
  • sensor fusion

Fingerprint

Dive into the research topics of 'Exploring Cross-fusion and Curriculum Learning for Multi-modal Human Detection on Drones'. Together they form a unique fingerprint.

Cite this