Unmanned Aerial Vehicles (UAVs) or drones are increasingly used in search and rescue missions to detect human presence. Existing systems primarily leverage vision-based methods, which are prone to fail under low-visibility or occlusion. Drone-based audio perception offers promise but suffers from extreme ego-noise that masks sounds indicating human presence.

In this project, we develop DroneAudioset (The dataset is publicly available at https://huggingface.co/datasets/ahlab-drone-project/DroneAudioSet/ under the MIT license), a comprehensive and largest drone audition dataset featuring 23.5 hours of annotated recordings, covering a wide range of signal-to-noise ratios (SNRs) from -60 dB to 0 dB, across various drone types, throttles, microphone configurations, as well as environments. The dataset enables development and systematic evaluation of noise suppression and classification methods for human-presence detection under challenging conditions, while also informing practical design considerations for drone audition systems, such as microphone placement trade-offs, and development of drone noise-aware audio processing.
This project aims towards enabling the design and deployment of drone-audition systems, and studying human operator and drone collaboration through audio modality for a swarm of drones.

PUBLICATIONS

DroneAudioset: An Audio Dataset for Drone-based Search and Rescue

Gupta, C.*, Ramesh, S.*, Sasikumar, P., Yeo, K. P., & Nanayakkara, S. C., "DroneAudioset: An Audio Dataset for Drone-based Search and Rescue," NeurIPS 2025.

Blind Drone

Unmanned Aerial Vehicles (UAVs) or drones are increasingly used in search and rescue missions to detect human presence. Existing systems primarily leverage vision-based methods, which are prone to fail under low-visibility or occlusion. Drone-based audio perception offers promise but suffers from extreme ego-noise that masks sounds indicating human presence.

In this project, we develop DroneAudioset (The dataset is publicly available at https://huggingface.co/datasets/ahlab-drone-project/DroneAudioSet/ under the MIT license), a comprehensive and largest drone audition dataset featuring 23.5 hours of annotated recordings, covering a wide range of signal-to-noise ratios (SNRs) from -60 dB to 0 dB, across various drone types, throttles, microphone configurations, as well as environments. The dataset enables development and systematic evaluation of noise suppression and classification methods for human-presence detection under challenging conditions, while also informing practical design considerations for drone audition systems, such as microphone placement trade-offs, and development of drone noise-aware audio processing.
This project aims towards enabling the design and deployment of drone-audition systems, and studying human operator and drone collaboration through audio modality for a swarm of drones.

PUBLICATIONS

DroneAudioset: An Audio Dataset for Drone-based Search and Rescue

Gupta, C.*, Ramesh, S.*, Sasikumar, P., Yeo, K. P., & Nanayakkara, S. C., "DroneAudioset: An Audio Dataset for Drone-based Search and Rescue," NeurIPS 2025.

Unmanned Aerial Vehicles (UAVs) or drones are increasingly used in search and rescue missions to detect human presence. Existing systems primarily leverage vision-based methods, which are prone to fail under low-visibility or occlusion. Drone-based audio perception offers promise but suffers from extreme ego-noise that masks sounds indicating human presence.

In this project, we develop DroneAudioset (The dataset is publicly available at https://huggingface.co/datasets/ahlab-drone-project/DroneAudioSet/ under the MIT license), a comprehensive and largest drone audition dataset featuring 23.5 hours of annotated recordings, covering a wide range of signal-to-noise ratios (SNRs) from -60 dB to 0 dB, across various drone types, throttles, microphone configurations, as well as environments. The dataset enables development and systematic evaluation of noise suppression and classification methods for human-presence detection under challenging conditions, while also informing practical design considerations for drone audition systems, such as microphone placement trade-offs, and development of drone noise-aware audio processing.
This project aims towards enabling the design and deployment of drone-audition systems, and studying human operator and drone collaboration through audio modality for a swarm of drones.

PUBLICATIONS

DroneAudioset: An Audio Dataset for Drone-based Search and Rescue

Gupta, C.*, Ramesh, S.*, Sasikumar, P., Yeo, K. P., & Nanayakkara, S. C., "DroneAudioset: An Audio Dataset for Drone-based Search and Rescue," NeurIPS 2025.

Blind Drone

Unmanned Aerial Vehicles (UAVs) or drones are increasingly used in search and rescue missions to detect human presence. Existing systems primarily leverage vision-based methods, which are prone to fail under low-visibility or occlusion. Drone-based audio perception offers promise but suffers from extreme ego-noise that masks sounds indicating human presence.

In this project, we develop DroneAudioset (The dataset is publicly available at https://huggingface.co/datasets/ahlab-drone-project/DroneAudioSet/ under the MIT license), a comprehensive and largest drone audition dataset featuring 23.5 hours of annotated recordings, covering a wide range of signal-to-noise ratios (SNRs) from -60 dB to 0 dB, across various drone types, throttles, microphone configurations, as well as environments. The dataset enables development and systematic evaluation of noise suppression and classification methods for human-presence detection under challenging conditions, while also informing practical design considerations for drone audition systems, such as microphone placement trade-offs, and development of drone noise-aware audio processing.
This project aims towards enabling the design and deployment of drone-audition systems, and studying human operator and drone collaboration through audio modality for a swarm of drones.

PUBLICATIONS

DroneAudioset: An Audio Dataset for Drone-based Search and Rescue

Gupta, C.*, Ramesh, S.*, Sasikumar, P., Yeo, K. P., & Nanayakkara, S. C., "DroneAudioset: An Audio Dataset for Drone-based Search and Rescue," NeurIPS 2025.