@inproceedings{2796d4f24a624d07be059b4db7488e2b,
title = "Single-Channel Speaker Distance Estimation in Reverberant Environments",
abstract = "We introduce the novel task of continuous-valued speaker distance estimation which focuses on estimating non-discrete distances between a sound source and microphone, based on audio captured by the microphone. A novel learning-based approach for estimating speaker distance in reverberant environments from a single omnidirectional microphone is proposed. Using common acoustic features, such as the magnitude and phase of the audio spectrogram, with a convolutional recurrent neural network results in errors on the order of centimeters in noiseless audios. Experiments are carried out by means of an image-source room simulator with convolved speeches from a public dataset. An ablation study is performed to demonstrate the effectiveness of the proposed feature set. Finally, a study of the impact of real background noise, extracted from the WHAM! dataset at different signal-to-noise ratios highlights the discrepancy between noisy and noiseless scenarios, underlining the difficulty of the problem.",
keywords = "Deep Learning, Distance estimation, Reverberation, Single-channel",
author = "Michael Neri and Archontis Politis and Daniel Krause and Marco Carli and Tuomas Virtanen",
note = "Publisher Copyright: {\textcopyright} 2023 IEEE.; IEEE Workshop on Applications of Signal Processing to Audio and Acoustics ; Conference date: 22-10-2023 Through 25-10-2023",
year = "2023",
doi = "10.1109/WASPAA58266.2023.10248087",
language = "English",
series = "IEEE Workshop on Applications of Signal Processing to Audio and Acoustics",
publisher = "IEEE",
booktitle = "Proceedings of the 2023 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, WASPAA 2023",
address = "United States",
}