TY - GEN
T1 - THE SPEAR CHALLENGE - REVIEW OF RESULTS
AU - Tourbabin, Vladimir
AU - Guiraud, Pierre
AU - Hafezi, Sina
AU - Naylor, Patrick A.
AU - Moore, Alastair H.
AU - Donley, Jacob
AU - Lunner, Thomas
N1 - Publisher Copyright:
© 2023 Vladimir Tourbabin et al. This is an open-access article distributed under the terms of the Creative Commons Attribution 3.0 Unported License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
PY - 2023/1/1
Y1 - 2023/1/1
N2 - Verbal communication can be challenging in the presence of acoustic noise. To tackle this problem, microphone arrays coupled with numerous processing methods have been studied in the past few decades. Recent interest in Augmented Reality (AR) applications gives rise to head-worn microphone arrays. This highlights additional important aspects such as motion of the capture device with respect to the scene, the need to preserve spatial characteristics in the processed sound, and the tightened constraints on latency and computational budgets. The SPeech Enhancement for Augmented Reality (SPEAR) Challenge, endorsed by the IEEE Challenges and Data Collection initiative, was organized in order to further ignite interest in this important problem and to obtain a better sense of the remaining technological gaps. The challenge is based on an adaptation of the recently published EasyCom dataset that contains noisy conversation recordings from a glasses form-factor AR device with 6 microphones along with the positional information and additional labeled modalities. A competitive evaluation of the challenge entrant algorithms was carried out by using both objective metrics and subjective listening tests. The current contribution is focused on providing an overview of the SPEAR challenge and highlighting some of the most important findings and outcomes.
AB - Verbal communication can be challenging in the presence of acoustic noise. To tackle this problem, microphone arrays coupled with numerous processing methods have been studied in the past few decades. Recent interest in Augmented Reality (AR) applications gives rise to head-worn microphone arrays. This highlights additional important aspects such as motion of the capture device with respect to the scene, the need to preserve spatial characteristics in the processed sound, and the tightened constraints on latency and computational budgets. The SPeech Enhancement for Augmented Reality (SPEAR) Challenge, endorsed by the IEEE Challenges and Data Collection initiative, was organized in order to further ignite interest in this important problem and to obtain a better sense of the remaining technological gaps. The challenge is based on an adaptation of the recently published EasyCom dataset that contains noisy conversation recordings from a glasses form-factor AR device with 6 microphones along with the positional information and additional labeled modalities. A competitive evaluation of the challenge entrant algorithms was carried out by using both objective metrics and subjective listening tests. The current contribution is focused on providing an overview of the SPEAR challenge and highlighting some of the most important findings and outcomes.
KW - augmented reality
KW - cocktail party problem
KW - microphone array processing
KW - multi-modal data
KW - speech enhancement
UR - http://www.scopus.com/inward/record.url?scp=85183720519&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85183720519
T3 - Proceedings of Forum Acusticum
BT - Forum Acusticum 2023 - 10th Convention of the European Acoustics Association, EAA 2023
PB - European Acoustics Association, EAA
T2 - 10th Convention of the European Acoustics Association, EAA 2023
Y2 - 11 September 2023 through 15 September 2023
ER -