Conference paper
Audio-visual scene analysis in reverberant multi-talker environments
Normal-hearing subjects are accurate in localizing sound sources even in reverberant multi-talker environments (e.g., Kopčo, 2010; Weller, 2016). Weller et al. (2016) showed that subjects can accurately analyse reverberant multi-talker scenes with up to four simultaneous talkers. While multi-talker scene analysis has mainly been investigated with only auditory information, the addition of visual information might influence the subjects’ perception.
To investigate the visual influence, audio-visual scenes with a varying number of talkers and degrees of reverberation were considered in the present study. The acoustic information was provided using a spherical loudspeaker array and the visual information was provided using head-tracked virtual reality glasses.
The visual information represented various possible talker locations and the subjects were asked to identify the number of talkers and their specific locations. For the identification of talkers, subjects had to label visual locations with headlines from the talker’s speech topic. It was hypothesized that the addition of visual information improves subjects’ ability to analyse complex auditory scenes, while the amount of reverberation impairs the overall performance.
Language: | English |
---|---|
Publisher: | Deutsche Gesellschaft für Akustik e.V. |
Year: | 2019 |
Pages: | 3890-3896 |
Proceedings: | 23rd International Congress on AcousticsInternational Congress on Acoustics |
ISBN: | 3939296155 and 9783939296157 |
Types: | Conference paper |
ORCIDs: | Ahrens, Axel and Dau, Torsten |