About

Log in?

DTU users get better search results including licensed content and discounts on order fees.

Anyone can log in and get personalized features such as favorites, tags and feeds.

Log in as DTU user Log in as non-DTU user No thanks

DTU Findit

Journal article

Computational speech segregation based on an auditory-inspired modulation analysis

From

Department of Electrical Engineering, Technical University of Denmark1

Hearing Systems, Department of Electrical Engineering, Technical University of Denmark2

A monaural speech segregation system is presented that estimates the ideal binary mask from noisy speech based on the supervised learning of amplitude modulation spectrogram (AMS) features. Instead of using linearly scaled modulation filters with constant absolute bandwidth, an auditory- inspired modulation filterbank with logarithmically scaled filters is employed.

To reduce the dependency of the AMS features on the overall background noise level, a feature normalization stage is applied. In addition, a spectro-temporal integration stage is incorporated in order to exploit the context information about speech activity present in neighboring time-frequency units.

In order to evaluate the generalization performance of the system to unseen acoustic conditions, the speech segregation system is trained with a limited set of low signal-to-noise ratio (SNR) conditions, but tested over a wide range of SNRs up to 20dB. A systematic evaluation of the system demonstrates that auditory-inspired modulation processing can substantially improve the mask estimation accuracy in the presence of stationary and fluctuating interferers

Language: English
Publisher: Acoustical Society of America
Year: 2014
Pages: 3350-3359
ISSN: 15208524 , 00014966 and 01630962
Types: Journal article
DOI: 10.1121/1.4901711
ORCIDs: May, Tobias and Dau, Torsten

DTU users get better search results including licensed content and discounts on order fees.

Log in as DTU user

Access

Analysis