Impact of single-microphone dereverberation on DNN-based meeting transcription systems
2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
MetadataShow full item record
Yoshioka, T., Chen, X., & Gales, M. (2014). Impact of single-microphone dereverberation on DNN-based meeting transcription systems. 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 5527-5531. https://doi.org/10.1109/ICASSP.2014.6854660
Over the past few decades, a range of front-end techniques have been proposed to improve the robustness of automatic speech recognition systems against environmental distortion. While these techniques are effective for small tasks consisting of carefully designed data sets, especially when used with a classical acoustic model, there has been limited evidence that they are useful for a state-of-the-art system with large scale realistic data. This paper focuses on reverberation as a type of distortion and investigates the degree to which dereverberation processing can improve the performance of various forms of acoustic models based on deep neural networks (DNNs) in a challenging meeting transcription task using a single distant microphone. Experimental results show that dereverberation improves the recognition performance regardless of the acoustic model structure and the type of the feature vectors input into the neural networks, providing additional relative improvements of 4.7% and 4.1% to our best configured speaker-independent and speaker-adaptive DNN-based systems, respectively.
Xie Chen was funded by Toshiba Research Europe Ltd, Cambridge Research Lab.
External DOI: https://doi.org/10.1109/ICASSP.2014.6854660
This record's URL: https://www.repository.cam.ac.uk/handle/1810/247409