Carnegie Mellon University
file.pdf (154.78 kB)

Improving Low-Resource CD-DNN-HMM Using Dropout and Multilingual DNN Training

Download (154.78 kB)
journal contribution
posted on 2013-08-01, 00:00 authored by Yajie Miao, Florian MetzeFlorian Metze

We investigate two strategies to improve the context-dependent deep neural network hidden Markov model (CD-DNN-HMM) in low-resource speech recognition. Although outperforming the conventional Gaussian mixture model (GMM) HMM on various tasks, CD-DNN-HMM acoustic modeling becomes challenging with limited transcribed speech, e.g., less than 10 hours. To resolve this issue, we firstly exploit dropout which prevents overfitting in DNN finetuning and improves model robustness under data sparseness. Then, the effectiveness of multilingual DNN training is evaluated when additional auxiliary languages are available. The hidden layer parameters of the target language are shared and learned over multiple languages. Experiments show that both strategies boost the recognition performance significantly. Combining them results in further reduction in word error rate, achieving 11.6% and 6.2% relative improvement on two limited data conditions.


Publisher Statement

Copyright 2013 ISCA