Using KL-divergence and multilingual information to improve ASR for under-resourced languages
Type of publication: | Conference paper |
Citation: | Imseng_ICASSP_2012 |
Booktitle: | Proceedings IEEE International Conference on Acoustics, Speech and Signal Processing |
Year: | 2012 |
Month: | March |
Pages: | 4869--4872 |
Location: | Kyoto |
Abstract: | Setting out from the point of view that automatic speech recognition (ASR) ought to benefit from data in languages other than the target language, we propose a novel Kullback-Leibler (KL) divergence based method that is able to exploit multilingual information in the form of universal phoneme posterior probabilities conditioned on the acoustics. We formulate a means to train a recognizer on several different languages, and subsequently recognize speech in a target language for which only a small amount of data is available. Taking the Greek SpeechDat(II) data as an example, we show that the proposed formulation is sound, and show that it is able to outperform a current state-of-the-art HMM/GMM system. We also use a hybrid Tandem-like system to further understand the source of the benefit. |
Keywords: | |
Projects |
Idiap SNSF-MULTI IM2 |
Authors | |
Added by: | [UNK] |
Total mark: | 0 |
Attachments
|
|
Notes
|
|
|