CONF
stephenson02a/IDIAP
Mixed Bayesian Networks with Auxiliary Variables for Automatic Speech Recognition
Stephenson, Todd Andrew
Magimai-Doss, Mathew
Bourlard, Hervé
EXTERNAL
https://publications.idiap.ch/attachments/papers/2002/todd-icpr2002.pdf
PUBLIC
https://publications.idiap.ch/index.php/publications/showcite/stephenson01c
Related documents
International Conference on Pattern Recognition (ICPR~2002)
4
293-296
2002
Quebec City, PQ, Canada
August 2002
Standard hidden Markov models (HMMs,',','),
as used in automatic speech recognition (ASR,',','),
calculate their emission probabilities by an artificial neural network (ANN) or a Gaussian distribution conditioned on the hidden state variable, considering the emissions independent of any other variable in the model. Recent work showed the benefit of conditioning the emission distributions on a discrete auxiliary variable, which is observed in training and hidden in recognition. Related work has shown the utility of conditioning the emission distributions on a continuous auxiliary variable. We apply mixed Bayesian networks (BNs) to extend these works by introducing a continuous auxiliary variable that is observed in training but is hidden in recognition. We find that an auxiliary pitch variable conditioned itself upon the hidden state can degrade performance unless the auxiliary variable is also hidden. The performance, furthermore, can be improved by making the auxiliary pitch variable independent of the hidden state.
REPORT
stephenson01c/IDIAP
Mixed Bayesian Networks with Auxiliary Variables for Automatic Speech Recognition
Stephenson, Todd Andrew
Magimai-Doss, Mathew
Bourlard, Hervé
EXTERNAL
https://publications.idiap.ch/attachments/reports/2001/rr01-45.pdf
PUBLIC
Idiap-RR-45-2001
2001
IDIAP
In ``International Conference on Pattern Recognition (ICPR~2002)'', 2002
Standard hidden Markov models (HMMs,',','),
as used in automatic speech recognition (ASR,',','),
calculate their emission probabilities by an artificial neural network (ANN) or a Gaussian distribution conditioned on the hidden state variable, considering the emissions independent of any other variable in the model. Recent work showed the benefit of conditioning the emission distributions on a discrete auxiliary variable, which is observed in training and hidden in recognition. Related work has shown the utility of conditioning the emission distributions on a continuous auxiliary variable. We apply mixed Bayesian networks (BNs) to extend these works by introducing a continuous auxiliary variable that is observed in training but is hidden in recognition. We find that an auxiliary pitch variable conditioned itself upon the hidden state can degrade performance unless the auxiliary variable is also hidden. The performance, furthermore, can be improved by making the auxiliary pitch variable independent of the hidden state.