CONF stephenson02a/IDIAP Mixed Bayesian Networks with Auxiliary Variables for Automatic Speech Recognition Stephenson, Todd Andrew Magimai-Doss, Mathew Bourlard, Hervé EXTERNAL https://publications.idiap.ch/attachments/papers/2002/todd-icpr2002.pdf PUBLIC https://publications.idiap.ch/index.php/publications/showcite/stephenson01c Related documents International Conference on Pattern Recognition (ICPR~2002) 4 293-296 2002 Quebec City, PQ, Canada August 2002 Standard hidden Markov models (HMMs,',','), as used in automatic speech recognition (ASR,',','), calculate their emission probabilities by an artificial neural network (ANN) or a Gaussian distribution conditioned on the hidden state variable, considering the emissions independent of any other variable in the model. Recent work showed the benefit of conditioning the emission distributions on a discrete auxiliary variable, which is observed in training and hidden in recognition. Related work has shown the utility of conditioning the emission distributions on a continuous auxiliary variable. We apply mixed Bayesian networks (BNs) to extend these works by introducing a continuous auxiliary variable that is observed in training but is hidden in recognition. We find that an auxiliary pitch variable conditioned itself upon the hidden state can degrade performance unless the auxiliary variable is also hidden. The performance, furthermore, can be improved by making the auxiliary pitch variable independent of the hidden state. REPORT stephenson01c/IDIAP Mixed Bayesian Networks with Auxiliary Variables for Automatic Speech Recognition Stephenson, Todd Andrew Magimai-Doss, Mathew Bourlard, Hervé EXTERNAL https://publications.idiap.ch/attachments/reports/2001/rr01-45.pdf PUBLIC Idiap-RR-45-2001 2001 IDIAP In ``International Conference on Pattern Recognition (ICPR~2002)'', 2002 Standard hidden Markov models (HMMs,',','), as used in automatic speech recognition (ASR,',','), calculate their emission probabilities by an artificial neural network (ANN) or a Gaussian distribution conditioned on the hidden state variable, considering the emissions independent of any other variable in the model. Recent work showed the benefit of conditioning the emission distributions on a discrete auxiliary variable, which is observed in training and hidden in recognition. Related work has shown the utility of conditioning the emission distributions on a continuous auxiliary variable. We apply mixed Bayesian networks (BNs) to extend these works by introducing a continuous auxiliary variable that is observed in training but is hidden in recognition. We find that an auxiliary pitch variable conditioned itself upon the hidden state can degrade performance unless the auxiliary variable is also hidden. The performance, furthermore, can be improved by making the auxiliary pitch variable independent of the hidden state.