CONF Moerland-98.2b/IDIAP Classification using localized mixtures of experts Moerland, Perry EXTERNAL https://publications.idiap.ch/attachments/papers/1999/moerland-localized98.pdf PUBLIC https://publications.idiap.ch/index.php/publications/showcite/moerland-98.2 Related documents Proceedings of the International Conference on Artificial Neural Networks (ICANN'99) 2 838-843 1999 London: IEE (IDIAP-RR 98-14) A mixture of experts consists of a gating network that learns to partition the input space and of experts networks attributed to these different regions. This paper focuses on the choice of the gating network. First, a localized gating network based on a mixture of linear latent variable models is proposed that extends a gating network introduced by Xu et al, based on Gaussian mixture models. It is shown that this localized mixture of experts model, can be trained with the Expectation Maximization algorithm. The localized model is compared on a set of classification problems, with mixtures of experts having single or multi-layer perceptrons as gating network. It is found that the standard mixture of experts with feed-forward networks as gate often outperforms the other models. REPORT Moerland-98.2/IDIAP Localized mixtures of experts Moerland, Perry Idiap-RR-14-1998 1998 IDIAP