Hierarchical Mixtures of Experts and the EM Algorithm
- 1 March 1994
- journal article
- research article
- Published by MIT Press in Neural Computation
- Vol. 6 (2) , 181-214
- https://doi.org/10.1162/neco.1994.6.2.181
Abstract
We present a tree-structured architecture for supervised learning. The statistical model underlying the architecture is a hierarchical mixture model in which both the mixture coefficients and the mixture components are generalized linear models (GLIM's). Learning is treated as a maximum likelihood problem; in particular, we present an Expectation-Maximization (EM) algorithm for adjusting the parameters of the architecture. We also develop an on-line learning algorithm in which the parameters are updated incrementally. Comparative simulation results are presented in the robot dynamics domain.This publication has 12 references indexed in Scilit:
- SmcHD1, containing a structural-maintenance-of-chromosomes hinge domain, has a critical role in X inactivationNature Genetics, 2008
- Neural Networks and the Bias/Variance DilemmaNeural Computation, 1992
- Multivariate Adaptive Regression SplinesThe Annals of Statistics, 1991
- A tree-structured adaptive network for function approximation in high-dimensional spacesIEEE Transactions on Neural Networks, 1991
- Adaptive Mixtures of Local ExpertsNeural Computation, 1991
- A general regression neural networkIEEE Transactions on Neural Networks, 1991
- Inferring decision trees using the minimum description lenght principleInformation and Computation, 1989
- Multilayer feedforward networks are universal approximatorsNeural Networks, 1989
- Induction of decision treesMachine Learning, 1986
- On the Convergence Properties of the EM AlgorithmThe Annals of Statistics, 1983