pith. machine review for the scientific record. sign in

arxiv: 1903.07756 · v1 · submitted 2019-03-18 · 💻 cs.LG · stat.ML

Recognition: unknown

Hierarchical Routing Mixture of Experts

Authors on Pith no claims yet
classification 💻 cs.LG stat.ML
keywords modelsexpertsdataregressionmodelfittedhierarchicalhrme
0
0 comments X
read the original abstract

In regression tasks the distribution of the data is often too complex to be fitted by a single model. In contrast, partition-based models are developed where data is divided and fitted by local models. These models partition the input space and do not leverage the input-output dependency of multimodal-distributed data, and strong local models are needed to make good predictions. Addressing these problems, we propose a binary tree-structured hierarchical routing mixture of experts (HRME) model that has classifiers as non-leaf node experts and simple regression models as leaf node experts. The classifier nodes jointly soft-partition the input-output space based on the natural separateness of multimodal data. This enables simple leaf experts to be effective for prediction. Further, we develop a probabilistic framework for the HRME model, and propose a recursive Expectation-Maximization (EM) based algorithm to learn both the tree structure and the expert models. Experiments on a collection of regression tasks validate the effectiveness of our method compared to a variety of other regression models.

This paper has not been read by Pith yet.

discussion (0)

Sign in with ORCID, Apple, or X to comment. Anyone can read and Pith papers without signing in.