Home Page




Editorial Board


Open Source Software




Frequently Asked Questions

Contact Us

RSS Feed

Bayesian Combination of Probabilistic Classifiers using Multivariate Normal Mixtures

Gregor Pirš, Erik Štrumbelj; 20(51):1−18, 2019.


Ensemble methods are a powerful tool, often outperforming individual prediction models. Existing Bayesian ensembles either do not model the correlations between sources, or they are only capable of combining non-probabilistic predictions. We propose a new model, which overcomes these disadvantages. Transforming the probabilistic predictions with the inverse additive logistic transformation allows us to model the correlations with multivariate normal mixtures. We derive an efficient Gibbs sampler for the proposed model and implement a regularization method to make it more robust. We compare our method to related work and the classical linear opinion pool. Empirical evaluation on several toy and real-world data sets, including a case study on air-pollution forecasting, shows that the method outperforms other methods, while being robust and easy to use.

© JMLR 2019. (edit, beta)