TY - GEN
T1 - Aggregating the Gaussian Experts' Predictions via Undirected Graphical Models
AU - Jalali, Hamed
AU - Kasneci, Gjergji
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Distributed Gaussian process (DGP) is a popular approach to scale Gaussian processes to big data which divides the training data into some subsets, performs local inference for each partition, and aggregates the results to acquire global prediction. To combine the local predictions, the conditional independence assumption is used which basically means there is a perfect diversity between the subsets. Although it keeps the aggregation tractable, it is often violated in practice and generally yields poor results. In this paper, we propose a novel approach for aggregating the Gaussian experts' predictions by Gaussian graphical model (GGM) where the target aggregation is defined as an unobserved latent variable and the local predictions are the observed variables. We first estimate the joint distribution of la-tent and observed variables using the Expectation-Maximization (EM) algorithm. The interaction between experts can be en-coded by the precision matrix of the joint distribution and the aggregated predictions are obtained based on the property of conditional Gaussian distribution. Using both synthetic and real datasets, our experimental evaluations illustrate that our new method outperforms other state-of-the-art DGP approaches.
AB - Distributed Gaussian process (DGP) is a popular approach to scale Gaussian processes to big data which divides the training data into some subsets, performs local inference for each partition, and aggregates the results to acquire global prediction. To combine the local predictions, the conditional independence assumption is used which basically means there is a perfect diversity between the subsets. Although it keeps the aggregation tractable, it is often violated in practice and generally yields poor results. In this paper, we propose a novel approach for aggregating the Gaussian experts' predictions by Gaussian graphical model (GGM) where the target aggregation is defined as an unobserved latent variable and the local predictions are the observed variables. We first estimate the joint distribution of la-tent and observed variables using the Expectation-Maximization (EM) algorithm. The interaction between experts can be en-coded by the precision matrix of the joint distribution and the aggregated predictions are obtained based on the property of conditional Gaussian distribution. Using both synthetic and real datasets, our experimental evaluations illustrate that our new method outperforms other state-of-the-art DGP approaches.
KW - Conditional Dependency
KW - Distributed Gaussian Process
KW - Gaussian Graph-ical Models
UR - http://www.scopus.com/inward/record.url?scp=85127570045&partnerID=8YFLogxK
U2 - 10.1109/BigComp54360.2022.00014
DO - 10.1109/BigComp54360.2022.00014
M3 - Conference contribution
AN - SCOPUS:85127570045
T3 - Proceedings - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022
SP - 23
EP - 26
BT - Proceedings - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022
A2 - Unger, Herwig
A2 - Kim, Young-Kuk
A2 - Hwang, Eenjun
A2 - Cho, Sung-Bae
A2 - Pareigis, Stephan
A2 - Kyandoghere, Kyamakya
A2 - Ha, Young-Guk
A2 - Kim, Jinho
A2 - Morishima, Atsuyuki
A2 - Wagner, Christian
A2 - Kwon, Hyuk-Yoon
A2 - Moon, Yang-Sae
A2 - Leung, Carson
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2022 IEEE International Conference on Big Data and Smart Computing, BigComp 2022
Y2 - 17 January 2022 through 20 January 2022
ER -