Abstract
In order to scale standard Gaussian process (GP) regression to large-scale datasets, aggregation models employ factorized training process and then combine predictions from distributed experts. The state-of-the-art aggregation models, however, either provide inconsistent predictions or require time-consuming aggregation process. We first prove the inconsistency of typical aggregations using disjoint or random data partition, and then present a consistent yet efficient aggregation model for large-scale GP. The proposed model inherits the advantages of aggregations, e.g., closed-form inference and aggregation, par- allelization and distributed computing. Furthermore, theoretical and empirical analyses reveal that the new aggregation model performs better due to the consistent predictions that converge to the true underlying function when the training size approaches infinity.
Original language | English |
---|---|
Title of host publication | Proceedings of Machine Learning Research |
Subtitle of host publication | International Conference on Machine Learning, 10-15 July 2018, Stockholmsmässan, Stockholm Sweden |
Editors | Jennifer Dy, Andreas Krause |
Place of Publication | Stroudsburg PA USA |
Publisher | International Machine Learning Society (IMLS) |
Pages | 3131-3140 |
Number of pages | 10 |
Volume | 80 |
ISBN (Electronic) | 9781510867963 |
Publication status | Published - 2018 |
Externally published | Yes |
Event | International Conference on Machine Learning 2018 - Stockholmsmässan, Stockholm, Sweden Duration: 10 Jul 2018 → 15 Jul 2018 Conference number: 35th |
Conference
Conference | International Conference on Machine Learning 2018 |
---|---|
Abbreviated title | ICML 2018 |
Country/Territory | Sweden |
City | Stockholm |
Period | 10/07/18 → 15/07/18 |