Sangam: A Confluence of Knowledge Streams

Parallel Gaussian process regression for big data: Low-rank representation meets markov approximation

Show simple item record

dc.contributor Massachusetts Institute of Technology. Department of Electrical Engineering and Computer Science
dc.contributor Jaillet, Patrick
dc.creator Low, Kian Hsiang
dc.creator Yu, Jiangbo
dc.creator Chen, Jie
dc.creator Jaillet, Patrick
dc.date 2018-06-12T17:40:35Z
dc.date 2018-06-12T17:40:35Z
dc.date 2015-01
dc.date.accessioned 2023-03-01T18:07:13Z
dc.date.available 2023-03-01T18:07:13Z
dc.identifier 0-262-51129-0
dc.identifier http://hdl.handle.net/1721.1/116273
dc.identifier Low, Kian Hsiang. "Parallel gaussian process regression for big data: low-rank representation meets markov approximation." AAAI'15 Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence, 25-30 January, 2015, Austin, Texas, ACM, 2015, pp. 2821-2827.
dc.identifier https://orcid.org/0000-0002-8585-6566
dc.identifier.uri http://localhost:8080/xmlui/handle/CUHPOERS/278821
dc.description The expressive power of a Gaussian process (GP) model comes at a cost of poor scalability in the data size. To improve its scalability, this paper presents a low-rank-cum-Markov approximation (LMA) of the GP model that is novel in leveraging the dual computational advantages stemming from complementing a low-rank approximate representation of the full-rank GP based on a support set of inputs with a Markov approximation of the resulting residual process; the latter approximation is guaranteed to be closest in the Kullback-Leibler distance criterion subject to some constraint and is considerably more refined than that of existing sparse GP models utilizing low-rank representations due to its more relaxed conditional independence assumption (especially with larger data). As a result, our LMA method can trade off between the size of the support set and the order of the Markov property to (a) incur lower computational cost than such sparse GP models while achieving predictive performance comparable to them and (b) accurately represent features/patterns of any scale. Interestingly, varying the Markov order produces a spectrum of LMAs with PIC approximation and full-rank GP at the two extremes. An advantage of our LMA method is that it is amenable to parallelization on multiple machines/cores, thereby gaining greater scalability. Empirical evaluation on three real-world datasets in clusters of up to 32 computing nodes shows that our centralized and parallel LMA methods are significantly more time-efficient and scalable than state-of-the-art sparse and full-rank GP regression methods while achieving comparable predictive performances.
dc.description Singapore-MIT Alliance in Research and Technology (SMART) (52 R-252-000-550-592)
dc.format application/pdf
dc.language en_US
dc.publisher Association for Computing Machinery
dc.relation http://dl.acm.org/citation.cfm?id=2886714
dc.relation AAAI'15 Proceedings of the Twenty-Ninth AAAI Conference on Artificial Intelligence
dc.rights Creative Commons Attribution-Noncommercial-Share Alike
dc.rights http://creativecommons.org/licenses/by-nc-sa/4.0/
dc.source MIT Web Domain
dc.title Parallel Gaussian process regression for big data: Low-rank representation meets markov approximation
dc.type Article
dc.type http://purl.org/eprint/type/ConferencePaper


Files in this item

Files Size Format View
Jaillet_Parallel gausian process.pdf 464.1Kb application/pdf View/Open

This item appears in the following Collection(s)

Show simple item record

Search DSpace


Advanced Search

Browse