The full text of this item is not available at this time because the student has placed this item under an embargo for a period of time. The Libraries are not authorized to provide a copy of this work during the embargo period, even for Texas A&M users with NetID.
Computationally Efficient Low-Rank Algorithms for Gaussian Process Regression
dc.contributor.advisor | Sarin, Vivek | |
dc.creator | Thomas, Emil | |
dc.date.accessioned | 2023-09-18T17:12:29Z | |
dc.date.created | 2022-12 | |
dc.date.issued | 2022-11-30 | |
dc.date.submitted | December 2022 | |
dc.identifier.uri | https://hdl.handle.net/1969.1/198739 | |
dc.description.abstract | Gaussian Process Regression (GPR) is a Bayesian non-parametric method widely used in machine learning for supervised learning. Compared to neural networks and support vector regressions, prediction using GPR provides one with a posterior distribution. Further, the training of GPR can be done efficiently by exact optimization of the marginal likelihood with implicit regularization. However, both the training and prediction require the inversion of an N ×N kernel matrix (time complexity O(N^3)), which limits the scalability of GPR for big datasets (N > 10^4). In this research, we develop algorithms to scale GPR without sacrificing accuracy. Our training algorithm uses the thin-QR decomposition of the low-rank matrix used in the Nyström approximation. The limitations of the Nyström method are eliminated by restricting the prediction in the subspace spanned by the orthogonal matrix of this decomposition. To improve the prediction accuracy, we propose two algorithms that augment the neighbors of a test point. These algorithms leverage the general structure of the problem through the low-rank approximation and improves its accuracy further by exploiting locality at each test input. Results on synthetic and real-world datasets show that our algorithms achieve an accuracy comparable to the full kernel matrix at lower ranks than other competing methods. Thus, our algorithms provide faster training and prediction while at the same time reducing storage requirements. | |
dc.format.mimetype | application/pdf | |
dc.language.iso | en | |
dc.subject | Gaussian Process | |
dc.subject | Nystrom Approximation | |
dc.subject | QR decomposition | |
dc.subject | Low-rank Approximation | |
dc.subject | Subset of data method | |
dc.subject | Bayesian non-parametric regression | |
dc.subject | Supervised learning | |
dc.title | Computationally Efficient Low-Rank Algorithms for Gaussian Process Regression | |
dc.type | Thesis | |
thesis.degree.department | Computer Science and Engineering | |
thesis.degree.discipline | Computer Engineering | |
thesis.degree.grantor | Texas A&M University | |
thesis.degree.name | Doctor of Philosophy | |
thesis.degree.level | Doctoral | |
dc.contributor.committeeMember | Walker, Duncan | |
dc.contributor.committeeMember | Mahapatra, Rabi | |
dc.contributor.committeeMember | Gildin, Eduardo | |
dc.type.material | text | |
dc.date.updated | 2023-09-18T17:12:30Z | |
local.embargo.terms | 2024-12-01 | |
local.embargo.lift | 2024-12-01 | |
local.etdauthor.orcid | 0000-0003-1347-4072 |
Files in this item
This item appears in the following Collection(s)
-
Electronic Theses, Dissertations, and Records of Study (2002– )
Texas A&M University Theses, Dissertations, and Records of Study (2002– )