کد مقاله | کد نشریه | سال انتشار | مقاله انگلیسی | نسخه تمام متن |
---|---|---|---|---|
534293 | 870244 | 2014 | 7 صفحه PDF | دانلود رایگان |
• Two on-line approaches for Gaussian process (GP) regression are proposed.
• The 1st approach (RGP) is computationally cheap with errors close to a full GP.
• The 2nd approach (RGP★) in addition performs on-line hyperparameter learning.
• RGP★ can outperform even off-line learning algorithms in terms of error.
• RGP★ is computationally cheaper than other on-line learners with lower error.
Two approaches for on-line Gaussian process regression with low computational and memory demands are proposed. The first approach assumes known hyperparameters and performs regression on a set of basis vectors that stores mean and covariance estimates of the latent function. The second approach additionally learns the hyperparameters on-line. For this purpose, techniques from nonlinear Gaussian state estimation are exploited. The proposed approaches are compared to state-of-the-art sparse Gaussian process algorithms.
Journal: Pattern Recognition Letters - Volume 45, 1 August 2014, Pages 85–91