Abstract
There is increasing interest in the problem of nonparametric regression with high-dimensional predictors. When the number of predictors D is large, one encounters a daunting problem in attempting to estimate aD-dimensional surface based on limited data. Fortunately, in many applications, the support of the data is concentrated on a d-dimensional subspace with d ≤ D. Manifold learning attempts to estimate this subspace. Our focus is on developing computationally tractable and theoretically supported Bayesian nonparametric regression methods in this context. When the subspace corresponds to a locally-Euclidean compact Riemannian manifold, we show that a Gaussian process regression approach can be applied that leads to the minimax optimal adaptive rate in estimating the regression function under some conditions. The proposed model bypasses the need to estimate the manifold, and can be implemented using standard algorithms for posterior computation in Gaussian processes. Finite sample performance is illustrated in a data analysis example.
Original language | English (US) |
---|---|
Pages (from-to) | 876-905 |
Number of pages | 30 |
Journal | Annals of Statistics |
Volume | 44 |
Issue number | 2 |
DOIs | |
State | Published - Apr 2016 |
Externally published | Yes |
Keywords
- Asymptotics
- Contraction rates
- Dimensionality reduction
- Gaussian process
- Manifold learning
- Nonparametric Bayes
- Subspace learning
ASJC Scopus subject areas
- Statistics and Probability
- Statistics, Probability and Uncertainty