Minimax-Optimal Rates for Sparse Additive Models over Kernel Classes via Convex Programming
CALIFORNIA UNIV BERKELEY DEPT OF STATISTICS
Pagination or Media Count:
Sparse additive models are families of d-variate functions that have the additive decomposition f Sum, j included in S, fj , where S is a unknown subset of cardinality s much less than d. We consider the case where each component function fj lies in a reproducing kernel Hilbert space, and analyze a simple kernel-based convex program for estimating the unknown function f. Working within a high-dimensional framework that allows both the dimension d and sparsity s to scale, we derive convergence rates in the L2P and L2Pn norms. These rates consist of two terms a subset selection term of the order s log dn , corresponding to the difficulty of finding the unknown s-sized subset, and an estimation error term of the order sVnsup 2, where Vnsup 2 is the optimal rate for estimating an univariate function within the RKHS. We complement these achievable results by deriving minimax lower bounds on the L2P error, thereby showing that our method is optimal up to constant factors for sub-linear sparsity s od. Thus, we obtain optimal minimax rates for many interesting classes of sparse additive models, including polynomials, splines, nite-rank kernel classes, as well as Sobolev smoothness classes.
- Numerical Mathematics
- Theoretical Mathematics