Abstract
Efforts have been directed at obtaining flexible learning procedures that optimally adapt to various possible characteristics of the data generating mechanism. A question that addresses the issue of how far one can go in this direction is: Given a regression procedure, however sophisticated it is, how many regression functions are estimated accurately? In this work, for a given sequence of prescribed estimation accuracy (in sample size), we give an upper bound (in terms of metric entropy) on the number of regression functions for which the accuracy is achieved. Interesting consequences on adaptive and sparse estimations are also given.
Original language | English (US) |
---|---|
Pages (from-to) | 636-643 |
Number of pages | 8 |
Journal | Journal of Machine Learning Research |
Volume | 2 |
State | Published - 2007 |
Event | 11th International Conference on Artificial Intelligence and Statistics, AISTATS 2007 - San Juan, Puerto Rico Duration: Mar 21 2007 → Mar 24 2007 |