Open Access
February 2004 Aggregating regression procedures to improve performance
Yuhong Yang
Author Affiliations +
Bernoulli 10(1): 25-47 (February 2004). DOI: 10.3150/bj/1077544602

Abstract

A fundamental question regarding combining procedures concerns the potential gain and how much one needs to pay for it in terms of statistical risk. Juditsky and Nemirovski considered the case where a large number of procedures are to be combined. We give upper and lower bounds for complementary cases. Under an $l_1$ constraint on the linear coefficients, it is shown that for pursuing the best linear combination of $n^\tau$ procedures, in terms of rate of convergence under the squared $L_2$ loss, one can pay a price of order $O$(log $n/n^{1-\tau})$ when $0 < \tau < \frac{1}{2}$ and a price of order $O$((log $n/n)^{\frac {1}{2}}$ when $\frac{1}{2} \leq \tau < \infty$. These rates cannot be improved or essentially improved in a uniform sense. This result suggests that one should be cautious in pursuing the best linear combination, because one may end up paying a high price for nothing when linear combination in fact does not help. We show that with care in aggregation, the final procedure can automatically avoid paying the high price for such a case and then behaves as well as the best candidate procedure.

Citation

Download Citation

Yuhong Yang. "Aggregating regression procedures to improve performance." Bernoulli 10 (1) 25 - 47, February 2004. https://doi.org/10.3150/bj/1077544602

Information

Published: February 2004
First available in Project Euclid: 23 February 2004

zbMATH: 1040.62030
MathSciNet: MR2044592
Digital Object Identifier: 10.3150/bj/1077544602

Keywords: adaptive estimation , aggregating procedures , linear combining , Nonparametric regression

Rights: Copyright © 2004 Bernoulli Society for Mathematical Statistics and Probability

Vol.10 • No. 1 • February 2004
Back to Top