Abstract
A fundamental question regarding combining procedures concerns the potential gain and how much one needs to pay for it in terms of statistical risk. Juditsky and Nemirovski considered the case where a large number of procedures are to be combined. We give upper and lower bounds for complementary cases. Under an $l_1$ constraint on the linear coefficients, it is shown that for pursuing the best linear combination of $n^\tau$ procedures, in terms of rate of convergence under the squared $L_2$ loss, one can pay a price of order $O$(log $n/n^{1-\tau})$ when $0 < \tau < \frac{1}{2}$ and a price of order $O$((log $n/n)^{\frac {1}{2}}$ when $\frac{1}{2} \leq \tau < \infty$. These rates cannot be improved or essentially improved in a uniform sense. This result suggests that one should be cautious in pursuing the best linear combination, because one may end up paying a high price for nothing when linear combination in fact does not help. We show that with care in aggregation, the final procedure can automatically avoid paying the high price for such a case and then behaves as well as the best candidate procedure.
Citation
Yuhong Yang. "Aggregating regression procedures to improve performance." Bernoulli 10 (1) 25 - 47, February 2004. https://doi.org/10.3150/bj/1077544602
Information