Open Access
August 2018 Optimal shrinkage of eigenvalues in the spiked covariance model
David Donoho, Matan Gavish, Iain Johnstone
Ann. Statist. 46(4): 1742-1778 (August 2018). DOI: 10.1214/17-AOS1601

Abstract

We show that in a common high-dimensional covariance model, the choice of loss function has a profound effect on optimal estimation.

In an asymptotic framework based on the spiked covariance model and use of orthogonally invariant estimators, we show that optimal estimation of the population covariance matrix boils down to design of an optimal shrinker $\eta$ that acts elementwise on the sample eigenvalues. Indeed, to each loss function there corresponds a unique admissible eigenvalue shrinker $\eta^{*}$ dominating all other shrinkers. The shape of the optimal shrinker is determined by the choice of loss function and, crucially, by inconsistency of both eigenvalues and eigenvectors of the sample covariance matrix.

Details of these phenomena and closed form formulas for the optimal eigenvalue shrinkers are worked out for a menagerie of 26 loss functions for covariance estimation found in the literature, including the Stein, Entropy, Divergence, Fréchet, Bhattacharya/Matusita, Frobenius Norm, Operator Norm, Nuclear Norm and Condition Number losses.

Citation

Download Citation

David Donoho. Matan Gavish. Iain Johnstone. "Optimal shrinkage of eigenvalues in the spiked covariance model." Ann. Statist. 46 (4) 1742 - 1778, August 2018. https://doi.org/10.1214/17-AOS1601

Information

Received: 1 March 2014; Revised: 1 May 2017; Published: August 2018
First available in Project Euclid: 27 June 2018

zbMATH: 06936477
MathSciNet: MR3819116
Digital Object Identifier: 10.1214/17-AOS1601

Subjects:
Primary: 62C20 , 62H25
Secondary: 90C22 , 90C25

Keywords: Bhattacharya/Matusita affinity , condition number loss , Covariance estimation , divergence loss , entropy loss , Fréchet distance , high-dimensional ssymptotics , optimal shrinkage , spiked covariance , Stein loss

Rights: Copyright © 2018 Institute of Mathematical Statistics

Vol.46 • No. 4 • August 2018
Back to Top