Open Access
February 2018 Pathwise coordinate optimization for sparse learning: Algorithm and theory
Tuo Zhao, Han Liu, Tong Zhang
Ann. Statist. 46(1): 180-218 (February 2018). DOI: 10.1214/17-AOS1547

Abstract

The pathwise coordinate optimization is one of the most important computational frameworks for high dimensional convex and nonconvex sparse learning problems. It differs from the classical coordinate optimization algorithms in three salient features: warm start initialization, active set updating and strong rule for coordinate preselection. Such a complex algorithmic structure grants superior empirical performance, but also poses significant challenge to theoretical analysis. To tackle this long lasting problem, we develop a new theory showing that these three features play pivotal roles in guaranteeing the outstanding statistical and computational performance of the pathwise coordinate optimization framework. Particularly, we analyze the existing pathwise coordinate optimization algorithms and provide new theoretical insights into them. The obtained insights further motivate the development of several modifications to improve the pathwise coordinate optimization framework, which guarantees linear convergence to a unique sparse local optimum with optimal statistical properties in parameter estimation and support recovery. This is the first result on the computational and statistical guarantees of the pathwise coordinate optimization framework in high dimensions. Thorough numerical experiments are provided to support our theory.

Citation

Download Citation

Tuo Zhao. Han Liu. Tong Zhang. "Pathwise coordinate optimization for sparse learning: Algorithm and theory." Ann. Statist. 46 (1) 180 - 218, February 2018. https://doi.org/10.1214/17-AOS1547

Information

Received: 1 August 2016; Revised: 1 January 2017; Published: February 2018
First available in Project Euclid: 22 February 2018

zbMATH: 06865109
MathSciNet: MR3766950
Digital Object Identifier: 10.1214/17-AOS1547

Subjects:
Primary: 62F30 , 90C26
Secondary: 62J12 , 90C52

Keywords: active set , global linear convergence , Nonconvex sparse learning , optimal statistical rates of convergence , oracle property , pathwise coordinate optimization , strong rule

Rights: Copyright © 2018 Institute of Mathematical Statistics

Vol.46 • No. 1 • February 2018
Back to Top