Open Access
2018 Online natural gradient as a Kalman filter
Yann Ollivier
Electron. J. Statist. 12(2): 2930-2961 (2018). DOI: 10.1214/18-EJS1468

Abstract

We cast Amari’s natural gradient in statistical learning as a specific case of Kalman filtering. Namely, applying an extended Kalman filter to estimate a fixed unknown parameter of a probabilistic model from a series of observations, is rigorously equivalent to estimating this parameter via an online stochastic natural gradient descent on the log-likelihood of the observations.

In the i.i.d. case, this relation is a consequence of the “information filter” phrasing of the extended Kalman filter. In the recurrent (state space, non-i.i.d.) case, we prove that the joint Kalman filter over states and parameters is a natural gradient on top of real-time recurrent learning (RTRL), a classical algorithm to train recurrent models.

This exact algebraic correspondence provides relevant interpretations for natural gradient hyperparameters such as learning rates or initialization and regularization of the Fisher information matrix.

Citation

Download Citation

Yann Ollivier. "Online natural gradient as a Kalman filter." Electron. J. Statist. 12 (2) 2930 - 2961, 2018. https://doi.org/10.1214/18-EJS1468

Information

Received: 1 June 2017; Published: 2018
First available in Project Euclid: 18 September 2018

zbMATH: 06942962
MathSciNet: MR3855360
Digital Object Identifier: 10.1214/18-EJS1468

Subjects:
Primary: 65K10 , 68T05
Secondary: 49M15 , 90C26 , 93E11 , 93E35

Keywords: Kalman filter , natural gradient , Statistical learning , Stochastic gradient descent

Vol.12 • No. 2 • 2018
Back to Top