Open Access
Translator Disclaimer
March 2016 Recursive Learning for Sparse Markov Models
Jie Xiong, Väinö Jääskinen, Jukka Corander
Bayesian Anal. 11(1): 247-263 (March 2016). DOI: 10.1214/15-BA949


Markov chains of higher order are popular models for a wide variety of applications in natural language and DNA sequence processing. However, since the number of parameters grows exponentially with the order of a Markov chain, several alternative model classes have been proposed that allow for stability and higher rate of data compression. The common notion to these models is that they cluster the possible sample paths used to predict the next state into invariance classes with identical conditional distributions assigned to the same class. The models vary in particular with respect to constraints imposed on legitime partitions of the sample paths. Here we consider the class of sparse Markov chains for which the partition is left unconstrained a priori. A recursive computation scheme based on Delaunay triangulation of the parameter space is introduced to enable fast approximation of the posterior mode partition. Comparisons with stochastic optimization, k-means and nearest neighbor algorithms show that our approach is both considerably faster and leads on average to a more accurate estimate of the underlying partition. We show additionally that the criterion used in the recursive steps for comparison of triangulation cell contents leads to consistent estimation of the local structure in the sparse Markov model.


Download Citation

Jie Xiong. Väinö Jääskinen. Jukka Corander. "Recursive Learning for Sparse Markov Models." Bayesian Anal. 11 (1) 247 - 263, March 2016.


Published: March 2016
First available in Project Euclid: 15 April 2015

zbMATH: 1359.62269
MathSciNet: MR3447098
Digital Object Identifier: 10.1214/15-BA949

Keywords: clustering , Delaunay triangulation , recursive learning , sequence analysis , sparse Markov chains

Rights: Copyright © 2016 International Society for Bayesian Analysis


Vol.11 • No. 1 • March 2016
Back to Top