Open Access
February, 1967 Some Invariance Principles for Functionals of a Markov Chain
David A. Freedman
Ann. Math. Statist. 38(1): 1-7 (February, 1967). DOI: 10.1214/aoms/1177699053

Abstract

The new results are (C), (D) and (F) below. Let $X_0, X_1, \cdots$ be a Markov chain with countable state space $I$ and stationary transitions. Suppose $I$ is a positive recurrent class, with stationary probability vector $p$. Let $f$ be a real-valued function on $I$. Fix a reference state $s \varepsilon I$, and let $0 \leqq t_1 < t_2 < \cdots$ be the times $n$ at which $X_n = s$. Let $Y_j = \sum \{f(X_n):t_j \leqq n < t_{j + 1}\}$ and $U_j = \sum \{|f(X_n)|:t_j \leqq n < t_{j + 1}\}.$ Let $V_m = \sum^m_{j = 1} Y_j$ and $S_n = \sum^n_{j = 0}f(X_j)$. For (C) and (D) below, assume (A) $\sum_{i \varepsilon I} p_if(i) = 0$; and (B) $U^2_j$ has finite expectation. Then: (C) Theorem. $n^{-\frac{1}{2}} \max \{|S_j - V_{jp_s}|: 1 \leqq j \leqq n\} \rightarrow 0$ in probability; and (D) Theorem. $(n \log \log n)^{-\frac{1}{2}} \max \{|S_j - V_{jp_s}|: 1 \leqq j \leqq n\} \rightarrow 0$ almost everywhere. For (F), do not assume (A) and (B), but assume (E) $Y_j$ differs from 0 with positive probability. Let $v_m$ (respectively, $s_n$) be 1 or 0 according as $V_m$ (respectively, $S_n$) is positive or non-positive. Then (F) Theorem. $n^{-1} \sum \{s_j : 1 \leqq j \leqq n\} - p^{-1}_sn^{-1} \sum \{v_j : 1 \leqq j \leqq np_s\} \rightarrow 0$ almost everywhere. I do not believe the convergence in (C) is a.e., but have no counter-example.

Citation

Download Citation

David A. Freedman. "Some Invariance Principles for Functionals of a Markov Chain." Ann. Math. Statist. 38 (1) 1 - 7, February, 1967. https://doi.org/10.1214/aoms/1177699053

Information

Published: February, 1967
First available in Project Euclid: 27 April 2007

zbMATH: 0233.60054
MathSciNet: MR211464
Digital Object Identifier: 10.1214/aoms/1177699053

Rights: Copyright © 1967 Institute of Mathematical Statistics

Vol.38 • No. 1 • February, 1967
Back to Top