Hoeffding's lemma for Markov Chains and its applications to statistical learning

02/01/2018
by   Jianqing Fan, et al.
0

We establish the counterpart of Hoeffding's lemma for Markov dependent random variables. Specifically, if a stationary Markov chain {X_i}_i > 1 with invariant measure π admits an L_2(π)-spectral gap 1-λ, then for any bounded functions f_i: x [a_i,b_i], the sum of f_i(X_i) is sub-Gaussian with variance proxy 1+λ/1-λ·∑_i (b_i-a_i)^2/4. The counterpart of Hoeffding's inequality immediately follows. Our results assume none of reversibility, countable state space and time-homogeneity of Markov chains. They are optimal in terms of the multiplicative coefficient (1+λ)/(1-λ), and cover Hoeffding's lemma and inequality for independent random variables as special cases with λ = 0. We illustrate the utility of these results by applying them to six problems in statistics and machine learning. They are linear regression, lasso regression, sparse covariance matrix estimation with Markov-dependent samples; Markov chain Monte Carlo estimation; respondence driven sampling; and multi-armed bandit problems with Markovian rewards.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset
Success!
Error Icon An error occurred

Sign in with Google

×

Use your Google Account to sign in to DeepAI

×

Consider DeepAI Pro