Reference textbook High-Dimensional Statistics: A Non-Asymptotic Viewpoint, Martin J. Wainwright
Integral identity: For X to be a non-negative R.V.
EX=E∫0∞1t<xdt=∫0∞E1t<xdt=∫0∞P(X>t)dt Markov's Inequality:For non-negative R.V. and t>0
EX=EX1X≥t+EX1x≤t≥Et1X≥t=tP(X≥t) Chebyshev's inequality: X:=∣X−μ∣2
Generalization: X:=∣X−μ∣k
P[∣X−μ∣≥t]≤tkE[∣X−μ∣k] Chernoff Bound#
Chernoff bound: X:=eλ(X−μ)
P[(X−μ)≥t]=P[eλ(X−μ)≥eλt]≤eλtE[eλ(X−μ)]logP[(X−μ)≥t]≤λ∈[0,b]inf{logE[eλ(X−μ)]−λt} k=0,1,2,..infδkE[∣X∣k]≤λ>0infeλδE[eλX].
Hoeffding's inequality of Rademacher RV#
Hoeffding's inequality for n independent Rademacher R.V.s, vector a∈Rn and t≥0:
P{i=1∑naiXi≥t}≤exp(−2∥a∥22t2) Proof:
P{i=1∑NaiXi≥t}=P{exp(λi=1∑NaiXi)≥exp(λt)}≤e−λtEexp(λi=1∑NaiXi)=e−λti=1∏NEexp(λaiXi) For Eexp(λaiXi)=2exp(λai)+exp(−λai)=cosh(λai) and cosh(x)≤exp(x2/2) for all x∈R
P{i=1∑NaiXi≥t}≤e−λti=1∏Nexp(λ2ai2/2)=exp(−λt+2λ2i=1∑Nai2)=exp(−λt+2λ2)( make λ=t) Chernoff's inequality for sum of Bernoulli rv#
P{SN≥t}≤e−μ(teμ)t