Bandit Algorithms

(Jeff_L) #1
5.3 The Cramer-Chernoff method and subgaussian random variables 75

A similar inequality holds for the left tail. By using the union bound
P(A∪B)≤P(A)+P(B)we also find thatP(|X|≥ε)≤ 2 exp(−ε^2 /(2σ^2 )).
An equivalent form of these bounds is:


P

(


X≥



2 σ^2 log(1/δ)

)


≤δ P

(


|X|≥



2 σ^2 log(2/δ)

)


≤δ.

This form is often more convenient and especially the latter, which for smallδ
shows that with overwhelming probabilityXtakes values in the interval
(



2 σ^2 log(2/δ),


2 σ^2 log(2/δ)

)


.


To study the tail behavior of ˆμ−μ, we need one more lemma.


Lemma5.4. Suppose thatXisσ-subgaussian andX 1 andX 2 are independent
andσ 1 andσ 2 -subgaussian respectively, then:


(a)E[X] = 0andV[X]≤σ^2.
(b)cXis|c|σ-subgaussian for al lc∈R.
(c)X 1 +X 2 is



σ 12 +σ^22 -subgaussian.

The proof of the lemma is left to the reader (Exercise 5.7). Combining
Lemma 5.4 and Theorem 5.3 leads to a straightforward bound on the tails
of ˆμ−μ.

Corollary5.5.Assume thatXi−μare independent,σ-subgaussian random
variables. Then for anyε≥ 0


P(ˆμ≥μ+ε)≤exp

(


−nε

2
2 σ^2

)


and P(ˆμ≤μ−ε)≤exp

(


−nε

2
2 σ^2

)


,


whereμˆ=^1 n


∑n
t=1Xt.
Proof By Lemma 5.4 it holds thatμˆ−μ=

∑n
i=1(Xi−μ)/nisσ/


n-subgaussian.
Then apply Theorem 5.3.


Forx > 0 it holds thatexp(−x)≤ 1 /(ex), which shows that the above
inequality is stronger than what we obtained via Chebyshev’s inequality except
whenεis very small. It is exponentially smaller ifnε^2 is large relative toσ^2. The
deviation form of the above result says that under the conditions of the result,
for anyδ∈[0,1], with probability at least 1−δ,


μ≤μˆ+


2 σ^2 log(1/δ)
n

. (5.6)


Symmetrically, it also follows that with probability at least 1−δ,

μ≥μˆ−


2 σ^2 log(1/δ)
n. (5.7)

Again, one can use a union bound to derive a two-sided inequality.


Example5.6.The following random variables are subgaussian:
Free download pdf