5.1. Convergence in Probability 3250 500 1000 1500 2000−0.2−0.10.00.10.20.3nxbarFigure 5.1.1:Realizations of the point estimatorXfor samples of size 10 to 2000
in steps of 10 which are drawn from aN(0,1) distribution.
sample variance is an unbiased estimator ofσ^2. We now show that it is a consistent
estimator ofσ^2. Recall Theorem 5.1.1 which shows thatXn
P
→μ. To show that the
sample variance converges in probability toσ^2 , assume further thatE[X 14 ]<∞,so
that Var(S^2 )<∞. Using the preceding results, we can show the following:
S^2 n=1
n− 1∑ni=1(Xi−Xn)^2 =n
n− 1(
1
n∑ni=1Xi^2 −X
2
n)→P 1 ·[E(X^2
1 )−μ(^2) ]=σ (^2).
Hence the sample variance is a consistent estimator ofσ^2. From the discussion
above, we have immediately thatSn
P
→σ; that is, the sample standard deviation is
a consistent estimator of the population standard deviation.
Unlike the last example, sometimes we can obtain the convergence by using the
distribution function. We illustrate this with the following example:
Example 5.1.2(Maximum of a Sample from a Uniform Distribution). Suppose
X 1 ,...,Xnis a random sample from a uniform(0,θ) distribution. Supposeθis
unknown. An intuitive estimate ofθis the maximum of the sample. LetYn=
max{X 1 ,...,Xn}. Exercise 5.1.4 shows that the cdf ofYnis
FYn(t)=
⎧
⎨
⎩
(^1) ( t>θ
t
θ
)n
0 <t≤θ
0 t≤ 0.
(5.1.1)