30.5 PROPERTIES OF DISTRIBUTIONS
|x−μ|≥c. From (30.48), we find that
σ^2 ≥∫|x−μ|≥c(x−μ)^2 f(x)dx≥c^2∫|x−μ|≥cf(x)dx. (30.49)The first inequality holds because both (x−μ)^2 andf(x) are non-negative for
allx, and the second inequality holds because (x−μ)^2 ≥c^2 over the range of
integration. However, the RHS of (30.49) is simply equal toc^2 Pr(|X−μ|≥c),
and thus we obtain the required inequality
Pr(|X−μ|≥c)≤σ^2
c^2.A similar derivation may be carried through for the case of a discrete random
variable. Thus, foranydistributionf(x) that possesses a variance we have, for
example,
Pr(|X−μ|≥ 2 σ)≤1
4and Pr(|X−μ|≥ 3 σ)≤1
9.30.5.4 MomentsThe mean (or expectation) ofXis sometimes called thefirst momentofX,since
it is defined as the sum or integral of the probability density function multiplied
by the first power ofx. By a simple extension thekth moment of a distribution
is defined by
μk≡E[Xk]={∑
jxk
jf(xj) for a discrete distribution,
∫
xkf(x)dx for a continuous distribution. (30.50)For notational convenience, we have introduced the symbolμkto denoteE[Xk],
thekth moment of the distribution. Clearly, the mean of the distribution is then
denoted byμ 1 , often abbreviated simply toμ, as in the previous subsection, as
this rarely causes confusion.
A useful result that relates the second moment, the mean and the variance ofa distribution is proved using the properties of the expectation operator:
V[X]=E[
(X−μ)^2]=E[
X^2 − 2 μX+μ^2]=E[
X^2]
− 2 μE[X]+μ^2=E[
X^2]
− 2 μ^2 +μ^2
=E[
X^2]
−μ^2. (30.51)In alternative notations, this result can be written
〈(x−μ)^2 〉=〈x^2 〉−〈x〉^2 or σ^2 =μ 2 −μ^21.