9.2.4 Sufficiency
Let X 1 ,X 2 ,...,Xn be a sample of a population X the distribution of which
depends on unknown parameter. If )i sastatisticsuch
that, for any other statistic
the conditional distribution of Z, given that Y =y does not depend on , then
Y is called a sufficient statistic for. If also , then Y is said to be a
sufficient estimator for.
In words, the definition for sufficiency states that, if Y is a sufficient statistic
for , all sample information concerning is contained in Y. A sufficient
statistic is thus of interest in that if it can be found for a parameter then an
estimator based on this statistic is able to make use of all the information that
the sample contains regarding the value of the unknown parameter. Moreover,
an important property of a sufficient estimator is that, starting with any
unbiased estimator of a parameter that is not a function of the sufficient
estimator, it is possible to find an unbiased estimator based on the sufficient
statistic that has a variance smaller than that of the initial estimator. Sufficient
estimators thus have variances that are smaller than any other unbiased esti-
mators that do not depend on sufficient statistics.
If a sufficient statistic for a parameter exists, Theorem 9.4, stated here
without proof, provides an easy way of finding it.
Theorem 9. 4: Fisher – N ey ma n f a ct o riz a t io n crit erio n.Let
be a statistic based on a sample of size n. Then Y is a sufficient statistic for
if and only if the jo int probability density function of X 1 ,X 2 ,..., and
can be factorized in the form
If X is discrete, we have
The sufficiency of the factorization criterion was first pointed out by Fisher
(1922). N eyman (1935) showed that it is also necessary.
Parameter Estimation 275
YhX 1 ,X 2 ,...,Xn
Zg
X 1 ;X 2 ;...;Xn;
EfYg
YhX 1 ,X 2 ,...,Xn)
Xn,fXx 1 ;)fXxn;),
Yn
j 1
fX
xj;g 1 h
x 1 ;...;xn;g 2
x 1 ;...;xn:
9 : 49
Yn
j 1
pX
xj;g 1 h
x 1 ;...;xn;g 2
x 1 ;...;xn:
9 : 50