In this post, I would like to write about Sufficiency. Here is the plan for this post.
Ch 7. Sufficiency
1. Measures of Quality of Estimators
Consider a point estimator \(Y_n = u(X_1, \cdots, X_n)\) based on a sample \(X_1, \cdots, X_n\). There is a several properties of point estimators:
\(Y_n\) is a consistent estimator of \(\theta\) if \(Y_n\) converges to \(\theta\) in probability; i.e., \(Y_n\) is close to \(\theta\) for large sample sizes.
\(Y_n\) is an unbiased estimator of \(\theta\) if \(E(Y_n\)) = \(\theta\). Note that maximum likelihood estimators may not be unbiased, although generally they are asymptotically unbiased.
2. A Sufficient Statistics for a Parameter
Definition 2 Let \(X_1, X_2, \cdots, X_n\) denote a random sample of size \(n\) from a distribution that has pdf or pmf \(f(x;\theta), \theta \in \Omega\). Let \(Y_1 = u_1(X_1, X_2, \cdots, X_n)\) be a statistic whose pdf or pmf is \(f_{Y_1}(y_1; \theta)\). Then \(Y_1\) is a sufficient statistic for \(\theta\) if and only if
\[\frac{f(x_1;\theta)f(x_2;\theta) \cdots f(x_n;\theta)}{f_{Y_1}[u_1(x_1,x_2,\cdots,x_n);\theta]} = H(x_1, x_2, \cdots, x_n) \]
where \(H(x_1, x_2, \cdots, x_n)\) does not depend upon \(\theta \in \Omega\).References:
- Introduction to Mathematical Statistics, Robert V. Hogg, Joeseph McKean, Allen T. Craig, Seventh Edition