Biased and Unbiased Point Estimates

Carson West

AP Stats Home

Biased and Unbiased Point Estimates

In statistics, we often want to estimate unknown population parameters (like the population mean $ \mu $ or population proportion $ p $ ) using data from a sample. A point estimate is a single value, calculated from sample data, that is used to estimate an unknown population parameter.

Unbiased Point Estimates

An estimator is considered unbiased if its sampling distribution has a mean that is equal to the true value of the parameter being estimated. In other words, an unbiased estimator does not systematically over- or under-estimate the true parameter over many samples. If we were to take an infinite number of samples and calculate the estimate for each, the average of those estimates would equal the true population parameter.

For an estimator $ \hat{\theta} $ of a parameter $ \theta $ , it is unbiased if:

$$ E[\hat{\theta}] = \theta $$
Where $ E[\hat{\theta}] $ represents the expected value (mean) of the sampling distribution of the estimator.

Examples of Unbiased Estimators:

Biased Point Estimates

An estimator is considered biased if its sampling distribution has a mean that is not equal to the true value of the parameter being estimated. This means the estimator consistently tends to over-estimate or under-estimate the true parameter. The bias of an estimator $ \hat{\theta} $ is defined as the difference between its expected value and the true parameter value:

$$ Bias(\hat{\theta}) = E[\hat{\theta}] - \theta $$
If $ Bias(\hat{\theta}) \neq 0 $ , the estimator is biased.

Examples of Biased Estimators:

Bias vs. Variability

It’s crucial to distinguish between bias and variability (or precision).

Potential Problems with Sampling can lead to both biased estimates (if the sampling method is flawed) and high variability (if the sample size is too small or the population is very diverse).

Consider a dartboard analogy:

Mean Squared Error (MSE)

Sometimes, we prefer a slightly biased estimator if it has significantly lower variability, leading to overall better performance. The Mean Squared Error (MSE) is a measure that combines both bias and variability:

$$ MSE(\hat{\theta}) = E[(\hat{\theta} - \theta)^2] = Bias(\hat{\theta})^2 + Var(\hat{\theta}) $$
Where $ Var(\hat{\theta}) $ is the variance of the estimator. An estimator with a smaller MSE is generally preferred, even if it has a small bias.

Summary Table

Estimator Parameter Estimated Unbiased? Notes Related Topic
Sample Mean $ \bar{x} $ Population Mean $ \mu $ Yes Average of sample means equals population mean Sampling Distributions for Sample Means
Sample Proportion $ \hat{p} $ Population Proportion $ p $ Yes Average of sample proportions equals population proportion Sampling Distributions for Sample Proportions
Sample Variance $ s_x^2 $ Population Variance $ \sigma^2 $ Yes (when using $ n-1 $ denominator) Using $ n $ in the denominator results in a biased (underestimated) variance Summary Statistics for a Quantitative Variable
Sample Standard Deviation $ s_x $ Population Standard Deviation $ \sigma $ No (generally biased) Tends to slightly underestimate $ \sigma $ , especially for small $ n $ Describing the Distribution of a Quantitative Variable