Académique Documents
Professionnel Documents
Culture Documents
Suppose we have an unknown population parameter, such as a population mean μ or a population proportion p, which we'd
like to estimate. We can't possibly survey the entire population. So, of course, we do what comes naturally and take a random
sample from the population, and use the resulting data to estimate the value of the population parameter. Of course, we want
the estimate to be "good" in some way.
Point estimation, is the process of finding an approximate value of some parameter – such as the mean (average) of a
population from random samples of the population.
UNBIASED ESTIMATOR
In statistics, estimation (or inference) refers to the process by which one makes inferences (e.g. draws conclusions) about a
population, based on information obtained from a sample.
A statistic is any measurable quantity calculated from a sample of data (e.g. the average). This is a stochastic variable as, for
a given population, it will in general vary from sample to sample.
An estimator is any quantity calculated from the sample data which is used to give information about an unknown quantity in
the population (the estimand).
An estimate is the particular value of an estimator that is obtained by a particular sample of data and used to indicate the
value of a parameter.
Example:
Definition
The bias of an estimator is the difference between the expectation value over its PDF (i.e. its mean value) and the population
value.
An unbiased estimator is an accurate statistic that’s used to approximate a population parameter. “Accurate” in this sense
means that it’s neither an overestimate nor an underestimate.
This essentially means that it is an unbiased estimator when the mean of the statistic’s sampling distribution is equal to the
population’s parameter.
Example:
1. Show that the sample mean is an unbiased estimator of a population mean. E(x-bar)= myu 𝐸(𝑋̅) = 𝜇
2. Show that the sample variance is an unbiased estimator of the population variance. 𝐸(𝑠 2 ) = 𝜎 2
Useful properties:
𝑬 ∑ 𝑿𝒊 = ∑ 𝑬(𝑿𝒊 )
𝑬(𝒄𝑿) = 𝒄𝑬(𝑿)
𝑬 (𝑪) = 𝑪
∑ 𝑪 = 𝒏𝑪
𝒊=𝟏
𝟐
̅̅̅̅𝟐 ) = 𝝈 + 𝝁𝟐
𝑬(𝑿
𝒏
Example:
Let X is a random variable with E(X)=100 and V(X)=15. Find the following
a. 𝑉(3𝑋 + 10)
b. 𝑉(−2𝑋)
Goodness of an Estimator
Three of the measures that we will use to assess the goodness of an estimator are its bias, its mean-square error and its standard
error.
1. If 𝜃̂ is an estimator of 𝜃, then the bias of 𝜃 is given by
𝐵(𝜃̂) = 𝐸(𝜃̂) − 𝜃
2. Mean-square error of 𝜃̂ is given by
2
𝑀𝑆𝐸(𝜃̂) = 𝐸(𝜃̂ − 𝜃)
2
𝑀𝑆𝐸(𝜃̂) = 𝑉𝑎𝑟(𝜃̂) + [𝐵(𝜃̂)]
3. If 𝜃̂ is an estimator of 𝜃, then the standard error of 𝜃̂ is simply its standard deviation.
𝑆𝐸(𝜃̂) = √𝑉𝑎𝑟(𝜃̂)
STATISTICAL INTERVALS
A confidence interval estimate for 𝜇 is an interval of the form 𝑙 ≤ 𝜇 ≤ 𝑢, where the end-points 𝑙 and 𝑢 are computed from
the sample data. Because different samples will produce different values of 𝑙 and 𝑢, these end-points are values of random
variables L and U, respectively. Supposed that we can determine values of L and U such that the following probability
statement is true:
𝑃{𝐿 ≤ 𝜇 ≤ 𝑈} = 1 − 𝛼
where 0 ≤ 𝛼 ≤ 1. There is a probability of 1 − 𝛼 of selecting a sample for which the confidence interval will contain the true
value 𝜇.
Example:
Ten measurements of impact energy (J) on specimens of A238 steel cut at 60 degree Celsius are as follows 64.1, 64.7,64.5,
64.6, 64.5, 64.3, 64.6, 64.8, 64.2, and 64.3. Assume that impact energy is normally distributed with 𝜎 = 1𝐽. We want to find a
95% confidence interval for 𝜇, the mean impact energy.
Example: Using the same example above. Find the lower one-sided 95% confidence interval for the mean impact energy.