Académique Documents
Professionnel Documents
Culture Documents
2&3-1
2&3-2
2&3-7
2&3-8
2&3-9
Small
657.4
19.4
238
Large
650.0
17.9
182
1. Estimation
Ysmall Ylarge = 657.4 650.0 = 7.4
where Ysmall =
nsmall
nsmall
and Ylarge =
i =1
1
nlarge
nlarge
i =1
th
2. Hypothesis testing
Difference-in-means test: compute the t-statistic,
t=
Ys Yl
ss2
ns
sl2
nl
Ys Yl
=
SE (Ys Yl )
(remember this?)
Y
(etc.)
districts; and ss2 =
i
s
ns 1 i =1
2&3-12
sY
small
657.4
19.4
238
large
650.0
17.9
182
t=
Ys Yl
ss2
ns
sl2
nl
657.4 650.0
19.42
238
17.92
182
7.4
=
= 4.05
1.83
|t| > 1.96, so reject (at the 5% significance level) the null
hypothesis that the two means are the same.
2&3-13
3. Confidence interval
A 95% confidence interval for the difference between the
means is,
(Ys Yl ) 1.96SE(Ys Yl )
= 7.4 1.961.83 = (3.8, 11.0)
Two equivalent statements:
1. The 95% confidence interval for doesnt include 0;
2. The hypothesis that = 0 is rejected at the 5% level.
2&3-14
2&3-15
Random variable Y
Numerical summary of a random outcome
Here, the numerical value of district average test
scores (or district STR), once we choose a
year/district to sample.
Population distribution of Y
The probabilities of different values of Y that occur in
the population, for ex. Pr[Y = 650] (when Y is
discrete)
or: The probabilities of sets of these values, for ex.
Pr[Y > 650] (when Y is continuous).
2&3-17
Conditional distributions
The distribution of Y, given value(s) of some other
random variable, X
Ex: the distribution of test scores, given that STR < 20
Moments of conditional distributions
conditional mean = mean of conditional distribution
= E(Y|X = x) (important notation)
conditional variance = variance of conditional
distribution
Example: E(Test scores|STR < 20), the mean of test
scores for districts with small class sizes
2&3-19
2&3-22
2&3-23
1.
2.
3.
4.
2&3-27
2&3-29
2&3-30
Variance:
= E (Yi Y )
n i =1
1 n n
= E 2 (Yi Y )(Y j Y )
n i =1 j =1
2&3-31
so
1 n n
var(Y ) = 2 E (Yi Y )(Y j Y )
n i =1 j =1
1 n n
= 2 cov(Yi ,Y j )
n i =1 j =1
1 n
1 n n
= 2 var(Yi ) + 2 cov(Yi ,Y j )
n i =1
n i =1 j =1, j i
1 n 2
= 2 Y +0
n i =1
Y2
n
2&3-32
Y2
n
Implications:
Y is an unbiased estimator of Y (that is, E(Y ) = Y)
var(Y ) is inversely proportional to n
spread of sampling distribution is proportional to
1/ n
in this sense, the sampling uncertainty arising from
using Y to make inferences about Y is proportional to
1/ n
2&3-33
2&3-35
Y2
n
2&3-39
Y E (Y )
var(Y )
2&3-40
Y E (Y )
var(Y )
2&3-41
consistency: Y Y
Y is the least squares estimator of Y; Y solves,
n
1.
2.
3.
4.
= PrH [|
0
Y Y ,0
Y / n
|>|
Y act Y ,0
Y / n
|]
Y
i
n 1 i =1
2
Y
Fact:
4
2&3-47
= PrH [|
0
PrH [|
0
Y Y ,0
Y / n
Y Y ,0
sY / n
|>|
|>|
Y act Y ,0
Y / n
Y act Y ,0
sY / n
|]
|]
(large n)
2&3-48
2&3-49
1.
2.
3.
4.
2&3-52
Summary:
From the assumptions of:
(1) simple random sampling of a population, that is,
{Yi, i =1,,n} are i.i.d.
(2) 0 < E(Y4) <
we developed, for large samples (large n):
Theory of estimation (sampling distribution of Y )
Theory of hypothesis testing (large-n distribution of tstatistic and computation of the p-value)
Theory of confidence intervals (constructed by
inverting test statistic)
Are assumptions (1) & (2) plausible in practice? Yes
2&3-53
But this is the slope of a line relating test score and STR
So somehow we need to estimate this slope
2&3-54