Académique Documents
Professionnel Documents
Culture Documents
9/11/2010
In this homework we'll use the notation used by Hayashi [?] in his book.
Specically we have :
xi1 β1 y1
xi = ... , β = ... , y = ...
xiK βK yn
0
1 x1 x1 1 ... x1 K
.. .. .. .
.
ε = . ,X = . = .
... .
n x0 n xn 1 ... xn K
1 Theoretical Exercises
1.1 Chapter 1 : Small Sample
Question 1
N
!1/p N
!1/p
X X
p p
|ei | = (|yi − xi β̃|)
i=1 i=1
N
!1/p N
!1/p
X X
min |ei |p = min (|yi − xi β̃|)p
β̃ β̃
i=1 i=1
1
N
X N
X
2
min |ei | = min (|yi − xi b|)2
β̃ β̃
i=1 i=1
0
= min e e
β̃
= min(y − x0 i b)0 (y − x0 i b)
β̃
2
We prefer the quadratic form, i.e. OLS estimation, because :
First, it leads to the only ecient unbiased estimators (in the Gauss-Markov
sens, i.e. the b regressor is BLUE) among all the linear estimation meth-
ods, and
From this last expression we see that DXβ = 0 because we have assumed
that β̂ was also unbiased.
Therefore, from (1) :
β̂ = Dε + b
β̂ − β = Dε + (b − β) (we subtract −β )
= Dε + (A(Xβ + ε) − β)
= Dε + (X0 X)−1 X0 Xβ + Aε − β
= Dε + Aε
= (D + A)ε
We want to determine wether var(β̂ − β|X) is larger than var(b − β|X) (so
the OLS estimator is indeed ecient) or not, therefore we calculate :
3
Before continuing we point out that :
0
DA0 = D (X0 X)−1 X0
b − β = Ay − β
= (X0 X)−1 X(Xβ + ε) − β
= β + Aε − β
= Aε
Therefore,
σ 2 (X0 X)−1 = var(b − β|x) < var(β̂ − β|X) = σ 2 (DD0 + (X0 X)−1 )
This concludes the proof : the OLS estimator b is more ecient than the
innity-norm estimator β̂ (if it exists).
Question 2
y = Xβ + ε
4
y = Xb + e
We want to demostrate the equality
e0 e = ε0 M ε
where M = In − X(X0 X)−1 X0 .
0 −1 0
We know that e = y − Xb, by substituting b (b = (X X) X y):
e = y − Xb = y − X(X0 X)−1 X0 y
= (In − X(X0 X)−1 X0 )y
= My
M y = y − Xb
= y − X(X0 X)−1 X0 (Xβ + ε)
= y − Xβ − X(X0 X)−1 X0 ε
Here, X(X0 X)−1 X0 X = X, because (X0 X)−1 (X0 X) = In . Also, X(X0 X)−1 X0
is the so called projection matrix, P which has the property that P X = X.
Then we substitute y − Xβ for ε and get:
= ε − X(X0 X)−1 X0 ε
= (In − X(X0 X)−1 X0 )ε
= Mε
e0 e = e0 M ε
= (M ε)0 M ε
= ε0 M 0 M ε
= ε0 M ε
5
Because (X(X0 X)−1 X0 )0 = X(X0 X)−1 X0 . Furthermore, (X0 X)−1 X0 X can-
cels out. We are left with:
Question 3
• First, we will recall why does q follow a chi-square distribution and link
it with its components, i.e. the errors terms.
E(ε) = E (E(ε|X))
= E(0)
=0
ε
Therefore ε ∼ N (0, σ 2 In ), which means that
σ ∼ N (0, In ). And because
2
M is idempotent and its rank is equal to N −K we have that q ∼ χ (N − K),
i.e. the marginal distribution of q is also chi-squared.
Question 4
e0 e
s2 =
n−K
2σ 4
We want to show that V ar(s2 |X) = n−K . We have already shown that
0
2 ee 2 2
q|X ∼ χ (n−K). Observe that q = σ 2 = (n−k)s /σ . Therefore, (n−k)s2 /σ 2
2
has a χ distribution with (n − K) degrees of freedom.
6
n−K 2
V ar s = 2(n − K)
σ2
The variance of the χ2 distribution with (n − K) degrees of freedom is
2(n-K).
−2
n−K
V ar(s2 ) = 2(n − k)
σ2
2σ 4
=
n−K
The Cramer-Rao bound for the variance term is 2σ 4 /n. The OLS estimator
2
does not reach this bound, but no unbiased estimator of σ reach this bound.
Question 1
We'll rst point out that the denition of this random experiment satises
one important axiom : the joint probability of every event, i.e. the sample
space probability, is equal to 1.
This was possible because these three events are mutually exclusive.
As convergence in Mean Squares implies convergence in probability, we will
show the former not to converge neither to α nor 0 concluding that zn doesn't
converge in probability neither.
As a reminder, if lim E[(zn − α)2 ] = 0, then zn →M S α.
n→∞
This last result was possible because the expectation is a linear operator
and neither α nor 2 are random variables. Next we have :
7
n−c c−1 αn − αc − nc + n
E(zn ) = α +n =
n n n
and
n−c c−1 α 2 − α 2 c + n2 c − n2
E(zn2 ) = α2 + n2 =
n n n
therefore
α 2 − α 2 c + n2 c − n2 αn − αc − nc + n
= − 2α + α2
n n
α2 − α2 c + n2 c − n2 − 2α2 n + 2α2 c + 2αnc − 2αn + α2 n
=
n
α2 α 2 c n2 c n2 α2 n 2αnc 2αn
= + + − − + −
n n n n n n n
α2 + α2 c
= + n(c − 1) − α(α − 2c + 2)
n
So the limit of this expectation as n goes to innity is :
α2 + α2 c
lim E[(zn2 − α)2 ] = lim + n(c − 1) − α(α − 2c + 2)
n→∞ n→∞ n
=∞
Then we can conclude that the {sn } doesn't converge in mean squares to
α, nor to 0, and because convergence in MS implies convergence in probability,
{sn } doesn't converge neither to α nor to 0 in probability.
Question 2
What we know :
√ 4n − c
n(θ̂ − θ) →d N (0, σ 2 ) αn =
n
n−1 1
p(βn = 0) = p(βn = c) =
n n
Where c is a constant and βn a random sequence.
√
Limit in distribution of αn n(θ̂ − θ)?
√
By applying lemma 2.4 c) in Hayashi [?, p.92] if n(θ̂ − θ) →d Θ (where Θ
√
is the limit in distribution) and αn →p α then αn n(θ̂ − θ) →d αΘ. And
√
as pointed out in the same page, if Θ ∼ N (0, σ) then αn n(θ̂ − θ) →d
Nr (0, ασ 2 α0 ), where r is the number of rows of α.
We therefore rst have to show that αn →p α, for that, we will show that
8
αn →ms α which implies the former convergence :
αn →ms α ⇒ αn →p α
E(αΘ) = 0
and
= E (αΘ)2 = ασ 2 α
P ((αΘ)2 = ασ 2 α) = 1
therefore
(αΘ)2 = ασ 2 α
√ √
αΘ = ασ α
Θ=σ
√
αn n(θ̂ − θ) →d ασ
√
n(θ̂ − θ) + βn →d σ + β̂
9
Therefore we need to know what is the limit in probability of β :
n−1 n−1 1
P (βn = 0) = ⇒ P (|βn − 0| > ε) = 1 − =
n n n
therefore
βn →p β̂ (where β̂ = 0)
and so
√
n(θ̂ − θ) + βn →d σ + β̂
√
n(θ̂ − θ) + βn →d σ
Limit in probability of θ̂ − θ ?
√
We know that θˆn is consistent (because n(θ̂−θ) →d N (0, σ 2 )), that means
that :
θˆn →p θ
θˆn − θ →p 0
Limit in probability of βn αn ?
We already know that αn →p α and βn → p 0 , therefore by applying lemma
2.2 we have :
αn βn →p 0
Limit in mean square of βn ?
The expected value of βn is :
n−1 1
E(βn ) = 0 × +c×
n n
c
=
n
and
c2
E(βn2 ) =
n
Therefore :
h i
E (βn − β̂)2 = E(βn2 − 2βn β̂ + β̂ 2 )
= E(βn2 ) − 2β̂E(β) + β̂ 2
c2 1
= − 2β̂ + β̂ 2
n n
By taking the limit :
h i c2 1
lim E (βn − β̂)2 = lim − 2β̂ + β̂ 2
n→∞ n→∞ n n
= 0 − 0 + β̂ 2
⇒ β̂ 2 = 0 ⇔ β̂ = 0
10
Therefore, βn converges Almost Surely to β̂ only if β̂ = 0, i.e. :
βn →AS 0
Question 3
z̄n →p µ
Limit in mean square of z̄n ?
We will suppose that :
E (z̄n − z̃)2 = E [z̄n − E(z̄n )]2 + 2[z̄n − E(z̄n )][E(z̄n ) − z̃] + [E(z̄n ) − z̃]2
⇒ z̄n →M S µ
11
Question 4
We know that :
E(yi ) = E(σi i )
2 2
)2
= E (c + αyi−1 + γyi−2
We immediatly see that the expected value of y depends on its past values
and a constant, therefore it is not equal to zero.
is yi a white noise process?
We will rst show that i is independent of (y1 , ...yi−2 ), then we will verify
that the denition of a martingal dierence sequence is satised :
We can rewrite y i = σ i i as :
2 2
yi = (c + αyi−1 + γyi−2 )1/2 i
We could substitute yi−1 and yi−2 by its past values up to the rst obser-
vationy1 , in that case, yi would become a function of y1 and {e3 , e4 , . . . , ei}
(because i ≥ 3), therefore i is independent of {y1 , . . . , yi−2 }.
With that in mind, we have :
12
Question 5
The random walk process yi = yi−1 +i is covariance stationary if E(yi ) doesn't
depend on t and cov(yi , yi−j ) exists for j = 0, ±1, ±2, . . . and depends only on
j but not on i.
E(yi ) = E(yi−1 + i )
= E(yi−1 ) + E(i )
= E(yi−2 + i−1 ) (because is IWN)
= E(yi−2 ) + E(i−2 )
.
.
.
= E(y0 )
= y0 (it doesn't depend on i because it's a constant)
.
.
.
i
X i−j
X
= E[(y0 + l )(y0 + k )] − y02
l=1 k=1
i−j i−j
" i i
#
X X X X
= E y0 y0 + y0 l + y 0 k + k l − y02
l=1 k=1 k=1 l=1
i
X i−j
X i−j
X
= y02 + y0 E(l ) + y0 E(k ) + E k l − y02
l=1 k=1 k=1,l=1
2
= (i − j)σ
13
2 Empirical exercise
We consider the regression
We note that family income is, not surprisingly, correlated with the edu-
cation of both the mother and father, with correlation coerricients of 0.448
and 0.427 respectively. Furthermore, we note that the education between the
mother and father is even stronger correlated, with a correlation coecient of
0.643.
Now we turn to our regression model, which we etimate using OLS. We re-
move the incomplete observations, and from an initial number of 1388 observa-
tions we are left with 1191 complete observations. The results from OxMetrics
are reported below.
14
---- PcGive 13.10 session started at 18:53:08 on 8-11-2010 ----
We see that neither f aminc and f atheduc are signicant at the 10% critical
value. motheduc is just below this limit, and not signicant at the 5% critical
value.
In order to construct the 95% condence interval, we note that
bk
−tα/2 (n − K) < < tα/2 (n − K)
SE(bk − β¯k
where 1−α is the condence interval you wish to use, in our case the 95%
condence interval, and β¯k is the value of the null hypothesis which you wish
to test, in our case ¯
βk = 0 . I will use the values from the normal distribution,
considering this as a good approximation of the t-distribution with 1185 degrees
of freedom. Then we have the following 95% condence intervals. First, the
constant:
7.1578
−1.96 < < 1.96
0.2330
Here, with a t-value of 30.7, clearly we reject the null. Then cigs,
−0.0372
−1.96 < < 1.96
0.0067
15
which gives a t-value of -5.40, and again, we reject the null with good
margin. Then for parity :
0.1127
−1.96 < < 1.96
0.0412
which a gives a t-value of 2.71 > 1.96 and again we reject the null. Then
for f aminc:
0.0035
−1.96 < < 1.96
0.0023
This time, we have a t-value of 1.53, and this time we cannot reject the
null. Next, motheduc:
0.0295
−1.96 < < 1.96
0.0177
The t-value this time is 1.67, and we do not reject the null.
−0.0232
−1.96 < < 1.96
0.0200
This is for f atheduc, and gives a t-value of -1.16, which lies inside the
interval, and we do not reject the null.
The rank of the matrix X is 6, so there is no perfect multicollinearity.
However, when excluding motheduc from the regression, we see that the part
R squared of f aminc, that is, the correlation of f aminc with the dependent
variable given the other explanatory variables, increases (see under). This
suggests that there might be a collinearity problem between these two variables.
0 0 0 0 1 0 0
R= ,r =
0 0 0 0 0 1 0
This gives us the following test statistics for the F-test:
16
r vector
0.00000 0.00000
LinRes F(2,1185) = 1.4373 [0.2380]
Heteroscedasticity coefficients:
Coefficient Std.Error t-value
cigs -0.032041 0.049584 -0.64620
parity 0.0045450 0.44633 0.010183
faminc 0.015506 0.026471 0.58575
motheduc 0.15605 0.23825 0.65501
fatheduc 0.29249 0.33937 0.86185
cigs^2 0.0010846 0.0019191 0.56514
parity^2 -0.0071292 0.091492 -0.077921
faminc^2 -0.00020918 0.00033770 -0.61941
motheduc^2 -0.0059738 0.0093890 -0.63626
fatheduc^2 -0.010163 0.012990 -0.78240
17
References
18