Nothing Special   »   [go: up one dir, main page]

Exact Inference

Download as pdf or txt
Download as pdf or txt
You are on page 1of 18

Single Linear Hypothesis

Multiple Linear Hypothesis


Confidence Intervals and Regions

Hypothesis Testing and Confidence Intervals under


Normality
Walter Sosa-Escudero
Econ 507. Econometric Analysis. Spring 2009

December 11, 2008

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Introduction
Classical linear model:
1

Linearity: Y = X + u.

Strict exogeneity: E(u|X) = 0

No Multicollinearity: (X) = K.

No heteroskedasticity/ serial correlation: V (u|X) = 2 X.

0 X)1 X 0 Y
OLS Estimator: = (XP
2
2
Estimator of : S = 2i=1 e2i /(n K).

Now we are interested in testing hypothesis about the unknown


vector of coefficients , or in constructing confidence intervals
We do not have enough information to perform this task!
Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Why? Consider the following example.


We are interested in testing H0 : j = 0 (significance).
j is not observed. But we can produce an estimator j
Classical approach: if H0 is correct, and j is good estimator
of j , then j ' 0. So we should reject H0 is j is sufficiently
different from 0.
What do we mean by j being sufficiently different from 0 ?
j is by construction a random variable, so it has to be a
probabilistic statement.
We need the distribution of j .
Can we get it based on the model as it is so far?

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Assumption 5: normality. u|X is normally distributed (it is a


vector, so this involves the multivariate normal).
Note that this together with the classical assumptions imply
u|X N (0, 2 In )
Remember that = + (X 0 X)1 X 0 u.Then
| X N , 2 (X 0 X)1

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Hypothesis about single coefficients


H0 : j = j0 vs. HA : j 6= j0
Let ais denote de (i, s) element of (X 0 X)1 .
Then, when H0 is true j j0 N (0, 2 ajj ) so
j j0
N (0, 1)
zj p
2 ajj
Note that 2 ajj = V (j ).
Special case: j0 = 0 significance hypothesis.
The distribution of zk does not depend on X.
If 2 is observed, reject if zj lies outside an acceptance region.

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

The problem is that 2 is not observed. Define:


j j0
tj p
S 2 ajj
which is zk with 2 replaced by its unbiased estimator S 2 .

Result: Under assumptions 1 to 5 and when H0 holds,


tj t(n K).

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Proof: First, two preliminary results:


If X N (0, 1) and Y 2 (m) and independent, then
X
p
tm
Y /m
If X N (0, Im ), then X 0 AX 2 ((A)).
Note that:
j j0
zj
zj
tj = p
=p
=q
e0 e/(nK)
S 2 ajj
S 2 / 2
2

We have already shown zj N (0, 1) so if we can show


e0 e/ 2 2 (n K) and independence, we are done.

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

2 numerator: Recall e = M u, so e0 e = u0 M M 0 u = u0 M u
since M is idempotent. Also, we have assumed
u|X N (0, 2 In ), then by our previous result
u0 M u/ 2 2 (n K) since (M ) = n K.
Independence: = + (X 0 X)1 X 0 u and e = M u are linear
functions of u, they are jointly normal given X. Now
e) = E[( )e0 ]
Cov(,
= E[(X 0 X)1 X 0 uu0 M ]
= (X 0 X)1 X 0 E(uu|X)M
= 2 (X 0 X)1 X 0 M = 0
using LIE and the spherical disturbances assumption.

Since zj depends solely on and the numerator only on e, then


the result follows.
Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Hypothesis about linear combinations of .

H0 : c0 r = 0 vs. HA : c0 r 6= 0, c <K , r <.


WLOG, supose K = 3 so,
Yi = 1 X1i + 2 X2i + 3 X3i + ui

i = 1, . . . , n

Consider the following hypotheses:


a) H0 : 2 = 3 , or H0 : 2 3 = 0. In this case
c = (0, 1, 1) and r = 0.
b) H0 : 2 + 3 = 1, so now c = (0, 1, 1) and r = 1.

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

To derive an appropiate test statistic note:


c0 r N (0, 2 c0 (X 0 X)c) N (0, 1)
So

c0 r
N (0, 1)
z=p
2 c0 (X 0 X)c)

And again, by the same argument as before, a feasible version is


c0 r
t= p
t(n K)
S 2 c0 (X 0 X)1 c)

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

As a simple exercise, the appropriate statistics for the cases


considered before are
a) c0 r = 2 3 and
d 1 , 2 ), so
2 c0 (X 0 X)1 c = V (2 ) + V (2 ) 2Cov(
t=

2 3
d 1 , 2 )
V (2 ) + V (2 ) 2Cov(

b) c0 r = 2 + 3 1, and
d 1 , 2 ), so
2 c0 (X 0 X)1 c = V (2 ) + V (2 ) + 2Cov(
t=

2 + 3 1
d 1 , 2 )
V (2 ) + V (2 ) + 2Cov(

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Multiple Linear Hypothesis


H0 : R r = 0, R is a q K matrix with (R) = q, and r <q
Example. In our previous case consider the multiple hypothesis
H0 : 2 = 0 : 3 = 0
These are actually two joint hypothesis about the coefficient vector
. In this case


 
0 1 0
0
R=
r=
0 0 1
0
with q = 2. r is the number of restrictions.
What is the full row rankrequirement, (R) = q, asking for?
Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Consider the following test statistic:



1
(R r)0 / q
(R r)0 R(X 0 X)1 R0
F =
S2
Result: under all assumptions and when H0 is true,
F F (q, n K).
= 2 R(X 0 X)1 R0 . Then,
Intuition: Note that V (R)
2
0
1
= S R(X X) R0 , so
V (R)
1

F = (R r)0 V (R|X)
(R r)0 / q

F is actually checking how large R r is.

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Proof: Again, let us start with two results


If X 2 (m) and Y 2 (n) and independent, then
(X/m)
F (m, n)
X/n
If X is an m random vector X N (0, ) with invertible,
then X 0 1 X 2 (m)
Reexpress the F statistic as follows:

1
(R r)0 R(X 0 X)1 R0
(R r) / q
W/r
=
F =
2
S
Q/(n K)

1
with W (R r)0 2 R(X 0 X)1 R0
(R r) and
0
2
Q = e e/ .
Now it is obvious what we have to do...
Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

We already know e0 e/ 2 2 (n K).


Now, under H0 : R r = 0 and our assumptions
R r | X N (0, 2 R(X 0 X)1 R0 )
It is easy to check that R(X 0 X)1 R0 is invertible (why? do it!),
hence
W | X 2 (r)
Now the desired result follows since W is a function of and Q a
function of e, which are independent.

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

An Alternative Formulation: Restricted Least Squares


Consider H0 : R r = 0, and the following estimator that arises
from the optimization problem
s.t. RR = r
R = argmin SSR()

R is the restricted least squares estimator. Lets adopt the


following notation
eR = Y X 0 R , SSRR = e0R eR
SSRU = e0 e

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Result:
F =

(SSRR SSRU ) / q
SSRU /(n K)

Intuition
SSRR SSRU 0. Why?
(SSRR SSRU ) is the loss in explanatory power by imposing
H0 as a restriction.
When H0 is true, the restriction should not matter so the loss
should be small.
Proof: see homework

Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

Single Linear Hypothesis


Multiple Linear Hypothesis
Confidence Intervals and Regions

Confidence Intervals and Regions


Let t/2 be the 1 /2 quantile of a Students t RV with (n K)
df. By construction

j j
< t/2 = 1
P r t/2 < q

V (j )


q
q
P r j t/2 V (j ) < j < j + t/2 V (j ) = 1
Then
j t/2

V (j )

is an 1 confidence interval for j . Note that it contains all the


values of j for which H0 : j = 0 is accepted.
Walter Sosa-Escudero

Hypothesis Testing and Confidence Intervals under Normality

You might also like