Sie sind auf Seite 1von 25

Econometrics I

Professor William Greene


Stern School of Business
Department of Economics

4-1/25 Part 4: Partial Regression and Correlation


Econometrics I

Part 4 – Partial Regression


and Partial Correlation

4-2/25 Part 4: Partial Regression and Correlation


Frisch-Waugh (1933) ‘Theorem’
Context: Model contains two sets of variables:
X = [ (1,time) : (other variables)]
= [X1 X2]
Regression model:
y = X11 + X22 +  (population)
= X1b1 + X2b2 + e (sample)
Problem: Algebraic expression for the second set
of least squares coefficients, b2

4-3/25 Part 4: Partial Regression and Correlation


Partitioned Solution
Method of solution (Why did F&W care? In 1933, matrix
computation was not trivial!)
Direct manipulation of normal equations produces

( XX )b = X y
 X1 X1 X1 X 2   X1 y 
X = [X1 , X 2 ] so X X =   and X y =  
X 
 2 1 X X 
2 2
X X 
 2  y
 X1 X1 X1 X 2   b1   X1 y 
(XX)b =     = 
X 
 2 1 X X 
2 2  2 
X b X 
 2  y
X1 X1b1  X 1 X 2b2  X 1 y
X2 X1b1  X 2 X 2b2  X 2 y ==> X 2 X 2b2  X 2 y - X 2 X 1b1
= X2 (y - X1b1 )
4-4/25 Part 4: Partial Regression and Correlation
Partitioned Solution
Direct manipulation of normal equations produces
b2 = (X2X2)-1X2(y - X1b1)
What is this? Regression of (y - X1b1) on X2
If we knew b1, this is the solution for b2.
Important result (perhaps not fundamental). Note the result
if X2X1 = 0.
Useful in theory: Probably
Likely in practice? Not at all.

4-5/25 Part 4: Partial Regression and Correlation


Partitioned Inverse
Use of the partitioned inverse result produces a
fundamental result: What is the southeast
element in the inverse of the moment matrix?

-1
 X1'X1 X1'X2 
 X 'X 
 2 1 X2'X2 

4-6/25 Part 4: Partial Regression and Correlation


Partitioned Inverse
The algebraic result is:
[ ]-1(2,2) = {[X2’X2] - X2’X1(X1’X1)-1 X1’X2}-1
= [X2’(I - X1(X1’X1)-1X1’)X2]-1
= [X2’M1X2]-1
 Note the appearance of an “M” matrix. How do
we interpret this result?
 Note the implication for the case in which X1 is a
single variable. (Theorem, p. 37)
 Note the implication for the case in which X1 is
the constant term. (p. 38)

4-7/25 Part 4: Partial Regression and Correlation


Frisch-Waugh (1933) Basic Result
Lovell (JASA, 1963) did the matrix algebra.
Continuing the algebraic manipulation:

b2 = [X2’M1X2]-1[X2’M1y].

This is Frisch and Waugh’s famous result - the “double residual


regression.”

How do we interpret this? A regression of residuals on residuals.

“We get the same result whether we (1) detrend the other variables
by using the residuals from a regression of them on a constant
and a time trend and use the detrended data in the regression or
(2) just include a constant and a time trend in the regression and
not detrend the data”

“Detrend the data” means compute the residuals from the


regressions of the variables on a constant and a time trend.

4-8/25 Part 4: Partial Regression and Correlation


Important Implications
 Isolating a single coefficient in a regression.
(Corollary 3.2.1, p. 37). The double residual
regression.
 Regression of residuals on residuals – ‘partialling’
out the effect of the other variables.
 It is not necessary to ‘partial’ the other Xs out of
y because M1 is idempotent. (This is a very
useful result.) (i.e., X2M1’M1y = X2M1y)
 (Orthogonal regression) Suppose X1 and X2 are
orthogonal; X1X2 = 0. What is M1X2?

4-9/25 Part 4: Partial Regression and Correlation


Applying Frisch-Waugh

Using gasoline data from Notes 3.


X = [1, year, PG, Y], y = G as before.
Full least squares regression of y on X.

Partitioned regression strategy:


1. Regress PG and Y on (1,Year) (detrend them) and compute
residuals PG* and Y*
2. Regress G on (1,Year) and compute residuals G*. (This step
is not actually necessary.)
3. Regress G* on PG* and Y*. (Should produce -11.9827 and
0.0478181.

4-10/25 Part 4: Partial Regression and Correlation


Detrending the Variables – Pg
Pg* are the residuals from this regression

4-11/25 Part 4: Partial Regression and Correlation


Regression of detrended G on detrended Pg
and detrended Y

4-12/25 Part 4: Partial Regression and Correlation


Partial Regression
Important terms in this context:
Partialing out the effect of X1.
Netting out the effect …

“Partial regression coefficients.”


To continue belaboring the point: Note the interpretation of partial
regression as “net of the effect of …”

This is the (very powerful) Frisch – Waugh Theorem. This is what is


meant by “controlling for the effect of X1.”

Now, follow this through for the case in which X1 is just a constant term,
column of ones.
What are the residuals in a regression on a constant. What is M1?
Note that this produces the result that we can do linear regression on data
in mean deviation form.

'Partial regression coefficients' are the same as 'multiple regression


coefficients.' It follows from the Frisch-Waugh theorem.

4-13/25 Part 4: Partial Regression and Correlation


Partial Correlation
Working definition. Correlation between sets of residuals.
Some results on computation: Based on the M matrices.
Some important considerations:
Partial correlations and coefficients can have signs and
magnitudes that differ greatly from gross correlations and
simple regression coefficients.

Compare the simple (gross) correlation of G and PG with the partial


correlation, net of the time effect. (Could you have predicted
the negative partial correlation?)
4 .5 0

CALC;list;Cor(g,pg)$
4 .0 0

3 .5 0

Result = .7696572 3 .0 0

PG
2 .5 0

2 .0 0

CALC;list;cor(gstar,pgstar)$ 1 .5 0

Result = -.6589938 1 .0 0

.5 0
70 80 90 100 110 120
G

4-14/25 Part 4: Partial Regression and Correlation


Partial Correlation

4-15/25 Part 4: Partial Regression and Correlation


THE Most Famous Application of
Frisch-Waugh: The Fixed Effects Model
A regression model with a dummy variable for
each individual in the sample, each observed Ti times.
There is a constant term for each individual.
yi = Xi + diαi + εi, for each individual
N columns

 y1   X1 d1 0 0 0
  X
 y2    2 0 d2 0 0   β 
ε N may be thousands. I.e., the
     α  regression has thousands of
    variables (coefficients).
 yN   X N 0 0 0 dN 
β
= [X, D]    ε
 α
= Zδ  ε

4-16/25 Part 4: Partial Regression and Correlation


Application – Health and Income
German Health Care Usage Data, 7,293 Individuals,
Varying Numbers of Periods
Variables in the file are
Data downloaded from Journal of Applied Econometrics Archive. This is an unbalanced
panel with N = 7,293 individuals. There are altogether n = 27,326 observations. The
number of observations ranges from 1 to 7 per family.
(Frequencies are: 1=1525, 2=2158, 3=825, 4=926, 5=1051, 6=1000, 7=987).
The dependent variable of interest is
DOCVIS = number of visits to the doctor in the observation period
HHNINC = household nominal monthly net income in German marks / 10000.
(4 observations with income=0 were dropped)
HHKIDS = children under age 16 in the household = 1; otherwise = 0
EDUC = years of schooling
AGE = age in years

We desire also to include a separate family effect (7293 of them) for each family. This
requires 7293 dummy variables in addition to the four regressors.

4-17/25 Part 4: Partial Regression and Correlation


Estimating the Fixed Effects Model
The FE model is a linear regression model
with many independent variables
1
b  X X X D   X y 
  DX DD  Dy 
 a    
Using the Frisch-Waugh theorem
b =[X MD X ] X MD y 
 1

4-18/25 Part 4: Partial Regression and Correlation


Fixed Effects Estimator (cont.)
MD  I  D(DD) 1 D
d1 0 0
 
0 d 0
D  2 
 ... 
 
 0 0 dN

d1d1 0 0 
  
0 d d 0
DD   2 2 
 ... 
 
 0 0 
dNdN 

4-19/25 Part 4: Partial Regression and Correlation


Fixed Effects Estimator (cont.)

MD  I  D(DD) 1 D
M1D 0 0 
 2 
0 M 0
MD   D  (The dummy variables are orthogonal)
 
 N

 0 0 MD 
MDi  I Ti  di (didi ) 1 d = I Ti  (1/Ti )did
1 - T1i - T1i ... - T1i 
 1 
 - Ti 1 - Ti ... - Ti 
1 1

= 
 ... ... ... ... 
- 1 - 1
... 1 - T1i 
 Ti Ti

4-20/25 Part 4: Partial Regression and Correlation


‘Within’ Transformations

  
MDi X i  I Ti  di (didi ) 1 d X i = I Ti  (1/Ti )did X i 
 X i  (1/Ti )di  dX i   X i  (1/Ti )di (Ti x i )  X i  di x i
(Ti  K)  (Ti  1) * (1  K)
X MD X = Ni=1 X iMDi X i ,
 X iMDi X i 
k,l
T
  t=1
i
(x it,k -x i.,k )(x it,l -x i.,l ) (k,l element)

X MD y = Ni=1 X iMDi y i ,


XM y 
i
i
D i k   t=1
i
T
(x it,k -x i.,k )(y it -y i. )

4-21/25 Part 4: Partial Regression and Correlation


Least Squares Dummy Variable Estimator

 b is obtained by ‘within’ groups least squares (group


mean deviations)
 Normal equations for a are D’Xb+D’Da=D’y
a = (D’D)-1D’(y – Xb) (see slide 5)

Ti
ai=(1/Ti )Σt=1 (yit -xitb)=ei

4-22/25 Part 4: Partial Regression and Correlation


A Fixed Effects Regression Constant terms

4-23/25 Part 4: Partial Regression and Correlation


What happens to a time invariant
variable such as gender?
f = a variable that is the same in every period (FEMALE)
T
f MD f  ni=1fM
i Di fi  i=1
n
 t=1
i
(fit -fi. )2
but fit = fi , so fit  fi. and (fit -fi. )  0 for all i and t
This is a simple case of multicollinearity.
f = diag(fi )*D

4-24/25 Part 4: Partial Regression and Correlation


4-25/25 Part 4: Partial Regression and Correlation

Das könnte Ihnen auch gefallen