Beruflich Dokumente
Kultur Dokumente
E(bj ) = j
Statement
K
regardless of the values of Xij . Now, let j=1 j j
be some linear combination of the coecients. Then the
mean squared error of the corresponding estimation is
y = X + ,
2
K
E
j (bj j ) ;
expanding to,
j=1
yi =
j Xij + i
i = 1, 2, . . . , n
j=1
E(i ) = 0,
V (i ) = 2 < ,
1
As it has been stated before, the condition of V ()
b
of y and X (where X denotes the transpose of X ) that V () is equivalent to the property that the best linear unt
t
minimizes the sum of squares of residuals (mispredic- biased estimator of l is l b (best in the sense that it has
minimum variance). To see this, let lt another linear
tion amounts):
unbiased estimator of lt .
= lt V ()l
= 2 lt (X X)1 l +lt DDt l
V (lt )
2
{z
}
|
K
n
n
V (lt )
2
b
(yi ybi ) =
yi
j Xij
.
+ (Dt l)t (Dt l) = V (lt )
+ ||Dt l|| V (lt )
= V (lt )
i=1
i=1
j=1
V (lt )
.
Therefore, V (lt )
=
Moreover, suppose that the equality holds ( V (lt )
t
t
V (l ) ). It happens if and only if D l = 0 . Remembering that, from the proof above, we have =
((X X)1 X + D)Y , then:
lt =lt (X X)1 X Y + lt DY = lt b + (Dt l)t Y = lt b
| {z }
=0
Proof
5 GaussMarkov theorem
in econometrics
0
= V (Cy) = CV (y)C = 2 CC
V ()
= 2 ((X X)1 X + D)(X(X X)1 + D )
as
5.1 Linearity
2
2
1
2
1
of
specied
in the model. The specication
DX (X X) + DD
= (X X) + (X X) (DX
|{z}) + |{z}
must
be
linear
in
its
parameters.
This does not mean that
0
0
there must be a linear relationship between the indepen2
1
2
= (X X) + DD .
|
{z
}
dent and dependent variables. The independent variables
V ()
can take non-linear forms as long as the parameters are
linear. The equation y = 0 + 1 x2 , qualies as linear
Since DD' is a positive semidenite matrix, V () exceeds while y = 0 + 12 x can be transformed to be linear by
by a positive semidenite matrix.
V ()
replacing 12 by another parameter, say . An equation
5.4
Spherical errors
Strict exogeneity
E[ i | X] = E[ i | x1 , . . . , xn ] = 0.
[
]T
where xi = xi1 xi2 . . . xik is the data vector
of regressors for the ith observation, and consequently
[
]T
xT
. . . xT
X = xT
is the data matrix or design
n
1
2
matrix.
E[ T | X] = Var[ | X] = .
..
0
2
..
.
...
...
..
.
0
0
2
.. = I
.
...
E[ xj1 i ]
E[ xj2 i ]
E[ xj i ] =
=0
..
.
E[ xjk i ]
all fori, j n
Heteroskedacity occurs when the amount of error is correlated with an independent variable. For example, in a
regression on food expenditure and income, the error is
correlated with income. Low income people generally
spend a similar amount on food, while high income people may spend a very large amount or as little as low income people spend. Heteroskedacity can also be caused
by changes in measurement practices. For example, as
statistical oces improve their data, measurement error
decreases, so the error term declines over time.
with 2 > 0
EXTERNAL LINKS
8 References
See also
6.1
Notes
9 External links
Earliest Known Uses of Some of the Words of
Mathematics: G (brief history and explanation of
the name)
Proof of the Gauss Markov theorem for multiple linear regression (makes use of matrix algebra)
A Proof of the Gauss Markov theorem using geometry
10
10.1
10.2
Images
10.3
Content license