Beruflich Dokumente
Kultur Dokumente
SE 037 439
O'Brien,-Trancis J.
AUTHOR
TITLE'
Jr.
Correlation
Cdefficient Has Limits ((Plus or Minus)) 1.
P r
'
PUB DATE
NOTE
32p.
EDRS PRICE
DESCRIPTORS
.
.11
8'2
IDENTIFIERS
ABSTRACT'
6
1
4
'
,P,
.
I
O
f,
****************************4,**************1***************************
*
*
Reproductions supplied by EDRS are the beet that can be made
* ,
.*
from the original document.
*****************************************v*****************************.fr
Has
Limits
-I:
reprOduCnOn Quality.
othcfal NIE
posafon or 'poky
-4'
..
r%-
CO '-.
%CO
orslinear
s
r--4
'HoWever,
proves this.
One of the
This paper will set forth in-detail a proof bf the limits of the sample
bivariate correlation coefficient. Since the proof requires only knowledge of
4
standing.the proof.
applied statistics; I know that the typical student can understand the
prodf as it is presented here.
The key fdr tinderstanding statistical proofs is a presentation
-
A review
1P4
..
and
.\
'2.
1.
'
usually required).
in detail im-
portant statistical proofsithey feel that some of the mystery and magic
of mathematics has been unveiled.
motivational
)
Some Preliminary
Concepts
1This
T
paper is one of a series contemplated.for publication. [See
O'Brien, 1982]. Eventually ,I hope to present a textbook of applied /stat'
tics proofs and derivations to supplement standard applied statistics
textbooks.
61"
Table 1
Measure
Unstandardizgd
Measure
(X -R)/S'
rR)/S
(X
3
r,
x2
Y1
='i
JY
= Z
(Y -Y)/S
Zx
1
Standardized
Unstandardized
',Standardized
(X -;) /S
x
1
Y2
= Z
(Y
3'
Y3
I
0
t*
x.
(g.1
-R1/S
= Z
(Yi-) /S
Yi
x.
.
= Z
Y
-X
(X -T) /S
.=Z
= Z
(Y
40
Y n'
'Sample
Size
n
zx
n
x
Y,
Sample
Mean
Sample
Variance
.,2*
3c
r'.
#.
.
-NOTE:
= Ti
,1= n
z.x
-= n
zy
..
' o,
Any of these sample size terms could be identified by,just.Che symbol -- such As
n.
We' will use h when it is not important to disiinguish..kmong the other sample'
size terms, but will Ilse the table' values above when itis necessary or important
to do so.
,,
Table
-Y
Measure
Measure
-4
Sample
Mean
1 =1
X=
X.
X
1=1
Y
n
TX.
nxX =
nx
Sum
Y.
n Y =
E
1=1
i=1
Y.
1
nx
i=1
Sample
Variance
S
x
nx
2
(n -1)S
x
x
Sum of
Squares ,
L....(Xi :X)
2,
.
(n
i=1
-1)S2 =
Y
),__.
i=1
(T.-(Y. -Y')-?
1
.
NOTES:
l ,
..
3.
.
Ilr
Standard Scores
It will be Kecalled that the standard score for an unstandardized
.
'
measure (raw score) is "the score-minus the mean divided by the standard
deviation".
x
=
1
X
Y -Y
1
Y1
Similarly,
Y-
.,
,
.
__
.
0-
6.
1
.)
xi
i=i
Z =
x
40k
z
0
n
cz
11 2
P
4
1=1
n
z
.
--.
E ) 2
...
-1
4,
..
16
7.
'I
n
z
1E: (z
2
i=1
-z 1
Yi Y
nz
-1
/.
If we square
scor4s.
each, standard score for the X measurein Table 1 and sum them, we obtain:
11,
4
.
Xi"
i=1
Z2
>..... Z2
Z
.
,,,-.
Z-2
-2
V
.
If we substitute the appropriate means and variances in .the right hand side
'n
zx
E
.i.71
------
z2
xi
;
0
.r'
(X
+
2
40 2
0(
-X).
(X -X)
n
2
C,
8.
Since
the
2
S
n ,
z ..
x
'
'
z-x4
..
(x -x)
1
(x
-xr
+...+ (x
)i)
n.
1 =1
.1
S,
!I
4'
,.,)
.e08
(X1 --co 2
h
z
1:
k=1,
Z2
S
i
i=1
.
x'
.
.
From Table 2, we know that we can substitute the sum of squares term
into the numerator on'the right hand side. This results in:
.5
(n Ll)S
x
n x-1
n71
,(n -1)Sy
Y
2
--.1a -1
,Y
4
(Recall that n 71 = n-1) .
41.
9.
As
Theu relationships between squared.z scores and sample size are very important'
,
Correlation Formulas
Unstandardized Form .
dized
i=1
r
3;
0;
xy
-7
-X)2
1=1
i=1
n -1
x
ma.
n -1,
Y
6...1.
Note that the numerator contains the term n-1 because it is not important
9
4'
10
if5a,
-1 =n-1)
Jr
10.
Standardized Form
The correlation of measure,Xsand measure Y in standard score form
is defined as follows:
1
n-1
rzx`zY
zx
z.
E(Z x.
-Z
i=1
i=1,
nz- -1
z y,
E.Z x.Z
d=1
171.
.
-1
rz z
x y
obtain:
-n
(n-1) r
x. , y .
z z
x y
i=1 "
ll:
.
.__
. .
'C'
,Inequalities,
to
it it necessaryto revie
one eifurther
form of inequality:
and
> B
<
A ',
) 1 or equivalently
}e obvious.
All'of this m
1 L 3.
For
example, if
.
N.
J
=1( (3)
(1)]
-3
/---
12
12.
4"
.
That is, the inequality sign is reversed. when multiplying by a negative number.
004:
,
For example:
1 - A >
-11(1-A) >
-(1-A)
(B)]
(0))
<
A-1 <
Example:
1 1
-Multiplying through by
4>Q
we obtain:
-1[ (1-1/4)
-(1-1/4)
347.1
<
^:
a
We have reviewed standard scores(z), correlkion formulas and
algebrait inequalities.
(
the
Table 3
x-z'
1=1
z
n-1
=
Y.
1=1
xy,
z z
x y
I
n
Z
1=1
,,
.y.
(n-1) r
'
(n-1) r
x y
(I-A)
z z
>.
(B)
<
xy
. \
.4
14.
Proof
-1
xy
<
xy
+1
.-011
-1
C
410t
+1
xy
The proof consists of two parts: one part shows the lower limit
4
o
of r
of'r
(i.e.,
xy
xy
xy
(i.e., e
xy
>
-1), and
g7
Proof that
xy
+1
algebraic manipulations
Dz
i=1
z.
xi
Yi
15
15.
e,
Yi
fi
Z
values starting at
and
...44.
xl, yi and conti uing down ,to the last pair of Z's (Zx ,Z
n
).
Yn
ob
Most students readily agree that the squared sum will be greater than O.
But can it ever be exactly equal to 0? Yes, theoretically it can.
Refer-
',
4
1
the same numerical value, then it is apparent that each difference will be
it is onlyreqUired
sense.
that
of 0 is also O.
value
Yi
be true in a mathematical
Notes
Refer,
examine the algebraic statement on the left side of the page. Then read the
Example:
Y.
i
1.41
-.68
1.41
-.68
.05
.05
etc.
40.
1.
That
+1
xy
Notes
2
-Z
Z(Z
yestatement
each term, we obtain an expansion of
1=1'
;t(Z2
2Z
1=1
Z
x. y.
3.
>
A2 + B2 - 2AB ,
(A-B.)
AK.
Squaring
from before.
Y1
3.
ti
n
2
+" Z- Y1
x.
1=1
1=1 1:4:
2 E
1=1
Z
x. y.
1
2:Z2
n-1
x.
i=1 1
17
n
I
n-1
.1.Z2
i=1
xZy
3.
3.
(n-1)r
(n-1)
z z
x y
i=1
.
So
f"
Making the
> 0
2(n -1)r
(n-1)
(n-1)
'xy
ree substitutions
2.(n -1)r
2(.1-1)
2(n-1)(17r
xy
>
>
xy
n-1).term
.FactOring the
Dividing each
2(n-1) which
equality s gn
positive b ca
2 (p-1). is always
'
2(n-1)(1-rxy ]
2(n-1)
2(n-1)'
(1-r
Here we make
se of multiplying an
a negative number. Let
ch side of the inequality
us multiply
which reverses
by -1_(see Ta le 3)
sign
and
reverses the 1-r
the inequalit
xy
inequality' b
13
-1[(1-r
xy
xy
>
'+i
xy
0]
-1
xy
Now, add
:+1
to each side
This gives us
END OF PROOF
xy
18.
Proof that
xy
-1
of
We will
true, namely:g.
.
",
2: (zx
)2
'o
y.
1=1
next page-
-1
4
0
10'
,0
-41Nha
xy
;..?
4
(2,
,. n
Zx
>
)2
i=1
(A+B)
o,
2AB
Ez
i=1
+ Z
x.
2Z
+
.
Yi e
>
Z
)
x. y.
1
1
....,
2 EZ Z
i=1 1
i=1111
, 0
y
i
i
.'
i
t+
(n-1)
2 (n-1) r
2(n-1)[1 + r
xy
xy
.>
Adding.
i%
1.
xy
.41"4"'to.each side
>
-1
Simplifying
xy 0
>
-1
23
22
ra
jolt
JO,
20.
,
.'..
...
+1.
See the
Appendix
'
tit
re
ri
,r
ti
24
v..
21.
J.
.4
APPENDIX
Fs
g.
Selsiled'Proofs2
1:
6C
'4604
the X meature:'
. n
zx '
Z. 41 i
Z
1=1
=
x
(X2-56
-i)
(X
n_z
S
-
..
..._
S
.
____CX_n ri)----
+ ... +
S
?(
in summation notation:
n
x
(ni:X)
1 ;
Zx
nz
47
1=1
22
7
'
:*
t_
Zx
n.
Sx
z
x-
Y.
e,
Since
x,
"")L
i=1
v Xlisequal
4,
x
n X
x
. -
X..-.1,.
i=1
,t--..:
to in X
X
P,
0,"'
..
1.
1
-
Zx
(n X
x
----
n-
n7()
.,
nz
Z
yi
i=1
Zy
(n
n1
zy
"Y.
- n
S1
.y
mean equal to 0.
0,
..,0
c + c + c + ..1- c
i=1
. ,nCL
.
.
.
26
.
23.
2.
By. definition, the variance fbr X meAsures in ,'standard score form is:
2::(Z
x.
i=1
,z
zX
-1
S.
n1
nz
*S---(Z
L___ x.
1=1
1
nz -1
.,
If we
..
x',
rewrite 'Z
'
'T
deviation:
z =
nz
x
i=1
2
S
Rearranging terms:
n
='
2-'
-1
z
S2
x
x
2
14EDX -X)
i=1
59
24.
the "sup of
x
This results in:
Since
nx = nz
2
S.
=
z
-1
:(n -1)(S )
X
X
(n -1) (S2)
Y
'Y
-1
nzy
2
,
Y
46.
Since n
= n
z
-Y'
2
In each case, the standard deviation for approp riate variance terms,
2
n.
Sz
=
X
=17-=
Sz
and
Tr
S
Y'
=
,-
=j1
S.
z
equal to 1.
a.
25.
0 3.
That
xy
rz
x y
:E=
,e 1
n-1
r
(z E
x
i=1
(Z E y)
tex
Yi
z z
x y
-am
z
2
-Z
x
i=1
n
^
i=1
-1
-1
x
,
Since Z
4i
n
(Zx ) (Z
n-1
r
y .
3.
3.
i=1
z z
x y
nz
2
(Z
Y.
i=1
nz -1
Y4
(z'xi
)2
= n -1
z
.111140114,
nz
and
r.
26.
S--(Z
r
.)(Z
x.
1.
y.
n-1 i=1
z z
x y
-1
.nz)c-1
-1
z
-1
n
z
r
z z
11-1
x y
xi yi
1=1
(Recall that this relationship was used in the proof for the limits of r ).
xy
Nowo expanding the z score terms:
r.--
,
r
z z
x y
cp
n-1
(Xi -T)
(y.- 1-i)
.4i=1 '(S ) (S
x
This is identical to :
n
1.
RHY i)'
n-1 i=1
30
441
27.
Recognizing that
S2
and \02
, we can write:
40P
'
)
(X -R).(Y.-i
1
n-1
1=1
2
(S
2
) (S
aw score
.
i Em-X)
n-1
.
z z
x y
r-
-i)
i=1
....
n
Y
2
21(X. -X)
.4*
i=1
i=1
.
n x-1
Therefore,
-Y)2
Y:
n
,110i46
=mow.
01/0.
.....
...-
n
x
xY
forms is identical.
31
28.
References
O'Brien, Francis J., Jr. A 15/15tf that t and 'F are i e tical: the general
case. ERIC Clearinghouse orisCience, Mathematics and Environmental`.
(Note: the ED number for
Education, Ohio State University, April, 1982.
was not available at the time of this publication).
this document
.7
-DOCUMENT'RESUME
A
SE 037 098
AUTIOR
TITLE
Case.
,24 Apr 82
PUB DATE
NOTE
EDR PRICE
DESCRIPTORS
20p.
.
'
ABSTRACT
MO'
***********************************************************************
Reproductions supplied by EDRS are the best thtt can be made
*
from the original document.
****************-*******************************************************
A Proof that t
onognaung a
-_
Francis
F.
root NIP
POSMon Or poky
.1t is well known that a researcher wilt) has. collected data from two
independent groupsmA,
This is because
For example,
the equivalent
= F or t = (F)
1/2
.
....
or psychology
of the relationship (when sample sizes are equal). (See, for example,
.
N.
First, many
prior statements of the proof for the general case (of unequal sample
sizes) are either abbreviated, mathematically inaccessible or incomplete
the literature
'
that many students can follow well-articulated proofs and desire to see them
worked out.
/
In this paper
three
Firsty
Third,
a plan will be
be able to follow the proof of the general case, and therefore, shOuld
Proof that t
= F: the Gen4ral
j
Case
2.
3.
4.
5_.
observe that t
t8 the squared t
3.
Table 1
Table Layout
oup 1
Group 2
n.
X11
12
2.
21
22
t..
31
41
n.
n.
2
32
3.
42
AP
+
X.
n.
1
n. 2X.2
'2
n.1 ,+
n.
2
X.
12
Ime
4.
s ?.
is not needed
nl
Total
Sample
Size
n.
Sample
Mean
i.
n.
x..
X.2
Sample
Variance
n..
s.
s.
41
s..
N-
fr
5.
4-
\21'
This is the expresion we will use as a basis oecomparison with the.
simplified F statistic to be obtained in step 4.
2
.
. o
Step 3.
the F statistic
It.\ \will be recalled that the F statistic is the ratio of two independent :
1
).
df
is:
SSb/df
SS
w/df
The genera1.form'ofdfb'is "the number of groups minus one" (i.e., dfb = J-1,
where J is the number of groups).
4
two groups is df
= 2
1 = 1.
= n
+ n.
.1
df
='n
'1
J.
2
= n.. - J).
is
+ n.
'will be used instead of\the correct "among": 44afocept the righteous indig.
fr
5.
(N
Step 3.
2
.
4%
1.
the F statistic
).
sums of sqdares
( SSb )
df
and df
w (within sums of squares degrees of freedom).
is:
SS.
b/df
FJ- b,n.. -J
SS
w/df
= J-1,
two groups-is df
= 2 - 1 = 1.
an Write
= n.1 + n.
J.
2
= n.. - J).
Since n.! = n.
+ n.
1
dfw,
44
ad
-6.
sb/i
F
1,n.
+ n.
1
- 2
2
ss
SS
SS
w
- 2)
+n.
(n.
1
familiar
That is:
n.
1nl+n. -2
+
1
n.
-X..)
(Z(
-1)s.
(n.
1
-1)s.
(n.
2
n.1 + n.2 2
'
7.
Table 2
L.
Restatement of t
and F
t2
F
41.
z
.-X.
(X.
1
(n: -.1)s.
(n.
+ n.2\
)s..
(A
-1)s.
:2
-R.:)2
+ (n. -1).
2
-2
,n.1
+ n.
(i.
+,n.
+
n.
(n.
)(n.
+ n.
n.
We will first
A full step-by-step
simplify the ,numerator
c..
4
4..
Notes,
n.
SS
n.
(;
1 "" 1
R..
(R..R..)2
n.
""=
n:1
+ n.
n.
1
n.2 4X
n.
n.2 /
(n.1X.1 + n.2X.2
(n.1 +n.2)X.1
+ n.
1
n.2
+n.
(n.
1
))-(.
- (n.1X.1
R.
n.
2
2!
+ n.
1 n.
1
].
n.
L
SS
P n.
n.
X.1
+ n.
1
Sat.
=
n.
n. 2X.21
+ n. X.
R.
1
AP
V.
+ n.
.n.
1
SS
- n. X..
tl
n.
1,
- n. X.
1
4/
n.1 + n.
2
+ n.
R.
n.
n.
2
n. R.
R.
1
n.
1
2X.2
2
Cancellipg.like teifms,
+ n.
n.
n.
n.
1
terms,
+ n.
n.
1
n.
SS
-R.
(R.
+ n.
n.
1
2
2
n.
rR. --)
n.1
n.2
10Na
SS
(h.
n.
2
1
(n.
(n.
2'
+ n.
1
/n.
qo(n.1) (n.2)
n.2(X.1 -X.2)2
n.
+ n.
-R.
(R.
1
1-
)2
1
-X,
(X .
b
(n.
SS
brackets,
(X.
+ n.
(n.
)(n.
(n.
+,n.
1
(R.2-R.1
1
(n.
SS
-X.
(X.
1)
y
lk
(n.1 + n.2)2
is the
411,
to obtain,
9.P
)(n..
(h.
1
(n
SS
+ n
+ n.21
(n
(X.1 1
)2
'2
and (n .
+ n .
1
) 2',
we obtain, 16-
2-
)(n. 2)
(n.
SS
(R.
1
-R.
)2
2
n.1 + n.,
4.
1_3
-14
'so
p
C
0/
11.
)(n.
(n.
1
(X.
n
'1
k
0
-1.)s..
(n.
).
-1)s.
(n.
-X.
1
(n.
n;
(R.
1
F
2
-1)s.
-R. )2
+ n.
Step 5.
observe that t
- 2
= F
t (n.
+ n.
n.
.
ENO OF PROOF.
-1)s2
(n.
n.
we obtain,
4-n.
+ n.
n.
(n.
).(n.
) (n.
form, signifying that they'are equal. Refer to Table 2 for this comparison.
This completes the entire proof in accordance with the five step plan.
15"
16
4
12.
t
Numerica1 Example
The analytib
proof using algebraic rules for the general case
.
41,
insight.
only.
Table 3
Group 1
"`Group 2
10
50
40
70
20 "
100
50
40
'QM
75
ieb
4
.
90
'Total
Sample
Size
10
Sample
Mean
07.5
65.0
54.5
Sample
Variance
957.5,
700.0
NOT NEEDED
13.
-,
value is:
47,.5 - 65.0
t
(6-1)957.5 + (4-1)700.0
10
6887.5
J.
t
10
24
-.9240
1%.
(-.9240)
.8537
(t
= .8537)
.We place the computed value in the margin for easy reference.
Now
+ 4(65.0-54.5)
(6-1)957.5 + (4-1)700.0
\_J
6 + 4
6(49) + 4(110.5)
F
6887,5
8
.8537
(F = .8537)
18
./
Thus, t
14.
= F.
=F is somewhat easier to
general case.
= n.
n.
R.2
R.
2.
X.. =
Thi5 value
de`
By making these two changes and by following the five step outline
used for the general case proof, the reader should be able to derive the
proof for the special case., One may.also wish to "make up" an.easy to work with
numerical data set to check
on the process.
I
0
Note
two
EdwardAll'en.L.
References
Glass, Gene V and Stanley, Julian_ C. Statistical Methods in Education
and Psychology. Englewood Cliffs, New Jersey: Prvitice-Hall, 1970.
Rocca, Anthony J. and Tweney, Ryan D. Analysis oalvariance and the "second discipline"
of Scientific isychology: a historicaraccount. Psychological
Bulletin, 1980, Vol.,87, No. 1, 166-184.
O
'fp
a MD
DOCUMENT RESUME
SE 051 972
ED 331 706
AUTHOR
TITLE
PUB DATE
NOTE
PUB TYPE
EDRS PRICE
DESCRIPTORS
ABSTRACT
This paper is the sixth in a series designed to
supplement the statistics training of students. The intended audience
is social science undergraduate and graduate students studying
applied statistics. The purpose of the applied statistics monographs
is to provide selected proofs and derivations of important
relationships or formulas that students do not find available and/or
comprehensible in journals, textbooks and similar sources. Derived is
the theoretical limits of the sample multivariate (or multiple)
correlation of one criterion (dependent variable) and any (finite)
number of predictors (independent variables). The proof given in this
paper involves deriving the individual terms of R. The lower limit
and upper limit of R are derived separately. (KR)
***********************************************************************
Reproductions supplied by EDRS are the best that can be made
from the original document.
**********************************************************************,
CC,
Cf'D
U S DEPARTMENT OF EDUCATION
(Mr* at F &rational Research and Improvement
EDUCATIONAL RESOURCES INFORMATION
CENTER IERIC)
cjit..1 L'i
C;j11)F11.11
RESOURCES
TO THE EDUCATIONAL
INFORMATION CENTER (ERIC)."
36 Linden Street
Middletown, RI 02840
March, 1991
1991
FRANCIS J. O'BRIEN, JR.
ALL RIGHTS RESERVED
Table of Contents
Page
Introduction
Introduction io Proof
Proof that 0 5 R 5 1
11
13
17
References
Table
1
Page
10
..0.7-,oEsoo
a ass....
ar*
_EE
1,
aet412Verata.thansao_o .ACM96011172
Introduction to Proof
+ Biy2 +
+ Bjryi +
+ Bryp
Rzrzi.
zi.
zp
sVtBjr 1
J=1
where
Rzy.zi,
Z.J,
Zp
ZY
zi. 42.
variables
5
B1 B2,
B,
B
p
variable*
r
r
..., r
y1, y2,
ryp
variable correlations.
Proof that 0 R 5 1
In this section we present a detailed proof that the limits of the
multiple R are 0/1. First, a review is given of the notation and
uecessary definitions as well as the relevant results that were derived
in O'Brien (1982c).
We can state the formal linear regression prediction equation for
p standardized predictors as follows:**
= B1 Z1 + B2Z2 + ...+ BiZi
...+ BZp
p+
* Technically, the beta weights (1) ) are called "standardized partial regression
cofficients". The formal notation in some standard textbooks is more elaborate than
ours (e.g., Hays, 1973 or Kendall and Stuart. 1967). As in previous papers, we have
minimized the reading of the symbdism to clarify the concepts in the development of
the proof.
**The coefficient "A" is not included for the reason given in O'Brien (1982c) ; i.e.. it
"drops out" in the least squares derivations and so may be ignored.
The multiple correlation (or just R for short) for this regression
model of p standardized predictors may be defined conceptually as:
=
Corray,
Cot(Zir
\IVaray)Vara0
'.vaaff
Table 1
Formula
Note
Sum
13 =0
i=1
summation is understood to
be across the sample for a
given predictorj.
Sum of Squares
Mean
= n-1
EZ 2
1=1
=0
summation is understood
to be across the sample
for a given predictorj.
Variance
n-1
= Vaa
r
Variance of jth
predictot j.
(Table 1 cont.)
Correlation
so rzAzy = rxy
General zero-order
correlation formula for
any two standardized
variables, Zx and Zy..
Table 2
rmulas
/92
p113,4
COU(Zy,
Var(Zy)
=1
Var(71)
= ;EI2 + 2 JUEhair
-pryj
Cord?,
, Zs7)
P
pr
+
2A 1.; ;Biro
P P2 j.; BiBrij
iBjryj
J=1
where
= dependent/independent variable
Pearson (zero-order) correlations, and
rg = Pearson correlations among the p
independent variables
ryi
to the proof for the 0/1 limits of R as developed in this paper. We now
demonstrate this proof. The development of the proof will consist of
two parts --one part will demonstrate the proof for the lower limit and
the other will show the proof for the upper limit. The lower limit is
now presented.
Vara0
n-1
Var(Z)
n-1
The reader will agree that the following algebraic inequality is a true
statement mathematically:
n-1
Zi.)
;Er +
1) P-1
;Iwo
.1"2 11`
Hence,
- Vad
= R2
or
Var(74) 0
Since the value of the square root of a variance term is, by definition,
positive, then
iJVar(Z)
or by substituting R2 ,
k0
4-1-21
Consequently,
R
O.
= a minimum.
E(Z
zi
Egy
)2
1=1
Ezyi&
24
1=1
1=1
1=1
2
Y1
1=1
Y1
1=1
(n-1)Vad 74)
+ BpZ)
1=1
1=1
21(BI7I7yi
+ B2Z2Zyi +
+ Bpyn)
i=1
2(B1EZ1Zy + B2 EZ2Zy +
i=1
Oe
+B
13
1:1
1=1
Z,i
X
1=1
From Table 1, it can be seen that any term of the form EZx Zy is equal to
1=1
(n- 1 )r
(r
xl
xY
I zyi
2EB 1
.=,
(n-l)r
+ B2(n-1)r
yl
2(n-l)iB r
Jyj
+ B (n- 1)r
P
yP1
.1=1
Collecting all terms together, we can now rewrite the least squares criterion
as:
14
15
)2
2(n-1)13 r
20
1=1
Upon factoring out n-1 and dividing it through the inequality, we have
2iBjry 2 0
E(Z, - 7.4- )2 = 1 +
z
1=1
.1=1
iB r
74).
Yi
.1=1
Thus,
(Zu.
- Zo
)2
1 + Var(
2Var(
20
i=1
Or
- Zi> )2
=1-
Var( 74)
i=1
Reversing the s% lse of the inequality, we can write the right hand side of
the above as:
Var(
5 1.
Vara& 5
Viii S 1 or
R S 1. Proof is completed.
References
Hays, W. L. Statistics for the Social Sciences (2nd ed.). NY: Holt,
17
is
DOCUMENT RESUME
TM 830 618
ED 235 205
AUTHOR
TITLE
INSTITUTION
PUB DATE.
NOTE
PUB TYPE
EDRS PRICE
DESCRIPTORS
IDENTIFIERS
MF01/PC03 Plus_Postage.
correlation; Higher Education; Instructional
Materials; *Mathematical Formulas; *Scores;
*StatittiCS; *Supplementary Reading Materials
Linear MOdelt; *Multiple Correlation Formula
ABSTRACT
**A********************************************************************
*
Reroductions supplied by-EDRS are the best that can be made
*
*
from the original document.
*********************************************************A*************
Scores
Raw
NORC
Sampling Department
902 Broadway
New Ygrk, NY 10010
June 24; 1983
1983
tvrt,.1 frmo
r10111.11i1),1
Minor
iwrson or orcianIzation
it
hantr..
11., b,,)
r11
,,(1,
re111o1,n.Z.,1
trtir)
',It
:
;)tifit,
SHEET
ERRATA
"A
GO-itettt-O
Noi4 ReadS
10, footnote,
4 lines down
13
17, footnote
nX1Y
hX Y
var(b2,x2)
var(b2x2)
Multiple R
multiple
b Ex2
29, f--JquationZx-Y
36)
2 lines from
bottom of text
mathematical calculus
Si
mathematical statistics
r
2
SY
Table of Contents
Page
Introduction
Overview of Derivation
Multiple Correlation
12
Derivation
15
19
28
30
36
Introduction
36
Plan
39
40
44
Alternate Procedure
47
48
51
'" 52
list of Tables
Page
Fable
16
ii
31
'
Research
Introduction
The purpose
My intended
1982c).
1-
Overview of Derivation
this formula for one raw score criterion (dependent variable) and
lany finite number of raw score predictors (independent variables);
Let us first state the formula we will derive and introduce the
notation used
RY
,x ,...,x,,...,x
=
P
hit
Sy S1 +
v 2
br-,S S
J YJ
+ ...+ b r S S
P YP y p
RY.X
1
x 2,".,.X
'13.r
:S S
J YJ Y
-where:
Y.A ix
x-
scores,
xl,x2,...,xj,...,xp
3.
...,b ,... b
b'ill
2
,r
vi,..., vp
...,S
;S-,0....S
_--
slope coefficients o
,--
regression weight-Si
We will
first present a deriVation for the simplest multivariate case: one criterion
and two predictors.
In addition, it
will motivate the use of summation when the algebra becomes complex.
The deriVation is then presented for the general case of 0 (finite)
predictors.
In that
use
The steps
are:
1.
2.
T.
4.
5.
6.
simplify.
we
4.
analysis theory.
et_al., 1982.
The intention
of regression analysis;
often selected to obtain the best linear "fit" for these conditions is
provided by the following equation:
+ 6-X. +
x
2
where:
Y
a,
b2
ter-MS,
1U
5.
tom.
That is, for each predictor, first find its mean and then subtract
For example,
X1
re-express
raw score'and
ThiS is dohe
As such, it is a matter
Y (or Y) as deviations: We could
=a
b (X
1-bx
1
1):
1
It is idealized
for prediction of actual or observed criterion scores.
of Y. In
in the sense that it assumes no error is made in the prediction
1-
Readers of the 1982c paper may wonder why on page 2 there6f the
raTscore regression model was stated in terms of gross raw score (and
As stated, it is not_necessary mathnot deviation score) predictors;
In
any
case, the major result_we are seeking in
ematically to re-express;
this paper is unaffected by the initial form of the predictors. The
derivation could be made witnout the translation of predictors into
deviation score form, but the result ould involve unnecessary and unwanted
complexities. Practically speaking, t .ispaper would have been very much
longer if re-expression was not done
6.
where:
A
e
Y - Y
"N_
The.least
1
r's
Y
_7
e
a minimum
1=1
A
7.
b X
(a + b x
1
,=ie
minimum
and b
in Y such that the squared error term reSultS in the least-poSSible value
Normal Equations
The reader
may wish to read Appendix A at this 'point (or take the next step
on taith).
The normal equations are derived from the least squares criterion
The basic idea that lay behind the technique for
using calculus.
equations for a, b
and b-,;,
na
Y
12.xl
x9Y
'Table
Statistic
Mean
X1
same
Variance
1
n-1
Standard Deviation
(where y = Y-Y)
Y and
)( y
(Xii) (Y-7
Correlation of
Y1
x
(n-1)S.S
(n-1)S S
1
(n-I)S-S2'
fi
15
Note:
(and Y'
9.
1.
In the first normal equation, we recognize that, on the right hand side:
x1- )
E(x
)2
1
or
As for Ex
S2
is:
(n-1)S1
correlation
between x
y.(xA)(x-a2)
x
or:
(n-1)S S
(n-1)S1S-7
(n=1)r12S
Ex
y(x x)
is
Now,
identical to'
or
`'1) (Y -Y)
(where y = Y-Y)
and Y (ryl or r
ly
Hence,
).
x Y
(n-1)S S
1
or
xiY
SYS
(n-1 ) r
y 1
This
PROOF:
1E:(
Now,
n3"-ci
xV
E(X1-7 )(Y-V)
/.(X Y
-)(1)Y
+ TclY )
sfiY +X Y
= yX1
- Y(nX
n7)
-1-4.
TC17
Therefore,
=5q)Y
= 2:(X
) (Y=T)
End of proof.
11.
For the:equation
3.
Simplifications:
translates into
)
x-Y
S -S-
y2 y"2
In addition,
5Ex9-
lx; =
(n-1 )r
12
(n=1)S7
na
b9(0)
(0)
1
(n-l)r
S S
v
a(0)
-=
(n =1
y2 y'.2
b2 (n -1
b (n-1)S
TO further simplify; eliminate zero terms, and for the last two. normal
equations, divide each term by (n-1).
na
EY
-9
r
yi y
y2 y
_;.
S
1
s'T
1
b r
2
b r
2
12
-2
h
12
(n=1)S2-
a(0)
S
2
12.
S S
YI Y
b r
S
I
b r
S S
y2 y 2
S S
19
TheFe are the normal equations we want to work with in the derivation
for two predictors;
Multiple Correlation
We are now ready to define the multiple correlation for one criterion
and two predictors.
RY.x
x2
By definition:
corr(Y,Y) =
2 2
cov(Y,Y)
\ivar(Y) var()
cov(Y,a +
1
var (Y)
x +bx
2 2
1
var (a + b xi + b2x2)
where:
y.x +x
I
IL)
Or
x2
among others.
13.
ahove
cov(Y,a)
var(a)
cov(Y,b1x1)
cov(Y,
Var(b X
var(b ;x
)
1
dvar(Y)
2cov(a,bix )
+ 2cov(a,b2x7)
x )
2cov(bixi,b2x9)
= b1
1
= 0
but
-x
cov(b-1 x1'
cov(x ,x
r12S1S2
r12
var(x ) var(x ?
1
20
14.
04-b1
cov(Ifx)+b2 cov(Y;x-)
2
X9
)
0 + bvar(x
1
1
b2var(x
0 4- 0 4- 2bib2cov(xi,x)
\II
As mentioned, by definition:
cov(Y;x ) =rSS
YI Y
cov(Y,x,)
cov(x ix
=
9
y2 V 2
S1
12 12
of the predictors.
var(X1 - 5Z
var(x1 )
var(X1) + var(R )
2cov(Xt:RI)
is a constant,
Since X
1
var
var(X1) + 0
Si
are
made:
Ry5x
brylSS
+
y
r-S
2 y2 y
x2
b12 S1 2
b22S2
S
2
+ 2bbr
SS 2
2 12
1
NI
convenience.
2i
It will
15.
Diarivation
for
b r
1
Ry.x ,x
S S
yl y
+ hr_SS
2 y2 y 2
S.
Table 2
No-tmai-Ettuations
S S
yl y
S- S
12
b r
2
y2 y 2
h S
12
S
1
b-S
2 2
Multiple Cornelia-Lo
brSS
)12 y 2
hr SS
yl y
RY.x
x
2
; 2
sy
term is
1--
The
NOTE:
a= Y
omitted
b S
2b1h2r12SIS2
Proof involves the substitution of the normal equation into the numerator of
the multiple R formula and simplifying. See text for details.
17.
each term into R and rearrange terms, we obtain the following reSultS:
1
112-
+ b9 S9
7_ 9
b-S-7
y
17-2-S
2b b
+bbrSS
')
12
S`
S
oi2 S")
2b
2- 2
b2 S2
b r
2b
1
b r_
(Hence,
-2-2
12
9 9
b-S-
2b b2 r12S1S2
+ b2r y 2 S yS 2
S S
S S
r
2
2-2
1pS2
2b
1--
\jr.k
18.
Simplifying:
2
b`1 S
2 9
b2S7
2b
r
1
12
S
1
Therefore,
br
S_S
1 yl y
+br2 y2SS
y 2
RY'xi'x2
END OF PROOF1
Obtain a "theap" proof in the analogous standard score regression model:. If variables
are in standard score form, then the standard deviations become unity;
the
1982c
1.
paper
R-
z .z,z 2
Y
y2
19.
the ti,;(-2 of summation which becomes necessary for the general case of
p predictors.
Y = a + b1x1 + b7x7 + b k
3
For
three predictors:
e.'
As in
Y)-
b x- = Le2
- b x
1
a minimum
+b
5.s.
yl y
brS
rSS
y2 y 2
r
b3r23S2S1
rS
b r
_S
S S
2
2 23 2
S
3
--+,
Recall that the value of
is determined in practice
derivation Since it "drops out" in covariance and variance operations of the multiple
R derivation.
The above normal equations are the ones we will make use of in the derivation
of the multiple R formula for three predictors.
on
on the followin'g
page.
Y.x
,x
1
+b
corr(Y b1 x
corr(YiY )
3,
cov(Y;Y)
3 3
2 2
var(Y) var(Y)
All of
cov(h x
h X )
+ 2cov
x2)
+ 2cov
33
hr SS-
y j 1
SS
b r
r -S S
2 y2 y 2
2
3
+2b
r
1
This is
as far as
easy reference;
ti
we can simplify
See Table 3.
S1 S
13
S2 S
2b2 b3 r
23
22
Table 3
Nor-Mal.--1:ytaLi-on,s -and Multiple Correlation Formula for Three RawScore- Pred-ict-orS
Normal Equations'
br SS
S-
-S
12
S2
13
br S
b-r12S1S2
y2
3 23 2
SS
b3 r
2
b r
1
b
13
r
2
S3
Multiple Correlation
brSS
1 _04_4_
242_
2
23
+
1
2b b2r 4S1S2
-r -S
rSS
3-0
2b
_2
b2
2b}b3r1
r
2
S
2 3
S
2
Again) we note that the term a (=Y) is omitted from normal equations and the multiple R.
NOTE:
perivation involves substituting the normal equations into the multiple R and simplifying. See
the text for details.
23.
We have stated the multiple regression model and least squares criterion;
the normal equations and the multiple R fbrMula.
and presented
tlie-fturriglato>/ r
cov(YiY) =
b-r -S S1
h-rSS-
hrS S-
2 y 2 y 2
3 y 3
S S ) +
r
13
(b'S'
+h
12
3 113 223
(1) b
r13S1S3) +
b
1
+b2'p-r S S
rlS
Y 3
r
2
+ b bjr jS2S
ijS-S
+ bjS-
S2 + b2
12
-S S
23
(bi jr jSiSj
jr2j
3
Now
let us write each parenthesised term on a separate line to form a covariance matrix:
./
cov(Y;Y) =
b b r
1)i s1
12'1
9-9
r
12
S
1
2'2
S3
S
2
2- 2
b S
33
1.1
squared terms along the northwest to southeast diagonal of the covariance matrix.
that we might express these terms in summation as follows:
3
2
bS
+b2 S 2
2 2
j=1
S S
2b b r
1
12
2b
r
1
of quantitie:
13
+b
2E 2:1),b.r-..
j
2
i(j
b.j b-r.
It is clear
25.
summation operator is
set to I;
(jr=2,3) giving i j=12 + 13. Now increment i to 2 and complete the limits of j
e.g.,
The subscripts that result from all of the summation operations are:
12 + 13 + 23.
26.
Thus, the nine covariance terms of the multiple R numerator can he written in all of
the following ways:
coy Y,Y)
+ bY
b21 S
-2-
+ 2b
S 5
2b
S S
3
r 3S S3
2b
1
2 2
b b.r..S.S.
j=i
j 1]
brS5 +brSS
2 y2 y 2
1
b r S S
3 y3 y 3
b.r .S S.
Y J
The second
I.22
S
S,
2 2
2 2
+ 2bbr-SS+ 2bbrSS + 2bbrSS
bS+bS+
b-S
2 3 23 2 3
3 3
2 2
3
1
12
-3--2vr- 2 2
b
bar -S S.
b j ij y j
S. S.
j=1
j=2 1=1
3 dI-
If we now form the ratio of covariance and variance terms for the multiple R, we
y7-b 2
-P
El
i=1 J
J72 i71 1 j
ij
,x
111.110,=1.1mmEMMINE
2,x3
b r
j=1 3
J =2i=
S.
ij I J
If we
make the same algebraic simplification we made for the two predictor derivation, we obtain:'
b b,r.
Jj=1
RY.x ,x ,x
2
j=
i=1
=
3
.S S:
J YJ. Y J
END OF PROOF
34
28.
Our
2.
3.
4.
5.
express the
6.
7.
Simplify
A
Y = a + blk1
2 2
P P
3 J
i.= 2
1(Y = Y)-
Le
a minimum
Sbstituting for Y:
1(
1x1
box
J
b-x-)-=
P P
e2
a minimum
,x
See
1
=
na
hl Exl
Ex
2
)c2
-f-...+ b.
x,
+...+ b
Ex
P
Ex Y
Ex
Ex x
Ex x
b
2
+...+ b
3
Ex x. +...+ b
j
Ex,x
p
Ex Y
2
a Ex
2
Ex3Y
b1 Ex x
Ex
a Ex3
Ex x
Ex x
b
3
+...+ b.
3
a Ex
Ex x
2 j
Ex x
1 p
3 p
S S
yl y
+brS
S2
11
12
h-rY
rY3SYS3
S S
YP
=
P
+ b3r3
S2
lj
h-r
-S
+...+
j 2) 2 j
1p 2p 2 p
.
b r
3 +
3 13
b2r23S2S3 + b3S
brSS+br
S 1S
1p
p
1
a3J3J
p3p3p
+ b r
3
+...+ b S
JP j P
P P
S S
3p
2 and 3 predictors, we
3 P
Ex
P
Ex x
Exx +...+ b
+...+ b
Ex x
It we apply the same logic and make the same substitutions we made for
Ex x
p
b3Ex3+...+b,Ex x. +...+ b
b2 Ex2x3
Ex
Ex x, +...+ b
We are no
lc tO-rs--an-dDeri-vatiori
See Table 4
the multiple R.
cov(Y,a +bk
bk
+b2 k
b,x, +...+bk )
P P
br SS +br SS + b-r-SS
1
Y1 Y
2 Y2 9 2
b.r_.S S. +...+br_SS
P YP Y
J YJ Y 3
3 Y3 Y
tN
SS
cov(Y,Y) = b (b,S'T
1
1 2
4--..,--1--
S S) + b (b r
p 1p
2\
S S
-I-. ..+
r '
p 2p
12
...+
p
-2
S2)
b (b-r SS + b- r` y ---i--...-1-- b
P
Multiply
IP
PP
P,-P
I)
each of the b, terms inside the parentheses and write each parenthesiied sum on a separate line:
A
cov(Y,Y) =
+ b2r
-S
b-r
v2 y 2
yS1
12
rSS+
b h
3 13
n ip
sp
3 23 2 3
1 p 1p
bbrS
2 3
22
,..+hbr SI S
13
1S
+1)22
S.
hbr SS.
PYPYP
Y3 Y
+ blb3r
+...+ b b,r
p 20 2 p
S S,
2j
22
+...+ b-r
33
3p 3 p
bb r- SS + hr SS
2
3 3p 3 p
P JP
.J 1:SL
S
I
37
P p
31.
Table 4
Pte44ctd-rs
Co-rtelationFormula
Normal.
Nortif-1--ktiat fori-S
rSS.b S.
yly1
b r
11
rSS .brS
y 2
12
S2
b3r13S1S
b3r3S2S3
12
2 2
+...+ b r
=b1 r13 S
ry3 Sy S
2232
r SS
r-
1 10p3
YP Y P
3J
b-r
b-r
2 20 2p
+...+ b 52
S:S
+;;;+
-S
S S, +...+ b r
P 3P 3 P
3 J
pp
jjp jp
3p 3 p
Mg-tiple COtteLat
RY.x
x ,x ,...,x
2
+,..+bix +.;A b x )
corr (Y; b x + b-
corr(Y;Y)
=
j
22
s3
coV(Yib Xi +
pp
j j
+ b3X3
cov(Y,Y)
Iv
var(Y),
a r 69 var(Y)
var(b x +
1
r
1
...+b,r .5 S +...+b r
x +...+b,x, ...+b x )
PP
S S
SY
b
P
S
2b,b,r,.
+.
j ij i j
:.+
2b
2b
P
b
S
p -1 p p -11P P-
+ 2b
121
r
3
...+
S S
13
YJ Y J
40
32.
To facilitate workihg with such a complex matrix, we will introduce summation at this point;
p x p = p2 terms.
2 2
1
to b SP P
It is evident that
Let X
TOTAL M A T R I X = p
= p + X
or
X
-2
= p -p
X = 0(0-1)
Thus, the entire covariance matrix consists of p diagonal terms and p(p-1)
off
terms.
to i t ;
.S
ij
In general,
There are
halve
S S
p(p-1)
the total p
a right triangle.
matrix
Thus;
of terms.
Suppose we
the p
2
3
terms.
upper
41
33.
mattix (0
2 0(0=1)
[
Examine the matrix of covariance terms for the three predictor case
for further clarification;
A
As explained,
9
there are p
and p(p-1)b.b r
ID'S'
=p2
or 2fP(p-1)/1
S S.
ij
terms;
terms in
2'9
2 2
P P
j=1 J
12
S.S )
r
JP J P
EbbrSS
2E
j=2
i=i
ij
1Fbr those reader_; familiar with combinatoric -; the following may assist
in clarifying the logic.
2
which are combined with all such terms
There is a t_ctal of p b.2 S
2
com-
l(p-1)(p-2)...1
11(p-1)!
terms are
P!
C)
terms
1,4S?
b,2 S
J
terms (pairs
2-
Thus:
4.m
Total number Of
bb,r:,S,S; term8
11
P(p=l)(P=2) (13-3)
2
P-
(: ;)
P(P-1)
2!(p-2)!
(p-2) (p-3)
4;
d9
P(P-1)
terms
34.
For
b-b-r--12 S-S1
'
b b
1
r
2
and
-8 S-
12
b b
10
9 1
and
S
9S 10 .
2E
P -9-2
cov(Y,Y) =
j=1 3
p-1
Ebbj rS-S.
1=1
Equivalentl,
cov(Y,Y
b r
SS +br 2SS
=...+ b,r ,S S.
y2 y 2
1 yl y
3 Y3 Y 3
p
E
j=1
b,1 r
.S
Y3 Y
S.
3
Thus,
cov(Y,Y)
i=
P P-1
E b.b,r..S.S,
E
j=2i=1 1 3 13 1 3
E b,r ,S S1 Y3 Y
3L-1
is simply Sy by definition.
The term,
var (Y)
can be manioulated
by covariance and variance rules to produce the folloWing (See Table 4):
35.
var(Y)
var
JJ
2
2
2 2
...+ b, S, +...+ b S
P P
J
J
+ 2b
+...+ 2b,b,r
S,+...+ 2b -b
1JJIJ
4r 14
r-
S -S
PP-1 PfP-PP-1
In summation notation:
var(Y)
b; S;
J
P=.1
Ebb,rS,S,
2E
13 1 j
i j
j.2 i=1
E 11,S;
-E
Ebbr ,S,S,
.x
x,,...,x,i...,x
1
21.2 i=1 1 J
2 2
E b.S
2 E
_i=1
il 1 J
17F-T7T7"-
E b;b,r,S,S,
j=1 j
j.2 i.1 1 j iJ 1 J
p
2-
p-1
2E
E b.S:
j.i
Ebb,r_S,S,
i 1 1.3
j =2i =l
P
\I'
44
JE
'
b,r
,S
YJ Y
S.
36.
Appendix
Normal Equations
in
Regression
AnalySiS
Introduction
The procedures
general, whenever
re-expression, or, in
For example,
+ b2x2
1
To derive normal equations for a given regression model requires knowledge of elementary differential calculus which makes
1
37.
= a+bx + b x 2
The raw score model includes an error component, and the error
made in prediction of the criterion (Y) with the above model may be
negative, zero or positive.
= Y + 6
Y = e
Ee
The problem with this approach is that the resulting sum on the left
"
side turns out to be exactly zerol;
E(Y-Y)
Ee
That is,
0.
positive errors cancel out negative errors leaving zero as the ove:ali
sum;
bad
is for
its
Proof.
2".(Y-a=bixib2y
E(Y-V-
E(Y-V)
IX,-
blExi
b2Ex2
47
38.
For these reasons, the most widely used and accepted procedure
for finding normal equations is based on the least squares criterion;
F.(Y
- Y)-
E(Y-a -b-
X-
= minimum
Ee
(The summation ranges from 1=1 to i=n o over the entire set of observations
least squares states: find numerical values for a,
In words;
13
and b2
2
which will make the prediction error the smallest possible numerical
upon subStitution.
amount
least squares
one
E(Y-7)2
n-1
Mean, Y,
The arithmetic
is used
the smallest possible value when the mean is used rather than any
other number (or combination of numbers)
Take an example:
.(Y-2)`
(Y-4)
(Y=10)2
(Y-'8)
(Y -11)2
8
10
11
--
(The n-1
(Y -Y)
Find_eath squared sum and compare it against
can be ignored since it is a constant and has no material bearing on
the result).. It will be seen that only
.
(Y-7)
48
-Cy.752
39.
The resulting
In essence, we solve
p+1 equations
Each equation is
Finally,
E(Y -a-bx
x2)
Ee
minimum.
b
1
2.
doing that;
Plan
We will now set down a plan for finding the normal equations.
A four phaSe
40.
normal equations.
^
Y
A.
State
B.
C.
D.
Let us apply the four phase plan firat to the two predictor case.
A.
Y = a
B.
b 1x1
E(Y - Y)2
C.
+ b2x2
E(Y - a -b x- - b x )2
Ee2
Ea
c)
EY
na
Zb-x
b x
b1E xi =
0
2
b2E x2
= 0
d)
EY =
na
+ biE x1
e)
EY =
na
+ b1(0)
b2E x2
+ b2(0)
50
41.
2.
are:
d)
'1
b2xlx2
Eb x2
1
b2 Exix2
:xi +
b2 Exix2
--10
a Z,X
Eb x x
Exl
since
EYxi = (n-OryiSySi
Exi = (n-1)S
and
Exix
and
(n- 1)r12SiS2
(recall that
Ex
(n-Or
0 + b (n-1)S2
_(n-1)r12
=,..0).
1
n-1), we obtin:
-S
Y1 Y
S-
r
2
12
42.
3.
b)
c)
d)
e)
f)
g)
a)
E(Y -a
- b2x2) =
b-1 x1
b)
bixIx2
ax2
d)
EY-x 2
- FA)-,2
e)
EYX2
aEx
f)
E Yx2
since
Ex
1
aEx
c)
-bE x22
?Klx2
b22x22
x1 x
2
2
-Eb2x;
EbiX1
i
u2x2
r
= (n=l)-S-S
y2 y 2
EYx
x=r12 SS 2
2
and
E4
and
(n-1)S1
(n-1)r2S)iS2
bl(1-1)r12S1S2
b2.(n-1)S2
S S
y2 y
b S
2 2
b
2
.0
43.
D.
We now recapitulate.
Y:
and b2 were:
b Ex
= n
EY
For
For
= aEx1
EYX
Ex-
Ex x
b,Ex1.2
For b
Ex
aEx
EYx-
2:
b2r1
S S
yl .y
b r
S S
y2 y 2
12
A
Y
Y+
1
x +bx
2 2
1
53
44.
We will
But
S S
yl y 1
The "multiple" R_in this case is the simple Pearson product correlation,
This is obtained from the second equation.
S
which is equal to
r
b
yl
SY
Thus,
b x
1
S-
4- ry,
xl
Applying the four phase plan gives the following results for the general case.
A.
hk-
b k
h_x_
+...+ 1),X,
P P
B.
:,;(Y
1)-
b,x,
J
P P
C. The procedures for fihding the normal equations for a and any
2.
Finding the normal equation for any b: term can he done in seven steps:
a)
b)
c)
d)
e)
,)
solve L
sunrit
46.
a)
h1
h)
22
-b x,x
j) ...p) p
22j
11j
= 0
pp
= b x x.- b x x
ax
c)
-bpxp)
b x
b-x_
E(Y - a
2--bj
22j
= 0
e)
P ) P
J J
b Ex?
b Ex x
aTx-
ZYx.
b Ex,x- = 0
J P
P
f)
8Ex, + b-Ex-x- + b
ZYx, =
Ilj
,x-x, +...+
2 j
_2
g)
S S
01-1 r
J
1 )
Dividing thrbUgh by
r .S S: =
yi y.3
n-1)
b r
1
13
Thus, the normal equations for any number of predictors in the regression model
consist of
8=Y
56
defined above.
yJyJ
JpJp
47.
Alternate -Proce-dure
,S S.
is a general result.
NoW a
Yj Y
much simpler procedure which makes use of this fatt Will be presented.
Recall that the simple correlation of any variable with itself is equal to 1.
II
jj
is eIual to
= 1.
PP
2
the
COV(X
X )
l'
or,
That is;
in general
cdv
= S2
COV(X ,X
.= S
2
2
1
Cov(x -ik )
is S11
in general
;
we cnn Write
coli(x,Oc.)=S.or
J
S.
Jj
F.
From these facts; it is possible to write down an entire set of normal equations for any nutter
of
predictors.
kids
If r
YJ v J
j=p.
ofixi= 2x2=2'
4 terms,
Thuil
,S S.
yj y j
S,
YJ Y J
br SS
,S S
j
b-r
b r
1
2 ij 2 j
lj
22j
Now; substitute the appropriate j value: j=1 for line 1; and j..=2 for line 2i
twice
as follows:
S S
Y1 Y
r
1
11
+ b r
S S
S S
11
yl y 1
S S
21 2
...
S=b1
OR
r-
S -S
y2 y
b r
S S
b r
2 2 2 22
S1 S2 +1,S2
12
OR
r_ S S
yl y
br- SS
br SS
S S
y2 y
12
12
+
2
The last set shows the subscripts of the correlations between predictors and criterion; and
the predictor standard deviations written so that the first Subscript is less than the second
subscript.
As mentioned
We will show
The long method could be used by applying the steps listed above for any
15,
the shorter method gives identical results, we will not work through the longer method.
We begin by writing out the 52 = 25 terms for the general r_ ,S S. normal ecivation.
nit i.,
YJ Y J
write out r
YJ Y J
br,SS, +
YJ Y
SS
S S
b r
I
2j 2 j
brSS
S S
lj
1j
S S
,SS,
bl
biSS3
3)j3
+ b3rS
S S
+
+ b4r 4i4i:
S S
2 2] 2 j
3j3
b -r
b rSS
b r
S S
3 3j
+ brS
b2r
J 33
+ br
b tSS
+brSS
223 2 j
b
4 4j 4 j
3 33 3
33j3j
I jr
+ br SS. +
+ br,SS
SS
b,
b r
S S
Yi Y
brS S
I j
5 5j 5 3
Substitute the ,appropri ate j value (j=1 for line 1, j=2 for line 2, etc.); set ri
etc;
=1 r22=2)
and Set S
=S
etc.
= S.
22
'
+ br SS
2
.S S.
yl y
S S2
y2 y
S S
r
1
12
biri3SrS3
S S
rY.
21
S2
+ br SS
3 31 3
41
b3r32S3S2
b- r-
5 51 5
b4r42S4
b4
b5r52S5S2
b2r23yj
+ b3S
23
S S
S S4,
S S
Y
Y 5
24 2 4
y4 y
+ br SS
SS
b1 r
b r -S
1
15
25
Y
4
+ bS-
SS
4
'5
60
S S
5 54 5
4 4
4.
+ b3 r,
S2 S
b2 r
+ b-r
b4r45S4S5
1,
"5"05
normal equations that would be obtained if the longer method were used to derive the
h(tMal equations.
The auLhor would be pleased to receive comment and reactions by readers of this paper
and others that appear in this series.
of
textbooks
need to bridge the gap between the standard applied statistics (and psychometrics)
currently on the market and mathematical_statistics. The mathematical sophistication
I would like to hear from readers (students, professors and others) regarding these
Are there proofs that you would like to
For example, are they clear?
papers.
see (statistics or psychometrics) in this format? Please remember, at this time I
am limiting my selections to those which can be presented with algebra.
I welcome comments on any level from readers of these papers. My mailing address is:
Franci J. O'Brien, Jr.
106 Morningide Drive, Apartr6nt# 5
61
10027
51.
Appendix B
ERRATA for " A derivation of the sample multiple Correlation formula for standard scores" ED 223 429
--Nor
rs
Correct to
Reads--.
concepts; notation
concepts, math
...
1_
3; footnote
1-
limit.all together;
limits 4togthser;
cov(Z,y1,B2Z2,B3Z3,.,
Z
J
cov(2,B Z
B Z
2
B.Z.
PP
PP
JJ
corr(Zy,BIZ1,B2z2,B3
22
Corr(yBIZ1
B2Z2
B-P Z-P )
B.Z,
B Z )
JJ%
PP,
add period, D.
demonstrate
demonstrate
31, line 2
consisdered
onSidered
first
erratum
firSt
first
corrected prose is underlined; but focMuIas are rewritten with applied corrections only,
,B Z
3 3
52.
REFERENCES
Philadelhpia: W.B.
Lindeman, Richard H., Ruth Gold and Peter Merenda. Bivariate and Multivariate
Chlcago: Scott, Foresman and Co., 1982.
Analysis
.
1iMitS
DOCUMENT RESUME
ED 280 896
TM 870 228
AUTHOR
TITLE
PUB DATE
NOTE
87
PUB TYPE
EDRS PRICE
DESCRIPTORS
IDENTIFIERS
ABSTRACT
This paper is pa4.-t of a series of applied statistics
monographs intended to provide s:applementary reading for applied
statistics students. In the present paper, derivations of the
unbiased standard error of estimate for both the raw score and
standard score linear models are presented. The derivations for raw
score linear models are presented in graduated steps of generality
for one, two, three, and any finite number of predictors. A brief
overview of regression analysis precedes the derivations. Appendices
include: (1) errata for a derivation of the sample multiple
correlation formula; and (2) a discussion of linear and nonlinear
regression models. (LMO)
**********************************************************************
Reproductions supplied by EDRS are the best that can be made
from the original document.
***********************************************************************
Fj O'Br
Tr.
1 987
RESERVED
4,
2nd equation
(X - X )
1
6,
CHANGE TO
NOW READS
PAGE
definition
2
Y.x ,x
of R
definition
of r
9,
3xY
91
9,
Ix
9
1
footnote b
...,x
2
9, definition
of
x
1
10,
ist equation
Y.x
Y.x1
Ix
Ex y
18,
2nd equation
2 2
2 2
2 2
+ b
(b S
1 2
2 2
2b b r
S S
1 2 91 9 2
-2(Ebr SS)
j'l
9j Y
2b b r
S S )
1 2 12 1 2
2 2
b S
2 2
(b S +
1 1
-2(Eb r
S S
j41 j 9j 9 j
PAGE
CORRECT TO
NOW. READS
34
'2
1 -
1 - 1
br SY
j 9j Y
br
j
j.
SS
9.1
9 j
Table of Contents
Page
Introduction
Overview of Derivation
14
20
29
36
Introduction
36
37
40
41
42
Notes
46
References
48
List of Tables
Page
Table
2.
17
3.
Functions of R
4.
5.
Functions of R
6.
Functions of R
1.
for Two
for Three
19
25
27
ii
34
Introduction
The present paper assumes familiarity with two previous papers in this
series (O'Brien, 19820, 1983a). Each paper formulated a detailed derivation
of the multiple correlation formula of one criterion and p predictors for the
linear model.
The first"
paper (1982c) presented a derivation of the
multiple R based on
1
Overview of Derivation
is linear in form.
The mathematical model we might select
to"fit" such a distribution is the simple linear equation:
A
a + b X
1 1
Where:
A
Y
i.e.,
x -
where
1
"T
A
Such deviations from Y
are considered errors of prediction.
We can
conceive a raw scor :? observation as consisting of a component predicted by the
model plus an error component. That is:
+ e
Where:
Y
A
the idealized mathematical model
actual raw score criterion (Y).
(Y)
to predict the
The error made in predicting the observed criterion score by the model
is
simply:
enY-Y
One of the goals of regression analysis is to minimize the prediction error
denoted by e above. It can be seen that if e.0, then the actual criterion is
perfectly predicted by the selected mathematical model.
That is to say, the
simple linear
equation fitted to the observed data points,
a + b x
1 1
in the distribution.
Geometrically,
terms.
The technique most often used in the social sciences -`o minimize the
error of prediction is the "least squares" procedure.
Essentially,
this
procedure seeks to maximize predictability by minimizing prediction error.
The least squares criterion or goal is summarized in the following
expression:
a minimum
Ie
(Y-a-b x )
(a+b x
1 1
minimum
(As an aside, "least squares" means we determine values for a and b such that
the squared error term results in the least possible value).
The Standard Error of Estimate
...
(Y-Y
(X
1-17-(a+b
.)
X )1
E(Y- a
b x )
1 1
n-2
where:
Y.x
1
is
in deviation form.
10
n-1 (1-r
x y
Y.x
n-2
S
1
where:
S
x Y
1
11
number
Y.x ,x
,...,x
X(Y- Y
.411".11
n-(p+1)
(1- R
(n-1)
Y.x
, x
n-(p+1)
where:
Y.x ,x
1
Y.x ,x
The standard error of estimate also can be de:ived for regression models
in which the variables have been expressed in standard score (Z) form.
The
unbiased sample standard error of estimate for a one predictor standard score
linear
is defined as
model
n-2
2.
E[zy
(A +
iij
n-2
where:
weight.
n-1 (1-r
Z .Z
Y
Z ,Z
3.
n-2
13
to:
where:
and Z
1
1
n-1
Z .Z ,Z
Y
,...Z
2
n-(p+1)
2
-
Z .Z
Y
...,Z
...,Z
where:
...,Z ,...,Z
.Z ,Z
.Z ,Z
Z
Y
14
Table 1
Regression Model
a+bx V+ r
1 1
91
1
1
Variance of Y:
r(Y-v)
n-1
n-1
2
Variance of X
n-1
n-1
Correlation
x
and
of
:
(n-1) S S
91
Y 1
i.1
r .(Y -a-b x )
1.1
i.1
1 1
n
N
minimum
i
9 or IX Y, where
Y,
X - 7
x
1
15
and 9 .Y-Y.
expression
10
(or r
).
92
yl
standard error
of estimate:
Substituting for
Y.
1:(Y -
a - b x )
1 1
Y.x
n-2
2
2
Y.x
a - b x )
1 1
(Y
1
n-2
a,
16
is
11
2
2
p-7)
Y.x
b x )]
1 1
n-2
to simplify the
Y-Y.
E (y-b x )
Y.x
1 1
n-2
2
2
Y.x
+ b
1
2
x
n-2
- 29b x )
1 1
12
2
2
Y.x
2
1
x 9)
-2b
n-2
(n-1) S
N:x
(n-1)S
1
91
(n-1)r
Y
1
S S
91 9 1
Thus:
Y.x
(n-1)S
S / S )
(r
91
(n-1)S
n-2
-2(r
S /5
91 9 1
(n-1)r
S S
9191
18
)]
13
2 2
(S /S ) S
(n-1)
y1
2r
y1
(n-2)
Simplifying:
(n-1)
(n-2)
[li
(n-1)
Y.x
91
9
1
2r
9
91
91
2
S
1.
(n-0
91
(n-2)
Taking the (positive) square root, the unbiased standard error of estimate
for one raw score predictor is:
Y.x
ENO OF PROOF
(n-1) 1
(n-2)
19
14
Y.x ,x
(n-1)
- R
Ii
.x ,x
Y
(n-3)
where:
Y.x
x
1
form)
A
Y.x ,x
1
(Y-Y
E(Y-Y
n-(p+1)
n-3
2
BY
-a -bx- bx
1 1
2 2
n-3
20
15
2
2
S
Y .x
SE(Y- a- bx -bx)
, x
2 2
1 1
r.w.gw.m.m.Mgmm.wa./Mr.
n-3
Substituting -sir
S
1
-bx -bx)
Z[Y-7)
Y.x ,x
1 1
2 2
n-3
in
Y.x ,x
1
n-3
[y+bx +bx
1
- 2yb x
2 2
1 1
-2yb x
+ 2b b x x ]
1 2 1 2
21
16
2
S
--
Y.x ,x
1
n-3
1
+
2b b
1 2
7x
( 1:9 + b
rx
" 1
+ b
5-
2 ' 2
x )
1 2
(n-1)S
L9
2
(n-1)S
7Ex
(n-1)S
2
EA2
S S
(n-1)r
91
9 1
92
S S
9 2
12
S S
1 2
(n-1)r
(n-1)r
24 2
17
Table 2
Substitution Equations for Two Predictor Raw Score Model
(n-1)S
9
2
(n-1)S
1
Ex
(n-1)S
2
Ex
y1
S S
y 1
y2
S S
y 2
12
S S
1 2
(n-1)r
(n-1)r
2
yx x
(n-1)r
1 2
E(Y-7) /(n-1)
"
Ey /(n-1)
Solving in terms of
(n-1)S
23
18
+ (n-1)b
1 1
+ (n-1)b S
fin-1)S
S
2
n-3
-2(n-1)b r S S
1 91 9 1
2(n-1)b r
S S
2 92 9 2
+ 2(n-1)b br SS
1 2 12 1 2
r2
(n-1)
Y.x ,x
(n-3)
2
+
(b
1
2
+
b
2
-2(brSS+brSS
1 91 9 1
2b br SS)
1 2 91 9 2
2
)1
2 92 9 2
R or R
Table 3 shows forms of R which will be used in the
next step. (Compare O'Brien, 1383a, pages 12-18, especially p. 18).
.
24
19
Table 3
2
Functions of R
bS +bS
2
+ 2bbrSS
1 2 12 1 2
br SS +br SS
1
'21
y 1
2 y2 y 2
Rearranging:
2 2
R S
b
1
+ 2bbr SS
Note: R
+b
1 2 12 1 2
2
a
Y.x ,x
1
25
br SS +br SS
1 y1 y 1
2 y2 y 2
20
Thus,
bS +bS +2bbr SS
R S
1 2 12 1 2
br SS +br SS
2 92 Y 2
1 91 Y 1
Y.x ,x
1
n-3
+ S
25
R
9
n-1
2-1
n-3
n-1
1- R
n-3
Y.x ,x
Taking the positive square root, the unbiased standard error of estimate for
two raw score predictors is:
(n-1)
Y.x ,x
1
-------
[1-
END OF PROOF
Y.x ,x
(n-3)
26
21
2
S
Y .x , x
1
, x
n-1
Y.x ,x
n-4
CY
Y.x ,x ,x
1
n- (p+1)
-bx -bx)
1
2 2
3 3
n-4
27
22
a- bx -bx -bx)
2
3 3
2 2
3.
Y.x ,x ,x
n-4
as 9:
2
S
Y .x
1
- bx -bx -bx)
3 3
2.2
1 1
, x
, x
and
n-4
[9
-bx - bx -bx
3 3
2 2
1 1
n-4
2
3.
Y.x ,x ,x
1
n-4
LI
2
+b
2
x
1 1
2
x
3.
- 29b x
2
+
29b x
2 2
2
x
- 296 x
3 3
+ 2b b x x + 2b b x x + 2b b x x
2 3 2 3
1 3 1 3
1 2 1 2
Bringing the summation operator inside:
28
23
Y.x ,x ,x
2
n-4
2
+b 5x +b yx
+b
1
rx
- 2b
1
- 2b
x y
2
2b brxx
1 2
1 2
x y
2b
2b b
1 3
+ 2b b cx x
1 3
2 31--
2 3
(n-1)
ru
2.
For any x
2
For any A lc
Y 4
EA X.
(n-1)r
S S
Y
j
For any x x
i
S S
(n-1)r
i
ij
29
24
2
2
Y.>: ,x ,x
1
En-1)5
3
n-4
(n-1) b
1
- 2(n-1)b r
+ (n-1) b
SS
- 2(n-1)b r
+ (n-1)b
2
1 91 9 1
2
S
SS
2 92 9 2
- 2(n-1)b r
SS
3 93 9 3
1 3 13 1 3
30
2 3 23 2 3
25
Table 4
Generalized Substitution Equations For Raw Score Model
(n-1) Ly
9
(n-1)
Ex
S S
(n-1)r
9j 9
S S
(n-1)r
CE:x x
j
ij
31_
26
n-1
Y.x ,x ,x
n-4
S +b S +b S
+(b
1
'3
2bbr SS
SS
.r2bbr
1 2 12 1 2
2bbr SS)
1 3 13 1 3
2 3 23 2 3
)1
3 93 9 3
2 92 9 2
13830:
n-1
maillma.11=1111Mlis
x ,x
Y.x
1
Sy
n-4
3
3
+
j1
+ 2 1 r bbr SS)
j2 1.1
ij
3
-
2 (' z br SS ) 1
ji
9j V
in summation notation.
32
27
Table 5
a
j'1
For Three
Functions of R
bbr SS
2 E.
j'2 i'1
ij
2: b r
j.1
S S
yj Y j
Rearranging:
2 2
RS
j1
Note:
j=2 in1
Y.x ,x ,x
a
gEbr SS
2 'Er bbr 5 $
33
ij
j'l
9j
28
Thus:
2
R S
b
j
bbr SS
ja2 ia1
ij
br SS
j
9j
Substituting:
n-1
Y.x ,x ,x
1
2 3
2
S
n-4
2
+
Simplifying:
34
2
R
2S
R
9
29
IiR
n-1 S
Y.x ,x ,x
1
n-4
or
n-1 S
Y.x ,x ,x
1
[1
Y.x ,x ,x
n-4
END OF PROOF
n-1[. - R
Y.x ,x ,x
Y.x ,x ,x
1
n-4
35
30
S
1
1 n-1
n-(p+1)
2
)
n-(p+1)
1: [Y- a-bixi-b2x2-...-b.x.-...-b
J
P P
n-(p+1)
I--
1 1
2 2
p p
n-(p+1)
Now replace a by V
1:
1 1
2 2
n-(p+1)
36
P P
31
X
Y .x
, x
n- (p+1)
E(9 +b
1
+...+ b
+...+ b
ii
2
x
29b x -...-2yb x
j
j
P P
- 29b x
2 2
- 29b x
1
+b
+ 2b b x x + 2b b x x +...+ 2b b x x
1 2 1 2
1 3 1 3
ijij
+...+ 2b
b x
x )
p-1 p p-1 p
Y .x
, x
n- (p+1)
rx
( E9 + b
1
+ 2b b
1 2
2
x
1 2
r.x
+...+
...-2b
...-2b ST' yx
sx
j/
+ 2b b
+...+ b
2
2b 7yx
1
- 2b \5,
1
x x
+..
P
b
1 3' 1 3
+...+2b
.! 4 i
b 1:x
p-1 p
37
x )
p-1 p
we
32
Y.x ,x
...,x
n-(p+1)
(n-1)S
(n-l)b
- 2(n-1)b r
S S
+...+
- 2(n-1)b r
S S
2 y2 y 2
1311
(n-1)b
(n-l)b
2(n-1)b r
''
S S
.-
JJJ
S S
2(n-1)b r
P YP
P
S S +...+
ij
b r
2(n-1)b
n-1
Y.x ,x
n-(p+1)
2
[S2
2
+
(b
S
1
b
2
+...+
+..+
ii
2b b r S S + 2b b r S S +...+ 2b b r
1 2 12 1 2
1 3 13 1 3
j
i
b r
2b
2
S
+...+
S S
ij
S )
SS +...+br SS)
j9j9j
P913913
SS +br SS +...+ b r
y1 y 1
222
38
33
n-1
2
S
n-(p+1)
P
S
9
2...
(L
j1
2(r
P
J*1
2
S
bbr SS)
j2
jiji j
br SS )1
j
9j
39
for p
34
Table 6
2
Functions of R
for p Predictors
jl
bbr SS
j2 i1
ij
r br SY
j1
9j Y
Rearranging:
RS
jal
Note:
E b
2
S
2 LEbbr SS
p2 i1
40
ij
br SS
j
jm1
9j 9
35
Thus:
2 2
RS
1E:b
9
p-1
+ 21bbr SS
j.2 i1
ij
'tbr SS
j=1
9j 9
n-1
Y .x
, x
n- (p+1)
n-1
x ,
2
S
Y.x
- 2R S
+ R S
9.x
n- (p+1) 9
x
2
Therefore:
n-1
s
x
Y .x
x
j
R
9
Y.x
n- (p+1)
END OF PROOF
41
x
1
36
Introduction
42
37
2
n-1.
[1-r
--9
n-2
2
r
Z ,Z
Y
91
1
43
38
n-1
1-r
Z ,Z
n-2
9
n-1
1-r
x y
n-2 L
1[7::
x y
n-2
44
39
n-1
,...Z 2
.Z ,Z
Y
n-(p+i)
Z .Z
Y
Z
2
(n-1)
[1- R
Y.x ,x
1
,...,x.
2
n -(p+1)
also is equal to 1.
45
40
may be helpful
in working through
46
Appendix A
Errata for "A derivation
of
ED 235 205
Correct to
Nol.irReads
10, footnote,
3 lines down
X Y
X Y
10, footnote,
4 lines down
n X Y
13
var (b ,x )
n X Y
1
var(b x )
2 2
16, footnote,
last 2 lines
17, footnote
Multiple R
multiple R
24, footnote 1
Omit this.
29, equation
x Y
b
p
x
p
bbr SS
bbr SS
2 p 2p 2 p
2 j 2j 2 j
34,2nd equation
x
p
change = to +
=...+b r S S
j yj Y j
mathematical calculus
2
mathematical statistics
2
47
Page number-at -top of text:-
4 2
Appendix
2
Y
f (X)
-3 < X
MM.
48
< 3
111
4 3
i(X)
a + bX
exp(b
+
1
e)
Where:
exp
b,,b
XI
2-
(1)
4 4
lnExp(b
lnF
b X
+
1
e).)
2
2
b X
2
(2)
lnF,
Let:
Ymb + bX
1
+e
(3)
To take a second example (also from Draper and Smith), consider the
following regression model:
X]
exp(-b X) - exp(-b
b - b
1
+ e
(4)
Where:
exp
b ,b
1
45
46
Notes
Page
See O'Brien (1983a, Appendix B) for an errata sheet.
references given in the errata pertain to the original pagination
(i.e., at the top of the page).
Errata for this paper are given in Appendix A of the present paper.
5
If it is understood that the summation limits range from the first
observation (im1) to the last (imn), then we can drop the summation
limits; n refers to the total number of observations for the
criterion and predictor(s). This sample size is the same
Later when the algebra
regardless of the number of predictors.
becomes more complex, we use summation limits extensively.
6
52
47
ZA
A+BZ
1 1
is:
+ e
where:
Z"
where S
(x -7 )/S
z
1
is the
1
standard deviation of X
1
53
48
References
Kendall, Maurice G.
and Alan Stuart. The Advanced Theory
of Statistics: Inference and Relationship, Vol. 2.
(2nd ed.). New York:
Hafner Publishing Co., 1967.
Nie, Norman H. et al. Statistical Package for the Social Sciences (2nd ed.).
NY: McGraw Hill Book Co., 1875
2
1882b.
54
.7
-DOCUMENT'RESUME
A
SE 037 098
AUTIOR
TITLE
Case.
,24 Apr 82
PUB DATE
NOTE
EDR PRICE
DESCRIPTORS
20p.
.
'
ABSTRACT
MO'
***********************************************************************
Reproductions supplied by EDRS are the best thtt can be made
*
from the original document.
****************-*******************************************************
A Proof that t
onognaung a
-_
Francis
F.
root NIP
POSMon Or poky
.1t is well known that a researcher wilt) has. collected data from two
independent groupsmA,
This is because
For example,
the equivalent
= F or t = (F)
1/2
.
....
or psychology
of the relationship (when sample sizes are equal). (See, for example,
.
N.
First, many
prior statements of the proof for the general case (of unequal sample
sizes) are either abbreviated, mathematically inaccessible or incomplete
the literature
'
that many students can follow well-articulated proofs and desire to see them
worked out.
/
In this paper
three
Firsty
Third,
a plan will be
be able to follow the proof of the general case, and therefore, shOuld
Proof that t
= F: the Gen4ral
j
Case
2.
3.
4.
5_.
observe that t
t8 the squared t
3.
Table 1
Table Layout
oup 1
Group 2
n.
X11
12
2.
21
22
t..
31
41
n.
n.
2
32
3.
42
AP
+
X.
n.
1
n. 2X.2
'2
n.1 ,+
n.
2
X.
12
Ime
4.
s ?.
is not needed
nl
Total
Sample
Size
n.
Sample
Mean
i.
n.
x..
X.2
Sample
Variance
n..
s.
s.
41
s..
N-
fr
5.
4-
\21'
This is the expresion we will use as a basis oecomparison with the.
simplified F statistic to be obtained in step 4.
2
.
. o
Step 3.
the F statistic
It.\ \will be recalled that the F statistic is the ratio of two independent :
1
).
df
is:
SSb/df
SS
w/df
The genera1.form'ofdfb'is "the number of groups minus one" (i.e., dfb = J-1,
where J is the number of groups).
4
two groups is df
= 2
1 = 1.
= n
+ n.
.1
df
='n
'1
J.
2
= n.. - J).
is
+ n.
'will be used instead of\the correct "among": 44afocept the righteous indig.
fr
5.
(N
Step 3.
2
.
4%
1.
the F statistic
).
sums of sqdares
( SSb )
df
and df
w (within sums of squares degrees of freedom).
is:
SS.
b/df
FJ- b,n.. -J
SS
w/df
= J-1,
two groups-is df
= 2 - 1 = 1.
an Write
= n.1 + n.
J.
2
= n.. - J).
Since n.! = n.
+ n.
1
dfw,
44
ad
-6.
sb/i
F
1,n.
+ n.
1
- 2
2
ss
SS
SS
w
- 2)
+n.
(n.
1
familiar
That is:
n.
1nl+n. -2
+
1
n.
-X..)
(Z(
-1)s.
(n.
1
-1)s.
(n.
2
n.1 + n.2 2
'
7.
Table 2
L.
Restatement of t
and F
t2
F
41.
z
.-X.
(X.
1
(n: -.1)s.
(n.
+ n.2\
)s..
(A
-1)s.
:2
-R.:)2
+ (n. -1).
2
-2
,n.1
+ n.
(i.
+,n.
+
n.
(n.
)(n.
+ n.
n.
We will first
A full step-by-step
simplify the ,numerator
c..
4
4..
Notes,
n.
SS
n.
(;
1 "" 1
R..
(R..R..)2
n.
""=
n:1
+ n.
n.
1
n.2 4X
n.
n.2 /
(n.1X.1 + n.2X.2
(n.1 +n.2)X.1
+ n.
1
n.2
+n.
(n.
1
))-(.
- (n.1X.1
R.
n.
2
2!
+ n.
1 n.
1
].
n.
L
SS
P n.
n.
X.1
+ n.
1
Sat.
=
n.
n. 2X.21
+ n. X.
R.
1
AP
V.
+ n.
.n.
1
SS
- n. X..
tl
n.
1,
- n. X.
1
4/
n.1 + n.
2
+ n.
R.
n.
n.
2
n. R.
R.
1
n.
1
2X.2
2
Cancellipg.like teifms,
+ n.
n.
n.
n.
1
terms,
+ n.
n.
1
n.
SS
-R.
(R.
+ n.
n.
1
2
2
n.
rR. --)
n.1
n.2
10Na
SS
(h.
n.
2
1
(n.
(n.
2'
+ n.
1
/n.
qo(n.1) (n.2)
n.2(X.1 -X.2)2
n.
+ n.
-R.
(R.
1
1-
)2
1
-X,
(X .
b
(n.
SS
brackets,
(X.
+ n.
(n.
)(n.
(n.
+,n.
1
(R.2-R.1
1
(n.
SS
-X.
(X.
1)
y
lk
(n.1 + n.2)2
is the
411,
to obtain,
9.P
)(n..
(h.
1
(n
SS
+ n
+ n.21
(n
(X.1 1
)2
'2
and (n .
+ n .
1
) 2',
we obtain, 16-
2-
)(n. 2)
(n.
SS
(R.
1
-R.
)2
2
n.1 + n.,
4.
1_3
-14
'so
p
C
0/
11.
)(n.
(n.
1
(X.
n
'1
k
0
-1.)s..
(n.
).
-1)s.
(n.
-X.
1
(n.
n;
(R.
1
F
2
-1)s.
-R. )2
+ n.
Step 5.
observe that t
- 2
= F
t (n.
+ n.
n.
.
ENO OF PROOF.
-1)s2
(n.
n.
we obtain,
4-n.
+ n.
n.
(n.
).(n.
) (n.
form, signifying that they'are equal. Refer to Table 2 for this comparison.
This completes the entire proof in accordance with the five step plan.
15"
16
4
12.
t
Numerica1 Example
The analytib
proof using algebraic rules for the general case
.
41,
insight.
only.
Table 3
Group 1
"`Group 2
10
50
40
70
20 "
100
50
40
'QM
75
ieb
4
.
90
'Total
Sample
Size
10
Sample
Mean
07.5
65.0
54.5
Sample
Variance
957.5,
700.0
NOT NEEDED
13.
-,
value is:
47,.5 - 65.0
t
(6-1)957.5 + (4-1)700.0
10
6887.5
J.
t
10
24
-.9240
1%.
(-.9240)
.8537
(t
= .8537)
.We place the computed value in the margin for easy reference.
Now
+ 4(65.0-54.5)
(6-1)957.5 + (4-1)700.0
\_J
6 + 4
6(49) + 4(110.5)
F
6887,5
8
.8537
(F = .8537)
18
./
Thus, t
14.
= F.
=F is somewhat easier to
general case.
= n.
n.
R.2
R.
2.
X.. =
Thi5 value
de`
By making these two changes and by following the five step outline
used for the general case proof, the reader should be able to derive the
proof for the special case., One may.also wish to "make up" an.easy to work with
numerical data set to check
on the process.
I
0
Note
two
EdwardAll'en.L.
References
Glass, Gene V and Stanley, Julian_ C. Statistical Methods in Education
and Psychology. Englewood Cliffs, New Jersey: Prvitice-Hall, 1970.
Rocca, Anthony J. and Tweney, Ryan D. Analysis oalvariance and the "second discipline"
of Scientific isychology: a historicaraccount. Psychological
Bulletin, 1980, Vol.,87, No. 1, 166-184.
O
'fp
a MD