Beruflich Dokumente
Kultur Dokumente
Multiple Regression
1. The model
2. The Least Squares Method of Determining Sample Regression Coefficients
3. Using Matrix Algebra to Solve the System of Least Squares Normal Equations
4. Standard Error of Estimate
5. Sampling Properties of the Least Squares Estimators
5.1. The Variances and Covariances of the Least Squares Estimators
6. Inferences about the Population Regression Parameters
6.1. Interval Estimates for the Coefficients of the Regression
6.2. Confidence Interval for the Mean Value of y for Given Values of
6.3. Prediction Interval for the Individual Value of y for Given Values of
6.4. Hypothesis Testing for a Single Coefficient
6.4.1. Two-Tail Test of Significance
6.4.2. One-Tailed Tests
6.4.2.1. Example 1: Testing for Price Elasticity of Demand
6.4.2.2. Example 2: Testing for Effectiveness of Advertising
6.4.2.3. Example 3: Testing for a Linear Combination of Coefficients
7. Extension of the Regression Model
7.1. The Optimal Level of Advertising
7.2. Interaction Variables
7.3. Log-Linear Model
8. Measuring Goodness of Fit2
8.1. Adjusted 2
1. The model
In multiple regression models we may include two or more independent variables. We will start with the
model that includes two independent variables 2 and 3 . The main features of and assumptions regarding
the simple regression model can be extended to multiple regression models. Here the dependent variable is
influenced by the variations in the independent variables through the three parameters 1 , 2 , and 3 . The
disturbance term u is still present, because the variations in y continue to be influenced by the random
component . Thus,
= 1 + 2 2 + 3 3 +
() = 1 + 2 2 + 3 3
= 1 + 2 2 + 3 3 +
To estimate the parameters from the sample data we need formulas to determine values for the estimators of
the population parameters. The estimators are the sample regression coefficients 1 , 2 , and 3 .
= 1 + 2 2 + 3 3
Substituting for in the residual equation above, squaring both sides, and then summing for all , we have,
= 1 2 2 3 3
2 = ( 1 2 2 3 3 )2
Taking three partial derivatives, one for each coefficient, and then setting them equal to zero, we obtain three
normal equations.
2
= 2( 1 2 2 3 3 ) = 0
1
2
= 22 ( 1 2 2 3 3 ) = 0
2
2
= 23 ( 1 2 2 3 3 ) = 0
3
1 2 2 3 3 = 0
2 1 2 2 22 3 2 3 = 0
3 1 3 2 2 3 3 23 = 0
To find the solutions for the three s, we can write the normal equations in the following way, by taking the
terms not involving the to the right-hand-side.
1 + (2 )2 + (3 )3 =
(2 )1 + (22 )2 + (2 3 )3 = 2
(3 )1 + (2 3 )2 + (32 )3 = 3
Here we have a system of three equations with three unknowns, 1 , 2 , and 3 . Now we need to develop a
method to find the solution for these unknowns. For this, a brief introduction to matrix algebra is called for.
(See the appendix at the end of this chapter.)
2 3 1
[2 2
2
2 3 ] [2 ] = [2 ]
3 2 3 32 3 3
where,
2 3 1
= [2 2
2
2 3 ] = [2 ] = [2 ]
3 2 3 32 3 3
Solutions for are obtained by finding the product of the inverse matrix of , , times .
Example:
We want to obtain a regression of (), in $1,000's, on (2 ), in dollars, and
(3 ), in $1,000's, of a fast food restaurant. The data is contained in the Excel file CH6 DATA in
worksheet tab burger.
Use Excel to compute the values for the elements of the matrix and :
2 3 75 426.5 138.3
[2 22 2 3 ] = [426.5 2445.707 787.381]
3 2 3 32 138.3 787.381 306.21
5803.1
[2 ] = [32847.7]
3 10789.6
Since = 1 , we need to find 1 . Now that we know what the inverse of a matrix means and how to find
it, we have Excel to do the hard work for us. In Excel use the array type function =(). You must
first highlight a block of 3 3 cells, then call up the function =MINVERSE(). When it asks for array, you
must lasso the block of cells containing the elements of the matrix and then press Ctrl-Shift-Enter keys
together. The result is the following:
2 = 7.908 implies that for each dollar increase in price (advertising held constant) monthly sales would fall
by $7,908. Or, a 10 increase in price would result in a decrease in monthly sales of $790.8. 3 = $1.863
implies that (holding price constant) for each additional $1,000 of advertising sales would increase by $1,863.
The following is the Excel regression output showing the estimated coefficients.
SUMMARY OUTPUT
Regression Statistics
Multiple R 0.669521
R Square 0.448258
Adjusted R Square 0.432932
Standard Error 4.886124
Observations 75
ANOVA
df SS MS F Significance F
Regression 2 1396.5389 698.26946 29.247859 5.04086E-10
Residual 72 1718.9429 23.874207
Total 74 3115.4819
2 ( )2
var() = =
2 2
The square root of the estimated variance was called the standard error of estimate, se(). Here with two
independent variables, the least squares residuals take the same form:
= = (1 + 2 2 + 3 3 )
2 ( )2
var() = =
where is number of parameters of the regression being estimated. Here = 3, for estimating 1 , 2 , and 3 .
Given the estimated regression equation in the last example,
the predicted value of (), for price 2 = $6.20 and advertising 3 = $3.0 (observation #10), is:
The observed value of sales is 10 = $76.4. Thus, the residual is calculated as = 76.4 75.473 = 0.927.
Computing such residuals for all 75 observations and finding the sum of squared residuals, the estimated
variance and the standard error of estimate are, respectively
( )2 1718.943
var() = = = 23.874
72
se() = 4.886
Once again, the linearity assumption is important for inferences about the parameters of the regression. If
the disturbance terms , and correspondingly the values, are normally distributed about the regression
plane, then linearly related coefficients are also normally distributed. If the disturbance terms are not
normal, then according to the central limit theorem, the coefficients will be approximately normal for large
samples.
From the covariance matrix on the extreme right hand side we have the following variances and covariances:
For 1 : , = 1 2, se(1 )
For 2 : , = 2 2, se(2 )
For 3 : , = 3 2, se(3 )
The standard error for each regression coefficient can be obtained from the variance-covariance matrix. The
95% interval estimates for the slope coefficients in the above example, based on the following information,
are given below.
, = 2 0.025,(3) se(2 )
, = 3 0.025,(3) se(3 )
An increase in advertising expenditure of $1,000 would increase sales between $501 and $3,225. This is a
relatively wide interval and, therefore, does not convey good information about the range of responses in
sales to advertising. This wide interval arises from the fact that the sampling variability of the 3 coefficient is
large. One way to reduce this variability is to increase the sample size. But this in many cases is impractical
because of the limitations on data availability. An alternative way is to introduce some kind of non-sample
information on the coefficient, to be explained later.
6.2. Interval Estimate for the Mean Value of y for Given Values of
The mean value of for given values of the independent variables is 0 , the predicted value of for the given
values x, x, ... The confidence interval for the mean value of y is therefore,
, = 0 2,() se(0 )
To determine se(0 ), start with var(0 ). For a model with two independent variables,
0 = 1 + 2 02 + 3 03
var(0 ) = var(1 + 2 02 + 3 03 )
2 2
var(0 ) = var(1 ) + 02 var(2 ) + 03 var(3 ) + 202 covar(1 , 2 ) + 203 covar(1 , 3 ) + 202 03 covar(2 , 3 )
To build a 95% interval for the mean value of in the model under consideration,
we have,
var(0 ) = 1.5407
0 0 =
Therefore,
se(0 ) = 5.0413
= 1 0
= 0.42 + 0.83
, = 0.025, se()
We are 95% confident the mean increase in sales will be between $3,239 and $6,068.
Generally, a computer output such as the Excel regression output provides the test statistic for the test of
null hypothesis 0 : = 0 against 1 : 0. If the null hypothesis is not rejected, then we conclude that the
independent variable does not influence . The test statistic for such a test is
= || =
se( )
And, since this is a two-tail test, the critical value, for a level of significance , is = 2,() . For example,
0 : 2 = 0 1 : 2 0
= 7.9081.096 = 7.215, and = 0.025,(72) = 1.993. Since > , we reject the null hypothesis and
conclude that revenue is related to price. Also note that the probability value for the test statistic is 2
P( > 7.215) 0. Since this is less than = 0.05, we reject the null. The test for 3 is as follows:
0 : 3 = 0 1 : 3 0
= 1.8630.683 = 2.726 and = 1.993. Since > , we reject the null hypothesis and conclude that
revenue is related to advertising expenditure. The probability value for the test statistic is 2 P( > 2.726) =
0.0080.
All these figures are presented in the Excel regression output shown above.
We will use the current example to provide an example of one-tailed test of hypothesis in regression.
Suppose we are interested to test the hypothesis that the demand is price inelastic against the alternative
hypothesis that demand is price elastic. According to the total-revenue test of elasticity, if demand is
inelastic, then a decrease in price would lead to a decrease in revenue. If demand is elastic, a decrease in
price will lead to an increase in revenue.
0 : 2 0 (When demand is price inelastic, price and revenue change in the same direction.)
1 : 2 < 0 (When demand is price elastic, price and revenue change in the opposite direction.)
Note that here we are giving the benefit of the doubt to inelasticity. If demand is elastic, we want to have
strong evidence of that. Also note that the regression result already provides some evidence that demand is
elastic (2 = 7.908 < 0). But is this evidence significant?
Since by the null hypothesis (2 )0 is equal to (or greater than) zero, the test statistic simplifies to =
2 se(2 ). However, since this is a one-tailed test, the critical value is = ,() .
This leads us to reject the null hypothesis and conclude that there is strong evidence that 2 is negative, and
hence demand is elastic.
We can also perform a test for the effectiveness of advertising. Does an increase in advertising expenditure
bring an increase in total revenue above that spent on advertising? That is, is 3 > 1? Again, the sample
regression provides that 3 = 1.863 > 1. But we want to prove if this evidence is significant. Thus,
0 : 3 1
1 : 3 > 1
3 (3 )0 1.862 1
= = = 1.263
se(2 ) 0.683
Since = 1.263 < = 0.05,(72) = 1.666, do not reject the null hypothesis. The p-value for the test is P( >
1.263) = 0.1053, which exceeds = 0.05. The test does not prove that advertising is effective. That is, 3 is
not significantly greater than 1.
Test the hypothesis that dropping price by $0.20 will be more effective for increasing sales revenue than
increasing advertising expenditure by $500, that is: 0.202 > 0.53
Note that the regression model provides that 0.22 = 1.582 > 0.53 = 0.931. The test is to prove that this
is significant. Therefore, the null and alternative hypotheses are:
0 : 0.22 0.53 0
1 : 0.22 0.53 > 0
0.22 0.53
=
se(0.22 0.52 )
The problem here is to find the standard error of the linear combination of the two coefficients in the
denominator. For that, first determine var(0.202 0.52 ):
Then,
(0.2)(7.9079) (0.5)(1.8626)
= = 1.622
0.4010
= 0.05,(72) = 1.666
Since = 1.622 < = 0.05,(72) = 1.666, do not reject the null hypothesis. The p-value for the test is P( >
1.622) = 0.055, which exceeds = 0.05. The test does not prove that 0.202 > 5003 .
= 1 + 2 2 + 3 3 + 4 32 +
E() = 1 + 2 2 + 3 3 + 4 32
Thus,
E()
= 3 + 24 3
3
We expect revenues to increase with each additional unit increase in advertising expenditure. Therefore,
3 > 0. We also expect the rate of increase in revenues to decrease with each additional unit increase in
advertising. Therefore, 4 < 0.
Once we point out the characteristics of the extended model, we can treat 32 as a new variable 4 . Using the
same data, the Excel regression output is show below.
ANOVA
df SS MS F Significance F
Regression 3 1583.397408 527.79914 24.459316 5.59996E-11
Residual 71 1532.084459 21.578654
Total 74 3115.481867
From the regression table, the coefficient of 2 , 4 = 2.768. It has the anticipated sign, and it is also
significantly different from zero (value = 0.0044 < 0.05). There is diminishing returns to advertising.
To interpret the role of the coefficients 3 = 12.151 and 4 = 2.768, consider the following table where
() is computed for a fixed value of 2 () = $5 and for various values of 3 (). Starting
from (3 )0 = $1.8 and increasing advertising by the increment of 3 = 0.10 to , 3 = 1.9, the predicted sales
increases by = $0.19. As advertising expenditure is increased by the same increment, the increment in
predicted sales decreases to 0.14 and then to 0.08.
(3 )0
1.8 84.42
1.9 84.61 0.19
2.0 84.75 0.14
2.1 84.83 0.08
What is the optimum level of advertising? Optimality in economics always implies marginal benefit of an
action be equal to its marginal cost. If marginal benefit exceeds the marginal cost, the action should be taken.
If marginal benefit is less than the marginal cost, the action should be curtailed. The optimum is, therefore,
where the two are equal.
The marginal benefit of advertising is the contribution of each additional one thousand dollar ($1) of
advertising expenditure to total revenue. Form the model,
= 1 + 2 2 + 3 3 + 4 32 +
Ignoring the marginal cost of additional sales, marginal cost is each additional $1 (thousand) spent on
advertising. Thus, the optimality requirement is = :
3 + 24 3 = $1
12.151 + 2(2.768)3 = 1
1 3
(3 ) =
24
13
Note that the sample statistic is obtained as a nonlinear combination of the two coefficients 3 and 4 .
24
Therefore, we cannot use the same formula that we used to find the variance of the linear combination of the
two coefficients. The (approximate) variance of the nonlinear combination of two random variables is
obtained using the delta method. Let
1 3
=
24
Then,
2 2
var() = ( ) var(3 ) + ( ) var(4 ) + 2 ( )( ) covar(3 , 4 )
3 4 3 4
1 1 3
= and =
3 24 4 242
Hence,
2
1 2 1 3 1 1 3
var() = ( ) var(3 ) + ( ) var(4 ) + 2 ( ) ( ) covar(3 , 4 )
24 242 24 242
We can obtain var(3 ) and var(4 ) by simply squaring the standard errors from the regression output.
Unfortunately, the Excel regression output does not provide the covariance value. We can, however, still use
Excel to compute covar(3 , 4 ). If you recall, using matrix algebra we can determine the variance-covariance
matrix: var() 1 . Adding 32 to the model, we have now have three independent variables. The solution for
this problem is simple because the matrix can be expanded to incorporate any number of independent
variables. For a 3-variable model we have,
Using Excel we can compute these quantities as the elements of the X matrix:
var() 1 = 21.579 1
Thus,
1 2
1 12.151 2 1 1 12.151
var() = ( ) (12.646) + ( 2
) (0.885) + 2 ( ) ( ) (3.289)
2 2.768 2 2.768 2 2.768 2 2.7682
se() = 0.12872
The 95% confidence interval for the optimal level of advertising then is,
, = 2,(4) se()
An example to illustrate the use of interaction variables is the life-cycle model of consumption behavior. Here
the model involves the response of consumption () to (2 ) and (3 ) of the consumer.
First consider the simple model without the interaction variables.
= 1 + 2 2 + 3 3 +
Using the data in the Excel file CH6 DATA in the tab pizza, the estimated regression equation is
Thus, holding constant, for each additional year of the expenditure on decreases by
$7.58:
2 = = 7.576
2
And, holding constant, for each additional $1,000 increase in expenditure on rises by
$1.83.
However, we would expect that people of different ages would not spend similar amount on pizza for each
additional $1,000 increase in income. It is reasonable to expect that older persons spend smaller amount of
additional income on pizza than younger persons. Thus, we expect there to be an interaction between age
and income variables. This gives rise to the interaction variable in the model, represented by the product of
the variables and : (2 3 ).
= 1 + 2 2 + 3 3 + 4 2 3 +
= 2 + 4 3 = 3 + 4 2
2 3
= 2.977 0.1233 = 6.98 0.1232
2 2
= 2.977 0.123(30) = 6.67 = 6.98 0.123(25) = 3.90
2 2
When income is $30,000, for each additional year, When age is 25, for each additional $1,000 income,
expenditure on pizza is reduced by $6.67 expenditure on pizza is increased by $3.90.
= 2.977 0.123(80) = 12.84 = 6.98 0.123(50) = 0.82
2 2
When income is $80,000, for each additional year, When age is 50, for each additional $1,000 income,
expenditure on pizza is reduced by $12.84 expenditure on pizza is increased by $0.82.
Lets start with a model where the percent change in () is a stochastic function of two independent
variables (2 ) and years of (3 )
ln() = 1 + 2 2 + 3 3 +
ln() = 1 + 2 2 + 3 3 + 4 2 3 +
ln() = 1.3923 + 0.094942 + 0.0063295143 0.00003642 3
1
= 2 + 4 3
2
2
= 0.09494 0.000036(5) = 0.09476 (9.476%)
At a higher level of , say 3 = 10 years, the percentage increase in for an additional year
of decreases slightly to 9.46%.
2
= 0.09494 0.000036(10) = 0.09457 (9.457%)
Note that these percentage changes in are the result of a very small change in the variable
. The results of a discrete change in the variable , where 2 = 1, are shown in the
calculations in the following table. The results show that at a higher level of , any additional
year of education has a slightly smaller impact on .
02 12 02 12
1.39232 1 1 1 1
(2 ) 0.09494 16 17 16 17
(2 ) 0.00633 5 5 10 10
(2 3 ) -0.000036 80 85 160 170
ln () 2.94 3.03 2.97 3.06
18.917 20.797 19.469 21.400
1.880 1.931
( )% 9.94% 9.92%
1
= 3 + 4 2
3
Holding constant at 2 = 8,
3
= 0.00633 0.000036(16) = 0.00575 (0.575%)
The results of a discrete change in the variable , where 3 = 1, are shown the calculation in
the following table.
0 1 0 1
1.39232 1 1 1 1
(2 ) 0.09494 8 8 16 16
(2 ) 0.00633 10 11 10 11
(2 3 ) -0.000036 80 88 160 176
ln () 2.212 2.218 2.969 2.975
9.136 9.192 19.469 19.585
0.055 0.116
( )% 0.617% 0.598%
The greater the number of years of education, the less valuable is an extra year of experience.
( )2
2 = =
( )2
These quantities are easily calculated and they are also shown in the ANOVA part of the regression output for
the example.
ANOVA
df SS
Regression 2 1396.539
Residual 72 1718.943
Total 74 3115.482
1396.539
2 = = 0.4483
3115.482
This implies that nearly 45% of the variations in (monthly sales) is explained by the variations in price and
advertising expenditure.
In Chapter 3 it was shown that 2 is a measure of goodness of fit, that is, how well the estimated regression
2
fits the data: 2 = . The same argument applies here. A high value means there is a close association
2
= 1 + 2 2 + 3 3
( )2
2 = 1
( )2
Now, if we add another variable to the regression model, then the quantity = ( )2 becomes smaller
and 2 = 1 becomes larger.
( )
2 = 1
( 1)
1718.943 72
RA2 1 = 0.4329
3115.482 74
This figure is shown in the computer regression output right below the regular 2 .
1Note that 2 = 1 . Divide the numerator and denominator of the quotient on the right-hand-side by their
respective degrees of freedom. Thus,
( )
2 = 1
( 1)
To show why 2 does not increase with the increase in (the number of independent variables), we can rewrite it as:
1 1
2 = 1 = 1 (1 2 )
1. Introduction
Matrix algebra enables us to:
write an equation system in a compact way,
develop a method to test the existence of a solution by evaluation of a determinant, and
devise a method to find the solution.
For example, consider the following equation system with three variables 1 , 2 , and 3 :
61 + 32 + 13 = 22
11 + 42 23 = 12
41 12 + 53 = 10
6 3 1 1 22
[1 4 2] [2 ] = [12]
4 1 5 3 10
The lead matrix on the left hand side is the coefficient matrix, denoted by . The lag matrix is the variable
matrix and the matrix on the right hand side is the matrix of the constant terms, .
6 3 1 1 22
= [1 4 2]
= [ 2] = [12]
4 1 5 3 10
Any given matrix can be defined by its dimension: The number of rows () and number of columns (). The
dimension of is 3 3, that of and are both 3 1.
6 8 8 5 68 85 2 3
=[ ] =[ ] =[ ]=[ ]
12 19 3 11 12 3 19 11 9 8
4 5 16 20
4[ ]=[ ]
6 12 24 48
A B AB
(32) ( 21) (31)
The following example shows how the multiplication rule applies to two matrices.
11 12 11 11 + 12 21
= [21 22 ] = [11 ] = [ 21 11 + 22 21 ]
31 32 21
31 11 + 32 21
5 4 5(7) + 4(5) 55
7
= [6 1] =[ ] = [6(7) + 1(5)] = [47]
5
2 9 2(7) + 9(5) 59
Note that if is used as the lead matrix and as the lag matrix, the two are no longer conformable for
multiplication. : 2 1 and : 3 2. The number of columns of the lead matrix is not the same as the
number of rows of the lag matrix. Even if switching the lead and lag matrices preserved the conformability,
still the resulting product matrix would not be the same. That is: . Matrix multiplication is not
commutative.
Another example,
2 5
1 2 1
A =[ ] B =[ 4 3]
23 3 1 4 32
2 1
13 1 22
4 2
AB = [ ] BA = [5 5 16]
22 6 16 33
5 5 2
Also note that we have used the matrix multiplication rule to write an equation system in the matrix format:
6 3 1 1 22 61 + 32 + 13 = 22
[1 4 2] [2 ] = [12] 11 + 42 23 = 12
4 1 5 3 10 41 12 + 53 = 10
1 0 0
1 0
2 = [ ] 3 = [0 1 0]
1 0
0 0 1
Pre or post multiplying a matrix by an identity matrix leaves the matrix unchanged.
= =
Example
2 4 6
= [3 2 3]
1 4 9
2 4 6 1 0 0 2 4 6
= [3 2 3] [0 1 0] = [3 2 3] =
1 4 9 0 0 1 1 4 9
1 0 0 2 4 6 2 4 6
= [0 1 0] [3 2 3] = [3 2 3] =
0 0 1 1 4 9 1 4 9
11 12
11 21 31
= [21 22 ] = [ 22 32 ]
12
31 32
For example,
5 4
5 6 2
= [6 1] = [ ]
4 1 9
2 9
1 = 1 =
If does not have an inverse, then it is called a singular matrix. Otherwise it is a nonsingular matrix.
12
|| = |11 22 | = 11 22 12 21
21
11 12 13
23 21 23 21 22
|| = |21 22 23 | = 11 | 22
32 33 | 12 |31 33 | + 13 |31 32 |
31 32 33
|| = 11 22 33 11 23 32 12 21 33 + 12 23 31 + 13 21 32 13 22 31
In the latter case, each element in the top row is multiplied by a sub determinant. The first sub determinant
22 23
| 33 |
32
multiplied by 11 , is obtained by eliminating the first row and the first column. The sub determinant
associated with 11 is called the minor of that element and is denoted by 11 . The minor of 11 , 11 , is
obtained by eliminating the first row and second column, and so on. If the sum of the subscript of the element
is odd, then the sign is negative. The calculation of the determinant of A then can be presented as:
|| = 11 11 12 12 + 13 13
2 1 3
5 6 4 6 4 5
|4 5 6| = 2 | || | + 3| | = 2(5 9 6 8) (4 9 6 7) + 3(4 8 5 7) = 9
8 9 7 9 7 8
7 8 9
Here the determinant is obtained by expanding the first row. The same determinant can be found by
expanding any other row or column. In the previous example, find the determinant by expanding the third
column:
|| = 13 13 23 23 + 33 33
2 1 3
4 5 2 1 2 1
|4 5 6| = 3 | | 6| | + 9| | = 3(4 8 5 7) 6(2 8 1 7) + 9(2 5 1 4) = 9
7 8 7 8 4 5
7 8 9
Now lets introduce a related concept to the minor, called the cofactor. A cofactor, denoted by , is a minor
with a prescribed algebraic sign attached. If the sum of the two subscripts in the minor is odd then the
cofactor is negative:
= (1)(+)
In short, the value of determinant || of order can be found by the expansion of any row or any column as
follows:
1 2 1 3
|| = | | = 1 4 2 3 = 2 || = | | = 1 4 2 3 = 2
3 4 2 4
2) The interchange of any two rows (or any two columns) will alter the sign, leaving the absolute value
of the determinant unchanged.
1 2 3 4
| | = 1 4 2 3 = 2 | |= 3241 =2
3 4 1 2
3) Multiplying any one row (or any one column) by a scalar will change the value of the determinant by
the multiple of that scalar.
1 2 15 25
| | = 1 4 2 3 = 2 | | = 5 4 10 3 = 10
3 4 3 4
4) The addition (subtraction) of a multiple of any row (or column) to another row (or column) will leave
the value of the determinant unchanged.
1 2
In the determinant| |, multiply the first row by 3 and add to the second row:
3 4
1 2 1 2
| |=| | = 2
33 46 0 2
5) If one row (or column) is a multiple of another row (or column), the value of the determinant is zero;
the determinant will vanish. In other words, if two rows (or columns) are linearly dependent, the
determinant vanishes.
In the following example the second row is the first row multiplied by 4.
1 2
| |=824 =0
4 8
A very important conclusion relating the existence of the determinant to the existence of a unique
solution for an equation system:
61 + 32 + 13 = 22
11 + 42 23 = 12
41 12 + 53 = 10
6 3 1
= [1 4 2]
4 1 5
This equation system has a unique solution because there is a non-vanishing determinant || = 52.
161 + 32 + 13 = 22
121 + 62 + 23 = 11
141 12 + 53 = 10
6 3 1
|| = |12 6 2| = 6(6 5 + 2) 3(12 5 2 4) + (1 12 6 4) = 0
4 1 5
The determinant vanishes because rows 1 and 2 are linearly dependent: The second row is first row
multiplied by 2. Thus the equation system will not have a unique solution.
6) The expansion of the determinant by alien cofactors (the cofactor of a wrong row or column)
always yields a value of zero.
Expand the following determinant by using the first row elements but the cofactors of the second row
elements
6 3 1
= |1 4 2|
4 1 5
11 = 6 21 = [3 5 (1) 1] = 16 11 21 = 96
12 = 3 22 = (6 5 1 4) = 26 12 22 = 78
12 = 1 23 = [6 (1) 3 4] = 18 13 23 = 18
= 11 21 + 12 22 + 13 23 = 96 + 78 + 18 = 0
=1
This last property of the determinants finally leads us to the method to find the inverse of a determinant.
Finding the inverse of A involves the following steps:
11 12 13
= [21 22 23 ]
31 32 33
11 12 13
= [21 22 23 ]
31 32 33
11 21 31
adjoint = [12 22 32 ]
13 23 33
3) Multiply by adjoint
11 12 13 11 21 31 1 1 1 2 1 3
= [21
22 23 ] [12 22 32 ] = [2 1 2 2 2 3 ]
31 32 33 13 23 33 3 1 3 2 3 3
Now note that the elements in principal diagonal of the product matrix simply provide the determinant of .
All other elements outside the principal diagonal are the expansions of the determinant by an alien cofactor.
Thus, they are all zeros.
1 1 1 2 1 3 || 0 0 1 0 1
[2 1 2 2 2 3 ] = [ 0 || 0 ] = || [0 1 0] = ||
3 1 3 2 3 3 0 0 || 0 0 1
Thus,
= ||
=
||
1
= 1
||
1 =
||
6 3 1
Find the inverse of = [1 4 2]
4 1 5
4 2 1 2 1 4
| | | | | |
1 5 4 5 4 1 18 13 17
3 1 6 1 6 3
= | | | | | | = [16 26 18]
1 5 4 5 4 1
3 1 6 1 6 3 10 13 21
[ |4 2| |
1 2
| |
1 4
|]
Find adjoint
18 16 10
= [13 26 13]
17 18 21
1 1 18 16 10
1 = = [13 26 13]
|| 52
17 18 21
1.1.5. How to use the Inverse of Matrix A to find the solutions for an equation system
Note that in the previous example the matrix A was the coefficient matrix of the equation system
61 + 32 + 13 = 22 6 3 1 1 22
11 + 42 23 = 12 [1 4 2] [2 ] = [12]
41 12 + 53 = 10 4 1 5 3 10
Now pre multiply both sides of the matrix notation of the equation system by 1 :
1 = 1
which results in
= 1
Thus, 1 = 2, 2 = 3, and 3 = 1.