Sie sind auf Seite 1von 45

CFA Level II Quantitative Methods

Correlation and Regression


www.irfanullah.co
Graphs, charts, tables, examples, and figures are copyright 2012, CFA Institute. Reproduced
and republished with permission from CFA Institute. All rights reserved.
1

Contents
Introduction
Correlation Analysis
Linear Regression

www.irfanullah.co

2. Correlation Analysis
Scatter Plots
Correlation Analysis
Calculating and Interpreting the Correlation Coefficient
Limitations of Correlation Analysis
Uses of Correlation Analysis
Testing the Significance of the Correlation Coefficient

www.irfanullah.co

2.1 Scatter Plots


A scatter plot is a graph that shows the relationship between the observations for two
data series in two dimensions.

www.irfanullah.co

2.2 Correlation Analysis


The correlation coefficient is a measure of how closely related two data series are. In particular, the
correlation coefficient measures the direction and extent of linear association between two variables. A
.

www.irfanullah.co

2.3 Calculating and Interpreting the Correlation


Coefficient

www.irfanullah.co

2.4 Limitations of Correlation Analysis


Two variables can have a strong non-linear relation and still have a very low
correlation

Correlation can be unreliable when outliers are present

Spurious correlation refers to the following situations:


correlation between two variables that reflects chance relationships in a particular data set
correlation induced by a calculation that mixes each of two variables with a third
correlation between two variables arising not from a direct relation between them but from
their relation to a third variable
www.irfanullah.co

2.5 Uses of Correlation


Example 1: Evaluating Economic Forecasts
Example 2: Style Analysis Correlation
Example 3: Exchange Rate Correlations
Example 4: Correlations among Stock Return Series
Example 5: Correlations of Debt and Equity Returns
Example 6: Correlations among Net Income, Cash Flow from Operations and Free
Cash Flow to the Firm

www.irfanullah.co

Example 1: Evaluating Economic Forecasts (1)

www.irfanullah.co

Example 2: Site Analysis Correlations

www.irfanullah.co

10

Example 3: Exchange Rate Correlations

www.irfanullah.co

11

Example 4: Correlations among Stock Return Series


Table 4 shows the correlation matrix of
monthly returns to three U.S. stock indexes
during the period January 1971 to
December 1999 and in three sub-periods
(the 1970s, 1980s, and 1990s).10 The largecap style is represented by the return to the
S&P 500 Index, the small-cap style is
represented by the return to the
Dimensional Fund Advisors U.S. Small-Stock
Index, and the broad-market returns are
represented by the return to the Wilshire
5000 Index.

www.irfanullah.co

12

Example 5: Correlations of Debt and Equity Returns

www.irfanullah.co

13

Example 6: Correlations among Net Income, Cash Flow


from Operations and Free Cash Flow to the Firm

www.irfanullah.co

14

2.6 Testing the Significance of the Correlation


Coefficient
Significance tests allow us to assess whether apparent relationships
between random variables are the result of chance. We will use
hypothesis testing.
H0: = 0 versus Ha: 0
As long as the two variables are distributed normally, we can test to
determine whether the null hypothesis should be rejected using the
sample correlation, r. The formula for the t-test is

www.irfanullah.co

15

Example 7: Testing the Correlation between Money


Supply Growth and Inflation

www.irfanullah.co

16

Example 8: Testing the Krona-Yen Return Correlation

www.irfanullah.co

17

Example 9: The Correlation between Bond Returns and TBill Returns

www.irfanullah.co

18

Example 10: Testing the Correlation between Net Income


and Free Cash Flow to the Firm

www.irfanullah.co

19

3. Linear Regression
Linear Regression with One Independent Variable
Assumptions of the Linear Regression Model
The Standard Error of Estimate
The Coefficient of Determination
Hypothesis Testing
Analysis of Variation in a Regression with One Independent Variable
Prediction Intervals
Limitations of Regression Analysis
www.irfanullah.co

20

3.1 Linear Regression with One Independent


Variable
Linear regression allows us to use one variable to make predictions about another, test hypotheses
about the relation between two variables, and quantify the strength of the relationship between the
two variables. Linear regression assumes a linear relationship between the dependent and the
independent variables.

Yi = b0 + b1Xi + i, i = 1, ..., n

Minimize

www.irfanullah.co

21

3.2 Assumption of Linear Regression


1. Relation between Y and X is linear in the parameters b0 and b1
2. X is not random
3. Expected value of error term is zero
4. Variance of the error term is constant

Homoskedasticity

5. Error term is uncorrelated across observations


6. Error term is normally distributed

www.arifirfanullah.com

22

Example 11: Evaluating Economic Forecasts (2)

www.irfanullah.co

23

www.irfanullah.co

24

3.3 The Standard Error of Estimate


The standard error of estimate (SEE) measures how well a given linear regression model captures the
relationship between the dependent and independent variables.

www.irfanullah.co

25

Example 12: Computing the Standard Error of Estimate

www.irfanullah.co

26

www.irfanullah.co

27

3.4 Coefficient of Determination


The SEE gives some indication of how certain we can be about a particular prediction of Y using the
regression equation, it still does not tell us how well the independent variable explains variation in the
dependent variable. The coefficient of determination does exactly this: It measures the fraction of the
total variation in the dependent variable that is explained by the independent variable.
Total variation = Unexplained variation + Explained variation

With only one independent variable R-squared


is the square of the correlation. The
correlation is also called the Multiple R

www.irfanullah.co

28

Low R2 and High SEE

High R2 and Low SEE

www.arifirfanullah.com

29

Example 13: Inflation Rate and Growth in Money Supply

www.irfanullah.co

30

www.irfanullah.co

31

3.5 Hypothesis Testing


Suppose we regress a stocks returns on a stock market indexs returns and find that the slope
coefficient (

) is 1.5 with a standard error (

) of 0.200. Assume we used 62 monthly observations in

our regression analysis. The hypothesized value of the parameter (

) is 1.0, the market average slope

coefficient.
Define the null and alternative hypothesis
Compute the test statistic
At a 0.05 level of significance should we reject H0

www.irfanullah.co

32

p-value
The p-value is the smallest level of significance at which the null hypothesis can be rejected. It allows
the reader to interpret the results rather than be told that a certain hypothesis has been rejected or
accepted. In most regression software packages, the p-values printed for regression coefficients apply
to a test of null hypothesis that the true parameter is equal to 0 against the alternative that the
parameter is not equal to 0, given the estimated coefficient and the standard error for that coefficient.

www.irfanullah.co

33

Example 14: Estimating Beta for General Motors Stock

www.irfanullah.co

34

www.irfanullah.co

35

Example 15: Explaining Company Value Based on Returns


to Invested Capital

www.irfanullah.co

36

Example 16: Testing whether Inflation Forecasts are


Unbiased

www.irfanullah.co

37

3.6 Analysis of Variance in a Regression with one


Independent Variable

Analysis of variance (ANOVA) is a statistical procedure for dividing the total variability of a variable
into components that can be attributed to different sources.

We use ANOVA to determine the usefulness of the independent variable or variables in explaining
variation in the dependent variable.

The F-statistic tests whether all the slope coefficients in a linear regression are equal to 0. In a
regression with one independent variable, this is a test of the null hypothesis H 0: b1 = 0 against the
alternative hypothesis Ha: b1 0.

www.irfanullah.co

38

F-Statistic

The F-statistic measures how well the regression equation explains the variation in the dependent
variable.
Ratio of the average regression sum of squares to the average sum of the squared errors
Average regression sum of squares (RSS) is the amount of variation in Y explained by the regression
equation; it is computed by dividing the regression sum of squares by the number of slope parameters
estimated (in this case, one)
Average sum of squared errors (SSE) is the sum of the squared residuals; it is computed by dividing the
sum of squared errors by the number of observations, n, minus the total number of parameters estimated
(in this case, two)

www.irfanullah.co

39

Example 17: Performance Evaluation: The Dreyfus


Appreciation Fund

www.irfanullah.co

40

3.7 Prediction Intervals


We use regression results to make predictions about a dependent variable. It is important to
understand how to compute confidence intervals around regression forecasts.

Two sources of uncertainty


Error term
Uncertainty in the estimated parameters

Estimated variance of prediction error is given by:

Once the variance of the prediction error is


known, it is easy to determine the confidence
interval around the prediction.
www.irfanullah.co

41

Example 18: Predicting the Ratio of Enterprise Value to


Invested Capital

www.irfanullah.co

42

3.8 Limitations of Regression


Relationships change over time
Difficulty to apply
Specification issues
Assumptions violations

Usefulness limited for investment application since all


participants can observe relationships

www.arifirfanullah.com

43

Summary
Correlation Analysis
Testing the significance of

Linear Regression

Assumptions
SEE
R-squared
Hypothesis testing
F-stat
Prediction intervals

Conclusion
Read summary
Review learning objectives
Examples
Practice problems
Practice from other sources

www.irfanullah.co

45

Das könnte Ihnen auch gefallen