Beruflich Dokumente
Kultur Dokumente
Contents
Introduction
Correlation Analysis
Linear Regression
www.irfanullah.co
2. Correlation Analysis
Scatter Plots
Correlation Analysis
Calculating and Interpreting the Correlation Coefficient
Limitations of Correlation Analysis
Uses of Correlation Analysis
Testing the Significance of the Correlation Coefficient
www.irfanullah.co
www.irfanullah.co
www.irfanullah.co
www.irfanullah.co
www.irfanullah.co
www.irfanullah.co
www.irfanullah.co
10
www.irfanullah.co
11
www.irfanullah.co
12
www.irfanullah.co
13
www.irfanullah.co
14
www.irfanullah.co
15
www.irfanullah.co
16
www.irfanullah.co
17
www.irfanullah.co
18
www.irfanullah.co
19
3. Linear Regression
Linear Regression with One Independent Variable
Assumptions of the Linear Regression Model
The Standard Error of Estimate
The Coefficient of Determination
Hypothesis Testing
Analysis of Variation in a Regression with One Independent Variable
Prediction Intervals
Limitations of Regression Analysis
www.irfanullah.co
20
Yi = b0 + b1Xi + i, i = 1, ..., n
Minimize
www.irfanullah.co
21
Homoskedasticity
www.arifirfanullah.com
22
www.irfanullah.co
23
www.irfanullah.co
24
www.irfanullah.co
25
www.irfanullah.co
26
www.irfanullah.co
27
www.irfanullah.co
28
www.arifirfanullah.com
29
www.irfanullah.co
30
www.irfanullah.co
31
coefficient.
Define the null and alternative hypothesis
Compute the test statistic
At a 0.05 level of significance should we reject H0
www.irfanullah.co
32
p-value
The p-value is the smallest level of significance at which the null hypothesis can be rejected. It allows
the reader to interpret the results rather than be told that a certain hypothesis has been rejected or
accepted. In most regression software packages, the p-values printed for regression coefficients apply
to a test of null hypothesis that the true parameter is equal to 0 against the alternative that the
parameter is not equal to 0, given the estimated coefficient and the standard error for that coefficient.
www.irfanullah.co
33
www.irfanullah.co
34
www.irfanullah.co
35
www.irfanullah.co
36
www.irfanullah.co
37
Analysis of variance (ANOVA) is a statistical procedure for dividing the total variability of a variable
into components that can be attributed to different sources.
We use ANOVA to determine the usefulness of the independent variable or variables in explaining
variation in the dependent variable.
The F-statistic tests whether all the slope coefficients in a linear regression are equal to 0. In a
regression with one independent variable, this is a test of the null hypothesis H 0: b1 = 0 against the
alternative hypothesis Ha: b1 0.
www.irfanullah.co
38
F-Statistic
The F-statistic measures how well the regression equation explains the variation in the dependent
variable.
Ratio of the average regression sum of squares to the average sum of the squared errors
Average regression sum of squares (RSS) is the amount of variation in Y explained by the regression
equation; it is computed by dividing the regression sum of squares by the number of slope parameters
estimated (in this case, one)
Average sum of squared errors (SSE) is the sum of the squared residuals; it is computed by dividing the
sum of squared errors by the number of observations, n, minus the total number of parameters estimated
(in this case, two)
www.irfanullah.co
39
www.irfanullah.co
40
41
www.irfanullah.co
42
www.arifirfanullah.com
43
Summary
Correlation Analysis
Testing the significance of
Linear Regression
Assumptions
SEE
R-squared
Hypothesis testing
F-stat
Prediction intervals
Conclusion
Read summary
Review learning objectives
Examples
Practice problems
Practice from other sources
www.irfanullah.co
45