Beruflich Dokumente
Kultur Dokumente
One-Way ANOVA
The one-way analysis of variance is used to test the claim that two or more population means are equal
< The one-way ANOVA uses an F statistic and is therefore often called the ANOVA F
This is an extension of the two independent samples t-test In fact, with only two groups the t-test and one-way ANOVA are equivalent and will always give the same p-value
t-tests vs ANOVA
The difference between the t-test and the ANOVA F is that the t-test works in the same units as the original scores and the ANOVA F works in squared units Thus, with only two groups: F = t2
Comparing more than two groups with ttests would require a series of t tests to evaluate all of the mean differences.
< Remember, a t test can compare only 2 means at a time
Although each t test can be done with a specific -level (risk of Type I error), the -levels accumulate over a series of tests so that the final experimentwise -level can be quite large
< We will assume k samples (groups) < The k samples are the levels of the factor
The one-way is because each value is classified in exactly one way (i.e. there is only one factor variable
Assumptions
< The data are randomly sampled < The variances of the populations are equal < The distribution of scores in each population are normal in shape
We will come back to these assumptions after going through the steps for the ANOVA
The alternative hypothesis is that at least one of the means is different from another
In other words, Ho: 1 2 ... k would not be an acceptable way to write the alternate hypothesis (this slightly contradicts Gravetter & Wallnau, but tecnically there is no way to test this specific alternative hypothesis with a one-way ANOVA)
One-Way ANOVA
A random sample of the students in each section was taken The test scores were recorded:
< Left: 82, 83, 97, 93, 55, 67, 53 < Middle: 83, 78, 68, 61, 77, 54, 69, 51, 63 < Right: 38, 59, 55, 66, 45, 52, 52, 61
One-Way ANOVA
The summary statistics for the grades of each section are shown in the table below Section
Sample size
One-Way ANOVA
Variation
< Variation is the sum of the squares of the deviations between a value and the mean of the value < Sum of Squares (SS) is the term used to represent this variation
< Note that the sum of the between group and within group SS equals the total SS
SST = SSB + SSW
One-Way ANOVA
One-Way ANOVA
One-Way ANOVA
One-Way ANOVA
The between group variation is the variation among the sample means Each individual variation is weighted by the sample size
One-Way ANOVA
df
MS
Degrees of Freedom, df
< A degree of freedom occurs for each value that can vary before the rest of the values are predetermined < For example, if you had six numbers that had an average of 40, you would know that the total had to be 240. Five of the six numbers could be anything, but once the first five are known, the last one is fixed so the sum is 240. The df would be 6-1=5 < The df is often one less than the number of values
The within group df is the sum of the individual dfs of each group
< The sample sizes are 7, 9, and 8 < dfW = 6 + 8 + 7 = 21 < Alternatively, dfW = N - k = 24 - 3 = 21
One-Way ANOVA
df
2 21 23
MS
One-Way ANOVA - MS
Variances
The variances are also called the Mean of the Squares and abbreviated by MS, often with an accompanying variable MSB or MSW They are an average squared deviation from the mean and are found by dividing the variation by the degrees of freedom MS = SS / df
Variation Variance = df
One-Way ANOVA - MS
MSB= 1901.52 / 2= 950.76 MS = 3386.32 / 21= 161.25 W MS = 5287.83 / 23= 229.91 T
< Notice that the MS(Total) is NOT the sum of MS(Between) and MS(Within). < This works for the SS(Total), but not the mean square MS(Total) < The MS(Total) is often not presented in an ANOVA summary table
df
MS 2 950.76
21 161.25 23 229.91
One-Way ANOVA
F test statistic
< An F test statistic is the ratio of two sample variances < Specifically, F is the ratio of the MSB to MSW
In other words, how variability in there in the group means relative to the variability within each group
One-Way ANOVA
df
MS 2 950.76
5.90
21 161.25 23 229.91
One-Way ANOVA
The F test statistic has an F distribution with dfB and dfW degrees of freedom
< Df SS MS F p < Section 2 1901.5 950.8 5.8961 0.009284 < Within 21 3386.3 161.3
One-Way ANOVA
df
MS 2 950.76
5.90 .009
21 161.25 23 229.91
One-Way ANOVA
There is enough evidence to support the claim that there is a difference in the mean scores of the left, middle, and right sections of the class. However, there are still a few important points to consider:
< What about effect sizes?? < How do we know which sections differ in terms of mean test scores?? < What about assumption violations?