Vous êtes sur la page 1sur 15

Lab

Stuff about Chi-Square?

Questions Intro

to Analysis of Variance (ANOVA)

Final

lab will be distributed on Thursday


Very similar to lab 3, but with different

data You will be expected to find appropriate variables for three major tests (correlation, t-test, chi-square test of independence) You will be expected to interpret the findings from each test (one short paragraph per test).

Males

Student 46 (40.97) 37 (42.03)

NotStudent 71 (76.02) 83 (77.97)

Total 117

Females

120

Total

83

154

237

Observed Expected
4

In

its simplest form, it is used to compare means for three or more categories.
Example: Income (metric) and Marital Status (many categories)

Relies

on the F-distribution
6

Just like the t-distribution and chi-square

distribution, there are several sampling

If we have a categorical variable with 3+ categories and a metric/scale variable, we could just run 3 t-tests. One problem is that the 3 tests would not be independent of each other (i.e., all of the information is known). As number of comparisons grow, likelihood of some differences are expected but do not necessarily indicate an overall difference. A better approach: compare the variability between groups (treatment variance + error) to the variability within the groups (error)

F=

MSbg MS wg

MS = mean square bg = between groups wg = within groups

The numerator and denominator have their own degrees of freedom


8

Generally,

an f-ratio is a measure of how different the means are relative to the variability within each sample values greater likelihood that the difference between means are not just due to chance alone
9

Larger

If there is no difference between the means, then the between-group sum of squares should = the within-group sum of squares.

F=

MSbg MS wg
10

A It

right-skewed distribution is a ratio of two chi-square distributions

11

F-test

for ANOVA is a one-tailed test.

12

http://tinyurl.com/271ANOVA

13

How do we know where the differences exist once we know that we have an overall difference between groups?
t-tests become important after an ANOVA so that

we can find out which pairs are significantly different (post-hoc tests).

Certain corrections can be applied to such post-

hoc t-tests so that we account for multiple comparisons (e.g., Bonferroni correction, which divides p-value by the number of comparisons being made)

There are many means comparisons test available

(Tukey, Sidak, Bonferroni, etc). All are basically

14

Conceptual Class

Intro to ANOVA

Example:

anova.do GSS96_small.dta

15

Vous aimerez peut-être aussi