(or SPSS where needed)
[by Bernhard Riecke, iSpaceLab.com]
Disclaimer: this is work in progress - in case you find any errors or have suggestions for improvement, please email me at b_r@sfu.ca
How to calculate effect sizes in JMP 1
Ok… but what does effect size even mean? 2
Why report effect sizes (and not just p=values)? 2
Which effect size should I use? 2
Independent measures (between-subject) t-test 3
Repeated measures (within-subject) t-test 5
Independent measures (between-subject) >= 1-Way ANOVA 6
Calculate effect sizes with Add-in 6
Repeated measures (within-subject) 1-Way ANOVA 7
>= 2-Way Independent measures (between-subject) 9
>= 2-Way Repeated measures (within-subject) or mixed ANOVA 9
Calculate effect sizes in JMP 10
The p-value of the inferential stats only tells you if the result is “statistically significant” in the sense that, e.g., for a p-value of p = .04 you’d expect that if you ran the same experiment 100 times all but 4 of these (100-4 = 96) would show the same effect. It does not tell you if the effect is meaningful, or of a size that matters. For this, the effect size is useful, and should always be included in your writeup.
That is, the effect size is a quantitative measure of the magnitude of the effect. The larger the effect size the larger the effect (e..g, of the independent on the dependent variable).
As a rule of thumb, and according to Cohen (1988, 1992), small/medium/large effects sizes refer to
*r is not measured on a linear scale, so r = 0.4 is NOT twice as big as r = 0.2
Note that different effect size measures can mean different things. E.g., Cohen's d is an effect size used for comparing 2 means (e.g., after a t-test), and indicates the standardized difference between two means - basically the size of the difference in standard deviations. E.g., a d = 2.2 indicates that the two groups differ by 2.2 standard deviations (or z-scores),
Table on magnitude of effect sizes for other effect size measures like Cohen's d or η2
The effect size describes how big the difference is between groups. A value of 0 means there's no effect, and a value of 1 would be a perfect effect (not a linear scale). Effect sizes are important because they go beyond the simplistic 'Does it work or not?' to the far more sophisticated, 'How well does it work in a range of contexts?'. So, for example, you could have a significance (p-value) of 0.01, which is highly significant. This means this effect probably didn't happen by chance. However, if the effect size is 0.15, then that means this effect is small and not that important.
Well, a significant p-value (p < .05 typically) tells us that an intervention works (or in general that the IV had an effect on the DV), whereas an effect size tells us how much it works (ie., how large the effect of the IV on the DV is). Note that the effect size does NOT increase with N (the sample size), unlike significance tests which become more significant for larger N.
This is discussed in various resources in more detail - e.g., the Analysis Factor explains that R squared, Eta Squared, Partial Eta Squared, and Omega Squared all have the intuitive interpretation of the proportion of the variance accounted for (compared to, e.g., Cohen’s d, which indicates the size of the difference in standard deviations or z-scores, and can be larger than 1). Some effect sizes such as eta or eta squared can be biased, and often omega squared it suggested as an alternative - see https://www.theanalysisfactor.com/effect-size/ or https://daniellakens.blogspot.com/2015/06/why-you-should-use-omega-squared.html for details.
So if your stats software provides omega squared, this might be the best option as a rule of thumb.
based on r := sqrt (t^2) / (sqrt (t^2 + df))
Here: r^2 = (-1.19^2) / (1.19^2 + 46) = .0298, hence
t(46) = -1.19, p = .24, r = .17
t(28.48) = -1.19, p = .24, r = .22
First: use Analyze > Fit Model to run ANOVA first (example)
Second: Add-ins > Calculate Effect Sizes > From Least Squares Report (Fit Model)
Third: you might see this window; click Yes
This should give you all the effect sizes needed in a separate table in JMP, for both t-test and ANOVAs
Formulas used:
Eta Squared
Partial Eta Squared
Omega Squared
Pairwise Effect Size: Cohen's d
based on r := sqrt (t^2) / (sqrt (t^2 + df))
First: convert your data to wide format
--> --> -->
Second: run Analyze > Specialized Modeling > Matched Pairs:
The provides the t and df needed to calculate r^2:
t(7) = -2.93, p = .02, r = .74
We don’t even have the SS in the JMP output to calculate the effect size…
→ go to SPSS or another software :-(
Laerd Statistics: one-way repeated measures ANOVA with SPSS
Analyze > General Linear Model > Repeated Measures
Define your levels and factors
Assign them (make sure they match)
Under “options” check the power, effect size etc. options if you need them:
Check sphericity:
Effect size partial eta squared and power are included in table below
Use G-F or H-F if sphericity is violated
Sample write up
Vection intensity ratings were higher for the leaning interface (M = 64.6, SD = 21.9) compared to the joystick (M = 49.7, SD = 26.4, F(1,15) = 10.406, p = .006, η2 = .410), see also Figure 2. The effect size of η2 = .410 indicates that 41% of the variability can be attributed to the factor interface, which is considered a large effect size (Cohen, 1988).
Laerd Statistics: two-way repeated measures ANOVA with SPSS
Lynda Tutorials: SPSS two categorical variables ANOVA
Laerd Statistics: mixed ANOVA with SPSS
Or just google for instructions, there’s plenty of tutorials out there.
Part 1: https://www.youtube.com/watch?v=qobtGqVBFig
Part 2: https://www.youtube.com/watch?v=zBIqSRC3-Z8
Part 3: https://www.youtube.com/watch?v=VzjAM0jEDtg
Under “options” check the power, effect size etc. options if you need them:
First of all, you don’t have to report effect sizes for post-hoc tests or planned contrasts.
It can still be interesting and useful though, for example if you’re interested in how much variability is explained when you compare two factor levels of one IV:
First: JMP > Analyze > Fit model
Second: Then Effect Details > Test Slices
Alternatively, if you’d like to use the t-test results:
Second: Effect Details > LS Means Student’s t
*just be sure to correct for Alpha inflation using, e.g, Bonferroni correction
Third: to see the actual t values to compute r or r^2, use the Detailed Comparison option:
E.g., https://www.simplypsychology.org/statistics.html has useful and easy to understand summaries
From https://imaging.mrc-cbu.cam.ac.uk/statswiki/FAQ/effectSize
Effect Size | Use | Small | Medium | Large | |||
Correlation inc Phi | 0.1 | 0.3 | 0.5 | ||||
r x c frequency tables | 0.1 (Min(r-1,c-1)=1), 0.07 (Min(r-1,c-1)=2), 0.06 (Min(r-1,c-1)=3) | 0.3 (Min(r-1,c-1)=1), 0.21 (Min(r-1,c-1)=2), 0.17 (Min(r-1,c-1)=3) | 0.5 (Min(r-1,c-1)=1), 0.35(Min(r-1,c-1)=2), 0.29 (Min(r-1,c-1)=3) | ||||
Comparing two proportions | 0.2 | 0.5 | 0.8 | ||||
η2 | Anova | 0.01 | 0.06 | 0.14 | |||
Anova; See Field (2013) | 0.01 | 0.06 | 0.14 | ||||
one-way MANOVA | 0.01 | 0.06 | 0.14 | ||||
Cohen's f | one-way an(c)ova (regression) | 0.10 | 0.25 | 0.40 | |||
η2 | Multiple regression | 0.02 | 0.13 | 0.26 | |||
κ2 | Mediation analysis | 0.01 | 0.09 | 0.25 | |||
Cohen's f | Multiple Regression | 0.14 | 0.39 | 0.59 | |||
Cohen's d | t-tests | 0.2 | 0.5 | 0.8 | |||
Cohen's ω | chi-square | 0.1 | 0.3 | 0.5 | |||
Odds Ratios | 2 by 2 tables | 1.5 | 3.5 | 9.0 | |||
Odds Ratios | 0.55 | 0.65 | 0.75 | ||||
Friedman test | 0.1 | 0.3 | 0.5 |
Also:Haddock et al (1998) state that 3π multiplied by the log of the odds ratio is a standardised difference equivalent t