## 16.2 Analysis of variance

Follow these steps when conducting a hypothesis test:

• State the null and alternative hypotheses
• Set an $$\alpha$$ level
• Identify the appropriate test & provide rationale
• Prepare an appropriately formatted table of descriptive statistics for the response variable grouped by the categories of the explanatory variable.
• Visualize the data with a good graph and interpret the graph
• Assumptions
• state the assumptions of the test
• use appropriate figures and / or tests to check whether the assumptions of the statistical test are met
• transform data to meet assumptions if required
• if assumptions can’t be met (e.g. after transformation), use non-parametric test and repeat steps above
• Conduct the test, and report the test statistic and associated P-value
• Calculate and include a confidence interval (e.g. for t-tests) or $$R^2$$ value (e.g. for ANOVA) when appropriate
• Optionally, for ANOVA, conduct a “post-hoc” comparison of means using “Tukey-Kramer post-hoc test”
• Draw the appropriate conclusion and communicate it clearly
• Your concluding statement should refer to an ANOVA table (if required), a good figure, and the $$R^2$$ value

Additional, OPTIONAL steps for ANOVA tests

These steps may be warranted depending on the context.

• An appropriately formatted ANOVA results table (definitely include this for any research projects you’re writing up)
• A good figure with results of post-hoc tests included (if these tests were conducted)

### 16.2.1 Hypothesis statements

The null hypothesis of ANOVA is that the population means $$\mu_i$$ are the same for all treatments. Thus, for our current example in which melatonin production was measured among treatment groups:

H0: Mean melatonin production is equal among all treatment groups ($$\mu_1 = \mu_2 = \mu_3$$).
HA: At least one treatment group’s mean is different from the others

OR

HA: Mean melatonin production is not equal among all treatment groups.

We’ll set $$\alpha = 0.05$$.

We’ll use a fixed-effects ANOVA, which uses the $$F$$ test statistic.

### 16.2.2 A table of descriptive statistics

In a previous tutorial we learned how to calculate descriptive statistics for a numeric variable grouped by a categorical variable. In the comparing 2 means tutorial we also learned how to use the t.test function to calculate confidence intervals for a numeric variable.

Let’s use all these skills to generate a table of summary statistics for the “shift” variable, grouped by “treatment”:

circadian.stats <- circadian %>%
group_by(treatment) %>%
summarise(
Count = n() - naniar::n_miss(shift),
Count_NA = naniar::n_miss(shift),
Mean = mean(shift, na.rm = TRUE),
SD = sd(shift, na.rm = TRUE),
SEM = SD/sqrt(Count),
Low_95_CL = t.test(shift, conf.level = 0.95)$conf.int, Up_95_CL = t.test(shift, conf.level = 0.95)$conf.int
)

Let’s have a look at the result, and we’ll use the kable function to make the table nicer:

kable(circadian.stats, digits = 4)
treatment Count Count_NA Mean SD SEM Low_95_CL Up_95_CL
Control 8 0 -0.3088 0.6176 0.2183 -0.8250 0.2075
Knee 7 0 -0.3357 0.7908 0.2989 -1.0671 0.3957
Eyes 7 0 -1.5514 0.7063 0.2670 -2.2047 -0.8982

### 16.2.3 Visualize the data

We are interested in visualizing a numeric response variable in relation to a categorical explanatory variable.

We learned in an earlier tutorial that we can use a stripchart, violin plot, or boxplot to visualize a numerical response variable in relation to a categorical variable.

It is commonplace to use a stripchart for small sample sizes in each group.

We’ll use code we learned in an earlier tutorial.

However, because we’re now not only describing and visualizing data, but also testing a hypothesis about differences in group means, it is best to add group means and error bars to our stripchart.

The stat_summary function from the ggplot2 package is what provides the error bars and group means:

?stat_summary

Here’s the code:

circadian %>%
ggplot(aes(x = treatment, y = shift)) +
geom_jitter(colour = "darkgrey", size = 3, shape = 1, width = 0.1) +
stat_summary(fun.data = mean_se, geom = "errorbar",
colour = "black", width = 0.1,
position = position_nudge(x = 0.15)) +
stat_summary(fun = mean, geom = "point",
colour = "black", size = 3,
position = position_nudge(x = 0.15)) +
xlab("Light treatment") +
ylab("Shift in circadian rhythm (h)") +
theme_classic() Figure 16.1: Stripchart of phase shifts in the circadian rhythm of melatonin production in 22 participants of an experiment. Solid circles denote group means, and bars +/- one standard error

We’ve added two sets of stat_summary functions, one for adding the group mean, and one for adding the error bars.

• The argument “fun = mean_se” adds the standard error of the group means
• the “geom = ‘errorbar’” tells ggplot to use an error bar format
• the “position = position_nudge(x = 0.15)” tells ggplot to nudge the error bar off to the side a bit, so it doesn’t obscure the data points
• the next stat_summary function is the same, but this time for the group means.

Clearly there is no difference between the control and knee treatment groups, but we’ll have to await the results of the ANOVA to see whether the mean of the “eyes” group is different.

### 16.2.4 Assumptions of ANOVA

The assumptions of ANOVA are:

• The measurements in every group represent a random sample from the corresponding population (NOTE: for an experimental study, the assumption is that subjects are randomly assigned to treatments)
• The response variable has a normal distribution in each population
• The variance of the response variable is the same in all populations (called the “homogeneity of variance” assumption)

Test for normality

The sample sizes per group are rather small, so graphical aids such as histograms or normal quantile plots will not be particularly helpful.

Instead, let’s rely on the fact that (i) the strip chart in Figure 2 does not show any obvious outliers in any of the groups, and (ii) the central limit theorem makes ANOVAs quite robust to deviations in normality, especially with larger sample sizes, but even with relatively small sample sizes such as these. So, we’ll proceed under the assumption that the measurements are normally distributed within the three populations.

Test for equal variances

Now we need to test the assumption of equal variance among the groups, using Levene’s Test as we learned in the checking assumptions tutorial.

variance.check <- leveneTest(shift ~ treatment, data = circadian)
variance.check
## Levene's Test for Homogeneity of Variance (center = median)
##       Df F value Pr(>F)
## group  2  0.1586 0.8545
##       19

We see that the P-value for the test is much greater than 0.05, so the null hypothesis of equal variance is not rejected. There is no evidence against the assumption that the variances are equal.

A reasonable statement:

“A Levene’s test showed no evidence against the assumption of equal variance (F = 0.16; P-value = 0.854).”

IF assumptions are not met

A later section discusses what to do if assumptions are NOT met.

### 16.2.5 Conduct the ANOVA test

There are two steps to conducting an ANOVA in R:

• Use the lm function (from the base package) (it stands for “linear model”) to create an object that holds the key output from the test

• Use the anova function (from the base package) on the preceding lm output object to generate the appropriate output for interpreting the results.

The lm function uses the syntax: Y ~ X.

Let’s do this step first:

circadian.lm <- lm(shift ~ treatment, data = circadian) 

Next the anova step:

circadian.lm.anova <- anova(circadian.lm)

Let’s have a look at what this produced:

circadian.lm.anova
## Analysis of Variance Table
##
## Response: shift
##           Df Sum Sq Mean Sq F value   Pr(>F)
## treatment  2 7.2245  3.6122  7.2894 0.004472 **
## Residuals 19 9.4153  0.4955
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1

This is not in the format of a traditional ANOVA table (as described in Chapter 15 of the text): it is missing the “Total” row, and it puts the degrees of freedom (df) column before the Sums of Squares. Also, the column headers could be more informative.

Nor is the object in “tidy” format.

Optionally, we can use the tidy function from the broom package to tidy the output as follows:

circadian.lm.anova.tidy <- circadian.lm.anova %>%
broom::tidy()

Have a look at the tidier (but still not ideal) output:

circadian.lm.anova.tidy
## # A tibble: 2 × 6
##   term         df sumsq meansq statistic   p.value
##   <chr>     <int> <dbl>  <dbl>     <dbl>     <dbl>
## 1 treatment     2 7.224 3.612      7.289  0.004472
## 2 Residuals    19 9.415 0.4955    NA     NA

#### 16.2.5.1 OPTIONAL: Generate nice ANOVA table

Let’s look again at the raw output from the anova function:

circadian.lm.anova
## Analysis of Variance Table
##
## Response: shift
##           Df Sum Sq Mean Sq F value   Pr(>F)
## treatment  2 7.2245  3.6122  7.2894 0.004472 **
## Residuals 19 9.4153  0.4955
## ---
## Signif. codes:  0 '***' 0.001 '**' 0.01 '*' 0.05 '.' 0.1 ' ' 1

In the code chunk below we create a new function called “create.anova.table”, which we can use to convert an “anova” object (like the one above) to an appropriately formatted ANOVA table.

When you run this code chunk, it creates the function in your workspace. Once it is created, you can use it.

You will need to run this chunk to create the function each time you start a new R session. AND this is a quick-and-dirty function that does NOT follow all the best practices in coding!

create.anova.table <- function(intable){
# first check that input table is of class "anova"
if(class(intable) != "anova") stop("Input object not of class 'anova'")
require(tidyverse) # ensure tidyverse is loaded
require(broom) # ensures broom is loaded
tidy.intable <- tidy(intable)
# "intable" is a the object from the anova output
temp.anova <- tidy.intable %>%
select("term", "sumsq", "df", "meansq", "statistic", "p.value")  # reorder columns
temp.anova <- temp.anova %>%
rename(Source = term, SS = sumsq, MS = meansq, F = statistic, P_val = p.value) # rename columns
totals <- data.frame(
Source = "Total",
df = sum(temp.anova$df), SS = sum(temp.anova$SS),
MS = NA,
F = NA,
P_val = NA
)  # calculate totals
nice.anova.table <- as_tibble(rbind(temp.anova, totals)) # generate table
return(nice.anova.table)
}

Let’s try out the function, using our original anova table:

nice.anova.table <- create.anova.table(circadian.lm.anova)

And present it:

kable(nice.anova.table, digits = 4,
caption = "ANOVA table for the circadian rythm experiment.",
lable = "niceanovapdf")
Table 16.1: ANOVA table for the circadian rythm experiment.
Source SS df MS F P_val
treatment 7.2245 2 3.6122 7.2894 0.0045
Residuals 9.4153 19 0.4955 NA NA
Total 16.6398 21 NA NA NA

Nice!

### 16.2.6 Calculate $$R^2$$ for the ANOVA

One measure that is typically reported with any “linear model” like ANOVA is the “variance explained” or coefficient of determination, denoted $$R^2$$.

This value indicates the fraction of the variation in Y that is explained by groups.

The remainder of the variation ($$1 - R^2$$) is “error”, or variation that is left unexplained by the groups.

To calculate this we use two steps, and we go back to our original “lm” object we created earlier:

circadian.lm.summary <- summary(circadian.lm)
circadian.lm.summary$r.squared ##  0.4341684 We’ll refer to this value in our concluding statement. Before we get there, we need to figure out which group mean(s) differ??. ### 16.2.7 Tukey-Kramer post-hoc test As described in Chapter 15 in the text, planned comparisons (aka planned contrasts) are ideal and most powerful, but unfortunately we often need to conduct unplanned comparisons to assess which groups differ in our ANOVA test. This is what we’ll learn here. We can guess from our stripchart (Figure 16.1) that it’s the “Eyes” treatment group that differs from the others, but we need a formal test. We could simply conduct three 2-sample t-tests on each of the three pair-wise comparisons, but then we would inflate our Type-I error rate, due to multiple-testing. The Tukey-Kramer “post-hoc” (unplanned) test adjusts our P-values correctly to account for multiple tests. For this test we use the TukeyHSD function in the base stats package (HSD stands for “Honestly Significant Difference”). ?TukeyHSD The one catch in using this function is that we need to re-do our ANOVA analysis using a different function… the aov function in lieu of the lm function. This is necessary because the TukeyHSD function is expecting a particular object format, which only the aov function provides. First, let’s re-do the ANOVA analysis using the aov function, creating a new object. NOTE This is only necessary if you are wishing to conduct the post-hoc Tukey HSD analysis (which we often are). circadian.aov <- aov(shift ~ treatment, data = circadian) Now we can pass this object to the TukeyHSD function, and use the appropriate confidence level (corresponding to 1 minus alpha): circadianTukey <- TukeyHSD(circadian.aov, conf.level = 0.95) And let’s make the output tidy: circadianTukey.tidy <- circadianTukey %>% broom::tidy() Have a look a the output: circadianTukey.tidy ## # A tibble: 3 × 7 ## term contrast null.value estimate conf.low conf.high adj.p.value ## <chr> <chr> <dbl> <dbl> <dbl> <dbl> <dbl> ## 1 treatment Knee-Control 0 -0.02696 -0.9525 0.8986 0.9970 ## 2 treatment Eyes-Control 0 -1.243 -2.168 -0.3171 0.007866 ## 3 treatment Eyes-Knee 0 -1.216 -2.172 -0.2598 0.01168 Now let’s reformat this table for producing a nice output. circadianTukey.tidy.table <- circadianTukey.tidy %>% select(-c(term, null.value)) %>% # take away these two columns rename(Contrast = "contrast", Difference = "estimate", Conf_low = "conf.low", Conf_high = "conf.high", Adj_Pval = "adj.p.value") Let’s see what we made: circadianTukey.tidy.table ## # A tibble: 3 × 5 ## Contrast Difference Conf_low Conf_high Adj_Pval ## <chr> <dbl> <dbl> <dbl> <dbl> ## 1 Knee-Control -0.02696 -0.9525 0.8986 0.9970 ## 2 Eyes-Control -1.243 -2.168 -0.3171 0.007866 ## 3 Eyes-Knee -1.216 -2.172 -0.2598 0.01168 The output clearly shows the pairwise comparisons and associated P-values, adjusted for multiple comparisons. It also shows the difference in means, and the lower and upper 95% confidence interval for the differences. We can see that the mean for the “Eyes” treatment group differs significantly (at $$\alpha = 0.05$$) from each of the other group means. One typically shows these results visually on the same figure used to display the data (here, Figure 16.1). ### 16.2.8 Visualizing post-hoc test results Here is our original figure, to which we’ll add the results of the post-hoc test: circadian %>% ggplot(aes(x = treatment, y = shift)) + geom_jitter(colour = "darkgrey", size = 3, shape = 1, width = 0.1) + stat_summary(fun.data = mean_se, geom = "errorbar", colour = "black", width = 0.1, position = position_nudge(x = 0.15)) + stat_summary(fun = mean, geom = "point", colour = "black", size = 3, position=position_nudge(x = 0.15)) + xlab("Light treatment") + ylab("Shift in circadian rhythm (h)") + theme_classic() Figure 16.2: Stripchart of phase shifts in the circadian rhythm of melatonin production in 22 participants of an experiment. Solid circles denote group means, and bars +/- one standard error To visualize the outcomes of the Tukey-Kramer post-hoc test, we superimpose a text letter alongside (or above) each group in the figure, such that groups sharing the same letter are not significantly different according to the Tukey-Kramer post-hoc test. For this we use the annotate function, and this requires that we know the coordinates on the graph where we wish to place our annotations. Based on our stripchart above, it looks like we need to use Y-value coordinates of around 1.1 to place our letters above each group. It may take some trial-and-error to figure this out. Here’s the code, with the annotate function added: ggplot(circadian, aes(x = treatment, y = shift)) + geom_jitter(colour = "darkgrey", size = 3, shape = 1, width = 0.1) + stat_summary(fun.data = mean_se, geom = "errorbar", colour = "black", width = 0.1, position=position_nudge(x = 0.15)) + stat_summary(fun = mean, geom = "point", colour = "black", size = 3, position=position_nudge(x = 0.15)) + ylim(c(-3.1, 1.2)) + annotate("text", x = 1:3, y = 1.1, label = c("a", "a", "b")) + labs(x = "Light treatment", y = "Shift in circadian rhythm (h)") + theme_classic() Figure 16.3: Stripchart showing the phase shift in the circadian rhythm of melatonin production in 22 experimental participants given alternative light treatments. Solid circles represent group means, and bars represent +/- one SE. Group means sharing the same letter are not significantly different according to the Tukey-Kramer post-hoc test (family-wise $$\alpha$$ = 0.05). Notice that we have added a line including the ylim function, so that we can specify the min and max y-axis limits, ensuring that our annotation letters fit on the graph. NOTE: The chunk option I included to get the figure heading shown above is as follows: {r fig.cap = "Stripchart showing the phase shift in the circadian rhythm of melatonin production in 22 experimental participants given alternative light treatments. Solid circles represent group means, and bars represent +/- one SE. Group means sharing the same letter are not significantly different according to the Tukey-Kramer post-hoc test (family-wise$\\alpha\$ = 0.05).", fig.width = 4, fig.height = 4.5}

The “family-wise” $$\alpha$$ statement means that the Tukey-Kramer test uses our initial $$\alpha$$ level, but ensures that the probability of making at least one Type-1 error throughout the course of testing all pairs of means is no greater than the originally stated $$\alpha$$ level.

TIP: Figure 16.3 is the kind of figure that should be referenced, typically along with ANOVA table 16.1, when communicating your results of an ANOVA. For instance, with the results of the Tukey-Kramer post-hoc tests superimposed on the figure, you can not only state that the null hypothesis is rejected, but you can also state which group(s) differ from which others.

We are now able to write our truly final concluding statement, below.

1. Annotating stripcharts

Pretend that our post-hoc test showed that the “Knee” group was not different from either the control or the Eyes group. Re-do the figure above, but this time place an “ab” above the Knee group on the figure. This is how you would indicate that the control and eyes grouped differed significantly from one-another, but neither differed significantly from the Knee group.

### 16.2.9 Concluding statement

With out ANOVA table and final figure in hand, we are ready for the concluding statement:

Shifts in circadian rhythm differ significantly among treatment groups (ANOVA; Table 16.1; $$R^2$$ = 0.43). As shown in Figure 16.3, the mean shift among the “Eyes” subjects was significantly lower than both of the other treatment groups.