`superb`

When the group’s variances are heterogeneous, it may results in error bars of differing length. It is then possible than one of the error bar includes the other mean, but that the second error bar does not. What can we conclude in such circumstances? This vignette explores this questions, examining Welch test solution, and Tryon adjustment along the way.

Consider this example where two groups have been examined. Herein, the sample sizes are equal, but what follows does not depends on this.

```
<- c( 56, 54, 73, 46, 59, 62, 55, 53, 77, 60, 69, 66, 63,
grp1 62, 53, 82, 74, 70, 65, 70, 72, 65, 56, 58, 83)
<- c( 51, 99, 194, 123, 40, 83, 87, 117, 46, 89, 61, 81, 53,
grp2 141, 52, 53, 39, 96, 14, 81, 63, 66, 80, 113, 82)
```

These two groups have means of 64.12 and 80.16. However, they also have quite differing standard deviations, 9.493 and 37.918. The second is over four times larger than the first, which denotes a very heterogeneous variance situation.

Let’s reunite the `score`

s of these two datasets in a
single data frame with a `group`

indicator so that we can do
a first plot:

```
<- data.frame( cbind(
dtaHetero id = 1:(length(grp1)+length(grp2)),
group = c(rep(1,length(grp1)), rep(2, length(grp2)) ),
score = c(grp1, grp2)
))head(dtaHetero)
```

```
## id group score
## 1 1 1 56
## 2 2 1 54
## 3 3 1 73
## 4 4 1 46
## 5 5 1 59
## 6 6 1 62
```

(just to make the plots look nicer, I prepare some directives that will be added to these:

```
library(superb) # to make the summary plot
library(ggplot2) # for all the graphic directives
library(gridExtra) # for grid.arrange
= list(
ornate xlab("Difference"),
scale_x_discrete(labels=c("Pre-\nTreatment","Post-\nTreatment")),
ylab("Score"),
coord_cartesian( ylim = c(40,+110) ),
theme_light(base_size = 10) +
theme( plot.subtitle = element_text(size=16))
)
```

). We are now ready to make a first plot. The adjustment
`purpose = "difference"`

is necessary as we plan to compare
the two results.

```
<- superbPlot(dtaHetero,
pt BSFactors = "group",
variables = "score",
adjustments = list(purpose = "difference"),
gamma = 0.95,
statistic = "mean",
errorbar = "CI",
plotStyle = "line",
lineParams = list(alpha = 0) #the line is made transparent
+ ornate
) pt
```

As seen, the two error bars deliver a different message: the left one suggests a huge difference between the means, the right one does not. How do we settle this?

Error bars are depiction of an error variance. However, they are actually based on the standard errors which are the square root of the error variances.

Standard errors cannot be sumed (or averaged), only error variances
can. Thus, when averaging the two error bars, it is their squares that
must be averaged (followed by a square root). This operation is called
the **mean in the square sense**. It is denoted
**mean* ** hereafter.

Given that the first error bar is given by \(\bar{X}_1 \pm \sqrt{2} \;t \; SE_1\) and the second, by \(\bar{X}_2 \pm \sqrt{2} \;t \; SE_2\), where \(\bar{X}_i\) is the average of the \(i^{\text{th}}\) group and \(SE_i\) is the standard error of the the \(i^{\text{th}}\) group, averaging the error bar lengths can be simplified to

\[ \DeclareMathOperator{\mean}{Mean} \begin{split} \mean{^*}( \text{ the two confidence intervals }) &= \mean{^*} \left( \sqrt{2} \;t \; SE_1, \sqrt{2} \; t \; SE_2 \right) \\ &= \sqrt{\mean \left( (\sqrt{2} \;t \;SE_1)^2, (\sqrt{2} \;t \;SE_2)^2 \right) } \\ &= \sqrt{\mean \left( 2 \;t^2 \;SE_1^2, 2 \;t^2 \;SE_2^2 \right) } \\ &= \sqrt{2} \;t\; \sqrt{\mean( SE_1^2, SE_2^2) } \\ &= \sqrt{2} \;t\; \sqrt{ \frac{SE_1^2 + SE_2^2}{2} } \\ &= t \;\sqrt{SE_1^2 + SE_2^2} \end{split} \]

Hence, asking if the **lag** between the two groups is
significant is actually asking

\[ \DeclareMathOperator{\mean}{Mean} \begin{split} \text{Reject $H_0$ if }& {\bar{X}_1-\bar{X}_2} > \mean^*( \text{ the two confidence intervals })\\ \text{Reject $H_0$ if }& {\bar{X}_1-\bar{X}_2} > t\; {\sqrt{SE_1^2 + SE_2^2}} \\ \text{Reject $H_0$ if }& \frac{\bar{X}_1-\bar{X}_2} {\sqrt{SE_1^2 + SE_2^2}} > t \\ \end{split} \]

which is precisely the definition of a Welch test (Welch, 1947).

In other word, when appraising the general length of the error bars, you are actually performing a Welch test. This test is arguably the only mean test we need (Delacre, Lakens, & Leys, 2017).

Actually, the above is not entirely true, as the degrees of freedom of the \(t\) multiplier is only based on the number of observations in each group (\(n-1\)).

To really have a Welch test, it is necessary to (1) pool the degrees of freedom of the groups together; (2) rectify this degree of freedom based on the amount of heterogeneity in the data.

To get the rectified degrees of freedom according to Welch, you can issue

```
# Welch's rectified degrees of freedom
<- WelchDegreeOfFreedom(dtaHetero, "score","group") wdf
```

in which `"score"`

is the column(s) containing the
dependent variable(s) and `"group"`

is the column(s)
containing the group identifiers. Here, the result is 26.9965 (compared
to \(n_1+n_2-2\), that is 48,
indicating a fairly large reduction in degree of freedom caused by a
fairly important heterogeneity.

In `superb`

, it is possible to override the default degree
of freedom using the error bar estimator
`errorbar = CIwithDF"`

and then in the `gamma`

argument, use a vector with both the confidence level desired and the
rectified degree of freedom, as in
`gamma = c(0.95, 26.9965)`

. Lets have a small plot of
this:

```
<- superbPlot(dtaHetero,
pw BSFactors = "group",
variables = "score",
adjustments = list(purpose = "difference"),
gamma = c(0.95, wdf), # new!
statistic = "mean",
errorbar = "CIwithDF", # new!
plotStyle = "halfwidthline",
lineParams = list(alpha = 0)
+ ornate
) pw
```

In this plot, I used a different layout, the
`"halfwidthline"`

which shows the mid-point of the error bar
with a small gap. We will see the reason for that in a minute.

Averaging visually two lengths is already difficult, if it needs to be done in the square sense, that may be impossible… When averaging in the square sense, the longer lines have a stronger influence on the means so that ignoring this fact, we would tend to underestimate the mean length.

The first solution is not to bother. The usual average and the average in the square sense returns sensibly the same length. Therefore, if your results are not ambiguous, there is no need to worry any more.

The second solution is to adjust the error bar length so that they are amenable to the usual average. This is actually what Tryon’s adjustment does (Tryon, 2001). When the variances are truly homogeneous, the correction for examining pair-wise comparision is an increase by a factor of \(\sqrt{2}\). When we deviate from homogeneity, Tryon’s adjustment tends to be a little larger than 1.41 to compensate for the underestimation introduced by regular averaging. This transformation is right on so that the usual mean of the two error bars, following a Tryon transformation, will precisely indicate whether the means are included in the mean error bar length.

```
<- superbPlot(dtaHetero,
pwt BSFactors = "group",
variables = "score",
adjustments = list(purpose = "tryon"), #new!
gamma = c(0.95, wdf),
statistic = "mean",
errorbar = "CIwithDF",
plotStyle = "halfwidthline",
lineParams = list(alpha = 0)
+ ornate
) pwt
```

Let’s return to the whole question that motivated this vignette.

If we used the difference-adjusted error bars, then we need to average in the square sense the error bars. Lets do that:

```
# get the summary statistics with superbData
<- superbData(dtaHetero,
t BSFactors = "group",
variables = "score",
adjustments = list(purpose = "difference"),
gamma = 0.95,
statistic = "mean", errorbar = "CI"
)# keep only the summary statistics:
<- t$summaryStatistics
t2
# the length is in column "upperwidth", for lines 1 and 2,
# so lets do the mean in the square sense:
<- sqrt( (t2$upperwidth[1]^2 + t2$upperwidth[2]^2)/2 ) tmean2
```

which equals 16.135. By comparison, here is the regular mean:

`<- (t2$upperwidth[1] + t2$upperwidth[2] )/2 tmean `

which equals 13.838. This shows that in the present case, the usual average of the error bars is a length too short compared to the adequate average (in the square sense).

We’ll add a line segment of that length in the subsequent plot. Let’s do the same following Tryon’s adjusment and Welch rectified degrees of freedom:

```
# get the summary statistics with superbData
<- superbData(dtaHetero,
wt BSFactors = "group",
variables = "score",
adjustments = list(purpose = "tryon"),
gamma = c(0.95, wdf),
statistic = "mean", errorbar = "CIwithDF"
)<- wt$summaryStatistics
wt2 <- (wt2$upperwidth[1]+wt2$upperwidth[2]) / 2 wmean
```

As seen, both length are very similar, 16.135 for the first, and 16.041 for the second. The small difference comes from the fact that in the Tryon-adjusted error bars, the degree of freedom were rectified by Welch’s formula rather than being unpooled.

With Tryon’s adjustment, the length can be averaged, which means that
half of each length totalizes the average length: this is where the
`halfwidthline`

error bars come handy: we see the half of
their length. Hence, if they are aligned, their sum is exactly equal to
the lag between the two means.

Lets represent all this:

```
# showing all three plots, with reference lines in red
grid.arrange(
+ labs(subtitle="Difference-adjusted 95% CI\n with default degree of freedom") +
pt geom_text( x = 1.15, y = mean(grp1)+(mean(grp2)-mean(grp1))/2, label = "power-2 mean", angle = 90) +
geom_text( x = 1.55, y = mean(grp1)+(mean(grp2)-mean(grp1))/2, label = "regular mean", angle = 90) +
geom_hline(yintercept = mean(grp1)+(mean(grp2)-mean(grp1))/2-tmean2/2, colour = "red", linewidth = 0.5, linetype=2) +
geom_hline(yintercept = mean(grp1)+(mean(grp2)-mean(grp1))/2+tmean2/2, colour = "red", linewidth = 0.5, linetype=2) +
# arrow for the power-2 mean
geom_segment(arrow = arrow(length =unit(0.4,"cm")), x=1.33, y=mean(grp1)+(mean(grp2)-mean(grp1))/2,
xend=1.33, yend=mean(grp1)+(mean(grp2)-mean(grp1))/2+tmean2/2) +
geom_segment(arrow = arrow(length =unit(0.4,"cm")), x=1.333, y=mean(grp1)+(mean(grp2)-mean(grp1))/2,
xend=1.33, yend=mean(grp1)+(mean(grp2)-mean(grp1))/2-tmean2/2) +
# arrow for the regular mean
geom_segment(arrow = arrow(length =unit(0.4,"cm")), x=1.66, y=mean(grp1)+(mean(grp2)-mean(grp1))/2,
xend=1.66, yend=mean(grp1)+(mean(grp2)-mean(grp1))/2+tmean/2) +
geom_segment(arrow = arrow(length =unit(0.4,"cm")), x=1.66, y=mean(grp1)+(mean(grp2)-mean(grp1))/2,
xend=1.66, yend=mean(grp1)+(mean(grp2)-mean(grp1))/2-tmean/2),
+ labs(subtitle="Difference-adjusted 95% CI\nwith df from Welch"),
pw
+ labs(subtitle="Tryon-adjusted 95% CI\nwith df from Welch") +
pwt geom_hline(yintercept = mean(grp1)+wt2$upperwidth[1]/2, colour = "red", linewidth = 0.5, linetype=2) +
geom_hline(yintercept = mean(grp2)+wt2$lowerwidth[2]/2, colour = "red", linewidth = 0.5, linetype=2),
ncol=3)
```

From the Tryon-adjusted confidence intervals, we see that the average length is precisely the lag between the means. In other word, the two groups are borderline significantly different. Still not convinced? Let’s do a Welch test:

```
t.test( grp1, grp2,
var.equal=FALSE)
```

```
##
## Welch Two Sample t-test
##
## data: grp1 and grp2
## t = -2.0518, df = 26.996, p-value = 0.05001
## alternative hypothesis: true difference in means is not equal to 0
## 95 percent confidence interval:
## -3.208072e+01 7.239399e-04
## sample estimates:
## mean of x mean of y
## 64.12 80.16
```

The difference is right on the limit of significance at the .05 level, mirroring the 95% confidence level used in the plot.

`superb`

is making error bars from separate error
estimations. Hence, whenever the sample sizes are quite unequal, the
variances are quite heterogeneous, or both, the inference performed by
eyes are actually the same as performing a Welch test of means. When
both samples sizes are equal and variances are homogeneouse, there is no
difference between a Welch test and a *t* test.

When the degree of freedom are rectified and Tryon’s ajustment is used, the center of both confidence intervals should be exactly aligned to indicate that their mean fills exactly the gap.

Although the Tryon/Welch pair lets you make very precise inferences, visually and practically speaking, the difference between the rectified Welch degree of freedom adjustments and a plot with the default degree of freedom are undetectable (as evidence by the first plot and the second plot of Figure 4 just above).

If you can perform average in the square sense, the difference adjustment is perfect. In doubt, consider applying the Tryon adjustment when the variances are cleary heterogeneous.

Delacre, M., Lakens, D., & Leys, C. (2017). Why psychologists should
by default use welch’s t-test instead of student’s t-test.
*International Review of Social Psychology*, *30*(1). https://doi.org/10.5334/irsp.82

Tryon, W. W. (2001). Evaluating statistical difference, equivalence, and
indeterminacy using inferential confidence intervals: An integrated
alternative method of conducting null hypothesis statistical tests.
*Psychological Methods*, *6*, 371–386. https://doi.org/10.1037/1082-989X.6.4.371

Welch, B. L. (1947). The generalization of student’s’ problem when
several different population variances are involved.
*Biometrika*, *34*(1/2), 28–35. https://doi.org/10.1093/biomet/34.1-2.28