Mediation (David A. Kenny)
Mediation (David A. Kenny)
PDF Power
David A. Kenny
May 4, 2021
MEDIATION
Introduction
The Four Steps
Indirect Effect
Power
Specification Error
Reporting Results
Additional Variables
Extensions
Causal Inference Approach
Links to Other Sites
References
Introduction
Consider a variable X that is
assumed to cause another variable Y. The variable X is called the causal variable and the variable that
it causes or Y is called the outcome.
In diagrammatic form, the unmediated model is
(These two
diagrams are essential to the understanding of this page. Please study
them carefully!) Path c' is
called the direct effect. The
mediator has been called an intervening or process variable. Complete mediation is the case in which
variable X no longer affects Y after
M has been controlled, making path c' zero. Partial mediation is the
case in which the path from X to Y is reduced in absolute size but is
still
different from zero when the mediator is introduced.
The Steps
Baron and Kenny (1986), Judd and
Kenny (1981), and James and Brett (1984) discussed four steps in establishing mediation:
Step 4: To
establish that M completely mediates the X-Y relationship, the effect of X on Y
controlling for M (path c') should be
zero (see
discussion below on significance testing). The effects in both
Steps 3 and 4 are estimated in the same equation.
Inconsistent Mediation
If c' were
opposite in sign to ab something that MacKinnon, Fairchild, and Fritz (2007)
refer to as inconsistent mediation,
then it
could be the case that Step 1 would not be met, but there is still
mediation. In this case the mediator acts like a suppressor
variable. One
example of inconsistent
mediation is the relationship between stress and mood as mediated by
coping. Presumably, the direct effect is
negative: more stress, the worse the mood. However, likely the effect of stress on coping is positive (more stress,
more coping) and the
effect of coping on mood is positive (more coping, better
mood), making the indirect effect positive. The total effect of stress on mood
then is likely to be very small
because the direct and indirect effects will tend to cancel each other out. Note too that with inconsistent
mediation
that typically the direct effect is even larger than the total effect.
or using
symbols
c = c' + ab
Note also
that the indirect effect equals the reduction of the effect of the causal
variable on the outcome or ab = c - c'. In contemporary
mediational analyses, the
indirect effect or ab is the measure
of the amount of mediation.
Most often the indirect effect is computed directly as the product of a and b. Below are discussed three different ways to test the
product of the two coefficients. Imai, Keele, and Tingly (2010) have re-proposed the use of c - c' as the measure of the indirect effect. They
make the claim that difference in coefficients is more robust to certain forms of specification error. It is unclear at this point if the
difference in coefficients approach will replace the product in coefficients approach. It is also noted here that the Causal Inference
Approach (Pearl, 2011) has developed a very general approach to measuring the indirect effect.
Below are described four tests of the indirect effect or ab. Read carefully as some of the tests have key drawbacks that should be
noted. One key issue concerns whether paths a and b are correlated: If path a is over-estimated, is path b also over-estimated? Paths a
and b are uncorrelated when multiple regression is used to estimate them, but are not for most other methods. The different tests make
different assumptions about this correlation.
Sobel Test
A test, first proposed by Sobel (1982), was initially often used. Some sources refer to this test as the delta method. It requires the
standard error of a or sa (which equals a/ta where ta is the t test of coefficient a) and the
standard error of b or sb. The Sobel test provides
an approximate estimate of the standard error of ab which equals to the square root of
b2sa2 + a2sb2
Bootstrapping
An increasingly popular method of
testing the indirect effect is bootstrapping (Bollen & Stine, 1990; Shrout &
Bolger, 2002).
Bootstrapping is a
non-parametric method based on resampling with replacement which is done many
times, e.g., 5000 times. From each
of
these samples the indirect effect is computed and a sampling distribution can
be empirically generated. Because the
mean of the
bootstrapped distribution will not exactly equal the indirect
effect a correction for bias can be made. With the distribution, a confidence
interval, a p value, or a standard error can be determined. Very typically a confidence interval is
computed and it is checked to determine if
zero is in the interval. If zero is not in the interval, then the
researcher can be confident that the indirect effect is different from
zero. Also
a Z value can determined by dividing the bootstrapped estimate by its standard error, but bootstrapped standard errors suffer the same
problem
as the Sobel standard errors and are not recommended. (Bootstrapping does not require the assumption that a and b are
uncorrelated.)
Fritz, Taylor, and MacKinnon (2012) have raised concerns that bias-corrected bootstrapping test is too liberal with alpha being
around .07. Actually not doing the bias correction seems to improve the Type I error rate. Hayes and Scharkow (2013) recommended
using the bias corrected bootstrap if power is the major concern, but if Type I error rate is the major concern, then the percentile bootstrap
should be used.
Several SEM programs (e.g., lavann, Mplus and Amos) can be used to
bootstrap. Also Hayes and Preacher have written SPSS and
SAS macros that can be downloaded for tests of indirect effects (click here to download the Hayes and Preacher macro; look for "SPSS
and SAS macros to accompany Preacher & Hayes (2004) paper on mediation.").
Power
Effect Size of the Indirect Effect and the Computation of Power
The indirect effect is the product of two effects.
One simple way, but not the only way, to determine the effect size is to measure
the product of the two effects, each turned into an effect size. The standard effect size for paths a and b is a partial correlation; that is, for
path a, it is the correlation between X and M, controlling for the covariates and any other Xs and for path b, it is the correlation between M
and Y, controlling for covariates and other Ms and Xs. One possible effect size for the indirect effect would be the product of the two partial
correlations.
(Preacher and Kelley (2011) discuss a similar measure of effect size which they refer to as the completely standardized
indirect effect, which uses betas, not partial correlations.)
There are two different strategies for determining small, medium, and large effect sizes. (Any designation of small, medium, or
larger is fundamentally arbitrary and depends on the particular application.)
First, following Shrout and Bolger (2002), the usual Cohen
(1988) standards of .1 for small, .3 for medium, and .5 for large could be used. Alternatively and I think more appropriately because an
indirect effect is a product of two effects, these values should be squared. Thus, a small effect size would be .01, medium would .09, and
large would be .25. Note that if X is a dichotomy, it makes sense to replace the correlation for path a with Cohen’s d. In this case the
effect size would be a d times an r and a small effect size would be .02, medium would .15, and large would be .40.
One strategy to compute the power of the test of the indirect effect is to use the joint test of significance . Thus, one computes the
power of test of paths a and b and then multiply their power to obtain the power of the test of the indirect effect. One can use the app that I
have written called MedPower to conduct a power analysis. (There is a webinars on power and a earlier version of the app called
PowMedR.) Alternatively and more generally, one could use a structural equation modeling to run a simulation to conduct a power
analysis.
Multicollinearity
If M is a successful mediator, it is
necessarily correlated with X due to path a. This correlation, called
collinearity, affects the
precision of the estimates of the last regression equation. If X were to explain all of the variance in M, then
there would be no unique
variance in M to explain Y. Given that path a is nonzero, the power of the tests of
the coefficients b and c’ is lowered. The effective
sample
size for the tests of coefficients b and c’ is approximately N(1 - r2) where N is
the total sample size and r is the
correlation between the
causal variable and the mediator, which is equal to
standardized a. So if M is a strong mediator (path a is large), to achieve equivalent
power, the sample size to test coefficients b and c' would have to be larger
than what it would be if M were a weak
mediator.
Multicollinearity is to be expected in a mediational analysis and it cannot be avoided.
Low Power
for Steps 1 and 4
As described by Kenny and Judd (2014), as well as others, the tests of c and c’ have relatively low power, especially in comparison
to the indirect effect. It can easily happen, that ab can be statistically
significant but c is not. For instance, if a = b = .4 and c’ = 0, making c
= .16, and N = 100, the power of the test of path a is .99, the power of the test of path b is .97 which makes the power of ab equal to about
.96, but the power of the test that c is only .36. Surprisingly,
it is very easy to have complete mediation, a statistically significant
indirect
effect, but no statistical evidence that X causes Y.
Because of the low power in the test of c’, one needs to be very careful about any claim of complete mediation based on the non-
significance of c’. In fact, several sources (e.g., Hayes, 2013) have argued that one should never make any claim of complete or partial
mediation. It seems more sensible to be careful about claims of complete mediation. One idea is establish first that is sufficient power to
test for partial mediation: The power of the test of c’. Also if the sample size is very large, then finding a significant value for c’ and so
"partial" mediation is not very informative. More informative, in the case of large N, is knowing the proportion of the total effect that is
mediated or ab/c.
Specification Error
Mediation is a hypothesis about a causal
network. (See Kraemer, Wilson, Fairburn, and Agras (2002) who attempt to
define
mediation without making causal assumptions.) The conclusions from
a mediation analysis are valid only if the causal assumptions are
valid (Judd
& Kenny, 2010). In this section, the three major assumptions of
mediation are discussed. Mediation analysis also makes all of
the
standard assumptions of the general linear model (i.e., linearity, normality,
homogeneity of error variance, and independence of
errors). It is strongly advised to check these assumptions
before conducting a mediational analysis. Clustering effects are discussed in
the Extensions section. What follows are sufficient conditions. That is, if the assumptions are met, the mediational model is identified.
However, there are sometimes special cases in which an assumption can be violated, yet the mediation effects are identified (Pearl, 2014).
Reverse
Causal Effects
The mediator may be caused by the outcome
variable (Y would cause M in the above diagram), what is commonly called a feedback
model. When the causal
variable is a manipulated variable, it cannot be caused by either the mediator
or the outcome. But because both
the mediator and the outcome variables
are not manipulated variables, they may cause each other.
Often it is advisable to interchange the
mediator and the outcome variable and have the outcome "cause" the
mediator. If the
results look similar to the specified mediational pattern
(i.e., the c' and b are about the same in the two models), one would be less
confident in the specified model. However, it should be realized that the direction of causation between M
and Y cannot be determined by
statistical analyses.
If it can be assumed that c' is zero, then reverse causal effects
can be estimated. That is, if it can be assumed that there is complete
mediation (X does not directly cause Y and so c’ is zero), the mediator may cause the outcome and the outcome
may cause the mediator
and the model can be estimated using instrumental
variable estimation.
Measurement
Error in the Mediator
If the mediator is measured with less
than perfect reliability, then the effects (b and c')
are likely biased. The effect of the mediator
on the outcome (path b) is likely
underestimated and the effect of the causal variable on the outcome (path c') is likely over-estimated if ab
is positive (which is typical). The
over-estimation of c' is exacerbated
to the extent to which path a is
large. In a parallel fashion, if X is
measured with less than perfect
reliability, then the effects (b and c') are likely biased. The effect of the
M on Y mediator on the outcome
(path b)
is likely over-estimated and the effect of the causal variable on the outcome
(path c') is likely
under-estimated. Moreover,
measurement
error in X attenuates the estimate of path a and c. Measurement error in Y does not bias
unstandardized estimates, but it
does bias standardized estimates, attenuating
them.
Omitted Variables
In this case, there is a variable that
causes both variables in the equation. These variables are called confounders in some
literatures and the assumption can be stated more formally and generally, see below. For example, there is a variable that causes both the
mediator and the outcome. This is the most
likely source of specification error and is difficult to find solutions to circumvent it.
[Unfortunately, this key assumption is
not directly discussed in Baron and Kenny (1986), but is discussed in
Judd and Kenny (1981).]
Although there has been some work on the omitted
variable problem, the only complete solution is to specify and measure such
variables
and control for their effects.
As stated above, the absence of omitted variables is not a necessary condition (Pearl, 2014). Consider the case that there is an
unmeasured confounding variable, C, that causes M and Y. The first situation (called the backdoor condition) is to find a measured variable
Z which is either a complete mediator of the C to M or the C to Y relationship either M ← Z ← C → Y or M ← C → Z → Y . By controlling for
Z, one can remove the confounding effect of C.The second situation (called the frontdoor condition) is to find a measured variable Z which
is a complete mediator of the M to Y relationship. The effect from M to Y is given by the product of the path from M to Z controlling for X
and the path from Z to Y controlling for X and M (i.e., the indirect effect from M to Y).
Brewer, Campbell, and Crano (1970) argued that in some cases when X is not manipulated, it might be that single unmeasured
variable can explain the covariation between all the variables. In fact, unless there is inconsistent mediation, a single latent variable can
always explain the covariation between X, M, and Y (i.e., a solution that converges with no Heywood cases).
The
Combined Effects of Measurement Error and an Omitted Variable
Fritz, Kenny, and MacKinnon (2017) point out that in the case in which X is manipulated and M has measurement error and there is a
confounder for M and Y, it can happen that the two biases can to some degree offset each other. That is, assuming a and b being positive,
measurement error in M in this case leads to under-estimation of b and over-estimation of c' whereas typically (but not always) a MY
confounder leads to over-estimation of b and under-estimation of c'. See the paper for details and an example.
The
Mediator as also a Moderator
Baron and Kenny (1986) and Kraemer et al. (2002) discuss the possibility that M might
interact with X to cause Y. Baron and
Kenny
(1986) refer to this as M being both a mediator and a moderator and
Kraemer et al. (2002) as a form of mediation. The X with M interaction
should be estimated and tested and added to the model if present. Judd and Kenny (1981) discuss how the meaning of path b changes
when this interaction is present. Also the Causal Inference Approach begins with the assumption that X and M interact and treats the
interaction as part of the mediation.. Because it is relatively easy to test this assumption, it is advisable to do so before conducting the
mediational analysis.
Design Considerations
One of the best
ways to increase the internal validity of mediational analysis is by the design
of study. Key considerations are
randomizing X (i.e., randomly assigning units to levels of X), the timing of
measurement of M and Y, and obtaining prior values of M and Y.
By randomizing X, it is known that both M and
Y do not cause X. By measuring M after X,
and Y after M, it is known that M does not cause
X and that Y does not cause X
or M. Finally by obtaining prior
measures of M and Y and control for them, we can reduce and perhaps
eliminate
the effects of omitted variables. The
reader should consult Cole and Maxwell (2003) about the difficulties of
estimating
mediational effect using a cross-sectional design. Also as mentioned earlier, it is possible to
randomize X, M, and Y (Smith, 1982). .
Sensitivity Analyses
As discussed above, it is usually assumed for mediation that there is perfect reliability for X and M, no omitted variables for the X to
M, X to Y, and M to Y relationships, and no causal effects from Y to X and M and from M to X. It is possible to determine what would
happen to the mediational paths one or more of these assumptions is violated by conducting sensitivity analyses. For instance, one might
find that allowing for measurement error in M (reliability of .8) that path b would be larger by 0.15 and that c' would be less by 0.10.
Alternatively, one might determine what was the value of reliability that would make c' equal zero.
One way to conduct a sensitivity analysis is to estimate the mediational model using Structural Equation Modeling. One fixes the
additional parameter to the value of interest. For example, an omitted variable is added that has a moderate effect on M and Y. One can
then use the estimates from this analysis in the sensitivity analysis. See the Sensitivity Analyses webinar (small charge) that I have created
which has more details. A convincing mediation analyses should be accompanied by a sensitivity analysis.
Reporting Results
These days, one does not report results from the steps, but rather reports the indirect effect and its confidence interval. Additionally
the paths a, b, c, and c', as well their statistical significance (or confidence interval) are reported. Although some have suggested not
drawing conclusions about complete versus partial mediation, it seems something that is reasonable, although one must be care about
issues of power. It is essential that one discuss the likelihood of meeting the assumptions of mediational analysis and ideally reporting on
the results of sensitivity analyses. See the paper by Muller et al. (2008) on reporting results of mediational analyses.
Additional Variables
Rarely in
mediation are there just the three variables of X, M, and Y. Discussed in this section is how to handle
additional variables
in a mediational model.
Multiple
Mediators
If there are
multiple mediators, they can be tested simultaneously or separately. The
advantage of doing them simultaneously is that
one learns if the mediation is
independent of the effect of the other mediators. One should make sure
that the different mediators are
conceptually distinct and not too highly
correlated. [Kenny et al. (1998) consider an example with two mediators.]
There is an interesting case of two mediators (see below) in which the indirect effects are opposite in sign. The sum of
indirect
effects for M1 and M2 would be zero. It might then be possible
that the total effect or c is near zero, because there
are two indirect effects
that work in the opposite direction. In this
case "no effect" would be mediated. I suggest that the case in which there are two indirect
effects of the same effect that are approximately equal in size but opposite in sign be called opposing mediation.
Multiple
Outcomes
If there are multiple outcomes, they can
be tested simultaneously or separately. If tested simultaneously, the
entire model can be
estimated by structural equation modeling. One might want to consider combining the
multiple outcomes into one or more latent
variables.
Multiple
Causal Variables
In this case there are multiple X
variables and each has an indirect effect on Y. The Hayes and Preacher bootstrapping macro can be
used to test
hypotheses about the linear combinations of indirect effects: For example, are
they equal? Do they sum to zero? One can
alternatively treat the multiple X
variables as a formative variable and
so if a single “super variable” can be used to summarize the indirect
effect. As seen below, the formative
variable X “mediates” the effect of X on M and Y. The model can be tested and it has k - 1 degrees of
freedom where k is the number of X variables. Thus, the degrees of freedom for the example
would be 1.
Covariates
There are often variables that do not
change that can cause or be correlated with the causal variable, mediator, and
outcome (e.g.,
age, gender, and ethnicity); these variables are commonly called covariates. They would generally be included in the M and Y equations, A
covariate would not be trimmed from one equations unless it is dropped from all of the
other equation. If a covariates interacts with X or M,
it would be called
a moderator variable.
Extensions
Mediated
Moderation and Moderated Mediation
Moderation means that the effect of a
variable on an outcome is altered (i.e., moderated) by a covariate. (To read
about moderation
click here.) Moderation
is usually captured by an interaction between the causal variable and the
covariate. If this moderation is mediated,
then we have the usual pattern
of mediation but the X variable is an interaction and the pattern would be
referred to as mediated
moderation. All the Baron and Kenny
steps would be repeated with the causal variable or X being an interaction, and
the two main effects
would be treated as "covariates." We could compute the total effect or the
original moderation effect, the direct effect or how much
moderation exists
after introducing the moderator, and the indirect effect or how much of the
total effect of the moderator is due to the
mediator.
Latent Variables
Some or all of the mediational variables
might be latent variables. Estimation
would be accomplished using a structural equation
modeling (SEM) program (e.g.,
LISREL, Amos, Eqs, or MPlus). Some programs provide measures and tests of
indirect effects with
bootstrapping. Also such programs are quite flexible in handling
multiple mediators and outcomes. The one complication is how to
handle
Step 1. That is, if two models are estimated, one with the mediator and
one without, the paths c and c’ are not directly comparable
because the
factor loadings would be different. It is then inadvisable to test the
relative fit of two structural models, one with the mediator
and one
without. Rather c, the total
effect, can be estimated using the formula of c' + ab. Most SEM programs give this estimate.
If
there are multiple mediators, Amos does not compute indirect effects for each
mediator. The reader should consult Macho
and
Ledermann (2011) for a method that does decompose the total indirect effect
into separate effects.
One
advantage of a latent variable model is that correlated measurement error in X,
M, and Y might be modeled. For instance,
if
some of the measures are self-report, their errors might be correlated.
Dichotomous
Variables
If either the mediator or the
outcome is a dichotomy, standard methods of estimation should not be used. (Having the causal
variable or X be a dichotomy is not
problematic.) If either the mediator or the
outcome is a dichotomy, the analysis would likely be
conducted using
logistic regression. One can
still use the Baron and Kenny steps. The Sobel test is problematic in that it assumes that a
and b are independent, which may not be the case. The one
complication is the computation of indirect effect the degree of mediation
because the coefficients need to be transformed. (To read about the
computation of indirect effects using logistic or probit regression
click here.) With dichotomous outcomes, it is advisable to
use a program like Mplus that can handle such variables.
Clustered
Data
Traditional mediation analyses presume
that the data are at just one level. However, sometimes the data are clustered in that
persons are in
classrooms or groups, or the same person is measured over time. With clustered data, multilevel modeling
should be
used. Estimation of mediation
within multilevel models can be very complicated, especially when the mediation
occurs at level one and
when that mediation is allowed to be random, i.e., vary
across level two units. The reader is referred to Krull and MacKinnon
(1999), Kenny,
Korchmaros, and Bolger (2003), and Bauer and Preacher (2006) for
a discussion of this topic. Recently, Preacher, Zyphur, and Zhang (2010)
have proposed that multilevel structural equation methods or MSEM can be used to estimate these models. Ledermann, Macho, and
Kenny (2011) discuss
mediational models for dyadic data.
Over-Time
Data
Over-time data can be treated as clustered data, but there are further complications due to temporal nature of the data. Among such
issues are correlated errors, lagged effects, and the outcome causing the mediator. One might consult Bolger and Laurenceau (2013) for
guidance. Also, Judd, Kenny, and McClelland, (2001) discuss a generalization of repeated measures analysis of variance test of mediation.
Using an SEM Program Instead of Multiple Regression
Traditionally the mediation model is estimated by estimating a series of multiple regression equations. However, there are
considerable advantages to estimate the program using a Structural Equation Modeling (SEM) program, such as Amos or Mplus. First, all
the coefficients are estimated in a single run. Second most SEM programs provide estimates of indirect effects and bootstrapping. Third,
SEM with FIML estimation can allow for a more complex model of missing data. Fourth, it is relatively easy to conduct sensitivity analyses
with a SEM program.
Normally with SEM, one computes a measure of fit. However, the mediational model is saturated and a no usual measure of fit is
possible. However, one can adopt the following strategy. Use a fit statistic an "Information" measure like the AIC or BIC. I actually prefer
the SABIC which, at least in my experience, performs better than the other two. With these measures a fit value can be obtained for a
saturated model. Then one can compute the measure for each of the following models: no direct effect, no effect from causal variable to
the mediator, and no effect from the mediator to outcome. Then the best fitting model is the one with lowest value.
A group of researchers have developed an approach that has several emphases that are different from the traditional SEM approach,
the approach that is emphasized on this page. The approach is commonly called the Causal Inference approach, and I provide here a brief
and relatively non-technical summary in which I attempt to explain the approach to those more familiar with Structural Equation Modeling.
Robins and Greenland (1992) conceptualized the approach and more recent papers within this tradition are Pearl (2001; 2011) and Imai et
al. (2010). Somewhat more accessible is the paper by Valeri and VanderWeele (2013). Unfortunately, SEMers know relatively little about
this approach and, I believe also that Causal Inference researchers fail to appreciate the insights of SEM.
The Causal Inference Approach uses the same basic causal structure (see diagram) as the SEM approach, albeit usually with
different symbols for variables and paths. The two key differences are that the relationships between variables need not be linear and the
variables need not be interval. In fact, typically the variables of X, Y, and M are presumed to be binary and that X and M are presumed to
interact to cause Y.
Similar to SEM, the Causal Inference approach attempts to develop a formal basis for causal inference in general and mediation in
particular. Typically counterfactuals or potential outcomes are used. The potential outcome for person i on Y for whom X = 1 would be
denoted as Yi(1). The potential outcome of Yi(0) can be defined even though person i did not score 0 on X. Thus, it is a potential outcome
or a counterfactual. The averages of these potential outcomes across persons are denoted as E[Y(0)] and E[Y(1)]. To an SEM modeler,
potential outcomes can be viewed as predicted values of a structural equation. Consider the "Step 1" structural equation:
Yi= d + cXi + ei
If for individual i for whom Xi equals 1, then Yi(1) = d + c + ei equals his or her score on Y. We can determine what the score of person i
would have been had his or her score on Xi been equal to 0, i.e., the potential outcome for person i, by taking the structural equation and
setting Xi to zero to yield d + ei. Although the term is new, potential outcomes are not really new to SEMers. They simply equal the
predicted value for endogenous variable, once we fix the values of its causal variables.
The Causal Inference approach also employs directed acyclic graphs or DAGs, which are similar to, though not identical to, path
diagrams. DAGs typically do not include disturbances but they are implicit. The curved lines of path diagrams between exogenous
variables are also not drawn but are implicit.
Assumptions
Earlier, the assumptions necessary for mediation were stated using structural equation modeling terms. Within the Causal Inference
approach, there are essentially the same assumptions, but they are stated somewhat differently. Note that the term confounder is used
where earlier the term omitted variable was used.
Condition 1: No unmeasured confounding of the XY relationship; that is, any variable that causes both X and Y must be included in the
model.
Condition 4: Variable X must not cause any known confounder of the MY relationship.
Note also that these assumptions are sufficient but not necessary. That is, if these conditions are met the mediational paths are identified,
but there are some special cases where mediational paths are identified even if the assumptions are violated (Pearl, 2013). For instance,
consider the case that M ← Z1 ← Z2 → Y but Z1 and not Z2 is measured and included in the model. Note that Z2 is a MY confounder and
thus violates Condition 2, but it is sufficient to control for only Z1.
The Causal Inference approach emphasizes sensitivity analyses: These are analyses that ask the question such as, “What would happen
to the results if there was a MY confounder that had both a moderate effect on M and Y?” SEMers would benefit by considering these
analyses more often.
E[Y(1)] - E[Y(0)]
This looks strange to an SEMer, but it is useful to remember effects can be viewed as a difference between what the outcome would be
when the causal variable differs by one unit. Consider path c in mediation. We can view c as the difference between what it would be
expected that Y would equal when X was 1 and equal to 0, the difference between the two potential outcomes, E[Y(1)] - E[Y(0)].
In the Causal Inference approach, there is the Controlled Direct Effect or CDE for the mediator equal to a particular value, denoted as
M (not to be confused with the variable M):
where M is a particular value of the mediator. Note that it is E[Y(1,M)] and not E[Y(1|M)], the expected value of Y given that X equals 1
"controlling for M." If X and M interact, the CDE(M) changes for different values of M. To obtain a single measure of the direct effect,
several different suggestions have been made. Although the suggestions are different, all of these measures are called "Natural." One
idea is to determine the Natural Direct Effect as follows
where M0 is M(0) which is the expected value on the mediator if X were to equal 0 (i.e., the potential outcome of M given X = 0). Thus, within
this approach, there needs to be a meaningful "baseline" value for X which becomes its zero value. For instance, if X is the variable
experimental group versus control group, then the control group would have a score of 0. However, if X is level of self-esteem, it might be
more arbitrary to define the zero value. The parallel Natural Indirect Effect is defined as
where M1 is M(1) or the potential outcome for M when X equals 1. The Total Effect becomes the sum of the two:
Note that both the CDE and the NDE would equal the regression slope or what was earlier called path c' if the model is linear,
assumptions are met, and there is no XM interaction affecting Y, the NIE would equal ab, and the TE would equal ab + c'. In the case in
which the specifications made by traditional mediation approach (e.g., linearity, no omitted variables, no XM interaction), the estimates
would be the same.
Here I give the general formulas for the NDE and NIE when X is an intervally measured based on Valeri & VanderWeele, (2013). If the
XM effect is added to the Y equation, that equation can be stated as
Y = iY + c'X + bM + dXM + EY
and the intercept in the M equation can be denoted as iM. The NDE is
where X0 is a theoretical baseline score on X or a "zero" score and X1 is a theoretical "improvement" score on X or "1" score.
When X is a dichotomy, it is fairly obvious what values to use for X0 and X1. However, when X is measured at the interval level of
measurement, there is no consensus as to what to use for the two values. Perhaps, one idea is to use one standard deviation below the
mean for X0 and one standard deviation above the mean for X1. Do note that the measure is not in "symmetric" in that if we flip the two
(make X0 be one standard deviation above the mean and X1 be one standard deviation below the mean, the new effects are not the old
effects with the opposite sign. Thus, X0needs to be a "theoretically" low value of X and it might even be the case that X0is greater than X1.
Strategy (pdf) O
Financial Wisdom
Links
A powerpoint presentation that summarizes much of this webpage.
Doing a mediation analysis and output a text description of the results using SPSS.
A description of an R mediation program by Tingley, Yamamoto, and Kosuke Imai that is especially useful for non-normal
variables.
Doing a mediation analysis and output a text description of the results using R.
To find out why
computing partial correlations to test mediation is problematic.
Power analyses for a mediation analysis using R.
Go to my moderation page.
A paper I have written called "Reflections on Mediation."
Dave MacKinnon’s mediation website.
Kris Preacher's papers and programs.
Andrew Hayes' papers and programs.
Mediation Facebook page.
View my webinars on mediation. (small charge is requested)
Please suggest new links!
References
Baron, R.
M., & Kenny, D. A. (1986). The moderator-mediator variable distinction
in social psychological research: Conceptual, strategic
and statistical
considerations. Journal of Personality and Social Psychology, 51,
1173-1182.
Bauer, D.
J., Preacher, K. J., & Gil, K. M. (2006). Conceptualizing and testing
random indirect effects and moderated mediation in multilevel
models: New
procedures and recommendations. Psychological Methods, 11, 142-163.
Bolger, N., & Laurenceau, J.-P. (2013). Intensive longitudinal methods: An introduction to diary and experience sampling research. New:
York, Guilford Press.
Brewer, M., Campbell, D. T., & Crano, W. (1970) Testing a single-factor model as an alternative to the misuse of partial correlations in
hypothesis-testing. Sociometry, 33, 1-11.
Cole,
D. A., & Maxwell, S. E. (2003). Testing mediational models with
longitudinal data: Questions and tips in the use of structural equation
modeling. Journal of Abnormal Psychology,
112, 558-577.
Edwards,
J. R., & Lambert L. S. (2007). Methods for integrating moderation and
mediation: A general analytical framework using moderated
path analysis. Psychological
Methods, 12, 1-22.
Frazier,
P. A., Tix, A. P., & Barron, K. E. (2004). Testing moderator and mediator
effects in counseling psychology research. Journal of
Counseling Psychology,
51, 115-134.
Fritz, M. S., Kenny, D. A., & MacKinnon, D. P. (2017). The opposing effects of simultaneously ignoring measurement error and omitting
confounders in a single-mediator model. Multivariate Behavioral Research, in press.
Fritz, M. S., Taylor, A. B., & MacKinnon, D. P. (2012). Explanation of two anomalous results in statistical mediation analysis. Multivariate
Behavioral Research, 47, 61-87.
Hayes, A. F. (2013). Introduction to mediation, moderation, and conditional process analysis: A regression-based approach. New York:
Guilford Press.
Hayes, A. F., & Scharkow, M. (2013). The relative trustworthiness of inferential tests of the indirect effect in statistical mediation analysis:
Does method really matter?
Psychological Science,
24, 1918-1927.
Hoyle, R.
H., & Kenny, D. A. (1999). Statistical power and tests of
mediation. In R. H. Hoyle (Ed.), Statistical strategies for small
sample
research.
Newbury
Park: Sage.
Hyman, H.
H. (1955). Survey design and analysis.
New
York:
Glencoe,
IL: The Free Press.
Imai,
K., Keele, L., & Tingley, D. (2010). A general approach to causal mediation
analysis. Psychological Methods, 15,
309-334.
James, L.
R., & Brett, J. M. (1984). Mediators, moderators and tests for
mediation. Journal of Applied Psychology, 69, 307-321.
Jose, P. E. (2013). Doing statistical mediation and moderation. New York: Guilford Press.
Judd, C.
M., & Kenny, D. A. (1981). Process analysis: Estimating mediation in
treatment evaluations. Evaluation Review, 5, 602-619.
Judd, C. M., Kenny, D. A., & McClelland, G. H. (2001). Estimating and testing mediation and moderation in within-subject designs.
Psychological Methods, 6, 115-134.
Kenny, D. A., & Judd, C. M. (2014). Power anomalies in testing mediation. Psychological Science,
25, 334-339.
Kenny, D.
A., Korchmaros, J. D., & Bolger, N. (2003). Lower level
mediation in multilevel models. Psychological Methods, 8, 115-128.
Kraemer H.
C., Wilson G. T., Fairburn C. G., & Agras W. S. (2002).
Mediators and moderators of treatment effects in randomized clinical
trials. Archives
of General Psychiatry, 59, 877-883.
Krull, J.
L. & MacKinnon, D. P. (1999). Multilevel mediation modeling in
group-based intervention studies. Evaluation Review, 23, 418-444.
Ledermann, T.,
Macho, S., & Kenny, D. A. (2011). Assessing mediation in dyadic data using
the Actor-Partner Interdependence Model.
Structural
Equation Modeling, 18, 595-612.
MacKinnon,
D. P., Fairchild, A. J., & Fritz, M. S. (2007). Mediation analysis. Annual
Review of Psychology, 58, 593-614.
MacKinnon, D. P., Lockwood, C. M., & Williams, J. (2004). Confidence limits for the indirect effect: Distribution of the product and
resampling methods. Multivariate Behavioral Research, 39, 99-128.
MacKinnon,
D. P., Warsi, G., & Dwyer, J. H. (1995). A simulation study of
mediated effect measures. Multivariate Behavioral Research, 30,
41-62.
Muller,
D., Judd, C. M., & Yzerbyt, V. Y. (2005). When moderation is mediated and
mediation is moderated. Journal of Personality and Social
Psychology, 89, 852-863.
Muthén, B. (2011). Applications of causally defined direct and indirect effects in mediation analysis using SEM in Mplus. Download at
www.statmodel.com/download/causalmediation.pdf.
Pearl J. (2001). Direct and indirect effects." In Proceedings of the Seventeenth Conference on Uncertainty in Artificial Intelligence, pp. 411-
420. Morgan Kaufmann, San Francisco, CA.
Pearl, J.
(2011). The causal mediation
formula -- A guide to the assessment of pathways and mechanisms. Prevention Science, 13, 426-
436.
Pearl, J.
(2014). Interpretation and identification of causal mediation. Psychological Methods, 19, 459-481.
Preacher,
K. J., & Kelley, K. (2011. Effect size measures for mediation
models: Quantitative strategies for
communicating indirect effects.
Psychological Methods, 16, 93-115.
Robins J. M., & Greenland S. (1992). Identifiability and exchangeability for direct and indirect effects. Epidemiology, 3, 143-155.
Shrout, P.
E., & Bolger, N. (2002). Mediation in experimental and
nonexperimental studies: New procedures and recommendations.
Psychological
Methods, 7, 422-445.
Smith, E.
(1982). Beliefs, attributions, and evaluations: Nonhierarchical models of
mediation in social cognition. Journal of Personality
and Social Psychology,
43, 248-259.
Sobel, M.
E. (1982). Asymptotic confidence intervals for indirect effects in
structural equation models. In S. Leinhardt (Ed.), Sociological
Methodology
1982 (pp. 290-312).
Washington
DC: American Sociological
Association.
Valeri, L., & VanderWeele, T.J. (2013). Mediation analysis allowing for exposure-mediator interactions and causal interpretation: theoretical
assumptions and implementation with SAS and SPSS macros. Psychological Methods, 18, 137-150.
VanderWeele, T.J. (2015). Explanation in causal inference: Methods for mediation and interaction. New York: Oxford University Press.
Wright, S. (1934). The method of path coefficients. Annals of Mathematical Statistics, 5, 161-215.