Variables (IV) Is Used To Estimate
Variables (IV) Is Used To Estimate
variables (IV) is used to estimate causal relationships when controlled experiments are not feasible
or when a treatment is not successfully delivered to every unit in a randomized experiment. [1]
Instrumental variable methods allow consistent estimation when the explanatory
variables (covariates) are correlated with the error terms of a regressionrelationship. Such
correlation may occur when the dependent variable causes at least one of the covariates ("reverse"
causation), when there are relevant explanatory variables which are omitted from the model, or
when the covariates are subject to measurement error. In this situation, ordinary linear
regressiongenerally produces biased and inconsistent estimates.[2] However, if an instrument is
available, consistent estimates may still be obtained. An instrument is a variable that does not itself
belong in the explanatory equation and is correlated with the endogenous explanatory variables,
conditional on the other covariates. In linear models, there are two main requirements for using an
IV:
The instrument must be correlated with the endogenous explanatory variables, conditional
on the other covariates.
The instrument cannot be correlated with the error term in the explanatory equation
(conditional on the other covariates), that is, the instrument cannot suffer from the same problem
as the original predicting variable.
Contents
[hide]
1 Definitions
2 Example
3 Applications
5 Estimation
7 Identification
8 Non-parametric analysis
10 Potential problems
13 References
14 Further reading
15 External links
Definitions[edit]
The theory of instrumental variables was first derived by Philip G. Wright, possibly in co-authorship
with his son Sewall Wright, in his 1928 book The Tariff on Animal and Vegetable Oils.[3][4] Traditionally,
[5]
However, this definition suffers from ambiguities in concepts such as "error term" and "independent
variable," and has led to confusion as to the meaning of the equation itself, which was wrongly
labeled "regression."[6]
General definitions of instrumental variables, using counterfactual and graphical formalism, were
given by Pearl (2000; p. 248).[7] The graphical definition requires that Z satisfy the following
conditions:
where
cut off.
The counterfactual definition requires that Z satisfies
where
stands for the value that Y would attain had X been x. and
If there are additional covariates W then the above definitions are modified so that Z qualifies as an
instrument if the given criteria hold conditional on W.
The essence of Pearl's definition is:
1. The equations of interest are "structural," not "regression."
2. The error term U stands for all exogenous factors that affect Y when X is held constant.
Example[edit]
Informally, in attempting to estimate the causal effect of some variable x on another y, an instrument
is a third variable z which affects y only through its effect on x. For example, suppose a researcher
wishes to estimate the causal effect of smoking on general health.[9] Correlation between health and
smoking does not imply that smoking causes poor health because other variables may affect both
health and smoking, or because health may affect smoking in addition to smoking causing health
problems. It is at best difficult and expensive to conduct controlled experiments on smoking status in
the general population. The researcher may proceed to attempt to estimate the causal effect of
smoking on health from observational data by using time series on the tax rate for tobacco products
as an instrument for smoking in a causal analysis. If tobacco taxes and state of health are correlated
then this may be viewed as evidence that smoking causes changes in health.
Because demonstrating that the third variable 'z' is causally related to 'y' exclusively via 'x' is an
experimental impossibility, and because the same limitations that prevent an experiment to
determine if there is a causal relationship between 'x' and 'y' will normally also prevent experiments
determining if there is a causal relationship between 'z' and 'y' (assumed to be mediated through 'x'),
correlational data is the only type of evidence that analysis by instrumental variable can provide, and
causal inference is not justified. The use of an instrumental variable produces additional evidence of
a statistical relationship (in this case between 'z' and 'y'), without providing evidence of what type of
relationship it is, and without providing direct evidence for the type of relationship between 'x' and 'y'.
Applications[edit]
IV methods are commonly used to estimate causal effects in contexts in which controlled
experiments are not available. Credibility of the estimates hinges on the selection of suitable
instruments. Good instruments are often created by policy changes. For example, the cancellation of
a federal student-aid scholarship program may reveal the effects of aid on some students' outcomes.
Other natural and quasi-natural experiments of various types are commonly exploited, for example,
Miguel, Satyanath, and Sergenti (2004) use weather shocks to identify the effect of changes in
economic growth (i.e., declines) on civil conflict. [10] Angrist andKrueger (2001) present a survey of the
history and uses of instrumental variable techniques. [11]
Figure 2:
Figure 3: Proximity does not qualify as an instrumental variable given Library Hours
Figure 4: Proximity qualifies as an instrumental variable but does not qualify as an instrumental variable given
Library Hours
Suppose that we wish to estimate the effect of a university tutoring program on GPA at a university
where the dormitories to which students are assigned is random. The relationship between attending
the tutoring program and GPA may be confounded by a number of factors. Students that attend the
tutoring program may care more about their grades or may be struggling with their work. (This
confounding is depicted in the Figures 1-3 on the right through the bidirected arc between Tutoring
Program and GPA.) Given that students are assigned to dormitories at random, the proximity of the
student's dorm to the tutoring program is a natural candidate for being an instrumental variable.
However, what if the tutoring program is located in the college library? Proximity may also cause
students to spend more time at the library, which in turn improves their GPA (see Figure 1). Using
the causal graph depicted in the Figure 2, we see that Proximity does not qualify as an instrumental
variable because it is d-connected to GPA through the path Proximity
in
Library Hours
GPA
. However, if we control for Library Hours by adding it as a covariate then Proximity becomes
an instrumental variable since Proximity is d-separated from GPA given Library Hours in
. Now,
suppose that we notice that a student's "natural ability" affects his or her number of hours in the
library as well as his or her GPA, as in Figure 3. Using the causal graph, we see that Library Hours is
a collider and conditioning on it opens the path Proximity
Library Hours
GPA. As a result,
Proximity cannot be used an instrumental variable. Finally, suppose that Library Hours does not
actually affect GPA because students who do not study in the library simply study elsewhere, as in
Figure 4. In this case, controlling for Library Hours still opens a spurious path from Proximity to GPA.
However, if we do not control for Library Hours and remove it as a covariate then Proximity can
again be used an instrumental variable.
Estimation[edit]
Suppose the data are generated by a process of the form
where
i indexes observations,
is an independent variable,
The parameter
of
of
other than
, and
are uncorrelated and that they are drawn from distributions with the same variance, that is,
Suppose also that a regression model of nominally the same form is proposed. Given a random
sample of T observations from this process, the ordinary least squares estimator is
where x, y and
certain regularity conditions the second term has an expected value conditional on x of zero
and converges to zero in the limit, so the estimator is unbiased and consistent. When x and
the other unmeasured, causal variables collapsed into the
the OLS estimator is generally biased and inconsistent for . In this case, it is valid to use
the estimates to predict values of y given values of x, but the estimate does not recover the
causal effect of x on y.
An instrumental variable z is one that is correlated with the independent variable but not with
the error term. Using the method of moments, take expectations conditional on z to find
The second term on the right-hand side is zero by assumption. Solve for
When z and
approaches zero in the limit, providing a consistent estimator. Put another way, the
causal effect of x on y can be consistently estimated from these data even
though x is not randomly assigned through experimental methods.
The approach generalizes to a model with multiple explanatory variables.
Suppose X is the T K matrix of explanatory variables resulting
from T observations on Kvariables. Let Z be a T K matrix of instruments. Then it
can be shown that the estimator
where
Note that the second expression collapses to the first when the number of
instruments is equal to the number of covariates in the equation of interest
(just-identified case).
Proof that GMM collapses to IV in the just-identified case
[show]
Which gives:
is