Instrumental Variables

For estimation of a classical linear regression model Y = Xβ + ε, recall the important Assumption 3:

E(ε|X) = 0

That is, E(εi|X) = 0, i=1,2,...,N. This implies:

There are many occasions such as omitted variables or errors in the explanatory variables, which result a clear violation of Assumption 3 for the classical linear regression model. The consequence is that the least squares estimator is biased, inconsistent and inefficient.

Instrumental Variables

A replacement for X, called Z, may be used to restore the consistency of least squares estimation. Z must satisfy the following conditions:

Z is called the instrumental variables for X. Z may include all or part of X variables as long as they are exogenous.

Instrumental Variable Estimation

The estimation consists of two steps as follows:

  1. Formulate a multivariate regression as X = + u, and estimate the parameter vector of δ as:
    d = (Z'Z)-1Z'X, and
    Xp = Zd = Z(Z'Z)-1Z'X.

  2. Using OLS to estimate the model Y = Xpβ + ε:
    b = (Xp'Xp)-1Xp'Y
    = [X'Z(Z'Z)-1Z'X]-1 X'Z(Z'Z)-1Z'Y

It is clear that the selection of instrumental variables is crucial for a successful estimation of the model parameters. In practice, in addition to the replacement for endogenous explanatory variables, the instrumental variables include the exogenous explanatory variables already in the model. Therefore, the instrumental variable estimation (IV) is summarized as:

Define W = Z(Z'Z)-1Z'X, and note that W'X = W'W.
b = (W'X)-1W'Y = [X'Z(Z'Z)-1Z'X]-1 X'Z(Z'Z)-1Z'Y
Var(b) = s2(W'X)-1 = s2[X'Z(Z'Z)-1Z'X]-1
where s2 = e'e/(N-K) and e = Y - Xb.

Two Special Cases

  1. If Z and X has the same number of columns, then
    b = (Z'X)-1Z'Y
    Var(b) = s2(Z'X)-1

  2. If Z = X, then it is OLS.

Robust Inference of IV Estimator

If the classical assumptions of homoscedasticity and no autocorrelation are not satisfied, in conjunction with the problem of endogenous independent variables (i.e., lagged dependent variable), the linear regression model is presented as follows:

Y = Xβ +ε, ε|Z ~ iid(0,Ω), Ω ≠ σ2I
b = (Xp'X)-1Xp'Y = [X'Z(Z'Z)-1Z'X]-1X'Z(Z'Z)-1Z'Y
Var(b) = (Xp'X)-1(XpXp)(Xp'X)-1
where XpXp = X'Z(Z'Z)-1(Z'ΩZ)(Z'Z)-1Z'X

The estimation of Var(b) depends on the estimation of the consistent estimator of Z'ΩZ = Z'E(εε')Z. A robust estimate of the variance-covariance matrix can be based on Newey-West estimator allowing general heteroscedasticity and autocorrelation. That is,

Σ = Z'ΩZ = S0 + ∑j=1,...,J[1-j/(J+1)](Sj+Sj')
where
S0 = (1/N)∑i=1,...,Nei2zizi'
Sj = (1/N)∑i=j+1,...,Neiei-jzizi-j'
Note: ei = Yi- Xib, i=1,...,N.

Generalized Method of Moments

Consider the model: Y = Xβ +ε, ε|Z ~ iid(0,Ω), a more general estimation approach is the Generalized Method of Moments (GMM). The key assumption is the exogeneity assumption of the intruments:

E(ε|Z) = 0

This implies the moment functions E(Ziεi) = E(Zi(Yi-Xiβ)) = 0. GMM estimator of β is obtained to minimize the objective function which is the weighted quadratic form of the moment functions:

Q(β) = (Z'ε/N)'W(Z'ε/N)

where W is the weighted matrix which is symmetric positive definite. We have,

b = [X'ZWZ'X]-1X'ZWZ'Y
Var(b) = [X'ZWZ'X]-1[X'Z(WΣW')Z'X][X'ZWZ'X]-1
Σ = E(Z'εε'Z) = Z'ΩZ = = S0 + ∑j=1,...,J[1-j/(J+1)](Sj+Sj')
as defined above, allowing for general heteroscedasticity and autocorrelation up to the J-th order.

If W = Σ-1 (optimal weighted matrix), then it is the optimal or efficient GMM estimator:

b = [X'ZΣ-1Z'X]-1X'ZΣ-1Z'Y
Var(b) = [X'ZΣ-1Z'X]-1

If W = (Z'Z)-1, then it is IV estimator.

If W = I, then it is Minimum Distance (MD) estimator.

Instrumental Variable Specification Tests


Copyright © Kuan-Pin Lin
Last updated: January 20, 2010