Yit = Xitbit + eit
Let bit = b and assume eit = ui + vt + eit where ui represents the individual or cross section differnence in intercept and vt is the time difference in intercept. Two-ways analysis includes both time and individual effects. For simplicity, we further assume vt = 0. That is, only the one-way individual effects will be analyzed:
eit = ui + eit
The component eit is a classical error term, with zero mean, homogeneous variance, and there is no serial correlation and no contemporary correlation. Also, eit is uncorrelated with the regressors Xit. That is,
Assume that the error component ui, the individual differnence, is fixed or nonstochastic (but it varies across individuals). Thus, the model error is simply eit = eit. The model is expressed as:
Yit = (Xitb + ui) + eit
where ui is interpreted as the change in the intercept. Therefore the individual effect is defined as ui plus the intercept. If Xit does not include the constant term, ui itself is the individual effect.
Random Effects Model
Assume that the error component ui, the individual differnence, is random and satisfies the following assumptions:
S = |
|
|
| = se2ITxT + su21TxT |
where i is a vector of T ones, 1 = ii' is a TxT matrix of ones, and I is a TxT identity matrix. Let e be a NT-element vector of the stacked errors e1, e2, ..., eN, e = [e1,e2, ..., eN]', then E(e) = 0 and E(ee') = INxNÄSTxT, where S is the TxT variance-covariance matrix defined above.
In summary with matrix notation, the random effects model is defined by
YNTx1 = XNTxKbKx1 + uNx1ÄiTx1 + eNTx1
The assumptions are:
|
|
| = |
|
|
| b + |
|
|
|
or, Y = Xb + e
Yit = (Xitb + ui) + eit (i=1,2,...,N; t=1,2,...,T).
|
|
| = |
|
|
| b + |
|
|
| + |
|
|
|
or, Y = Xb + e = Xb + uÄi + e
For each i, define NT´1 vector Di with the element:
Dij = | 1 if (i-1)´T+1 £ j £ i´T |
0 otherwise |
Then D = [D1, D2, ..., DN-1] is NT´(N-1) matrix of N-1 dummy variables. Ordinary least squares can be used to estimate the model with dummy variables as follows:
Y = Xb + uÄi + e = Xb + Dd + e
Since X includes a constant term, one less dummy variables are included for estimation and the estimated d measures the individual change from the intercept.
Let Ymi = (St=1,2,...,TYit)/T, Xmi = (St=1,2,...,TXit)/T, and emi = (St=1,2,...,Teit)/T. Then the within estimates of the model can be obtained by estimating the mean deviation model:
(Yit - Ymi) = (Xit - Xmi)b + (eit - emi)
Or, equivalently
Yit = Xitb + (Ymi - Xmib) + (eit - emi)
Note that the constant term drops out due to mean deviation transformation. The degree of freedom for estimating the above mean deviation model is NT-K-1 (K is the number of explanatory variables including constant term). Therefore, the estimated individual effects of the model is ui = Ymi - Xmib. The variance-covariance matrix of individual effects is estimated as follows:
Var(ui) = v/T + Xmi [Var(b)] Xmi'
where v is the estimated variance of the mean deviation regression corrected for the degree of freedom NT-N-K (instead of NT-K-1). That is,
v = Si=1,2,...,NSt=1,2,...,T (eit - emi)2 / (NT-N-K).
It may be of interest to estimate the between parameters of the model by estimating
Ymi = Xmib + ui + emi
which is related to the estimated individual effects from the within estimates.
Based on the dummy variable approach, this is a Wald F-test for the joint significance of the parameters associated with dummy variables representing the individual effects. If the null hypothesis d = 0 can not be rejected, then there is no fixed effects in the model.
Based on the deviation approach, the equivalent test statistic is computed from the restricted (pooled model) and unrestricted (mean deviation model) sum of squared residuals RSSR and RSSUR, repectively. That is,
| ~ F(N-1, NT-N-K) |
Y = Xb + e
where e = [e1,e2,...,eN]', ei = [ei1,ei2,...,eiT]', and the random error components eit = ui + eit. By assumptions, E(e) = 0, and E(ee') = IÄS. The Generalized Least Squares estimates of b is
b | = [X'(IÄS-1)X]-1 X'(IÄS-1)Y |
= [åi=1,...,NXi'S-1Xi]-1 [åi=1,...,NXi'S-1Yi] |
where
S = |
|
|
| = se2I + su21 |
and
S-1 = (se2I + su21)-1 = (1/se2)[I+(su2/se2)1]-1 = (1/se2){I-[su2/(Tsu2+se2)]1}
Since S-1 can be derived from the estimated variance components se2 and su2, in practice the model is estimated using the following partial deviation approach.
Let v = se2.
Ymi = Xmib + (ui + emi)
where the error structure of ui + emi satisfies:
E(ui + emi) = 0
E((ui + emi)2) = su2 + se2/T
E((ui + emi)(uj + emj)) = 0, for i¹j
Let v1 = T su2 + se2 = T su2 + v.
If v1 > v, then define w = 1 - (v/v1)½.
In case of v1 £ v, the estimate of su2 becomes negative. The alternative is to use (v0-v) for su2, where v0 is the estimated variance s2 obtained from the pooled model:
Yit = Xitb + eit
v0 is a consistent estimator of su2 + se2, where the estimator of se2 is v (obtained from the estimated fixed effect model, see Step 1). Then the consistent estimator of su2 is (v0-v). If v0 £ v, we need to use large sample variances to construct the estimator of su2:
v0 = [(NT-K-1)/NT] s2
v = [(NT-N-K)/NT] se2
Let v1 = T (v0-v) + v, and define w = 1 - (v/v1)½.
Y*it = Yit - w Ymi
X*it = Xit - w Xmi
Then the model for estimation is:
Y*it = X*itb + e*it
where e*it = (1-w) ui + eit - w emi.
Or, equivalently
Yit = Xitb + w (Ymi - Xmib) + e*it
It is easy to validate that
E(e*it) = 0
E(e*2it) = se2
E(e*ite*it) = 0
for t¹t
E(e*ite*jt) = 0
for i¹j
The least squares estimate of [w (Ymi - Xmib)] is interpreted as the change of individual effects.
Assume the panel data model Y = Xe + e ~ Normal(0,IÄS)
or, ei ~ Normal (0,S), i = 1,2,...,N
where S = se2I + su21
Write the probability density of ei as
f(ei) = (1/(2p)½)T |S|-½ exp(-½ei'S-1ei)
The corresponding log-likelihood function is
lli(b,se2,su2) = -T/2 ln(2p) -½ ln|S| -½ (Yi-Xib)'S-1(Yi-Xib)
Because S-1 = (1/se2){I-[su2/(Tsu2+se2)]1} (see above) and
|S| = |se2I+su21| = (se2)T|I+(su2/se2)1| = (se2)T(Tsu2/se2+1)
(Note: |I+(su2/se2)1| = product of eigenvalues: (Tsu2/se2)+1,1,1,...,1)
Therefore
lli(b,se2,su2) =
-T/2 ln(2pse2) -½ ln(Tsu2/se2+1)
-½ (1/se2)
{[åt=1,...,T(Yit-Xitb)2] -
[su2/(Tsu2+se2)]
[åt=1,...,T(Yit-Xitb)]2}
The maximum likelihood estimator of the parameters vector (b,se2,su2) is obtained by maximizing the total log-likelihood defined by
ll(b,se2,su2) = åi=1,...,N lli(b,se2,su2)
To test for no correlation relationship of the error terms eit = ui + eit and eit = ui + eit, that is su2 = 0, the following Breusch-Pagan LM test statistic based on the estimated residuals of the restricted (pooled) model, eit (i=1,2,...N, t=1,2,...,T), is distributed as Chi-square with one degree of freedom:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
Note that emi = St=1,2,...,Teit/T.
H = (brandom-bfixed)'[Var(brandom)-Var(bfixed)]-1(brandom-bfixed)
The rejection of the null hypothesis implies the fixed effects model. However, not rejecting E(X'u) = 0 could suggest either a random effects or fixed effects model.
Firm | Within Estimates Fixed Effects | Within Estimates Random Effects | ML Estimates Random Effects |
---|---|---|---|
1 | 9.7059 (0.19323) | 9.6378 (0.18313) | 9.6319 |
2 | 9.6647 (0.19908) | 9.5979 (0.18716) | 9.5860 |
3 | 9.4970 (0.22505) | 9.4408 (0.20686) | 9.4055 |
4 | 9.8905 (0.24185) | 9.7780 (0.21918) | 9.7892 |
5 | 9.7300 (0.26102) | 9.6299 (0.23371) | 9.6194 |
6 | 9.7930 (0.26374) | 9.6831 (0.23544) | 9.6798 |
Yi = Xibi + ei
bi = b + ui
where Yi = [Yi1,Yi2,...,YiT]', Xi = [Xi1,Xi2,...,XiT]', and ei = [ei1,ei2,...,eiT]'. We note that not only the intercept but also the slope parameters are random across individuals. The assumptions of the model are:
The model for estimation is
Yi = Xib +
(Xiui + ei), or
Yi = Xib + wi
where wi = Xiui +
ei, and
The stacked (pooled) model is
Y = Xb + w
where w = [w1,...,wN]', and
E(w) = 0NTx1
Var(w) = E(ww') = V = |
|
|
|
GLS is used to estimate the model. That is,
b* = (X'V-1X)-1X'V-1Y
Var(b*) = (X'V-1X)-1
The computation is based on the following steps (Swamy, 1971):
The individual parameter vectors may be predicted as follows:
bi* = (G+Vi)-1[G-1b*+Vi-1bi]
= Aib* + (I-Ai)bi,
where Ai = (G+Vi)-1G-1.
Var(bi*) = [Ai | I-Ai] |
|
|
|
|
|
|
Yit = Xitbi + eit (i=1,2,...,N; t=1,2,...,T).
Let Yi = [Yi1,Yi2,...,YiT]', Xi = [Xi1,Xi2,...,XiT]', and ei = [ei1,ei2,...,eiT]', the stacked N equations (T observations each) system is Y = Xb + e, or
|
|
| = |
|
|
|
|
|
| + |
|
|
|
Notice that not only the intercept but also the slope terms of the estimated parameters are different across individuals. The error structure of the model is summarized as follows:
Parameter restrictions can be built into the matrix X and the corresponding parameter vector b. The model is estimated using techniques for systems of regression equations.
The system estimation techniques such as 3SLS and FIML should be used for parameter estimation. It is called the Seemingly Unrelated Regression Estimation (SURE) in the current context. Denote b and S as the estimated b and S, respectively. Then,
b = [X'(S-1ÄI)X]-1X'(S-1ÄI)Y
Var(b) = [X'(S-1ÄI)X]-1, and
S = ee'/T, where e = Y-Xb is the estimated error e.