Top Qs
Timeline
Chat
Perspective
Seemingly unrelated regressions
Concept in statistical mathematics From Wikipedia, the free encyclopedia
Remove ads
In econometrics, the seemingly unrelated regressions (SUR)[1]: 306 [2]: 279 [3]: 332 or seemingly unrelated regression equations (SURE)[4][5]: 2 model, proposed by Arnold Zellner in (1962), is a generalization of a linear regression model that consists of several regression equations, each having its own dependent variable and potentially different sets of exogenous explanatory variables. Each equation is a valid linear regression on its own and can be estimated separately, which is why the system is called seemingly unrelated,[3]: 332 although some authors suggest that the term seemingly related would be more appropriate,[1]: 306 since the error terms are assumed to be correlated across the equations.
The model can be estimated equation-by-equation using standard ordinary least squares (OLS). Such estimates are consistent, however generally not as efficient as the SUR method, which amounts to feasible generalized least squares with a specific form of the variance-covariance matrix. Two important cases when SUR is in fact equivalent to OLS are when the error terms are in fact uncorrelated between the equations (so that they are truly unrelated) and when each equation contains exactly the same set of regressors on the right-hand-side.
The SUR model can be viewed as either the simplification of the general linear model where certain coefficients in matrix are restricted to be equal to zero, or as the generalization of the general linear model where the regressors on the right-hand-side are allowed to be different in each equation. The SUR model can be further generalized into the simultaneous equations model, where the right-hand side regressors are allowed to be the endogenous variables as well.
Remove ads
The model
Summarize
Perspective
Suppose there are m regression equations
Here i represents the equation number, r = 1, …, R is the individual observation, and we are taking the transpose of the column vector. The number of observations R is assumed to be large, so that in the analysis we take R → , whereas the number of equations m remains fixed.
Each equation i has a single response variable yir, and a ki-dimensional vector of regressors xir. If we stack observations corresponding to the i-th equation into R-dimensional vectors and matrices, then the model can be written in vector form as
where yi and εi are R×1 vectors, Xi is a R×ki matrix, and βi is a ki×1 vector.
Finally, if we stack these m vector equations on top of each other, the system will take the form [4]: eq. (2.2)
1 |
The assumption of the model is that error terms εir are independent across observations, but may have cross-equation correlations within observations. Thus, we assume that E[ εir εis | X ] = 0 whenever r ≠ s, whereas E[ εir εjr | X ] = σij. Denoting Σ = [σij] the m×m skedasticity matrix of each observation, the covariance matrix of the stacked error terms ε will be equal to [4]: eq. (2.4) [3]: 332
where IR is the R-dimensional identity matrix and ⊗ denotes the matrix Kronecker product.
Remove ads
Estimation
Summarize
Perspective
The SUR model is usually estimated using the feasible generalized least squares (FGLS) method. This is a two-step method where in the first step we run ordinary least squares regression for (1). The residuals from this regression are used to estimate the elements of matrix :[6]: 198
In the second step we run generalized least squares regression for (1) using the variance matrix :
This estimator is unbiased in small samples assuming the error terms εir have symmetric distribution; in large samples it is consistent and asymptotically normal with limiting distribution[6]: 198
Other estimation techniques besides FGLS were suggested for SUR model:[7] the maximum likelihood (ML) method under the assumption that the errors are normally distributed; the iterative generalized least squares (IGLS), where the residuals from the second step of FGLS are used to recalculate the matrix , then estimate again using GLS, and so on, until convergence is achieved; the iterative ordinary least squares (IOLS) scheme, where estimation is performed on equation-by-equation basis, but every equation includes as additional regressors the residuals from the previously estimated equations in order to account for the cross-equation correlations, the estimation is run iteratively until convergence is achieved. Kmenta and Gilbert (1968) ran a Monte-Carlo study and established that all three methods—IGLS, IOLS and ML—yield numerically equivalent results, they also found that the asymptotic distribution of these estimators is the same as the distribution of the FGLS estimator, whereas in small samples neither of the estimators was more superior than the others.[8] Zellner and Ando (2010) developed a direct Monte Carlo method for the Bayesian analysis of SUR model.[9]
Remove ads
Equivalence to OLS
There are two important cases when the SUR estimates turn out to be equivalent to the equation-by-equation OLS. These cases are:
- When the matrix Σ is known to be diagonal, that is, there are no cross-equation correlations between the error terms. In this case the system becomes not seemingly but truly unrelated.
- When each equation contains exactly the same set of regressors, that is X1 = X2 = … = Xm. That the estimates turn out to be numerically identical to OLS estimates follows from Kruskal's tree theorem,[1]: 313 or can be shown via the direct calculation.[6]: 197
Statistical packages
- In R, SUR can be estimated using the package “systemfit”.[10][11][12][13]
- In SAS, SUR can be estimated using the
syslin
procedure.[14] - In Stata, SUR can be estimated using the
sureg
andsuest
commands.[15][16][17] - In Limdep, SUR can be estimated using the
sure
command [18] - In Python, SUR can be estimated using the command
SUR
in the “linearmodels” package.[19] - In gretl, SUR can be estimated using the
system
command.
Remove ads
See also
References
Further reading
Wikiwand - on
Seamless Wikipedia browsing. On steroids.
Remove ads