Alternative algorithms for the estimation of dynamic factor, mimic and varying coefficient regression models (Q790575): Difference between revisions
From MaRDI portal
Set profile property. |
Set OpenAlex properties. |
||
Property / full work available at URL | |||
Property / full work available at URL: https://doi.org/10.1016/0304-4076(83)90066-0 / rank | |||
Normal rank | |||
Property / OpenAlex ID | |||
Property / OpenAlex ID: W2072041292 / rank | |||
Normal rank |
Revision as of 18:21, 19 March 2024
scientific article
Language | Label | Description | Also known as |
---|---|---|---|
English | Alternative algorithms for the estimation of dynamic factor, mimic and varying coefficient regression models |
scientific article |
Statements
Alternative algorithms for the estimation of dynamic factor, mimic and varying coefficient regression models (English)
0 references
1983
0 references
A state-space model is considered consisting of two sets of equations: transition equations: \(x_ t=\phi_ tx_{t-1}+\gamma_ tz_ t+G_ tv_ t\); \(\phi_ t:\) (\(j\times j)\); \(\gamma_ t:\) (\(j\times k)\); \(G_ t:\) (jxm); \(x_ t:\) unobservable (\(j\times 1)\) state vector; \(z_ t:\) exogenous or lagged dependent (\(k\times 1)\) vector; \(v_ t:\) (\(m\times 1)\) disturbance. The measurement equations are: \(y_ t=\alpha_ tx_ t+\beta_ tz_ t+e_ t\), where \(y_ t\) is measured; \(e_ t:\) disturbance; \(\left( \begin{matrix} v_ t\\ e_ t\end{matrix} \right)\sim NI(0\), \(\left( \begin{matrix} Q_ t\\ 0\end{matrix} \begin{matrix} 0\\ R_ t\end{matrix} \right))\); and the parameters \(\phi\),\(\gamma\),G,\(\alpha\),\(\beta\),Q,R are assumed here time invariant. The log-likelihood function of the unknown parameters is: \(L(\theta)=-\sum^{T}_{t=1}(\log | H_ t| +\eta^ T_ tH_ t^{-1}\eta_ t)\) where the ''innovations'' \(\eta_ t=y_ t-E(y_ t| y_{t-1},...,y_ 1,z_ t,...,z_ 1)\), whereas the matrix \(H_ t\) denotes the variance of \(\eta_ t.\) A scoring method of estimation is given using first derivatives of L (a zig-zag approach is suggested: using k th iteration values \(\alpha^ k,\phi^ k,R^ k,Q^ k\) to determine \(\beta^ k\) and \(\gamma^ k\), these values being then used to find \(\alpha^{k+1},\phi^{k+1},R^{k+1},Q^{k+1}\), and so on until convergence). The EM consists of two steps: Estimation and maximization are iterated until convergence: more precisely, in the presence of unobservables, the estimation step constructs estimates of the sufficient statistics of the model conditional on the observables, unobservables being estimated based on the parameter values at the current step of the iteration whereas at the maximization step, the likelihood function is maximized based both on the observables and these estimated values. This EM method is presented under special linear restrictions on \(\alpha\),\(\beta\),\(\phi\),\(\gamma\). The max. step uses SUR calculations (regression), so neither first nor second derivatives need be computed. In the neighbourhood of the EM solution, the scoring method may be useful to pinpoint the maximum and to estimate the information matrix. An economic application is given (estimation of a common factor in wage rate data from several industries in Los Angeles).
0 references
varying coefficient regression models
0 references
estimation of dynamic unobserved component models
0 references
method of scoring
0 references
EM algorithm
0 references
derivative-free method
0 references
Kalman filter
0 references