Economic Models - Convex Optimization
Economic Models - Convex Optimization Economic Models - Convex Optimization
Time Varying Responses 47 Combining Eqs. (14) and (14a) we have BX ′ =−ƔU ′ + R ′ ⇒ X ′ =−B −1 ƔU ′ + B −1 R ′ ⇒ X ′ = U ′ + W ′ (16) where W ′ = B −1 R ′ Denoting the ith column of matrix X ′ by x i and the ith column of matrix W ′ by w i , we can write ⎡ ⎤ u 1i 0 ··· 0 u 2i 0 ··· 0 u gi 0 ··· 0 0 u 1i ··· 0 0 u 2i ··· 0 0 u gi ··· 0 x i = · · · · · · · · · ⎢ · · · · · · · · · ⎥ ⎣ · · · · · · · · · ⎦ 0 0 ··· u 1i 0 0 ··· u 2i ··· 0 ... u gi π + w i (17) where u ij is the element of the jth column and ith row of matrix U. The vector π ∈ E ng , as mentioned earlier, consists of the g column of matrix . Equation (17) can be written in a compact form, as x i = H i π + w i , i = 1, 2,...,N (17a) where x i ∈ E n ,w i ∈ E n and the observation matrix H i is defined on E n × E ng . In a time-invariant econometric model, the coefficients vector π is assumed random with constant expectation overtime, so that π i+1 = π i , for all i. (18) In a time-varying and stochastic model we can have π i+1 = π i + ε i (18a) where ε i ∈ E ng is the noise. Based on these, we can re-write Eq. (17a) as x i+1 = H i+1 π i+1 + w i+1 i = 0, 1,...,N − 1. (19) We make the following assumptions. (a) The vector x i+1 and matrix H i+1 can he measured exactly for all i. (b) The noises ε i and w i+1 are independent discrete white noises with known statistics, i.e., E(ε i ) = 0; E(w i+1 ) = 0 E(ε i w ′ i+1 ) = 0
48 Dipak R. Basu and Alexis Lazaridis E(c i ε ′ i ) = Q 1δ ij , E(w i w ′ i ) = Q 2δ ij . where δ ij is the Kronecker delta, and The above co-variance matrices, assumed to be positive definite. (c) The state vector is normally distributed with a finite co-variance matrix. (d) Regarding Eqs. (18a) and (19), the Jacobians of the transformation of ε i into π i+1 and of w i+1 into x i+1 are unities. Hence, the corresponding conditional probability densities are: p(π i+1 |π i ) = p(ε i ) p(x i+1 |π i+1 ) = p(w i+1 ). Under the above assumptions and given Eqs. (18a) and (19), the problem set is to evaluate and E(π i+1 |x i+1 ) = π ∗ i+1 cov (π i+1 |x i+1 ) = S i+1 (the error co-variance matrix) where x i+1 = x 1 ,x 2 ,x 3 ,...,x i+1 . The solution to this problem (Basu and Lazaridis, 1986; Lazaridis, 1980) is given by the following set of recursive equations, as it is briefly shown in Appendix A. π ∗ i+1 = π∗ i + K i+1(x i+1 − H i+1 π ∗ i ) (20) K i+1 = S i+1 H i+1 ′ Q−1 2 (21) Si+1 −1 i+1 + H i+1 ′ Q−1 2 H i+1 (22) Pi+1 −1 1 + S i ) −1 . (23) The recursive process is initiated by regarding K 0 and H 0 as null matrices and computing π ∗ 0 and S 0 from π0 ∗ =ˆπ i.e., the reduced form coefficients (columns of matrix ˆ) S 0 = P 0 = ˜. The reduced form coefficients, along with their co-variance matrices, can be updated by this recursive process and at each stage the set of Riccati
- Page 21 and 22: xx Introduction type of model is ve
- Page 23 and 24: This page intentionally left blank
- Page 25 and 26: This page intentionally left blank
- Page 27 and 28: 4 Olav Bjerkholt Among these was al
- Page 29 and 30: 6 Olav Bjerkholt method, searching
- Page 31 and 32: 8 Olav Bjerkholt context, but one o
- Page 33 and 34: 10 Olav Bjerkholt be modified in th
- Page 35 and 36: 12 Olav Bjerkholt through a number
- Page 37 and 38: 14 Olav Bjerkholt memoranda. These
- Page 39 and 40: 16 Olav Bjerkholt hopelessly ineffe
- Page 41 and 42: 18 Olav Bjerkholt References Arrow,
- Page 43 and 44: This page intentionally left blank
- Page 45 and 46: This page intentionally left blank
- Page 47 and 48: 24 Alexis Lazaridis A straightforwa
- Page 49 and 50: 26 Alexis Lazaridis where ⎛ ⎞ j
- Page 51 and 52: 28 Alexis Lazaridis Estimating the
- Page 53 and 54: 30 Alexis Lazaridis 5.1. Testing fo
- Page 55 and 56: 32 Alexis Lazaridis by applying the
- Page 57 and 58: 34 Alexis Lazaridis Nevertheless, t
- Page 59 and 60: 36 Alexis Lazaridis Table 2. Residu
- Page 61 and 62: 38 Alexis Lazaridis Table 3. Critic
- Page 63 and 64: 40 Alexis Lazaridis Figure 4. The s
- Page 65 and 66: 42 Alexis Lazaridis Perron, P and T
- Page 67 and 68: 44 Dipak R. Basu and Alexis Lazarid
- Page 69: 46 Dipak R. Basu and Alexis Lazarid
- Page 73 and 74: 50 Dipak R. Basu and Alexis Lazarid
- Page 75 and 76: 52 Dipak R. Basu and Alexis Lazarid
- Page 77 and 78: 54 Dipak R. Basu and Alexis Lazarid
- Page 79 and 80: 56 Dipak R. Basu and Alexis Lazarid
- Page 81 and 82: 58 Dipak R. Basu and Alexis Lazarid
- Page 83 and 84: 60 Dipak R. Basu and Alexis Lazarid
- Page 85 and 86: 62 Dipak R. Basu and Alexis Lazarid
- Page 87 and 88: 64 Dipak R. Basu and Alexis Lazarid
- Page 89 and 90: 66 Dipak R. Basu and Alexis Lazarid
- Page 91 and 92: This page intentionally left blank
- Page 93 and 94: 70 Andrew Hughes Hallet have, there
- Page 95 and 96: 72 Andrew Hughes Hallet other. The
- Page 97 and 98: 74 Andrew Hughes Hallet success in
- Page 99 and 100: 76 Andrew Hughes Hallet with a hori
- Page 101 and 102: 78 Andrew Hughes Hallet control inf
- Page 103 and 104: 80 Andrew Hughes Hallet Table 2. Fi
- Page 105 and 106: 82 Andrew Hughes Hallet country 13
- Page 107 and 108: 84 Andrew Hughes Hallet Given this
- Page 109 and 110: 86 Andrew Hughes Hallet 5.4. Instit
- Page 111 and 112: 88 Andrew Hughes Hallet Substitutin
- Page 113 and 114: 90 Andrew Hughes Hallet This is alw
- Page 115 and 116: 92 Andrew Hughes Hallet (d) Since E
- Page 117 and 118: 94 Andrew Hughes Hallet Table 3. Ex
- Page 119 and 120: 96 Andrew Hughes Hallet restraint m
48 Dipak R. Basu and Alexis Lazaridis<br />
E(c i ε ′ i ) = Q 1δ ij ,<br />
E(w i w ′ i ) = Q 2δ ij .<br />
where δ ij is the Kronecker delta, and<br />
The above co-variance matrices, assumed to be positive definite.<br />
(c) The state vector is normally distributed with a finite co-variance matrix.<br />
(d) Regarding Eqs. (18a) and (19), the Jacobians of the transformation of<br />
ε i into π i+1 and of w i+1 into x i+1 are unities. Hence, the corresponding<br />
conditional probability densities are:<br />
p(π i+1 |π i ) = p(ε i )<br />
p(x i+1 |π i+1 ) = p(w i+1 ).<br />
Under the above assumptions and given Eqs. (18a) and (19), the problem<br />
set is to evaluate<br />
and<br />
E(π i+1 |x i+1 ) = π ∗ i+1<br />
cov (π i+1 |x i+1 ) = S i+1<br />
(the error co-variance matrix)<br />
where x i+1 = x 1 ,x 2 ,x 3 ,...,x i+1 .<br />
The solution to this problem (Basu and Lazaridis, 1986; Lazaridis,<br />
1980) is given by the following set of recursive equations, as it is briefly<br />
shown in Appendix A.<br />
π ∗ i+1 = π∗ i + K i+1(x i+1 − H i+1 π ∗ i ) (20)<br />
K i+1 = S i+1 H i+1 ′ Q−1 2<br />
(21)<br />
Si+1 −1 i+1 + H i+1 ′ Q−1 2 H i+1 (22)<br />
Pi+1 −1 1 + S i ) −1 . (23)<br />
The recursive process is initiated by regarding K 0 and H 0 as null matrices<br />
and computing π ∗ 0 and S 0 from<br />
π0 ∗ =ˆπ i.e., the reduced form coefficients (columns of matrix ˆ)<br />
S 0 = P 0 = ˜.<br />
The reduced form coefficients, along with their co-variance matrices,<br />
can be updated by this recursive process and at each stage the set of Riccati