International Business Cycles: An Unobserved Components Approach
Chapter 2. Unobserved Component models Esther Ruiz 2006-2007 PhD Program in Business Administration...
-
Upload
cassie-buckson -
Category
Documents
-
view
216 -
download
3
Transcript of Chapter 2. Unobserved Component models Esther Ruiz 2006-2007 PhD Program in Business Administration...
Chapter 2. Unobserved Component models
Esther Ruiz2006-2007
PhD Program in Business Administration and Quantitative Analysis
Financial Econometrics
2.1 Description and propertiesUnobserved component models assume that the
variables of interest are made of components with a direct interpretation that cannot be directly observed
Applications in finance:“Fads” model of Potterba and Summers (1998).
There are two types of traders: informed (μt) and uninformed (εt). The observed price is yt
ttt
ttty
1
Models for Ex ante interest differentials proposed by Cavaglia (1992): We observe the ex post interest differential which is equal to the ex ante interest differential plus the cross-country differential in inflation
ttt
ttt
yLy
yy
**
*
)( 1
Factor models simplify the computation of the covariance matrix in mean-variance portfolio allocation and are central in two asset pricing theories: CAPM and APT
Nk
rErffE
f
f
f
v
v
v
r
r
r
tttttttt
Nt
t
t
kt
t
t
Nkt
kt
kt
tN
t
t
Nt
t
t
Nt
t
t
11
2
1
2
1
2
1
1
21
11
2
1
2
1
|,|
.........
...
...
...
...
.........
''
Term structure of interest rates model proposed by Rossi (2004): The observed yields are given by the theoretical rates implied by a no arbitrage condition plus a stochastic disturbance
tttttt
nt
t
t
uCrBA
y
y
y
)(
...
)(
)(
2
1
tt
u
t
t
t cu
r
tbba
tatbta
u
r
exp~
~expexp~exp
01
1
Modelling volatility
There are two main types of models to represent the dynamic evolution of volatilities:
i) GARCH models that assume the volatility is a non-linear funcion of past returns
2
122
122
1
111
21
21
2
0
ttttttttt
tttttttt
ttt
ttt
YEYEYyVar
YEYEYyE
y
y
|||
|||
σ is the one-step ahead (conditional) variance and, therefore, can be observed given observations up to time t-1.
As a result, classical inference procedures can be implemented.
Example: Consider the following GARCH(1,1) model for the IBEX35 returns
The returns corresponding to the first two days in the sample are: 0.21and -0.38
21
21
2 889010300190 ttt y ...
951
142889038010300190380210
0380210
142382889021010300190210
0210
2213
213
212
12
.
.*..*..).,.|(
).,.|(
..*..*..).|(
).|(
yyyVar
yyyE
yyVar
yyE
In this case, there are not unobserved components but consider the model for fundamental prices with GARCH errors
In this case, the variances of the noises cannot be observed with information available at time t-1
ttt
ttty
1
2110
2110
ttttt
ttttt
hh
,
,
ii) Stochastic volatility models assume that the volatility represents the arrival of new information into the market and, consequently, it is unobserved
Both models are able to represent
a) Excess kurtosis
b) Autocorrelations of squares: small and persistent
ttty *
ttt )log()log( 21
2
Although the properties of SV models are more attractive and closer to the empirical properties observed in real financial returns, their estimation is more complicated because the volatility, σt, cannot be observed one-step-ahead.
2.2 State space modelsThe Kalman filter allows the estimation of the underlying
unobserved components.
To implement the Kalman filter we are writting the unobserved model of interest in a general form known as “state space model”.
The state space model is given by
where the matrices Zt, Ht, Tt and Qt can evolve over time as far as they are known at time t-1.
),(,
),(,
tttttt
tttttt
QNT
HNZy
0
0
1
Consider, for example, the random walk plus noise model proposed to represent fundamental prices in the market.
In this case, the measurement equation is given by
Therefore, Zt=1, the state αt is the underlying level μt and Ht=
The transition equation is given by
and Tt=1 and Qt=
ttttttt yZy
ttttttt T 11
2
2
Unobserved component models depend on several disturbances. Provided de model is linear, the components can be combined to give a model with a single disturbance: reduced form.
The reduced form is an ARIMA model with restrictions in the parameters.
Consider the random walk plus noise model
In this case
The mean and variance of are given by
ttt
ttty
1
ttty
ty
0)()( ttt EyE
222 2)()( ttt EyVar
The autocorrelation function is given by
signal to noise ratio.The reduced form is an IMA(1,1) model with
negative parameter where
2,0
1,2
1
2)(
22
2
h
hq
h
2
2
q
2
242 qqq
When, q=0, reduces to a non-invertible MA(1) model, i.e. yt is a white noise process. On the other hand, as q increases, the autocorrelations of order one, and consequently, θ , decreases. In the limit, if , is a white noise and yt is a random walk.
ty
02 ty
Although we are focusing on univariate series, the results are valid for multivariate systems. The Kalman filter is made up of two sets of equations:
i) Prediction equations: one-step ahead predictions of the states and their corresponding variances
For example, in the random walk plus noise model:
111 tttttt aTYEa |/
11 ttt mm /
tttttttttt QTPTYaEP '
// | 1111
211 ttt PP /
ii) Updated equations: Each new observation changes (updates) our estimates of the states obtained using past information
where
1
111
111
ttttttttttttttt
tttttttttt
PZFZPPYaaEP
FZPaYEa
/'
//
'//
|)')((
|
11 tttttttt aZyyy //ˆ
tttttt HZPZEF '
/ 12
The updating equations can be derived using the properties of the multivariate normal distribution.
Consider the distribution of αt and yt conditional on past information up to and including time t-1.
The conditional mean and variance are: ttttttt caTYEa 111/ |
tttttttt daZYyEy 1/11/ |ˆ
The conditional covariance can be easily derived by writting
tttttttttt aZdaZy )( 1/1/
'
1/
'1/1/
11/1/
11 )))')(((()')(()|,(
ttt
ttttttttt
ttttttttt
ttt
ZP
ZaaEdaZyaEYyCov
tttt
ttttt
tttt
ttt
t
t
FPZ
ZPP
daZ
aNY
y 1/
'1/1/
1/
1/1 ,|
Consider, once more the random walk plus noise model. In this case,
21
11
11
1
1
ttt
t
ttttt
tttt
ttttt
Pf
fP
PP
myf
Pmm
/
//
//
/ )(
The Kalman filter needs some initial conditions for the state and its covariance matrix at time t=0. There are several alternatives. One of the simplest consists on assuming that the state at time zero es equal to its marginal mean and P0 is the marginal variance.
However, when the state is not stationary this solution is not factible. In this case, one can initiallize the filter by assuming what is known as a diffuse prior (we do not have any information about what happens at time zero: m0=0 and P0=∞.
Now we are in the position to run the Kalman filter. Consider, for example, the random walk plus noise model and that we want to obtain estimates of the underlying level of the series. In this case, the equations are given by
2001
001 0
PP
mm
/
/
222
0
202
0011
101011
11220
20
0111
01011
1
P
PPPFPPP
yyP
Pmy
FP
mm
)(
)(
///
//
/
Consider, for example, that we have observations of a time series generated by a random walk with and
1.14, 0.59, 1.58,….
12 42
8305615
68014159065
141
615
541
141
1
141
2
2
2
12
12
1
1
.)/(
.)..(.
.
.
/
/
P
m
f
P
m
P
m
The Kalman filter gives:
i) One-step ahead and updated estimates of the unobserved states and their associated mean squared errors: at/t-1, Pt/t-1, at and Pt
ii) One-step ahead estimates of yt
iii) One-step ahead errors (innovations) and their variances, νt and Ft
111 ttttttt aZYyEy // |ˆ
Smoothing algorithmsThere are also other algorithms known as smoothing
algorithms that generate estimates of the unobserved states based on the whole sample
The smoothers are very useful because they generate estimates of the disturbances associated with each of the components of the model: auxiliary residuals.
111
11
11
ttttt
tttTtttTt
ttTtttTtTt
PTPP
PPPPPP
aTaPaYEa
/'*
*'//
*/
/*
/
)(
)(|
For example, in the random walk plus noise model:
)(
)(
//
/
//
/
2121
222
11
111
TTTTT
TTTT
TTTTT
TTTT
aaPP
aa
aaPP
aa
The auxiliary residuals are useful to:
i) identify outliers in different components; Harvey and Koopman (1992)
ii) test whether the components are heteroscedastic; Broto and Ruiz (2005a,b)
This test is based on looking at the differences between the autocorrelations of squares and the squared autocorrelations of each of the auxiliary residuals; Maravall (1983).
PredictionOne of the main objectives when dealing with time series
analysis is the prediction of future values of the series of interest. This can easily be done in the context of state space models by running the prediction equations without the updating equations:
In the context of the random walk plus noise model:
ZTQTZZTPTZYaaEP
aTZYEa
k
j
jt
jkT
ktkTkTkTkTkT
Tk
tTkTkT
1
0
''
'
''|)')((
|
21 kPP
mm
TTT
TTkT
/
/
Estimation of the parametersUp to now, we have assumed that the parameters of the
state space model are known. However, in practice, we need to estimate them. In Gaussian state space models, the estimation can be done by Maximum Likelihood. In this case, we can write
The expression of the log-likelihood is then given by
T
t
T
t t
tt F
FT
L1 1
2
2
1
2
12
2
||log)log(log
ttttttttt aZaZy )( // 11
),(| / tttttt FaZNYy 11
The asymptotic properties of the ML estimator are the usual ones as far as the parameters lie on the interior of the parameter space. However, in many models of interest, the parameters are variances, and it is of interest to know whether they are zero (we have deterministic components). In some cases, the asymptotic distribution could still be related with the Normal but is modified as to take into account of the boundary; see Harvey (1989).
If the model is not conditionally Gaussian, then maximizing the Gaussian log-likelihood, we obtain what is known as the Quasi-Maximum Likelihood (QML) estimator. In this case, the estimator looses its eficiency. Furthermore, droping the Normality assumption tends to affect the asymptotic distribution of all the parameters. In this case, the asymptotic distribution is given by
where
11 IJJT )ˆ(
'logL
EJ2
'loglogLL
EI
Unobserved component models for financial time seriesWe are considering two particular applications of
unobserved component models with financial data:
i) Dealing with stochastic volatility
ii) Heteroscedastic components
Stochastic volatility modelsUnderstanding and modelling stock volatility is necessary to
traders for hedging against risk, in options pricing theory or as a simple risk measure in many asset pricing models.
The simplest stochastic volatility model is given by
ttt
ttty
)log()log(
*
21
2
Taking logarithms of squared returns, we obtain a linear although non-Gaussian state space model
Because, log(yt)2 is not truly Gaussian, the Kalman filter yields minimium mean square linear estimators (MMSLE) of ht and future observations rather than minimum mean square estimators (MMSE).
ttt
ttt
ttt
hh
hy
y
1
2
2222
)log(
)log()log()log()log( *
If y1, y2,…,yT is the returns series, we transform the data by
The Kalman filter is given by
2)ˆ/)log((*ytt syyy
)(),(
,
)/(,
///
*//
/
//
1
01011011
1
01011
2011
20
201001
2200
1
10
fP
PPhyfP
hh
Pf
PPhh
Ph
Unobserved heteroscedastic componentsUnobserved component models with heteroscedastic
disturbances have been extensively used in the analysis of financial time series; for example, multivariate models with common factors where both the idiosyncratic and common factors may be heteroscedastic as in Harvey, Ruiz and Sentana (1992) or univariate models where the components may be heteroscedastic as in Broto and Ruiz (2005b).
To simplify the exposition, we are focusing on the random walk plus noise model with ARCH(1) disturbances given by
where
ttt
ttty
1
2110
2110
ttttt
ttttt
hh
,
,
The model can be written in state space form as follows
where
tt
t
t
t
tt
tty
1
1
00
01
01
1
1
12110
21111
110
21
110
2
1
tttttt
tt
tt
tt PmmyEEEH ˆ
12110
2111
110
21
110
2
1
ttttt
tt
tt
tt PEEEQ ˆˆˆ
The model is not conditionally Gaussian, since knowledge of past observations does not, in general, imply knowledge of past disturbances. Nevertheless we can proceed on the basis that the model can be treated as though it were conditionally Gaussian, and we will refer to the Kalman filter as being quasi-optimal.
QML estimates of the Random walk plus noise parameters:
Diagnostics of innovations
4821
11102
2
.ˆ
.ˆ
*.)(
.)(
7814510
8921510
2
Q
Q
Diagnostics of auxiliary residuals
*
*
*
*
.)(
.)(
.)(
.)(
8114810
8052310
7040110
6740810
2
2
Q
Q
Q
Q
QML estimates of the Random walk plus noise parameters:
Diagnostics of innovations
*.)(
.)(
5419310
8982010
2
Q
Q
3527
21502
2
.ˆ
.ˆ
Diagnostics of auxiliary residuals
*
*
*
*
.)(
.)(
.)(
.)(
7319610
8722210
7043110
3951510
2
2
Q
Q
Q
Q