Bayesian inference for the mixed conditional heteroskedasticity model

Size: px
Start display at page:

Download "Bayesian inference for the mixed conditional heteroskedasticity model"

Transcription

1 Econometrics Journal (2007), volume 0, pp doi: 0./j X x Bayesian inference for the mixed conditional heteroskedasticity model L. BAUWENS AND J.V.K. ROMBOUTS CORE and Department of Economics, Université Catholique de Louvain Institute of Applied Economics at HEC Montréal, CIRANO, CIRPEE, CORE and CREF First version received: November 2005; final version accepted: February 2007 Summary We estimate by Bayesian inference the mixed conditional heteroskedasticity model of Haas et al. (2004a Journal of Financial Econometrics 2, 2 50). We construct a Gibbs sampler algorithm to compute posterior and predictive densities. The number of mixture components is selected by the marginal likelihood criterion. We apply the model to the SP500 daily returns. Key words: Bayesian inference, Finite mixture, ML estimation, Value at risk.. INTRODUCTION Finite mixture models (see e.g. McLachlan and Peel 2000) are more and more used in statistics and econometrics. Their main advantage lies in the flexibility they provide in model specification, compared to the use of a more simple distribution. On the other hand, these models are more difficult to estimate than corresponding models without a mixture, but their estimation becomes more and more feasible as computational power increases. However, computational power is not sufficient, one needs also good algorithms. Maximum likelihood estimation of mixture models is not at all as easy as for non-mixture models, and not very reliable in some cases. The EM algorithm was initially developed in this perspective (see Dempster et al. 977). Bayesian estimation is also very efficient for mixture models (see Marin et al. 2005; Geweke and Keane 2005). Conditionally heteroskedastic models are very widespread for modelling time-series of financial returns. The most used class of model is the GARCH family (see e.g. Bollerslev et al. 994) for a survey. A lot of research has been devoted to refine the dynamic specification of the conditional variance equation, for which the benchmark is the linear GARCH specification of Bollerslev (986). The conditional distribution of the model error term is chosen by most researchers among the normal, Student-t, skewed versions of these, and the GED distribution (see Nelson 99). Empirical models typically include around five parameters to fit time-series of a few thousand observations. This may be considered as a powerful way to represent the data. Such models fit the most important stylized facts of financial returns, which are volatility clustering and fat tails. However, a typical result of the estimation of such models is that the conditional variance is almost integrated of the order one and therefore very persistent, at least for relatively long time series at the daily frequency. Several authors have argued that this could be an artefact of structural changes (see e.g. Diebold 986; Mikosch and Starica 2004). Furthermore, it has also. Published by Blackwell Publishing Ltd, 9600 Garsington Road, Oxford OX4 2DQ, UK and 350 Main Street, Malden, MA, 0248, USA.

2 Bayesian inference 409 been observed that volatility is less persistent around crisis periods than during normal periods. Such empirical regularities can be captured by using a finite mixture approach. Finite mixture GARCH models have been recently developed by Haas et al. (2004a), who build on the results of Wong and Li (2000, 200), Haas et al. (2004b) and Alexander and Lazar (2004). All these authors use ML estimation, while Bauwens et al. (2004) propose a particular two component mixture GARCH model and estimate it by Bayesian inference. Bayesian estimation of GARCH models has been studied by Geweke (989), Kleibergen and Dijk (993) and Bauwens and Lubrano (998). Note that finite mixtures are different from continuous mixtures. An example of a continuous mixture GARCH model is a GARCH equation combined with a Student-t distribution for the error term, since the latter distribution is a continuous mixture of normal distributions whose variance follows an inverted-gamma distribution. Thus, a t-garch model results in fatter tails than a Gaussian GARCH but it does not increase the flexibility of the conditional variance equation, whereas a finite mixture GARCH model permits this. Bayesian inference for the mixed normal GARCH model of Haas et al. (2004a) is the subject of this paper. The model is defined in Section 2. In Section 3, we explain how this model can be estimated in the Bayesian framework. We design a Gibbs sampler, and discuss how to obtain predictive densities and how to choose the number of components of the mixture. In Section 4, we apply the approach to returns of the SP500 index. 2. MIXED CONDITIONAL HETEROSKEDASTICITY Haas et al. (2004a) define a mixture model on a demeaned series = y t E(y t F t ) where F t is the information set up to time t and the conditional mean does not depend on the components of the mixture. They call this model (diagonal) MN-GARCH (MN for mixed normal). The conditional CDF of is the K-component mixture ( ) K µ k F( F t ) = π k () hk,t where k= h k,t = ω k + α k ɛ 2 t + β kh k,t (2) and ( ) is the standard Gaussian cdf. Note that the parameter π k is positive for all k and K k= π k =, which is imposed by setting π K = K k= π k. The other Greek letters denote the other parameters. The zero mean assumption on is ensured by the restriction K µ K = k= π k µ k π K. Haas et al. (2004a) also consider a more general model where the h k,t s are GARCH(p k, q k ) and more importantly may depend on other h j,t s, k j (contrary to the diagonal specification defined above). The weak stationarity condition for a (diagonal) MN-GARCH model is [ K ( )] π K k α k β k β k > 0, (4) β k k= k=

3 40 L. Bauwens and J.V.K. Rombouts where β k = β k. Its unconditional variance is then given by E ( ɛ 2 t ) = c + K k= π kω k / β k K k= π k( α k β k )/ β k (5) where c = K k= π kµ 2 k. One can check that the process may be stationary even if some components are not stationary provided that these components have sufficiently low corresponding component weights. Strict stationarity conditions are not known for this model. 3. BAYESIAN INFERENCE We specify the conditional mean E(Y t F t )asanar(p) model with a constant term. The model is then written as y t = ρ 0 + ρ y t + +ρ p y t p +, (6) where follows the MN-GARCH specification defined by (). We replace in the sequel (6) by the shorter notation y t = ρ x t +, (7) where ρ = (ρ 0, ρ,..., ρ p ) and x t = (, y t,..., y t p ). The likelihood of the MN-GARCH model for T observations is given by L( y) = T t= k= K π k φ(y t µ k + ρ x t,θ k ) (8) where is the vector regrouping the parameters ρ and π k, µ k, θ k for k =,..., K, y = (y, y 2,..., y T ) and φ( µ k + ρ x t, θ k ) denotes a normal density with mean µ k + ρ x t and variance h k,t that depends on θ k = (ω k, α k, β k ). A direct evaluation of the likelihood function is difficult because it consists of a product of sums. To alleviate this evaluation, we introduce for each observation a state variable S t {, 2,..., K } that takes the value k if the observation y t belongs to component k. The vector S T contains the state variables for the T observations. We assume that the state variables are independent given the group probabilities, and the probability that S t is equal to k is equal to π k : ϕ(s T π) = T ϕ(s t π) = where π = (π, π 2,..., π K ). Given S T and y the likelihood function is L( S T, y) = t= T π St, (9) t= T π St φ(y t µ St + ρ x t,θ St ), (0) t= which is easier to evaluate than (8). Since S T is not observed we treat it as a parameter of the model. This technique is called data augmentation, (see Tanner and Wong (987) for more details). Although the augmented model contains more parameters, inference becomes easier by making use of Markov chain Monte Carlo (MCMC) methods. In this paper, we implement a Gibbs

4 Bayesian inference 4 sampling algorithm that allows to sample from the posterior distribution by sampling from its conditional posterior densities, which are called blocks. The blocks of the Gibbs sampler, and the prior densities, are explained in the next subsections, using the parameter vectors ρ, π, θ = (θ, θ 2,..., θ k ) and µ = (µ, µ 2,..., µ K ). The joint posterior distribution is given by ϕ(s T,ρ,µ,θ,π y) ϕ(ρ) ϕ(µ) ϕ(θ) ϕ(π) T π St φ(y t µ St + ρ x t,θ St ), () where ϕ(ρ), ϕ(µ), ϕ(θ), ϕ(π) are the corresponding prior densities. Thus, we assume prior independence between ρ, π, µ and θ. We define these prior densities below when we explain the different blocks of the Gibbs sampler. t= 3.. Sampling S T from ϕ(s T ρ, µ, θ, π, y) Given ρ, µ, θ, π and y, the posterior density of S T is proportional to L( S T, y). It turns out that the S t s are mutually independent, so that we can write the relevant conditional posterior density as ϕ(s T ρ,µ,θ,π, y) = T ϕ(s t ρ,µ,θ,π, y). (2) As the sequence {S t } t= T is equivalent to a multinomial process, we simply have to sample from a discrete distribution where the K probabilities are given by P(S t = k ρ,µ,θ,π, y) = t= π k φ(y t µ k + ρ x t,θ k ) K j= π, (k =,...,K ). jφ(y t µ j + ρ x t,θ j ) To sample S t, we draw one observation from a uniform distribution on (0, ) and decide which group k to take according to Sampling π from ϕ(π S T, ρ, µ, θ, y) The full conditional posterior density of π depends only on S T and y and is given by ϕ(π S T, y) = ϕ(π S T ) ϕ(π) K k= π x k k, (4) where x k is the number of times that S t = k. The prior ϕ(π) is chosen to be a Dirichlet distribution, Di(a 0, a 20 a K 0 ) with parameter vector a 0 = (a 0, a 20 a K 0 ) (see the Appendix for more details). As a consequence, ϕ(π S T, y) is also a Dirichlet distribution, Di(a, a 2 a K ) with a k = a k0 + x k, k =, 2,..., K Sampling µ from ϕ(µ S T, ρ, π, θ, y) We show in the Appendix that the conditional distribution of µ = (µ,µ 2,...,µ K ) is Gaussian with a non-diagonal covariance matrix. Once µ has been drawn, the last mean µ K is obtained from.

5 42 L. Bauwens and J.V.K. Rombouts 3.4. Sampling ρ from ϕ(ρ S T, µ, π, θ, y) Given that the conditional variances h k,t depend on ρ, the conditional posterior distribution for this block does not belong to a family that can be easily simulated. We can, for example, employ the Metropolis Hastings algorithm. For the latter, we use a Gaussian proposal q( ) the functional form of which is given in the Appendix. The acceptance probability at iteration n + for candidate ρ has the form ( ϕ(s T,ρ,µ,θ,π y) q(ρ ; ρ = ρ (n) ) ) min ϕ(s T,ρ (n),µ,θ,π y) q(ρ (n) ; ρ = ρ ),. (5) Apart from ρ, the other parameters in the posterior ϕ( ) are fixed at their latest draw Sampling θ from ϕ(θ S T, ρ, µ, π, y) By assuming prior independence between the θ k s, i.e. ϕ(θ) = K k= ϕ (θ k ), it follows that ϕ(θ S T,ρ,µ,π,y) = ϕ(θ S T,ρ,µ,y) = ϕ(θ ρ,µ, ỹ )ϕ ( θ 2 ρ,µ 2, ỹ 2) ϕ ( θ K ρ,µ K, ỹ K ) (6) where ỹ k ={y t S t = k} and ϕ ( θ k ρ,µ k, ỹ k) ϕ(θ k ) t S t =k φ(y t µ k + ρ x t,θ k ). (7) Since we condition on the state variables, we can simulate each block θ k separately. We do this with the griddy-gibbs sampler (see the Appendix, and for further details, see Bauwens et al. 999). Note that intervals of values for ω k, α k and β k must be defined. The choice of these bounds needs to be finely tuned in order to cover the range of the parameter over which the posterior is relevant. For the deterministic integration we used 33 points, which proved to be enough according to several experiments Label switching In mixture models, the labelling of the components is arbitrary and one can shuffle the labels without changing the likelihood function. The latter has as many modes as there are permutations of the regime labels. In the Bayesian framework, one can run an algorithm that explores all the modes, which may not be easy and take a lot of computing time, or impose an identification condition through the prior information. The solution used by Haas et al. (2004a) in the ML framework is to impose that π >π 2 > >π K but this solution is destroying the result that the full conditional posterior of π is Dirichlet and thus the sampling of π would be more difficult. We choose the solution of imposing that the component specific parameters have sufficiently different prior densities (e.g. through non-overlapping supports, but this is an extreme solution that is not necessary).

6 Bayesian inference Predictive densities Predictive densities are essential for financial applications such as portfolio optimization and risk management. Unlike prediction in the classical framework, predictive densities take into account parameter uncertainty by construction. The predictive density of y T + is given by f (y T + y) = f (y T +, y) ϕ( y) d, (8) where f (y T +, y) = K k= π kφ(y T + µ k + ρ x T +,θ k ) as implied by (). An analytical solution to (8) is not available but extending the algorithm of Geweke (989), it can be approximated by N ( N K π ( j) k φ ( y T + µ ( j) k + ρ ( j) x T +,θ ( j) k, y )) (9) k= j= where the superscript (j) indexes the draws generated with the Gibbs sampler and N is the number of draws. Therefore, simultaneously with the Gibbs sampler, we repeat N times the following two-step algorithm Step : simulate ( j) ϕ ( y). This is done by the Gibbs sampler. Step 2: simulate y ( j) T + f (y T + ( j), y). Go to step. Extending the idea used for y T +, the predictive density for y T +s may be written as [ f (y T +s y) =... f (y T +s y T +s,...,y T +, y, ) f (y T +s y T +s 2,...,y T +, y, )... f (y T + y, )dy T +s dy T +s 2 dy T + ]ϕ( y) d (20) for which draws can be obtained by extending the above algorithm to a (s+)-step algorithm. The draw of y T + serves as conditioning information to draw y T +2, both realisations serve to draw y T +3, etc. All these draws are easily generated from the finite mixture of normal densities. A non-bayesian procedure typically proceeds by conditioning on a point estimate of, which ignores the estimation uncertainty Marginal likelihood The marginal likelihood of y, also called predictive density, is useful for selecting the number of components K in the mixture. It is well defined when the prior density is integrable, which is the case for the prior we use in this paper. For example, Bayes factors are ratios of marginal likelihoods (see Kass and Raftery 995 for a detailed explanation). The marginal likelihood is defined as the integral of the likelihood with respect to the prior density m(y) = L( y)ϕ( ) d. (2)

7 44 L. Bauwens and J.V.K. Rombouts Since this is the normalizing constant in Bayes theorem, we can also write L( y)ϕ( ) m(y) =. (22) ϕ( y) Note that (22) is an identity that holds for every. Deterministic numerical integration of (2) is computationally too demanding for the finite mixture model of this paper. Instead, we calculate the marginal likelihood by the Laplace approximation (see Tierney and Kadane 986). To explain this, let us define exp(h( )) = L( y)ϕ( ). The Laplace approximation is based on a secondorder Taylor expansion of h( ) around the posterior mode ˆ = arg max ln φ( y), so that the first-order term in the expansion vanishes: h( ) h( ˆ ) + 2 ( ˆ ) 2 h( ) = ˆ ( ˆ ). (23) Therefore, the marginal likelihood can be computed as ( ) exp h( )d exp(h( ˆ )) exp 2 ( ˆ ) 2 h( ) = ˆ ( ˆ ) d (24) or m(y) = L( ˆ y) ϕ( ˆ ) (2π) k/2 ( ˆ ) /2, (25) where k is the dimension of and [ ] ( ˆ ) = 2 ln L( y) ϕ( ) = ˆ. (26) We choose the model with the highest marginal likelihood value. Another possibility to choose the number of components is to treat K as an additional parameter in the model as is done in Richardson and Green (997) who make use of the reversible jump MCMC methods. In this way, the prior information on the number of components can be taken explicitly into account by specifying for example a Poisson distribution on K in such a way that it favours a small number of components. 4. APPLICATION TO S&P500 DATA We fit the two component mixture model to daily S& P500 percentage return data from 0/03/994 to 09/06/2005 (3047 observations). Descriptive statistics are given in Table. Panel (a) of Figure displays the sample path of the returns (the other panels are explained further down in this section). It is clear that excess kurtosis and volatility clustering are present in the data. We analyzed whether a dynamic specification for the conditional mean is necessary and we found evidence for an autoregressive model of order three. Thus, we estimate the model defined by (6) with p = 3 and by () with K = 2. The ML estimates and the Bayesian first two marginal posterior moments are given in Table 2. The parameters a k0 of the Dirichlet prior for π are all equal to, which means that the prior density for the probability π is uniform on (0, ). The prior densities for the other parameters are all independent. For the parameters in ρ and µ, these prior densities are flat on wide intervals (their bounds need not be specified). For the GARCH parameters, the densities are uniform on finite

8 Bayesian inference 45 Table. Descriptive statistics S&P 500 returns. Observations 3047 Mean Standard deviation.07 Maximum 5.58 Minimum 7. Skewness 0. Kurtosis 6.74 Statistics for SP500 percentage daily returns from 0/03/994 to 09/06/2005. intervals given by <ω < 0.004, <α < 0.08, 0.89 <β < 0.99, <ω 2 < 0.6, 0.08 <α 2 < 0.65, 0.73 <β 2 < These values are the bounds used in the griddy-gibbs sampler part of the algorithm described in Section 3.5. The posterior marginal distributions of all the parameters are given in Figure 2. The x-axes for the GARCH parameters are the prior intervals reported above. Note that the posterior marginals of ω and ω 2 are somewhat truncated at zero given that they are restricted to be positive. A scatterplot of the α and β draws for both components is given in Figure 3. A clear conclusion from the figures is that the data are much less informative on the explosive regime than on the stable one. We checked the convergence of the Gibbs sampler for all parameters with CUMSUM plots of the draws (see Bauwens et al. 999 for details). From Table 2, we conclude that the ML and Bayesian parameter estimates are close to each other. The posterior standard deviations (SD) are in most cases a little bit smaller than the ML standard errors (SE) computed from the Hessian matrix evaluated at the ML estimates. These differences come to some extent from the use of finite intervals as support of some prior densities. The estimated probability is about 0.83 for the first component which is driven by a persistent stationary GARCH process (α + β = 0.98). The second component of the mixture has an explosive conditional variance (α 2 + β 2 =.7) with a probability of about 0.7. To illustrate the interest of the Bayesian estimation of the two component model, we report in panel (b) of Figure the sample path of the posterior means of the state variables (mean states), i.e. for each observation we count the proportion of the Gibbs sampler generated state values that correspond to the explosive regime. The mean of these proportions is equal to 0.6 which is close to the probability of being in the second component of the mixture. Panel (c) of the figure contains the scatter plot of these mean states and the corresponding returns. From these graphs, one can identify a clear association between the explosive regime and the extreme returns, especially for the negative returns. The asymmetric shape of this relation can be interpreted as the leverage effect, i.e. the association of large negative returns of a given value to a higher volatility than in the case of positive returns having the same absolute value. As a comparison, we report estimates of the single component mixture model, i.e. the conventional GARCH(,) model. The ML estimates and the Bayesian first two marginal posterior moments are given in Table 3. The process looks like integrated in variance given that α + β is estimated at This may be interpreted as a compromise between the less persistent and explosive components of the mixture model. We obtained a similar result when we estimated the GARCH(,) model with data simulated from a two component mixture. Thus, the observation

9 46 L. Bauwens and J.V.K. Rombouts (a) S&P 500 returns (b) Mean states (c) Scatterplot returns and mean states Figure. Information on states.

10 Bayesian inference 47 Table 2. Estimation results S&P 500. MLE Bayes Estimate SE Mean SD ρ ρ ρ ρ π µ ω α β ω α β Results for the AR two component normal mixture GARCH(,) model. Sample of 3047 observations from 0/03/994 to 09/06/2005. that a quasi-integrated GARCH model (α + β ) is obtained in many empirical results can be explained by a lack of flexibility of this model. In Table 4, we report the marginal likelihood and the Bayesian information criterion (BIC) values for the single-and two-component models. The results indicate a strong preference for the two-component model. As for any time series model, prediction is essential. As we explain in Section 3.7, Bayesian inference allows to obtain predictive densities that by construction incorporate parameter uncertainty. Furthermore, they can be easily computed together with the posterior densities during the application of the Gibbs sampler for the model parameters. We report in Figure 4 the computed predictive densities for a horizon up to five days out of sample (September 7, 2005 until September, 2005). Eyeballing the graphs, we see that the left tail of the predictive densities is fatter for the two component model than for the simple GARCH model. In Table 5, we report the skewness and kurtosis coefficients, plus the value-at-risk (VaR) at per cent for the five days. Judging from the skewness and kurtosis values, the single-component model yields close to normal predictive densities, while the two component model produces predictive densities with fatter tails and negative skewness. Because of the fatter left tail of the two component model predictive densities, their VaR are smaller than for the one component model. We also computed a sequence of one step ahead VaR s from the end of the sample until September 2006 (250 new observations). We computed the failure rates for the, 5, 0, 90, 95 and 99 VaR levels. Likelihood ratio tests for each VaR level, for both models, do not reject, hence both models are able to fit the tails of the distribution well. This similar performance can be explained by the fact that the distribution of the returns in the covered forecast period is very close to being normal, that is very symmetrical and with low excess kurtosis.

11 48 L. Bauwens and J.V.K. Rombouts (a) ρ 0 (b) ρ (c) ρ 2 (d) ρ (e) π (f) µ (g) ω and ω (h) α and α (i) β and β 2 Figure 2. Posterior densities (kernel estimates from Gibbs output) for two component normal mixture GARCH(,) model.

12 Bayesian inference 49 Figure 3. MCMC draws in α-β space. Table 3. Estimation results (one component) S&P 500. MLE Bayes Estimate SD mean SE ρ ρ ρ ρ ω α β Results for Gaussian GARCH(,) model. Sample of 3047 observations from 0/03/994 to 09/06/ CONCLUSION We have shown how a certain type of mixture GARCH model can be estimated by Bayesian inference. ML estimation is typically not easy because of the complexity of the likelihood function. In Bayesian estimation, this is taken care of by enlarging the parameter space with state variables, so that a Gibbs sampling algorithm is easy to implement. Despite a higher computing time, the Bayesian solution is reliable since estimation does not fail, while this may happen in MLE. Moreover, the Gibbs algorithm delivers automatically posterior results on the state variables, which can be used for interpreting the nature of the second regime, as we illustrate in Section 4. Finally, the Gibbs algorithm can be extended to include the computation of predictive densities,

13 420 L. Bauwens and J.V.K. Rombouts Table 4. Model choice criteria S&P500 data. K Marginal log-lik. Maximized log-lik. No. par. BIC K is the number of components of the normal mixture GARCH(,) model (a) T+ (b) T (c) T+3 (d) T (e) T+5 Figure 4. Kernel density estimates of predictive densities for September 7 to, 2005 (dotted line for the two component model, solid line for single component model).

14 Bayesian inference 42 Table 5. Features of predictive densities h One component Two components Skewness Kurtosis VaR h is the post-sample prediction horizon. VaR is the per cent value-at-risk quantile. which takes care of estimation uncertainty. Prediction in the ML approach is typically done by conditioning on the ML estimate and therefore ignores estimation uncertainty. Bayesian estimation of other types of mixture GARCH models can be handled in a similar way as in this paper. A bivariate mixture GARCH model is estimated by Bauwens et al. (2006). ACKNOWLEDGMENTS We thank Viorel Maxim for research assistance and Arie Preminger and anonymous referees for useful comments. Bauwens s work was supported in part by the European Community s Human Potential Programme under contract HPRN-CT , MICFINMA and by a FSR grant from UCL. Rombouts s work was supported by a HEC Montréal Fonds de démarrage and by the Centre for Research on e-finance. This text presents research results of the Belgian Program on Interuniversity Poles of Attraction initiated by the Belgian State, Prime Minister s Office, Science Policy Programming. The scientific responsibility is assumed by the authors. REFERENCES Alexander, C. and E. Lazar (2004). Normal mixture GARCH(,): Applications to exchange rate modelling. Journal of Applied Econometrics 2, Bauwens, L., C. Bos, R. van Oest and H. van Dijk (2004). Adaptive radial-based direction sampling: A class of flexible and robust Monte Carlo integration methods. Journal of Econometrics 23,

15 422 L. Bauwens and J.V.K. Rombouts Bauwens, L., C. Hafner and J. Rombouts (2006). Multivariate mixed normal conditional heteroskedasticity. CORE Discussion Paper 2006/2, Louvain-La-Neuve. Computational Statistics and Data Analysis (Forthcoming). Bauwens, L. and M. Lubrano (998). Bayesian inference on GARCH models using the Gibbs sampler. Econometrics Journal, C Bauwens, L., M. Lubrano and J. Richard (999). Bayesian Inference in Dynamic Econometric Models. Oxford: Oxford University Press. Bollerslev, T. (986). Generalized Autoregressive Conditional Heteroskedasticity. Journal of Econometrics 3, Bollerslev, T., R. Engle and D. Nelson (994). ARCH Models. in Handbook of Econometrics, ed. by R. Engle and D. McFadden, chap. 4, pp North Holland Press, Amsterdam. Dempster, A., N. Laird and D. Rubin (977). Maximum likelihood for incomplete data via the EM algorithm (with discussion). Journal of the Royal Statistical Society Series B 39, 38. Diebold, F. (986). Comment on Modeling the Persistence of Conditional Variances. Econometric Reviews 5, 5 6. Geweke, J. (989). Exact Predictive Densitites in Linear Models with ARCH Disturbances. Journal of Econometrics 40, Geweke, J. and M. Keane (2005). Smoothly mixing regressions. Working Paper, University of Iowa. Journal of Econometrics (Forthcoming). Haas, M., S. Mittnik and M. Paolella (2004a). Mixed normal conditional heteroskedasticity. Journal of Financial Econometrics 2, Haas, M., S. Mittnik and M. Paolella (2004b). A new approach to Markov-Switching GARCH models. Journal of Financial Econometrics 2, Kass, R. and A. Raftery (995). Bayes factors. Journal of the American Statistical Association 90, Kleibergen, F. and H. V. Dijk (993). Non-stationarity in GARCH Models: A Bayesian Analysis. Journal of Applied Econometrics 8, S4 6. Marin, J., K. Mengersen and C. Robert (2005). Bayesian Modelling and Inference on Mixtures of Distributions, Handbook of Statistics 25.D. Dey and C.R. Rao (eds), Elsevier-Sciences. McLachlan, G. and D. Peel (2000). Finite Mixture Models. New York Wiley Interscience. Mikosch, T. and C. Starica (2004). Nonstationarities in financial time series, the long-range dependence, and the IGARCH Effects. Review of Economics and Statistics 86, Nelson, D. (99). Conditional heteroskedasticity in asset returns: A new approach. Econometrica 59, Richardson, S. and P. Green (997). On Bayesian Analysis of Mixtures with an Unknown Number of Components. Journal of the Royal Statistical Society, Series B 59, Tanner, M. and W. Wong (987). The calculation of posterior distributions by data augmentation. Journal of the American Statistical Association 82, Tierney, L. and J. Kadane (986). Accurate Approximations for posterior moments and marginal densities. Journal of the American Statistical Association 8, Wilks, S. (962). Mathematical Statistics. New York: Wiley. Wong, C. and W. Li (2000). On a mixture autoregressive model. Journal of the Royal Statistical Society, Series B, 62, Wong, C. and W. Li (200). On a mixture autoregressive conditional heteroscedastic model. Journal of the American Statistical Association 96,

16 Bayesian inference 423 The Dirichlet density function is given by f Di (π a, a 2 a K ) = APPENDIX The Dirichlet distribution Ɣ(A) K k= Ɣ(a k) K k= π a k k SK (π) (A.) where a k > 0(k =,...,K ), A = K i= a i and S K ={π k, k =,...,K π k > 0 k, K k= π k = }. The first two moments are given by E(π i a) = a i, V (π A i a) = a i (A a i ) and cov(π A 2 (A+) i,π j a) = a i a j, A 2 (A+) respectively. We sample a Dirichlet distribution by sampling K independent gamma random variables, X k G(a k, ), and transforming them to X i π i = i =,...,K X + + X K π K = π π 2 π K. It follows that (π,..., π K ) Di(a,..., a K ). Other properties of the Dirichlet distribution can be found in Wilks (962). (k) Proof that ϕ( µ S T,ρ,π,θ,y) is Gaussian We illustrate this for K = 3. Minus two times the log-kernels for the first two components are given by ( ) 2 µ k = c k + µ 2 k 2µ k (k =, 2), (A.2) hk,t h k,t h k,t (k) where (k) means summation over all t for which s t = k, and c k does not depend on µ k. The third mixture component contributes in the following way: ( ) ɛt + π µ + π 2 2 ( µ 2 = c 3 + µ 2 h3,t π + 2µ π ) 2 The sum of (A.2) and (A.3) can be written compactly as (k) ( + µ µ 2 π 2 π 2 ) π π 2 µ µ 2 π 2 3. (A.3) ( µ µ) A( µ µ) + c, (A.4) where c is a constant not depending on µ, bydefining the matrix A as ( ) 2 () h,t + π π π 2 π 2 3 π π 2 π3 2 (2) h 2,t + ( π2 ) 2, (A.5)

17 424 L. Bauwens and J.V.K. Rombouts and the vector µ as A b, where b = [ π π 2 () (2) h,t h 2,t ]. (A.6) Minus one half times the first term of (30) is the log-kernel of a bivariate Gaussian density with mean µ and covariance matrix A. For K components, where π = (π,...,π K ), and ( A = diag () b =,..., h,t π π K (K ) π K π K (K ) h (K ) K,t h K,t (). h K,t (K ) ) π π + h,t π 2 K h K,t (K ). h K,t, (A.7) (A.8) Gaussian proposal for ϕ(ρ S T, µ, π, θ, y) We illustrate this for K = 3. We condition on the previous draw of ρ, denoted by ρ, and we compute h k,t conditional on ρ and therefore use the notation h k,t. Minus two times the log-kernels for the first two components is given by ( y t ρ x t µ k (k) h k,t ) 2 = c k + ρ (k) x t x t h k,t ρ 2ρ (k) x t y t h k,t + 2µ k ρ (k) x t h k,t (k =, 2), (A.9) where (k) means summation over all t for which s t = k, and c k does not depend on ρ. The third mixture component contributes in the following way: ( ) yt ρ x t + π µ + π 2 2 µ 2 = c 3 + ρ x t x t ρ 2ρ h x t y t π 2µ ρ 3,t h x t 3,t 2µ 2 π 2 ρ x t. (A.0) The sum of (A.9) for k = and 2 and (A.0) can be written compactly as where c is a constant not depending on ρ. The matrix A is defined by (ρ ρ) A(ρ ρ) + c, (A.) 3 k= (k) x t x t h k,t and the vector ρ is equal to A b, where b = 3 k= (k) x t y t h k,t + 2 k= µ k [ (k) ] x t π k x t. h k,t

18 Bayesian inference 425 Minus one half times the first term of (37) is the log-kernel of a multivariate Gaussian density with mean ρ and covariance matrix A. To generalize the last two formulas to K > 3, replace 3 by K and 2 by K. Griddy-Gibbs sampler for ϕ(θ k ρ,µ k, ỹ k ) The algorithm works as follows at iteration n + (for lighter notations, we drop the index k and the conditioning variables ρ, µ k and ỹ k ): () Using (7), compute κ(ω α n, β n ), the kernel of the conditional posterior density of ω given the values of α and β sampled at iteration n, over a grid (ω, ω 2, ω G ), to obtain the vector G κ = (κ, κ 2,,κ G ). (2) By a deterministic integration rule using M points, compute G f = (0, f 2,..., f G ) where ωi f i = κ(ω α n,β n ) dω, i = 2,...,G. (A.2) ω Generate u U(0, f G ) and invert f (ω α (n), β (n) ) by numerical interpolation to get a draw ω (n+) ϕ(ω α (n), β (n) ). (4) Repeat Steps 3 for ϕ (α ω (n+), β n ) and ϕ(β ω (n+), α (n+) ).

CORE DISCUSSION PAPER 2005/85 BAYESIAN INFERENCE FOR THE MIXED CONDITIONAL HETEROSKEDASTICITY MODEL. December 1, Abstract

CORE DISCUSSION PAPER 2005/85 BAYESIAN INFERENCE FOR THE MIXED CONDITIONAL HETEROSKEDASTICITY MODEL. December 1, Abstract CORE DISCUSSION PAPER 2005/85 BAYESIAN INFERENCE FOR THE MIXED CONDITIONAL HETEROSKEDASTICITY MODEL L. Bauwens and J.V.K. Rombouts December, 2005 Abstract We estimate by Bayesian inference the mixed conditional

More information

Stock index returns density prediction using GARCH models: Frequentist or Bayesian estimation?

Stock index returns density prediction using GARCH models: Frequentist or Bayesian estimation? MPRA Munich Personal RePEc Archive Stock index returns density prediction using GARCH models: Frequentist or Bayesian estimation? Ardia, David; Lennart, Hoogerheide and Nienke, Corré aeris CAPITAL AG,

More information

Volatility. Gerald P. Dwyer. February Clemson University

Volatility. Gerald P. Dwyer. February Clemson University Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use

More information

Econometrics. Journal. Theory and inference for a Markov switching GARCH model

Econometrics. Journal. Theory and inference for a Markov switching GARCH model The Econometrics Journal Econometrics Journal (2010), volume 13, pp. 218 244. doi: 10.1111/j.1368-423X.2009.00307.x Theory and inference for a Markov switching GARCH model LUC BAUWENS,ARIE PREMINGER AND

More information

August 13, 2007, revised February 21, 2008

August 13, 2007, revised February 21, 2008 CORE DISCUSSION PAPER 2007/55 THEORY AND INFERENCE FOR A MARKOV SWITCHING GARCH MODEL Luc Bauwens 1, Arie Preminger, 2 and Jeroen V.K. Rombouts 3 August 13, 2007, revised February 21, 2008 Abstract We

More information

Session 5B: A worked example EGARCH model

Session 5B: A worked example EGARCH model Session 5B: A worked example EGARCH model John Geweke Bayesian Econometrics and its Applications August 7, worked example EGARCH model August 7, / 6 EGARCH Exponential generalized autoregressive conditional

More information

Modeling conditional distributions with mixture models: Theory and Inference

Modeling conditional distributions with mixture models: Theory and Inference Modeling conditional distributions with mixture models: Theory and Inference John Geweke University of Iowa, USA Journal of Applied Econometrics Invited Lecture Università di Venezia Italia June 2, 2005

More information

Analytical derivates of the APARCH model

Analytical derivates of the APARCH model Analytical derivates of the APARCH model Sébastien Laurent Forthcoming in Computational Economics October 24, 2003 Abstract his paper derives analytical expressions for the score of the APARCH model of

More information

Generalized Autoregressive Score Models

Generalized Autoregressive Score Models Generalized Autoregressive Score Models by: Drew Creal, Siem Jan Koopman, André Lucas To capture the dynamic behavior of univariate and multivariate time series processes, we can allow parameters to be

More information

Econ 423 Lecture Notes: Additional Topics in Time Series 1

Econ 423 Lecture Notes: Additional Topics in Time Series 1 Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes

More information

2007/97. Mixed exponential power asymmetric conditional heteroskedasticity. Mohammed Bouaddi and Jeroen V.K. Rombouts

2007/97. Mixed exponential power asymmetric conditional heteroskedasticity. Mohammed Bouaddi and Jeroen V.K. Rombouts 2007/97 Mixed exponential power asymmetric conditional heteroskedasticity Mohammed Bouaddi and Jeroen V.K. Rombouts CORE DISCUSSION PAPER 2007/97 Mixed exponential power asymmetric conditional heteroskedasticity

More information

Markov Chain Monte Carlo Methods

Markov Chain Monte Carlo Methods Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

Bayesian Semiparametric GARCH Models

Bayesian Semiparametric GARCH Models Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods

More information

GARCH Models. Eduardo Rossi University of Pavia. December Rossi GARCH Financial Econometrics / 50

GARCH Models. Eduardo Rossi University of Pavia. December Rossi GARCH Financial Econometrics / 50 GARCH Models Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 50 Outline 1 Stylized Facts ARCH model: definition 3 GARCH model 4 EGARCH 5 Asymmetric Models 6

More information

Bayesian Semiparametric GARCH Models

Bayesian Semiparametric GARCH Models Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods

More information

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture

More information

Gaussian kernel GARCH models

Gaussian kernel GARCH models Gaussian kernel GARCH models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics 7 June 2013 Motivation A regression model is often

More information

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix Infinite-State Markov-switching for Dynamic Volatility Models : Web Appendix Arnaud Dufays 1 Centre de Recherche en Economie et Statistique March 19, 2014 1 Comparison of the two MS-GARCH approximations

More information

13. Estimation and Extensions in the ARCH model. MA6622, Ernesto Mordecki, CityU, HK, References for this Lecture:

13. Estimation and Extensions in the ARCH model. MA6622, Ernesto Mordecki, CityU, HK, References for this Lecture: 13. Estimation and Extensions in the ARCH model MA6622, Ernesto Mordecki, CityU, HK, 2006. References for this Lecture: Robert F. Engle. GARCH 101: The Use of ARCH/GARCH Models in Applied Econometrics,

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

The Bayesian Approach to Multi-equation Econometric Model Estimation

The Bayesian Approach to Multi-equation Econometric Model Estimation Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation

More information

Multivariate Asset Return Prediction with Mixture Models

Multivariate Asset Return Prediction with Mixture Models Multivariate Asset Return Prediction with Mixture Models Swiss Banking Institute, University of Zürich Introduction The leptokurtic nature of asset returns has spawned an enormous amount of research into

More information

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract Bayesian analysis of a vector autoregressive model with multiple structural breaks Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus Abstract This paper develops a Bayesian approach

More information

Bayesian semiparametric GARCH models

Bayesian semiparametric GARCH models ISSN 1440-771X Australia Department of Econometrics and Business Statistics http://www.buseco.monash.edu.au/depts/ebs/pubs/wpapers/ Bayesian semiparametric GARCH models Xibin Zhang and Maxwell L. King

More information

Modeling conditional distributions with mixture models: Applications in finance and financial decision-making

Modeling conditional distributions with mixture models: Applications in finance and financial decision-making Modeling conditional distributions with mixture models: Applications in finance and financial decision-making John Geweke University of Iowa, USA Journal of Applied Econometrics Invited Lecture Università

More information

Index. Pagenumbersfollowedbyf indicate figures; pagenumbersfollowedbyt indicate tables.

Index. Pagenumbersfollowedbyf indicate figures; pagenumbersfollowedbyt indicate tables. Index Pagenumbersfollowedbyf indicate figures; pagenumbersfollowedbyt indicate tables. Adaptive rejection metropolis sampling (ARMS), 98 Adaptive shrinkage, 132 Advanced Photo System (APS), 255 Aggregation

More information

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS

ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

Accounting for Missing Values in Score- Driven Time-Varying Parameter Models

Accounting for Missing Values in Score- Driven Time-Varying Parameter Models TI 2016-067/IV Tinbergen Institute Discussion Paper Accounting for Missing Values in Score- Driven Time-Varying Parameter Models André Lucas Anne Opschoor Julia Schaumburg Faculty of Economics and Business

More information

Bayesian Modeling of Conditional Distributions

Bayesian Modeling of Conditional Distributions Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings

More information

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract Bayesian Estimation of A Distance Functional Weight Matrix Model Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies Abstract This paper considers the distance functional weight

More information

POSTERIOR ANALYSIS OF THE MULTIPLICATIVE HETEROSCEDASTICITY MODEL

POSTERIOR ANALYSIS OF THE MULTIPLICATIVE HETEROSCEDASTICITY MODEL COMMUN. STATIST. THEORY METH., 30(5), 855 874 (2001) POSTERIOR ANALYSIS OF THE MULTIPLICATIVE HETEROSCEDASTICITY MODEL Hisashi Tanizaki and Xingyuan Zhang Faculty of Economics, Kobe University, Kobe 657-8501,

More information

The Metropolis-Hastings Algorithm. June 8, 2012

The Metropolis-Hastings Algorithm. June 8, 2012 The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings

More information

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications

ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty

More information

ADVANCED FINANCIAL ECONOMETRICS PROF. MASSIMO GUIDOLIN

ADVANCED FINANCIAL ECONOMETRICS PROF. MASSIMO GUIDOLIN Massimo Guidolin Massimo.Guidolin@unibocconi.it Dept. of Finance ADVANCED FINANCIAL ECONOMETRICS PROF. MASSIMO GUIDOLIN a.a. 14/15 p. 1 LECTURE 3: REVIEW OF BASIC ESTIMATION METHODS: GMM AND OTHER EXTREMUM

More information

1 Phelix spot and futures returns: descriptive statistics

1 Phelix spot and futures returns: descriptive statistics MULTIVARIATE VOLATILITY MODELING OF ELECTRICITY FUTURES: ONLINE APPENDIX Luc Bauwens 1, Christian Hafner 2, and Diane Pierret 3 October 13, 2011 1 Phelix spot and futures returns: descriptive statistics

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

VAR models with non-gaussian shocks

VAR models with non-gaussian shocks VAR models with non-gaussian shocks Ching-Wai (Jeremy) Chiu Haroon Mumtaz Gabor Pinter September 27, 2016 Motivation and aims Density of the residuals from a recursive VAR(13) (1960m1-2015m6) Motivation

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

Switching Regime Estimation

Switching Regime Estimation Switching Regime Estimation Series de Tiempo BIrkbeck March 2013 Martin Sola (FE) Markov Switching models 01/13 1 / 52 The economy (the time series) often behaves very different in periods such as booms

More information

Discussion of Predictive Density Combinations with Dynamic Learning for Large Data Sets in Economics and Finance

Discussion of Predictive Density Combinations with Dynamic Learning for Large Data Sets in Economics and Finance Discussion of Predictive Density Combinations with Dynamic Learning for Large Data Sets in Economics and Finance by Casarin, Grassi, Ravazzolo, Herman K. van Dijk Dimitris Korobilis University of Essex,

More information

STAT 425: Introduction to Bayesian Analysis

STAT 425: Introduction to Bayesian Analysis STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 2) Fall 2017 1 / 19 Part 2: Markov chain Monte

More information

Markov Switching Regular Vine Copulas

Markov Switching Regular Vine Copulas Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS057) p.5304 Markov Switching Regular Vine Copulas Stöber, Jakob and Czado, Claudia Lehrstuhl für Mathematische Statistik,

More information

Modeling Ultra-High-Frequency Multivariate Financial Data by Monte Carlo Simulation Methods

Modeling Ultra-High-Frequency Multivariate Financial Data by Monte Carlo Simulation Methods Outline Modeling Ultra-High-Frequency Multivariate Financial Data by Monte Carlo Simulation Methods Ph.D. Student: Supervisor: Marco Minozzo Dipartimento di Scienze Economiche Università degli Studi di

More information

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January

More information

The profit function system with output- and input- specific technical efficiency

The profit function system with output- and input- specific technical efficiency The profit function system with output- and input- specific technical efficiency Mike G. Tsionas December 19, 2016 Abstract In a recent paper Kumbhakar and Lai (2016) proposed an output-oriented non-radial

More information

Revisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data

Revisiting linear and non-linear methodologies for time series prediction - application to ESTSP 08 competition data Revisiting linear and non-linear methodologies for time series - application to ESTSP 08 competition data Madalina Olteanu Universite Paris 1 - SAMOS CES 90 Rue de Tolbiac, 75013 Paris - France Abstract.

More information

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity

Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity International Mathematics and Mathematical Sciences Volume 2011, Article ID 249564, 7 pages doi:10.1155/2011/249564 Research Article The Laplace Likelihood Ratio Test for Heteroscedasticity J. Martin van

More information

Lecture Notes based on Koop (2003) Bayesian Econometrics

Lecture Notes based on Koop (2003) Bayesian Econometrics Lecture Notes based on Koop (2003) Bayesian Econometrics A.Colin Cameron University of California - Davis November 15, 2005 1. CH.1: Introduction The concepts below are the essential concepts used throughout

More information

D I S C U S S I O N P A P E R 2009/61. On marginal likelihood computation in change-point models. Luc Bauwens and Jeroen V.K.

D I S C U S S I O N P A P E R 2009/61. On marginal likelihood computation in change-point models. Luc Bauwens and Jeroen V.K. 2009/61 On marginal likelihood computation in change-point models Luc Bauwens and Jeroen V.K. Rombouts D I S C U S S I O N P A P E R Center for Operations Research and Econometrics Voie du Roman Pays,

More information

ISSN Article. Selection Criteria in Regime Switching Conditional Volatility Models

ISSN Article. Selection Criteria in Regime Switching Conditional Volatility Models Econometrics 2015, 3, 289-316; doi:10.3390/econometrics3020289 OPEN ACCESS econometrics ISSN 2225-1146 www.mdpi.com/journal/econometrics Article Selection Criteria in Regime Switching Conditional Volatility

More information

Partially Censored Posterior for Robust and Efficient Risk Evaluation.

Partially Censored Posterior for Robust and Efficient Risk Evaluation. Preliminary Draft. Please do not cite, circulate or quote without the authors permission Partially Censored Posterior for Robust and Efficient Risk Evaluation. Agnieszka Borowska (a,b), Lennart Hoogerheide

More information

Inference in VARs with Conditional Heteroskedasticity of Unknown Form

Inference in VARs with Conditional Heteroskedasticity of Unknown Form Inference in VARs with Conditional Heteroskedasticity of Unknown Form Ralf Brüggemann a Carsten Jentsch b Carsten Trenkler c University of Konstanz University of Mannheim University of Mannheim IAB Nuremberg

More information

eqr094: Hierarchical MCMC for Bayesian System Reliability

eqr094: Hierarchical MCMC for Bayesian System Reliability eqr094: Hierarchical MCMC for Bayesian System Reliability Alyson G. Wilson Statistical Sciences Group, Los Alamos National Laboratory P.O. Box 1663, MS F600 Los Alamos, NM 87545 USA Phone: 505-667-9167

More information

Heteroskedasticity in Time Series

Heteroskedasticity in Time Series Heteroskedasticity in Time Series Figure: Time Series of Daily NYSE Returns. 206 / 285 Key Fact 1: Stock Returns are Approximately Serially Uncorrelated Figure: Correlogram of Daily Stock Market Returns.

More information

A Comparison of Bayesian Model Selection based on MCMC with an application to GARCH-Type Models

A Comparison of Bayesian Model Selection based on MCMC with an application to GARCH-Type Models A Comparison of Bayesian Model Selection based on MCMC with an application to GARCH-Type Models Tatiana Miazhynskaia Sylvia Frühwirth-Schnatter Georg Dorffner Report No. 83 November, 23 November, 23 SFB

More information

Bayesian Inference for DSGE Models. Lawrence J. Christiano

Bayesian Inference for DSGE Models. Lawrence J. Christiano Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.

More information

Outlier detection in ARIMA and seasonal ARIMA models by. Bayesian Information Type Criteria

Outlier detection in ARIMA and seasonal ARIMA models by. Bayesian Information Type Criteria Outlier detection in ARIMA and seasonal ARIMA models by Bayesian Information Type Criteria Pedro Galeano and Daniel Peña Departamento de Estadística Universidad Carlos III de Madrid 1 Introduction The

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis Sílvia Gonçalves and Benoit Perron Département de sciences économiques,

More information

Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling

Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling 1 / 27 Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling Melih Kandemir Özyeğin University, İstanbul, Turkey 2 / 27 Monte Carlo Integration The big question : Evaluate E p(z) [f(z)]

More information

GARCH processes probabilistic properties (Part 1)

GARCH processes probabilistic properties (Part 1) GARCH processes probabilistic properties (Part 1) Alexander Lindner Centre of Mathematical Sciences Technical University of Munich D 85747 Garching Germany lindner@ma.tum.de http://www-m1.ma.tum.de/m4/pers/lindner/

More information

Modelling and forecasting of offshore wind power fluctuations with Markov-Switching models

Modelling and forecasting of offshore wind power fluctuations with Markov-Switching models Modelling and forecasting of offshore wind power fluctuations with Markov-Switching models 02433 - Hidden Markov Models Pierre-Julien Trombe, Martin Wæver Pedersen, Henrik Madsen Course week 10 MWP, compiled

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Marginal Specifications and a Gaussian Copula Estimation

Marginal Specifications and a Gaussian Copula Estimation Marginal Specifications and a Gaussian Copula Estimation Kazim Azam Abstract Multivariate analysis involving random variables of different type like count, continuous or mixture of both is frequently required

More information

Eco517 Fall 2013 C. Sims MCMC. October 8, 2013

Eco517 Fall 2013 C. Sims MCMC. October 8, 2013 Eco517 Fall 2013 C. Sims MCMC October 8, 2013 c 2013 by Christopher A. Sims. This document may be reproduced for educational and research purposes, so long as the copies contain this notice and are retained

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition. Christian P. Robert The Bayesian Choice From Decision-Theoretic Foundations to Computational Implementation Second Edition With 23 Illustrations ^Springer" Contents Preface to the Second Edition Preface

More information

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano

Dynamic Factor Models and Factor Augmented Vector Autoregressions. Lawrence J. Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Lawrence J Christiano Dynamic Factor Models and Factor Augmented Vector Autoregressions Problem: the time series dimension of data is relatively

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

GARCH Models Estimation and Inference

GARCH Models Estimation and Inference GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in

More information

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous

More information

Online Appendix for: A Bounded Model of Time Variation in Trend Inflation, NAIRU and the Phillips Curve

Online Appendix for: A Bounded Model of Time Variation in Trend Inflation, NAIRU and the Phillips Curve Online Appendix for: A Bounded Model of Time Variation in Trend Inflation, NAIRU and the Phillips Curve Joshua CC Chan Australian National University Gary Koop University of Strathclyde Simon M Potter

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Simulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris

Simulation of truncated normal variables. Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris Simulation of truncated normal variables Christian P. Robert LSTA, Université Pierre et Marie Curie, Paris Abstract arxiv:0907.4010v1 [stat.co] 23 Jul 2009 We provide in this paper simulation algorithms

More information

Label Switching and Its Simple Solutions for Frequentist Mixture Models

Label Switching and Its Simple Solutions for Frequentist Mixture Models Label Switching and Its Simple Solutions for Frequentist Mixture Models Weixin Yao Department of Statistics, Kansas State University, Manhattan, Kansas 66506, U.S.A. wxyao@ksu.edu Abstract The label switching

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Working Papers in Econometrics and Applied Statistics

Working Papers in Econometrics and Applied Statistics T h e U n i v e r s i t y o f NEW ENGLAND Working Papers in Econometrics and Applied Statistics Finite Sample Inference in the SUR Model Duangkamon Chotikapanich and William E. Griffiths No. 03 - April

More information

Cross-sectional space-time modeling using ARNN(p, n) processes

Cross-sectional space-time modeling using ARNN(p, n) processes Cross-sectional space-time modeling using ARNN(p, n) processes W. Polasek K. Kakamu September, 006 Abstract We suggest a new class of cross-sectional space-time models based on local AR models and nearest

More information

Doing Bayesian Integrals

Doing Bayesian Integrals ASTR509-13 Doing Bayesian Integrals The Reverend Thomas Bayes (c.1702 1761) Philosopher, theologian, mathematician Presbyterian (non-conformist) minister Tunbridge Wells, UK Elected FRS, perhaps due to

More information

Using all observations when forecasting under structural breaks

Using all observations when forecasting under structural breaks Using all observations when forecasting under structural breaks Stanislav Anatolyev New Economic School Victor Kitov Moscow State University December 2007 Abstract We extend the idea of the trade-off window

More information

Introduction to ARMA and GARCH processes

Introduction to ARMA and GARCH processes Introduction to ARMA and GARCH processes Fulvio Corsi SNS Pisa 3 March 2010 Fulvio Corsi Introduction to ARMA () and GARCH processes SNS Pisa 3 March 2010 1 / 24 Stationarity Strict stationarity: (X 1,

More information

Study Notes on the Latent Dirichlet Allocation

Study Notes on the Latent Dirichlet Allocation Study Notes on the Latent Dirichlet Allocation Xugang Ye 1. Model Framework A word is an element of dictionary {1,,}. A document is represented by a sequence of words: =(,, ), {1,,}. A corpus is a collection

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Kobe University Repository : Kernel

Kobe University Repository : Kernel Kobe University Repository : Kernel タイトル Title 著者 Author(s) 掲載誌 巻号 ページ Citation 刊行日 Issue date 資源タイプ Resource Type 版区分 Resource Version 権利 Rights DOI URL Note on the Sampling Distribution for the Metropolis-

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

Bayesian inference for factor scores

Bayesian inference for factor scores Bayesian inference for factor scores Murray Aitkin and Irit Aitkin School of Mathematics and Statistics University of Newcastle UK October, 3 Abstract Bayesian inference for the parameters of the factor

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Advanced Statistics and Data Mining Summer School

More information

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY

Time Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7

More information

Lecture 6: Univariate Volatility Modelling: ARCH and GARCH Models

Lecture 6: Univariate Volatility Modelling: ARCH and GARCH Models Lecture 6: Univariate Volatility Modelling: ARCH and GARCH Models Prof. Massimo Guidolin 019 Financial Econometrics Winter/Spring 018 Overview ARCH models and their limitations Generalized ARCH models

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

Bayesian Inference. Chapter 4: Regression and Hierarchical Models Bayesian Inference Chapter 4: Regression and Hierarchical Models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Reminder of some Markov Chain properties:

Reminder of some Markov Chain properties: Reminder of some Markov Chain properties: 1. a transition from one state to another occurs probabilistically 2. only state that matters is where you currently are (i.e. given present, future is independent

More information

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering Lecturer: Nikolay Atanasov: natanasov@ucsd.edu Teaching Assistants: Siwei Guo: s9guo@eng.ucsd.edu Anwesan Pal:

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information