ADVANCED FINANCIAL ECONOMETRICS PROF. MASSIMO GUIDOLIN
|
|
- Kristin Wilkerson
- 6 years ago
- Views:
Transcription
1 Massimo Guidolin Dept. of Finance ADVANCED FINANCIAL ECONOMETRICS PROF. MASSIMO GUIDOLIN a.a. 14/15 p. 1 LECTURE 3: REVIEW OF BASIC ESTIMATION METHODS: GMM AND OTHER EXTREMUM ESTIMATORS; SIMULATION-BASED ESTIMATORS
2 OVERVIEW 1) Definition of Generalized Method of Moment Estimators: Unconditional and conditional moment restrictions 2) Extremum estimators: asymptotic normality 3) Efficient GMM tests 4) Goodness of fit tests in GMM (extremum estimators) 5) Sequential, partial estimation 6) Simulation-based estimators (brief introduction): SMM 7) Simulation-based estimators (brief introduction): MCMC 8) Some notes on SNP auxiliary models a.a. 14/15 p. 2
3 THE GMM: DEFINITION In modern econometrics, GMM is the leading case of limited information estimator We want to estimate a parameter vector θ 0 in the admissible parameter space Φ R K and is to be based on a sample, a sub-vector of the complete set of variables appearing in a DAPM. The restrictions on the distribution of to be used in estimating θ 0 are summarized as a set of restrictions on the moments of functions of z t The moment restrictions may be conditional or unconditional In the unconditional case, is satisfied uniquely by θ 0, where h is an M-dim. vector, M K h may define standard central or noncentral moments of returns, the orthogonality of forecast errors to variables, etc. a.a. 14/15 p. 3
4 THE GMM: UNCONDITIONAL MOMENTS Definition 4 [Just identified GMM, K = M]: Because the function defined as H 0 (θ) E[h(z t ; θ)] satisfies H 0 (θ 0 ) = 0, a natural estimation is to replace H 0 by its sample counterpart, and choose θ T that sets H T (θ T ) = 0 M If H T converges to its population counterpart as T gets large by a LLN, H T (θ) H 0 (θ), for all θ Φ, then under regularity conditions we should expect that θ T θ 0 Next suppose that M > K: then there is in general no unique way of solving for the K unknowns using the M equations H T (θ) = 0 Definition 5 [Over-identified GMM, K < M]: Let {a T : T 1} a.a. 14/15 p. 4
5 THE GMM: UNCONDITIONAL MOMENTS be a sequence of s M matrices of rank s, K s M, and consider the function. Then the GMM criterion function is a quadratic form:. denotes the Euclidean norm, x = (x x) 1/2 Quadratic form follows from GMM estimators are optimal, in the sense of being asymptotically most efficient, when they can be represented as the solution to (*) for appropriate choice of W T Let s now focus on stronger, conditional moment restrictions: W T To construct an estimator of θ 0, we choose K sample moment a.a. 14/15 p. 5 (*)
6 THE GMM: CONDITIONAL MOMENTS equations in the K unknowns θ Because h(z t+n ; θ 0 ) is orthogonal to any random variable in I t, we have much more flexibility in choosing these moment conditions We can afford to introduce a class of K M full-rank instrument matrices A t with elements in I t. For any A t, we use (**) A family of GMM estimators indexed by A t A, θ A T is the solutions to the corresponding sample moment equations: If the sample mean of A t h(z t+n ; θ) converges to its population counterpart, for all θ Φ, and A t and h are chosen so that θ 0 is the unique element of Φ satisfying (**), then we might reasonably expect θ A T to converge to θ 0 as T a.a. 14/15 p. 6
7 THE GMM: CONDITIONAL MOMENTS The large sample distribution of θ A T depends, in general, on the choice of A t The conditional GMM estimator, is not the extreme value of a criterion function; it is instead the solution to K moment equations in K unknowns, and θ T solves the sample counterpart of these equations The class of θ A T conditional GMM estimators offers more flexibility in choosing the weights on h; as a result, they are often more efficient than their unconditional counterparts This means that they allow us to exploit more information about the distribution of z t than (*) in the estimation of θ 0 Example 2 [Linear projections, aka regressions ] If we define then by construction, δ 0 satisfies a.a. 14/15 p. 7
8 THE GMM: CONDITIONAL MOMENTS Why is this notable/remarkable? Because the celebrated orthogonal projection theorem tells us that the unique solution to a standard projection (== population regression) problem is given by the δ 0 R K satisfying Notice that this means that projection/regressions are just a special case of GMM based on unconditional moment restrictions Example 1 (cont d) [CIR model] Because the conditional mean E[r t+ r t ] for any finite interval is given by then it is natural to base GMM estimation on the condition and instruments can then be set to any g(r t ) so that estimation is based on a.a. 14/15 p. 8
9 THE GMM: CONDITIONAL MOMENTS Example 2 [Two-factor SV model] Assume that v is SV and therefore unobserved: The interest rate is no longer Markov given its past history Although the variance of r t conditioned on r t 1 is not known in closed form, nor is the form of density of r t conditioned on J-histories of r, is correct Hence there are two possible estimation strategies: (i) approximate the log-likelihood function; (ii) GMM based on either unconditional or conditional restrictions from Example 1 However, this GMM estimator ignores entirely the structure of the volatility process; thus, not only are we unable to recover any information about the parameters of volatility, but knowledge of the functional form of the volatility equation is ignored a.a. 14/15 p. 9
10 EXTREMUM ESTIMATORS Substantially more information about f(r t r t 1 ; θ 0 ) can be used in estimation, but to accomplish this we have to extend the GMM estimation strategy to allow for unobserved state variables Under some technical conditions on the sequence of weight matrices {W T }, one can show that the GMM for θ 0 is (strongly) consistent for θ Definition 6 [Extremum Estimators] EEs are estimators obtained by either maximizing or minimizing a criterion function over the admissible parameter space Obviously, the specific properties of EEs will depend on the adopted criterion function In fact, ML, QML, GMM (unconditional and conditional), and regression (LP) estimators are all cases of EEs As already mentioned in lecture 2, EEs share a common and a.a. 14/15 p. 10
11 EXTREMUM ESTIMATORS: ASYMPTOTIC NORMALITY important statistical property, they have an asymptotically normal distribution Suppose that θ T is strongly consistent for θ 0. To show asymptotic normality of θ T, we focus on FOCs for the maximization or minimization of Q T, the sample mean of the function D(z t ; θ): Thus, the function D(z t ; θ) representing the FOCs for Q 0, is Hansen (1982, ECMA) has argued that in some ways, EEs are a.a. 14/15 p. 11 w/
12 EXTREMUM ESTIMATORS: ASYMPTOTIC NORMALITY just generalizations of GMM to generic instrument matrix A and to generic restrictions often written as scores or orthogonality conditions One of the common mis-conceptions at this point is that it is a simple application of the classical CLT may deliver asymptotic normality However, the classical CLT is based on the strong assumption of IIDness of z t : in FEc, the assumption of independence is typically too strong It rules out persistence in the state variables and time-varying conditional volatilities Also the assumption that {X t } is a stationary and ergodic time series, which is much weaker than an IID assumption, is not sufficient to establish a CLT, N The problem is that an ergodic time series can be highly a.a. 14/15 p. 12
13 EXTREMUM ESTIMATORS: ASYMPTOTIC NORMALITY persistent, so that the X t and X s, for s t are too highly correlated for to converge to a normal A weaker assumption than IIDness that however delivers a CLT is instead that z t follows a martingale difference sequence, i.e. (referred to a generic time series {X t }) that with probability one, i.e., X t is mean -independent If {X t } is also stationary and ergodic and E[X 2 1] is finite (exists), then a CLT applies to Formally stated, asymptotic normality consists of: However, the specific structure of Σ 0 will depend on the estimator under consideration a.a. 14/15 p. 13
14 EXTREMUM ESTIMATORS: ASYMPTOTIC NORMALITY 1 Maximum Likelihood Estimator (MLE): Because and A t = I K The second equality derives from the fact that from the MLE FOCs one has Differentiating under the integral sign and using the chain rule: a.a. 14/15 p. 14
15 EXTREMUM ESTIMATORS: ASYMPTOTIC NORMALITY Putting everything together, we have: In actual implementations, the asymptotic covariance can be estimated either as the inverse of the sample mean of the outer product of the likelihood scores, or as minus the inverse of the sample mean of the secondderivative matrix evaluated at b T ML Asymptotically, the two objects are identical, but in small a.a. 14/15 p. 15
16 EXTREMUM ESTIMATORS: ASYMPTOTIC NORMALITY these do not have to be the same The asymptotic covariance matrix of b T ML is the Cramer-Rao lower bound, the inverse of the Hessian matrix Even though the MLE may be biased in small samples, as T gets large, it is the most efficient estimator in the sense of having the smallest asymptotic covariance matrix among all consistent estimators of β 0 2 GMM: Because and assuming that A T converges in probability to A 0 = d 0 W 0, we have: If the probability limit of the distance matrix defining the GMM a.a. 14/15 p. 16
17 EXTREMUM ESTIMATORS: ASYMPTOTIC NORMALITY criterion function is chosen to be W 0 = Σ -1 0, then where d 0 = 3 QMLE: It is essentially the same as the MLE case, but now and where a.a. 14/15 p. 17
18 EFFICIENT GMM ESTIMATOR Under QMLE, Σ 0 and d 0 are different from those under MLE In particular, no further simplification obtains so that In many circumstances, a researcher estimating a GMM will have considerable latitude in choosing either A 0 or h(z t, θ), or both Therefore, a natural question is: Which is the optimal GMM estimator among all admissible estimators? A natural answer is simply: the most efficient in the sense of having the smallest asymptotic covariance matrix among all estimators that exploit the same information about the distribution of a.a. 14/15 p. 18
19 EFFICIENT GMM ESTIMATOR Notice that the choice of a weighting matrix A t makes sense iff M > K, when there are more moment conditions than parameters In the case of the GMM fixed on unconditional moment restrictions (sometimes called fixed ), it is easy to show that To relate this observation back to the standard GMM criterion function expressed as a quadratic form in H T (θ) because A 0 =d 0 W 0, the optimal GMM estimator is obtained by setting W 0 = Σ -1 0 As Σ 0 is the asymptotic covariance matrix of the sample moment H T (θ 0 ), this choice of W 0 gives the most weight to those moment a.a. 14/15 p. 19
20 EFFICIENT GMM ESTIMATOR conditions that are most precisely estimated in the sense of having a small (asymptotic) variance In the case of the conditional moment restriction-based GMM, the optimal choice of the weights becomes time-varying: and plugging into A t =d t W t gives: Notice that MLE represents a special case of optimal GMM in which the optimality derives from the choice of the moment conditions(s) and not of the weighting matrix (i.e., A t = I K t): a.a. 14/15 p. 20
21 GMM GOODNESS OF FIT TESTS In minimizing Q T (θ) over the choice of θ Θ, the GMM is chosen to set K linear combinations of the M sample moment conditions H T to zero (the K first-order conditions): Yet, if the model is correctly specified, all M sample moment equations H T (θ T ) should be close to zero. Therefore one can construct a goodness-of-fit test of the model by examining whether linear combinations of H T (θ T ) that are not set to zero in estimation are in fact close to 0 It turns out that the min value of the GMM criterion function, scaled by sample size, TQ T (θ T ) is a goodness-of-fit test based Under the null hypothesis that the model is correctly specified, TQ T (θ T ) is asymptotically distributed as a chisquare distribution with M K degrees of freedom a.a. 14/15 p. 21
22 SEQUENTIAL ESTIMATORS (ANOTHER CASE OF QML) Notice that which is a weighted average of sample moment conditions With complex estimation problems, it is often convenient to break down the problem in multiple steps to exploit computational advantages For instance, given, first the sub-vector θ 0 is estimated using a subset of the available moment equations, and then the sub-vector 0 is estimated in a second stage using additional moment conditions that also depend on θ 0 An important issue with sequential estimation is whether the asymptotic distribution of λ T is affected by first-stage estimation of θ 0 The answer is generally yes ; however, it is not always yes and, fortunately, there is a simple way to check... a.a. 14/15 p. 22
23 SEQUENTIAL ESTIMATORS (ANOTHER CASE OF QML) Consider the GMM problem where where θ 0 has dimension K 1, h 1 is a M 1 -vector function, λ 0 has dimension K 2, and h 2 is a M 2 -vector function A first-stage estimator θ T of θ 0 is obtained by solving for some K 1 M 1 matrix A 1T with probability limit A 10 Similarly, the second-stage estimator λ T of λ 0 is obtained as the solution to Taking mean-value expansions and solving for (θ T θ 0 ) and a.a. 14/15 p. 23 (**)
24 SEQUENTIAL ESTIMATORS (ANOTHER CASE OF QML) (λ T λ 0 ) gives suitable intermediate values Mean value expansion: The asymptotic distribution of λ T that solves the system is the same as that of the distribution of the λ T that solves (**) iff: Under this condition, the correct limiting distribution of λ T is a.a. 14/15 p. 24
25 SEQUENTIAL ESTIMATORS (ANOTHER CASE OF QML) obtained by treating θ 0 as if it were known: there is no effect on the (limiting) distribution of λ T of pre-estimation of θ 0 in a first stage The condition is a useful test for whether sequential estimation affects the inference in the second stage Example 3 [Two-stage estimation of GARCH models] Consider r t+1 = + 2 t+1 t+1 t+1 IID N(0,1) 2 t+1 = + (r t - ) t Typically, estimation of [ ] is performed by MLE This is consistent with this lecture as we know that MLE is just a special case of GMM However, it is not infrequent to see papers in which first is estimated by OLS and in a second step, using e t+1 = r t+1 -, a.a. 14/15 p. 25
26 SEQUENTIAL ESTIMATORS (ANOTHER CASE OF QML) then MLE is applied to the GARCH model to estimate [ ], as if were a known value Problem: Also should be estimated by GLS and not OLS, where GLS is equivalent to MLE in this simple case Because a quick look at the log-likelihood function reveals that the portion that contains [ ] shows that enters also this second component, i.e., clearly such a procedure is not equivalent to full-mle For several reasons, the implementation of GMM and MLE in the analysis of DAPMs may be computationally demanding, if not infeasible One such circumstance is when there are unobserved state a.a. 14/15 p. 26 variables, e.g., in the case of stochastic volatility:
27 SEQUENTIAL ESTIMATORS (ANOTHER CASE OF QML) Unobserved! Because of the unobservability of volatility, discretely sampled returns, {r t }, are not Markov conditioned on their own history and the form of the conditional dstr. of r t is unknown Apart from a few special cases, the moments of r t, expressed as functions of the unknown parameters, are also unknown This problem may be rendered even more challenging by the presence of jumps in returns or volatility (At least) two solutions among estimation methods: (1) Simulated Method of Moments; (2) Monte Carlo Markov Chain (MCMC) methods Both estimators are applicable to DAPMs without latent a.a. 14/15 p. 27 variables, but they are most useful when latent variables (e.g.,
28 SIMULATED METHOD OF MOMENTS a.a. 14/15 p. 28 jumps) when these methods often dominate others, both in their tractability and efficiency SMM extends the GMM estimator to a class of DAPM for which moment restrictions do not have analytic representations in terms of observable variables and unknown parameters SMM are not just GMM applied to simulated data: the reason is that functions of the current value of the simulated state depend on the unknown parameter vector: _ through the structure of the model (as in any GMM problem) and _ indirectly, through generation of data by simulation The feedback effect of the latter dependence on the transition law of the simulated state implies that the (first-momentcontinuity) conditions used to establish the uniform conver-
29 SIMULATED METHOD OF MOMENTS gence of sample to population criterion functions in GMM fail Assume that a given R N -valued state process {Y t } t=1 is generated by the difference equation where the parameter vector β 0 Θ R K and { t } is an IID sequence of R p -valued random variables The number of shocks, p, need not equal the dimension of the state vector, N Letting Z t (Y t, Y t 1,..., Y t l+1 ) for some positive integer l <, the estimation of β 0 is based on the moments of the observation function g(z t, β), g : R N Θ R M Moments of the observed series are calculated as sample moments of the observed g* t g(z t, β) The function H may be known or determined implicitly by the a.a. 14/15 p. 29
30 SIMULATED METHOD OF MOMENTS numerical solution of a discrete-time model or by a discretetime approximation of a continuous-time model Example 2 [Cont d] Consider A standard (Euler) discretization scheme, gives Bivariate N(0, I 2 ) If the function mapping β into E[g(Z t,β)] is known and independent of t, the GMM estimator is applicable a.a. 14/15 p. 30 h(z t,β) h(z t,β)
31 SIMULATED METHOD OF MOMENTS Unfortunately, the form of g(z t,β) is known only for special cases, like when κ = 0 or when The SMM circumvents the requirement that g(z t, β) is known by making the much weaker assumption that we have access to an R p -valued sequence These random variables are identical in distribution to, and independent of, { t } The idea is that for any R N -valued initial point Y 0 and any parameter vector β Θ, the simulated state process {Y β t} can be constructed inductively by letting Y β t = Y 0 and Y β t+1 = H(Y β t, Likewise, the simulated observation process is constructed by g β t = g(z β t, β), where The SMM of β 0 is then the parameter vector b T that best matches the sample moments of the actual and simulated observation processes, {g* t } and {g b T t } a.a. 14/15 p. 31,β)
32 SIMULATED METHOD OF MOMENTS Let T: Z Z define the simulation sample size T (T) generated for a given sample size T of actual observations, where T (T) as T For any β, let denote the difference in sample moments. If {g* t } and {g bt t} satisfy a LLN and under adequate identification conditions, In the over-identified case, one also selects a sequence W = {W T } of M M positive-semidefinite matrices such that b T is: Although mechanically this is the case, SMM does not simply extend the GMM by replacing the population moment E[g(Z t,β)] with its sample counterpart, calculated with simulated data a.a. 14/15 p. 32
33 SIMULATED METHOD OF MOMENTS What are the main differences between GMM and SMM? 1 The key difference is the parameter dependency of the simulated series {g β t} because this depends on β not only directly, but indirectly through the dependence of the entire past history of the simulated process {Y β t} on β 2 Unless the environment is IID, simulations require initial conditions for the state Y t : even if the transition function of the Markov process {Y t } is stationary, the simulated process {Y β t} is generally non-stationary This derives from the fact that the initial simulated state Y β 0 is typically not drawn from the ergodic distribution of the process Practically, one can leave out an initial portion of the simulated state variables {Y β t} to mitigate transient effects What are the key statistical properties of SMM? As usual, a.a. 14/15 p. 33
34 SIMULATED METHOD OF MOMENTS they are (weakly) consistent and asymptotically normal Weak consistency: plim b T = β 0 The proof of consistency is interesting because strong consistency is difficult to obtain Duffie and Singleton (1993, ECMA) give conditions on H that guarantee that the compounding effects of simulation on the properties of estimators damp out over time, and use these conditions to prove strong consistency However, their damping conditions are not satisfied by many of diffusion DAPM Because the simulated state process is usually not initialized with a draw from its ergodic distribution, one needs a condition that allows the use of an arbitrary initial state, knowing that the state process converges rapidly to its stationary distribution Such a condition is geometric ergodicity: a condition ensuring a.a. 14/15 p. 34
35 SIMULATED METHOD OF MOMENTS that the simulated state process satisfies a LLN with an asymptotic distribution that is invariant to initial conditions Let P t x denote the t-step transition probability for a timehomogeneous Markov process {X t }, i.e., the distribution of X t given the initial point X 0 = x; {X t } is ρ-ergodic, for some ρ (0, 1], if there is a measure π s.t., for every initial point x, Total variation norm If {X t } is ρ-ergodic for ρ < 1, then {X t } is geometrically ergodic In calculating asymptotic distributions, geometric ergodicity can substitute for stationarity since it means that the process converges geometrically to its stationary distribution As usual, the condition is imposed on {Y β t} because it delivers a (S)LLN for the simulated series {g β t} a.a. 14/15 p. 35 Ergodic dstr.
36 SIMULATED METHOD OF MOMENTS Because of the (S)LLN, the criterion function Q T (β) converges to the asymptotic criterion function Q 0 : Θ R defined by Q 0 (β) = G (β) W 0 G (β) almost surely What is the optimal sequence W = {W T } of M M SPD matrices? Similarly to GMM, this happens when W T W 0 = Σ 0-1 Σ 0 is a function solely of the moments of {g* t } and hence the data, and not of β or on the moments of the simulated process {g β t} When it comes to asymptotic normality, one needs to justify the assumption of uniform continuity of the observations as a function of β: with simulations, a perturbation of β affects not only the current observation, but also transitions btw past states, a dependence that compounds over time Formally, one has: a.a. 14/15 p. 36
37 SIMULATED METHOD OF MOMENTS N(0, ) as As 0, AsyVar(b T ) (D 0 Σ -1 0 D 0 ) -1, the covariance matrix obtained when an analytic expression for E[g(Z t,β)] is known Knowledge of E[g(Z t,β)] increases efficiency, but if the simulated sample size T (T) is chosen to be large relative to the size T of the sample, then there is essentially no loss in efficiency Typically, in applications of the SMM to asset pricing, it is assumed that T is large and 0 So far, we stated that one has freedom to pick moments that enter the definition of and that at best SMM may require an optimal choice of the weighting matrix, W T W 0 = Σ -1 0 a.a. 14/15 p. 37
38 KLIC-BASED MOMENT SELECTION: EMM In principle, one would want to choose moment equations that capture some of the known features of the data, such as persistence, conditional heteroskedasticity, and non-normality Gallant and Tauchen (1996, ECT) proposed a clever application of SMM that allows one to easily capture these features Their approach is based on the concept of minimization of the Kullback-Leibler information criterion (KLIC): where DGP process and is the density of the actual but unknown is just an approximation The KLIC can be interpreted as a measure of our ignorance about the true structure of the DGP However, minimizing KLIC for given p has clear implications for the denominator: one should maximize as in MLE a.a. 14/15 p. 38
39 KLIC-BASED MOMENT SELECTION: EMM Gallant and Tauchen s idea is simple and yet powerful: let Y t denote the observed subvector of the state process Y t What s the difference? In example 2, Y t = r t and Y t = (r t, v t ) because stochastic volatility is not observable Let f (Y t Y t-1 ; δ) be a conditional density function of the data that captures parametrically the features of the data that one is interested in representing by {g t } Applying ML to gives δ T s.t. There is no presumption that the density f is the true conditional density of Y t or that δ T is a consistent estimator of any of the parameters of the true DGP process for Y t δ T is a consistent estimator of δ 0 that minimizes the KLIC At this point, GT s proposal is simple: having chosen f and estimated δ T by MLE, let the score of this log-likelihood be the a.a. 14/15 p. 39
40 KLIC-BASED MOMENT SELECTION: EMM vector of moments used to estimate β 0 : and the corresponding function for simulated data is In simulating Y β t, it is generally necessary to simulate the entire vector Y β t and then select the Y β t in the moment conditions: The sample moments entering the SMM criterion depend only on simulated data a.a. 14/15 p. 40
41 KLIC-BASED MOMENT SELECTION: EMM Of course simulation is common but not needed for this KLICdriven method to work, i.e., this can be applied to GMM too However, you know already that when in a GMM you set the (unconditional) moment restrictions to equal the scores, then MLE obtains (as well as the Cramer-Rao lower bound) One wonders about the efficiency of SMM when the moment conditions are picked this way Gallant and Tauchen write about Efficient Method of Moments when the auxiliary model is picked in a special way (see below, GT s seminonparametric framework) Several studies have examined the small-sample properties of the SMM and compared them to the properties of standard GMM and (when feasible) ML estimators Properties of the SMM depend on the choice of auxiliary model, number of moments, and sample size a.a. 14/15 p. 41
42 KLIC-BASED MOMENT SELECTION: EMM Chumacero (1997, SNDE): SMM is more efficient and often shows less bias than the GMM; however, tests of overidentifying restrictions using SMM tend to reject the models too often Andersen, Chung and Sorensen (1999, JoE): SMM performed well vs. GMM based on less systematic choice of instruments For their models and parameters the overall goodness-of-fit chisquare statistics from simulated moments led to reliable inference Example 4 [Matching simple features of the data, model-free case] If Y t is an observed scalar process and we care for firstorder serial correlation and conditional heteroskedasticity that depends on lagged squared projection errors, then set The FOCs vs. a 0, 0 and 1 are: a.a. 14/15 p. 42
43 KLIC-BASED MOMENT SELECTION: EMM In this case the components of G T ( ) are: An alternative estimation strategy for diffusion models, including models with latent state variables, is the method of Markov chain Monte Carlo (MCMC) Its conceptual foundations draw upon Bayesian theory MCMC generates estimates not just of the parameters of the model, but also of the latent volatility, jump times, and jump sizes, i.e., of latent variables Under approximate ML, QML, GMM, and SMM, a time series of values of latent variables (e.g., stochastic volatility) is usually computed after estimation using filtering methods a.a. 14/15 p. 43
44 MONTE CARLO MARKOV CHAIN ESTIMATION MCMC allows separation and quantification of estimation risk and model specification risk, and infrequent observations or missing data are easily accommodated The basic idea is to combine a prior distribution over the unknown parameters with the conditional density of the state vector to obtain a joint posterior distribution of the parameters and the state conditional on the observables From this joint posterior distribution the marginal posterior distributions of the states and parameters can be computed Mean or median, standard deviation, quantiles, and so on, of the posterior distribution of the parameters can be computed Denote as Θ the parameter vector of interest, X be a vector of (possibly latent) state variables, and Y denote the vector of observed asset prices or yields The MCMC algorithm constructs a Markov chain that converges a.a. 14/15 p. 44
45 MONTE CARLO MARKOV CHAIN ESTIMATION to the joint distribution p(θ, X Y ). From this distribution, one can determine both p(θ Y ) (which gives the parameter estimates) and p(x Y) (which provides estimates of the unobserved states) Key to this construction is the Clifford-Hammersley theorem: under a positivity condition, knowing p(θ X, Y) and p(x Θ, Y) is equivalent to knowing p(x, Θ Y) What gives the MCMC algorithm its traction is that the first two distributions are often much easier to characterize than the joint distribution p(x,θ Y ) When feasible to simulate from both densities, MCMC algorithm uses a Gibbs sampler: given realizations up to g 1, X g t is drawn from p(x t Θ g 1,Y ); Θ g is drawn from p(θ X g 1 t, Θ g-1,y) When direct sampling from the joint density p(x, Y ) is not a.a. 14/15 p. 45 feasible, researchers have replaced Gibbs sampling with
46 MONTE CARLO MARKOV CHAIN ESTIMATION Metropolis-Hastings sampling Suppose that simulation from the conditional density p(x Θ, Y) is not feasible and let The basic idea is to start with a distribution q(x g+1 X g ) that is known and from which samples can be easily drawn Then a single Gibbs sampling step is replaced by the two steps: Example 5 [Simple Brownian motion model] Consider that can be discretized into a.a. 14/15 p. 46
47 MONTE CARLO MARKOV CHAIN ESTIMATION Of course, the log-lik fnct of this model is known in closed form and the first line of action consists of estimating one MLE However, adopt a Bayesian approach under independent priors, with objects of interest and From Bayes rule: posteriors Likelihoods Typical choices of conjugate priors are p( S ) normal and p(σ 2 S) inverted gamma; is a normal, b/c The MCMC draws from known densities (i.e., it is a Gibbs sampler) priors a.a. 14/15 p. 47
48 GALLANT AND TAUCHEN S SNP MODEL A flexible family of auxiliary models is constructed as follows: let Y,t 1 be the linear projection of Y t onto L μ lags of Y; allow for ARCH -like errors by transforming the innovations in this autoregression by a matrix R Y,t 1 with elements that are linear in the absolute values of L r past values of Let z t the standardized Y t, and approximate the conditional density of Y t by Standard normal What is a Hermite polynomial? Hermite polynomial in z Coeffs. that can be made fnc. of time The Hermite expansion serves to introduce nonnormality in the conditional distribution of Y t by scaling the conditional a.a. 14/15 p. 48
49 GALLANT AND TAUCHEN S SNP MODEL normal density by a polynomial in lagged values of Y t SNP auxiliary models give another advantage: using Wald tests of individual moment conditions, you can test the null hypotheses that elements of mean score vector from the auxiliary model, G (β 0 ), are zero using the sample scores Rejection of the null that a particular mean score is zero would suggest that the DAPM does not adequately describe the features of the conditional distribution of Y governed by the associated parameter in the auxiliary model a.a. 14/15 p. 49
Follow links for Class Use and other Permissions. For more information send to:
COPYRIGH NOICE: Kenneth. Singleton: Empirical Dynamic Asset Pricing is published by Princeton University Press and copyrighted, 00, by Princeton University Press. All rights reserved. No part of this book
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationSTATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN
Massimo Guidolin Massimo.Guidolin@unibocconi.it Dept. of Finance STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN SECOND PART, LECTURE 2: MODES OF CONVERGENCE AND POINT ESTIMATION Lecture 2:
More informationMaximum Likelihood (ML) Estimation
Econometrics 2 Fall 2004 Maximum Likelihood (ML) Estimation Heino Bohn Nielsen 1of32 Outline of the Lecture (1) Introduction. (2) ML estimation defined. (3) ExampleI:Binomialtrials. (4) Example II: Linear
More informationLecture 2: Univariate Time Series
Lecture 2: Univariate Time Series Analysis: Conditional and Unconditional Densities, Stationarity, ARMA Processes Prof. Massimo Guidolin 20192 Financial Econometrics Spring/Winter 2017 Overview Motivation:
More informationGARCH Models Estimation and Inference
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in
More informationStat 5101 Lecture Notes
Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random
More informationThe Metropolis-Hastings Algorithm. June 8, 2012
The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings
More informationSTA 4273H: Statistical Machine Learning
STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate
More informationFollow links for Class Use and other Permissions. For more information send to:
COPYRIGH NOICE: Kenneth J. Singleton: Empirical Dynamic Asset Pricing is published by Princeton University Press and copyrighted, 00, by Princeton University Press. All rights reserved. No part of this
More informationAnswers and expectations
Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E
More informationGARCH Models Estimation and Inference
Università di Pavia GARCH Models Estimation and Inference Eduardo Rossi Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationAn estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic
Chapter 6 ESTIMATION OF THE LONG-RUN COVARIANCE MATRIX An estimate of the long-run covariance matrix, Ω, is necessary to calculate asymptotic standard errors for the OLS and linear IV estimators presented
More information1 Outline. 1. Motivation. 2. SUR model. 3. Simultaneous equations. 4. Estimation
1 Outline. 1. Motivation 2. SUR model 3. Simultaneous equations 4. Estimation 2 Motivation. In this chapter, we will study simultaneous systems of econometric equations. Systems of simultaneous equations
More informationGARCH Models Estimation and Inference. Eduardo Rossi University of Pavia
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia Likelihood function The procedure most often used in estimating θ 0 in ARCH models involves the maximization of a likelihood function
More informationVolatility. Gerald P. Dwyer. February Clemson University
Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use
More informationECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS
ECO 513 Fall 2009 C. Sims HIDDEN MARKOV CHAIN MODELS 1. THE CLASS OF MODELS y t {y s, s < t} p(y t θ t, {y s, s < t}) θ t = θ(s t ) P[S t = i S t 1 = j] = h ij. 2. WHAT S HANDY ABOUT IT Evaluating the
More informationEstimation of Dynamic Regression Models
University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.
More informationLecture 6: Univariate Volatility Modelling: ARCH and GARCH Models
Lecture 6: Univariate Volatility Modelling: ARCH and GARCH Models Prof. Massimo Guidolin 019 Financial Econometrics Winter/Spring 018 Overview ARCH models and their limitations Generalized ARCH models
More informationA Course on Advanced Econometrics
A Course on Advanced Econometrics Yongmiao Hong The Ernest S. Liu Professor of Economics & International Studies Cornell University Course Introduction: Modern economies are full of uncertainties and risk.
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY PREFACE xiii 1 Difference Equations 1.1. First-Order Difference Equations 1 1.2. pth-order Difference Equations 7
More informationNonlinear GMM. Eric Zivot. Winter, 2013
Nonlinear GMM Eric Zivot Winter, 2013 Nonlinear GMM estimation occurs when the GMM moment conditions g(w θ) arenonlinearfunctionsofthe model parameters θ The moment conditions g(w θ) may be nonlinear functions
More informationEconometric Analysis of Cross Section and Panel Data
Econometric Analysis of Cross Section and Panel Data Jeffrey M. Wooldridge / The MIT Press Cambridge, Massachusetts London, England Contents Preface Acknowledgments xvii xxiii I INTRODUCTION AND BACKGROUND
More informationECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications
ECONOMICS 7200 MODERN TIME SERIES ANALYSIS Econometric Theory and Applications Yongmiao Hong Department of Economics & Department of Statistical Sciences Cornell University Spring 2019 Time and uncertainty
More informationSTA 4273H: Sta-s-cal Machine Learning
STA 4273H: Sta-s-cal Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 2 In our
More informationTime Series Analysis. James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY
Time Series Analysis James D. Hamilton PRINCETON UNIVERSITY PRESS PRINCETON, NEW JERSEY & Contents PREFACE xiii 1 1.1. 1.2. Difference Equations First-Order Difference Equations 1 /?th-order Difference
More informationStock index returns density prediction using GARCH models: Frequentist or Bayesian estimation?
MPRA Munich Personal RePEc Archive Stock index returns density prediction using GARCH models: Frequentist or Bayesian estimation? Ardia, David; Lennart, Hoogerheide and Nienke, Corré aeris CAPITAL AG,
More informationMultivariate GARCH models.
Multivariate GARCH models. Financial market volatility moves together over time across assets and markets. Recognizing this commonality through a multivariate modeling framework leads to obvious gains
More informationDynamic Discrete Choice Structural Models in Empirical IO
Dynamic Discrete Choice Structural Models in Empirical IO Lecture 4: Euler Equations and Finite Dependence in Dynamic Discrete Choice Models Victor Aguirregabiria (University of Toronto) Carlos III, Madrid
More informationBayesian Estimation of DSGE Models 1 Chapter 3: A Crash Course in Bayesian Inference
1 The views expressed in this paper are those of the authors and do not necessarily reflect the views of the Federal Reserve Board of Governors or the Federal Reserve System. Bayesian Estimation of DSGE
More informationSTA414/2104 Statistical Methods for Machine Learning II
STA414/2104 Statistical Methods for Machine Learning II Murat A. Erdogdu & David Duvenaud Department of Computer Science Department of Statistical Sciences Lecture 3 Slide credits: Russ Salakhutdinov Announcements
More informationTime Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley
Time Series Models and Inference James L. Powell Department of Economics University of California, Berkeley Overview In contrast to the classical linear regression model, in which the components of the
More informationEconometrics of Panel Data
Econometrics of Panel Data Jakub Mućk Meeting # 6 Jakub Mućk Econometrics of Panel Data Meeting # 6 1 / 36 Outline 1 The First-Difference (FD) estimator 2 Dynamic panel data models 3 The Anderson and Hsiao
More informationIntroduction to Machine Learning CMU-10701
Introduction to Machine Learning CMU-10701 Markov Chain Monte Carlo Methods Barnabás Póczos & Aarti Singh Contents Markov Chain Monte Carlo Methods Goal & Motivation Sampling Rejection Importance Markov
More informationBayesian Inference and MCMC
Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the
More informationQuick Review on Linear Multiple Regression
Quick Review on Linear Multiple Regression Mei-Yuan Chen Department of Finance National Chung Hsing University March 6, 2007 Introduction for Conditional Mean Modeling Suppose random variables Y, X 1,
More informationLecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem
Lecture 5: Unit Roots, Cointegration and Error Correction Models The Spurious Regression Problem Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Stochastic vs. deterministic
More informationEstimation, Inference, and Hypothesis Testing
Chapter 2 Estimation, Inference, and Hypothesis Testing Note: The primary reference for these notes is Ch. 7 and 8 of Casella & Berger 2. This text may be challenging if new to this topic and Ch. 7 of
More informationA Bayesian perspective on GMM and IV
A Bayesian perspective on GMM and IV Christopher A. Sims Princeton University sims@princeton.edu November 26, 2013 What is a Bayesian perspective? A Bayesian perspective on scientific reporting views all
More informationBayesian Regression Linear and Logistic Regression
When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we
More informationBTRY 4090: Spring 2009 Theory of Statistics
BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)
More informationA Robust Approach to Estimating Production Functions: Replication of the ACF procedure
A Robust Approach to Estimating Production Functions: Replication of the ACF procedure Kyoo il Kim Michigan State University Yao Luo University of Toronto Yingjun Su IESR, Jinan University August 2018
More informationThe Linear Regression Model
The Linear Regression Model Carlo Favero Favero () The Linear Regression Model 1 / 67 OLS To illustrate how estimation can be performed to derive conditional expectations, consider the following general
More informationEcon 583 Final Exam Fall 2008
Econ 583 Final Exam Fall 2008 Eric Zivot December 11, 2008 Exam is due at 9:00 am in my office on Friday, December 12. 1 Maximum Likelihood Estimation and Asymptotic Theory Let X 1,...,X n be iid random
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationModeling conditional distributions with mixture models: Applications in finance and financial decision-making
Modeling conditional distributions with mixture models: Applications in finance and financial decision-making John Geweke University of Iowa, USA Journal of Applied Econometrics Invited Lecture Università
More informationBayesian Inference for DSGE Models. Lawrence J. Christiano
Bayesian Inference for DSGE Models Lawrence J. Christiano Outline State space-observer form. convenient for model estimation and many other things. Bayesian inference Bayes rule. Monte Carlo integation.
More information11. Further Issues in Using OLS with TS Data
11. Further Issues in Using OLS with TS Data With TS, including lags of the dependent variable often allow us to fit much better the variation in y Exact distribution theory is rarely available in TS applications,
More informationTAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω
ECO 513 Spring 2015 TAKEHOME FINAL EXAM (1) Suppose the univariate stochastic process y is ARMA(2,2) of the following form: y t = 1.6974y t 1.9604y t 2 + ε t 1.6628ε t 1 +.9216ε t 2, (1) where ε is i.i.d.
More informationMarkov chain Monte Carlo
1 / 26 Markov chain Monte Carlo Timothy Hanson 1 and Alejandro Jara 2 1 Division of Biostatistics, University of Minnesota, USA 2 Department of Statistics, Universidad de Concepción, Chile IAP-Workshop
More informationThis note introduces some key concepts in time series econometrics. First, we
INTRODUCTION TO TIME SERIES Econometrics 2 Heino Bohn Nielsen September, 2005 This note introduces some key concepts in time series econometrics. First, we present by means of examples some characteristic
More informationECON3327: Financial Econometrics, Spring 2016
ECON3327: Financial Econometrics, Spring 2016 Wooldridge, Introductory Econometrics (5th ed, 2012) Chapter 11: OLS with time series data Stationary and weakly dependent time series The notion of a stationary
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationBayesian Semiparametric GARCH Models
Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods
More informationGMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p
GMM and SMM Some useful references: 1. Hansen, L. 1982. Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p. 1029-54. 2. Lee, B.S. and B. Ingram. 1991 Simulation estimation
More informationSequential Monte Carlo Methods for Bayesian Computation
Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter
More informationLecture 9: Markov Switching Models
Lecture 9: Markov Switching Models Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Defining a Markov Switching VAR model Structure and mechanics of Markov Switching: from
More informationChapter 1. GMM: Basic Concepts
Chapter 1. GMM: Basic Concepts Contents 1 Motivating Examples 1 1.1 Instrumental variable estimator....................... 1 1.2 Estimating parameters in monetary policy rules.............. 2 1.3 Estimating
More informationMonte Carlo Methods. Leon Gu CSD, CMU
Monte Carlo Methods Leon Gu CSD, CMU Approximate Inference EM: y-observed variables; x-hidden variables; θ-parameters; E-step: q(x) = p(x y, θ t 1 ) M-step: θ t = arg max E q(x) [log p(y, x θ)] θ Monte
More informationFinancial Econometrics
Financial Econometrics Estimation and Inference Gerald P. Dwyer Trinity College, Dublin January 2013 Who am I? Visiting Professor and BB&T Scholar at Clemson University Federal Reserve Bank of Atlanta
More informationIntroduction to Machine Learning
Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin
More informationLecture 8: Multivariate GARCH and Conditional Correlation Models
Lecture 8: Multivariate GARCH and Conditional Correlation Models Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Three issues in multivariate modelling of CH covariances
More informationDSGE Methods. Estimation of DSGE models: GMM and Indirect Inference. Willi Mutschler, M.Sc.
DSGE Methods Estimation of DSGE models: GMM and Indirect Inference Willi Mutschler, M.Sc. Institute of Econometrics and Economic Statistics University of Münster willi.mutschler@wiwi.uni-muenster.de Summer
More informationThe Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations
The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture
More informationShort Questions (Do two out of three) 15 points each
Econometrics Short Questions Do two out of three) 5 points each ) Let y = Xβ + u and Z be a set of instruments for X When we estimate β with OLS we project y onto the space spanned by X along a path orthogonal
More informationWhen is a copula constant? A test for changing relationships
When is a copula constant? A test for changing relationships Fabio Busetti and Andrew Harvey Bank of Italy and University of Cambridge November 2007 usetti and Harvey (Bank of Italy and University of Cambridge)
More informationMarkov Chain Monte Carlo
Markov Chain Monte Carlo Recall: To compute the expectation E ( h(y ) ) we use the approximation E(h(Y )) 1 n n h(y ) t=1 with Y (1),..., Y (n) h(y). Thus our aim is to sample Y (1),..., Y (n) from f(y).
More informationA Course in Applied Econometrics Lecture 18: Missing Data. Jeff Wooldridge IRP Lectures, UW Madison, August Linear model with IVs: y i x i u i,
A Course in Applied Econometrics Lecture 18: Missing Data Jeff Wooldridge IRP Lectures, UW Madison, August 2008 1. When Can Missing Data be Ignored? 2. Inverse Probability Weighting 3. Imputation 4. Heckman-Type
More informationEcon 423 Lecture Notes: Additional Topics in Time Series 1
Econ 423 Lecture Notes: Additional Topics in Time Series 1 John C. Chao April 25, 2017 1 These notes are based in large part on Chapter 16 of Stock and Watson (2011). They are for instructional purposes
More informationChapter 2. GMM: Estimating Rational Expectations Models
Chapter 2. GMM: Estimating Rational Expectations Models Contents 1 Introduction 1 2 Step 1: Solve the model and obtain Euler equations 2 3 Step 2: Formulate moment restrictions 3 4 Step 3: Estimation and
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationSTATISTICS SYLLABUS UNIT I
STATISTICS SYLLABUS UNIT I (Probability Theory) Definition Classical and axiomatic approaches.laws of total and compound probability, conditional probability, Bayes Theorem. Random variable and its distribution
More informationDSGE-Models. Limited Information Estimation General Method of Moments and Indirect Inference
DSGE-Models General Method of Moments and Indirect Inference Dr. Andrea Beccarini Willi Mutschler, M.Sc. Institute of Econometrics and Economic Statistics University of Münster willi.mutschler@uni-muenster.de
More informationLikelihood-free MCMC
Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte
More informationMultivariate Time Series: VAR(p) Processes and Models
Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with
More informationEconomics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models
University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationMarkov Chain Monte Carlo methods
Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As
More informationSTA 294: Stochastic Processes & Bayesian Nonparametrics
MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationGeneralized Autoregressive Score Models
Generalized Autoregressive Score Models by: Drew Creal, Siem Jan Koopman, André Lucas To capture the dynamic behavior of univariate and multivariate time series processes, we can allow parameters to be
More informationTesting Restrictions and Comparing Models
Econ. 513, Time Series Econometrics Fall 00 Chris Sims Testing Restrictions and Comparing Models 1. THE PROBLEM We consider here the problem of comparing two parametric models for the data X, defined by
More informationOnline appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US
Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus
More informationMFE Financial Econometrics 2018 Final Exam Model Solutions
MFE Financial Econometrics 2018 Final Exam Model Solutions Tuesday 12 th March, 2019 1. If (X, ε) N (0, I 2 ) what is the distribution of Y = µ + β X + ε? Y N ( µ, β 2 + 1 ) 2. What is the Cramer-Rao lower
More informationGreene, Econometric Analysis (6th ed, 2008)
EC771: Econometrics, Spring 2010 Greene, Econometric Analysis (6th ed, 2008) Chapter 17: Maximum Likelihood Estimation The preferred estimator in a wide variety of econometric settings is that derived
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is
More informationEconometrics Summary Algebraic and Statistical Preliminaries
Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationFurther Evidence on Simulation Inference for Near Unit-Root Processes with Implications for Term Structure Estimation
Further Evidence on Simulation Inference for Near Unit-Root Processes with Implications for Term Structure Estimation Peter Fuleky and Eric Zivot January 9, 2010 Abstract We study indirect estimation methods
More informationMA Advanced Econometrics: Applying Least Squares to Time Series
MA Advanced Econometrics: Applying Least Squares to Time Series Karl Whelan School of Economics, UCD February 15, 2011 Karl Whelan (UCD) Time Series February 15, 2011 1 / 24 Part I Time Series: Standard
More informationSpring 2017 Econ 574 Roger Koenker. Lecture 14 GEE-GMM
University of Illinois Department of Economics Spring 2017 Econ 574 Roger Koenker Lecture 14 GEE-GMM Throughout the course we have emphasized methods of estimation and inference based on the principle
More informationWhat s New in Econometrics. Lecture 15
What s New in Econometrics Lecture 15 Generalized Method of Moments and Empirical Likelihood Guido Imbens NBER Summer Institute, 2007 Outline 1. Introduction 2. Generalized Method of Moments Estimation
More informationEcon 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE
Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE Eric Zivot Winter 013 1 Wald, LR and LM statistics based on generalized method of moments estimation Let 1 be an iid sample
More informationBayesian inference for multivariate extreme value distributions
Bayesian inference for multivariate extreme value distributions Sebastian Engelke Clément Dombry, Marco Oesting Toronto, Fields Institute, May 4th, 2016 Main motivation For a parametric model Z F θ of
More informationWeek 5 Quantitative Analysis of Financial Markets Characterizing Cycles
Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036
More informationLecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications
Lecture 3: Autoregressive Moving Average (ARMA) Models and their Practical Applications Prof. Massimo Guidolin 20192 Financial Econometrics Winter/Spring 2018 Overview Moving average processes Autoregressive
More information1. The Multivariate Classical Linear Regression Model
Business School, Brunel University MSc. EC550/5509 Modelling Financial Decisions and Markets/Introduction to Quantitative Methods Prof. Menelaos Karanasos (Room SS69, Tel. 08956584) Lecture Notes 5. The
More informationThe Instability of Correlations: Measurement and the Implications for Market Risk
The Instability of Correlations: Measurement and the Implications for Market Risk Prof. Massimo Guidolin 20254 Advanced Quantitative Methods for Asset Pricing and Structuring Winter/Spring 2018 Threshold
More information