Marginal Specifications and a Gaussian Copula Estimation

Size: px
Start display at page:

Download "Marginal Specifications and a Gaussian Copula Estimation"

Transcription

1 Marginal Specifications and a Gaussian Copula Estimation Kazim Azam Abstract Multivariate analysis involving random variables of different type like count, continuous or mixture of both is frequently required in economics. A Copula based methodology can be adopted for such data, where the association among the random variables is independent to their specific marginal distributions. Depending upon the chosen marginal specifications, copula estimation proceeds. A semi-parametric copula estimation, where the marginals are specified empirically performs very well, but for discrete data it s appropriateness is questioned (see Genest et al. (995)). Hoff (7) proposes a methodology where the marginal distributions are left completely unspecified and the copula parameters are estimated based on the order statistics of the observed data. We conduct an analysis to determine the effect on the estimates of a gaussian copula due to various marginal specifications. Employing a bayesian framework, we find that treating marginal distribution as unknown outperforms empirically distributed margins and misspecified margins in terms of biasedness and mean square error in small samples. JEL Classification: C, C, C5. Introduction Copula based method to conduct multivariate analysis is gaining popularity within the field of economics. It provides a framework, which is general across different type of data analysis, unlike other joint non-linear modelling of non-normal data where problems are dealt with caseby-case. The copula framework is invariant to applications in finance or micro data analysis. Embrechts et al. (999) show how for a VaR analysis, the assumption of multivariate normal fails to capture joint observations in the tails, and hence apply copula methods. Cherubini et al. () provides further financial applications. Munkin and Trivedi (999), using discrete micro data show how generally joint modelling is troublesome, and the problem increases when the

2 marginal distributions belong to different parametric families. Cameron et al. (), analyse a selection model with discrete outcomes in a copula framework. Among others see Cameron et al. (998) and Chib and Winkelmann ()for copula analysis with discretely varied marginal distributions. Hoff (7) applies a multivariate gaussian copula to estimate the correlation between an individuals income, degree, number of children etc. The separation of the joint distribution through a copula allows the marginals to belong to different parametric families, or even be non-parametric in which case we have a semiparametric copula estimation problem. Genest et al. (995), for continuous margins shows that by maximising the log pseudo-likelihood based on the normalized ranks, we obtain consistent and asymptotically normal estimates for the copula parameters. Such a semi-parametric specification is attractive, as unlike a parametric distribution it requires no parameters to be estimated. For financial applications such an option is welcoming, due to the marginals exhibiting high kurtosis and skewness which is troublesome for parametric marginals to capture. In case the data varies discretely, then regardless of employing a parametric or non-parametric marginal, we face difficulties. Trivedi and Zimmer (6) state for discrete margins, the copula maximizations often runs into computational problems, like algorithm convergence. They propose to employ a continuation transformation to the discrete variable and then base likelihood estimation on continuous copula families. Genest and Ne slehová (7) show that using rankbased estimators, the ties observed in the ranks would have to be dealt with first, and given low count data the step-size of an empirical distribution is large. Such problems make such an estimator quite biased. Pitt et al. (6) propose a bayesian sampling scheme for continuous and discrete margins in a fully parametric gaussian copula, where some of the issues regarding discrete margins are dealt with. Alternatively, for discrete or mixture of continuous-discrete data, Hoff (7) proposes a method where the marginals are left unspecified. Copula estimation is based on the order statistics of the observed data using bayesian techniques. The inference on the copula parameters is based on a summary statistic which is not a function of the nuisance marginal parameters. We aim to analyse the performance of the method proposed by Hoff, as compared to em-

3 ploying an empirical, or misspecified (continuous transformation) distribution for the marginals. We compute the biasedness and Mean Square Error (MSE) for these estimators in a gaussian copula framework with mixture of continuous-discrete margins. We merge the bayesian framework of Hoff (7) and Pitt et al. (6), to estimate the copula parameters. The sampling scheme is separated by first drawing the unknown quantities related to the marginal distributions conditional upon the copula parameters, followed by sampling the copula parameters conditional upon the marginals. We see by leaving the marginals unspecified in Hoff s method, produces less bias as compared to assuming empirically distributed or misspecified margins. The difference is larger in small samples and for correlation between continuous-discrete mixture margins. The bias approaches zero for large samples, except when misspecified margins are used. The mean square error also exhibits similar patterns, where Hoff s method based estimator has the smallest value as compared to other two, and equals the empirically distributed margins value for large samples. In section, we first provide the copula setup and provide details of the various marginal specifications which can be employed. That follows by setting out the bayesian sampling scheme in section 3. The Data Generating Process (DGP) is explained in section. Section 5 will give details of various marginal specifications we use to estimate the copula. Section 6, will describe the simulation over the DGP and quantities we are computing. Then finally we discuss the results from the simulation before concluding. Gaussian Copula Setup The definition of a copula can be best given by referring to Sklar s theorem (959), which states, if H is the multivariate distribution of dimension p, then it can partitioned into a copula C and the marginal distributions F,..., F p for the random variables Y,..., Y p given as H(y,..., y p ) = C(F (y ),..., F p (y p )), where C[, ] p [, ]. The copula distribution can also be stated as C(u,..., u p ) = pr (U u,..., U p u p ), 3

4 where U are the Probability Integral Transformations (PIT) of Y, obtained through the marginal distributions. There is a wide selection of copula families available, to capture different patterns of dependency among the random variables. Nelsen (7) and Joe (997)) provide a detailed coverage of copula theory the various families available. As our question investigates the effect of different specification of the marginal distribution on efficiency of a copula estimation, rather than how marginal specifications effect different copulas, we choose the most frequently used copula, namely the Gaussian copula. Using a gaussian copula along with normal margins, is essentially equivalent to a multivariate normal distribution. The gaussian copula can be defined as C(u,..., u p ) = Φ p {Φ (u ),..., Φ (u p )}, where Φ is the standard normal Cumulative Distribution Function (CDF), and Φ p is the CDF of a multivariate normal vector of dimension p. Let us denote a standard normal variable as z j with zero mean and variance one, which is computed as z j = Φ (u j ), for j =,..., p. Let z = (z,..., z p ), then we can define the multivariate normal with zero mean and the covariance matrix equal to the correlation matrix Θ as z N p (, Θ). Song () states that gaussian copula density equals Θ / exp( z Θ z)exp( zz ). Till now, we have only mentioned that u = (u,..., u p ) are obtained through PIT of the observed data, which in general copula methodology implies, applying the marginal distribution. F j, for the j th component could either be a parametric or a non-parametric marginal distribution. If a known parametric distribution is chosen, it will have some parameters associated with it. These parameters will need to be estimated along with the gaussian copula parameters (i.e. correlation matrix). For given values of the marginal parameters, the corresponding standard normals z j can be computed. If a non-parametric specification is preferred, either

5 due to the lack of knowledge about y j or the limitations a parametric distribution can have. Then the corresponding z j can be obtained through the empirical distribution of the observed data, without having to estimate any marginal parameters. We simplify the problem by not having mixture of marginal specifications in a given multivariate analysis. That is, if F j is specified to be parametric, then F \j (i.e. all other marginals distributions except F j ) will also be parametric, and vice versa for the case of non-parametric specifications. We now present the various marginal specifications used with gaussian copula to be considered in the simulation, in detail.. Parametric Copula Specification Let n be the total number of observations given as y,..., y n, for i =,..., n, where each y i is a (p ) vector. a Then the fully parametric gaussian copula estimation problem is given as z i N p (, Θ), y ij = F ij {Φ(z ij) β j }, for all i and j, where F ij is the CDF function for either a continuous or discrete random variable, and β j is the parameter vector associated with the j th component. For a component j, the marginal distribution F ij is fixed over all the i s, and hence could also be simply stated as F j. As F j could either be corresponding to a continuous or discrete random variable, the mapping from y ij to u ij will vary. In case j th component is continuous, F j will be a one-to-one function. Given a value of β j, then z ij can be easily be computed. But if the j th component is discrete, F j will be a many-to-one function. Then given a value for β j, we cannot directly impute the corresponding z ij. We will have to consider them as auxiliary variables, and will have to be simulated along with the copula and the marginal parameters. Our estimation problem here is similar to Pitt et al. (6), but we do not account for presence of covariates in the marginal specification. 5

6 . Semi-Parametric Copula Specification If the z ij are computed through assuming employing a non-parametric marginal distribution, namely an empirical distribution. Then along with a parametric copula, the estimation problem is on a semi-parametric based specification. In such a setup, there are no marginal parameters which need to be estimated, hence by employing rank based transformations over all the i s for each component j, z ij can be obtained. If all the F j s are corresponding to continuous random variables, then an estimator based on the normalized ranks is consistent and asymptotically normal (see Genest et al. (995)). However, these properties and the estimator becomes biased, if all or some of the marginals are discrete. The size of the bias depends upon the possible variation in the discrete data, the worst case being a component j is a binary random variable. The underlying problem is the ties observed in the rank of the observed data. There are various methods to deal with ties, like splitting, ignoring or adjusting the ties. Genest et al. () show through simulation that a method based on splitting the ties produces the smallest bias in the estimation of Θ. Hoff (7) presents a semi-parametric copula estimation technique, which unlike the above explained method, treats all the z ij as auxiliary variables. No assumption is made regarding F j, and it is treated as completely unknown. This method is applicable to discrete, continuous and mixture of both data. The methodology is equivalent to employing the standard semiparametric technique above, where z ij are known, if the margins are continuous and the sample size n is large. The benefit of not having to estimate the marginal parameters, comes at a cost of having less information from the available data. Let us see the exact specifications for both of the cases, when z (i.e. all the standard normals known) are completely known, and when as Hoff (7) we treat them as latent variables in a semi-parameteric setup... Empirical Distribution F j Given empirical distributions are used for all the marginals in a multivariate gaussian copula, then there are no parameters associated to any components. Also given z are completely known, then the modelling specification becomes 6

7 z i N p (, Θ), y ij = F ij {Φ(z ij)}, F j (y mj ) = n n+ i= (y ij y mj ), for all i and j. F j denotes the empirical distribution, used instead of a parametric F j for all j. The third equation is just the empirical CDF, and we divide the ranks by n + to avoid boundary values. As mentioned, we only need to estimate the correlation matrix Θ, and in case any of the random variable is discrete then decide how to deal with ties. As Genest et al. (995) show splitting the ties produces the smallest bias, so we randomly split them... Unknown F j Here, unlike employing an empirical CDF and splitting the ties in the rank of the observed data to obtain z, we treat them as completely unknown. There is no assumption made regarding any F j, and they are all completely unknown. The only information we have regarding F j is that for all the components it is a non-decreasing function. We also know the ranks for each i in a given component j, hence if the rank of y ij is k, then we can write corresponding order statistic as y (k) j, such that y ij = y (k). From this information we can infer that the unobserved j z ij corresponding to y ij, will have the same rank k. Hence we can write this more formally as y (k ) j z (k ) j (y ij = y (k) j ) y (k+) j, < (z ij = z (k) j ) < z (k+) j, Note that we do not have strict inequality for the observed data, which is to accommodate ties in the ranks. From the above equalities, we know for certain that z ij has to lie in the interval dictated by the largest order statistic which is smaller than z (k) j and the smallest order statistic which is greater than z (k) j. Based on this information, we set out the same gaussian copula specification as z i N p (, Θ), for all i. When the interval where z ij lies in becomes smaller due to having continuous margins and large n, then the uncertainty regarding the true value of z ij is less and the methodology is similar to simply applying the empirical distribution F j for each component. 7

8 In the next section we describe the bayesian sampling scheme to estimate Θ, for all of the marginal specifications described above. 3 Bayesian Estimation As described previously, we aim to estimate the gaussian copula parameters, namely the correlation matrix denoted as Θ. If all the margins are parametrically specified, then each component j will have a parameter vector β j associated to it. Let β denote the vector containing all the marginal parameters from all the components, β = (β,..., β p ). In a parametric setting, if any of the random variables is discrete, we also need to estimate the standard normals z ij, for that component. When we apply the empirical distribution to obtain the corresponding z, then the only parameter needed to be estimated is the correlation matrix Θ, however if F is taken to be unknown, then we do not observe z, and will have to sampled along with Θ. Given this setting we can partition the bayesian sampling scheme into two parts. First β = (β,..., β p ) and z = (z,..., z p ) are sampled conditional upon Θ, where needed. Secondly, we sample Θ conditional upon β and z. 3. First Stage p(β, z Θ) This stage can be skipped and move to the second stage, if a semi-parametric specification is employed with empirically distributed margins F. But if the margins are specified parametrically F, or if the marginal distributions are not known, then we first need to sample β and z. 3.. Parametric Margins In case the marginals are all parametrically specified, then we sample in this order. Sample from p(β j y.,j, z.,\j, Θ), where y.,j denotes all the observations n for the given component j, and z.,\j is all the observations from all the other components exempt j.. If j th margin s distribution F j is continuous, then compute z ij = Φ {F ij (y ij β j )}. If F j is a discrete distreibution, sample z ij from p(z ij β j, y ij, z i,\j ; Θ), for all i. 8

9 The above two steps are repeated for each j, in turn. Pitt et al. (6) provide details in respect to the form of the conditional density of β j. As it is not possible to sample directly from the conditional density. They propose a Metropolis-Hasting algorithm, where the proposal density is approximated to a multivariate t distribution, with mean equal to the mode β j of log p(β j y.,j, z.,\j, Θ), this can be found by quasi-newton Raphson scheme. The variance of the t distribution is equated to the negative inverse of the second derivative of the log conditional density, computed at the mode. The degrees-of-freedom is chosen, such that the proposal density can dominate the true density in the tails. Such a method is similar to a Laplace-type proposal (see Chib and Greenberg (998) and, Chib and Winkelmann ()). A new proposed value βj is then evaluated in a Metropolis-Hasting step. In case if the component j has a discrete distribution, z ij are sampled after sampling β j, from a univariate gaussian distribution, where the mean and the variance take into account the standard normals from the other components z i,\j and the correlation among them, given through Θ. We refer the interested reader for full details of the briefly explained above algorithm to Pitt et al. (6), page Unspecified Marginals On the other hand if a semi-parametric copula approach is adopted, where no assumption regarding F j is made, then there no β j to sample, but only z needs to sampled in this stage. Here we follow the approach set out by Hoff (7), Where we sample z ij from z ij p(z ij Θ, z i,\j, y (k) j ), for all i and j, where the conditional density of z ij is conditioned on the correlation matrix Θ and all the other standard normals from each j. The conditioning of y k j, implies knowing the rank of z ij through y ij, using that we can determine the interval (z (k ) j, z (k+) j ) where z ij lies in. The sampling scheme iterates over all the i s in a given j, and then moves on to the next j. Similar to the case, where z are sampled if the F is known and is a discrete distribution, here z ij are sampled from a univariate gaussian distribution, with the mean and variance considering the correlation among the variables and the other components z i,\j. The major difference in sampling the z here is that the truncation is dictated by the order statistics, whereas in the parametric case, 9

10 the truncation is given by the CDF of the discrete parametric distribution, evaluated at y ij and y ij (see Pitt et al. (6)). Hence this scheme is adopted regardless of the true distribution of j th being continuous or discrete. Depending upon whichever marginal specifications are chosen, before moving on to the second stage we should obtain z, to proceed with sampling Θ. 3. Second Stage In this stage, we no longer care about the assumptions specified on the marginal distributions. All we require are z from the previous stage, to sample Θ. Hence, either parametrically defined marginal distribution or non-parameterically, this scheme for Θ is invariant. We can write the posterior of Θ as p(θ z) p(θ) p(z Θ). Similar to Hoff (7), we assume a semi-conjugate prior for the gaussian copula. The prior p(θ) is defined as, let the prior of V be given as an inverse-wishart distribution (ν, ν V ), parametrized such that E[V ] = V, where ν is the degrees-of-freedom and ν V the scale matrix. Let Θ be equal in distribution to correlation matrix given as Θ [i,j] = V [i,j] V[i,i] V [j,j]. The posterior of V can then shown to be proportional to V z inverse-wishart(ν + n, ν V + z z), from which a sample of V can be obtained, and then Θ computed from the above transformation. We could have followed Pitt et al. (6) with their sampling scheme for Θ, but choose not, as our focus in not on an efficient sampling scheme, but to check the effects of the marginal specifications on copula estimation.

11 Data Generating Process In this section we show how to simulate data from a multivariate gaussian copula and provide details about the Data Generating Process (DGP). This simulated data will then be used to test various marginal specifications and their effect on a gaussian copula estimation. For some correlation matrix Θ and β, a set of generated y can be sampled as follow. Sample z from N p (z ; Θ),. Obtain u = Φ(z), 3. Compute y ij = F j (u ij β j ), for all i and j, where u = (u,..., u p ), and each u i is (n ) vector. Step 3 above implies, we need to be able to compute the inverse CDF of the chosen parametric marginal distribution. Let us then set out the DGP, which will be used through the rest of the paper. We choose p = 3 and alter n such that it ranges from small sample (n = ) to large sample (n = 5). z N ; ,..6 u = Φ(z), y., = F (u.,.5) F (y.,.5) = Exponential(y., λ ), y., = F (u., 6) F (y., 6) = Poisson(y., λ ), y.,3 = F 3 (u.,3.6) F (y.,.6) = Bernoulli(y., λ 3 ). So the true DGP is a mixture of continuous and discrete marginals. This DGP will stay fixed through out the simulation. Using the simulated y, we assume various marginal specifications and compare them in terms of estimating the true correlation matrix Θ.

12 5 Marginal Specifications Now we state the various marginal specifications we will employ, in order to estimate the gaussian copula parameters. For ease of reference, we can refer to them as Marginal Specifications (MS), So various specifications will be defined as MS, MS etc. Their detail is as follow MS All three marginals (F, F and F 3 ) are assumed to be completely unknown. Using the order statistics of the observed data, first z, and then the correlation matrix is sampled. This is as described previously, the method proposed by Hoff (7). MS Assume all the margins are empirically distributed, implying compute z i,j = F j (y i,j ), for all i and j. MS3 Perform a continuation transformation for the two discrete margins, then let z ij = Φ {F (y ij β j )} = ln N (y ij µ j, σ j ), for all i and j. Hence all margins are log normally distributed. So we only specify three different marginal specifications. The first two correspond to semiparametric copula estimation, and the last to a fully parametric copula estimation. We decided to consider misspecified margins, as in very small sample it is interesting to see how well do they perform in estimating the copula parameters. MS3 takes the discrete marginals and adds a random independent term between [, ] with the observed values, to make them continuous. This is an approach stated in Trivedi and Zimmer (6), to avoid computational problems generally encountered in likelihood estimation. This transformation along with assuming log normal distribution induces a misspecification. For the first margin (originally exponential in the DGP), is also misspecified by assuming a log normal too. Next, let us look at the MCMC and the simulation over the DGP in more detail.

13 6 Simulation 6. Setup We will in essence perform a Gibbs sampling type of algorithm over the two stages defined in section 3. We perform 55 iterations over the sampling scheme to obtain the posterior density of Θ, of which the first 5 are discarded for burn-in of the MCMC chain. We will not be discussing the posterior density obtained of the marginal parameters, as our focus is on the correlation matrix. From the posterior density we compute the posterior mean E(Θ y). To analyse the properties of the various marginal specifications and their effect on the estimation of Θ, we have to obtain a distribution for the posterior mean, hence we employ Monte carlo over the DGP. We choose the size of the Monte carlo simulation to be 5, which is sufficient to conduct inference on. At each Monte carlo iteration we obtain a new sample of y through the same DGP, which can be denoted as {y} s, where s =,..., S. We can define the general simulation structure as for s =,..., S, Sample {y} s from the DGP, For each, MS, MS & MS3, ] obtain E [Θ {y} s, end, end. After ignoring the first 5 iterations for burn-in, the autocorrelation for all the parameters is lower than. after three lags. The trace plot for all the parameters also shows, that the MCMC chain mixes well and dominates in the tails of the true distribution. 6. Biasedness & Variance After obtaining the distribution of the posterior mean of Θ, for all the three marginal specifications, we would like to compare them in terms of their biasedness and variance to the true 3

14 correlation matrix Θ. We compute two quantities of interest for all the marginal specifications. ] First, we compute the bias where we compute the difference of E [Θ {y} s from Θ. Secondly, we compute the MSE, which combines the variance and bias of the estimator. Both are given as ] Bias = S S s= [Θ {y} E s Θ, Mean Squared Error (MSE) = S S s= [ E[Θ {y} s ] Θ]. We will compare the biasedness across all three estimators (specifications), and as our interest is in determining the performance of MS compared to the other specifications, we will compute the MSE ratio of MS with respect to MS and MS3 ω = MSE M MSE M, ω 3 = MSE M MSE M3. These quantities will be computed for all the off-diagonal entries of the correlation matrix Θ (i.e. lower correlation parameters). The whole simulation is repeated for different values n. 7 Results 7. Bias Let us first take a look at the biasedness for the three marginal specifications. B MS refers to the bias values from MS in table, and similarly denoting the other marginal specifications. We see all the MS under-predict the true Θ. Θ [i,j] represent an entry in row i and column j of the correlation matrix Θ. The bias for MS is lower than MS and MS3 for small sample n =. The difference is particularly larger for Θ [,] (correlation between continuous and discrete margin) in MS and MS, hence Hoff s method has a smaller bias for mixture of distributions, which is more prominent for small sample. It is interesting to see that the misspecified model MS3 has smaller bias as compared to MS for n =, but as n increases, MS becomes less biased as compared to MS3. This is simply due to the bias created by

15 small sample common across all marginal specifications. As we increase the sample size, the difference between B MS and B MS converges to zero, but the rate convergence is slower for mixture of margins (continuous and discrete). We see the biasedness in MS3, does not drop as dramatically as MS and MS. This is especially true for Θ [,3], where two discrete margins are misspecified, which shows continuation transformation not to be a very efficient technique. The misspecification of an exponential margin to a log-normal still seems to create smaller bias, when associated with discrete margins transformed. Overall, we see Hoff s method has a lower bias, as compared to computing z through an empirical distribution. B MS Table : Computed bias for all marginal specifications n= n=5 n=5 n= n=5 n=5 Θ [,] Θ [,3] Θ [,3] B MS Θ [,] Θ [,3] Θ [,3] B MS3 Θ [,] Θ [,3] Θ [,3] The difference becomes smaller as n increases, as the step-size in an empirical distribution and the interval wherez ij lies (Hoff s method) both become smaller. This is especially the case when the bias is examined for two continuously distributed margins. In our case, we either have mixture of continuous and discrete, or both discrete margins then Hoff s method has smaller bias. 5

16 7. MSE In table 7., we compute the MSE ratio of MS with respect to the other marginal specifications. We denote the ratios as ω and ω 3 with respect to MS and MS3 respectively. Similar to the case of biasedness of the marginal specifications, here in ω for the case of Θ [,] indicates the MSE is smaller for MS than MS, as compared to the other correlation parameters in Θ. Again, this ratio reaches the value one, as the sample size is increased. Over all the values of n, see Hoff s method produces the smallest MSE, as both ω and ω 3 are both less than one. ω 3, actually decreases further as n increases, further pointing out the inappropriateness of using misspecified margins. Another interesting result is how the ratio of ω for Θ [,3] and Θ [,3] differ. MS has smaller MSE compared to MS, when a continuous and high count margin (poisson), or a discrete-discrete margin are considered. But for a combination of a continuous-low count margin (binary), the MSE difference is smaller. This is due to the ranking problems a binary variable induces. Similar to the bias, MSE ratio for MS and MS reach value of as the sample size increases. ω Table : MSE ratio n= n=5 n=5 n= n=5 n=5 Θ [,] Θ [,3] Θ [,3] ω 3 Θ [,] Θ [,3] Θ [,3]

17 7.3 MS Kernel We present some kernel density plots for the first marginal specification, assuming F to be completely unknown. Figure - 5 are density plots for the posterior mean over the DGP. We can clearly see the dispersion around the mean (red line) decreases as n increases. Theta[,] denotes the correlation parameter between the first and the second random variable. It is interesting to see the dispersion for Theta[,3] and Theta[,3] is relatively more as compared to that for Theta[,], through all sample sizes. The black line represents the true correlation from the DGP. Figure : Posterior of E(Θ y), n = Theta[,] Theta[,3] Theta[,3] 7

18 Figure : Posterior of E(Θ y), n = Theta[,] Theta[,3] Theta[,3] Figure 3: Posterior of E(Θ y), n = Theta[,].5.5 Theta[,3] Theta[,3] 8

19 Figure : Posterior of E(Θ y), n = Theta[,] Theta[,3] Theta[,3] Figure 5: Posterior of E(Θ y), n = Theta[,].5.5 Theta[,] Theta[,] 9

20 8 Conclusion Copula based method provide flexible multivariate analysis for margins of different types and dependency patterns, which are not best described by elliptical distributions. Marginals can be specified parametrically and along with a parametric copula, the estimation problem is fully parametric. Alternatively, a non-parametric distribution can be used for the marginals which leads to a semi-parametric copula estimation problem. For random variables of continuous type, a semi-parametric copula estimation is shown to efficient and asymptotically normal (see Genest et al. (995)), and hence both approach are equivalent. But for multivariate analysis of discrete or mixture of continuous-discrete data, empirically computed marginals are not appropriate. Even in the parametric case, the exact knowledge of the marginal distribution is not always available and transformation to a continuous distribution is also not fruitful. Hoff (7) proposes a method, where the marginal parameters are not required to be estimated and by simply obtaining the information contained in the order statistics, we can estimate the copula parameters. Such a method is useful, as it considers the uncertainty in the mapping through the CDF from the observed data, and the lack of knowledge about the true marginal distribution. We specified various marginal specifications, to estimate a multivariate gaussian copula. Borrowing the bayesian estimation framework for a fully parametric copula specification, and for a semi-parametric copula estimation from Pitt et al. (6) and Hoff (7) respectively. To check for the performance of various marginal specifications, we provided a simulation setup and described a DGP of mixture of marginals type. We specified three different marginal specifications to estimate the gaussian copula. For the first case, the marginals were considered unknown (Hoff s method). Second, the PIT was performed through empirical distributions, and finally a completely misspecified margins were employed. The results showed how Hoff s method outperformed the other two specifications, especially in small samples, where the uncertainty and inappropriateness for the other methods is large. The bias for Hoff s method is smaller in all the samples, but for large samples it becomes equal to applying empirical distribution. The misspecified method has the largest bias, in large samples. Also in terms of mean squared error, Hoff s method has smaller MSE

21 for small samples, compared to the other two. Overall, even though for continuous and large sample both approaches are equivalent, but in case we are dealing with discrete data, Hoff s method performs better than empirically assumed margins, and should be considered for multivariate analysis of either discrete or continuousdiscrete data in a copula framework. An initial continuation transformation of discrete data before estimating the copula, is also not very efficient. References A. C. Cameron, T. Li, P. K. Trivedi, and D. M. Zimmer. Modelling the differences in counted outcomes using bivariate copula models with application to mismeasured counts. Econometrics Journal, 7():566 58, December. S. Chib and E. Greenberg. Analysis of multivariate probit models. Biometrika, pages 37 36, 998. S. Chib and R. Winkelmann. Markov chain monte carlo analysis of correlated count data. Journal of Business & Economic Statistics, 9():8 35,. P. Embrechts, A. McNeil, and D. Straumann. Correlation and dependence in risk management: Properties and pitfalls. In Risk Management: Value At Risk And Beyond, pages Cambridge University Press, 999. C. Genest and J. Ne slehová. A primer on copulas for count data. Astin Bulletin, 37():75, 7. C. Genest, K. Ghoudi, and L.-P. Rivest. A semiparametric estimation procedure of dependence parameters in multivariate families of distributions. Biometrika, 8(3):53 55, 995. C. Genest, J. Ne slehová, and N. Ben Ghorbal. Estimators based on kendall s tau in multivariate copula models. Australian & New Zealand Journal of Statistics, 53():57 77,. ISSN 67-8X. P. D. Hoff. Extending the rank likelihood for semiparametric copula estimation. Ann. Appl. Stat., ():65 83, 7. H. Joe. Multivariate Models and Dependence Concepts. Chapman & Hall/CRC, 997. M. K. Munkin and P. K. Trivedi. Simulated maximum likelihood estimation of multivariate mixed-poisson regression models, with application. Econometrics Journal, ():9 8, 999. R. B. Nelsen. An Introduction to Copulas. Springer, 7. M. Pitt, D. Chan, and R. Kohn. Efficient bayesian inference for gaussian copula regression models. Biometrika, 93(3):537 55, September 6. P. X.-K. Song. Multivariate dispersion models generated from gaussian copula. Scandinavian Journal of Statistics, 7():35 3,. P. K. Trivedi and D. M. Zimmer. Copula Modeling: An Introduction for Practitioners. Foundations and Trends in Econometrics, ():, 6. ISSN

A Goodness-of-fit Test for Copulas

A Goodness-of-fit Test for Copulas A Goodness-of-fit Test for Copulas Artem Prokhorov August 2008 Abstract A new goodness-of-fit test for copulas is proposed. It is based on restrictions on certain elements of the information matrix and

More information

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal

More information

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff David Gerard Department of Statistics University of Washington gerard2@uw.edu May 2, 2013 David Gerard (UW)

More information

Financial Econometrics and Volatility Models Copulas

Financial Econometrics and Volatility Models Copulas Financial Econometrics and Volatility Models Copulas Eric Zivot Updated: May 10, 2010 Reading MFTS, chapter 19 FMUND, chapters 6 and 7 Introduction Capturing co-movement between financial asset returns

More information

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline. Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Markov Chain Monte Carlo methods

Markov Chain Monte Carlo methods Markov Chain Monte Carlo methods By Oleg Makhnin 1 Introduction a b c M = d e f g h i 0 f(x)dx 1.1 Motivation 1.1.1 Just here Supresses numbering 1.1.2 After this 1.2 Literature 2 Method 2.1 New math As

More information

Modelling Dropouts by Conditional Distribution, a Copula-Based Approach

Modelling Dropouts by Conditional Distribution, a Copula-Based Approach The 8th Tartu Conference on MULTIVARIATE STATISTICS, The 6th Conference on MULTIVARIATE DISTRIBUTIONS with Fixed Marginals Modelling Dropouts by Conditional Distribution, a Copula-Based Approach Ene Käärik

More information

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline.

Dependence. MFM Practitioner Module: Risk & Asset Allocation. John Dodson. September 11, Dependence. John Dodson. Outline. MFM Practitioner Module: Risk & Asset Allocation September 11, 2013 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y

More information

Robustness of a semiparametric estimator of a copula

Robustness of a semiparametric estimator of a copula Robustness of a semiparametric estimator of a copula Gunky Kim a, Mervyn J. Silvapulle b and Paramsothy Silvapulle c a Department of Econometrics and Business Statistics, Monash University, c Caulfield

More information

Report and Opinion 2016;8(6) Analysis of bivariate correlated data under the Poisson-gamma model

Report and Opinion 2016;8(6)   Analysis of bivariate correlated data under the Poisson-gamma model Analysis of bivariate correlated data under the Poisson-gamma model Narges Ramooz, Farzad Eskandari 2. MSc of Statistics, Allameh Tabatabai University, Tehran, Iran 2. Associate professor of Statistics,

More information

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

MULTIDIMENSIONAL POVERTY MEASUREMENT: DEPENDENCE BETWEEN WELL-BEING DIMENSIONS USING COPULA FUNCTION

MULTIDIMENSIONAL POVERTY MEASUREMENT: DEPENDENCE BETWEEN WELL-BEING DIMENSIONS USING COPULA FUNCTION Rivista Italiana di Economia Demografia e Statistica Volume LXXII n. 3 Luglio-Settembre 2018 MULTIDIMENSIONAL POVERTY MEASUREMENT: DEPENDENCE BETWEEN WELL-BEING DIMENSIONS USING COPULA FUNCTION Kateryna

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Gaussian kernel GARCH models

Gaussian kernel GARCH models Gaussian kernel GARCH models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics 7 June 2013 Motivation A regression model is often

More information

Semi-parametric predictive inference for bivariate data using copulas

Semi-parametric predictive inference for bivariate data using copulas Semi-parametric predictive inference for bivariate data using copulas Tahani Coolen-Maturi a, Frank P.A. Coolen b,, Noryanti Muhammad b a Durham University Business School, Durham University, Durham, DH1

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Bayesian GLMs and Metropolis-Hastings Algorithm

Bayesian GLMs and Metropolis-Hastings Algorithm Bayesian GLMs and Metropolis-Hastings Algorithm We have seen that with conjugate or semi-conjugate prior distributions the Gibbs sampler can be used to sample from the posterior distribution. In situations,

More information

Bayesian spatial hierarchical modeling for temperature extremes

Bayesian spatial hierarchical modeling for temperature extremes Bayesian spatial hierarchical modeling for temperature extremes Indriati Bisono Dr. Andrew Robinson Dr. Aloke Phatak Mathematics and Statistics Department The University of Melbourne Maths, Informatics

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Bayesian inference for multivariate extreme value distributions

Bayesian inference for multivariate extreme value distributions Bayesian inference for multivariate extreme value distributions Sebastian Engelke Clément Dombry, Marco Oesting Toronto, Fields Institute, May 4th, 2016 Main motivation For a parametric model Z F θ of

More information

Multivariate negative binomial models for insurance claim counts

Multivariate negative binomial models for insurance claim counts Multivariate negative binomial models for insurance claim counts Peng Shi (Northern Illinois University) and Emiliano A. Valdez (University of Connecticut) 9 November 0, Montréal, Quebec Université de

More information

Bayesian Semiparametric GARCH Models

Bayesian Semiparametric GARCH Models Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods

More information

Chapter 1. Bayesian Inference for D-vines: Estimation and Model Selection

Chapter 1. Bayesian Inference for D-vines: Estimation and Model Selection Chapter 1 Bayesian Inference for D-vines: Estimation and Model Selection Claudia Czado and Aleksey Min Technische Universität München, Zentrum Mathematik, Boltzmannstr. 3, 85747 Garching, Germany cczado@ma.tum.de

More information

Bayesian Semiparametric GARCH Models

Bayesian Semiparametric GARCH Models Bayesian Semiparametric GARCH Models Xibin (Bill) Zhang and Maxwell L. King Department of Econometrics and Business Statistics Faculty of Business and Economics xibin.zhang@monash.edu Quantitative Methods

More information

The Bayesian Approach to Multi-equation Econometric Model Estimation

The Bayesian Approach to Multi-equation Econometric Model Estimation Journal of Statistical and Econometric Methods, vol.3, no.1, 2014, 85-96 ISSN: 2241-0384 (print), 2241-0376 (online) Scienpress Ltd, 2014 The Bayesian Approach to Multi-equation Econometric Model Estimation

More information

Using Estimating Equations for Spatially Correlated A

Using Estimating Equations for Spatially Correlated A Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Behaviour of multivariate tail dependence coefficients

Behaviour of multivariate tail dependence coefficients ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 22, Number 2, December 2018 Available online at http://acutm.math.ut.ee Behaviour of multivariate tail dependence coefficients Gaida

More information

Estimation of Copula Models with Discrete Margins (via Bayesian Data Augmentation) Michael S. Smith

Estimation of Copula Models with Discrete Margins (via Bayesian Data Augmentation) Michael S. Smith Estimation of Copula Models with Discrete Margins (via Bayesian Data Augmentation) Michael S. Smith Melbourne Business School, University of Melbourne (Joint with Mohamad Khaled, University of Queensland)

More information

ABC methods for phase-type distributions with applications in insurance risk problems

ABC methods for phase-type distributions with applications in insurance risk problems ABC methods for phase-type with applications problems Concepcion Ausin, Department of Statistics, Universidad Carlos III de Madrid Joint work with: Pedro Galeano, Universidad Carlos III de Madrid Simon

More information

Efficient estimation of a semiparametric dynamic copula model

Efficient estimation of a semiparametric dynamic copula model Efficient estimation of a semiparametric dynamic copula model Christian Hafner Olga Reznikova Institute of Statistics Université catholique de Louvain Louvain-la-Neuve, Blgium 30 January 2009 Young Researchers

More information

Bayesian Nonparametric Regression for Diabetes Deaths

Bayesian Nonparametric Regression for Diabetes Deaths Bayesian Nonparametric Regression for Diabetes Deaths Brian M. Hartman PhD Student, 2010 Texas A&M University College Station, TX, USA David B. Dahl Assistant Professor Texas A&M University College Station,

More information

Bayesian Methods in Multilevel Regression

Bayesian Methods in Multilevel Regression Bayesian Methods in Multilevel Regression Joop Hox MuLOG, 15 september 2000 mcmc What is Statistics?! Statistics is about uncertainty To err is human, to forgive divine, but to include errors in your design

More information

Songklanakarin Journal of Science and Technology SJST R1 Sukparungsee

Songklanakarin Journal of Science and Technology SJST R1 Sukparungsee Songklanakarin Journal of Science and Technology SJST-0-0.R Sukparungsee Bivariate copulas on the exponentially weighted moving average control chart Journal: Songklanakarin Journal of Science and Technology

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

Riemann Manifold Methods in Bayesian Statistics

Riemann Manifold Methods in Bayesian Statistics Ricardo Ehlers ehlers@icmc.usp.br Applied Maths and Stats University of São Paulo, Brazil Working Group in Statistical Learning University College Dublin September 2015 Bayesian inference is based on Bayes

More information

Online Appendix to: Marijuana on Main Street? Estimating Demand in Markets with Limited Access

Online Appendix to: Marijuana on Main Street? Estimating Demand in Markets with Limited Access Online Appendix to: Marijuana on Main Street? Estating Demand in Markets with Lited Access By Liana Jacobi and Michelle Sovinsky This appendix provides details on the estation methodology for various speci

More information

Markov Chain Monte Carlo Methods

Markov Chain Monte Carlo Methods Markov Chain Monte Carlo Methods John Geweke University of Iowa, USA 2005 Institute on Computational Economics University of Chicago - Argonne National Laboaratories July 22, 2005 The problem p (θ, ω I)

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model

Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model UNIVERSITY OF TEXAS AT SAN ANTONIO Hastings-within-Gibbs Algorithm: Introduction and Application on Hierarchical Model Liang Jing April 2010 1 1 ABSTRACT In this paper, common MCMC algorithms are introduced

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Simulation of Tail Dependence in Cot-copula

Simulation of Tail Dependence in Cot-copula Int Statistical Inst: Proc 58th World Statistical Congress, 0, Dublin (Session CPS08) p477 Simulation of Tail Dependence in Cot-copula Pirmoradian, Azam Institute of Mathematical Sciences, Faculty of Science,

More information

A New Generalized Gumbel Copula for Multivariate Distributions

A New Generalized Gumbel Copula for Multivariate Distributions A New Generalized Gumbel Copula for Multivariate Distributions Chandra R. Bhat* The University of Texas at Austin Department of Civil, Architectural & Environmental Engineering University Station, C76,

More information

Imputation Algorithm Using Copulas

Imputation Algorithm Using Copulas Metodološki zvezki, Vol. 3, No. 1, 2006, 109-120 Imputation Algorithm Using Copulas Ene Käärik 1 Abstract In this paper the author demonstrates how the copulas approach can be used to find algorithms for

More information

The Instability of Correlations: Measurement and the Implications for Market Risk

The Instability of Correlations: Measurement and the Implications for Market Risk The Instability of Correlations: Measurement and the Implications for Market Risk Prof. Massimo Guidolin 20254 Advanced Quantitative Methods for Asset Pricing and Structuring Winter/Spring 2018 Threshold

More information

Bayesian Inference and MCMC

Bayesian Inference and MCMC Bayesian Inference and MCMC Aryan Arbabi Partly based on MCMC slides from CSC412 Fall 2018 1 / 18 Bayesian Inference - Motivation Consider we have a data set D = {x 1,..., x n }. E.g each x i can be the

More information

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Tihomir Asparouhov 1, Bengt Muthen 2 Muthen & Muthen 1 UCLA 2 Abstract Multilevel analysis often leads to modeling

More information

Multivariate Distributions

Multivariate Distributions IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate

More information

Multivariate Statistics

Multivariate Statistics Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical

More information

Counts using Jitters joint work with Peng Shi, Northern Illinois University

Counts using Jitters joint work with Peng Shi, Northern Illinois University of Claim Longitudinal of Claim joint work with Peng Shi, Northern Illinois University UConn Actuarial Science Seminar 2 December 2011 Department of Mathematics University of Connecticut Storrs, Connecticut,

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

EVANESCE Implementation in S-PLUS FinMetrics Module. July 2, Insightful Corp

EVANESCE Implementation in S-PLUS FinMetrics Module. July 2, Insightful Corp EVANESCE Implementation in S-PLUS FinMetrics Module July 2, 2002 Insightful Corp The Extreme Value Analysis Employing Statistical Copula Estimation (EVANESCE) library for S-PLUS FinMetrics module provides

More information

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract

Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies. Abstract Bayesian Estimation of A Distance Functional Weight Matrix Model Kazuhiko Kakamu Department of Economics Finance, Institute for Advanced Studies Abstract This paper considers the distance functional weight

More information

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis Summarizing a posterior Given the data and prior the posterior is determined Summarizing the posterior gives parameter estimates, intervals, and hypothesis tests Most of these computations are integrals

More information

Semiparametric Gaussian Copula Models: Progress and Problems

Semiparametric Gaussian Copula Models: Progress and Problems Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle European Meeting of Statisticians, Amsterdam July 6-10, 2015 EMS Meeting, Amsterdam Based on

More information

Likelihood-free MCMC

Likelihood-free MCMC Bayesian inference for stable distributions with applications in finance Department of Mathematics University of Leicester September 2, 2011 MSc project final presentation Outline 1 2 3 4 Classical Monte

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Lecture 8: The Metropolis-Hastings Algorithm

Lecture 8: The Metropolis-Hastings Algorithm 30.10.2008 What we have seen last time: Gibbs sampler Key idea: Generate a Markov chain by updating the component of (X 1,..., X p ) in turn by drawing from the full conditionals: X (t) j Two drawbacks:

More information

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract

Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus. Abstract Bayesian analysis of a vector autoregressive model with multiple structural breaks Katsuhiro Sugita Faculty of Law and Letters, University of the Ryukyus Abstract This paper develops a Bayesian approach

More information

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Econometrics Workshop UNC

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning MCMC and Non-Parametric Bayes Mark Schmidt University of British Columbia Winter 2016 Admin I went through project proposals: Some of you got a message on Piazza. No news is

More information

Markov Switching Regular Vine Copulas

Markov Switching Regular Vine Copulas Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS057) p.5304 Markov Switching Regular Vine Copulas Stöber, Jakob and Czado, Claudia Lehrstuhl für Mathematische Statistik,

More information

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA Intro: Course Outline and Brief Intro to Marina Vannucci Rice University, USA PASI-CIMAT 04/28-30/2010 Marina Vannucci

More information

Three-Stage Semi-parametric Estimation of T-Copulas: Asymptotics, Finite-Samples Properties and Computational Aspects

Three-Stage Semi-parametric Estimation of T-Copulas: Asymptotics, Finite-Samples Properties and Computational Aspects Three-Stage Semi-parametric Estimation of T-Copulas: Asymptotics, Finite-Samples Properties and Computational Aspects Dean Fantazzini Moscow School of Economics, Moscow State University, Moscow - Russia

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Practical Bayesian Quantile Regression. Keming Yu University of Plymouth, UK

Practical Bayesian Quantile Regression. Keming Yu University of Plymouth, UK Practical Bayesian Quantile Regression Keming Yu University of Plymouth, UK (kyu@plymouth.ac.uk) A brief summary of some recent work of us (Keming Yu, Rana Moyeed and Julian Stander). Summary We develops

More information

Semiparametric Gaussian Copula Models: Progress and Problems

Semiparametric Gaussian Copula Models: Progress and Problems Semiparametric Gaussian Copula Models: Progress and Problems Jon A. Wellner University of Washington, Seattle 2015 IMS China, Kunming July 1-4, 2015 2015 IMS China Meeting, Kunming Based on joint work

More information

arxiv: v1 [stat.me] 27 Feb 2017

arxiv: v1 [stat.me] 27 Feb 2017 arxiv:1702.08148v1 [stat.me] 27 Feb 2017 A Copula-based Imputation Model for Missing Data of Mixed Type in Multilevel Data Sets Jiali Wang 1, Bronwyn Loong 1, Anton H. Westveld 1,2, and Alan H. Welsh 3

More information

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework HT5: SC4 Statistical Data Mining and Machine Learning Dino Sejdinovic Department of Statistics Oxford http://www.stats.ox.ac.uk/~sejdinov/sdmml.html Maximum Likelihood Principle A generative model for

More information

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US Gerdie Everaert 1, Lorenzo Pozzi 2, and Ruben Schoonackers 3 1 Ghent University & SHERPPA 2 Erasmus

More information

Research Division Federal Reserve Bank of St. Louis Working Paper Series

Research Division Federal Reserve Bank of St. Louis Working Paper Series Research Division Federal Reserve Bank of St Louis Working Paper Series Kalman Filtering with Truncated Normal State Variables for Bayesian Estimation of Macroeconomic Models Michael Dueker Working Paper

More information

Working Papers in Econometrics and Applied Statistics

Working Papers in Econometrics and Applied Statistics T h e U n i v e r s i t y o f NEW ENGLAND Working Papers in Econometrics and Applied Statistics Finite Sample Inference in the SUR Model Duangkamon Chotikapanich and William E. Griffiths No. 03 - April

More information

Timevarying VARs. Wouter J. Den Haan London School of Economics. c Wouter J. Den Haan

Timevarying VARs. Wouter J. Den Haan London School of Economics. c Wouter J. Den Haan Timevarying VARs Wouter J. Den Haan London School of Economics c Wouter J. Den Haan Time-Varying VARs Gibbs-Sampler general idea probit regression application (Inverted Wishart distribution Drawing from

More information

ST 740: Markov Chain Monte Carlo

ST 740: Markov Chain Monte Carlo ST 740: Markov Chain Monte Carlo Alyson Wilson Department of Statistics North Carolina State University October 14, 2012 A. Wilson (NCSU Stsatistics) MCMC October 14, 2012 1 / 20 Convergence Diagnostics:

More information

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix

Labor-Supply Shifts and Economic Fluctuations. Technical Appendix Labor-Supply Shifts and Economic Fluctuations Technical Appendix Yongsung Chang Department of Economics University of Pennsylvania Frank Schorfheide Department of Economics University of Pennsylvania January

More information

The Metropolis-Hastings Algorithm. June 8, 2012

The Metropolis-Hastings Algorithm. June 8, 2012 The Metropolis-Hastings Algorithm June 8, 22 The Plan. Understand what a simulated distribution is 2. Understand why the Metropolis-Hastings algorithm works 3. Learn how to apply the Metropolis-Hastings

More information

Copulas. MOU Lili. December, 2014

Copulas. MOU Lili. December, 2014 Copulas MOU Lili December, 2014 Outline Preliminary Introduction Formal Definition Copula Functions Estimating the Parameters Example Conclusion and Discussion Preliminary MOU Lili SEKE Team 3/30 Probability

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Multivariate Non-Normally Distributed Random Variables

Multivariate Non-Normally Distributed Random Variables Multivariate Non-Normally Distributed Random Variables An Introduction to the Copula Approach Workgroup seminar on climate dynamics Meteorological Institute at the University of Bonn 18 January 2008, Bonn

More information

Gaussian Process Vine Copulas for Multivariate Dependence

Gaussian Process Vine Copulas for Multivariate Dependence Gaussian Process Vine Copulas for Multivariate Dependence José Miguel Hernández-Lobato 1,2 joint work with David López-Paz 2,3 and Zoubin Ghahramani 1 1 Department of Engineering, Cambridge University,

More information

Correlation: Copulas and Conditioning

Correlation: Copulas and Conditioning Correlation: Copulas and Conditioning This note reviews two methods of simulating correlated variates: copula methods and conditional distributions, and the relationships between them. Particular emphasis

More information

Statistical Methods in Particle Physics Lecture 1: Bayesian methods

Statistical Methods in Particle Physics Lecture 1: Bayesian methods Statistical Methods in Particle Physics Lecture 1: Bayesian methods SUSSP65 St Andrews 16 29 August 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan

More information

Bayesian Regression Linear and Logistic Regression

Bayesian Regression Linear and Logistic Regression When we want more than point estimates Bayesian Regression Linear and Logistic Regression Nicole Beckage Ordinary Least Squares Regression and Lasso Regression return only point estimates But what if we

More information

Cross-sectional space-time modeling using ARNN(p, n) processes

Cross-sectional space-time modeling using ARNN(p, n) processes Cross-sectional space-time modeling using ARNN(p, n) processes W. Polasek K. Kakamu September, 006 Abstract We suggest a new class of cross-sectional space-time models based on local AR models and nearest

More information

Lecture 5: Spatial probit models. James P. LeSage University of Toledo Department of Economics Toledo, OH

Lecture 5: Spatial probit models. James P. LeSage University of Toledo Department of Economics Toledo, OH Lecture 5: Spatial probit models James P. LeSage University of Toledo Department of Economics Toledo, OH 43606 jlesage@spatial-econometrics.com March 2004 1 A Bayesian spatial probit model with individual

More information

POSTERIOR ANALYSIS OF THE MULTIPLICATIVE HETEROSCEDASTICITY MODEL

POSTERIOR ANALYSIS OF THE MULTIPLICATIVE HETEROSCEDASTICITY MODEL COMMUN. STATIST. THEORY METH., 30(5), 855 874 (2001) POSTERIOR ANALYSIS OF THE MULTIPLICATIVE HETEROSCEDASTICITY MODEL Hisashi Tanizaki and Xingyuan Zhang Faculty of Economics, Kobe University, Kobe 657-8501,

More information

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Review. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Review DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Probability and statistics Probability: Framework for dealing with

More information

Risk Measures with Generalized Secant Hyperbolic Dependence. Paola Palmitesta. Working Paper n. 76, April 2008

Risk Measures with Generalized Secant Hyperbolic Dependence. Paola Palmitesta. Working Paper n. 76, April 2008 Risk Measures with Generalized Secant Hyperbolic Dependence Paola Palmitesta Working Paper n. 76, April 2008 Risk Measures with Generalized Secant Hyperbolic Dependence Paola Palmitesta University of

More information

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations

The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations The Mixture Approach for Simulating New Families of Bivariate Distributions with Specified Correlations John R. Michael, Significance, Inc. and William R. Schucany, Southern Methodist University The mixture

More information

The sbgcop Package. March 9, 2007

The sbgcop Package. March 9, 2007 The sbgcop Package March 9, 2007 Title Semiparametric Bayesian Gaussian copula estimation Version 0.95 Date 2007-03-09 Author Maintainer This package estimates parameters of

More information

A Fully Nonparametric Modeling Approach to. BNP Binary Regression

A Fully Nonparametric Modeling Approach to. BNP Binary Regression A Fully Nonparametric Modeling Approach to Binary Regression Maria Department of Applied Mathematics and Statistics University of California, Santa Cruz SBIES, April 27-28, 2012 Outline 1 2 3 Simulation

More information

A simple graphical method to explore tail-dependence in stock-return pairs

A simple graphical method to explore tail-dependence in stock-return pairs A simple graphical method to explore tail-dependence in stock-return pairs Klaus Abberger, University of Konstanz, Germany Abstract: For a bivariate data set the dependence structure can not only be measured

More information

Gibbs Sampling in Latent Variable Models #1

Gibbs Sampling in Latent Variable Models #1 Gibbs Sampling in Latent Variable Models #1 Econ 690 Purdue University Outline 1 Data augmentation 2 Probit Model Probit Application A Panel Probit Panel Probit 3 The Tobit Model Example: Female Labor

More information