ON MULTIPLE TRY SCHEMES AND THE PARTICLE METROPOLIS-HASTINGS ALGORITHM

Size: px
Start display at page:

Download "ON MULTIPLE TRY SCHEMES AND THE PARTICLE METROPOLIS-HASTINGS ALGORITHM"

Transcription

1 ON MULTIPLE TRY SCHEMES AN THE PARTICLE METROPOLIS-HASTINGS ALGORITHM L. Martino, F. Leisen, J. Coraner University of Helsinki, Helsinki (Finlan). University of Kent, Canterbury (UK). ABSTRACT Markov Chain Monte Carlo (MCMC) algorithms an Sequential Monte Carlo (SMC) methos (a.k.a., particle filters) are well-known Monte Carlo methoologies, wiely use in ifferent fiels for Bayesian inference an stochastic optimization. The Multiple Try Metropolis (MTM) algorithm is an extension of the stanar Metropolis- Hastings (MH) algorithm in which the next state of the chain is chosen among a set of caniates, accoring to certain weights. The Particle MH (PMH) algorithm is another avance MCMC technique specifically esigne for scenarios where the multiimensional target ensity can be easily factorize as multiplication of conitional ensities. PMH combines jointly SMC an MCMC approaches. Both, MTM an PMH, have been wiely stuie an applie in literature. PMH variants have been often applie for the joint purpose of tracking ynamic variables an tuning constant parameters in a state space moel. Furthermore, PMH can be also consiere as an alternative particle smoothing metho. In this work, we investigate connections, similarities an ifferences among MTM schemes an PMH methos. This stuy allows the esign of novel efficient schemes for filtering an smoothing purposes in state space moels. More specially, one of them, calle particle Multiple Try Metropolis (P-MTM), obtains very promising results in ifferent numerical simulations. Keywors: Bayesian Inference; Particle Filter; Particle smoother; Markov Chain Monte Carlo (MCMC); Multiple Try Metropolis; Particle MCMC.. INTROUCTION Monte Carlo statistical methos are powerful tools for numerical inference an stochastic optimization [34, 2. Markov Chain Monte Carlo (MCMC) [5, 9, 20, 34 an Sequential Monte Carlo (SMC) algorithms (a.k.a., particle filters) [2, 7, 0, 29 are classical Monte Carlo techniques employe in orer to approximate an otherwiseincalculable (analytically) integral involving a complicate target probability ensity function (pf) [2, 20 (the posterior istribution in Bayesian inference). MCMC algorithms prouce a Markov chain with a stationary istribution that coincies with the target pf, whereas SMC methos (a.k.a., particle filters) yiel an approximation of the target measure by weighte samples. The Multiple Try Metropolis (MTM) metho of [22, [2, Chapter 5 is an avance MCMC technique, which is extension of the well-known Metropolis-Hastings (MH) algorithm [28, 3. In MTM, the next state of the chain is selecte among a set of caniates accoring to some suitable weights. This enables the MTM sampler to make large step-size jumps without a lowering of the acceptance rate an thus the exploration of a larger portion of the state space is facilitate. A famous special case of MTM, well-known in molecular simulation fiel, is the orientational This work has been supporte by the AoF grant 2570, an by the European Union 7th Framework Programme through the Marie Curie Initial Training Network Machine Learning for Personalize Meicine MLPM202, Grant No

2 bias Monte Carlo technique [. ue to its goo performance, several generalizations of the basic MTM scheme [22 can be foun in literature: with correlate caniates [26, 33, more general form of the weights an ifferent frameworks [8, 27, 3, 39, with aaptive an interacting proposal pfs [4. Interesting an relate stuies about MTM or the use of multiple auxiliary variables for builing acceptance probabilities within a MH approach can be foun in [3, 37, 25. Inepenently from the erivation of the MTM schemes, the class of Particle MCMC methos (P-MCMC) has been propose [, 6, 30, 38 in literature. P-MCMC methos are specifically esigne to solve inference problems in state space moels, combining SMC an the MCMC approaches. In this work, we focus on a specific P-MCMC metho, calle Particle Metropolis-Hastings (PMH) algorithm [, 6. The iea behin PMH is first to approximate the target measure with a elta approximation base on weighte samples obtaining by a particle filter, an then to use this approximation as proposal pf within MH technique. The Particle Marginal MH technique is a variant of the stanar PMH metho esigne in orer to estimate jointly the sequence of hien states an the static parameters of the moel [38. Note that the stanar PMH scheme can be interprete as a particle smoother [9, 0, 2, 6, 7, 36 since, at each iteration of PMH, the weighte paths obtaine by a particle filter are employe in orer to upate the previous estimation accoring to a suitable MH-type rule, taking into account all the receive observations. The authors in [ iscuss the relationships of P-MCMC with other existing techniques. They mention an escribe precisely the relationship with the so-calle configurational bias Monte Carlo metho [35,[2, Chapter 5. This technique is also strictly connecte to the MTM scheme. The authors also allue quickly to the MTM metho [22. However, the relationship between MTM an PMH eserves a more careful look. In this work, we show that MTM an PMH algorithms are strictly connecte. PMH can be interprete as an MTM using an inepenent proposal pf which generates correlate caniates, rawn an weighte sequentially through a particle filter. In orer to clarify the relationship between MTM an PMH, we recall the batch importance sampling (IS) an sequential importance sampling (SIS) methos an point out some relevant consierations about ifferent estimators of the marginal likelihoo (a.k.a, Bayesian evience) [9. Furthermore, we also introuce a variant of the MTM scheme with inepenent proposal pf. These observations are essential in orer to show the connection between MTM an PMH. This exhaustive stuy allows the esign of novel more efficient schemes, taking avantage of the ifferent alternatives an analyses alreay provie in literature about both techniques [37, 27, 30, 38. Hence, we propose novel possible MTM an PMH schemes. One of them, calle Particle MTM (P-MTM), combines the stanar PMH an MTM kernels, mixing properly the main avantages of both: the sequential construction of the ifferent tries by PMH, an the possibility given an MTM scheme of consiering several caniates perturbing the previous state of the chain. P-MTM provie excellent performance as we show with numerical results. In one simulation, we test P-MTM as a particle smoother in orer to make inference of a sequence of hien states in a stochastic volatility moel, obtaining very favorable results. The paper is structure as follows. In Section 2, we recall some require concepts about importance sampling an resampling techniques. The MTM methos are escribe in Section 3 whereas the escription of the PMH algorithms an their relationship with the MTM schemes is provie in Section 4. Novel schemes are iscusse in Section 5. Section 6 is evote to the numerical simulations an, in Section 7, we provie some conclusions. 2. IMPORTANCE SAMPLING In many applications, we esire to infer a vector of unknown parameters, x R ζ, give a set of observe ata, y R Y. In these cases, one is intereste in approximating ifferent moments the posterior ensity π(x y) that, Let us enote π(x : y :) the posterior of the hien states x : given the receive observations y :. PMH schemes can be also employe for improving the approximations of the marginal posteriors π(x y :) with, jointly with the complete posterior π(x : y :). PMH can be interprete as a way to combine properly several inepenent runs of a particle filter (after obtaining all the observations y :), generating a ergoic Markov chain with invariant istribution π(x : y :) [9.

3 hereafter, we simply enote as π(x). More specifically, in this work, we enote the variable of interest as x = x : = [x, x 2,..., x = X R ζ, where x X R ζ for all =,...,. The target ensity is inicate as π(x) = Z π(x), where Z = π(x)x, () is often known as marginal likelihoo (a.k.a., Bayesian inference). In many application, we are only able to evaluate π(x) since Z is unknown. Moreover, in general, we are not able to raw ranom samples from π(x). Monte Carlo techniques employ a simpler proposal ensity, enote as q(x), with support X R ζ, 2 for generating possible ranom caniates. Then, these caniates are filtere using some suitable proceure, in orer to prouce a particle approximation of π(x) an also to provie an estimation of Z. 2.. Batch an Sequential Importance Sampling A well-known Monte Carlo technique is the importance sampling (IS) metho. IS provies an approximation with weighte samples of the measure of π. More specifically, N samples x (),..., x (N) are rawn from a proposal pf q(x) an then they are weighte as w (n) = π(x(n) ) q(x (n), n =,..., N, (2) ) where the super-inex n in w (n) enotes the corresponing particle an the subinex refers to the imension of x, i.e., x = x : = [x,..., x. Thus, the particle approximation is π (x) = w (n) δ(x x(n) ), (3) where we have enote the normalize weights as w (n) = w(n) i= w(i). An estimation of Z is given by Ẑ = N w (n). (4) In high imensional spaces (x = X R ζ ), an equivalent sequential proceure, calle sequential importance sampling (SIS), is preferre to the previous batch approach. Recall that x = x : = [x,..., x, we can observe that a target pf π(x) can always be expresse as π(x) π(x) = γ (x ) γ (x x : ) (5) using the chain rule [32 where γ (x ) is a marginal pf an γ (x x : ) are conitional pfs. We also consier the joint probability of the partial vector x : = [x,..., x, X =2 π (x : ) = Z π (x : ) π (x : ) = γ (x ) γ j (x j x :j ), (6) 2 For the sake of simplicity, in the observations of the rest of the work, we consier the proposal function q(x) be normalize, i.e., q(x)x =. j=2

4 where Z = π (x : )x :, (7) X an, clearly, we have π (x : ) = π(x). In many applications, the target appears irectly ecompose as in Eq. (5), e.g., as in state-space moels. However, in general, one nees to marginalize several times the target π(x) for obtaining analytically the conitional pfs γ (x x : ), =,...,. Given the target in Eq. (5), we can also consier a proposal pf ecompose in the same fashion q(x) = q (x )q 2 (x 2 x ) q (x x : 2 )q (x x : ). In a batch IS scheme, given an n-th sample x (n) = x (n) : q(x), we assign the importance weight w (n) = π(x(n) ) q(x (n) ) = γ (n) (x )γ 2(x (n) 2 x(n) ) γ (x (n) x(n) : ) q (x (n) )q 2(x (n) 2 x(n) ) q (x (n) x(n) : ). The previous expression suggests a recursive proceure for computing the importance weights: starting with w (n) = π(x(n) ) q(x (n) where we have set ) an then w (n) = w (n) β(n), = β (n) j, =,...,, j= β (n) = w (n) an β (n) = γ (x (n) x(n) : ) q (x (n) x(n) (8) : ), (9) for = 2,...,. Thus, given N samples x (),..., x (N), finally we obtain the particle approximations of the sequence of pfs π (x : ) as π (x : ) = w (n) an an estimator of each normalizing constant Z is given by Ẑ = N δ(x : x (n) ), =,...,, (0) w (n) = N : β (n) j j=. () However, an alternative equivalent formulation is often use [ N Z =, (2) = = j= j= [ Ẑj j= w (n) j β(n) j [ w(n) j w(n) j Ẑ j, (3) = Ẑ Ẑ 2... Ẑ = Ẑ 0 Ẑ Ẑ Ẑ, (4) where, for simplicity, we have set Ẑ0 =. A alternative erivation of the (final) estimator Z is given in Appenix A. Remark. In SIS, there are two equivalent formulations, Ẑ in Eq. (4) an Z in Eq. (2) of estimator of Z.

5 2.2. Sequential Importance Resampling (SIR) Sequential Importance Resampling (SIR) [2, 34 combines the sequential construction of the importance weights as in SIS with the application of resampling steps [7, 8. Namely, when some pre-establishe criterion is fulfille [7, 8, 23, N inepenent particles are rawn accoring to the probability mass π (x : ). Then, the resample particles are propagate for proviing the next approximation π + (x :+ ). More specifically, let us consier that a resampling step is performe at the -th iteration. Hence, N samples x (j) : are rawn from π (x : ), an then the corresponing weights are set to the same value [7, 8. A proper choice [24 is to set the unnormalize importance weights w (n) = Ẑ, j =,..., N. (5) i.e., w () = w (2) that w (x (n) : ) = N following weights =... = w (N), for all j =,..., N. One reason why this is a goo choice, for instance, is that efining the, equal for each resample particle x (n) :. Hence, after a resampling step, we have ξ (n) ) = { (n) w, without resampling at -th iteration, Ẑ, with resampling at -th iteration. (6) then, in any case, N N ξ(n) = Ẑ, as expecte. Therefore, the weight recursion for SIR becomes ξ (n) = ξ (n) β(n), where ξ(n) = { (n) ξ, without res. at ( )-th iter., Ẑ, with res. at ( )-th iter. (7) See Appenix A for further etails. Remark 2. With the recursive efinition of the weights ξ (n) Ẑ = N ξ (n) β(n), Z = in Eq. (7), the two estimators [ j= ξ (n) j β(n) j (8) (n) where ξ j = ξ(n) j, are both vali an equivalent estimators of Z [24. i= ξ(i) j For instance, if the resampling is applie at each iteration, observe that they become an Z = [ N j= [ Ẑ = Ẑ N β (n) = β (n) j j= [ N, (9) β (n) j, (20) an clearly coincie. Note that, w.r.t. the estimator in Eq. () (for SIS, i.e., without resampling), the operations of prouct an sum are inverte. Figure 2 epicts ifferent examples of generation of weighte samples x (n) with or without employing resampling steps. More specifically, Figure 2 shows the components x, (n)..., x (n) of each sample, with = 0. Remark 2 is necessary to to escribe exhaustively the relationship between MTM an PMH algorithms. Below, we recall the MTM schemes an iscuss a novel suitable variant in orer to link MTM to PMH.

6 Table. Generic MTM algorithm.. Choose a initial state x 0 an the total number of iterations K. 2. For k =,..., K: (a) raw N samples from x (i) q(x x k ), i =,..., N. (b) Choose one sample x {x (),..., x (N) } with probability proportional to the importance weights Namely, raw a sample x from w (i) = π(x(i) ), i =,..., N. q(x (i) x k ) π (x) = w (n) δ(x x(n) ). (c) raw N auxiliary samples z (j) q(x x ), j =,..., N, an set z (N) = x k. () Compute the importance weights also for the auxiliary points, ρ (i) = π(z(i) ), i =,..., N. q(z (i) x ) (e) Set x k = x with probability α = otherwise, with probability α, set x k = x k. i= w(i), i= ρ(i) 3. MULTIPLE TRY METROPOLIS (MTM) ALGORITHMS The Multiple Try Metropolis (MTM) algorithm [22 is an avance MCMC technique, where N caniates are generate each iterations. Accoring to some suitable weights, one caniate is chosen an accepte as new state with a suitable probability α. The MTM steps with a generic proposal q(x x k ), epening on the previous state, are summarize in Table where we have enote a b = min[a, b. For N =, the MTM algorithm becomes the stanar Metropolis-Hastings (MH) metho [2, 34. We consier importance weights for facilitating the comparison with other techniques. However, ifferent kin of weights coul be applie [22, 27. The MTM metho generates a reversible Markov chain that converges to π(x) [22, 27. If the proposal pf is inepenent from the previous state of the chain, i.e., q(x), the algorithm can be simplifie. inee, the steps 2c an 2 can be remove in the MTM scheme. Namely, one oes not nee to generate the auxiliary samples at step 2c. Inee, in this case, we coul irectly set z (j) = x (j), j =,..., N. The simplifie MTM algorithm (I-MTM) is given in Table. A graphical representation of a MTM scheme is provie in Figure, with = an N = 2. Alternative version of the I-MTM metho (I-MTM2). In this work, we highlight that the I-MTM metho can be esigne in an alternative way. With a proposal pf inepenent from the previous state, we have seen that we can set z (j) = x (j), j =,..., N, because each x (j) is itself rawn from q(x). With the same argument, we can also use the samples generate in the previous iteration of the algorithm as auxiliary points, since all the samples are generate inepenently from the same proposal pf. Namely, the alternative suitable version of the I-MTM is summarize in Table 3. Note that, in this case, we can write the acceptance probability α as α = Ẑ Ẑ (k )

7 q(x x k ) w () w (2) x = x (2) (2) q(z x ) () =^ w() + w (2) () + (2) x () x x (2) k z (2) = x z () k Forwar Backwar Test x k = x or x k = x k Fig.. Sketch of a generic MTM metho with = an N = 2 tries. In this example, the secon caniate is selecte as x = x (2). It has been selecte with probability w () an z (2) = x k. = w() w () +w(2). The auxiliary points are z () q(z x ) Table 2. MTM with inepenent proposal (I-MTM).. Choose a initial state x 0 an the total number of iterations K. 2. For k =,..., K: (a) raw N samples from x (i) q(x), i =,..., N. (b) Choose one sample x {x (),..., x (N) } with probability proportional to the importance weights = π(x(i) ), i =,..., N. q(x (i) ) w (i) Moreover, we enote as w an w,k, the weights corresponing to x an x k, respectively. (c) Set x k = x with probability α = i= w(i) i= w(i) w + w,k = i= w(i), (2) i= ρ(i) where the values ρ (i) enote the importance weights of {z(),..., z (N) } = {x (),..., x (N) } \ {x } {x k }. Otherwise, set x k = x k. where Ẑ an Ẑ(k ) are both estimators of Z. 4. PARTICLE METROPOLIS-HASTINGS ALGORITHM AN ITS RELATIONSHIP WITH MTM Consier that we are able to factorize the target ensity as π(x) π(x) = γ (x )γ 2 (x 2 x ) γ (x x : ). The Particle Metropolis Hastings (PMH) metho [ is another MCMC technique propose inepenently from the MTM algorithm, specifically esigne for being applie in this framework. The complete escription is provie in Table 4. At each iteration, a particle filter is run in orer to provie a particle approximation by N weighte samples

8 Table 3. Alternative I-MTM algorithm (I-MTM2).. Choose a initial state x 0, the total number of iterations K an obtain an estimation Ẑ(0) Z. 2. For k =,..., K: (a) Choose one sample x {x (),..., x (N) } with probability proportional to the importance weights (b) Set x k = x an Ẑ(k) = Ẑ = N w (i) i= w(i) = π(x(i) ), i =,..., N. q(x (i) ) with probability α = N N i= w(i) = Ẑ (k ) Ẑ Ẑ (k ) otherwise, with probability α, set x k = x k an Ẑ(k) = Ẑ(k ). Table 4. Particle Metropolis-Hastings (PMH) algorithm.. Choose a initial state x 0, the total number of iterations K an obtain an estimation Ẑ(0) Z. 2. For k =,..., K: (a) Using a proposal pf of type q(x) = q (x )q 2(x 2 x ) q (x x : ), we employ SIR (see Section 2.2) for rawing with N particles an weighting properly them, {x (i), w (i) }N i=. Namely, we obtain a particle approximation of the measure of target pf π (x) = i= Furthermore, we also obtain Ẑ in Eq. (), or Z in Eq. (2). w (i) δ(x x(i) ). (b) raw x π(x), i.e., choose a particle x = {x (),..., x (N) } with probability w (i), i =,..., N. (c) Set x k = x an Ẑ(k) = Ẑ with probability α = Ẑ, (22) Ẑ (k ) otherwise set x k = x k an Ẑ(k) = Ẑ(k ). of the measure of the target. Then, a sample among the N weighte particles is chosen by one resampling steps. This selecte sample is then accepte or rejecte as next state of the chain accoring to an MH-type acceptance probability, which involves two estimators of Z. Both estimators Ẑ an Z can be use in PMH (although the original algorithm is escribe with the use of Z [), if the resample particles are properly weighte as shown in Eq. (5) [24. A generalization of PMH for hanling both ynamic an fixe parameters (as hien states an parameters of the moel, respectively), calle Particle Marginal MH algorithm, is escribe in Appenix B. Smoothing. PMH is a particle smoother since the outputs of ifferent run of a particle filter (i.e., at each run, N weighte paths obtaine with a SIR proceure) are further processe through the MH acceptance functions, taking into account all the receive observations.

9 6 4 2 x (5) x (2) x (3) = x (3) : x () x (4) (a) Batch-IS or SIS with q (x ) = N (x 0, 2) an N = (b) Batch-IS or SIS with q (x x ) an N = (c) SIR using q (x x ) an resampling at the iterations = 4, 8 (with N = 40). Fig. 2. Examples of application of the IS technique. We consier as target ensity a multivariate Gaussian pf, π(x) = 0 = N (x 2, 2 ). In each figure, every component of ifferent particles are represente, so that each particle x (i) forms a path. The normalize weights w n corresponing to each figure are also shown. The line-with of each path is proportional to the corresponing weight w n. The particle corresponing to the greatest weight is always epicte in black. (a) Batch IS or SIS with N = 5 particles an q(x) = = N (x 0, 2). (b) Batch IS or SIS with N = 40 particles an q(x) = N (x 2, ) =2 N (x x, ). (c) SIR with N = 40 particles an q(x) = N (x 2, ) =2 N (x x, ) an resampling steps at the iterations = 4, Relationship between MTM an PMH A simple look at the alternative version of the MTM technique with inepenent proposal (I-MTM2), introuce in Section 3, an the PMH metho, shows that are strictly relate. Inee, the structure of the two algorithms coincies. The connections an ifferences are liste below: The main ifference lies that the caniates in PMH are generate sequentially, using a SMC proceure (i.e., by a particle filter). If the resampling steps in the SMC are not applie them I-MTM2 an PMH are exactly the same algorithm, where the caniates are rawn in a batch setting or sequential way. Namely, I-MTM2 generates irectly x (i) = [x (i) from a pf q(x) in the space x whereas PMH raws sequentially each component x (i),..., x(i) of x from q (x x (i) : ) (see also Figure 4, for further clarifications). The use of resampling steps is the main ifference between the generation proceures of PMH an I-MTM2. Owing to the use of the resampling, the caniates {x (),..., x (N) } propose by PMH are not inepenent, ifferently from the MTM schemes consiere in this work. Without resampling, the generate samples x (i) = x (i) : woul be inepenent as in I-MTM2. The generation of correlate samples can be also consiere in MTM methos, as simply shown for instance in [5, without jeoparizing the ergoicity of the chain. Thus, more precisely, PMH can be consiere as an I-MTM2 scheme using correlate samples (e.g., as in [5), an where the caniates are generate sequentially. For clarifying this point, in Figure 2 we show ifferent particles weighte with IS weights (the line-with of each path is proportional to the corresponing normalize weight w n ). More specifically, we represent each component of x (n), =,..., = 0 of each particle x(n) = x (n) :0 with n =,..., N {5, 40}. The target ensity is a multivariate Gaussian pf, π(x) = 0 = N (x 2, 2 ), i.e., with expecte value µ = 2, for =,..., 0. Figures 2(a)-(b) correspons to the application of IS with two ifferent proposal pfs an without resampling. In Figure 2(a), the components x (n) are inepenent. In Figure 2(b), the components x (n)

10 within each sample x (n) are correlate, but the samples x (n), n =,..., N, are still inepenent. In Figure 2(c) two resampling are also applie at the iterations = 4, 8, generating correlation among the particles x (n), n =,..., N, as well. Figure 2(c) correspons to the sample generation in PMH. In the correspoing stanar formulations, I-MTM2 uses the estimator Ẑ in Eq. (4) whereas PMH has been propose using Z, given in Eq. (2). However, they are equivalent formulation of an estimator of the normalizing constant Z (see Remark 2), if the resample particles are proper weighte [24. The PMH approach is preferable in high imension, when the target can be factorize as in Eq. (5), since the use of the resampling steps can provie a better proposal generation proceure. 5. NOVEL SCHEMES The previous consierations also allow us to esign novel PMH schemes. For instance, we can easily suggest an alternative proper acceptance probability, α = NẐ NẐ w + w. (23),k We enote as var-pmh the variant of the PMH technique which uses the probability α above, instea of the probability α in Eq. (22). Namely, var-pmh is ientical with the PMH metho in Table 4, replacing Eq. (22) with Eq. (23). The var-pmh structure is equivalent (within a sequential framework) to I-MTM of Table 2, in the same fashion as PMH in Table 4 is equivalent to I-MTM2 of Table 3. State epenent PMH (S-PMH). Moreover, we can also exten the stanar PMH metho employing a state-epenent proposal pf (epenent from the previous state), instea of an inepenent proposal (namely, inepenent from the previous state) as in Table 4. This novel scheme, enote as S-PMH, is outline in Table 5. In this case, the generation of a backwar path is require at step 2c. Hence, in S-PMH, we have this aitional computation cost. However, the generate backwar paths coul be also recycle for estimating the hien states (nevertheless, this requires an eserves more specific analysis). The valiity of S-PMH is ensure since it correspons to the MTM scheme in Table. In S-PMH, the approximation π is provie consiering with correlate samples ue to the resampling, unlike in MTM. However, it oes not jeoparize the ergoicity (as shown, e.g., in [5). Furthermore, we consier the use of resampling steps only at certain 0 R K pre-establishe iterations,,..., R. If R = 0, no resampling it is applie so that we obtain a stanar MTM scheme with a sequential generation of the tries. If R = K, the resampling is applie at each iteration, so that we have a bootstrap filter for generating the samples [7, 9. Figure 3(a) shows a sketch of the ifferent schemes iscusse in this work. The MTM schemes are given on the left sie, whereas the corresponing PMH approaches are provie on the right. The boxes with ashe contours represent the novel schemes introuce in this work. As an example of proposal in S-PMH scheme, we can consier q (s s :, x :,k ) = q (s s, x,k ), so that the complete proposal is q(s x k ) = q (s x,k ) =2 q (s s, x,k ). However, it is not straightforwar to choose an tune properly the components q (s s, x,k ). Particle Multiple Try Metropolis (P-MTM) algorithm. A simpler an robust scheme consists in performing alternatively a stanar PMH kernel (or var-pmh), enote as K P MH (x t x t ) an an MTM kernel with a ranom walk proposal pf, enote as K MT M (x t x t ). Namely, Particle Multiple Try Metropolis (P-MTM) algorithm is forme by the following steps:

11 . Choose an initial state x 0, an set t =. 2. While t < T : (a) Generate a new state using one step of PMH, i.e., x t K P MH (x x t ), generating sequentially with an inepenent proposal pf q(x) = q (x ) =2 q (x x ), an set t = t +. (b) Generate a new state using one step of MTM, x t K MT M (x x t ), using a ranom walk proposal pf, i.e., q(x x t ), in orer to raw the N caniates, an set t = t +. The two kernels are linke together an, since each kernel leaves invariant the target pf π, it is to show that the prouct of two vali kernels also leaves invariant π (see Appenix C). P-MTM exploits the avantage of selecting N particles component by component using resampling steps (as in PMH) an, jointly, the possibility of proposing N caniates taking in account the previous state of the chain. Figure 3(b) provies a graphical representation of P-MTM an Figure 4 gives a sketch of the two generation approaches jointly use in P-MTM. Gen. MTM S-PMH Particle MTM algorithm I-MTM I-MTM2 variant of PMH PMH K PMH (x t x t ) K MTM (x t+ x t ) (a) (b) Fig. 3. (a) Graphical representation of the MTM methos an the corresponing PMH schemes. The boxes with ashe contours contain the novel schemes presente in this work. (b) Graphical sketch of the P-MTM algorithm. 5Y x (n) q(x) =q (x ) q(x x ) =2 x 2,k x (n) q(x x k ) x k x 5,k x (n) x,k x (n) x 3,k x 4,k (a) Sequential generation of a caniate x (n) (b) Generation with a ranom walk proposal of x (n). Fig. 4. Examples of generation of one caniate x (n) (with = 5) (a) with a sequential approach as in PMH, without consiering resampling steps; (b) from a ranom walk proposal q(x x k ) which takes into account the previous state of the chain x k, for instance, q(x x k ) = 5 = q (x x,k ) (if each component is rawn inepenent from the others).

12 6.. Comparison among ifferent particle schemes 6. NUMERICAL SIMULATIONS In orer to the the ifferent techniques, we consier a multiimensional Gaussian target ensity, π(x) = π(x,..., x ) = N (x µ, σ 2 ), (24) with x = x : R, = 0, with µ :3 = 2, µ 4:7 = 4, µ 8:0 =, an σ = 2. We apply I-MTM, I-MTM2, PMH, Var-PMH (using the acceptance probability in Eq. (23)) an P-MTM for estimating the vector µ :0. In each metho, we employ Gaussian-piece proposal pf for the sequential proposal construction of the N caniates, i.e. = q(x x ) = N (x x, σ 2 p), setting σ p = 2. For the ranom walk MTM (RW-MTM) part within P-MTM, we consier a Gaussian proposal q rw (x x k ) = = N (x x,k, σrw) 2 with σ rw = an x k = [x,k,..., x,k is the previous state of the chain. For all the PMH schemes, we consier to perform resampling at each iteration (in I-MTM an I-MTM2, clearly no resampling is applie). We test the techniques consiering ifferent value of number of particles N an number of iterations of the chain K. We compute the MSE in estimating the vector µ :0, averaging over 500 inepenent simulations. The starting particles, =, are chosen ranomly x (i) N (x; 2, 4), for i =,..., N, at each run an for each metho. Figures 5(a)-(b) show the MSE as function of number of iterations K in semilog scale, keeping fixe the number of tries N = 3. Figure 5(a) reports the results of the MTM schemes whereas Figure 5(b) reports the results of the PMH schemes. Figure 5(c) epicts the MSE in the estimation of µ :0 of function of N, for the PMH methos. These results show that the use of an acceptance probability of type in Eq. (2)-(23) provie smaller MSE. This is more evient for small number of caniates N. Namely, the use of acceptance probability in Eq. (23) within a PMH is preferable since provies better performance. When N grows, the performance of PMH an var-pmh methos becomes similar, since the acceptance probability approaches, in both cases. P-MTM provies excellent results, outperforming the other schemes. The MSE vanishes to zero when N increases, as expecte, confirming the valiity of the novel schemes. The results also shows that performing resampling at each iteration is not optimal an that a smaller rate of resampling steps coul improve the performance [7, 9. Figure 5() epicts 35 ifferent states x k = x :0,k at ifferent iteration inices k, obtaine with var-pmh (N = 000 an K = 000) an the values µ :0 shown in ashe line Inference in a stochastic volatility moel We consier a stochastic volatility moel where the hien state x R at the -th iteration follows an AR() process an represents the log-volatility [4 of a financial time series at time t N. The state space moel is given by { x = αx + u, y = exp ( x 2 ) v, =,...,. (25) where α = 0.9 is the AR parameter, an u an v enote inepenent zero-mean Gaussian ranom variables with variances σu 2 = an σv 2 = 0.5, respectively. Note that v is a multiplicative noise. We observe a sequence y : R, an we esire to infer the hien states x :, analyzing the joint posterior π(x : y : ). A classical particle filtering (PF) approach provies a particle approximation of the posterior π(x : y : ) an, ue to the Monte Carlo approach, also of the marginal posteriors π(x y : ) with (use for smoothing). Unfortunately, when <<, the stanar PF strategy in general fails: the marginal

13 I MTM2 I MTM 0 PMH var PMH P MTM MSE MSE K (a) MTM schemes K (b) PMH schemes. 0 PMH var PMH P MTM 4 2 MSE N (c) PMH schemes () ifferent States of var-pmh. Fig. 5. (a)-(b) MSE versus number of iterations K of the chain in semilog scale, fixing the number of particles N = 3. (a) I-MTM (soli line) an I-MTM2 (ashe line). (b) PMH (soli line), var-pmh (triangles an ashe line) an P-MTM (stars an ashe line). (c) MSE versus N of the chain in semilog scale for PMH (soli line), var- PMH ( triangles an ashe line) an P-MTM (stars an ashe line). () ifferent states x k = x :0,k at ifferent iteration inices k, obtaine with var-pmh (N = 000 an K = 000). The values µ :0 are shown in ashe line (µ :3 = 2, µ 4:7 = 4 an µ 8:0 = ).

14 istribution π(x y : ) (use for filtering) occupies a privilege role within the particle filter framework as it is better characterize than any of the other marginal posterior istributions, as pointe out in [9. PMH schemes can be also employe for improving the approximations of the marginal posteriors π(x y : ) with as well as of the complete posterior π(x : y : ). Inee, PMH can be interprete as a way to combine properly several inepenent runs of a particle filter (after obtaining all the observations y : ), generating a ergoic Markov chain with invariant istribution π(x : y : ). We test the stanar PMH technique [ an the Particle MTM (P-MTM) metho escribe in Section 5. In both cases, for the sequential proposal part, we employ a stanar bootstrap particle filter (i.e., using the transition probability p(x x ) given by the test moel as proposal an with resampling at each iteration). For the stanar MTM block in P-MTM we use q(x x k ) = N (x x,k, σp), 2 = as proposal pf, with σ p = 0.5. We have average the results over 500 inepenent runs. At each run, we generate a ifferent of observations y : from the moel, setting = 00. Clearly, both schemes are compare with the same number of evaluations of the posteriors (i.e., for instance fixing N, PMH performs K iterations of the chain, an P-MTM is forme by K 2 of PMH steps an K 2 of MTM steps, as escribe in Section 5). First, we consier N = 0 an varies the number of total iterations of the algorithms, K from K = 2 to K = 500. The results are shown in Fig. 6(a). Then, we fix K = 50 an test ifferent values of N, from N = 0 to N = 000. The results are shown in Fig. 6(b). We can observe that always P-MTM outperforms PMH obtaining smaller values of MSE (with the exception of one value with a very small number of iterations, K = 2, in Fig. 6(a)) P MTM PMH P MTM PMH MSE MSE K N (a) MSE versus K. (b) MSE versus N. Fig. 6. (a) MSE versus number of iterations K of the generate chain in semi-log scale, fixing the number of particles (N = 0). (b) MSE versus number of particles N in semi-log scale, fixing the number of iterations of the algorithms (K = 50). 7. CONCLUSIONS In this work, we have highlighte the strong connection between MTM an PMH algorithms. Specifically, PMH can be interprete as an MTM using correlate caniates, rawn an weighte sequentially through a particle filter.

15 We have employe this consieration in orer to introuce novel MTM an PMH schemes, mixing properly both approaches. For instance, the Particle MTM (P-MTM) algorithm combines the stanar PMH an MTM kernels, using the sequential construction of the ifferent tries obtaine by PMH, an the possibility of perturbing the previous state of the chain using a MTM with ranom-walk proposal ensity. P-MTM is a very efficient technique for filtering an smoothing in state space moel. We have teste P-MTM, jointly with other techniques, in ifferent numerical simulations obtaining excellent results. As future line, we plan to esign the marginal version of P-MTM for the joint purpose of inferring ynamic an static parameters in a state space framework. References [ C. Anrieu, A. oucet, an R. Holenstein. Particle Markov chain Monte Carlo methos. J. R. Statist. Soc. B, 72(3): , 200. [2 M. S. Arulumpalam, S. Maskell, N. Goron, an T. Klapp. A tutorial on particle filters for online nonlinear/non- Gaussian Bayesian tracking. IEEE Transactions Signal Processing, 50(2):74 88, February [3 M. Béar, R. ouc, an E. Mouline. Scaling analysis of multiple-try MCMC methos. Stochastic Processes an their Applications, 22: , 202. [4 R. Casarin, R. Craiu, an F. Leisen. Interacting multiple try algorithms with ifferent proposal istributions. Statistics an Computing, 23(2):85 200, 203. [5 R. V. Craiu an C. Lemieux. Acceleration of the Multiple-Try Metropolis algorithm using antithetic an stratifie sampling. Statistics an Computing, 7(2):09 20, [6 J. ahlin, F. Linsten, an T. B. Schn. Particle Metropolis Hastings using Langevin ynamics. In 203 IEEE International Conference on Acoustics, Speech an Signal Processing, pages , May 203. [7 P. M. jurić, J. H. Kotecha, J. Zhang, Y. Huang, T. Ghirmai, M. F. Bugallo, an J. Míguez. Particle filtering. IEEE Signal Processing Magazine, 20(5):9 38, September [8 A. oucet, N. e Freitas, an N. Goron, eitors. Sequential Monte Carlo Methos in Practice. Springer, New York (USA), 200. [9 A. oucet an A. M. Johansen. A tutorial on particle filtering an smoothing: fifteen years later. technical report, [0 W. Fong, S. J. Gosill, A. oucet, an M. West. Monte Carlo smoothing with application to auio signal enhancement. IEEE Transactions on Signal Processing, 50(2): , February [. Frenkel an B. Smit. Unerstaning molecular simulation: from algorithms to applications. Acaemic Press, San iego, 996. [2 S. Gosill, A. oucet, an M. West. Monte Carlo smoothing for nonlinear time series. Journal of the American Statistical Association, 99(465):56 68, March [3 W. K. Hastings. Monte Carlo sampling methos using Markov chains an their applications. Biometrika, 57():97 09, 970. [4 E. Jacquier, N. G. Polson, an P. E. Rossi. Bayesian analysis of stochastic volatility moels. Journal of Business an Economic Statistics, 2(4):37 389, October 994.

16 [5 L. Jing an P. Vaakkepat. Interacting MCMC particle filter for tracking maneuvering target. igital Signal Processing, 20:56 574, August 200. [6 G. Kitagawa. Monte Carlo filter an smoother for non-gaussian nonlinear state-space moels. J. Comput. Graph. Statist., : 25, 996. [7 G. Kitagawa an S. Sato. Monte Carlo smoothing an self-organising state-space moel. In A. oucet, N. e Freitas, an N. Goron, eitors, Sequential Monte Carlo Methos in Practice, chapter 9, pages Springer, 200. [8 G. Kobayashi an H. Kozumi. Generalize multiple-point Metropolis algorithms for approximate bayesian computation. Journal of Statistical Computation an Simulation, 85(4), 205. [9 J. R. Larocque an P. Reilly. Reversible jump MCMC for joint etection an estimation of sources in colore noise. IEEE Transactions on Signal Processing, 50(2), February 998. [20 F. Liang, C. Liu, an R. Caroll. Avance Markov Chain Monte Carlo Methos: Learning from Past Samples. Wiley Series in Computational Statistics, Englan, 200. [2 J. S. Liu. Monte Carlo Strategies in Scientific Computing. Springer, [22 J. S. Liu, F. Liang, an W. H. Wong. The Multiple-Try metho an local optimization in Metropolis sampling. Journal of the American Statistical Association, 95(449):2 34, March [23 L. Martino, V. Elvira, an F. Louzaa. Alternative Effective Sample Size measures for Importance Sampling. IEEE Workshop on Statistical Signal Processing (SSP), June 206. [24 L. Martino, V. Elvira, an F. Louzaa. Weighting a resample particle in Sequential Monte Carlo. IEEE Workshop on Statistical Signal Processing (SSP), June 206. [25 L. Martino an F. Louzaa. Issues in the Multiple Try Metropolis mixing. Computational Statistics (to appear), pages 4, 206. [26 L. Martino, V. P. el Olmo, an J. Rea. A multi-point Metropolis scheme with generic weight functions. Statistics & Probability Letters, 82(7): , 202. [27 L. Martino an J. Rea. On the flexibility of the esign of Multiple Try Metropolis schemes. Computational Statistics, 28(6), [28 N. Metropolis, A. Rosenbluth, M. Rosenbluth, A. Teller, an E. Teller. Equations of state calculations by fast computing machines. Journal of Chemical Physics, 2:087 09, 953. [29 J. Míguez, T. Ghirmai, M. F. Bugallo, an P. M. jurić. A sequential Monte Carlo technique for blin synchronization an etection in frequency-flat Rayleigh faing wireless channels. Signal Processing, 84(): , November [30 I. Nevat, G. W. Peters, an J. Yuan. Channel tracking in relay systems via particle MCMC. In IEEE Vehicular Technology Conference, pages 5, Sept 20. [3 S. Panolfi, F. Bartolucci, an N. Friel. A generalize multiple-try version of the reversible jump algorithm. Computational Statistics & ata Analysis (available online), 203. [32 A. Papoulis. Probability, Ranom Variables an Stochastic Processes. McGraw-Hill Series in Electrical Engineering, 984.

17 [33 Z. S. Qin an J. S. Liu. Multi-Point Metropolis metho with application to hybri Monte Carlo. Journal of Computational Physics, 72: , 200. [34 C. P. Robert an G. Casella. Monte Carlo Statistical Methos. Springer, [35 J. I. Siepmann an. Frenkel. Configurational bias Monte Carlo: a new sampling scheme for flexible chains. Molecular Physics, 75():59 70, 992. [36 P. Stavropoulos an. M. Titterington. Improve particle filters an smoothing. In A. oucet, N. e Freitas, an N. Goron, eitors, Sequential Monte Carlo Methos in Practice, chapter 4, pages Springer, 200. [37 G. Storvik. On the flexibility of Metropolis-Hastings acceptance probabilities in auxiliary variable proposal generation. Scaninavian Journal of Statistics, 38(2): , February 20. [38 T. Vu, B. N. Vo, an R. Evans. A particle marginal Metropolis-Hastings multi-target tracker. IEEE Transactions on Signal Processing, 62(5): , Aug 204. [39 Y. Zhang an W. Zhang. Improve generic acceptance function for multi-point Metropolis algorithms. 2n International Conference on Electronic an Mechanical Engineering an Information Technology, 202. A. ALTERNATIVE FORMULATION OF THE ESTIMATOR OF Z In SIS approach, there are two possible equivalent formulations of the estimators of Z, the first one Ẑ in Eqs. (4)- () an the secon one Z given in Eq. (2). This alternative formulation can be also erive as follows. Consier the following integrals, Z = π (x : )x : Ẑ = w (n) X N, (26) an Clearly, we can write X γ (x x : ) π (x : )x : = X γ (x x : ) π (x : )x : = = π (x : ) X π (x : ) π (x : )x :, (27) = Z Z. (28) γ (x x : ) X q (x x : ) q (x x : ) π (x : )x :, X β (x x : )q (x x : ) π (x : )x :, (29) where we have set β (x x : ) = γ (x x : ) q (x x : ). Replacing π (x : ) with π (x : ) given in Eq. (0), β (x x : )q (x x : ) π (x : )x : = X = = w (n) w (n) X β (x x : )q (x x : )δ(x : x (n) : )x :, X β (x x (n) : )q (x x (n) : )x.

18 Hence, using again Monte Carlo for approximating each integral within the sum, i.e., given N samples x (n) q (x x (n) : ), n =,..., N (one sample for each ifferent q ( x (n) : )), an enoting β(n) = β (x (n) x(n) : ), we obtain X β (x x : )q (x x : ) π (x : )x : = where we have use w (n) = = = = w (n) β(n), (30) i= w(i) i= w(i) N w(n) N i= w(i) w (n) β(n), w (n), = Ẑ Ẑ w(n), the recursive expression of the weights, w (n) i= w(i) = w (n) estimator in Eq. (26). Finally, we can obtain, setting Ẑ0 =, [ Ẑ Ẑ 2 Z = = Ẑ Ẑ Ẑ Ẑ = Ẑ Ẑ 2 Ẑ = that is exactly the estimator in Eq. (2). A.. Application of resampling = i= w (x (i) : )β (x (i) x(i) : ) β(n) Z Z, (3), an Ẑ is the Z, (32) Let us consier to approximate the integral in Eq. (30) via importance sampling, assuming in this case to raw N samples, x () :,..., x(n) :, from q (x x : ) π (x : ), hence we can write Moreover using Eq. (27), we have N X β (x x : )q (x x : ) π (x : )x : N β(n) Z Z. β (n). (33) B. PARTICLE MARGINAL METROPOLIS-HASTINGS (PM-MH) ALGORITHM The Particle Marginal Metropolis-Hastings (PM-MH) algorithm is a simple extension of the PMH metho for the combine sampling of ynamic an fixe unknown parameters, enote as x an θ, respectively. Let us consier the following state space moel { q (x x, θ), (34) l (y x, θ) where q represents a transition probability, an l is the likelihoo function. The parameter θ Θ is consiere also unknown so that the inference problem consists in inferring (x :, θ) given the sequence of receive measurements y :. With respect to the notation use in Section 2., we have γ (x θ) = l (y x, θ)q (x θ), an γ (x x :, θ) = l (y x, θ)q (x x, θ),

19 with = 2,...,. Hence, consiering also a prior p(θ) over θ, an x = x :, y = y :, the complete target is π(x, θ y) = π(x y, θ)p(θ y), (35) = π(x y, θ) p(y θ)p(θ), (36) p(y) = π(x, y θ) p(θ) p(y), (37) [ p(θ) = l (y x, θ)q (x θ) l (y x, θ)q (x x, θ) p(y). (38) =2 We can evaluate π(x, y θ) π(x y, θ), it is not an issue using a self-normalize IS approach for approximating π(x y, θ). However, we cannot evaluate p(θ y), p(y θ) an p(y). Let us consier to apply a stanar MH metho for sampling from π(x, θ y). We assume possible to raw samples [x, θ as proposal pf q(θ, x θ k ) = q θ (θ θ k ) π(x y, θ ), where k =,..., K is the iteration of the chain an π(x y, θ) is the posterior of x. Assume hypothetically that it is possible to raw from q(θ k, x k θ k ), we obtain the following acceptance probability α = π(x, θ y)q(θ k, x k θ ) π(x k, θ k y)q(θ, x θ k ), (39) = π(x, θ y)q θ (θ k θ ) π(x k y, θ k ) π(x k, θ k y)q θ (θ θ k ) π(x y, θ ). (40) Then, since π(x, θ y) = π(x y, θ)p(θ y), we can replace it into the expression above α = π(x y, θ )p(θ y)q θ (θ k θ ) π(x k y, θ k ) π(x k y, θ k )p(θ k y)q θ (θ θ k ) π(x y, θ ), (4) = p(θ y)q θ (θ k θ ) p(θ k y)q θ (θ θ k ), (42) = p(y θ )p(θ )q θ (θ k θ ) p(y θ k )p(θ k )q θ (θ θ k ). (43) The problem is that, in general, we are not able to evaluate the likelihoo function Z(θ) = p(y θ) = π(x, y θ)x. However, we can approximate Z(θ) via importance sampling. Thus, the iea is to use the approximate proposal pf q(θ, x θ k ) = q θ (θ θ k ) π(x y, θ ), where π is a particle approximation of π obtaine by SIR an, at the same, we get the estimation Ẑ(θ ). Therefore, the PM-MH algorithm can be summarize as following:. For k =,..., K : (a) raw θ q θ (θ θ k ) an then x π(x y, θ ) via SIR. (b) Set [θ k, x k = [θ, x with probability otherwise set [θ k, x k = [θ k, x k. Ẑ(θ )p(θ )q θ (θ k θ ) α = Ẑ(θ k )p(θ k )q θ (θ θ k ) Given the observations provie in this work, PM-MH can be seen as a combination of a MH metho w.r.t. θ an a MTM-type metho w.r.t. x.

20 C. INVARIANT ENSITY OF P-MTM Let us consier two MCMC kernels, K P MH (y x) an K MT M (z y) with x, y, z R x, corresponing to the PMH an MTM steps in the P-MTM scheme, respectively. We assume π( ) is the invariant ensity of both chains. The two kernels have been esigne such that K P MH (y x) π(x)x = π(y), K MT M (z y) π(y)y = π(z). In P-MTM, the kernels K P MH, K MT M are use sequentially. Namely, first a sample is rawn from y K P MH (y x) an then z K MT M (z y ). The complete transition probability from z to x is given by K P MT M (z x) = K MT M (z y)k P MH (y x)y. (44) The target π is also invariant w.r.t. K P MT M (z x) [34, 2, 20. Inee, we can write K P MT M (z x) π(x)x = [ = K MT M (z y)k P MH (y x)y π(x)x, [ = K MT M (z y) K P MH (y x) π(x)x y, = K MT M (z y) π(y)y, = π(z), (45) which is precisely the efinition of invariant pf of K P MT M (z x). Clearly, we can invert the orer of the application of the kernels, i.e., using first K MT M an then K P MH. Thus, since the two are connecte sequentially, also the intermeiate steps are istribute as π(z), after a burn-in perio.

21 Table 5. State epenent PMH (S-PMH). Choose a initial state x 0, the total number of iterations K. 2. For k =,..., K: (a) Using a proposal pf of type q(s x k ) = q (s x,k )q 2(s 2 s, x :2,k ) q (s s :, x :,k ), (46) we employ SIR (see Section 2.2) for rawing with N particles, x (i), an weighting properly them, {x (i), w (i) )}N i=. The resampling steps are applie at R fixe an pre-establishe iterations (0 R K), < 2 <... < R. Thus, we obtain a particle approximation of the measure of target pf π (x) = i= Furthermore, we also obtain ẐX in Eq. () or Z X as in Eq. (2). w (i) δ(x x(i) ). (b) raw x π(x), i.e., choose a particle x = {x (),..., x (N) } with probability w (i), i =,..., N. (c) raw N particles z (),..., z (N ) via SIR using q(z x ) as in Eq. (46), applying resampling at the same iterations, < 2 <... < R, use in the generation of x (i) s. Moreover, set z N = x. () Compute where ρ (i) = π(z(i) ), i =,..., N. q(z (i) x ) (e) Set x k = x with probability otherwise set x k = x k. Ẑ Z = N α = i= ρ (i). Ẑ X Ẑ Z,

WEIGHTING A RESAMPLED PARTICLE IN SEQUENTIAL MONTE CARLO. L. Martino, V. Elvira, F. Louzada

WEIGHTING A RESAMPLED PARTICLE IN SEQUENTIAL MONTE CARLO. L. Martino, V. Elvira, F. Louzada WEIGHTIG A RESAMPLED PARTICLE I SEQUETIAL MOTE CARLO L. Martino, V. Elvira, F. Louzaa Dep. of Signal Theory an Communic., Universia Carlos III e Mari, Leganés (Spain). Institute of Mathematical Sciences

More information

A Review of Multiple Try MCMC algorithms for Signal Processing

A Review of Multiple Try MCMC algorithms for Signal Processing A Review of Multiple Try MCMC algorithms for Signal Processing Luca Martino Image Processing Lab., Universitat e València (Spain) Universia Carlos III e Mari, Leganes (Spain) Abstract Many applications

More information

WEIGHTING A RESAMPLED PARTICLES IN SEQUENTIAL MONTE CARLO (EXTENDED PREPRINT) L. Martino, V. Elvira, F. Louzada

WEIGHTING A RESAMPLED PARTICLES IN SEQUENTIAL MONTE CARLO (EXTENDED PREPRINT) L. Martino, V. Elvira, F. Louzada WEIGHTIG A RESAMLED ARTICLES I SEQUETIAL MOTE CARLO (ETEDED RERIT) L. Martino, V. Elvira, F. Louzaa Dep. of Signal Theory an Communic., Universia Carlos III e Mari, Leganés (Spain). Institute of Mathematical

More information

Group Importance Sampling for particle filtering and MCMC

Group Importance Sampling for particle filtering and MCMC Group Importance Sampling for particle filtering an MCMC Luca Martino, Víctor Elvira, Gustau Camps-Valls Image Processing Laboratory, Universitat e València (Spain). IMT Lille Douai CRISTAL (UMR 989),

More information

Introduction. A Dirichlet Form approach to MCMC Optimal Scaling. MCMC idea

Introduction. A Dirichlet Form approach to MCMC Optimal Scaling. MCMC idea Introuction A Dirichlet Form approach to MCMC Optimal Scaling Markov chain Monte Carlo (MCMC quotes: Metropolis et al. (1953, running coe on the Los Alamos MANIAC: a feasible approach to statistical mechanics

More information

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x)

The derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x) Y. D. Chong (2016) MH2801: Complex Methos for the Sciences 1. Derivatives The erivative of a function f(x) is another function, efine in terms of a limiting expression: f (x) f (x) lim x δx 0 f(x + δx)

More information

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013

Survey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013 Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing

More information

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012

Lecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012 CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration

More information

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors

Math Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors Math 18.02 Notes on ifferentials, the Chain Rule, graients, irectional erivative, an normal vectors Tangent plane an linear approximation We efine the partial erivatives of f( xy, ) as follows: f f( x+

More information

KNN Particle Filters for Dynamic Hybrid Bayesian Networks

KNN Particle Filters for Dynamic Hybrid Bayesian Networks KNN Particle Filters for Dynamic Hybri Bayesian Networs H. D. Chen an K. C. Chang Dept. of Systems Engineering an Operations Research George Mason University MS 4A6, 4400 University Dr. Fairfax, VA 22030

More information

Generalization of the persistent random walk to dimensions greater than 1

Generalization of the persistent random walk to dimensions greater than 1 PHYSICAL REVIEW E VOLUME 58, NUMBER 6 DECEMBER 1998 Generalization of the persistent ranom walk to imensions greater than 1 Marián Boguñá, Josep M. Porrà, an Jaume Masoliver Departament e Física Fonamental,

More information

Linear First-Order Equations

Linear First-Order Equations 5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)

More information

Hyperbolic Moment Equations Using Quadrature-Based Projection Methods

Hyperbolic Moment Equations Using Quadrature-Based Projection Methods Hyperbolic Moment Equations Using Quarature-Base Projection Methos J. Koellermeier an M. Torrilhon Department of Mathematics, RWTH Aachen University, Aachen, Germany Abstract. Kinetic equations like the

More information

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions

Computing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5

More information

Schrödinger s equation.

Schrödinger s equation. Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of

More information

Introduction to Markov Processes

Introduction to Markov Processes Introuction to Markov Processes Connexions moule m44014 Zzis law Gustav) Meglicki, Jr Office of the VP for Information Technology Iniana University RCS: Section-2.tex,v 1.24 2012/12/21 18:03:08 gustav

More information

Least-Squares Regression on Sparse Spaces

Least-Squares Regression on Sparse Spaces Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction

More information

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments

Time-of-Arrival Estimation in Non-Line-Of-Sight Environments 2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor

More information

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE

TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS. Yannick DEVILLE TEMPORAL AND TIME-FREQUENCY CORRELATION-BASED BLIND SOURCE SEPARATION METHODS Yannick DEVILLE Université Paul Sabatier Laboratoire Acoustique, Métrologie, Instrumentation Bât. 3RB2, 8 Route e Narbonne,

More information

Switching Time Optimization in Discretized Hybrid Dynamical Systems

Switching Time Optimization in Discretized Hybrid Dynamical Systems Switching Time Optimization in Discretize Hybri Dynamical Systems Kathrin Flaßkamp, To Murphey, an Sina Ober-Blöbaum Abstract Switching time optimization (STO) arises in systems that have a finite set

More information

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k

Robust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine

More information

Introduction to the Vlasov-Poisson system

Introduction to the Vlasov-Poisson system Introuction to the Vlasov-Poisson system Simone Calogero 1 The Vlasov equation Consier a particle with mass m > 0. Let x(t) R 3 enote the position of the particle at time t R an v(t) = ẋ(t) = x(t)/t its

More information

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.

. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences. S 63 Lecture 8 2/2/26 Lecturer Lillian Lee Scribes Peter Babinski, Davi Lin Basic Language Moeling Approach I. Special ase of LM-base Approach a. Recap of Formulas an Terms b. Fixing θ? c. About that Multinomial

More information

Kernel adaptive Sequential Monte Carlo

Kernel adaptive Sequential Monte Carlo Kernel adaptive Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) December 7, 2015 1 / 36 Section 1 Outline

More information

Effective Sample Size for Importance Sampling based on discrepancy measures

Effective Sample Size for Importance Sampling based on discrepancy measures Effective Sample Size for Importance Sampling based on discrepancy measures L. Martino, V. Elvira, F. Louzada Universidade de São Paulo, São Carlos (Brazil). Universidad Carlos III de Madrid, Leganés (Spain).

More information

Learning the hyper-parameters. Luca Martino

Learning the hyper-parameters. Luca Martino Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth

More information

Proof of SPNs as Mixture of Trees

Proof of SPNs as Mixture of Trees A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a

More information

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1

d dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1 Lecture 5 Some ifferentiation rules Trigonometric functions (Relevant section from Stewart, Seventh Eition: Section 3.3) You all know that sin = cos cos = sin. () But have you ever seen a erivation of

More information

SYNCHRONOUS SEQUENTIAL CIRCUITS

SYNCHRONOUS SEQUENTIAL CIRCUITS CHAPTER SYNCHRONOUS SEUENTIAL CIRCUITS Registers an counters, two very common synchronous sequential circuits, are introuce in this chapter. Register is a igital circuit for storing information. Contents

More information

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters

Exercises Tutorial at ICASSP 2016 Learning Nonlinear Dynamical Models Using Particle Filters Exercises Tutorial at ICASSP 216 Learning Nonlinear Dynamical Models Using Particle Filters Andreas Svensson, Johan Dahlin and Thomas B. Schön March 18, 216 Good luck! 1 [Bootstrap particle filter for

More information

Analyzing Tensor Power Method Dynamics in Overcomplete Regime

Analyzing Tensor Power Method Dynamics in Overcomplete Regime Journal of Machine Learning Research 18 (2017) 1-40 Submitte 9/15; Revise 11/16; Publishe 4/17 Analyzing Tensor Power Metho Dynamics in Overcomplete Regime Animashree Ananumar Department of Electrical

More information

Lagrangian and Hamiltonian Mechanics

Lagrangian and Hamiltonian Mechanics Lagrangian an Hamiltonian Mechanics.G. Simpson, Ph.. epartment of Physical Sciences an Engineering Prince George s Community College ecember 5, 007 Introuction In this course we have been stuying classical

More information

A Course in Machine Learning

A Course in Machine Learning A Course in Machine Learning Hal Daumé III 12 EFFICIENT LEARNING So far, our focus has been on moels of learning an basic algorithms for those moels. We have not place much emphasis on how to learn quickly.

More information

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics

This module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics This moule is part of the Memobust Hanbook on Methoology of Moern Business Statistics 26 March 2014 Metho: Balance Sampling for Multi-Way Stratification Contents General section... 3 1. Summary... 3 2.

More information

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback

An Optimal Algorithm for Bandit and Zero-Order Convex Optimization with Two-Point Feedback Journal of Machine Learning Research 8 07) - Submitte /6; Publishe 5/7 An Optimal Algorithm for Banit an Zero-Orer Convex Optimization with wo-point Feeback Oha Shamir Department of Computer Science an

More information

An introduction to Sequential Monte Carlo

An introduction to Sequential Monte Carlo An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods

More information

Adaptive Rejection Sampling with fixed number of nodes

Adaptive Rejection Sampling with fixed number of nodes Adaptive Rejection Sampling with fixed number of nodes L. Martino, F. Louzada Institute of Mathematical Sciences and Computing, Universidade de São Paulo, Brazil. Abstract The adaptive rejection sampling

More information

Improving Estimation Accuracy in Nonrandomized Response Questioning Methods by Multiple Answers

Improving Estimation Accuracy in Nonrandomized Response Questioning Methods by Multiple Answers International Journal of Statistics an Probability; Vol 6, No 5; September 207 ISSN 927-7032 E-ISSN 927-7040 Publishe by Canaian Center of Science an Eucation Improving Estimation Accuracy in Nonranomize

More information

A Modification of the Jarque-Bera Test. for Normality

A Modification of the Jarque-Bera Test. for Normality Int. J. Contemp. Math. Sciences, Vol. 8, 01, no. 17, 84-85 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1988/ijcms.01.9106 A Moification of the Jarque-Bera Test for Normality Moawa El-Fallah Ab El-Salam

More information

Math 1B, lecture 8: Integration by parts

Math 1B, lecture 8: Integration by parts Math B, lecture 8: Integration by parts Nathan Pflueger 23 September 2 Introuction Integration by parts, similarly to integration by substitution, reverses a well-known technique of ifferentiation an explores

More information

Gaussian processes with monotonicity information

Gaussian processes with monotonicity information Gaussian processes with monotonicity information Anonymous Author Anonymous Author Unknown Institution Unknown Institution Abstract A metho for using monotonicity information in multivariate Gaussian process

More information

Robust Low Rank Kernel Embeddings of Multivariate Distributions

Robust Low Rank Kernel Embeddings of Multivariate Distributions Robust Low Rank Kernel Embeings of Multivariate Distributions Le Song, Bo Dai College of Computing, Georgia Institute of Technology lsong@cc.gatech.eu, boai@gatech.eu Abstract Kernel embeing of istributions

More information

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION

LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische

More information

Cascaded redundancy reduction

Cascaded redundancy reduction Network: Comput. Neural Syst. 9 (1998) 73 84. Printe in the UK PII: S0954-898X(98)88342-5 Cascae reunancy reuction Virginia R e Sa an Geoffrey E Hinton Department of Computer Science, University of Toronto,

More information

Adaptive Rejection Sampling with fixed number of nodes

Adaptive Rejection Sampling with fixed number of nodes Adaptive Rejection Sampling with fixed number of nodes L. Martino, F. Louzada Institute of Mathematical Sciences and Computing, Universidade de São Paulo, São Carlos (São Paulo). Abstract The adaptive

More information

A simple model for the small-strain behaviour of soils

A simple model for the small-strain behaviour of soils A simple moel for the small-strain behaviour of soils José Jorge Naer Department of Structural an Geotechnical ngineering, Polytechnic School, University of São Paulo 05508-900, São Paulo, Brazil, e-mail:

More information

THE VAN KAMPEN EXPANSION FOR LINKED DUFFING LINEAR OSCILLATORS EXCITED BY COLORED NOISE

THE VAN KAMPEN EXPANSION FOR LINKED DUFFING LINEAR OSCILLATORS EXCITED BY COLORED NOISE Journal of Soun an Vibration (1996) 191(3), 397 414 THE VAN KAMPEN EXPANSION FOR LINKED DUFFING LINEAR OSCILLATORS EXCITED BY COLORED NOISE E. M. WEINSTEIN Galaxy Scientific Corporation, 2500 English Creek

More information

Relative Entropy and Score Function: New Information Estimation Relationships through Arbitrary Additive Perturbation

Relative Entropy and Score Function: New Information Estimation Relationships through Arbitrary Additive Perturbation Relative Entropy an Score Function: New Information Estimation Relationships through Arbitrary Aitive Perturbation Dongning Guo Department of Electrical Engineering & Computer Science Northwestern University

More information

Non-Linear Bayesian CBRN Source Term Estimation

Non-Linear Bayesian CBRN Source Term Estimation Non-Linear Bayesian CBRN Source Term Estimation Peter Robins Hazar Assessment, Simulation an Preiction Group Dstl Porton Down, UK. probins@stl.gov.uk Paul Thomas Hazar Assessment, Simulation an Preiction

More information

Table of Common Derivatives By David Abraham

Table of Common Derivatives By David Abraham Prouct an Quotient Rules: Table of Common Derivatives By Davi Abraham [ f ( g( ] = [ f ( ] g( + f ( [ g( ] f ( = g( [ f ( ] g( g( f ( [ g( ] Trigonometric Functions: sin( = cos( cos( = sin( tan( = sec

More information

The total derivative. Chapter Lagrangian and Eulerian approaches

The total derivative. Chapter Lagrangian and Eulerian approaches Chapter 5 The total erivative 51 Lagrangian an Eulerian approaches The representation of a flui through scalar or vector fiels means that each physical quantity uner consieration is escribe as a function

More information

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion

Hybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion Hybri Fusion for Biometrics: Combining Score-level an Decision-level Fusion Qian Tao Raymon Velhuis Signals an Systems Group, University of Twente Postbus 217, 7500AE Enschee, the Netherlans {q.tao,r.n.j.velhuis}@ewi.utwente.nl

More information

Nonlinear Adaptive Ship Course Tracking Control Based on Backstepping and Nussbaum Gain

Nonlinear Adaptive Ship Course Tracking Control Based on Backstepping and Nussbaum Gain Nonlinear Aaptive Ship Course Tracking Control Base on Backstepping an Nussbaum Gain Jialu Du, Chen Guo Abstract A nonlinear aaptive controller combining aaptive Backstepping algorithm with Nussbaum gain

More information

Math 342 Partial Differential Equations «Viktor Grigoryan

Math 342 Partial Differential Equations «Viktor Grigoryan Math 342 Partial Differential Equations «Viktor Grigoryan 6 Wave equation: solution In this lecture we will solve the wave equation on the entire real line x R. This correspons to a string of infinite

More information

The Recycling Gibbs Sampler for Efficient Learning

The Recycling Gibbs Sampler for Efficient Learning The Recycling Gibbs Sampler for Efficient Learning L. Martino, V. Elvira, G. Camps-Valls Universidade de São Paulo, São Carlos (Brazil). Télécom ParisTech, Université Paris-Saclay. (France), Universidad

More information

APPROXIMATE SOLUTION FOR TRANSIENT HEAT TRANSFER IN STATIC TURBULENT HE II. B. Baudouy. CEA/Saclay, DSM/DAPNIA/STCM Gif-sur-Yvette Cedex, France

APPROXIMATE SOLUTION FOR TRANSIENT HEAT TRANSFER IN STATIC TURBULENT HE II. B. Baudouy. CEA/Saclay, DSM/DAPNIA/STCM Gif-sur-Yvette Cedex, France APPROXIMAE SOLUION FOR RANSIEN HEA RANSFER IN SAIC URBULEN HE II B. Bauouy CEA/Saclay, DSM/DAPNIA/SCM 91191 Gif-sur-Yvette Ceex, France ABSRAC Analytical solution in one imension of the heat iffusion equation

More information

TIME-DELAY ESTIMATION USING FARROW-BASED FRACTIONAL-DELAY FIR FILTERS: FILTER APPROXIMATION VS. ESTIMATION ERRORS

TIME-DELAY ESTIMATION USING FARROW-BASED FRACTIONAL-DELAY FIR FILTERS: FILTER APPROXIMATION VS. ESTIMATION ERRORS TIME-DEAY ESTIMATION USING FARROW-BASED FRACTIONA-DEAY FIR FITERS: FITER APPROXIMATION VS. ESTIMATION ERRORS Mattias Olsson, Håkan Johansson, an Per öwenborg Div. of Electronic Systems, Dept. of Electrical

More information

Kernel Sequential Monte Carlo

Kernel Sequential Monte Carlo Kernel Sequential Monte Carlo Ingmar Schuster (Paris Dauphine) Heiko Strathmann (University College London) Brooks Paige (Oxford) Dino Sejdinovic (Oxford) * equal contribution April 25, 2016 1 / 37 Section

More information

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation

Thermal conductivity of graded composites: Numerical simulations and an effective medium approximation JOURNAL OF MATERIALS SCIENCE 34 (999)5497 5503 Thermal conuctivity of grae composites: Numerical simulations an an effective meium approximation P. M. HUI Department of Physics, The Chinese University

More information

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling

Collapsed Gibbs and Variational Methods for LDA. Example Collapsed MoG Sampling Case Stuy : Document Retrieval Collapse Gibbs an Variational Methos for LDA Machine Learning/Statistics for Big Data CSE599C/STAT59, University of Washington Emily Fox 0 Emily Fox February 7 th, 0 Example

More information

A Novel Decoupled Iterative Method for Deep-Submicron MOSFET RF Circuit Simulation

A Novel Decoupled Iterative Method for Deep-Submicron MOSFET RF Circuit Simulation A Novel ecouple Iterative Metho for eep-submicron MOSFET RF Circuit Simulation CHUAN-SHENG WANG an YIMING LI epartment of Mathematics, National Tsing Hua University, National Nano evice Laboratories, an

More information

Influence of weight initialization on multilayer perceptron performance

Influence of weight initialization on multilayer perceptron performance Influence of weight initialization on multilayer perceptron performance M. Karouia (1,2) T. Denœux (1) R. Lengellé (1) (1) Université e Compiègne U.R.A. CNRS 817 Heuiasyc BP 649 - F-66 Compiègne ceex -

More information

Markov chain Monte Carlo methods for visual tracking

Markov chain Monte Carlo methods for visual tracking Markov chain Monte Carlo methods for visual tracking Ray Luo rluo@cory.eecs.berkeley.edu Department of Electrical Engineering and Computer Sciences University of California, Berkeley Berkeley, CA 94720

More information

Acute sets in Euclidean spaces

Acute sets in Euclidean spaces Acute sets in Eucliean spaces Viktor Harangi April, 011 Abstract A finite set H in R is calle an acute set if any angle etermine by three points of H is acute. We examine the maximal carinality α() of

More information

The Exact Form and General Integrating Factors

The Exact Form and General Integrating Factors 7 The Exact Form an General Integrating Factors In the previous chapters, we ve seen how separable an linear ifferential equations can be solve using methos for converting them to forms that can be easily

More information

Chapter 6: Energy-Momentum Tensors

Chapter 6: Energy-Momentum Tensors 49 Chapter 6: Energy-Momentum Tensors This chapter outlines the general theory of energy an momentum conservation in terms of energy-momentum tensors, then applies these ieas to the case of Bohm's moel.

More information

6 General properties of an autonomous system of two first order ODE

6 General properties of an autonomous system of two first order ODE 6 General properties of an autonomous system of two first orer ODE Here we embark on stuying the autonomous system of two first orer ifferential equations of the form ẋ 1 = f 1 (, x 2 ), ẋ 2 = f 2 (, x

More information

Expected Value of Partial Perfect Information

Expected Value of Partial Perfect Information Expecte Value of Partial Perfect Information Mike Giles 1, Takashi Goa 2, Howar Thom 3 Wei Fang 1, Zhenru Wang 1 1 Mathematical Institute, University of Oxfor 2 School of Engineering, University of Tokyo

More information

YVES F. ATCHADÉ, GARETH O. ROBERTS, AND JEFFREY S. ROSENTHAL

YVES F. ATCHADÉ, GARETH O. ROBERTS, AND JEFFREY S. ROSENTHAL TOWARDS OPTIMAL SCALING OF METROPOLIS-COUPLED MARKOV CHAIN MONTE CARLO YVES F. ATCHADÉ, GARETH O. ROBERTS, AND JEFFREY S. ROSENTHAL Abstract. We consier optimal temperature spacings for Metropolis-couple

More information

Implicit Differentiation

Implicit Differentiation Implicit Differentiation Thus far, the functions we have been concerne with have been efine explicitly. A function is efine explicitly if the output is given irectly in terms of the input. For instance,

More information

3.7 Implicit Differentiation -- A Brief Introduction -- Student Notes

3.7 Implicit Differentiation -- A Brief Introduction -- Student Notes Fin these erivatives of these functions: y.7 Implicit Differentiation -- A Brief Introuction -- Stuent Notes tan y sin tan = sin y e = e = Write the inverses of these functions: y tan y sin How woul we

More information

MODELLING DEPENDENCE IN INSURANCE CLAIMS PROCESSES WITH LÉVY COPULAS ABSTRACT KEYWORDS

MODELLING DEPENDENCE IN INSURANCE CLAIMS PROCESSES WITH LÉVY COPULAS ABSTRACT KEYWORDS MODELLING DEPENDENCE IN INSURANCE CLAIMS PROCESSES WITH LÉVY COPULAS BY BENJAMIN AVANZI, LUKE C. CASSAR AND BERNARD WONG ABSTRACT In this paper we investigate the potential of Lévy copulas as a tool for

More information

The Press-Schechter mass function

The Press-Schechter mass function The Press-Schechter mass function To state the obvious: It is important to relate our theories to what we can observe. We have looke at linear perturbation theory, an we have consiere a simple moel for

More information

Lower Bounds for the Smoothed Number of Pareto optimal Solutions

Lower Bounds for the Smoothed Number of Pareto optimal Solutions Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.

More information

Equilibrium in Queues Under Unknown Service Times and Service Value

Equilibrium in Queues Under Unknown Service Times and Service Value University of Pennsylvania ScholarlyCommons Finance Papers Wharton Faculty Research 1-2014 Equilibrium in Queues Uner Unknown Service Times an Service Value Laurens Debo Senthil K. Veeraraghavan University

More information

A simple tranformation of copulas

A simple tranformation of copulas A simple tranformation of copulas V. Durrleman, A. Nikeghbali & T. Roncalli Groupe e Recherche Opérationnelle Créit Lyonnais France July 31, 2000 Abstract We stuy how copulas properties are moifie after

More information

26.1 Metropolis method

26.1 Metropolis method CS880: Approximations Algorithms Scribe: Dave Anrzejewski Lecturer: Shuchi Chawla Topic: Metropolis metho, volume estimation Date: 4/26/07 The previous lecture iscusse they some of the key concepts of

More information

MEASURES WITH ZEROS IN THE INVERSE OF THEIR MOMENT MATRIX

MEASURES WITH ZEROS IN THE INVERSE OF THEIR MOMENT MATRIX MEASURES WITH ZEROS IN THE INVERSE OF THEIR MOMENT MATRIX J. WILLIAM HELTON, JEAN B. LASSERRE, AND MIHAI PUTINAR Abstract. We investigate an iscuss when the inverse of a multivariate truncate moment matrix

More information

Topic Modeling: Beyond Bag-of-Words

Topic Modeling: Beyond Bag-of-Words Hanna M. Wallach Cavenish Laboratory, University of Cambrige, Cambrige CB3 0HE, UK hmw26@cam.ac.u Abstract Some moels of textual corpora employ text generation methos involving n-gram statistics, while

More information

05 The Continuum Limit and the Wave Equation

05 The Continuum Limit and the Wave Equation Utah State University DigitalCommons@USU Founations of Wave Phenomena Physics, Department of 1-1-2004 05 The Continuum Limit an the Wave Equation Charles G. Torre Department of Physics, Utah State University,

More information

A Path Planning Method Using Cubic Spiral with Curvature Constraint

A Path Planning Method Using Cubic Spiral with Curvature Constraint A Path Planning Metho Using Cubic Spiral with Curvature Constraint Tzu-Chen Liang an Jing-Sin Liu Institute of Information Science 0, Acaemia Sinica, Nankang, Taipei 5, Taiwan, R.O.C., Email: hartree@iis.sinica.eu.tw

More information

Analytic Scaling Formulas for Crossed Laser Acceleration in Vacuum

Analytic Scaling Formulas for Crossed Laser Acceleration in Vacuum October 6, 4 ARDB Note Analytic Scaling Formulas for Crosse Laser Acceleration in Vacuum Robert J. Noble Stanfor Linear Accelerator Center, Stanfor University 575 San Hill Roa, Menlo Park, California 945

More information

Lecture 2: Correlated Topic Model

Lecture 2: Correlated Topic Model Probabilistic Moels for Unsupervise Learning Spring 203 Lecture 2: Correlate Topic Moel Inference for Correlate Topic Moel Yuan Yuan First of all, let us make some claims about the parameters an variables

More information

Delocalization of boundary states in disordered topological insulators

Delocalization of boundary states in disordered topological insulators Journal of Physics A: Mathematical an Theoretical J. Phys. A: Math. Theor. 48 (05) FT0 (pp) oi:0.088/75-83/48//ft0 Fast Track Communication Delocalization of bounary states in isorere topological insulators

More information

Florian Heiss; Viktor Winschel: Estimation with Numerical Integration on Sparse Grids

Florian Heiss; Viktor Winschel: Estimation with Numerical Integration on Sparse Grids Florian Heiss; Viktor Winschel: Estimation with Numerical Integration on Sparse Gris Munich Discussion Paper No. 2006-15 Department of Economics University of Munich Volkswirtschaftliche Fakultät Luwig-Maximilians-Universität

More information

A NONLINEAR SOURCE SEPARATION APPROACH FOR THE NICOLSKY-EISENMAN MODEL

A NONLINEAR SOURCE SEPARATION APPROACH FOR THE NICOLSKY-EISENMAN MODEL 6th European Signal Processing Conference EUSIPCO 28, Lausanne, Switzerlan, August 25-29, 28, copyright by EURASIP A NONLINEAR SOURCE SEPARATION APPROACH FOR THE NICOLSKY-EISENMAN MODEL Leonaro Tomazeli

More information

Monte Carlo Methods with Reduced Error

Monte Carlo Methods with Reduced Error Monte Carlo Methos with Reuce Error As has been shown, the probable error in Monte Carlo algorithms when no information about the smoothness of the function is use is Dξ r N = c N. It is important for

More information

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs

Lectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs Lectures - Week 10 Introuction to Orinary Differential Equations (ODES) First Orer Linear ODEs When stuying ODEs we are consiering functions of one inepenent variable, e.g., f(x), where x is the inepenent

More information

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks

A PAC-Bayesian Approach to Spectrally-Normalized Margin Bounds for Neural Networks A PAC-Bayesian Approach to Spectrally-Normalize Margin Bouns for Neural Networks Behnam Neyshabur, Srinah Bhojanapalli, Davi McAllester, Nathan Srebro Toyota Technological Institute at Chicago {bneyshabur,

More information

Bayesian analysis of massive datasets via particle filters

Bayesian analysis of massive datasets via particle filters Bayesian analysis of massive atasets via particle filters Greg Rigeway RAND PO Box 2138 Santa Monica, CA 90407-2138 gregr@ran.org Davi Maigan Department of Statistics 477 Hill Center Rutgers University

More information

Diophantine Approximations: Examining the Farey Process and its Method on Producing Best Approximations

Diophantine Approximations: Examining the Farey Process and its Method on Producing Best Approximations Diophantine Approximations: Examining the Farey Process an its Metho on Proucing Best Approximations Kelly Bowen Introuction When a person hears the phrase irrational number, one oes not think of anything

More information

LDA Collapsed Gibbs Sampler, VariaNonal Inference. Task 3: Mixed Membership Models. Case Study 5: Mixed Membership Modeling

LDA Collapsed Gibbs Sampler, VariaNonal Inference. Task 3: Mixed Membership Models. Case Study 5: Mixed Membership Modeling Case Stuy 5: Mixe Membership Moeling LDA Collapse Gibbs Sampler, VariaNonal Inference Machine Learning for Big Data CSE547/STAT548, University of Washington Emily Fox May 8 th, 05 Emily Fox 05 Task : Mixe

More information

Concentration of Measure Inequalities for Compressive Toeplitz Matrices with Applications to Detection and System Identification

Concentration of Measure Inequalities for Compressive Toeplitz Matrices with Applications to Detection and System Identification Concentration of Measure Inequalities for Compressive Toeplitz Matrices with Applications to Detection an System Ientification Borhan M Sananaji, Tyrone L Vincent, an Michael B Wakin Abstract In this paper,

More information

Radar Sensor Management for Detection and Tracking

Radar Sensor Management for Detection and Tracking Raar Sensor Management for Detection an Tracking Krüger White, Jason Williams, Peter Hoffensetz Defence Science an Technology Organisation PO Box 00, Einburgh, SA Australia Email: Kruger.White@sto.efence.gov.au,

More information

Dot trajectories in the superposition of random screens: analysis and synthesis

Dot trajectories in the superposition of random screens: analysis and synthesis 1472 J. Opt. Soc. Am. A/ Vol. 21, No. 8/ August 2004 Isaac Amiror Dot trajectories in the superposition of ranom screens: analysis an synthesis Isaac Amiror Laboratoire e Systèmes Périphériques, Ecole

More information

arxiv: v2 [cond-mat.stat-mech] 11 Nov 2016

arxiv: v2 [cond-mat.stat-mech] 11 Nov 2016 Noname manuscript No. (will be inserte by the eitor) Scaling properties of the number of ranom sequential asorption iterations neee to generate saturate ranom packing arxiv:607.06668v2 [con-mat.stat-mech]

More information

How to Minimize Maximum Regret in Repeated Decision-Making

How to Minimize Maximum Regret in Repeated Decision-Making How to Minimize Maximum Regret in Repeate Decision-Making Karl H. Schlag July 3 2003 Economics Department, European University Institute, Via ella Piazzuola 43, 033 Florence, Italy, Tel: 0039-0-4689, email:

More information

Controlled sequential Monte Carlo

Controlled sequential Monte Carlo Controlled sequential Monte Carlo Jeremy Heng, Department of Statistics, Harvard University Joint work with Adrian Bishop (UTS, CSIRO), George Deligiannidis & Arnaud Doucet (Oxford) Bayesian Computation

More information

Optimal Signal Detection for False Track Discrimination

Optimal Signal Detection for False Track Discrimination Optimal Signal Detection for False Track Discrimination Thomas Hanselmann Darko Mušicki Dept. of Electrical an Electronic Eng. Dept. of Electrical an Electronic Eng. The University of Melbourne The University

More information

arxiv: v1 [math.co] 29 May 2009

arxiv: v1 [math.co] 29 May 2009 arxiv:0905.4913v1 [math.co] 29 May 2009 simple Havel-Hakimi type algorithm to realize graphical egree sequences of irecte graphs Péter L. Erős an István Miklós. Rényi Institute of Mathematics, Hungarian

More information