A Review of Multiple Try MCMC algorithms for Signal Processing
|
|
- Logan Sharp
- 6 years ago
- Views:
Transcription
1 A Review of Multiple Try MCMC algorithms for Signal Processing Luca Martino Image Processing Lab., Universitat e València (Spain) Universia Carlos III e Mari, Leganes (Spain) Abstract Many applications in signal processing require the estimation of some parameters of interest given a set of observe ata. More specifically, Bayesian inference nees the computation of a-posteriori estimators which are often expresse as complicate multiimensional integrals. Unfortunately, analytical expressions for these estimators cannot be foun in most real-worl applications, an Monte Carlo methos are the only feasible approach. A very powerful class of Monte Carlo techniques is forme by the Markov Chain Monte Carlo (MCMC) algorithms. They generate a Markov chain such that its stationary istribution coincies with the target posterior ensity. In this work, we perform a thorough review of MCMC methos using multiple caniates in orer to select the next state of the chain, at each iteration. With respect to the classical Metropolis-Hastings metho, the use of multiple try techniques foster the exploration of the sample space. We present ifferent Multiple Try Metropolis schemes, Ensemble MCMC methos, Particle Metropolis-Hastings algorithms an the Delaye Rejection Metropolis technique. We highlight limitations, benefits, connections an ifferences among the ifferent methos, an compare them by numerical simulations. Keywors: Markov Chain Monte Carlo, Multiple Try Metropolis, Particle Metropolis- Hastings, Particle Filtering, Monte Carlo methos, Bayesian inference 1 Introuction 5 10 Bayesian methos have become very popular in signal processing over the last years [1, 2, 3, 4]. They require the application of sophisticate Monte Carlo techniques, such as Markov chain Monte Carlo (MCMC) an particle filters, for the efficient computation of a-posteriori estimators [5, 6, 2]. More specifically, the MCMC algorithms generate a Markov chain such that its stationary istribution coincies with the posterior probability ensity function (pf) [7, 8, 9]. Typically, the only requirement is to be able to evaluate the target function, where the knowlege of the normalizing constant is usually not neee. The most popular MCMC metho is unoubtely the Metropolis-Hastings (MH) algorithm [10, 11]. The MH technique is a very simple metho, easy to be applie: this is the reason of 1
2 its success. In MH, at each iteration, one new caniate is generate from a proposal pf an then is properly compare with the previous state of the chain, in orer to ecie the next state. However, the performance of MH are often not satisfactory. For instance, when the posterior is multimoal, or when the imension of the space increases, the correlation among the generate samples is usually high an, as a consequence, the variance of the resulting estimators grows. To spee up the convergence an reuce the burn-in perio of the MH chain, several extensions have been propose in literature. In this work, we provie an exhaustive review of more sophisticate MCMC methos that, at each iteration, consier ifferent caniates as possible new state of the chain. More specifically, at each iteration ifferent samples are compare by certain weights an then one of them is selecte as possible future state. The main avantage of these algorithms is that they foster the exploration of a larger portion of the sample space, ecreasing the correlation among the states of the generate chain. In this work, we escribe ifferent algorithms of this family, inepenently introuce in literature. The main contribution is to present them uner the same the framework an notation, remarking ifferences, relationships, limitations an strengths. All the iscusse techniques yiel an ergoic chain converging to the posterior ensity of interest (in the following, referre also as target pf). The first scheme of this MCMC class, calle Orientational Bias Monte Carlo (OBMC) [12, Chapter 13], was propose in the context of molecular simulation. Later on a more general algorithm, calle Multiple Try Metropolis (MTM), was introuce [13]. 1 The MTM algorithm has been extensively stuie an generalize in ifferent ways [14, 15, 16, 17, 18]. Other techniques, alternative to the MTM schemes, are the so-calle the Ensemble MCMC (EnMCMC) methos [19, 20, 21, 22]. They follow a similar approach to MTM but employ a ifferent acceptance function for selecting the next state of the chain. With respect to (w.r.t.) a generic MTM scheme, EnMCMC oes not require any generation of auxiliary samples (as in a MTM scheme employing a generic proposal pf) an hence, in this sense, EnMCMC are less costly. In all the previous techniques, the caniates are rawn in a batch way an compare jointly. In the Delaye Rejection Metropolis (DRM) algorithm [23, 24, 25], in case of rejection of the novel possible state, the authors suggest to perform an aitional acceptance test consiering a new caniate. If this caniate is again rejecte, the proceure can be iterate until reaching a esire number of attempts. The main benefit of DRM is that the proposal pf can be improve at each intermeiate stage. However, the acceptance function progressively becomes more complex so that the implementation of DRM for a great number of attempts is not straightforwar (compare to the implementation of a MTM scheme with a generic number of tries). In the last years, other Monte Carlo methos which combine particle filtering an MCMC have become very popular in the signal processing community. For instance, this is the case of the Particle Metropolis Hastings (PMH) an the Particle Marginal Metropolis Hastings (PMMH) algorithms, which have been wiely use in signal processing in orer to make inference an smoothing about ynamical an static parameters in state space moels [26, 27]. PMH can be interprete as a MTM scheme where the ifferent caniates are generate an weighte by the use of a particle filter [28, 29]. In this work, we present PMH an PMMH an iscuss 1 MTM inclues OBMC as a special case (see Section 4.1.1). 2
3 55 60 their connections an ifferences with the classical MTM approach. Furthermore, we escribe a suitable proceure for recycling some caniates in the final Monte Carlo estimators, calle Group Metropolis Sampling (GMS) [29, 30]. The GMS scheme can be also seen as a way of generating a chain of sets of weighte samples. Finally, note that other similar an relate techniques can be foun within the so-calle ata augmentation approach [31, 32]. The remaining of the paper is organize as follows. Section 2 recalls the problem statement an some backgroun material, introucing also the require notation. The basis of MCMC an the Metropolis-Hastings (MH) algorithm are presente in Section 3. Section 4 is the core of the work, which escribes the ifferent MCMC using multiple caniates. Section 6 provies some numerical results, applying ifferent techniques in a hyperparameter tuning problem for a Gaussian Process regression moel, an in a localization problem consiering a wireless sensor network. Some conclusions are given in Section 7. 2 Problem statement an preliminaries 65 In many signal processing applications, the goal consists in inferring a variable of interest, θ = [θ 1,..., θ D ] D R D, given a set of observations or measurements, y R P. In the Bayesian framework, the total knowlege about the parameters, after the ata have been observe, is represente by the posterior probability ensity function (pf) [8, 9], i.e., π(θ) = p(θ y) = l(y θ)g(θ), Z(y) = 1 π(θ), (1) Z where l(y θ) enotes the likelihoo function (i.e., the observation moel), g(θ) is the prior probability ensity function (pf) an Z = Z(y) is the marginal likelihoo (a.k.a., Bayesian evience) [33, 34] an π(θ) = l(y θ)g(θ). In general Z is unknown, an it is possible to evaluate only the unnormalize target function, π(θ) π(θ). The analytical stuy of the posterior ensity π(θ) is often unfeasible an integrals involving π(θ) are typically intractable [33, 35, 34]. For instance, one might be intereste in the estimation of I = E π [f(θ)] = f(θ) π(θ)θ, (2) D = 1 f(θ)π(θ)θ, (3) Z where f(θ) is a generic integrable function w.r.t. π. Dynamic an static parameters. In some specific application, the variable of interest θ can be split in two isjoint parts, θ = [x, λ], where one, x, is involve into a ynamical system (for instance, x is the hien state in a state-space moel) an the other, λ, is a static parameter (for instance, an unknown parameter of the moel). The strategies for making inference about x an λ shoul take into account the ifferent nature of the two parameters (e.g., see Section 4.2.2). The main notation an acronyms are summarize in Tables 1-2. D 3
4 Table 1: Summary of the main notation. θ = [θ 1,..., θ D ] Variable of interest, θ D R D. y Observe measurements/ata. θ = [x, λ] x : ynamic parameters; λ : static parameters. π(θ) ormalize posterior pf, π(θ) = p(θ y). π(θ) Unnormalize posterior function, π(θ) π(θ). π(θ) Particle approximation of π(θ). I Integral of interest, in Eq. (2). Î, Ĩ Estimators of I. Z Marginal likelihoo; normalizing constant of π(θ). Ẑ, Z Estimators of the marginal likelihoo Z. Table 2: Summary of the main acronyms. MCMC Markov Chain Monte Carlo MH Metropolis-Hastings I-MH Inepenent Metropolis-Hastings MTM Multiple Try Metropolis I-MTM Inepenent Multiple Try Metropolis I-MTM2 Inepenent Multiple Try Metropolis (version 2) PMH Particle Metropolis-Hastings PMMH Particle Marginal Metropolis-Hastings GMS Group Metropolis Sampling EnMCMC Ensemble MCMC I-EnMCMC Inepenent Ensemble MCMC DRM Delaye Rejection Metropolis IS Importance Sampling SIS Sequential Importance Sampling SIR Sequential Importance Resampling 2.1 Monte Carlo integration In many practical scenarios, the integral I cannot be compute in a close form, an numerical approximations are typically require. Many eterministic quarature methos are available in the literature [36, 37]. However, as the imension D of the inference problem grows (θ R D ), the eterministic quarature schemes become less efficient. In this case, a common approach consists of approximating the integral I in Eq. (2) by using Monte Carlo (MC) quarature [8, 9]. amely, consiering T inepenent an ientically istribute (i.i..) samples rawn from the posterior 4
5 target pf, i.e. θ 1,..., θ T π(θ), 2 we can buil the consistent estimator Î T = 1 T T f(θ t ). (4) t=1 85 Î T converges in probability to I ue to the weak law of large numbers. The approximation above ÎT is known as a irect (or ieal) Monte Carlo estimator if the samples θ t are inepenent an ientically istribute (i.i..) from π. Unfortunately, in many practical applications, irect methos for rawing inepenent samples from π(θ) are not available. Therefore, ifferent approaches are require, such as the Markov chain Monte Carlo (MCMC) techniques. 3 Markov chain Monte Carlo (MCMC) methos A MCMC algorithm generates an ergoic Markov chain with invariant (a.k.a., stationary) ensity given by the posterior pf π(θ) [7, 9]. Specifically, given a starting state θ 0, a sequence of correlate samples is generate, θ 0 θ 1 θ 2... θ T. Even if the samples are now correlate, the estimator Ĩ T = 1 T f(θ t ) (5) T t=1 is consistent, regarless the starting vector θ (0) [9]. 3 With respect to the irect Monte Carlo approach using i.i.. samples, the application of an MCMC algorithm entails a loss of efficiency of the estimator ĨT, since the samples are positively correlate, in general. In other wors, to achieve a given variance obtaine with the irect Monte Carlo estimator, it is necessary to generate more samples. Thus, in orer to improve the performance of an MCMC technique we have to ecrease the correlation among the states of the chain The Metropolis-Hastings (MH) algorithm One of the most popular an wiely applie MCMC algorithm is the Metropolis-Hastings (MH) metho [11, 7, 9]. Recall that we are able to evaluate point-wise a function proportional to the target, i.e., π(θ) π(θ). A proposal ensity (a pf which is easy to raw from) is enote as q(θ θ t 1 ) > 0, with θ, θ t 1 R D. In Table 3, we escribe the stanar MH algorithm in etail. The algorithm returns the sequence of states {θ 1, θ 2,..., θ t,..., θ T } (or a subset of them removing the burn-in perio if an estimation of its length is available). We can see that the next state θ t can be the propose sample θ (with probability α) or the previous state θ t 1 (with probability 1 α). Uner some mil regularity conitions, when t grows, the pf of the current state θ t converges to the target ensity π(θ) [9]. The MH algorithm satisfies the so-calle etaile 2 In this work, for simplicity, we use the same notation for enoting a ranom variable or one realization of a ranom variable. 3 Recall we are assuming that the Markov chain is ergoic an hence the starting value is forgotten. 4 For the sake of simplicity, we use all the generate states in the final estimators, without removing any burn-in perio [9]. 5
6 Table 3: The MH algorithm 1. Initialization: Choose an initial state θ FOR t = 1,..., T : (a) Draw a sample θ q(θ θ t 1 ). (b) Accept the new state, θ t = θ, with probability [ α(θ t 1, θ ) = min 1, Otherwise, set θ t = θ t Return: {θ t } T t=1. π(θ )q(θ t 1 θ ) π(θ t 1 )q(θ θ t 1 ) ], (6) balance conition which is sufficient to guarantee that the output chain is ergoic an has π as stationary istribution [7, 9]. ote that the acceptance probability α can be rewritten as [ ] [ ] α(θ t 1, θ π(θ )q(θ t 1 θ ) ) = min 1, = min 1, w(θ θ t 1 ). (7) π(θ t 1 )q(θ θ t 1 ) w(θ t 1 θ ) 110 π(θ t 1) q(θ t 1 θ ) where we have enote w(θ θ t 1 ) = π(θ ) an w(θ q(θ θ t 1 ) t 1 θ ) = in a similar fashion of the importance sampling weights of θ an θ t 1 [9]. If the proposal pf is inepenent from the previous state, i.e., q(θ θ t 1 ) = q(θ), the acceptance function epens on the ratio of the importance weights w(θ ) = π(θ ) an w(θ q(θ ) t 1) = π(θ t 1) q(θ t 1, as shown in Table 4. We refer to this ) special MH case as the Inepenent MH (I-MH) algorithm. It is strictly relate to other techniques escribe in the following (e.g., see Section 4.2.1). 4 MCMC using multiple caniates In the stanar MH technique escribe above, at each iteration one new sample θ is generate to be teste with the previous state θ t 1 by the acceptance probability α(θ t 1, θ ). Other generalize MH schemes generate several caniates at each iteration to be teste as new possible state. In all these schemes, an extene acceptance probability α is properly esigne in orer to guarantee the ergoicity of the chain. Figure 1 provies a graphical representation of the ifference between MH an the techniques using several caniates. Below, we escribe the most important examples of this class of MCMC algorithms. In most of them, a single MH-type test is performe at each iteration whereas in other methos a sequence of tests is employe. Furthermore, most of these techniques use an Importance Sampling (IS) approximation of the target ensity [8, 9] in orer to improve the proposal proceure employe 6
7 Table 4: The Inepenent MH (I-MH) algorithm 1. Initialization: Choose an initial state θ FOR t = 1,..., T : (a) Draw a sample θ q(θ). (b) Accept the new state, θ t = θ, with probability [ α(θ t 1, θ ) = min 1, Otherwise, set θ t = θ t Return: {θ t } T t=1. w(θ ) w(θ t 1 ) ], (8) t 1 TEST 0 t 1 (1) (2) ()... TEST t (a) Stanar MH t (b) MCMC using multiple tries Figure 1: Graphical representation of the classical MH metho an the MCMC schemes using ifferent caniates at each iteration within a MH-type algorithm. amely, they buil an IS approximation, an then raw one sample from this approximation (resampling step). Finally, the selecte sample is compare with the previous state of the chain, θ t 1, accoring to a suitable generalize acceptance probability α. It can be prove that all the methoologies presente in this work yiel a ergoic chain with the posterior π as invariant ensity. 4.1 The Multiple Try Metropolis (MTM) algorithm The Multiple Try Metropolis (MTM) algorithms are examples of this class of methos, where samples θ (1), θ (2),..., θ () (calle also tries or caniates ) are rawn from the proposal pf q(θ), at each iteration [13, 14, 15, 16, 17, 38, 39]. Then, one of them is selecte accoring to some suitable weights. Finally, the selecte caniate is accepte or rejecte as new state accoring to a generalize probability function α. The MTM algorithm is given in Table 5. For the sake of simplicity, we have consiere the use of the importance weights w(θ θ t 1 ) =, but there is not a unique possibility, as also π(θ) q(θ θ t 1 ) 7
8 shown below [13, 38]. In its general form, when the proposal epens on the previous state of the chain q(θ θ t 1 ), the MTM requires the generation of 1 auxiliary samples, v (i), which are employe in the computation of the acceptance function α. They are neee in orer to guarantee the ergoicity. Inee, the resulting MTM kernel satisfies the etaile balance conition, so that the chain is reversible [13, 38]. ote that for = 1, we have θ (j) = θ (1), v (1) = θ t 1 an the acceptance probability of the MTM metho becomes ( ) α(θ t 1, θ (1) ) = min 1, w(θ(1) θ t 1 ), w(v (1) θ (1) ) ( = min 1, w(θ(1) θ t 1 ) w(θ t 1 θ (1) ) ) ( ) = min 1, π(θ(1) )q(θ t 1 θ (1) ), (9) π(θ t 1 )q(θ (1) θ t 1 ) that is the acceptance probability of the classical MH technique. Several variants have been stuie, for instance, with correlate tries an consiering the use of ifferent proposal pfs [38, 40]. Remark 1. The MTM metho in Table 5 nees at step 2 the generation of 1 auxiliary samples an at step 2e the computation of their weights (an, as a consequence, 1 aitional evaluation of the target pf are require), that are only employe in the computation of the acceptance function α Generic form of the weights The importance weights are not the unique possible choice. It is possible to show that the MTM algorithm generates an ergoic chain with invariant ensity π, if the weight function w(θ θ t 1 ) is chosen with the form w(θ θ t 1 ) = π(θ)q(θ t 1 θ)ξ(θ t 1, θ), (13) where ξ(θ t 1, θ) = ξ(θ, θ t 1 ), θ, θ t 1 D. 1 For instance, choosing ξ(θ t 1, θ) =, we obtain the importance weights w(θ θ q(θ θ t 1 )q(θ t 1 θ) t 1) = π(θ) q(θ θ t 1 use above. If we set ξ(θ ) t 1, θ) = 1, we have w(θ θ t 1 ) = π(θ)q(θ t 1 θ). Another interesting example can be employe if the proposal is symmetric, i.e., q(θ θ t 1 ) = q(θ t 1 θ). In 1 this case, we can choose ξ(θ t 1, θ) = an then w(θ θ q(θ t 1 θ) t 1) = w(θ) = π(θ), i.e., the weights only epen on the value of the target ensity at θ. Thus, MTM contains the Orientational Bias Monte Carlo (OBMC) scheme [12, Chapter 13] as a special case, when a symmetric proposal pf is employe, an then one caniate is chosen with weights proportional to the target ensity, i.e., w(θ θ t 1 ) = π(θ) Inepenent Multiple Try Metropolis (I-MTM) schemes The MTM metho escribe in Table 5 requires to raw 2 1 samples at each iteration ( caniates an 1 auxiliary samples) an 1 are only use in the acceptance probability function. The generation of the auxiliary points v (1),..., v (j 1), v (j+1),..., v () q(θ θ (j) ), 8
9 Table 5: The MTM algorithm with importance sampling weights. 1. Initialization: Choose an initial state θ FOR t = 1,..., T : (a) Draw θ (1), θ (2),..., θ () q(θ θ t 1 ). (b) Compute the importance weights w(θ (n) θ t 1 ) = π(θ(n) ), with n = 1,...,. (10) q(θ (n) θ t 1 ) (c) Select one sample θ (j) {θ (1),..., θ () }, accoring to the probability mass function w n = w(θ(n) θ t 1 ) P. i=1 w(θ(i) θ t 1 ) () Draw 1 auxiliary samples from q(θ θ (j) ), enote as v (1),..., v (j 1), v (j+1),..., v () q(θ θ (j) ), an set v (j) = θ t 1. (e) Compute the weights of the auxiliary samples w(v (n) θ (j) ) = π(v(n) ), with n = 1,...,. (11) q(v (n) θ (j) ) (f) Set θ t = θ (j) with probability α(θ t 1, θ (j) ) = ( ) min 1, θ t 1 ), w(v(n) θ (j) ) (12) otherwise, set θ t = θ t Return: {θ t } T t=1. can be avoie if the proposal pf is inepenent from the previous state, i.e., q(θ θ t 1 ) = q(θ). Inee, in this case, we shoul raw 1 samples again from q(θ) at the step 2 of Table 5. Since we have alreay rawn samples from q(θ) at step 2a of Table 5, we can set v (1) = θ (1),..., v (j 1) = θ (j 1), v (j) = θ (j+1)... v ( 1) = θ (), (14) without jeoparizing the ergoicity of the chain (recall that v (j) = θ t 1 ). Hence, we can avoi step 2 an the acceptance function can be rewritten as ( α(θ t 1, θ (j) ) = min 1, w(θ(j) ) + ),n j w(θ(n) ) w(θ t 1 ) +. (15),n j w(θ(n) ) 9
10 The I-MTM algorithm is provie in Table Remark 2. An I-MTM metho requires only new evaluations of the target pf at each iteration, instea of 2 1 new evaluations in the generic MTM scheme in Table 5. ote that we can also write α(θ t 1, θ (j) ) as ( ) α(θ t 1, θ (j) ) = min 1, Ẑ1 Ẑ 2, (16) where we have enote Ẑ 1 = 1 w(θ (n) ), Ẑ 2 = 1 ( w(θ t 1 ) +,n j ) w(θ (n) ). (17) Table 6: The Inepenent Multiple Try Metropolis (I-MTM) algorithm. 1. Initialization: Choose an initial state θ FOR t = 1,..., T : (a) Draw θ (1), θ (2),..., θ () q(θ). (b) Compute the importance weights (c) Select one sample θ (j) function w n = w(θ(n) ) w(θ (n) ) = π(θ(n) ), with n = 1,...,. (18) q(θ (n) ) P i=1 w(θ(i) ). {θ (1),..., θ () }, accoring to the probability mass () Set θ t = θ (j) with probability ( α(θ t 1, θ (j) ) = min 1, w(θ(j) ) + ),n j w(θ(n) ) w(θ t 1 ) +,,n j w(θ(n) ) ( ) otherwise, set θ t = θ t Return: {θ t } T t=1. = min 1, Ẑ1 Ẑ 2, (19) 10
11 Alternative version (I-MTM2). From the IS theory, we know that Ẑ1 = 1 w(θ(n) ) is an unbiase estimator of the normalizing constant Z of the target π (a.k.a, Bayesian evience or marginal likelihoo). It suggests to replace Ẑ2 with other unbiase estimators of Z (without jeoparizing the ergoicity of the chain). For instance, instea of recycling the samples generate in the same iteration as auxiliary points as in Eq. (14), we coul reuse samples generate in the previous iteration t 1. This alternative version of I-MTM metho (I-MTM2) is given in Table 7. ote that, in both cases I-MTM an I-MTM2, the selecte caniate θ (j) is rawn from the following particle approximation of the target π, π(θ θ (1:) ) = i=1 w(θ (i) )δ(θ θ (i) ), w i = w(θ (i) ) = w(θ (i) ) w(θ(n) ), (20) 170 i.e., θ (j) π(θ θ (1:) ). The acceptance probability α use in I-MTM2 can be also justifie consiering a proper IS weighting of a resample particle [41] an using the expression (7) relate to the stanar MH metho, as iscusse in [29]. Figure 2 provies a graphical representation of the I-MTM schemes. Table 7: Alternative version of I-MTM metho (I-MTM2). 1. Initialization: Choose an initial state θ 0, an obtain an initial approximation Ẑ 0 Z. 2. FOR t = 1,..., T : (a) Draw θ (1), θ (2),..., θ () q(θ). (b) Compute the importance weights w(θ (n) ) = π(θ(n) ), with n = 1,...,. (21) q(θ (n) ) (c) Select one sample θ (j) {θ (1),..., θ () }, accoring to the probability mass function w n = 1 Z b w(θ(n) ) where Ẑ = 1 i=1 w(θ(i) ). () Set θ t = θ (j) an Ẑt = Ẑ with probability α(θ t 1, θ (j) ) = min otherwise, set θ t = θ t 1 an Ẑt = Ẑt 1. ( 1, Ẑ Ẑ t 1 ), (22) 11
12 Generation (1),..., () q( ) Approximation (1) b ( ) (2) ( 1) ()... Resampling Acceptance Test (j) b ( ) ( t 1, (j) ) t Figure 2: Graphical representation of the I-MTM schemes Reusing caniates in parallel I-MTM chains Let us consier to run C inepenent parallel chains yiele by an I-MTM scheme. In this case, we have C evaluations of the target function π an C resampling steps performe at each iteration (so that we have CT total target evaluations an CT total resampling steps). In literature, ifferent authors have suggeste to recycle the caniates, θ (1), θ (2),..., θ () q(θ), in orer to reuce the number of evaluations of the target pf [42]. The iea is to performs C-times the resampling proceure consiering the same set of caniates, θ (1), θ (2),..., θ () (a similar approach was propose in [20]). Each resample caniate is then teste as possible future state of one chain. In this scenario, The number of target evaluations per iteration is only (hence, the total number of evaluation of π is T ). However, the resulting C parallel chains are no longer inepenent, an there is a lose of performance w.r.t. the inepenent chains. There exists also the possibility of reucing the total number of resampling steps, as suggeste in the Block Inepenent MTM scheme [42] (but the epenence among the chains grows even more). 4.2 Particle Metropolis-Hastings (PMH) metho Assume that the variable of interest is forme by only a ynamical variable, i.e., θ = x = x 1:D = [x 1..., x D ] (see Section 2). This is the case of inferring a hien state in state-space moel, for instance. More generally, let assume that we are able to factorize the target ensity as π(x) π(x) = γ 1 (x 1 )γ 2 (x 2 x 1 ) γ D (x D x 1:D 1 ), (23) D = γ 1 (x 1 ) γ (x x 1 ). (24) =2 The Particle Metropolis Hastings (PMH) metho [43] is an efficient MCMC technique, propose inepenently from the MTM algorithm, specifically esigne for being applie in this framework. Inee, we can take avantage of the factorization of the target pf an consier a proposal pf ecompose in the same fashion q(x) = q 1 (x 1 )q 2 (x 2 x 1 ) q D (x D x 1:D 1 ) = q 1 (x 1 ) D q (x x 1 ). =2 12
13 Then, as in a batch IS scheme, given an n-th sample x (n) = x (n) 1:D we assign the importance weight w(x (n) ) = w (n) D = π(x(n) ) q(x (n) ) = γ 1(x (n) q(x) with x(n) q (x x 1 ), 1 )γ 2 (x (n) 2 x (n) 1 ) γ D (x (n) D x(n) 1:D 1 ) q 1 (x (n) 1 )q 2 (x (n) 2 x (n) 1 ) q D (x (n) D x(n) 1:D 1 ). (25) The previous expression suggests a recursive proceure for computing the importance weights: starting with w (n) 1 = π(x(n) 1 ) an then ) where we have set q(x (n) 1 w (n) = w (n) 1 β(n) = j=1 β (n) 1 = w (n) 1 an β (n) β (n) j, = 1,..., D, (26) = γ (x (n) x(n) 1: 1 ) q (x (n) x(n) 1: 1 ), (27) for = 2,..., D. This metho is usually referre as Sequential Importance Sampling (SIS). If resampling steps are also employe at some iteration, the metho is calle Sequential Importance Resampling (SIR), a.k.a., particle filtering (PF) (see Appenix B). PMH uses a SIR approach for proviing the particle approximation π(x x (1:) ) = i=1 w(i) D δ(x x(i) ) where w (i) w (i) D = w(i) D P w(n) D D = w(x(i) ), obtaine using Eq. (26) (with a proper weighting of a resample particle [41, 29]). Then, one particle is rawn from this approximation, i.e., with a probability proportional to the corresponing normalize weight. Estimation of the marginal likelihoo Z in particle filtering. SIR combines the SIS approach with the application of resampling proceures. In SIR, a consistent estimator of Z is given by [ D ] Z =, where w (i) 1 = =1 w (n) 1 β(n) Due to the application of the resampling, in SIR the stanar estimator an w (i) 1. (28) w(n) 1 Ẑ = 1 w (n) D = 1 w(x (n) ), (29) 190 is a possible alternative only if a proper weighting of the resample particles is applie [29, 41] (otherwise, it is not an estimator of Z). If a proper weighting of a resample particle is employe, both Z an Ẑ are equivalent estimators of Z [41, 29, 28]. Without the use of resampling steps (i.e., in SIS), Z an Ẑ are always equivalent estimators [29]. See also Appenix B. 195 The complete escription of PMH is provie in Table 8 consiering the use of Z. At each iteration, a particle filter is run in orer to provie an approximation by weighte samples of the 13
14 Table 8: Particle Metropolis-Hastings (PMH) algorithm. 1. Initialization: Choose a initial state x 0, obtain an initial estimation Z 0 Z. 2. For t = 1,..., T : (a) We employ a SIR approach for rawing with particles an weighting them, {x (i), w (i) D } i=1, i.e., we obtain sequentially a particle approximation π(x) = i=1 w(i) D δ(x x(i) ) where x (i) = [x (i) 1,..., x (i) D ]. Furthermore, we also obtain Z as in Eq. (28). (b) Draw x π(x x (1:) ), i.e., choose a particle x = {x (1),..., x () } with probability, i = 1,...,. w (i) D (c) Set x t = x an Z t = Z with probability [ α = min otherwise set x t = x t 1 an Z t = Z t 1. () Return: {x t } T t=1 where x t = [x 1,t,..., x D,t ]. 1, Z Z t 1 ], (30) 200 measure of the target. Then, a sample among the weighte particles is chosen by one resampling step. This selecte sample is then accepte or rejecte as next state of the chain accoring to an MH-type acceptance probability, which involves two estimators of marginal likelihoo Z. PMH is also relate to other popular metho in molecular simulation calle Configurational Bias Monte Carlo (CBMC) [44] Relationship among I-MTM2, PMH an I-MH A simple look at I-MTM2 an PMH shows that they are strictly relate [28]. Inee, the structure of the two algorithms coincies. The main ifference lies that the caniates in PMH are generate sequentially, using a SIR scheme. If no resampling steps are applie, then I-MTM2 an PMH are exactly the same algorithm, where the caniates are rawn in a batch setting or sequential way. Hence, the application of resampling steps is the main ifference between the generation proceures of PMH an I-MTM2. Owing to the use of resampling, the caniates {x (1),..., x () } propose by PMH are not inepenent (ifferently from I-MTM2). As an example, Figure 3 shows = 40 particles (with D = 10) generate an weighte by SIS an SIR proceures (each path is a generate particle x (i) = x (i) 1:10). The generation of correlate samples can be also consiere 14
15 in MTM methos without jeoparizing the ergoicity of the chain, as simply shown for instance in [16], for instance. Another ifference is the use of Z or Ẑ. However, if a proper weighting of a resample particle is employe, both estimators coincie [29, 41, 28]. Furthermore, both I-MTM2 an PMH can be consiere as I-MH schemes where a proper importance sampling weighting of a resample particle is employe [41]. amely, I-MTM2 an PMH are equivalent to an I-MH technique using the following complete proposal pf, [ ] q(θ) = π(θ θ (1:) ) q(θ (i) ) θ (1:), (31) D where π is given in Eq. (20), i.e., θ (j) q(θ), an then consiering the generalize (proper) IS weighting, w(θ (j) ) = Ẑ, w(θ t 1 ) = Ẑt 1 [29, 41]. For further etails see Appenix A. i= (a) Batch-IS or SIS. (b) SIR with resampling at = 4, 8. Figure 3: Graphical representation of SIS an SIR. We consier as target ensity a multivariate Gaussian pf, π(x) = 10 =1 (x 2, 1 2 ). In each figure, every component of ifferent particles are represente, so that each particle x (i) = x (i) 1:D forms a path (with D = 10). We set = 40. The normalize weights w(i) D = w(x(i) ) corresponing to each figure are also shown in the bottom. The line-with of each path is proportional to the corresponing weight w (i) D. The particle corresponing to the greatest weight is always epicte in black. The proposal pfs use are q 1 (x 1 ) = (x 1 2, 1) an q(x x 1 ) = (x x 1, 1) for 2. (a) Batch IS or SIS. (b) SIR resampling steps at the iterations = 4, Particle Marginal Metropolis-Hastings (PMMH) metho Assume now that the variable of interest if forme by both ynamical an static variables, i.e., θ = [x, λ]. For instance, this is the case of inferring both, an hien state x in state-space moel, an static parameters λ of the moel. The Particle Marginal Metropolis-Hastings (PMMH) technique is a extension of PMH which aresses this problem. 15
16 210 Let us consier x = x 1:D = [x 1, x 2,..., x D ] R x, an an aitional moel parameter λ R λ to be inferre as well (θ = [x, λ] R D, with D = x + λ ). Assuming a prior pf g λ (λ) over λ, an a factorize complete posterior pf π(θ) = π(x, λ), π(x, λ) π(x, λ) = g λ (λ)π(x λ), where π(x λ) = γ 1 (x 1 λ) D =2 γ (x x 1: 1, λ). For a specific value of λ, we can use a particle filter approach, obtaining the approximation π(x λ) = w(n) D δ(x x(n) ) an the estimator Z(λ), as escribe above. The PMMH technique is then summarize in Table 9. The pf q λ (λ λ t 1 ) enotes the proposal ensity for generating possible values of λ. Observe that, with the specific choice q λ (λ λ t 1 ) = g λ (λ), then the acceptance function becomes [ ] Z(λ ) α = min 1,. (32) Z(λ t 1 ) ote also that PMMH w.r.t. to λ can be interprete as MH metho where an the posterior cannot be evaluate point-wise. Inee, Z(λ) approximates the marginal likelihoo p(y λ) [45]. Table 9: Particle Marginal MH (PMMH) algorithm 1. Initialization: Choose the initial states x 0, λ 0, an an initial approximation Z 0 (λ) Z(λ) p(y λ). 2. For t = 1,..., T : (a) Draw λ q λ (λ λ t 1 ). (b) Given λ, run a particle filter obtaining π(x λ ) = w(n) D δ(x x(n) ) an Z(λ ), as in Eq. (28). (c) () Draw x π(x λ, x (1:) ), i.e., choose a particle x = {x (1),..., x () } with probability w (i) D, i = 1,...,. (e) Set λ t = λ, x t = x, with probability [ ] Z(λ )g λ (λ )q λ (λ t 1 λ ) α = min 1,. (33) Z(λ t 1 )g λ (λ t 1 )q λ (λ λ t 1 ) Otherwise, set λ t = λ an x t = x t Return: {x t } T t=1 an {λ t } T t=1. 16
17 Group Metropolis Sampling The auxiliary weighte samples in the I-MTM schemes (i.e., the 1 samples rawn at each iteration that are not selecte to be compare with the previous state θ t 1 ) can be recycle proviing a consistent an more efficient estimators [41, 29]. The so-calle Group Metropolis Sampling (GMS) metho is shown in Table 10. GMS yiels a sequence of sets of weighte samples S t = {θ n,t, ρ n,t }, for t = 1,..., T, where we have enote with ρ n,t the importance weights assigne to the samples θ n,t (see Figure 4). All the samples are then employe for a joint particle approximation of the target. Alternatively, GMS can irectly provie an approximation of a specific moment of the target pf (i.e., given a particular function f). The estimator of this specific moment provie by GMS is Ĩ T = 1 T T t=1 ρ n,t i=1 ρ f(θ n,t ) = 1 i,t T T t=1 Ĩ (t). (34) Unlike in the I-MTM schemes, no resampling steps are performe in GMS. However, we can recover an I-MTM chain from the GMS output applying one resampling step when S t S t 1, i.e., ρ n,t θ t θ t = i=1 ρ δ(θ θ n,t ), if S t S t 1, i,t (38) θ t 1, if S t = S t 1, for t = 1,..., T. More specifically, {θ t } T t=1 is a Markov chain obtaine by one run of an I-MTM2 technique. The consistency of the GMS estimators is iscusse in Appenix C. GMS can be also interprete as an iterative IS scheme where an IS approximation of samples is built at each iteration an compare with the previous IS approximation. This proceure is iterate T times an all the accepte IS estimators Ĩ(t) are finally combine to provie a unique global approximation of T samples. ote that the temporal combination of the IS estimators is obtaine ynamically by the ranom repetitions ue to the rejections in the acceptance test. Remark 3. The complete weighting proceure in GMS can be interprete as the composition of two weighting schemes: (a) by an IS approach builing {ρ n,t } an (b) by the possible ranom repetitions ue to the rejections in the acceptance test. Figure 4 epicts a graphical representation of the GMS outputs as chain of sets S t = {θ n,t, ρ n,t }. 4.4 Ensemble MCMC algorithms Another alternative proceure, often referre as Ensemble MCMC (EnMCMC) methos (a.k.a., calle Locally weighte MCMC), involving several tries at each iteration [19, 21]. Relate techniques has been propose inepenently in ifferent works [20, 22]. First, let us efine the joint proposal ensity q(θ (1),..., θ () θ t ) : D R, (39) an, consiering +1 possible elements, S = {θ (1),..., θ (), θ (+1) }, we efine the D matrix Θ k = [θ (1),..., θ (k 1), θ (k+1),..., θ (+1) ], (40) 17
18 Table 10: Group Metropolis Sampling 1. Initialization: Choose an initial state θ 0 an an initial approximation Ẑ0 Z. 2. FOR t = 1,..., T : (a) Draw θ (1), θ (2),..., θ () q(θ). (b) Compute the importance weights w(θ (n) ) = π(θ(n) ), with n = 1,...,, (35) q(θ (n) ) efine S = {θ (n), w(θ (n) )} ; an compute Ẑ = 1 w(θ(n) ). (c) Set S t = S, i.e., an Ẑt = Ẑ, with probability S t = { θ n,t = θ (n), ρ n,t = w(θ (n) ) }, α(s t 1, S ) = min Otherwise, set S t = S t 1 an Ẑt = Ẑt Return: All the sets {S t } T t=1, or {Ĩ(t) }T t=1 where Ĩ (t) = [ 1, Ẑ Ẑ t 1 ]. (36) ρ n,t i=1 ρ i,t g(θ n,t ), (37) an ĨT = 1 T T t=1 Ĩ(t). S t 1 S t S t+1 Figure 4: Chain of sets S t = {θ n,t, ρ n,t } generate by the GMS metho (graphical representation with = 4). 18
19 with columns all the vectors in S with the exception of θ (k). For simplicity, in the followings we abuse of the notation writing q(θ (1),..., θ () θ t ) = q(θ +1 θ t ), for instance. One simple example of joint proposal pf is q(θ (1),..., θ () θ t ) = q(θ (n) θ t ), (41) i.e., consiering inepenence among θ (n) s (an having the same marginal proposal pf q). More sophisticate joint proposal ensities can be employe. A generic EnMCMC algorithm is outline in Table 11. Table 11: Generic EnMCMC algorithm 1. Initialization: Choose an initial state θ FOR t = 1,..., T : (a) Draw θ (1),..., θ () q(θ (1),..., θ () θ t 1 ). (b) Set θ (+1) = θ t 1. (c) Set θ t = θ (j), resampling θ (j) within the set { θ (1),..., θ (), θ (+1) = θ t 1 }, forme by + 1 samples, accoring to the probability mass function α(θ t 1, θ (j) ) = π( θ (j) )q(θ j θ (j) ) +1 l=1 π( θ (l) )q(θ l θ (l) ), for j = 1,..., + 1 an Θ j is efine in Eq. (40). 3. Return: {θ t } T t= Remark 4. ote that with respect to the generic MTM metho, EnMCMC oes not require to raw auxiliary samples an weights. Therefore, in EnMCMC a smaller number of evaluation of target is require w.r.t. a generic MTM scheme Inepenent Ensemble MCMC In this section, we present an interesting special case, which employs a single proposal pf q(θ) inepenent on the previous state of the chain, i.e., q(θ (1),..., θ () θ t ) = q(θ (1),..., θ () ) = q(θ (n) ). (42) 19
20 In this case, the technique can be simplifie as shown below. At each iteration, the algorithm escribe in Table 12 generates new samples θ (1), θ (2),..., θ () an then resample the new state θ t within a set of + 1 samples, {θ (1),..., θ (), θ t+1 } (which inclues the previous state), accoring to the probabilities α(θ t 1, θ (j) ) = w j = w(θ (j) ), j = 1,..., + 1, (43) i=1 w(θ(i) ) + w(θ t 1 ) 245 where w(θ) = π(θ) q(θ) enotes the importance sampling weight. ote that Eq. (43) for = 1 becomes α(θ t 1, θ (j) ) = = = w(θ (j) ) w(θ (i) ) + w(θ t 1 ), π(θ (j) ) q(θ (j) ) π(θ (j) ) + π(θ t 1) q(θ (j) ) q(θ t 1 ), π(θ (j) )q(θ t 1 ) π(θ (j) )q(θ t 1 ) + π(θ t 1 )q(θ (j) ), (44) that is the Barker s acceptance function (see [46, 47]). Table 12: EnMCMC with an inepenent proposal pf (I-EnMCMC). 1. Initialization: Choose an initial state θ FOR t = 1,..., T : (a) Draw θ (1), θ (2),..., θ () q(θ), an set θ (+1) = θ t 1. (b) Compute the importance weights w(θ (n) ) = π(θ(n) ), with n = 1,..., + 1. (45) q(θ (n) ) (c) Set θ t = θ (j), resampling θ (j) within the set {θ (1),..., θ (), θ t+1 } forme by + 1 samples, accoring to the probability mass function α(θ t 1, θ (j) ) = w j = w(θ (j) ) i=1 w(θ(i) ) + w(θ t 1 ). 3. Return: {θ t } T t=1. As iscusse in [42, Appenix B], [48, Appenix C], [41], the ensity of a resample caniate becomes closer an closer to π as grows, i.e.,. Hence, the performance of I-EnMCMC 20
21 250 clearly improves with (see Appenix A). The I-EnMCMC algorithm prouces an ergoic chain with invariant ensity π, by resampling + 1 samples at each iteration ( new samples θ (1),..., θ () from q an setting θ (+1) = θ t 1 ). Figure 5 summarizes the steps of I-EnMCMC. Approximation Generation (1),..., () q( ) (1) (+1) = (2) ( 1) t 1 ()... b ( ) Resampling (j) b ( ) t Figure 5: Graphical representation of the steps of the I-EnMCMC scheme. 4.5 Delaye Rejection Metropolis (DRM) Sampling An alternative use of ifferent caniates in one iteration of a Metropolis-type metho is given in [23, 24, 25]. The iea behin the propose algorithm, calle Delaye Rejection Metropolis (DRM) algorithm, is the following. As in a stanar MH metho, at each iteration, one sample is propose θ (1) q 1 (θ θ t 1 ) an accepte with probability ( ) α 1 (θ t 1, θ (1) ) = min 1, π(θ(1) )q 1 (θ t 1 θ (1) ). π(θ t 1 )q 1 (θ (1) θ t 1 ) 255 If θ (1) is accepte then θ t = θ (1) an the chain is move forwar. If θ (1) is rejecte, the DRM metho suggests of rawing another samples θ (2) q 2 (θ θ (1), θ t 1 ) (consiering a ifferent proposal pf q 2 taking into account possibly the previous caniate θ (1) ) an accepte with a suitable acceptance probability α 2 (θ t 1, θ (2) ) = ( = min 1, π(θ (2) )q 1 (θ (1) θ (2) )q 2 (θ t 1 θ (1), θ (2) )(1 α 1 (θ (2), θ (1) )) π(θ t 1 )q 1 (θ (1) θ t 1 )q 2 (θ (2) θ (1), θ t 1 )(1 α 1 (θ t 1, θ (1) )) ) The acceptance function α 2 (θ t 1, θ (2) ) is esigne in orer to ensure the ergoicity of the chain. If θ (2) is rejecte we can set θ t = θ t 1 an perform another iteration of the algorithm, or continue with this iterative strategy rawing θ (3) q 3 (θ θ (2), θ (1), θ t 1 ) an test it with a proper probability α 3 (θ t 1, θ (3) ). The DRM algorithm with only 2 acceptance stages is outline in Table 13 an summarize in Figure 6. Remark 5. ote that the proposal pf can be improve at each intermeiate stage (θ (1) q 1 (θ θ t 1 ), θ (2) q 2 (θ θ (1), θ t 1 ) etc.), using the information provie by the previous generate samples an the corresponing target evaluations. The iea behin DRM of creating a path of intermeiate points, then improving the proposal pf, an hence fostering larger jumps have been also consiere in other works [14, 49]. 21
22 Table 13: Delaye Rejection Metropolis algorithm with 2 acceptance steps. 1. Initialization: Choose an initial state θ FOR t = 1,..., T : (a) Draw θ (1) q 1 (θ θ t 1 ) an u 1 U([0, 1]). (b) Define the probability (c) If u 1 α 1 (θ t 1, θ (1) ), set θ t = θ (1). ( ) α 1 (θ t 1, θ (1) ) = min 1, π(θ(1) )q 1 (θ t 1 θ (1) ), (46) π(θ t 1 )q 1 (θ (1) θ t 1 ) () Otherwise, if u 1 > α 1 (θ t 1, θ (1) ), o: (1) Draw θ (2) q 2 (θ θ (1), θ t 1 ) an u 2 U([0, 1]). (2) Given the function ψ(θ t 1, θ (2) θ (1) ) = π(θ t 1 )q 1 (θ (1) θ t 1 ) q 2 (θ (2) θ (1), θ t 1 )(1 α 1 (θ t 1, θ (1) )), an the probability ( ) α 2 (θ t 1, θ (2) ) = min 1, ψ(θ(2), θ t 1 θ (1) ). (47) ψ(θ t 1, θ (2) θ (1) ) (3) If u 2 α 2 (θ t 1, θ (2) ), set θ t = θ (2). (4) Otherwise, if u 2 > α 1 (θ t 1, θ (2) ), set θ t = θ t Return: {θ t } T t=1. Generation Acceptance Test Generation Acceptance Test o (2) q 1 ( (1), t 1 ) ( t 1, (2) ) t (1) q 1 ( t 1 ) ( t 1, (1) ) Yes t = (1) Figure 6: Graphical representation of the DRM scheme with two acceptance tests. 22
23 5 Summary: computational cost, ifferences an connections The performance of the algorithms escribe above improves as grows, in general: the correlation among samples vanishes to zero, an the acceptance rate of new state approaches one (see Section 6.1). 5 As increases, they become similar an similar to an exact sampler rawing inepenent samples irectly from the target ensity (for MTM, PMH an EnMCMC schemes the explanation is given in Appenix A). However, this occurs at the expense of an aitional computational cost. In Table 14, we summarize the total number of target evaluations, E, an the total number of samples use in the final estimators, Q (without consiering to remove any burn-in perio). The generic MTM algorithm has the greatest number of target evaluations. However, a ranom-walk proposal pf can be use in a generic MTM algorithm an, in general, it fosters the exploration of the state space. In this sense, the generic EnMCMC seems to be preferable w.r.t. MTM, since E = T an the ranom-walk proposal can be applie. A isavantage of the EnMCMC schemes is that their acceptance function seems worse in terms of Peskun s orering [50] (see numerical results in Section 6.1). amely, fixing the number of tries, the target π the proposal q pfs, the MTM schemes seem to provie greater acceptance rates than the corresponing EnMCMC techniques. This is theoretically prove for = 1 [50], an the ifference vanishes to zero as grows. The GMS technique, like other strategies [20, 42], has been propose to recycle samples or re-use target evaluations, in orer to increase Q (see also Section 4.1.3). Table 14: Total number of target evaluations, E, an total number of samples, Q, use in the final estimators. Algorithm MTM I-MTM I-MTM2 PMH E 2T 1 T T T Q T T T T Algorithm GMS EnMCMC I-EnMCMC DMR E T T T T Q T T T T 290 In PMH, the components of the ifferent tries are rawn sequentially an they are correlate ue to the application of the resampling steps. In DMR, each caniate is rawn in a batch way (all the components jointly) but the ifferent caniates are rawn in a sequential manner (see Figure 6), θ (1) then θ (2) etc. The benefit of this strategy is that the proposal pf can be improve consiering the previous generate tries. Hence, if the proposal takes into account the previous samples, DMR generates correlate caniates as well. The main isavantage of DRM is that 5 Generally, an acceptance rate close to 1 is not an evience of goo performance for an MCMC algorithm. However, for the techniques tackle in this work, the situation is ifferent: as grows, the proceure use for proposing a novel possible state (involving tries, resampling steps etc.) becomes better an better, yieling a better approximation of the target pf. See Appenix A for further etails. 23
24 the implementation for a generic > 2 is not straightforwar. I-MTM an I-MTM2 iffers for the acceptance function employe. Furthermore, The main ifference between the I-MTM2 an PMH schemes is the use of resampling steps uring the generation the ifferent tries. For this reason, the caniates of PMH are correlate (unlike in I-MTM2). I-MTM2 an PMH can be interprete as I-MH methos using a sophisticate proposal ensity q(θ) in Eq. (31), an an extene IS weighting proceure is employe. ote that, inee, q cannot be evaluate pointwise, hence a stanar IS weighting strategy cannot be employe. 6 umerical Experiments We test ifferent MCMC using multiple caniates in ifferent numerical experiments. In the first example, an exhaustive comparison among several techniques with an inepenent proposal is given. We have consiere ifferent number of tries, length of the chain, parameters of the proposal pfs an also ifferent imension of the inference problem. In the secon numerical simulation, we compare ifferent particle methos. The thir one regars the hyperparameter selection for a Gaussian Process (GP) regression moel. The last two examples are localization problems in a wireless sensor network (WS): in the fourth one some parameters of the WS are also tune, whereas in last example a real ata analysis is performe. 6.1 A first comparison of efficiency In orer to compare the performance of ifferent techniques, in this section we consier a multimoal, multiimensional Gaussian target ensity. More specifically, we have π(θ) = 1 3 (θ µ i, Σ i ), θ R D, (48) 3 i=1 where µ 1 = [µ 1,1,..., µ 1,D ], µ 2 = [µ 2,1,..., µ 2,D ], µ 3 = [µ 3,1,..., µ 3,D ], with µ 1, = 3, µ 2, = 0, µ 3, = 2 for all = 1,..., D. Moreover, the covariance matrices are iagonal, Σ i = δ i I D (where I D is the D D ientity matrix), with δ i = 0.5 for i = 1, 2, 3. Hence, given a ranom variable Θ π(θ), we know analytically that E[Θ] = [ θ 1,..., θ D ] with θ = 1 for all, an 3 iag{cov[θ]} = [ξ 1,..., ξ D ] with ξ = 85 for all = 1,..., D. 18 We apply I-MTM, IMTM2 an I-EnMCMC in orer to estimate all the expecte values an all the variances of the marginal target pfs. amely, for a given imension D, we have to estimate all { θ } D =1 an {ξ } D =1, hence 2D values. The results are average over 3000 inepenent runs. At each run, we compute an average square error obtaine in the estimation of the 2D values an then calculate the Mean Square Error (MSE) average over the 3000 runs. For all the techniques, we consier a Gaussian proposal ensity q(θ) = (θ µ, σ 2 I D ) with µ = [µ 1 = 0,..., µ D = 0] (inepenent from the previous state) an ifferent values of σ are consiere Description of the experiments We perform several experiments varying the number of tries,, the length of the generate chain, T, the imension of the inference problem, D, an the scale parameter of the proposal pf, σ. In 24
25 325 Figures 7(a)-(b)-(c)-()-(e), we show the MSE (obtaine by ifferent techniques) as function of, T, D an σ, respectively. In Figure 7(), we only consier I-MTM with {1, 5, 100, 1000} in orer to show the effect of using ifferent tries in ifferent imensions D. ote that I-MTM with = 1 coincies with I-MH. Let us enote as φ(τ) the auto-correlation function of the states of the generate chain. Figures 8(a)-(b)-(c) epicts the normalize auto-correlation function φ(τ) = φ(τ) (recall that φ(0) φ(τ) φ(0) for τ 0) at ifferent lags τ = 1, 2, 3, respectively. Furthermore, given the efinition of the Effective Sample Size (ESS) [51, Chapter 4], ESS = T 1 + 2, (49) τ=1 φ(τ) 330 in Figure 8(), we show of the ratio ESS (approximate; cutting off the series in the enominator T at lag τ = 10), as function of. 6 Finally, in Figures 9(a)-(b), we provie the Acceptance Rate (AR) of a new state (i.e., the expecte number of accepte jumps to a novel state), as function of an D, respectively Comment on the results Figures 7(a)-(b), 8 an 9(a), clearly show that the performance improves as grows, for all the algorithms. The MSE values an the correlation ecrease, an the ESS an the AR grow. I-MTM seems to provie the best performance. Recall that for = 1, I-EnMCMC becomes an I-MH with Baker s acceptance function an I-MTM becomes the I-MH in Table 4 [47]. For = 1, the results confirm the Peskun s orering about the acceptance function for a MH metho [50, 47]. Observing the results, the Peskun s orering appears vali also for the multiple try case, > 1. I-MTM2 seems to have worse performance than I-MTM for all. With respect to I-EnMCMC, I-MTM2 performs better for smaller. The ifference among the MSE values obtaine by the samplers becomes smaller as grows, as shown in Figure 7(a)-(b) (note that in the first one D = 1, in the other D = 10, an the range of is ifferent). The comparison among I-MTM, I-MTM2 an I-EnMCMC seems not to be affecte by changing T an σ, as epicte in Figures 7(c)-(e). amely, the MSE values change but the orering of the methos (e.g., best an worst) seems to epen mainly on. Obviously, for greater D, more tries are require in orer to obtain goo performance (see Figures 7() an9(b), for instance). ote that for, I-MTM, I-MTM2 an I-EnMCMC perform similarly to an exact sampler rawing T inepenent samples from π(θ): the correlation φ(τ) among the samples approaches zero (for all τ), ESS approaches T an AR approaches 1. 6 Since the MCMC algorithms yiel positive correlate sequences of states, we have ESS < T in general. 25
Group Importance Sampling for particle filtering and MCMC
Group Importance Sampling for particle filtering an MCMC Luca Martino, Víctor Elvira, Gustau Camps-Valls Image Processing Laboratory, Universitat e València (Spain). IMT Lille Douai CRISTAL (UMR 989),
More informationWEIGHTING A RESAMPLED PARTICLE IN SEQUENTIAL MONTE CARLO. L. Martino, V. Elvira, F. Louzada
WEIGHTIG A RESAMPLED PARTICLE I SEQUETIAL MOTE CARLO L. Martino, V. Elvira, F. Louzaa Dep. of Signal Theory an Communic., Universia Carlos III e Mari, Leganés (Spain). Institute of Mathematical Sciences
More informationON MULTIPLE TRY SCHEMES AND THE PARTICLE METROPOLIS-HASTINGS ALGORITHM
ON MULTIPLE TRY SCHEMES AN THE PARTICLE METROPOLIS-HASTINGS ALGORITHM L. Martino, F. Leisen, J. Coraner University of Helsinki, Helsinki (Finlan). University of Kent, Canterbury (UK). ABSTRACT Markov Chain
More informationWEIGHTING A RESAMPLED PARTICLES IN SEQUENTIAL MONTE CARLO (EXTENDED PREPRINT) L. Martino, V. Elvira, F. Louzada
WEIGHTIG A RESAMLED ARTICLES I SEQUETIAL MOTE CARLO (ETEDED RERIT) L. Martino, V. Elvira, F. Louzaa Dep. of Signal Theory an Communic., Universia Carlos III e Mari, Leganés (Spain). Institute of Mathematical
More informationSurvey Sampling. 1 Design-based Inference. Kosuke Imai Department of Politics, Princeton University. February 19, 2013
Survey Sampling Kosuke Imai Department of Politics, Princeton University February 19, 2013 Survey sampling is one of the most commonly use ata collection methos for social scientists. We begin by escribing
More informationThe total derivative. Chapter Lagrangian and Eulerian approaches
Chapter 5 The total erivative 51 Lagrangian an Eulerian approaches The representation of a flui through scalar or vector fiels means that each physical quantity uner consieration is escribe as a function
More informationThe Recycling Gibbs Sampler for Efficient Learning
The Recycling Gibbs Sampler for Efficient Learning L. Martino, V. Elvira, G. Camps-Valls Universidade de São Paulo, São Carlos (Brazil). Télécom ParisTech, Université Paris-Saclay. (France), Universidad
More informationLATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION
The Annals of Statistics 1997, Vol. 25, No. 6, 2313 2327 LATTICE-BASED D-OPTIMUM DESIGN FOR FOURIER REGRESSION By Eva Riccomagno, 1 Rainer Schwabe 2 an Henry P. Wynn 1 University of Warwick, Technische
More informationGaussian processes with monotonicity information
Gaussian processes with monotonicity information Anonymous Author Anonymous Author Unknown Institution Unknown Institution Abstract A metho for using monotonicity information in multivariate Gaussian process
More informationComputing Exact Confidence Coefficients of Simultaneous Confidence Intervals for Multinomial Proportions and their Functions
Working Paper 2013:5 Department of Statistics Computing Exact Confience Coefficients of Simultaneous Confience Intervals for Multinomial Proportions an their Functions Shaobo Jin Working Paper 2013:5
More informationKNN Particle Filters for Dynamic Hybrid Bayesian Networks
KNN Particle Filters for Dynamic Hybri Bayesian Networs H. D. Chen an K. C. Chang Dept. of Systems Engineering an Operations Research George Mason University MS 4A6, 4400 University Dr. Fairfax, VA 22030
More informationMonte Carlo Methods with Reduced Error
Monte Carlo Methos with Reuce Error As has been shown, the probable error in Monte Carlo algorithms when no information about the smoothness of the function is use is Dξ r N = c N. It is important for
More informationTime-of-Arrival Estimation in Non-Line-Of-Sight Environments
2 Conference on Information Sciences an Systems, The Johns Hopkins University, March 2, 2 Time-of-Arrival Estimation in Non-Line-Of-Sight Environments Sinan Gezici, Hisashi Kobayashi an H. Vincent Poor
More informationThis module is part of the. Memobust Handbook. on Methodology of Modern Business Statistics
This moule is part of the Memobust Hanbook on Methoology of Moern Business Statistics 26 March 2014 Metho: Balance Sampling for Multi-Way Stratification Contents General section... 3 1. Summary... 3 2.
More information. Using a multinomial model gives us the following equation for P d. , with respect to same length term sequences.
S 63 Lecture 8 2/2/26 Lecturer Lillian Lee Scribes Peter Babinski, Davi Lin Basic Language Moeling Approach I. Special ase of LM-base Approach a. Recap of Formulas an Terms b. Fixing θ? c. About that Multinomial
More informationGeneralization of the persistent random walk to dimensions greater than 1
PHYSICAL REVIEW E VOLUME 58, NUMBER 6 DECEMBER 1998 Generalization of the persistent ranom walk to imensions greater than 1 Marián Boguñá, Josep M. Porrà, an Jaume Masoliver Departament e Física Fonamental,
More informationProof of SPNs as Mixture of Trees
A Proof of SPNs as Mixture of Trees Theorem 1. If T is an inuce SPN from a complete an ecomposable SPN S, then T is a tree that is complete an ecomposable. Proof. Argue by contraiction that T is not a
More informationExpected Value of Partial Perfect Information
Expecte Value of Partial Perfect Information Mike Giles 1, Takashi Goa 2, Howar Thom 3 Wei Fang 1, Zhenru Wang 1 1 Mathematical Institute, University of Oxfor 2 School of Engineering, University of Tokyo
More informationarxiv: v1 [math.co] 29 May 2009
arxiv:0905.4913v1 [math.co] 29 May 2009 simple Havel-Hakimi type algorithm to realize graphical egree sequences of irecte graphs Péter L. Erős an István Miklós. Rényi Institute of Mathematics, Hungarian
More informationLinear First-Order Equations
5 Linear First-Orer Equations Linear first-orer ifferential equations make up another important class of ifferential equations that commonly arise in applications an are relatively easy to solve (in theory)
More informationMath 342 Partial Differential Equations «Viktor Grigoryan
Math 342 Partial Differential Equations «Viktor Grigoryan 6 Wave equation: solution In this lecture we will solve the wave equation on the entire real line x R. This correspons to a string of infinite
More information6 General properties of an autonomous system of two first order ODE
6 General properties of an autonomous system of two first orer ODE Here we embark on stuying the autonomous system of two first orer ifferential equations of the form ẋ 1 = f 1 (, x 2 ), ẋ 2 = f 2 (, x
More informationTopic Modeling: Beyond Bag-of-Words
Hanna M. Wallach Cavenish Laboratory, University of Cambrige, Cambrige CB3 0HE, UK hmw26@cam.ac.u Abstract Some moels of textual corpora employ text generation methos involving n-gram statistics, while
More informationHow to Minimize Maximum Regret in Repeated Decision-Making
How to Minimize Maximum Regret in Repeate Decision-Making Karl H. Schlag July 3 2003 Economics Department, European University Institute, Via ella Piazzuola 43, 033 Florence, Italy, Tel: 0039-0-4689, email:
More informationTopic 7: Convergence of Random Variables
Topic 7: Convergence of Ranom Variables Course 003, 2016 Page 0 The Inference Problem So far, our starting point has been a given probability space (S, F, P). We now look at how to generate information
More informationIntroduction to the Vlasov-Poisson system
Introuction to the Vlasov-Poisson system Simone Calogero 1 The Vlasov equation Consier a particle with mass m > 0. Let x(t) R 3 enote the position of the particle at time t R an v(t) = ẋ(t) = x(t)/t its
More informationAnalyzing Tensor Power Method Dynamics in Overcomplete Regime
Journal of Machine Learning Research 18 (2017) 1-40 Submitte 9/15; Revise 11/16; Publishe 4/17 Analyzing Tensor Power Metho Dynamics in Overcomplete Regime Animashree Ananumar Department of Electrical
More informationSYNCHRONOUS SEQUENTIAL CIRCUITS
CHAPTER SYNCHRONOUS SEUENTIAL CIRCUITS Registers an counters, two very common synchronous sequential circuits, are introuce in this chapter. Register is a igital circuit for storing information. Contents
More informationIntroduction to Markov Processes
Introuction to Markov Processes Connexions moule m44014 Zzis law Gustav) Meglicki, Jr Office of the VP for Information Technology Iniana University RCS: Section-2.tex,v 1.24 2012/12/21 18:03:08 gustav
More informationNonlinear Adaptive Ship Course Tracking Control Based on Backstepping and Nussbaum Gain
Nonlinear Aaptive Ship Course Tracking Control Base on Backstepping an Nussbaum Gain Jialu Du, Chen Guo Abstract A nonlinear aaptive controller combining aaptive Backstepping algorithm with Nussbaum gain
More informationInfluence of weight initialization on multilayer perceptron performance
Influence of weight initialization on multilayer perceptron performance M. Karouia (1,2) T. Denœux (1) R. Lengellé (1) (1) Université e Compiègne U.R.A. CNRS 817 Heuiasyc BP 649 - F-66 Compiègne ceex -
More informationParameter estimation: A new approach to weighting a priori information
Parameter estimation: A new approach to weighting a priori information J.L. Mea Department of Mathematics, Boise State University, Boise, ID 83725-555 E-mail: jmea@boisestate.eu Abstract. We propose a
More informationRobust Low Rank Kernel Embeddings of Multivariate Distributions
Robust Low Rank Kernel Embeings of Multivariate Distributions Le Song, Bo Dai College of Computing, Georgia Institute of Technology lsong@cc.gatech.eu, boai@gatech.eu Abstract Kernel embeing of istributions
More informationCascaded redundancy reduction
Network: Comput. Neural Syst. 9 (1998) 73 84. Printe in the UK PII: S0954-898X(98)88342-5 Cascae reunancy reuction Virginia R e Sa an Geoffrey E Hinton Department of Computer Science, University of Toronto,
More informationAnalysis of tandem fluid queues
Analysis of tanem flui queues Małgorzata M. O Reilly School of Mathematics University of Tasmania Tas 7001, Australia malgorzata.oreilly@utas.eu.au Werner Scheinhart Department of Applie Mathematics University
More informationTIME-DELAY ESTIMATION USING FARROW-BASED FRACTIONAL-DELAY FIR FILTERS: FILTER APPROXIMATION VS. ESTIMATION ERRORS
TIME-DEAY ESTIMATION USING FARROW-BASED FRACTIONA-DEAY FIR FITERS: FITER APPROXIMATION VS. ESTIMATION ERRORS Mattias Olsson, Håkan Johansson, an Per öwenborg Div. of Electronic Systems, Dept. of Electrical
More informationSwitching Time Optimization in Discretized Hybrid Dynamical Systems
Switching Time Optimization in Discretize Hybri Dynamical Systems Kathrin Flaßkamp, To Murphey, an Sina Ober-Blöbaum Abstract Switching time optimization (STO) arises in systems that have a finite set
More informationThe Recycling Gibbs Sampler for Efficient Learning
The Recycling Gibbs Sampler for Efficient Learning Luca Martino, Victor Elvira, Gustau Camps-Valls Image Processing Laboratory, Universitat de València (Spain). Department of Signal Processing, Universidad
More informationA Course in Machine Learning
A Course in Machine Learning Hal Daumé III 12 EFFICIENT LEARNING So far, our focus has been on moels of learning an basic algorithms for those moels. We have not place much emphasis on how to learn quickly.
More informationLower Bounds for the Smoothed Number of Pareto optimal Solutions
Lower Bouns for the Smoothe Number of Pareto optimal Solutions Tobias Brunsch an Heiko Röglin Department of Computer Science, University of Bonn, Germany brunsch@cs.uni-bonn.e, heiko@roeglin.org Abstract.
More informationHyperbolic Moment Equations Using Quadrature-Based Projection Methods
Hyperbolic Moment Equations Using Quarature-Base Projection Methos J. Koellermeier an M. Torrilhon Department of Mathematics, RWTH Aachen University, Aachen, Germany Abstract. Kinetic equations like the
More informationCapacity Analysis of MIMO Systems with Unknown Channel State Information
Capacity Analysis of MIMO Systems with Unknown Channel State Information Jun Zheng an Bhaskar D. Rao Dept. of Electrical an Computer Engineering University of California at San Diego e-mail: juzheng@ucs.eu,
More informationA. Incorrect! The letter t does not appear in the expression of the given integral
AP Physics C - Problem Drill 1: The Funamental Theorem of Calculus Question No. 1 of 1 Instruction: (1) Rea the problem statement an answer choices carefully () Work the problems on paper as neee (3) Question
More informationMath Notes on differentials, the Chain Rule, gradients, directional derivative, and normal vectors
Math 18.02 Notes on ifferentials, the Chain Rule, graients, irectional erivative, an normal vectors Tangent plane an linear approximation We efine the partial erivatives of f( xy, ) as follows: f f( x+
More informationTable of Common Derivatives By David Abraham
Prouct an Quotient Rules: Table of Common Derivatives By Davi Abraham [ f ( g( ] = [ f ( ] g( + f ( [ g( ] f ( = g( [ f ( ] g( g( f ( [ g( ] Trigonometric Functions: sin( = cos( cos( = sin( tan( = sec
More informationarxiv: v2 [math.st] 29 Oct 2015
EXPONENTIAL RANDOM SIMPLICIAL COMPLEXES KONSTANTIN ZUEV, OR EISENBERG, AND DMITRI KRIOUKOV arxiv:1502.05032v2 [math.st] 29 Oct 2015 Abstract. Exponential ranom graph moels have attracte significant research
More informationd dx But have you ever seen a derivation of these results? We ll prove the first result below. cos h 1
Lecture 5 Some ifferentiation rules Trigonometric functions (Relevant section from Stewart, Seventh Eition: Section 3.3) You all know that sin = cos cos = sin. () But have you ever seen a erivation of
More informationChapter 6: Energy-Momentum Tensors
49 Chapter 6: Energy-Momentum Tensors This chapter outlines the general theory of energy an momentum conservation in terms of energy-momentum tensors, then applies these ieas to the case of Bohm's moel.
More informationSchrödinger s equation.
Physics 342 Lecture 5 Schröinger s Equation Lecture 5 Physics 342 Quantum Mechanics I Wenesay, February 3r, 2010 Toay we iscuss Schröinger s equation an show that it supports the basic interpretation of
More informationIntroduction. A Dirichlet Form approach to MCMC Optimal Scaling. MCMC idea
Introuction A Dirichlet Form approach to MCMC Optimal Scaling Markov chain Monte Carlo (MCMC quotes: Metropolis et al. (1953, running coe on the Los Alamos MANIAC: a feasible approach to statistical mechanics
More informationAPPROXIMATE SOLUTION FOR TRANSIENT HEAT TRANSFER IN STATIC TURBULENT HE II. B. Baudouy. CEA/Saclay, DSM/DAPNIA/STCM Gif-sur-Yvette Cedex, France
APPROXIMAE SOLUION FOR RANSIEN HEA RANSFER IN SAIC URBULEN HE II B. Bauouy CEA/Saclay, DSM/DAPNIA/SCM 91191 Gif-sur-Yvette Ceex, France ABSRAC Analytical solution in one imension of the heat iffusion equation
More informationA simple model for the small-strain behaviour of soils
A simple moel for the small-strain behaviour of soils José Jorge Naer Department of Structural an Geotechnical ngineering, Polytechnic School, University of São Paulo 05508-900, São Paulo, Brazil, e-mail:
More informationA. Exclusive KL View of the MLE
A. Exclusive KL View of the MLE Lets assume a change-of-variable moel p Z z on the ranom variable Z R m, such as the one use in Dinh et al. 2017: z 0 p 0 z 0 an z = ψz 0, where ψ is an invertible function
More informationTHE VAN KAMPEN EXPANSION FOR LINKED DUFFING LINEAR OSCILLATORS EXCITED BY COLORED NOISE
Journal of Soun an Vibration (1996) 191(3), 397 414 THE VAN KAMPEN EXPANSION FOR LINKED DUFFING LINEAR OSCILLATORS EXCITED BY COLORED NOISE E. M. WEINSTEIN Galaxy Scientific Corporation, 2500 English Creek
More informationImproving Estimation Accuracy in Nonrandomized Response Questioning Methods by Multiple Answers
International Journal of Statistics an Probability; Vol 6, No 5; September 207 ISSN 927-7032 E-ISSN 927-7040 Publishe by Canaian Center of Science an Eucation Improving Estimation Accuracy in Nonranomize
More informationRobust Forward Algorithms via PAC-Bayes and Laplace Distributions. ω Q. Pr (y(ω x) < 0) = Pr A k
A Proof of Lemma 2 B Proof of Lemma 3 Proof: Since the support of LL istributions is R, two such istributions are equivalent absolutely continuous with respect to each other an the ivergence is well-efine
More informationThermal conductivity of graded composites: Numerical simulations and an effective medium approximation
JOURNAL OF MATERIALS SCIENCE 34 (999)5497 5503 Thermal conuctivity of grae composites: Numerical simulations an an effective meium approximation P. M. HUI Department of Physics, The Chinese University
More informationThe derivative of a function f(x) is another function, defined in terms of a limiting expression: f(x + δx) f(x)
Y. D. Chong (2016) MH2801: Complex Methos for the Sciences 1. Derivatives The erivative of a function f(x) is another function, efine in terms of a limiting expression: f (x) f (x) lim x δx 0 f(x + δx)
More informationBayesian analysis of massive datasets via particle filters
Bayesian analysis of massive atasets via particle filters Greg Rigeway RAND PO Box 2138 Santa Monica, CA 90407-2138 gregr@ran.org Davi Maigan Department of Statistics 477 Hill Center Rutgers University
More informationSemiclassical analysis of long-wavelength multiphoton processes: The Rydberg atom
PHYSICAL REVIEW A 69, 063409 (2004) Semiclassical analysis of long-wavelength multiphoton processes: The Ryberg atom Luz V. Vela-Arevalo* an Ronal F. Fox Center for Nonlinear Sciences an School of Physics,
More informationRelative Entropy and Score Function: New Information Estimation Relationships through Arbitrary Additive Perturbation
Relative Entropy an Score Function: New Information Estimation Relationships through Arbitrary Aitive Perturbation Dongning Guo Department of Electrical Engineering & Computer Science Northwestern University
More informationA NONLINEAR SOURCE SEPARATION APPROACH FOR THE NICOLSKY-EISENMAN MODEL
6th European Signal Processing Conference EUSIPCO 28, Lausanne, Switzerlan, August 25-29, 28, copyright by EURASIP A NONLINEAR SOURCE SEPARATION APPROACH FOR THE NICOLSKY-EISENMAN MODEL Leonaro Tomazeli
More informationLearning the hyper-parameters. Luca Martino
Learning the hyper-parameters Luca Martino 2017 2017 1 / 28 Parameters and hyper-parameters 1. All the described methods depend on some choice of hyper-parameters... 2. For instance, do you recall λ (bandwidth
More informationEquilibrium in Queues Under Unknown Service Times and Service Value
University of Pennsylvania ScholarlyCommons Finance Papers Wharton Faculty Research 1-2014 Equilibrium in Queues Uner Unknown Service Times an Service Value Laurens Debo Senthil K. Veeraraghavan University
More informationu!i = a T u = 0. Then S satisfies
Deterministic Conitions for Subspace Ientifiability from Incomplete Sampling Daniel L Pimentel-Alarcón, Nigel Boston, Robert D Nowak University of Wisconsin-Maison Abstract Consier an r-imensional subspace
More informationMODELLING DEPENDENCE IN INSURANCE CLAIMS PROCESSES WITH LÉVY COPULAS ABSTRACT KEYWORDS
MODELLING DEPENDENCE IN INSURANCE CLAIMS PROCESSES WITH LÉVY COPULAS BY BENJAMIN AVANZI, LUKE C. CASSAR AND BERNARD WONG ABSTRACT In this paper we investigate the potential of Lévy copulas as a tool for
More informationA Path Planning Method Using Cubic Spiral with Curvature Constraint
A Path Planning Metho Using Cubic Spiral with Curvature Constraint Tzu-Chen Liang an Jing-Sin Liu Institute of Information Science 0, Acaemia Sinica, Nankang, Taipei 5, Taiwan, R.O.C., Email: hartree@iis.sinica.eu.tw
More informationModeling the effects of polydispersity on the viscosity of noncolloidal hard sphere suspensions. Paul M. Mwasame, Norman J. Wagner, Antony N.
Submitte to the Journal of Rheology Moeling the effects of polyispersity on the viscosity of noncolloial har sphere suspensions Paul M. Mwasame, Norman J. Wagner, Antony N. Beris a) epartment of Chemical
More informationarxiv:cond-mat/ v1 31 Jan 2003
Nosé-Hoover sampling of quantum entangle istribution functions D. Mentrup an J. Schnack 1 arxiv:con-mat/030165 v1 31 Jan 003 Abstract Universität Osnabrück, Fachbereich Physik Barbarastr. 7, D-49069 Osnabrück
More informationThe Exact Form and General Integrating Factors
7 The Exact Form an General Integrating Factors In the previous chapters, we ve seen how separable an linear ifferential equations can be solve using methos for converting them to forms that can be easily
More informationA Modification of the Jarque-Bera Test. for Normality
Int. J. Contemp. Math. Sciences, Vol. 8, 01, no. 17, 84-85 HIKARI Lt, www.m-hikari.com http://x.oi.org/10.1988/ijcms.01.9106 A Moification of the Jarque-Bera Test for Normality Moawa El-Fallah Ab El-Salam
More informationHarmonic Modelling of Thyristor Bridges using a Simplified Time Domain Method
1 Harmonic Moelling of Thyristor Briges using a Simplifie Time Domain Metho P. W. Lehn, Senior Member IEEE, an G. Ebner Abstract The paper presents time omain methos for harmonic analysis of a 6-pulse
More informationLecture Introduction. 2 Examples of Measure Concentration. 3 The Johnson-Lindenstrauss Lemma. CS-621 Theory Gems November 28, 2012
CS-6 Theory Gems November 8, 0 Lecture Lecturer: Alesaner Mąry Scribes: Alhussein Fawzi, Dorina Thanou Introuction Toay, we will briefly iscuss an important technique in probability theory measure concentration
More informationarxiv: v1 [hep-lat] 19 Nov 2013
HU-EP-13/69 SFB/CPP-13-98 DESY 13-225 Applicability of Quasi-Monte Carlo for lattice systems arxiv:1311.4726v1 [hep-lat] 19 ov 2013, a,b Tobias Hartung, c Karl Jansen, b Hernan Leovey, Anreas Griewank
More informationLayered Adaptive Importance Sampling
Noname manuscript No (will be inserted by the editor) Layered Adaptive Importance Sampling L Martino V Elvira D Luengo J Corander Received: date / Accepted: date Abstract Monte Carlo methods represent
More informationLectures - Week 10 Introduction to Ordinary Differential Equations (ODES) First Order Linear ODEs
Lectures - Week 10 Introuction to Orinary Differential Equations (ODES) First Orer Linear ODEs When stuying ODEs we are consiering functions of one inepenent variable, e.g., f(x), where x is the inepenent
More informationState observers and recursive filters in classical feedback control theory
State observers an recursive filters in classical feeback control theory State-feeback control example: secon-orer system Consier the riven secon-orer system q q q u x q x q x x x x Here u coul represent
More informationAn introduction to Sequential Monte Carlo
An introduction to Sequential Monte Carlo Thang Bui Jes Frellsen Department of Engineering University of Cambridge Research and Communication Club 6 February 2014 1 Sequential Monte Carlo (SMC) methods
More informationNon-Linear Bayesian CBRN Source Term Estimation
Non-Linear Bayesian CBRN Source Term Estimation Peter Robins Hazar Assessment, Simulation an Preiction Group Dstl Porton Down, UK. probins@stl.gov.uk Paul Thomas Hazar Assessment, Simulation an Preiction
More informationDelay Limited Capacity of Ad hoc Networks: Asymptotically Optimal Transmission and Relaying Strategy
Delay Limite Capacity of A hoc Networks: Asymptotically Optimal Transmission an Relaying Strategy Eugene Perevalov Lehigh University Bethlehem, PA 85 Email: eup2@lehigh.eu Rick Blum Lehigh University Bethlehem,
More informationInter-domain Gaussian Processes for Sparse Inference using Inducing Features
Inter-omain Gaussian Processes for Sparse Inference using Inucing Features Miguel Lázaro-Greilla an Aníbal R. Figueiras-Vial Dep. Signal Processing & Communications Universia Carlos III e Mari, SPAIN {miguel,arfv}@tsc.uc3m.es
More informationTrack Initialization from Incomplete Measurements
Track Initialiation from Incomplete Measurements Christian R. Berger, Martina Daun an Wolfgang Koch Department of Electrical an Computer Engineering, University of Connecticut, Storrs, Connecticut 6269,
More informationLecture 2 Lagrangian formulation of classical mechanics Mechanics
Lecture Lagrangian formulation of classical mechanics 70.00 Mechanics Principle of stationary action MATH-GA To specify a motion uniquely in classical mechanics, it suffices to give, at some time t 0,
More informationDot trajectories in the superposition of random screens: analysis and synthesis
1472 J. Opt. Soc. Am. A/ Vol. 21, No. 8/ August 2004 Isaac Amiror Dot trajectories in the superposition of ranom screens: analysis an synthesis Isaac Amiror Laboratoire e Systèmes Périphériques, Ecole
More informationMEASURES WITH ZEROS IN THE INVERSE OF THEIR MOMENT MATRIX
MEASURES WITH ZEROS IN THE INVERSE OF THEIR MOMENT MATRIX J. WILLIAM HELTON, JEAN B. LASSERRE, AND MIHAI PUTINAR Abstract. We investigate an iscuss when the inverse of a multivariate truncate moment matrix
More informationThe Press-Schechter mass function
The Press-Schechter mass function To state the obvious: It is important to relate our theories to what we can observe. We have looke at linear perturbation theory, an we have consiere a simple moel for
More informationEntanglement is not very useful for estimating multiple phases
PHYSICAL REVIEW A 70, 032310 (2004) Entanglement is not very useful for estimating multiple phases Manuel A. Ballester* Department of Mathematics, University of Utrecht, Box 80010, 3508 TA Utrecht, The
More information26.1 Metropolis method
CS880: Approximations Algorithms Scribe: Dave Anrzejewski Lecturer: Shuchi Chawla Topic: Metropolis metho, volume estimation Date: 4/26/07 The previous lecture iscusse they some of the key concepts of
More informationError Floors in LDPC Codes: Fast Simulation, Bounds and Hardware Emulation
Error Floors in LDPC Coes: Fast Simulation, Bouns an Harware Emulation Pamela Lee, Lara Dolecek, Zhengya Zhang, Venkat Anantharam, Borivoje Nikolic, an Martin J. Wainwright EECS Department University of
More informationMath 1B, lecture 8: Integration by parts
Math B, lecture 8: Integration by parts Nathan Pflueger 23 September 2 Introuction Integration by parts, similarly to integration by substitution, reverses a well-known technique of ifferentiation an explores
More informationBalancing Expected and Worst-Case Utility in Contracting Models with Asymmetric Information and Pooling
Balancing Expecte an Worst-Case Utility in Contracting Moels with Asymmetric Information an Pooling R.B.O. erkkamp & W. van en Heuvel & A.P.M. Wagelmans Econometric Institute Report EI2018-01 9th January
More informationHybrid Fusion for Biometrics: Combining Score-level and Decision-level Fusion
Hybri Fusion for Biometrics: Combining Score-level an Decision-level Fusion Qian Tao Raymon Velhuis Signals an Systems Group, University of Twente Postbus 217, 7500AE Enschee, the Netherlans {q.tao,r.n.j.velhuis}@ewi.utwente.nl
More informationLogarithmic spurious regressions
Logarithmic spurious regressions Robert M. e Jong Michigan State University February 5, 22 Abstract Spurious regressions, i.e. regressions in which an integrate process is regresse on another integrate
More informationLeast-Squares Regression on Sparse Spaces
Least-Squares Regression on Sparse Spaces Yuri Grinberg, Mahi Milani Far, Joelle Pineau School of Computer Science McGill University Montreal, Canaa {ygrinb,mmilan1,jpineau}@cs.mcgill.ca 1 Introuction
More informationLecture 10 Notes, Electromagnetic Theory II Dr. Christopher S. Baird, faculty.uml.edu/cbaird University of Massachusetts Lowell
Lecture 10 Notes, Electromagnetic Theory II Dr. Christopher S. Bair, faculty.uml.eu/cbair University of Massachusetts Lowell 1. Pre-Einstein Relativity - Einstein i not invent the concept of relativity,
More informationarxiv: v5 [cs.lg] 28 Mar 2017
Equilibrium Propagation: Briging the Gap Between Energy-Base Moels an Backpropagation Benjamin Scellier an Yoshua Bengio * Université e Montréal, Montreal Institute for Learning Algorithms March 3, 217
More informationTransmission Line Matrix (TLM) network analogues of reversible trapping processes Part B: scaling and consistency
Transmission Line Matrix (TLM network analogues of reversible trapping processes Part B: scaling an consistency Donar e Cogan * ANC Eucation, 308-310.A. De Mel Mawatha, Colombo 3, Sri Lanka * onarecogan@gmail.com
More informationA Novel Decoupled Iterative Method for Deep-Submicron MOSFET RF Circuit Simulation
A Novel ecouple Iterative Metho for eep-submicron MOSFET RF Circuit Simulation CHUAN-SHENG WANG an YIMING LI epartment of Mathematics, National Tsing Hua University, National Nano evice Laboratories, an
More informationNumber of wireless sensors needed to detect a wildfire
Number of wireless sensors neee to etect a wilfire Pablo I. Fierens Instituto Tecnológico e Buenos Aires (ITBA) Physics an Mathematics Department Av. Maero 399, Buenos Aires, (C1106ACD) Argentina pfierens@itba.eu.ar
More informationPolynomial Inclusion Functions
Polynomial Inclusion Functions E. e Weert, E. van Kampen, Q. P. Chu, an J. A. Muler Delft University of Technology, Faculty of Aerospace Engineering, Control an Simulation Division E.eWeert@TUDelft.nl
More information