Nonlinear Filtering: Interacting Particle Resolution

Size: px
Start display at page:

Download "Nonlinear Filtering: Interacting Particle Resolution"

Transcription

1 onlinear Filtering: Interacting Particle Resolution P. Del Moral Laboratoire de Statistiques et Probabilités, CRS UMR C55830, Bat 1R1, Université Paul Sabatier, 118 Route de arbonne, Toulouse Cedex, France Markov Processes and Related Fields, Volume 2 umber 4, Received April 2, 1996, revised October 3, 1996 Abstract. This paper covers stochastic particle methods for the numerical solution of the nonlinear filtering equations based on the simulation of interacting particle systems. The main contribution of this paper is to prove convergence of such approximations to the optimal filter, thus yielding what seemed to be the first convergence results for such approximations of the nonlinear filtering equations. This new treatment has been influenced primarily by the development of genetic algorithms J. H. Holland [11], R. Cerf [2] and secondarily by the papers of H. Kunita and L. Stettner [12, 13]. Such interacting particle resolutions encompass genetic algorithms. Incidentally, our models provide essential insight for the analysis of genetic algorithms with a non-homogeneous fitness function with respect to time. Keywords: nonlinear filtering, Bayesian estimation, Monte-Carlo sampling, interacting particle systems, sampling-resampling, evolutionary processes, genetic algorithms AMS Subject Classification: 60F10, 60G35, 60J10, 62D05, 62E25, 62F12, 62G05, 62L20, 62M05, 92D15, 93E10, 93E11 1. Introduction The basic model for the general nonlinear filtering problem consists of a nonlinear plant X with state noise W and nonlinear observation Y with observation noise V. Let X, Y be the Markov process taking values in S R and defined by the system: { X ν, K FX Y 1.1 Y n = h X n + V n, n 1, where S = R d, d 1, h : S R and V n are independent random variables having a density g n with respect to the Lebesgue measure. The signal process X

2 2 P. Del Moral that we consider, is assumed to be a temporally homogeneous Markov process on S with transition probability kernel K and initial probability measure ν. We assume that the observation noise V and the state plant X are independent. For simplicity, the observation process Y is taken real valued, the extension to vector observation processes is straightforward. One can study filtering problems in more general settings. We choose not to do so here, and prefer to focus on the central ideas. The methods used in this paper can be extended, if desired. The filtering problem is concerned with estimating a functional of the state process using the information contained in the observation process Y. The information is encoded in the filtration defined by the sigma algebra generated by the observations Y 1,..., Y n. Let f be an integrable Borel test function from S into R. The best estimate of X n given the observations up to time n, is the conditional expectation π n f def = EfX n Y n, Y n def = Y 1,..., Y n. With the notable exception of the linear-gaussian situation, general optimal filters have no finitely recursive solution [3]. This paper covers stochastic particle methods for the numerical solution of the nonlinear filtering equations based on the simulation of interacting particle systems. Such algorithms are an extension of the Sampling/Resampling S/R principles introduced by Gordon, Salmon and Smith in [10] and independently by Del Moral, oyer, Rigal and Salut in [4] and [5]. Several examples of practical problems that can be solved using these methods, are given in [1] and [8], including problems in Radar/Sonar signal processing and GPS/IS integration. Such particle nonlinear filters will differ from the others [6, 7] in the way they store and update the information that is accumulated through the resampling of the positions. We start by giving some general notation and we recall some basic facts related to the theory introduced by Kunita and Stettner [12, 13]. In Section 3 we introduce the interacting particle approximation and we design a natural stochastic basis for the convergence study. In Section 4 we describe recursive formulas for the conditional distributions and the context that we are interested in. We propose a natural framework which allows to explicitly formulate mean error bounds in terms of the likelihood functions related to the resampling/selection procedure. The hardest point of our program is contained in Section 5: the convergence study of our algorithm requires a specific development because of the difficulty to compute mean error estimates which are essential to perform convergence rates. The key idea is to introduce in the mean square error estimates a martingale with unit mean, using the functions g n. This martingale approach simplifies drastically! The evaluation of the convergence rates is discussed in Section 6. This last section contains our main result: we prove that the interacting particle filters converge to the conditional distribution, as the number of particles tends to infinity. The convergence rate estimates arise quite naturally from the results and associated methodologies of Sections 5 and 6.

3 onlinear filtering: interacting particle resolution 3 2. on linear filtering equation 2.1. General notations Before starting the description of the nonlinear filtering equation, let us first introduce some general notation. Let CS be the space of bounded continuous functions on S with norm f = sup x S fx. Let PS be the space of all probability measures on S in which the weak topology is induced: lim µ n = µ in PS for all f CS lim f µ n = f µ. n + n + Let µ PS, f CS and, let K 1 and K 2 be two Markov kernels. We will use the standard notation µk 1 dy = µdx K 1 x, dy K 1 K 2 x, dz = K 1 x, dy K 2 y, dz K 1 fx = K 1 x, dy fy µf = µdx fx. With m PS 2 we associate two measures m and m PS as follows, for all f CS : mf = mdx 1, dx 2 fx 2, mf = mdx 1, dx 2 fx 1. With a Markov kernel K and a measure µ PS we associate a measure µ K PS 2 by setting for all f CS 2 : µ Kf = µdx 1 Kx 1, dx 2 fx 1, x 2. Finally we denote by CPS the space of bounded continuous functions F : PS R Recursive filters and Bayes formula In this section we describe recursive expressions for the conditional distribution of X n and X n, X n+1 given the observations Y n = Y 1,..., Y n. Let Ω = Ω 1 Ω 2, F n, P be the canonical space for the signal observation pair X, Y. Therefore P is the probability measure on Ω corresponding to the filtering model FX Y, when ν is the probability measure of X 0 ; the marginal of P on Ω 1 is the law of X; V n = Y n hx n is a sequence of independent random variables with densities g n.

4 4 P. Del Moral We use E to denote expectation with respect to P on Ω. To clarify the presentation, we will also write m n+1 = π n K. 2.1 Using Bayes Theorem, we see that the conditional distribution of X n given the observations up to time n is given by π n = ρ n π n 1, Y n n 1 π 0 = ν, 2.2 where ρ n µ, yf = fx gn y hx µ Kdx gn y hx µ Kdx 2.3 for all f CS, µ PS, y R and n 1. Much more is true. Following Kunita and Stettner [12, 13], the above description enables us to consider the conditional distributions π n as a σy n, P-Markov process with infinite dimensional state space PS and transition probability kernel Π n defined by Π n F µ = F ρ n µ, y g n y hz µ Kdz dy for any bounded continuous function F : PS R and µ PS. In other words, with some obvious abuse of notation dpy n, x n, x n 1 π n 1 = g n y n hx n dy n π n 1 dx n 1 Kx n 1, dx n 2.4 py n /π n 1 = g n y n hx n π n 1 Kdx n 2.5 ow, the construction of the recursive expression for the conditional distribution of X n, X n+1 given the observations Y n, is a fairly immediate consequence of 2.1 and 2.2. Using the above notation one easily gets m n+1 = Φ n m n, Y n n 1 m 0 = ν K, 2.6 where Φ n m, yf = fx1, x 2 g n y hx 1 mdx 0, dx 1 Kx 1, dx 2 gn y hz 1 mdz 0, dz for all f CS 2, µ PS and y R. 3. Interacting particle systems 3.1. Description of the algorithm The particle system under study will be a Markov chain with state space S, where 1 is the size of the system. The -tuple of elements of S, i.e. the

5 onlinear filtering: interacting particle resolution 5 points of the set S, are called systems of particles and they will be mostly denoted by the letters x, y, z. Given the observations Y = y, we denote by x n, x n+1 n 0 the PS 2 -Markov process defined by the transition probabilities P [y] { x0, x 1 dz 0, z 1 } = p=1 m 0 dz p 0, dzp 1 P [y] { xn, x n+1 dz 0, z 1 x n 1, x n = x 0, x 1 } = p=1 Φ n 1 δ x i 0,x i 1, y n dz p 0, zp By the very definition of m 0 and Φ n we also have the following. Initial Particle System P [y] { x 0 = dx} = p=1 νdx p, Sampling/Exploration P [y] {x n = dx x n 1 = z} = p=1 Kz p, dx p, Resampling/Selection P [y] { x n = dx x n = z} = p=1 g n y n hz i j=1 g ny n hz j δ z idxp. Finally one gets a sequence of particle systems x n 1 = x 1 n 1,..., x n 1 x n = x 1 n,..., x n x n = x 1 n,..., x n. It is essential to remark that the particles x i n are chosen randomly and independently from the population {x 1 n,..., x n } by the law πn defined by the likelihood functions and the present measurement Y n. amely π n = g n y n hx i n j=1 g ny n hx j n δ x i n 3.2 Then it moves to x i n+1 using the transition probability kernel K. In other words, the S -valued Markov chain x n = x 1 n,..., x n is obtained through overlapping another chain x n = x 1 n,..., x n, representing the successive particles obtained

6 6 P. Del Moral by exploring the probability space with the transitions K. More precisely, the motion of particles is decomposed into two stages: x n Sampling/Exploration Resampling/Selection x n x n The algorithm constructed in this way, will be called an interacting particle filter. The terminology interacting is intended to emphasize, that the particles are not independent and it differs from the particle resolutions introduced in [7] and [6]. Remark 3.1. The formulated algorithm permits generalisation in the case that the selecting procedure is used from time to time. In practical situations, a simple way to do this consists of introducing a resampling schedule. For instance, we may choose to resample the particles, when fifty percent of the weights is lower than A p for a convenient choice of A > 0 and p 2. To point out the connection with genetic algorithms and to so emphasize the role of the likelihood functions g n, assume further that 1. the state space S is finite; 2. the Markov transition kernels K l x, z are governed by a parameter l with K l x, z 1 x z, as l tends to infinity; 3. a noise observation V n, also governed by a parameter l, with distribution dp V n v = exp V v log l dv ; exp V u log l du 4. homogeneous series of observations with respect to time for all n 1 Y n = y V y hx def = fx. The corresponding Exploration and Selection mechanisms are governed by a parameter l and they take the following form: P r { x l n = dx x l n = z } = l P r { x n = x l x n 1 = z} = p=1 p=1 K l z p, x p l fzi j=1 l fzj 1 z ixp. For this very special situation, Cerf [2] gives several conditions on the rate of decrease of the perturbations 1/ log ln, to ensure that all particles x ln,i n

7 onlinear filtering: interacting particle resolution 7 visit the set of global maxima of the fitness function f in finite time, when the number of particles is greater than a critical value. Unfortunately there is a critical lack of theoretical results on the convergence of such algorithms for numerically solving the nonlinear filtering equations. The crucial question is of course, whether the empirical random measure 1 δ x i n converges to the conditional distribution π n, when the size of the particle system is growing. This is positively answered in Theorem 6.2 Section 6. We will show that for every bounded Borel test function f : S R and for n 1 lim E 1 + f x i n π n f = We give a new and detailed analysis of this problem. These results are largely recent, although the question of local convergence occurs in [5]. Unlike genetic algorithms it should be emphasized that we are not necessarily trying to exactly recover the unknown state variables. The conditional distribution gives the conditional minimum variance estimate, but the error does not in general converge to zero as time tends to infinity see Kunita [12]. Such particle nonlinear filters differ from those introduced in [7] and [6]. It clearly encompasses the genetic algorithms introduced by Holland [11] and recently developed by R. Cerf [2]. An advantage of this procedure is, that it simultaneously explores the probability space using the a priori Markov kernel and updates the information that is accumulated through resampling of the positions. In introducing such adaptation/selection laws, the particle system acquires certain configurations, so as to represent an estimate of the conditional distribution. They provide a natural procedure for a system of particles to sense its environment through their likelihood functions and the observations. It is clear that such particle resolutions could be formulated in an natural way in other contexts as neural networks, model parameter identification, optimal control [9] The associated Markov process ext we shall proceed to model such interacting particle procedure and the nonlinear filtering problem on a natural stochastic basis. In the preceding paragraph we have remarked on the fact that the particles coincide with the support of random measures that estimate the conditional distribution. A convenient tool for analysing the modelling of such interacting particle systems is the split-

8 8 P. Del Moral ting transition kernel PS 2 C { 1 δ x i : x i S 2} PS 2, 3.5 defined for every F CPS 2, η PS 2 and 1, by C F η = F m C η, dm def = S 2 1 F δ x i ηdx 1... ηdx. 3.6 The interpretation of C is the following: by starting with a measure η PS 2, the next measure is the result of sampling independent random variables x i with common law η η C m = 1 δ x i. Given a series of observations Y = y, we observe that the random empirical measures m n+1 = 1 δ x i n,xi n+1 are the result of sampling independent random variables with common law Φ n m n, y n = π n K = g n y n hx i n j=1 g ny n hx j n δ x i n K. The above observation enables us to consider the empirical measures m n canonical PS 2 -valued Markov process Ω, β n, P [y], by setting as a P [y] {m 1 dη} = C Φ n m 0, y n, dη P [y] {m n+1 dη m n = µ} = C Φ n µ, y n, dη. 3.7 ow we design a stochastic basis for the convergence of our particle approximations. To capture all randomness we list all outcomes into the canonical space defined as follows. 1. Recall Ω, F n, P is the canonical space for the signal observation pair X, Y. 2. We define Ω = Ω Ω and F n = β n F n and, for every ω def = ω 1, ω 2, ω 3 Ω, we define m n ω = ω 1 n, X n ω = ω 2 n, Y n ω = ω 3 n.

9 onlinear filtering: interacting particle resolution 9 3. For every A β n and B F n, we define P as follows: P{A B} def = P [Y ω] A dpω. 3.8 B As usual we use Ẽ to denote the expectation with respect to P, and Ẽ[y] to denote the expectation with respect to P [y]. With some obvious abuse of notation we have d pm 1,..., m n, y 1,..., y n n = C Φk 1 m k 1, y k 1, dm k dp y1,..., y n, k=1 with the convention Φ 0 m 0, y 0 = ν K. 4. General recursive formulas In this section we shall adopt an unconventional model for the conditional distributions. The choice is dictated by our desire to have very simple relationships between the likelihood functions and the conditional distributions. Initially, this will require a quite different setup from the one used in Sections 2 and 3, but in the end of our investigations will resemble more and more the models presented in Section 2. The setting is the same as in Section 3.1. In particular, unless otherwise stated, we assume the observation data to be a fixed series of real numbers Y = y. This assumption enables us to consider the conditional distribution π n as a probability parametrised by the observation parameters y 1,..., y n,.... Therefore, when the context is unambiguous, we will often write for brevity g n x instead of g n y n hx n. Using this notation, an alternative notation for 2.2 is the recursive formula π n f = π n 1K f g n π n 1 Kg n, 4.1. Likelihood functions analysis The proof of the convergence 3.4 involves for all f CS. 1. the Maximum Log-likelihood functions given by pyn,..., y p+1 x p+1 2 V n/p y = sup log dpxp+1 x p S py n,..., y p+1 y p x p, 4.1 S p + 1 n n V n y = V p/p 1 y, 4.2 p=1

10 10 P. Del Moral where py n,..., y p+1 y p denotes the density under P of the distribution of Y n,..., Y p+1 conditionally on Y p = Y 1,..., Y p, py n,..., y p+1 x p+1 denotes the density under P of the distribution of Y n,..., Y p+1 conditionally on X p+1, dpx p+1 x p = Kx p, dx p+1 ; 2. the conditional expectations for 0 p n 1, given by f p n x p = E fx n X p, Y p+1,..., Y n x p, y p+1,..., y n. 4.3 The functions V n/p represent Log-likelihood functions on the observation process Y and they are closely related to the resampling/selection mechanism of our algorithm. This relationship is due to the fact, that the selection mechanism is formulated in terms of the fitness functions and of py n x n = g n y n hx n 4.4 py n,..., y p x p = py n,..., y p+1 x p+1 dpx p+1 x p. 4.5 S Example Assume here that state and observation processes X, Y are given by the linear dynamics X n = A X n 1 + W n 4.6 Y n = C X n + V n, 4.7 where X n R, Y n R, Y 0 = 0, A and C are real numbers and X 0, W n and V n are normally distributed with means 0 and respective non-negative covariances Q 0, Q and R. The conditional densities of Y n given X n and Y n 1 = Y 1,..., Y n 1, are given by 1 py n x n = g n y n C x n = exp 1 2π R 2 y n C x n 2 R 1 py n y n 1 = 1 2π C2 P n/n 1 + R exp 1 2 y n C A X n 1 2 C 2 P n/n 1 + R 1, with the well known measurement update equations X n K n P n/n 1 P n 1/n 1 def = EX n Y n = A X n 1 + K n Y n C A X n 1 def = C P n/n 1 C 2 P n/n 1 + R 1 def = E X n EX n /Y n 1 2 = A 2 P n 1/n 1 + Q def = E X n 1 EX n 1 /Y n 1 2 = P 1 n 1/n 2 + CR 1 C 1.

11 onlinear filtering: interacting particle resolution 11 In this very special situation [ C 2 P n/n 1 + R V n/n 1 Y log exp Y n CA R X n 1 2 C 2 P n/n 1 + R 1]. Further manipulations yield [ n V n/p Y log k=p C 2 P k/k 1 + R C 2 S k p + R with S 0 = 0 and, for A 1, exp Y k CA X k 1 2 C 2 P k/k 1 + R 1], E X k E X k /X p 2 = S k p = 1 Ak p 1 A Q. 2. Let X be a discrete time Markov process belonging to a finite discrete set S and assume that the observation noise V n is a sequence of independent, random variables with dp V n v = exp U n exp Un v dv U n : S R +. In this hidden Markov model, V n/n 1 Y 2 sup x S U n Y n hx. The functions 4.1, 4.3 and 4.5 will be used in our analysis of the convergence 3.4. This analysis requires a specific development, because of the difficulty of estimating mean errors. For instance we have Ẽ [y] 1 2 f x i n = Ẽ[y] m n+1 f 2 = 1 Ẽ[y]π n f Ẽ[y] π n f 2 = 1 Ẽ[y] π n f π n f 2 + Ẽ[y] π n f 2. Thus, Chebyshev s inequality and convenient estimates of Ẽ [y] π n f 2 = Ẽ[y] g n y n hx i n 2 j=1 g ny n hx j n fxi n 4.8 would give a complete answer concerning convergence in L 0 P [y]. Unfortunately, it is difficult to estimate 4.8 or even Ẽ[y] π n f. Accordingly, in order to point out the connections between 4.3 and 4.5 and to so emphasize the role of the Log-likelihood functions V n/p, we first introduce a natural framework in which the relationship between 4.3 and 4.5 is made explicit. The recursive expressions described in the next section, will be used repeatedly for the convergence 3.4 in the last part of the paper.

12 12 P. Del Moral 4.2. Recursive formulas The main purpose of this section is to introduce recursive expressions for 4.3 and 4.5. The technical approach presented here, is to work with the same given sequence of observations Y = y. The conditional distributions π n and the conditional expectations f n p 1 p n will be formulated as a probability and a sequence of functions parametrised by the observation parameters y. The recursive formulas described in this section, will show how the sequence of observations scales the updates of both π n and f n p. Our constructions will be explicit and the recursions will have a simple form. First, let us give some details on the use of Bayes formula in our setting. With some obvious abuse of notation, we have for all 1 p n the recursion py n,..., y p x p = py p x p py n,..., y p+1 x p = py p x p py n,..., y p+1 x p+1 dpx p+1 x p py n,..., y p x p py n,..., y p y p 1 = py p x p py p y p 1 pyn,..., y p+1 x p+1 py p+1 y p dpx p+1 x p py n,..., y 1 = py n y n 1 py n 1 y n 2 py 2 y 1 py 1. To clarify the presentation, we introduce for all 0 p n the following definitions 1. g p n x p = py n,..., y p x p 2. g n/p 1 x p = py n,..., y p x p /py n,..., y p y p 1. With a slight abuse of notation we will often write g n x n instead of g n n x n = py n x n. It is now easily checked from the above remarks, that g n/p 1 x p = g n p x p π p dz p 1 Kz p 1, dz p g n p z p 4.9 g n p x p = g p x p Kx p, dx p+1 g n p+1 x p g n/p 1 x p = g p/p 1 x p Kx p, dx p+1 g n/p x p Summarising, we have the backward recursive formulas. Proposition 4.1. For every 0 p < n 1 g p n = g p K g p+1 n g n/p 1 = g n p π p 1 Kg n p g n/p 1 = g p/p 1 Kg n/p. 4.12

13 onlinear filtering: interacting particle resolution 13 We will adopt the conventions π 1 K = ν and g 0 0 = 1. Moreover, by the very definition of π n and π n, one gets π n f = π n f = For brevity we will write fz gn y n hz m n dz gn y n hz m n dz g n y n hx i n j=1 g ny n hx j n fxi n π n f = m n g n f m n g n, π n f = m n g n f m n g n. Continuing in the same vein, we derive the conditional expectations f p n, introduced in 4.3. Using Bayes rule we have for all 1 p n the following basic equation dpx n, x p x p 1, y p,..., y n = dpx n x p, y p+1,..., y n dpx p x p 1, y p,..., y n = dpx n x p, y p+1,..., y n py n,..., y p x p py n,..., y p x p 1 dpx p x p 1. By the same line of arguments, for all 1 p n, we get dpx n, x p, x p 1 y n = dpx n x p, y p+1,..., y n dpx p, x p 1 y n = dpx n x p, y p+1,..., y n py n,..., y p x p py n,..., y p y p 1 dpx p x p 1 dpx p 1 y p 1. Thus we arrive at πp 1 dx p 1 Kx p 1, dx p f n p x p g n p x p π n f =, 4.14 πp 1 dx p 1 Kx p 1, dx p g n p x p for all f CS, 1 p n. Additionally we have f n p 1 Kxp 1, dx p f n p x p g n p x p x p 1 =, 4.15 Kxp 1, dx p g n p x p for all f CS, 1 p n. Summarising, our conditional expectations f p n can be described as follows.

14 14 P. Del Moral Proposition 4.2. For every f CS, the conditional expectations f n p 1 p n satisfy the recursive formula. For every f CS and 1 p n f p 1 n def = p p Kf n g n/p 1 Kf n g p = Kg n/p 1 Kg n p n, for all 1 p n Moreover, for every f CS and 1 p n π n f = π p 1 K f n p g n p π p 1 K g n p = π p 1 K f n p g n/p 1 π p 1 K g n/p 1 We will adopt the convention f 1 n 5. Mean square estimates def = νf 0 n g 0 n /νg 0 n. In this section we analyse the structure of the Log-likelihood functions V n/p, whilst pointing out explicit bounds. Our next objective is to estimate the convergence rate and mean errors. We shall do this now, beginning with some lemmas that will be used repeatedly in this section. Lemma 5.1. Let f : S R be any integrable Borel test function, and let n 0 and 1. We have P-a.e. m n+1 f β n = π n f m n+1 f β n = π n Kf 5.1 Proof. m n+1 f 2 1 β n = π n f π n f m n+1 f 2 1 β n = π n Kf π n Kf It suffices to note that m n+1 f β n m n+1 f 2 /β n = = mf C πn K, dm = πn f mf 2 C πn K, dm m n+1 f β n m n+1 f 2 β n = 1 π n f π n f 2 = mf C πn K, dm = πn Kf = mf 2 C πn K, dm = 1 π n Kf π n Kf 2.

15 onlinear filtering: interacting particle resolution 15 ext we derive a technical recursive formula in p, for the expressions m p+1 g n/p f n p+1 β p 0 p n 1. Lemma 5.2. Let f : S R be any integrable Borel test function and let n 0 and 1. For every 0 p n 1 we have the recursion m p+1 g n/p f n p+1 β p = πp K g n/p f n p+1 = m p g n/p 1 f n p m, P-a.e. 5.4 p g p/p 1 Proof. Using the inductive definition of f n p, g n/p and Lemma 5.1, we have m p+1 g n/p f p+1 n β p = π p Kg n/p f p+1 = m p = m p n gp/p 1 Kg n/p f n p+1 m p g p/p 1 gn/p 1 f n p m. p g p/p Martingale approach For estimating mean square errors, the key idea is to introduce a P, F n - martingale U, using the functions g n/n 1 and the random measures m n. More precisely, we define Un as follows. Definition 5.1. We denote by U the stochastic process defined by U 0 = 1, U n = m n g n/n 1 U n 1, for all n In other words, U n = n k=1 1 g k/k 1 x i k. Lemma 5.3. Un Un = 1. is a P, F n -martingale with ẼU n = 1 and P-a.s. Proof. The first statement follows by recalling that g n/n 1 x = g nx π n 1 Kg n = g ny n hx py n Y n 1.

16 16 P. Del Moral This gives P-a.s. ẼU n F n 1 = U n 1 = U n 1 mg n/n 1 C π n 1 K, dm dpy n Y n 1 g n y n hz π n 1Kdz dy n = U n 1 and the first assertion follows. To prove Ẽ[Y ]Un = 1, the above discussion goes through with some minor changes. Indeed, P-a.s. Un = ] Ẽ[Y m n g n/n 1 Un 1 = Ẽ[Y ]Ẽ[Y ] m n g n/n 1 β n 1 U n ow, using the inductive definition of g n/p, we obtain Then 5.6 gives m n g n/n 1 β n 1 = m n 1g n/n 2 /m n 1g n 1/n 2. U n = Ẽ[Y ]m n 1g n/n 2 U n 2 = Ẽ[Y ]Ẽ[Y ] m n 1g n/n 2 β n 2 U n 2. This procedure can be repeated. Using the recursive formulas described in Section 4.2, we note that Un = Ẽ[Y ]Ẽ[Y ] m p g n/p 1 β p 1 Up 1 Then m p g n/p 1 β p 1 = m p 1g n/p 2 /m p 1g p 1/p 2. U n = Ẽ[Y ]Ẽ[Y ] m p 1g n/p 2 β p 2 U p 2. Using backward induction in p, the result follows from the fact that m 1 g n/0 = νkg n/0 = 1. The analysis of U is a powerful tool to study the convergence rate of our interacting particle filter. That is, introducing the process U makes the calculation of mean errors possible, then these estimates will be reinterpreted back. As a typical example we have the following lemma. Lemma 5.4. For any integrable Borel test function f : S R, we have U n m n+1 f π n f = 0, P-a.e. 5.7

17 onlinear filtering: interacting particle resolution 17 Proof. Using Lemma 5.3, this is equivalent to proving U n m n+1 f = π n f. ow, using Lemma 5.1 and the fact that f n n = f, note that Un m n+1f = Ẽ[Y ]Ẽ[Y ] m n+1f β n Un Then m n+1 f β n = π n f = m n g n/n 1 f/m n g n/n 1. Un m n+1f = Ẽ[Y ] U n 1 m n g n/n 1 f n n. ow, using backward induction in 1 p n 1 and the recursive formulas described in Section 4.2, we have U p m p+1g n/p f n p+1 = Ẽ[Y ] U p Ẽ [y] m p+1 g n/p f n p+1 β p m p+1 g n/p f p+1 n β p Hence = m p g n/p 1 f p n /m n 1g p/p 1. U n m n+1f = Ẽ[Y ] U p 1 m p g n/p 1 f n p. The result finally follows from the recursive formulas described in Section 4.2 and the fact that n m 1 g n/0 f n 1 = νkg n/0 f n 1 = νkg n/0 f 1 νkg n/0 = π n f Mean square estimates To prove the convergence 3.4, it clearly suffices to prove that lim U n = 0 = lim f x i n π n f Un. 5.8 The main purpose of this section is to provide a way for estimating the rate of convergence of 5.8 in terms of the log-likelihood functions V n/p. Such computations are at the heart of the development. They will drastically simplify the evaluation of the convergence rates discussed in the last section. We first quote the following result. Proposition 5.1. For every series of observations y R,

18 18 P. Del Moral 1. for all 0 p < n, V n/p y V n/n 1 y + V n 1/n 2 y + + V p+1/p y V n y; 2. for every 0 k n and 0 p 0 < < p k 1 < p k = n, we have k l=1 V pl /p l 1 y V n y. Proof. The proof is a straightforward computation using that g n/p = Kg n/p+1 g p+1/p implies V n/p y = log Kgn/p 2 V n/p+1 y + V p+1/p y. ext we want to develop explicit formulas for expressing the effect of the population size and the observation likelihood functions on the mean square errors 5.8. The following propositions will be used repeatedly in the last section. Proposition 5.2. For every 1 and n 0 we have P-a.e. 1. U n U n 1 2 n k= k 1 n k 0 p 0< <p k =n 1 1 k exp l=1 n e VnY. V pl /p l 1 Y, Proof. Inequality 2 is clearly a consequence of inequality 1 and Proposition 5.1. Let us prove inequality 1. Using Lemma 5.3 U n 1 2 = Ẽ[Y ] U n 2 1. Hence, with the standard convention = 0, it is sufficient to prove that U n 2 n k=0 1 1 k 1 n k k exp V pl /p l 1 Y p 0< <p k =n l=1 The proof of 5.9 is based on backward and forward induction in n and maximisation techniques. Using Lemma 5.1 and the recursion in Lemma 5.2, we

19 onlinear filtering: interacting particle resolution 19 have U n 2 = Ẽ[Y ] Ẽ[Y ] m n g n/n 1 2 β n 1 U n 1 2 m n g n/n 1 2 β n 1 1 π n 1Kgn/n π n 1Kg n/n ev n/n 1Y m n 1 g n/n 2 2. m n 1g n 1/n 2 Thus, U n 2 1 ev n/n 1Y U n Ẽ[Y ] Ẽ[Y ] m n 1 g n/n 2 2 β n 2 U n 2 2. By the same line of arguments, for every 1 p n 1 m p+1 g n/p 2 β p m 1 g n/0 2 β 0 1 ev n/py m p g n/p 1 m p g p/p 1 1 ev n/0y νkgn/ Cascading in the above expression U n n 1 + n 1 q=0 1 1 n 1 qe V n/q Y U q 2. Suppose the inequalities 5.9 have been proved for every q n 1, that is then U q 2 q k=0 1 1 k 1 q k k exp V pl /p l 1 Y, 0 p 0< <p k =q l=1 U n k+1 n n 1 + k=0 0 p 0< <p k <n 1 1 n 1+k exp V n/pk Y + k l=1 V pl /p l 1 Y, and the result follows. The same techniques then establish the following result.

20 20 P. Del Moral Proposition 5.3. For any bounded Borel test function f : S R, 1 and n 0, we have P-a.e. Therefore f x i n π n f Un 0 p 0< <p k 1 p k =n 4 f 2 2 f x i n π n f Un 4 f 2 1 n+1 k=1 l=1 1 1 k 1 n+1 k k exp V pl /p l 1 Y n+1 exp V n Y Proof. Let us prove the first statement. For any bounded Borel test function f : S R, using the recursive formulas described in Section 4.2, we have Then 1 2 f x i nun m n+1 f 2 β n = Ẽ[Y ] m n+1 fun 2 = Ẽ[Y ] Ẽ[Y ] m n+1 f 2 β n U n 2 1 f m n g n/n 1 f 2. m n g n/n 1 m n+1 fun 2 1 f 2 Ẽ[Y ] U n Ẽ[Y ] m n g n/n 1 fun Arguing as above, for every 1 p n 1 we get m p+1g n/p f p+1 2 n U p = Ẽ[Y ] Ẽ[Y ] m p+1 g n/p f n p+1 2 βp U p 2 m p+1 gn/p f n p+1 2 βp m 1 1 f 2 e V n/py m p g n/p 1 f p m p g p/p 1 gn/0 f n 1 2 β0 1 f 2 e V n/0y π n f 2. n 2

21 onlinear filtering: interacting particle resolution 21 Using the above inequalities with expression 5.12, we conclude that m n+1 f U n 2 1 f 2 n k=0 1 1 n k e V n/k Y U k n+1 πn f 2, with the convention V n/n = 0. Then, using formula 5.9 we easily obtain m n+1 f π n f U n 2 4 f 2 4 f 2 n l=0 n+1 l=1 1 1 l 1 n l exp 0 p 0< <p l p l+1 =n l+1 s=1 V pl /p l 1 Y 1 1 l 1 n+1 l l exp V pl /p l 1 Y. 0 p 0< <p l 1 p l =n s=1 The second statement follows from the first inequality in Proposition 5.1 and the fact that n l n+1 l. 6. Convergence theorems We are now ready to prove the convergence of the interacting particle approximation described in Section 3. The following theorem is our main result, and it states the relevant consequences of the mean square error estimates stated in Propositions 5.2 and 5.3. Theorem 6.1. For every 1, n 0, any bounded Borel test function f : S R, T > 0, and 0 < ε < 1/2, we have { 1 sup P [Y ] n [0,T ] } f x i n π n f > ε AT, f, Y ε T +1 with AT, f, Y = 2 sup 4 f 2, 1 e V T Y. Moreover, 1 sup n [0,T ] f x i n π n f BT, f, Y 1 with BT, f, Y = 4 f e V T Y /2. P-a.e., T +1 1/2 P-a.e., 6.2

22 22 P. Del Moral Proof. Using Propositions 5.2 and 5.3 and the Cauchy Schwarz inequality, we have 1 = Ẽ[Y ] f x i n π n f 1 f x i n π n f BT, f, Y 1 U n 1 1 n+1 1/2 1 + Ẽ[Y ] f x i n π n f 1 Un and the inequality 6.2 follows. To prove 6.1, write A = {U n 1 ε} B = Using Proposition 5.2 and the fact that { 1 } f x i n π n f > ε. U n 1 ε, for all 0 < ε < 1 implies we have U n 1 ε, P [Y ] {Un 1 ε} P [Y ] { Un 1 ε} 1 1 ε 2 Ẽ[Y ] U n ε n e VnY. Therefore, P [Y ] {A} 1 1 ε n+1 e VnY. On the other hand, proposition 5.3 gives the inequalities P [Y ] {B A} 1 ε 2 Ẽ[Y ] 1 f x i n π n f 2 1 A 1 1 ] 1 εε 2 Ẽ[Y 4 f 2 1 εε f x i n π n f Un n+1 e VnY.

23 onlinear filtering: interacting particle resolution 23 Finally, the inequality P [Y ] {B} P [Y ] {B A} + P [Y ] {A} 6.3 and the fact that 0 < ε < 1/2 implies ε < 1 ε, complete the proof. Corollary 6.1. For every 2, n 0, for any integrable Borel test function f : S R, T > 0, and 0 < ε < 1/2, we have { 1 sup P [Y ] n [0,T ] } f x i n π n f > ε where Aε, T, f, Y = 4 T AT, f, Y /ε 4. Aε, T, f, Y, P-a.e., 6.4 Proof. From the inequality e y 1 + y, it follows that n+1 n + 1 log 1 1 = n + 1 log 1 and that log x x 1. Thus, we conclude that n+1 n + 1 2n n 1 for all 2 and for all n 1. This completes the proof. Theorem 6.2. Let Y = y R be a series of observations such that V n y < +. For every bounded Borel test function f : S R 1 and n 0, we have for all p > 0, Proof. Using the inequality for all ε > 0, sup n [0,T ] Ẽ [y] 1 Ẽ [y] 1 and Corollary 6.1, the result follows. p f x i n π n f ε p + 2 f p P[y] { 1 p f x i n π n f Consider the following assumption on the observation process: } f x i n π n f > ε V for all n 0 Ẽ V n Y def = αn < This assumption enables us to estimate the convergence rate of our approximations in spaces L p P with p > 0.

24 24 P. Del Moral Theorem 6.3. Assume V is satisfied. For every bounded Borel test function f : S R 1, n 0, 0 < ε < 1/2 and M > 0, we have { 1 sup P n [0,T ] } f x i n π n f > ε 1 M with AT, M, f, ε = 8 sup 4 f, 1 T e M αt /ε 4. Proof. Using Theorem 6.1 { 1 P } f x i n π n f ε P V n Y Mαn with Af = 2 sup 4 f, 1, we get { 1 P } f x i n π n f ε 1 1 M AT, M, f, ε +, 6.7 Af emαn ε n+1, Af emαn + ε n+1. The arguments used in the proof of Corollary 6.1 complete the proof. Finally, the following corollary is derived by the same reasoning as in Theorem 6.2. Corollary 6.2. Assume V is satisfied. For every bounded Borel test function f : S R, 1 and n 0, we have for all p > 0, Acknowledgements 1 sup Ẽ n [0,T ] p f x i n π n f I am deeply grateful to Laure Coutin and Pierre Vandekerkhove for their careful reading of the original manuscript. I also thank the anonymous referee for his comments and suggestions that greatly improved the presentation of the paper. References [1] H. Carvalho, P. Del Moral, A. Monin and G. Salut 1996 Optimal nonlinear filtering in GPS/IS integration. To appear in IEEE Trans. Aerospace and Electron. Syst.

25 onlinear filtering: interacting particle resolution 25 [2] R. Cerf 1994 Une théorie asymptotique des Algorithmes Génétiques. Université Montpellier II, Ph.D. Sciences et techniques du Languedoc. [3] M. Chaleyat-Maurel and D. Michel 1983 Des résultats de non existence de filtres de dimension finie. C.R. Acad. Sci. Paris, Ser. I, t [4] P. Del Moral, G. Rigal, J.C. oyer and G. Salut 1993 Traitement non-linéaire du signal par reseau particulaire: application radar. Quatorzième colloque GRETSI, Juan les Pins, Sept [5] P. Del Moral 1994 Résolution particulaire des problèmes d estimation et d optimisation non-linéaires. Thèse de l Université Paul Sabatier, Rapport LAAS 94269, Juin [6] P. Del Moral and G. Salut 1995 Filtrage non-linéaire: résolution particulaire à la Monte-Carlo. C.R. Acad. Sci. Paris, Ser. I, t [7] P. Del Moral 1995 on linear filtering using random particles. Probab. Theory Appl [8] P. Del Moral, J.C. oyer and G. Salut 1995 Résolution particulaire et traitement non-linéaire du signal: Application Radar/Sonar. Traitement du signal, Septembre [9] P. Del Moral and G. Salut 1996 Particle interpretation of nonlinear filtering and optimization. Russian Journal of Mathematical Physics. In press. [10].J. Gordon, D.J. Salmon and A.F.M. Smith 1993 ovel approach to non-linear/non-gaussian Bayesian state estimation. IEE. [11] J.H. Holland 1975 Adaptation in atural and Artificial Systems. University of Michigan Press, Ann Arbor. [12] H. Kunita 1971 Asymptotic behavior of nonlinear filtering errors of Markov processes. Journal of Multivariate Analysis 1 4, [13] L. Stettner 1989 On Invariant Measures of Filtering Processes. Stochastic Differential Systems, Lecture otes in Control and Information Sciences, 126, Springer Verlag.

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS

SUPPLEMENT TO PAPER CONVERGENCE OF ADAPTIVE AND INTERACTING MARKOV CHAIN MONTE CARLO ALGORITHMS Submitted to the Annals of Statistics SUPPLEMENT TO PAPER CONERGENCE OF ADAPTIE AND INTERACTING MARKO CHAIN MONTE CARLO ALGORITHMS By G Fort,, E Moulines and P Priouret LTCI, CNRS - TELECOM ParisTech,

More information

Introduction. log p θ (y k y 1:k 1 ), k=1

Introduction. log p θ (y k y 1:k 1 ), k=1 ESAIM: PROCEEDINGS, September 2007, Vol.19, 115-120 Christophe Andrieu & Dan Crisan, Editors DOI: 10.1051/proc:071915 PARTICLE FILTER-BASED APPROXIMATE MAXIMUM LIKELIHOOD INFERENCE ASYMPTOTICS IN STATE-SPACE

More information

A Note on Auxiliary Particle Filters

A Note on Auxiliary Particle Filters A Note on Auxiliary Particle Filters Adam M. Johansen a,, Arnaud Doucet b a Department of Mathematics, University of Bristol, UK b Departments of Statistics & Computer Science, University of British Columbia,

More information

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989),

1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer 11(2) (1989), Real Analysis 2, Math 651, Spring 2005 April 26, 2005 1 Real Analysis 2, Math 651, Spring 2005 Krzysztof Chris Ciesielski 1/12/05: sec 3.1 and my article: How good is the Lebesgue measure?, Math. Intelligencer

More information

Particle Filters: Convergence Results and High Dimensions

Particle Filters: Convergence Results and High Dimensions Particle Filters: Convergence Results and High Dimensions Mark Coates mark.coates@mcgill.ca McGill University Department of Electrical and Computer Engineering Montreal, Quebec, Canada Bellairs 2012 Outline

More information

A NON-PARAMETRIC TEST FOR NON-INDEPENDENT NOISES AGAINST A BILINEAR DEPENDENCE

A NON-PARAMETRIC TEST FOR NON-INDEPENDENT NOISES AGAINST A BILINEAR DEPENDENCE REVSTAT Statistical Journal Volume 3, Number, November 5, 155 17 A NON-PARAMETRIC TEST FOR NON-INDEPENDENT NOISES AGAINST A BILINEAR DEPENDENCE Authors: E. Gonçalves Departamento de Matemática, Universidade

More information

Lecture 7: Optimal Smoothing

Lecture 7: Optimal Smoothing Department of Biomedical Engineering and Computational Science Aalto University March 17, 2011 Contents 1 What is Optimal Smoothing? 2 Bayesian Optimal Smoothing Equations 3 Rauch-Tung-Striebel Smoother

More information

3 Integration and Expectation

3 Integration and Expectation 3 Integration and Expectation 3.1 Construction of the Lebesgue Integral Let (, F, µ) be a measure space (not necessarily a probability space). Our objective will be to define the Lebesgue integral R fdµ

More information

Faithful couplings of Markov chains: now equals forever

Faithful couplings of Markov chains: now equals forever Faithful couplings of Markov chains: now equals forever by Jeffrey S. Rosenthal* Department of Statistics, University of Toronto, Toronto, Ontario, Canada M5S 1A1 Phone: (416) 978-4594; Internet: jeff@utstat.toronto.edu

More information

SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE

SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE SEPARABILITY AND COMPLETENESS FOR THE WASSERSTEIN DISTANCE FRANÇOIS BOLLEY Abstract. In this note we prove in an elementary way that the Wasserstein distances, which play a basic role in optimal transportation

More information

Computational Graphs, and Backpropagation

Computational Graphs, and Backpropagation Chapter 1 Computational Graphs, and Backpropagation (Course notes for NLP by Michael Collins, Columbia University) 1.1 Introduction We now describe the backpropagation algorithm for calculation of derivatives

More information

Contraction properties of Feynman-Kac semigroups

Contraction properties of Feynman-Kac semigroups Journées de Statistique Marne La Vallée, January 2005 Contraction properties of Feynman-Kac semigroups Pierre DEL MORAL, Laurent MICLO Lab. J. Dieudonné, Nice Univ., LATP Univ. Provence, Marseille 1 Notations

More information

Geometric ρ-mixing property of the interarrival times of a stationary Markovian Arrival Process

Geometric ρ-mixing property of the interarrival times of a stationary Markovian Arrival Process Author manuscript, published in "Journal of Applied Probability 50, 2 (2013) 598-601" Geometric ρ-mixing property of the interarrival times of a stationary Markovian Arrival Process L. Hervé and J. Ledoux

More information

Convergence of the Ensemble Kalman Filter in Hilbert Space

Convergence of the Ensemble Kalman Filter in Hilbert Space Convergence of the Ensemble Kalman Filter in Hilbert Space Jan Mandel Center for Computational Mathematics Department of Mathematical and Statistical Sciences University of Colorado Denver Parts based

More information

On Reparametrization and the Gibbs Sampler

On Reparametrization and the Gibbs Sampler On Reparametrization and the Gibbs Sampler Jorge Carlos Román Department of Mathematics Vanderbilt University James P. Hobert Department of Statistics University of Florida March 2014 Brett Presnell Department

More information

A Note on the Central Limit Theorem for a Class of Linear Systems 1

A Note on the Central Limit Theorem for a Class of Linear Systems 1 A Note on the Central Limit Theorem for a Class of Linear Systems 1 Contents Yukio Nagahata Department of Mathematics, Graduate School of Engineering Science Osaka University, Toyonaka 560-8531, Japan.

More information

Brownian Motion and Stochastic Calculus

Brownian Motion and Stochastic Calculus ETHZ, Spring 17 D-MATH Prof Dr Martin Larsson Coordinator A Sepúlveda Brownian Motion and Stochastic Calculus Exercise sheet 6 Please hand in your solutions during exercise class or in your assistant s

More information

Inverse Brascamp-Lieb inequalities along the Heat equation

Inverse Brascamp-Lieb inequalities along the Heat equation Inverse Brascamp-Lieb inequalities along the Heat equation Franck Barthe and Dario Cordero-Erausquin October 8, 003 Abstract Adapting Borell s proof of Ehrhard s inequality for general sets, we provide

More information

Universal examples. Chapter The Bernoulli process

Universal examples. Chapter The Bernoulli process Chapter 1 Universal examples 1.1 The Bernoulli process First description: Bernoulli random variables Y i for i = 1, 2, 3,... independent with P [Y i = 1] = p and P [Y i = ] = 1 p. Second description: Binomial

More information

Gaussian Processes. 1. Basic Notions

Gaussian Processes. 1. Basic Notions Gaussian Processes 1. Basic Notions Let T be a set, and X : {X } T a stochastic process, defined on a suitable probability space (Ω P), that is indexed by T. Definition 1.1. We say that X is a Gaussian

More information

CONDITIONAL ERGODICITY IN INFINITE DIMENSION. By Xin Thomson Tong and Ramon van Handel Princeton University

CONDITIONAL ERGODICITY IN INFINITE DIMENSION. By Xin Thomson Tong and Ramon van Handel Princeton University CONDITIONAL ERGODICITY IN INFINITE DIMENSION By Xin Thomson Tong and Ramon van Handel Princeton University The goal of this paper is to develop a general method to establish conditional ergodicity of infinite-dimensional

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes

Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Alea 4, 117 129 (2008) Convergence rates in weighted L 1 spaces of kernel density estimators for linear processes Anton Schick and Wolfgang Wefelmeyer Anton Schick, Department of Mathematical Sciences,

More information

University of Toronto Department of Statistics

University of Toronto Department of Statistics Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704

More information

Sequential Monte Carlo Methods for Bayesian Computation

Sequential Monte Carlo Methods for Bayesian Computation Sequential Monte Carlo Methods for Bayesian Computation A. Doucet Kyoto Sept. 2012 A. Doucet (MLSS Sept. 2012) Sept. 2012 1 / 136 Motivating Example 1: Generic Bayesian Model Let X be a vector parameter

More information

Sobolev Spaces. Chapter 10

Sobolev Spaces. Chapter 10 Chapter 1 Sobolev Spaces We now define spaces H 1,p (R n ), known as Sobolev spaces. For u to belong to H 1,p (R n ), we require that u L p (R n ) and that u have weak derivatives of first order in L p

More information

STABILITY AND UNIFORM APPROXIMATION OF NONLINEAR FILTERS USING THE HILBERT METRIC, AND APPLICATION TO PARTICLE FILTERS 1

STABILITY AND UNIFORM APPROXIMATION OF NONLINEAR FILTERS USING THE HILBERT METRIC, AND APPLICATION TO PARTICLE FILTERS 1 The Annals of Applied Probability 0000, Vol. 00, No. 00, 000 000 STABILITY AND UNIFORM APPROXIMATION OF NONLINAR FILTRS USING TH HILBRT MTRIC, AND APPLICATION TO PARTICL FILTRS 1 By François LeGland and

More information

Empirical Processes: General Weak Convergence Theory

Empirical Processes: General Weak Convergence Theory Empirical Processes: General Weak Convergence Theory Moulinath Banerjee May 18, 2010 1 Extended Weak Convergence The lack of measurability of the empirical process with respect to the sigma-field generated

More information

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011

LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS. S. G. Bobkov and F. L. Nazarov. September 25, 2011 LARGE DEVIATIONS OF TYPICAL LINEAR FUNCTIONALS ON A CONVEX BODY WITH UNCONDITIONAL BASIS S. G. Bobkov and F. L. Nazarov September 25, 20 Abstract We study large deviations of linear functionals on an isotropic

More information

The Hopf argument. Yves Coudene. IRMAR, Université Rennes 1, campus beaulieu, bat Rennes cedex, France

The Hopf argument. Yves Coudene. IRMAR, Université Rennes 1, campus beaulieu, bat Rennes cedex, France The Hopf argument Yves Coudene IRMAR, Université Rennes, campus beaulieu, bat.23 35042 Rennes cedex, France yves.coudene@univ-rennes.fr slightly updated from the published version in Journal of Modern

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1)

n [ F (b j ) F (a j ) ], n j=1(a j, b j ] E (4.1) 1.4. CONSTRUCTION OF LEBESGUE-STIELTJES MEASURES In this section we shall put to use the Carathéodory-Hahn theory, in order to construct measures with certain desirable properties first on the real line

More information

Independence of some multiple Poisson stochastic integrals with variable-sign kernels

Independence of some multiple Poisson stochastic integrals with variable-sign kernels Independence of some multiple Poisson stochastic integrals with variable-sign kernels Nicolas Privault Division of Mathematical Sciences School of Physical and Mathematical Sciences Nanyang Technological

More information

MATHS 730 FC Lecture Notes March 5, Introduction

MATHS 730 FC Lecture Notes March 5, Introduction 1 INTRODUCTION MATHS 730 FC Lecture Notes March 5, 2014 1 Introduction Definition. If A, B are sets and there exists a bijection A B, they have the same cardinality, which we write as A, #A. If there exists

More information

The Stein and Chen-Stein methods for functionals of non-symmetric Bernoulli processes

The Stein and Chen-Stein methods for functionals of non-symmetric Bernoulli processes The Stein and Chen-Stein methods for functionals of non-symmetric Bernoulli processes Nicolas Privault Giovanni Luca Torrisi Abstract Based on a new multiplication formula for discrete multiple stochastic

More information

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University

Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University Lecture Notes in Advanced Calculus 1 (80315) Raz Kupferman Institute of Mathematics The Hebrew University February 7, 2007 2 Contents 1 Metric Spaces 1 1.1 Basic definitions...........................

More information

OPTIMAL TRANSPORTATION PLANS AND CONVERGENCE IN DISTRIBUTION

OPTIMAL TRANSPORTATION PLANS AND CONVERGENCE IN DISTRIBUTION OPTIMAL TRANSPORTATION PLANS AND CONVERGENCE IN DISTRIBUTION J.A. Cuesta-Albertos 1, C. Matrán 2 and A. Tuero-Díaz 1 1 Departamento de Matemáticas, Estadística y Computación. Universidad de Cantabria.

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

Convergence at first and second order of some approximations of stochastic integrals

Convergence at first and second order of some approximations of stochastic integrals Convergence at first and second order of some approximations of stochastic integrals Bérard Bergery Blandine, Vallois Pierre IECN, Nancy-Université, CNRS, INRIA, Boulevard des Aiguillettes B.P. 239 F-5456

More information

A Backward Particle Interpretation of Feynman-Kac Formulae

A Backward Particle Interpretation of Feynman-Kac Formulae A Backward Particle Interpretation of Feynman-Kac Formulae P. Del Moral Centre INRIA de Bordeaux - Sud Ouest Workshop on Filtering, Cambridge Univ., June 14-15th 2010 Preprints (with hyperlinks), joint

More information

A connection between the stochastic heat equation and fractional Brownian motion, and a simple proof of a result of Talagrand

A connection between the stochastic heat equation and fractional Brownian motion, and a simple proof of a result of Talagrand A connection between the stochastic heat equation and fractional Brownian motion, and a simple proof of a result of Talagrand Carl Mueller 1 and Zhixin Wu Abstract We give a new representation of fractional

More information

Review of Probability Theory

Review of Probability Theory Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving

More information

Tools from Lebesgue integration

Tools from Lebesgue integration Tools from Lebesgue integration E.P. van den Ban Fall 2005 Introduction In these notes we describe some of the basic tools from the theory of Lebesgue integration. Definitions and results will be given

More information

Properties of an infinite dimensional EDS system : the Muller s ratchet

Properties of an infinite dimensional EDS system : the Muller s ratchet Properties of an infinite dimensional EDS system : the Muller s ratchet LATP June 5, 2011 A ratchet source : wikipedia Plan 1 Introduction : The model of Haigh 2 3 Hypothesis (Biological) : The population

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing

More information

Concentration inequalities for Feynman-Kac particle models. P. Del Moral. INRIA Bordeaux & IMB & CMAP X. Journées MAS 2012, SMAI Clermond-Ferrand

Concentration inequalities for Feynman-Kac particle models. P. Del Moral. INRIA Bordeaux & IMB & CMAP X. Journées MAS 2012, SMAI Clermond-Ferrand Concentration inequalities for Feynman-Kac particle models P. Del Moral INRIA Bordeaux & IMB & CMAP X Journées MAS 2012, SMAI Clermond-Ferrand Some hyper-refs Feynman-Kac formulae, Genealogical & Interacting

More information

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,

More information

Auxiliary Particle Methods

Auxiliary Particle Methods Auxiliary Particle Methods Perspectives & Applications Adam M. Johansen 1 adam.johansen@bristol.ac.uk Oxford University Man Institute 29th May 2008 1 Collaborators include: Arnaud Doucet, Nick Whiteley

More information

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535

Additive functionals of infinite-variance moving averages. Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Additive functionals of infinite-variance moving averages Wei Biao Wu The University of Chicago TECHNICAL REPORT NO. 535 Departments of Statistics The University of Chicago Chicago, Illinois 60637 June

More information

Lecture 21 Representations of Martingales

Lecture 21 Representations of Martingales Lecture 21: Representations of Martingales 1 of 11 Course: Theory of Probability II Term: Spring 215 Instructor: Gordan Zitkovic Lecture 21 Representations of Martingales Right-continuous inverses Let

More information

Entropy and Ergodic Theory Lecture 15: A first look at concentration

Entropy and Ergodic Theory Lecture 15: A first look at concentration Entropy and Ergodic Theory Lecture 15: A first look at concentration 1 Introduction to concentration Let X 1, X 2,... be i.i.d. R-valued RVs with common distribution µ, and suppose for simplicity that

More information

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes

Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Pointwise convergence rates and central limit theorems for kernel density estimators in linear processes Anton Schick Binghamton University Wolfgang Wefelmeyer Universität zu Köln Abstract Convergence

More information

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan

Monte-Carlo MMD-MA, Université Paris-Dauphine. Xiaolu Tan Monte-Carlo MMD-MA, Université Paris-Dauphine Xiaolu Tan tan@ceremade.dauphine.fr Septembre 2015 Contents 1 Introduction 1 1.1 The principle.................................. 1 1.2 The error analysis

More information

EXISTENCE AND REGULARITY RESULTS FOR SOME NONLINEAR PARABOLIC EQUATIONS

EXISTENCE AND REGULARITY RESULTS FOR SOME NONLINEAR PARABOLIC EQUATIONS EXISTECE AD REGULARITY RESULTS FOR SOME OLIEAR PARABOLIC EUATIOS Lucio BOCCARDO 1 Andrea DALL AGLIO 2 Thierry GALLOUËT3 Luigi ORSIA 1 Abstract We prove summability results for the solutions of nonlinear

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

An almost sure invariance principle for additive functionals of Markov chains

An almost sure invariance principle for additive functionals of Markov chains Statistics and Probability Letters 78 2008 854 860 www.elsevier.com/locate/stapro An almost sure invariance principle for additive functionals of Markov chains F. Rassoul-Agha a, T. Seppäläinen b, a Department

More information

Series of Error Terms for Rational Approximations of Irrational Numbers

Series of Error Terms for Rational Approximations of Irrational Numbers 2 3 47 6 23 Journal of Integer Sequences, Vol. 4 20, Article..4 Series of Error Terms for Rational Approximations of Irrational Numbers Carsten Elsner Fachhochschule für die Wirtschaft Hannover Freundallee

More information

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing Advances in Dynamical Systems and Applications ISSN 0973-5321, Volume 8, Number 2, pp. 401 412 (2013) http://campus.mst.edu/adsa Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

CHAPTER 6. Differentiation

CHAPTER 6. Differentiation CHPTER 6 Differentiation The generalization from elementary calculus of differentiation in measure theory is less obvious than that of integration, and the methods of treating it are somewhat involved.

More information

Alternative Characterizations of Markov Processes

Alternative Characterizations of Markov Processes Chapter 10 Alternative Characterizations of Markov Processes This lecture introduces two ways of characterizing Markov processes other than through their transition probabilities. Section 10.1 describes

More information

DEGREE AND SOBOLEV SPACES. Haïm Brezis Yanyan Li Petru Mironescu Louis Nirenberg. Introduction

DEGREE AND SOBOLEV SPACES. Haïm Brezis Yanyan Li Petru Mironescu Louis Nirenberg. Introduction Topological Methods in Nonlinear Analysis Journal of the Juliusz Schauder Center Volume 13, 1999, 181 190 DEGREE AND SOBOLEV SPACES Haïm Brezis Yanyan Li Petru Mironescu Louis Nirenberg Dedicated to Jürgen

More information

Regularity of the density for the stochastic heat equation

Regularity of the density for the stochastic heat equation Regularity of the density for the stochastic heat equation Carl Mueller 1 Department of Mathematics University of Rochester Rochester, NY 15627 USA email: cmlr@math.rochester.edu David Nualart 2 Department

More information

ERRATA: Probabilistic Techniques in Analysis

ERRATA: Probabilistic Techniques in Analysis ERRATA: Probabilistic Techniques in Analysis ERRATA 1 Updated April 25, 26 Page 3, line 13. A 1,..., A n are independent if P(A i1 A ij ) = P(A 1 ) P(A ij ) for every subset {i 1,..., i j } of {1,...,

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 7 Sequential Monte Carlo methods III 7 April 2017 Computer Intensive Methods (1) Plan of today s lecture

More information

ON COMPOUND POISSON POPULATION MODELS

ON COMPOUND POISSON POPULATION MODELS ON COMPOUND POISSON POPULATION MODELS Martin Möhle, University of Tübingen (joint work with Thierry Huillet, Université de Cergy-Pontoise) Workshop on Probability, Population Genetics and Evolution Centre

More information

4 Derivations of the Discrete-Time Kalman Filter

4 Derivations of the Discrete-Time Kalman Filter Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof N Shimkin 4 Derivations of the Discrete-Time

More information

Alternate Characterizations of Markov Processes

Alternate Characterizations of Markov Processes Chapter 10 Alternate Characterizations of Markov Processes This lecture introduces two ways of characterizing Markov processes other than through their transition probabilities. Section 10.1 addresses

More information

An essay on the general theory of stochastic processes

An essay on the general theory of stochastic processes Probability Surveys Vol. 3 (26) 345 412 ISSN: 1549-5787 DOI: 1.1214/1549578614 An essay on the general theory of stochastic processes Ashkan Nikeghbali ETHZ Departement Mathematik, Rämistrasse 11, HG G16

More information

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales.

Lecture 2. We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. Lecture 2 1 Martingales We now introduce some fundamental tools in martingale theory, which are useful in controlling the fluctuation of martingales. 1.1 Doob s inequality We have the following maximal

More information

Supermodular ordering of Poisson arrays

Supermodular ordering of Poisson arrays Supermodular ordering of Poisson arrays Bünyamin Kızıldemir Nicolas Privault Division of Mathematical Sciences School of Physical and Mathematical Sciences Nanyang Technological University 637371 Singapore

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Mean-field dual of cooperative reproduction

Mean-field dual of cooperative reproduction The mean-field dual of systems with cooperative reproduction joint with Tibor Mach (Prague) A. Sturm (Göttingen) Friday, July 6th, 2018 Poisson construction of Markov processes Let (X t ) t 0 be a continuous-time

More information

ON THE ZERO-ONE LAW AND THE LAW OF LARGE NUMBERS FOR RANDOM WALK IN MIXING RAN- DOM ENVIRONMENT

ON THE ZERO-ONE LAW AND THE LAW OF LARGE NUMBERS FOR RANDOM WALK IN MIXING RAN- DOM ENVIRONMENT Elect. Comm. in Probab. 10 (2005), 36 44 ELECTRONIC COMMUNICATIONS in PROBABILITY ON THE ZERO-ONE LAW AND THE LAW OF LARGE NUMBERS FOR RANDOM WALK IN MIXING RAN- DOM ENVIRONMENT FIRAS RASSOUL AGHA Department

More information

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS

OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS APPLICATIONES MATHEMATICAE 29,4 (22), pp. 387 398 Mariusz Michta (Zielona Góra) OPTIMAL SOLUTIONS TO STOCHASTIC DIFFERENTIAL INCLUSIONS Abstract. A martingale problem approach is used first to analyze

More information

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS

(2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS (2m)-TH MEAN BEHAVIOR OF SOLUTIONS OF STOCHASTIC DIFFERENTIAL EQUATIONS UNDER PARAMETRIC PERTURBATIONS Svetlana Janković and Miljana Jovanović Faculty of Science, Department of Mathematics, University

More information

Wiener Measure and Brownian Motion

Wiener Measure and Brownian Motion Chapter 16 Wiener Measure and Brownian Motion Diffusion of particles is a product of their apparently random motion. The density u(t, x) of diffusing particles satisfies the diffusion equation (16.1) u

More information

The Uniform Integrability of Martingales. On a Question by Alexander Cherny

The Uniform Integrability of Martingales. On a Question by Alexander Cherny The Uniform Integrability of Martingales. On a Question by Alexander Cherny Johannes Ruf Department of Mathematics University College London May 1, 2015 Abstract Let X be a progressively measurable, almost

More information

A NEW NONLINEAR FILTER

A NEW NONLINEAR FILTER COMMUNICATIONS IN INFORMATION AND SYSTEMS c 006 International Press Vol 6, No 3, pp 03-0, 006 004 A NEW NONLINEAR FILTER ROBERT J ELLIOTT AND SIMON HAYKIN Abstract A discrete time filter is constructed

More information

Defining the Integral

Defining the Integral Defining the Integral In these notes we provide a careful definition of the Lebesgue integral and we prove each of the three main convergence theorems. For the duration of these notes, let (, M, µ) be

More information

The Gaussian free field, Gibbs measures and NLS on planar domains

The Gaussian free field, Gibbs measures and NLS on planar domains The Gaussian free field, Gibbs measures and on planar domains N. Burq, joint with L. Thomann (Nantes) and N. Tzvetkov (Cergy) Université Paris Sud, Laboratoire de Mathématiques d Orsay, CNRS UMR 8628 LAGA,

More information

Learning Static Parameters in Stochastic Processes

Learning Static Parameters in Stochastic Processes Learning Static Parameters in Stochastic Processes Bharath Ramsundar December 14, 2012 1 Introduction Consider a Markovian stochastic process X T evolving (perhaps nonlinearly) over time variable T. We

More information

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms

08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms (February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops

More information

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9

MAT 570 REAL ANALYSIS LECTURE NOTES. Contents. 1. Sets Functions Countability Axiom of choice Equivalence relations 9 MAT 570 REAL ANALYSIS LECTURE NOTES PROFESSOR: JOHN QUIGG SEMESTER: FALL 204 Contents. Sets 2 2. Functions 5 3. Countability 7 4. Axiom of choice 8 5. Equivalence relations 9 6. Real numbers 9 7. Extended

More information

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline. Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,

More information

On a Class of Multidimensional Optimal Transportation Problems

On a Class of Multidimensional Optimal Transportation Problems Journal of Convex Analysis Volume 10 (2003), No. 2, 517 529 On a Class of Multidimensional Optimal Transportation Problems G. Carlier Université Bordeaux 1, MAB, UMR CNRS 5466, France and Université Bordeaux

More information

Bayesian estimation of Hidden Markov Models

Bayesian estimation of Hidden Markov Models Bayesian estimation of Hidden Markov Models Laurent Mevel, Lorenzo Finesso IRISA/INRIA Campus de Beaulieu, 35042 Rennes Cedex, France Institute of Systems Science and Biomedical Engineering LADSEB-CNR

More information

9 Brownian Motion: Construction

9 Brownian Motion: Construction 9 Brownian Motion: Construction 9.1 Definition and Heuristics The central limit theorem states that the standard Gaussian distribution arises as the weak limit of the rescaled partial sums S n / p n of

More information

SOLUTIONS OF SEMILINEAR WAVE EQUATION VIA STOCHASTIC CASCADES

SOLUTIONS OF SEMILINEAR WAVE EQUATION VIA STOCHASTIC CASCADES Communications on Stochastic Analysis Vol. 4, No. 3 010) 45-431 Serials Publications www.serialspublications.com SOLUTIONS OF SEMILINEAR WAVE EQUATION VIA STOCHASTIC CASCADES YURI BAKHTIN* AND CARL MUELLER

More information

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 Last compiled: November 6, 213 1. Conditional expectation Exercise 1.1. To start with, note that P(X Y = P( c R : X > c, Y c or X c, Y > c = P( c Q : X > c, Y

More information

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product

Finite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )

More information

k-protected VERTICES IN BINARY SEARCH TREES

k-protected VERTICES IN BINARY SEARCH TREES k-protected VERTICES IN BINARY SEARCH TREES MIKLÓS BÓNA Abstract. We show that for every k, the probability that a randomly selected vertex of a random binary search tree on n nodes is at distance k from

More information

PREDICTABLE REPRESENTATION PROPERTY OF SOME HILBERTIAN MARTINGALES. 1. Introduction.

PREDICTABLE REPRESENTATION PROPERTY OF SOME HILBERTIAN MARTINGALES. 1. Introduction. Acta Math. Univ. Comenianae Vol. LXXVII, 1(28), pp. 123 128 123 PREDICTABLE REPRESENTATION PROPERTY OF SOME HILBERTIAN MARTINGALES M. EL KADIRI Abstract. We prove as for the real case that a martingale

More information

Citation Osaka Journal of Mathematics. 41(4)

Citation Osaka Journal of Mathematics. 41(4) TitleA non quasi-invariance of the Brown Authors Sadasue, Gaku Citation Osaka Journal of Mathematics. 414 Issue 4-1 Date Text Version publisher URL http://hdl.handle.net/1194/1174 DOI Rights Osaka University

More information

The Codimension of the Zeros of a Stable Process in Random Scenery

The Codimension of the Zeros of a Stable Process in Random Scenery The Codimension of the Zeros of a Stable Process in Random Scenery Davar Khoshnevisan The University of Utah, Department of Mathematics Salt Lake City, UT 84105 0090, U.S.A. davar@math.utah.edu http://www.math.utah.edu/~davar

More information

Extreme Value Analysis and Spatial Extremes

Extreme Value Analysis and Spatial Extremes Extreme Value Analysis and Department of Statistics Purdue University 11/07/2013 Outline Motivation 1 Motivation 2 Extreme Value Theorem and 3 Bayesian Hierarchical Models Copula Models Max-stable Models

More information

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit

Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Convergence of Square Root Ensemble Kalman Filters in the Large Ensemble Limit Evan Kwiatkowski, Jan Mandel University of Colorado Denver December 11, 2014 OUTLINE 2 Data Assimilation Bayesian Estimation

More information

Nonparametric inference for ergodic, stationary time series.

Nonparametric inference for ergodic, stationary time series. G. Morvai, S. Yakowitz, and L. Györfi: Nonparametric inference for ergodic, stationary time series. Ann. Statist. 24 (1996), no. 1, 370 379. Abstract The setting is a stationary, ergodic time series. The

More information

Self-normalized laws of the iterated logarithm

Self-normalized laws of the iterated logarithm Journal of Statistical and Econometric Methods, vol.3, no.3, 2014, 145-151 ISSN: 1792-6602 print), 1792-6939 online) Scienpress Ltd, 2014 Self-normalized laws of the iterated logarithm Igor Zhdanov 1 Abstract

More information