Asymptotics for posterior hazards

Size: px
Start display at page:

Download "Asymptotics for posterior hazards"

Transcription

1 Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and Igor Prünster (University of Turin)

2 Outline Life-testing model with mixture hazard rate Mixture hazard rate Kernels and completely random measures Asymptotic issues Posterior consistency Weak consistency Sufficient condition for Kullback-Leibler support Prior probability of uniform balls of the true hazard rate CLTs for functionals of the posterior hazard rate Posterior distribution Limit theorems for shifted measures Applications

3 Mixture hazard rate Let Y be a positive absolutely continuous random variable (lifetime) with random hazard rate of the form h(t) = k(t, x) µ(dx), X k is a kernel, jointly measurable application R + X R +, µ is a random mixing measure on the space X, modeled as a completely random measure. The cumulative hazard is given by H(t) = t h(s)ds and is required to 0 satisfy H(t) for t a.s. Then f (t) = h(t) exp( H(t)) defines a random density function to be used in Bayesian nonparametric inference. See Dykstra and Laud (1981) and Lo and Weng (1989).

4 Kernels As examples,we consider the class of mixture hazard rates that arise under the following choice of kernels: Dykstra-Laud (DL) kernel (monotone increasing hazard rates) k(t, x) = I (0 x t) rectangular (rect) kernel with bandwidth τ > 0 k(t, x) = I ( t x τ) Ornstein-Uhlenbeck (OU) kernel with κ > 0 k(t, x) = 2κ e κ(t x) I (0 x t) exponential (exp) kernel (monotone decreasing hazard rates) k(t, x) = x 1 e t x.

5 Completely random measures Let M be the space of boundedly finite measures on X. Definition. A random element µ taking values in M such that (i) µ( ) = 0 (ii) for any disjoint sets (A i ) i 1 : µ(a 1 ), µ(a 2 ),... are mutually independent µ( j 1 A j ) = j 1 µ(a j) is said to be a completely random measure (CRM) on X. Remark. A CRM can always be represented as a linear functional of a Poisson random measure. In particular, µ is uniquely characterized by its Laplace functional [ E e R X g(x) µ(dx)] R = e R + X [1 e v g(x) ]ν(dv,dx) where ν stands for the intensity of the Poisson random measure underlying µ.

6 CRMs select almost surely discrete measures: hence, µ can always be represented as i 1 J iδ Xi. Note that if ν(dv, dx) = ρ(dv)λ(dx) the law of the J i s and the X i s are independent. In this case µ is termed homogeneous. A non-homogeneous CRM has ν(dv, dx) = ρ(dv x)λ(dx). Non-homogeneous CRMs arise in the posterior. If ν(r +, dx) = for any x, then the CRM jumps infinitely often on any bounded set A. However, recall that µ(a) < a.s. We consider CRMs with Poisson intensity of the following form: ν(dv, dx) = 1 Γ(1 σ) e γ(x)v dv λ(dx), σ [0, 1), γ( ) > 0. v 1+σ γ constant generalized gamma measure (Brix, 1999) σ = 0 weighted gamma measure (Lo & Weng, 1989).

7 Asymptotic issues Posterior consistency. First generate independent data from a true fixed density f 0, then check whether the sequence of posterior distributions of f accumulates in some suitable neighborhood of f 0. Which conditions on f 0 and µ are needed under the four kernels considered in order to achieve consistency? CLTs for functionals of the posterior hazard. For a fixed T > 0, we consider random objects like (i) H(T ), (ii) T 1 T 0 h(t) 2 dt (path-second moment) and (iii) T 1 T 0 [ h(t) H(T )/T ] 2 dt (path-variance). Let Y = (Y 1,..., Y n ) be a set of observations. We are interested in establishing Central Limit Theorems of the type η(t ) [ H(T ) τ(t )] N(0, σ 2 ) as T + Y law for appropriate positive functions τ(t ) and η(t ) and variance σ 2. How are τ( ), η( ) and σ 2 influenced by the observed data?

8 Weak consistency Denote by P 0 the probability distribution associated with f 0 and by P 0 the infinite product measure. Also, let Π be the prior distribution of the random density function f and Π n the posterior distribution. Definition. If, for any ɛ > 0 and as n, Π n (A ɛ (f 0 )) 1 a.s. P 0, where A ɛ (f 0 ) is a ɛ neighborhood of f 0 in the weak topology, then Π is said to be weak consistent at f 0. Remark. A sufficient condition for weak consistency requires a prior Π to assign positive probability to Kullback Leibler (K-L) neighborhoods of f 0 ( K-L condition): Π ( f F : log(f 0 /f )f 0 < ɛ ) > 0 for any ɛ > 0, where log(f 0 /f )f 0 denotes the K-L divergence between f 0 and f, whereas F is the space of density functions defined on R +.

9 General consistency result Assume X = R + and postulate the existence of a boundedly finite measure µ 0 on R + such that the true density f 0 has hazard h 0 (t) = R + k(t, x)µ 0 (dx) where k is the kernel used for constructing f via the mixture hazard h. We translate the K-L condition w.r.t. f 0 into a condition of positive prior probability assigned to uniform neighborhoods of h 0 on [0, T ]. THEOREM Weak consistency Assume (i) h 0 (t) > 0 for any t 0, (ii) R + g(t) f 0 (t)dt <, where g(t) = max{e[ H(t)], t}. Then, a sufficient condition for Π to be weakly consistent at f 0 is that Π {h : sup 0<t T h(t) h 0 (t) } < δ > 0 for any finite T and positive δ. Note that condition (ii) is related to the asymptotic characterization of functionals of the random hazard rate (see later).

10 Relaxing the condition h 0 (0) > 0 For (DL) and (OU) kernels one may actually have h 0 (0) = 0. Problems then arise in establish the K-L condition since log(h 0 / h) (hence log(f 0 / f )) can become arbitrarily large around 0. In order to cover also the case h 0 (0) = 0, we resort to the short time behaviour of µ, which determines the way h vanishes in 0 for the (DL) and (OU) cases. PROPOSITION 1 Short time behaviour Weak consistency holds for (DL) and (OU) kernels also with h 0 (0) = 0 provided there exist α, r > 0 such that: (a) lim t 0 h 0 (t)/t α = 0; (b) lim inf t 0 µ((0, t])/t r = a.s. In particular (b) holds if µ is a CRM belonging to the generalized gamma family with σ (0, 1) and λ(dx) = dx. Note that the powers α and r do not need to satisfy any relation.

11 Uniform balls of the true hazard rate In order to derive explicit consistency results for the mixture hazard models based on the four kernels considered, we need to establish: { Π h : sup 0<t T h(t) h 0 (t) } < δ > 0. To this aim, the following result is essential. Denote by G the space of nondecreasing càdlàg functions on R + such that G(0) = 0 for any G G. Note that any CRM induces a distribution on G and that, for G 0 (x) = µ 0 ( [0, x] ), G0 G. LEMMA 1 Support of CRMs Let µ be a CRM on R +, satisfying ν(r +, dx) = for any x X, and denote by Q the distribution induced on G. Then, for any G 0 G, any finite M and η > 0, { } Q G G : sup x M G(x) G 0 (x) < η > 0. Different techniques are required for the (exp) kernel w.r.t. the other three kernels (different support of k(t, x) seen as a function of x).

12 PROPOSITION 2 Weak consistency for (DL), (rect) and (OU) Let µ be a generalized gamma CRM with σ (0, 1) and λ(dx) = dx. Then weak consistency holds for the (DL), (rect) and (OU) kernels provided that: (i) h 0 (t) > 0 for any t > 0; (ii) t i f 0 (t)dt < with i = 1 for (rect) and (OU), i = 2 for (DL). Note that Condition (i) is automatically satisfied if µ 0 is absolutely continuous. PROPOSITION 3 Weak consistency for (exp) Weak consistency holds for the (exp) kernel provided that: (i) h(0) < a.s. and h 0 (0) < ; (ii) t f 0 (t)dt <. The only relevant assumption consists in asking for both h and h 0 not to explode in 0. Note that h 0 (t) > 0 is automatically satisfied.

13 CLTs for linear and quadratic functionals CLTs provide a synthetic picture of the random hazard, like information about the trend, oscillation and overall variance for large time horizons. In Peccati and Pruenster (2006) functionals of the prior hazard rate are considered. Such results can serve as a guide for prior specification. Here we investigate the asymptotic behaviour of functionals of the posterior random hazard rate given a fixed number of observations. We found out that, in all the considered special cases, the CLTs associated with the posterior hazard rate are the same as for the prior ones, and this for any number of observations. Although consistency implies that a given model can be asymptotically directed towards any deterministic target, the overall structure of a posterior hazard rate is systematically determined by the prior choice.

14 Posterior distribution We exploit an explicit posterior representation of the random mixture hazard model of James (2005). Let P be the random probability measure associated with f (t) and let (Y n ) n 1 be a sequence of exchangeable observations from P. Set Y = (Y 1,..., Y n ). Define a random probability measure on X R + as P(dx, dt) = e R t R 0 X k(s,y) µ(dy) ds k(t, x) µ(dx) dt. By introducing the latent variables X = (X 1,..., X n ), we describe µ Y in terms of µ Y, X mixed over X Y. Since the X i s may feature ties, we denote by X = (X1,..., X k ) the k n distinct latent variables. Let n j be the frequency of Xj and C j = {r : x r = xj }. The likelihood function is then given by L (µ; y, x) = e R P n R yi X i=1 0 k(t,x)dtµ(dx) k j=1 µ(dx j ) n j i C j k(y i, xj )

15 Given X and Y, the conditional distribution of µ Y, X coincides with the distribution of the random measure µ n d = µ n + k i=1 J i δ X j = µ n + n µ n is a CRM with intensity measure ν n (dv, dx) := e v P R n yi i=1 0 k(t,x)dt ρ(dv x) λ(dx) n := k i=1 J i δ X j with, for i = 1,..., k, X i a fixed points of discontinuity with corresponding jump J i which are, conditionally on X and Y, independent of µ n. The conditional distribution X Y, say P X Y, can be expressed as the conditional distributions of X p, Y and p Y, where p is a partitions of the integers {1,..., n}. We derive CLTs for the posterior random hazard given X and Y : h n(t) = X k(t, x) µn (dx) + k i=1 J ik(t, X i ), and, then, average over the distribution of X Y.

16 Limit theorems for shifted measures Peccati and Pruenster (2006) provide sufficient conditions to have that linear and quadratic functionals associated with a random mixture hazard without fixed atoms verify a CLT. Linear functional. For given X and Y, define H n(t ) = T 0 h n(t)dt = H n (T ) + k i=1 J i T 0 k(t, X i )dt Assume there exists a (deterministic) function T C n (k, T ) and σn(k) 2 > 0 such that, as T +, Cn(k, 2 T ) [ s ] T 2 R + X 0 k(t, x)dt ν n (ds, dx) σn(k), 2 Cn(k, 3 T ) [ s ] T 3 R + X 0 k(t, x)dt ν n (ds, dx) 0, Then C n (k, T ) [ Hn (T ) E[ H ] Y n law (T )] N(0, σn(k)), 2 where σ 2 n(k) may depend on Y.

17 Assume moreover that lim C n(k, T ) k j=1 J j T + T 0 k(t, X j )dt = m n ( n, k), (1) for some quantity m n ( n, k) 0 that may depend on X. Then C n (k, T ) [ H n(t ) E[ H ] n Y law (T )], X N ( m n ( n, k), σn(k) 2 ). If (1) holds for X = (x 1,..., x n ) X n in a set of P X Y probability 1, then we can state the result as η(t ) [ ] H(T law ) τ(t ) Y Z X,Y d P X Y, X where n η(t ) = C n (k, T ); τ(t ) = E[ H n (T ) Y ]; Z X,Y is a Gaussian random variable with mean m n ( n, k) and variance σ n (k). The mixture in X n Z X,Y d P X Y is over the quantity m n ( n, k) that may change according with the realization of X.

18 Applications We derive CLTs for functionals of posterior hazards based on the four kernels combined with generalized gamma CRMs. In general m n ( n, k) = 0, not depending on Y, X, that is the part of the posterior with fixed points of discontinuity affects the asymptotic behaviour just in the limit. Then, the convergence is simply expressed in terms of a Gaussian random variable. LINEAR CLT 1 (OU) kernel [ ] T 1 2 H(T ) 2 κ γ (1 σ) law T Y N(0, 2γ (1 σ) ) LINEAR CLT 2 (DL) kernel [ T 3 2 H(T ) (2γ (1 σ) ) 1 T 2] ( ) law Y 1 σ N 0, 3γ 2 σ Most important: C n (k, T ) = C 0 (k, T ), E[ H(T )] E[ H(T ) Y ] and σ n (k) = σ 0 (k), that is we get the same trend, oscillation around the trend and overall variance we get with the prior cumulative hazard.

19 One might think that the lack of dependence on the data of posterior functionals may be due to the boundedness of the support of the (DL), (OU) and (rect) kernels, as it happens that ν n = ν in R + [Y (n), ), where Y (n) is the largest observed lifetime. Consider then the (exp) kernel. In particular, we set λ(dx)=x 1 2 e 1 x (2 π) 1 in the Poisson intensity ν(dv, dx) of generalized gamma CRM, which gives us h(0) < a.s. and a prior mean centered on a quasi Weibull hazard, E[ h(t)] = γ 1+σ / t + 1. LINEAR CLT 3 (exp) kernel [ T 1 2 H(T ) (γ 1+σ ) 1 T 1/2] ( ) law Y N 0, (2 2)(1 σ) γ 2 σ Again, the oscillation rate, the trend and the overall variance coincides with the prior ones, that is they do not depend on the observation Y.

20 Concluding remarks We provide a comprehensive investigation of posterior weak consistency of the mixture hazard model. L 1 consistency would be the natural successive target. The investigation of CLTs for functionals represents a completely new line of research and has interest in its own. The coincidence of asymptotic behaviour of the posterior and the prior hazard means that the data do not influence the behaviour of the model for times larger than the largest observed lifetime Y (n) ; and, this, despite the fact that these models are consistent. The fact that the overall variance is not influenced by the data is somehow counterintuitive: since the contribution of the CRM vanishes in the limit, one would expect the variance to become smaller and smaller as more data come in. Since this does not happen, our findings show a clear evidence that the choice of the CRM really matters whatever the size of the dataset and provide a guideline in incorporating prior knowledge appropriately into the model.

21 Appendix For Further Reading Brix, A. (1999) Generalized gamma measures and shot-noise Cox processes. Adv. Appl. Prob. 31, Dykstra, R.L. and Laud, P. (1981) A Bayesian nonparametric approach to reliability. Ann. Statist. 9, James, L.F. (2005) Bayesian Poisson process partition calculus with an application to Bayesian Lévy moving averages. Ann. Statist. 33, Lo, A.Y. and Weng, C.S. (1989) On a class of Bayesian nonparametric estimates. II. Hazard rate estimates. Ann. Inst. Statist. Math. 41, Peccati, G. and Prünster, I. (2006) Linear and quadratic functionals of random hazard rates: an asymptotic analysis. Preprint.

Asymptotics for posterior hazards

Asymptotics for posterior hazards Asymptotics for posterior hazards Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with P. Di Biasi and G. Peccati Workshop on Limit Theorems and Applications Paris, 16th January

More information

On the posterior structure of NRMI

On the posterior structure of NRMI On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline

More information

Truncation error of a superposed gamma process in a decreasing order representation

Truncation error of a superposed gamma process in a decreasing order representation Truncation error of a superposed gamma process in a decreasing order representation Julyan Arbel Inria Grenoble, Université Grenoble Alpes julyan.arbel@inria.fr Igor Prünster Bocconi University, Milan

More information

Bayesian semiparametric analysis of short- and long- term hazard ratios with covariates

Bayesian semiparametric analysis of short- and long- term hazard ratios with covariates Bayesian semiparametric analysis of short- and long- term hazard ratios with covariates Department of Statistics, ITAM, Mexico 9th Conference on Bayesian nonparametrics Amsterdam, June 10-14, 2013 Contents

More information

Truncation error of a superposed gamma process in a decreasing order representation

Truncation error of a superposed gamma process in a decreasing order representation Truncation error of a superposed gamma process in a decreasing order representation B julyan.arbel@inria.fr Í www.julyanarbel.com Inria, Mistis, Grenoble, France Joint work with Igor Pru nster (Bocconi

More information

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals

Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico

More information

Bayesian Nonparametrics: some contributions to construction and properties of prior distributions

Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Annalisa Cerquetti Collegio Nuovo, University of Pavia, Italy Interview Day, CETL Lectureship in Statistics,

More information

On the Truncation Error of a Superposed Gamma Process

On the Truncation Error of a Superposed Gamma Process On the Truncation Error of a Superposed Gamma Process Julyan Arbel and Igor Prünster Abstract Completely random measures (CRMs) form a key ingredient of a wealth of stochastic models, in particular in

More information

Dependent hierarchical processes for multi armed bandits

Dependent hierarchical processes for multi armed bandits Dependent hierarchical processes for multi armed bandits Federico Camerlenghi University of Bologna, BIDSA & Collegio Carlo Alberto First Italian meeting on Probability and Mathematical Statistics, Torino

More information

Poisson random measure: motivation

Poisson random measure: motivation : motivation The Lévy measure provides the expected number of jumps by time unit, i.e. in a time interval of the form: [t, t + 1], and of a certain size Example: ν([1, )) is the expected number of jumps

More information

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS

PROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

arxiv: v2 [math.st] 27 May 2014

arxiv: v2 [math.st] 27 May 2014 Full Bayesian inference with hazard mixture models Julyan Arbel a, Antonio Lijoi b,a, Bernardo Nipoti c,a, arxiv:145.6628v2 [math.st] 27 May 214 a Collegio Carlo Alberto, via Real Collegio, 3, 124 Moncalieri,

More information

Normalized kernel-weighted random measures

Normalized kernel-weighted random measures Normalized kernel-weighted random measures Jim Griffin University of Kent 1 August 27 Outline 1 Introduction 2 Ornstein-Uhlenbeck DP 3 Generalisations Bayesian Density Regression We observe data (x 1,

More information

STAT Sample Problem: General Asymptotic Results

STAT Sample Problem: General Asymptotic Results STAT331 1-Sample Problem: General Asymptotic Results In this unit we will consider the 1-sample problem and prove the consistency and asymptotic normality of the Nelson-Aalen estimator of the cumulative

More information

Bayesian nonparametric models for bipartite graphs

Bayesian nonparametric models for bipartite graphs Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers

More information

Math 341: Probability Seventeenth Lecture (11/10/09)

Math 341: Probability Seventeenth Lecture (11/10/09) Math 341: Probability Seventeenth Lecture (11/10/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams

More information

Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting

Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting Anne Philippe Laboratoire de Mathématiques Jean Leray Université de Nantes Workshop EDF-INRIA,

More information

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition

Filtrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,

More information

Step-Stress Models and Associated Inference

Step-Stress Models and Associated Inference Department of Mathematics & Statistics Indian Institute of Technology Kanpur August 19, 2014 Outline Accelerated Life Test 1 Accelerated Life Test 2 3 4 5 6 7 Outline Accelerated Life Test 1 Accelerated

More information

Full Bayesian inference with hazard mixture models

Full Bayesian inference with hazard mixture models ISSN 2279-9362 Full Bayesian inference with hazard mixture models Julyan Arbel Antonio Lijoi Bernardo Nipoti No. 381 December 214 www.carloalberto.org/research/working-papers 214 by Julyan Arbel, Antonio

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x

More information

Introduction to self-similar growth-fragmentations

Introduction to self-similar growth-fragmentations Introduction to self-similar growth-fragmentations Quan Shi CIMAT, 11-15 December, 2017 Quan Shi Growth-Fragmentations CIMAT, 11-15 December, 2017 1 / 34 Literature Jean Bertoin, Compensated fragmentation

More information

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t )

Lecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t ) LECURE NOES 21 Lecture 4 7. Sufficient statistics Consider the usual statistical setup: the data is X and the paramter is. o gain information about the parameter we study various functions of the data

More information

NEW FUNCTIONAL INEQUALITIES

NEW FUNCTIONAL INEQUALITIES 1 / 29 NEW FUNCTIONAL INEQUALITIES VIA STEIN S METHOD Giovanni Peccati (Luxembourg University) IMA, Minneapolis: April 28, 2015 2 / 29 INTRODUCTION Based on two joint works: (1) Nourdin, Peccati and Swan

More information

On prediction and density estimation Peter McCullagh University of Chicago December 2004

On prediction and density estimation Peter McCullagh University of Chicago December 2004 On prediction and density estimation Peter McCullagh University of Chicago December 2004 Summary Having observed the initial segment of a random sequence, subsequent values may be predicted by calculating

More information

On pathwise stochastic integration

On pathwise stochastic integration On pathwise stochastic integration Rafa l Marcin Lochowski Afican Institute for Mathematical Sciences, Warsaw School of Economics UWC seminar Rafa l Marcin Lochowski (AIMS, WSE) On pathwise stochastic

More information

Regular Variation and Extreme Events for Stochastic Processes

Regular Variation and Extreme Events for Stochastic Processes 1 Regular Variation and Extreme Events for Stochastic Processes FILIP LINDSKOG Royal Institute of Technology, Stockholm 2005 based on joint work with Henrik Hult www.math.kth.se/ lindskog 2 Extremes for

More information

Lecture Characterization of Infinitely Divisible Distributions

Lecture Characterization of Infinitely Divisible Distributions Lecture 10 1 Characterization of Infinitely Divisible Distributions We have shown that a distribution µ is infinitely divisible if and only if it is the weak limit of S n := X n,1 + + X n,n for a uniformly

More information

Bayesian Modeling of Conditional Distributions

Bayesian Modeling of Conditional Distributions Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings

More information

Convergence of Feller Processes

Convergence of Feller Processes Chapter 15 Convergence of Feller Processes This chapter looks at the convergence of sequences of Feller processes to a iting process. Section 15.1 lays some ground work concerning weak convergence of processes

More information

Densities for the Navier Stokes equations with noise

Densities for the Navier Stokes equations with noise Densities for the Navier Stokes equations with noise Marco Romito Università di Pisa Universitat de Barcelona March 25, 2015 Summary 1 Introduction & motivations 2 Malliavin calculus 3 Besov bounds 4 Other

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Asymptotic properties of maximum likelihood estimator for the growth rate for a jump-type CIR process

Asymptotic properties of maximum likelihood estimator for the growth rate for a jump-type CIR process Asymptotic properties of maximum likelihood estimator for the growth rate for a jump-type CIR process Mátyás Barczy University of Debrecen, Hungary The 3 rd Workshop on Branching Processes and Related

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012

Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012 Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM

More information

Two viewpoints on measure valued processes

Two viewpoints on measure valued processes Two viewpoints on measure valued processes Olivier Hénard Université Paris-Est, Cermics Contents 1 The classical framework : from no particle to one particle 2 The lookdown framework : many particles.

More information

Statistical inference on Lévy processes

Statistical inference on Lévy processes Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline

More information

Appendix A. Sequences and series. A.1 Sequences. Definition A.1 A sequence is a function N R.

Appendix A. Sequences and series. A.1 Sequences. Definition A.1 A sequence is a function N R. Appendix A Sequences and series This course has for prerequisite a course (or two) of calculus. The purpose of this appendix is to review basic definitions and facts concerning sequences and series, which

More information

Probability and Measure

Probability and Measure Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability

More information

Gaussian processes for inference in stochastic differential equations

Gaussian processes for inference in stochastic differential equations Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017

More information

S1. Proofs of the Key Properties of CIATA-Ph

S1. Proofs of the Key Properties of CIATA-Ph Article submitted to INFORMS Journal on Computing; manuscript no. (Please, provide the mansucript number!) S-1 Online Supplement to Modeling and Simulation of Nonstationary and Non-Poisson Processes by

More information

New Dirichlet Mean Identities

New Dirichlet Mean Identities Hong Kong University of Science and Technology Isaac Newton Institute, August 10, 2007 Origins CIFARELLI, D. M. and REGAZZINI, E. (1979). Considerazioni generali sull impostazione bayesiana di problemi

More information

Foundations of Nonparametric Bayesian Methods

Foundations of Nonparametric Bayesian Methods 1 / 27 Foundations of Nonparametric Bayesian Methods Part II: Models on the Simplex Peter Orbanz http://mlg.eng.cam.ac.uk/porbanz/npb-tutorial.html 2 / 27 Tutorial Overview Part I: Basics Part II: Models

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

Translation Invariant Experiments with Independent Increments

Translation Invariant Experiments with Independent Increments Translation Invariant Statistical Experiments with Independent Increments (joint work with Nino Kordzakhia and Alex Novikov Steklov Mathematical Institute St.Petersburg, June 10, 2013 Outline 1 Introduction

More information

The information complexity of sequential resource allocation

The information complexity of sequential resource allocation The information complexity of sequential resource allocation Emilie Kaufmann, joint work with Olivier Cappé, Aurélien Garivier and Shivaram Kalyanakrishan SMILE Seminar, ENS, June 8th, 205 Sequential allocation

More information

Consistency of the maximum likelihood estimator for general hidden Markov models

Consistency of the maximum likelihood estimator for general hidden Markov models Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models

More information

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics CS 540: Machine Learning Lecture 2: Review of Probability & Statistics AD January 2008 AD () January 2008 1 / 35 Outline Probability theory (PRML, Section 1.2) Statistics (PRML, Sections 2.1-2.4) AD ()

More information

GARCH processes continuous counterparts (Part 2)

GARCH processes continuous counterparts (Part 2) GARCH processes continuous counterparts (Part 2) Alexander Lindner Centre of Mathematical Sciences Technical University of Munich D 85747 Garching Germany lindner@ma.tum.de http://www-m1.ma.tum.de/m4/pers/lindner/

More information

Order book resilience, price manipulation, and the positive portfolio problem

Order book resilience, price manipulation, and the positive portfolio problem Order book resilience, price manipulation, and the positive portfolio problem Alexander Schied Mannheim University Workshop on New Directions in Financial Mathematics Institute for Pure and Applied Mathematics,

More information

1. Stochastic Processes and filtrations

1. Stochastic Processes and filtrations 1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S

More information

MATH 5640: Fourier Series

MATH 5640: Fourier Series MATH 564: Fourier Series Hung Phan, UMass Lowell September, 8 Power Series A power series in the variable x is a series of the form a + a x + a x + = where the coefficients a, a,... are real or complex

More information

Math 430/530: Advanced Calculus I Chapter 1: Sequences

Math 430/530: Advanced Calculus I Chapter 1: Sequences Math 430/530: Advanced Calculus I Chapter 1: Sequences Fall 2018 Sequences A sequence is just a list of real numbers a 1, a 2,... with a definite order. We write a = (a n ) n=1 for a sequence a with n

More information

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007

More information

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012

Gaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012 Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature

More information

Random Process Lecture 1. Fundamentals of Probability

Random Process Lecture 1. Fundamentals of Probability Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus

More information

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem

Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem

More information

Dependent mixture models: clustering and borrowing information

Dependent mixture models: clustering and borrowing information ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,

More information

Frontier estimation based on extreme risk measures

Frontier estimation based on extreme risk measures Frontier estimation based on extreme risk measures by Jonathan EL METHNI in collaboration with Ste phane GIRARD & Laurent GARDES CMStatistics 2016 University of Seville December 2016 1 Risk measures 2

More information

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop

These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop Music and Machine Learning (IFT68 Winter 8) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop

More information

Nonparametric regression with martingale increment errors

Nonparametric regression with martingale increment errors S. Gaïffas (LSTA - Paris 6) joint work with S. Delattre (LPMA - Paris 7) work in progress Motivations Some facts: Theoretical study of statistical algorithms requires stationary and ergodicity. Concentration

More information

Evgeny Spodarev WIAS, Berlin. Limit theorems for excursion sets of stationary random fields

Evgeny Spodarev WIAS, Berlin. Limit theorems for excursion sets of stationary random fields Evgeny Spodarev 23.01.2013 WIAS, Berlin Limit theorems for excursion sets of stationary random fields page 2 LT for excursion sets of stationary random fields Overview 23.01.2013 Overview Motivation Excursion

More information

Discussion of On simulation and properties of the stable law by L. Devroye and L. James

Discussion of On simulation and properties of the stable law by L. Devroye and L. James Stat Methods Appl (2014) 23:371 377 DOI 10.1007/s10260-014-0269-4 Discussion of On simulation and properties of the stable law by L. Devroye and L. James Antonio Lijoi Igor Prünster Accepted: 16 May 2014

More information

The strictly 1/2-stable example

The strictly 1/2-stable example The strictly 1/2-stable example 1 Direct approach: building a Lévy pure jump process on R Bert Fristedt provided key mathematical facts for this example. A pure jump Lévy process X is a Lévy process such

More information

Harnack Inequalities and Applications for Stochastic Equations

Harnack Inequalities and Applications for Stochastic Equations p. 1/32 Harnack Inequalities and Applications for Stochastic Equations PhD Thesis Defense Shun-Xiang Ouyang Under the Supervision of Prof. Michael Röckner & Prof. Feng-Yu Wang March 6, 29 p. 2/32 Outline

More information

Estimation of the long Memory parameter using an Infinite Source Poisson model applied to transmission rate measurements

Estimation of the long Memory parameter using an Infinite Source Poisson model applied to transmission rate measurements of the long Memory parameter using an Infinite Source Poisson model applied to transmission rate measurements François Roueff Ecole Nat. Sup. des Télécommunications 46 rue Barrault, 75634 Paris cedex 13,

More information

Slice sampling σ stable Poisson Kingman mixture models

Slice sampling σ stable Poisson Kingman mixture models ISSN 2279-9362 Slice sampling σ stable Poisson Kingman mixture models Stefano Favaro S.G. Walker No. 324 December 2013 www.carloalberto.org/research/working-papers 2013 by Stefano Favaro and S.G. Walker.

More information

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.

More information

Weak quenched limiting distributions of a one-dimensional random walk in a random environment

Weak quenched limiting distributions of a one-dimensional random walk in a random environment Weak quenched limiting distributions of a one-dimensional random walk in a random environment Jonathon Peterson Cornell University Department of Mathematics Joint work with Gennady Samorodnitsky September

More information

The parabolic Anderson model on Z d with time-dependent potential: Frank s works

The parabolic Anderson model on Z d with time-dependent potential: Frank s works Weierstrass Institute for Applied Analysis and Stochastics The parabolic Anderson model on Z d with time-dependent potential: Frank s works Based on Frank s works 2006 2016 jointly with Dirk Erhard (Warwick),

More information

ON ADDITIVE TIME-CHANGES OF FELLER PROCESSES. 1. Introduction

ON ADDITIVE TIME-CHANGES OF FELLER PROCESSES. 1. Introduction ON ADDITIVE TIME-CHANGES OF FELLER PROCESSES ALEKSANDAR MIJATOVIĆ AND MARTIJN PISTORIUS Abstract. In this note we generalise the Phillips theorem [1] on the subordination of Feller processes by Lévy subordinators

More information

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector

Mathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean

More information

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:

More information

Cumulants of a convolution and applications to monotone probability theory

Cumulants of a convolution and applications to monotone probability theory Cumulants of a convolution and applications to monotone probability theory arxiv:95.3446v2 [math.pr] 24 May 29 Takahiro Hasebe JSPS Research Fellow (DC), Graduate School of Science, Kyoto University, Kyoto

More information

Temporal point processes: the conditional intensity function

Temporal point processes: the conditional intensity function Temporal point processes: the conditional intensity function Jakob Gulddahl Rasmussen December 21, 2009 Contents 1 Introduction 2 2 Evolutionary point processes 2 2.1 Evolutionarity..............................

More information

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3

Brownian Motion. 1 Definition Brownian Motion Wiener measure... 3 Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................

More information

University of Toronto Department of Statistics

University of Toronto Department of Statistics Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704

More information

On some distributional properties of Gibbs-type priors

On some distributional properties of Gibbs-type priors On some distributional properties of Gibbs-type priors Igor Prünster University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics Workshop ICERM, 21st September 2012 Joint work with: P. De Blasi,

More information

Statistics (1): Estimation

Statistics (1): Estimation Statistics (1): Estimation Marco Banterlé, Christian Robert and Judith Rousseau Practicals 2014-2015 L3, MIDO, Université Paris Dauphine 1 Table des matières 1 Random variables, probability, expectation

More information

On oscillating systems of interacting neurons

On oscillating systems of interacting neurons Susanne Ditlevsen Eva Löcherbach Nice, September 2016 Oscillations Most biological systems are characterized by an intrinsic periodicity : 24h for the day-night rhythm of most animals, etc... In other

More information

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.

In terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3. 1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

Bayesian Nonparametrics

Bayesian Nonparametrics Bayesian Nonparametrics Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent

More information

P-adic Functions - Part 1

P-adic Functions - Part 1 P-adic Functions - Part 1 Nicolae Ciocan 22.11.2011 1 Locally constant functions Motivation: Another big difference between p-adic analysis and real analysis is the existence of nontrivial locally constant

More information

Information geometry for bivariate distribution control

Information geometry for bivariate distribution control Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic

More information

Nonparametric estimation of log-concave densities

Nonparametric estimation of log-concave densities Nonparametric estimation of log-concave densities Jon A. Wellner University of Washington, Seattle Seminaire, Institut de Mathématiques de Toulouse 5 March 2012 Seminaire, Toulouse Based on joint work

More information

The information complexity of best-arm identification

The information complexity of best-arm identification The information complexity of best-arm identification Emilie Kaufmann, joint work with Olivier Cappé and Aurélien Garivier MAB workshop, Lancaster, January th, 206 Context: the multi-armed bandit model

More information

Understanding Regressions with Observations Collected at High Frequency over Long Span

Understanding Regressions with Observations Collected at High Frequency over Long Span Understanding Regressions with Observations Collected at High Frequency over Long Span Yoosoon Chang Department of Economics, Indiana University Joon Y. Park Department of Economics, Indiana University

More information

STOCHASTIC GEOMETRY BIOIMAGING

STOCHASTIC GEOMETRY BIOIMAGING CENTRE FOR STOCHASTIC GEOMETRY AND ADVANCED BIOIMAGING 2018 www.csgb.dk RESEARCH REPORT Anders Rønn-Nielsen and Eva B. Vedel Jensen Central limit theorem for mean and variogram estimators in Lévy based

More information

LAN property for ergodic jump-diffusion processes with discrete observations

LAN property for ergodic jump-diffusion processes with discrete observations LAN property for ergodic jump-diffusion processes with discrete observations Eulalia Nualart (Universitat Pompeu Fabra, Barcelona) joint work with Arturo Kohatsu-Higa (Ritsumeikan University, Japan) &

More information

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )

Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( ) Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical

More information

Log Gaussian Cox Processes. Chi Group Meeting February 23, 2016

Log Gaussian Cox Processes. Chi Group Meeting February 23, 2016 Log Gaussian Cox Processes Chi Group Meeting February 23, 2016 Outline Typical motivating application Introduction to LGCP model Brief overview of inference Applications in my work just getting started

More information

Large Deviations for Small-Noise Stochastic Differential Equations

Large Deviations for Small-Noise Stochastic Differential Equations Chapter 22 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large

More information

The small ball property in Banach spaces (quantitative results)

The small ball property in Banach spaces (quantitative results) The small ball property in Banach spaces (quantitative results) Ehrhard Behrends Abstract A metric space (M, d) is said to have the small ball property (sbp) if for every ε 0 > 0 there exists a sequence

More information

Malliavin calculus and central limit theorems

Malliavin calculus and central limit theorems Malliavin calculus and central limit theorems David Nualart Department of Mathematics Kansas University Seminar on Stochastic Processes 2017 University of Virginia March 8-11 2017 David Nualart (Kansas

More information

Large Deviations for Small-Noise Stochastic Differential Equations

Large Deviations for Small-Noise Stochastic Differential Equations Chapter 21 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large

More information

Advanced Calculus Math 127B, Winter 2005 Solutions: Final. nx2 1 + n 2 x, g n(x) = n2 x

Advanced Calculus Math 127B, Winter 2005 Solutions: Final. nx2 1 + n 2 x, g n(x) = n2 x . Define f n, g n : [, ] R by f n (x) = Advanced Calculus Math 27B, Winter 25 Solutions: Final nx2 + n 2 x, g n(x) = n2 x 2 + n 2 x. 2 Show that the sequences (f n ), (g n ) converge pointwise on [, ],

More information

Maximum likelihood estimation of a log-concave density based on censored data

Maximum likelihood estimation of a log-concave density based on censored data Maximum likelihood estimation of a log-concave density based on censored data Dominic Schuhmacher Institute of Mathematical Statistics and Actuarial Science University of Bern Joint work with Lutz Dümbgen

More information