Asymptotics for posterior hazards
|
|
- Julie Craig
- 5 years ago
- Views:
Transcription
1 Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and Igor Prünster (University of Turin)
2 Outline Life-testing model with mixture hazard rate Mixture hazard rate Kernels and completely random measures Asymptotic issues Posterior consistency Weak consistency Sufficient condition for Kullback-Leibler support Prior probability of uniform balls of the true hazard rate CLTs for functionals of the posterior hazard rate Posterior distribution Limit theorems for shifted measures Applications
3 Mixture hazard rate Let Y be a positive absolutely continuous random variable (lifetime) with random hazard rate of the form h(t) = k(t, x) µ(dx), X k is a kernel, jointly measurable application R + X R +, µ is a random mixing measure on the space X, modeled as a completely random measure. The cumulative hazard is given by H(t) = t h(s)ds and is required to 0 satisfy H(t) for t a.s. Then f (t) = h(t) exp( H(t)) defines a random density function to be used in Bayesian nonparametric inference. See Dykstra and Laud (1981) and Lo and Weng (1989).
4 Kernels As examples,we consider the class of mixture hazard rates that arise under the following choice of kernels: Dykstra-Laud (DL) kernel (monotone increasing hazard rates) k(t, x) = I (0 x t) rectangular (rect) kernel with bandwidth τ > 0 k(t, x) = I ( t x τ) Ornstein-Uhlenbeck (OU) kernel with κ > 0 k(t, x) = 2κ e κ(t x) I (0 x t) exponential (exp) kernel (monotone decreasing hazard rates) k(t, x) = x 1 e t x.
5 Completely random measures Let M be the space of boundedly finite measures on X. Definition. A random element µ taking values in M such that (i) µ( ) = 0 (ii) for any disjoint sets (A i ) i 1 : µ(a 1 ), µ(a 2 ),... are mutually independent µ( j 1 A j ) = j 1 µ(a j) is said to be a completely random measure (CRM) on X. Remark. A CRM can always be represented as a linear functional of a Poisson random measure. In particular, µ is uniquely characterized by its Laplace functional [ E e R X g(x) µ(dx)] R = e R + X [1 e v g(x) ]ν(dv,dx) where ν stands for the intensity of the Poisson random measure underlying µ.
6 CRMs select almost surely discrete measures: hence, µ can always be represented as i 1 J iδ Xi. Note that if ν(dv, dx) = ρ(dv)λ(dx) the law of the J i s and the X i s are independent. In this case µ is termed homogeneous. A non-homogeneous CRM has ν(dv, dx) = ρ(dv x)λ(dx). Non-homogeneous CRMs arise in the posterior. If ν(r +, dx) = for any x, then the CRM jumps infinitely often on any bounded set A. However, recall that µ(a) < a.s. We consider CRMs with Poisson intensity of the following form: ν(dv, dx) = 1 Γ(1 σ) e γ(x)v dv λ(dx), σ [0, 1), γ( ) > 0. v 1+σ γ constant generalized gamma measure (Brix, 1999) σ = 0 weighted gamma measure (Lo & Weng, 1989).
7 Asymptotic issues Posterior consistency. First generate independent data from a true fixed density f 0, then check whether the sequence of posterior distributions of f accumulates in some suitable neighborhood of f 0. Which conditions on f 0 and µ are needed under the four kernels considered in order to achieve consistency? CLTs for functionals of the posterior hazard. For a fixed T > 0, we consider random objects like (i) H(T ), (ii) T 1 T 0 h(t) 2 dt (path-second moment) and (iii) T 1 T 0 [ h(t) H(T )/T ] 2 dt (path-variance). Let Y = (Y 1,..., Y n ) be a set of observations. We are interested in establishing Central Limit Theorems of the type η(t ) [ H(T ) τ(t )] N(0, σ 2 ) as T + Y law for appropriate positive functions τ(t ) and η(t ) and variance σ 2. How are τ( ), η( ) and σ 2 influenced by the observed data?
8 Weak consistency Denote by P 0 the probability distribution associated with f 0 and by P 0 the infinite product measure. Also, let Π be the prior distribution of the random density function f and Π n the posterior distribution. Definition. If, for any ɛ > 0 and as n, Π n (A ɛ (f 0 )) 1 a.s. P 0, where A ɛ (f 0 ) is a ɛ neighborhood of f 0 in the weak topology, then Π is said to be weak consistent at f 0. Remark. A sufficient condition for weak consistency requires a prior Π to assign positive probability to Kullback Leibler (K-L) neighborhoods of f 0 ( K-L condition): Π ( f F : log(f 0 /f )f 0 < ɛ ) > 0 for any ɛ > 0, where log(f 0 /f )f 0 denotes the K-L divergence between f 0 and f, whereas F is the space of density functions defined on R +.
9 General consistency result Assume X = R + and postulate the existence of a boundedly finite measure µ 0 on R + such that the true density f 0 has hazard h 0 (t) = R + k(t, x)µ 0 (dx) where k is the kernel used for constructing f via the mixture hazard h. We translate the K-L condition w.r.t. f 0 into a condition of positive prior probability assigned to uniform neighborhoods of h 0 on [0, T ]. THEOREM Weak consistency Assume (i) h 0 (t) > 0 for any t 0, (ii) R + g(t) f 0 (t)dt <, where g(t) = max{e[ H(t)], t}. Then, a sufficient condition for Π to be weakly consistent at f 0 is that Π {h : sup 0<t T h(t) h 0 (t) } < δ > 0 for any finite T and positive δ. Note that condition (ii) is related to the asymptotic characterization of functionals of the random hazard rate (see later).
10 Relaxing the condition h 0 (0) > 0 For (DL) and (OU) kernels one may actually have h 0 (0) = 0. Problems then arise in establish the K-L condition since log(h 0 / h) (hence log(f 0 / f )) can become arbitrarily large around 0. In order to cover also the case h 0 (0) = 0, we resort to the short time behaviour of µ, which determines the way h vanishes in 0 for the (DL) and (OU) cases. PROPOSITION 1 Short time behaviour Weak consistency holds for (DL) and (OU) kernels also with h 0 (0) = 0 provided there exist α, r > 0 such that: (a) lim t 0 h 0 (t)/t α = 0; (b) lim inf t 0 µ((0, t])/t r = a.s. In particular (b) holds if µ is a CRM belonging to the generalized gamma family with σ (0, 1) and λ(dx) = dx. Note that the powers α and r do not need to satisfy any relation.
11 Uniform balls of the true hazard rate In order to derive explicit consistency results for the mixture hazard models based on the four kernels considered, we need to establish: { Π h : sup 0<t T h(t) h 0 (t) } < δ > 0. To this aim, the following result is essential. Denote by G the space of nondecreasing càdlàg functions on R + such that G(0) = 0 for any G G. Note that any CRM induces a distribution on G and that, for G 0 (x) = µ 0 ( [0, x] ), G0 G. LEMMA 1 Support of CRMs Let µ be a CRM on R +, satisfying ν(r +, dx) = for any x X, and denote by Q the distribution induced on G. Then, for any G 0 G, any finite M and η > 0, { } Q G G : sup x M G(x) G 0 (x) < η > 0. Different techniques are required for the (exp) kernel w.r.t. the other three kernels (different support of k(t, x) seen as a function of x).
12 PROPOSITION 2 Weak consistency for (DL), (rect) and (OU) Let µ be a generalized gamma CRM with σ (0, 1) and λ(dx) = dx. Then weak consistency holds for the (DL), (rect) and (OU) kernels provided that: (i) h 0 (t) > 0 for any t > 0; (ii) t i f 0 (t)dt < with i = 1 for (rect) and (OU), i = 2 for (DL). Note that Condition (i) is automatically satisfied if µ 0 is absolutely continuous. PROPOSITION 3 Weak consistency for (exp) Weak consistency holds for the (exp) kernel provided that: (i) h(0) < a.s. and h 0 (0) < ; (ii) t f 0 (t)dt <. The only relevant assumption consists in asking for both h and h 0 not to explode in 0. Note that h 0 (t) > 0 is automatically satisfied.
13 CLTs for linear and quadratic functionals CLTs provide a synthetic picture of the random hazard, like information about the trend, oscillation and overall variance for large time horizons. In Peccati and Pruenster (2006) functionals of the prior hazard rate are considered. Such results can serve as a guide for prior specification. Here we investigate the asymptotic behaviour of functionals of the posterior random hazard rate given a fixed number of observations. We found out that, in all the considered special cases, the CLTs associated with the posterior hazard rate are the same as for the prior ones, and this for any number of observations. Although consistency implies that a given model can be asymptotically directed towards any deterministic target, the overall structure of a posterior hazard rate is systematically determined by the prior choice.
14 Posterior distribution We exploit an explicit posterior representation of the random mixture hazard model of James (2005). Let P be the random probability measure associated with f (t) and let (Y n ) n 1 be a sequence of exchangeable observations from P. Set Y = (Y 1,..., Y n ). Define a random probability measure on X R + as P(dx, dt) = e R t R 0 X k(s,y) µ(dy) ds k(t, x) µ(dx) dt. By introducing the latent variables X = (X 1,..., X n ), we describe µ Y in terms of µ Y, X mixed over X Y. Since the X i s may feature ties, we denote by X = (X1,..., X k ) the k n distinct latent variables. Let n j be the frequency of Xj and C j = {r : x r = xj }. The likelihood function is then given by L (µ; y, x) = e R P n R yi X i=1 0 k(t,x)dtµ(dx) k j=1 µ(dx j ) n j i C j k(y i, xj )
15 Given X and Y, the conditional distribution of µ Y, X coincides with the distribution of the random measure µ n d = µ n + k i=1 J i δ X j = µ n + n µ n is a CRM with intensity measure ν n (dv, dx) := e v P R n yi i=1 0 k(t,x)dt ρ(dv x) λ(dx) n := k i=1 J i δ X j with, for i = 1,..., k, X i a fixed points of discontinuity with corresponding jump J i which are, conditionally on X and Y, independent of µ n. The conditional distribution X Y, say P X Y, can be expressed as the conditional distributions of X p, Y and p Y, where p is a partitions of the integers {1,..., n}. We derive CLTs for the posterior random hazard given X and Y : h n(t) = X k(t, x) µn (dx) + k i=1 J ik(t, X i ), and, then, average over the distribution of X Y.
16 Limit theorems for shifted measures Peccati and Pruenster (2006) provide sufficient conditions to have that linear and quadratic functionals associated with a random mixture hazard without fixed atoms verify a CLT. Linear functional. For given X and Y, define H n(t ) = T 0 h n(t)dt = H n (T ) + k i=1 J i T 0 k(t, X i )dt Assume there exists a (deterministic) function T C n (k, T ) and σn(k) 2 > 0 such that, as T +, Cn(k, 2 T ) [ s ] T 2 R + X 0 k(t, x)dt ν n (ds, dx) σn(k), 2 Cn(k, 3 T ) [ s ] T 3 R + X 0 k(t, x)dt ν n (ds, dx) 0, Then C n (k, T ) [ Hn (T ) E[ H ] Y n law (T )] N(0, σn(k)), 2 where σ 2 n(k) may depend on Y.
17 Assume moreover that lim C n(k, T ) k j=1 J j T + T 0 k(t, X j )dt = m n ( n, k), (1) for some quantity m n ( n, k) 0 that may depend on X. Then C n (k, T ) [ H n(t ) E[ H ] n Y law (T )], X N ( m n ( n, k), σn(k) 2 ). If (1) holds for X = (x 1,..., x n ) X n in a set of P X Y probability 1, then we can state the result as η(t ) [ ] H(T law ) τ(t ) Y Z X,Y d P X Y, X where n η(t ) = C n (k, T ); τ(t ) = E[ H n (T ) Y ]; Z X,Y is a Gaussian random variable with mean m n ( n, k) and variance σ n (k). The mixture in X n Z X,Y d P X Y is over the quantity m n ( n, k) that may change according with the realization of X.
18 Applications We derive CLTs for functionals of posterior hazards based on the four kernels combined with generalized gamma CRMs. In general m n ( n, k) = 0, not depending on Y, X, that is the part of the posterior with fixed points of discontinuity affects the asymptotic behaviour just in the limit. Then, the convergence is simply expressed in terms of a Gaussian random variable. LINEAR CLT 1 (OU) kernel [ ] T 1 2 H(T ) 2 κ γ (1 σ) law T Y N(0, 2γ (1 σ) ) LINEAR CLT 2 (DL) kernel [ T 3 2 H(T ) (2γ (1 σ) ) 1 T 2] ( ) law Y 1 σ N 0, 3γ 2 σ Most important: C n (k, T ) = C 0 (k, T ), E[ H(T )] E[ H(T ) Y ] and σ n (k) = σ 0 (k), that is we get the same trend, oscillation around the trend and overall variance we get with the prior cumulative hazard.
19 One might think that the lack of dependence on the data of posterior functionals may be due to the boundedness of the support of the (DL), (OU) and (rect) kernels, as it happens that ν n = ν in R + [Y (n), ), where Y (n) is the largest observed lifetime. Consider then the (exp) kernel. In particular, we set λ(dx)=x 1 2 e 1 x (2 π) 1 in the Poisson intensity ν(dv, dx) of generalized gamma CRM, which gives us h(0) < a.s. and a prior mean centered on a quasi Weibull hazard, E[ h(t)] = γ 1+σ / t + 1. LINEAR CLT 3 (exp) kernel [ T 1 2 H(T ) (γ 1+σ ) 1 T 1/2] ( ) law Y N 0, (2 2)(1 σ) γ 2 σ Again, the oscillation rate, the trend and the overall variance coincides with the prior ones, that is they do not depend on the observation Y.
20 Concluding remarks We provide a comprehensive investigation of posterior weak consistency of the mixture hazard model. L 1 consistency would be the natural successive target. The investigation of CLTs for functionals represents a completely new line of research and has interest in its own. The coincidence of asymptotic behaviour of the posterior and the prior hazard means that the data do not influence the behaviour of the model for times larger than the largest observed lifetime Y (n) ; and, this, despite the fact that these models are consistent. The fact that the overall variance is not influenced by the data is somehow counterintuitive: since the contribution of the CRM vanishes in the limit, one would expect the variance to become smaller and smaller as more data come in. Since this does not happen, our findings show a clear evidence that the choice of the CRM really matters whatever the size of the dataset and provide a guideline in incorporating prior knowledge appropriately into the model.
21 Appendix For Further Reading Brix, A. (1999) Generalized gamma measures and shot-noise Cox processes. Adv. Appl. Prob. 31, Dykstra, R.L. and Laud, P. (1981) A Bayesian nonparametric approach to reliability. Ann. Statist. 9, James, L.F. (2005) Bayesian Poisson process partition calculus with an application to Bayesian Lévy moving averages. Ann. Statist. 33, Lo, A.Y. and Weng, C.S. (1989) On a class of Bayesian nonparametric estimates. II. Hazard rate estimates. Ann. Inst. Statist. Math. 41, Peccati, G. and Prünster, I. (2006) Linear and quadratic functionals of random hazard rates: an asymptotic analysis. Preprint.
Asymptotics for posterior hazards
Asymptotics for posterior hazards Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with P. Di Biasi and G. Peccati Workshop on Limit Theorems and Applications Paris, 16th January
More informationOn the posterior structure of NRMI
On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline
More informationTruncation error of a superposed gamma process in a decreasing order representation
Truncation error of a superposed gamma process in a decreasing order representation Julyan Arbel Inria Grenoble, Université Grenoble Alpes julyan.arbel@inria.fr Igor Prünster Bocconi University, Milan
More informationBayesian semiparametric analysis of short- and long- term hazard ratios with covariates
Bayesian semiparametric analysis of short- and long- term hazard ratios with covariates Department of Statistics, ITAM, Mexico 9th Conference on Bayesian nonparametrics Amsterdam, June 10-14, 2013 Contents
More informationTruncation error of a superposed gamma process in a decreasing order representation
Truncation error of a superposed gamma process in a decreasing order representation B julyan.arbel@inria.fr Í www.julyanarbel.com Inria, Mistis, Grenoble, France Joint work with Igor Pru nster (Bocconi
More informationFunctional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals
Functional Limit theorems for the quadratic variation of a continuous time random walk and for certain stochastic integrals Noèlia Viles Cuadros BCAM- Basque Center of Applied Mathematics with Prof. Enrico
More informationBayesian Nonparametrics: some contributions to construction and properties of prior distributions
Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Annalisa Cerquetti Collegio Nuovo, University of Pavia, Italy Interview Day, CETL Lectureship in Statistics,
More informationOn the Truncation Error of a Superposed Gamma Process
On the Truncation Error of a Superposed Gamma Process Julyan Arbel and Igor Prünster Abstract Completely random measures (CRMs) form a key ingredient of a wealth of stochastic models, in particular in
More informationDependent hierarchical processes for multi armed bandits
Dependent hierarchical processes for multi armed bandits Federico Camerlenghi University of Bologna, BIDSA & Collegio Carlo Alberto First Italian meeting on Probability and Mathematical Statistics, Torino
More informationPoisson random measure: motivation
: motivation The Lévy measure provides the expected number of jumps by time unit, i.e. in a time interval of the form: [t, t + 1], and of a certain size Example: ν([1, )) is the expected number of jumps
More informationPROBABILITY: LIMIT THEOREMS II, SPRING HOMEWORK PROBLEMS
PROBABILITY: LIMIT THEOREMS II, SPRING 218. HOMEWORK PROBLEMS PROF. YURI BAKHTIN Instructions. You are allowed to work on solutions in groups, but you are required to write up solutions on your own. Please
More informationBayesian estimation of the discrepancy with misspecified parametric models
Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012
More informationCovariance function estimation in Gaussian process regression
Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian
More informationarxiv: v2 [math.st] 27 May 2014
Full Bayesian inference with hazard mixture models Julyan Arbel a, Antonio Lijoi b,a, Bernardo Nipoti c,a, arxiv:145.6628v2 [math.st] 27 May 214 a Collegio Carlo Alberto, via Real Collegio, 3, 124 Moncalieri,
More informationNormalized kernel-weighted random measures
Normalized kernel-weighted random measures Jim Griffin University of Kent 1 August 27 Outline 1 Introduction 2 Ornstein-Uhlenbeck DP 3 Generalisations Bayesian Density Regression We observe data (x 1,
More informationSTAT Sample Problem: General Asymptotic Results
STAT331 1-Sample Problem: General Asymptotic Results In this unit we will consider the 1-sample problem and prove the consistency and asymptotic normality of the Nelson-Aalen estimator of the cumulative
More informationBayesian nonparametric models for bipartite graphs
Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers
More informationMath 341: Probability Seventeenth Lecture (11/10/09)
Math 341: Probability Seventeenth Lecture (11/10/09) Steven J Miller Williams College Steven.J.Miller@williams.edu http://www.williams.edu/go/math/sjmiller/ public html/341/ Bronfman Science Center Williams
More informationConstruction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting
Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting Anne Philippe Laboratoire de Mathématiques Jean Leray Université de Nantes Workshop EDF-INRIA,
More informationFiltrations, Markov Processes and Martingales. Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition
Filtrations, Markov Processes and Martingales Lectures on Lévy Processes and Stochastic Calculus, Braunschweig, Lecture 3: The Lévy-Itô Decomposition David pplebaum Probability and Statistics Department,
More informationStep-Stress Models and Associated Inference
Department of Mathematics & Statistics Indian Institute of Technology Kanpur August 19, 2014 Outline Accelerated Life Test 1 Accelerated Life Test 2 3 4 5 6 7 Outline Accelerated Life Test 1 Accelerated
More informationFull Bayesian inference with hazard mixture models
ISSN 2279-9362 Full Bayesian inference with hazard mixture models Julyan Arbel Antonio Lijoi Bernardo Nipoti No. 381 December 214 www.carloalberto.org/research/working-papers 214 by Julyan Arbel, Antonio
More informationMultiple Random Variables
Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x
More informationIntroduction to self-similar growth-fragmentations
Introduction to self-similar growth-fragmentations Quan Shi CIMAT, 11-15 December, 2017 Quan Shi Growth-Fragmentations CIMAT, 11-15 December, 2017 1 / 34 Literature Jean Bertoin, Compensated fragmentation
More informationLecture 4. f X T, (x t, ) = f X,T (x, t ) f T (t )
LECURE NOES 21 Lecture 4 7. Sufficient statistics Consider the usual statistical setup: the data is X and the paramter is. o gain information about the parameter we study various functions of the data
More informationNEW FUNCTIONAL INEQUALITIES
1 / 29 NEW FUNCTIONAL INEQUALITIES VIA STEIN S METHOD Giovanni Peccati (Luxembourg University) IMA, Minneapolis: April 28, 2015 2 / 29 INTRODUCTION Based on two joint works: (1) Nourdin, Peccati and Swan
More informationOn prediction and density estimation Peter McCullagh University of Chicago December 2004
On prediction and density estimation Peter McCullagh University of Chicago December 2004 Summary Having observed the initial segment of a random sequence, subsequent values may be predicted by calculating
More informationOn pathwise stochastic integration
On pathwise stochastic integration Rafa l Marcin Lochowski Afican Institute for Mathematical Sciences, Warsaw School of Economics UWC seminar Rafa l Marcin Lochowski (AIMS, WSE) On pathwise stochastic
More informationRegular Variation and Extreme Events for Stochastic Processes
1 Regular Variation and Extreme Events for Stochastic Processes FILIP LINDSKOG Royal Institute of Technology, Stockholm 2005 based on joint work with Henrik Hult www.math.kth.se/ lindskog 2 Extremes for
More informationLecture Characterization of Infinitely Divisible Distributions
Lecture 10 1 Characterization of Infinitely Divisible Distributions We have shown that a distribution µ is infinitely divisible if and only if it is the weak limit of S n := X n,1 + + X n,n for a uniformly
More informationBayesian Modeling of Conditional Distributions
Bayesian Modeling of Conditional Distributions John Geweke University of Iowa Indiana University Department of Economics February 27, 2007 Outline Motivation Model description Methods of inference Earnings
More informationConvergence of Feller Processes
Chapter 15 Convergence of Feller Processes This chapter looks at the convergence of sequences of Feller processes to a iting process. Section 15.1 lays some ground work concerning weak convergence of processes
More informationDensities for the Navier Stokes equations with noise
Densities for the Navier Stokes equations with noise Marco Romito Università di Pisa Universitat de Barcelona March 25, 2015 Summary 1 Introduction & motivations 2 Malliavin calculus 3 Besov bounds 4 Other
More informationGaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008
Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:
More informationAsymptotic properties of maximum likelihood estimator for the growth rate for a jump-type CIR process
Asymptotic properties of maximum likelihood estimator for the growth rate for a jump-type CIR process Mátyás Barczy University of Debrecen, Hungary The 3 rd Workshop on Branching Processes and Related
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationPhenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 2012
Phenomena in high dimensions in geometric analysis, random matrices, and computational geometry Roscoff, France, June 25-29, 202 BOUNDS AND ASYMPTOTICS FOR FISHER INFORMATION IN THE CENTRAL LIMIT THEOREM
More informationTwo viewpoints on measure valued processes
Two viewpoints on measure valued processes Olivier Hénard Université Paris-Est, Cermics Contents 1 The classical framework : from no particle to one particle 2 The lookdown framework : many particles.
More informationStatistical inference on Lévy processes
Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline
More informationAppendix A. Sequences and series. A.1 Sequences. Definition A.1 A sequence is a function N R.
Appendix A Sequences and series This course has for prerequisite a course (or two) of calculus. The purpose of this appendix is to review basic definitions and facts concerning sequences and series, which
More informationProbability and Measure
Chapter 4 Probability and Measure 4.1 Introduction In this chapter we will examine probability theory from the measure theoretic perspective. The realisation that measure theory is the foundation of probability
More informationGaussian processes for inference in stochastic differential equations
Gaussian processes for inference in stochastic differential equations Manfred Opper, AI group, TU Berlin November 6, 2017 Manfred Opper, AI group, TU Berlin (TU Berlin) inference in SDE November 6, 2017
More informationS1. Proofs of the Key Properties of CIATA-Ph
Article submitted to INFORMS Journal on Computing; manuscript no. (Please, provide the mansucript number!) S-1 Online Supplement to Modeling and Simulation of Nonstationary and Non-Poisson Processes by
More informationNew Dirichlet Mean Identities
Hong Kong University of Science and Technology Isaac Newton Institute, August 10, 2007 Origins CIFARELLI, D. M. and REGAZZINI, E. (1979). Considerazioni generali sull impostazione bayesiana di problemi
More informationFoundations of Nonparametric Bayesian Methods
1 / 27 Foundations of Nonparametric Bayesian Methods Part II: Models on the Simplex Peter Orbanz http://mlg.eng.cam.ac.uk/porbanz/npb-tutorial.html 2 / 27 Tutorial Overview Part I: Basics Part II: Models
More informationPattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions
Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite
More informationTranslation Invariant Experiments with Independent Increments
Translation Invariant Statistical Experiments with Independent Increments (joint work with Nino Kordzakhia and Alex Novikov Steklov Mathematical Institute St.Petersburg, June 10, 2013 Outline 1 Introduction
More informationThe information complexity of sequential resource allocation
The information complexity of sequential resource allocation Emilie Kaufmann, joint work with Olivier Cappé, Aurélien Garivier and Shivaram Kalyanakrishan SMILE Seminar, ENS, June 8th, 205 Sequential allocation
More informationConsistency of the maximum likelihood estimator for general hidden Markov models
Consistency of the maximum likelihood estimator for general hidden Markov models Jimmy Olsson Centre for Mathematical Sciences Lund University Nordstat 2012 Umeå, Sweden Collaborators Hidden Markov models
More informationCS 540: Machine Learning Lecture 2: Review of Probability & Statistics
CS 540: Machine Learning Lecture 2: Review of Probability & Statistics AD January 2008 AD () January 2008 1 / 35 Outline Probability theory (PRML, Section 1.2) Statistics (PRML, Sections 2.1-2.4) AD ()
More informationGARCH processes continuous counterparts (Part 2)
GARCH processes continuous counterparts (Part 2) Alexander Lindner Centre of Mathematical Sciences Technical University of Munich D 85747 Garching Germany lindner@ma.tum.de http://www-m1.ma.tum.de/m4/pers/lindner/
More informationOrder book resilience, price manipulation, and the positive portfolio problem
Order book resilience, price manipulation, and the positive portfolio problem Alexander Schied Mannheim University Workshop on New Directions in Financial Mathematics Institute for Pure and Applied Mathematics,
More information1. Stochastic Processes and filtrations
1. Stochastic Processes and 1. Stoch. pr., A stochastic process (X t ) t T is a collection of random variables on (Ω, F) with values in a measurable space (S, S), i.e., for all t, In our case X t : Ω S
More informationMATH 5640: Fourier Series
MATH 564: Fourier Series Hung Phan, UMass Lowell September, 8 Power Series A power series in the variable x is a series of the form a + a x + a x + = where the coefficients a, a,... are real or complex
More informationMath 430/530: Advanced Calculus I Chapter 1: Sequences
Math 430/530: Advanced Calculus I Chapter 1: Sequences Fall 2018 Sequences A sequence is just a list of real numbers a 1, a 2,... with a definite order. We write a = (a n ) n=1 for a sequence a with n
More informationParametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory
Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationRandom Process Lecture 1. Fundamentals of Probability
Random Process Lecture 1. Fundamentals of Probability Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2016 1/43 Outline 2/43 1 Syllabus
More informationNonlinear representation, backward SDEs, and application to the Principal-Agent problem
Nonlinear representation, backward SDEs, and application to the Principal-Agent problem Ecole Polytechnique, France April 4, 218 Outline The Principal-Agent problem Formulation 1 The Principal-Agent problem
More informationDependent mixture models: clustering and borrowing information
ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,
More informationFrontier estimation based on extreme risk measures
Frontier estimation based on extreme risk measures by Jonathan EL METHNI in collaboration with Ste phane GIRARD & Laurent GARDES CMStatistics 2016 University of Seville December 2016 1 Risk measures 2
More informationThese slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop
Music and Machine Learning (IFT68 Winter 8) Prof. Douglas Eck, Université de Montréal These slides follow closely the (English) course textbook Pattern Recognition and Machine Learning by Christopher Bishop
More informationNonparametric regression with martingale increment errors
S. Gaïffas (LSTA - Paris 6) joint work with S. Delattre (LPMA - Paris 7) work in progress Motivations Some facts: Theoretical study of statistical algorithms requires stationary and ergodicity. Concentration
More informationEvgeny Spodarev WIAS, Berlin. Limit theorems for excursion sets of stationary random fields
Evgeny Spodarev 23.01.2013 WIAS, Berlin Limit theorems for excursion sets of stationary random fields page 2 LT for excursion sets of stationary random fields Overview 23.01.2013 Overview Motivation Excursion
More informationDiscussion of On simulation and properties of the stable law by L. Devroye and L. James
Stat Methods Appl (2014) 23:371 377 DOI 10.1007/s10260-014-0269-4 Discussion of On simulation and properties of the stable law by L. Devroye and L. James Antonio Lijoi Igor Prünster Accepted: 16 May 2014
More informationThe strictly 1/2-stable example
The strictly 1/2-stable example 1 Direct approach: building a Lévy pure jump process on R Bert Fristedt provided key mathematical facts for this example. A pure jump Lévy process X is a Lévy process such
More informationHarnack Inequalities and Applications for Stochastic Equations
p. 1/32 Harnack Inequalities and Applications for Stochastic Equations PhD Thesis Defense Shun-Xiang Ouyang Under the Supervision of Prof. Michael Röckner & Prof. Feng-Yu Wang March 6, 29 p. 2/32 Outline
More informationEstimation of the long Memory parameter using an Infinite Source Poisson model applied to transmission rate measurements
of the long Memory parameter using an Infinite Source Poisson model applied to transmission rate measurements François Roueff Ecole Nat. Sup. des Télécommunications 46 rue Barrault, 75634 Paris cedex 13,
More informationSlice sampling σ stable Poisson Kingman mixture models
ISSN 2279-9362 Slice sampling σ stable Poisson Kingman mixture models Stefano Favaro S.G. Walker No. 324 December 2013 www.carloalberto.org/research/working-papers 2013 by Stefano Favaro and S.G. Walker.
More informationComputer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression
Group Prof. Daniel Cremers 4. Gaussian Processes - Regression Definition (Rep.) Definition: A Gaussian process is a collection of random variables, any finite number of which have a joint Gaussian distribution.
More informationWeak quenched limiting distributions of a one-dimensional random walk in a random environment
Weak quenched limiting distributions of a one-dimensional random walk in a random environment Jonathon Peterson Cornell University Department of Mathematics Joint work with Gennady Samorodnitsky September
More informationThe parabolic Anderson model on Z d with time-dependent potential: Frank s works
Weierstrass Institute for Applied Analysis and Stochastics The parabolic Anderson model on Z d with time-dependent potential: Frank s works Based on Frank s works 2006 2016 jointly with Dirk Erhard (Warwick),
More informationON ADDITIVE TIME-CHANGES OF FELLER PROCESSES. 1. Introduction
ON ADDITIVE TIME-CHANGES OF FELLER PROCESSES ALEKSANDAR MIJATOVIĆ AND MARTIJN PISTORIUS Abstract. In this note we generalise the Phillips theorem [1] on the subordination of Feller processes by Lévy subordinators
More informationMathematical Institute, University of Utrecht. The problem of estimating the mean of an observed Gaussian innite-dimensional vector
On Minimax Filtering over Ellipsoids Eduard N. Belitser and Boris Y. Levit Mathematical Institute, University of Utrecht Budapestlaan 6, 3584 CD Utrecht, The Netherlands The problem of estimating the mean
More informationComputer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression
Group Prof. Daniel Cremers 9. Gaussian Processes - Regression Repetition: Regularized Regression Before, we solved for w using the pseudoinverse. But: we can kernelize this problem as well! First step:
More informationCumulants of a convolution and applications to monotone probability theory
Cumulants of a convolution and applications to monotone probability theory arxiv:95.3446v2 [math.pr] 24 May 29 Takahiro Hasebe JSPS Research Fellow (DC), Graduate School of Science, Kyoto University, Kyoto
More informationTemporal point processes: the conditional intensity function
Temporal point processes: the conditional intensity function Jakob Gulddahl Rasmussen December 21, 2009 Contents 1 Introduction 2 2 Evolutionary point processes 2 2.1 Evolutionarity..............................
More informationBrownian Motion. 1 Definition Brownian Motion Wiener measure... 3
Brownian Motion Contents 1 Definition 2 1.1 Brownian Motion................................. 2 1.2 Wiener measure.................................. 3 2 Construction 4 2.1 Gaussian process.................................
More informationUniversity of Toronto Department of Statistics
Norm Comparisons for Data Augmentation by James P. Hobert Department of Statistics University of Florida and Jeffrey S. Rosenthal Department of Statistics University of Toronto Technical Report No. 0704
More informationOn some distributional properties of Gibbs-type priors
On some distributional properties of Gibbs-type priors Igor Prünster University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics Workshop ICERM, 21st September 2012 Joint work with: P. De Blasi,
More informationStatistics (1): Estimation
Statistics (1): Estimation Marco Banterlé, Christian Robert and Judith Rousseau Practicals 2014-2015 L3, MIDO, Université Paris Dauphine 1 Table des matières 1 Random variables, probability, expectation
More informationOn oscillating systems of interacting neurons
Susanne Ditlevsen Eva Löcherbach Nice, September 2016 Oscillations Most biological systems are characterized by an intrinsic periodicity : 24h for the day-night rhythm of most animals, etc... In other
More informationIn terms of measures: Exercise 1. Existence of a Gaussian process: Theorem 2. Remark 3.
1. GAUSSIAN PROCESSES A Gaussian process on a set T is a collection of random variables X =(X t ) t T on a common probability space such that for any n 1 and any t 1,...,t n T, the vector (X(t 1 ),...,X(t
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationBayesian Nonparametrics
Bayesian Nonparametrics Peter Orbanz Columbia University PARAMETERS AND PATTERNS Parameters P(X θ) = Probability[data pattern] 3 2 1 0 1 2 3 5 0 5 Inference idea data = underlying pattern + independent
More informationP-adic Functions - Part 1
P-adic Functions - Part 1 Nicolae Ciocan 22.11.2011 1 Locally constant functions Motivation: Another big difference between p-adic analysis and real analysis is the existence of nontrivial locally constant
More informationInformation geometry for bivariate distribution control
Information geometry for bivariate distribution control C.T.J.Dodson + Hong Wang Mathematics + Control Systems Centre, University of Manchester Institute of Science and Technology Optimal control of stochastic
More informationNonparametric estimation of log-concave densities
Nonparametric estimation of log-concave densities Jon A. Wellner University of Washington, Seattle Seminaire, Institut de Mathématiques de Toulouse 5 March 2012 Seminaire, Toulouse Based on joint work
More informationThe information complexity of best-arm identification
The information complexity of best-arm identification Emilie Kaufmann, joint work with Olivier Cappé and Aurélien Garivier MAB workshop, Lancaster, January th, 206 Context: the multi-armed bandit model
More informationUnderstanding Regressions with Observations Collected at High Frequency over Long Span
Understanding Regressions with Observations Collected at High Frequency over Long Span Yoosoon Chang Department of Economics, Indiana University Joon Y. Park Department of Economics, Indiana University
More informationSTOCHASTIC GEOMETRY BIOIMAGING
CENTRE FOR STOCHASTIC GEOMETRY AND ADVANCED BIOIMAGING 2018 www.csgb.dk RESEARCH REPORT Anders Rønn-Nielsen and Eva B. Vedel Jensen Central limit theorem for mean and variogram estimators in Lévy based
More informationLAN property for ergodic jump-diffusion processes with discrete observations
LAN property for ergodic jump-diffusion processes with discrete observations Eulalia Nualart (Universitat Pompeu Fabra, Barcelona) joint work with Arturo Kohatsu-Higa (Ritsumeikan University, Japan) &
More informationMathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio ( )
Mathematical Methods for Neurosciences. ENS - Master MVA Paris 6 - Master Maths-Bio (2014-2015) Etienne Tanré - Olivier Faugeras INRIA - Team Tosca November 26th, 2014 E. Tanré (INRIA - Team Tosca) Mathematical
More informationLog Gaussian Cox Processes. Chi Group Meeting February 23, 2016
Log Gaussian Cox Processes Chi Group Meeting February 23, 2016 Outline Typical motivating application Introduction to LGCP model Brief overview of inference Applications in my work just getting started
More informationLarge Deviations for Small-Noise Stochastic Differential Equations
Chapter 22 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large
More informationThe small ball property in Banach spaces (quantitative results)
The small ball property in Banach spaces (quantitative results) Ehrhard Behrends Abstract A metric space (M, d) is said to have the small ball property (sbp) if for every ε 0 > 0 there exists a sequence
More informationMalliavin calculus and central limit theorems
Malliavin calculus and central limit theorems David Nualart Department of Mathematics Kansas University Seminar on Stochastic Processes 2017 University of Virginia March 8-11 2017 David Nualart (Kansas
More informationLarge Deviations for Small-Noise Stochastic Differential Equations
Chapter 21 Large Deviations for Small-Noise Stochastic Differential Equations This lecture is at once the end of our main consideration of diffusions and stochastic calculus, and a first taste of large
More informationAdvanced Calculus Math 127B, Winter 2005 Solutions: Final. nx2 1 + n 2 x, g n(x) = n2 x
. Define f n, g n : [, ] R by f n (x) = Advanced Calculus Math 27B, Winter 25 Solutions: Final nx2 + n 2 x, g n(x) = n2 x 2 + n 2 x. 2 Show that the sequences (f n ), (g n ) converge pointwise on [, ],
More informationMaximum likelihood estimation of a log-concave density based on censored data
Maximum likelihood estimation of a log-concave density based on censored data Dominic Schuhmacher Institute of Mathematical Statistics and Actuarial Science University of Bern Joint work with Lutz Dümbgen
More information