New Dirichlet Mean Identities

Size: px
Start display at page:

Download "New Dirichlet Mean Identities"

Transcription

1 Hong Kong University of Science and Technology Isaac Newton Institute, August 10, 2007

2 Origins CIFARELLI, D. M. and REGAZZINI, E. (1979). Considerazioni generali sull impostazione bayesiana di problemi non parametrici. Le medie associative nel contesto del processo aleatorio di Dirichlet I, II. Riv. Mat. Sci. Econom. Social 2, Later there is the important 1990 Annals of Statistics paper.

3 1 It is a cornerstone work in Bayesian NP and I believe one of the most important contributions in the theory, and indeed application, of random processes. 2 Why? Dirichlet means arise everywhere.

4 Some things to keep in mind Hard and very technical stuff But Cifarelli and Regazzini already did the hard work so we do not have to Our Task is really how to figure out how to use their results. Not so much on how to re-prove them. For some recent non-trivial applications look me up on the math Arxiv I never really do hard things. This is not a Joke

5 Theme Prendere due piccioni con una fava

6 Dirichlet process Let X be a non-negative random variable with cumulative distribution function F X. Note furthermore for a measurable set C, we use the notation F X (C) to mean the probability that X is in C. One may define a Dirichlet process random probability measure, say P θ, on [0, ) with total mass parameter θ and prior parameter F X, via its finite dimensional distribution as follows; for any disjoint partition on [0, ), say (C 1,..., C k ), the distribution of the random vector. (P θ (C 1 ),..., P θ (C k )) is a k-variate Dirichet distribution with parameters (θf X (C 1 ),..., θf X (C k )).

7 Hence for each C, P θ (C) = 0 I(x C)P θ (dx) has a beta distribution with parameters (θf X (C), θ(1 F X (C))).

8 Equivalently setting θf X (C i ) = θ i for i = 1,..., k, ( ) (P θ (C 1 ),..., P θ (C k )) = d Gθi ; i = 1,..., k G θ where (G θi ) are independent random variables with gamma(θ i, 1) distributions and G θ = G θ1 + + G θk has a gamma(θ, 1) distribution.

9 This means that one can define the Dirichlet process via the normalization of an independent increment gamma process on [0, ), say γ θ ( ), as P θ ( ) = γ θ( ) γ θ ([0, )) where γ θ (C i ) d = G θi and whose almost surely finite total random mass is γ θ ([0, )) d = G θ.

10 A very important aspect of this construction is the fact that G θ is independent of P θ, and hence any functional P θ (g) = 0 g(x)p θ (dx)

11 Furthermore θ > 0, G θ P θ (g) = d 0 g(x)γ θ (dx) = γ θ (g) See for instance Lijoi and Regazzini- AOP

12 Cifarelli and Regazzini/Markov Krein ID Now, E[e λg θp θ (g) ] = E[(1 + λp θ (g)) θ ] The Laplace transform of a random variable representable as γ θ (g) is equivalent to the Cauchy-Stieltjes transform of order θ of a random variable representable as P θ (g)

13 Something to keep in mind Lets think about random variables on the real line rather than random processes in some Banach space

14 A more general statement: Psychologically easier to prove???!!! Let M be a positive random variable independent of G θ, and define the random variable R θ d = Gθ M then from your first course in probability E[e λr θ ] = E[(1 + λm) θ ]

15 Question Are the results on the last two slides equivalent?

16 Answer-Questions γ θ (g) is always infinitely divisible Scale mixtures of gamma random variables are not always infinitely divisible. Scale mixtures of gamma random variables of index 0 < θ 1 are always infinitely divisible. If R θ is infinitely divisible is it true then?

17 Classical Question Recall that E[e λr θ ] = E[(1 + λm) θ ] In general :How to find the density of R θ? In general :How to find the density of M?

18 Classical but painful answer Utilize Classical Inversion formula You have to deal with the Analysis of Complex functions No simpler answer in general

19 Cifarelli and Regazzini provide answers for M d = P θ (g)

20 My Solutions Call up my Italian friends!! (Lijoi and Prünster) Build on existing results

21 Important? γ θ (g) constitute a large class of Infinitely divisible random variables. Id random variables generate EVERY positive Lévy process Levy processes are now the building blocks in many diverse fields, physics, finance, genetics, machine learning. All these models can be treated from a Bayesian NP point of view Levy processes are connected with special functions One wants to get the laws of certain functionals, BUT this is hard Simple example-finance: One may want to use a Barndorff-Nielsen OU SV process for option pricing without resorting to a series representation.

22 some more 1 Many important quantities can be represented as P θ (g) 2 Curiously James, Lijoi and Pruenster[JLP], show that for θ > 0 every linear functional of a two parameter (α, θ) Poisson Dirichlet random probability measure can be represented as a Dirichlet mean of order θ.

23 Define a Dirichlet mean of order θ indexed by F X as M θ (F X ) = 0 xp θ (dx)

24 Say that d T θ = Gθ M θ (F X ) is a GGC(θ, F X ) random variable which satisfies E[e λt θ ] = E[(1 + λm θ (F X )) θ ] = e θψ F X (λ)

25 where ψ FX (λ) = 0 log(1 + λx)f X (dx) = E[log(1 + λx)].

26 Key expressions for density Cifarelli and Regazzini(1990) 1 Φ FX (t) = 2 furthermore, define, 0 log( t x )I(t x)f X (dx) = E[log( t X )I(t X)] θ (t F X ) = 1 π sin(πθf X(t))e θφ F X (t).

27 Then from [CR] 1 The cdf of M θ (F X ) for all θ > 0 is expressible as x 2 when θ = 1, the density is, 0 (x t) θ 1 θ (t F X )dt ξ FX (x) = 1 (x F X ) = 1 π sin(πf X(x))e Φ F (x) X. 3 Density formulae for θ > 1 are described as ξ θfx (x) = (θ 1) x 0 (x t) θ 2 θ (t F X )dt.

28 Result James, Lijoi and Prünster(2006) Recognizing a connection of [CR] s result for the cdf to Abel transforms, [JLP] reasoned that it is rather straightforward to establish the following 1 for all θ > 0 ξ θfx (x) = 2 is the derivative of x 0 (x t) θ 1 (t F X )dt

29 In general the only really nice expression for the density is when θ = 1 ξ FX (x) = 1 (x F X ) = 1 π sin(πf X(x))e Φ F X (x).

30 But notice that θ (t F X ) = 1 π sin(πθf X(t))e θφ F X (t) > 0. for 0 < θ 1 and all t > 0.

31 Comments on possible random variables that are Dirichlet means 1 The formulae we just saw are general expressions and don t look too familiar 2 However, we (James, Roynette and Yor-Coming Soon) show that there are many many familiar random variables that are Dirichlet means 3 This includes Generalized Inverse Gaussian, Positive Stable, Pareto, Uniform and at the very least a large class of random variables that are generalized gamma convolutions 4 Note if for instance U is uniform[0,1] then for 0 < θ 2, there exist F Xθ such that U d = M θ (F Xθ )

32 A useful identity: Hjort and Ongaro 05 Let D = (D 1,..., D k ) denote a Dirichlet (θ 1,..., θ k ) random vector such that θ = k i=1 θ i. Then M θ (F X ) d = k D i M θi (F X ) i=1 where M θi (F X ) are independent and independent of D

33 Goal Use the analytic work used to obtain results for M θ (F X ) to obtain results for an entire family of Dirichlet mean random variables {M θ (F Zc ) : c > 0} As a byproduct we get results for T θ and Lévy processes derived from T θ

34 Simple examples are of course the choices Z c = X + c and Z c = cx, which, due to the linearity properties of mean functionals, results easily in the identities in law M θ (F X+c ) = c + M θ (F X ) and M θ (F cx ) = cm θ (F X )

35 General DP Mean Identities I introduced two simple ideas to address some of these points. The idea to construct such things was in part influenced quite a bit by some results of Pitman and Yor (1997) On length of excursions of bessel processes and others. Concerning multiplication by beta random variables and so on. The two results can be described as 1 Beta Scaling of DP Mean functionals 2 Gamma Tilting 3 Actually there are two more

36 Beta Scaling of DP mean functionals Let 0 < σ 1 and θ > 0. Let β θσ,θ(1 σ) denote a beta random variable. Then consider the following simple random variable β θσ,θ(1 σ) M θσ (F X ) Remember the density of M θσ (F X ) is not NICE except for θσ = 1.

37 Let Y σ denote a Bernoulli random variable with success probability σ. Consider the independent product XY σ with cdf F XYσ (x) = σf X (x) + (1 σ)i(x 0)

38 Result 1: change of total mass Then β θσ,θ(1 σ) M θσ (F X ) d = M θ (F XYσ )

39 Why? ψ FXYσ (λ) = but is equal to 0 σψ FX (λ) = σ log(1 + λw)f XYσ (dw) = E[log(1 + λxy σ )]. 0 log(1 + λx)f X (dx) = σe[log(1 + λx)].

40 One to many For every fixed θ how many mean functionals of order θ did we just create? Answer: An uncountable number indexed by 0 < σ 1 (M θ (F XYσ ) : 0 < σ 1)

41 One to many explicit densities How many mean functionals of order θ = 1 did we just create? Answer: An uncountable number indexed by 0 < σ 1 (M 1 (F XYσ ) : 0 < σ 1)

42 Φ FXYσ (x) = E[log( x XY σ )I(XY σ x)] = σφ FX (x) + (1 σ) log(x)

43 Result 2 The density of β σ,1 σ M σ (F X ) d = M 1 (F XYσ ) is ξ FXYσ (x) = xσ 1 π sin(πf XY σ (x))e σφ F X (x) Note you do not need to know the density of M σ (F X ).

44 Back to GGC Recall but Hence T θσ = G θσ M θσ (F X ) G θσ d = Gθ [β θσ,θ(1 σ) ] T θσ = G θσ M θσ (F X ) = G θ [β θσ,θ(1 σ) M θσ (F X )]

45 Hence i.e. it is a GGC(θ, F XYσ ) T θσ = G θσ M θσ (F X ) = G θ M θ (F XYσ )

46 Hence i.e. it is a GGC(1, F XYσ ) T σ = G σ M σ (F X ) = G 1 M 1 (F XYσ )

47 Exploiting infinite divisibility Let 0 < σ k = 1 Then k=1 T 1 d = k=1 T σk where T σk are independent and each has distribution d T σk = G1 M 1 (F XYσk )

48 GGC species sampling models Let (Z k ) denote iid random elements with some common distribution H. Then set P k = T σk /T 1 Define a random probability measure as, P k δ Zk (dx) = P(dx) k=1 THESE ARE NOT NRMI

49 That is, for any finite k (P 1,..., P k ) d = (G 1,i M 1 (F XYσi )/[T + S]; i = 1,..., k)

50 where T = and for σ k = 1 k i=1 σ i k G 1,i M 1 (F XYσi ) i=1 S d = G 1 M 1 (F XYσ k ) independent of T.

51 For fun one could choose for i = 1, 2,... σ i = λ i 1 e λ /(i 1)!

52 Fidi of GGC NRM Let µ( ) denote a completely random measure derived from T 1, such that E[µ(A)] = H(A) and µ(x ) = T 1 Then for any A, set H(A) = σ, and one has that µ(a) = G 1 M 1 (F XYσ ) moreover for a disjoint partition of some space X, (C 1,..., C N ) The fidi of µ( ) is composed of independent components so that µ(c i ) = G 1 M 1 (F XYσi )

53 but writing T 1 = N µ(c i ) k=1 This gives the fidi of the random probability measure, P( ) = µ( )/T 1 i.e. (P(C 1 ),..., P(C N ).) These NRM have appeared in James Lijoi and Pruenster (2005) but not the fidi part.

54 Now conditioning on T 1 = t, this gives the fidi of the corresponding conditional Poisson Kingman class, L((P(C 1 ),..., P(C N ) T 1 = t) Now mix over an arbitrary density γ(t). Hence I just created an uncountable number of random probability measures with explicit fidi s.

55 Approximating GGC NRM s Now suppose that for general θ, let µ θ ( ) denote a CRM derived from T θ. Then form the NRM by, P θ ( ) = µ θ ( )/T θ Now write T θ = N k=1 T θ/n,k where T θ/n,k are iid equal in distribution to T θ/n

56 Now we are interested in choosing N > θ, in this case the joint distribution of (T θ/n,k /T θ ; k N) can be computed and certainly easily simulated. Since Furthermore as N T θ/n d = G1 M 1 (F XYθ/N ) N k=1 T θ/n,k T θ δ Zk ( ) P θ ( ) We can also use the left hand side as a sieve by letting N grow with the data.

57 Special Known Case G θ/n,k iid gamma(θ/n, 1) then N k=1 where D θ is a Dirichlet process G θ/n,k G θ δ Zk ( ) D θ ( )

On the posterior structure of NRMI

On the posterior structure of NRMI On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline

More information

Truncation error of a superposed gamma process in a decreasing order representation

Truncation error of a superposed gamma process in a decreasing order representation Truncation error of a superposed gamma process in a decreasing order representation B julyan.arbel@inria.fr Í www.julyanarbel.com Inria, Mistis, Grenoble, France Joint work with Igor Pru nster (Bocconi

More information

Dependent hierarchical processes for multi armed bandits

Dependent hierarchical processes for multi armed bandits Dependent hierarchical processes for multi armed bandits Federico Camerlenghi University of Bologna, BIDSA & Collegio Carlo Alberto First Italian meeting on Probability and Mathematical Statistics, Torino

More information

A marginal sampler for σ-stable Poisson-Kingman mixture models

A marginal sampler for σ-stable Poisson-Kingman mixture models A marginal sampler for σ-stable Poisson-Kingman mixture models joint work with Yee Whye Teh and Stefano Favaro María Lomelí Gatsby Unit, University College London Talk at the BNP 10 Raleigh, North Carolina

More information

Discussion of On simulation and properties of the stable law by L. Devroye and L. James

Discussion of On simulation and properties of the stable law by L. Devroye and L. James Stat Methods Appl (2014) 23:371 377 DOI 10.1007/s10260-014-0269-4 Discussion of On simulation and properties of the stable law by L. Devroye and L. James Antonio Lijoi Igor Prünster Accepted: 16 May 2014

More information

Asymptotics for posterior hazards

Asymptotics for posterior hazards Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and

More information

Normalized kernel-weighted random measures

Normalized kernel-weighted random measures Normalized kernel-weighted random measures Jim Griffin University of Kent 1 August 27 Outline 1 Introduction 2 Ornstein-Uhlenbeck DP 3 Generalisations Bayesian Density Regression We observe data (x 1,

More information

Bayesian Nonparametrics: some contributions to construction and properties of prior distributions

Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Annalisa Cerquetti Collegio Nuovo, University of Pavia, Italy Interview Day, CETL Lectureship in Statistics,

More information

On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation

On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation Antonio LIJOI, Igor PRÜNSTER, andstepheng.walker The past decade has seen a remarkable development in the area of Bayesian

More information

Bayesian nonparametric models for bipartite graphs

Bayesian nonparametric models for bipartite graphs Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers

More information

Chapter 5. Chapter 5 sections

Chapter 5. Chapter 5 sections 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

Truncation error of a superposed gamma process in a decreasing order representation

Truncation error of a superposed gamma process in a decreasing order representation Truncation error of a superposed gamma process in a decreasing order representation Julyan Arbel Inria Grenoble, Université Grenoble Alpes julyan.arbel@inria.fr Igor Prünster Bocconi University, Milan

More information

Statistics for scientists and engineers

Statistics for scientists and engineers Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3

More information

On Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process

On Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process On Simulations form the Two-Parameter arxiv:1209.5359v1 [stat.co] 24 Sep 2012 Poisson-Dirichlet Process and the Normalized Inverse-Gaussian Process Luai Al Labadi and Mahmoud Zarepour May 8, 2018 ABSTRACT

More information

A Brief Overview of Nonparametric Bayesian Models

A Brief Overview of Nonparametric Bayesian Models A Brief Overview of Nonparametric Bayesian Models Eurandom Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin Also at Machine

More information

Foundations of Nonparametric Bayesian Methods

Foundations of Nonparametric Bayesian Methods 1 / 27 Foundations of Nonparametric Bayesian Methods Part II: Models on the Simplex Peter Orbanz http://mlg.eng.cam.ac.uk/porbanz/npb-tutorial.html 2 / 27 Tutorial Overview Part I: Basics Part II: Models

More information

Probability Distributions Columns (a) through (d)

Probability Distributions Columns (a) through (d) Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)

More information

Bayesian Nonparametrics: Dirichlet Process

Bayesian Nonparametrics: Dirichlet Process Bayesian Nonparametrics: Dirichlet Process Yee Whye Teh Gatsby Computational Neuroscience Unit, UCL http://www.gatsby.ucl.ac.uk/~ywteh/teaching/npbayes2012 Dirichlet Process Cornerstone of modern Bayesian

More information

Northwestern University Department of Electrical Engineering and Computer Science

Northwestern University Department of Electrical Engineering and Computer Science Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability

More information

1 Probability and Random Variables

1 Probability and Random Variables 1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in

More information

Slice sampling σ stable Poisson Kingman mixture models

Slice sampling σ stable Poisson Kingman mixture models ISSN 2279-9362 Slice sampling σ stable Poisson Kingman mixture models Stefano Favaro S.G. Walker No. 324 December 2013 www.carloalberto.org/research/working-papers 2013 by Stefano Favaro and S.G. Walker.

More information

Stat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory. Charles J. Geyer School of Statistics University of Minnesota

Stat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory. Charles J. Geyer School of Statistics University of Minnesota Stat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory Charles J. Geyer School of Statistics University of Minnesota 1 Asymptotic Approximation The last big subject in probability

More information

Bayesian nonparametric latent feature models

Bayesian nonparametric latent feature models Bayesian nonparametric latent feature models Indian Buffet process, beta process, and related models François Caron Department of Statistics, Oxford Applied Bayesian Statistics Summer School Como, Italy

More information

Dependent mixture models: clustering and borrowing information

Dependent mixture models: clustering and borrowing information ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,

More information

The Poisson-Dirichlet Distribution: Constructions, Stochastic Dynamics and Asymptotic Behavior

The Poisson-Dirichlet Distribution: Constructions, Stochastic Dynamics and Asymptotic Behavior The Poisson-Dirichlet Distribution: Constructions, Stochastic Dynamics and Asymptotic Behavior Shui Feng McMaster University June 26-30, 2011. The 8th Workshop on Bayesian Nonparametrics, Veracruz, Mexico.

More information

Numerical Methods with Lévy Processes

Numerical Methods with Lévy Processes Numerical Methods with Lévy Processes 1 Objective: i) Find models of asset returns, etc ii) Get numbers out of them. Why? VaR and risk management Valuing and hedging derivatives Why not? Usual assumption:

More information

3 Continuous Random Variables

3 Continuous Random Variables Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random

More information

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014 Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions

More information

Construction of Dependent Dirichlet Processes based on Poisson Processes

Construction of Dependent Dirichlet Processes based on Poisson Processes 1 / 31 Construction of Dependent Dirichlet Processes based on Poisson Processes Dahua Lin Eric Grimson John Fisher CSAIL MIT NIPS 2010 Outstanding Student Paper Award Presented by Shouyuan Chen Outline

More information

Octavio Arizmendi, Ole E. Barndorff-Nielsen and Víctor Pérez-Abreu

Octavio Arizmendi, Ole E. Barndorff-Nielsen and Víctor Pérez-Abreu ON FREE AND CLASSICAL TYPE G DISTRIBUTIONS Octavio Arizmendi, Ole E. Barndorff-Nielsen and Víctor Pérez-Abreu Comunicación del CIMAT No I-9-4/3-4-29 (PE /CIMAT) On Free and Classical Type G Distributions

More information

Recitation 2: Probability

Recitation 2: Probability Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions

More information

15-388/688 - Practical Data Science: Basic probability. J. Zico Kolter Carnegie Mellon University Spring 2018

15-388/688 - Practical Data Science: Basic probability. J. Zico Kolter Carnegie Mellon University Spring 2018 15-388/688 - Practical Data Science: Basic probability J. Zico Kolter Carnegie Mellon University Spring 2018 1 Announcements Logistics of next few lectures Final project released, proposals/groups due

More information

HCM PROPERTY AND THE HALF-CAUCHY DISTRIBUTION PIERRE B O S C H (LILLE 1)

HCM PROPERTY AND THE HALF-CAUCHY DISTRIBUTION PIERRE B O S C H (LILLE 1) PROBABILITY AND MATHEMATICAL STATISTICS Vol. 35, Fasc. 2 25), pp. 9 2 HCM PROPERTY AND THE HALF-CAUCHY DISTRIBUTION BY PIERRE B O S C H LILLE ) Abstract. Let Z α and Z α be two independent positive α-stable

More information

CS Lecture 19. Exponential Families & Expectation Propagation

CS Lecture 19. Exponential Families & Expectation Propagation CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces

More information

Lecture 6 Basic Probability

Lecture 6 Basic Probability Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic

More information

Review of Probabilities and Basic Statistics

Review of Probabilities and Basic Statistics Alex Smola Barnabas Poczos TA: Ina Fiterau 4 th year PhD student MLD Review of Probabilities and Basic Statistics 10-701 Recitations 1/25/2013 Recitation 1: Statistics Intro 1 Overview Introduction to

More information

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours

MATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours MATH2750 This question paper consists of 8 printed pages, each of which is identified by the reference MATH275. All calculators must carry an approval sticker issued by the School of Mathematics. c UNIVERSITY

More information

DS-GA 1002 Lecture notes 2 Fall Random variables

DS-GA 1002 Lecture notes 2 Fall Random variables DS-GA 12 Lecture notes 2 Fall 216 1 Introduction Random variables Random variables are a fundamental tool in probabilistic modeling. They allow us to model numerical quantities that are uncertain: the

More information

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =

E[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) = Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of

More information

On the Truncation Error of a Superposed Gamma Process

On the Truncation Error of a Superposed Gamma Process On the Truncation Error of a Superposed Gamma Process Julyan Arbel and Igor Prünster Abstract Completely random measures (CRMs) form a key ingredient of a wealth of stochastic models, in particular in

More information

Expectation is a positive linear operator

Expectation is a positive linear operator Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 6: Expectation is a positive linear operator Relevant textbook passages: Pitman [3]: Chapter

More information

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero

STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Week 1 Quantitative Analysis of Financial Markets Distributions A

Week 1 Quantitative Analysis of Financial Markets Distributions A Week 1 Quantitative Analysis of Financial Markets Distributions A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October

More information

arxiv: v1 [math.st] 28 Feb 2015

arxiv: v1 [math.st] 28 Feb 2015 Are Gibbs type priors the most natural generalization of the Dirichlet process? P. De Blasi 1, S. Favaro 1, A. Lijoi 2, R.H. Mena 3, I. Prünster 1 and M. Ruggiero 1 1 Università degli Studi di Torino and

More information

Conditional densities, mass functions, and expectations

Conditional densities, mass functions, and expectations Conditional densities, mass functions, and expectations Jason Swanson April 22, 27 1 Discrete random variables Suppose that X is a discrete random variable with range {x 1, x 2, x 3,...}, and that Y is

More information

1: PROBABILITY REVIEW

1: PROBABILITY REVIEW 1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following

More information

1 Review of Probability

1 Review of Probability 1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x

More information

SECTION 9.2: ARITHMETIC SEQUENCES and PARTIAL SUMS

SECTION 9.2: ARITHMETIC SEQUENCES and PARTIAL SUMS (Chapter 9: Discrete Math) 9.11 SECTION 9.2: ARITHMETIC SEQUENCES and PARTIAL SUMS PART A: WHAT IS AN ARITHMETIC SEQUENCE? The following appears to be an example of an arithmetic (stress on the me ) sequence:

More information

Continuous Random Variables 1

Continuous Random Variables 1 Continuous Random Variables 1 STA 256: Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 32 Continuous Random Variables: The idea Probability is area

More information

Lesson 21 Not So Dramatic Quadratics

Lesson 21 Not So Dramatic Quadratics STUDENT MANUAL ALGEBRA II / LESSON 21 Lesson 21 Not So Dramatic Quadratics Quadratic equations are probably one of the most popular types of equations that you ll see in algebra. A quadratic equation has

More information

Subordinated Brownian Motion: Last Time the Process Reaches its Supremum

Subordinated Brownian Motion: Last Time the Process Reaches its Supremum Sankhyā : The Indian Journal of Statistics 25, Volume 77-A, Part, pp. 46-64 c 24, Indian Statistical Institute Subordinated Brownian Motion: Last Time the Process Reaches its Supremum Stergios B. Fotopoulos,

More information

Stationary distributions of non Gaussian Ornstein Uhlenbeck processes for beam halos

Stationary distributions of non Gaussian Ornstein Uhlenbeck processes for beam halos N Cufaro Petroni: CYCLOTRONS 2007 Giardini Naxos, 1 5 October, 2007 1 Stationary distributions of non Gaussian Ornstein Uhlenbeck processes for beam halos CYCLOTRONS 2007 Giardini Naxos, 1 5 October Nicola

More information

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017 CPSC 340: Machine Learning and Data Mining MLE and MAP Fall 2017 Assignment 3: Admin 1 late day to hand in tonight, 2 late days for Wednesday. Assignment 4: Due Friday of next week. Last Time: Multi-Class

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture

More information

STAT 3610: Review of Probability Distributions

STAT 3610: Review of Probability Distributions STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

= 2 x y 2. (1)

= 2 x y 2. (1) COMPLEX ANALYSIS PART 5: HARMONIC FUNCTIONS A Let me start by asking you a question. Suppose that f is an analytic function so that the CR-equation f/ z = 0 is satisfied. Let us write u and v for the real

More information

Two common difficul:es with HW 2: Problem 1c: v = r n ˆr

Two common difficul:es with HW 2: Problem 1c: v = r n ˆr Two common difficul:es with HW : Problem 1c: For what values of n does the divergence of v = r n ˆr diverge at the origin? In this context, diverge means becomes arbitrarily large ( goes to infinity ).

More information

Bayesian Nonparametric Models for Ranking Data

Bayesian Nonparametric Models for Ranking Data Bayesian Nonparametric Models for Ranking Data François Caron 1, Yee Whye Teh 1 and Brendan Murphy 2 1 Dept of Statistics, University of Oxford, UK 2 School of Mathematical Sciences, University College

More information

Lecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs

Lecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs Lecture Notes 3 Multiple Random Variables Joint, Marginal, and Conditional pmfs Bayes Rule and Independence for pmfs Joint, Marginal, and Conditional pdfs Bayes Rule and Independence for pdfs Functions

More information

Things to remember when learning probability distributions:

Things to remember when learning probability distributions: SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Chapter 5 continued. Chapter 5 sections

Chapter 5 continued. Chapter 5 sections Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

CSC321 Lecture 18: Learning Probabilistic Models

CSC321 Lecture 18: Learning Probabilistic Models CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling

More information

MATH 353 LECTURE NOTES: WEEK 1 FIRST ORDER ODES

MATH 353 LECTURE NOTES: WEEK 1 FIRST ORDER ODES MATH 353 LECTURE NOTES: WEEK 1 FIRST ORDER ODES J. WONG (FALL 2017) What did we cover this week? Basic definitions: DEs, linear operators, homogeneous (linear) ODEs. Solution techniques for some classes

More information

Stochastic Gradient Descent in Continuous Time

Stochastic Gradient Descent in Continuous Time Stochastic Gradient Descent in Continuous Time Justin Sirignano University of Illinois at Urbana Champaign with Konstantinos Spiliopoulos (Boston University) 1 / 27 We consider a diffusion X t X = R m

More information

Introduction to Applied Bayesian Modeling. ICPSR Day 4

Introduction to Applied Bayesian Modeling. ICPSR Day 4 Introduction to Applied Bayesian Modeling ICPSR Day 4 Simple Priors Remember Bayes Law: Where P(A) is the prior probability of A Simple prior Recall the test for disease example where we specified the

More information

Compound Random Measures

Compound Random Measures Compound Random Measures Jim Griffin (joint work with Fabrizio Leisen) University of Kent Introduction: Two clinical studies 3 CALGB8881 3 CALGB916 2 2 β 1 1 β 1 1 1 5 5 β 1 5 5 β Infinite mixture models

More information

Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures

Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures 17th Europ. Conf. on Machine Learning, Berlin, Germany, 2006. Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures Shipeng Yu 1,2, Kai Yu 2, Volker Tresp 2, and Hans-Peter

More information

3 Algebraic Methods. we can differentiate both sides implicitly to obtain a differential equation involving x and y:

3 Algebraic Methods. we can differentiate both sides implicitly to obtain a differential equation involving x and y: 3 Algebraic Methods b The first appearance of the equation E Mc 2 in Einstein s handwritten notes. So far, the only general class of differential equations that we know how to solve are directly integrable

More information

Random Variables (Continuous Case)

Random Variables (Continuous Case) Chapter 6 Random Variables (Continuous Case) Thus far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions

More information

Discrete Random Variables

Discrete Random Variables Discrete Random Variables We have a probability space (S, Pr). A random variable is a function X : S V (X ) for some set V (X ). In this discussion, we must have V (X ) is the real numbers X induces a

More information

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1

ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 Last compiled: November 6, 213 1. Conditional expectation Exercise 1.1. To start with, note that P(X Y = P( c R : X > c, Y c or X c, Y > c = P( c Q : X > c, Y

More information

The Central Limit Theorem: More of the Story

The Central Limit Theorem: More of the Story The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November 2015 1 / 33 Central Limit Theorem Theorem (Central Limit

More information

CPSC 340: Machine Learning and Data Mining

CPSC 340: Machine Learning and Data Mining CPSC 340: Machine Learning and Data Mining MLE and MAP Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due tonight. Assignment 5: Will be released

More information

Priors for Random Count Matrices with Random or Fixed Row Sums

Priors for Random Count Matrices with Random or Fixed Row Sums Priors for Random Count Matrices with Random or Fixed Row Sums Mingyuan Zhou Joint work with Oscar Madrid and James Scott IROM Department, McCombs School of Business Department of Statistics and Data Sciences

More information

MOMENTS AND CUMULANTS OF THE SUBORDINATED LEVY PROCESSES

MOMENTS AND CUMULANTS OF THE SUBORDINATED LEVY PROCESSES MOMENTS AND CUMULANTS OF THE SUBORDINATED LEVY PROCESSES PENKA MAYSTER ISET de Rades UNIVERSITY OF TUNIS 1 The formula of Faa di Bruno represents the n-th derivative of the composition of two functions

More information

BMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution

BMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random

More information

STAT 430/510: Lecture 15

STAT 430/510: Lecture 15 STAT 430/510: Lecture 15 James Piette June 23, 2010 Updates HW4 is up on my website. It is due next Mon. (June 28th). Starting today back at section 6.4... Conditional Distribution: Discrete Def: The conditional

More information

Conditional distributions

Conditional distributions Conditional distributions Will Monroe July 6, 017 with materials by Mehran Sahami and Chris Piech Independence of discrete random variables Two random variables are independent if knowing the value of

More information

REAL ANALYSIS I Spring 2016 Product Measures

REAL ANALYSIS I Spring 2016 Product Measures REAL ANALSIS I Spring 216 Product Measures We assume that (, M, µ), (, N, ν) are σ- finite measure spaces. We want to provide the Cartesian product with a measure space structure in which all sets of the

More information

4. CONTINUOUS RANDOM VARIABLES

4. CONTINUOUS RANDOM VARIABLES IA Probability Lent Term 4 CONTINUOUS RANDOM VARIABLES 4 Introduction Up to now we have restricted consideration to sample spaces Ω which are finite, or countable; we will now relax that assumption We

More information

Weighted Exponential Distribution and Process

Weighted Exponential Distribution and Process Weighted Exponential Distribution and Process Jilesh V Some generalizations of exponential distribution and related time series models Thesis. Department of Statistics, University of Calicut, 200 Chapter

More information

Real Analysis Problems

Real Analysis Problems Real Analysis Problems Cristian E. Gutiérrez September 14, 29 1 1 CONTINUITY 1 Continuity Problem 1.1 Let r n be the sequence of rational numbers and Prove that f(x) = 1. f is continuous on the irrationals.

More information

Meixner matrix ensembles

Meixner matrix ensembles Meixner matrix ensembles W lodek Bryc 1 Cincinnati April 12, 2011 1 Based on joint work with Gerard Letac W lodek Bryc (Cincinnati) Meixner matrix ensembles April 12, 2011 1 / 29 Outline of talk Random

More information

MAS113 Introduction to Probability and Statistics. Proofs of theorems

MAS113 Introduction to Probability and Statistics. Proofs of theorems MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a

More information

Nonparametric Bayesian Methods (Gaussian Processes)

Nonparametric Bayesian Methods (Gaussian Processes) [70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent

More information

Math 40510, Algebraic Geometry

Math 40510, Algebraic Geometry Math 40510, Algebraic Geometry Problem Set 1, due February 10, 2016 1. Let k = Z p, the field with p elements, where p is a prime. Find a polynomial f k[x, y] that vanishes at every point of k 2. [Hint:

More information

Simulating Random Variables

Simulating Random Variables Simulating Random Variables Timothy Hanson Department of Statistics, University of South Carolina Stat 740: Statistical Computing 1 / 23 R has many built-in random number generators... Beta, gamma (also

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite

More information

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Random variables. DS GA 1002 Probability and Statistics for Data Science. Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities

More information

Multivariate Distributions (Hogg Chapter Two)

Multivariate Distributions (Hogg Chapter Two) Multivariate Distributions (Hogg Chapter Two) STAT 45-1: Mathematical Statistics I Fall Semester 15 Contents 1 Multivariate Distributions 1 11 Random Vectors 111 Two Discrete Random Variables 11 Two Continuous

More information

Statistics 3657 : Moment Generating Functions

Statistics 3657 : Moment Generating Functions Statistics 3657 : Moment Generating Functions A useful tool for studying sums of independent random variables is generating functions. course we consider moment generating functions. In this Definition

More information

Introduction to Probabilistic Machine Learning

Introduction to Probabilistic Machine Learning Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning

More information

Bayesian Interpretations of Regularization

Bayesian Interpretations of Regularization Bayesian Interpretations of Regularization Charlie Frogner 9.50 Class 15 April 1, 009 The Plan Regularized least squares maps {(x i, y i )} n i=1 to a function that minimizes the regularized loss: f S

More information

STAT 830 Bayesian Estimation

STAT 830 Bayesian Estimation STAT 830 Bayesian Estimation Richard Lockhart Simon Fraser University STAT 830 Fall 2011 Richard Lockhart (Simon Fraser University) STAT 830 Bayesian Estimation STAT 830 Fall 2011 1 / 23 Purposes of These

More information

Chapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS

Chapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Chapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Gamma Distribution Weibull Distribution Lognormal Distribution Sections 4-9 through 4-11 Another exponential distribution example

More information