New Dirichlet Mean Identities
|
|
- Philippa Cain
- 6 years ago
- Views:
Transcription
1 Hong Kong University of Science and Technology Isaac Newton Institute, August 10, 2007
2 Origins CIFARELLI, D. M. and REGAZZINI, E. (1979). Considerazioni generali sull impostazione bayesiana di problemi non parametrici. Le medie associative nel contesto del processo aleatorio di Dirichlet I, II. Riv. Mat. Sci. Econom. Social 2, Later there is the important 1990 Annals of Statistics paper.
3 1 It is a cornerstone work in Bayesian NP and I believe one of the most important contributions in the theory, and indeed application, of random processes. 2 Why? Dirichlet means arise everywhere.
4 Some things to keep in mind Hard and very technical stuff But Cifarelli and Regazzini already did the hard work so we do not have to Our Task is really how to figure out how to use their results. Not so much on how to re-prove them. For some recent non-trivial applications look me up on the math Arxiv I never really do hard things. This is not a Joke
5 Theme Prendere due piccioni con una fava
6 Dirichlet process Let X be a non-negative random variable with cumulative distribution function F X. Note furthermore for a measurable set C, we use the notation F X (C) to mean the probability that X is in C. One may define a Dirichlet process random probability measure, say P θ, on [0, ) with total mass parameter θ and prior parameter F X, via its finite dimensional distribution as follows; for any disjoint partition on [0, ), say (C 1,..., C k ), the distribution of the random vector. (P θ (C 1 ),..., P θ (C k )) is a k-variate Dirichet distribution with parameters (θf X (C 1 ),..., θf X (C k )).
7 Hence for each C, P θ (C) = 0 I(x C)P θ (dx) has a beta distribution with parameters (θf X (C), θ(1 F X (C))).
8 Equivalently setting θf X (C i ) = θ i for i = 1,..., k, ( ) (P θ (C 1 ),..., P θ (C k )) = d Gθi ; i = 1,..., k G θ where (G θi ) are independent random variables with gamma(θ i, 1) distributions and G θ = G θ1 + + G θk has a gamma(θ, 1) distribution.
9 This means that one can define the Dirichlet process via the normalization of an independent increment gamma process on [0, ), say γ θ ( ), as P θ ( ) = γ θ( ) γ θ ([0, )) where γ θ (C i ) d = G θi and whose almost surely finite total random mass is γ θ ([0, )) d = G θ.
10 A very important aspect of this construction is the fact that G θ is independent of P θ, and hence any functional P θ (g) = 0 g(x)p θ (dx)
11 Furthermore θ > 0, G θ P θ (g) = d 0 g(x)γ θ (dx) = γ θ (g) See for instance Lijoi and Regazzini- AOP
12 Cifarelli and Regazzini/Markov Krein ID Now, E[e λg θp θ (g) ] = E[(1 + λp θ (g)) θ ] The Laplace transform of a random variable representable as γ θ (g) is equivalent to the Cauchy-Stieltjes transform of order θ of a random variable representable as P θ (g)
13 Something to keep in mind Lets think about random variables on the real line rather than random processes in some Banach space
14 A more general statement: Psychologically easier to prove???!!! Let M be a positive random variable independent of G θ, and define the random variable R θ d = Gθ M then from your first course in probability E[e λr θ ] = E[(1 + λm) θ ]
15 Question Are the results on the last two slides equivalent?
16 Answer-Questions γ θ (g) is always infinitely divisible Scale mixtures of gamma random variables are not always infinitely divisible. Scale mixtures of gamma random variables of index 0 < θ 1 are always infinitely divisible. If R θ is infinitely divisible is it true then?
17 Classical Question Recall that E[e λr θ ] = E[(1 + λm) θ ] In general :How to find the density of R θ? In general :How to find the density of M?
18 Classical but painful answer Utilize Classical Inversion formula You have to deal with the Analysis of Complex functions No simpler answer in general
19 Cifarelli and Regazzini provide answers for M d = P θ (g)
20 My Solutions Call up my Italian friends!! (Lijoi and Prünster) Build on existing results
21 Important? γ θ (g) constitute a large class of Infinitely divisible random variables. Id random variables generate EVERY positive Lévy process Levy processes are now the building blocks in many diverse fields, physics, finance, genetics, machine learning. All these models can be treated from a Bayesian NP point of view Levy processes are connected with special functions One wants to get the laws of certain functionals, BUT this is hard Simple example-finance: One may want to use a Barndorff-Nielsen OU SV process for option pricing without resorting to a series representation.
22 some more 1 Many important quantities can be represented as P θ (g) 2 Curiously James, Lijoi and Pruenster[JLP], show that for θ > 0 every linear functional of a two parameter (α, θ) Poisson Dirichlet random probability measure can be represented as a Dirichlet mean of order θ.
23 Define a Dirichlet mean of order θ indexed by F X as M θ (F X ) = 0 xp θ (dx)
24 Say that d T θ = Gθ M θ (F X ) is a GGC(θ, F X ) random variable which satisfies E[e λt θ ] = E[(1 + λm θ (F X )) θ ] = e θψ F X (λ)
25 where ψ FX (λ) = 0 log(1 + λx)f X (dx) = E[log(1 + λx)].
26 Key expressions for density Cifarelli and Regazzini(1990) 1 Φ FX (t) = 2 furthermore, define, 0 log( t x )I(t x)f X (dx) = E[log( t X )I(t X)] θ (t F X ) = 1 π sin(πθf X(t))e θφ F X (t).
27 Then from [CR] 1 The cdf of M θ (F X ) for all θ > 0 is expressible as x 2 when θ = 1, the density is, 0 (x t) θ 1 θ (t F X )dt ξ FX (x) = 1 (x F X ) = 1 π sin(πf X(x))e Φ F (x) X. 3 Density formulae for θ > 1 are described as ξ θfx (x) = (θ 1) x 0 (x t) θ 2 θ (t F X )dt.
28 Result James, Lijoi and Prünster(2006) Recognizing a connection of [CR] s result for the cdf to Abel transforms, [JLP] reasoned that it is rather straightforward to establish the following 1 for all θ > 0 ξ θfx (x) = 2 is the derivative of x 0 (x t) θ 1 (t F X )dt
29 In general the only really nice expression for the density is when θ = 1 ξ FX (x) = 1 (x F X ) = 1 π sin(πf X(x))e Φ F X (x).
30 But notice that θ (t F X ) = 1 π sin(πθf X(t))e θφ F X (t) > 0. for 0 < θ 1 and all t > 0.
31 Comments on possible random variables that are Dirichlet means 1 The formulae we just saw are general expressions and don t look too familiar 2 However, we (James, Roynette and Yor-Coming Soon) show that there are many many familiar random variables that are Dirichlet means 3 This includes Generalized Inverse Gaussian, Positive Stable, Pareto, Uniform and at the very least a large class of random variables that are generalized gamma convolutions 4 Note if for instance U is uniform[0,1] then for 0 < θ 2, there exist F Xθ such that U d = M θ (F Xθ )
32 A useful identity: Hjort and Ongaro 05 Let D = (D 1,..., D k ) denote a Dirichlet (θ 1,..., θ k ) random vector such that θ = k i=1 θ i. Then M θ (F X ) d = k D i M θi (F X ) i=1 where M θi (F X ) are independent and independent of D
33 Goal Use the analytic work used to obtain results for M θ (F X ) to obtain results for an entire family of Dirichlet mean random variables {M θ (F Zc ) : c > 0} As a byproduct we get results for T θ and Lévy processes derived from T θ
34 Simple examples are of course the choices Z c = X + c and Z c = cx, which, due to the linearity properties of mean functionals, results easily in the identities in law M θ (F X+c ) = c + M θ (F X ) and M θ (F cx ) = cm θ (F X )
35 General DP Mean Identities I introduced two simple ideas to address some of these points. The idea to construct such things was in part influenced quite a bit by some results of Pitman and Yor (1997) On length of excursions of bessel processes and others. Concerning multiplication by beta random variables and so on. The two results can be described as 1 Beta Scaling of DP Mean functionals 2 Gamma Tilting 3 Actually there are two more
36 Beta Scaling of DP mean functionals Let 0 < σ 1 and θ > 0. Let β θσ,θ(1 σ) denote a beta random variable. Then consider the following simple random variable β θσ,θ(1 σ) M θσ (F X ) Remember the density of M θσ (F X ) is not NICE except for θσ = 1.
37 Let Y σ denote a Bernoulli random variable with success probability σ. Consider the independent product XY σ with cdf F XYσ (x) = σf X (x) + (1 σ)i(x 0)
38 Result 1: change of total mass Then β θσ,θ(1 σ) M θσ (F X ) d = M θ (F XYσ )
39 Why? ψ FXYσ (λ) = but is equal to 0 σψ FX (λ) = σ log(1 + λw)f XYσ (dw) = E[log(1 + λxy σ )]. 0 log(1 + λx)f X (dx) = σe[log(1 + λx)].
40 One to many For every fixed θ how many mean functionals of order θ did we just create? Answer: An uncountable number indexed by 0 < σ 1 (M θ (F XYσ ) : 0 < σ 1)
41 One to many explicit densities How many mean functionals of order θ = 1 did we just create? Answer: An uncountable number indexed by 0 < σ 1 (M 1 (F XYσ ) : 0 < σ 1)
42 Φ FXYσ (x) = E[log( x XY σ )I(XY σ x)] = σφ FX (x) + (1 σ) log(x)
43 Result 2 The density of β σ,1 σ M σ (F X ) d = M 1 (F XYσ ) is ξ FXYσ (x) = xσ 1 π sin(πf XY σ (x))e σφ F X (x) Note you do not need to know the density of M σ (F X ).
44 Back to GGC Recall but Hence T θσ = G θσ M θσ (F X ) G θσ d = Gθ [β θσ,θ(1 σ) ] T θσ = G θσ M θσ (F X ) = G θ [β θσ,θ(1 σ) M θσ (F X )]
45 Hence i.e. it is a GGC(θ, F XYσ ) T θσ = G θσ M θσ (F X ) = G θ M θ (F XYσ )
46 Hence i.e. it is a GGC(1, F XYσ ) T σ = G σ M σ (F X ) = G 1 M 1 (F XYσ )
47 Exploiting infinite divisibility Let 0 < σ k = 1 Then k=1 T 1 d = k=1 T σk where T σk are independent and each has distribution d T σk = G1 M 1 (F XYσk )
48 GGC species sampling models Let (Z k ) denote iid random elements with some common distribution H. Then set P k = T σk /T 1 Define a random probability measure as, P k δ Zk (dx) = P(dx) k=1 THESE ARE NOT NRMI
49 That is, for any finite k (P 1,..., P k ) d = (G 1,i M 1 (F XYσi )/[T + S]; i = 1,..., k)
50 where T = and for σ k = 1 k i=1 σ i k G 1,i M 1 (F XYσi ) i=1 S d = G 1 M 1 (F XYσ k ) independent of T.
51 For fun one could choose for i = 1, 2,... σ i = λ i 1 e λ /(i 1)!
52 Fidi of GGC NRM Let µ( ) denote a completely random measure derived from T 1, such that E[µ(A)] = H(A) and µ(x ) = T 1 Then for any A, set H(A) = σ, and one has that µ(a) = G 1 M 1 (F XYσ ) moreover for a disjoint partition of some space X, (C 1,..., C N ) The fidi of µ( ) is composed of independent components so that µ(c i ) = G 1 M 1 (F XYσi )
53 but writing T 1 = N µ(c i ) k=1 This gives the fidi of the random probability measure, P( ) = µ( )/T 1 i.e. (P(C 1 ),..., P(C N ).) These NRM have appeared in James Lijoi and Pruenster (2005) but not the fidi part.
54 Now conditioning on T 1 = t, this gives the fidi of the corresponding conditional Poisson Kingman class, L((P(C 1 ),..., P(C N ) T 1 = t) Now mix over an arbitrary density γ(t). Hence I just created an uncountable number of random probability measures with explicit fidi s.
55 Approximating GGC NRM s Now suppose that for general θ, let µ θ ( ) denote a CRM derived from T θ. Then form the NRM by, P θ ( ) = µ θ ( )/T θ Now write T θ = N k=1 T θ/n,k where T θ/n,k are iid equal in distribution to T θ/n
56 Now we are interested in choosing N > θ, in this case the joint distribution of (T θ/n,k /T θ ; k N) can be computed and certainly easily simulated. Since Furthermore as N T θ/n d = G1 M 1 (F XYθ/N ) N k=1 T θ/n,k T θ δ Zk ( ) P θ ( ) We can also use the left hand side as a sieve by letting N grow with the data.
57 Special Known Case G θ/n,k iid gamma(θ/n, 1) then N k=1 where D θ is a Dirichlet process G θ/n,k G θ δ Zk ( ) D θ ( )
On the posterior structure of NRMI
On the posterior structure of NRMI Igor Prünster University of Turin, Collegio Carlo Alberto and ICER Joint work with L.F. James and A. Lijoi Isaac Newton Institute, BNR Programme, 8th August 2007 Outline
More informationTruncation error of a superposed gamma process in a decreasing order representation
Truncation error of a superposed gamma process in a decreasing order representation B julyan.arbel@inria.fr Í www.julyanarbel.com Inria, Mistis, Grenoble, France Joint work with Igor Pru nster (Bocconi
More informationDependent hierarchical processes for multi armed bandits
Dependent hierarchical processes for multi armed bandits Federico Camerlenghi University of Bologna, BIDSA & Collegio Carlo Alberto First Italian meeting on Probability and Mathematical Statistics, Torino
More informationA marginal sampler for σ-stable Poisson-Kingman mixture models
A marginal sampler for σ-stable Poisson-Kingman mixture models joint work with Yee Whye Teh and Stefano Favaro María Lomelí Gatsby Unit, University College London Talk at the BNP 10 Raleigh, North Carolina
More informationDiscussion of On simulation and properties of the stable law by L. Devroye and L. James
Stat Methods Appl (2014) 23:371 377 DOI 10.1007/s10260-014-0269-4 Discussion of On simulation and properties of the stable law by L. Devroye and L. James Antonio Lijoi Igor Prünster Accepted: 16 May 2014
More informationAsymptotics for posterior hazards
Asymptotics for posterior hazards Pierpaolo De Blasi University of Turin 10th August 2007, BNR Workshop, Isaac Newton Intitute, Cambridge, UK Joint work with Giovanni Peccati (Université Paris VI) and
More informationNormalized kernel-weighted random measures
Normalized kernel-weighted random measures Jim Griffin University of Kent 1 August 27 Outline 1 Introduction 2 Ornstein-Uhlenbeck DP 3 Generalisations Bayesian Density Regression We observe data (x 1,
More informationBayesian Nonparametrics: some contributions to construction and properties of prior distributions
Bayesian Nonparametrics: some contributions to construction and properties of prior distributions Annalisa Cerquetti Collegio Nuovo, University of Pavia, Italy Interview Day, CETL Lectureship in Statistics,
More informationOn Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation
On Consistency of Nonparametric Normal Mixtures for Bayesian Density Estimation Antonio LIJOI, Igor PRÜNSTER, andstepheng.walker The past decade has seen a remarkable development in the area of Bayesian
More informationBayesian nonparametric models for bipartite graphs
Bayesian nonparametric models for bipartite graphs François Caron Department of Statistics, Oxford Statistics Colloquium, Harvard University November 11, 2013 F. Caron 1 / 27 Bipartite networks Readers/Customers
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationTruncation error of a superposed gamma process in a decreasing order representation
Truncation error of a superposed gamma process in a decreasing order representation Julyan Arbel Inria Grenoble, Université Grenoble Alpes julyan.arbel@inria.fr Igor Prünster Bocconi University, Milan
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More informationOn Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process
On Simulations form the Two-Parameter arxiv:1209.5359v1 [stat.co] 24 Sep 2012 Poisson-Dirichlet Process and the Normalized Inverse-Gaussian Process Luai Al Labadi and Mahmoud Zarepour May 8, 2018 ABSTRACT
More informationA Brief Overview of Nonparametric Bayesian Models
A Brief Overview of Nonparametric Bayesian Models Eurandom Zoubin Ghahramani Department of Engineering University of Cambridge, UK zoubin@eng.cam.ac.uk http://learning.eng.cam.ac.uk/zoubin Also at Machine
More informationFoundations of Nonparametric Bayesian Methods
1 / 27 Foundations of Nonparametric Bayesian Methods Part II: Models on the Simplex Peter Orbanz http://mlg.eng.cam.ac.uk/porbanz/npb-tutorial.html 2 / 27 Tutorial Overview Part I: Basics Part II: Models
More informationProbability Distributions Columns (a) through (d)
Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)
More informationBayesian Nonparametrics: Dirichlet Process
Bayesian Nonparametrics: Dirichlet Process Yee Whye Teh Gatsby Computational Neuroscience Unit, UCL http://www.gatsby.ucl.ac.uk/~ywteh/teaching/npbayes2012 Dirichlet Process Cornerstone of modern Bayesian
More informationNorthwestern University Department of Electrical Engineering and Computer Science
Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability
More information1 Probability and Random Variables
1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in
More informationSlice sampling σ stable Poisson Kingman mixture models
ISSN 2279-9362 Slice sampling σ stable Poisson Kingman mixture models Stefano Favaro S.G. Walker No. 324 December 2013 www.carloalberto.org/research/working-papers 2013 by Stefano Favaro and S.G. Walker.
More informationStat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory. Charles J. Geyer School of Statistics University of Minnesota
Stat 5101 Lecture Slides: Deck 7 Asymptotics, also called Large Sample Theory Charles J. Geyer School of Statistics University of Minnesota 1 Asymptotic Approximation The last big subject in probability
More informationBayesian nonparametric latent feature models
Bayesian nonparametric latent feature models Indian Buffet process, beta process, and related models François Caron Department of Statistics, Oxford Applied Bayesian Statistics Summer School Como, Italy
More informationDependent mixture models: clustering and borrowing information
ISSN 2279-9362 Dependent mixture models: clustering and borrowing information Antonio Lijoi Bernardo Nipoti Igor Pruenster No. 32 June 213 www.carloalberto.org/research/working-papers 213 by Antonio Lijoi,
More informationThe Poisson-Dirichlet Distribution: Constructions, Stochastic Dynamics and Asymptotic Behavior
The Poisson-Dirichlet Distribution: Constructions, Stochastic Dynamics and Asymptotic Behavior Shui Feng McMaster University June 26-30, 2011. The 8th Workshop on Bayesian Nonparametrics, Veracruz, Mexico.
More informationNumerical Methods with Lévy Processes
Numerical Methods with Lévy Processes 1 Objective: i) Find models of asset returns, etc ii) Get numbers out of them. Why? VaR and risk management Valuing and hedging derivatives Why not? Usual assumption:
More information3 Continuous Random Variables
Jinguo Lian Math437 Notes January 15, 016 3 Continuous Random Variables Remember that discrete random variables can take only a countable number of possible values. On the other hand, a continuous random
More informationProbability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014
Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions
More informationConstruction of Dependent Dirichlet Processes based on Poisson Processes
1 / 31 Construction of Dependent Dirichlet Processes based on Poisson Processes Dahua Lin Eric Grimson John Fisher CSAIL MIT NIPS 2010 Outstanding Student Paper Award Presented by Shouyuan Chen Outline
More informationOctavio Arizmendi, Ole E. Barndorff-Nielsen and Víctor Pérez-Abreu
ON FREE AND CLASSICAL TYPE G DISTRIBUTIONS Octavio Arizmendi, Ole E. Barndorff-Nielsen and Víctor Pérez-Abreu Comunicación del CIMAT No I-9-4/3-4-29 (PE /CIMAT) On Free and Classical Type G Distributions
More informationRecitation 2: Probability
Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions
More information15-388/688 - Practical Data Science: Basic probability. J. Zico Kolter Carnegie Mellon University Spring 2018
15-388/688 - Practical Data Science: Basic probability J. Zico Kolter Carnegie Mellon University Spring 2018 1 Announcements Logistics of next few lectures Final project released, proposals/groups due
More informationHCM PROPERTY AND THE HALF-CAUCHY DISTRIBUTION PIERRE B O S C H (LILLE 1)
PROBABILITY AND MATHEMATICAL STATISTICS Vol. 35, Fasc. 2 25), pp. 9 2 HCM PROPERTY AND THE HALF-CAUCHY DISTRIBUTION BY PIERRE B O S C H LILLE ) Abstract. Let Z α and Z α be two independent positive α-stable
More informationCS Lecture 19. Exponential Families & Expectation Propagation
CS 6347 Lecture 19 Exponential Families & Expectation Propagation Discrete State Spaces We have been focusing on the case of MRFs over discrete state spaces Probability distributions over discrete spaces
More informationLecture 6 Basic Probability
Lecture 6: Basic Probability 1 of 17 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 6 Basic Probability Probability spaces A mathematical setup behind a probabilistic
More informationReview of Probabilities and Basic Statistics
Alex Smola Barnabas Poczos TA: Ina Fiterau 4 th year PhD student MLD Review of Probabilities and Basic Statistics 10-701 Recitations 1/25/2013 Recitation 1: Statistics Intro 1 Overview Introduction to
More informationMATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours
MATH2750 This question paper consists of 8 printed pages, each of which is identified by the reference MATH275. All calculators must carry an approval sticker issued by the School of Mathematics. c UNIVERSITY
More informationDS-GA 1002 Lecture notes 2 Fall Random variables
DS-GA 12 Lecture notes 2 Fall 216 1 Introduction Random variables Random variables are a fundamental tool in probabilistic modeling. They allow us to model numerical quantities that are uncertain: the
More informationE[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =
Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of
More informationOn the Truncation Error of a Superposed Gamma Process
On the Truncation Error of a Superposed Gamma Process Julyan Arbel and Igor Prünster Abstract Completely random measures (CRMs) form a key ingredient of a wealth of stochastic models, in particular in
More informationExpectation is a positive linear operator
Department of Mathematics Ma 3/103 KC Border Introduction to Probability and Statistics Winter 2017 Lecture 6: Expectation is a positive linear operator Relevant textbook passages: Pitman [3]: Chapter
More informationSTATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero
STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationWeek 1 Quantitative Analysis of Financial Markets Distributions A
Week 1 Quantitative Analysis of Financial Markets Distributions A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October
More informationarxiv: v1 [math.st] 28 Feb 2015
Are Gibbs type priors the most natural generalization of the Dirichlet process? P. De Blasi 1, S. Favaro 1, A. Lijoi 2, R.H. Mena 3, I. Prünster 1 and M. Ruggiero 1 1 Università degli Studi di Torino and
More informationConditional densities, mass functions, and expectations
Conditional densities, mass functions, and expectations Jason Swanson April 22, 27 1 Discrete random variables Suppose that X is a discrete random variable with range {x 1, x 2, x 3,...}, and that Y is
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More information1 Review of Probability
1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x
More informationSECTION 9.2: ARITHMETIC SEQUENCES and PARTIAL SUMS
(Chapter 9: Discrete Math) 9.11 SECTION 9.2: ARITHMETIC SEQUENCES and PARTIAL SUMS PART A: WHAT IS AN ARITHMETIC SEQUENCE? The following appears to be an example of an arithmetic (stress on the me ) sequence:
More informationContinuous Random Variables 1
Continuous Random Variables 1 STA 256: Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 32 Continuous Random Variables: The idea Probability is area
More informationLesson 21 Not So Dramatic Quadratics
STUDENT MANUAL ALGEBRA II / LESSON 21 Lesson 21 Not So Dramatic Quadratics Quadratic equations are probably one of the most popular types of equations that you ll see in algebra. A quadratic equation has
More informationSubordinated Brownian Motion: Last Time the Process Reaches its Supremum
Sankhyā : The Indian Journal of Statistics 25, Volume 77-A, Part, pp. 46-64 c 24, Indian Statistical Institute Subordinated Brownian Motion: Last Time the Process Reaches its Supremum Stergios B. Fotopoulos,
More informationStationary distributions of non Gaussian Ornstein Uhlenbeck processes for beam halos
N Cufaro Petroni: CYCLOTRONS 2007 Giardini Naxos, 1 5 October, 2007 1 Stationary distributions of non Gaussian Ornstein Uhlenbeck processes for beam halos CYCLOTRONS 2007 Giardini Naxos, 1 5 October Nicola
More informationCPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017
CPSC 340: Machine Learning and Data Mining MLE and MAP Fall 2017 Assignment 3: Admin 1 late day to hand in tonight, 2 late days for Wednesday. Assignment 4: Due Friday of next week. Last Time: Multi-Class
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.
Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Introduction to Markov chain Monte Carlo The Gibbs Sampler Examples Overview of the Lecture
More informationSTAT 3610: Review of Probability Distributions
STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More information= 2 x y 2. (1)
COMPLEX ANALYSIS PART 5: HARMONIC FUNCTIONS A Let me start by asking you a question. Suppose that f is an analytic function so that the CR-equation f/ z = 0 is satisfied. Let us write u and v for the real
More informationTwo common difficul:es with HW 2: Problem 1c: v = r n ˆr
Two common difficul:es with HW : Problem 1c: For what values of n does the divergence of v = r n ˆr diverge at the origin? In this context, diverge means becomes arbitrarily large ( goes to infinity ).
More informationBayesian Nonparametric Models for Ranking Data
Bayesian Nonparametric Models for Ranking Data François Caron 1, Yee Whye Teh 1 and Brendan Murphy 2 1 Dept of Statistics, University of Oxford, UK 2 School of Mathematical Sciences, University College
More informationLecture Notes 3 Multiple Random Variables. Joint, Marginal, and Conditional pmfs. Bayes Rule and Independence for pmfs
Lecture Notes 3 Multiple Random Variables Joint, Marginal, and Conditional pmfs Bayes Rule and Independence for pmfs Joint, Marginal, and Conditional pdfs Bayes Rule and Independence for pdfs Functions
More informationThings to remember when learning probability distributions:
SPECIAL DISTRIBUTIONS Some distributions are special because they are useful They include: Poisson, exponential, Normal (Gaussian), Gamma, geometric, negative binomial, Binomial and hypergeometric distributions
More informationGaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008
Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationCSC321 Lecture 18: Learning Probabilistic Models
CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling
More informationMATH 353 LECTURE NOTES: WEEK 1 FIRST ORDER ODES
MATH 353 LECTURE NOTES: WEEK 1 FIRST ORDER ODES J. WONG (FALL 2017) What did we cover this week? Basic definitions: DEs, linear operators, homogeneous (linear) ODEs. Solution techniques for some classes
More informationStochastic Gradient Descent in Continuous Time
Stochastic Gradient Descent in Continuous Time Justin Sirignano University of Illinois at Urbana Champaign with Konstantinos Spiliopoulos (Boston University) 1 / 27 We consider a diffusion X t X = R m
More informationIntroduction to Applied Bayesian Modeling. ICPSR Day 4
Introduction to Applied Bayesian Modeling ICPSR Day 4 Simple Priors Remember Bayes Law: Where P(A) is the prior probability of A Simple prior Recall the test for disease example where we specified the
More informationCompound Random Measures
Compound Random Measures Jim Griffin (joint work with Fabrizio Leisen) University of Kent Introduction: Two clinical studies 3 CALGB8881 3 CALGB916 2 2 β 1 1 β 1 1 1 5 5 β 1 5 5 β Infinite mixture models
More informationVariational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures
17th Europ. Conf. on Machine Learning, Berlin, Germany, 2006. Variational Bayesian Dirichlet-Multinomial Allocation for Exponential Family Mixtures Shipeng Yu 1,2, Kai Yu 2, Volker Tresp 2, and Hans-Peter
More information3 Algebraic Methods. we can differentiate both sides implicitly to obtain a differential equation involving x and y:
3 Algebraic Methods b The first appearance of the equation E Mc 2 in Einstein s handwritten notes. So far, the only general class of differential equations that we know how to solve are directly integrable
More informationRandom Variables (Continuous Case)
Chapter 6 Random Variables (Continuous Case) Thus far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions
More informationDiscrete Random Variables
Discrete Random Variables We have a probability space (S, Pr). A random variable is a function X : S V (X ) for some set V (X ). In this discussion, we must have V (X ) is the real numbers X induces a
More informationADVANCED PROBABILITY: SOLUTIONS TO SHEET 1
ADVANCED PROBABILITY: SOLUTIONS TO SHEET 1 Last compiled: November 6, 213 1. Conditional expectation Exercise 1.1. To start with, note that P(X Y = P( c R : X > c, Y c or X c, Y > c = P( c Q : X > c, Y
More informationThe Central Limit Theorem: More of the Story
The Central Limit Theorem: More of the Story Steven Janke November 2015 Steven Janke (Seminar) The Central Limit Theorem:More of the Story November 2015 1 / 33 Central Limit Theorem Theorem (Central Limit
More informationCPSC 340: Machine Learning and Data Mining
CPSC 340: Machine Learning and Data Mining MLE and MAP Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due tonight. Assignment 5: Will be released
More informationPriors for Random Count Matrices with Random or Fixed Row Sums
Priors for Random Count Matrices with Random or Fixed Row Sums Mingyuan Zhou Joint work with Oscar Madrid and James Scott IROM Department, McCombs School of Business Department of Statistics and Data Sciences
More informationMOMENTS AND CUMULANTS OF THE SUBORDINATED LEVY PROCESSES
MOMENTS AND CUMULANTS OF THE SUBORDINATED LEVY PROCESSES PENKA MAYSTER ISET de Rades UNIVERSITY OF TUNIS 1 The formula of Faa di Bruno represents the n-th derivative of the composition of two functions
More informationBMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution
Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random
More informationSTAT 430/510: Lecture 15
STAT 430/510: Lecture 15 James Piette June 23, 2010 Updates HW4 is up on my website. It is due next Mon. (June 28th). Starting today back at section 6.4... Conditional Distribution: Discrete Def: The conditional
More informationConditional distributions
Conditional distributions Will Monroe July 6, 017 with materials by Mehran Sahami and Chris Piech Independence of discrete random variables Two random variables are independent if knowing the value of
More informationREAL ANALYSIS I Spring 2016 Product Measures
REAL ANALSIS I Spring 216 Product Measures We assume that (, M, µ), (, N, ν) are σ- finite measure spaces. We want to provide the Cartesian product with a measure space structure in which all sets of the
More information4. CONTINUOUS RANDOM VARIABLES
IA Probability Lent Term 4 CONTINUOUS RANDOM VARIABLES 4 Introduction Up to now we have restricted consideration to sample spaces Ω which are finite, or countable; we will now relax that assumption We
More informationWeighted Exponential Distribution and Process
Weighted Exponential Distribution and Process Jilesh V Some generalizations of exponential distribution and related time series models Thesis. Department of Statistics, University of Calicut, 200 Chapter
More informationReal Analysis Problems
Real Analysis Problems Cristian E. Gutiérrez September 14, 29 1 1 CONTINUITY 1 Continuity Problem 1.1 Let r n be the sequence of rational numbers and Prove that f(x) = 1. f is continuous on the irrationals.
More informationMeixner matrix ensembles
Meixner matrix ensembles W lodek Bryc 1 Cincinnati April 12, 2011 1 Based on joint work with Gerard Letac W lodek Bryc (Cincinnati) Meixner matrix ensembles April 12, 2011 1 / 29 Outline of talk Random
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More informationNonparametric Bayesian Methods (Gaussian Processes)
[70240413 Statistical Machine Learning, Spring, 2015] Nonparametric Bayesian Methods (Gaussian Processes) Jun Zhu dcszj@mail.tsinghua.edu.cn http://bigml.cs.tsinghua.edu.cn/~jun State Key Lab of Intelligent
More informationMath 40510, Algebraic Geometry
Math 40510, Algebraic Geometry Problem Set 1, due February 10, 2016 1. Let k = Z p, the field with p elements, where p is a prime. Find a polynomial f k[x, y] that vanishes at every point of k 2. [Hint:
More informationSimulating Random Variables
Simulating Random Variables Timothy Hanson Department of Statistics, University of South Carolina Stat 740: Statistical Computing 1 / 23 R has many built-in random number generators... Beta, gamma (also
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationCS 361: Probability & Statistics
October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite
More informationRandom variables. DS GA 1002 Probability and Statistics for Data Science.
Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities
More informationMultivariate Distributions (Hogg Chapter Two)
Multivariate Distributions (Hogg Chapter Two) STAT 45-1: Mathematical Statistics I Fall Semester 15 Contents 1 Multivariate Distributions 1 11 Random Vectors 111 Two Discrete Random Variables 11 Two Continuous
More informationStatistics 3657 : Moment Generating Functions
Statistics 3657 : Moment Generating Functions A useful tool for studying sums of independent random variables is generating functions. course we consider moment generating functions. In this Definition
More informationIntroduction to Probabilistic Machine Learning
Introduction to Probabilistic Machine Learning Piyush Rai Dept. of CSE, IIT Kanpur (Mini-course 1) Nov 03, 2015 Piyush Rai (IIT Kanpur) Introduction to Probabilistic Machine Learning 1 Machine Learning
More informationBayesian Interpretations of Regularization
Bayesian Interpretations of Regularization Charlie Frogner 9.50 Class 15 April 1, 009 The Plan Regularized least squares maps {(x i, y i )} n i=1 to a function that minimizes the regularized loss: f S
More informationSTAT 830 Bayesian Estimation
STAT 830 Bayesian Estimation Richard Lockhart Simon Fraser University STAT 830 Fall 2011 Richard Lockhart (Simon Fraser University) STAT 830 Bayesian Estimation STAT 830 Fall 2011 1 / 23 Purposes of These
More informationChapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
Chapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Gamma Distribution Weibull Distribution Lognormal Distribution Sections 4-9 through 4-11 Another exponential distribution example
More information