Advanced Sequential Monte Carlo Methods

Size: px
Start display at page:

Download "Advanced Sequential Monte Carlo Methods"

Transcription

1 Advaced Sequetial Mote Carlo Methods Araud Doucet Departmets of Statistics & Computer Sciece Uiversity of British Columbia A.D. () / 35

2 Geeric Sequetial Mote Carlo Scheme At time =, sample q () ad set w Resample o, W (i) to obtai ew particles also deoted At time 2 sample X (i ) q j X (i ) : compute w X (i ) Resample : = γ : γ : q. : o :, W (i) to obtai ew particles also deoted = γ. q : o o A.D. () 2 / 35

3 Sequetial Mote Carlo for Hidde Markov Models At time =, sample q () ad set w = µ g y j Resample q, W (i). y At time 2 sample X (i ) q j y, X (i ) compute w X (i ) : = f X (i) q Resample o to obtai ew particles also deoted g y jx (i) X (i) y,. o :, W (i) to obtai ew particles also deoted : o o A.D. () 3 / 35

4 Resamplig ca drastically improve the performace of SIS i models havig good mixig properties; e.g. state-space models: this ca be veri ed experimetally ad theoretically. Resamplig does ot solve all our problems; oly the SMC approximatios of the most recet margials π (x L+: ) are reliable; i.e. we ca have uiform (i time) covergece bouds. A.D. () 4 / 35

5 A Limited Framework? It seems that there is ot much to do to improve over this SMC scheme. Give a sequece of distributios π (x : ), use your favourite resamplig scheme ad the oly degree of freedom is essetially q (x j x : ). We kow that the best choice is q opt (x j x : ) = π (x j x : ) so how ca we do ay better??? Aswer: Modify the sequece of target distributios ad the associated proposals i a sesible way. A.D. () 4 / 35

6 Advaced SMC Methods Auxiliary particle lter Resample-move algorithm Block samplig strategy A.D. () 5 / 35

7 Auxiliary Particle Filter This is a very popular strategy itroduced by Pitt & Shephard (999). It was origially itroduced usig auxiliary variables but presetatio here is completely di eret... Iitial Remark: The stadard SMC algorithm appears very ie ciet whe q (x j y, x ) = p (x j y, x ) ) w (x : ) = p (y j x ) A.D. () 6 / 35

8 Stadard Algorithm vs Alterative Strategy Stadard Strategy Sample X (i) p j y, ad compute w : = p y j. o Resample :, W (i) to obtai ew particles also deoted Alterative Strategy Compute w : = p o y j. Resample :, W (i) to obtai ew particles also deoted o : approx. dist. p (x : j y : ). Sample X (i) p j y, : We swap the samplig ad resamplig steps; this yields more diverse particles at time ; hece ituitively a better estimate. A.D. () 7 / 35 o.

9 Auxiliary Particle Filter We ca oly swap the samplig ad resamplig steps whe q (x j y, x ) = p (x j y, x ) as the resultig weight is idepedet of x ; i.e. we have w (x : ) = p (y j x ). If we caot sample from p (x j y, x ) ad/or do ot kow p (y j x ), we ca simply propose to use approximatios bp (x j y, x ) ad bp (y j x ) ad correct for the bias. The APF is thus essetially a look-ahead strategy where o we try to aticipate the quality of our curret particles : with respect to y. A.D. () 8 / 35

10 Algorithm Assumig you have W (i) o, : approx. p (x : j y : ). Compute W (i) W (i) bp y j. o Resample :, W (i) to obtai ew particles also deoted o : approx. dist. ep (x : j y : ) p (x : j y : ) bp (y j x ). Sample X (i) bp j y, ad reweight W (i) p : y : p : y : bp y j bp X (i) y, f X (i) g y j X (i) bp y j bp X (i) y, A.D. () 9 / 35

11 Iterpretatio as a stadard SMC algorithm It is easy to check that this algorithm is othig but a stadard SMC for π (x : ) p (x : j y : ) bp (y + j x ) We do ot target p (x : j y : ) directly so it is ecessary to use IS to correct for the discrepacy betwee this target ad π (x : ) bp (x j y, x ) the distributio of the particles obtaied after the samplig step W (i) p : p y : bp : y : y j bp y, All the covergece results for stadard SMC ca thus straightforwardly be exteded to the APF. Perhaps surprisigly, the APF does ot domiate uiformly the stadard SMC scheme eve if q (x j y, x ) = p (x j y, x ). This is because it is just a oe-step optimizatio procedure. A.D. () / 35

12 Desig Issues I the literature, it is ofte suggested to approximate Z p (y j x ) = g (y j x ) f (x j x ) dx via bp (y j x ) = g (y j µ (x )) where µ (x ) is the mode, mea or media of f (x j x ). Typically, people ted to build a approximatio bp (x j y, x ) idepedetly of the approximatio bp (y j x ). A simpler ad better way cosists of buildig a approximatio bp (x, y j x ) = bp (x, y j x ) bp (y j x ) of p (x, y j x ) = g (y j x ) f (x j x ) such that p (x, y j x ) bp (x, y j x ) < C < A.D. () / 35

13 Limitatios The algorithms described earlier su er from several limitatios. Eve if the optimal importace distributio p (x j y, x ) ca be used, this does ot guaratee that the SMC algorithms will be e ciet. Ideed, if the variace of p (y j x ) is high, the the variace of the resultig approximatio will be high. Hece it will be ecessary to resample very frequetly ad the approximatio bp (x : j y : ) of the joit distributio p (x : j y : ) will be ureliable. Oe major problem with the approaches discussed above is that oly the variables X i are sampled at time but the path values X i : remai xed. A obvious way to improve upo these algorithms would ivolve ot oly samplig X i at time, but also modifyig the values of the paths over a xed lag X i L+: for L > i light of the ew observatio y ; L beig xed or upper bouded. A.D. () 2 / 35

14 Resample-Move The Resample-Move algorithm (Gilks & Berzuii, JRSS B, 2) is a stadard approach to mitigate this problem. Like MCMC, it relies upo Markov kerels with appropriate ivariat distributios. Whilst MCMC uses such kerels to geerate collectios of correlated samples, the Resample-Move algorithm uses them i withi a SMC algorithm as a pricipled way to jitter the particle locatios ad thus to reduce degeeracy. A Markov kerel K (x: j x :) of ivariat distributio p (x : j y : ) is a Markov trasitio kerel with the property that Z p (x : j y : ) K x: x: dx: = p x: y:. For such a kerel, if X : p (x : j y : ) ad X : j X : K (x : j X : ) the X : p (x :j y : ). Eve if X : is ot dist.accordig to p (x : j y : ) the, after applyig K, X : ca oly have a dist. closer to p (x : j y : ) i TV tha that of X :. A.D. () 3 / 35

15 Examples of Markov Kerels For example, we could cosider the followig Gibbs sampler: set x : L = x : L the sample x L+ from p x L+ j y :, x : L, x L+2:, sample x L+2 from p x L+2 j y :, x : L+, x L+3: ad so o util we sample x from p x j y :, x : ; that is K x: x : = δx: L x: L k= L+ p xk y :, x:k, x k+: ad we write, with a slight abuse of otatio, the o-degeerate compoet of the MCMC kerel K x L+: x :. It is straightforward to verify that this kerel is p (x : j y : )-ivariat. A.D. () 4 / 35

16 If it is ot possible to sample from p xk j y :, x:k, x k+: = p x k j y k, xk, x k+, we ca istead employ a Metropolis-Hastigs (MH) strategy ad sample a cadidate accordig to some proposal q xk j y k, xk, x k :k+ ad accept it with the usual MH acceptace probability mi, p (x :k, x k+:j y : ) q x k j y k, xk, x k, x! k+ p x:k, x k+: y : q x k j y k, xk, x k :k+ = mi, g (y k j xk ) f (x k+j xk ) f x k j x k q xk j y k, xk, x k, x! k+ g (y k j x k ) f (x k+ j x k ) f x k j xk q x k j y k, xk, x k :k+ These kerels ca be ergodic oly if L = so that all of the compoets of x : are updated. However, i our cotext we will use o-ergodic kerels as we restrict ourselves to updatig the variables X L+: for some xed or bouded L. A.D. () 5 / 35

17 Resample Move Assumig we have access to X i : approx. dist. p (x : j y : ) the at times L Sample X i q(x j y, X i ) ad set X : i X : i, X i. Compute the weights W i g(y jx)f i ( XjX i i ). q( Xjy i,x i ) Resample W, i X: i N :o, X i. to obtai N ew equally-weighted particles Sample X i L+: K (x L+: j X i :) ad set X: i X i : L, X i L+:. A.D. () 6 / 35

18 Iterpretatio as a stadard SMC algorithm We ca justify isertig MCMC trasitios withi a SMC algorithm as follows. Give a target distributio π, a istrumetal distributio µ ad a π-ivariat Markov kerel K, the followig geeralizatio of the IS idetity is trivially true: Z π(y)ϕ(y)dy = ZZ for ay Markov kerel L. π(y)l(xj y) µ(x)k (yj x) µ(x)k (yj x) ϕ(y)dxdy This approach correspods to IS o a elarged space usig µ(x)k (yj x) as the proposal distributio for a target π(y)l(xj y) ad the estimatig a fuctio ϕ (x, y) = ϕ(y). A.D. () 7 / 35

19 I particular, for the time-reversal kerel associated with K L(xj y) = π(x)k (yj x), π (y) we have the importace weight π(y)l(xj y) µ(x)k (yj x) = π (x) µ(x). This iterpretatio of such a approach illustrates its de ciecy: the importace weights deped oly upo the locatio before the MCMC move while the sample depeds upo the locatio after the move. Eve if the kerel was perfectly mixig, leadig to a collectio of iid samples from the target distributio, some of these samples would be elimiated ad some replicated i the resamplig step. Resamplig after a MCMC step will always lead to greater sample diversity tha performig the steps i the other order (ad this algorithm ca be justi ed directly by the ivariace property). A.D. () 8 / 35

20 It is possible to reformulate this algorithm as a speci c applicatio of the geeric SMC algorithm. To simplify otatio we write q (x j x ) for q (x j y, x ) ad to clarify our argumet, it is ecessary to add a superscript to the variables; e.g. X p k correspods to the pth time the radom variable X k is sampled; i this ad the followig sectio, this superscript does ot deote the particle idex. Resample move is a geeric SMC algorithm associated to π x :L+,..., x :L+ L+, x :L :2 L,..., x = p x L+,...., x L+ L+, x L L,..., x 2 y : L x, x 2,..., x L L+ x 2, x 3,..., x L+ L+ K 2 x 2, x2 x 3, x2 2 L x x 2 where L is the time-reversal kerel associated with K. A.D. () 9 / 35

21 If o resamplig is used the we have q x :L+,..., x :L+ L+, x :L :2 L,..., x = q x K x 2 x q2 x2 x 2 K2 x 3, x2 2 x 2, x2 q x x 2 K x L+ L+,..., x 3, x 2 x: L+ L, x L L+,..., x 2, x. This sequece of target distributios admits the lterig distributios of iterest as margials. The clear theoretical advatage of usig MCMC moves is that the use of eve o-ergodic MCMC kerels fk g ca oly improve the mixig properties of fπ g compared to the atural sequece of lterig distributios; this explais why these algorithms outperform a stadard particle lter for a give umber of particles. A.D. () 2 / 35

22 Block Samplig Resample-Move su ers from a major drawback: it does allows us to reitroduce some diversity amog the set of particles after the resamplig step over a lag of legth L >, the importace weights have the same expressio as for the stadard particle lter. This strategy does ot sigi catly decrease the umber of resamplig steps compared to a stadard approach. It ca partially mitigate the problem associated with resamplig, but it does ot prevet these resamplig steps i the rst place. A alterative approach block samplig approach cosists of directly samplig the compoets x L+: at time ; the previously-sampled values of the compoets x L+: sampled are simply discarded. A.D. () 2 / 35

23 Why it is ot trivial... The basic idea is trivial but ot applicable... Cosider you have X : p (x : j y : ) ad at time you sample X L+: q ( j X L, y L+: ) the the joit distributio of X :, X L+: p (x : j y : ) q x L+: x L, y L+: If we discard X L+:, the the distributio of X : L, X L+: is Z p (x : j y : ) q x L+: x L, y L+: dx L+: = p (x : L j y : ) q x L+: x L, y L+:. We typically do ot kow p (x : L j y : ) up to a ormalizig costat so we caot use IS! A.D. () 22 / 35

24 Exteded Importace Samplig The idea cosists of usig a exteded target distributio p x : L, x L+: y: q(x L+: jy L+:, x L ) whose margial is by costructio p x : L, x L+: y :. We the use IS o this exteded state-space ad compute the weights p x : L, x L+: y : q(x L+: jy L+:, x L ) p (x : j y : ) q x L+: x L, y L+: A.D. () 23 / 35

25 Algorithm Settigs If we kew how to compute p (x : L j y : ) the the IS distributio miimizig the variace of the importace weight is q opt (x L+: j x L, y L+: ) = p (x L+: j y L+:, x L ) ad the importace weight is Z p (y L+: j x L ) = f (x k j x k ) g (y k j x k ) dx L+:. k= L+ This optimal weight has a variace which typically decreases expoetially fast with L (uder mixig assumptios). As this distributio is typically ot available ad/or p (y L+: j x L ) caot be computed, we eed to use a approximatio. A.D. () 24 / 35

26 The distributio q(x L+: jy L+:, x L ) miimizig the variace of the weights is simply = q opt (x L+: jy L+:, x L ) p (x : j y : ) q x L+: x L, y R L+: p (x: j y : ) q x L+: x L, y L+: dx L+: So if we pick q opt (x L+: j x L, y L+: ) = p (x L+: j y L+:, x L ) the q opt (x L+: jy L+:, x L ) = p(x L+: jy L+:, x L ) This suggests oce more usig a approximatio of this desity A.D. () 25 / 35

27 Block Samplig SMC Assumig we have access to the at time L o X i : approx. dist. p (x : j y : ) Sample X i L+: q(x L+:jy L+:, X i : ). Compute the weights p X i W i : L, X i L+:, y : bp(x i L+: jy L+:, X i L) p X i :, y : bp(x i L+: jy L+:, X i. L) o Resample W, i X i : L, X i L+: to obtai N ew equally weighted o particles X i :. A.D. () 26 / 35

28 Iterpretatio as a stadard SMC algorithm Oce more this is just a special case of the geeric SMC algorithm. To simply otatio we write q (x L+: j x L ) for q (x L+: j y L+:, x L ) ad to clarify our argumet we add a superscript to the variables; e.g. X p k correspods to the pth time the radom variable X k is sampled. The block samplig algoritmh correspods to π x :L,..., x :L L+, x :L = p x L : L+, x L q 2 x 2, x 2 q x. L+2,..., x L+2,..., x y : q x L L+,..., x If o resamplig is used, a path is sampled accordig to q x :L,..., x :L L+, x :L L+2,..., x = q x q2 x 2, x2 q x L L+,..., x x L L x L. L A.D. () 27 / 35

29 Applicatio to Bearig-oly Trackig Target model X = where V i.i.d. N (, Σ). T T C A X + V The state vector X = X X 2 X 3 X 4 T is such that X (resp. X 3 ) correspods to the horizotal (resp. vertical) positio of the target whereas X 2 (resp. X 4 ) correspods to the horizotal (resp. vertical) velocity. Oe oly receives observatios of the bearigs of the target from a sesor located at the origi + W where W i.i.d. N Y = ta X 3 X, 4 ; i.e. the observatios are almost oiseless. A.D. () 28 / 35

30 We build a approximatio bp (x L+: j y L+:, x L ) of the optimal importace distributio usig the EKF ad the forward lterig/backward samplig formula. We compare the stadard bootstrap lter, two resample-move algorithms where the SISR algorithm for L = usig the EKF proposal is used followed by: (i) oe at a time Metropolis-Hastigs (MH) moves usig a approximatio of the full coditioals p (x k j y k, x k, x k+ ) as a proposal over a lag L = (algorithm RML()); ad (ii) usig the EKF proposal for L = (algorithm RMFL()), the block samplig algorithms for L =, 2, 5 ad which are usig the EKF proposal deoted SMC-EKF(L). Systematic resamplig is performed wheever the ESS goes below N/2. A.D. () 29 / 35

31 Compariso i terms of resamplig steps Filter Avge. # Resamplig steps Bootstrap 46.7 SMC-EKF() 44.6 RML() 45.2 RMFL() 43.3 SMC-EKF(2) 34.9 SMC-EKF(5) 4.6 SMC-EKF().3 Table : Average umber of resamplig steps for simulatios, time istaces per simulatio usig N = particles. A.D. () 3 / 35

32 9 Bootstrap RMFL() EKF(5) EKF() Figure: Average umber of uique particles X (i ) approximatig p ( x j y : ) plotted agaist time (x-axis) A.D. () 3 / 35

33 Applicatio to Stochastic Volatility We cosider a stadard SV model X = φx + σv, X N Y = β exp (X /2) W, σ 2, φ 2, i.i.d. i.i.d. where V N (, ) ad W N (, ). We propose to build approximatio of p (x L+: j y L+:, x L ) usig the fact that log Y 2 = log β 2 + X + log W 2. We approximate o-gaussia oise term log W 2 with a Gaussia oise of similar mea ad variace ad hece obtai a liear Gaussia model approximatio. We the use the KF to build our proposal. The performace of our algorithms are assessed through computer simulatios based o varyig samples sizes to attai a approximately equal computatioal cost. A.D. () 32 / 35

34 Filter # Particles Avge. # Resamplig Steps Bootstrap SMC-EKF() SMC-EKF(2) 4 8. SMC-EKF(5) 6.6 SMC-EKF().45 Table 2: Average umber of resamplig steps for simulatios usig 5 time istaces per simulatio. A.D. () 33 / 35

35 2 8 6 Boots trap EKF() EKF(5) EKF() Figure: Average umber of uique particles X (i ) approximatig p ( x j y : ) plotted agaist time (x-axis) A.D. () 34 / 35

36 P osterior desity P osterior desity Figure: Empirical measure approximatios of p ( x j y :945 ) at times =, 3, 6, 9 for Bootstrap (top left), SMC-EKF() (top right), SMC-EKF(5) (bottom left), SMC-EKF() (bottom right) A.D. () 35 / 35 P osterior desity P osterior desity Time 5 Sample space Time 5 Sample space Time 5 Sample space Time 5 Sample space 5

Sequential Monte Carlo Methods - A Review. Arnaud Doucet. Engineering Department, Cambridge University, UK

Sequential Monte Carlo Methods - A Review. Arnaud Doucet. Engineering Department, Cambridge University, UK Sequetial Mote Carlo Methods - A Review Araud Doucet Egieerig Departmet, Cambridge Uiversity, UK http://www-sigproc.eg.cam.ac.uk/ ad2/araud doucet.html ad2@eg.cam.ac.uk Istitut Heri Poicaré - Paris - 2

More information

Efficient Block Sampling Strategies for Sequential Monte Carlo Methods

Efficient Block Sampling Strategies for Sequential Monte Carlo Methods Efficiet Block Samplig Strategies for Sequetial Mote Carlo Methods Araud DOUCET, Mark BRIERS, ad Stéphae SÉNÉCA Sequetial Mote Carlo SMC) methods are a powerful set of simulatio-based techiques for samplig

More information

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n. Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator

More information

Monte Carlo Integration

Monte Carlo Integration Mote Carlo Itegratio I these otes we first review basic umerical itegratio methods (usig Riema approximatio ad the trapezoidal rule) ad their limitatios for evaluatig multidimesioal itegrals. Next we itroduce

More information

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise First Year Quatitative Comp Exam Sprig, 2012 Istructio: There are three parts. Aswer every questio i every part. Questio I-1 Part I - 203A A radom variable X is distributed with the margial desity: >

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

Output Analysis and Run-Length Control

Output Analysis and Run-Length Control IEOR E4703: Mote Carlo Simulatio Columbia Uiversity c 2017 by Marti Haugh Output Aalysis ad Ru-Legth Cotrol I these otes we describe how the Cetral Limit Theorem ca be used to costruct approximate (1 α%

More information

A Note on Effi cient Conditional Simulation of Gaussian Distributions. April 2010

A Note on Effi cient Conditional Simulation of Gaussian Distributions. April 2010 A Note o Effi ciet Coditioal Simulatio of Gaussia Distributios A D D C S S, U B C, V, BC, C April 2010 A Cosider a multivariate Gaussia radom vector which ca be partitioed ito observed ad uobserved compoetswe

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

1 Introduction to reducing variance in Monte Carlo simulations

1 Introduction to reducing variance in Monte Carlo simulations Copyright c 010 by Karl Sigma 1 Itroductio to reducig variace i Mote Carlo simulatios 11 Review of cofidece itervals for estimatig a mea I statistics, we estimate a ukow mea µ = E(X) of a distributio by

More information

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1 EECS564 Estimatio, Filterig, ad Detectio Hwk 2 Sols. Witer 25 4. Let Z be a sigle observatio havig desity fuctio where. p (z) = (2z + ), z (a) Assumig that is a oradom parameter, fid ad plot the maximum

More information

Distribution of Random Samples & Limit theorems

Distribution of Random Samples & Limit theorems STAT/MATH 395 A - PROBABILITY II UW Witer Quarter 2017 Néhémy Lim Distributio of Radom Samples & Limit theorems 1 Distributio of i.i.d. Samples Motivatig example. Assume that the goal of a study is to

More information

1 of 7 7/16/2009 6:06 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 6. Order Statistics Defiitios Suppose agai that we have a basic radom experimet, ad that X is a real-valued radom variable

More information

Simulation. Two Rule For Inverting A Distribution Function

Simulation. Two Rule For Inverting A Distribution Function Simulatio Two Rule For Ivertig A Distributio Fuctio Rule 1. If F(x) = u is costat o a iterval [x 1, x 2 ), the the uiform value u is mapped oto x 2 through the iversio process. Rule 2. If there is a jump

More information

Frequentist Inference

Frequentist Inference Frequetist Iferece The topics of the ext three sectios are useful applicatios of the Cetral Limit Theorem. Without kowig aythig about the uderlyig distributio of a sequece of radom variables {X i }, for

More information

Chapter 2 The Monte Carlo Method

Chapter 2 The Monte Carlo Method Chapter 2 The Mote Carlo Method The Mote Carlo Method stads for a broad class of computatioal algorithms that rely o radom sampligs. It is ofte used i physical ad mathematical problems ad is most useful

More information

Lecture 19: Convergence

Lecture 19: Convergence Lecture 19: Covergece Asymptotic approach I statistical aalysis or iferece, a key to the success of fidig a good procedure is beig able to fid some momets ad/or distributios of various statistics. I may

More information

Chapter 12 EM algorithms The Expectation-Maximization (EM) algorithm is a maximum likelihood method for models that have hidden variables eg. Gaussian

Chapter 12 EM algorithms The Expectation-Maximization (EM) algorithm is a maximum likelihood method for models that have hidden variables eg. Gaussian Chapter 2 EM algorithms The Expectatio-Maximizatio (EM) algorithm is a maximum likelihood method for models that have hidde variables eg. Gaussia Mixture Models (GMMs), Liear Dyamic Systems (LDSs) ad Hidde

More information

Basics of Probability Theory (for Theory of Computation courses)

Basics of Probability Theory (for Theory of Computation courses) Basics of Probability Theory (for Theory of Computatio courses) Oded Goldreich Departmet of Computer Sciece Weizma Istitute of Sciece Rehovot, Israel. oded.goldreich@weizma.ac.il November 24, 2008 Preface.

More information

Entropy and Ergodic Theory Lecture 5: Joint typicality and conditional AEP

Entropy and Ergodic Theory Lecture 5: Joint typicality and conditional AEP Etropy ad Ergodic Theory Lecture 5: Joit typicality ad coditioal AEP 1 Notatio: from RVs back to distributios Let (Ω, F, P) be a probability space, ad let X ad Y be A- ad B-valued discrete RVs, respectively.

More information

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Convergence of random variables. (telegram style notes) P.J.C. Spreij Covergece of radom variables (telegram style otes).j.c. Spreij this versio: September 6, 2005 Itroductio As we kow, radom variables are by defiitio measurable fuctios o some uderlyig measurable space

More information

Approximations and more PMFs and PDFs

Approximations and more PMFs and PDFs Approximatios ad more PMFs ad PDFs Saad Meimeh 1 Approximatio of biomial with Poisso Cosider the biomial distributio ( b(k,,p = p k (1 p k, k λ: k Assume that is large, ad p is small, but p λ at the limit.

More information

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10 DS 00: Priciples ad Techiques of Data Sciece Date: April 3, 208 Name: Hypothesis Testig Discussio #0. Defie these terms below as they relate to hypothesis testig. a) Data Geeratio Model: Solutio: A set

More information

Lecture 33: Bootstrap

Lecture 33: Bootstrap Lecture 33: ootstrap Motivatio To evaluate ad compare differet estimators, we eed cosistet estimators of variaces or asymptotic variaces of estimators. This is also importat for hypothesis testig ad cofidece

More information

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters

More information

Series III. Chapter Alternating Series

Series III. Chapter Alternating Series Chapter 9 Series III With the exceptio of the Null Sequece Test, all the tests for series covergece ad divergece that we have cosidered so far have dealt oly with series of oegative terms. Series with

More information

Lecture Stat Maximum Likelihood Estimation

Lecture Stat Maximum Likelihood Estimation Lecture Stat 461-561 Maximum Likelihood Estimatio A.D. Jauary 2008 A.D. () Jauary 2008 1 / 63 Maximum Likelihood Estimatio Ivariace Cosistecy E ciecy Nuisace Parameters A.D. () Jauary 2008 2 / 63 Parametric

More information

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain

Since X n /n P p, we know that X n (n. Xn (n X n ) Using the asymptotic result above to obtain an approximation for fixed n, we obtain Assigmet 9 Exercise 5.5 Let X biomial, p, where p 0, 1 is ukow. Obtai cofidece itervals for p i two differet ways: a Sice X / p d N0, p1 p], the variace of the limitig distributio depeds oly o p. Use the

More information

Properties and Hypothesis Testing

Properties and Hypothesis Testing Chapter 3 Properties ad Hypothesis Testig 3.1 Types of data The regressio techiques developed i previous chapters ca be applied to three differet kids of data. 1. Cross-sectioal data. 2. Time series data.

More information

c. Explain the basic Newsvendor model. Why is it useful for SC models? e. What additional research do you believe will be helpful in this area?

c. Explain the basic Newsvendor model. Why is it useful for SC models? e. What additional research do you believe will be helpful in this area? 1. Research Methodology a. What is meat by the supply chai (SC) coordiatio problem ad does it apply to all types of SC s? Does the Bullwhip effect relate to all types of SC s? Also does it relate to SC

More information

A statistical method to determine sample size to estimate characteristic value of soil parameters

A statistical method to determine sample size to estimate characteristic value of soil parameters A statistical method to determie sample size to estimate characteristic value of soil parameters Y. Hojo, B. Setiawa 2 ad M. Suzuki 3 Abstract Sample size is a importat factor to be cosidered i determiig

More information

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ.

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ. 2 5. Weighted umber of late jobs 5.1. Release dates ad due dates: maximimizig the weight of o-time jobs Oce we add release dates, miimizig the umber of late jobs becomes a sigificatly harder problem. For

More information

Optimally Sparse SVMs

Optimally Sparse SVMs A. Proof of Lemma 3. We here prove a lower boud o the umber of support vectors to achieve geeralizatio bouds of the form which we cosider. Importatly, this result holds ot oly for liear classifiers, but

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15 17. Joit distributios of extreme order statistics Lehma 5.1; Ferguso 15 I Example 10., we derived the asymptotic distributio of the maximum from a radom sample from a uiform distributio. We did this usig

More information

7.1 Convergence of sequences of random variables

7.1 Convergence of sequences of random variables Chapter 7 Limit Theorems Throughout this sectio we will assume a probability space (, F, P), i which is defied a ifiite sequece of radom variables (X ) ad a radom variable X. The fact that for every ifiite

More information

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 016 MODULE : Statistical Iferece Time allowed: Three hours Cadidates should aswer FIVE questios. All questios carry equal marks. The umber

More information

Mixtures of Gaussians and the EM Algorithm

Mixtures of Gaussians and the EM Algorithm Mixtures of Gaussias ad the EM Algorithm CSE 6363 Machie Learig Vassilis Athitsos Computer Sciece ad Egieerig Departmet Uiversity of Texas at Arligto 1 Gaussias A popular way to estimate probability desity

More information

CS284A: Representations and Algorithms in Molecular Biology

CS284A: Representations and Algorithms in Molecular Biology CS284A: Represetatios ad Algorithms i Molecular Biology Scribe Notes o Lectures 3 & 4: Motif Discovery via Eumeratio & Motif Represetatio Usig Positio Weight Matrix Joshua Gervi Based o presetatios by

More information

Bayesian Methods: Introduction to Multi-parameter Models

Bayesian Methods: Introduction to Multi-parameter Models Bayesia Methods: Itroductio to Multi-parameter Models Parameter: θ = ( θ, θ) Give Likelihood p(y θ) ad prior p(θ ), the posterior p proportioal to p(y θ) x p(θ ) Margial posterior ( θ, θ y) is Iterested

More information

Machine Learning Brett Bernstein

Machine Learning Brett Bernstein Machie Learig Brett Berstei Week 2 Lecture: Cocept Check Exercises Starred problems are optioal. Excess Risk Decompositio 1. Let X = Y = {1, 2,..., 10}, A = {1,..., 10, 11} ad suppose the data distributio

More information

Application to Random Graphs

Application to Random Graphs A Applicatio to Radom Graphs Brachig processes have a umber of iterestig ad importat applicatios. We shall cosider oe of the most famous of them, the Erdős-Réyi radom graph theory. 1 Defiitio A.1. Let

More information

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample.

Statistical Inference (Chapter 10) Statistical inference = learn about a population based on the information provided by a sample. Statistical Iferece (Chapter 10) Statistical iferece = lear about a populatio based o the iformatio provided by a sample. Populatio: The set of all values of a radom variable X of iterest. Characterized

More information

A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS

A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS J. Japa Statist. Soc. Vol. 41 No. 1 2011 67 73 A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS Yoichi Nishiyama* We cosider k-sample ad chage poit problems for idepedet data i a

More information

Regression with an Evaporating Logarithmic Trend

Regression with an Evaporating Logarithmic Trend Regressio with a Evaporatig Logarithmic Tred Peter C. B. Phillips Cowles Foudatio, Yale Uiversity, Uiversity of Aucklad & Uiversity of York ad Yixiao Su Departmet of Ecoomics Yale Uiversity October 5,

More information

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence Chapter 3 Strog covergece As poited out i the Chapter 2, there are multiple ways to defie the otio of covergece of a sequece of radom variables. That chapter defied covergece i probability, covergece i

More information

Estimation of the Mean and the ACVF

Estimation of the Mean and the ACVF Chapter 5 Estimatio of the Mea ad the ACVF A statioary process {X t } is characterized by its mea ad its autocovariace fuctio γ ), ad so by the autocorrelatio fuctio ρ ) I this chapter we preset the estimators

More information

There is no straightforward approach for choosing the warmup period l.

There is no straightforward approach for choosing the warmup period l. B. Maddah INDE 504 Discrete-Evet Simulatio Output Aalysis () Statistical Aalysis for Steady-State Parameters I a otermiatig simulatio, the iterest is i estimatig the log ru steady state measures of performace.

More information

REGRESSION WITH QUADRATIC LOSS

REGRESSION WITH QUADRATIC LOSS REGRESSION WITH QUADRATIC LOSS MAXIM RAGINSKY Regressio with quadratic loss is aother basic problem studied i statistical learig theory. We have a radom couple Z = X, Y ), where, as before, X is a R d

More information

2 Markov Chain Monte Carlo Sampling

2 Markov Chain Monte Carlo Sampling 22 Part I. Markov Chais ad Stochastic Samplig Figure 10: Hard-core colourig of a lattice. 2 Markov Chai Mote Carlo Samplig We ow itroduce Markov chai Mote Carlo (MCMC) samplig, which is a extremely importat

More information

Entropy Rates and Asymptotic Equipartition

Entropy Rates and Asymptotic Equipartition Chapter 29 Etropy Rates ad Asymptotic Equipartitio Sectio 29. itroduces the etropy rate the asymptotic etropy per time-step of a stochastic process ad shows that it is well-defied; ad similarly for iformatio,

More information

A goodness-of-fit test based on the empirical characteristic function and a comparison of tests for normality

A goodness-of-fit test based on the empirical characteristic function and a comparison of tests for normality A goodess-of-fit test based o the empirical characteristic fuctio ad a compariso of tests for ormality J. Marti va Zyl Departmet of Mathematical Statistics ad Actuarial Sciece, Uiversity of the Free State,

More information

Lecture 3: August 31

Lecture 3: August 31 36-705: Itermediate Statistics Fall 018 Lecturer: Siva Balakrisha Lecture 3: August 31 This lecture will be mostly a summary of other useful expoetial tail bouds We will ot prove ay of these i lecture,

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

Algorithms for Clustering

Algorithms for Clustering CR2: Statistical Learig & Applicatios Algorithms for Clusterig Lecturer: J. Salmo Scribe: A. Alcolei Settig: give a data set X R p where is the umber of observatio ad p is the umber of features, we wat

More information

Sequences. Notation. Convergence of a Sequence

Sequences. Notation. Convergence of a Sequence Sequeces A sequece is essetially just a list. Defiitio (Sequece of Real Numbers). A sequece of real umbers is a fuctio Z (, ) R for some real umber. Do t let the descriptio of the domai cofuse you; it

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS MASSACHUSTTS INSTITUT OF TCHNOLOGY 6.436J/5.085J Fall 2008 Lecture 9 /7/2008 LAWS OF LARG NUMBRS II Cotets. The strog law of large umbers 2. The Cheroff boud TH STRONG LAW OF LARG NUMBRS While the weak

More information

6.3 Testing Series With Positive Terms

6.3 Testing Series With Positive Terms 6.3. TESTING SERIES WITH POSITIVE TERMS 307 6.3 Testig Series With Positive Terms 6.3. Review of what is kow up to ow I theory, testig a series a i for covergece amouts to fidig the i= sequece of partial

More information

Advanced Stochastic Processes.

Advanced Stochastic Processes. Advaced Stochastic Processes. David Gamarik LECTURE 2 Radom variables ad measurable fuctios. Strog Law of Large Numbers (SLLN). Scary stuff cotiued... Outlie of Lecture Radom variables ad measurable fuctios.

More information

Kolmogorov-Smirnov type Tests for Local Gaussianity in High-Frequency Data

Kolmogorov-Smirnov type Tests for Local Gaussianity in High-Frequency Data Proceedigs 59th ISI World Statistics Cogress, 5-30 August 013, Hog Kog (Sessio STS046) p.09 Kolmogorov-Smirov type Tests for Local Gaussiaity i High-Frequecy Data George Tauche, Duke Uiversity Viktor Todorov,

More information

Expectation-Maximization Algorithm.

Expectation-Maximization Algorithm. Expectatio-Maximizatio Algorithm. Petr Pošík Czech Techical Uiversity i Prague Faculty of Electrical Egieerig Dept. of Cyberetics MLE 2 Likelihood.........................................................................................................

More information

Stochastic Simulation

Stochastic Simulation Stochastic Simulatio 1 Itroductio Readig Assigmet: Read Chapter 1 of text. We shall itroduce may of the key issues to be discussed i this course via a couple of model problems. Model Problem 1 (Jackso

More information

Estimation for Complete Data

Estimation for Complete Data Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of

More information

Statisticians use the word population to refer the total number of (potential) observations under consideration

Statisticians use the word population to refer the total number of (potential) observations under consideration 6 Samplig Distributios Statisticias use the word populatio to refer the total umber of (potetial) observatios uder cosideratio The populatio is just the set of all possible outcomes i our sample space

More information

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING Lectures MODULE 5 STATISTICS II. Mea ad stadard error of sample data. Biomial distributio. Normal distributio 4. Samplig 5. Cofidece itervals

More information

Infinite Sequences and Series

Infinite Sequences and Series Chapter 6 Ifiite Sequeces ad Series 6.1 Ifiite Sequeces 6.1.1 Elemetary Cocepts Simply speakig, a sequece is a ordered list of umbers writte: {a 1, a 2, a 3,...a, a +1,...} where the elemets a i represet

More information

Lecture 11 and 12: Basic estimation theory

Lecture 11 and 12: Basic estimation theory Lecture ad 2: Basic estimatio theory Sprig 202 - EE 94 Networked estimatio ad cotrol Prof. Kha March 2 202 I. MAXIMUM-LIKELIHOOD ESTIMATORS The maximum likelihood priciple is deceptively simple. Louis

More information

Lecture 01: the Central Limit Theorem. 1 Central Limit Theorem for i.i.d. random variables

Lecture 01: the Central Limit Theorem. 1 Central Limit Theorem for i.i.d. random variables CSCI-B609: A Theorist s Toolkit, Fall 06 Aug 3 Lecture 0: the Cetral Limit Theorem Lecturer: Yua Zhou Scribe: Yua Xie & Yua Zhou Cetral Limit Theorem for iid radom variables Let us say that we wat to aalyze

More information

Module 1 Fundamentals in statistics

Module 1 Fundamentals in statistics Normal Distributio Repeated observatios that differ because of experimetal error ofte vary about some cetral value i a roughly symmetrical distributio i which small deviatios occur much more frequetly

More information

Generalized Semi- Markov Processes (GSMP)

Generalized Semi- Markov Processes (GSMP) Geeralized Semi- Markov Processes (GSMP) Summary Some Defiitios Markov ad Semi-Markov Processes The Poisso Process Properties of the Poisso Process Iterarrival times Memoryless property ad the residual

More information

Introduction to Extreme Value Theory Laurens de Haan, ISM Japan, Erasmus University Rotterdam, NL University of Lisbon, PT

Introduction to Extreme Value Theory Laurens de Haan, ISM Japan, Erasmus University Rotterdam, NL University of Lisbon, PT Itroductio to Extreme Value Theory Laures de Haa, ISM Japa, 202 Itroductio to Extreme Value Theory Laures de Haa Erasmus Uiversity Rotterdam, NL Uiversity of Lisbo, PT Itroductio to Extreme Value Theory

More information

Regression with quadratic loss

Regression with quadratic loss Regressio with quadratic loss Maxim Ragisky October 13, 2015 Regressio with quadratic loss is aother basic problem studied i statistical learig theory. We have a radom couple Z = X,Y, where, as before,

More information

Understanding Samples

Understanding Samples 1 Will Moroe CS 109 Samplig ad Bootstrappig Lecture Notes #17 August 2, 2017 Based o a hadout by Chris Piech I this chapter we are goig to talk about statistics calculated o samples from a populatio. We

More information

Chapter 6 Sampling Distributions

Chapter 6 Sampling Distributions Chapter 6 Samplig Distributios 1 I most experimets, we have more tha oe measuremet for ay give variable, each measuremet beig associated with oe radomly selected a member of a populatio. Hece we eed to

More information

The random version of Dvoretzky s theorem in l n

The random version of Dvoretzky s theorem in l n The radom versio of Dvoretzky s theorem i l Gideo Schechtma Abstract We show that with high probability a sectio of the l ball of dimesio k cε log c > 0 a uiversal costat) is ε close to a multiple of the

More information

Elements of Statistical Methods Lots of Data or Large Samples (Ch 8)

Elements of Statistical Methods Lots of Data or Large Samples (Ch 8) Elemets of Statistical Methods Lots of Data or Large Samples (Ch 8) Fritz Scholz Sprig Quarter 2010 February 26, 2010 x ad X We itroduced the sample mea x as the average of the observed sample values x

More information

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 22

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 22 CS 70 Discrete Mathematics for CS Sprig 2007 Luca Trevisa Lecture 22 Aother Importat Distributio The Geometric Distributio Questio: A biased coi with Heads probability p is tossed repeatedly util the first

More information

CSE 527, Additional notes on MLE & EM

CSE 527, Additional notes on MLE & EM CSE 57 Lecture Notes: MLE & EM CSE 57, Additioal otes o MLE & EM Based o earlier otes by C. Grat & M. Narasimha Itroductio Last lecture we bega a examiatio of model based clusterig. This lecture will be

More information

Lecture 7: Properties of Random Samples

Lecture 7: Properties of Random Samples Lecture 7: Properties of Radom Samples 1 Cotiued From Last Class Theorem 1.1. Let X 1, X,...X be a radom sample from a populatio with mea µ ad variace σ

More information

Exponential Families and Bayesian Inference

Exponential Families and Bayesian Inference Computer Visio Expoetial Families ad Bayesia Iferece Lecture Expoetial Families A expoetial family of distributios is a d-parameter family f(x; havig the followig form: f(x; = h(xe g(t T (x B(, (. where

More information

Kernel density estimator

Kernel density estimator Jauary, 07 NONPARAMETRIC ERNEL DENSITY ESTIMATION I this lecture, we discuss kerel estimatio of probability desity fuctios PDF Noparametric desity estimatio is oe of the cetral problems i statistics I

More information

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic

More information

Element sampling: Part 2

Element sampling: Part 2 Chapter 4 Elemet samplig: Part 2 4.1 Itroductio We ow cosider uequal probability samplig desigs which is very popular i practice. I the uequal probability samplig, we ca improve the efficiecy of the resultig

More information

1 Inferential Methods for Correlation and Regression Analysis

1 Inferential Methods for Correlation and Regression Analysis 1 Iferetial Methods for Correlatio ad Regressio Aalysis I the chapter o Correlatio ad Regressio Aalysis tools for describig bivariate cotiuous data were itroduced. The sample Pearso Correlatio Coefficiet

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2018 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

A survey on penalized empirical risk minimization Sara A. van de Geer

A survey on penalized empirical risk minimization Sara A. van de Geer A survey o pealized empirical risk miimizatio Sara A. va de Geer We address the questio how to choose the pealty i empirical risk miimizatio. Roughly speakig, this pealty should be a good boud for the

More information

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes.

Direction: This test is worth 150 points. You are required to complete this test within 55 minutes. Term Test 3 (Part A) November 1, 004 Name Math 6 Studet Number Directio: This test is worth 10 poits. You are required to complete this test withi miutes. I order to receive full credit, aswer each problem

More information

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1.

Econ 325/327 Notes on Sample Mean, Sample Proportion, Central Limit Theorem, Chi-square Distribution, Student s t distribution 1. Eco 325/327 Notes o Sample Mea, Sample Proportio, Cetral Limit Theorem, Chi-square Distributio, Studet s t distributio 1 Sample Mea By Hiro Kasahara We cosider a radom sample from a populatio. Defiitio

More information

Journal of Multivariate Analysis. Superefficient estimation of the marginals by exploiting knowledge on the copula

Journal of Multivariate Analysis. Superefficient estimation of the marginals by exploiting knowledge on the copula Joural of Multivariate Aalysis 102 (2011) 1315 1319 Cotets lists available at ScieceDirect Joural of Multivariate Aalysis joural homepage: www.elsevier.com/locate/jmva Superefficiet estimatio of the margials

More information

Empirical Process Theory and Oracle Inequalities

Empirical Process Theory and Oracle Inequalities Stat 928: Statistical Learig Theory Lecture: 10 Empirical Process Theory ad Oracle Iequalities Istructor: Sham Kakade 1 Risk vs Risk See Lecture 0 for a discussio o termiology. 2 The Uio Boud / Boferoi

More information

MA131 - Analysis 1. Workbook 9 Series III

MA131 - Analysis 1. Workbook 9 Series III MA3 - Aalysis Workbook 9 Series III Autum 004 Cotets 4.4 Series with Positive ad Negative Terms.............. 4.5 Alteratig Series.......................... 4.6 Geeral Series.............................

More information

MAS275 Probability Modelling

MAS275 Probability Modelling MAS275 Probability Modellig 6 Poisso processes 6.1 Itroductio Poisso processes are a particularly importat topic i probability theory. The oe-dimesioal Poisso process, which most of this sectio will be

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

The Poisson Process *

The Poisson Process * OpeStax-CNX module: m11255 1 The Poisso Process * Do Johso This work is produced by OpeStax-CNX ad licesed uder the Creative Commos Attributio Licese 1.0 Some sigals have o waveform. Cosider the measuremet

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

Ada Boost, Risk Bounds, Concentration Inequalities. 1 AdaBoost and Estimates of Conditional Probabilities

Ada Boost, Risk Bounds, Concentration Inequalities. 1 AdaBoost and Estimates of Conditional Probabilities CS8B/Stat4B Sprig 008) Statistical Learig Theory Lecture: Ada Boost, Risk Bouds, Cocetratio Iequalities Lecturer: Peter Bartlett Scribe: Subhrasu Maji AdaBoost ad Estimates of Coditioal Probabilities We

More information

Comparison of Resampling Schemes for Particle Filtering

Comparison of Resampling Schemes for Particle Filtering Compariso of Resamplig Schemes for article Filterig Radal Douc Ecole olytechique 928 alaiseau, Frace douc at cmapx.polytechique.fr Eric Moulies GET Télécom aris 46 rue Barrault, 75634 aris, Frace moulies

More information

Randomized Algorithms I, Spring 2018, Department of Computer Science, University of Helsinki Homework 1: Solutions (Discussed January 25, 2018)

Randomized Algorithms I, Spring 2018, Department of Computer Science, University of Helsinki Homework 1: Solutions (Discussed January 25, 2018) Radomized Algorithms I, Sprig 08, Departmet of Computer Sciece, Uiversity of Helsiki Homework : Solutios Discussed Jauary 5, 08). Exercise.: Cosider the followig balls-ad-bi game. We start with oe black

More information

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3

(A sequence also can be thought of as the list of function values attained for a function f :ℵ X, where f (n) = x n for n 1.) x 1 x N +k x N +4 x 3 MATH 337 Sequeces Dr. Neal, WKU Let X be a metric space with distace fuctio d. We shall defie the geeral cocept of sequece ad limit i a metric space, the apply the results i particular to some special

More information

Vector Quantization: a Limiting Case of EM

Vector Quantization: a Limiting Case of EM . Itroductio & defiitios Assume that you are give a data set X = { x j }, j { 2,,, }, of d -dimesioal vectors. The vector quatizatio (VQ) problem requires that we fid a set of prototype vectors Z = { z

More information