Maximum Likelihood Parameter Estimation in General State-Space Models using Particle Methods. -

Size: px
Start display at page:

Download "Maximum Likelihood Parameter Estimation in General State-Space Models using Particle Methods. -"

Transcription

1 Maximum Likelihood Parameter Estimatio i Geeral State-Space Models usig Particle Methods George Poyiadjis, 2 Araud Doucet ad Sumeetpal S. Sigh Departmet of Egieerig, Uiversity of Cambridge, Cambridge CB2 PZ, UK. 2 Departmets of Computer Sciece ad Statistics, Uiversity of British Columbia, Vacouver, BC, Caada. gp243,sss4@cam.ac.uk - araud@stat.ubc.ca KEY WORDS: particle filter ; parameter estimatio ; geeral state-space model; filter derivative ; hah-jorda decompositio. Abstract A large umber of time series ca be described by oliear, o-gaussia state-space models. While state estimatio for these models is ow routiely performed usig particle filters, maximum likelihood estimatio of the model parameters is much more challegig. I this paper, we preset ew umerical methods to approximate the derivative of the optimal filter. We use this to perform batch ad recursive maximum likelihood parameter estimatio ad trackig by maximizig the likelihood through a gradiet ascet method. We geeralize the method to iclude the secod derivative of the optimal filter. This provides estimates of the Hessia of the likelihood ad ca be used to accelerate the gradiet ascet method.. Itroductio May time series problems arisig i statistics, egieerig ad applied scieces are cocered with the estimatio of the state of a dyamic model whe oly iaccurate observatios are available. Most real-world problems are oliear ad o-gaussia, therefore optimal state estimatio i such problems does ot admit a closed form solutio. Recetly, there has bee a surge of iterest i Sequetial Mote Carlo SMC methods, also kow as particle filterig methods, to perform sequetial state estimatio i o-liear o-gaussia models [7], [8], [9], [2], [5]. SMC methods are a set of simulatio-based techiques that recursively geerate ad update a set of weighted samples, which provide approximatios to the posterior probability distributios of iterest. Uder the assumptio that the model parameters are kow, umerous SMC algorithms have bee proposed over the last decade; see [8] for a review. I real-world applicatios however, the model parameters deoted θ, are ofte ukow ad also eed to be estimated from the data. Maximum likelihood ML parameter estimatio usig SMC methods still remais a ope problem, despite various earlier attempts i the literature. The majority of the proposed SMC-based parameter estimatio methods rely o augmetig the hidde state to iclude the ukow parameter ad castig the problem as a filterig oe [], [6], [2]. Static parameter estimatio with SMC is the implemeted by either itroducig artificial dyamics for the parameters or MCMC rejuveatio steps. The latter method is more elegat, sice the model of iterest is ot artificially altered. However, the MCMC steps rely o sufficiet statistics that are based o a approximatio of the path posterior desity p θ x : Y : of the hidde process up to time give the observatios up to time. This desity caot be properly approximated usig SMC methods, for a fixed umber of particles, ad the sufficiet statistics degrade over time due to error accumulatio []. I this paper we preset a origial maximum likelihood method that is based o a direct particle approximatio of the derivative of the optimal filter. Previous attempts to approximate the filter derivative usig particle methods - e.g. [5] [] ad [6] - were based implicitly o the sequece of path desities {p θ x : Y : }. As i the case of filterig-based parameter estimatio, the approximatio errors they produce icrease with the data legth. The methods we proposed here to approximate the filter derivative are based o the sequece of margial distributios {p θ x Y : } ad hece do ot suffer from the aforemetioed problem. We use the filter derivative approximatio to compute the log-likelihood gradiet ad we combie it with a gradiet ascet algorithm to geerate maximum likelihood estimates of the model parameters. The approach is geeralized to compute a particle approximatio to the secod derivative of the filter. This leads to a estimate of the Hessia of the likelihood that ca be used to scale the gradiet compoets ad accelerate the covergece of the gradiet algorithm. Additioally it may allow us to compute cofidece regios for the estimated parameters. The rest of the paper is orgaized as follows: I Sectio 2 the statistical model of iterest is preseted ad the optimal filter ad its first ad secod derivatives are described. I Sectio 3 we review the particle filter algorithm ad derive particle methods for the derivatives of the filter. I For the rest of this paper we will adopt the followig otatio: for ay sequece {z k } ad radom process {Z k }, we defie z i:j = z i, z i+,..., z j ad Z i:j = Z i, Z i+,..., Z j, respectively.

2 Sectio 4 we describe how the first ad secod filter derivative approximatios ca be used to perform ML parameter estimatio i a recursive ad a batch maer. Sectio 5 presets simulatio results showig the performace of the proposed algorithm. Fially i Sectio 6 we discuss the results ad provide some cocludig remarks. 2. Optimal Filter ad its Derivatives 2. State-Space Models Let {X } ad {Y } be R x ad R y -valued stochastic processes defied o a measurable space Ω, F. These stochastic processes deped o a parameter θ Θ, where Θ is a ope subset of R θ. The process {X } is a uobserved hidde Markov process of iitial desity µ; i.e. X µ, ad a Markov trasitio desity f θ x x; i.e. X + X = x f θ x. Although {X } is ukow, it is partially observed through the observatio process {Y }. It is assumed that the observatios coditioed upo {X } are idepedet with margial desity g θ y x ; i.e. Y X = x g θ x. 2 This class of models icludes may oliear ad o- Gaussia time series models such as X + = ϕ θ X, V +, Y = ψ θ X, W where {V } ad {W } are mutually idepedet sequeces of idepedet radom variables ad ϕ θ, ψ θ determie the evolutio of the state ad observatio processes. 2.2 Optimal Filter Derivatives Assume for the time beig that θ is kow. I such a case, sequetial iferece about the hidde process X is typically based o the sequece of joit posterior distributios {p θ x : Y : }. This summarizes all the relevat iformatio available about X :, up to time. Usig a importace samplig approach with a arbitrary importat desity q θ x Y, x, whose support icludes the support of g θ Y x f θ x x, it ca be easily show that the joit posterior desity satisfies the recursio p θ x : Y : = α θ x :, Y p θ Y Y :: q θ x Y, x p θ x : Y :, where the importace weights are give by 3 α θ x :, Y = g θ Y x f θ x x. 4 q θ x Y, x I most problems, we are iterested i the margial p θ x Y :, which is kow as the filterig desity. This ca be expressed as p θ x Y : α θ x :, Y q θ x Y, x p θ x Y : dx I applicatios such as parameter estimatio ad stochastic cotrol, we are ofte iterested i optimizig differet performace criteria that require a approximatio of the filter derivatives. I the cotext of parameter estimatio, we will cosider the first two derivatives of the optimal filter with respect to θ, amely p θ x Y : ad 2 p θ x Y :. To simplify the otatio, let p θ x Y : 5 ξx, Y : ξx, Y : dx 6 where ξ θ x, Y : = g θ Y x f θ x x p θ x Y : dx = g θ Y x p θ x Y :. Uder regularity assumptios, the first ad secod derivative of 6 leads to the followig recursios 2, ad 7 p θ x Y : = ξ θx, Y : ξθ x, Y : dx ξθ x, Y : dx p θ x Y : 8 ξθ x, Y : dx 2 p θ x Y : = 2 ξ θ x, Y : 2 p θ x Y : ξθ x, Y : dx ξθ x, Y : dx 2 ξ θ x, Y : dx p θ x Y :, ξθ x, Y : dx ξθ x, Y : dx where ξ θ x, Y : = g θ Y x 9 f θ x x p θ x Y : [ log g θ Y x + log f θ x x ] dx + g θ Y x f θ x x p θ x Y : dx 2 The st derivative p θ x Y : is a θ vector, where the i th etry is give by p θx Y :. The 2d derivative θ 2 p i θ x Y : is a θ θ matrix, where the i, j th etry is give by 2 p θ x Y :. θ i θ j 2

3 ad 2 ξ θ x, Y : = g θ Y x f θ x x { [ log g θ Y x + log f θ x x ] log g θ Y x + 2 log f θ x x } p θ x Y : dx + 2g θ Y x f θ x x [ log g θ Y x + log f θ x x ] p θ x Y : dx + g θ Y x f θ x x 2 p θ x Y : dx. Except i some simple cases, o closed-form expressio ca be obtaied for either of the above recursios ad oe typically resorts to umerical approximatios. Our mai objective i this paper is to derive particle methods to approximate p θ x Y : ad 2 p θ x Y :. 3. Particle Methods for the Filter Derivatives 3. Particle Filters Particle methods are widely used to umerically approximate the filterig recursio i 3 by meas of a weighted empirical distributio of a set of N samples, termed as particles. This empirical distributio is propagated sequetially as follows. [ Assume that at time ] a set of particles X :N : X :,..., XN : with correspodig weights ã :N ã,..., ãn are available, with [ ] j= ãj =. We further assume that this weighted particle set is distributed approximately accordig to the joit desity p θ x : Y :. A stadard way to approximate the joit desity at the ext time step is to exted the path usig X i q θ x Y : = ã i q θ Y, X i. 2 Samplig from 2 is achieved by first samplig the discrete idex i usig a stadard resamplig algorithm, such as stratified or multiomial resamplig. The, the ew particle X i is geerated accordig to X i q θ Y, X ϕi, 3 where ϕ i is the discrete idex obtaied from the resamplig mechaism. Note[ that the ew ] set of equally weighted particles X :N : = X :,..., XN :, with X i : = X ϕi :, Xi will be approximately distributed accordig to the joit desity q θ x Y, x p θ x : Y :. Substitutio of this approximatio ito 3 leads to the updated empirical distributio where p θ x : Y : = a i = α θ X ϕi, Xi, Y ã i δ ad ã i = x : X i : a i j= aj, 4. 5 I practice, a particle approximatio of the filterig desity p θ x Y : i 5 is obtaied by margializatio of Filter derivative approximatios The filter derivatives p θ x Y : ad 2 p θ x Y : are siged measures; i.e. they ca take positive ad egative values ad itegrate to zero. Empirical approximatios of these measures usig particle methods are still possible, provided that oe uses the same set of particles as i the filter approximatio, but with differet weights. This idea that was first itroduced i [5], computes the particle approximatios p θ x Y : ad 2 p θ x Y : by propagatig the weighted particles o the path space ad margializig the expressios p θ x : Y : = 2 p θ x : Y : = ã i β i δ x : X i :, 6 ã i λ i δ x : X i :, 7 where β i, λ i ca be positive or egative. As already metioed, the drawback of this approach stems from the fact that it relies o the path, which is a space of growig dimesios. Cosequetly, as the legth of the path icreases, the variace of pθ x : Y :, 2 p θ x Y : will icrease ad the approximatios of their margials will degrade severely. Aother effect that deteriorates the performace of a path-based particle algorithm results from the arragemet of the particle mass o the state space. The derivative of a probability measure is a siged measure ν that ca be expressed as ν = c π π 2, where π, π 2 are two probability measures ad c is a oegative costat. This approach, kow as weak derivative decompositio, allows a arbitrary umber of possible decompositios for a give siged measure. The path-based particle method discussed i the previous sectio decomposes the derivative siged measures by approximatig the two probability measures π ad π 2 by a set of positively ad egatively weighted particles o overlappig regios 3

4 of the state space. To see this, cosider a poit x ad a eighborhood of it, B x, for which the sig measure satisfies ν >. If we use the particle represetatio of 6 to approximate ν, a estimate of I Bx x ν x dx becomes equal to ãi β i I Bx X i. While this is a valid approximatio, we may have that for two particles k ad l belogig to B x, the weights are ot of the same sig, i.e., ã k β k < while ã l β l >. I such a case we say that the particles mix, as illustrated i the top plot of Figure for the case where R x = R θ = R. This implies that may particles with opposite sigs ca ed up approximatig regios of the state space that have low total mass see for example low mass regio aroud the value x = i the top plot of Figure. This effect builds up due to the sequetial ature of the algorithm ad the implemetatio becomes less accurate ad iefficiet as the data legth icreases. We propose here a origial method to approximate the optimal filter derivatives that is based o a direct poitwise approximatio of 8 ad 9 ad hece does ot suffer from the limitatios discussed i the previous paragraphs. This method essetially itegrates aalytically a discrete approximatio of the latet variable ad will therefore have a lower variace. From a weak derivatives poit of view, this is equivalet to a a particle implemetatio of a Hah- Jorda decompositio, which esures that the probability measures of the decompositio are cocetrated o disjoit regios of the state. As a result, the algorithm does ot suffer from mixig of the positively ad egatively weighted particles, as illustrated i the bottom plot of Figure. 2 Path based particle approximatio of siged measure True siged measure 2 Particle Approx.+ve Particle Approx. ve Poitwise particle approximatio of siged measure Hah Jorda decompositio Figure : Top plot: Histogram represetatio of a pathbased particle approximatio of pθ x Y : w.r.t. a oedimesioal parameter θ. Bottom plot: Poit-wise particle approximatio of the same siged measure that maitais the positive ad egative weights o separate regios of the state support Hah-Jorda decompositio. 3.3 Particle algorithm I this sectio we describe the proposed sequetial method to approximate the first two derivatives of the optimal filter. Assume that at time, we have particle approximatios of p θ x Y :, p θ x Y : ad 2 p θ x Y : of the form p θ x Y : = p θ x Y : = 2 p θ x Y : = ã i δ x X i, 8 ã i βi δ x X i, 9 ã i λi δ x X i. 2 Substitutio of these ito 7, ad leads to the followig poit-wise approximatios ad ξ θ x, Y : = ξ θ x, Y : = k= ã k g θy x f θ x X k, 2 k= ã k g θy x f θ x X k ] [ log g θ Y x + log f θ x X k + β k, 22 2 ξ θ x, Y : = k= ã k g θy x f θ x X k { [ ] 2 log g θ Y x + log f θ x X k + 2 log g θ Y x + 2 log f θ x X k +2β k [ log g θ Y x + log f θ x X k ] } + λ k 23 As i the stadard particle filter, we geerate a set of particles X i, for i =,..., N, usig 2. Evaluatig the poit-wise approximatios i 2, 22 ad 23 at poits yields the followig particle approximatios X i ξ θ x, Y : = N ξ θ x, Y : = N a i δ i X ρ i δ i X x, 24 x ad 25 4

5 where a i 2 ξ θ X i,y : q θ X i 2 ξ θ x, Y : = N = ξ θx i,y: q θ X i, ρ i Y : π i δ i X = ξ θ X i,y: q θ X i x, 26 Y : ad π i =. Substitutio of the last three approximatios ito 6, 8 ad 9 Y : gives where ã i = p θ x Y : = p θ x Y : = 2 p θ x Y : = a i j= aj ã i δ i X ã i β i, ã i β i = ã i λ i π i = 2ã i j= aj β i ã i δ i X λ i δ i X ρ i j= aj j= ρj j= aj x, 27 x ad 28 x, 29 ã i ã i j= ρj j= aj j= πj j= aj Note that 27 gives a filterig byproduct of the algorithm. Compared to stadard path-based particle filters, the poit-wise particle filter p θ x Y : requires O N 2 operatios istead of O N. However, for a fixed umber of particles N, it will outperform path-based methods due to the aalytical itegratio ivolved. 4. ML Parameter Estimatio Let us ow cosider the geeral state space model of sectio 2., where the model parameter θ is ukow. We will assume that the model that geerates the observatio sequece {Y } evolves accordig to a true but ukow static parameter θ, i.e. X X = x f θ. x 3 Y X = x g θ. x. 3 Our objective is to idetify θ based o {Y }. We propose here two gradiet algorithms to perform maximum likelihood estimatio. These are based o a gradiet ascet method that utilizes the estimates of the derivatives of the filter that were preseted i the previous sectio. The first method is a recursive algorithm that updates the parameter estimate as soo as a ew observatio is received. This is based o the maximizatio of a average log-likelihood criterio ad requires a large umber of observatios to be available. A batch versio of the algorithm is also preseted. This directly maximizes the log-likelihood of some available set of observatios Y :.,. 4. Recursive ML A stadard approach to Recursive ML RML estimatio cosiders a series of log-likelihood fuctios {log p θ Y :k } k, where log p θ Y :k = k = log p θy Y : [7]. The expressio p θ Y Y : is kow as the predictive likelihood ad ca be writte as p θ Y Y : = g θ Y x f θ x x p θ x Y : dx : 32 Uder suitable regularity coditios described i [2] it ca be show that the average log-likelihood coverges to the followig limit lim k k + k log p θ Y Y : = l θ, 33 = where l θ is give by l θ = log R y PR x g θ y x µ x dx λ θ,θ dy, dµ. Here P R x is the space of probability distributios o R x ad λ θ,θ dy, dµ is the joit ivariat distributio of the couple Y, p θ x Y :. Note that λ θ,θ is a fuctio of both θ ad θ, sice the observatio compoet evolves accordig to the true parameter θ, while the predictio filter compoet evolves accordig to θ. Followig the approach used i [4] for fiite state space models, it ca be show that l θ admits θ as a global maximum. The fuctio l θ does ot have a aalytical expressio ad we do ot have access to it. Nevertheless, idetificatio of θ ca still be achieved based o the ergodicity property i 33, which provides us with a set of accessible fuctios log p θ Y Y : that coverge to l θ. Oe way to exploit this i order to maximize l θ, is to use a Stochastic Approximatio SA algorithm to update the parameter estimate at time usig the recursio θ = θ + γ log p θ: Y Y :, 34 where θ is the parameter estimate at time ad log p θ Y Y : deotes the gradiet of log p θ Y Y : 3. Provided that the step size θ is a positive o-icreasig sequece, such that γ = ad γ 2 <, it ca be show that θ will coverge to the set of global or local maxima of l θ. 3 The SA requires a estimate of log p θ Y Y : with θ held fixed. I our problem, θ caot be fixed, sice we are estimatig it recursively. However, sice θ chages slowly, a stadard approach [8] is to reuse the previous particle calculatios that were based o θ : 2 ad use the parameter estimate θ at time. 5

6 The remaiig step i the developmet of the algorithm is to obtai a umerical approximatio to log p θ Y Y :. This follows directly from the the expressios for p θ x Y : ad p θ x Y : i Sectio 3.3, sice compariso of 7 ad 32 gives p θ Y Y : = ξθ x, Y : dx. Usig the particle approximatios of ξ θ x, Y : ad ξ θ x, Y : i 24 ad 25 we obtai log p θ Y Y : = p θ Y Y : p θ Y Y : 4.. Adaptive SA = ξθ x, Y : dx ξθ x, Y : dx = j= ρj j= aj Adaptive steps The SA of 34 ca be thought of as a stochastic geeralizatio of the steepest descet method. Faster covergece ca be achieved if oe employs a Newtoia method that is based o a estimate of the Hessia of the objective fuctio ad leads to a asymptotically optimal search directio [3]. I geeral, estimatio of the Hessia is o-trivial ad fiite differece approximatios are typically used to approximate it [9]. I our framework, the Hessia of the log-likelihood ca be straightforwardly estimated usig the particle approximatios of the optimal filter ad its first ad secod derivatives i 27, 28 ad 29. More specifically, the θ θ Hessia matrix estimate at time will be give by 2 log p θ Y Y : = 2 p θ Y Y : p θ Y Y : = j= πi j= aj 2 pθ Y Y : p θ Y Y : 2 j= ρj. j= aj This allows oe to compute the asymptotic value of the Hessia usig, for example, a recursio of the form H = H + Ĥ H, 35 + where Ĥ = 2 log p θ Y Y :. This is simply a recursive calculatio of the sample mea H up to time. By costructio, the true Hessia is a egative defiite, symmetric matrix whose iverse ca provide a adaptive step i 34. Direct iversio of the estimated value H will be possible oly if this matrix is egative defiite. I practice this is usually esured by projectig H oto the set of egative defiite matrices - see [4] ad [9] for details. The Newto-type SA recursio that ca replace 34 will take the form. This adaptive SA is particularly attractive, i terms of covergece acceleratio, i the termial phase of the algorithm, where the steepest descet-type method slows dow. Cofidece Regios From a practical poit of view, it is ofte desirable to assess the accuracy of the parameter estimate by meas of cofidece itervals or more geerally cofidece regios. I priciple, cetral limit theorems that have bee established for a umber of stadard SA algorithms allow the computatio of cofidece regios for the estimates - see [3] for detailed results. Oe of the difficulties is that the covariace matrix of the limitig multivariate ormal distributio depeds o the iverse of the Hessia of the objective fuctio. The proposed method provides estimates for these quatity through RML Parameter Estimatio Algorithm The Recursive ML estimatio is summarized as follows:. Samplig Step For i =,..., N, sample 4 X i q θ: Y : = N ãi q θ 2. Weight Calculatio Compute a i ρ i π i = ξ θ: X i,y:, q θ: X i Y : = ξ θ: X i,y : q θ: X i = 2 ξ θ: X i,y: q θ: X i Y : ad Compute the weights ã i ã i β i ã i λ i = ρi j= aj = πi j= aj Y, X i Y : usig 2, 22 ad 23. = ai j= aj ã i j= ρj j= aj 2ã i 3. Parameter Update Step Ĥ = j= πi j= aj H = H + + θ = θ + γ H N β i 2 j= ρj j= aj, ad j= ρj j= aj ã i j= πj j= aj Ĥ H, H = Ψ H j= ρj j= aj Remark : The fuctio Ψ H is a mappig to the set of egative defiite matrices, based o diagoal modificatios to the Hessia. Remark 2: This algorithm guaratees that αi β i = αi λ i =. Remark 3: Eve if θ Θ, it is possible that for the updated value we have θ / Θ. A stadard approach to θ = θ + γ H log p θ: Y Y : Note that i this approach the resamplig step is icluded whe we sample from q θ x Y :. 6

7 prevet such divergece is to reproject the parameter value iside Θ = θ [ ] µ= θ mi µ, θµ max. 4.2 Batch ML I cases where a set of observatios Y : is available, we describe here a batch off-lie versio BML of the previous algorithm. This algorithm maximizes the log-likelihood log p θ Y : usig a SA recursio at iteratio m give by θ m = θ m + γ m log p θm Y :, 37 where log p θm Y : is a estimate of the derivative of the log-likelihood evaluated at poit θ m. This estimate ca be obtaied usig a modified versio of the RML method as follows: At iteratio m, we ru the RML algorithm from time to time by omittig the parameter update step ad keepig the parameter value fixed at the curret estimated value θ m. At the ed of the ru, a Mote Carlo estimate of the derivative of the log-likelihood ca be computed as log p θm Y : = k= p θm Y k Y :k = p θm Y k Y :k k= j= ρj k j= aj k where Y : =. This is used to update the parameter to θ m, as give by Numerical Study The RML ad BML algorithms were tested, based o artificial ad real observatios. 5. Liear Gaussia State Space Model We first cosider the followig scalar liear Gaussia state space model σ 2 V X + = φx + σ V V +, X N, φ 2 Y = X + σ W W i.i.d. i.i.d. where V N, ad W N,. We are iterested i estimatig the parameter θ φ, σ V, σ W. I such a model, the optimal filter is give by the Kalma filter ad exact expressios for the first ad secod derivative of the filter ca be obtaied. This allows us to compare our umerical methods with the groud truth. The RML algorithm was implemeted usig the optimal importace desity q θ x Y, x g θ Y x f θ x x ad N = particles. Figure 2 displays the aalytical posterior desity ad its derivatives with respect to φ ad compares them with the particle approximatios we obtaied. The aalytical ad umerical values.4.2 px Y : φ px Y : φ px Y : Particle Approx. ve Particle Approx. ve True measure Figure 2: Liear Gaussia state space example: Aalytical optimal filter ad its first ad secod derivative w.r.t. φ ad the particle approximatios obtaied usig the proposed method. of the score vector log p θ Y Y : ad the Hessia matrix 2 log p θ Y Y : were compared up to =. These were almost idistiguishable. A example of the compariso results obtaied for the compoet 2 log p θ Y Y : φ 2 is show i Figure log p φ Y Y : Aalytical Results Particle Approxim Figure 3: Aalytical ad umerical results for 2 log p θ Y Y : / φ 2 for the liear Gaussia state space model usig N =. 7

8 5.2 Stochastic Volatility Model The RML algorithm was implemeted usig the followig Stochastic Volatility model σ 2 X + = φx + σv +, X N, φ 2 Y = β exp X /2 W i.i.d. i.i.d. where V N, ad W N,. We are iterested i estimatig the true parameter θ σ, φ, β =.35,.85,.65 from simulated data, where Θ =, Ξ,, Ξ with Ξ =. We use q θ x Y, x = f θ x x ad N = particles. As it ca be see for the results i Figure 4, the estimate coverged to a value θ i the eighborhood of the true parameter. We the applied our BML method to the poud/dollar daily exchage rates; see [9]. This time series cosists of 945 data poits. The parameter estimates for M = iteratios usig N = particles are show i Figure 5. Our results are cosistet with results obtaied i [9] Time steps iteratios Figure 5: Sequece of BML parameter estimates for θ m = σ m, φ m, β m ad N =. From top to bottom: φ m, β m ad σ m True φ Estimate.5.5 Time steps x Figure 4: Sequece of RML parameter estimates for θ = σ, φ, β ad N =. From top to bottom: φ, β ad σ. The true values were θ =.35,.85, Parameter trackig A uique advatage of the RML algorithm of sectio 4., is its ability to track variatios i θ. A stadard approach to track a time-varyig parameter is to set the step-size to a small positive umber γ, istead of a decreasig sequece γ [7]. The choice of value for γ will be a trade-off betwee trackig capability large γ ad low estimatio oise aroud the parameter small γ. A example of the trackig performace of the RML algorithm based o the liear Gaussia state space model, havig a time-varyig drift parameter φ, is show i Figure 6. x 4 Figure 6: RML algorithm trackig performace for timevaryig φ usig N = particles. 6. Discussio This paper has preseted origial particle methods to estimate the first ad secod derivative of the optimal filter i geeral state-space models. The methods use o-stadard particle methods to approximate the Hah-Jorda decompositio of the resultat siged measures. This allows the calculatio of accurate approximatios to the score vector ad the Hessia matrix of the log-likelihood with respect to the model parameters. Based o this, we propose a recursive ad a batch algorithm to perform ML parameter estimatio usig a gradiet ascet method. The Hessia estimate ca be used as a adaptive step-size i the gradiet ascet recursio to provide faster covergece of the algorithm. 8

9 The computatioal cost of the proposed particle methods for the filter derivatives is quadratic i the umber of particles. Fast computatio methods ca however be employed to address this issue [3]. The proposed methods ca also be exteded to the case where it is possible to itegrate aalytically a subset of the state variables, such as the class of partially observed liear Gaussia state-space models ad coditioally liear Gaussia state-space models [2], [8]. Such extesios ca provide efficiet particle methods that reduce the variace of the Mote Carlo estimates. Refereces [] Adrieu C., Doucet A. ad Tadic V.B. 25 Olie parameter estimatio i geeral state space models, Proc. IEEE CDC/ECC [2] Adrieu, C. ad Doucet, A. 22. Particle filterig for partially observed gaussia state space models. J. Royal Statist. Soc. B, 64, [3] Beveiste, A., Métivier, M. ad Priouret, P. 99. Adaptive Algorithms ad Stochastic Approximatio. New York: Spriger-Verlag. [4] Bertsekas D. 999, Noliear Programmig, 2d Editio, Athea Scietific. [5] Cérou F., LeGlad F. ad Newto N.J. 2 Stochastic particle methods for liear taget equatios. i Optimal Cotrol ad PDE s - Iovatios ad Applicatios eds. J. Mealdi, E. Rofma & A. Sulem, pp , IOS Press, Amsterdam. [6] Doucet, A. ad Tadić, V.B. 23. Parameter estimatio i geeral state-space models usig particle methods. A. Ist. Stat. Math., 55, [7] Doucet A., Godsill S.J. ad Adrieu C. 2 O sequetial Mote Carlo samplig methods for Bayesia filterig, Statist. Comput., vol., pp [8] Doucet, A., de Freitas, J.F.G. ad Gordo N.J. eds. 2. Sequetial Mote Carlo Methods i Practice. New York: Spriger-Verlag. [] Guyader, A., LeGlad, F. ad Oudjae, N. 23 A particle implemetatio of the recursive MLE for partially observed diffusios. Proceedigs of the 3th IFAC Symposium o System Idetificatio, [2] Kitagawa G. 996 Mote Carlo filter ad smoother for o-gaussia oliear state space models. J. Comput. Graph. Statist., 5, -25. [3] Klaas, M., Lag, D., Hamze, F. ad de Freitas, N. 24 Fast probability propagatio: Beyod beliefs. Techical report, CS Departmet, Uiversity of British Columbia. [4] LeGlad F. ad Mevel L. 997 Recursive idetificatio i hidde Markov models, Proc. 36th IEEE Cof. Decisio ad Cotrol, pp [5] Liu J.S. ad Che R. 998 Sequetial Mote Carlo methods for dyamic systems, J. Am. Statist. Ass., vol. 93, pp [6] Liu J. ad West M. 2 Combied parameter ad state estimatio i simulatio-based filterig, I Sequetial Mote Carlo Methods i Practice eds Doucet A., de Freitas J.F.G. ad Gordo N.J. New York: Spriger-Verlag,. [7] Ljug L. ad Söderström T. 983, Theory ad Practice of Recursive Idetificatio, MIT Press, Cambridge. [8] Pflug G.C. 996, Optimizatio of Stochastic Models. Kluwer. [9] Spall J. C. 2, Adaptive stochastic approximatio by the simultaeous perturbatio method, IEEE Tras. Autom. Cotr., vol. 45, pp [2] Storvik G. 22, Particle filters i state space models with the presece of ukow static parameters, IEEE. Tras. Sigal Processig, vol. 5, pp [2] Tadić V.B. ad Doucet A. 25 Expoetial forgettig ad geometric ergodicity for optimal filterig i geeral state-space models. Stochastic Processes ad Their Applicatios, vol. 5, pp [9] Durbi, J. ad Koopma, S. J. 2. Time series aalysis of o-gaussia observatios based o state space models from both classical ad Bayesia perspectives with discussio. J. R. Statist. Soc. B, 62, [] Fearhead P. 22 MCMC, sufficiet statistics ad particle filter, J. Comp. Graph. Stat., vol., pp

Gradient-free Maximum Likelihood Parameter Estimation with Particle Filters

Gradient-free Maximum Likelihood Parameter Estimation with Particle Filters Proceedigs of the 2006 America Cotrol Coferece Mieapolis, Miesota, USA, Jue 4-6, 2006 ThB082 Gradiet-free Maximum Likelihood Parameter Estimatio with Particle Filters George Poyiadjis, Sumeetpal S Sigh

More information

Sequential Monte Carlo Methods - A Review. Arnaud Doucet. Engineering Department, Cambridge University, UK

Sequential Monte Carlo Methods - A Review. Arnaud Doucet. Engineering Department, Cambridge University, UK Sequetial Mote Carlo Methods - A Review Araud Doucet Egieerig Departmet, Cambridge Uiversity, UK http://www-sigproc.eg.cam.ac.uk/ ad2/araud doucet.html ad2@eg.cam.ac.uk Istitut Heri Poicaré - Paris - 2

More information

Lecture 2: Monte Carlo Simulation

Lecture 2: Monte Carlo Simulation STAT/Q SCI 43: Itroductio to Resamplig ethods Sprig 27 Istructor: Ye-Chi Che Lecture 2: ote Carlo Simulatio 2 ote Carlo Itegratio Assume we wat to evaluate the followig itegratio: e x3 dx What ca we do?

More information

Expectation-Maximization Algorithm.

Expectation-Maximization Algorithm. Expectatio-Maximizatio Algorithm. Petr Pošík Czech Techical Uiversity i Prague Faculty of Electrical Egieerig Dept. of Cyberetics MLE 2 Likelihood.........................................................................................................

More information

A Note on Effi cient Conditional Simulation of Gaussian Distributions. April 2010

A Note on Effi cient Conditional Simulation of Gaussian Distributions. April 2010 A Note o Effi ciet Coditioal Simulatio of Gaussia Distributios A D D C S S, U B C, V, BC, C April 2010 A Cosider a multivariate Gaussia radom vector which ca be partitioed ito observed ad uobserved compoetswe

More information

Exponential Families and Bayesian Inference

Exponential Families and Bayesian Inference Computer Visio Expoetial Families ad Bayesia Iferece Lecture Expoetial Families A expoetial family of distributios is a d-parameter family f(x; havig the followig form: f(x; = h(xe g(t T (x B(, (. where

More information

Lainiotis filter implementation. via Chandrasekhar type algorithm

Lainiotis filter implementation. via Chandrasekhar type algorithm Joural of Computatios & Modellig, vol.1, o.1, 2011, 115-130 ISSN: 1792-7625 prit, 1792-8850 olie Iteratioal Scietific Press, 2011 Laiiotis filter implemetatio via Chadrasehar type algorithm Nicholas Assimais

More information

Chapter 2 The Monte Carlo Method

Chapter 2 The Monte Carlo Method Chapter 2 The Mote Carlo Method The Mote Carlo Method stads for a broad class of computatioal algorithms that rely o radom sampligs. It is ofte used i physical ad mathematical problems ad is most useful

More information

Monte Carlo Integration

Monte Carlo Integration Mote Carlo Itegratio I these otes we first review basic umerical itegratio methods (usig Riema approximatio ad the trapezoidal rule) ad their limitatios for evaluatig multidimesioal itegrals. Next we itroduce

More information

3/8/2016. Contents in latter part PATTERN RECOGNITION AND MACHINE LEARNING. Dynamical Systems. Dynamical Systems. Linear Dynamical Systems

3/8/2016. Contents in latter part PATTERN RECOGNITION AND MACHINE LEARNING. Dynamical Systems. Dynamical Systems. Linear Dynamical Systems Cotets i latter part PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 13: SEQUENTIAL DATA Liear Dyamical Systems What is differet from HMM? Kalma filter Its stregth ad limitatio Particle Filter Its simple

More information

Statistical Inference Based on Extremum Estimators

Statistical Inference Based on Extremum Estimators T. Rotheberg Fall, 2007 Statistical Iferece Based o Extremum Estimators Itroductio Suppose 0, the true value of a p-dimesioal parameter, is kow to lie i some subset S R p : Ofte we choose to estimate 0

More information

Lecture 11 and 12: Basic estimation theory

Lecture 11 and 12: Basic estimation theory Lecture ad 2: Basic estimatio theory Sprig 202 - EE 94 Networked estimatio ad cotrol Prof. Kha March 2 202 I. MAXIMUM-LIKELIHOOD ESTIMATORS The maximum likelihood priciple is deceptively simple. Louis

More information

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n.

Resampling Methods. X (1/2), i.e., Pr (X i m) = 1/2. We order the data: X (1) X (2) X (n). Define the sample median: ( n. Jauary 1, 2019 Resamplig Methods Motivatio We have so may estimators with the property θ θ d N 0, σ 2 We ca also write θ a N θ, σ 2 /, where a meas approximately distributed as Oce we have a cosistet estimator

More information

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator

Economics 241B Relation to Method of Moments and Maximum Likelihood OLSE as a Maximum Likelihood Estimator Ecoomics 24B Relatio to Method of Momets ad Maximum Likelihood OLSE as a Maximum Likelihood Estimator Uder Assumptio 5 we have speci ed the distributio of the error, so we ca estimate the model parameters

More information

Lecture 19: Convergence

Lecture 19: Convergence Lecture 19: Covergece Asymptotic approach I statistical aalysis or iferece, a key to the success of fidig a good procedure is beig able to fid some momets ad/or distributios of various statistics. I may

More information

THE KALMAN FILTER RAUL ROJAS

THE KALMAN FILTER RAUL ROJAS THE KALMAN FILTER RAUL ROJAS Abstract. This paper provides a getle itroductio to the Kalma filter, a umerical method that ca be used for sesor fusio or for calculatio of trajectories. First, we cosider

More information

Product measures, Tonelli s and Fubini s theorems For use in MAT3400/4400, autumn 2014 Nadia S. Larsen. Version of 13 October 2014.

Product measures, Tonelli s and Fubini s theorems For use in MAT3400/4400, autumn 2014 Nadia S. Larsen. Version of 13 October 2014. Product measures, Toelli s ad Fubii s theorems For use i MAT3400/4400, autum 2014 Nadia S. Larse Versio of 13 October 2014. 1. Costructio of the product measure The purpose of these otes is to preset the

More information

Chandrasekhar Type Algorithms. for the Riccati Equation of Lainiotis Filter

Chandrasekhar Type Algorithms. for the Riccati Equation of Lainiotis Filter Cotemporary Egieerig Scieces, Vol. 3, 00, o. 4, 9-00 Chadrasekhar ype Algorithms for the Riccati Equatio of Laiiotis Filter Nicholas Assimakis Departmet of Electroics echological Educatioal Istitute of

More information

Advanced Sequential Monte Carlo Methods

Advanced Sequential Monte Carlo Methods Advaced Sequetial Mote Carlo Methods Araud Doucet Departmets of Statistics & Computer Sciece Uiversity of British Columbia A.D. () / 35 Geeric Sequetial Mote Carlo Scheme At time =, sample q () ad set

More information

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1

EECS564 Estimation, Filtering, and Detection Hwk 2 Solns. Winter p θ (z) = (2θz + 1 θ), 0 z 1 EECS564 Estimatio, Filterig, ad Detectio Hwk 2 Sols. Witer 25 4. Let Z be a sigle observatio havig desity fuctio where. p (z) = (2z + ), z (a) Assumig that is a oradom parameter, fid ad plot the maximum

More information

Bayesian Methods: Introduction to Multi-parameter Models

Bayesian Methods: Introduction to Multi-parameter Models Bayesia Methods: Itroductio to Multi-parameter Models Parameter: θ = ( θ, θ) Give Likelihood p(y θ) ad prior p(θ ), the posterior p proportioal to p(y θ) x p(θ ) Margial posterior ( θ, θ y) is Iterested

More information

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ.

w (1) ˆx w (1) x (1) /ρ and w (2) ˆx w (2) x (2) /ρ. 2 5. Weighted umber of late jobs 5.1. Release dates ad due dates: maximimizig the weight of o-time jobs Oce we add release dates, miimizig the umber of late jobs becomes a sigificatly harder problem. For

More information

R. van Zyl 1, A.J. van der Merwe 2. Quintiles International, University of the Free State

R. van Zyl 1, A.J. van der Merwe 2. Quintiles International, University of the Free State Bayesia Cotrol Charts for the Two-parameter Expoetial Distributio if the Locatio Parameter Ca Take o Ay Value Betwee Mius Iity ad Plus Iity R. va Zyl, A.J. va der Merwe 2 Quitiles Iteratioal, ruaavz@gmail.com

More information

Random Variables, Sampling and Estimation

Random Variables, Sampling and Estimation Chapter 1 Radom Variables, Samplig ad Estimatio 1.1 Itroductio This chapter will cover the most importat basic statistical theory you eed i order to uderstad the ecoometric material that will be comig

More information

Properties and Hypothesis Testing

Properties and Hypothesis Testing Chapter 3 Properties ad Hypothesis Testig 3.1 Types of data The regressio techiques developed i previous chapters ca be applied to three differet kids of data. 1. Cross-sectioal data. 2. Time series data.

More information

Topic 9: Sampling Distributions of Estimators

Topic 9: Sampling Distributions of Estimators Topic 9: Samplig Distributios of Estimators Course 003, 2016 Page 0 Samplig distributios of estimators Sice our estimators are statistics (particular fuctios of radom variables), their distributio ca be

More information

Introduction to Optimization Techniques. How to Solve Equations

Introduction to Optimization Techniques. How to Solve Equations Itroductio to Optimizatio Techiques How to Solve Equatios Iterative Methods of Optimizatio Iterative methods of optimizatio Solutio of the oliear equatios resultig form a optimizatio problem is usually

More information

Lecture 33: Bootstrap

Lecture 33: Bootstrap Lecture 33: ootstrap Motivatio To evaluate ad compare differet estimators, we eed cosistet estimators of variaces or asymptotic variaces of estimators. This is also importat for hypothesis testig ad cofidece

More information

A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS

A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS J. Japa Statist. Soc. Vol. 41 No. 1 2011 67 73 A RANK STATISTIC FOR NON-PARAMETRIC K-SAMPLE AND CHANGE POINT PROBLEMS Yoichi Nishiyama* We cosider k-sample ad chage poit problems for idepedet data i a

More information

Efficient Block Sampling Strategies for Sequential Monte Carlo Methods

Efficient Block Sampling Strategies for Sequential Monte Carlo Methods Efficiet Block Samplig Strategies for Sequetial Mote Carlo Methods Araud DOUCET, Mark BRIERS, ad Stéphae SÉNÉCA Sequetial Mote Carlo SMC) methods are a powerful set of simulatio-based techiques for samplig

More information

The Choquet Integral with Respect to Fuzzy-Valued Set Functions

The Choquet Integral with Respect to Fuzzy-Valued Set Functions The Choquet Itegral with Respect to Fuzzy-Valued Set Fuctios Weiwei Zhag Abstract The Choquet itegral with respect to real-valued oadditive set fuctios, such as siged efficiecy measures, has bee used i

More information

Stat 421-SP2012 Interval Estimation Section

Stat 421-SP2012 Interval Estimation Section Stat 41-SP01 Iterval Estimatio Sectio 11.1-11. We ow uderstad (Chapter 10) how to fid poit estimators of a ukow parameter. o However, a poit estimate does ot provide ay iformatio about the ucertaity (possible

More information

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss

ECE 901 Lecture 12: Complexity Regularization and the Squared Loss ECE 90 Lecture : Complexity Regularizatio ad the Squared Loss R. Nowak 5/7/009 I the previous lectures we made use of the Cheroff/Hoeffdig bouds for our aalysis of classifier errors. Hoeffdig s iequality

More information

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors

ECONOMETRIC THEORY. MODULE XIII Lecture - 34 Asymptotic Theory and Stochastic Regressors ECONOMETRIC THEORY MODULE XIII Lecture - 34 Asymptotic Theory ad Stochastic Regressors Dr. Shalabh Departmet of Mathematics ad Statistics Idia Istitute of Techology Kapur Asymptotic theory The asymptotic

More information

6.3 Testing Series With Positive Terms

6.3 Testing Series With Positive Terms 6.3. TESTING SERIES WITH POSITIVE TERMS 307 6.3 Testig Series With Positive Terms 6.3. Review of what is kow up to ow I theory, testig a series a i for covergece amouts to fidig the i= sequece of partial

More information

NEW FAST CONVERGENT SEQUENCES OF EULER-MASCHERONI TYPE

NEW FAST CONVERGENT SEQUENCES OF EULER-MASCHERONI TYPE UPB Sci Bull, Series A, Vol 79, Iss, 207 ISSN 22-7027 NEW FAST CONVERGENT SEQUENCES OF EULER-MASCHERONI TYPE Gabriel Bercu We itroduce two ew sequeces of Euler-Mascheroi type which have fast covergece

More information

A statistical method to determine sample size to estimate characteristic value of soil parameters

A statistical method to determine sample size to estimate characteristic value of soil parameters A statistical method to determie sample size to estimate characteristic value of soil parameters Y. Hojo, B. Setiawa 2 ad M. Suzuki 3 Abstract Sample size is a importat factor to be cosidered i determiig

More information

Optimally Sparse SVMs

Optimally Sparse SVMs A. Proof of Lemma 3. We here prove a lower boud o the umber of support vectors to achieve geeralizatio bouds of the form which we cosider. Importatly, this result holds ot oly for liear classifiers, but

More information

Monte Carlo Optimization to Solve a Two-Dimensional Inverse Heat Conduction Problem

Monte Carlo Optimization to Solve a Two-Dimensional Inverse Heat Conduction Problem Australia Joural of Basic Applied Scieces, 5(): 097-05, 0 ISSN 99-878 Mote Carlo Optimizatio to Solve a Two-Dimesioal Iverse Heat Coductio Problem M Ebrahimi Departmet of Mathematics, Karaj Brach, Islamic

More information

Entropy Rates and Asymptotic Equipartition

Entropy Rates and Asymptotic Equipartition Chapter 29 Etropy Rates ad Asymptotic Equipartitio Sectio 29. itroduces the etropy rate the asymptotic etropy per time-step of a stochastic process ad shows that it is well-defied; ad similarly for iformatio,

More information

4. Partial Sums and the Central Limit Theorem

4. Partial Sums and the Central Limit Theorem 1 of 10 7/16/2009 6:05 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 4. Partial Sums ad the Cetral Limit Theorem The cetral limit theorem ad the law of large umbers are the two fudametal theorems

More information

Stochastic Simulation

Stochastic Simulation Stochastic Simulatio 1 Itroductio Readig Assigmet: Read Chapter 1 of text. We shall itroduce may of the key issues to be discussed i this course via a couple of model problems. Model Problem 1 (Jackso

More information

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15

17. Joint distributions of extreme order statistics Lehmann 5.1; Ferguson 15 17. Joit distributios of extreme order statistics Lehma 5.1; Ferguso 15 I Example 10., we derived the asymptotic distributio of the maximum from a radom sample from a uiform distributio. We did this usig

More information

Clustering. CM226: Machine Learning for Bioinformatics. Fall Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar.

Clustering. CM226: Machine Learning for Bioinformatics. Fall Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar. Clusterig CM226: Machie Learig for Bioiformatics. Fall 216 Sriram Sakararama Ackowledgmets: Fei Sha, Ameet Talwalkar Clusterig 1 / 42 Admiistratio HW 1 due o Moday. Email/post o CCLE if you have questios.

More information

Distribution of Random Samples & Limit theorems

Distribution of Random Samples & Limit theorems STAT/MATH 395 A - PROBABILITY II UW Witer Quarter 2017 Néhémy Lim Distributio of Radom Samples & Limit theorems 1 Distributio of i.i.d. Samples Motivatig example. Assume that the goal of a study is to

More information

Regression with an Evaporating Logarithmic Trend

Regression with an Evaporating Logarithmic Trend Regressio with a Evaporatig Logarithmic Tred Peter C. B. Phillips Cowles Foudatio, Yale Uiversity, Uiversity of Aucklad & Uiversity of York ad Yixiao Su Departmet of Ecoomics Yale Uiversity October 5,

More information

3. Z Transform. Recall that the Fourier transform (FT) of a DT signal xn [ ] is ( ) [ ] = In order for the FT to exist in the finite magnitude sense,

3. Z Transform. Recall that the Fourier transform (FT) of a DT signal xn [ ] is ( ) [ ] = In order for the FT to exist in the finite magnitude sense, 3. Z Trasform Referece: Etire Chapter 3 of text. Recall that the Fourier trasform (FT) of a DT sigal x [ ] is ω ( ) [ ] X e = j jω k = xe I order for the FT to exist i the fiite magitude sese, S = x [

More information

CSE 527, Additional notes on MLE & EM

CSE 527, Additional notes on MLE & EM CSE 57 Lecture Notes: MLE & EM CSE 57, Additioal otes o MLE & EM Based o earlier otes by C. Grat & M. Narasimha Itroductio Last lecture we bega a examiatio of model based clusterig. This lecture will be

More information

Goodness-Of-Fit For The Generalized Exponential Distribution. Abstract

Goodness-Of-Fit For The Generalized Exponential Distribution. Abstract Goodess-Of-Fit For The Geeralized Expoetial Distributio By Amal S. Hassa stitute of Statistical Studies & Research Cairo Uiversity Abstract Recetly a ew distributio called geeralized expoetial or expoetiated

More information

( θ. sup θ Θ f X (x θ) = L. sup Pr (Λ (X) < c) = α. x : Λ (x) = sup θ H 0. sup θ Θ f X (x θ) = ) < c. NH : θ 1 = θ 2 against AH : θ 1 θ 2

( θ. sup θ Θ f X (x θ) = L. sup Pr (Λ (X) < c) = α. x : Λ (x) = sup θ H 0. sup θ Θ f X (x θ) = ) < c. NH : θ 1 = θ 2 against AH : θ 1 θ 2 82 CHAPTER 4. MAXIMUM IKEIHOOD ESTIMATION Defiitio: et X be a radom sample with joit p.m/d.f. f X x θ. The geeralised likelihood ratio test g.l.r.t. of the NH : θ H 0 agaist the alterative AH : θ H 1,

More information

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise

First Year Quantitative Comp Exam Spring, Part I - 203A. f X (x) = 0 otherwise First Year Quatitative Comp Exam Sprig, 2012 Istructio: There are three parts. Aswer every questio i every part. Questio I-1 Part I - 203A A radom variable X is distributed with the margial desity: >

More information

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering CEE 5 Autum 005 Ucertaity Cocepts for Geotechical Egieerig Basic Termiology Set A set is a collectio of (mutually exclusive) objects or evets. The sample space is the (collectively exhaustive) collectio

More information

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence

Chapter 3. Strong convergence. 3.1 Definition of almost sure convergence Chapter 3 Strog covergece As poited out i the Chapter 2, there are multiple ways to defie the otio of covergece of a sequece of radom variables. That chapter defied covergece i probability, covergece i

More information

Efficient GMM LECTURE 12 GMM II

Efficient GMM LECTURE 12 GMM II DECEMBER 1 010 LECTURE 1 II Efficiet The estimator depeds o the choice of the weight matrix A. The efficiet estimator is the oe that has the smallest asymptotic variace amog all estimators defied by differet

More information

5 : Exponential Family and Generalized Linear Models

5 : Exponential Family and Generalized Linear Models 0-708: Probabilistic Graphical Models 0-708, Sprig 206 5 : Expoetial Family ad Geeralized Liear Models Lecturer: Matthew Gormley Scribes: Yua Li, Yichog Xu, Silu Wag Expoetial Family Probability desity

More information

RAINFALL PREDICTION BY WAVELET DECOMPOSITION

RAINFALL PREDICTION BY WAVELET DECOMPOSITION RAIFALL PREDICTIO BY WAVELET DECOMPOSITIO A. W. JAYAWARDEA Departmet of Civil Egieerig, The Uiversit of Hog Kog, Hog Kog, Chia P. C. XU Academ of Mathematics ad Sstem Scieces, Chiese Academ of Scieces,

More information

Double Stage Shrinkage Estimator of Two Parameters. Generalized Exponential Distribution

Double Stage Shrinkage Estimator of Two Parameters. Generalized Exponential Distribution Iteratioal Mathematical Forum, Vol., 3, o. 3, 3-53 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/.9/imf.3.335 Double Stage Shrikage Estimator of Two Parameters Geeralized Expoetial Distributio Alaa M.

More information

Output Analysis and Run-Length Control

Output Analysis and Run-Length Control IEOR E4703: Mote Carlo Simulatio Columbia Uiversity c 2017 by Marti Haugh Output Aalysis ad Ru-Legth Cotrol I these otes we describe how the Cetral Limit Theorem ca be used to costruct approximate (1 α%

More information

Math 155 (Lecture 3)

Math 155 (Lecture 3) Math 55 (Lecture 3) September 8, I this lecture, we ll cosider the aswer to oe of the most basic coutig problems i combiatorics Questio How may ways are there to choose a -elemet subset of the set {,,,

More information

32 estimating the cumulative distribution function

32 estimating the cumulative distribution function 32 estimatig the cumulative distributio fuctio 4.6 types of cofidece itervals/bads Let F be a class of distributio fuctios F ad let θ be some quatity of iterest, such as the mea of F or the whole fuctio

More information

Investigating the Significance of a Correlation Coefficient using Jackknife Estimates

Investigating the Significance of a Correlation Coefficient using Jackknife Estimates Iteratioal Joural of Scieces: Basic ad Applied Research (IJSBAR) ISSN 2307-4531 (Prit & Olie) http://gssrr.org/idex.php?joural=jouralofbasicadapplied ---------------------------------------------------------------------------------------------------------------------------

More information

Notes On Median and Quantile Regression. James L. Powell Department of Economics University of California, Berkeley

Notes On Median and Quantile Regression. James L. Powell Department of Economics University of California, Berkeley Notes O Media ad Quatile Regressio James L. Powell Departmet of Ecoomics Uiversity of Califoria, Berkeley Coditioal Media Restrictios ad Least Absolute Deviatios It is well-kow that the expected value

More information

Lecture 9: September 19

Lecture 9: September 19 36-700: Probability ad Mathematical Statistics I Fall 206 Lecturer: Siva Balakrisha Lecture 9: September 9 9. Review ad Outlie Last class we discussed: Statistical estimatio broadly Pot estimatio Bias-Variace

More information

The Expectation-Maximization (EM) Algorithm

The Expectation-Maximization (EM) Algorithm The Expectatio-Maximizatio (EM) Algorithm Readig Assigmets T. Mitchell, Machie Learig, McGraw-Hill, 997 (sectio 6.2, hard copy). S. Gog et al. Dyamic Visio: From Images to Face Recogitio, Imperial College

More information

THE SOLUTION OF NONLINEAR EQUATIONS f( x ) = 0.

THE SOLUTION OF NONLINEAR EQUATIONS f( x ) = 0. THE SOLUTION OF NONLINEAR EQUATIONS f( ) = 0. Noliear Equatio Solvers Bracketig. Graphical. Aalytical Ope Methods Bisectio False Positio (Regula-Falsi) Fied poit iteratio Newto Raphso Secat The root of

More information

Kolmogorov-Smirnov type Tests for Local Gaussianity in High-Frequency Data

Kolmogorov-Smirnov type Tests for Local Gaussianity in High-Frequency Data Proceedigs 59th ISI World Statistics Cogress, 5-30 August 013, Hog Kog (Sessio STS046) p.09 Kolmogorov-Smirov type Tests for Local Gaussiaity i High-Frequecy Data George Tauche, Duke Uiversity Viktor Todorov,

More information

1 of 7 7/16/2009 6:06 AM Virtual Laboratories > 6. Radom Samples > 1 2 3 4 5 6 7 6. Order Statistics Defiitios Suppose agai that we have a basic radom experimet, ad that X is a real-valued radom variable

More information

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5

CS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 5 CS434a/54a: Patter Recogitio Prof. Olga Veksler Lecture 5 Today Itroductio to parameter estimatio Two methods for parameter estimatio Maimum Likelihood Estimatio Bayesia Estimatio Itroducto Bayesia Decisio

More information

MAT1026 Calculus II Basic Convergence Tests for Series

MAT1026 Calculus II Basic Convergence Tests for Series MAT026 Calculus II Basic Covergece Tests for Series Egi MERMUT 202.03.08 Dokuz Eylül Uiversity Faculty of Sciece Departmet of Mathematics İzmir/TURKEY Cotets Mootoe Covergece Theorem 2 2 Series of Real

More information

CS284A: Representations and Algorithms in Molecular Biology

CS284A: Representations and Algorithms in Molecular Biology CS284A: Represetatios ad Algorithms i Molecular Biology Scribe Notes o Lectures 3 & 4: Motif Discovery via Eumeratio & Motif Represetatio Usig Positio Weight Matrix Joshua Gervi Based o presetatios by

More information

Estimation for Complete Data

Estimation for Complete Data Estimatio for Complete Data complete data: there is o loss of iformatio durig study. complete idividual complete data= grouped data A complete idividual data is the oe i which the complete iformatio of

More information

Modeling and Estimation of a Bivariate Pareto Distribution using the Principle of Maximum Entropy

Modeling and Estimation of a Bivariate Pareto Distribution using the Principle of Maximum Entropy Sri Laka Joural of Applied Statistics, Vol (5-3) Modelig ad Estimatio of a Bivariate Pareto Distributio usig the Priciple of Maximum Etropy Jagathath Krisha K.M. * Ecoomics Research Divisio, CSIR-Cetral

More information

Bayesian Control Charts for the Two-parameter Exponential Distribution

Bayesian Control Charts for the Two-parameter Exponential Distribution Bayesia Cotrol Charts for the Two-parameter Expoetial Distributio R. va Zyl, A.J. va der Merwe 2 Quitiles Iteratioal, ruaavz@gmail.com 2 Uiversity of the Free State Abstract By usig data that are the mileages

More information

Precise Rates in Complete Moment Convergence for Negatively Associated Sequences

Precise Rates in Complete Moment Convergence for Negatively Associated Sequences Commuicatios of the Korea Statistical Society 29, Vol. 16, No. 5, 841 849 Precise Rates i Complete Momet Covergece for Negatively Associated Sequeces Dae-Hee Ryu 1,a a Departmet of Computer Sciece, ChugWoo

More information

Quantile regression with multilayer perceptrons.

Quantile regression with multilayer perceptrons. Quatile regressio with multilayer perceptros. S.-F. Dimby ad J. Rykiewicz Uiversite Paris 1 - SAMM 90 Rue de Tolbiac, 75013 Paris - Frace Abstract. We cosider oliear quatile regressio ivolvig multilayer

More information

Dimension-free PAC-Bayesian bounds for the estimation of the mean of a random vector

Dimension-free PAC-Bayesian bounds for the estimation of the mean of a random vector Dimesio-free PAC-Bayesia bouds for the estimatio of the mea of a radom vector Olivier Catoi CREST CNRS UMR 9194 Uiversité Paris Saclay olivier.catoi@esae.fr Ilaria Giulii Laboratoire de Probabilités et

More information

Lecture 8: Solving the Heat, Laplace and Wave equations using finite difference methods

Lecture 8: Solving the Heat, Laplace and Wave equations using finite difference methods Itroductory lecture otes o Partial Differetial Equatios - c Athoy Peirce. Not to be copied, used, or revised without explicit writte permissio from the copyright ower. 1 Lecture 8: Solvig the Heat, Laplace

More information

Similarity Solutions to Unsteady Pseudoplastic. Flow Near a Moving Wall

Similarity Solutions to Unsteady Pseudoplastic. Flow Near a Moving Wall Iteratioal Mathematical Forum, Vol. 9, 04, o. 3, 465-475 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/0.988/imf.04.48 Similarity Solutios to Usteady Pseudoplastic Flow Near a Movig Wall W. Robi Egieerig

More information

Singular Continuous Measures by Michael Pejic 5/14/10

Singular Continuous Measures by Michael Pejic 5/14/10 Sigular Cotiuous Measures by Michael Peic 5/4/0 Prelimiaries Give a set X, a σ-algebra o X is a collectio of subsets of X that cotais X ad ad is closed uder complemetatio ad coutable uios hece, coutable

More information

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 016 MODULE : Statistical Iferece Time allowed: Three hours Cadidates should aswer FIVE questios. All questios carry equal marks. The umber

More information

Ω ). Then the following inequality takes place:

Ω ). Then the following inequality takes place: Lecture 8 Lemma 5. Let f : R R be a cotiuously differetiable covex fuctio. Choose a costat δ > ad cosider the subset Ωδ = { R f δ } R. Let Ωδ ad assume that f < δ, i.e., is ot o the boudary of f = δ, i.e.,

More information

Chapter 6 Principles of Data Reduction

Chapter 6 Principles of Data Reduction Chapter 6 for BST 695: Special Topics i Statistical Theory. Kui Zhag, 0 Chapter 6 Priciples of Data Reductio Sectio 6. Itroductio Goal: To summarize or reduce the data X, X,, X to get iformatio about a

More information

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015

ECE 8527: Introduction to Machine Learning and Pattern Recognition Midterm # 1. Vaishali Amin Fall, 2015 ECE 8527: Itroductio to Machie Learig ad Patter Recogitio Midterm # 1 Vaishali Ami Fall, 2015 tue39624@temple.edu Problem No. 1: Cosider a two-class discrete distributio problem: ω 1 :{[0,0], [2,0], [2,2],

More information

Convergence of random variables. (telegram style notes) P.J.C. Spreij

Convergence of random variables. (telegram style notes) P.J.C. Spreij Covergece of radom variables (telegram style otes).j.c. Spreij this versio: September 6, 2005 Itroductio As we kow, radom variables are by defiitio measurable fuctios o some uderlyig measurable space

More information

Math 113 Exam 3 Practice

Math 113 Exam 3 Practice Math Exam Practice Exam will cover.-.9. This sheet has three sectios. The first sectio will remid you about techiques ad formulas that you should kow. The secod gives a umber of practice questios for you

More information

1 Introduction to reducing variance in Monte Carlo simulations

1 Introduction to reducing variance in Monte Carlo simulations Copyright c 010 by Karl Sigma 1 Itroductio to reducig variace i Mote Carlo simulatios 11 Review of cofidece itervals for estimatig a mea I statistics, we estimate a ukow mea µ = E(X) of a distributio by

More information

It should be unbiased, or approximately unbiased. Variance of the variance estimator should be small. That is, the variance estimator is stable.

It should be unbiased, or approximately unbiased. Variance of the variance estimator should be small. That is, the variance estimator is stable. Chapter 10 Variace Estimatio 10.1 Itroductio Variace estimatio is a importat practical problem i survey samplig. Variace estimates are used i two purposes. Oe is the aalytic purpose such as costructig

More information

The picture in figure 1.1 helps us to see that the area represents the distance traveled. Figure 1: Area represents distance travelled

The picture in figure 1.1 helps us to see that the area represents the distance traveled. Figure 1: Area represents distance travelled 1 Lecture : Area Area ad distace traveled Approximatig area by rectagles Summatio The area uder a parabola 1.1 Area ad distace Suppose we have the followig iformatio about the velocity of a particle, how

More information

ENGI 4421 Confidence Intervals (Two Samples) Page 12-01

ENGI 4421 Confidence Intervals (Two Samples) Page 12-01 ENGI 44 Cofidece Itervals (Two Samples) Page -0 Two Sample Cofidece Iterval for a Differece i Populatio Meas [Navidi sectios 5.4-5.7; Devore chapter 9] From the cetral limit theorem, we kow that, for sufficietly

More information

Research Article A New Second-Order Iteration Method for Solving Nonlinear Equations

Research Article A New Second-Order Iteration Method for Solving Nonlinear Equations Abstract ad Applied Aalysis Volume 2013, Article ID 487062, 4 pages http://dx.doi.org/10.1155/2013/487062 Research Article A New Secod-Order Iteratio Method for Solvig Noliear Equatios Shi Mi Kag, 1 Arif

More information

Berry-Esseen bounds for self-normalized martingales

Berry-Esseen bounds for self-normalized martingales Berry-Essee bouds for self-ormalized martigales Xiequa Fa a, Qi-Ma Shao b a Ceter for Applied Mathematics, Tiaji Uiversity, Tiaji 30007, Chia b Departmet of Statistics, The Chiese Uiversity of Hog Kog,

More information

Orthogonal Gaussian Filters for Signal Processing

Orthogonal Gaussian Filters for Signal Processing Orthogoal Gaussia Filters for Sigal Processig Mark Mackezie ad Kiet Tieu Mechaical Egieerig Uiversity of Wollogog.S.W. Australia Abstract A Gaussia filter usig the Hermite orthoormal series of fuctios

More information

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4 MATH 30: Probability ad Statistics 9. Estimatio ad Testig of Parameters Estimatio ad Testig of Parameters We have bee dealig situatios i which we have full kowledge of the distributio of a radom variable.

More information

On an Application of Bayesian Estimation

On an Application of Bayesian Estimation O a Applicatio of ayesia Estimatio KIYOHARU TANAKA School of Sciece ad Egieerig, Kiki Uiversity, Kowakae, Higashi-Osaka, JAPAN Email: ktaaka@ifokidaiacjp EVGENIY GRECHNIKOV Departmet of Mathematics, auma

More information

Surveying the Variance Reduction Methods

Surveying the Variance Reduction Methods Iteratioal Research Joural of Applied ad Basic Scieces 2013 Available olie at www.irjabs.com ISSN 2251-838X / Vol, 7 (7): 427-432 Sciece Explorer Publicatios Surveyig the Variace Reductio Methods Arash

More information

Chapter 12 EM algorithms The Expectation-Maximization (EM) algorithm is a maximum likelihood method for models that have hidden variables eg. Gaussian

Chapter 12 EM algorithms The Expectation-Maximization (EM) algorithm is a maximum likelihood method for models that have hidden variables eg. Gaussian Chapter 2 EM algorithms The Expectatio-Maximizatio (EM) algorithm is a maximum likelihood method for models that have hidde variables eg. Gaussia Mixture Models (GMMs), Liear Dyamic Systems (LDSs) ad Hidde

More information

Mathematical Modeling of Optimum 3 Step Stress Accelerated Life Testing for Generalized Pareto Distribution

Mathematical Modeling of Optimum 3 Step Stress Accelerated Life Testing for Generalized Pareto Distribution America Joural of Theoretical ad Applied Statistics 05; 4(: 6-69 Published olie May 8, 05 (http://www.sciecepublishiggroup.com/j/ajtas doi: 0.648/j.ajtas.05040. ISSN: 6-8999 (Prit; ISSN: 6-9006 (Olie Mathematical

More information

VARIABLE RATE PARTICLE FILTERS FOR TRACKING APPLICATIONS. Simon Godsill and Jaco Vermaak

VARIABLE RATE PARTICLE FILTERS FOR TRACKING APPLICATIONS. Simon Godsill and Jaco Vermaak VARIABLE RATE PARTICLE FILTERS FOR TRACKING APPLICATIONS Simo Godsill ad Jaco Vermaak Sigal Processig Group Departmet of Egieerig, Uiversity of Cambridge, U.K. sjg@eg.cam.ac.uk, jv211@eg.cam.ac.uk ABSTRACT

More information

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization

ECE 901 Lecture 14: Maximum Likelihood Estimation and Complexity Regularization ECE 90 Lecture 4: Maximum Likelihood Estimatio ad Complexity Regularizatio R Nowak 5/7/009 Review : Maximum Likelihood Estimatio We have iid observatios draw from a ukow distributio Y i iid p θ, i,, where

More information

Lecture 3 The Lebesgue Integral

Lecture 3 The Lebesgue Integral Lecture 3: The Lebesgue Itegral 1 of 14 Course: Theory of Probability I Term: Fall 2013 Istructor: Gorda Zitkovic Lecture 3 The Lebesgue Itegral The costructio of the itegral Uless expressly specified

More information