IMATI-Milano Department

Similar documents
Bayesian analysis of a parametric semi-markov process applied to seismic data

CPSC 540: Machine Learning

LECTURE 15 Markov chain Monte Carlo

Computational statistics

Basic math for biology

Bayesian Methods for Machine Learning

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01

STA 4273H: Statistical Machine Learning

Answers and expectations

16 : Markov Chain Monte Carlo (MCMC)

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Graphical Models and Kernel Methods

The Metropolis-Hastings Algorithm. June 8, 2012

Markov Chain Monte Carlo (MCMC)

Modeling daily precipitation in Space and Time

Bayesian non-parametric model to longitudinally predict churn

An introduction to Bayesian statistics and model calibration and a host of related topics

Curve Fitting Re-visited, Bishop1.2.5

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

BAYESIAN METHODS FOR VARIABLE SELECTION WITH APPLICATIONS TO HIGH-DIMENSIONAL DATA

Hidden Markov Models for precipitation

Markov Chain Monte Carlo methods

ABC methods for phase-type distributions with applications in insurance risk problems

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

Extreme Value Analysis and Spatial Extremes

Advanced Machine Learning

Introduction to Computational Biology Lecture # 14: MCMC - Markov Chain Monte Carlo

Principles of Bayesian Inference

Non-Parametric Bayes

Approximate Bayesian Computation: a simulation based approach to inference

Monte Carlo Methods. Leon Gu CSD, CMU

arxiv: v3 [stat.me] 4 Apr 2014

Lecture 2 APPLICATION OF EXREME VALUE THEORY TO CLIMATE CHANGE. Rick Katz

Density Estimation. Seungjin Choi


Markov Chain Monte Carlo methods

Bayesian Nonparametrics

Markov Chain Monte Carlo Methods for Stochastic

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Markov Chain Monte Carlo Methods for Stochastic Optimization

The Particle Filter. PD Dr. Rudolph Triebel Computer Vision Group. Machine Learning for Computer Vision

Multistate Modeling and Applications

Step-Stress Models and Associated Inference

Introduction to Machine Learning CMU-10701

Particle Filtering Approaches for Dynamic Stochastic Optimization

Parameter Estimation

ECE276A: Sensing & Estimation in Robotics Lecture 10: Gaussian Mixture and Particle Filtering

Hmms with variable dimension structures and extensions

MCMC and Gibbs Sampling. Kayhan Batmanghelich

Metropolis-Hastings Algorithm


Stat 516, Homework 1

Bayesian Image Segmentation Using MRF s Combined with Hierarchical Prior Models

CS Lecture 19. Exponential Families & Expectation Propagation

A Bayesian methodology for systemic risk assessment in financial networks

Bayesian nonparametrics

Appendix F. Computational Statistics Toolbox. The Computational Statistics Toolbox can be downloaded from:

Markov Chain Monte Carlo

Evaluating the value of structural heath monitoring with longitudinal performance indicators and hazard functions using Bayesian dynamic predictions

Bagging During Markov Chain Monte Carlo for Smoother Predictions

Introduction to Bayesian Statistics 1

Dynamic models. Dependent data The AR(p) model The MA(q) model Hidden Markov models. 6 Dynamic models

Modelling trends in the ocean wave climate for dimensioning of ships

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods

Ronald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California

Statistics for extreme & sparse data

Introduction to Reliability Theory (part 2)

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

Characterizing Travel Time Reliability and Passenger Path Choice in a Metro Network

Likelihood-free MCMC

Theory of Stochastic Processes 8. Markov chain Monte Carlo

STOCHASTIC MODELING OF ENVIRONMENTAL TIME SERIES. Richard W. Katz LECTURE 5

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Infinite-State Markov-switching for Dynamic. Volatility Models : Web Appendix

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

CSC 2541: Bayesian Methods for Machine Learning

Monte Carlo in Bayesian Statistics

Variability within multi-component systems. Bayesian inference in probabilistic risk assessment The current state of the art

\ fwf The Institute for Integrating Statistics in Decision Sciences

Bayesian model selection for computer model validation via mixture model estimation

17 : Markov Chain Monte Carlo

STA 4273H: Statistical Machine Learning

13: Variational inference II

Bayesian linear regression

Construction of an Informative Hierarchical Prior Distribution: Application to Electricity Load Forecasting

DRAFT: A2.1 Activity rate Loppersum

Petr Volf. Model for Difference of Two Series of Poisson-like Count Data

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Spatial Statistics with Image Analysis. Outline. A Statistical Approach. Johan Lindström 1. Lund October 6, 2016

Doing Bayesian Integrals

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

Models for models. Douglas Nychka Geophysical Statistics Project National Center for Atmospheric Research

Bayesian phylogenetics. the one true tree? Bayesian phylogenetics

Generating synthetic rainfall using a disaggregation model

Statistical Inference and Methods

Principles of Bayesian Inference

Bayesian Point Process Modeling for Extreme Value Analysis, with an Application to Systemic Risk Assessment in Correlated Financial Markets

Transcription:

IMATI-Milano Department Main research topics: Data analysis Decision analysis Evaluation of seismic risk Industrial statistics Multivariate approximation Nonparametric Bayesian inference Reliability analysis Robustness of Bayesian analysis State-space modeling Stochastic models and parameter estimation in population dynamics Astrosiesta INAF Milano, Feb. 20th, 2009 1 / 32

Stochastic analysis of natural hazards Bruno Betrò, Antonella Bodini, Carla Brambilla, Renata Rotondi, Elisa Varini CNR - IMATI, Milano Astrosiesta INAF Milano, Feb. 20th, 2009 2 / 32

Natural hazards in Italy Natural phenomena of serious concern in Italy earthquakes landslides caused by extreme rainfalls Mathematical modelization requires the development of stochastic models Astrosiesta INAF Milano, Feb. 20th, 2009 3 / 32

Earthquake occurrence analysis Astrosiesta INAF Milano, Feb. 20th, 2009 4 / 32

Modelization of large earthquake sequences Renewal processes are considered appropriate models for sequences of large earthquakes, as one can assume that the stress accumulation process restarts after each event. The renewal model implies that the times between large seismic events can be considered as realizations of i.i.d. random variables T 1, T 2,.... If F is the common distribution function of the T i, the interest is in computing the occurrence probability at time t of an event in the next u years given the date t last of the last event before t F(t + u t last ) F(t t last ) 1 F(t t last ) Astrosiesta INAF Milano, Feb. 20th, 2009 5 / 32

Most used distributions exponential distribution, hazard function h(t) = f(t) = 1 F(t) λ, f density function gamma distribution h(t) = ba t a 1 e bt Γ(a) Γ(a, bt) lognormal distribution decreasing for a < 1, increasing for a > 1 f(t) h(t) = ( ) initially increasing, then decreasing, log t ξ 1 Φ σ 0 Weibull distribution h(t) = ca c t c 1 decreasing for c < 1, increasing if c > 1 if h( ) is multimodal? Astrosiesta INAF Milano, Feb. 20th, 2009 6 / 32

Nonparametric estimation of F Renewal processes are clearly a simplification of the real physical process but they can lead to useful results if F is properly estimated. Nonparametric methods are adaptive to anomalous behaviour in the data set We do not make any assumption on the functional form of the distribution F of the inter-event times but consider this distribution as a random function modelled by a mixture of Polya trees (Lavine, Ann. Statist., 20, 1225-1235 (1992)) Astrosiesta INAF Milano, Feb. 20th, 2009 7 / 32

The Bayesian approach X, Y absolutely continuous r.v. s; Bayes formula: f X Y (x y) = f X,Y(x, y) f Y (y) = f Y X (y x)f X (x) fy X (y u)f X (u) du in Bayesian inference, Bayes formula used for combining information from observations x expressed by the likelihood L(θ; x) = f(x θ), f(x θ) density of r.v. X = (X 1,..., X n ) a priori available information about unknown θ, assumed summarizable in a density function π(θ) (a priori density); θ is seen as a r.v. with density π(θ), f(x θ) is seen as a conditional density, a posteriori density of θ π(θ x) = f(x θ)π(θ) f(x u)π(u) du Astrosiesta INAF Milano, Feb. 20th, 2009 8 / 32

The Bayesian nonparametric approach if we don t want to specify the form of f(x θ) up to an unknown parameter, we can model f or the corresponding distribution function F as a stochastic process whose trajectories are densities or distribution functions (random distribution function, random probability measure). e.g., if Y(x), x R is a (right) continuous non-decreasing stochastic process such that Y( ) = 0 and Y( ) =, then F(x) = 1 exp( Y(x)) has trajectories satisfying conditions characterizing distribution functions. extending the parametric Bayesian approach to this situation makes it possible to obtain a posteriori information on F on the basis of a sample, e.g. E(F(x) x 1,..., x n ) Astrosiesta INAF Milano, Feb. 20th, 2009 9 / 32

Polya Trees E m = {ǫ : ǫ binary string of length m}, E = m=0 Em, E 0 = X separable measurable space (e.g. R n ) Π = {π m ; m = 0, 1, 2,...} nested partitions of X π 0 = X, π 1 = {B 0, B 1 }, B 0 B 1 =, B 0 B 1 = X π 2 = {B 00,B 01, B 10, B 11 },B 00 B 01 =,B 00 B 01 = B 0 B 10 B 11 =,B 10 B 11 = B 1 B ǫ0 ǫ E m ր, B ǫ π m ց B ǫ1 π m+1 Astrosiesta INAF Milano, Feb. 20th, 2009 10 / 32

Polya Trees (continued) A random probability measure P on X is said to have a Polya tree distribution with parameters (Π, A), if there exist nonnegative numbers A = {α ǫ, ǫ E } and random variables Y = {Y ǫ, ǫ E } s.t. all the random variables in Y are independent ǫ E, Y ǫ has a Beta distribution with parameters α ǫ,0 and α ǫ,1 m = 1, 2,... and ǫ E m P(B ǫ1,...,ǫ m ) = m j=1;ǫ j =0 Y ǫ 1,ǫ 2,...,ǫ j 1 m ( ) j=1;ǫ j =1 1 Yǫ1,ǫ 2,...,ǫ j 1 Astrosiesta INAF Milano, Feb. 20th, 2009 11 / 32

The role of E(P) Define the probability measure Q = E(P), by Q(B) = E(P(B)) for any measurable set B it is easy to compute Q(B ǫ ) B ǫ π m=0 m Q can be extended to the measurable sets generated by π m=0 m if the r.v. s X 1,X 2,... are a sample from P, i.e. given P, they are i.i.d. with distribution P, then P(X i B) = Q(B) Q is determined once Π, A are given; in the case X = R, a distribution function G(x) is given and the partition construction is lead by G; usual choices for αǫ 1,...,ǫ m : m 2, 2 m, k m (k > 1) Astrosiesta INAF Milano, Feb. 20th, 2009 12 / 32

Predictive distribution P X 1 = x 1 has still a PT distribution; simple updating rule: it is enough to add 1 to every α ǫ s.t. x 1 B ǫ exploiting the updating rule it is easy to compute P X 2 = x 2, X 1 = x 1 and so on if X = R, then it is easy to compute E(P((, x)) x 1,..., x n ) = E(F(x) x 1,..., x n ) i.e. a Bayesian estimate of the (unknown) distribution function of the observations x 1,..., x n Astrosiesta INAF Milano, Feb. 20th, 2009 13 / 32

Choice of G: the Generalized gamma distribution According to the information provided by the literature on the possible shape of the inter-event time distribution for strong earthquakes G is taken as a Generalized gamma distribution (X = R + ) distribution with density g(t; η, ξ, ρ) = ηξρ t ρη 1 exp( ξt η ), η,ξ, ρ > 0 Γ(ρ) this class of distributions properly includes usual distributions η = ρ = 1 exponential η = 1 gamma ρ = 1 Weibull lognormal ρ Astrosiesta INAF Milano, Feb. 20th, 2009 14 / 32

Mixtures of Polya Trees Mixtures of PT instead of single PT s have the advantage of decreasing influence of the partition scheme and of the parameters of G Given a random variable U (index) with mixing distribution H s.t. for each u we have P U = u PT (Π u, A u ) the distribution of a random measure P is said to be a mixture of Polya Trees if, for any measurable set S of probability measures on X Pr[P S] = Pr[P S u] H(du) parameters of G random vector u = (η, ξ, ρ), hierarchical structure: G(t u) H(u) = Gamma(η δ; g) Gamma(ξ β; a) Exp(ρ; b) δ Gamma(h, f) β Gamma(c, d) Astrosiesta INAF Milano, Feb. 20th, 2009 15 / 32

Estimation via Markov Chain Monte Carlo (MCMC) t = t 1,..., t n inter-event times update the current state z = (η, ξ, ρ, δ, β), performing an iteration of MCMC computation using Metropolis-Hastings within Gibbs sampling such that π(z i z i, t) is the equilibrium distribution of the Markov chain; repeat 500,000 times discarding the first 100,000 states (burn-in) every 50 iterations sample from the full conditional distribution of P z, t by sampling from the Polya tree generate values for the variates Y ǫ s, and hence obtain probabilities {p 1,..., p 2 m} of belonging to the sets at the level m draw samples of 50 inter-event times according to those probabilities use the simulated inter-event times to get a density estimate after a kernel smoothing Astrosiesta INAF Milano, Feb. 20th, 2009 16 / 32

Hazard maps of Italy Data sets: earthquakes with M w (Moment Magnitude) 5.3, occurred after 1600 up to 2002, drawn from the catalogue CPTI04; the inter-event times are calculated between shocks recorded in each of the seismogenic areas (DISS) belonging to the same tectonically homogeneous macroregion (MR) = we get 8 data sets used to estimate 8 density functions For each area the probability is obtained that an event occurs in the interval (t, t + u) given the date t 0 of the last event F(t + u t 0 ) F(t t 0 ) 1 F(t t 0 ) for each MR parameters of H(u) estimated through data from the other MR s and some information available in the seismic literature Astrosiesta INAF Milano, Feb. 20th, 2009 17 / 32

Homogeneous regions and seismogenic areas Astrosiesta INAF Milano, Feb. 20th, 2009 18 / 32

estimated densities nonparametric estimate of density functions Western Alps N = 5 α = j^2 Eastern Alps N = 11 α = j^2 density function 0.0 0.02 0.04 0 50 100 150 200 250 300 density function 0.0 0.010 0.020 0 50 100 150 200 250 300 interevent time interevent time Central Eastern Apennines N = 32 α = j^2 Central Western Apennines N = 32 α = j^2 density function 0.0 0.01 0.03 0 50 100 150 200 250 300 density function 0.0 0.02 0.04 0 50 100 150 200 250 300 interevent time interevent time Southern Eastern Apennines N = 9 α = j^2 Southern Western Apennines N = 12 α = j^2 density function 0.0 0.02 0.04 0 50 100 150 200 250 300 density function 0.0 0.005 0.015 0 50 100 150 200 250 300 interevent time interevent time Calabrian Arc N = 29 α = j^2 Sicily N = 15 α = j^2 density function 0.0 0.02 0.04 0 50 100 150 200 250 300 density function 0.0 0.005 0.015 0 50 100 150 200 250 300 interevent time interevent time Astrosiesta INAF Milano, Feb. 20th, 2009 19 / 32

Occurrence probabilities Probability of occurrence before 2013 α = j 2 8 10 12 14 16 18 46 46 44 42 40 38 44 42 40 38 Prob 0.323 0.300 0.276 0.253 0.229 0.206 0.182 0.159 0.135 0.112 0.088 0.065 36 8 10 12 14 16 18 36 Astrosiesta INAF Milano, Feb. 20th, 2009 20 / 32

Modelization of rainfall patterns Astrosiesta INAF Milano, Feb. 20th, 2009 21 / 32

The problem 6 9 Dec. 2004, Villagrande. Peak of 500 mm in 12 h. Astrosiesta INAF Milano, Feb. 20th, 2009 22 / 32

Astrosiesta INAF Milano, Feb. 20th, 2009 23 / 32

Aim of the work Because of high frequency of heavy or extreme rainfall events, which usually occur in a sudden way very local phenomenon = failure of GCM s Then characterize the occurrence of extreme events in the seasonal rainfall path highlight a reasonable, although necessarily simplified, rainfall mechanism derive hydrogeological risk indexes, flash flood thresholds, ecc. Astrosiesta INAF Milano, Feb. 20th, 2009 24 / 32

The study area 4 pluviometric stations of the Governmental Hydrographic Service: VILLAGRANDE ARZANA GAIRO JERZU Daily rainfall data Standard Period from 1961 1990 (WMO) Season from September January no other data are available many missing data from 1990 2000 and changes in the location of the pluviometric stations Astrosiesta INAF Milano, Feb. 20th, 2009 25 / 32

Data analysis From September January, 1961-1990 ARZANA GAIRO JERZU VILLAG. Altitude (m a.s.l.) 674 784 550 679 mean n. of wet days (nr) 45.9 38.3 45.9 30.0 mean daily rainfall (mm) (µ) 3.86 3.47 3.26 3.55 Std. Deviation (mm) (σ) 14.15 12.74 10.89 13.81 mean maximum (mm) 120.1 100.5 92.3 133.4 mean Cumulate (mm) 591.0 531.1 499.2 542.5 mean n. of events > 40mm (nr) 3.9 3.1 3.1 3.4 Outlier threshold (mm) (µ + 10σ) 145.2 130.9 112.2 141.6 Total n. of events µ + 10σ (nr) 9 9 7 7 Total n. of events 100 mm (nr) 22 16 9 15 complete records (nr) 22 24 23 16 Astrosiesta INAF Milano, Feb. 20th, 2009 26 / 32

Hidden Markov Models (HMMs): A graphical definition {X t } t rainfall process, {C t } hidden process rainfall X 1 X 2... X t 1 X t hidden states C 1 C 2... C t 1 C t Correspondence between hidden states and the concept of discrete weather state (Bárdossy & Plate, 1992) In B & P: states defined a priori (GCM s output) In HMMs: states inferred from data Astrosiesta INAF Milano, Feb. 20th, 2009 27 / 32

HMMs: A formal definition X t = (X t1,..., X tq ) r.v., q rain stations; x ti R + 0 C t {1,..., m} hidden process X 1:T := (X 1,...,X T ), C 1:T := (C 1,..., C T ) L( ) distribution of {C t } homogeneous, first order Markov Chain L(X t X 1:t 1, C 1:t ) = L(X t C t ) = i L(X ti C t ) L(X t C t ) does not depend on t L(X ti C t = c) = w ic δ 0 + (1 w ic )F( θ ic ) Charles et al. (1999) Astrosiesta INAF Milano, Feb. 20th, 2009 28 / 32

The adopted model F = mixture of Weibull distributions: W(x; α,β) = β α ( x α ) β 1 exp[ ( x α) β ], x > 0; α > 0, β > 0 Weibull distr. is an extreme value distribution is a transformation of the exponential distribution mixtures can capture different types of extreme values L(X ti C t = c) = w ic δ 0 + (1 w ic ) K γ k=1 kw ( (α k ic, βk ic )) γ k > K 0, γ k=1 k = 1, K = 1, 2, 3. Astrosiesta INAF Milano, Feb. 20th, 2009 29 / 32

Estimation and diagnostics MVNHMM toolbox, Kirshner (2005, 2007), http://www.cs.ualberta.ca/ sergey/mvnhmm/ (EM algorithm) X W(α, β) ( X α) β exp(1) β = 2, by fitting a Weibull distribution to annual maxima in each station: JUST A TRICK!! BIC + cross validation, for model selection goodness of fit, by comparing empirical and estimated relevant quantities Astrosiesta INAF Milano, Feb. 20th, 2009 30 / 32

Estimated model 6 hidden states L(X ti C t = c) = w ic δ 0 + (1 w ic ) 2 k=1 γ kw ( (α k ic, 2)) w ic C=1 C=2 C=3 C=4 C=5 C=6 Arzana 0.80 0.04 0.05 0.09 1.00 0.29 Gairo 0.79 0.05 0.17 0.35 1.00 0.49 Jerzu 0.69 0.02 0.04 0.13 1.00 0.16 Villagrande 0.93 0.06 0.10 0.35 1.00 0.76 Astrosiesta INAF Milano, Feb. 20th, 2009 31 / 32

Interpretation of states In terms of weather states: 5 = high pressure system 2 = moist currents from South East 3 = moist currents from South East, less intense phenomena, except for Gairo & Villagrande 4 = moderate rainfall 6 = rainfall from absent to weak 1 = negligeable rainfall, apart from Gairo Note that states appear to be well separated. Astrosiesta INAF Milano, Feb. 20th, 2009 32 / 32