Variability within multi-component systems. Bayesian inference in probabilistic risk assessment The current state of the art

Similar documents
Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Guidelines for Analysis of Data Related to Aging of Nuclear Power Plant Components and Systems

7. Estimation and hypothesis testing. Objective. Recommended reading

eqr094: Hierarchical MCMC for Bayesian System Reliability

Introduction to Bayesian Methods

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

Linear Regression. Data Model. β, σ 2. Process Model. ,V β. ,s 2. s 1. Parameter Model

Penalized Loss functions for Bayesian Model Choice

Reliability Growth in JMP 10

Foundations of Statistical Inference

Principles of Bayesian Inference

Principles of Bayesian Inference

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

10. Exchangeability and hierarchical models Objective. Recommended reading

Bayesian Inference. Chapter 4: Regression and Hierarchical Models

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017

Ronald Christensen. University of New Mexico. Albuquerque, New Mexico. Wesley Johnson. University of California, Irvine. Irvine, California

Introduction to Reliability Theory (part 2)

Data Analysis and Uncertainty Part 2: Estimation

STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01

Seminar über Statistik FS2008: Model Selection

Time-varying failure rate for system reliability analysis in large-scale railway risk assessment simulation

Evaluating the value of structural heath monitoring with longitudinal performance indicators and hazard functions using Bayesian dynamic predictions

Bayesian Networks 2:

Modelling Operational Risk Using Bayesian Inference

Represent processes and observations that span multiple levels (aka multi level models) R 2

Metropolis-Hastings Algorithm

Principles of Bayesian Inference

Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3

DIC, AIC, BIC, PPL, MSPE Residuals Predictive residuals

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Review. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Uncertain Inference and Artificial Intelligence

Estimation of Operational Risk Capital Charge under Parameter Uncertainty

Introduction to Bayesian Statistics with WinBUGS Part 4 Priors and Hierarchical Models

Bayesian Methods for Estimating the Reliability of Complex Systems Using Heterogeneous Multilevel Information

Bayesian Model Diagnostics and Checking

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

Bayesian Graphical Models

Principles of Bayesian Inference

Stat 451 Lecture Notes Markov Chain Monte Carlo. Ryan Martin UIC

Key Words: Lifetime Data Analysis (LDA), Probability Density Function (PDF), Goodness of fit methods, Chi-square method.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Using Model Selection and Prior Specification to Improve Regime-switching Asset Simulations

Estimation of Quantiles

COPYRIGHTED MATERIAL CONTENTS. Preface Preface to the First Edition

BAYESIAN MODELING OF DYNAMIC SOFTWARE GROWTH CURVE MODELS

Chapter 5: Generalized Linear Models

Bivariate Degradation Modeling Based on Gamma Process

Departamento de Economía Universidad de Chile

Bayesian Methods for Machine Learning

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

POMP inference via iterated filtering

Bayesian Prediction of Code Output. ASA Albuquerque Chapter Short Course October 2014

e 4β e 4β + e β ˆβ =0.765

Spatial discrete hazards using Hierarchical Bayesian Modeling

Quantile POD for Hit-Miss Data

A Probabilistic Framework for solving Inverse Problems. Lambros S. Katafygiotis, Ph.D.

Master of Science in Statistics A Proposal

Obnoxious lateness humor

36-463/663: Hierarchical Linear Models

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Non-Parametric Bayes

CS 361: Probability & Statistics

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

Bayesian Inference in Astronomy & Astrophysics A Short Course

Bayesian Inference. Chapter 1. Introduction and basic concepts

Bayesian Reliability Analysis: Statistical Challenges from Science-Based Stockpile Stewardship

ABC methods for phase-type distributions with applications in insurance risk problems

Bayesian linear regression

David Giles Bayesian Econometrics

Introduction to repairable systems STK4400 Spring 2011

Generalized Bayesian Inference with Sets of Conjugate Priors for Dealing with Prior-Data Conflict

Beta statistics. Keywords. Bayes theorem. Bayes rule

Density Estimation. Seungjin Choi

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Modelling geoadditive survival data

A Very Brief Summary of Bayesian Inference, and Examples

LISA Short Course Series Generalized Linear Models (GLMs) & Categorical Data Analysis (CDA) in R. Liang (Sally) Shan Nov. 4, 2014

Data Mining Chapter 4: Data Analysis and Uncertainty Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Bayesian Statistics. State University of New York at Buffalo. From the SelectedWorks of Joseph Lucke. Joseph F. Lucke

Bayesian Learning (II)

Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory

an introduction to bayesian inference

New Bayesian methods for model comparison

Use of Bayesian multivariate prediction models to optimize chromatographic methods

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Down by the Bayes, where the Watermelons Grow

Model comparison: Deviance-based approaches

\ fwf The Institute for Integrating Statistics in Decision Sciences

Bayesian Modeling of Accelerated Life Tests with Random Effects

TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1

Example: Ground Motion Attenuation

New Statistical Methods That Improve on MLE and GLM Including for Reserve Modeling GARY G VENTER

Bayesian Inference and MCMC

7. Estimation and hypothesis testing. Objective. Recommended reading

One-parameter models

CS 361: Probability & Statistics

Transcription:

PhD seminar series Probabilistics in Engineering : g Bayesian networks and Bayesian hierarchical analysis in engeering g Conducted by Prof. Dr. Maes, Prof. Dr. Faber and Dr. Nishijima Variability within multi-component systems Bayesian inference in probabilistic risk assessment The current state of the art Shuoyun Zhang ETH Zurich 1/47

Introduction Model the world Data Models Real Word Knowledge Models are not precise Data are not sufficient Natural variability Knowledge is subjective 2/47

Introduction Uncertain and parametric model Bayesian theorem as a inference tool Data Prior Knowledge New Knowledge π 1 ( θ x) = Θ ( ) 0 ( ) ( ) ( ) f x θ π θ f x θ π θ dθ 0 3/47

Introduction MCMC and the software package MCMC: Markov Chain Monte Carlo Software: Winbugs Openbugs Bayesian inference is applied broadly e.g. behavioral science, finance, human health, process control and ecological 4/47

Content Hierarchical modeling of variability (3) Modeling of time-dependent reliability (4) Modeling of random duration (5) Treatment of uncertain and missing data (6) Bayesian Regression models (7) Model selection and validation (8) Summary and conclusion (9,10) 5/47

Hierarchical modeling of variability Hierarchical Bayesian modeling π ( θ) = π ( ) ( ) 1 θ ϕ π2 ϕ dϕ Φ First-stage prior hyperprior π1 ( θ ϕ) is the population variability in θ for a given value ϕ π 2 ( ϕ ) is the distribution representing the uncertainty in ϕ 6/47

Hierarchical modeling of variability Illustrative example: Poisson process Possion process: ( λ ) ( λt) λt e f x =, x= 0,1 x!! x We have a prior belief that there is a source-to-source variability in λ Assuming π ( λϕ ) = π ( λαβ, ) ~ Gamma ( αβ, ) 1 1 7/47

Hierarchical modeling of variability Illustrative example Bayesian Network π ( λ α β ) 1 i, ( ) ( ) ( ) π θ = π1 θ ϕ π2 ϕ dϕ Φ ( ) f x λ i i * λ n n π ( ) ( ) ( ) ( ) ( 2 α, β x, t f α, β, x, t = 1,, ) 0 0 π λ 0 i α β π α β f xi λi dλ1 dλn i= 1 i= 1 n * πλ ( i x, t) = π1( λ, ) 1(, ) 2(,, ) 0 0 0 i α β π λ α β π α β xt dαdβ i= 1 * dλ1dλ2 dλi 1dλi+ 1 dλndλ = π λ 1 i xt,, α, β π α 2, β xt, d α d β ( ) ( ) * ( * π λ xt, ) π ( ) ( 1 λ α, β, xt, π2 α, β xtd, ) αdβ = 8/47

Hierarchical modeling of variability Illustrative example Winbugs Doodle model name: mu[i] type: logical link: identity value: lamda[i]*t[i] alpha beta lamda.star lamda[i] t[i] mu[i] x[i] for(i IN 1 : N) mean 2.5 10 6 90% credible interval 8 6 ( 3.9 10,8.0 10 ) π ( λ * xt, ) 9/47

Hierarchical modeling of variability Illustrative example: Emergency Diesel Generators (EDG) The failure probability p 1 of No.1 disel is of interest. (binomial distribution) x~ Bin( p, n ) π ( p αβ) = Beta( αβ) 1,, 10/47

Hierarchical modeling of variability Illustrative example: Emergency Diesel Generators name: p.star type: stochastic density: dbeta a alpha b beta lower bound upper bound alpha beta p.star p.fts[i] n[i] x[i] for(i IN 1 : N) Bayesian Network Winbugs Doodle model Result: 11/47

Content Hierarchical modeling of variability Modeling of time-dependent reliability (4) Modeling of random duration Treatment of uncertain and missing data Bayesian Regression models Model selection and validation Summary and conclusion 12/47

Modeling of time-dependent reliability Modeling time trends Modeling failure with repair 13/47

Modeling time trends Motivation It is sometimes the case that the usual Poisson and binomial models are rendered invalid because the parameter of interest ( λ or p, respectively) is not constant over time. Principal idea Generalized Linear Model (GLM) for λ, use log ( λ ) = a+ bt for use Logit function log p p, = a + bt 1 p b = 0 if, there is no time trend if b > 0, p or λ is increasing over time. 14/47

Modeling time trends Example: valve leakage if there appears to be any systematic time trend in p 15/47

Modeling time trends Example: valve leakage To check the clouds in our mind, we don t use any GLM model firstly, instead we update the Jeffreys prior with the data for each year, the result is: It looks likes that there is an increasing trend with time, but significant uncertainty in the individual estimates clouds this conclusion. 16/47

Modeling time trends Example: valve leakage Take the GLM to check, logit function Restult: b Density is highest for b > 0, suggesting an increasing trend in p 17/47

Modeling of time-dependent reliability Modeling time trends Modeling failure with repair 18/47

Modeling failure with repair Motivation Most of repair model is focus on same as new assumption, under the assumption of that the stochastic point process being observed is a renewal process. Less work has addressed the more reasonable assumption that repairs make the component same as old. Under the same as new assumption, the times between failures are independently, and identically distributed (iid). Under the same as old assumption for repair, the interarrival times are not iid; the distribution for the ith time is dependent upon the i-1th time. 19/47

Modeling failure with repair Principal idea for Poisson process: ( ) t λ () t dt 0 t ( ) = exp λ ( ) x ( 0 ) f x t dt x! ( ) If λ t is constant, it is the homogenous Poisson process (HPP) ; if λ ( t ) is dependent on time, it is the nonhomogenous Poisson process (NHPP) ; ( ) if λ t is increasing with time, the times between failures are decreasing with time; the component is aging or wearing out. If λ ( t ) is decreasing with time, the times between failures are increasing with time, the component is experiencing reliability growth. 20/47

Modeling failure with repair Principal idea Common form to model λ t include power-law process: λ ( t) α t = β β α 1 the loglinear model ( t) exp ( a bt) λ = + and the linear model λ ( t) = a+ bt ( ) 21/47

Content Hierarchical modeling of variability Modeling of time-dependent reliability Modeling of random duration (5) Treatment of uncertain and missing data Bayesian Regression models Model selection and validation Summary and conclusion 22/47

Modeling of random duration Motivation The simplest stochastic model for applications where time is the random variable of interest is the exponential distribution, which means that time-independent rate. There are numerous applications for time-dependent case, e.g. the rate of recovering offsite ac power at a commercial nuclear plant is often a decreasing function of time after power is lost, analyst is led to models that allow for time-dependent recovery rates, e.g. Weibull or lognormal distribution. Bayesian inference is more complicated when the likelihood function is other than exponential. The difficulty of Bayesian approach has led analyst to use frequentist t methods, such as MLE. The advent of Winbugs allows a fully Bayesian approach to the problem to be implemented. 23/47

Modeling of random duration Illustrative Example: Recovery of offsite ac power Use Weibull and lognormal distribution to model the rocovery time 24/47

Modeling of random duration Illustrative Example: Recovery of offsite ac power Use Weibull and lognormal distribution to model the model the rocovery time. We are interested in parameters: Weibull distribution: α (shape), (scale) Lognormal distribution: μ, σ β 25/47

Modeling of random duration Illustrative Example: Recovery of offsite ac power MLE result: Bayesian result: for Weibull for Lognormal 26/47

Modeling of random duration Illustrative Example: Recovery of offsite ac power Bayesian result: for Weibull distribution It is shown that the posterior distribution of has significant probability mass centered about 1.0, indicating that an exponential distribution might be a reasonable model. α 27/47

Content Hierarchical modeling of variability Modeling of time-dependent reliability Modeling of random duration Treatment of uncertain and missing data (6) Bayesian Regression models Model selection and validation Summary and conclusion 28/47

Treatment of uncertain and missing data Motivation It is common in risk applications to encounter situations in which the observed data, which would normally enter into Bayes theorem via the likelihood function, are either missing or the exact values are not known with certainty. Bayesian framework supplies a good tool for this case. We can simply assign the parameter of interest with a distribution that quantifies our available information, again reinforcing the idea that the Bayesian methodology encodes information via probability distribution. 29/47

Treatment of uncertain and missing data Illustrative example: motor-operated valves failing to open on demandd The number of demands is nominally 381, but could have been as high as about 440, and as low as 275. Assume that we have seen four failures to open. How to estimate the probability of failure p? 30/47

Treatment of uncertain and missing data Illustrative example: motor-operated valves failing to open on demandd n ( ) ~ U 275,440 Assign the demand number Give the failure probability bilit a Jeffrey prior p ~ Beta 0505 0.5, 0.5 Result: mean: 0.0128401284 95% credible interval(0.003738, 0.02804) ( ) Comparison result: ( n use 381) mean: 0.01179 95% credible interval(0.003604, 003604 0.02483) 02483) 4 = 0.01050 381 31/47

Treatment of uncertain and missing data Illustrative example: motor-operated valves failing to open on demandd If we don t know the failure number exactly, we take it as a uncertain value, and assume it could be 3, 4, 5, or 6, with value and assume it could be 3 4 5 or 6 with ( ) ( ) Pr( 6) = 0.05 and. Pr 3 = 0.1, Pr 4 = 0.7, Pr(5) = 0.15 32/47

Treatment of uncertain and missing data Illustrative example: motor-operated valves failing to open on demandd. N nupper x, n π f ( i p ) ( p) π avg ( pxn, ) = π ndn Pr x n nupper i lower i= 1 f ( x, ) ( ) ( ) 0 n i p n π p π n dndp lower Result: If n ( ) ~ U 275,440 mean: 0.01371 95% credible interval: (0.003759, 0.0302) ( ) ( ) If n=381 mean: 0.0122 95% credible interval: (0.003495, 0.02609) 33/47

Content Hierarchical modeling of variability Modeling of time-dependent reliability Modeling of random duration Treatment of uncertain and missing data Bayesian Regression models (7) Model selection and validation Summary and conclusion 34/47

Bayesian Regression models Regression model The unknown parameters denoted as β; this may be a scalar or a vector of length k The independent variables, X The dependent variable, Y. ( β, ) Y = f X Regression model could be used to predict the value of interest which will provide information to decision-maker prior. Bayesian Regression Model The analysis was frequentist in nature, probability distributions representing epistemic uncertainty in the input parameters were not available. The Bayesian perspective is needed to operate the regression process. 35/47

Bayesian Regression models Illustrative example: predict model for ill-fated launch of Challenger in January 1986 There are six O-rings on the shuttle, during each launch, the number of distress events defined as erosion or blow-by of a primary field O-ring, is modeled d as binomial i with parameters p and 6: X ~ Binomial ( p,6) Use regression model: logit(p) = a+b *temp + c*press To predict the distress events at 31 F which is the approximate temperature re for the disastrous s launch of the Challenger. 36/47

Bayesian Regression models Illustrative example: predict model for ill-fated launch of Challenger in January 1986 logit(p) = a+b *temp + c*press Result: the model predicts about 4 distress events at 31 F logit(p) = a+b *temp 37/47

Content Hierarchical modeling of variability Modeling of time-dependent reliability Modeling of random duration Treatment of uncertain and missing data Bayesian Regression models Model selection and validation (8) Summary and conclusion 38/47

Model selection and validation Motivation Bayesian posterior predictive statistics and Bayesian p- value Deviance information criterion (DIC) 39/47

Model selection and validation Motivation If the selected prior distribution is appropriate? How to check it? If there are some random models for candidate to model the variability, which one is more appropriate? How to check it? 40/47

Model selection and validation To solve these problem, summary statistics derived from the posterior predictive distribution provides a very good tool, 2 such as Bayesian χ and Cramer-von Mises statistic, which lead to a Bayesian analog of p-value. The deviance information criterion (DIC), a Bayesian analog of a penalized likelihood measure also provides a good framework for the model selection. 41/47

Bayesian posterior predictive statistics and Bayesian p-value Bayesian 2 χ x use the observed values of to form the observed statistic: ti ti ( x ) 2 2 obs, i μi xobs = 2 σ i i x generate replicate values of from its posterior predictive distribution to construct an analogous statistic: ( x ) 2 2 rep, i μi xrep = 2 i σ i 2 μ is the ith mean value, σ i is the ith variance. i ( p ) 2 2 If the model is very ygood, p-value χ χ should be around 0.5. Pr rep obs If we want to judge the model validation, we can set the cutoff, we will reject a model with a p-value to below the cutoff. We also can select a model from many candidate models whose p-value is closest to 0.5 42/47

Bayesian posterior predictive statistics and Bayesian p-value Cramer-von Mises statistics x Rank the observed data x as () 1 ( 2) ( N) obs xobs xobs xobs construct statistic 2 2 i 1 ( ) Dobs = Fobs, i where, Pr ( ) i Fobs i = X < xobs i 2N generate replicate values of x rep from its posterior predictive distribution, ( ) ( ) ( ) then rank the replicate value as x 1 2 N to construct an rep xrep xrep analogous statistic: D rep = Frep, i i 2 i 1 2N 2 where ( ), Pr ( ) i rep i = < rep F X x p-value is Pr ( Drep Dobs ) 43/47

Bayesian posterior predictive statistics and Bayesian p-value Illustrative example: source to source variability λ assume the constant t assume variable λ P-value 44/47

Model selection and validation Deviance information criterion (DIC) The DIC is a hierarchical modeling generalization of the AIC (Akaike Information Criterion) and BIC (Bayesian Information Criterion). It is particularly useful in Bayesian model selection problems where the posterior distributions of the models have been obtained by (MCMC) simulation. Define the deviance as D ( θ ) = 2log ( p ( y θ ) ), p ( y θ ) is the likelihood function. The expectation D = E is a measure of how well the θ D( θ ) model fits the data; the larger this is, the worse the fit. The effective number of parameters of the model is computed as pd = D D( θ ). θ is the expectation of θ. The deviance information criterion is calculated as: ( ) DIC = p + D = 2D D θ D Rule: The small the DIC is, the better the model fit. DIC can be get from Winbugs. 45/47

Summary and conclusion A variety of other model of the world are amenable to inference via Bayes theorem, including Bayesian belief networks(bbn), influence diagram, and fault tree. Uncertain parameters (in the epistemic sense) are inputs to the models used to infer the values of future observations, leading to an increase in scientific knowledge. The advent of MCMC-based sampling methods, coupled with easy-to- use software and powerful computers, allows us to encode information via Bayes Theorem for a large variety of problems, domains, model types, data sets, and complications. It is best done via Bayesian inference with modern computational tools, which eliminate the need for the approximations and ad hoc approaches of the past. 46/47

47/47