Sequential Importance Sampling for Rare Event Estimation with Computer Experiments

Similar documents
Polynomial chaos expansions for structural reliability analysis

Dynamic System Identification using HDMR-Bayesian Technique

Gaussian Processes for Computer Experiments

Combining probability and possibility to respect the real state of knowledge on uncertainties in the evaluation of safety margins

Model Calibration under Uncertainty: Matching Distribution Information

Basics of Uncertainty Analysis

Gradient Enhanced Universal Kriging Model for Uncertainty Propagation in Nuclear Engineering

Monte Carlo Studies. The response in a Monte Carlo study is a random variable.

Research Collection. Basics of structural reliability and links with structural design codes FBH Herbsttagung November 22nd, 2013.

component risk analysis

Stat 5101 Lecture Notes

Introduction to emulators - the what, the when, the why

MODEL COMPARISON CHRISTOPHER A. SIMS PRINCETON UNIVERSITY

Sequential approaches to reliability estimation and optimization based on kriging

Neutron inverse kinetics via Gaussian Processes

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Parametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012

Model comparison. Christopher A. Sims Princeton University October 18, 2016

Statistical Analysis of the Whole Body Specific Absorption Rate using Human Body Characteristics

BEST ESTIMATE PLUS UNCERTAINTY SAFETY STUDIES AT THE CONCEPTUAL DESIGN PHASE OF THE ASTRID DEMONSTRATOR

Stat 451 Lecture Notes Numerical Integration

Risk Estimation and Uncertainty Quantification by Markov Chain Monte Carlo Methods

Probability and Estimation. Alan Moses

Bayesian Modeling of Accelerated Life Tests with Random Effects

Step-Stress Models and Associated Inference

VCMC: Variational Consensus Monte Carlo

Structural Reliability

Uncertainty Quantification and Validation Using RAVEN. A. Alfonsi, C. Rabiti. Risk-Informed Safety Margin Characterization.

Estimation of Quantiles

Introduction to Probabilistic Machine Learning

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Logistic Regression Review Fall 2012 Recitation. September 25, 2012 TA: Selen Uguroglu

Problem Selected Scores

On Bayesian Computation

Inference for P(Y<X) in Exponentiated Gumbel Distribution

Computer Practical: Metropolis-Hastings-based MCMC

Probabilistic assessment of danger zones using a surrogate model of CFD simulations

Application of Copulas as a New Geostatistical Tool

J. Sadeghi E. Patelli M. de Angelis

Hierarchical sparse Bayesian learning for structural health monitoring. with incomplete modal data

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Fast Numerical Methods for Stochastic Computations

CPSC 540: Machine Learning

Authors : Eric CHOJNACKI IRSN/DPAM/SEMIC Jean-Pierre BENOIT IRSN/DSR/ST3C. IRSN : Institut de Radioprotection et de Sûreté Nucléaire

Efficient Aircraft Design Using Bayesian History Matching. Institute for Risk and Uncertainty

Overfitting, Bias / Variance Analysis

Density Estimation: ML, MAP, Bayesian estimation

Non-Bayesian Classifiers Part II: Linear Discriminants and Support Vector Machines

Uncertainty Quantification in Performance Evaluation of Manufacturing Processes

Statistical Inference Using Progressively Type-II Censored Data with Random Scheme

Random Eigenvalue Problems Revisited

An Integrated Approach for Characterization of Uncertainty in Complex Best Estimate Safety Assessment

Organization. I MCMC discussion. I project talks. I Lecture.

Statistical learning. Chapter 20, Sections 1 3 1

S. Freitag, B. T. Cao, J. Ninić & G. Meschke

Inverse Uncertainty Quantification using the Modular Bayesian Approach based on Gaussian Process, Part 2: Application to TRACE

Reliable Condition Assessment of Structures Using Uncertain or Limited Field Modal Data

Calibration and uncertainty quantification using multivariate simulator output

Covariance function estimation in Gaussian process regression

Bayesian analysis in nuclear physics

Statistical Measures of Uncertainty in Inverse Problems

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

Parameter Estimation Method Using Bayesian Statistics Considering Uncertainty of Information for RBDO

However, reliability analysis is not limited to calculation of the probability of failure.

Gaussian Models

Part 1: Expectation Propagation

CSC 2541: Bayesian Methods for Machine Learning

Bayesian Regression Linear and Logistic Regression

Introduction to Design and Analysis of Computer Experiments

Machine Learning 2017

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

F denotes cumulative density. denotes probability density function; (.)

Neutral Bayesian reference models for incidence rates of (rare) clinical events

Uncertainty Propagation and Global Sensitivity Analysis in Hybrid Simulation using Polynomial Chaos Expansion

PART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics

Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics

Default Priors and Effcient Posterior Computation in Bayesian

Monitoring Wafer Geometric Quality using Additive Gaussian Process

Uncertainty in energy system models

ECE521 week 3: 23/26 January 2017

Bayesian Reliability Analysis: Statistical Challenges from Science-Based Stockpile Stewardship

Design for Reliability and Robustness through probabilistic Methods in COMSOL Multiphysics with OptiY

New Insights into History Matching via Sequential Monte Carlo

Uncertainty Quantification of EBR-II Loss of Heat Sink Simulations with SAS4A/SASSYS-1 and DAKOTA

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

Inferring from data. Theory of estimators

Probabilistic numerics for deep learning

Statistical learning. Chapter 20, Sections 1 4 1

Fast Likelihood-Free Inference via Bayesian Optimization

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation

Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study

Bayesian Dynamic Linear Modelling for. Complex Computer Models

How to Validate Stochastic Finite Element Models from Uncertain Experimental Modal Data Yves Govers

Monte Carlo Simulation

Stochastic optimization - how to improve computational efficiency?

PILCO: A Model-Based and Data-Efficient Approach to Policy Search

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

Algorithms for Uncertainty Quantification

The Role of Uncertainty Quantification in Model Verification and Validation. Part 6 NESSUS Exercise Review and Model Calibration

Transcription:

Sequential Importance Sampling for Rare Event Estimation with Computer Experiments Brian Williams and Rick Picard LA-UR-12-22467 Statistical Sciences Group, Los Alamos National Laboratory

Abstract Importance sampling often drastically improves the variance of percentile and quantile estimators of rare events. We propose a sequential strategy for iterative refinement of importance distributions for sampling uncertain inputs to a computer model to estimate quantiles of model output or the probability that the model output exceeds a fixed or random threshold. A framework is introduced for updating a model surrogate to maximize its predictive capability for rare event estimation with sequential importance sampling. Examples of the proposed methodology involving materials strength and nuclear reactor applications will be presented. Slide 1

Outline Interested in rare event estimation Outputs obtained from computational model Uncertainties in operating conditions and physics variables Physics variables calibrated wrt reference experimental data In particular, quantile or percentile estimation One of q α or α is specified and the other is to be inferred q α may be random when inferring α Sequential importance sampling for improved inference Oversample region of parameter space producing rare events of interest Sequentially refine importance distributions for improved inference Slide 2

Examples Risk of released airborne load colliding with aircraft For control variable settings of interest, inference on probability of danger Structural reliability benchmark Inference on probability of failure of a four-branch series system Catastrophe due to pyroclastic flows Inference on probability of catastrophe within t years Volume of storage in sewer system required to meet certain environmental standards Inference on 95 th quantile Pressurizer failure in nuclear reactor loop Inference on probability of dangerous peak coolant temperature Slide 3

Percentile Estimates q α fixed PDF of operating conditions PDF of physics variables Monte Carlo estimate of α and associated uncertainty obtained via samples {(x (j), θ (j) ), j=1,,n} from the product density f(x) π(θ) Random threshold q α having PDF h(q) and CDF H(q) Brute force Improvement via Rao-Blackwellization Slide 4

Percentile α fixed Quantile Estimates Obtain samples {(x (j), θ (j) ), j=1,,n} from the product density f(x) π(θ) Evaluate computational model on samples, {η(x (j), θ (j) ), j=1,,n} Sort code evaluations and choose an appropriate order statistic from the list e.g. if α = 0.001 and N=10,000, choose the 11-th largest value Quantile estimates from simulations are not unique, e.g. any value between the 10-th and 11-th largest values may be chosen Uncertainty obtained e.g. by inverting binomial confidence intervals Slide 5

Importance Sampling Sample from alternative distribution g(x,θ), evaluate computational model, and weight each output Importance weight Importance density percentile estimator: quantile estimator: Smallest value of q α for which Importance density g(x,θ) ideally chosen to increase the likelihood of observing desired rare events Such knowledge may initially be vague: how to proceed? Slide 6 minimum variance importance density

A subset of variables may be responsible for generating rare events Relevant x r, θ r Irrelevant x i, θ i Relevant (Sensitive) Variables Minimum variance importance distribution Importance distribution g(x,θ) becomes Slide 7

Example: PTW Material Strength Model activation energy T/ T m (ρ) strain rate yield stress saturation stress T = temperature T m (ρ) = melting temp. atomic vibration time strain stress Slide 8 Model parameters θ Calibrated to experimental data Multivariate Normal Input conditions x Independent normal distributions

Adaptive Importance Sampling Choose g(x,θ) by iterative refinement 1. Sample from initial importance density g (1) (x,θ) = f(x) π(θ) 2. Given target probability or quantile level, determine which parameters are sensitive for producing large output values e.g. for target 0.0001 quantile, examine the 100 largest output order statistics calculated from 10 6 samples and compare the corresponding parameter samples to random parameter samples 1-in-10,000 1-in-100,000 Random Slide 9

Adaptive Importance Sampling 3. For sensitive parameters in (2), adopt a family of importance densities (e.g. Beta, (truncated) Normal, etc.). Fit the parameters of this family to the selected largest order statistics (e.g. method of moments, MLE). 4. Keep the conditional distribution from (1) for the insensitive parameters, given values of the sensitive parameters. 5. Combine the distributions of (3) and (4) to obtain g (2) (x,θ). 6. Repeat (2)-(5) until the updated importance distribution g (k) (x,θ) stabilizes in its approximation of the minimum variance importance distribution. Means, variances, and covariances of sensitive parameters are now estimated via their conditional distributions based on a random sample from g (k) (x,θ), e.g. Slide 10

Results: Variance Reduction Four sensitive PTW variables: Define variance reduction factor: VRF = Brute Force Variance Variance of IS quantity VRF > 1 : Importance Sampling Favored Target Quantile: 0.0001 0.001 0.0001 0.00001 0.000001 4-D VRF 3 3231 2245 225 10-D VRF 0.3 1237 204 20 Slide 11

Results: Adaptive Importance Sampling Four sensitive PTW variables: Results of iterative process: Minimum Variance Importance Dist n Convergence often achieved rapidly! Slide 12

Normalized Importance Sampling Estimates In many applications, π(θ) is a posterior distribution with unknown normalizing constant: π(θ) = c norm k(θ) Modified weights and estimators: Although, w distribution has a sample mean (SD) of 0.007 (4.1) based on 10 6 samples 98% of weights less than sample mean Slide 13 Target Quantile: 0.0001 0.001 0.0001 0.00001 0.000001 4-D q 1512.6 1566.0 1610.5 1649.0 10-D q 1509.9 1566.0 1610.4 1648.7 Normalized q 1615.7 1653.8 1679.0 1718.8

Many Applications Will Require Code Surrogates Many computational models run too slowly for direct use in brute force or importance sampling based rare event inference Use training runs to develop a statistical surrogate model for the complex code (i.e., the emulator) based on a Gaussian process Deterministic code is interpolated with zero uncertainty Kriging Predict outputs evaluated at training runs z 1,..., z n correlations between prediction site z and training runs z 1,..., z n pairwise correlations between training runs z 1,..., z n Kriging Variance Slide 14

Previous Work Oakley, J. (2004). Estimating percentiles of uncertain computer code outputs, Applied Statistics, 53, 83-93. Quantile estimation with α = 0.05 Assuming a Gaussian Process model for unsampled code output: Generate a random function η (i) (.) from the posterior process Use Monte Carlo methods to estimate q 0.05,(i) Repeat previous steps to obtain a sample q 0.05,(1),,q 0.05,(K) Sequential design to improve emulation in relevant region of input space First stage: space-filling to learn about η(.) globally Second stage: IMSE criterion with weight function chosen as a density estimate from the K inputs corresponding to q 0.05,(1),,q 0.05,(K) Slide 15

Surrogate-based Percentile/Quantile Estimation Gaussian Process model with plug-in covariance parameter estimates, e.g. Process-based inference Slide 16

Surrogate Bias May Affect Inference Results Example: Approximation of PTW model based on 55 runs Importance distributions and efficiencies based on surrogate are similar to those based on direct model calls Mean Standard Deviation T p s 0 T p s 0 True 82 2.27 1.66 0.0105 21 2.1 0.18 0.00045 Surr. 95 4.25 1.73 0.0104 23 1.6 0.17 0.00047 However, quantile estimates from surrogate can be significantly different than the true values relative to Monte Carlo Uncertainty 1566.0 (true) vs. 1549.3 (surrogate) for target quantile of 0.0001 Slide 17

Convergence Assessment Through Iterative Refinement of Surrogate Choose design augmentation that minimizes integrated mean square error with respect to the currently estimated importance distributions for sensitive parameters A version of targeted IMSE (timse) Estimate correlation parameters based on current design D 0 For sensitive input i, select w i (z i ) to be its importance distribution Slide 18

Uncorrelated Importance Sampling Importance distribution g (k) (x,θ) is multivariate Gaussian IMSE calculation: Obtaining uncorrelated variables Relevant (sensitive) quantities Insensitive quantities Transformations Similar for insensitive variables Slide 19

Quantile Estimates May Converge Quickly Target quantile 0.0001 single-stage design Quantile bias at convergence consistent with average surrogate error Surrogate quality in importance region may benefit from higher frequency updates +5; +10 Slide 20

Or Quantile Estimates May Converge Slowly 0.00001 0.0001 0.001 Quantile bias at convergence consistent with average surrogate error Surrogate quality in importance region benefits from starting sequential design early Slide 21

Scenario: Pressurizer failure, followed by pump trip and initiation of SCRAM (insertion of control rods) Goal: Understand behavior of peak coolant temperature (PCT) in the reactor Interested in probability that PCT exceeds 700 o K Example: VR2plus Model Slide 22

VR2plus Details Single thermal-hydraulics loop with 21 components Working coolant is water at 16MPa and 600 o K, single-phase flow Nominal power output of this reactor is 15MW Calculations performed with reactor safety analysis code R7 (INL) Input Parameter Min Max Description PumpTripPre 15.6 MPa 15.7 MPa Min. pump pressure causing trip PumpStopTime 10 s 100 s Relaxation time of pump phase-out PumpPow 0.0 0.4 Pump end power SCRAMtemp 625 o K 635 o K Max. temp. causing SCRAM CRinject 0.025 0.24 Position of CR at end of SCRAM CRtime 10 s 50 s Relaxation time of CR system Slide 23 Input parameters assigned independent Uniform distributions on their ranges

VR2plus Analysis Quantile inference Target 0.0001 quantile of PCT distribution Percentile inference PCT > 700 o K Importance distribution Independent Beta distributions for sensitive parameters Independent Uniform distributions for insensitive parameters PumpPow and CRinject are sensitive Slide 24

Pressurizer Failure: Quantile Inference Importance Distributions Max = 0.4 Min = 0.025 Slide 25 Again, convergence benefits seen with higher frequency updates

Pressurizer Failure: Percentile Inference Importance Distributions VRF = 42 10,000 sample Monte Carlo estimate and upper confidence limit quantile inference cut-offs Slide 26 Additional runs required to reduce IMSE in greater volume of input space

Discussion Independent importance distributions? Can negatively impact variance reduction factors Large range of eigenvalues in calibrated physics variable correlation matrix Partial least squares to identify orthogonal directions? Sparse grid quadrature for importance-sampled sensitive variables? Unknown normalizing constant for π(θ) Bayesian variational methods? Alternative design criteria to achieve faster convergence of quantile estimates? Alternative surrogate treatment(s) for faster bias reduction in important region of input space? Slide 27

Recent Work Auffray, et al. (2012), Bounding rare event probabilities in computer experiments. Goal: Obtain sharp upper bound on Approximate importance region Importance sample (x (j),θ (j) ) from Perform code calculations η(x (j),θ (j) ) and T = #{ η(x (j),θ (j) ) > q } Stochastic upper bound on π q : Slide 28

Discussion Importance sampling methods extend straightforwardly to problem of bounding a rare event probability Estimate by applying sequential importance sampling to the function with respect to quantile q, and computing Pr( ˆR π )= 1 N 1 N h η(x (j),θ (j) )>q κ dv w(x ar( η(x (j),θ ))i (j), θ (j) ) (j) j=1 Calculate T as in Auffray, et al. (2012) Slide 29

Discussion Bect, J. et al. (2011), Sequential design of computer experiments for the estimation of a probability of failure, Statistics and Computing. Considered one-step look ahead (myopic) criterion Criterion can be bounded (exact value not available) One bound is IMSE of 1 [η(x,θ) > q] Can also use g(x,θ ) Slide 30

Conclusions Importance sampling improves UQ of percentile and quantile estimates relative to brute force approach Benefits of importance sampling increase as percentiles become more extreme Iterative refinement improves importance distributions in relatively few iterations Surrogates are necessary for slow running codes Sequential design improves surrogate quality in region of parameter space indicated by importance distributions Importance distributions and VRFs stabilize quickly, while quantile estimates may converge slowly Slide 31