Engineering Risk Benefit Analysis

Similar documents
Probability and Random Variable Primer

Stochastic Structural Dynamics

Expected Value and Variance

Chapter 1. Probability

Lecture 3: Probability Distributions

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

CS-433: Simulation and Modeling Modeling and Probability Review

PhysicsAndMathsTutor.com

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

j) = 1 (note sigma notation) ii. Continuous random variable (e.g. Normal distribution) 1. density function: f ( x) 0 and f ( x) dx = 1

A be a probability space. A random vector

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

Module 2. Random Processes. Version 2 ECE IIT, Kharagpur

Limited Dependent Variables

Statistics and Quantitative Analysis U4320. Segment 3: Probability Prof. Sharyn O Halloran

xp(x µ) = 0 p(x = 0 µ) + 1 p(x = 1 µ) = µ

Statistical analysis using matlab. HY 439 Presented by: George Fortetsanakis

CS 798: Homework Assignment 2 (Probability)

: 5: ) A

6. Stochastic processes (2)

6. Stochastic processes (2)

SELECTED PROOFS. DeMorgan s formulas: The first one is clear from Venn diagram, or the following truth table:

Machine learning: Density estimation

1.4. Experiments, Outcome, Sample Space, Events, and Random Variables

Introduction to Random Variables

Artificial Intelligence Bayesian Networks

ENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition

A random variable is a function which associates a real number to each element of the sample space

Probability Theory (revisited)

Department of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

Continuous Time Markov Chain

Appendix B. Criterion of Riemann-Stieltjes Integrability

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

Marginal Effects in Probit Models: Interpretation and Testing. 1. Interpreting Probit Coefficients

Learning from Data 1 Naive Bayes

First Year Examination Department of Statistics, University of Florida

Chapter 1: Introduction to Probability

Analysis of Discrete Time Queues (Section 4.6)

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

More metrics on cartesian products

Decision-making and rationality

9.913 Pattern Recognition for Vision. Class IV Part I Bayesian Decision Theory Yuri Ivanov

Conjugacy and the Exponential Family

EGR 544 Communication Theory

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

7. Multivariate Probability

Notes prepared by Prof Mrs) M.J. Gholba Class M.Sc Part(I) Information Technology

Predictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore

Maximum Likelihood Estimation (MLE)

Goodness of fit and Wilks theorem

The Geometry of Logit and Probit

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Convergence of random processes

Rules of Probability

Lecture 21: Numerical methods for pricing American type derivatives

10.34 Fall 2015 Metropolis Monte Carlo Algorithm

Math 426: Probability MWF 1pm, Gasson 310 Homework 4 Selected Solutions

The Feynman path integral

Statistics and Probability Theory in Civil, Surveying and Environmental Engineering

Simulation and Random Number Generation

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Lecture 12: Discrete Laplacian

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup

Bayesian Networks. Course: CS40022 Instructor: Dr. Pallab Dasgupta

CS286r Assign One. Answer Key

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

First day August 1, Problems and Solutions

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Lecture Notes on Linear Regression

Homework 9 for BST 631: Statistical Theory I Problems, 11/02/2006

Linear Regression Analysis: Terminology and Notation

Distributions /06. G.Serazzi 05/06 Dimensionamento degli Impianti Informatici distrib - 1

7. Multivariate Probability

Numerical Heat and Mass Transfer

APPENDIX A Some Linear Algebra

Statistical Inference. 2.3 Summary Statistics Measures of Center and Spread. parameters ( population characteristics )

Digital Signal Processing

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.265/15.070J Fall 2013 Lecture 12 10/21/2013. Martingale Concentration Inequalities and Applications

Gaussian Mixture Models

The Multiple Classical Linear Regression Model (CLRM): Specification and Assumptions. 1. Introduction

Computing MLE Bias Empirically

U-Pb Geochronology Practical: Background

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

Multiple Choice. Choose the one that best completes the statement or answers the question.

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

} Often, when learning, we deal with uncertainty:

PROBABILITY PRIMER. Exercise Solutions

ECE 534: Elements of Information Theory. Solutions to Midterm Exam (Spring 2006)

Notes on Frequency Estimation in Data Streams

Basically, if you have a dummy dependent variable you will be estimating a probability.

Probability review. Adopted from notes of Andrew W. Moore and Eric Xing from CMU. Copyright Andrew W. Moore Slide 1

Strong Markov property: Same assertion holds for stopping times τ.

Randomness and Computation

Transcription:

Engneerng Rsk Beneft Analyss.55, 2.943, 3.577, 6.938, 0.86, 3.62, 6.862, 22.82, ESD.72, ESD.72 RPRA 2. Elements of Probablty Theory George E. Apostolaks Massachusetts Insttute of Technology Sprng 2007 RPRA 2. Elements of Probablty Theory

Probablty: Axomatc Formulaton The probablty of an event A s a number that satsfes the followng axoms (Kolmogorov): 0 P(A) P(certan event) For two mutually exclusve events A and B: P(A or B) P(A) + P(B) RPRA 2. Elements of Probablty Theory 2

Relatve-frequency nterpretaton Imagne a large number n of repettons of the experment of whch A s a possble outcome. If A occurs k tmes, then ts relatve frequency s: It s postulated that: lm n k n P(A) k n RPRA 2. Elements of Probablty Theory 3

Degree-of-belef (Bayesan) nterpretaton o need for dentcal trals. The concept of lkelhood s prmtve,.e., t s meanngful to compare the lkelhood of two events. P(A) < P(B) smply means that the assessor judges B to be more lkely than A. Subjectve probabltes must be coherent,.e., must satsfy the mathematcal theory of probablty and must be consstent wth the assessor s knowledge and belefs. RPRA 2. Elements of Probablty Theory 4

Basc rules of probablty: egaton S Venn Dagram E E E E S P(E) + P(E) P(S) P(E) P(E) RPRA 2. Elements of Probablty Theory 5

RPRA 2. Elements of Probablty Theory 6 Basc rules of probablty: Unon ( ) ( ) ( ) + + + + I U K j j A P A A P A P A P ( ) A P P UA Rare-Event Approxmaton:

Unon (cont d) For two events: P(A B) P(A) + P(B) P(AB) For mutually exclusve events: P(A B) P(A) + P(B) RPRA 2. Elements of Probablty Theory 7

Example: Far De Sample Space: {, 2, 3, 4, 5, 6} (dscrete) Far : The outcomes are equally lkely (/6). P(even) P(2 4 6) ½ (mutually exclusve) RPRA 2. Elements of Probablty Theory 8

RPRA 2. Elements of Probablty Theory 9 Unon of mnmal cut sets + + + + j j T M M M M X ) (... Rare-event approxmaton: ( ) ( ) T M P X P From RPRA, slde 5 + + + + j j T M P... P(M M ) P(M ) ) (X P ) (

Upper and lower bounds + T ) P(M ) P(MM j) +... + ( P(M ) j + P (X ) P ( X ) T P ( M ) The frst term,.e., sum, gves an upper bound. P(X T ) P(M ) P(MM j) j + The frst two terms, gve a lower bound. RPRA 2. Elements of Probablty Theory 0

Condtonal probablty P ( A B) P P ( AB) ( B) ( AB) P( A B) P( B) P( B / A) P( A) P For ndependent events: P ( AB) P( A) P( B) ( A) P( B) P B/ Learnng that A s true has no mpact on our probablty of B. RPRA 2. Elements of Probablty Theory

Example: 2-out-of-4 System 2 M X X 2 X 3 M 2 X 2 X 3 X 4 M 3 X 3 X 4 X M 4 X X 2 X 4 3 4 X T ( M ) ( M 2 ) ( M 3 ) ( M 4 ) X T (X X 2 X 3 + X 2 X 3 X 4 + X 3 X 4 X + X X 2 X 4 ) - 3X X 2 X 3 X 4 RPRA 2. Elements of Probablty Theory 2

2-out-of-4 System (cont d) P(X T ) P(X X 2 X 3 + X 2 X 3 X 4 + X 3 X 4 X + X X 2 X 4 ) 3P(X X 2 X 3 X 4 ) Assume that the components are ndependent and nomnally dentcal wth falure probablty q. Then, P(X T ) 4q 3 3q 4 Rare-event approxmaton: P(X T ) 4q 3 RPRA 2. Elements of Probablty Theory 3

Updatng probabltes () The events, H,..., are mutually exclusve and exhaustve,.e., H H j Ø, for j, H S, the sample space. Ther probabltes are P(H ). Gven an event E, we can always wrte H H2 E H 3 P(E) P(E/ H )P(H ) RPRA 2. Elements of Probablty Theory 4

Updatng probabltes (2) Evdence E becomes avalable. What are the new (updated) probabltes P(H /E)? Start wth the defnton of condtonal probabltes, slde. P(EH ) P(E / H )P(H ) P(H / E)P(E) P(H / E) P(E / H )P(H P(E) ) Usng the expresson on slde 4 for P(E), we get RPRA 2. Elements of Probablty Theory 5

Bayes Theorem Lkelhood of the Evdence P Posteror Probablty ( H E) P ( E H ) P( H ) P ( E H ) P( H ) Pror Probablty RPRA 2. Elements of Probablty Theory 6

Example: Let s Make A Deal Suppose that you are on a TV game show and the host has offered you what's behnd any one of three doors. You are told that behnd one of the doors s a Ferrar, but behnd each of the other two doors s a Yugo. You select door A. At ths tme, the host opens up door B and reveals a Yugo. He offers you a deal. You can keep door A or you can trade t for door C. What do you do? RPRA 2. Elements of Probablty Theory 7

Let s Make A Deal: Soluton () Settng up the problem n mathematcal terms: A {The Ferrar s behnd Door A} B {The Ferrar s behnd Door B} C {The Ferrar s behnd Door C} The events A, B, C are mutually exclusve and exhaustve. P(A) P(B) P(C) /3 E {The host opens door B and a Yugo s behnd t} What s P(A/E)? Bayes Theorem RPRA 2. Elements of Probablty Theory 8

P Let s Make A Deal: Soluton (2) ( A E ) P( E A ) P( A ) P( E A ) P( A ) + P( E B ) P( B ) + P( E C ) P( C ) But P(E/B) 0 (A Yugo s behnd door B). P(E/C) (The host must open door B, f the Ferrar s behnd door C; he cannot open door A under any crcumstances). RPRA 2. Elements of Probablty Theory 9

Let s Make A Deal: Soluton (3) Let P(A/E) x and P(E/A) p Bayes' theorem gves: x p + p Therefore For P(E/A) p /2 (the host opens door B randomly, f the Ferrar s behnd door A) P(A/E) x /3 P(A) (the evdence has had no mpact) RPRA 2. Elements of Probablty Theory 20

Let s Make A Deal: Soluton (4) Snce P(A/E) + P(C/E) P(C/E) - P(A/E) 2/3 The player should swtch to door C For P(E/A) p (the host always opens door B, f the Ferrar s behnd door A) P(A/E) /2 P(C/E) /2, swtchng to door C does not offer any advantage. RPRA 2. Elements of Probablty Theory 2

Random Varables Sample Space: The set of all possble outcomes of an experment. Random Varable: A functon that maps sample ponts onto the real lne. Example: For a de S {,2,3,4,5,6} For the con: S {H, T} {0, } RPRA 2. Elements of Probablty Theory 22

Events 3.6-0 2 3 4 5 6 We say that {X x} s an event, where x s any number on the real lne. For example (de experment): {X 3.6} {, 2, 3} { or 2 or 3} {X 96} S (the certan event) {X -62} (the mpossble event) RPRA 2. Elements of Probablty Theory 23

Sample Spaces The SS for the de s an example of a dscrete sample space and X s a dscrete random varable (DRV). A SS s dscrete f t has a fnte or countably nfnte number of sample ponts. A SS s contnuous f t has an nfnte (and uncountable) number of sample ponts. The correspondng RV s a contnuous random varable (CRV). Example: {T t} {falure occurs before t} RPRA 2. Elements of Probablty Theory 24

Cumulatve Dstrbuton Functon (CDF) The cumulatve dstrbuton functon (CDF) s F(x) Pr[X x] Ths s true for both DRV and CRV. Propertes:. F(x) s a non-decreasng functon of x. 2. F(- ) 0 3. F( ) RPRA 2. Elements of Probablty Theory 25

CDF for the De Experment F(x) / 6 2 3 4 5 6 x RPRA 2. Elements of Probablty Theory 26

Probablty Mass Functon (pmf) For DRV: probablty mass functon P( X x ) p F P(S) ( x) p for all x x, p normalzaton Example: For the de, p /6 and F(2.3) P( 2) p 2 + 6 6 3 RPRA 2. Elements of Probablty Theory 27 6 p

Probablty Densty Functon (pdf) f(x)dx P{x < X < x+dx} ( x) f P(S) df ( x) dx F( ) f(s)ds F(x) x RPRA 2. Elements of Probablty Theory 28 f(s)ds normalzaton

f Example of a pdf () Determne k so that 2 ( x) kx, for 0 x ( x) 0, f otherwse s a pdf. Answer: The normalzaton condton gves: kx dx k 0 2 3 RPRA 2. Elements of Probablty Theory 29

Example of a pdf (2) F(x) F ( x) 3 x 0.67 0.42 F(0.875) F(0.75) 0.875 0.75 3x 2 dx 0.67-0.42 0.25 f(x) 3 P{0.75 < X < 0.875} 0.75 0.875 RPRA 2. Elements of Probablty Theory 30

Moments Expected (or mean, or average) value [ ] E X m xf j x ( ) x j p dx Varance (standard devaton σ ) E [ ] ( ) 2 X m σ 2 j CRV DRV ( ) ( ) j x ( x m) j m 2 f 2 x p dx RPRA 2. Elements of Probablty Theory 3 j CRV DRV

Percentles Medan: The value x m for whch F(x m ) 0.50 For CRV we defne the 00γ percentle as that value of x for whch x γ f x ( ) dx γ RPRA 2. Elements of Probablty Theory 32

Example m 3x dx 0. 75 σ F 2 0 3 ( x 0 75) 3. x dx 0. 0375 σ 0. 94 0 2 3 ( x ) x 0. 5 x 0 79 m m m. 3 x 05 0. 05 0. 05 x0. 0. 37 2 3 0. 0. x 95 0. 95 x0. 95 98 RPRA 2. Elements of Probablty Theory 33