Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester

Similar documents
Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403 Probability Distributions II: More Properties of PDFs and PMFs

Physics 509: Error Propagation, and the Meaning of Error Bars. Scott Oser Lecture #10

Physics 403. Segev BenZvi. Numerical Methods, Maximum Likelihood, and Least Squares. Department of Physics and Astronomy University of Rochester

Statistics and Data Analysis

Statistical Methods in Particle Physics

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester

Bayesian rules of probability as principles of logic [Cox] Notation: pr(x I) is the probability (or pdf) of x being true given information I

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

 ± σ A ± t A ˆB ± σ B ± t B. (1)

Physics 509: Propagating Systematic Uncertainties. Scott Oser Lecture #12

Strong Lens Modeling (II): Statistical Methods

Bayes Decision Theory

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

arxiv:hep-ex/ v1 2 Jun 2000

18 Bivariate normal distribution I

Statistical Methods in Particle Physics

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO

arxiv: v1 [physics.data-an] 3 Jun 2008

Linear Methods for Prediction

Statistics. Lecture 2 August 7, 2000 Frank Porter Caltech. The Fundamentals; Point Estimation. Maximum Likelihood, Least Squares and All That

DS-GA 1003: Machine Learning and Computational Statistics Homework 7: Bayesian Modeling

Density Estimation. Seungjin Choi

Introduction to Data Analysis

Physics 403. Segev BenZvi. Credible Intervals, Confidence Intervals, and Limits. Department of Physics and Astronomy University of Rochester

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Consider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y.

Introduction to Statistics and Data Analysis

A Probability Review

Modern Methods of Data Analysis - WS 07/08

Gaussian Processes (10/16/13)

Parameter Estimation

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1

Bayesian Regression Linear and Logistic Regression

Nonparameteric Regression:

Data Fitting - Lecture 6

Error propagation. Alexander Khanov. October 4, PHYS6260: Experimental Methods is HEP Oklahoma State University

Measurement Error and Linear Regression of Astronomical Data. Brandon Kelly Penn State Summer School in Astrostatistics, June 2007

Rowan University Department of Electrical and Computer Engineering

Introduction to Bayesian Data Analysis

Computer Vision Group Prof. Daniel Cremers. 4. Gaussian Processes - Regression

conditional cdf, conditional pdf, total probability theorem?

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That

Statistics notes. A clear statistical framework formulates the logic of what we are doing and why. It allows us to make precise statements.

ECE531 Homework Assignment Number 6 Solution

Lecture 10: The quality of the Estimate

ECON 3150/4150, Spring term Lecture 6

The Treatment of Numerical Experimental Results

Prediction of Data with help of the Gaussian Process Method

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6

10. Composite Hypothesis Testing. ECE 830, Spring 2014

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester

Notes on Statistical Methods

PHYSTAT2003, SLAC, Stanford, California, September 8-11, 2003

Lecture 3: Pattern Classification

Introduction to Maximum Likelihood Estimation

Statistics, Data Analysis, and Simulation SS 2015

A Convolution Method for Folding Systematic Uncertainties into Likelihood Functions

Hypothesis Testing - Frequentist

GAUSSIAN PROCESS REGRESSION

Introduc)on to Bayesian Methods

Primer on statistics:

A Bayesian Treatment of Linear Gaussian Regression

Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics

University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries

Statistical and Learning Techniques in Computer Vision Lecture 2: Maximum Likelihood and Bayesian Estimation Jens Rittscher and Chuck Stewart

g-priors for Linear Regression

Fractional Hot Deck Imputation for Robust Inference Under Item Nonresponse in Survey Sampling

L2: Review of probability and statistics

Lecture : Probabilistic Machine Learning

ECON 4160, Autumn term Lecture 1

Statistical Data Analysis Stat 3: p-values, parameter estimation

CSC 2541: Bayesian Methods for Machine Learning

Introduction to Probability and Stocastic Processes - Part I

F denotes cumulative density. denotes probability density function; (.)

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm

Introduction to Normal Distribution

Bayesian Learning. HT2015: SC4 Statistical Data Mining and Machine Learning. Maximum Likelihood Principle. The Bayesian Learning Framework

Multiple Random Variables

Confidence Limits and Intervals 3: Various other topics. Roger Barlow SLUO Lectures on Statistics August 2006

How to build an automatic statistician

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Multiple Random Variables

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

1 Introduction. Systems 2: Simulating Errors. Mobile Robot Systems. System Under. Environment

Random Variables. P(x) = P[X(e)] = P(e). (1)

Computer Vision Group Prof. Daniel Cremers. 9. Gaussian Processes - Regression

Bayesian Interpretations of Regularization

Lecture 5: Moment Generating Functions

Lightweight Probabilistic Deep Networks Supplemental Material

Intro to Probability. Andrei Barbu

Bayesian analysis in nuclear physics

Solutions to Homework Set #5 (Prepared by Lele Wang) MSE = E [ (sgn(x) g(y)) 2],, where f X (x) = 1 2 2π e. e (x y)2 2 dx 2π

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Machine Learning. Gaussian Mixture Models. Zhiyao Duan & Bryan Pardo, Machine Learning: EECS 349 Fall

2: Distributions of Several Variables, Error Propagation

Transcription:

Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester

Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals from ln L Marginal and Joint Confidence Regions 2 Propagation of Uncertainties Error Propagation Formula Using the Covariance Breakdown of Error Propagation Averaging Correlated Measurements with Least Squares Asymmetric Error Bars 3 Bayesian Approach: Using the Complete PDF Transformation of Variables Breakdown of the Error Propagation Formula Segev BenZvi (UR) PHY 403 2 / 35

Maximum Likelihood and Method of Least Squares Suppose we measure data x and we want to find the posterior of the model parameters θ. If our priors on the parameters are uniform then p(θ x, I ) p(x θ, I ) p(θ I ) = p(x θ, I ) = L(x θ) In this case finding the best estimate ˆθ is equivalent to maximizing the likelihood L If {x i } are independent measurements with Gaussian errors then ( ) 1 N (f (x i ) x i ) 2 p(x θ, I ) = L(x θ) = exp (2πΣ) N/2 Least Squares: equivalent to maximizing ln L, except you minimize i=1 2σ 2 i χ 2 = N (f (x i ) x i ) 2 σ 2 i=1 i Segev BenZvi (UR) PHY 403 3 / 35

Obtaining Uncertainty Intervals from ln L and χ 2 For Gaussian uncertainties we can obtain 1σ, 2σ, and 3σ intervals using the rules Error ln L χ 2 1σ 0.5 1 2σ 2 4 3σ 4.5 9 Even without Gaussian errors this can work reasonably well. But, a safe alternative is simulation of ln L with Monte Carlo Segev BenZvi (UR) PHY 403 4 / 35

Marginal and Joint Confidence Regions The curves χ 2 = 1.00, 2.71, 6.63 project onto 1D intervals containing 68.3%, 90%, and 99% of normally distributed data Note that it s the intervals, not the ellipses themselves, that contain 68.3%. The ellipse that contains 68% of the 2D space is χ 2 = 2.30 [1] Segev BenZvi (UR) PHY 403 5 / 35

Joint Confidence Intervals If we want multi-dimensional error ellipses that contain 68.3%, 95.4%, and 99.7% of the data, we use these contours in ln L: joint parameters Range p 1 2 3 4 5 6 1σ 68.3% 0.50 1.15 1.76 2.36 2.95 3.52 2σ 95.4% 2.00 3.09 4.01 4.85 5.65 6.4 3σ 99.7% 4.50 5.90 7.10 8.15 9.10 10.05 Or these in χ 2 [1]: joint parameters Range p 1 2 3 4 5 6 1σ 68.3% 1.00 2.30 3.53 4.72 5.89 7.04 2σ 95.4% 4.00 6.17 8.02 9.70 11.3 12.8 3σ 99.7% 9.00 11.8 14.2 16.3 18.2 20.1 Segev BenZvi (UR) PHY 403 6 / 35

Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals from ln L Marginal and Joint Confidence Regions 2 Propagation of Uncertainties Error Propagation Formula Using the Covariance Breakdown of Error Propagation Averaging Correlated Measurements with Least Squares Asymmetric Error Bars 3 Bayesian Approach: Using the Complete PDF Transformation of Variables Breakdown of the Error Propagation Formula Segev BenZvi (UR) PHY 403 7 / 35

Propagation of Uncertainties We know that measurements (or fit parameters) x have uncertainties, and these uncertainties need to be propagated when you calculate functions of measured quantities f (x) From undergraduate lab courses you know the formula [2] σ 2 f N ( f i=1 x i ) 2 σ 2 x i Question: what does this formula assume about the uncertainties on x = (x 1, x 2,..., x N )? Question: what does this formula assume about the PDFs of the {x i } (if anything)? Question: what does this formula assume about f? Segev BenZvi (UR) PHY 403 8 / 35

Propagation of Uncertainties Let s start with a set of N random variables x. E.g., the {x i } could be parameters from a fit We want to calculate a function f (x), but suppose we don t know the PDFs of the {x i }, just best estimates of their means ˆx and the covariance matrix V Linearize the problem: expand f (x) to first order about the means of the x i : N f f (x) f (ˆx) + x i (x i ˆx i ) x=ˆx i=1 The name of the game: calculate the expectation and variance of f (x) to derive the error propagation formula. To first order, E [f (x)] f (ˆx) Segev BenZvi (UR) PHY 403 9 / 35

Error Propagation Formula Get the variance by calculating the expectation of f 2 : N E [f 2 (x)] f 2 f (ˆx) + 2f (ˆx) x i E (x i ˆx i ) i=1 x=ˆx ( N ) + E f N x i (x i ˆx i ) f x=ˆx x j (x j ˆx j ) x=ˆx = f 2 (ˆx) + i=1 N i,j=1 f f x i x j V ij x=ˆx j=1 Since var (f ) = σ 2 f = E (f 2 ) E (f ) 2, we find that σ 2 f N i,j=1 f f x i x j V ij x=ˆx Segev BenZvi (UR) PHY 403 10 / 35

Error Propagation Formula For a set of m functions f 1 (x),..., f m (x), we have a covariance matrix cov (f k, f l ) = U kl N f k f l x i x j V ij x=ˆx i,j=1 Writing the matrix of derivatives as A ij = f i / x j, the covariance matrix can be written U = AVA For uncorrelated x i, V is diagonal and so σ 2 f N f x i σi 2 x=ˆx i=1 This is the form you re used to from elementary courses. Segev BenZvi (UR) PHY 403 11 / 35

Propagation of Uncertainties for Two Variables Let x = (x, y). The general form of σ 2 f is σ 2 f = ( ) f 2 ( ) f 2 σx 2 + σy 2 + 2 f f x y x y ρσ xσ y The final cross term is ignored altogether in lab courses, but it s important! Since the correlation between x and y can be negative, you can overestimate the uncertainty in f by failing to include it Don t forget the assumptions underlying this expression: 1. Gaussian uncertainties with known covariance matrix 2. f is approximately linear in the range (x ± σ x, y ± σ y ) If the assumptions are violated, the error propagation formula breaks down Segev BenZvi (UR) PHY 403 12 / 35

Interpolation of Linear Fit Example Example LS fit: best estimators ˆm = 2.66 ± 0.10, ˆb = 2.05 ± 0.51, cov (m, b) = 0.10 = ρ = 0.94 y(5.5) = 16.68 ± 0.75 without using the correlation. With the correlation, y(5.5) = 16.68 ± 0.19. Segev BenZvi (UR) PHY 403 13 / 35

Breakdown of Error Propagation Example Imagine two independent variables x and y with ˆx = 10 ± 1 and ŷ = 10 ± 1. The variance in the ratio f = x 2 /y is [ ( ) x 2 ( ) ] x 4 σf 2 = 4 σx 2 + σy 2 y y For ˆx = ŷ = 10 and σ 2 x = σ 2 y = 1, σ 2 f = 4 ( ) 10 2 (1) 2 + 10 x=ˆx ( ) 10 4 (1) 2 = 5 10 But, suppose ŷ = 1. Then the uncertainty blows up σ 2 f = 4 ( ) 10 2 (1) 2 + 1 ( ) 10 4 (1) 2 = 10400 1 Segev BenZvi (UR) PHY 403 14 / 35

Breakdown of Error Propagation What happened? If ŷ = 1, then y can be very close to zero when f (x, y) is expanded about the mean, so f can blow up and become non-linear Note: be careful even when the error propagation assumptions of small uncertainties and linearity apply; the resulting distribution could still be non-gaussian. Example: x/y, with ˆx = 5 ± 1 and ŷ = 1 ± 0.5: In this case, reporting a central value and RMS for f = x/y is clearly inadequate Segev BenZvi (UR) PHY 403 15 / 35

Case Study: Polarization Asymmetry Example Early evidence supporting the Standard Model of particle physics came from observing the difference in cross sections σ R and σ L for inelastic scattering of right- and left-handed polarized electrons on a deuterium target [3] The experiment studied the polarization asymmetry defined by α = σ R σ L σ R + σ L Must be careful about using the error on α to conclude whether or not α is consistent with zero More robust approach: check whether or not σ R σ L alone is consistent with zero Segev BenZvi (UR) PHY 403 16 / 35

Averaging Correlated Measurements using Least Squares Imagine we have a set of measurements x i ± σ i of some true value λ. Since λ is the same for all measurements, we can minimize χ 2 = N (x i λ) 2 i=1 The LS estimator for λ is the weighted average σ 2 i ˆλ = yi /σ 2 i 1/σ 2 i, var (ˆλ) = For correlated measurements, we can write 1 1/σ 2 i ˆλ = χ 2 = N w i x i, w i = i=1 N (x i λ)(v 1 ) ij (x j λ) i,j=1 N j=1 (V 1 ) ij N k,l=1 (V 1 ) kl, var (ˆλ) = N w i V ij w j i,j=1 Segev BenZvi (UR) PHY 403 17 / 35

Example: Averaging Correlated Measurements Example We measure a length with two rulers made of different materials (and different coefficients of thermal expansion). Both are calibrated to be accurate at T = T 0 but otherwise have a temperature dependence y i = L i + c i (T T 0 ) We know the c i and the uncertainties, T, and L 1 and L 2 from the calibration. We want to combine measurements and get ŷ. The variances and covariance are var (y i ) = σ 2 i = σ 2 L i + c 2 i σ 2 T cov (y 1, y 2 ) = E (y 1 y 2 ) ŷ 2 = c 1 c 2 σ 2 T Solve for ŷ with the weighted mean derived using least squares Segev BenZvi (UR) PHY 403 18 / 35

Example: Averaging Correlated Measurements Example Plug in the following values: T 0 = 25, T = 23 ± 2, and Ruler c i L i y i 1 0.1 2.0 ± 0.1 1.80 ± 0.22 2 0.2 2.3 ± 0.1 1.90 ± 0.41 Solving, we find the weighted average is ŷ = 1.75 ± 0.19 So the effect of the correlation is that the weighted average is less than either of the two individual measurements. Does that make sense? Segev BenZvi (UR) PHY 403 19 / 35

Averaging Correlated Measurements Horizontal bands: lengths L i from two rulers Slanted: lengths y i corrected for T If L 1 and L 2 are known accurately, but y 1 and y 2 differ, then the true temperature must be different than the measured value of T The χ 2 favors reducing ŷ until y 1 (T ) and y 2 (T ) intersect If the correction T σ T, some assumption is probably wrong. This would be reflected as a large value of χ 2 and a small p-value Segev BenZvi (UR) PHY 403 20 / 35

Asymmetric Uncertainties You will often encounter published data with asymmetric error bars σ + and σ, e.g., if the author found an error interval with the maximimum likelihood method What do you do if you have no further information about the form of the likelihood, which is almost never published? Suggestion due to Barlow [4, 5]: parameterize the likelihood as ln L = 1 (ˆx x) 2 2 σ(x) 2 where σ(x) = σ + σ (x ˆx). Requiring it to go through the 1/2 points gives ln L = 1 ( ) (ˆx x)(σ + + σ ) 2 2σ + σ + (σ + σ )(x ˆx) When σ + = σ this reduces to an expression that gives the usual ln L = 1/2 rule Segev BenZvi (UR) PHY 403 21 / 35

Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals from ln L Marginal and Joint Confidence Regions 2 Propagation of Uncertainties Error Propagation Formula Using the Covariance Breakdown of Error Propagation Averaging Correlated Measurements with Least Squares Asymmetric Error Bars 3 Bayesian Approach: Using the Complete PDF Transformation of Variables Breakdown of the Error Propagation Formula Segev BenZvi (UR) PHY 403 22 / 35

Full Bayesian Approach Transformation of Variables In the Bayesian universe, you would ideally know the complete PDF and use that to propagate uncertainties In this case, if we have some p(x I ) and we define y = f (x), then we need to map p(x I ) to p(y I ) Consider a small interval δx around x such that p(x + δx/2 x < δx/2 I ) p(x = x I ) δx y = f (x) maps x to y = f (x ) and δx to δy. The range of y values in y ± δy/2 is equivalent to a variation in x between x ± δx/2, and so p(x = x I ) δx = p(y = y I ) δy In the limit δx 0, this yields the PDF transformation rule p(x I ) = p(y I ) dy dx Segev BenZvi (UR) PHY 403 23 / 35

Transformation of Variables More than One Variable For more than one variable, p({x i } I ) δx 1... δx m = p({y i } I ) δ m vol({y i }) where δ m vol({y i }) is an m-dimensional volume in y mapped out by the hypercube δx 1... δx m The m-dimensional equivalent of the 1D transformation rule is p({x i } I ) = p({y i } I ) (y 1,..., y m ) (x 1,..., x m ) where the rightmost expression is the Jacobian matrix of partial derivatives dy i /dx j Segev BenZvi (UR) PHY 403 24 / 35

Polar Coordinates Example For x = R cos θ and y = R sin θ, (x, y) (R, θ) = cos θ R sin θ sin θ R cos θ = R(cos2 θ + sin 2 θ) = R Therefore, p(r, θ I ) is related to p(x, y I ) by p(r, θ I ) = p(x, y I ) R You saw this earlier in the semester with the Rayleigh distribution: p(x, y I ) = 1 { 2πσ 2 exp x 2 + y 2 } 2σ 2 = p(r, θ I ) = R } { 2πσ 2 exp R2 2σ 2 We have just equated the volume elements dx dy = R dr dθ. Segev BenZvi (UR) PHY 403 25 / 35

Application to Simple Problems If we want to estimate a sum like z = x + y or a ratio z = x/y, we integrate the joint PDF p(x, y I ) along the shaded strips defined by δ(z f (x, y)): The explicit marginalization is p(z I ) = dx dy p(z x, y, I ) p(x, y I ) = dx dy δ(z f (x, y)) p(x, y I ) Segev BenZvi (UR) PHY 403 26 / 35

Sum of Two Random Variables The sum z = x + y requires that we marginalize p(z I ) = dx dy δ(z (x + y)) p(x, y I ) If we are given x = ˆx ± σ x and y = ŷ ± σ y, then we can assume x and y are independent and factor the joint PDF into separate PDFs by the product rule: p(z I ) = dx p(x I ) dy p(y I ) δ(z x y) = dx p(x I ) p(y = z x I ) Assuming Gaussian PDFs for x and y, } 1 (x ˆx)2 p(z I ) = dx exp { 2πσ x σ y 2σx 2 exp { } (z x ŷ)2 2σy 2 Segev BenZvi (UR) PHY 403 27 / 35

Sum of Two Random Variables After some rearranging of terms and changes of variables, we can express } } 1 (x ˆx)2 (z x ŷ)2 p(z I ) = dx exp { 2πσ x σ y 2σx 2 exp { 2σy 2 as where p(z I ) = } 1 (z ẑ)2 exp { 2πσz 2σz 2 ẑ = ˆx + ŷ and σ 2 z = σ 2 x + σ 2 y Hence, we see how the quadrature sum rule for adding uncertainties derives directly from the assumption of Gaussian errors. Note that for a difference z = x y, the uncertainties still add in quadrature but ẑ = ˆx ŷ, as you d expect Segev BenZvi (UR) PHY 403 28 / 35

Case Study: Amplitude of a Bragg Peak in Crystallography Isn t this serious overkill given that we have the error propagation formula? Unfortunately, recall that the formula can break down Example In crystallography, one measures a Bragg peak A =  ± σ A The peak is related to the structure factor A = F 2 We want to estimate f = F = A. From the propagation formula, σ A f =  ± 2  Problem: suppose  < 0, which is an allowed measurement due to reflections Now we re in trouble, because the error propagation formula requires us to take the square root of a negative number Segev BenZvi (UR) PHY 403 29 / 35

Solution with Full PDF Let s write down the full posterior PDF p(a {data}, I ) p({data} A, I ) p(a I ) By applying the error propagation formula, we assumed A is distributed like a Gaussian, so { } (A Â)2 p({data} A, I ) exp 2σA 2 Since A < 0 is a problem, let s define the prior to force A into a physical region: { constant A 0 p(a I ) = 0 otherwise When  < 0, the prior will truncate the Gaussian likelihood Segev BenZvi (UR) PHY 403 30 / 35

Solution with Full PDF Truncating the PDF violates the error propagation formula, because it depends on a Taylor expansion about a central maximum There is no such restriction on the formal change of variables to f : p(f {data}, I ) = p(a {data}, I ) da df The Jacobian is da/df = 2f, with f = F 0, so { } (A Â)2 p(f {data}, I ) f exp 2σA 2 for f 0 Find ˆf by maximizing ln p, and σ 2 f from σ 2 f = ( 2 ln p/ f 2 ) 1 : [ ] 1 2ˆf 2 = Â + Â 2 + 2σA 2, 1 σ2 f = ˆf + 2(3ˆf 2 Â) 2 σa 2 Segev BenZvi (UR) PHY 403 31 / 35

Asymptotic Agreement of PDF and Error Propagation When  > 0 and  σ A, the expressions for f and σf 2 2ˆf 2 =  +  2 + 2σA 2 ˆf =  [ σ 2 f = 1 ˆf + 2(3ˆf 2 Â) 2 σa 2 ] 1 σ2 A 4 For example, if A = 9 ± 1, the posterior PDFs of A and f look very similar to the Gaussian PDF implied by the error propagation formula: are Segev BenZvi (UR) PHY 403 32 / 35

Asymptotic Agreement of PDF and Error Propagation If A = 1 ± 9, the error propagation formula (dashed) begins to blow up compared to the full PDF: If A = 20 ± 9, the error propagation formula can t even be applied. The posterior PDF looks like a Rayleigh distribution: Segev BenZvi (UR) PHY 403 33 / 35

Summary The standard error propagation formula applies when uncertainties are Gaussian and f (x) can be approximated by a first-order Taylor expansion (linearized) Most undergraduate courses emphasize only uncorrelated uncertainties, but you need to account for correlations Often authors will report asymmetric error bars, implying non-gaussian uncertainties, without giving the form of the PDF. In this case there are some approximations to the likelihood that you can try to use Standard error propagation breaks down when the errors are asymmetric or f (x) can t be linearized The general case is to use the full PDF to construct a new uncertainty interval on your best estimator. It s a pain (and often overkill) but it is always correct and can help you when standard error propagation fails Segev BenZvi (UR) PHY 403 34 / 35

References I [1] W. Press et al. Numerical Recipes in C. New York: Cambridge University Press, 1992. URL: http://www.nr.com. [2] John R. Taylor. An Introduction to Error Analysis. Sausalito, CA: University Science Books, 1997. [3] C.Y. Prescott et al. Parity Nonconservation in Inelastic Electron Scattering. In: Phys.Lett. B77 (1978), pp. 347 352. [4] Roger Barlow. Asymmetric Errors. In: PHYSTAT 2003. SLAC, 2003, pp. 250 255. [5] Roger Barlow. Asymmetric Statistical Errors. In: (2004), pp. 56 59. arxiv: physics/0406120 [physics]. Segev BenZvi (UR) PHY 403 35 / 35