Uncertainty and Bias UIUC, 403 Advanced Physics Laboratory, Fall 2014

Similar documents
Error Analysis. V. Lorenz L. Yang, M. Grosse Perdekamp, D. Hertzog, R. Clegg PHYS403 Spring 2016

Modern Methods of Data Analysis - WS 07/08

Normal Distributions Rejection of Data + RLC Circuits. Lecture 4 Physics 2CL Summer 2011

Data and Error analysis

Lab 1: Measurement, Uncertainty, and Uncertainty Propagation

PHY 123 Lab 1 - Error and Uncertainty and the Simple Pendulum

Laboratory 3: Acceleration due to gravity

EM Waves in Media. What happens when an EM wave travels through our model material?

Uncertainty, Error, and Precision in Quantitative Measurements an Introduction 4.4 cm Experimental error

PHYSICS 2150 EXPERIMENTAL MODERN PHYSICS. Lecture 3 Rejection of Data; Weighted Averages

PHY 101L - Experiments in Mechanics

Introduction to Statistics and Data Analysis

Achilles: Now I know how powerful computers are going to become!

θ Beam Pivot F r Figure 1. Figure 2. STATICS (Force Vectors, Tension & Torque) MBL-32 (Ver. 3/20/2006) Name: Lab Partner: Lab Partner:

EXPERIMENTAL UNCERTAINTY

Hypothesis testing I. - In particular, we are talking about statistical hypotheses. [get everyone s finger length!] n =

Speed of waves. Apparatus: Long spring, meter stick, spring scale, stopwatch (or cell phone stopwatch)

Errors: What they are, and how to deal with them

Measurements and Data Analysis

Measurements and Data Analysis An Introduction

Dealing with the assumption of independence between samples - introducing the paired design.

Meas ure ment: Uncertainty and Error in Lab Measurements

Advanced Statistical Methods. Lecture 6

The Treatment of Numerical Experimental Results

Measurement and Measurement Errors

Module 1: Introduction to Experimental Techniques Lecture 6: Uncertainty analysis. The Lecture Contains: Uncertainity Analysis

Physics 509: Error Propagation, and the Meaning of Error Bars. Scott Oser Lecture #10

Junior Laboratory. PHYC 307L, Spring Webpage:

BRIDGE CIRCUITS EXPERIMENT 5: DC AND AC BRIDGE CIRCUITS 10/2/13

PHYSICS 2150 LABORATORY LECTURE 1

Significant Figures and an Introduction to the Normal Distribution

Introduction to Uncertainty and Treatment of Data

Physics 509: Bootstrap and Robust Parameter Estimation

Intermediate Lab PHYS 3870

37-6 Watching the electrons (matter waves)

MEASUREMENT UNCERTAINTIES

Part 1: You are given the following system of two equations: x + 2y = 16 3x 4y = 2

MEASUREMENTS AND ERRORS (OR EXPERIMENTAL UNCERTAINTIES)

Data analysis and Geostatistics - lecture VI

Lab 1: Measurement and Uncertainty

Course Project. Physics I with Lab

Lab 12: Periodic Motion

Introduction to the General Physics Laboratories

Descriptive Statistics (And a little bit on rounding and significant digits)

Measurement and Uncertainty

Take the measurement of a person's height as an example. Assuming that her height has been determined to be 5' 8", how accurate is our result?

Experiment 4 Free Fall

Experiment 2. F r e e F a l l

MITOCW watch?v=7q32wnm4dew

Lecture 2: Reporting, Using, and Calculating Uncertainties 2. v = 6050 ± 30 m/s. v = 6047 ± 3 m/s

Systematic uncertainties in statistical data analysis for particle physics. DESY Seminar Hamburg, 31 March, 2009

Experiment 2. Reaction Time. Make a series of measurements of your reaction time. Use statistics to analyze your reaction time.

Error Analysis in Experimental Physical Science Mini-Version

Measurement: The Basics

1 SIMPLE PENDULUM 1 L (1.1)

ents & Uncertainties Significant Figures 1.005, Round best to the experimental want to meters and use 300 m 2. significant figures because of

PHYSICS 15a, Fall 2006 SPEED OF SOUND LAB Due: Tuesday, November 14

MITOCW ocw f99-lec09_300k

This is particularly true if you see long tails in your data. What are you testing? That the two distributions are the same!

Statistical Methods in Particle Physics

Department of Electrical- and Information Technology. ETS061 Lecture 3, Verification, Validation and Input

Chapter 1 Review of Equations and Inequalities

Searches at the LHC and the discovery of the Higgs Boson

Data and Error Analysis

Math Lab 8: Electric Fields Integrating Continuous Charge Distributions II Due noon Thu. Feb. 1 in class

EXPERIMENT 2 Reaction Time Objectives Theory

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg

SPH3U1 Lesson 03 Introduction. 6.1 Expressing Error in Measurement

How Measurement Error Affects the Four Ways We Use Data

Lecture 12: Quality Control I: Control of Location

1 SIMPLE PENDULUM 1 L (1.1)

MITOCW R11. Double Pendulum System

The Inductive Proof Template

Pre-Lab: Primer on Experimental Errors

Talk Science Professional Development

6.041SC Probabilistic Systems Analysis and Applied Probability, Fall 2013 Transcript Tutorial:A Random Number of Coin Flips

Statistical Methods for Astronomy

Statistical Methods in Particle Physics

PARADISE VALLEY COMMUNITY COLLEGE PHYSICS COLLEGE PHYSICS I LABORATORY

What is the correct framework for Quantum Field Theories?

Sampling. Where we re heading: Last time. What is the sample? Next week: Lecture Monday. **Lab Tuesday leaving at 11:00 instead of 1:00** Tomorrow:

EXPERIMENT: REACTION TIME

MITOCW MITRES18_005S10_DiffEqnsGrowth_300k_512kb-mp4

UNIT 5 SESSION 3: FORCE, MASS AND ACCELERATION

Class 7 Preclass Quiz on MasteringPhysics

ERROR ANALYSIS ACTIVITY 1: STATISTICAL MEASUREMENT UNCERTAINTY AND ERROR BARS

Statistics and Data Analysis

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1

Electromagnetism lab project

Sequential Logic (3.1 and is a long difficult section you really should read!)

Precision Correcting for Random Error

PHYS 228 Template Example

LABORATORY VII MECHANICAL OSCILLATIONS

Error propagation. Alexander Khanov. October 4, PHYS6260: Experimental Methods is HEP Oklahoma State University

6.867 Machine Learning

Correlation. We don't consider one variable independent and the other dependent. Does x go up as y goes up? Does x go down as y goes up?

1 Introduction to Generalized Least Squares

Numbers and Uncertainty

that relative errors are dimensionless. When reporting relative errors it is usual to multiply the fractional error by 100 and report it as a percenta

Transcription:

Uncertainty and Bias UIUC, 403 Advanced Physics Laboratory, Fall 2014 Liang Yang* There are three kinds of lies: lies, damned lies and statistics. Benjamin Disraeli If your experiment needs statistics, you ought to have done a better experiment. Lord Ernest Rutherford I can live with doubt and uncertainty and not knowing. I think it is much more interesting to live not knowing than to have answers that might be wrong. - Richard Feynman * Talk based on lectures by Matthias Perdekamp and David Herzog 1

How to report your measurement results? Always include uncertainty estimates in your results. Have the correct number of significant digits. Examples: Origin fit result, V = 0.122349 m/s, σ s = 0.01298 m/s You should report V = 0.122 ± 0.013 m/s Measurement result, T = 3.745 x 10-3 s, σ T = 0.0798 ms You should report T = (3.75 ± 0.08)x10-3 s 2

Significant Digits in Design unit: mm x.x ± 0.4 x.xx ± 0.20 x.xxx ± 0.100 Keeping correct number of significant digits and specify precision is critical in designs. 3

Uncertainty Propagation Example formula from Wikipedia In most cases for this class, variables are uncorrelated, therefore the correlation term can be safely ignored. Before using the formula, you should check if your assumption about variable correlation is warranted. 4

Systematic vs. Statistical Uncertainties Systematic uncertainty Uncertainties associated with imperfect knowledge of measurement apparatus, other physical quantities needed for the measurement, or the physical model used to interpret the data. Generally correlated between measurements. Cannot be reduced by multiple measurements. Better calibration, or measurement of other variable can reduce the uncertainty. 5

Systematic vs. Statistical Uncertainties Systematic uncertainty Uncertainties associated with imperfect knowledge of measurement apparatus, other physical quantities needed for the measurement, or the physical model used to interpret the data. Generally correlated between measurements. Cannot be reduced by multiple measurements. Better calibration, or measurement of other variable can reduce the uncertainty. Statistical Uncertainty Uncertainties due to stochastic fluctuations Generally there is no correlation between successive measurements. Multiple measurements can be used to reduce to estimate the uncertainty. 5

Gaussian Distribution Most statistic uncertainties have the characteristic that, as the sample size becomes large, the shape of the histogram tends to that of the normal distribution. One can fit the data to a Gaussian, or calculate the mean and variance. The assumption is not always true! 6

Bias, Data Selection? What if an experiment doesn't give the result you expected? What if it gives a result that you just know is wrong in some way? Don't you keep trying until you get the "right" result? Note: relevant here in our modern physics lab course where the right result was, in general, published a long time ago 7

How common is data rejection? Answer: Common Realities of complex experiments Stuff goes wrong Equipment malfunctions People make mistakes Burden on the physicist Record everything Responsibility of physicist Develop a result-unbiased algorithm for data rejection Make decisions before you look at the results Keep answer in a blind or unbiased space You can rarely use the result to determine inclusion 8

Rejection of Data from J. Taylor, Ch. 6 of An Introduction to Error Analysis Consider 6 measurements of a pendulum period : 3.8, 3.5, 3.9, 3.9, 3.4, 1.8 Should the last measurement be rejected? Yes: If some aspect of the experiment was changed... new slow stopwatch, etc. No: Never! You must always keep all data!! (diehards; beware) Maybe? The usual case. You don t know why, but something may have made this measurement bad. How do you set judge you in set an up a judgement unbiased manner that is? unbiased? An elementary test to indicate if you might have a problem with your data First, compute some simple x statistics: = 3.8 Mean of measurements: 3.4 1 2 σx = xi x = 0. Standard deviation: ( ) 8 Is the 1.8 measurement anomalous? It differs by 2 σ 2.2 (1.6) σ from the mean. 10 N 0.73

Rejection of Data from J. Taylor, Ch. 6 of An Introduction to Error Analysis Consider 6 measurements of a pendulum period : 3.8, 3.5, 3.9, 3.9, 3.4, 1.8 Should the last measurement be rejected? Yes: If some aspect of the experiment was changed... new slow stopwatch, etc. No: Never! You must always keep all data!! (diehards; beware) Maybe? The usual case. You don t know why, but something may have made this measurement bad. How do you set judge you in set an up a judgement unbiased manner that is? unbiased? An elementary test to indicate if you might have a problem with your data x = 3.8 First, compute some simple statistics: Mean of measurements: x = 3.4 1 2 Standard deviation: σx = ( xi x) = 0. 8 Is the 1.8 measurement anomalous? It differs by 2 σ 2.2 (1.6) σ from the mean. N 0.73

Chauvenet s Criterion The probability (assuming a Gaussian distribution) is 0.05 for this to be an acceptable measurement. What s wrong with that? We would even expect that 1 out of 20 measurements would fall outside of the 2 σ bound. But, we only made 6 measurements. So, we expect that only 1/(20/6) = 0.3 measurements should fall outside the 2s bound. Now, it is a bit of personal taste. Is this unreasonable? Chauvenet s criterion is the following: If the suspect measurement has a lower probability than 1/2, you should reject it. Our measurement has 0.27 so it goes. New results: Mean = 3.7 Standard deviation = 0.2! much smaller! The probability that 1 or more measurements are outside 2 sigma for 6 measurements is 1-0.95 6 =0.27 10

Our case study: A very simple first step Data set (picked off the graph by hand) 11.5, 5.5, 4.0, 8.0, 7.6, 1.5, 10.2, 0.5 (note, at same beam intensity!) Mean: = 6.1 Standard deviation = 4.0 List of deviations in sigma: 1.35, -0.15, -0.53, 0.48, 0.38, -1.15, 1.03, -1.40 (these are the bad guys) Data Points prob in 1 prob in 8 (8,11.5) 0.09 0.53 (44,1.5) 0.07 0.44 (65.0.5) 0.15 0.72

Let s look at our data What are the uncertainties? Can we attach the power fluctuations to particular data points? Why should we trust the theory prediciton? It could be simply wrong 12

Let s look at our data Assume we find the errors to be +/- 2.5 independent of beam intensity Are the data compatible with a constant behavior. Not sure chi2/ndf is 2.5 13

Let s look at our data Assume we find the errors to be +/- 2.5 independent of beam intensity Are the data compatible with a polynomial. Not sure chi2/ndf is 2.4 In absence of slow control data for beam and experimental apparatus data cannot be rejected! 14

Repeat: Is all data good data? NO! Write down everything in the logbook; take your time; use sentences; record numbers (values); glitch in the power? note the time temperature cold or hot? comment about it somebody reset the system? note it please and when Record (electronically if possible) everything reasonable as parallel information to the main data set temperatures; voltages; generally called slow controls You WILL (almost certainly) have to go back and hunt for this documentation when something possibly anomalous arises and it will 15

Some additional points Data rejection does exist and is necessary. If you can document a problem, then it is easy to discard There still may be some data you would like to throw out. this is tricky and takes some carefully prepared, bias-free statistical tests to justify Theory curves can be misleading* and should generally (always?) be avoided when dealing with issues of data rejection You must also think in reverse. How self-consistent is your data set? There are then many sophisticated tests of the data set itself You will be expected to demonstrate this in many cases 16

Summary (for your report) Always include uncertain estimates for all your measurements if applicable. (use correct number of significant digits) Compare your results with published values if applicable: is your measurements agree with the published values within uncertainty? If not, is your estimate of systematic or statistical uncertainty correct? Is there other factors that can influence your result that you forgot to consider? If you need to reject certain sets or points of data, you should describe the reason that data should not be included. The reason should be based on changes in environment, setup, etc, and not solely result driven. 17