Parameter Estimation and Fitting to Data

Similar documents
Parameter Estimation and Hypothesis Testing

Statistical Methods for Astronomy

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Hypothesis testing (cont d)

Composite Hypotheses and Generalized Likelihood Ratio Tests

Statistical Data Analysis Stat 3: p-values, parameter estimation

10. Composite Hypothesis Testing. ECE 830, Spring 2014

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Modern Methods of Data Analysis - WS 07/08

Intermediate Lab PHYS 3870

Statistics and Data Analysis

Institute of Actuaries of India

1 Degree distributions and data

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LINEAR REGRESSION ANALYSIS. MODULE XVI Lecture Exercises

y Xw 2 2 y Xw λ w 2 2

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Frequentist-Bayesian Model Comparisons: A Simple Example

Statistical Methods for Particle Physics Lecture 3: Systematics, nuisance parameters

TMA 4275 Lifetime Analysis June 2004 Solution

Practice Problems Section Problems

Scatter plot of data from the study. Linear Regression

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Simple and Multiple Linear Regression

Statistical Methods in Particle Physics

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Poisson distribution and χ 2 (Chap 11-12)

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

MS&E 226: Small Data

Scatter plot of data from the study. Linear Regression

Psychology 282 Lecture #4 Outline Inferences in SLR

32. STATISTICS. 32. Statistics 1

Negative binomial distribution and multiplicities in p p( p) collisions

LECTURE 5 HYPOTHESIS TESTING

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

32. STATISTICS. 32. Statistics 1

Recall the Basics of Hypothesis Testing

If we want to analyze experimental or simulated data we might encounter the following tasks:

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Basics of Experimental Design. Review of Statistics. Basic Study. Experimental Design. When an Experiment is Not Possible. Studying Relations

A732: Exercise #7 Maximum Likelihood

COM336: Neural Computing

Dr. Junchao Xia Center of Biophysics and Computational Biology. Fall /1/2016 1/46

Generalized Linear Models (1/29/13)

Hypothesis testing: theory and methods

[y i α βx i ] 2 (2) Q = i=1

MATH5745 Multivariate Methods Lecture 07

Introduction to Error Analysis

STATISTICS OF OBSERVATIONS & SAMPLING THEORY. Parent Distributions

Python Analysis. PHYS 224 September 25/26, 2014

PHYS 6710: Nuclear and Particle Physics II

Notes on the Multivariate Normal and Related Topics

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

Python Analysis. PHYS 224 October 1/2, 2015

Lecture 2. G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1

Computational functional genomics

Least Squares Regression

SPRING 2007 EXAM C SOLUTIONS

Probability and Statistics Notes

ISQS 5349 Final Exam, Spring 2017.

Some Statistics. V. Lindberg. May 16, 2007

Linear Regression and Its Applications

IEOR 165 Lecture 13 Maximum Likelihood Estimation

Business Statistics. Lecture 10: Correlation and Linear Regression

HYPOTHESIS TESTING: FREQUENTIST APPROACH.

Least Squares Regression

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit

Statistical Methods for Particle Physics Lecture 4: discovery, exclusion limits

Simple Linear Regression for the Climate Data

Modeling Data with Linear Combinations of Basis Functions. Read Chapter 3 in the text by Bishop

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. - Statistics and Error Analysis -

Some Statistical Tools for Particle Physics

Lecture 5: GPs and Streaming regression

Direction: This test is worth 250 points and each problem worth points. DO ANY SIX

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Chapter 9: Hypothesis Testing Sections

Intro to Probability. Andrei Barbu

8 Basics of Hypothesis Testing

MS&E 226: Small Data

FYST17 Lecture 8 Statistics and hypothesis testing. Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons

ECON 4160, Autumn term Lecture 1

Six Sigma Black Belt Study Guides

Comparing two independent samples

Lecture 21: October 19

ECONOMETRIC THEORY. MODULE VI Lecture 19 Regression Analysis Under Linear Restrictions

Lecture 3: Basic Statistical Tools. Bruce Walsh lecture notes Tucson Winter Institute 7-9 Jan 2013

E. Santovetti lesson 4 Maximum likelihood Interval estimation

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

Statistics for the LHC Lecture 1: Introduction

Lecture 1 Basic Statistical Machinery

Statistical Distribution Assumptions of General Linear Models

Chapter 14 Simple Linear Regression (A)

Transcription:

Parameter Estimation and Fitting to Data Parameter estimation Maximum likelihood Least squares Goodness-of-fit Examples Elton S. Smith, Jefferson Lab 1

Parameter estimation

Properties of estimators 3

An estimator for the mean 4

An estimator for the variance 5

The Likelihood function via example We have a data set given by N data pairs (x i, y i ±σ i ) graphically represented below. The goal is to determine the fixed, but unknown, µ = f(x). σ is known or estimated from the data. 6

Gaussian probabilities (least-squares) We assume that at a fixed value of x i, we have made a measurement y i and that the measurement was drawn from a Gaussian probability distribution with mean y(x i ) = a + bx i and variance σ i. f (y i ;a,b) = e (y i y(x i )) 1 σ i L = N 1 πσ i πσ i e (y i a bx i ) σ i y(x i ) = a + bx i χ = ln L + k = N (y i a bx i ) σ i 7

χ minimization χ a = N (y i a bx i ) = 0 σ i χ b = N x i (y i a bx i ) = 0 σ i a N 1 σ i + b N x i σ i = N y i σ i a N x i σ i + b N x i σ i = N x i y i σ i 8

Solution for linear fit For simplicity, assume constant σ = σ i. Then solve two simultaneous equations for unknowns: a = y i x i x i x i y i ( ) N x i x i b = N x y x i i i y i N x i x i ( ) Parameter uncertainties can be estimated from the curvature of the χ function. V[a] χ a V[b] χ b 1 θ = ˆ θ 1 = = θ = ˆ θ σ x i ( ) N x i x i Nσ ( ) N x i x i 9

Parameter uncertainties In a graphical method the uncertainty in the parameter estimator θ 0 is obtained by changing χ by one unit. χ (θ 0 ±σ θ ) = χ (θ 0 ) + 1 In general, using maximum likelihood lnl(θ 0 ±σ θ ) = lnl(θ 0 ) 1/ 10

The Likelihood function via example What does the fit look like? ROOT fit to pol1 Additional information about the fit: χ and probability Were the assumptions for the fit valid? This question is addressed by exploring the significance and the goodness of the fit 11

Testing significance/goodness of fit Quantify the level of agreement between the data and a hypothesis without explicit reference to alternative hypotheses. This is done by defining a goodness-of-fit statistic, and the goodness-of-fit is quantified using the p-value. For the case when the χ is the goodness-of-fit statistic, then the p-value is given by p = f (χ ;n d )dχ χ obs The p-value is a function of the observed value χ obs and is therefore itself a random variable. If the hypothesis used to compute the p-value is true, then p will be uniformly distributed between zero and one. 1

χ distribution Gaussian-like 13

p-value for χ distribution 14

Using the goodness-of-fit Data generated using Y(x) = 6.0 + 0. x, σ = 0.5 Compare three different polynomial fits to the same data. y(x) = p 0 + p 1 x y(x) = p 0 y(x) = p 0 + p 1 x + p x 15

χ /DOF vs degrees of freedom 16

More about the likelihood method Recall likelihood for least squares: L = N 1 πσ i e (y i a bx i ) σ i But the probability density depends on application L = N f (y i ; parameters) Proceed as before maximizing lnl (χ has minus sign). The values of the estimated parameters might not be very different, but the uncertainties can be greatly affected. 17

Applications L = N 1 πσ i e (y i a bx i ) σ i a) σ i = constant b) σ i = y i c) σ i = Y(x) Poisson distribution (see PDG Eq. 3.1) L = N ( a bx i) n i e (a bx i ) n i! Stirling s approx ln(n!) ~ n ln(n) - n ln L = N i 1 [( a + bx ) i n i + n i ln[n i /(a + bx i )]] 18

Exercise 3 - Linear Fits Assume a parent distribution of the form y(x) = a + bx, a=5, b=1 Assume one experiment collects a data set of ten points of the form (x i, y i ±σ), i=0,1,,...9, with the measurements y i following a Gaussian distribution with a fixed width σ=0.5. Invent the data points y i for one experiment. Fit the data y i to the form y = a + bx. Determine y and the uncertainty of y as a function of x from the fit. 19

Linear Fit one experiment Fit for one experiment showing the fitted parameters Uncertainty on σ y can be computed using σ y =σ a +x σ b +xσ ab 0

Linear Fit 1

Linear Fit Covariance Matrix

Fitted Results to 1000 experiments For each fit, plot the fitted value of the intercept and slope. Fit the distributions to Gaussian functions Mean = 5.004 ± 0.010 Sigma = 0.303 What is the relation between these two? Mean = 0.9997 ± 0.0018 Sigma = 0.0566 3

Plot difference between fitted and true values Fit Gaussian to slices y(x)-y fit σ ab =0 Uncertainty on σ y can be computed using σ y =σ a +x σ b +xσ ab Correlation term is important 4

Statistical tests In addition to estimating parameters, one often wants to assess the validity of statements concerning the underlying distribution Hypothesis tests provide a rule for accepting or rejecting hypotheses depending on the outcome of an experiment. [Comparison of H 0 vs H 1 ] In goodness-of-fit tests one gives the probability to obtain a level of incompatibility with a certain hypothesis that is greater than or equal to the level observed in the actual data. [How good is my assumed H 0?] Formulation of the relevant question is critical to deciding how to answer it. 5

Summary of second lecture Parameter estimation Illustrated the method of maximum likelihood using the least squares assumption Use of goodness-of-fit statistics to determine the validity of underlying assumptions used to determine parent parameters 6