Introduction to least square fits for Geant 4 Simulation and ROOT Analysis of a Silicon Beam Telescope Mar , DESY Hamburg Olaf Behnke, DESY

Size: px
Start display at page:

Download "Introduction to least square fits for Geant 4 Simulation and ROOT Analysis of a Silicon Beam Telescope Mar , DESY Hamburg Olaf Behnke, DESY"

Transcription

1 Introduction to least square fits for Geant 4 Simulation and ROOT Analysis of a Silicon Beam Telescope Mar 4-6 2, DESY Hamburg Olaf Behnke, DESY Literature: Roger Barlow: Statistics, A Guide To The Use Of Statistical Methods In The Physical Sciences Wiley & Sons, 994 Jay Orear: Notes on Statistics for Physicists, Revised, 958, http : //

2 y (cm) Example: for possible discovery Z µ + µ need precise muon track fits ZeVis Zeus Run Event 366 date: time: 4::4 5 XY View Introductory Track fit example Track: coordinates y i measured in n detector layers at fixed positions x i perform track fit Typical Assumptions: Measurements with gaussian uncertainties Linear(ized) model, here: y = a +a x+a 2 x 2 (but could also use exact track helix model) Construct χ 2 : χ 2 = [y i (a +a x+a 2 x 2 )] 2 i σi 2 Determine a, a, a 2 by finding χ 2 minimum (normal equations) Check consistency: use χ 2 and χ 2 -fit probability reject outliers Analyse results: parameters, errors and correlations (error ellipses), track trajectory error band calculate momentum (error propagation) x (cm)

3 y (cm) Example: for possible discovery Z µ + µ need precise muon track fits ZeVis Zeus Run Event 366 date: time: 4::4 5 XY View Introductory Track fit example Track: coordinates y i measured in n detector layers at fixed positions x i perform track fit Typical Assumptions: Measurements with gaussian uncertainties Linear(ized) model, here: y = a +a x+a 2 x 2 (but could also use exact track helix model) Construct χ 2 : χ 2 = [y i (a +a x+a 2 x 2 )] 2 i σi 2 Determine a, a, a 2 by finding χ 2 minimum (normal equations) Check consistency: use χ 2 and χ 2 -fit probability reject outliers Analyse results: parameters, errors and correlations (error ellipses), track trajectory error band calculate momentum (error propagation) x (cm)

4 4 Lecture Part Least square χ 2 -fit method introduction Fit of a constant χ 2 min as consistency check

5 Method of least squares fit - Intro Example: Particle trajectory measurement y (cm).4.2 Det Det 2 Det 3 Det 4 n-measurements y i ± σ i at fixed x i.8 Model: y = f(x,a).6 here: y = ax x (cm) how to determine a? Idea: for correct a one expects: y i f(x i,a) < σ i 5

6 Method of least squares fit - Intro Example: Particle trajectory measurement y (cm).4.2 Det Det 2 Det 3 Det 4 n-measurements y i ± σ i at fixed x i.8 Model: y = f(x,a).6 here: y = ax x (cm) how to determine a? Idea: for correct a one expects: y i f(x i,a) < σ i 6

7 7 Method of least squares fit - Intro χ 2 = n i= (y i f(x i,a)) 2 σ 2 i Minimum w.r.t a determine estimator â from dχ2 da = dχ2 da = 2 n i= (y i f(x i,a)) σ 2 i df(x i,a) da = In general not analytically solvable. use iterative (numerical) methods (MINUIT, Mathematica)

8 8 Method of least squares fit - Intro χ 2 = n i= (y i f(x i,a)) 2 σ 2 i Minimum w.r.t a determine estimator â from dχ2 da = dχ2 da a=â = 2 n i= (y i f(x i,a)) σ 2 i df(x i,a) da = In general not analytically solvable. use iterative (numerical) methods (MINUIT, Mathematica)

9 9 Method of least squares fit Most general case y i, y j correlated measurem. with cov. V ij m fitparameters a χ2 = n i,j= (y i f(x i, a))v ij (y j f(x j, a)) = ( y f( a)) t V ( y f( a)) (vector-notation)

10 Method of least squares fit Most general case y i, y j correlated measurem. with cov. V ij m fitparameters a χ2 = n i,j= (y i f(x i, a))v ij (y j f(x j, a)) = ( y f( a)) t V ( y f( a)) (vector-notation)

11 Example for two correlated measurements y (cm) y =.3+-. Measure track in two detector layers with global position uncertainty. y 2 =-.+-. V =.2 + σ 2 corr σ 2 corr -. σ 2 corr. 2 + σ 2 corr σ corr x (cm)

12 Fit of a constant y (cm).4.3 Det Det 2 Det 3 Measure position of horizontally flying particle x (cm) Averaging of n measurements y i ± σ i χ 2 = n (y i a) 2 σ 2 i i 2

13 3 Fit of a constant (one measurement) Idiot example of one measurement y ± σ : y (cm) Det χ 2 = (y a) 2 σ 2 Min.χ 2 dχ 2 : da = -.2 Estimated value: â = y Error propagation: σâ = σ x (cm)

14 True and inverse probability densities for one measurement with gaussian uncertainty: â = y, σâ = σ True probability density to observe â for given true value a o : p = e (â a o) 2 2σ 2 2πσ But what if we don t know a? Estimate inverse probability density for a o from the measurement â ± σâ : p = 2πσâ e (â ao)2 2σ â 2 a o â Note: this is not a real prob. density! from now on we will use a as synonym for a! 4

15 Fit of a constant (one measurement) Inverse probability density for true a: p p e (a â)2 2σ â 2 with χ 2 = (a â)2 σâ 2 p e χ2 /2 σ σ 2 â = 2 d 2 χ 2 da 2 a=â â σ χ 2 (â ± σâ) = â Note: These two relations hold for a large class of one parameter χ 2 -fitproblems! 5

16 6 Fit of a constant - many measurements Probability for true value a to observe measurements y i, with i =,n: p(y,y 2,...,y n a) n = e 2 n i= (y i a) 2 σ 2 i i= = e χ2 2 (y i a)2 2σ e i 2 but we don t know true a, so let s turn the whole thing around to estimate probability density for true a from the measurements

17 Fit of a constant - many measurements Recalling p(y,y 2,...,y n a) = e χ2 /2 Expand χ 2 around its minimum at â: χ 2 = χ 2 (â) + dχ2 da a=â }{{} = (a â) + 2 = χ 2 (â)+h (a â) 2 with H = 2 d 2 χ 2 da 2 a=â d 2 χ 2 da 2 a=â (a â) 2 Hesse matrix (for one par. a number) p(y,y 2,...,y n a) e χ2(â) 2 }{{} Fit consistency e 2 H (â a)2 }{{} gaussian density interpreted as inverse probability density for true a: Gaussian distribution around â with width σ = H /2 7

18 Fit of a constant - many measurements Recalling p(y,y 2,...,y n a) = e χ2 /2 Expand χ 2 around its minimum at â: χ 2 = χ 2 (â) + dχ2 da a=â }{{} = (a â) + 2 = χ 2 (â)+h (a â) 2 with H = 2 d 2 χ 2 da 2 a=â d 2 χ 2 da 2 a=â (a â) 2 Hesse matrix (for one par. a number) p(y,y 2,...,y n a) e χ2(â) 2 }{{} Fit consistency e 2 H (â a)2 }{{} gaussian density interpreted as inverse probability density for true a: Gaussian distribution around â with width σ = H /2 8

19 Generalisation to any one-parameter (linear) fit χ 2 (a) = χ 2 (a â)2 (â) + σâ 2 χ 2 (â ± σâ) = χ 2 (â) + = χ 2 + min χ 2 (a).5 σ χ 2 min+ Read error directly from χ 2 curve.5 χ 2 min ˆa a-σ a+σ a 9

20 Mini-exercise Averaging of two meas. via χ 2 parabolas Two measurements y and y 2 of the observable a are represented in the figure by χ 2 parabolas: χ 2 i = (y i a) 2 /σ 2 i ; i =, 2 Determine (yes by eye!) from the two χ 2 curves the values y, σ and y 2, σ 2 Draw the total χ 2, i.e. the sum of the two parabolas (yes, do it by hand :-)) and determine â and σâ χ 2 2 (use χ 2 min and χ 2 = χ 2 min + ) How much is the error σâ reduced compared to σ and σ 2? Relax your eyes and hands ;-) χ

21 2 Averaging several measurements n measurements y i ± σ i (Quiz question: Why is n Σy i not the best average? (Note: σ σ 2, etc.) y (cm).4.3 Det DUT Det 3 Measure position of horizontally flying particle x (cm)

22 Averaging several measurements n measurements y i ± σ i : dχ 2 da = = n i= χ 2 = n i= 2(y i a) σ 2 i â = n σ 2 â = 2 i= (y i a) 2 σ 2 i = 2 [ yi σ 2 i n i= ] / n i= d 2 χ 2 da 2 = n i= [ σ 2 i y i σ 2 i σ 2 i ] + 2a n i= σ 2 i 22

23 Averaging several measurements n measurements y i ± σ i : dχ 2 da = = n i= χ 2 = n i= 2(y i a) σ 2 i â = n σ 2 â = 2 i= (y i a) 2 σ 2 i = 2 [ yi σ 2 i n i= ] / n i= d 2 χ 2 da 2 = n i= [ σ 2 i y i σ 2 i σ 2 i ] + 2a n i= σ 2 i 23

24 Averaging - just reformulated Single measurements contribute with weight G i = σ i 2 Define G s := n G i ; Hesse matrix H = d 2 χ 2 2 = G da 2 s i= â = n i= G i n i= G i y i = G s σâ from simple error propagation: σâ 2 = n ( ) 2 dâ dy i σ 2 i = n = G 2 s i= n i= i= G i = G s = n i= /σ 2 i n i= ( Gi G s ) 2 σ 2 i G i y i ; Corollar: least square fitting is nothing else than a clever mapping of measurements to the fitparameters and obtaining fitparameter uncertainties using error propagation 24

25 The role of the Hesse matrix illustrated for weighted average (just a number) H = 2 d 2 χ 2 da 2 = n i= σ 2 i + σ 2 σ 2 2 H grows with each measurement σ 2 σ2 2 σ3 2 σ 4 2 σ 5 2 H is counting the information from the measurements Finally V = H Note: all this holds also for fits with many parameters 25

26 Mini exercise weighted average y (cm) Track trajectory y =.+-.2 Weighted average of two measurements: ŷ = σŷ = ( ).2 2 = = y 2 = x (cm) 26

27 Mini exercise weighted average y (cm) Track trajectory y =.+-.2 Weighted average of two measurements: ŷ = σŷ = ( ).2 2 = = y 2 = x (cm) 27

28 28 Mini summary of what we have learnt One parameter fits: Least square expression for independent measurements: χ 2 = n (y i f(x i,a)) 2 i= σ i 2 get estimator â from minimum χ 2 dχ 2 /da a=â = True physics parameters have a definite value, so true probability densities exist only for the measurements, fitting means estimating (inverse) probability densities for the true parameters = d 2 χ 2 (general relation) σâ 2 2 da 2 a=â χ 2 (â ± σâ) = (general relation) Averaging several measurements can be easily done graphically by adding individual χ 2 parabolas Least square fitting is nothing else than clever mapping of measurements to fitparameters; errors of fitparameters can be obtained from simple errorpropagation The Hesse matrix H = 2 measurements d 2 χ 2 da 2 a=â counts the information from the

29 29 Consistency of measurements Recall inverse probability density for averaging n measurements: p(y,y 2,...,y n a) e χ2 (â) 2 }{{} Fit consistency Now lets have a closer look at the first term e 2 H (â a)2 }{{} gaussian density

30 Consistency of measurements Example: Two measurements y ± σ and y 2 ± σ 2 ; the true value a be known, are the measurements consistent with a?: χ 2 = 2 (y a) 2 + (y 2 a) 2 = 2 χ 2 = 8 σ 2 σ2 2 χ 2 is a measure of consistency But how should χ 2 be distributed? 3

31 Consistency of measurements Example: Two measurements y ± σ and y 2 ± σ 2 ; the true value a be known, are the measurements consistent with a?: χ 2 = 2 (y a) 2 + (y 2 a) 2 = 2 χ 2 = 8 σ 2 σ2 2 χ 2 is a measure of consistency But how should χ 2 be distributed? 3

32 32 χ 2 for two measurements and known true value Expected density for (y,y 2 ) (simple case a = ;σ = σ 2 = ): f(y,y 2 ) = 2π e y2 /2 e y2 2 /2 = 2π e r2 /2 with r = y 2 + y2 2 = χ 2 Probability to find value between r and r + dr: f(r)dr = 2πr 2π e r2 /2 dr = re r2 /2 dr z = r 2 : f(z)dz = f(r) dr dz dz = 2 e z/2 dz introduces χ 2 -distribution for z = χ 2 and two imensions (ndf=2):

33 33 χ 2 for two measurements and known true value Expected density for (y,y 2 ) (simple case a = ;σ = σ 2 = ): f(y,y 2 ) = 2π e y2 /2 e y2 2 /2 = 2π e r2 /2 with r = y 2 + y2 2 = χ 2 Probability to find value between r and r + dr: f(r)dr = 2πr 2π e r2 /2 dr = re r2 /2 dr z = r 2 : f(z)dz = f(r) dr dz dz = 2 e z/2 dz introduces χ 2 -distribution for z = χ 2 and two imensions (ndf=2):

34 34 χ 2 for two measurements and known true value Expected density for (y,y 2 ) (simple case a = ;σ = σ 2 = ): f(y,y 2 ) = 2π e y2 /2 e y2 2 /2 = 2π e r2 /2 with r = y 2 + y2 2 = χ 2 Probability to find value between r and r + dr: f(r)dr = 2πr 2π e r2 /2 dr = re r2 /2 dr z = r 2 : f(z)dz = f(r) dr dz dz = 2 e z/2 dz introduces χ 2 -distribution for z = χ 2 and two dimensions (ndf=2): f(z, 2) = 2 e z/2

35 χ 2 -function for n degrees of freedom maps the χ 2 in n dimensions into probability density for χ 2. χ 2 distribution f(χ 2,n) = Γ(n/2)2 n/2 (χ2 ) n/2 e χ2 /2 with Γ(n/2) = n dof = n dof =2 n dof =3 n dof =4 n dof =5 dte t t n/2 Properties: f(χ 2, n)dχ 2 = χ 2 = n V (χ 2 ) = 2n; σ(χ 2 ) = 2n χ 2 /n = V (χ 2 /n) = 2; σ(χ 2 /n) = 2/n χ 2 35

36 36 χ 2 distributions for various n χ 2 distribution n dof = n dof =2 n dof =3 n dof =4 n dof = χ 2 distribution n dof = n dof =2 n dof =3 n dof =4 n dof =5 χ 2 distr χ χ χ 2 /n dof distribution n dof = n dof =2 n dof =3 n dof =4 n dof = χ 2 /n dof distribution n dof = n dof =2 n dof =3 n dof =4 n dof =5 χ 2 /n distr χ 2 /n dof χ 2 /n dof

37 37 f(χ 2, 2) function and prob(χ 2, 2) f(χ 2, 2) = 2 e χ2 /2 prob(χ 2, 2) = χ 2 f(χ 2, 2)dχ 2 Probability to observe for repeated experiments a χ 2 greater or equal than the current one is a common measure for consistency Here prob(χ 2 = 3, 2) = TMath :: Prob(3, 2) = 22% }{{} ROOT F unction

38 prob(χ 2, n)-function for n degrees of freedom prob(χ 2,n) = χ 2 f(χ 2,n)dχ 2 = Γ(n/2) χ 2 /2 dte t t n/2 prob(χ 2,n) n dof = n dof =2 n dof =3 n dof =4 n dof = χ 2 Note: for repeated experiments expect the observed values of prob(χ 2,n) to be flatly distributed over interval [, ] 38

39 χ 2 for averaging measurements y (cm) Weighted Average x (cm) The figure shows the result of a fit of a constant using n measurements. When repeating the fit many times the resulting χ 2 min distribution should follow a χ 2 distribution with n degrees of freedom. One degree of freedom is sacrificed to determine the weighted average. A prove for this (for n = 2) is given in the appendix. 39

40 Mini exercise χ 2 and probability y (cm) The figure shows the result of a fit of a constant. Determine the total χ 2 (from reading the figure) and the χ 2 - probability x (cm) 4

41 Fits with problems: outliers Toy simulations of constant fits through data points No outliers % random outliers (σ) Exemplary fit track.3.2. a Entries Mean 4.95 RMS χ / ndf 2.7 / 9 p.3733 ±.362 track a Entries Mean RMS χ / ndf / 9 p.373 ± χ 2 min distribution for 2 experiments chi2 distr chi2d Entries 2 Mean RMS chi2 distr chi2d Entries 2 Mean 6.5 RMS prob(χ 2 min, 9) distribution for 2 experiments prob(chi2) distr pchi2d Entries 2 Mean.596 RMS.292 prob(chi2) distr pchi2d Entries 2 Mean.2372 RMS χ 2 min and prob(χ2 min,n dof) highly sensitive to wrong measurements 4

42 CDF I D I CDF II World average of W boson mass or how to arrive at a good χ 2 LEP II NuTeV χ 2 min =.8, n dof = 4, probability = M W [GeV] CDF I D I CDF II LEP II NuTeV Taking out NuTeV result: χ 2 min =.7, n dof = 3, probability =.64 Outlier rejection, is this allowed? M W [GeV] CDF I D I CDF II LEP II NuTeV M W [GeV] Scaling all errors by S = χ 2 min /n dof =.64 χ 2 min = 4., n dof = 4, probability =.4 Standard procedure by Particle Data group destroying the hard work of many experimentalists 42

43 Mini summary of what we have learnt The χ 2 min of a fit is a consistency check Expect χ 2 min/n dof for good fits 43 if χ 2 min/n dof significantly larger than one then suspect data could contain outliers or errors are (generally) underestimated the fitfunction might not be the correct model for the data for repeated experiments (e.g. many track fits) expect for good fits mean value of χ 2 min/n dof distribution and flat prob(χ 2 min, n dof ) distribution in interval [,]

44 44 Lecture part 2 General solution for linear least square fits (normal equations) Straight line fits

45 Our study object y (cm) Det Det 2 Det 3 The main trajectories we will study in this workshop are straight lines: y i = a + a x i This is a classical linear least square fit problem x (cm) 45

46 y vector of n measurements Linear least square fits y (x ). y n (x n ) with cov-matrix V Linear model y : = A a, a vector of m fitparameters Example: y = a ; a = (a ); A =.. In general: A = A( x), but no dependence on a Master formula : χ 2 = ( y A a) t V ( y A a) a. a m to be minimised w.r.t a obtain estimators ˆ a and covariance matrix Vˆ a 46

47 y vector of n measurements Linear least square fits y (x ). y n (x n ) with cov-matrix V Linear model y : = A a, a vector of m fitparameters Example: y = a ; a = (a ); A =.. In general: A = A( x), but no dependence on a Master formula : χ 2 = ( y A a) t V ( y A a) a. a m to be minimised w.r.t a obtain estimators ˆ a and covariance matrix Vˆ a 47

48 Examples for linear least square fits Linear means that y depends linearly on the fitparameters a i. a = ( a a ) ; A = x.. x n a = (a); A = e x.. e x n Watch out: function can be highly non-linear in x 48

49 49 General solution via normal equations χ 2 = ( y A a) t V ( y A a) = y t V y 2 a t AV y + a t A t V A a Min. χ 2 dχ2 d a t = 2A t V y + 2A t V A a = Solution: ˆ a = (A t V A) A t V y = H A t V y = UA t V y with U = H = Cov(ˆ a) Normal equations: Powerful & simple linear algebra to solve fit!

50 5 General solution via normal equations χ 2 = ( y A a) t V ( y A a) = y t V y 2 a t AV y + a t A t V A a Min. χ 2 dχ2 d a t = 2A t V y + 2A t V A a = Solution: ˆ a = (A t V A) A t V y = H A t V y = UA t V y with U = H = Cov(ˆ a) Normal equations: Powerful & simple linear algebra to solve fit!

51 5 Straight line fit through n detector layers y (cm).4.3 Det Det 2 Det 3 χ 2 = n i= y = A a; a = (y i a a x i ) 2 ( ) a a σ 2 x ; A=.. x n ; A t = (.. x.. x n Apply normal equations: ˆ a = (A t V A) A t V y = σ 2 (A t A) i = = ( U =.. i x i x x x 2 ( i x i i x2 i ) ( y xy ) ( i y i i x iy i = σ 2 â cov(â, â ) cov(â, â ) σ 2 â x 2 x 2 ) ) ; V = σ 2. σ 2 y = (A t A) A t y σ 2At ) ( ) ( N Nx = Nx Nx 2 ( x 2 x x ) ( y xy = (A t V A) = σ2 NV [x] Ny Nxy ) ) ( = V [x] x 2 y xxy xy + xy ( ) x 2 x x x (cm) )

52 52 Straight line fit through n detector layers y (cm).4.3 Det Det 2 Det 3 χ 2 = n i= y = A a; a = (y i a a x i ) 2 ( ) a a σ 2 x ; A=.. x n ; A t = (.. x.. x n Apply normal equations: ˆ a = (A t V A) A t V y = σ 2 (A t A) i = = ( U =.. i x i x x x 2 ( i x i i x2 i ) ( y xy ) ( i y i i x iy i = σ 2 â cov(â, â ) cov(â, â ) σ 2 â x 2 x 2 ) ) ; V = σ 2. σ 2 y = (A t A) A t y σ 2At ) ( ) ( N Nx = Nx Nx 2 ( x 2 x x ) ( y xy = (A t V A) = σ2 NV [x] Ny Nxy ) ) ( = V [x] x 2 y xxy xy + xy ( ) x 2 x x x (cm) )

53 53 Straight line fit through n detector layers y (cm).4.3 Det Det 2 Det 3 χ 2 = n i= y = A a; a = (y i a a x i ) 2 ( ) a a σ 2 x ; A=.. x n ; A t = (.. x.. x n Apply normal equations: ˆ a = (A t V A) A t V y = σ 2 (A t A) i = = ( U =.. i x i x x x 2 ( i x i i x2 i ) ( y xy ) ( i y i i x iy i = σ 2 â cov(â, â ) cov(â, â ) σ 2 â x 2 x 2 ) ) ; V = σ 2. σ 2 y = (A t A) A t y σ 2At ) ( ) ( N Nx = Nx Nx 2 ( x 2 x x ) ( y xy ) = (A t V A) = σ2 NV [x] Ny Nxy = V [x] ( ( x 2 x x ) x 2 y xxy xy + xy ) x (cm) )

54 54 Straight line fit through n detector layers y (cm).4.3 Det Det 2 Det 3 χ 2 = n i= y = A a; a = (y i a a x i ) 2 ( ) a a σ 2 x ; A=.. x n ; A t = (.. x.. x n Apply normal equations: ˆ a = (A t V A) A t V y = σ 2 (A t A) i = = ( U =.. i x i x x x 2 ( i x i i x2 i ) ( y xy ) ( i y i i x iy i = σ 2 â cov(â, â ) cov(â, â ) σ 2 â x 2 x 2 ) ) ; V = σ 2. σ 2 y = (A t A) A t y σ 2At ) ( ) ( N Nx = Nx Nx 2 ( x 2 x x ) ( y xy = (A t V A) = σ2 NV [x] Ny Nxy ) ) ( = V [x] x 2 y xxy xy + xy ( ) x 2 x x x (cm) )

55 55 Mini exercise: straight line track-fit The covariance formula ( σ 2 â cov(â, â ) cov(â, â ) σ 2 â ) = σ2 NV [x] ( x 2 x x ) is valid for e.g. a straight line track fit in N detectors of resolution σ: Determine the improvements on the slope error σ a by: a) Doubling the number of detector layers N within the same interval in x b) Distributing the detector layers over an interval in x twice as large c) Buying detectors with measurement uncertainties σ reduced by a factor two

56 Computer exercise straight line trajectory fit Physics example: A muon track is measured in four layers of streamer tube detectors at x positions of 4., 5., 6. and 7. (in cm), with a measurement precision for y of.5 cm. The goal is to determine its trajectory assuming a straight line. Macro StraightLineFit.C, accessible at http : // õbehnke/stat/gean/straightlinef it.c 56 fits a straight line track trajectory through four measured points. Steering parameters in the macro: xmin, xmax = Interval of the trajectory displayed Output: Tasks: Histogram data (it s of the type TGraphErrors) Plots are drawn of the fitted histogram with error bands error ellipse of the two fitparameters a) Run the macro as it is by.x StraightLineFit.C and fill the fit results for p, p, their errors and correlation into the table below b) Precision of trajectory: Evaluate (by eye) from the shown error bands at which point roughly the trajectory is known best and with which precision (fill the results in the table below) c) Precision of extrapolated trajectory: Evaluate the precision of the extrapolated trajectory at x = (Hint: Change xmax to large value and run the macro again) d) Effect of shift of x coordinate origin: Shift all four xv al points in the macro (simply by overwriting by hand) by a constant value -5.5, set xmin = 4. and xmax = 4. and run the macro again. Fill the fit results in the table. Can you explain why the correlation of p and p has changed? e) Apply a very precise vertex constraint at the origin: Change N to 5 and add a new first point to the measurement points list with xv al =., xerr =., yv al =. and yerr =. (just by hand). Run the macro again and write down the fitted results in the table. How much are the parameter errors reduced by adding this extra point?

57 Task results sheet 57 Task a) Task b) Straight line fit trough four points p = p = corr = x-best precision = y-error = Task c) y-error(x = ) = Task d) Task e) Shifting all x values by -5.5: p = p = corr = Adding vertex constraint at x = : p = p = corr =

58 Mini summary of what we have learnt 58 Linear least square problems: y = A a, y is a linear function of the fitparameters a but can be a linear or nonlinear function of the continuous parameter x. The normal equations are a powerful tool to solve linear least square fit problems ˆ a = (A t V A) A t V y, cov(ˆ a) = (A t V A) Straight line fits are a typical application and there are many others (e.g. parabolas, higher order polynoms, etc.)

59 59 Appendix Content: Proof that χ 2 min for averaging two measurements follows χ 2 -distribution with one degree of freedom Linear least square fits: Covariance matrix of fit parameters

60 χ 2 for two measurements with unknown true value χ 2 min = (y â) 2 σ 2 + (y 2 â) 2 σ 2 2 χ 2 min = G ; â = σ 2 + σ 2 2 ( y σ 2 ( y (G y + G 2 y 2 ) G + G 2 + y ) 2 σ2 2 = G y + G 2 y 2 G + G 2 (with G i := /σ 2 i )) ) 2 ( + G 2 y 2 (G y + G 2 y 2 ) G + G 2 ( ) 2 ( ) 2 (G2 y G 2 y 2 ) (G y 2 G y ) = G + G 2 G + G 2 G + G 2 G G 2 2 = (G + G 2 ) 2(y y 2 ) 2 G 2 G 2 + (G + G 2 ) 2(y y 2 ) 2 = G G 2 (G + G 2 ) (G + G 2 ) 2 (y y 2 ) 2 = G G 2 (y y 2 ) 2 G + G 2 = (y y 2 ) 2 = /G + /G 2 σ 2 + (y y 2 ) 2 σ2 2 ) 2 = y y 2 σ 2+σ2 2 should follow (errorpropagation!) gauss distribution e 2 2 χ 2 = 2 follows -dim χ 2 distr.! One degree of freedom sacrificed for determination of â. General: n-measurements with one unknown a follows χ 2 distribution with n degrees of freedom 6

61 6 Linear least square fits: Covariance Matrix Proof that Covariance matrix U of fit parameters ˆ a is given by U = H Use Normal Equations: ˆ a = B y with B = H A t V Then apply errorpropagation: U = BV B t = H A t V V V AH = H A t V AH = H HH = H

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg Statistics for Data Analysis PSI Practical Course 2014 Niklaus Berger Physics Institute, University of Heidelberg Overview You are going to perform a data analysis: Compare measured distributions to theoretical

More information

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit Statistics Lent Term 2015 Prof. Mark Thomson Lecture 2 : The Gaussian Limit Prof. M.A. Thomson Lent Term 2015 29 Lecture Lecture Lecture Lecture 1: Back to basics Introduction, Probability distribution

More information

Python Analysis. PHYS 224 October 1/2, 2015

Python Analysis. PHYS 224 October 1/2, 2015 Python Analysis PHYS 224 October 1/2, 2015 Goals Two things to teach in this lecture 1. How to use python to fit data 2. How to interpret what python gives you Some references: http://nbviewer.ipython.org/url/media.usm.maine.edu/~pauln/

More information

Parameter Estimation and Fitting to Data

Parameter Estimation and Fitting to Data Parameter Estimation and Fitting to Data Parameter estimation Maximum likelihood Least squares Goodness-of-fit Examples Elton S. Smith, Jefferson Lab 1 Parameter estimation Properties of estimators 3 An

More information

Python Analysis. PHYS 224 September 25/26, 2014

Python Analysis. PHYS 224 September 25/26, 2014 Python Analysis PHYS 224 September 25/26, 2014 Goals Two things to teach in this lecture 1. How to use python to fit data 2. How to interpret what python gives you Some references: http://nbviewer.ipython.org/url/media.usm.maine.edu/~pauln/

More information

Modern Methods of Data Analysis - WS 07/08

Modern Methods of Data Analysis - WS 07/08 Modern Methods of Data Analysis Lecture V (12.11.07) Contents: Central Limit Theorem Uncertainties: concepts, propagation and properties Central Limit Theorem Consider the sum X of n independent variables,

More information

Event Reconstruction: Tracking

Event Reconstruction: Tracking Event Reconstruction: Tracking Few points Most everyone did not get the randomness part of homework one correctly. If you want to generate a random number from a distribution you can just generate a random

More information

1 Random and systematic errors

1 Random and systematic errors 1 ESTIMATION OF RELIABILITY OF RESULTS Such a thing as an exact measurement has never been made. Every value read from the scale of an instrument has a possible error; the best that can be done is to say

More information

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising Review: Optimal Average/Scaling is equivalent to Minimise χ Two 1-parameter models: Estimating < > : Scaling a pattern: Two equivalent methods: Algebra of Random Variables: Optimal Average and Optimal

More information

Statistics, Data Analysis, and Simulation SS 2015

Statistics, Data Analysis, and Simulation SS 2015 Statistics, Data Analysis, and Simulation SS 2015 08.128.730 Statistik, Datenanalyse und Simulation Dr. Michael O. Distler Mainz, 27. April 2015 Dr. Michael O. Distler

More information

Simple Linear Regression for the MPG Data

Simple Linear Regression for the MPG Data Simple Linear Regression for the MPG Data 2000 2500 3000 3500 15 20 25 30 35 40 45 Wgt MPG What do we do with the data? y i = MPG of i th car x i = Weight of i th car i =1,...,n n = Sample Size Exploratory

More information

Introduction to Statistics and Data Analysis

Introduction to Statistics and Data Analysis Introduction to Statistics and Data Analysis RSI 2005 Staff July 15, 2005 Variation and Statistics Good experimental technique often requires repeated measurements of the same quantity These repeatedly

More information

Experimental Uncertainty (Error) and Data Analysis

Experimental Uncertainty (Error) and Data Analysis Experimental Uncertainty (Error) and Data Analysis Advance Study Assignment Please contact Dr. Reuven at yreuven@mhrd.org if you have any questions Read the Theory part of the experiment (pages 2-14) and

More information

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Propagation of Uncertainties. Department of Physics and Astronomy University of Rochester Physics 403 Propagation of Uncertainties Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Maximum Likelihood and Minimum Least Squares Uncertainty Intervals

More information

Inclusive Deep-Inelastic Scattering at HERA

Inclusive Deep-Inelastic Scattering at HERA Inclusive Deep-Inelastic Scattering at HERA Vladimir Chekelian (MPI for Physics, Munich) on behalf of the H and ZEUS Collaborations Completion of the HERA inclusive DIS cross section measurements: HERA

More information

1 Inner Product and Orthogonality

1 Inner Product and Orthogonality CSCI 4/Fall 6/Vora/GWU/Orthogonality and Norms Inner Product and Orthogonality Definition : The inner product of two vectors x and y, x x x =.., y =. x n y y... y n is denoted x, y : Note that n x, y =

More information

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising

Algebra of Random Variables: Optimal Average and Optimal Scaling Minimising Review: Optimal Average/Scaling is equivalent to Minimise χ Two 1-parameter models: Estimating < > : Scaling a pattern: Two equivalent methods: Algebra of Random Variables: Optimal Average and Optimal

More information

7.0 Lesson Plan. Regression. Residuals

7.0 Lesson Plan. Regression. Residuals 7.0 Lesson Plan Regression Residuals 1 7.1 More About Regression Recall the regression assumptions: 1. Each point (X i, Y i ) in the scatterplot satisfies: Y i = ax i + b + ɛ i where the ɛ i have a normal

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics Lecture 10 December 17, 01 Silvia Masciocchi, GSI Darmstadt Winter Semester 01 / 13 Method of least squares The method of least squares is a standard approach to

More information

Approximate Linear Relationships

Approximate Linear Relationships Approximate Linear Relationships In the real world, rarely do things follow trends perfectly. When the trend is expected to behave linearly, or when inspection suggests the trend is behaving linearly,

More information

STATISTICS OF OBSERVATIONS & SAMPLING THEORY. Parent Distributions

STATISTICS OF OBSERVATIONS & SAMPLING THEORY. Parent Distributions ASTR 511/O Connell Lec 6 1 STATISTICS OF OBSERVATIONS & SAMPLING THEORY References: Bevington Data Reduction & Error Analysis for the Physical Sciences LLM: Appendix B Warning: the introductory literature

More information

An introduction to plotting data

An introduction to plotting data An introduction to plotting data Eric D. Black California Institute of Technology v2.0 1 Introduction Plotting data is one of the essential skills every scientist must have. We use it on a near-daily basis

More information

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx

f X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx INDEPENDENCE, COVARIANCE AND CORRELATION Independence: Intuitive idea of "Y is independent of X": The distribution of Y doesn't depend on the value of X. In terms of the conditional pdf's: "f(y x doesn't

More information

2 Functions of random variables

2 Functions of random variables 2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

Least squares: introduction to the network adjustment

Least squares: introduction to the network adjustment Least squares: introduction to the network adjustment Experimental evidence and consequences Observations of the same quantity that have been performed at the highest possible accuracy provide different

More information

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6 MTH739U/P: Topics in Scientific Computing Autumn 16 Week 6 4.5 Generic algorithms for non-uniform variates We have seen that sampling from a uniform distribution in [, 1] is a relatively straightforward

More information

Advanced Engineering Statistics - Section 5 - Jay Liu Dept. Chemical Engineering PKNU

Advanced Engineering Statistics - Section 5 - Jay Liu Dept. Chemical Engineering PKNU Advanced Engineering Statistics - Section 5 - Jay Liu Dept. Chemical Engineering PKNU Least squares regression What we will cover Box, G.E.P., Use and abuse of regression, Technometrics, 8 (4), 625-629,

More information

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Homework 11/Solutions. (Section 6.8 Exercise 3). Which pairs of the following vector spaces are isomorphic?

Homework 11/Solutions. (Section 6.8 Exercise 3). Which pairs of the following vector spaces are isomorphic? MTH 9-4 Linear Algebra I F Section Exercises 6.8,4,5 7.,b 7.,, Homework /Solutions (Section 6.8 Exercise ). Which pairs of the following vector spaces are isomorphic? R 7, R, M(, ), M(, 4), M(4, ), P 6,

More information

EE/CpE 345. Modeling and Simulation. Fall Class 10 November 18, 2002

EE/CpE 345. Modeling and Simulation. Fall Class 10 November 18, 2002 EE/CpE 345 Modeling and Simulation Class 0 November 8, 2002 Input Modeling Inputs(t) Actual System Outputs(t) Parameters? Simulated System Outputs(t) The input data is the driving force for the simulation

More information

Review of Statistics

Review of Statistics Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and

More information

The Mu3e Experiment - Goal

The Mu3e Experiment - Goal Christian Graf 14.3.2016 The Mu3e Experiment A Calibration Scheme for the Mu3e Tile Detector Master Thesis Presentation for MPI Munich The Mu3e Experiment - Goal p µ DECAY MODES Fraction (Γ i /Γ) Confidence

More information

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That

Statistics. Lecture 4 August 9, 2000 Frank Porter Caltech. 1. The Fundamentals; Point Estimation. 2. Maximum Likelihood, Least Squares and All That Statistics Lecture 4 August 9, 2000 Frank Porter Caltech The plan for these lectures: 1. The Fundamentals; Point Estimation 2. Maximum Likelihood, Least Squares and All That 3. What is a Confidence Interval?

More information

Experiment 4 Free Fall

Experiment 4 Free Fall PHY9 Experiment 4: Free Fall 8/0/007 Page Experiment 4 Free Fall Suggested Reading for this Lab Bauer&Westfall Ch (as needed) Taylor, Section.6, and standard deviation rule ( t < ) rule in the uncertainty

More information

Experimental Uncertainty (Error) and Data Analysis

Experimental Uncertainty (Error) and Data Analysis E X P E R I M E N T 1 Experimental Uncertainty (Error) and Data Analysis INTRODUCTION AND OBJECTIVES Laboratory investigations involve taking measurements of physical quantities, and the process of taking

More information

The Multivariate Normal Distribution 1

The Multivariate Normal Distribution 1 The Multivariate Normal Distribution 1 STA 302 Fall 2017 1 See last slide for copyright information. 1 / 40 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2

More information

Introduction to Error Analysis

Introduction to Error Analysis Introduction to Error Analysis Part 1: the Basics Andrei Gritsan based on lectures by Petar Maksimović February 1, 2010 Overview Definitions Reporting results and rounding Accuracy vs precision systematic

More information

Uncertainty in Physical Measurements: Module 5 Data with Two Variables

Uncertainty in Physical Measurements: Module 5 Data with Two Variables : Often data have two variables, such as the magnitude of the force F exerted on an object and the object s acceleration a. In this Module we will examine some ways to determine how one of the variables,

More information

Error Analysis in Experimental Physical Science Mini-Version

Error Analysis in Experimental Physical Science Mini-Version Error Analysis in Experimental Physical Science Mini-Version by David Harrison and Jason Harlow Last updated July 13, 2012 by Jason Harlow. Original version written by David M. Harrison, Department of

More information

Multivariate Random Variable

Multivariate Random Variable Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate

More information

Linear Regression (continued)

Linear Regression (continued) Linear Regression (continued) Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 6, 2017 1 / 39 Outline 1 Administration 2 Review of last lecture 3 Linear regression

More information

1 TIME OF ARRIVAL LOCATION TECHNIQUE

1 TIME OF ARRIVAL LOCATION TECHNIQUE Contents 1 TIME OF ARRIVAL LOCATION TECHNIQUE 2 1.1 TOA Principals...................................... 2 1.2 Time of Arrival Location Error Analysis........................ 8 1.3 Determining the Best

More information

The Importance of High-Precision Hadronic Calorimetry to Physics

The Importance of High-Precision Hadronic Calorimetry to Physics The Importance of High-Precision Hadronic Calorimetry to Physics John Hauptman, Iowa State University IAS Program in High Energy Physics The Hong Kong University of Science and Technology 4-30 January

More information

Error propagation. Alexander Khanov. October 4, PHYS6260: Experimental Methods is HEP Oklahoma State University

Error propagation. Alexander Khanov. October 4, PHYS6260: Experimental Methods is HEP Oklahoma State University Error propagation Alexander Khanov PHYS660: Experimental Methods is HEP Oklahoma State University October 4, 017 Why error propagation? In many cases we measure one thing and want to know something else

More information

Comments on χ 2 minimisation

Comments on χ 2 minimisation Comments on χ 2 minimisation Volker Blobel University of Hamburg September 23 1. Introduction Parameter estimation Calorimeter calibration Common normalisation errors 2. Standard methods Least squares

More information

Lecture 9 SLR in Matrix Form

Lecture 9 SLR in Matrix Form Lecture 9 SLR in Matrix Form STAT 51 Spring 011 Background Reading KNNL: Chapter 5 9-1 Topic Overview Matrix Equations for SLR Don t focus so much on the matrix arithmetic as on the form of the equations.

More information

PHYSICS 2150 EXPERIMENTAL MODERN PHYSICS. Lecture 3 Rejection of Data; Weighted Averages

PHYSICS 2150 EXPERIMENTAL MODERN PHYSICS. Lecture 3 Rejection of Data; Weighted Averages PHYSICS 15 EXPERIMENTAL MODERN PHYSICS Lecture 3 Rejection of Data; Weighted Averages PREVIOUS LECTURE: GAUSS DISTRIBUTION 1.5 p(x µ, )= 1 e 1 ( x µ ) µ=, σ=.5 1. µ=3, σ=.5.5 µ=4, σ=1 4 6 8 WE CAN NOW

More information

Introduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves

Introduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves Introduction to Computer Graphics (Lecture No 07) Ellipse and Other Curves 7.1 Ellipse An ellipse is a curve that is the locus of all points in the plane the sum of whose distances r1 and r from two fixed

More information

EE/CpE 345. Modeling and Simulation. Fall Class 9

EE/CpE 345. Modeling and Simulation. Fall Class 9 EE/CpE 345 Modeling and Simulation Class 9 208 Input Modeling Inputs(t) Actual System Outputs(t) Parameters? Simulated System Outputs(t) The input data is the driving force for the simulation - the behavior

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

Uncertainty in Physical Measurements: Module 5 Data with Two Variables

Uncertainty in Physical Measurements: Module 5 Data with Two Variables : Module 5 Data with Two Variables Often data have two variables, such as the magnitude of the force F exerted on an object and the object s acceleration a. In this Module we will examine some ways to

More information

SKILL BUILDER TEN. Graphs of Linear Equations with Two Variables. If x = 2 then y = = = 7 and (2, 7) is a solution.

SKILL BUILDER TEN. Graphs of Linear Equations with Two Variables. If x = 2 then y = = = 7 and (2, 7) is a solution. SKILL BUILDER TEN Graphs of Linear Equations with Two Variables A first degree equation is called a linear equation, since its graph is a straight line. In a linear equation, each term is a constant or

More information

A fast introduction to the tracking and to the Kalman filter. Alberto Rotondi Pavia

A fast introduction to the tracking and to the Kalman filter. Alberto Rotondi Pavia A fast introduction to the tracking and to the Kalman filter Alberto Rotondi Pavia The tracking To reconstruct the particle path to find the origin (vertex) and the momentum The trajectory is usually curved

More information

Unit 6 - Simple linear regression

Unit 6 - Simple linear regression Sta 101: Data Analysis and Statistical Inference Dr. Çetinkaya-Rundel Unit 6 - Simple linear regression LO 1. Define the explanatory variable as the independent variable (predictor), and the response variable

More information

Practical Statistics

Practical Statistics Practical Statistics Lecture 1 (Nov. 9): - Correlation - Hypothesis Testing Lecture 2 (Nov. 16): - Error Estimation - Bayesian Analysis - Rejecting Outliers Lecture 3 (Nov. 18) - Monte Carlo Modeling -

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics Lecture 3 October 29, 2012 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2012 / 13 Outline Reminder: Probability density function Cumulative

More information

Mathematics of Physics and Engineering II: Homework problems

Mathematics of Physics and Engineering II: Homework problems Mathematics of Physics and Engineering II: Homework problems Homework. Problem. Consider four points in R 3 : P (,, ), Q(,, 2), R(,, ), S( + a,, 2a), where a is a real number. () Compute the coordinates

More information

Sections 7.1, 7.2: Sums, differences, products of polynomials CHAPTER 7: POLYNOMIALS

Sections 7.1, 7.2: Sums, differences, products of polynomials CHAPTER 7: POLYNOMIALS Sections 7.1, 7.2: Sums, differences, products of polynomials CHAPTER 7: POLYNOMIALS Quiz results Average 73%: high h score 100% Problems: Keeping track of negative signs x = + = + Function notation f(x)

More information

Linear Algebra, Summer 2011, pt. 2

Linear Algebra, Summer 2011, pt. 2 Linear Algebra, Summer 2, pt. 2 June 8, 2 Contents Inverses. 2 Vector Spaces. 3 2. Examples of vector spaces..................... 3 2.2 The column space......................... 6 2.3 The null space...........................

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics. Probability Distributions Prof. Dr. Klaus Reygers (lectures) Dr. Sebastian Neubert (tutorials) Heidelberg University WS 07/8 Gaussian g(x; µ, )= p exp (x µ) https://en.wikipedia.org/wiki/normal_distribution

More information

The Delta Method and Applications

The Delta Method and Applications Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order

More information

STA 302f16 Assignment Five 1

STA 302f16 Assignment Five 1 STA 30f16 Assignment Five 1 Except for Problem??, these problems are preparation for the quiz in tutorial on Thursday October 0th, and are not to be handed in As usual, at times you may be asked to prove

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Lecture 9: March 26, 2014

Lecture 9: March 26, 2014 COMS 6998-3: Sub-Linear Algorithms in Learning and Testing Lecturer: Rocco Servedio Lecture 9: March 26, 204 Spring 204 Scriber: Keith Nichols Overview. Last Time Finished analysis of O ( n ɛ ) -query

More information

Statistical Methods for Astronomy

Statistical Methods for Astronomy Statistical Methods for Astronomy Probability (Lecture 1) Statistics (Lecture 2) Why do we need statistics? Useful Statistics Definitions Error Analysis Probability distributions Error Propagation Binomial

More information

Error Analysis. V. Lorenz L. Yang, M. Grosse Perdekamp, D. Hertzog, R. Clegg PHYS403 Spring 2016

Error Analysis. V. Lorenz L. Yang, M. Grosse Perdekamp, D. Hertzog, R. Clegg PHYS403 Spring 2016 Error Analysis V. Lorenz L. Yang, M. Grosse Perdekamp, D. Hertzog, R. Clegg PHYS403 Spring 2016 Reporting measurement results Always include uncertainty estimates in your results Have the correct number

More information

The Multivariate Normal Distribution 1

The Multivariate Normal Distribution 1 The Multivariate Normal Distribution 1 STA 302 Fall 2014 1 See last slide for copyright information. 1 / 37 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2

More information

Experiment 2. F r e e F a l l

Experiment 2. F r e e F a l l Suggested Reading for this Lab Experiment F r e e F a l l Taylor, Section.6, and standard deviation rule in Taylor handout. Review Chapters 3 & 4, Read Sections 8.1-8.6. You will also need some procedures

More information

Statistics and Data Analysis

Statistics and Data Analysis Statistics and Data Analysis The Crash Course Physics 226, Fall 2013 "There are three kinds of lies: lies, damned lies, and statistics. Mark Twain, allegedly after Benjamin Disraeli Statistics and Data

More information

Water tank. Fortunately there are a couple of objectors. Why is it straight? Shouldn t it be a curve?

Water tank. Fortunately there are a couple of objectors. Why is it straight? Shouldn t it be a curve? Water tank (a) A cylindrical tank contains 800 ml of water. At t=0 (minutes) a hole is punched in the bottom, and water begins to flow out. It takes exactly 100 seconds for the tank to empty. Draw the

More information

18.02 Multivariable Calculus Fall 2007

18.02 Multivariable Calculus Fall 2007 MIT OpenourseWare http://ocw.mit.edu 8.02 Multivariable alculus Fall 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms. 8.02 Lecture 8. hange of variables.

More information

Lecture 2: Review of Probability

Lecture 2: Review of Probability Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................

More information

Systems of Nonlinear Equations and Inequalities: Two Variables

Systems of Nonlinear Equations and Inequalities: Two Variables Systems of Nonlinear Equations and Inequalities: Two Variables By: OpenStaxCollege Halley s Comet ([link]) orbits the sun about once every 75 years. Its path can be considered to be a very elongated ellipse.

More information

CMS Note Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland

CMS Note Mailing address: CMS CERN, CH-1211 GENEVA 23, Switzerland Available on CMS information server CMS OTE 997/064 The Compact Muon Solenoid Experiment CMS ote Mailing address: CMS CER, CH- GEEVA 3, Switzerland 3 July 997 Explicit Covariance Matrix for Particle Measurement

More information

Lecture 1: Review of linear algebra

Lecture 1: Review of linear algebra Lecture 1: Review of linear algebra Linear functions and linearization Inverse matrix, least-squares and least-norm solutions Subspaces, basis, and dimension Change of basis and similarity transformations

More information

ECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1

ECE 650 Lecture 4. Intro to Estimation Theory Random Vectors. ECE 650 D. Van Alphen 1 EE 650 Lecture 4 Intro to Estimation Theory Random Vectors EE 650 D. Van Alphen 1 Lecture Overview: Random Variables & Estimation Theory Functions of RV s (5.9) Introduction to Estimation Theory MMSE Estimation

More information

Regression and Statistical Inference

Regression and Statistical Inference Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF

More information

STA 2201/442 Assignment 2

STA 2201/442 Assignment 2 STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution

More information

Simple Linear Regression for the Climate Data

Simple Linear Regression for the Climate Data Prediction Prediction Interval Temperature 0.2 0.0 0.2 0.4 0.6 0.8 320 340 360 380 CO 2 Simple Linear Regression for the Climate Data What do we do with the data? y i = Temperature of i th Year x i =CO

More information

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO LECTURE NOTES FYS 4550/FYS9550 - EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I PROBABILITY AND STATISTICS A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO Before embarking on the concept

More information

Global χ 2 approach to the Alignment of the ATLAS Silicon Tracking Detectors

Global χ 2 approach to the Alignment of the ATLAS Silicon Tracking Detectors Global χ 2 approach to the Alignment of the ALAS Silicon racking Detectors P. Brückman University of Oxford DWB, Keble Road Oxford OX1 3RH, UK and Institute of Nuclear Physics PAN ul. Radzikowskiego 152

More information

CMS Internal Note. The content of this note is intended for CMS internal use and distribution only

CMS Internal Note. The content of this note is intended for CMS internal use and distribution only Available on CMS information server CMS IN 2003/xxxx CMS Internal Note The content of this note is intended for CMS internal use and distribution only August 26, 2003 Expected signal observability at future

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

Chapter 7. Linear Regression (Pt. 1) 7.1 Introduction. 7.2 The Least-Squares Regression Line

Chapter 7. Linear Regression (Pt. 1) 7.1 Introduction. 7.2 The Least-Squares Regression Line Chapter 7 Linear Regression (Pt. 1) 7.1 Introduction Recall that r, the correlation coefficient, measures the linear association between two quantitative variables. Linear regression is the method of fitting

More information

Chapter 14. Linear least squares

Chapter 14. Linear least squares Serik Sagitov, Chalmers and GU, March 5, 2018 Chapter 14 Linear least squares 1 Simple linear regression model A linear model for the random response Y = Y (x) to an independent variable X = x For a given

More information

Lecture 11: Simple Linear Regression

Lecture 11: Simple Linear Regression Lecture 11: Simple Linear Regression Readings: Sections 3.1-3.3, 11.1-11.3 Apr 17, 2009 In linear regression, we examine the association between two quantitative variables. Number of beers that you drink

More information

Data Analysis and Fitting for Physicists

Data Analysis and Fitting for Physicists Data Analysis and Fitting for Physicists Peter Young (Dated: September 17, 2012) Contents I. Introduction 2 II. Averages and error bars 2 A. Basic Analysis 2 B. Advanced Analysis 8 1. Traditional method

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

ECON 4160, Autumn term Lecture 1

ECON 4160, Autumn term Lecture 1 ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least

More information

Chapter 8. Linear Regression. Copyright 2010 Pearson Education, Inc.

Chapter 8. Linear Regression. Copyright 2010 Pearson Education, Inc. Chapter 8 Linear Regression Copyright 2010 Pearson Education, Inc. Fat Versus Protein: An Example The following is a scatterplot of total fat versus protein for 30 items on the Burger King menu: Copyright

More information

Lecture 18: Simple Linear Regression

Lecture 18: Simple Linear Regression Lecture 18: Simple Linear Regression BIOS 553 Department of Biostatistics University of Michigan Fall 2004 The Correlation Coefficient: r The correlation coefficient (r) is a number that measures the strength

More information

General Form: y = a n x n + a n 1 x n a 2 x 2 + a 1 x + a 0

General Form: y = a n x n + a n 1 x n a 2 x 2 + a 1 x + a 0 Families of Functions Prepared by: Sa diyya Hendrickson Name: Date: Definition: function A function f is a rule that relates two sets by assigning to some element (e.g. x) in a set A exactly one element

More information

Physics 509: Error Propagation, and the Meaning of Error Bars. Scott Oser Lecture #10

Physics 509: Error Propagation, and the Meaning of Error Bars. Scott Oser Lecture #10 Physics 509: Error Propagation, and the Meaning of Error Bars Scott Oser Lecture #10 1 What is an error bar? Someone hands you a plot like this. What do the error bars indicate? Answer: you can never be

More information

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests http://benasque.org/2018tae/cgi-bin/talks/allprint.pl TAE 2018 Benasque, Spain 3-15 Sept 2018 Glen Cowan Physics

More information

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel Lecture Notes 4 Vector Detection and Estimation Vector Detection Reconstruction Problem Detection for Vector AGN Channel Vector Linear Estimation Linear Innovation Sequence Kalman Filter EE 278B: Random

More information

ECE Lecture #9 Part 2 Overview

ECE Lecture #9 Part 2 Overview ECE 450 - Lecture #9 Part Overview Bivariate Moments Mean or Expected Value of Z = g(x, Y) Correlation and Covariance of RV s Functions of RV s: Z = g(x, Y); finding f Z (z) Method : First find F(z), by

More information

Unit 6 - Introduction to linear regression

Unit 6 - Introduction to linear regression Unit 6 - Introduction to linear regression Suggested reading: OpenIntro Statistics, Chapter 7 Suggested exercises: Part 1 - Relationship between two numerical variables: 7.7, 7.9, 7.11, 7.13, 7.15, 7.25,

More information

Preliminary Statistics. Lecture 3: Probability Models and Distributions

Preliminary Statistics. Lecture 3: Probability Models and Distributions Preliminary Statistics Lecture 3: Probability Models and Distributions Rory Macqueen (rm43@soas.ac.uk), September 2015 Outline Revision of Lecture 2 Probability Density Functions Cumulative Distribution

More information