Physics 6720 Introduction to Statistics April 4, 2017

Size: px
Start display at page:

Download "Physics 6720 Introduction to Statistics April 4, 2017"

Transcription

1 Physics 6720 Introduction to Statistics April 4, Statistics of Counting Often an experiment yields a result that can be classified according to a set of discrete events, giving rise to an integer count or set of integer counts as a result. For example the measurement of a radioactive decay may yield the number of counts in a detector over a period of time. The measurement of the scattering of a beam of particles from a target may yield so may counts over a particular range of deflection angles over a period of time. Any series of measurements, the result of which can be classified into histogram bins, produces counts. In this section we will discuss briefly the statistics of counting, with particular emphasis on Poisson statistics. 1.1 Binomial Distribution We begin with the binomial distribution. Here we consider an experiment that is repeated many times. There are two possible outcomes: A and B. The probability for outcome A is p and the probability for outcome B is 1 p. We assume that each experiment has the same probability for each outcome and that there is no correlation between the outcome of one experiment and that of another. We may then ask the question, out of N repetitions of the experiment, what is the probability that we get A exactly k times? For example, suppose N = 4 and k = 3. The answer is found by first asking for the probability for a particular sequence of outcomes AABA, for example. The probability is just the product of the probabilities for each event: pp(1 p)p. This statement makes use of the fact that there is no correlation between one experiment and another. Since our question doesn t ask for a particular order of outcomes, but just any order that yields 3 A s out of 4 trials, we then ask how many different ways there are of getting 3 A s. We can enumerate them: BAAA, ABAA, AABA, and AAAB. Since the probability for each is the same, the probability for any of them is four times the probability for just one of them. So the probability is 4p 3 (1 p) to get 3 A s out of 4 trials. The general 1

2 expression is called the binomial distribution. The probability for getting k A s (and N k B s) out of N trials is P (k, N) = N! k!(n k)! pk (1 p) N k. (1) Notice that the binomial probabilities generate the binomial series, which adds up to 1, as it should: 1.2 Poisson Distribution N P (k, N) = [p + (1 p)] N = 1. (2) k=0 The Poisson distribution applies in cases in which the probability for getting A is very small compared with other possible outcomes. In that case we would use the binomial formula with a value of k much smaller than N. For example, suppose we were counting radioactive decays as a function of time and we observe the decays over a time interval dt that is much smaller than the decay lifetime, so the amount of radioactive material available for decay does not change noticeably during the time of observation. If the decay rate is λ and we consider just one single atom, the probability that it decays in a time interval dt is p = λdt (true as long as this is very small). Call this event A. If it doesn t decay (probability 1 p) we call it event B. If we now consider N atoms we can use the binomial distribution to give us the probability that that k atoms out of N atoms decay in the time interval dt. We expect that on average there will be k = pn decays. Let s find the probability for getting k events in the limit of large N, if the expected (average) number k is constant as we take the limit. Notice that to keep k constant we have to decrease p as we increase N. This means we are decreasing the time interval dt as we increase N. To get the probability, we start with the binomial distribution, substitute p = k/n and take the limit P (k, k) N! = lim N k!(n k)! ( k/n) k (1 k/n) N k (3) After some algebra, using the Stirling approximation for the factorial and the Taylor expansion for the exponential function, we get the Poisson distribution: P (k, k) = k k e k. (4) k! 2

3 This distribution is normalized to 1 as well. The sum generates the Taylor series for the exponential function: P (k, k) = e ke k = 1. (5) k=0 We will return to a discussion of properties of the Poisson distribution after discussing the Gaussian normal distribution. 2 Normal Distribution 2.1 Populations and Their Means and Standard Deviations In order to develop confidence in the result of a measurement of a single quantity, such as the length of a table top, we often repeat the measurement process a number of times. The results of the measurement vary because of difficulties in reading the meter stick scale to the last tenth of a millimeter, and for other reasons. Suppose we repeated the measurement N times, getting a list of values x i. Our best guess for the true value is usually the average of these values: N x = x i /N, (6) i=1 which is also called the mean value of this sample set of observations. In our notation, x indicates our best, imperfect estimate of the true value x. If we repeat the measurement an infinite number of times, ideally, the mean value should approach the true value of the measurement. The statistical way to describe what is happening is that our set of N measurements is a sample of N values taken from an infinite population. The true population mean is given by x = lim N N x i /N. (7) i=1 We might ask of this infinite population, what is the probability of getting a value of x in the range (x, x+dx) when we make a measurement? This probability is expressed in terms of a probability function P (x) as P (x)dx. The factor dx is necessary because as the interval width dx gets smaller, the probability of getting a value in that tiny range must get smaller in proportion to dx. 3

4 If we make enough measurements, we can begin to construct this probability function, but usually we don t make enough measurements to know it very well. So we often assume for want of any better reason that the probability is given by the Gaussian distribution function (normal distribution) P (x) = exp[ (x x)2 /2σ 2 ] 2πσ (8) In this expression the true mean of the population is x and σ is the true standard deviation. This probability is normalized so that P (x)dx = 1. (9) i.e. the probability of measuring any value of x is 1. The Gaussian distribution is peaked at x = x and falls off on either side of x over a distance in x that is controlled by the value of σ. If σ is large, the fall off is slow and the most probable values of x are in a broad range around x; if σ is small, the fall off is rapid, and the most probable values of x are narrowly clustered around x. A property of the Gaussian distribution is that the probability of making a measurement and getting a value in the range x σ and x + σ is about 68%. (This value is found by calculating the integral under the probability distribution from x σ to x + σ.) Thus in common usage, we say that for a single measured value of x, the result is x ± σ. The standard deviation of a quantity is sometimes called the error in that quantity, so we say the error in a single measurement is σ. The statement that x lies in the range x ± σ is a statement we can make with 68% confidence. That means the result of a measurement is likely to be outside this range 32% of the times we repeat the experiment. A measure of the width of this peak is given by Var(x) = σ 2 = (x x) 2 P (x)dx (10) This is just the average of (x x) 2 over the population. If we made an infinite number of measurements, we would be able to determine the two parameters x and σ of the distribution exactly. With a finite set of measurements, however, we can estimate them. To estimate the mean value, we simply compute the average of the measurements x i : N x = x = x i /N. (11) i=1 4

5 Notice that we have put a star on x to distinguish the estimate from the true value x. The sample also permits an estimate of this population standard deviation σ. It is just N Var(x 1, x 2,... ) = σ 2 = (x i x ) 2 /(N 1). (12) i=1 The quantity σ is the estimated standard deviation, and its square is called the estimated variance of x from the mean value x, or just the estimated variance of x. 1 Another useful formula is obtained by expanding the square on the right side to give (N 1)σ 2 = x 2 i 2 x x i + N x 2 = N( x 2 x 2 ). (13) The x 2 means the average of x 2 i. In other words the estimated variance is just the difference between the average of the squares and the square of the average times N/(N 1). As an exercise, try writing a program that reads a list of values x i and calculates x and σ. 2.2 The Error in the Estimated Mean So we see that if we have a finite data sample, we can get an estimate of the true values of x and σ. But how far is our estimate x from the true value x? This is the central question of every measurement, because it tells us how much confidence we may put in our result. Measurements without error ranges are meaningless! For example, there is really no meaning to the statement that the length of the table top is 3 meters, because the associated error might be a kilometer. There is meaning only if we can associate an error with this figure and say, for example, that the length is 3.00 meters with an error of plus or minus 0.01 meter. Now suppose we make N measurements to make up one data sample on one day and make another N measurements to make a second sample on the next, and so collect a large number of samples. We determine the estimated mean value x for each sample. What is the probability distribution 1 This expression for σ 2 is based on the sample mean x, and so is biased. To compensate for the bias, we divide by N 1 instead of N. 5

6 for this estimated mean value? Note that it is not the same as the probability distribution of the population. One way to see this is to realize that if we take larger and larger samples almost all of our values would be expected to be closer and closer to the true mean x. In fact, a famous theorem of statistics, called the central limit theorem states that the probability distribution of the mean value approaches a Gaussian normal distribution as the sample size increases, regardless of whether the underlying population distribution P (x) is itself Gaussian. The standard deviation of the mean value is estimated by σ mean = σ / N, (14) where σ is given by Eq (13). As the sample size grows, σ stabilizes, and the standard deviation of the mean shrinks as 1/ N, so that the distribution of sample means x gets sharper around the true value x. So as a result of measuring one sample, we estimate the true mean value to be x ± σ mean = x ± σ / N. This is a practical formula. With it we need only make N measurements, then estimate the population mean from Eq. (11) and the population standard deviation from Eq. (12). Then we compute the error in the mean from Eq. (14). Please bear in mind the difference between σ, which is the estimate of the error in a single measurement, i.e. the population standard deviation, and σmean, which is the estimate of the error in our estimated mean value x. 2.3 Systematic Error The error we have been discussing so far is a statistical error. It is an error that can be made smaller by simply making more and more measurements of the same type. Another error that occurs all-too-frequently is a systematic error. For example, in measuring the table top, it may happen that our meter stick was slightly miscalibrated, so it gave consistently large results. We would not be able to correct for such an error by repeating the measurement. We would instead have to recalibrate the meter stick. Sometimes we aren t able to do the recalibration, but are assured by the manufacturer or by some other means that the meter stick agrees with a precise standard to within an error of, say m. We might then quote the result of a measurement of the table top by saying it is ± 0.002(stat) ± 0.005(syst) meters, thereby identifying separately the two sources of error. 6

7 3 Properties of the Poisson Distribution 3.1 Mean and Variance Figure 1: Poisson distribution with mean value 5 The Poisson distribution for k = 5 is shown in Fig. 1. Notice that it peaks at k = 5. Let us determine the mean and variance for the Poisson distribution. The mean is just k = kp (k, k). (15) A little algebra gives k=0 k = k. (16) This result is naturally what we would expect, of course. The variance is given 7

8 by Var(k) = k 2 k 2 = k 2 P (k, k) k 2 (17) k=0 A little algebra shows that the first term is just k( k + 1) so Var(k) = k. (18) This result says that the standard deviation is approximately k. Actually we have to be careful about using the term standard deviation for the Poisson distribution, unless k is large. For small k the shape is not very much like a Gaussian, but for large k the shape approximates a Gaussian reasonably well. 3.2 Bayes Theorem and Maximum Likelihood So far we have been thinking of the probability for getting a result k if we know that the mean value should be k. Now suppose we make a measurement and get k counts, but we don t know anything about k, except that it must be nonnegative, of course. We may turn the question around and ask what is the most likely value for k, given the result of our measurement. To make this turned-around idea more concrete, we use the concept of conditional probability. We say that the Poisson distribution P (k, k) tells us the probability that we get k, on the condition that the mean value is k. The notation P (A C) denotes the probability for getting A, given that C occurs or C is true. Thus we could write P (k k) = P (k, k) = k k e k. (19) k! Now the reverse question is, What is the probability that the mean value is k, given that we just made a measurement and got k?. This probability would be denoted P ( k k). Now a trivial but important theorem due to Bayes states that P (A C)P (C) = P (C A)P (A) (20) where P (C) is the a priori probability for C to occur, regardless of whether the event A occurs, and P (A) is the a priori probability for A to occur, regardless of whether the event C occurs. From this theorem we conclude that P ( k k) = P (k k)p ( k)/p (k) (21) 8

9 So we need to know P ( k) and P (k) to make progress. The first is the a priori probability for getting a particular value for k. If we don t know anything about k, except that it is nonnegative, then we must say that any nonnegative value whatsoever is equally probable. Thus without benefit of knowing the outcome of the measurement, we say P ( k) is constant, independent of k for nonnegative k, and it is zero for negative k. So the rhs of this equation reduces simply to P ( k k) = N k k e k (22) k! where the normalization factor N = P ( k)/p (k) can be determined by requiring that the total probability for having any k is 1. In fact it turns out that N = 1, so P ( k k) = k k e k. (23) k! This distribution is called the likelihood function for the parameter k. Notice that we are now thinking of the rhs as a continuous function of k with fixed k. This result is very remarkable, since a single measurement is giving us the whole probability distribution! Recall that if we were to measure the length of a table top, even if we started by assuming we were going to get a Gaussian distribution, a single measurement would allow us only to guess x and would tell us nothing about σ. To get σ takes at least two measurements, and even then we would be putting ourselves at the mercy of the gods of statistics for taking a chance with only two measurements. If we weren t so rash as to assume a Gaussian, we would have to make many measurements of the length of the table top to get the probability distribution in the measured length. We now ask, what is the most probable value of k, given that we just found k? This is the value with maximum likelihood. If we examine the probability distribution, we see that it peaks at k, just as we might have expected. We may then ask, what is the error in the determination of this value. This is a tricky question, because the Poisson distribution is not shaped like a Gaussian distribution. However, for large k it looks more and more like a Gaussian. Expanding the log of the Poisson distribution for large k and fixed k k gives P ( k k) exp[ (k k) 2 /(2 k)]/ 2π k (24) so for large k the error is σ k = k. (25) 9

10 To summarize, a single measurement yields the entire probability distribution. For large enough k we can say that k = k ± k. (26) To see how Bayesian statistics works, suppose we repeated the experiment and got a new value k. What is the probability distribution for k in light of the new result? Now things have changed, since the a priori probability for k is no longer constant because we already made one measurement and got k. Instead we have P ( k) = k k e k (27) k! so P ( k k ) = N k k e k k k e k. (28) k! k! Notice that the likelihood function is now the product of the individual likelihood functions. A more systematic notation would write this function as P ( k k, k ), i.e. the probability for k having a particular value, given that we made two measurements and found k and k. The normalization factor N is obtained by requiring the total probability to be 1. The most likely value of k is easily shown to be just the average k = (k + k )/2, (29) as we should have expected. The Bayesian approach insists that we fold together all of our knowledge about a parameter in constructing its likelihood function. Thus a generalization of these results would state that the likelihood function for the parameter set C, given the independently measured results A 1, A 2, A 3, etc. is just P (C A 1, A 2, A 3,... ) = NP (A 1 C)P (A 2 C)P (A 3 C)..., (30) where N is a normalization factor. Again, this is just the product of the separate likelihood functions. The result is completely general and applies to any probability distribution, not just a Poisson distribution. We will use this result in discussing χ 2 fits to data as a maximum likelihood search. 10

Chapter 8: An Introduction to Probability and Statistics

Chapter 8: An Introduction to Probability and Statistics Course S3, 200 07 Chapter 8: An Introduction to Probability and Statistics This material is covered in the book: Erwin Kreyszig, Advanced Engineering Mathematics (9th edition) Chapter 24 (not including

More information

Introduction to Probability

Introduction to Probability Introduction to Probability Salvatore Pace September 2, 208 Introduction In a frequentist interpretation of probability, a probability measure P (A) says that if I do something N times, I should see event

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics March 14, 2018 CS 361: Probability & Statistics Inference The prior From Bayes rule, we know that we can express our function of interest as Likelihood Prior Posterior The right hand side contains the

More information

Chapter 4: An Introduction to Probability and Statistics

Chapter 4: An Introduction to Probability and Statistics Chapter 4: An Introduction to Probability and Statistics 4. Probability The simplest kinds of probabilities to understand are reflected in everyday ideas like these: (i) if you toss a coin, the probability

More information

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester Physics 403 Choosing Priors and the Principle of Maximum Entropy Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Odds Ratio Occam Factors

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite

More information

Physics Sep Example A Spin System

Physics Sep Example A Spin System Physics 30 7-Sep-004 4- Example A Spin System In the last lecture, we discussed the binomial distribution. Now, I would like to add a little physical content by considering a spin system. Actually this

More information

Some Statistics. V. Lindberg. May 16, 2007

Some Statistics. V. Lindberg. May 16, 2007 Some Statistics V. Lindberg May 16, 2007 1 Go here for full details An excellent reference written by physicists with sample programs available is Data Reduction and Error Analysis for the Physical Sciences,

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

2.3 Estimating PDFs and PDF Parameters

2.3 Estimating PDFs and PDF Parameters .3 Estimating PDFs and PDF Parameters estimating means - discrete and continuous estimating variance using a known mean estimating variance with an estimated mean estimating a discrete pdf estimating a

More information

Statistics, Data Analysis, and Simulation SS 2013

Statistics, Data Analysis, and Simulation SS 2013 Statistics, Data Analysis, and Simulation SS 213 8.128.73 Statistik, Datenanalyse und Simulation Dr. Michael O. Distler Mainz, 23. April 213 What we ve learned so far Fundamental

More information

PHYSICS 2150 LABORATORY

PHYSICS 2150 LABORATORY PHYSICS 2150 LABORATORY Professor John Cumalat TAs: Adam Green John Houlton Lab Coordinator: Scott Pinegar Lecture 6 Feb. 17, 2015 ANNOUNCEMENT The problem set will be posted on the course website or you

More information

STATISTICS OF OBSERVATIONS & SAMPLING THEORY. Parent Distributions

STATISTICS OF OBSERVATIONS & SAMPLING THEORY. Parent Distributions ASTR 511/O Connell Lec 6 1 STATISTICS OF OBSERVATIONS & SAMPLING THEORY References: Bevington Data Reduction & Error Analysis for the Physical Sciences LLM: Appendix B Warning: the introductory literature

More information

1 Some Statistical Basics.

1 Some Statistical Basics. Q Some Statistical Basics. Statistics treats random errors. (There are also systematic errors e.g., if your watch is 5 minutes fast, you will always get the wrong time, but it won t be random.) The two

More information

MASSACHUSETTS INSTITUTE OF TECHNOLOGY PHYSICS DEPARTMENT

MASSACHUSETTS INSTITUTE OF TECHNOLOGY PHYSICS DEPARTMENT G. Clark 7oct96 1 MASSACHUSETTS INSTITUTE OF TECHNOLOGY PHYSICS DEPARTMENT 8.13/8.14 Junior Laboratory STATISTICS AND ERROR ESTIMATION The purpose of this note is to explain the application of statistics

More information

Lecture 2: Discrete Probability Distributions

Lecture 2: Discrete Probability Distributions Lecture 2: Discrete Probability Distributions IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge February 1st, 2011 Rasmussen (CUED) Lecture

More information

Probability Distributions - Lecture 5

Probability Distributions - Lecture 5 Probability Distributions - Lecture 5 1 Introduction There are a number of mathematical models of probability density functions that represent the behavior of physical systems. In this lecture we explore

More information

26, 24, 26, 28, 23, 23, 25, 24, 26, 25

26, 24, 26, 28, 23, 23, 25, 24, 26, 25 The ormal Distribution Introduction Chapter 5 in the text constitutes the theoretical heart of the subject of error analysis. We start by envisioning a series of experimental measurements of a quantity.

More information

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit Statistics Lent Term 2015 Prof. Mark Thomson Lecture 2 : The Gaussian Limit Prof. M.A. Thomson Lent Term 2015 29 Lecture Lecture Lecture Lecture 1: Back to basics Introduction, Probability distribution

More information

Confidence Intervals

Confidence Intervals Quantitative Foundations Project 3 Instructor: Linwei Wang Confidence Intervals Contents 1 Introduction 3 1.1 Warning....................................... 3 1.2 Goals of Statistics..................................

More information

Central limit theorem. Paninski, Intro. Math. Stats., October 5, probability, Z N P Z, if

Central limit theorem. Paninski, Intro. Math. Stats., October 5, probability, Z N P Z, if Paninski, Intro. Math. Stats., October 5, 2005 35 probability, Z P Z, if P ( Z Z > ɛ) 0 as. (The weak LL is called weak because it asserts convergence in probability, which turns out to be a somewhat weak

More information

Statistical Methods for Astronomy

Statistical Methods for Astronomy Statistical Methods for Astronomy Probability (Lecture 1) Statistics (Lecture 2) Why do we need statistics? Useful Statistics Definitions Error Analysis Probability distributions Error Propagation Binomial

More information

Chapter 4. Repeated Trials. 4.1 Introduction. 4.2 Bernoulli Trials

Chapter 4. Repeated Trials. 4.1 Introduction. 4.2 Bernoulli Trials Chapter 4 Repeated Trials 4.1 Introduction Repeated indepentent trials in which there can be only two outcomes are called Bernoulli trials in honor of James Bernoulli (1654-1705). As we shall see, Bernoulli

More information

Introduction to Error Analysis

Introduction to Error Analysis Introduction to Error Analysis Part 1: the Basics Andrei Gritsan based on lectures by Petar Maksimović February 1, 2010 Overview Definitions Reporting results and rounding Accuracy vs precision systematic

More information

Statistics 100A Homework 5 Solutions

Statistics 100A Homework 5 Solutions Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to

More information

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables ECE 6010 Lecture 1 Introduction; Review of Random Variables Readings from G&S: Chapter 1. Section 2.1, Section 2.3, Section 2.4, Section 3.1, Section 3.2, Section 3.5, Section 4.1, Section 4.2, Section

More information

Fourier and Stats / Astro Stats and Measurement : Stats Notes

Fourier and Stats / Astro Stats and Measurement : Stats Notes Fourier and Stats / Astro Stats and Measurement : Stats Notes Andy Lawrence, University of Edinburgh Autumn 2013 1 Probabilities, distributions, and errors Laplace once said Probability theory is nothing

More information

Statistical Methods in Particle Physics. Lecture 2

Statistical Methods in Particle Physics. Lecture 2 Statistical Methods in Particle Physics Lecture 2 October 17, 2011 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2011 / 12 Outline Probability Definition and interpretation Kolmogorov's

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Introduction to Statistical Data Analysis Lecture 4: Sampling

Introduction to Statistical Data Analysis Lecture 4: Sampling Introduction to Statistical Data Analysis Lecture 4: Sampling James V. Lambers Department of Mathematics The University of Southern Mississippi James V. Lambers Statistical Data Analysis 1 / 30 Introduction

More information

Practical Statistics

Practical Statistics Practical Statistics Lecture 1 (Nov. 9): - Correlation - Hypothesis Testing Lecture 2 (Nov. 16): - Error Estimation - Bayesian Analysis - Rejecting Outliers Lecture 3 (Nov. 18) - Monte Carlo Modeling -

More information

MA131 - Analysis 1. Workbook 4 Sequences III

MA131 - Analysis 1. Workbook 4 Sequences III MA3 - Analysis Workbook 4 Sequences III Autumn 2004 Contents 2.3 Roots................................. 2.4 Powers................................. 3 2.5 * Application - Factorials *.....................

More information

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm

Midterm Examination. STA 215: Statistical Inference. Due Wednesday, 2006 Mar 8, 1:15 pm Midterm Examination STA 215: Statistical Inference Due Wednesday, 2006 Mar 8, 1:15 pm This is an open-book take-home examination. You may work on it during any consecutive 24-hour period you like; please

More information

Maximum-Likelihood fitting

Maximum-Likelihood fitting CMP 0b Lecture F. Sigworth Maximum-Likelihood fitting One of the issues I want to address in this lecture is the fitting of distributions dwell times. We want to find the best curve to draw over a histogram,

More information

Math Bootcamp 2012 Miscellaneous

Math Bootcamp 2012 Miscellaneous Math Bootcamp 202 Miscellaneous Factorial, combination and permutation The factorial of a positive integer n denoted by n!, is the product of all positive integers less than or equal to n. Define 0! =.

More information

Special distributions

Special distributions Special distributions August 22, 2017 STAT 101 Class 4 Slide 1 Outline of Topics 1 Motivation 2 Bernoulli and binomial 3 Poisson 4 Uniform 5 Exponential 6 Normal STAT 101 Class 4 Slide 2 What distributions

More information

1 Measurement Uncertainties

1 Measurement Uncertainties 1 Measurement Uncertainties (Adapted stolen, really from work by Amin Jaziri) 1.1 Introduction No measurement can be perfectly certain. No measuring device is infinitely sensitive or infinitely precise.

More information

Fitting a Straight Line to Data

Fitting a Straight Line to Data Fitting a Straight Line to Data Thanks for your patience. Finally we ll take a shot at real data! The data set in question is baryonic Tully-Fisher data from http://astroweb.cwru.edu/sparc/btfr Lelli2016a.mrt,

More information

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg Statistics for Data Analysis PSI Practical Course 2014 Niklaus Berger Physics Institute, University of Heidelberg Overview You are going to perform a data analysis: Compare measured distributions to theoretical

More information

Modern Methods of Data Analysis - WS 07/08

Modern Methods of Data Analysis - WS 07/08 Modern Methods of Data Analysis Lecture V (12.11.07) Contents: Central Limit Theorem Uncertainties: concepts, propagation and properties Central Limit Theorem Consider the sum X of n independent variables,

More information

1.1.1 Algebraic Operations

1.1.1 Algebraic Operations 1.1.1 Algebraic Operations We need to learn how our basic algebraic operations interact. When confronted with many operations, we follow the order of operations: Parentheses Exponentials Multiplication

More information

Chapter 11 - Sequences and Series

Chapter 11 - Sequences and Series Calculus and Analytic Geometry II Chapter - Sequences and Series. Sequences Definition. A sequence is a list of numbers written in a definite order, We call a n the general term of the sequence. {a, a

More information

Statistics, Data Analysis, and Simulation SS 2017

Statistics, Data Analysis, and Simulation SS 2017 Statistics, Data Analysis, and Simulation SS 2017 08.128.730 Statistik, Datenanalyse und Simulation Dr. Michael O. Distler Mainz, 27. April 2017 Dr. Michael O. Distler

More information

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions

Introduction to Statistical Data Analysis Lecture 3: Probability Distributions Introduction to Statistical Data Analysis Lecture 3: Probability Distributions James V. Lambers Department of Mathematics The University of Southern Mississippi James V. Lambers Statistical Data Analysis

More information

Primer on statistics:

Primer on statistics: Primer on statistics: MLE, Confidence Intervals, and Hypothesis Testing ryan.reece@gmail.com http://rreece.github.io/ Insight Data Science - AI Fellows Workshop Feb 16, 018 Outline 1. Maximum likelihood

More information

Uncertainty and Graphical Analysis

Uncertainty and Graphical Analysis Uncertainty and Graphical Analysis Introduction Two measures of the quality of an experimental result are its accuracy and its precision. An accurate result is consistent with some ideal, true value, perhaps

More information

Take the measurement of a person's height as an example. Assuming that her height has been determined to be 5' 8", how accurate is our result?

Take the measurement of a person's height as an example. Assuming that her height has been determined to be 5' 8, how accurate is our result? Error Analysis Introduction The knowledge we have of the physical world is obtained by doing experiments and making measurements. It is important to understand how to express such data and how to analyze

More information

First-Order Differential Equations

First-Order Differential Equations CHAPTER 1 First-Order Differential Equations 1. Diff Eqns and Math Models Know what it means for a function to be a solution to a differential equation. In order to figure out if y = y(x) is a solution

More information

Statistics and Data Analysis

Statistics and Data Analysis Statistics and Data Analysis The Crash Course Physics 226, Fall 2013 "There are three kinds of lies: lies, damned lies, and statistics. Mark Twain, allegedly after Benjamin Disraeli Statistics and Data

More information

Week 2: Review of probability and statistics

Week 2: Review of probability and statistics Week 2: Review of probability and statistics Marcelo Coca Perraillon University of Colorado Anschutz Medical Campus Health Services Research Methods I HSMP 7607 2017 c 2017 PERRAILLON ALL RIGHTS RESERVED

More information

Chapter 13 - Inverse Functions

Chapter 13 - Inverse Functions Chapter 13 - Inverse Functions In the second part of this book on Calculus, we shall be devoting our study to another type of function, the exponential function and its close relative the Sine function.

More information

Statistics, Data Analysis, and Simulation SS 2015

Statistics, Data Analysis, and Simulation SS 2015 Statistics, Data Analysis, and Simulation SS 2015 08.128.730 Statistik, Datenanalyse und Simulation Dr. Michael O. Distler Mainz, 27. April 2015 Dr. Michael O. Distler

More information

Line Broadening. φ(ν) = Γ/4π 2 (ν ν 0 ) 2 + (Γ/4π) 2, (3) where now Γ = γ +2ν col includes contributions from both natural broadening and collisions.

Line Broadening. φ(ν) = Γ/4π 2 (ν ν 0 ) 2 + (Γ/4π) 2, (3) where now Γ = γ +2ν col includes contributions from both natural broadening and collisions. Line Broadening Spectral lines are not arbitrarily sharp. There are a variety of mechanisms that give them finite width, and some of those mechanisms contain significant information. We ll consider a few

More information

Probability Method in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur

Probability Method in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur Probability Method in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur Lecture No. # 34 Probability Models using Discrete Probability Distributions

More information

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS

POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS POISSON PROCESSES 1. THE LAW OF SMALL NUMBERS 1.1. The Rutherford-Chadwick-Ellis Experiment. About 90 years ago Ernest Rutherford and his collaborators at the Cavendish Laboratory in Cambridge conducted

More information

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory Chapter 1 Statistical Reasoning Why statistics? Uncertainty of nature (weather, earth movement, etc. ) Uncertainty in observation/sampling/measurement Variability of human operation/error imperfection

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics Lecture 3 October 29, 2012 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2012 / 13 Outline Reminder: Probability density function Cumulative

More information

04. Random Variables: Concepts

04. Random Variables: Concepts University of Rhode Island DigitalCommons@URI Nonequilibrium Statistical Physics Physics Course Materials 215 4. Random Variables: Concepts Gerhard Müller University of Rhode Island, gmuller@uri.edu Creative

More information

CS 361: Probability & Statistics

CS 361: Probability & Statistics February 26, 2018 CS 361: Probability & Statistics Random variables The discrete uniform distribution If every value of a discrete random variable has the same probability, then its distribution is called

More information

1 Measurement Uncertainties

1 Measurement Uncertainties 1 Measurement Uncertainties (Adapted stolen, really from work by Amin Jaziri) 1.1 Introduction No measurement can be perfectly certain. No measuring device is infinitely sensitive or infinitely precise.

More information

Poisson Statistics. Department of Physics University of Cape Town Course III Laboratory January 13, 2015

Poisson Statistics. Department of Physics University of Cape Town Course III Laboratory January 13, 2015 Poisson Statistics Department of Physics University of Cape Town Course III Laboratory January 13, 2015 Abstract The goal of a physics experiment is the extraction of physically meaningful parameters from

More information

University of California, Berkeley, Statistics 134: Concepts of Probability. Michael Lugo, Spring Exam 1

University of California, Berkeley, Statistics 134: Concepts of Probability. Michael Lugo, Spring Exam 1 University of California, Berkeley, Statistics 134: Concepts of Probability Michael Lugo, Spring 2011 Exam 1 February 16, 2011, 11:10 am - 12:00 noon Name: Solutions Student ID: This exam consists of seven

More information

T has many other desirable properties, and we will return to this example

T has many other desirable properties, and we will return to this example 2. Introduction to statistics: first examples 2.1. Introduction. The basic problem of statistics is to draw conclusions about unknown distributions of random variables from observed values. These conclusions

More information

Solving with Absolute Value

Solving with Absolute Value Solving with Absolute Value Who knew two little lines could cause so much trouble? Ask someone to solve the equation 3x 2 = 7 and they ll say No problem! Add just two little lines, and ask them to solve

More information

18.175: Lecture 13 Infinite divisibility and Lévy processes

18.175: Lecture 13 Infinite divisibility and Lévy processes 18.175 Lecture 13 18.175: Lecture 13 Infinite divisibility and Lévy processes Scott Sheffield MIT Outline Poisson random variable convergence Extend CLT idea to stable random variables Infinite divisibility

More information

Central Limit Theorem and the Law of Large Numbers Class 6, Jeremy Orloff and Jonathan Bloom

Central Limit Theorem and the Law of Large Numbers Class 6, Jeremy Orloff and Jonathan Bloom Central Limit Theorem and the Law of Large Numbers Class 6, 8.5 Jeremy Orloff and Jonathan Bloom Learning Goals. Understand the statement of the law of large numbers. 2. Understand the statement of the

More information

Section 5.4. Ken Ueda

Section 5.4. Ken Ueda Section 5.4 Ken Ueda Students seem to think that being graded on a curve is a positive thing. I took lasers 101 at Cornell and got a 92 on the exam. The average was a 93. I ended up with a C on the test.

More information

Why is the field of statistics still an active one?

Why is the field of statistics still an active one? Why is the field of statistics still an active one? It s obvious that one needs statistics: to describe experimental data in a compact way, to compare datasets, to ask whether data are consistent with

More information

Random Networks. Complex Networks CSYS/MATH 303, Spring, Prof. Peter Dodds

Random Networks. Complex Networks CSYS/MATH 303, Spring, Prof. Peter Dodds Complex Networks CSYS/MATH 303, Spring, 2011 Prof. Peter Dodds Department of Mathematics & Statistics Center for Complex Systems Vermont Advanced Computing Center University of Vermont Licensed under the

More information

Introduction to Statistics and Error Analysis II

Introduction to Statistics and Error Analysis II Introduction to Statistics and Error Analysis II Physics116C, 4/14/06 D. Pellett References: Data Reduction and Error Analysis for the Physical Sciences by Bevington and Robinson Particle Data Group notes

More information

Lecture 8 Hypothesis Testing

Lecture 8 Hypothesis Testing Lecture 8 Hypothesis Testing Taylor Ch. 6 and 10.6 Introduction l The goal of hypothesis testing is to set up a procedure(s) to allow us to decide if a mathematical model ("theory") is acceptable in light

More information

Numerical Methods Lecture 7 - Statistics, Probability and Reliability

Numerical Methods Lecture 7 - Statistics, Probability and Reliability Topics Numerical Methods Lecture 7 - Statistics, Probability and Reliability A summary of statistical analysis A summary of probability methods A summary of reliability analysis concepts Statistical Analysis

More information

Achilles: Now I know how powerful computers are going to become!

Achilles: Now I know how powerful computers are going to become! A Sigmoid Dialogue By Anders Sandberg Achilles: Now I know how powerful computers are going to become! Tortoise: How? Achilles: I did curve fitting to Moore s law. I know you are going to object that technological

More information

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur

Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Lecture No. # 36 Sampling Distribution and Parameter Estimation

More information

Math 50: Final. 1. [13 points] It was found that 35 out of 300 famous people have the star sign Sagittarius.

Math 50: Final. 1. [13 points] It was found that 35 out of 300 famous people have the star sign Sagittarius. Math 50: Final 180 minutes, 140 points. No algebra-capable calculators. Try to use your calculator only at the end of your calculation, and show working/reasoning. Please do look up z, t, χ 2 values for

More information

Quantum Mechanics- I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras

Quantum Mechanics- I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras Quantum Mechanics- I Prof. Dr. S. Lakshmi Bala Department of Physics Indian Institute of Technology, Madras Lecture - 6 Postulates of Quantum Mechanics II (Refer Slide Time: 00:07) In my last lecture,

More information

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Random variables. DS GA 1002 Probability and Statistics for Data Science. Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities

More information

Generating Functions

Generating Functions Generating Functions Karen Ge May, 07 Abstract Generating functions gives us a global perspective when we need to study a local property. We define generating functions and present its applications in

More information

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems

More information

California State Science Fair

California State Science Fair California State Science Fair How to Estimate the Experimental Uncertainty in Your Science Fair Project Part 2 -- The Gaussian Distribution: What the Heck is it Good For Anyway? Edward Ruth drruth6617@aol.com

More information

Conditional distributions (discrete case)

Conditional distributions (discrete case) Conditional distributions (discrete case) The basic idea behind conditional distributions is simple: Suppose (XY) is a jointly-distributed random vector with a discrete joint distribution. Then we can

More information

Sequences and infinite series

Sequences and infinite series Sequences and infinite series D. DeTurck University of Pennsylvania March 29, 208 D. DeTurck Math 04 002 208A: Sequence and series / 54 Sequences The lists of numbers you generate using a numerical method

More information

Discrete Binary Distributions

Discrete Binary Distributions Discrete Binary Distributions Carl Edward Rasmussen November th, 26 Carl Edward Rasmussen Discrete Binary Distributions November th, 26 / 5 Key concepts Bernoulli: probabilities over binary variables Binomial:

More information

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0.

CSE 103 Homework 8: Solutions November 30, var(x) = np(1 p) = P r( X ) 0.95 P r( X ) 0. () () a. X is a binomial distribution with n = 000, p = /6 b. The expected value, variance, and standard deviation of X is: E(X) = np = 000 = 000 6 var(x) = np( p) = 000 5 6 666 stdev(x) = np( p) = 000

More information

Physics 403. Segev BenZvi. Credible Intervals, Confidence Intervals, and Limits. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Credible Intervals, Confidence Intervals, and Limits. Department of Physics and Astronomy University of Rochester Physics 403 Credible Intervals, Confidence Intervals, and Limits Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Summarizing Parameters with a Range Bayesian

More information

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming

More information

EXPERIMENTAL UNCERTAINTY

EXPERIMENTAL UNCERTAINTY 3 EXPERIMENTAL UNCERTAINTY I am no matchmaker, as you well know, said Lady Russell, being much too aware of the uncertainty of all human events and calculations. --- Persuasion 3.1 UNCERTAINTY AS A 95%

More information

Basic counting techniques. Periklis A. Papakonstantinou Rutgers Business School

Basic counting techniques. Periklis A. Papakonstantinou Rutgers Business School Basic counting techniques Periklis A. Papakonstantinou Rutgers Business School i LECTURE NOTES IN Elementary counting methods Periklis A. Papakonstantinou MSIS, Rutgers Business School ALL RIGHTS RESERVED

More information

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO LECTURE NOTES FYS 4550/FYS9550 - EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I PROBABILITY AND STATISTICS A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO Before embarking on the concept

More information

Statistics of Radioactive Decay

Statistics of Radioactive Decay Statistics of Radioactive Decay Introduction The purpose of this experiment is to analyze a set of data that contains natural variability from sample to sample, but for which the probability distribution

More information

PHY 123 Lab 1 - Error and Uncertainty and the Simple Pendulum

PHY 123 Lab 1 - Error and Uncertainty and the Simple Pendulum To print higher-resolution math symbols, click the Hi-Res Fonts for Printing button on the jsmath control panel. PHY 13 Lab 1 - Error and Uncertainty and the Simple Pendulum Important: You need to print

More information

Parameter estimation Conditional risk

Parameter estimation Conditional risk Parameter estimation Conditional risk Formalizing the problem Specify random variables we care about e.g., Commute Time e.g., Heights of buildings in a city We might then pick a particular distribution

More information

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Experiment 2 Random Error and Basic Statistics

Experiment 2 Random Error and Basic Statistics PHY9 Experiment 2: Random Error and Basic Statistics 8/5/2006 Page Experiment 2 Random Error and Basic Statistics Homework 2: Turn in at start of experiment. Readings: Taylor chapter 4: introduction, sections

More information

Error propagation. Alexander Khanov. October 4, PHYS6260: Experimental Methods is HEP Oklahoma State University

Error propagation. Alexander Khanov. October 4, PHYS6260: Experimental Methods is HEP Oklahoma State University Error propagation Alexander Khanov PHYS660: Experimental Methods is HEP Oklahoma State University October 4, 017 Why error propagation? In many cases we measure one thing and want to know something else

More information

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1 Lecture 5 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Week 4: Chap. 3 Statistics of Radioactivity

Week 4: Chap. 3 Statistics of Radioactivity Week 4: Chap. 3 Statistics of Radioactivity Vacuum Technology General use of Statistical Distributions in Radiation Measurements -- Fluctuations in Number --- distribution function models -- Fluctuations

More information

CME 106: Review Probability theory

CME 106: Review Probability theory : Probability theory Sven Schmit April 3, 2015 1 Overview In the first half of the course, we covered topics from probability theory. The difference between statistics and probability theory is the following:

More information

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER /2018

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER /2018 ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER 1 2017/2018 DR. ANTHONY BROWN 1. Arithmetic and Algebra 1.1. Arithmetic of Numbers. While we have calculators and computers

More information

Originality in the Arts and Sciences: Lecture 2: Probability and Statistics

Originality in the Arts and Sciences: Lecture 2: Probability and Statistics Originality in the Arts and Sciences: Lecture 2: Probability and Statistics Let s face it. Statistics has a really bad reputation. Why? 1. It is boring. 2. It doesn t make a lot of sense. Actually, the

More information