Lecture 1 Probability and Statistics

Similar documents
Lecture 1 Probability and Statistics

Physics 3700 Probability, Statistics, & Data Analysis

FACULTY OF MATHEMATICAL STUDIES MATHEMATICS FOR PART I ENGINEERING. Lectures

CEE 522 Autumn Uncertainty Concepts for Geotechnical Engineering

IE 230 Seat # Name < KEY > Please read these directions. Closed book and notes. 60 minutes.

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Quick Review of Probability

Random Variables, Sampling and Estimation

Quick Review of Probability

Expectation and Variance of a random variable

Lecture 6 Chi Square Distribution (χ 2 ) and Least Squares Fitting

Discrete Mathematics for CS Spring 2008 David Wagner Note 22

It is always the case that unions, intersections, complements, and set differences are preserved by the inverse image of a function.

Probability and statistics: basic terms

Understanding Samples

Chapter 6 Sampling Distributions

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36

Lecture 4. Random variable and distribution of probability

Hypothesis Testing. Evaluation of Performance of Learned h. Issues. Trade-off Between Bias and Variance

Lecture 5. Random variable and distribution of probability

The standard deviation of the mean

Lecture 7: Properties of Random Samples

DS 100: Principles and Techniques of Data Science Date: April 13, Discussion #10

Variance of Discrete Random Variables Class 5, Jeremy Orloff and Jonathan Bloom

Final Review for MATH 3510

Topic 9: Sampling Distributions of Estimators

Lecture 2: Monte Carlo Simulation

Analysis of Experimental Data

NO! This is not evidence in favor of ESP. We are rejecting the (null) hypothesis that the results are

f X (12) = Pr(X = 12) = Pr({(6, 6)}) = 1/36

Topic 9: Sampling Distributions of Estimators

Chapter 6 Principles of Data Reduction

Topic 9: Sampling Distributions of Estimators

MATH 320: Probability and Statistics 9. Estimation and Testing of Parameters. Readings: Pruim, Chapter 4

PH 425 Quantum Measurement and Spin Winter SPINS Lab 1

Analysis of Experimental Measurements

Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Math 152. Rumbos Fall Solutions to Review Problems for Exam #2. Number of Heads Frequency

Discrete Mathematics for CS Spring 2005 Clancy/Wagner Notes 21. Some Important Distributions

Probability, Expectation Value and Uncertainty

Poisson approximations

Lecture 19: Convergence

Statisticians use the word population to refer the total number of (potential) observations under consideration

n outcome is (+1,+1, 1,..., 1). Let the r.v. X denote our position (relative to our starting point 0) after n moves. Thus X = X 1 + X 2 + +X n,

Discrete Mathematics for CS Spring 2007 Luca Trevisan Lecture 22

Approximations and more PMFs and PDFs

Infinite Sequences and Series

STA Learning Objectives. Population Proportions. Module 10 Comparing Two Proportions. Upon completing this module, you should be able to:

AMS570 Lecture Notes #2

4. Partial Sums and the Central Limit Theorem

Introduction to probability Stochastic Process Queuing systems. TELE4642: Week2

4. Basic probability theory

Note: we can take the Real and Imaginary part of the Schrödinger equation and write it in a way similar to the electromagnetic field. p = n!

ANALYSIS OF EXPERIMENTAL ERRORS

As stated by Laplace, Probability is common sense reduced to calculation.

Data Analysis and Statistical Methods Statistics 651

Massachusetts Institute of Technology


Advanced Stochastic Processes.

WHAT IS THE PROBABILITY FUNCTION FOR LARGE TSUNAMI WAVES? ABSTRACT

Lecture 6 Simple alternatives and the Neyman-Pearson lemma

Closed book and notes. No calculators. 60 minutes, but essentially unlimited time.

Statistics 511 Additional Materials

The variance of a sum of independent variables is the sum of their variances, since covariances are zero. Therefore. V (xi )= n n 2 σ2 = σ2.

Discrete Mathematics and Probability Theory Summer 2014 James Cook Note 15

Lecture Chapter 6: Convergence of Random Sequences

CS 330 Discussion - Probability

Binomial Distribution

Let us give one more example of MLE. Example 3. The uniform distribution U[0, θ] on the interval [0, θ] has p.d.f.

MASSACHUSETTS INSTITUTE OF TECHNOLOGY 6.436J/15.085J Fall 2008 Lecture 19 11/17/2008 LAWS OF LARGE NUMBERS II THE STRONG LAW OF LARGE NUMBERS

Limit Theorems. Convergence in Probability. Let X be the number of heads observed in n tosses. Then, E[X] = np and Var[X] = np(1-p).

STAT Homework 1 - Solutions

ACCESS TO SCIENCE, ENGINEERING AND AGRICULTURE: MATHEMATICS 1 MATH00030 SEMESTER / Statistics

Stat 319 Theory of Statistics (2) Exercises

Introduction to Probability I: Expectations, Bayes Theorem, Gaussians, and the Poisson Distribution. 1

Sample questions. 8. Let X denote a continuous random variable with probability density function f(x) = 4x 3 /15 for

Stat 421-SP2012 Interval Estimation Section

Empirical Distributions

Sample Size Estimation in the Proportional Hazards Model for K-sample or Regression Settings Scott S. Emerson, M.D., Ph.D.

Module 1 Fundamentals in statistics

EE 4TM4: Digital Communications II Probability Theory

5. Likelihood Ratio Tests

Computing Confidence Intervals for Sample Data

Parameter, Statistic and Random Samples

Discrete Mathematics and Probability Theory Spring 2012 Alistair Sinclair Note 15

Activity 3: Length Measurements with the Four-Sided Meter Stick

7.1 Convergence of sequences of random variables

Chapter 18 Summary Sampling Distribution Models

CH.25 Discrete Random Variables

STATS 200: Introduction to Statistical Inference. Lecture 1: Course introduction and polling

Introduction to Probability and Statistics Twelfth Edition

Lecture 2: Poisson Sta*s*cs Probability Density Func*ons Expecta*on and Variance Es*mators

Exponential Families and Bayesian Inference

DISTRIBUTION LAW Okunev I.V.

11 Correlation and Regression

Chapter 22. Comparing Two Proportions. Copyright 2010 Pearson Education, Inc.

Lesson 10: Limits and Continuity

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

Distribution of Random Samples & Limit theorems

NOTES ON DISTRIBUTIONS

Transcription:

Wikipedia: Lecture 1 Probability ad Statistics Bejami Disraeli, British statesma ad literary figure (1804 1881): There are three kids of lies: lies, damed lies, ad statistics. popularized i US by Mark Twai the statemet shows the persuasive power of umbers use of statistics to bolster weak argumets tedecy of people to disparage statistics that do ot support their positios The purpose of P3700: how to uderstad the statistical ucertaity of observatio/measuremet how to use statistics to argue agaist a weak argumet (or bolster a weak argumet?) how to argue agaist people disparagig statistics that do ot support their positios how to lie with statistics? K.K. Ga L1: Probability ad Statistics 1

Why there is statistical ucertaity? You sell 7 cryogeic equipmet last moth You kow how to cout ad 7 is the exact umber of equipmet sold there is o ucertaity o 7! However if you used the statistics of 7 to predict the future sale or compare with past sale there is a ucertaity o 7 the umber of equipmet sold could be 5, 8, or 10! must iclude the ucertaity i the calculatio What is the ucertaity o 7? Lecture 2: 7 = 2.6 there is a 68% chace that the expected umber of equipmet sold per moth is 4.4-9.6 However the umber of equipmet sold per moth is a discrete umber there is a ~68% chace that the expected umber of equipmet sold per moth is 4-10 should use Poisso statistics as i Lecture 2 for more precise predictio K.K. Ga L1: Probability ad Statistics 2

Itroductio: Uderstadig of may physical pheomea deped o statistical ad probabilistic cocepts: Statistical Mechaics (physics of systems composed of may parts: gases, liquids, solids.) 1 mole of aythig cotais 6x10 23 particles (Avogadro's umber) impossible to keep track of all 6x10 23 particles eve with the fastest computer imagiable resort to learig about the group properties of all the particles partitio fuctio: calculate eergy, etropy, pressure... of a system Quatum Mechaics (physics at the atomic or smaller scale) wavefuctio = probability amplitude probability of a electro beig located at (x,y,z) at a certai time. Uderstadig/iterpretatio of experimetal data deped o statistical ad probabilistic cocepts: how do we extract the best value of a quatity from a set of measuremets? how do we decide if our experimet is cosistet/icosistet with a give theory? how do we decide if our experimet is iterally cosistet? how do we decide if our experimet is cosistet with other experimets? I this course we will cocetrate o the above experimetal issues! K.K. Ga L1: Probability ad Statistics 3

Defiitio of probability: Suppose we have N trials ad a specified evet occurs r times. example: rollig a dice ad the evet could be rollig a 6. defie probability (P) of a evet (E) occurrig as: P(E) = r/n whe N examples: six sided dice: P(6) = 1/6 coi toss: P(heads) = 0.5 P(heads) should approach 0.5 the more times you toss the coi. for a sigle coi toss we ca ever get P(heads) = 0.5! by defiitio probability is a o-egative real umber bouded by 0 P 1 if P = 0 the the evet ever occurs if P = 1 the the evet always occurs sum (or itegral) of all probabilities if they are mutually exclusive must = 1. evets are idepedet if: P(AÇB) = P(A)P(B) Ǻitersectio, Ⱥ uio coi tosses are idepedet evets, the result of ext toss does ot deped o previous toss. evets are mutually exclusive (disjoit) if: P(AÇB) = 0 or P(AÈB) = P(A) + P(B) i coi tossig, we either get a head or a tail. K.K. Ga L1: Probability ad Statistics 4

Probability ca be a discrete or a cotiuous variable. Discrete probability: P ca have certai values oly. examples: tossig a six-sided dice: P(x i ) = P i here x i = 1, 2, 3, 4, 5, 6 ad P i = 1/6 for all x i. tossig a coi: oly 2 choices, heads or tails. for both of the above discrete examples (ad i geeral) whe we sum over all mutually exclusive possibilities: P( x i ) =1 Cotiuous probability: P ca be ay umber betwee 0 ad 1. defie a probability desity fuctio, pdf, f ( x) f ( x)dx = dp( x α x + dx) with a cotiuous variable probability for x to be i the rage a x b is: just like the discrete case the sum of all probabilities must equal 1. + f x dx =1 i P(a x b) = ( ) b a ( ) f x dx f(x) is ormalized to oe. probability for x to be exactly some umber is zero sice: x=a ( ) f x dx = 0 x=a Notatio: x i is called a radom variable K.K. Ga L1: Probability ad Statistics 5

Examples of some commo P(x) s ad f(x) s: Discrete = P(x) biomial Poisso Cotiuous = f(x) uiform, i.e. costat Gaussia expoetial chi square How do we describe a probability distributio? mea, mode, media, ad variace for a cotiuous distributio, these quatities are defied by: Mea Mode Media Variace average most probable 50% poit width of distributio + a + f x µ = xf (x)dx = 0 0.5 = f (x)dx σ 2 = f (x) x µ ( ) x x = a ( ) 2 dx for a discrete distributio, the mea ad variace are defied by: µ = 1 x i σ 2 = 1 (x i µ) 2 K.K. Ga L1: Probability ad Statistics 6

Some cotiuous pdf: Probability is the area uder the curves! mode media mea σ symmetric distributio (gaussia) For a Gaussia pdf, the mea, mode, ad media are all at the same x. Asymmetric distributio showig the mea, media ad mode For most pdfs, the mea, mode, ad media are at differet locatios. K.K. Ga L1: Probability ad Statistics 7

Calculatio of mea ad variace: example: a discrete data set cosistig of three umbers: {1, 2, 3} average (µ) is just: x µ = i 1+ 2+ 3 = = 2 3 complicatio: suppose some measuremet are more precise tha others. if each measuremet x i have a weight w i associated with it: µ = x i w i / w i variace (s 2 ) or average squared deviatio from the mea is just: σ 2 = 1 (x i µ) 2 variace describes the width of the pdf! s is called the stadard deviatio rewrite the above expressio by expadig the summatios: σ 2 = 1 % x 2 i + µ 2 ( ' 2µ x i * & ) = 1 = 1 x 2 i + µ 2 2µ 2 x 2 i µ 2 = x 2 x 2 < > º average weighted average i the deomiator would be -1 if we determied the average (µ) from the data itself. K.K. Ga L1: Probability ad Statistics 8

usig the defiitio of µ from above we have for our example of {1,2,3}: σ 2 = 1 x 2 i µ 2 = 4.67 2 2 = 0.67 the case where the measuremets have differet weights is more complicated: σ 2 = µ is the weighted mea w i (x i µ) 2 2 / w i = w i x i / w i µ 2 if we calculated µ from the data, s 2 gets multiplied by a factor /(-1). example: a cotiuous probability distributio, f (x) = si 2 x for 0 x 2π has two modes! has same mea ad media, but differ from the mode(s). f(x) is ot properly ormalized: ormalized pdf: f (x) = si 2 x / 2π 0 2π 0 si 2 xdx = π 1 si 2 xdx = 1 π si2 x K.K. Ga L1: Probability ad Statistics 9

for cotiuous probability distributios, the mea, mode, ad media are calculated usig either itegrals or derivatives: µ = 1 π 2π 0 x si 2 xdx = π mode : x si2 x = 0 π 2, 3π 2 α media : 1 π si2 xdx = 1 0 2 α = π example: Gaussia distributio fuctio, a cotiuous probability distributio p(x) = (x µ ) 1 2 σ 2π e 2σ 2 gaussia K.K. Ga L1: Probability ad Statistics 10

Accuracy ad Precisio: Accuracy: The accuracy of a experimet refers to how close the experimetal measuremet is to the true value of the quatity beig measured. Precisio: This refers to how well the experimetal result has bee determied, without regard to the true value of the quatity beig measured. just because a experimet is precise it does ot mea it is accurate!! accurate but ot precise precise but ot accurate K.K. Ga L1: Probability ad Statistics 11

Measuremet Errors (Ucertaities) Use results from probability ad statistics as a way of idicatig how good a measuremet is. most commo quality idicator: relative precisio = [ucertaity of measuremet]/measuremet example: we measure a table to be 10 iches with ucertaity of 1 ich. relative precisio = 1/10 = 0.1 or 10% (% relative precisio) ucertaity i measuremet is usually square root of variace: s = stadard deviatio usually calculated usig the techique of propagatio of errors (Lecture 4). Statistics ad Systematic Errors Results from experimets are ofte preseted as: N ± XX ± YY N: value of quatity measured (or determied) by experimet. XX: statistical error, usually assumed to be from a Gaussia distributio. with the assumptio of Gaussia statistics we ca say (calculate) somethig about how well our experimet agrees with other experimets ad/or theories. Expect a 68% chace that the true value is betwee N - XX ad N + XX. YY: systematic error. Hard to estimate, distributio of errors usually ot kow. examples: mass of proto = 0.9382769 ± 0.0000027 GeV (oly statistical error give) mass of W boso = 80.8 ± 1.5 ± 2.4 GeV K.K. Ga L1: Probability ad Statistics 12

What s the differece betwee statistical ad systematic errors? N ± XX ± YY statistical errors are radom i the sese that if we repeat the measuremet eough times: XX -> 0 systematic errors do ot -> 0 with repetitio. examples of sources of systematic errors: voltmeter ot calibrated properly a ruler ot the legth we thik is (meter stick might really be < meter!) because of systematic errors, a experimetal result ca be precise, but ot accurate! How do we combie systematic ad statistical errors to get oe estimate of precisio? big problem! two choices: s tot = XX + YY add them liearly s tot = (XX 2 + YY 2 ) 1/2 add them i quadrature widely accepted practice if XX ad YY are ot correlated errors ot of same origi, e.g. from the same voltmeter smaller s tot! Some other ways of quotig experimetal results lower limit: the mass of particle X is > 100 GeV upper limit: the mass of particle X is < 100 GeV +4 asymmetric errors: mass of particle X = 100 3 GeV K.K. Ga L1: Probability ad Statistics 13

How to preset your measured values: Do t quote ay measuremet to more tha three sigificat digits three sigificat digits meas you measure a quatity to 1 part i a thousad or 0.1% precisio difficult to achieve 0.1% precisio acceptable to quote more tha three sigificat digits if you have a large data sample (e.g. large simulatios) Do t quote ay ucertaity to more tha two sigificat digits Measuremet ad ucertaity should have the same umber of digits 991± 57 0.231 ± 0.013 (5.98± 0.43)x10-5 follow this rule i the lab report! K.K. Ga L1: Probability ad Statistics 14