Fourier and Stats / Astro Stats and Measurement : Stats Notes

Size: px
Start display at page:

Download "Fourier and Stats / Astro Stats and Measurement : Stats Notes"

Transcription

1 Fourier and Stats / Astro Stats and Measurement : Stats Notes Andy Lawrence, University of Edinburgh Autumn Probabilities, distributions, and errors Laplace once said Probability theory is nothing but common sense reduced to calculation. This is true, but only if we keep a clear head... Stepping through the basics of probability is laborious, but its worth it, because its all too easy to fall into simple traps when doing probability calculations. 1.1 Probability The concept of probability. There is an ongoing debate over the meaning of probabilities - whether they should be seen as the frequency of occurrence of something - the fraction of times a six is rolled - or the degree of belief in something. To some extent it doesn t matter as long as the quantities obey the same calculus of probabilities (see below), but adherents to these two views ask subtly different questions, as we shall see later. It is easier to follow the mathematical trail if we take the frequentist view for now, and consider the alternative view later. Experiments, outcomes, and events. We can refer to any process of observation or measurement as an experiment - rolling a die, or measuring the mass of an electron. The results obtained are outcomes - a four rolled, or a mass of 509keV. The set of all possible outcomes is called the sample space. For tossing a coin, that is just the two outcomes S = {H, T }. An event is a more general result, typically a combination of the elemental outcomes. For example, suppose we roll two dice. The sample space is the set of all outcomes which is S = {(x, y) x = 1, 2..6, y = 1, 2..6} i.e. the set of all 36 pairs of values x, y such that x can be any of 1 6 and likewise for y. But suppose we are interested in how often we get a given total T. The set of all ways we can achieve the event T = 7 is S 7 = {(1, 6), (2, 5), (3, 4), (4, 3), (5, 2), (6, 1)} This is 6 out of the 36 members of the full set of outcomes, so the probability of getting the event T = 7 is P (T = 7) = 1/6. We have assumed that all the outcomes have equal probability; but the various possible events T do not have equal probability. There is only one way of getting a 12, and no ways of getting 1. 1

2 Stats 1 : Probabilities 2 Figure 1: The difference between exclusive and non-exclusive events. Events A and B are mutually exclusive, whereas events C and D have elemental outcomes in common. 1.2 The calculus of probabilities A probability is a number between 0 and 1. The combined probability of all the possible outcomes in the sample space S could be expressed as P (S) = 1. If there are N possible outcomes and each outcome x has the same probability, then P (x) = 1/N. If the outcomes are not all equally likely, then the probability of a given outcome is the fraction of times that outcome occurs. If an event E could occur n different ways out of the N elemental outcomes, then the probability of E is P (E) = n/n. How do we combine probabilities for multiple events? Everybody remembers that you add probabilities if you can have one or the other, and multiply them if you want both; but in fact you have to be very careful about whether events are exclusive, and whether they are independent. As well as being important for getting probability problems right, it turns out to be important later when we look at statistical inference. Exclusive and non-exclusive events. What is the probability of getting either event A or event B? This depends on whether the events are mutually exclusive i.e. whether they can both occur. Suppose we roll two dice. If event A is a total of 7 and B is a total of 5 then A and B cannot both occur. However if B is the event the two numbers include a 1 then A and B can both occur. Rolling 1 and 6 is an example of both events occurring; rolling 2 and 5 is an example of only A occurring; and rolling 1 and 4 is an example of only B occurring. The situation is illustrated graphically by the Venn diagram in Fig. 1. Events A and B are mutually exclusive, and A B = A + B and so P (A or B) = P (A) + P (B) Events C and D are not exclusive, and C D = C + D C D, and so P (C or D) = P (C) + P (D) P (C and D) Of course this second version is generally true; but for exclusive events P (C and D) = 0.

3 Stats 1 : Probabilities 3 Dependence and conditional probabilities. Suppose we pick balls from an urn. (The favourite hobby of statisticians.) Let us say that it contains two red balls and eight black balls. The probability of picking a red ball is 1/5. If we replace the ball and pick again, the probability of picking a red ball a second time is the same, 1/5. The two events are independent. Suppose however we do not replace the first ball. Then the first pick matters - if it was black, the probability of red in the second pick is 2/9; but if the first pick was red, our probability is 1/9. The two events are dependent, or alternatively we can say that the second event is conditional on the first. In general for two events A and B in a sample space S P (A and B) = P (A, B) = P (A) P (B A) where the expression P (B A) is the conditional probability and can be read as the probability of B given A. If the two events are independent then P (B A) = P (B) where P (B) is just the overall fraction of occurrences of B in our sample space S. Bayes theorem. The conditionality works both ways, so we can write This formula can be re-arranged to say P (A, B) = P (A B).P (B) = P (B A) P (A) P (B A) = P (A B) P (B) P (A) In this form, it is known as Bayes Theorem. As it stands, this is a fairly innocuous re-statement of the laws of probability. However later on when we consider parameter estimation and model fitting, it will have very important consequences. 1.3 Probability distributions Random variables. A random variable is a number associated with the outcomes of chance experiments, whose value is different for each measurement or trial. The total T that we roll with a pair of dice is a random variable, as is the height h of a randomly chosen person. The total rolled with the dice is an example of a discrete random variable. It can only take on particular values. The height of a person is an example of a continuous random variable. It can take an infinite number of possible values, but typically within a restricted range - the height of a person cannot be negative for example. The set of all the possible values of the variable is the sample space we discussed above - whether that is a finite set of discrete values, or a countably infinite set of discrete values, or an uncountably infinite set of real values. Probability distributions. Every value of the random variable has an associated probability. For a discrete variable such as T this is the discrete set of values P (T ) associated with T. This mapping from T to P (T ) is known as a probability distribution. For several events A, B, C there will be a probability associated with all of A, B, and C etc occurring which is then a multivariate probability distribution, P (A, B, C...). For a continous variable such as h any specific value strictly speaking has an infinitesimal probability; we have instead to define the density of probability p(h) within a small region, in the usual way such that p(h)dh is the probability of occurrence of a value of in the range h to h + dh. Any such function f(x) is then a probability density function or PDF. When we come to hypothesis testing, it is often important to know the integrated probability over some range. It is customary to refer to the integrated probability for a value less than some value x, i.e.

4 Stats 1 : Probabilities 4 as the distribution function for x. F (x) = x f(x)dx Just like with discrete events, sometimes we want to define a multivariate PDF. For example if there are two random variables x and y we can define p(x, y) such that p(x, y)dxdy is the probability that in a specific trial the two variables fall in the range x to x + dx and y to y + dy. Following the terminology for discrete probabilities, we can write the probability of x given y as p(x y). You can visualise this if you think of p(x, y) as a two dimensional surface, and take a slice through it along x at a fixed value of y. Likewise if we simply write p(x), this is the probability distribution for x if we allow all values of y, and could be calculated as p(x) = p(x, y)dy. This is referred to as marginalisation over y. If x and y are independent then p(x, y) = p(x)p(y). Finally, we can also express the continuous version of Bayes theorem as p(y x) = p(x y)p(y) p(x) Parent vs sample distributions. The logic of a given physical situation implies a very definite probability distribution. If we roll two dice, we know exactly what proportion of the time we will get a total of 2, what proportion a total of 7, etc. For the heights of people, or the masses of galaxies, the situation is much more complicated, and we may not actually know what the true probability distribution is, but our assumption in the physical sciences is that some such underlying process does exist, even if we do not know what it is. However, when we conduct a series of trials, for example rolling two dice fifty times, the actual observed frequency of 2s and 7s will usually not exactly match the expected values. The intrinsic expected distribution is known as the parent or population distribution. The discrete observed distribution is known as the sample distribution, because the observed values are sampled from the underlying true distribution. One of the key problems of physical science is that often we don t know what the underlying process is, but would like to find out. All we have is the sample distribution, and wish to use this to try to understand what is going on. This is the problem of statistical inference, which will occupy us for parts Characterising sample distributions The parent distribution will be defined by some precise mathematical expression. By contrast the sample distribution is simply a list of values x i. For a situation that is intrinsically discrete such as the rolls of a die, there can be many repeat values - five 2s, eight 3s, etc. For a situation that is intrinsically continuous, we can visualise the probability density by binning the values - that is, counting the number of values that fall within a set of discrete ranges - and plotting a histogram. We would like some standard measures to characterise the distribution of sample values without having to know the parent distribution. Measures of location. First we might ask what is the typical value? There are three common ways of translating this qualitative concept into something rigorous. The first is to use the histogram and estimate the most probable value or mode, where the local density of x i values is highest. This is intuitive but sensitive to how you bin the histogram. Another method is the median - the value for which half the values are above and half below. This has the great virtue of being completely robust against transformations of x, but it is hard to work with mathematically. The commonest estimate is the arithmetic mean or average :

5 Stats 1 : Probabilities 5 x = xi N For symmetrical parent distributions, the mean, mode and median will all be the same on average. Measures of dispersion. Next we want to know the spread of values. We could start by calculating the deviation of each point from the mean, x i x. The mean value of this will typically be zero for many parent distributions, which is not helpful. We could find the average of the absolute value, but this is hard to work with mathematically. Instead it is usual to define the sample variance s 2 = (xi x) 2 N The square root of the variance, s, is then the standard deviation. The ratio s/ x is sometimes called the coefficient of variation. Moments of the sample distribution. A very general method of characterising a distribution is by defining its moments. The nth moment of a distribution about an arbitrary point x a is defined as µ a n = 1 N N (x i x a ) n i=1 If we take the first moment about the point x a = 0 we see that µ 0 1 = x. If the point x a is taken to be the mean x then the moments are called the central moments and written without the superscript. We then get µ 0 = 1 µ 1 = 0 µ 2 = s 2 Skewness and kurtosis. The first two moments (not counting the zeroth moment) correspond to the ideas of location and dispersion. The third moment is related to the skewness, or deviation from a symmetrical form, and the fourth moment is related to the kurtosis, or degree of peakiness. To compare different distributions with each other, it is usual to normalise these moments to the dispersion, but there are several different conventions for doing this. Two common definitions are: skewness α 3 = µ 3 µ 3 2 kurtosis α 4 = µ 4 µ Expected values The idea of moments is closely related to the more general concept of expectation. The expected value of a random variable is simply the average value of all its possible values, with due respect to the probability of occurrence of each value. We can define this concept for either discrete or continous random variables: E[X] = X X P (X) or E[x] = x p(x)dx

6 Stats 1 : Probabilities 6 Where P (X) is the set of probabilities of the discrete random variable X, and p(x) is the PDF for the continuous random variable x. You will often see expected values written as x rather than E[x]. Because a function of a random variable is also a random variable, we can calculate the expected value of any function of x E[f(x)] = f(x) p(x)dx Parent distribution moments. The moments of a random variable x can now be defined as the expected values of x n for various n: m n = E[x n ]. (In comparing with discussion above w.r.t. sample distributions, this corresponds to the moment about x = 0.) The zeroth moment m 0 = + p(x)dx = 1 as long as p(x) is properly normalised. The first moment m 1 = E[x] is the mean, which is the expectation value of x. The standard second moment is m 2 = E[x 2 ], but it is more useful to define the centred moment obtained by shifting the origin of x to the mean : µ n E [(x E(x)) n ]. Then the second centred moment is ( µ 2 E (x E(x)) 2) = σ 2 Conventionally we use the symbol σ 2 for the variance of the (theoretical) parent distribution, and s 2 for the variance of the (observed) sample distribution. We can define higher moments in the same fashion. The algebra of expectations. When manipulating moments of distributions, its handy to use some simple rules for how expectation values combine: E[X + Y ] = E[X] + E[Y ] E[X Y ] = E[X] E[Y ] E[aX + b] = ae[x] + b E[aX + by ] = ae[x] + be[y ] Note also that if b is a constant E(b) = b; and because an expected value is itself a constant, having integrated over x, then E(E(x)) is just E(x). A useful result which can be obtained is that ( σ 2 = E (x E (x)) 2) (( = E x 2 2xE (x) + E (x) 2)) = E ( x 2) E (x) 2 In other words, the variance can be obtained from the mean of the square minus the square of the mean. Relation between sample and parent moments. For sample distributions, statistics such as the mean and variance are useful objective summaries of the properties of the observed sample, but they can also be seen as estimates of the corresponding quantities in the parent distribution. (We will look at this more closley in part 3). However, there are in principle different ways of using the sample values to estimate the properties of the parent distribution, so we need to be a little careful. For example, according to the maximum likelihood method which we will discuss later, the best estimate of the parent variance is not s 2 but s 2 N/(N 1). In parts 3 and 5 we will look more closely about how you estimate the parameters of a model.

7 Stats 1 : Probabilities Error analysis From high school physics onwards, we are taught that we should associate uncertainties or errors with our measurements. What does this actually mean? Mistakes and biases. In normal English, an error is some kind of actual mistake. This can certainly apply to physical experiments. If we believe we are measuring a lump of Sodium, and it s actually a lump of Potassium, we will certainly get an incorrect value. There might also be some kind of fixed bias. We measure some effect with a voltmeter, but a problem with our circuitry means that all our measurements have 0.03V added to them. Random errors. Mistakes and biases can in principle be corrected. However, central to the idea of error analysis is that a range of measured values is unavoidable, because the act of measurement is itself a random process. We assume that some property X has a true value, but the measured value x is a random variable. Each time we run the experiment and produce a value of x, it is drawn from some probability distribution p(x) related to the true value X. We should carefully distinguish the idea of randomness in the measurement process from the idea of randomness in actual physical properties. People have a variety of heights, and so the height h of a person selected at random from a large population has a spread. If we measure the height of one specific person, we should in principle get a single definite value. However, when we measure the height of that specific person repeatedly, we actually get a small spread of results, because of randomness in the measurement process. The standard deviation of this error distribution is what we normally use to quantify the sense of error. Later in part 3 we will sharpen this idea and talk about the idea of confidence regions. Systematic vs random errors. If a sample is drawn from the underlying parent error distribution, we can use the sample mean as an estimate of the mean of the parent distribution, which in turn we take to represent the true value. The more measurements we make, the more precise will be our estimate of the parent mean. The error on 30 measurements is smaller than the error on 3 measurements. (We will sharpen that statement in parts 3 and 5). However, if we are dominated by systematic errors, the precision does not improve - the error stays just as big, however many measurements we make. Sometimes both a random error and a systematic error applies, and it is healthy if we quote each separately, rather than a combined error. We can also question the assumption that the mean of the parent error distribution is the same as the true value of the physical quantity concerned. This captures the normal English concept of bias but in a statistical sense. Because of the possibility of bias, accuracy - the difference between true and measured values - is not the same as precision. As we make more measurements and the random error reduces, the accuracy may not improve by much. Evaluating errors. Sometimes we know what is causing the randomness in the measurement process. For example, when an experiment involves counting things, such as the number of photons detected from a star, the number counted is subject to Poisson statistics, which we will look at in part 2. Often however, we do not know the source of error; we then need to estimate the error empirically by making a sequence of measurements and calculating the sample variance. Of course, as Donald Rumsfeld might have said, our problem is not just the known known errors, or the known unknown errors, but the unknown unknown errors. Propagation of errors. Once we have estimated an error on some measured quantity x, or perhaps two quantities x and y, we would like to know what this implies for some derived quantity z - for example in our equipment we actually measure a voltage and a time delay, but use some formula to calculate the value of electron mass that these measurements imply. We can use the usual rules of calculus to transform from one variable to another, and the definitions of expected values to calculate the second moment of the desired variable. The result depends on whether x and y are dependent or independent variables. We will defer a detailed discussion of the issue of dependence to part 4. Meanwhile, without proof, we can state the result for two independent variables.

8 Stats 1 : Probabilities 8 If z = f(x, y) is some function f of random variables x and y, and we know the variance of each of these, what is the variance of z? If the variables are independent then σ 2 z = ( ) f 2 σx 2 + x ( ) f 2 σy 2 y This formula is useful in many places, but in particular it shows us how to propagate errors. We can use the propagation formula to see how this works for a variety of different mathematical relationships. Some examples are : f = ax ± by σ 2 f = a2 σ 2 x + b 2 σ 2 y f = xy or x/y (σ f /f) 2 = (σ x /x) 2 + (σ y /y) 2 f = ax ±b f = a ln (±bx) f = ae ±bx (σ f /f) = b (σ x /x) σ f = a (σ x /x) (σ f /f) = bσ x f = a ±bx (σ f /f) = b ln(aσ x )

STA Module 4 Probability Concepts. Rev.F08 1

STA Module 4 Probability Concepts. Rev.F08 1 STA 2023 Module 4 Probability Concepts Rev.F08 1 Learning Objectives Upon completing this module, you should be able to: 1. Compute probabilities for experiments having equally likely outcomes. 2. Interpret

More information

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems

More information

STAT 430/510 Probability Lecture 7: Random Variable and Expectation

STAT 430/510 Probability Lecture 7: Random Variable and Expectation STAT 430/510 Probability Lecture 7: Random Variable and Expectation Pengyuan (Penelope) Wang June 2, 2011 Review Properties of Probability Conditional Probability The Law of Total Probability Bayes Formula

More information

Lecture 1: Probability Fundamentals

Lecture 1: Probability Fundamentals Lecture 1: Probability Fundamentals IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge January 22nd, 2008 Rasmussen (CUED) Lecture 1: Probability

More information

Lecture 10: Bayes' Theorem, Expected Value and Variance Lecturer: Lale Özkahya

Lecture 10: Bayes' Theorem, Expected Value and Variance Lecturer: Lale Özkahya BBM 205 Discrete Mathematics Hacettepe University http://web.cs.hacettepe.edu.tr/ bbm205 Lecture 10: Bayes' Theorem, Expected Value and Variance Lecturer: Lale Özkahya Resources: Kenneth Rosen, Discrete

More information

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix

More information

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019 Lecture 10: Probability distributions DANIEL WELLER TUESDAY, FEBRUARY 19, 2019 Agenda What is probability? (again) Describing probabilities (distributions) Understanding probabilities (expectation) Partial

More information

Confidence Intervals

Confidence Intervals Quantitative Foundations Project 3 Instructor: Linwei Wang Confidence Intervals Contents 1 Introduction 3 1.1 Warning....................................... 3 1.2 Goals of Statistics..................................

More information

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results

More information

Probability. Table of contents

Probability. Table of contents Probability Table of contents 1. Important definitions 2. Distributions 3. Discrete distributions 4. Continuous distributions 5. The Normal distribution 6. Multivariate random variables 7. Other continuous

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

Probability theory basics

Probability theory basics Probability theory basics Michael Franke Basics of probability theory: axiomatic definition, interpretation, joint distributions, marginalization, conditional probability & Bayes rule. Random variables:

More information

Probability Dr. Manjula Gunarathna 1

Probability Dr. Manjula Gunarathna 1 Probability Dr. Manjula Gunarathna Probability Dr. Manjula Gunarathna 1 Introduction Probability theory was originated from gambling theory Probability Dr. Manjula Gunarathna 2 History of Probability Galileo

More information

STATISTICS 1 REVISION NOTES

STATISTICS 1 REVISION NOTES STATISTICS 1 REVISION NOTES Statistical Model Representing and summarising Sample Data Key words: Quantitative Data This is data in NUMERICAL FORM such as shoe size, height etc. Qualitative Data This is

More information

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015 Probability Refresher Kai Arras, University of Freiburg Winter term 2014/2015 Probability Refresher Introduction to Probability Random variables Joint distribution Marginalization Conditional probability

More information

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc. Chapter 14 From Randomness to Probability Copyright 2012, 2008, 2005 Pearson Education, Inc. Dealing with Random Phenomena A random phenomenon is a situation in which we know what outcomes could happen,

More information

Week 12-13: Discrete Probability

Week 12-13: Discrete Probability Week 12-13: Discrete Probability November 21, 2018 1 Probability Space There are many problems about chances or possibilities, called probability in mathematics. When we roll two dice there are possible

More information

2. A Basic Statistical Toolbox

2. A Basic Statistical Toolbox . A Basic Statistical Toolbo Statistics is a mathematical science pertaining to the collection, analysis, interpretation, and presentation of data. Wikipedia definition Mathematical statistics: concerned

More information

MAT Mathematics in Today's World

MAT Mathematics in Today's World MAT 1000 Mathematics in Today's World Last Time We discussed the four rules that govern probabilities: 1. Probabilities are numbers between 0 and 1 2. The probability an event does not occur is 1 minus

More information

Basic Probability. Introduction

Basic Probability. Introduction Basic Probability Introduction The world is an uncertain place. Making predictions about something as seemingly mundane as tomorrow s weather, for example, is actually quite a difficult task. Even with

More information

CME 106: Review Probability theory

CME 106: Review Probability theory : Probability theory Sven Schmit April 3, 2015 1 Overview In the first half of the course, we covered topics from probability theory. The difference between statistics and probability theory is the following:

More information

1 Random variables and distributions

1 Random variables and distributions Random variables and distributions In this chapter we consider real valued functions, called random variables, defined on the sample space. X : S R X The set of possible values of X is denoted by the set

More information

Some Concepts of Probability (Review) Volker Tresp Summer 2018

Some Concepts of Probability (Review) Volker Tresp Summer 2018 Some Concepts of Probability (Review) Volker Tresp Summer 2018 1 Definition There are different way to define what a probability stands for Mathematically, the most rigorous definition is based on Kolmogorov

More information

STAT Chapter 3: Probability

STAT Chapter 3: Probability Basic Definitions STAT 515 --- Chapter 3: Probability Experiment: A process which leads to a single outcome (called a sample point) that cannot be predicted with certainty. Sample Space (of an experiment):

More information

Advanced Herd Management Probabilities and distributions

Advanced Herd Management Probabilities and distributions Advanced Herd Management Probabilities and distributions Anders Ringgaard Kristensen Slide 1 Outline Probabilities Conditional probabilities Bayes theorem Distributions Discrete Continuous Distribution

More information

Lecture 2: Review of Probability

Lecture 2: Review of Probability Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................

More information

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg Statistics for Data Analysis PSI Practical Course 2014 Niklaus Berger Physics Institute, University of Heidelberg Overview You are going to perform a data analysis: Compare measured distributions to theoretical

More information

Counting principles, including permutations and combinations.

Counting principles, including permutations and combinations. 1 Counting principles, including permutations and combinations. The binomial theorem: expansion of a + b n, n ε N. THE PRODUCT RULE If there are m different ways of performing an operation and for each

More information

MATH MW Elementary Probability Course Notes Part I: Models and Counting

MATH MW Elementary Probability Course Notes Part I: Models and Counting MATH 2030 3.00MW Elementary Probability Course Notes Part I: Models and Counting Tom Salisbury salt@yorku.ca York University Winter 2010 Introduction [Jan 5] Probability: the mathematics used for Statistics

More information

Probability, Entropy, and Inference / More About Inference

Probability, Entropy, and Inference / More About Inference Probability, Entropy, and Inference / More About Inference Mário S. Alvim (msalvim@dcc.ufmg.br) Information Theory DCC-UFMG (2018/02) Mário S. Alvim (msalvim@dcc.ufmg.br) Probability, Entropy, and Inference

More information

Statistics 100A Homework 5 Solutions

Statistics 100A Homework 5 Solutions Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to

More information

Review of Basic Probability

Review of Basic Probability Review of Basic Probability Erik G. Learned-Miller Department of Computer Science University of Massachusetts, Amherst Amherst, MA 01003 September 16, 2009 Abstract This document reviews basic discrete

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

If the objects are replaced there are n choices each time yielding n r ways. n C r and in the textbook by g(n, r).

If the objects are replaced there are n choices each time yielding n r ways. n C r and in the textbook by g(n, r). Caveat: Not proof read. Corrections appreciated. Combinatorics In the following, n, n 1, r, etc. will denote non-negative integers. Rule 1 The number of ways of ordering n distinguishable objects (also

More information

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2011

A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2011 A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2011 Reading Chapter 5 (continued) Lecture 8 Key points in probability CLT CLT examples Prior vs Likelihood Box & Tiao

More information

Probability, CLT, CLT counterexamples, Bayes. The PDF file of this lecture contains a full reference document on probability and random variables.

Probability, CLT, CLT counterexamples, Bayes. The PDF file of this lecture contains a full reference document on probability and random variables. Lecture 5 A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 2015 http://www.astro.cornell.edu/~cordes/a6523 Probability, CLT, CLT counterexamples, Bayes The PDF file of

More information

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3 Probability Paul Schrimpf January 23, 2018 Contents 1 Definitions 2 2 Properties 3 3 Random variables 4 3.1 Discrete........................................... 4 3.2 Continuous.........................................

More information

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the

More information

Chapter 4: An Introduction to Probability and Statistics

Chapter 4: An Introduction to Probability and Statistics Chapter 4: An Introduction to Probability and Statistics 4. Probability The simplest kinds of probabilities to understand are reflected in everyday ideas like these: (i) if you toss a coin, the probability

More information

Probability and distributions. Francesco Corona

Probability and distributions. Francesco Corona Probability Probability and distributions Francesco Corona Department of Computer Science Federal University of Ceará, Fortaleza Probability Many kinds of studies can be characterised as (repeated) experiments

More information

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming

More information

Probability - Lecture 4

Probability - Lecture 4 1 Introduction Probability - Lecture 4 Many methods of computation physics and the comparison of data to a mathematical representation, apply stochastic methods. These ideas were first introduced in the

More information

2) There should be uncertainty as to which outcome will occur before the procedure takes place.

2) There should be uncertainty as to which outcome will occur before the procedure takes place. robability Numbers For many statisticians the concept of the probability that an event occurs is ultimately rooted in the interpretation of an event as an outcome of an experiment, others would interpret

More information

Probability and Independence Terri Bittner, Ph.D.

Probability and Independence Terri Bittner, Ph.D. Probability and Independence Terri Bittner, Ph.D. The concept of independence is often confusing for students. This brief paper will cover the basics, and will explain the difference between independent

More information

review session gov 2000 gov 2000 () review session 1 / 38

review session gov 2000 gov 2000 () review session 1 / 38 review session gov 2000 gov 2000 () review session 1 / 38 Overview Random Variables and Probability Univariate Statistics Bivariate Statistics Multivariate Statistics Causal Inference gov 2000 () review

More information

Probability and Estimation. Alan Moses

Probability and Estimation. Alan Moses Probability and Estimation Alan Moses Random variables and probability A random variable is like a variable in algebra (e.g., y=e x ), but where at least part of the variability is taken to be stochastic.

More information

Chapter 8: An Introduction to Probability and Statistics

Chapter 8: An Introduction to Probability and Statistics Course S3, 200 07 Chapter 8: An Introduction to Probability and Statistics This material is covered in the book: Erwin Kreyszig, Advanced Engineering Mathematics (9th edition) Chapter 24 (not including

More information

Notes 12 Autumn 2005

Notes 12 Autumn 2005 MAS 08 Probability I Notes Autumn 005 Conditional random variables Remember that the conditional probability of event A given event B is P(A B) P(A B)/P(B). Suppose that X is a discrete random variable.

More information

Probability and Probability Distributions. Dr. Mohammed Alahmed

Probability and Probability Distributions. Dr. Mohammed Alahmed Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten

More information

1 Probability and Random Variables

1 Probability and Random Variables 1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in

More information

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014 Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions

More information

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces.

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces. Probability Theory To start out the course, we need to know something about statistics and probability Introduction to Probability Theory L645 Advanced NLP Autumn 2009 This is only an introduction; for

More information

RVs and their probability distributions

RVs and their probability distributions RVs and their probability distributions RVs and their probability distributions In these notes, I will use the following notation: The probability distribution (function) on a sample space will be denoted

More information

Bayesian Inference. Introduction

Bayesian Inference. Introduction Bayesian Inference Introduction The frequentist approach to inference holds that probabilities are intrinsicially tied (unsurprisingly) to frequencies. This interpretation is actually quite natural. What,

More information

01 Probability Theory and Statistics Review

01 Probability Theory and Statistics Review NAVARCH/EECS 568, ROB 530 - Winter 2018 01 Probability Theory and Statistics Review Maani Ghaffari January 08, 2018 Last Time: Bayes Filters Given: Stream of observations z 1:t and action data u 1:t Sensor/measurement

More information

Probability Rules. MATH 130, Elements of Statistics I. J. Robert Buchanan. Fall Department of Mathematics

Probability Rules. MATH 130, Elements of Statistics I. J. Robert Buchanan. Fall Department of Mathematics Probability Rules MATH 130, Elements of Statistics I J. Robert Buchanan Department of Mathematics Fall 2018 Introduction Probability is a measure of the likelihood of the occurrence of a certain behavior

More information

Example A. Define X = number of heads in ten tosses of a coin. What are the values that X may assume?

Example A. Define X = number of heads in ten tosses of a coin. What are the values that X may assume? Stat 400, section.1-.2 Random Variables & Probability Distributions notes by Tim Pilachowski For a given situation, or experiment, observations are made and data is recorded. A sample space S must contain

More information

Discrete Probability Refresher

Discrete Probability Refresher ECE 1502 Information Theory Discrete Probability Refresher F. R. Kschischang Dept. of Electrical and Computer Engineering University of Toronto January 13, 1999 revised January 11, 2006 Probability theory

More information

Recitation 2: Probability

Recitation 2: Probability Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions

More information

Review of Statistics

Review of Statistics Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and

More information

CS 246 Review of Proof Techniques and Probability 01/14/19

CS 246 Review of Proof Techniques and Probability 01/14/19 Note: This document has been adapted from a similar review session for CS224W (Autumn 2018). It was originally compiled by Jessica Su, with minor edits by Jayadev Bhaskaran. 1 Proof techniques Here we

More information

University of Regina. Lecture Notes. Michael Kozdron

University of Regina. Lecture Notes. Michael Kozdron University of Regina Statistics 252 Mathematical Statistics Lecture Notes Winter 2005 Michael Kozdron kozdron@math.uregina.ca www.math.uregina.ca/ kozdron Contents 1 The Basic Idea of Statistics: Estimating

More information

Probability calculus and statistics

Probability calculus and statistics A Probability calculus and statistics A.1 The meaning of a probability A probability can be interpreted in different ways. In this book, we understand a probability to be an expression of how likely it

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Random variables (discrete)

Random variables (discrete) Random variables (discrete) Saad Mneimneh 1 Introducing random variables A random variable is a mapping from the sample space to the real line. We usually denote the random variable by X, and a value that

More information

Statistics notes. A clear statistical framework formulates the logic of what we are doing and why. It allows us to make precise statements.

Statistics notes. A clear statistical framework formulates the logic of what we are doing and why. It allows us to make precise statements. Statistics notes Introductory comments These notes provide a summary or cheat sheet covering some basic statistical recipes and methods. These will be discussed in more detail in the lectures! What is

More information

PhysicsAndMathsTutor.com. Advanced/Advanced Subsidiary. You must have: Mathematical Formulae and Statistical Tables (Blue)

PhysicsAndMathsTutor.com. Advanced/Advanced Subsidiary. You must have: Mathematical Formulae and Statistical Tables (Blue) Write your name here Surname Other names Pearson Edexcel International Advanced Level Centre Number Statistics S1 Advanced/Advanced Subsidiary Candidate Number Friday 20 January 2017 Afternoon Time: 1

More information

Basics on Probability. Jingrui He 09/11/2007

Basics on Probability. Jingrui He 09/11/2007 Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability

More information

1 What are probabilities? 2 Sample Spaces. 3 Events and probability spaces

1 What are probabilities? 2 Sample Spaces. 3 Events and probability spaces 1 What are probabilities? There are two basic schools of thought as to the philosophical status of probabilities. One school of thought, the frequentist school, considers the probability of an event to

More information

Introduction to Statistics and Data Analysis

Introduction to Statistics and Data Analysis Introduction to Statistics and Data Analysis RSI 2005 Staff July 15, 2005 Variation and Statistics Good experimental technique often requires repeated measurements of the same quantity These repeatedly

More information

2.3 Estimating PDFs and PDF Parameters

2.3 Estimating PDFs and PDF Parameters .3 Estimating PDFs and PDF Parameters estimating means - discrete and continuous estimating variance using a known mean estimating variance with an estimated mean estimating a discrete pdf estimating a

More information

Examples of frequentist probability include games of chance, sample surveys, and randomized experiments. We will focus on frequentist probability sinc

Examples of frequentist probability include games of chance, sample surveys, and randomized experiments. We will focus on frequentist probability sinc FPPA-Chapters 13,14 and parts of 16,17, and 18 STATISTICS 50 Richard A. Berk Spring, 1997 May 30, 1997 1 Thinking about Chance People talk about \chance" and \probability" all the time. There are many

More information

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner

Fundamentals. CS 281A: Statistical Learning Theory. Yangqing Jia. August, Based on tutorial slides by Lester Mackey and Ariel Kleiner Fundamentals CS 281A: Statistical Learning Theory Yangqing Jia Based on tutorial slides by Lester Mackey and Ariel Kleiner August, 2011 Outline 1 Probability 2 Statistics 3 Linear Algebra 4 Optimization

More information

Brief Review of Probability

Brief Review of Probability Brief Review of Probability Nuno Vasconcelos (Ken Kreutz-Delgado) ECE Department, UCSD Probability Probability theory is a mathematical language to deal with processes or experiments that are non-deterministic

More information

2007 Winton. Empirical Distributions

2007 Winton. Empirical Distributions 1 Empirical Distributions 2 Distributions In the discrete case, a probability distribution is just a set of values, each with some probability of occurrence Probabilities don t change as values occur Example,

More information

Statistical Concepts

Statistical Concepts Statistical Concepts Ad Feelders May 19, 2015 1 Introduction Statistics is the science of collecting, organizing and drawing conclusions from data. How to properly produce and collect data is studied in

More information

ACE 562 Fall Lecture 2: Probability, Random Variables and Distributions. by Professor Scott H. Irwin

ACE 562 Fall Lecture 2: Probability, Random Variables and Distributions. by Professor Scott H. Irwin ACE 562 Fall 2005 Lecture 2: Probability, Random Variables and Distributions Required Readings: by Professor Scott H. Irwin Griffiths, Hill and Judge. Some Basic Ideas: Statistical Concepts for Economists,

More information

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ).

1. When applied to an affected person, the test comes up positive in 90% of cases, and negative in 10% (these are called false negatives ). CS 70 Discrete Mathematics for CS Spring 2006 Vazirani Lecture 8 Conditional Probability A pharmaceutical company is marketing a new test for a certain medical condition. According to clinical trials,

More information

Probability and statistics; Rehearsal for pattern recognition

Probability and statistics; Rehearsal for pattern recognition Probability and statistics; Rehearsal for pattern recognition Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských

More information

Randomized Algorithms

Randomized Algorithms Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours

More information

7.1 What is it and why should we care?

7.1 What is it and why should we care? Chapter 7 Probability In this section, we go over some simple concepts from probability theory. We integrate these with ideas from formal language theory in the next chapter. 7.1 What is it and why should

More information

PROBABILITY THEORY REVIEW

PROBABILITY THEORY REVIEW PROBABILITY THEORY REVIEW CMPUT 466/551 Martha White Fall, 2017 REMINDERS Assignment 1 is due on September 28 Thought questions 1 are due on September 21 Chapters 1-4, about 40 pages If you are printing,

More information

1 INFO Sep 05

1 INFO Sep 05 Events A 1,...A n are said to be mutually independent if for all subsets S {1,..., n}, p( i S A i ) = p(a i ). (For example, flip a coin N times, then the events {A i = i th flip is heads} are mutually

More information

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

STAT2201. Analysis of Engineering & Scientific Data. Unit 3 STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random

More information

Unit 4 Probability. Dr Mahmoud Alhussami

Unit 4 Probability. Dr Mahmoud Alhussami Unit 4 Probability Dr Mahmoud Alhussami Probability Probability theory developed from the study of games of chance like dice and cards. A process like flipping a coin, rolling a die or drawing a card from

More information

Topic 3: The Expectation of a Random Variable

Topic 3: The Expectation of a Random Variable Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation

More information

Dept. of Linguistics, Indiana University Fall 2015

Dept. of Linguistics, Indiana University Fall 2015 L645 Dept. of Linguistics, Indiana University Fall 2015 1 / 34 To start out the course, we need to know something about statistics and This is only an introduction; for a fuller understanding, you would

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:

More information

02 Background Minimum background on probability. Random process

02 Background Minimum background on probability. Random process 0 Background 0.03 Minimum background on probability Random processes Probability Conditional probability Bayes theorem Random variables Sampling and estimation Variance, covariance and correlation Probability

More information

Basic Probability Reference Sheet

Basic Probability Reference Sheet February 27, 2001 Basic Probability Reference Sheet 17.846, 2001 This is intended to be used in addition to, not as a substitute for, a textbook. X is a random variable. This means that X is a variable

More information

CONTENTS OF DAY 2. II. Why Random Sampling is Important 10 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE

CONTENTS OF DAY 2. II. Why Random Sampling is Important 10 A myth, an urban legend, and the real reason NOTES FOR SUMMER STATISTICS INSTITUTE COURSE 1 2 CONTENTS OF DAY 2 I. More Precise Definition of Simple Random Sample 3 Connection with independent random variables 4 Problems with small populations 9 II. Why Random Sampling is Important 10 A myth,

More information

Conditional probabilities and graphical models

Conditional probabilities and graphical models Conditional probabilities and graphical models Thomas Mailund Bioinformatics Research Centre (BiRC), Aarhus University Probability theory allows us to describe uncertainty in the processes we model within

More information

Bivariate distributions

Bivariate distributions Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient

More information

Introduction to Bayesian Statistics

Introduction to Bayesian Statistics School of Computing & Communication, UTS January, 207 Random variables Pre-university: A number is just a fixed value. When we talk about probabilities: When X is a continuous random variable, it has a

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

Conditional Probability

Conditional Probability Conditional Probability Idea have performed a chance experiment but don t know the outcome (ω), but have some partial information (event A) about ω. Question: given this partial information what s the

More information

Single Maths B: Introduction to Probability

Single Maths B: Introduction to Probability Single Maths B: Introduction to Probability Overview Lecturer Email Office Homework Webpage Dr Jonathan Cumming j.a.cumming@durham.ac.uk CM233 None! http://maths.dur.ac.uk/stats/people/jac/singleb/ 1 Introduction

More information

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations EECS 70 Discrete Mathematics and Probability Theory Fall 204 Anant Sahai Note 5 Random Variables: Distributions, Independence, and Expectations In the last note, we saw how useful it is to have a way of

More information