Statistical Methods for Astronomy

Similar documents
AST 418/518 Instrumentation and Statistics

AST 418/518 Instrumentation and Statistics

Statistical Methods for Astronomy

Practical Statistics

Statistics notes. A clear statistical framework formulates the logic of what we are doing and why. It allows us to make precise statements.

Hypothesis testing (cont d)

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

Probability calculus and statistics

Introduction to Bayesian Learning. Machine Learning Fall 2018

Statistical Methods in Particle Physics. Lecture 2

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Lecture 5: Bayes pt. 1

Statistics, Probability Distributions & Error Propagation. James R. Graham

Lecture 1: Probability Fundamentals

Statistical Methods for Astronomy

Chapter 8: An Introduction to Probability and Statistics

Advanced Herd Management Probabilities and distributions

Basics on Probability. Jingrui He 09/11/2007

Bayesian Learning. Reading: Tom Mitchell, Generative and discriminative classifiers: Naive Bayes and logistic regression, Sections 1-2.

Data, Estimation and Inference

Physics 6720 Introduction to Statistics April 4, 2017

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

Deep Learning for Computer Vision

STATISTICS OF OBSERVATIONS & SAMPLING THEORY. Parent Distributions

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

Chapter 4: An Introduction to Probability and Statistics

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Probability and Probability Distributions. Dr. Mohammed Alahmed

Probability, Entropy, and Inference / More About Inference

Notes on probability : Exercise problems, sections (1-7)

2. A Basic Statistical Toolbox

Probability, Statistics, and Bayes Theorem Session 3

Probability Density Functions

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

Statistical Methods in Particle Physics

Intro to Probability. Andrei Barbu

the time it takes until a radioactive substance undergoes a decay

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit

Probability: Why do we care? Lecture 2: Probability and Distributions. Classical Definition. What is Probability?

CSE 312 Final Review: Section AA

(3) Review of Probability. ST440/540: Applied Bayesian Statistics

Probability theory and inference statistics! Dr. Paola Grosso! SNE research group!! (preferred!)!!

Lecture 3. January 7, () Lecture 3 January 7, / 35

Computational Biology Lecture #3: Probability and Statistics. Bud Mishra Professor of Computer Science, Mathematics, & Cell Biology Sept

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Hypothesis Testing, Bayes Theorem, & Parameter Estimation

MATH 19B FINAL EXAM PROBABILITY REVIEW PROBLEMS SPRING, 2010

Modern Methods of Data Analysis - SS 2009

Lecture 2: Probability and Distributions

Fourier and Stats / Astro Stats and Measurement : Stats Notes

n N CHAPTER 1 Atoms Thermodynamics Molecules Statistical Thermodynamics (S.T.)

Lecture 1: Review of Probability

Physics 509: Non-Parametric Statistics and Correlation Testing

P (A B) P ((B C) A) P (B A) = P (B A) + P (C A) P (A) = P (B A) + P (C A) = Q(A) + Q(B).

Probability theory basics

Introduction to Statistics and Error Analysis

LECTURE 5. Introduction to Econometrics. Hypothesis testing

Machine Learning. CS Spring 2015 a Bayesian Learning (I) Uncertainty

Introduction to Statistics and Error Analysis II

Why Try Bayesian Methods? (Lecture 5)

Lecture 3: Probability

Math Review Sheet, Fall 2008

Lecture 7: Confidence interval and Normal approximation

Uncertain Knowledge and Bayes Rule. George Konidaris

Week 1 Quantitative Analysis of Financial Markets Distributions A

CONTINUOUS RANDOM VARIABLES

Chapter 1 Principles of Probability

Priors, Total Probability, Expectation, Multiple Trials

Journeys of an Accidental Statistician

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!

What is Probability? Probability. Sample Spaces and Events. Simple Event

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Lecture 2. Conditional Probability

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester

FYST17 Lecture 8 Statistics and hypothesis testing. Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons

Chapter Three. Hypothesis Testing

Review of probability

Introduction to Error Analysis

CS395T Computational Statistics with Application to Bioinformatics

Algorithms for Uncertainty Quantification

3rd IIA-Penn State Astrostatistics School July, 2010 Vainu Bappu Observatory, Kavalur

Introduction to Statistical Methods for High Energy Physics

4th IIA-Penn State Astrostatistics School July, 2013 Vainu Bappu Observatory, Kavalur

Evaluating Classifiers. Lecture 2 Instructor: Max Welling

With Question/Answer Animations. Chapter 7

Bayesian Inference: Concept and Practice

An example to illustrate frequentist and Bayesian approches

Data Modeling & Analysis Techniques. Probability & Statistics. Manfred Huber

Introduction to Machine Learning

Lecture 2: Repetition of probability theory and statistics

Discrete distribution. Fitting probability models to frequency data. Hypotheses for! 2 test. ! 2 Goodness-of-fit test

Sociology 6Z03 Review II

Contents. Decision Making under Uncertainty 1. Meanings of uncertainty. Classical interpretation

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

2.4 Conditional Probability

Numerical Methods for Data Analysis

Probability - Lecture 4

Transcription:

Statistical Methods for Astronomy Probability (Lecture 1) Statistics (Lecture 2) Why do we need statistics? Useful Statistics Definitions Error Analysis Probability distributions Error Propagation Binomial Distribution Least Squares Poisson Distribution chi-squared Gaussian Distribution Significance Bayes Theorem Comparison Statistics Central Limit theorem

Useful References Data Reduction and Error Analysis, Bevington and Robinson Practical Statistics for Astronomers, Wall and Jenkins Numerical Recipes, Press et al. Understanding Data Better with Bayesian and Global Statistical Methods, Press, 1996 (on astro-ph)

What use are statistical methods? Help you make a decision! Is a signal in a set of observations meaningful? Do the data fit our model of the phenomenon under study? Simulate Observations Plan size of sample, etc. What would happen if we repeated the observations? Compare different observations Are two sets of data consistent with each other? Are the observations truly independent?

Process of Decision Making Ask a Question Take Data Reduce Data Derive Statistics describing data Reflect on what is needed Probability Distribution Error Analysis Does the Statistic answer your question? No Hypothesis Testing Yes Simulation Publish! 4

Some definitions Statistic a number or set of numbers that describe a set of data. Probability distribution the relative chances for different outcomes for your set of data. Sample distribution Set of data that allow us to estimate useful values of the object under study. Parent distribution Presumed probability distribution of the data that one would measure if an infinite data set were acquired. Mean - the 1st moment of a distribution, which gives information about the most likely value one will observe. Variance the 2 nd moment of a distribution, which gives information about the range of values one will observe.

Typical Statistics If we have n data points, we typically want to know things like the value of a typical data point, or how much the variation in the data is: Mean µ = j x j n Variance s 2 = (x j µ) 2 n 1 j 6

Probability Distributions Probability distributions (P(x)) can be any strange function (or non-analytical curve) that can be imagined as long as prob(a <x<b)= P (x)dx =1 b a P (x)dx P(x) is a single, non-negative value for all real x.

Mean and Variance of Probability Distributions Mean Variance Discrete: n j P (n j ) (n j ) 2 P (n j ) µ 2 Continuous: µ = j µ = σ 2 = j xp (x)dx σ 2 = (n j µ) 2 P (n j )= j (x µ) 2 P (x)dx = x 2 P (x)dx µ 2 σ 2 = x 2 x 2

The Binomial distribution You are observing something that has a probability, p, of occurring in a single observation. You observe it N times. Want chance of obtaining n successes. For one, particular sequence of observations the probability is: P 1 (n) =p n (1 p) N n There are many sequences which yield n successes: N! P (n) = n!(n n)! pn (1 p) N n N = p n (1 p) N n n Mean Np Variance Np(1-p) Often said N choose n

The Poisson Distribution Consider the binomial case where p 0, but Np µ. The binomial distribution, then becomes: P (n) =µ n e µ n! Mean Np=µ Variance Np(1-p)~Np=µ

Gaussian Distribution The limiting case of the Poisson distribution for large µ is the Gaussian, or normal distribution P (x)dx = 1 σ 2π e (x µ) 2 2σ 2 dx Mean Variance µ σ^2 Large µ Poisson distributions are Gaussian with σ^2=µ, but Gaussian distributions do not necessarily follow this.

Gaussian Distribution The Gaussian distribution is often used (sometimes incorrectly) to express confidence. P ( x <µ+ σ ) > 0.68 P ( x <µ+2σ ) > 0.95 P ( x <µ+3σ ) > 0.997 12

Mean and Variance of Distributions Distribution Mean Variance Binomial Np Np(1-p) Poisson µ µ Gaussian µ σ 2 Uniform [a,b) (a+b)/2 (b-a)/12

Two approaches to discussing the problem: Knowing the distribution allows us to predict what we will observe. We often know what we have observed and want to determine what that tells us about the distribution.

Frequentist Approach I hypothesize that there are an equal number of red and white balls in a box. I see I have drawn 6 red balls out of 10 total trials. Prediction A box with equal number of balls will have a mean of 5 red balls with a standard deviation of 1.6. Based on this I cannot reject my original hypothesis. 15

Bayesian Approach I hypothesize that there are an equal number of red and white balls in a box. I see I have drawn 6 red balls out of 10 total trials. Odds on what is in the box. There is a 24% chance that my hypothesis is correct. 16

Approaches to Statistics Frequentist approaches will calculate statistics that a given distribution would have produced, and confirms or rejects a hypothesis. These are computationally easy, but often solve the inverse of the problem we want. Locked into a distribution (typically Gaussian) Bayesian approaches use both the data and any prior information to develop a posterior distribution. Allows calculation of parameter uncertainty more directly. More easily incorporates outside information.

Conditional Probability If two events, A and B, are related, then if we know B the probability of A happening is: Reversing the events, we get: P (A B) = P (A and B) P (B) P (B A) = P (BandA) P (A) P(B A) should be read as probability of B given A Now, P(A and B) = P (B and A) which gives us the important equality: P (B A) = P (A B) P (B) P (A) This is Bayes Formula. 18

Bayes Theorem Bayes formula is used to merge data with prior information. P (B A) = P (A B) P (B) P (A) A is typically the data, B the statistic we want to know. P(B) is the prior information we may know about the experiment. P(data) is just a normalization constant P (B data) P (data B) P (B)

Example of Bayes Theorem A game show host invites you to choose one of three doors for a chance to win a car (behind one) or a goat (behind the other two). After you choose a door, the host opens another door to reveal a goat. Should you switch your choice? 20

Using Bayes' theorem Assume we are looking for faint companions, and expect them to be around 1% of the stars we observe. From putting in fake companions we know that we can detect objects in the data 90% of the time. From the same tests, we know that we see false planets 3% of the observations. What is the probability that an object we see is actually a planet? P (planet + det.) = P (planet) =0.01 P (noplanet) =0.99 P (+det. planet) =0.9 P ( det. planet) =0.1 P (+det. noplanet) =0.03 P (+det planet)p (planet) P (+det) P (+det.) =P (+det planet)p (planet)+p (+det noplanet)p (noplanet) P (planet + det.) = 0.9 0.01 0.9 0.01 + 0.03 0.99 =0.23

General Bayesian Guidance Focuses on probability rather than accept/reject. Bayesian approaches allow you to calculate probabilities the parameters have a range of values in a more straightforward way. A common concern about Bayesian statistics is that it is subjective. This is not necessarily a problem. Bayesian techniques are generally more computationally intensive, but this is rarely a drawback for modern computers.

Why are Gaussian statistics so pervasive? Even an unusual probability distribution will converge to a Gaussian distribution, for a large enough number, N, of samplings. Referred to as the Central Limit Theorem From statisticalengineering.com

Odd Distributions This works for any unusual distributions that an individual random number may be drawn from: