Statistical Methods in Particle Physics. Lecture 2

Similar documents
Introduction to Statistical Methods for High Energy Physics

Statistische Methoden der Datenanalyse. Kapitel 1: Fundamentale Konzepte. Professor Markus Schumacher Freiburg / Sommersemester 2009

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests

Lectures on Statistical Data Analysis

Statistical Methods for Particle Physics (I)

Statistical Methods in Particle Physics Lecture 1: Bayesian methods

YETI IPPP Durham

The language of random variables

RWTH Aachen Graduiertenkolleg

Statistical Data Analysis 2017/18

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Single Maths B: Introduction to Probability

Lecture 1: Probability Fundamentals

Statistical Methods in Particle Physics

Modern Methods of Data Analysis - SS Introduction

Statistical Methods in Particle Physics

Lecture 5: Bayes pt. 1

2) There should be uncertainty as to which outcome will occur before the procedure takes place.

Econ 325: Introduction to Empirical Economics

Recitation 2: Probability

2. A Basic Statistical Toolbox

Deep Learning for Computer Vision

Priors, Total Probability, Expectation, Multiple Trials

L2: Review of probability and statistics

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Machine Learning. Bayes Basics. Marc Toussaint U Stuttgart. Bayes, probabilities, Bayes theorem & examples

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

With Question/Answer Animations. Chapter 7

Statistical Methods for Astronomy

Topics in Statistical Data Analysis for HEP Lecture 1: Bayesian Methods CERN-JINR European School of High Energy Physics Bautzen, June 2009

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

An introduction to Bayesian reasoning in particle physics

LECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO

Chapter 1 (Basic Probability)

Probability Theory and Applications

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Math Review Sheet, Fall 2008

Probability - Lecture 4

Probability deals with modeling of random phenomena (phenomena or experiments whose outcomes may vary)

CMPSCI 240: Reasoning about Uncertainty

Statistics for the LHC Lecture 1: Introduction

Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak

Machine Learning: Probability Theory

Probability Density Functions

Fourier and Stats / Astro Stats and Measurement : Stats Notes

Probability Review. Yutian Li. January 18, Stanford University. Yutian Li (Stanford University) Probability Review January 18, / 27

Probability. Lecture Notes. Adolfo J. Rumbos

BACKGROUND NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2016 PROBABILITY A. STRANDLIE NTNU AT GJØVIK AND UNIVERSITY OF OSLO

Advanced Herd Management Probabilities and distributions

Physics 6720 Introduction to Statistics April 4, 2017

STA Module 4 Probability Concepts. Rev.F08 1

Random Variables. Definition: A random variable (r.v.) X on the probability space (Ω, F, P) is a mapping

Lecture 2. Conditional Probability

Lecture 2: Review of Basic Probability Theory

18.600: Lecture 3 What is probability?

Preliminary statistics

Bayesian Inference: What, and Why?

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Lecture notes for probability. Math 124

Uncertainty. Variables. assigns to each sentence numerical degree of belief between 0 and 1. uncertainty

Statistical Theory 1

4. Conditional Probability

STT When trying to evaluate the likelihood of random events we are using following wording.

A Probability Primer. A random walk down a probabilistic path leading to some stochastic thoughts on chance events and uncertain outcomes.

Probability Theory. Prof. Dr. Martin Riedmiller AG Maschinelles Lernen Albert-Ludwigs-Universität Freiburg.

Grundlagen der Künstlichen Intelligenz

1 INFO Sep 05

E. Santovetti lesson 4 Maximum likelihood Interval estimation

AST 418/518 Instrumentation and Statistics

Probability and Probability Distributions. Dr. Mohammed Alahmed

Basic Probabilistic Reasoning SEG

2.3 Estimating PDFs and PDF Parameters

Probability, Entropy, and Inference / More About Inference

Lecture 3 Probability Basics

Chapter 14. From Randomness to Probability. Copyright 2012, 2008, 2005 Pearson Education, Inc.

Basic notions of probability theory

Statistics for the LHC Lecture 2: Discovery

Discrete Probability and State Estimation

Frequentist Statistics and Hypothesis Testing Spring

Refresher on Discrete Probability

Statistical Methods for Astronomy

Chapter 2. Conditional Probability and Independence. 2.1 Conditional Probability

Probability & Random Variables

Physics 403. Segev BenZvi. Choosing Priors and the Principle of Maximum Entropy. Department of Physics and Astronomy University of Rochester

Basic notions of probability theory

Discrete Mathematics and Probability Theory Spring 2014 Anant Sahai Note 10

2. AXIOMATIC PROBABILITY

Formal Modeling in Cognitive Science

Introduction Probability. Math 141. Introduction to Probability and Statistics. Albyn Jones

Formal Modeling in Cognitive Science Lecture 19: Application of Bayes Theorem; Discrete Random Variables; Distributions. Background.

STAT Chapter 3: Probability

Bayesian Updating: Discrete Priors: Spring

Statistics 1 - Lecture Notes Chapter 1

1 Preliminaries Sample Space and Events Interpretation of Probability... 13

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg

What are the Findings?

Probability- describes the pattern of chance outcomes

Math 105 Course Outline

Probability Space: Formalism Simplest physical model of a uniform probability space:

Chapter 4. Probability Theory. 4.1 Probability theory Events

Transcription:

Statistical Methods in Particle Physics Lecture 2 October 17, 2011 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2011 / 12

Outline Probability Definition and interpretation Kolmogorov's axioms Bayes' theorem Random variables Probability density functions (pdf) Cumulative distribution function (cdf) 2

Uncertainty In particle physics there are various elements of uncertainty: theory is not deterministic quantum mechanics random measurement errors present even without quantum effects things we could know in principle but don t e.g. from limitations of cost, time,... We can quantify the uncertainty using PROBABILITY 3

Some ingredients Set of elements S S Subsets A, B, of set S A B : A or B (union of A and B) A A B : A and B (intersection of A and B) A : not A (complementary of A) B A B : A contained in B P(A): probability of A A B A B 4

A definition of probability S A B Kolmogorov's axioms (1933) 1. For all A S, P A 0 2. P(S) = 1 3. If A B=0, thenp A B =P A P B Positive definite Normalized Additive 5

Further properties, independence From Kolmogorov's axioms we can derive further properties: P A =1 P A P A A =1 P 0 =0 If A B, thenp A P B P A B =P A P B P A B Subsets A and B are said independent if P A B =P A P B N.B. Do not confuse with disjoint subsets i.e. A B=0 6

Conditional probability Define conditional probability of A, given B (with P(B) 0 ) P A B = P A B P B Example: rolling dice: P n 3 neven = P n 3 neven P even = 1/6 3/6 = 1 3 If A and B independent (see previous page): P A B = P A P B P B = P A 7

Bayes' theorem From the definition of conditional probability we have: P A B = P A B P B and P B A = P B A P A But P A B =P B A, so: P A B = P B A P A P B Bayes' Theorem First published (posthumously) by the Reverend Thomas Bayes (1702 1761) An essay towards solving a problem in the doctrine of chances, Philos. Trans. R. Soc. 53c(1763) 370; reprinted in Biometrika, 45 (1958) 293. 8

The law of total probability Consider a subset B of the sample space S, divided into disjoint subsets A i such that i A i = S : S B A i B = B S = B i A i = i B A i P B = P i B A i = i P B A i B A i P B = i P B A i P A i Law of total probability Bayes' theorem becomes: P A B = P B A P A i P B A i P A i 9

Example: rare disease (1) Suppose the probability (for anyone) to have the disease A is: P A = 0.001 P no A = 0.999 Consider a test for that disease. The result can be 'pos' or 'neg' : P pos A = 0.98 P neg A = 0.02 probabilities to (in)correctly Identify an infected person P pos no A = 0.03 P neg no A = 0.97 probabilities to (in)correctly Identify an infected person Suppose your result is 'pos'. How worried should you be? 10

Example: rare disease (2) The probability to have the disease A, given a 'pos' result is: P A pos = 11

Example: rare disease (2) The probability to have the disease A, given a 'pos' result is: P A pos = = P pos A P A P pos A P A P pos no A P no A 0.98 0.001 0.98 0.001 0.03 0.999 = 0.032 i.e. you re probably OK! Your viewpoint: my degree of belief that I have disease A is 3.2% Your doctor s viewpoint: 3.2% of people like this will have disease A 12

Interpretation of probability 1. Interpretation of probability as RELATIVE FREQUENCY (frequentist approach): A, B,... are outcomes of a repeatable experiment: P A = lim n times outcome is A n See quantum mechanics, particle scattering, radioactive decays 2. SUBJECTIVE PROBABILITY A, B,... are hypotheses (statements that are true or false) P(A) = degree of belief that A is true In particle physics, frequency interpretation often most useful, but subjective probability can provide a more natural treatment of nonrepeatable phenomena (systematic uncertainties, probability that higgs exists...) 13

Frequentist statistics In frequentist statistics, probabilities are associated only with the data, i.e., outcomes of repeatable observations Any given experiment can be considered as one of an infinite sequence of possible repetitions of the same experiment, each capable of producing statistically independent results: Perform experiment N times in identical trials; assume event E occurs k times, then P E = lim k/n N BUT: Does the limit converge? How large needs N to be? What means identical condition? Can 'similar' be sufficient? Not applicable for single events 14

Bayesian probability In Bayesian statistics, use subjective probability for hypotheses (degree of belief that an hypothesis is true): Probability of the data assuming hypothesis H (the likelihood) Prior probability (before seeing the data) P H x = P x H H P x H H dh Posterior probability (after seeing the data) Normalization involves sum over all possible hypothesis Bayes theorem has an if-then character: If your prior probabilities were π(h), then it says how these probabilities should change in the light of the data. No general prescription for priors (subjective!) 15

Back to Bayes' theorem P A B = P B A P A P B Now take: A = theory, B = data: Likelihood Prior P theory data = P data theory P theory P data Posterior Evidence 16

Exercises 1. Show that: P A B =P A P B P A B 2. A beam of particles consists of a fraction 10-4 electrons and the rest photons. The particles pass through a double-layered detector which gives signals in either zero, one or both layers. The probabilities of these outcomes for electrons (e) and photons (γ) are: P 0 e =0.001 P 1 e =0.01 P 2 e =0.989 P 0 =0.99899 P 1 =0.001 P 2 =10 5 (a) what is the probability for a particle detected in one layer only to be a photon? (b) what is the probability for a particle detected in both layers to be an electron? 17

Exercises 3. Detector for particle identification In proton-proton collisions we have: 90% pions, 10% kaons 1. Kaon identification: 95% efficient 2. Pion misidentification: 6% Question: if the particle identification indicates a kaon, what is the probability that it is a real kaon / a real pion? 18

Hints 1. Express AUB as the union of three disjoint sets 2. 3. P A B = P B A P A P B A P A P B no A [1 P A ] 19

Random variables A random variable is a variable whose value results from a measurement on some type of random process. Formally, it is a function from a probability space, typically to the real numbers, which is measurable. Intuitively, a random variable is a numerical description of the outcome of an experiment e.g., the possible results of rolling two dice: (1, 1), (1, 2), etc. Random variables can be classified as: discrete (a random variable that may assume a finite number of values) continuous (a variable that may assume any numerical value in an interval or collection of intervals). 20

Probability density functions Suppose outcome of experiment is continuous value x: P x found in [x, x dx] = f x dx With: f(x) = probability density function (pdf) f x dx=1 Normalization (x must be somewhere) Note: f(x) 0 f(x) is NOT a probability! It has dimension 1/x! 21

Probability density functions Otherwise, for a discrete outcome x i, with i=1,2,... : P x i =p i i P x i =1 Probability mass function x must take on one of its possible values 22

Cumulative distribution function (cdf) Given a pdf f(x'), probability to have outcome less then or equal to x, is: x f x ' dx ' = F x Cumulative distribution function 23

Cumulative distribution function (cdf) x f x ' dx ' = F x Cumulative distribution function F(x) is a continuously non-decreasing function F(- )= 0, F( )=1 For well behaved distributions: pdf : f x = F x x 24

Exercise 1. Given the probability density function: f x = { 1-x for x in [0,2] 0 elsewhere - compute the cdf F(x) - what is the probability to find x > 1.5? - what is the probability to find x in [0.5,1]? 25

Histograms A probability density function can be obtained from a histogram at the limit of: Infinite data sample Zero bin width Normalized to unit area f x = N x n x N = number of entries Δx = bin width 26

Wrapping up lecture 2, next time This lecture: Abstract properties of probability Axioms Interpretations of probability Bayes' theorem Random variables Probability density function Cumulative distribution function Next time: Expectation values Error propagation Catalog of pdfs 27