Statistical Methods for Particle Physics Lecture 3: Systematics, nuisance parameters

Similar documents
Statistical Methods for Particle Physics Lecture 4: discovery, exclusion limits

Introductory Statistics Course Part II

Some Statistical Tools for Particle Physics

Statistical Methods for Particle Physics Lecture 3: systematic uncertainties / further topics

Recent developments in statistical methods for particle physics

Statistical Methods for Particle Physics

Statistics for the LHC Lecture 2: Discovery

Some Topics in Statistical Data Analysis

Statistical Methods for Discovery and Limits in HEP Experiments Day 3: Exclusion Limits

Discovery significance with statistical uncertainty in the background estimate

Systematic uncertainties in statistical data analysis for particle physics. DESY Seminar Hamburg, 31 March, 2009

Minimum Power for PCL

Statistical Data Analysis Stat 5: More on nuisance parameters, Bayesian methods

Statistical Methods in Particle Physics Lecture 2: Limits and Discovery

Asymptotic formulae for likelihood-based tests of new physics

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests

Statistical Methods in Particle Physics Day 4: Discovery and limits

arxiv: v3 [physics.data-an] 24 Jun 2013

Statistics Challenges in High Energy Physics Search Experiments

Hypothesis testing (cont d)

Discovery Potential for the Standard Model Higgs at ATLAS

Overview of Statistical Methods in Particle Physics

Wald s theorem and the Asimov data set

Statistics Forum News

Statistics for the LHC Lecture 1: Introduction

Statistical Models with Uncertain Error Parameters (G. Cowan, arxiv: )

Introduction to Likelihoods

Confidence Limits and Intervals 3: Various other topics. Roger Barlow SLUO Lectures on Statistics August 2006

Statistical Methods in Particle Physics

Statistical Tools in Collider Experiments. Multivariate analysis in high energy physics

Statistical Methods in Particle Physics Lecture 1: Bayesian methods

arxiv: v1 [hep-ex] 9 Jul 2013

Statistical Methods for Particle Physics Tutorial on multivariate methods

Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate methods

Second Workshop, Third Summary

Statistical Methods for Particle Physics (I)

Statistical Methods for Particle Physics Project on H ττ and multivariate methods

Primer on statistics:

Lecture 5. G. Cowan Lectures on Statistical Data Analysis Lecture 5 page 1

Lecture 2. G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1

Statistical issues for Higgs Physics. Eilam Gross, Weizmann Institute of Science

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Summary of Columbia REU Research. ATLAS: Summer 2014

Search and Discovery Statistics in HEP

Practical Statistics for Particle Physics

Statistics for Particle Physics. Kyle Cranmer. New York University. Kyle Cranmer (NYU) CERN Academic Training, Feb 2-5, 2009

Introduction to Statistical Methods for High Energy Physics

Negative binomial distribution and multiplicities in p p( p) collisions

Use of the likelihood principle in physics. Statistics II

32. STATISTICS. 32. Statistics 1

Physics 403. Segev BenZvi. Credible Intervals, Confidence Intervals, and Limits. Department of Physics and Astronomy University of Rochester

The Higgs boson discovery. Kern-und Teilchenphysik II Prof. Nicola Serra Dr. Annapaola de Cosa Dr. Marcin Chrzaszcz

Lectures on Statistical Data Analysis

FYST17 Lecture 8 Statistics and hypothesis testing. Thanks to T. Petersen, S. Maschiocci, G. Cowan, L. Lyons

Parameter Estimation and Fitting to Data

Search for tt(h bb) using the ATLAS detector at 8 TeV

E. Santovetti lesson 4 Maximum likelihood Interval estimation

32. STATISTICS. 32. Statistics 1

Advanced Statistics Course Part I

Recommendations for presentation of error bars

Statistics for Particle Physics. Kyle Cranmer. New York University. Kyle Cranmer (NYU) CERN Academic Training, Feb 2-5, 2009

Combined Higgs Results

Journeys of an Accidental Statistician

CMS Internal Note. The content of this note is intended for CMS internal use and distribution only

The profile likelihood ratio and the look elsewhere effect in high energy physics

32. STATISTICS. 32. Statistics 1

Topics in Statistical Data Analysis for HEP Lecture 1: Bayesian Methods CERN-JINR European School of High Energy Physics Bautzen, June 2009

Constructing Ensembles of Pseudo-Experiments

Final status of L3 Standard Model Higgs searches and latest results from LEP wide combinations. Andre Holzner ETHZ / L3

P Values and Nuisance Parameters

Physics 736. Experimental Methods in Nuclear-, Particle-, and Astrophysics. - Statistics and Error Analysis -

Bayesian Inference in Astronomy & Astrophysics A Short Course

Modern Methods of Data Analysis - WS 07/08

Higgs Boson Searches with ATLAS based on 2010 Data

Introduction to the Terascale: DESY 2015

Search for SM Higgs Boson at CMS

Error analysis for efficiency

arxiv: v1 [physics.data-an] 7 Mar 2012

RooStatsCms: a tool for analyses modelling, combination and statistical studies

Statistical Methods for Astronomy

PoS(ICHEP2012)238. Search for B 0 s µ + µ and other exclusive B decays with the ATLAS detector. Paolo Iengo

BABAR results on Lepton Flavour Violating (LFV) searches for τ lepton + pseudoscalar mesons & Combination of BABAR and BELLE LFV limits

Measurement And Uncertainty

Bayes Factors for Discovery

Two Early Exotic searches with dijet events at ATLAS

Statistical Data Analysis 2017/18

Statistical Methods for Particle Physics Lecture 2: multivariate methods

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

arxiv: v3 [hep-ph] 16 Feb 2017

Practical Statistics part II Composite hypothesis, Nuisance Parameters

Statistics and Data Analysis

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester

YETI IPPP Durham

RooFit/RooStats Analysis Report

Lecture 24. Introduction to Error Analysis. Experimental Nuclear Physics PHYS 741

Statistical Inference

Search for H ± and H ±± to other states than τ had ν in ATLAS

arxiv: v1 [physics.data-an] 2 Mar 2011

Handling uncertainties in background shapes: the discrete profiling method

Transcription:

Statistical Methods for Particle Physics Lecture 3: Systematics, nuisance parameters http://benasque.org/2018tae/cgi-bin/talks/allprint.pl TAE 2018 Centro de ciencias Pedro Pascual Benasque, Spain 3-15 September 2018 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan G. Cowan TAE 2018 / Statistics Lecture 3 1

Outline Lecture 1: Introduction and review of fundamentals Probability, random variables, pdfs Parameter estimation, maximum likelihood Introduction to statistical tests Lecture 2: More on statistical tests Discovery, limits Bayesian limits Lecture 3: Framework for full analysis Nuisance parameters and systematic uncertainties Tests from profile likelihood ratio Lecture 4: Further topics More parameter estimation, Bayesian methods Experimental sensitivity G. Cowan TAE 2018 / Statistics Lecture 3 2

Approximate confidence intervals/regions from the likelihood function Suppose we test parameter value(s) θ = (θ 1,..., θ n ) using the ratio Lower λ(θ) means worse agreement between data and hypothesized θ. Equivalently, usually define so higher t θ means worse agreement between θ and the data. p-value of θ therefore G. Cowan need pdf TAE 2018 / Statistics Lecture 3 3

Confidence region from Wilks theorem Wilks theorem says (in large-sample limit and providing certain conditions hold...) Assuming this holds, the p-value is chi-square dist. with # d.o.f. = # of components in θ = (θ 1,..., θ n ). where To find boundary of confidence region set p θ = α and solve for t θ : G. Cowan TAE 2018 / Statistics Lecture 3 4

Confidence region from Wilks theorem (cont.) i.e., boundary of confidence region in θ space is where For example, for 1 α = 68.3% and n = 1 parameter, and so the 68.3% confidence level interval is determined by Same as recipe for finding the estimator s standard deviation, i.e., is a 68.3% CL confidence interval. G. Cowan TAE 2018 / Statistics Lecture 3 5

Example of interval from ln L For n = 1 parameter, CL = 0.683, Q α = 1. Exponential example, now with only 5 events: Parameter estimate and approximate 68.3% CL confidence interval: G. Cowan TAE 2018 / Statistics Lecture 3 6

Multiparameter case For increasing number of parameters, CL = 1 α decreases for confidence region determined by a given G. Cowan TAE 2018 / Statistics Lecture 3 7

Multiparameter case (cont.) Equivalently, Q α increases with n for a given CL = 1 α. G. Cowan TAE 2018 / Statistics Lecture 3 8

Systematic uncertainties and nuisance parameters In general our model of the data is not perfect: model: truth: Can improve model by including additional adjustable parameters. x Nuisance parameter systematic uncertainty. Some point in the parameter space of the enlarged model should be true. Presence of nuisance parameter decreases sensitivity of analysis to the parameter of interest (e.g., increases variance of estimate). G. Cowan TAE 2018 / Statistics Lecture 3 9

p-values in cases with nuisance parameters Suppose we have a statistic q θ that we use to test a hypothesized value of a parameter θ, such that the p-value of θ is But what values of ν to use for f (q θ θ, ν)? Fundamentally we want to reject θ only if p θ < α for all ν. exact confidence interval But in general for finite data samples this is not true; one may be unable to reject some θ values if all values of ν must be considered (resulting interval for θ overcovers ). G. Cowan TAE 2018 / Statistics Lecture 3 10

Profile construction ( hybrid resampling ) Approximate procedure is to reject θ if p θ α where the p-value is computed assuming the value of the nuisance parameter that best fits the data for the specified θ: double hat notation means profiled value, i.e., parameter that maximizes likelihood for the given θ. The resulting confidence interval will have the correct coverage for the points (θ, ˆν(θ)). Elsewhere it may under- or overcover, but this is usually as good as we can do (check with MC if crucial or small sample problem). G. Cowan TAE 2018 / Statistics Lecture 3 11

Large sample distribution of the profile likelihood ratio (Wilks theorem, cont.) Suppose problem has likelihood L(θ, ν), with parameters of interest nuisance parameters Want to test point in θ-space. Define profile likelihood ratio: and define q θ = -2 ln λ(θ)., where profiled values of ν Wilks theorem says that distribution f (q θ θ,ν) approaches the chi-square pdf for N degrees of freedom for large sample (and regularity conditions), independent of the nuisance parameters ν. G. Cowan TAE 2018 / Statistics Lecture 3 12

Prototype search analysis Search for signal in a region of phase space; result is histogram of some variable x giving numbers: Assume the n i are Poisson distributed with expectation values where strength parameter signal background G. Cowan TAE 2018 / Statistics Lecture 3 13

Prototype analysis (II) Often also have a subsidiary measurement that constrains some of the background and/or shape parameters: Assume the m i are Poisson distributed with expectation values Likelihood function is nuisance parameters (θ s, θ b,b tot ) G. Cowan TAE 2018 / Statistics Lecture 3 14

The profile likelihood ratio Base significance test on the profile likelihood ratio: maximizes L for specified µ maximize L Define critical region of test of µ by the region of data space that gives the lowest values of λ(µ). Important advantage of profile LR is that its distribution becomes independent of nuisance parameters in large sample limit. G. Cowan TAE 2018 / Statistics Lecture 3 15

Test statistic for discovery Suppose relevant alternative to background-only (µ = 0) is µ 0. So take critical region for test of µ = 0 corresponding to high q 0 and ˆµ > 0 (data characteristic for µ 0). That is, to test background-only hypothesis define statistic i.e. here only large (positive) observed signal strength is evidence against the background-only hypothesis. Note that even though here physically µ 0, we allow ˆµ to be negative. In large sample limit its distribution becomes Gaussian, and this will allow us to write down simple expressions for distributions of our test statistics. G. Cowan TAE 2018 / Statistics Lecture 3 16

Cowan, Cranmer, Gross, Vitells, arxiv:1007.1727, EPJC 71 (2011) 1554 Distribution of q 0 in large-sample limit Assuming approximations valid in the large sample (asymptotic) limit, we can write down the full distribution of q 0 as The special case µ = 0 is a half chi-square distribution: In large sample limit, f(q 0 0) independent of nuisance parameters; f(q 0 µ ) depends on nuisance parameters through σ. G. Cowan TAE 2018 / Statistics Lecture 3 17

p-value for discovery Large q 0 means increasing incompatibility between the data and hypothesis, therefore p-value for an observed q 0,obs is use e.g. asymptotic formula From p-value get equivalent significance, G. Cowan TAE 2018 / Statistics Lecture 3 18

Cowan, Cranmer, Gross, Vitells, arxiv:1007.1727, EPJC 71 (2011) 1554 Cumulative distribution of q 0, significance From the pdf, the cumulative distribution of q 0 is found to be The special case µ = 0 is The p-value of the µ = 0 hypothesis is Therefore the discovery significance Z is simply G. Cowan TAE 2018 / Statistics Lecture 3 19

Cowan, Cranmer, Gross, Vitells, arxiv:1007.1727, EPJC 71 (2011) 1554 Monte Carlo test of asymptotic formula µ = param. of interest b = nuisance parameter Here take s known, τ = 1. Asymptotic formula is good approximation to 5σ level (q 0 = 25) already for b ~ 20. G. Cowan TAE 2018 / Statistics Lecture 3 20

How to read the p 0 plot The local p 0 means the p-value of the background-only hypothesis obtained from the test of µ = 0 at each individual m H, without any correct for the Look-Elsewhere Effect. The Expected (dashed) curve gives the median p 0 under assumption of the SM Higgs (µ = 1) at each m H. ATLAS, Phys. Lett. B 716 (2012) 1-29 The blue band gives the width of the distribution (±1σ) of significances under assumption of the SM Higgs. G. Cowan TAE 2018 / Statistics Lecture 3 21

Cowan, Cranmer, Gross, Vitells, arxiv:1007.1727, EPJC 71 (2011) 1554 Test statistic for upper limits For purposes of setting an upper limit on µ use where I.e. when setting an upper limit, an upwards fluctuation of the data is not taken to mean incompatibility with the hypothesized µ: From observed q µ find p-value: Large sample approximation: 95% CL upper limit on µ is highest value for which p-value is not less than 0.05. G. Cowan TAE 2018 / Statistics Lecture 3 22

Cowan, Cranmer, Gross, Vitells, arxiv:1007.1727, EPJC 71 (2011) 1554 Monte Carlo test of asymptotic formulae Consider again n ~ Poisson (µs + b), m ~ Poisson(τb) Use q µ to find p-value of hypothesized µ values. E.g. f (q 1 1) for p-value of µ =1. Typically interested in 95% CL, i.e., p-value threshold = 0.05, i.e., q 1 = 2.69 or Z 1 = q 1 = 1.64. Median[q 1 0] gives exclusion sensitivity. Here asymptotic formulae good for s = 6, b = 9. G. Cowan TAE 2018 / Statistics Lecture 3 23

How to read the green and yellow limit plots For every value of m H, find the upper limit on µ. Also for each m H, determine the distribution of upper limits µ up one would obtain under the hypothesis of µ = 0. The dashed curve is the median µ up, and the green (yellow) bands give the ± 1σ (2σ) regions of this distribution. ATLAS, Phys. Lett. B 716 (2012) 1-29 G. Cowan TAE 2018 / Statistics Lecture 3 24

Extra slides G. Cowan TAE 2018 / Statistics Lecture 3 25

Low sensitivity to µ It can be that the effect of a given hypothesized µ is very small relative to the background-only (µ = 0) prediction. This means that the distributions f(q µ µ) and f(q µ 0) will be almost the same: G. Cowan TAE 2018 / Statistics Lecture 3 26

Having sufficient sensitivity In contrast, having sensitivity to µ means that the distributions f(q µ µ) and f(q µ 0) are more separated: That is, the power (probability to reject µ if µ = 0) is substantially higher than α. Use this power as a measure of the sensitivity. G. Cowan TAE 2018 / Statistics Lecture 3 27

Spurious exclusion Consider again the case of low sensitivity. By construction the probability to reject µ if µ is true is α (e.g., 5%). And the probability to reject µ if µ = 0 (the power) is only slightly greater than α. This means that with probability of around α = 5% (slightly higher), one excludes hypotheses to which one has essentially no sensitivity (e.g., m H = 1000 TeV). Spurious exclusion G. Cowan TAE 2018 / Statistics Lecture 3 28

Ways of addressing spurious exclusion The problem of excluding parameter values to which one has no sensitivity known for a long time; see e.g., In the 1990s this was re-examined for the LEP Higgs search by Alex Read and others and led to the CL s procedure for upper limits. Unified intervals also effectively reduce spurious exclusion by the particular choice of critical region. G. Cowan TAE 2018 / Statistics Lecture 3 29

The CL s procedure In the usual formulation of CL s, one tests both the µ = 0 (b) and µ > 0 (µs+b) hypotheses with the same statistic Q = -2ln L s+b /L b : f (Q s+b) f (Q b) p b p s+b G. Cowan TAE 2018 / Statistics Lecture 3 30

The CL s procedure (2) As before, low sensitivity means the distributions of Q under b and s+b are very close: f (Q s+b) f (Q b) p b p s+b G. Cowan TAE 2018 / Statistics Lecture 3 31

The CL s procedure (3) The CL s solution (A. Read et al.) is to base the test not on the usual p-value (CL s+b ), but rather to divide this by CL b (~ one minus the p-value of the b-only hypothesis), i.e., Define: f (Q s+b) f (Q b) Reject s+b hypothesis if: 1-CL b = p b CL s+b = p s+b Increases effective p-value when the two distributions become close (prevents exclusion if sensitivity is low). G. Cowan TAE 2018 / Statistics Lecture 3 32

Setting upper limits on µ = σ/σ SM Carry out the CLs procedure for the parameter µ = σ/σ SM, resulting in an upper limit µ up. In, e.g., a Higgs search, this is done for each value of m H. At a given value of m H, we have an observed value of µ up, and we can also find the distribution f(µ up 0): ±1σ (green) and ±2σ (yellow) bands from toy MC; Vertical lines from asymptotic formulae. G. Cowan TAE 2018 / Statistics Lecture 3 33