Lecture 21: Order statistics

Similar documents
Lecture 3 Gaussian Probability Distribution

Section 11.5 Estimation of difference of two proportions

Numerical Analysis: Trapezoidal and Simpson s Rule

Chapter 9: Inferences based on Two samples: Confidence intervals and tests of hypotheses

CS667 Lecture 6: Monte Carlo Integration 02/10/05

Continuous Random Variables

THE EXISTENCE-UNIQUENESS THEOREM FOR FIRST-ORDER DIFFERENTIAL EQUATIONS.

Normal Distribution. Lecture 6: More Binomial Distribution. Properties of the Unit Normal Distribution. Unit Normal Distribution

Math 426: Probability Final Exam Practice

8 Laplace s Method and Local Limit Theorems

For the percentage of full time students at RCC the symbols would be:

Lecture 12: Numerical Quadrature

38.2. The Uniform Distribution. Introduction. Prerequisites. Learning Outcomes

Chapter 5 : Continuous Random Variables

AP Calculus Multiple Choice: BC Edition Solutions

Section 17.2 Line Integrals

The Regulated and Riemann Integrals

Review of Calculus, cont d

Tests for the Ratio of Two Poisson Rates

Joint distribution. Joint distribution. Marginal distributions. Joint distribution

1. Gauss-Jacobi quadrature and Legendre polynomials. p(t)w(t)dt, p {p(x 0 ),...p(x n )} p(t)w(t)dt = w k p(x k ),

Theoretical foundations of Gaussian quadrature

1 Probability Density Functions

Chapter 0. What is the Lebesgue integral about?

1 Linear Least Squares

Math 113 Fall Final Exam Review. 2. Applications of Integration Chapter 6 including sections and section 6.8

Chapter 4 Contravariance, Covariance, and Spacetime Diagrams

Lecture 3. Limits of Functions and Continuity

We partition C into n small arcs by forming a partition of [a, b] by picking s i as follows: a = s 0 < s 1 < < s n = b.

Expectation and Variance

Abstract inner product spaces

Best Approximation. Chapter The General Case

Review of Gaussian Quadrature method

3.4 Numerical integration

Lecture 14: Quadrature

X Z Y Table 1: Possibles values for Y = XZ. 1, p

APPLICATIONS OF THE DEFINITE INTEGRAL

Improper Integrals. Type I Improper Integrals How do we evaluate an integral such as

Data Assimilation. Alan O Neill Data Assimilation Research Centre University of Reading

n f(x i ) x. i=1 In section 4.2, we defined the definite integral of f from x = a to x = b as n f(x i ) x; f(x) dx = lim i=1

Section 6.1 Definite Integral

f(a+h) f(a) x a h 0. This is the rate at which

Monte Carlo method in solving numerical integration and differential equation

Overview of Calculus I

A REVIEW OF CALCULUS CONCEPTS FOR JDEP 384H. Thomas Shores Department of Mathematics University of Nebraska Spring 2007

Space Curves. Recall the parametric equations of a curve in xy-plane and compare them with parametric equations of a curve in space.

INTRODUCTION TO INTEGRATION

ODE: Existence and Uniqueness of a Solution

Properties of Integrals, Indefinite Integrals. Goals: Definition of the Definite Integral Integral Calculations using Antiderivatives

a < a+ x < a+2 x < < a+n x = b, n A i n f(x i ) x. i=1 i=1

Numerical integration

Predict Global Earth Temperature using Linier Regression

Discrete Least-squares Approximations

NUMERICAL INTEGRATION

Review of Riemann Integral

Line Integrals. Partitioning the Curve. Estimating the Mass

LECTURE NOTE #12 PROF. ALAN YUILLE

Problem. Statement. variable Y. Method: Step 1: Step 2: y d dy. Find F ( Step 3: Find f = Y. Solution: Assume

Tutorial 4. b a. h(f) = a b a ln 1. b a dx = ln(b a) nats = log(b a) bits. = ln λ + 1 nats. = log e λ bits. = ln 1 2 ln λ + 1. nats. = ln 2e. bits.

UNIFORM CONVERGENCE. Contents 1. Uniform Convergence 1 2. Properties of uniform convergence 3

MAA 4212 Improper Integrals

Unit #9 : Definite Integral Properties; Fundamental Theorem of Calculus

Problem Set 9. Figure 1: Diagram. This picture is a rough sketch of the 4 parabolas that give us the area that we need to find. The equations are:

63. Representation of functions as power series Consider a power series. ( 1) n x 2n for all 1 < x < 1

Suppose we want to find the area under the parabola and above the x axis, between the lines x = 2 and x = -2.

The area under the graph of f and above the x-axis between a and b is denoted by. f(x) dx. π O

P 3 (x) = f(0) + f (0)x + f (0) 2. x 2 + f (0) . In the problem set, you are asked to show, in general, the n th order term is a n = f (n) (0)

Solution for Assignment 1 : Intro to Probability and Statistics, PAC learning

Math 135, Spring 2012: HW 7

7 - Continuous random variables

CMDA 4604: Intermediate Topics in Mathematical Modeling Lecture 19: Interpolation and Quadrature

Midpoint Approximation

Advanced Calculus: MATH 410 Notes on Integrals and Integrability Professor David Levermore 17 October 2004

Orthogonal Polynomials

Math 32B Discussion Session Session 7 Notes August 28, 2018

Orthogonal Polynomials and Least-Squares Approximations to Functions

The steps of the hypothesis test

Section 6: Area, Volume, and Average Value

Non-Linear & Logistic Regression

Linear Inequalities: Each of the following carries five marks each: 1. Solve the system of equations graphically.

Exam 2, Mathematics 4701, Section ETY6 6:05 pm 7:40 pm, March 31, 2016, IH-1105 Instructor: Attila Máté 1

HW3, Math 307. CSUF. Spring 2007.

Pi evaluation. Monte Carlo integration

Lecture 1. Functional series. Pointwise and uniform convergence.

along the vector 5 a) Find the plane s coordinate after 1 hour. b) Find the plane s coordinate after 2 hours. c) Find the plane s coordinate

1B40 Practical Skills

Czechoslovak Mathematical Journal, 55 (130) (2005), , Abbotsford. 1. Introduction

SYDE 112, LECTURES 3 & 4: The Fundamental Theorem of Calculus

The Fundamental Theorem of Calculus. The Total Change Theorem and the Area Under a Curve.

Jackson 2.26 Homework Problem Solution Dr. Christopher S. Baird University of Massachusetts Lowell

Method: Step 1: Step 2: Find f. Step 3: = Y dy. Solution: 0, ( ) 0, y. Assume

Math 1B, lecture 4: Error bounds for numerical methods

Review of Probability Distributions. CS1538: Introduction to Simulations

Goals: Determine how to calculate the area described by a function. Define the definite integral. Explore the relationship between the definite

Z b. f(x)dx. Yet in the above two cases we know what f(x) is. Sometimes, engineers want to calculate an area by computing I, but...

Riemann is the Mann! (But Lebesgue may besgue to differ.)

p-adic Egyptian Fractions

Properties of the Riemann Integral

Lecture 6: Singular Integrals, Open Quadrature rules, and Gauss Quadrature

Chapter 6 Notes, Larson/Hostetler 3e

Transcription:

Lecture : Order sttistics Suppose we hve N mesurements of sclr, x i =, N Tke ll mesurements nd sort them into scending order x x x 3 x N Define the mesured running integrl S N (x) = 0 for x < x = i/n for x i x < x i+ = for x x N 474

Order sttistics So S N (x) looks like this: /Ν /Ν 0 x x x N 475

Order sttistics We would like to compre this with the predicted x function F(x ) = P(x < x H 0 ) = f x H ) dx ( 0 /Ν /Ν 0 x x x N 476

Order sttistics The most fmous sttistic is D N mx S N (x) F(x) D N /Ν /Ν 0 x x x N 477

Kolmogorov-Smirnov test Turns out the probbility function for N / D N is distribution-free nd clculble for N lrge ( 0 or so) α = P(N / D N > z) = = ( ) exp( z ) z> few exp( z ) 478

Kolmogorov-Smirnov test D N nd P(N ½ D N > z) form the K-S test which is Independent of binning Vlid for smll/medium smple sizes (lso lrge, of course) Loses sensitivity ner endpoints (x or x N ), since both integrls tend to sme vlue (0 or ) (this shortcoming is ddressed in some of the vrints discussed in the book) Like χ, cn be generlized to the cse of compring two mesured distributions to see if they re drwn from different smples 479

Kolmogorov-Smirnov test Compring two different smples, our mesured running integrls re S M nd S N Defining D NM = mx S M (x) S N (x), the probbility of (MN/[M+N]) ½ D NM exceeds given vlue z is gin α = = P([NM/(N = ( ) (i.e. (MN/[M+N]) ½ D NM replces N ½ D N ) + M)] / exp( D z NM ) > z) 480

Crmer-Von Mises-Smirnov test Less well known is the sttistic W = ( S ( x) F( x)) f ( x) dx = ( S ( x) F( x)) df N 0 N The quntity NW is lso distribution free nd hs clculble probbility function α = P(NW > z) = = ( ) / (4 + ) / exp( [4 + ] /6z) K / 4 ([4 + ] /6z) Binomil coeff. Modified Bessel fn 48

Crmer-Von Mises-Smirnov test This test works for very smll smples (N 3 or so) is more sensitive thn K-S ner the endpoints is the most powerful -D test cnnot be esily generlized to compre -D mesured distributions cn be written in more esily evluted form: N (i ) W = + F( xi ) N N i= N 48

Modeling dt So fr, we hve discussed only hypothesis testing. Often, we ssume some composite hypothesis nd wish to constrin its dustble prmeters by experiment Suppose there re m unknown prmeters, (with =,m) nd N dt points 483

Modeling dt The dt re typiclly of two kinds: ) We vry some externl vrible x, nd mke N mesurements of some quntity y to obtin the results y i (x i ) ± σ i (with i =,N) where σ i is the Gussin uncertinty on y i Exmple: we dust the voltge x i = V i cross diode to N vlues nd mesure the current y i = I i ±σ i for ech vlue 484

Modeling dt The dt re typiclly of two kinds: ) We mesure N vlues of the rndom vrible x i in n experiment Exmple: mesurement of individul photon energies in X-ry stronomy 485

Modeling dt In both cses, we would like to determine the prmeter set tht best pproximtes the dt Need n estimtor which mesures how well the model mtches the dt For experiments of the first type, we commonly use the oint Gussin probbility function P( y ) exp( ½χ ( )) = N i= πσ i We wnt to mximize P(y ) by minimizing χ () (nd cn lso see if the model fits by clculting the significnce) 486

Confidence intervls Suppose we hve crried out the minimiztion of χ () The vlue of t the minimum yields our BEST ESTIMATE of the free prmeters, i cll this * Question: wht re our confidence limits on the i? 487

Confidence intervls Becuse our probbility distribution is symmetric under interchnge of nd y i ( x, ) i, we cn use y i P( y) exp( ½χ ( )) = P( y ) = N πσ i= i 488

489 Confidence intervls Let s use Tylor expnsion for χ (): where nd the probbility density is * ) ( k k C = χ Zero *) ( *) ( *) ( *) *)( ( *) ( *) ( ) ( *, * C T k k k + = + + = χ χ χ χ χ *)] ( *) ½( exp[ ) ½ exp( ) ( C P T = χ

Confidence intervls Define Δ = * Then Δχ = χ χ min with Δχ = Δ T C Δ C = Δ Δ T / Δχ which is clled the error mtrix 490

Confidence intervls The digonl elements of the error mtrix (C ) ii tell us (Δ i ) / Δχ for ech prmeter provided tht the nd order Tylor expnsion is dequte Thus, the Byesin probbility for given vlue of Δ i is P(Δ i y) exp( ½Δχ ) exp( ½(Δ i ) / (C ) ii ). Gussin with men 0 nd stndrd devition σ i = (C ) ii 49

The -D problem Let s consider the -D problem in which we hve ust two free prmeters nd The dt yield Δχ (, ) = χ (, ) χ 0 = [c (Δ ) + c (Δ ) +c (Δ Δ )] +. Note tht contours of fixed Δχ re ellipses in the (, ) plne 49

The -D problem Picture: Δχ = *+σ * * σ * σ * *+σ 493

The -D problem We find tht σ = (C ) = c /(c c c ) σ = (C ) = c /(c c c ) The 68% confidence limit on is * ± σ The 68% confidence limit on is * ± σ 494

The -D problem The oint probbility tht nd lie within the Δχ = ellipse is NOT 68%, however. The oint probbility content within the ellipse comes out to be exp( ½Δχ ) = 0.39 nd the ellipse with 68% probbility content hs Δχ =.90 495

The -D problem Esy to prove this for the cse where c = c = c, nd c = c = 0 Δχ = *+σ * * σ σ = σ = σ * σ * *+σ 496

The -D problem In this cse, the oint probbility is P exp( ½Δχ ) d d = exp( ½c( Δ = exp( ½c( r + Δ ))πrdr exp( ½Δχ ) d( Δχ ) )) d d = π exp( ½c( r )) dr nd thus the probbility tht Δχ exceeds some vlue, z, is simply e ½ z 497

Non Gussinity So fr we hve ssumed tht the nd order Tylor expnsion is dequte out to some vlue of Δχ This isn t necessrily the true for the desired vlue of Δχ In tht cse, we hve to plot the ctul contours nd get our confidence limits from them Leds to non-symmetric confidence intervls 498

Non Gussinity Recent exmple: limits on the pressure nd HD/H bundnce rtio in interstellr gs clouds observed with the Spitzer Spce Telescope Δχ = nd 499

Summry: sttisticl tsks nd tools Tsks Sttistic Test/method Hypothesis testing (reection criterion) ) Simple hypothesis H versus simple hypothesis H 0 b) Simple hypothesis H 0 lone Likelihood rtio, P(t H )/P(t H 0 ) Chi-squred Neymn-Person P(t H )/P(t H 0 ) > k P(χ > χ 0 H 0 ) < α c) Composite hypothesis H 0 lone d) Simple hypothesis H 0 lone, where the dt re the counts of smll number of events e) Hypothesis tht two dt sets re drw for the sme distribution Modeling of dt (prmeter fitting) Minimum χ Order sttistic, D N Order sttistic, W Chi-squred Order sttistic, D NM Chi-squred P(χ > χ min H 0 ) < α Kolmogorov-Smirnov Crmer-Von Mises- Smirnov P(χ > χ 0 H 0 ) < α Kolmogorov-Smirnov χ minimiztion with c.l. from Δχ contours or error mtrix 500