ECE531 Lecture 13: Sequential Detection of Discrete-Time Signals

Size: px
Start display at page:

Download "ECE531 Lecture 13: Sequential Detection of Discrete-Time Signals"

Transcription

1 ECE531 Lecture 13: Sequential Detection of Discrete-Time Signals D. Richard Brown III Worcester Polytechnic Institute 30-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

2 Introduction All of the detection problems we have considered in this class have involved a fixed sample size n with the decision rule only being applied after all n samples are available. s k a + Y k n-sample detector decisions W k We now consider the sequential detection problem where we collect samples one at a time until we have enough observations to generate a final/terminal decision. There are two approaches to this problem: Bayes: Assign a relative cost to taking an observation and try to minimize the overall average cost. N-P: Keep taking observations until a desired tradeoff can be obtained between P fp and P D. Good decision rules achieve the desired tradeoff with fewer observations. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

3 Mathematical Model for Sequential Detection We consider only simple binary hypotheses here: H 0 : X = x 0 and H 1 : X = x 1. We assume that the scalar observations are i.i.d. random variables with density p 0 (z) under H 0 and p 1 (z) under H 1. The set of possible observation sequences can be written as Y = {y : y = {y k } k=1,y k R} A sequential decision rule (SDR) is a pair of sequences φ = {φ k } k=0 (stopping rule) δ = {δ k } k=0 (terminal decision rule) where φ k : R k {0,1} and δ k : R k {0,1}. Note that the SDR begins at index k = 0 whereas the first sample occurs at index k = 1. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

4 Sequential Decision Rules SDR procedure: If φn (y 1,...,y n ) = 0, we take another sample. If φn (y 1,...,y n ) = 1, we stop sampling and make the terminal decision δ n (y 1,..., y n ). The stopping time is the random variable N(φ) = min{n : φ n (Y 1,...,Y n ) = 1} Under our assumption of simple binary hypotheses, the performance metrics of interest are the conditional risks R 0 (δ) and R 1 (δ) as well as the expected stopping times under each state E 0 [N(φ)] := E[N(φ) X = x 0 ] and E 1 [N(φ)] := E[N(φ) X = x 1 ]. We would like to minimize all of these quantities but each can be traded off against the other. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

5 Bayes Formulation for Sequential Detection We assume the UCA. If we assign a cost of c to each observation, the conditional risks can be written as and R 0 (φ,δ) = Prob(decide H 1 X = x 0 ) + ce 0 [N(φ)] = E 0 [δ(y 1,...,Y N )] + ce 0 [N(φ)] R 1 (φ,δ) = Prob(decide H 0 X = x 1 ) + ce 1 [N(φ)] = 1 E 1 [δ(y 1,...,Y N )] + ce 1 [N(φ)]. Given a prior π 0 = Prob(X = x 0 ), π 1 = Prob(X = x 1 ), we can combine the conditional risks into a single Bayes risk r(φ,δ,π 1 ) = (1 π 1 )R 0 (φ,δ) + π 1 R 1 (φ,δ). We seek a SDR (φ,δ) that minimizes r(φ,δ,π 1 ). Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

6 Step 0: No Samples Taken Yet We have three choices: decide 0, decide 1, or take a sample. If φ 0 = 1, we are choosing to make a decision without any observations. Is this a good idea? Actually, it might be if the cost of an observation is very high, the prior π 1 is very close to zero or one, or the observations are very noisy. If we stop, the risks of each possible decision rule are r(φ 0 = 1, δ 0 = 0, π 1 ) = π 1 r(φ 0 = 1, δ 0 = 1, π 1 ) = 1 π 1 If we choose to take a sample, the minimum Bayes risk will then be V (π 1 ) = min (1 π 1 )R 0 (φ, δ) + π 1 R 1 (φ, δ) c φ : φ 0 = 0 δ V (π 1 ) is the minimum expected risk over all SDRs that take at least one sample. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

7 Graphical Intuituion: Minimum Risk Function 1 π 1 π 1 1 π 1 π 1 Bayes risk V (π 1 ) V (π 1 ) c π L π U c c prior π 1 prior π 1 Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

8 Step 0: No Samples Taken Yet It should be clear that our decision to take a sample at step 0 is determined by whether the quantity V (π 1 ) is smaller than both π 1 and 1 π 1. Lemma When the costs C 00 = C 11 = 0, then V (0) = V (1) = c and V (x) is a continous concave function on x [0,1]. Let P = {x [0,1] : V (x) min{x,1 x}}. If P is empty, then we definitely should not take a sample. We should just decide 1 when π 1 0.5, otherwise we should decide 0. If P is not empty then P = [π L,π U ] [0,1] and the best action at step 0 is now clear: π 1 π U : stop and decide 1 (φ 0 = 1,δ 0 = 1) π 1 π L : stop and decide 0 (φ 0 = 1,δ 0 = 0) π L < π 1 < π U : take a sample (φ 0 = 0,δ 0 =?) Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

9 Minimum Risk Function: Binary Signals in Gaussian Noise no observations no observations min risk one observation min risk two observations min risk three observations min risk four observations V * 0.3 Bayes risk prior π 1 Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

10 Example: Binary Signaling The previous slide plots the risk for the HT problem: H 0 : Y k N( 1,σ 2 ) H 1 : Y k N(1,σ 2 ) where σ 2 = 0.6 is known and c = According to the minimum risk curves on the previous slide, if our prior was π 1 = 0.5, we can minimize the risk by taking two samples. So the optimum SDR should take two samples, right? What if our first sample turned out to be y 1 = 10? Should we take another sample? The sample y 1 = 10 is very strong evidence in favor of H 1. Is another sample likely to lower our risk after observing y 1 = 10? A sequential decision rule does not decide in advance how many samples it should take. The decision to take another sample is made dynamically and depends on the outcomes of the previous samples. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

11 Step 1: One Sample Taken: What Should We Do? Given no samples, we had a prior probability π 1 = Prob(X = x 1 ). If we decided to take one sample and observed Y 1 = y 1, we can compute a posterior probability π 1 (y 1 ) := Prob(X = x 1 Y 1 = y 1 ) π 1 p 1 (y 1 ) = π 0 p 0 (y 1 ) + π 1 p 1 (y 1 ) L(y 1 ) = 1 π 1 π 1 + L(y 1 ) We have three choices at this point: decide 0, decide 1, or take another sample. Since we ve assumed i.i.d. samples and the cost of taking another sample is the same, the problem at step n > 0 is the same problem that we faced in step 0, the only difference being the updated prior. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

12 Example: Coin Flipping Suppose we have the following simple binary HT problem: H 0 : coin is fair H 1 : coin is unfair with Prob(H) = q > 0.5 Let π 1 = Prob(state is H 1 ) and assume the UCA. If we have take no samples, the risks will be r(δ 1,π) = π 1 (always decide fair) r(δ 2,π) = 1 π 1 (always decide unfair) If we take one sample, the risks will be r(δ 3, π) = 1 π1 2 r(δ 4, π) = 1 π1 2 + π 1(1 q) + c (y 1 =T, decide fair; y 1 =H, decide unfair) + π 1q + c (y 1 =T, decide unfair; y 1 =H, decide fair) To find V, we have to keep doing this for 2, 3, 4,... samples. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

13 Example: Coin Flipping: c = 0.05 and q = rule 1 rule 2 rule 3 + c rule 4 + c Bayes risk prior π 1 Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

14 Example: Coin Flipping: c = 0.05 and q = Bayes risk prior π 1 Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

15 Example: Coin Flipping: c = 0.05 Suppose we have a prior π 1 = 0.5. What should we do? Take a sample since taking one sample and using rule 3 has a lower Bayes risk at this prior (including the cost of the sample). The coin is flipped and you observe a head. What should you do? π 1 (y 1 = H) = π 1 p 1 (y 1 = H) π 0 p 0 (y 1 = H) + π 1 p 1 (y 1 = H) = Now what should we do? It looks like we need to take another sample. The coin is flipped and you observe a tail. What should you do? π 1({y 1, y 2} = {H,T}) = π 1(y 1 = H)p 1(y 2 = T) π 0(y 1 = H)p 0(y 2 = T) + π 1(y 1 = H)p 1(y 2 = T) Now what should we do? It is close. Let s stop and decide H 0. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

16 Example: Binary Signals in Gaussian Noise no observations no observations min risk one observation min risk two observations min risk three observations min risk four observations V * Bayes risk posterior probability π 1 (y) prior π number of samples taken a 1 = a 0 = 1; σ 2 = 0.6; c = 0.05 Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

17 Bayesian Formulation: Remarks When c > 0, we stop taking samples when the prior has been updated to a point where the expected minimum risk of taking one or more samples exceeds the risk of just making a decision now. If each sample were free, we could keep taking samples and make the minimum risk V (π 1 ) go to zero over all priors, e.g. coin flipping: Bayes risk prior π 1 Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

18 Neyman-Pearson Formulation Samples no longer have an explicit cost. We have already investigated the case when the number of samples is fixed. We can design decision rules to achieve a two-way tradeoff between P fp and P D. Now, we would like to design a sequential decision rule that achieves a desired three-way tradeoff between P fp, P D, and the expected number of samples. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

19 Sequential Probability Ratio Test Given an n-sample observation {y k } n k=1, and a prior π 1, the posterior can be written as where π 1 ({y k } n k=1 ) = π 1 p 1 ({y k } n k=1 ) (1 π 1 )p 0 ({y k } n k=1 ) + π 1p 1 ({y k } n k=1 ) L({y k } n k=1 = ) 1 π 1 π 1 + L({y k } n k=1 ) = λ n 1 π 1 + λ n λ n := L({y k } n k=1 ) = p 1({y k } n k=1 ) p 0 ({y k } n k=1 ) = n k=1 π 1 p 1 (y k ) n p 0 (y k ) = L(y k ) π 1 ({y k } n k=1 ) is a strictly monotone increasing function of λ n when 0 < π 1 < 1. Hence testing π 1 ({y k } n k=1 ) π U and π 1 ({y k } n k=1 ) π L is equivalent to testing λ n B and λ n A. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32 k=1

20 Sequential Probability Ratio Test (SPRT) A Sequential Probability Ratio Test, denoted as SPRT(A,B), is a sequential decision rule of the form (1,1) λ n B (stop and decide 1) (φ n ({y k } n k=1 ),δ n({y k } n k=1 )) = (1,0) λ n A (stop and decide 0) (0,?) A < λ n < B (take a sample) How are A and B related to π L and π U? Set π 1 ({y k } n k=1 ) equal to π L or π U and solve for λ n... A = B = (1 π 1)π L π 1 (1 π L ) (1 π 1)π U π 1 (1 π U ) Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

21 Wald-Wolfowitz Theorem Theorem Let (φ,δ) be an SPRT with parameters (A,B) and let (φ,δ ) be any sequential decision rule. If P fp (φ,δ ) P D (φ,δ ) P fp (φ,δ) and P D (φ,δ) then E j [N(φ )] E j [N(φ)], j = 0 and 1. The theorem says that no sequential decision rule can achieve a better false positive probability and a better detection probability than an SPRT without also requiring more samples, on average, than the SPRT. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

22 Wald-Wolfowitz Theorem: SPRT vs. n-sample Detector We can reformulate our standard n-sample detector as a sequential decision rule (φ,δ ) with fixed observation size n: φ n({y k } n k=1) = {0,..., 0, 1} where the 1 occurs in the nth position δ n({y k } n k=1) = {?,...,?, δ({y k } n k=1)} where the δ occurs in the nth position Suppose we design an n-sample N-P decision rule δ with P fp = α and P D = β. What is E j [N(φ )] for j = 0 and 1? Now consider any sequential decision rule of the SPRT form that achieves P fp α and P D β. Then the Wald-Wolfowitz theorem says that E j [N(φ)] n for j = 0 and 1. Hence, the expected number of samples required by an SPRT (under each hypothesis) is no larger than that of a fixed decision rule with the same performance. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

23 How to Select A and B to Achieve the Desired Tradeoff We define the sets S 1 n = {y = {y k } k=1 : j = 1,...,n 1,A < λ j < B and λ n B} S 0 n = {y = {y k } k=1 : j = 1,...,n 1,A < λ j < B and λ n A} Note that S i n is the set of all observation sequences on which SPRT(A,B) decides H i after exactly n samples. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

24 How to Select A and B to Achieve the Desired Tradeoff The probability of false positive of SPRT(A,B) is then P fp = = Prob({y k } n k=1 S1 n X = x 0) n=0 n=0 Sn 1 j=1 B 1 n=0 n p 0 (y j )dy Sn 1 j=1 n p 1 (y j )dy = B 1 Prob({y k } n k=1 S1 n X = x 1) n=0 = B 1 P D where the inequality results from the fact that λ n = Q n j=1 p 1(y j ) Q n j=1 p 0(y j ) B. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

25 How to Select A and B to Achieve the Desired Tradeoff The probability of missed detection (false negative) of SPRT(A,B) can also be written as P fn = Prob({y k } n k=1 S0 n X = x 1 ) = n=0 n=0 A = A Sn 0 j=1 n=0 n p 1 (y j )dy Sn 0 j=1 n p 0 (y j )dy Prob({y k } n k=1 S0 n X = x 0 ) n=0 = A(1 P fp ) where the inequality results from the fact that λ n = Q n j=1 p 1(y j ) Q n j=1 p 0(y j ) A. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

26 How to Select A and B to Achieve the Desired Tradeoff Putting it all together, we can write two linear inequality constraints BP fp + P fn 1 AP fp + P fn A or, in terms of P D = 1 P fn, BP fp P D 0 AP fp P D A 1 These inequalities constrain P fp and P D in terms of our SPRT design parameters A and B. Note that we certainly want B > A and typically A < 1 and B > 1. Inequalities are necessary here since λ n may not exactly hit the boundary (A or B) but may jump over it. Under the assumption that each sample provides only a small amount of information, the amount of overshooot, e.g. λ n = B + ǫ or λ n = A ǫ, will usually be small. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

27 Linear Inequality Constraints B A PD Pfp Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

28 Solving for A and B in terms of P fp and P D Our inequalities on A and B in terms of P fp and P D : A B 1 P D 1 P fp P D P fp Under the small overshoot assumption, we can assume that these inequalities are approximate equalities. Hence, if we require a N-P decision rule with P fp = α and P D = β, we should choose the SPRT parameters A B 1 β 1 α β α Note that the lines y = (1 A) + Ax and y = Bx intersect at the point (α,β) when A and B are chosen according to (1) and (2). Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32 (1) (2)

29 Example: Sequential Detection of Unfair Coins Suppose we have the following simple binary HT problem: H 0 : coin is fair H 1 : coin is unfair with Prob(H) = q > 0.5 Suppose we require P fp = 0.1 and P D = 0.8. We can solve for the SPRT parameters A = 2 9 B = 8 Hence, our SPRT should take samples until λ n 8, in which case we decide H 1, or until λ n 2/9, in which case we decide H 0. Note that λ n = p 1 ({y j } n j=1 )/p 0({y j } n j=1 ) = qk (1 q) n k /0.5 n where k is the number of heads and n is the number of samples/flips. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

30 Example: Sequential Detection of Unfair Coins q = likelihood ratio number of samples taken Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

31 Example: Sequential Detection of Unfair Coins q = likelihood ratio number of samples taken Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

32 Final Remarks Two approaches to sequential detection: Bayes: Assign a cost of c to each sample and a prior π to the states. Take samples until enough evidence is obtained such that the expected cost of additional samples is larger than the expected risk reduction. N-P: Three-way tradeoff between PD, P fp, and the number of samples. Please read example III.D.2 for a comparison of SPRT and fixed-sample-size (FSS) detectors. Comments on page 110 of your textbook are also enlightening. Even though the average number of samples for an SPRT is usually much less than a FSS detector, the actual number of samples for a particular sample sequence is unbounded! Sequential decision rules have limited application when the samples are not i.i.d. Worcester Polytechnic Institute D. Richard Brown III 30-Apr / 32

ECE531 Lecture 4b: Composite Hypothesis Testing

ECE531 Lecture 4b: Composite Hypothesis Testing ECE531 Lecture 4b: Composite Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute 16-February-2011 Worcester Polytechnic Institute D. Richard Brown III 16-February-2011 1 / 44 Introduction

More information

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters

ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters D. Richard Brown III Worcester Polytechnic Institute 26-February-2009 Worcester Polytechnic Institute D. Richard Brown III 26-February-2009

More information

ECE531 Lecture 2a: A Mathematical Model for Hypothesis Testing (Finite Number of Possible Observations)

ECE531 Lecture 2a: A Mathematical Model for Hypothesis Testing (Finite Number of Possible Observations) ECE531 Lecture 2a: A Mathematical Model for Hypothesis Testing (Finite Number of Possible Observations) D. Richard Brown III Worcester Polytechnic Institute 26-January-2011 Worcester Polytechnic Institute

More information

ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations

ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 7 Neyman

More information

ECE531 Lecture 2b: Bayesian Hypothesis Testing

ECE531 Lecture 2b: Bayesian Hypothesis Testing ECE531 Lecture 2b: Bayesian Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute 29-January-2009 Worcester Polytechnic Institute D. Richard Brown III 29-January-2009 1 / 39 Minimizing

More information

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing

ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing ECE531 Screencast 11.4: Composite Neyman-Pearson Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 8 Basics Hypotheses H 0

More information

ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules

ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules ECE531 Screencast 11.5: Uniformly Most Powerful Decision Rules D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 9 Monotone Likelihood Ratio

More information

Introduction to Bayesian Statistics

Introduction to Bayesian Statistics Bayesian Parameter Estimation Introduction to Bayesian Statistics Harvey Thornburg Center for Computer Research in Music and Acoustics (CCRMA) Department of Music, Stanford University Stanford, California

More information

Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1)

Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Detection problems can usually be casted as binary or M-ary hypothesis testing problems. Applications: This chapter: Simple hypothesis

More information

ECE531 Homework Assignment Number 2

ECE531 Homework Assignment Number 2 ECE53 Homework Assignment Number 2 Due by 8:5pm on Thursday 5-Feb-29 Make sure your reasoning and work are clear to receive full credit for each problem 3 points A city has two taxi companies distinguished

More information

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary

Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary ECE 830 Spring 207 Instructor: R. Willett Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we saw that the likelihood

More information

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?

2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)? ECE 830 / CS 76 Spring 06 Instructors: R. Willett & R. Nowak Lecture 3: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we

More information

ECE531 Lecture 8: Non-Random Parameter Estimation

ECE531 Lecture 8: Non-Random Parameter Estimation ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction

More information

COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference

COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference COMP 551 Applied Machine Learning Lecture 19: Bayesian Inference Associate Instructor: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~jpineau/comp551 Unless otherwise noted, all material posted

More information

Sequential Decisions

Sequential Decisions Sequential Decisions A Basic Theorem of (Bayesian) Expected Utility Theory: If you can postpone a terminal decision in order to observe, cost free, an experiment whose outcome might change your terminal

More information

Detection theory 101 ELEC-E5410 Signal Processing for Communications

Detection theory 101 ELEC-E5410 Signal Processing for Communications Detection theory 101 ELEC-E5410 Signal Processing for Communications Binary hypothesis testing Null hypothesis H 0 : e.g. noise only Alternative hypothesis H 1 : signal + noise p(x;h 0 ) γ p(x;h 1 ) Trade-off

More information

Decentralized Sequential Hypothesis Testing. Change Detection

Decentralized Sequential Hypothesis Testing. Change Detection Decentralized Sequential Hypothesis Testing & Change Detection Giorgos Fellouris, Columbia University, NY, USA George V. Moustakides, University of Patras, Greece Outline Sequential hypothesis testing

More information

DETECTION theory deals primarily with techniques for

DETECTION theory deals primarily with techniques for ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for

More information

Lecture 2: From Linear Regression to Kalman Filter and Beyond

Lecture 2: From Linear Regression to Kalman Filter and Beyond Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation

More information

ECE531: Principles of Detection and Estimation Course Introduction

ECE531: Principles of Detection and Estimation Course Introduction ECE531: Principles of Detection and Estimation Course Introduction D. Richard Brown III WPI 22-January-2009 WPI D. Richard Brown III 22-January-2009 1 / 37 Lecture 1 Major Topics 1. Web page. 2. Syllabus

More information

Sequential Decisions

Sequential Decisions Sequential Decisions A Basic Theorem of (Bayesian) Expected Utility Theory: If you can postpone a terminal decision in order to observe, cost free, an experiment whose outcome might change your terminal

More information

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm Math 408 - Mathematical Statistics Lecture 29-30. Testing Hypotheses: The Neyman-Pearson Paradigm April 12-15, 2013 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 1 / 12 Agenda Example:

More information

BAYESIAN DECISION THEORY

BAYESIAN DECISION THEORY Last updated: September 17, 2012 BAYESIAN DECISION THEORY Problems 2 The following problems from the textbook are relevant: 2.1 2.9, 2.11, 2.17 For this week, please at least solve Problem 2.3. We will

More information

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Applied Mathematical Sciences, Vol. 4, 2010, no. 62, 3083-3093 Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Julia Bondarenko Helmut-Schmidt University Hamburg University

More information

Detection theory. H 0 : x[n] = w[n]

Detection theory. H 0 : x[n] = w[n] Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal

More information

Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate methods

Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate methods Statistical Methods for Particle Physics Lecture 2: statistical tests, multivariate methods www.pp.rhul.ac.uk/~cowan/stat_aachen.html Graduierten-Kolleg RWTH Aachen 10-14 February 2014 Glen Cowan Physics

More information

hypothesis testing 1

hypothesis testing 1 hypothesis testing 1 Does smoking cause cancer? competing hypotheses (a) No; we don t know what causes cancer, but smokers are no more likely to get it than nonsmokers (b) Yes; a much greater % of smokers

More information

ECE531 Lecture 10b: Maximum Likelihood Estimation

ECE531 Lecture 10b: Maximum Likelihood Estimation ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So

More information

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing

ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing Robert Vanderbei Fall 2014 Slides last edited on November 24, 2014 http://www.princeton.edu/ rvdb Coin Tossing Example Consider two coins.

More information

INTRODUCTION TO BAYESIAN INFERENCE PART 2 CHRIS BISHOP

INTRODUCTION TO BAYESIAN INFERENCE PART 2 CHRIS BISHOP INTRODUCTION TO BAYESIAN INFERENCE PART 2 CHRIS BISHOP Personal Healthcare Revolution Electronic health records (CFH) Personal genomics (DeCode, Navigenics, 23andMe) X-prize: first $10k human genome technology

More information

Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box Durham, NC 27708, USA

Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box Durham, NC 27708, USA Testing Simple Hypotheses R.L. Wolpert Institute of Statistics and Decision Sciences Duke University, Box 90251 Durham, NC 27708, USA Summary: Pre-experimental Frequentist error probabilities do not summarize

More information

Bayesian Updating with Discrete Priors Class 11, Jeremy Orloff and Jonathan Bloom

Bayesian Updating with Discrete Priors Class 11, Jeremy Orloff and Jonathan Bloom 1 Learning Goals ian Updating with Discrete Priors Class 11, 18.05 Jeremy Orloff and Jonathan Bloom 1. Be able to apply theorem to compute probabilities. 2. Be able to define the and to identify the roles

More information

Cost-Aware Bayesian Sequential Decision-Making for Domain Search and Object Classification

Cost-Aware Bayesian Sequential Decision-Making for Domain Search and Object Classification Cost-Aware Bayesian Sequential Decision-Making for Domain Search and Object Classification Y. Wang, I. I. Hussein, D. R. Brown III, and R. S. Erwin Abstract This paper focuses on the development of a cost-aware

More information

Statistics for the LHC Lecture 1: Introduction

Statistics for the LHC Lecture 1: Introduction Statistics for the LHC Lecture 1: Introduction Academic Training Lectures CERN, 14 17 June, 2010 indico.cern.ch/conferencedisplay.py?confid=77830 Glen Cowan Physics Department Royal Holloway, University

More information

Introduction to Statistical Inference

Introduction to Statistical Inference Structural Health Monitoring Using Statistical Pattern Recognition Introduction to Statistical Inference Presented by Charles R. Farrar, Ph.D., P.E. Outline Introduce statistical decision making for Structural

More information

Detection and Estimation Chapter 1. Hypothesis Testing

Detection and Estimation Chapter 1. Hypothesis Testing Detection and Estimation Chapter 1. Hypothesis Testing Husheng Li Min Kao Department of Electrical Engineering and Computer Science University of Tennessee, Knoxville Spring, 2015 1/20 Syllabus Homework:

More information

Bayesian Data Analysis

Bayesian Data Analysis An Introduction to Bayesian Data Analysis Lecture 1 D. S. Sivia St. John s College Oxford, England April 29, 2012 Some recommended books........................................................ 2 Outline.......................................................................

More information

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix

More information

Ways to make neural networks generalize better

Ways to make neural networks generalize better Ways to make neural networks generalize better Seminar in Deep Learning University of Tartu 04 / 10 / 2014 Pihel Saatmann Topics Overview of ways to improve generalization Limiting the size of the weights

More information

Detection and Estimation Theory

Detection and Estimation Theory ESE 524 Detection and Estimation Theory Joseph A. O Sullivan Samuel C. Sachs Professor Electronic Systems and Signals Research Laboratory Electrical and Systems Engineering Washington University 2 Urbauer

More information

JOINT DETECTION-STATE ESTIMATION AND SECURE SIGNAL PROCESSING

JOINT DETECTION-STATE ESTIMATION AND SECURE SIGNAL PROCESSING Virginia Commonwealth University VCU Scholars Compass Theses and Dissertations Graduate School 2016 JOINT DETECTION-STATE ESTIMATION AND SECURE SIGNAL PROCESSING Mengqi Ren Virginia Commonwealth University

More information

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2

Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Logistics CSE 446: Point Estimation Winter 2012 PS2 out shortly Dan Weld Some slides from Carlos Guestrin, Luke Zettlemoyer & K Gajos 2 Last Time Random variables, distributions Marginal, joint & conditional

More information

How Much Evidence Should One Collect?

How Much Evidence Should One Collect? How Much Evidence Should One Collect? Remco Heesen October 10, 2013 Abstract This paper focuses on the question how much evidence one should collect before deciding on the truth-value of a proposition.

More information

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Classical Hypothesis Testing: The Likelihood Ratio Test. Department of Physics and Astronomy University of Rochester Physics 403 Classical Hypothesis Testing: The Likelihood Ratio Test Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Bayesian Hypothesis Testing Posterior Odds

More information

Multivariate statistical methods and data mining in particle physics

Multivariate statistical methods and data mining in particle physics Multivariate statistical methods and data mining in particle physics RHUL Physics www.pp.rhul.ac.uk/~cowan Academic Training Lectures CERN 16 19 June, 2008 1 Outline Statement of the problem Some general

More information

Bayesian Methods: Naïve Bayes

Bayesian Methods: Naïve Bayes Bayesian Methods: aïve Bayes icholas Ruozzi University of Texas at Dallas based on the slides of Vibhav Gogate Last Time Parameter learning Learning the parameter of a simple coin flipping model Prior

More information

Statistical Inference

Statistical Inference Statistical Inference Robert L. Wolpert Institute of Statistics and Decision Sciences Duke University, Durham, NC, USA Spring, 2006 1. DeGroot 1973 In (DeGroot 1973), Morrie DeGroot considers testing the

More information

Sequential Detection. Changes: an overview. George V. Moustakides

Sequential Detection. Changes: an overview. George V. Moustakides Sequential Detection of Changes: an overview George V. Moustakides Outline Sequential hypothesis testing and Sequential detection of changes The Sequential Probability Ratio Test (SPRT) for optimum hypothesis

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These

More information

7. Statistical estimation

7. Statistical estimation 7. Statistical estimation Convex Optimization Boyd & Vandenberghe maximum likelihood estimation optimal detector design experiment design 7 1 Parametric distribution estimation distribution estimation

More information

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,

If there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell, Recall The Neyman-Pearson Lemma Neyman-Pearson Lemma: Let Θ = {θ 0, θ }, and let F θ0 (x) be the cdf of the random vector X under hypothesis and F θ (x) be its cdf under hypothesis. Assume that the cdfs

More information

Bayesian Learning (II)

Bayesian Learning (II) Universität Potsdam Institut für Informatik Lehrstuhl Maschinelles Lernen Bayesian Learning (II) Niels Landwehr Overview Probabilities, expected values, variance Basic concepts of Bayesian learning MAP

More information

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior:

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior: Pi Priors Unobservable Parameter population proportion, p prior: π ( p) Conjugate prior π ( p) ~ Beta( a, b) same PDF family exponential family only Posterior π ( p y) ~ Beta( a + y, b + n y) Observed

More information

Appendix A : Introduction to Probability and stochastic processes

Appendix A : Introduction to Probability and stochastic processes A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of

More information

Detection Theory. Chapter 3. Statistical Decision Theory I. Isael Diaz Oct 26th 2010

Detection Theory. Chapter 3. Statistical Decision Theory I. Isael Diaz Oct 26th 2010 Detection Theory Chapter 3. Statistical Decision Theory I. Isael Diaz Oct 26th 2010 Outline Neyman-Pearson Theorem Detector Performance Irrelevant Data Minimum Probability of Error Bayes Risk Multiple

More information

A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I. kevin small & byron wallace

A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I. kevin small & byron wallace A.I. in health informatics lecture 2 clinical reasoning & probabilistic inference, I kevin small & byron wallace today a review of probability random variables, maximum likelihood, etc. crucial for clinical

More information

Machine Learning CMPT 726 Simon Fraser University. Binomial Parameter Estimation

Machine Learning CMPT 726 Simon Fraser University. Binomial Parameter Estimation Machine Learning CMPT 726 Simon Fraser University Binomial Parameter Estimation Outline Maximum Likelihood Estimation Smoothed Frequencies, Laplace Correction. Bayesian Approach. Conjugate Prior. Uniform

More information

Change Detection Algorithms

Change Detection Algorithms 5 Change Detection Algorithms In this chapter, we describe the simplest change detection algorithms. We consider a sequence of independent random variables (y k ) k with a probability density p (y) depending

More information

Lecture 22: Error exponents in hypothesis testing, GLRT

Lecture 22: Error exponents in hypothesis testing, GLRT 10-704: Information Processing and Learning Spring 2012 Lecture 22: Error exponents in hypothesis testing, GLRT Lecturer: Aarti Singh Scribe: Aarti Singh Disclaimer: These notes have not been subjected

More information

Statistical Methods for Particle Physics (I)

Statistical Methods for Particle Physics (I) Statistical Methods for Particle Physics (I) https://agenda.infn.it/conferencedisplay.py?confid=14407 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan

More information

STOCHASTIC PROCESSES, DETECTION AND ESTIMATION Course Notes

STOCHASTIC PROCESSES, DETECTION AND ESTIMATION Course Notes STOCHASTIC PROCESSES, DETECTION AND ESTIMATION 6.432 Course Notes Alan S. Willsky, Gregory W. Wornell, and Jeffrey H. Shapiro Department of Electrical Engineering and Computer Science Massachusetts Institute

More information

On the Optimality of Likelihood Ratio Test for Prospect Theory Based Binary Hypothesis Testing

On the Optimality of Likelihood Ratio Test for Prospect Theory Based Binary Hypothesis Testing 1 On the Optimality of Likelihood Ratio Test for Prospect Theory Based Binary Hypothesis Testing Sinan Gezici, Senior Member, IEEE, and Pramod K. Varshney, Life Fellow, IEEE Abstract In this letter, the

More information

Notes on Decision Theory and Prediction

Notes on Decision Theory and Prediction Notes on Decision Theory and Prediction Ronald Christensen Professor of Statistics Department of Mathematics and Statistics University of New Mexico October 7, 2014 1. Decision Theory Decision theory is

More information

ECE531 Lecture 10b: Dynamic Parameter Estimation: System Model

ECE531 Lecture 10b: Dynamic Parameter Estimation: System Model ECE531 Lecture 10b: Dynamic Parameter Estimation: System Model D. Richard Brown III Worcester Polytechnic Institute 02-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 02-Apr-2009 1 / 14 Introduction

More information

Fundamentals of Statistical Signal Processing Volume II Detection Theory

Fundamentals of Statistical Signal Processing Volume II Detection Theory Fundamentals of Statistical Signal Processing Volume II Detection Theory Steven M. Kay University of Rhode Island PH PTR Prentice Hall PTR Upper Saddle River, New Jersey 07458 http://www.phptr.com Contents

More information

Machine Learning: Homework Assignment 2 Solutions

Machine Learning: Homework Assignment 2 Solutions 10-601 Machine Learning: Homework Assignment 2 Solutions Professor Tom Mitchell Carnegie Mellon University January 21, 2009 The assignment is due at 1:30pm (beginning of class) on Monday, February 2, 2009.

More information

7. Statistical estimation

7. Statistical estimation 7. Statistical estimation Convex Optimization Boyd & Vandenberghe maximum likelihood estimation optimal detector design experiment design 7 1 Parametric distribution estimation distribution estimation

More information

CS 630 Basic Probability and Information Theory. Tim Campbell

CS 630 Basic Probability and Information Theory. Tim Campbell CS 630 Basic Probability and Information Theory Tim Campbell 21 January 2003 Probability Theory Probability Theory is the study of how best to predict outcomes of events. An experiment (or trial or event)

More information

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test. Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the

More information

Frequentist Statistics and Hypothesis Testing Spring

Frequentist Statistics and Hypothesis Testing Spring Frequentist Statistics and Hypothesis Testing 18.05 Spring 2018 http://xkcd.com/539/ Agenda Introduction to the frequentist way of life. What is a statistic? NHST ingredients; rejection regions Simple

More information

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses

Hypothesis Testing. Testing Hypotheses MIT Dr. Kempthorne. Spring MIT Testing Hypotheses Testing Hypotheses MIT 18.443 Dr. Kempthorne Spring 2015 1 Outline Hypothesis Testing 1 Hypothesis Testing 2 Hypothesis Testing: Statistical Decision Problem Two coins: Coin 0 and Coin 1 P(Head Coin 0)

More information

Lecture 2. G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1

Lecture 2. G. Cowan Lectures on Statistical Data Analysis Lecture 2 page 1 Lecture 2 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,

More information

Chapter Three. Hypothesis Testing

Chapter Three. Hypothesis Testing 3.1 Introduction The final phase of analyzing data is to make a decision concerning a set of choices or options. Should I invest in stocks or bonds? Should a new product be marketed? Are my products being

More information

Bayesian RL Seminar. Chris Mansley September 9, 2008

Bayesian RL Seminar. Chris Mansley September 9, 2008 Bayesian RL Seminar Chris Mansley September 9, 2008 Bayes Basic Probability One of the basic principles of probability theory, the chain rule, will allow us to derive most of the background material in

More information

IN HYPOTHESIS testing problems, a decision-maker aims

IN HYPOTHESIS testing problems, a decision-maker aims IEEE SIGNAL PROCESSING LETTERS, VOL. 25, NO. 12, DECEMBER 2018 1845 On the Optimality of Likelihood Ratio Test for Prospect Theory-Based Binary Hypothesis Testing Sinan Gezici, Senior Member, IEEE, and

More information

Point Estimation. Vibhav Gogate The University of Texas at Dallas

Point Estimation. Vibhav Gogate The University of Texas at Dallas Point Estimation Vibhav Gogate The University of Texas at Dallas Some slides courtesy of Carlos Guestrin, Chris Bishop, Dan Weld and Luke Zettlemoyer. Basics: Expectation and Variance Binary Variables

More information

ECE531: Principles of Detection and Estimation Course Introduction

ECE531: Principles of Detection and Estimation Course Introduction ECE531: Principles of Detection and Estimation Course Introduction D. Richard Brown III WPI 15-January-2013 WPI D. Richard Brown III 15-January-2013 1 / 39 First Lecture: Major Topics 1. Administrative

More information

On Sequential Watermark Detection

On Sequential Watermark Detection On Sequential Watermark Detection R. Chandramouli and Nasir D. Memon Abstract This paper introduces a new watermark detection paradigm based on sequential detection. It is shown that this framework leads

More information

Lecture 3. STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher

Lecture 3. STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher Lecture 3 STAT161/261 Introduction to Pattern Recognition and Machine Learning Spring 2018 Prof. Allie Fletcher Previous lectures What is machine learning? Objectives of machine learning Supervised and

More information

Hypothesis testing (cont d)

Hypothesis testing (cont d) Hypothesis testing (cont d) Ulrich Heintz Brown University 4/12/2016 Ulrich Heintz - PHYS 1560 Lecture 11 1 Hypothesis testing Is our hypothesis about the fundamental physics correct? We will not be able

More information

Bayesian Decision Theory

Bayesian Decision Theory Introduction to Pattern Recognition [ Part 4 ] Mahdi Vasighi Remarks It is quite common to assume that the data in each class are adequately described by a Gaussian distribution. Bayesian classifier is

More information

Some Probability and Statistics

Some Probability and Statistics Some Probability and Statistics David M. Blei COS424 Princeton University February 12, 2007 D. Blei ProbStat 01 1 / 42 Who wants to scribe? D. Blei ProbStat 01 2 / 42 Random variable Probability is about

More information

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing

Lecture 7. Union bound for reducing M-ary to binary hypothesis testing Lecture 7 Agenda for the lecture M-ary hypothesis testing and the MAP rule Union bound for reducing M-ary to binary hypothesis testing Introduction of the channel coding problem 7.1 M-ary hypothesis testing

More information

10. Composite Hypothesis Testing. ECE 830, Spring 2014

10. Composite Hypothesis Testing. ECE 830, Spring 2014 10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters

More information

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet. Stat 535 C - Statistical Computing & Monte Carlo Methods Arnaud Doucet Email: arnaud@cs.ubc.ca 1 CS students: don t forget to re-register in CS-535D. Even if you just audit this course, please do register.

More information

Probability Density Functions and the Normal Distribution. Quantitative Understanding in Biology, 1.2

Probability Density Functions and the Normal Distribution. Quantitative Understanding in Biology, 1.2 Probability Density Functions and the Normal Distribution Quantitative Understanding in Biology, 1.2 1. Discrete Probability Distributions 1.1. The Binomial Distribution Question: You ve decided to flip

More information

SEQUENTIAL TESTING OF SIMPLE HYPOTHESES ABOUT COMPOUND POISSON PROCESSES. 1. Introduction (1.2)

SEQUENTIAL TESTING OF SIMPLE HYPOTHESES ABOUT COMPOUND POISSON PROCESSES. 1. Introduction (1.2) SEQUENTIAL TESTING OF SIMPLE HYPOTHESES ABOUT COMPOUND POISSON PROCESSES SAVAS DAYANIK AND SEMIH O. SEZER Abstract. One of two simple hypotheses is correct about the unknown arrival rate and jump distribution

More information

Decentralized Detection In Wireless Sensor Networks

Decentralized Detection In Wireless Sensor Networks Decentralized Detection In Wireless Sensor Networks Milad Kharratzadeh Department of Electrical & Computer Engineering McGill University Montreal, Canada April 2011 Statistical Detection and Estimation

More information

Linear Models for Regression CS534

Linear Models for Regression CS534 Linear Models for Regression CS534 Example Regression Problems Predict housing price based on House size, lot size, Location, # of rooms Predict stock price based on Price history of the past month Predict

More information

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use?

Today. Statistical Learning. Coin Flip. Coin Flip. Experiment 1: Heads. Experiment 1: Heads. Which coin will I use? Which coin will I use? Today Statistical Learning Parameter Estimation: Maximum Likelihood (ML) Maximum A Posteriori (MAP) Bayesian Continuous case Learning Parameters for a Bayesian Network Naive Bayes Maximum Likelihood estimates

More information

Information Theory and Hypothesis Testing

Information Theory and Hypothesis Testing Summer School on Game Theory and Telecommunications Campione, 7-12 September, 2014 Information Theory and Hypothesis Testing Mauro Barni University of Siena September 8 Review of some basic results linking

More information

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014

Bayes Formula. MATH 107: Finite Mathematics University of Louisville. March 26, 2014 Bayes Formula MATH 07: Finite Mathematics University of Louisville March 26, 204 Test Accuracy Conditional reversal 2 / 5 A motivating question A rare disease occurs in out of every 0,000 people. A test

More information

The Logit Model: Estimation, Testing and Interpretation

The Logit Model: Estimation, Testing and Interpretation The Logit Model: Estimation, Testing and Interpretation Herman J. Bierens October 25, 2008 1 Introduction to maximum likelihood estimation 1.1 The likelihood function Consider a random sample Y 1,...,

More information

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 60, NO. 9, SEPTEMBER

IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 60, NO. 9, SEPTEMBER IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 60, NO. 9, SEPTEMBER 2012 4509 Cooperative Sequential Spectrum Sensing Based on Level-Triggered Sampling Yasin Yilmaz,StudentMember,IEEE, George V. Moustakides,SeniorMember,IEEE,and

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

Lecture 8: Information Theory and Statistics

Lecture 8: Information Theory and Statistics Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang

More information

Linear Models for Regression CS534

Linear Models for Regression CS534 Linear Models for Regression CS534 Prediction Problems Predict housing price based on House size, lot size, Location, # of rooms Predict stock price based on Price history of the past month Predict the

More information

Advanced statistical methods for data analysis Lecture 1

Advanced statistical methods for data analysis Lecture 1 Advanced statistical methods for data analysis Lecture 1 RHUL Physics www.pp.rhul.ac.uk/~cowan Universität Mainz Klausurtagung des GK Eichtheorien exp. Tests... Bullay/Mosel 15 17 September, 2008 1 Outline

More information

Bayesian Statistics. State University of New York at Buffalo. From the SelectedWorks of Joseph Lucke. Joseph F. Lucke

Bayesian Statistics. State University of New York at Buffalo. From the SelectedWorks of Joseph Lucke. Joseph F. Lucke State University of New York at Buffalo From the SelectedWorks of Joseph Lucke 2009 Bayesian Statistics Joseph F. Lucke Available at: https://works.bepress.com/joseph_lucke/6/ Bayesian Statistics Joseph

More information

ECE353: Probability and Random Processes. Lecture 3 - Independence and Sequential Experiments

ECE353: Probability and Random Processes. Lecture 3 - Independence and Sequential Experiments ECE353: Probability and Random Processes Lecture 3 - Independence and Sequential Experiments Xiao Fu School of Electrical Engineering and Computer Science Oregon State University E-mail: xiao.fu@oregonstate.edu

More information