F2E5216/TS1002 Adaptive Filtering and Change Detection. Course Organization. Lecture plan. The Books. Lecture 1
|
|
- Alberta Jordan
- 5 years ago
- Views:
Transcription
1 Adaptive Filtering and Change Detection Bo Wahlberg (KTH and Fredrik Gustafsson (LiTH Course Organization Lectures and compendium: Theory, Algorithms, Applications, Evaluation Toolbox and manual: Algorithms, Evaluation Implementations Lecture The goal of the course is to get an understanding for the theory of model based filtering and change detection, with particular attention to applications. Projects: Application Evaluation of Algorithms Exercises: To each part available from homepage Examination, 6 credits: Three-day take-home exam. Lecture, 2005 Lecture, The Books Steven Kay: Selected parts of Fundamentals of statistical signal processing: Volume II, detection theory. Prentice Hall, 998. Only Lecture Fredrik Gustafsson: Adaptive filtering and change detection. John Wiley & Sons, 200. Lecture plan Lecture : Detection theory from Kay s book. Lecture 2: Applications. Detection of mean changes: likelihood principles; Chapters, 2, 3. Lecture 3: Detection of mean changes: The parallel filter concept. Chapters 3,4, Appendix 4. Lecture 4: Parameter estimation: recursive least squares. Residual based detection. Chapter 5 Lecture 5: Parameter estimation: parallel filters and filter bank change detection. Chapter 6-7 Lecture 6: State estimation: Kalman, EKF and particle filters. Residual based detection. Chapter 8 Lecture 7: State estimation: parallel filters and filter bank change detection. Chapters 9, 0. Given by Fredrik Lecture 8: State estimation: the parity space approach. Chapter Lecture 9: State estimation: parity space diagnosability and PCA. Lecture, Lecture,
2 Probability Theory Lecture Stochastic variable: X Basic Probability Theory Basic Detection Theory Neyman -Pearson s theorem Basic Hypothesis tests Generalized Likelihood Ratio Tests Pages in Kay Probability density function (PDF: Mean: E(X = Variance: V ar(x = xp(xdx p(x (x E(X 2 p(xdx Cummulative distribution function: Φ(x = P(X < x = Right tail probability: Q(x = P(X > x = x p(xdx x p(xdx Lecture, Lecture, p(x 2πσ 2 e (x µ2 2σ 2 E(x µ Var(x σ 2 Gaussian or Normal N(µ, σ 2 Q(x 2π(x µ/σ e (x µ2 2σ 2 E(x n (n σ 2 Matlab Q(x=0.5*erfc((x-mu/sigma/sqrt(2 Exponential Exp(µ p(x µ e µ, x 0 E(x µ Var(x µ 2 Q(x e x µ Lecture, Lecture,
3 Chi-squared with ν degrees χ 2 ν p(x 2 ν/2 Γ(ν/2 x ν/2 e x/2,x > 0 Γ(u t u e t dt, Γ(n = (n! 0 Γ(u = (u Γ(u, Γ(0.5 = π E(x ν Var(x 2ν Q(x e x/2 ν/2 (x/2 k k=0,ν 2 and even k! Comments χ 2 2 = Exp(2. If x = ν i= x2 i, x i N(0,, then x χ 2 ν approximately Gaussian for large ν Non-central chi-squared with ν degrees p(x I r (u E(x Var(x Comments 2 k=0 χ 2 ν(λ ( x λν 2 4 e (x+λ/2 I ν/2 ( λx,x > 0 ν + λ 2ν + 4λ (u/2 2k+r Bessel function k!γ(r+k+ If x = ν i= x2 i, x i N(µ i,, then x χ 2 ν( i µ i Lecture, Lecture, Detection Principles Given: Data x[n] with known distribution under the null hypothesis H 0 and the alternative hypothesis H, respectively. General test: Form a test statistic T(x. Decide H 0 Decide H Basic design parameters: if T(x < γ (threshold if T(x > γ Probability of false alarm (Type I error P FA = P(H H 0 = x:t(x>γ p(x H 0dx Probability of detection (Power P D = P(H H = x:t(x>γ p(x H dx Miss P M = P(H 0 H = P D (Type II error The only hard result in detection theory Neyman-Pearson s theorem: To maximize the power P D for a given probability of false alarm P FA = α, take T(x = p(x H p(x H 0. The test is called LRT, the Likelihood Ratio Test. The threshold is found from P FA = x:t(x>γ p(x H 0 dx = α Lecture, 2005 Lecture,
4 Proof: Receiver Operating Characteristics Define the LRT test T = {, if, p(x H > γp(x H 0 0, if, p(x H < γp(x H 0 P D LRT chance line Let 0 T denote any test with less or equal P FA. Key inequality (Check the signs! (T T (p(x H γp(x H 0 dx 0 Re-arrange the integral: (P D P D γ(p FA P FA 0, where the last equality follows from the false alarm assumption. This shows T optimize the power! One has to be a bit more careful if p(x H = γp(x H 0 has a non-zero probability P FA The slope of the LRT ROC curve is γ! The curve is concave. The LRT maximizes any performance criterion that values low false alarm probabilities and/or high detection probabilities. Lecture, Lecture, Proof of Slope [ ] n E(T(x n p(x H H = p(x H dx = p(x H 0 [ ] n+ p(x H p(x H 0 dx = E(T(x n+ H 0 p(x H 0 which can be written t n g(t H dt = t n+ g(t H 0 dt, n Sufficient Statistics Replace x with the simpler sufficient statistics t, T(x = p(x H p(x H 0 = g(t H g(t H 0 where g(t H i is the PDF for the likelihood ratio under H i. This means that g(t H = tg(t H 0. Now and P D = g(t H dt, P F A = g(t H 0 dt γ γ dp D = dp D dp F A dγ / dp F A = g(t H dγ g(t H 0 = γ Example: x = (x 0...x N, x i N(m,σ 2, take t = N N i=0 x i Observe that γ 0 Lecture, Lecture,
5 Basic Hypothesis Test Is there a known mean A in the observed signal? H 0 :x[n] = w[n] H :x[n] = A + w[n] The more general problem of detecting a given signal H 0 :x[n] = w[n] H :x[n] = s[n] + w[n] can be rewritten to unknown mean. Example : Communication The signal s(t = 2A sin(ωt is transmitted on the carrier ω and y(t = s(t + w(t is received. Let x = T where w N(0, σ2 2T. T 0 = A + T y(t sin(ωt T 0 w(t sin(ωt = A + w, under H Lecture, Lecture, Example 2: Matched filter The signal s[n] with energy E may be contained in y[n]. Let where w N(0, Eσ2 N. x = N N n=0 y[n]s[n] = E + w under H T(x = Neyman-Pearson N e n=0 (x[n] A2 (2πσ 2 N/2 2σ 2 e (2πσ 2 N/2 N n=0 (x[n]2 2σ 2 Taking logarithm and simplifying give T(x = N N n=0 = e NA 2 N n=0 2Ax[n] 2σ 2 > γ. x[n] > σ2 NA log(γ + A 2 = γ. Lecture, Lecture,
6 ( Now T(x N ( N 0, σ2 N A, σ2 N under H 0 under H ( so P FA = P( T(x > γ H 0 = Q γ σ2 /N σ 2 γ = N Q (P FA ( ( P D = P( T(x γ A NA > γ H = Q = Q Q 2 (P FA σ2 /N σ 2 The last expression shows that P D is a function of the desired P FA and SNR, NA 2 /σ 2, only. Standard plots: ROC: Receiver Operating Characteristics plots P D versus P FA with γ being the parameter. Just guessing H with probability P FA gives P D = P FA. All clever tests give a curve above this straight line. Neyman-Pearson gives the upper bound curve for all tests. Detection performance: P D versus SNR for a given P FA. Again this performance is maximized for the Neyman-Pearson test. BER: Bit-Error Rate. In communication, both hypotheses are equal, and the design is to get P D = P FA. A BER-plot shows P D = P FA versus SNR. Lecture, Lecture, P D ROC N=5 σ= MC= LRT 0. T(x=max(x Wild guess P FA Typical Matlab code for ROC curves: gamma=[-0.:0.:4]; MC=000; N=5; A=; sigma=; x=sigma*randn(n,mc; T=mean(x; T2=max(x; for k=:length(gamma PD(k=mean(T+A>gamma(k; PD2(k=mean(T2+A>gamma(k; PFA(k=mean(T>gamma(k; PFA2(k=mean(T2>gamma(k; end Simple line search to design γ and calculate P D given P FA : PFAdesign=0.; ind=find(pfa<pfadesign; gamma=gamma(ind(, PDdesign=PD(ind(, ind=find(pfa2<pfadesign; gamma2=gamma(ind(, PD2design=PD2(ind( interp better to use. Lecture, Lecture,
7 Maximum Likelihood Estimation Probability density function for the observation x parameterized by the unknown parameter θ: p(x,θ ˆθ ML = arg maxp(x observed,θ θ Example: If x = (x 0...x N, x i N(θ,σ 2, then ˆθ ML = N N i=0 x i Some Properties Asymptotic distribution: N(ˆθML θ 0 N(0, [I(θ 0 ] where the Fisher information matrix equals I(θ = E [ ] [ d d log p(x,θ dθ dθ Asymptotically the best possible estimator (reach the Cramér-Rao bound ] T log p(x,θ Lecture, Lecture, The GLRT for unknown A Some more common special case: H 0 :x p 0 (x θ 0 H :x p (x θ General principle: Plug in the maximum likelihood estimate of A under H, and let L(x = max θ p (x θ max θ0 p 0 (x θ 0 = p (x ˆθ ML p 0 (x ˆθ ML Generalized Likelihood Ratio Test (GLRT Decide H 0 Decide H if L(x < γ if L(x > γ 0 H 0 : θ = θ 0 H : θ θ 0 GLRT becomes L(x = max θ p(x θ p(x θ 0 General result for asymptotic distribution: 2 log L(x { χ 2 n under H 0 χ 2 n(λ under H. = λ = (θ θ 0 T I(θ 0 (θ θ 0 where I(θ is Fisher s information matrix p(x ˆθ ML p(x θ 0 Lecture, Lecture,
8 Consider Example: H 0 : x[n] = w[n] H : x[n] = A + w[n] with unknown A and Gaussian noise with known variance σ 2. Then 2 log L(x = NÂ2 σ 2 Â = N N n=0 ( N x[n] ( N { χ 2 2 log L(x under H 0 χ 2 ( NA2 under H σ 2. 0, σ2 N A, σ2 N under H 0 under H The MLRT for unknown A H 0 : x p 0 (x θ 0 H : x p (x θ General principle: Use a prior on θ and integrate out its influence in the LRT. This is the Bayesian approach. L(x = θ p (x θ p(θ dθ θ 0 p 0 (x θ 0 p(θ 0 dθ 0 This is the Marginalized Likelihood Ratio Test (MLRT. Lecture, Lecture, Test: Mixed M/G LRT approach Nuisance parameters: classical approaches H 0 :θ = θ 0,η H :θ θ 0,η Here η is an unknown nuisance parameter. General principle using the Bayesian approach: marginalize nuisance parameters, estimate the other parameters. The LRT becomes L(x = max θ p(x θ,ηdη η p(x θ = p(x ˆθ ML,ηdη η η 0,ηdη p(x θ η 0,ηdη Non-Bayesian approaches for the test H 0 : θ = θ 0,η H : θ θ 0,η GLRT ML estimate of nuisance parameters: L(x = p(x ˆθ ML, ˆη ML p(x θ 0, ˆη ML Lecture, Lecture,
9 General result for asymptotic distribution: { χ 2 2 log L(x dim(θ under H 0 χ 2 dim(θ (λ under H. λ = (θ θ 0 ( T I θ,θ I θ,η Iη,ηI η,θ (θ θ 0 }{{} [I ] θ,θ Use Wald test L(x = (ˆθ θ 0 T [I ( θ,η H ] θ,θ (ˆθ θ 0 where I(θ is Fisher s information matrix partitioned as I = ( I θ,θ I η,θ I θ,η I η,η where θ,η H is the ML estimate under H. Same performance as GLRT. evaluated at the true θ 0,η 0 (also θ denotes the true value under H. Note that the underbraced expression [I ] θ,θ is the Schur complement to compute the upper left corner of the matrix inverse. Lecture, Lecture, Locally Most Powerful (LMP test Use L(x = d log(p(x θ,η dθ Rao test where ˆη H 0 is the ML estimate under H 0. Same performance as GLRT. T [I d log(p(x θ,η (θ 0, ˆη H 0 ] θ,θ θ 0,ˆη H 0 dθ Note that θ does not have to be estimated here, as it has in the Wald test. θ0,ˆη H 0 The Locally Most Powerful (LMP test is a special case of the Rao test for the case of scalar parameter, no nuisance parameter and one-sided test: Take the test statistic L(x = dlog(p(x θ dθ I(θ0 θ0 H 0 :θ = θ 0 H :θ > θ 0 { N(0, under H0 N( I(θ 0 (θ θ 0, under H for small changes θ θ 0 > 0. The derivation is based on local analysis by Taylor expansion. Lecture, Lecture,
10 Bayesian extensions Priors: Assume priors p(h 0 and p(h, respectively. Bayesian decision rule, compare the a posteriori distribution of each hypothesis, p(h x H 0 H p(h 0 x p(x H p(h p(x p(x H p(x H 0 H 0 p(x H 0 p(h 0 H p(x H 0 p(h H p(h 0 = γ Multiple tests The use of prior enables multiple hypotheses tests: ˆk = arg max k p(h k x = arg maxp(x H k p(h k k which shows that Bayes rule provides a way to design the threshold in the LRT test. Lecture, Lecture, Define a cost to each decision Bayes Risk C ij = P(H i H j that is, C 0 is the false alarm cost and C 0, is the missed detection cost. Then p(x H p(x H 0 H 0 (C 0 C 00 p(h H (C 0 C p(h 0 = γ minimizes the Bayes risk or expected cost R = E(C = i=0 C ij P(H i H j p(h j j=0 Model: Y = Φ T θ + E where Cov(E = C and p(y θ = Linear regressions (2π N/2 det(c e 2 (Y ΦT θ T C (Y Φ T θ. Lecture, Lecture,
11 Test: Known parameter { H0 : θ = 0 H : θ = θ Neyman-Pearson gives LRT: ( p(y θ log P(Y θ = 0 = 0.5(Y Φ T θ T C (Y Φ T θ + 0.5Y T C Y = Y T C Φ T θ 0.5θ T ΦC Φ T θ The last term in independent of data, so take Constant in noise: L(Y = Y T C Φ T θ N(θ T ΦC Φ T θ,θ T ΦC Φ T θ, ( P D = Q Q (P FA θ T ΦC Φ T θ Correlation detector or Matched Filter Step Pre-whiten the data: Ȳ = C /2 Y, Φ = C /2 Φ Ȳ = Φ T θ + Ē Cov(Ē = I Step 2: L = Ȳ T M, where M = Φ T θ Correlation detector: L = N i= ȳ[i]m[i] Matched filter: Convolution l(t = t i=0 h[i]ȳ[t i] Impulse response h[i] = m[n i], (matched to the signal Test statistics: L = l(n Lecture, Lecture, Test ML estimate Unknown parameter H 0 : θ = 0 H : θ 0 ˆθ = (ΦΦ T ΦY GLRT ( p(y ˆθ L(x = 2 log P(Y θ = 0 = (Y Φ T ˆθ T C (Y Φ T ˆθ + Y T C Y = 2Y T C Φ T ˆθ ˆθT ΦC Φ T ˆθ =... = Y T C Φ T (ΦC Φ T ΦC Y χ 2 dim(θ See Chapter 6. in Gustafsson for more calculations on the linear model. Lecture, Lecture,
12 Exercises Check the lecture notes and report errors to me Groups of at most five persons I want one TEX version of problems + solution per group Kay:.,.2,.4, 2.2, 3., 3.8, 3.3 Lecture,
Detection theory 101 ELEC-E5410 Signal Processing for Communications
Detection theory 101 ELEC-E5410 Signal Processing for Communications Binary hypothesis testing Null hypothesis H 0 : e.g. noise only Alternative hypothesis H 1 : signal + noise p(x;h 0 ) γ p(x;h 1 ) Trade-off
More informationLecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics. 1 Executive summary
ECE 830 Spring 207 Instructor: R. Willett Lecture 5: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we saw that the likelihood
More information2. What are the tradeoffs among different measures of error (e.g. probability of false alarm, probability of miss, etc.)?
ECE 830 / CS 76 Spring 06 Instructors: R. Willett & R. Nowak Lecture 3: Likelihood ratio tests, Neyman-Pearson detectors, ROC curves, and sufficient statistics Executive summary In the last lecture we
More information10. Composite Hypothesis Testing. ECE 830, Spring 2014
10. Composite Hypothesis Testing ECE 830, Spring 2014 1 / 25 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve unknown parameters
More informationDetection theory. H 0 : x[n] = w[n]
Detection Theory Detection theory A the last topic of the course, we will briefly consider detection theory. The methods are based on estimation theory and attempt to answer questions such as Is a signal
More informationDETECTION theory deals primarily with techniques for
ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for
More informationIntroduction to Statistical Inference
Structural Health Monitoring Using Statistical Pattern Recognition Introduction to Statistical Inference Presented by Charles R. Farrar, Ph.D., P.E. Outline Introduce statistical decision making for Structural
More informationIf there exists a threshold k 0 such that. then we can take k = k 0 γ =0 and achieve a test of size α. c 2004 by Mark R. Bell,
Recall The Neyman-Pearson Lemma Neyman-Pearson Lemma: Let Θ = {θ 0, θ }, and let F θ0 (x) be the cdf of the random vector X under hypothesis and F θ (x) be its cdf under hypothesis. Assume that the cdfs
More informationChapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1)
Chapter 2. Binary and M-ary Hypothesis Testing 2.1 Introduction (Levy 2.1) Detection problems can usually be casted as binary or M-ary hypothesis testing problems. Applications: This chapter: Simple hypothesis
More informationLecture 8: Information Theory and Statistics
Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 23, 2015 1 / 50 I-Hsiang
More informationECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters
ECE531 Lecture 6: Detection of Discrete-Time Signals with Random Parameters D. Richard Brown III Worcester Polytechnic Institute 26-February-2009 Worcester Polytechnic Institute D. Richard Brown III 26-February-2009
More informationDetection Theory. Composite tests
Composite tests Chapter 5: Correction Thu I claimed that the above, which is the most general case, was captured by the below Thu Chapter 5: Correction Thu I claimed that the above, which is the most general
More informationPATTERN RECOGNITION AND MACHINE LEARNING
PATTERN RECOGNITION AND MACHINE LEARNING Slide Set 3: Detection Theory January 2018 Heikki Huttunen heikki.huttunen@tut.fi Department of Signal Processing Tampere University of Technology Detection theory
More informationChapter 2 Signal Processing at Receivers: Detection Theory
Chapter Signal Processing at Receivers: Detection Theory As an application of the statistical hypothesis testing, signal detection plays a key role in signal processing at receivers of wireless communication
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More informationComposite Hypotheses and Generalized Likelihood Ratio Tests
Composite Hypotheses and Generalized Likelihood Ratio Tests Rebecca Willett, 06 In many real world problems, it is difficult to precisely specify probability distributions. Our models for data may involve
More informationFundamentals of Statistical Signal Processing Volume II Detection Theory
Fundamentals of Statistical Signal Processing Volume II Detection Theory Steven M. Kay University of Rhode Island PH PTR Prentice Hall PTR Upper Saddle River, New Jersey 07458 http://www.phptr.com Contents
More informationDetection and Estimation Theory
ESE 524 Detection and Estimation Theory Joseph A. O Sullivan Samuel C. Sachs Professor Electronic Systems and Signals Research Laboratory Electrical and Systems Engineering Washington University 2 Urbauer
More informationLecture 7 Introduction to Statistical Decision Theory
Lecture 7 Introduction to Statistical Decision Theory I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 20, 2016 1 / 55 I-Hsiang Wang IT Lecture 7
More informationBrandon C. Kelly (Harvard Smithsonian Center for Astrophysics)
Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics) Probability quantifies randomness and uncertainty How do I estimate the normalization and logarithmic slope of a X ray continuum, assuming
More informationDetection Theory. Chapter 3. Statistical Decision Theory I. Isael Diaz Oct 26th 2010
Detection Theory Chapter 3. Statistical Decision Theory I. Isael Diaz Oct 26th 2010 Outline Neyman-Pearson Theorem Detector Performance Irrelevant Data Minimum Probability of Error Bayes Risk Multiple
More informationFinal Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.
1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given. (a) If X and Y are independent, Corr(X, Y ) = 0. (b) (c) (d) (e) A consistent estimator must be asymptotically
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationSTOCHASTIC PROCESSES, DETECTION AND ESTIMATION Course Notes
STOCHASTIC PROCESSES, DETECTION AND ESTIMATION 6.432 Course Notes Alan S. Willsky, Gregory W. Wornell, and Jeffrey H. Shapiro Department of Electrical Engineering and Computer Science Massachusetts Institute
More information10-704: Information Processing and Learning Fall Lecture 24: Dec 7
0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of
More informationDetection and Estimation Theory
ESE 524 Detection and Estimation Theory Joseph A. O Sullivan Samuel C. Sachs Professor Electronic Systems and Signals Research Laboratory Electrical and Systems Engineering Washington University 211 Urbauer
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationCh. 5 Hypothesis Testing
Ch. 5 Hypothesis Testing The current framework of hypothesis testing is largely due to the work of Neyman and Pearson in the late 1920s, early 30s, complementing Fisher s work on estimation. As in estimation,
More informationLecture 21: October 19
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 21: October 19 21.1 Likelihood Ratio Test (LRT) To test composite versus composite hypotheses the general method is to use
More informationIntroduction to Detection Theory
Introduction to Detection Theory Detection Theory (a.k.a. decision theory or hypothesis testing) is concerned with situations where we need to make a decision on whether an event (out of M possible events)
More informationECE531 Lecture 8: Non-Random Parameter Estimation
ECE531 Lecture 8: Non-Random Parameter Estimation D. Richard Brown III Worcester Polytechnic Institute 19-March-2009 Worcester Polytechnic Institute D. Richard Brown III 19-March-2009 1 / 25 Introduction
More informationModule 2. Random Processes. Version 2, ECE IIT, Kharagpur
Module Random Processes Version, ECE IIT, Kharagpur Lesson 9 Introduction to Statistical Signal Processing Version, ECE IIT, Kharagpur After reading this lesson, you will learn about Hypotheses testing
More informationECE531 Lecture 2b: Bayesian Hypothesis Testing
ECE531 Lecture 2b: Bayesian Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute 29-January-2009 Worcester Polytechnic Institute D. Richard Brown III 29-January-2009 1 / 39 Minimizing
More information40.530: Statistics. Professor Chen Zehua. Singapore University of Design and Technology
Singapore University of Design and Technology Lecture 9: Hypothesis testing, uniformly most powerful tests. The Neyman-Pearson framework Let P be the family of distributions of concern. The Neyman-Pearson
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationMathematics Ph.D. Qualifying Examination Stat Probability, January 2018
Mathematics Ph.D. Qualifying Examination Stat 52800 Probability, January 2018 NOTE: Answers all questions completely. Justify every step. Time allowed: 3 hours. 1. Let X 1,..., X n be a random sample from
More informationDetecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf
Detecting Parametric Signals in Noise Having Exactly Known Pdf/Pmf Reading: Ch. 5 in Kay-II. (Part of) Ch. III.B in Poor. EE 527, Detection and Estimation Theory, # 5c Detecting Parametric Signals in Noise
More informationsimple if it completely specifies the density of x
3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely
More informationDetection and Estimation Theory
ESE 524 Detection and Estimation Theory Joseph A. O Sullivan Samuel C. Sachs Professor Electronic Systems and Signals Research Laboratory Electrical and Systems Engineering Washington University 2 Urbauer
More informationEIE6207: Maximum-Likelihood and Bayesian Estimation
EIE6207: Maximum-Likelihood and Bayesian Estimation Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak
More informationEconomics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,
Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem
More informationMathematical statistics
October 1 st, 2018 Lecture 11: Sufficient statistic Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation
More informationEIE6207: Estimation Theory
EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: Steven M.
More informationParameter Estimation
1 / 44 Parameter Estimation Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay October 25, 2012 Motivation System Model used to Derive
More informationLECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.
Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the
More informationParametric Techniques Lecture 3
Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to
More informationTopic 3: Hypothesis Testing
CS 8850: Advanced Machine Learning Fall 07 Topic 3: Hypothesis Testing Instructor: Daniel L. Pimentel-Alarcón c Copyright 07 3. Introduction One of the simplest inference problems is that of deciding between
More informationECE531 Homework Assignment Number 6 Solution
ECE53 Homework Assignment Number 6 Solution Due by 8:5pm on Wednesday 3-Mar- Make sure your reasoning and work are clear to receive full credit for each problem.. 6 points. Suppose you have a scalar random
More informationSTAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015
STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis
More informationUNIFORMLY MOST POWERFUL CYCLIC PERMUTATION INVARIANT DETECTION FOR DISCRETE-TIME SIGNALS
UNIFORMLY MOST POWERFUL CYCLIC PERMUTATION INVARIANT DETECTION FOR DISCRETE-TIME SIGNALS F. C. Nicolls and G. de Jager Department of Electrical Engineering, University of Cape Town Rondebosch 77, South
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond January 18, 2017 Contents 1 Batch and Recursive Estimation 2 Towards Bayesian Filtering 3 Kalman Filter and Bayesian Filtering and Smoothing
More informationSYDE 372 Introduction to Pattern Recognition. Probability Measures for Classification: Part I
SYDE 372 Introduction to Pattern Recognition Probability Measures for Classification: Part I Alexander Wong Department of Systems Design Engineering University of Waterloo Outline 1 2 3 4 Why use probability
More informationECE531 Lecture 4b: Composite Hypothesis Testing
ECE531 Lecture 4b: Composite Hypothesis Testing D. Richard Brown III Worcester Polytechnic Institute 16-February-2011 Worcester Polytechnic Institute D. Richard Brown III 16-February-2011 1 / 44 Introduction
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationHypothesis Testing - Frequentist
Frequentist Hypothesis Testing - Frequentist Compare two hypotheses to see which one better explains the data. Or, alternatively, what is the best way to separate events into two classes, those originating
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More informationEECS564 Estimation, Filtering, and Detection Exam 2 Week of April 20, 2015
EECS564 Estimation, Filtering, and Detection Exam Week of April 0, 015 This is an open book takehome exam. You have 48 hours to complete the exam. All work on the exam should be your own. problems have
More informationBayesian Decision and Bayesian Learning
Bayesian Decision and Bayesian Learning Ying Wu Electrical Engineering and Computer Science Northwestern University Evanston, IL 60208 http://www.eecs.northwestern.edu/~yingwu 1 / 30 Bayes Rule p(x ω i
More informationLecture 8: Signal Detection and Noise Assumption
ECE 830 Fall 0 Statistical Signal Processing instructor: R. Nowak Lecture 8: Signal Detection and Noise Assumption Signal Detection : X = W H : X = S + W where W N(0, σ I n n and S = [s, s,..., s n ] T
More informationECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations
ECE531 Screencast 9.2: N-P Detection with an Infinite Number of Possible Observations D. Richard Brown III Worcester Polytechnic Institute Worcester Polytechnic Institute D. Richard Brown III 1 / 7 Neyman
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationGARCH Models Estimation and Inference
GARCH Models Estimation and Inference Eduardo Rossi University of Pavia December 013 Rossi GARCH Financial Econometrics - 013 1 / 1 Likelihood function The procedure most often used in estimating θ 0 in
More informationSequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process
Applied Mathematical Sciences, Vol. 4, 2010, no. 62, 3083-3093 Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Julia Bondarenko Helmut-Schmidt University Hamburg University
More informationLecture 8: Information Theory and Statistics
Lecture 8: Information Theory and Statistics Part II: Hypothesis Testing and Estimation I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw December 22, 2015
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationECE 275B Homework # 1 Solutions Version Winter 2015
ECE 275B Homework # 1 Solutions Version Winter 2015 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2
More informationParametric Techniques
Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationDirection: This test is worth 250 points and each problem worth points. DO ANY SIX
Term Test 3 December 5, 2003 Name Math 52 Student Number Direction: This test is worth 250 points and each problem worth 4 points DO ANY SIX PROBLEMS You are required to complete this test within 50 minutes
More informationMathematical statistics
October 18 th, 2018 Lecture 16: Midterm review Countdown to mid-term exam: 7 days Week 1 Chapter 1: Probability review Week 2 Week 4 Week 7 Chapter 6: Statistics Chapter 7: Point Estimation Chapter 8:
More informationClassical Estimation Topics
Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min
More informationσ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =
Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,
More informationChapter 7. Hypothesis Testing
Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function
More informationIntroduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak
Introduction to Systems Analysis and Decision Making Prepared by: Jakub Tomczak 1 Introduction. Random variables During the course we are interested in reasoning about considered phenomenon. In other words,
More informationBayesian Decision Theory
Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian
More informationIntroduction to Estimation Methods for Time Series models Lecture 2
Introduction to Estimation Methods for Time Series models Lecture 2 Fulvio Corsi SNS Pisa Fulvio Corsi Introduction to Estimation () Methods for Time Series models Lecture 2 SNS Pisa 1 / 21 Estimators:
More informationParameter Estimation in a Moving Horizon Perspective
Parameter Estimation in a Moving Horizon Perspective State and Parameter Estimation in Dynamical Systems Reglerteknik, ISY, Linköpings Universitet State and Parameter Estimation in Dynamical Systems OUTLINE
More informationECE 275B Homework # 1 Solutions Winter 2018
ECE 275B Homework # 1 Solutions Winter 2018 1. (a) Because x i are assumed to be independent realizations of a continuous random variable, it is almost surely (a.s.) 1 the case that x 1 < x 2 < < x n Thus,
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationHypothesis Testing. Econ 690. Purdue University. Justin L. Tobias (Purdue) Testing 1 / 33
Hypothesis Testing Econ 690 Purdue University Justin L. Tobias (Purdue) Testing 1 / 33 Outline 1 Basic Testing Framework 2 Testing with HPD intervals 3 Example 4 Savage Dickey Density Ratio 5 Bartlett
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More informationUniversity of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout 2:. The Multivariate Gaussian & Decision Boundaries
University of Cambridge Engineering Part IIB Module 3F3: Signal and Pattern Processing Handout :. The Multivariate Gaussian & Decision Boundaries..15.1.5 1 8 6 6 8 1 Mark Gales mjfg@eng.cam.ac.uk Lent
More informationExponential Families
Exponential Families David M. Blei 1 Introduction We discuss the exponential family, a very flexible family of distributions. Most distributions that you have heard of are in the exponential family. Bernoulli,
More informationDA Freedman Notes on the MLE Fall 2003
DA Freedman Notes on the MLE Fall 2003 The object here is to provide a sketch of the theory of the MLE. Rigorous presentations can be found in the references cited below. Calculus. Let f be a smooth, scalar
More informationLecture 2: From Linear Regression to Kalman Filter and Beyond
Lecture 2: From Linear Regression to Kalman Filter and Beyond Department of Biomedical Engineering and Computational Science Aalto University January 26, 2012 Contents 1 Batch and Recursive Estimation
More informationUniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence
Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Valen E. Johnson Texas A&M University February 27, 2014 Valen E. Johnson Texas A&M University Uniformly most powerful Bayes
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationLecture 22: Error exponents in hypothesis testing, GLRT
10-704: Information Processing and Learning Spring 2012 Lecture 22: Error exponents in hypothesis testing, GLRT Lecturer: Aarti Singh Scribe: Aarti Singh Disclaimer: These notes have not been subjected
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationFoundations of Statistical Inference
Foundations of Statistical Inference Jonathan Marchini Department of Statistics University of Oxford MT 2013 Jonathan Marchini (University of Oxford) BS2a MT 2013 1 / 27 Course arrangements Lectures M.2
More informationML Testing (Likelihood Ratio Testing) for non-gaussian models
ML Testing (Likelihood Ratio Testing) for non-gaussian models Surya Tokdar ML test in a slightly different form Model X f (x θ), θ Θ. Hypothesist H 0 : θ Θ 0 Good set: B c (x) = {θ : l x (θ) max θ Θ l
More informationLecture 2 Machine Learning Review
Lecture 2 Machine Learning Review CMSC 35246: Deep Learning Shubhendu Trivedi & Risi Kondor University of Chicago March 29, 2017 Things we will look at today Formal Setup for Supervised Learning Things
More informationAn Introduction to Signal Detection and Estimation - Second Edition Chapter III: Selected Solutions
An Introduction to Signal Detection and Estimation - Second Edition Chapter III: Selected Solutions H. V. Poor Princeton University March 17, 5 Exercise 1: Let {h k,l } denote the impulse response of a
More informationEstimation and Detection
Estimation and Detection Lecture : Detection Theory Unknown Parameters Dr. ir. Richard C. Hendriks //05 Previous Lecture H 0 : T (x) < H : T (x) > Using detection theory, rules can be derived on how to
More information