Basic concepts in estimation
|
|
- Kristian Chapman
- 5 years ago
- Views:
Transcription
1 Basic concepts in estimation Random and nonrandom parameters Definitions of estimates ML Maimum Lielihood MAP Maimum A Posteriori LS Least Squares MMS Minimum Mean square rror Measures of quality of estimates Unbiased, variance, consistency,the Cramer-Rao lower bound, Fisher information, fficiency
2 The problem of parameter estimation Parameter is usually assumed to be time-invariant quantity, vs. time varying parameter Given measurements (),.. to estimate parameter, in presence of disturbances w() ( ) h[,, w( ) ],, L, find a function of observation { ( ) } [ ] or ˆ Z ˆ Z ˆ, Δ Z This function is called estimator, and the value is called estimate The estimation error ~ Δ ˆ
3 The problem of parameter estimation Models for estimation of a parameter. Nonrandom, unnown constant, unnown true value, non-bayesian or Fisher approach. Random, the parameter is a random variable with a prior (a priori) pdf p(), the Bayesian approach The Bayesian approach from the prior pdf, one can obtain posterior (a posteriori) pdf using Bayes formula ( Z ) p p p Z p( Z ) p p( Z ) c c normaliation constant the posterior pdf can be used in several ways to estimate
4 The problem of parameter estimation The Non-Bayesian approach no the prior pdf associated with parameter, one cannot define posterior pdf for it One has the pdf of the measurements conditioned on the parameter, Lielihood Function LF of the parameter Λ Z Δ ( p Z or Λ p Z ) Δ as a measure of how liely the parameter value is for the observations, LF serves as a measure of evidence from the data.
5 ML and MAP estimators Maimiation of the LF relative to produces Maimum Lielihood stimator (ML) ( Z ) arg ma Λ arg ma p( Z ) ML ˆ Z ML, being a function of random observations Z, is a random variable ML is the solution of lielihood equation dλ Z d d p( Z ) d
6 ML and MAP estimators Maimum A Posteriori stimator (MAP), for a random parameter, follows from the maimiation of the posterior pdf ( Z ) arg ma p( Z ) arg ma[ p( Z ) p ] MAP ˆ MAP estimate is a random variable Normaliation constant c (from Bayes formula) is irrelevant for the maimiation
7 ML versus MAP with Gaussian prior Consider the single measurement w where p ( w N, ) Assume that is unnown constant, no a prior information Λ ( p N, ) e π ; then ˆ ML ( Z ) arg ma Λ since the pea or mode occurs at
8 ML versus MAP with Gaussian prior Assume now that a prior information about is Gaussian, N ;, p which is independent of w, then the posterior pdf of, on the observation, is p ( ) p ( ) p p p c π where c is the normaliation constant independent of c e ( )
9 ML versus MAP with Gaussian prior rearrangement of the eponent
10 ML versus MAP with Gaussian prior the posterior pdf of, on the observation, becomes where [ ], ; ξ π ξ e N p p p p ξ Δ Δ p MAP ξ ma arg ˆ
11 ML versus MAP with Gaussian prior note that MAP estimator is a weighted combination of. the ML estimator, pea of the lielihood function. the pea of the a prior pdf of the parameter ˆ MAP arg ma p( ) ξ the weighting is inversely proportional to their variance directly proportional to information, inverse variance
12 The sufficient statistics and the lielihood equation If the LF can be decomposed Λ p( Z ) f g( Z ) then the ML depends only on function g(z), called sufficient statistics, [ ] f ( Z ) Δ, summaries the information contained in the data set about ample: scalar measurement, mutually independent ( ) w w( ),,..., ( ) ( N, N, )
13 The sufficient statistics and the lielihood equation LF of in terms of is then sufficient statistics () [ ] [ ] [ ] Z g e Z g f ce Z f Z f Z g f e ce ce ce N p Z p Λ Δ Δ Δ,,,,, ;,..., { } Z
14 The sufficient statistics and the lielihood equation Lielihood equation dλ df g Z, d d since logarithm is monotonic transformation, equivalent loglielihood function can be used d ln Λ d d ln f [ g( Z ), ] d [ ] ( ) which yields ˆ ML
15 Least Squares and Minimum Mean Square rror estimation Another common estimation procedure for for nonrandom parameters is Least Squares method (LS) the LS of is ( ) h(, ) w( ),,..., ˆ LS arg min [ ( ) h(, ) ] Nonlinear LS-problem, linear LS-problem treated in detail no assumptions about the measurement errors
16 Least Squares and Minimum Mean Square rror estimation if the measurement errors are independent Gaussian random variables p w N, then LS coincides with the ML, which is shown: ( ( ) ) N h(, ) Lielihood Function of is Λ Δ ( p Z ) p[ (),..., ] N ( ) ; h(, ) (, ) ( ) (, ), p,..., ce ( ( ) h(, ) )
17 Least Squares and Minimum Mean Square rror estimation ˆ ML arg ma ln Λ arg ma arg min ˆ LS [ ( ) h(, ) ] [ ( ) h(, ) ] under Gaussian assumption LS method is a disguised ML approach
18 Least Squares and Minimum Mean Square rror estimation For random parameter the counterpart of LS is the Minimum Mean Square rror (MMS) stimator d ˆ ( Z) arg min ˆ [( ˆ ) Z ] MMS [ ( ˆ ) Z ] ( ˆ ) dˆ ˆ MMS ( Z) [ Z ] ( ˆ [ Z ]) Δ [ Z ] p( Z ) d the solution is the conditional mean of
19 Least Squares and Minimum Mean Square rror estimation Some LS-estimators Single measurement of the unnown parameter if the noise is Gaussian Several measurement of the unnown parameter, Gaussian independent noise terms sample mean sample average [ ] ML LS w ˆ arg min ˆ, [ ] [ ] ML LS w ˆ arg ma ln arg ma arg min ˆ,..., ), ( ) ( Λ
20 Least Squares and Minimum Mean Square rror estimation MMS versus MAP stimator Single measurement of the unnown random parameter with a Gaussian priori pdf, noise is also Gaussian ( ξ ( )) w, p e π the mean of this Gaussian pdf is ξ(), which is also the mode of this pdf. Thus [ ] MAP ˆ MMS ˆ ξ MMS estimator coincides with MAP estimator due to the the fact that mean and mode of Gaussian coincide MMS MAP N ; ˆ, N ; ˆ p,
21 Unbiased estimators For a nonrandom parameter, an estimator is unbiased if [ ] ˆ,, ( Z p Z ) pdf where is the true value of the parameter. Bayesian case, is random with a prior pdf p() [ ] [ ] ˆ, Z General definition with estimation error [ ~ ] ~ Δ ˆ;
22 Unbiasedness of ML and MAP estimator Single measurement, ML estimator unbiased Single measurement, MAP estimator, prior pdf p() unbiased [ ] [] [ ] [ ] [ ] ˆ ˆ, w w w ML ML [ ] [ ] [] [ ] w w MAP MAP Δ ) ( ˆ ) ( ˆ, ξ ξ
23 ln Λ Bias in the ML estimation of two parameters stimating (ML) of unnown mean and variance of a set of measurements mean ( ( ) ) Λ [ ], p,...,, e ( π ) ML ˆ ( ) sample variance (, ) 3 ( ( ) ) ( ) ( ML ˆ ) 3 [ ] ML ˆ ( ) ML ˆ ( ) ( )
24 mean Bias in the ML estimation of two parameters [ ] ML ˆ ( ) sample mean estimator unbiased sample variance {[ ] } ML ˆ ( ) ( ) sample variance estimator biased, (asymptotically unbiased) In order to be unbiased [ ] ML ˆ ( ) ( )
25 The variance of an estimator Non-Bayesian case, ML or LS var[ ˆ ( Z )] Δ { ˆ ( Z ) [ ˆ ( Z )]} if estimator unbiased var[ ˆ ( Z )] { ˆ ( Z ) } if estimator is biased then Mean Square rror (MS) Bayesian case, MAP Unconditional (MS) MS MS [ ˆ ( Z )] { ˆ ( Z ) } MS [ ˆ ( Z )] [ ˆ ( Z ) ] [ ˆ ( Z )] { ˆ ( Z ) } [ { }] Z MS ˆ ( Z ) [ [ Z ] inside bracets is conditional MS
26 The variance of an estimator For the MMS estimator, the conditional MS is [ ] ] ˆ Z Z ( Z ) [ ] ] Z ( Z ) MMS var that is, the conditional variance of given Z. Averaging over Z yields [ ] [ var( Z )] [ ( Z )] which is the unconditional variance of the estimate average squared error over all possible observations ˆ MMS
27 The variance of an estimator General definition with the estimation error ~ Δ ˆ the epected value of the square of the estimation error is the estimator s variance or MS [ ] ~ MS var ( ˆ ) ( ˆ ) biased or unibiased The square root of the variance (or MS) ˆ var( ˆ ) is its standard error, also called standard deviation of the estimation error if ˆ if ˆ unbiased
28 The variance of an estimator Comparison of variances of an ML and MAP estimator Single observation This is due to the availability of of the a priori information [ ] { } { } ˆ ˆ var Δ ML ML [ ] { } [ ] ML MAP MAP w w ˆ var ) ( ˆ ˆ var <
29 Consistency and efficiency of estimators Consistency is here an asymptotic property An estimator of a nonrandom parameter is consistent estimator, if the estimate converges to the true value in some sense Using convergence in the mean square criterion then, Over Z lim ˆ, Z is the condition for consistency in the mean square sense for a random parameter [ ] ] lim [ ] ] ˆ, Z [ ~ (, )] Z lim Over Z and For unbiased estimators
30 Consistency and efficiency of estimators The Cramer-Rao Lower Bound (CRLB) and the Fisher Information Matri (FIM) CRLB: The mean square error corresponding to the estimator of a parameter cannot be smaller than a certain quantity Scalar case Scalar nonrandom parameter with an unbiased estimator [ ] ] ˆ, Z Δ J Λ J is the Fisher Information Λ Λ() is the lielihood function the true value of the parameter. J where
31 Consistency and efficiency of estimators The Cramer-Rao Lower Bound (CRLB) and the Fisher Information Matri (FIM) Scalar case For scalar random parameter with an unbiased estimator J is the Fisher Information If the estimator variance is equal to CRLB, then such an estimator is called efficient. [ ] [ ] Δ,, J where, ˆ Z p Z p J Z
32 Consistency and efficiency of estimators The Cramer-Rao Lower Bound (CRLB) and the Fisher Information Matri (FIM) Vector case For nonrandom vector parameter, the CRLB states that the covariance matri of an unbiased estimator is bounded from below: [ ˆ ][ ˆ ] ] T Z Z Δ J [ ] T T ln Λ [ ln Λ ] ln Λ J is the Fisher Information Matri (FIM) Λ() is the lielihood function the true value of the parameter J where [ [ ] ] T A B C Δ A B pos. def.
2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More informationVariations. ECE 6540, Lecture 10 Maximum Likelihood Estimation
Variations ECE 6540, Lecture 10 Last Time BLUE (Best Linear Unbiased Estimator) Formulation Advantages Disadvantages 2 The BLUE A simplification Assume the estimator is a linear system For a single parameter
More informationEstimation Tasks. Short Course on Image Quality. Matthew A. Kupinski. Introduction
Estimation Tasks Short Course on Image Quality Matthew A. Kupinski Introduction Section 13.3 in B&M Keep in mind the similarities between estimation and classification Image-quality is a statistical concept
More informationCramér-Rao Bounds for Estimation of Linear System Noise Covariances
Journal of Mechanical Engineering and Automation (): 6- DOI: 593/jjmea Cramér-Rao Bounds for Estimation of Linear System oise Covariances Peter Matiso * Vladimír Havlena Czech echnical University in Prague
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More informationEstimators as Random Variables
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until
More informationLecture Note 2: Estimation and Information Theory
Univ. of Michigan - NAME 568/EECS 568/ROB 530 Winter 2018 Lecture Note 2: Estimation and Information Theory Lecturer: Maani Ghaffari Jadidi Date: April 6, 2018 2.1 Estimation A static estimation problem
More informationEstimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators
Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let
More informationParametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory
Statistical Inference Parametric Inference Maximum Likelihood Inference Exponential Families Expectation Maximization (EM) Bayesian Inference Statistical Decison Theory IP, José Bioucas Dias, IST, 2007
More informationEstimation and Detection
stimation and Detection Lecture 2: Cramér-Rao Lower Bound Dr. ir. Richard C. Hendriks & Dr. Sundeep P. Chepuri 7//207 Remember: Introductory xample Given a process (DC in noise): x[n]=a + w[n], n=0,,n,
More informationf-domain expression for the limit model Combine: 5.12 Approximate Modelling What can be said about H(q, θ) G(q, θ ) H(q, θ ) with
5.2 Approximate Modelling What can be said about if S / M, and even G / G? G(q, ) H(q, ) f-domain expression for the limit model Combine: with ε(t, ) =H(q, ) [y(t) G(q, )u(t)] y(t) =G (q)u(t) v(t) We know
More informationBrief Review on Estimation Theory
Brief Review on Estimation Theory K. Abed-Meraim ENST PARIS, Signal and Image Processing Dept. abed@tsi.enst.fr This presentation is essentially based on the course BASTA by E. Moulines Brief review on
More informationEstimation Theory Fredrik Rusek. Chapters
Estimation Theory Fredrik Rusek Chapters 3.5-3.10 Recap We deal with unbiased estimators of deterministic parameters Performance of an estimator is measured by the variance of the estimate (due to the
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationMotivation CRLB and Regularity Conditions CRLLB A New Bound Examples Conclusions
A Survey of Some Recent Results on the CRLB for Parameter Estimation and its Extension Yaakov 1 1 Electrical and Computer Engineering Department University of Connecticut, Storrs, CT A Survey of Some Recent
More informationTechnique for Numerical Computation of Cramér-Rao Bound using MATLAB
Technique for Numerical Computation of Cramér-Rao Bound using MATLAB Hing Cheung So http://www.ee.cityu.edu.hk/~hcso Department of Electronic Engineering City University of Hong Kong H. C. So Page 1 The
More informationParametric Models. Dr. Shuang LIANG. School of Software Engineering TongJi University Fall, 2012
Parametric Models Dr. Shuang LIANG School of Software Engineering TongJi University Fall, 2012 Today s Topics Maximum Likelihood Estimation Bayesian Density Estimation Today s Topics Maximum Likelihood
More informationChapter 3: Maximum-Likelihood & Bayesian Parameter Estimation (part 1)
HW 1 due today Parameter Estimation Biometrics CSE 190 Lecture 7 Today s lecture was on the blackboard. These slides are an alternative presentation of the material. CSE190, Winter10 CSE190, Winter10 Chapter
More informationELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE)
1 ELEG 5633 Detection and Estimation Minimum Variance Unbiased Estimators (MVUE) Jingxian Wu Department of Electrical Engineering University of Arkansas Outline Minimum Variance Unbiased Estimators (MVUE)
More informationSTATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN
Massimo Guidolin Massimo.Guidolin@unibocconi.it Dept. of Finance STATISTICS/ECONOMETRICS PREP COURSE PROF. MASSIMO GUIDOLIN SECOND PART, LECTURE 2: MODES OF CONVERGENCE AND POINT ESTIMATION Lecture 2:
More informationDetection & Estimation Lecture 1
Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven
More informationSGN Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection
SG 21006 Advanced Signal Processing: Lecture 8 Parameter estimation for AR and MA models. Model order selection Ioan Tabus Department of Signal Processing Tampere University of Technology Finland 1 / 28
More informationModule 2. Random Processes. Version 2, ECE IIT, Kharagpur
Module Random Processes Version, ECE IIT, Kharagpur Lesson 9 Introduction to Statistical Signal Processing Version, ECE IIT, Kharagpur After reading this lesson, you will learn about Hypotheses testing
More informationPARAMETER ESTIMATION AND ORDER SELECTION FOR LINEAR REGRESSION PROBLEMS. Yngve Selén and Erik G. Larsson
PARAMETER ESTIMATION AND ORDER SELECTION FOR LINEAR REGRESSION PROBLEMS Yngve Selén and Eri G Larsson Dept of Information Technology Uppsala University, PO Box 337 SE-71 Uppsala, Sweden email: yngveselen@ituuse
More informationReview Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the
Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then
More informationDetection & Estimation Lecture 1
Detection & Estimation Lecture 1 Intro, MVUE, CRLB Xiliang Luo General Course Information Textbooks & References Fundamentals of Statistical Signal Processing: Estimation Theory/Detection Theory, Steven
More informationDensity Estimation: ML, MAP, Bayesian estimation
Density Estimation: ML, MAP, Bayesian estimation CE-725: Statistical Pattern Recognition Sharif University of Technology Spring 2013 Soleymani Outline Introduction Maximum-Likelihood Estimation Maximum
More informationSignal detection theory
Signal detection theory z p[r -] p[r +] - + Role of priors: Find z by maximizing P[correct] = p[+] b(z) + p[-](1 a(z)) Is there a better test to use than r? z p[r -] p[r +] - + The optimal
More informationOn the convergence of the iterative solution of the likelihood equations
On the convergence of the iterative solution of the likelihood equations R. Moddemeijer University of Groningen, Department of Computing Science, P.O. Box 800, NL-9700 AV Groningen, The Netherlands, e-mail:
More informationAdvanced Signal Processing Introduction to Estimation Theory
Advanced Signal Processing Introduction to Estimation Theory Danilo Mandic, room 813, ext: 46271 Department of Electrical and Electronic Engineering Imperial College London, UK d.mandic@imperial.ac.uk,
More informationEIE6207: Maximum-Likelihood and Bayesian Estimation
EIE6207: Maximum-Likelihood and Bayesian Estimation Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak
More informationSystem Identification
System Identification Lecture : Statistical properties of parameter estimators, Instrumental variable methods Roy Smith 8--8. 8--8. Statistical basis for estimation methods Parametrised models: G Gp, zq,
More informationModern Methods of Data Analysis - WS 07/08
Modern Methods of Data Analysis Lecture VIc (19.11.07) Contents: Maximum Likelihood Fit Maximum Likelihood (I) Assume N measurements of a random variable Assume them to be independent and distributed according
More informationLecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis
Lecture 3 1 Probability (90 min.) Definition, Bayes theorem, probability densities and their properties, catalogue of pdfs, Monte Carlo 2 Statistical tests (90 min.) general concepts, test statistics,
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationThe Joint MAP-ML Criterion and its Relation to ML and to Extended Least-Squares
3484 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 48, NO. 12, DECEMBER 2000 The Joint MAP-ML Criterion and its Relation to ML and to Extended Least-Squares Arie Yeredor, Member, IEEE Abstract The joint
More informationResearch Article Improved Estimators of the Mean of a Normal Distribution with a Known Coefficient of Variation
Probability and Statistics Volume 2012, Article ID 807045, 5 pages doi:10.1155/2012/807045 Research Article Improved Estimators of the Mean of a Normal Distribution with a Known Coefficient of Variation
More informationEfficient Monte Carlo computation of Fisher information matrix using prior information
Efficient Monte Carlo computation of Fisher information matrix using prior information Sonjoy Das, UB James C. Spall, APL/JHU Roger Ghanem, USC SIAM Conference on Data Mining Anaheim, California, USA April
More informationEstimation Theory Fredrik Rusek. Chapters 6-7
Estimation Theory Fredrik Rusek Chapters 6-7 All estimation problems Summary All estimation problems Summary Efficient estimator exists All estimation problems Summary MVU estimator exists Efficient estimator
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationTiming Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL
T F T I G E O R G A I N S T I T U T E O H E O F E A L P R O G R ESS S A N D 1 8 8 5 S E R V L O G Y I C E E C H N O Timing Recovery at Low SNR Cramer-Rao bound, and outperforming the PLL Aravind R. Nayak
More informationFrames and Phase Retrieval
Frames and Phase Retrieval Radu Balan University of Maryland Department of Mathematics and Center for Scientific Computation and Mathematical Modeling AMS Short Course: JMM San Antonio, TX Current Collaborators
More informationBayesian Methods / G.D. Hager S. Leonard
Bayesian Methods Recall Robot Localization Given Sensor readings z 1, z 2,, z t = z 1:t Known control inputs u 0, u 1, u t = u 0:t Known model t+1 t, u t ) with initial 1 u 0 ) Known map z t t ) Compute
More informationTerminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the
More informationy Xw 2 2 y Xw λ w 2 2
CS 189 Introduction to Machine Learning Spring 2018 Note 4 1 MLE and MAP for Regression (Part I) So far, we ve explored two approaches of the regression framework, Ordinary Least Squares and Ridge Regression:
More informationStatistics: Learning models from data
DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial
More informationMachine Learning Lecture 2
Machine Perceptual Learning and Sensory Summer Augmented 6 Computing Announcements Machine Learning Lecture 2 Course webpage http://www.vision.rwth-aachen.de/teaching/ Slides will be made available on
More informationRegression #3: Properties of OLS Estimator
Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with
More informationChapter 3. Point Estimation. 3.1 Introduction
Chapter 3 Point Estimation Let (Ω, A, P θ ), P θ P = {P θ θ Θ}be probability space, X 1, X 2,..., X n : (Ω, A) (IR k, B k ) random variables (X, B X ) sample space γ : Θ IR k measurable function, i.e.
More informationMaximum Likelihood Diffusive Source Localization Based on Binary Observations
Maximum Lielihood Diffusive Source Localization Based on Binary Observations Yoav Levinboo and an F. Wong Wireless Information Networing Group, University of Florida Gainesville, Florida 32611-6130, USA
More informationIntroduction to Maximum Likelihood Estimation
Introduction to Maximum Likelihood Estimation Eric Zivot July 26, 2012 The Likelihood Function Let 1 be an iid sample with pdf ( ; ) where is a ( 1) vector of parameters that characterize ( ; ) Example:
More informationParameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!
Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses
More informationEstimation techniques
Estimation techniques March 2, 2006 Contents 1 Problem Statement 2 2 Bayesian Estimation Techniques 2 2.1 Minimum Mean Squared Error (MMSE) estimation........................ 2 2.1.1 General formulation......................................
More informationST5215: Advanced Statistical Theory
Department of Statistics & Applied Probability Wednesday, October 5, 2011 Lecture 13: Basic elements and notions in decision theory Basic elements X : a sample from a population P P Decision: an action
More informationMachine Learning 4771
Machine Learning 4771 Instructor: Tony Jebara Topic 11 Maximum Likelihood as Bayesian Inference Maximum A Posteriori Bayesian Gaussian Estimation Why Maximum Likelihood? So far, assumed max (log) likelihood
More informationProof In the CR proof. and
Question Under what conditions will we be able to attain the Cramér-Rao bound and find a MVUE? Lecture 4 - Consequences of the Cramér-Rao Lower Bound. Searching for a MVUE. Rao-Blackwell Theorem, Lehmann-Scheffé
More informationA Few Notes on Fisher Information (WIP)
A Few Notes on Fisher Information (WIP) David Meyer dmm@{-4-5.net,uoregon.edu} Last update: April 30, 208 Definitions There are so many interesting things about Fisher Information and its theoretical properties
More informationCSE 559A: Computer Vision
CSE 559A: Computer Vision Fall 208: T-R: :30-pm @ Lopata 0 Instructor: Ayan Chakrabarti (ayan@wustl.edu). Course Staff: Zhihao ia, Charlie Wu, Han Liu http://www.cse.wustl.edu/~ayan/courses/cse559a/ Sep
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More informationRegression, Ridge Regression, Lasso
Regression, Ridge Regression, Lasso Fabio G. Cozman - fgcozman@usp.br October 2, 2018 A general definition Regression studies the relationship between a response variable Y and covariates X 1,..., X n.
More informationCSE 559A: Computer Vision Tomorrow Zhihao's Office Hours back in Jolley 309: 10:30am-Noon
CSE 559A: Computer Vision ADMINISTRIVIA Tomorrow Zhihao's Office Hours back in Jolley 309: 0:30am-Noon Fall 08: T-R: :30-pm @ Lopata 0 This Friday: Regular Office Hours Net Friday: Recitation for PSET
More information6. Vector Random Variables
6. Vector Random Variables In the previous chapter we presented methods for dealing with two random variables. In this chapter we etend these methods to the case of n random variables in the following
More informationParameter Estimation
Parameter Estimation Consider a sample of observations on a random variable Y. his generates random variables: (y 1, y 2,, y ). A random sample is a sample (y 1, y 2,, y ) where the random variables y
More informationThe problem we are interest in this lecture is that of estimating the value of an n{
Chapter 4 Introduction to Estimation Theory 4. Concepts of Probabilistic Estimation The problem we are interest in this lecture is that of estimating the value of an n{ dimensional vector of parameters
More informationLecture 3: Pattern Classification. Pattern classification
EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and
More informationBayesian estimation of chaotic signals generated by piecewise-linear maps
Signal Processing 83 2003 659 664 www.elsevier.com/locate/sigpro Short communication Bayesian estimation of chaotic signals generated by piecewise-linear maps Carlos Pantaleon, Luis Vielva, David Luengo,
More information10-704: Information Processing and Learning Fall Lecture 24: Dec 7
0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of
More informationProbabilistic Graphical Models
Probabilistic Graphical Models Lecture Notes Fall 2009 November, 2009 Byoung-Ta Zhang School of Computer Science and Engineering & Cognitive Science, Brain Science, and Bioinformatics Seoul National University
More information5. Discriminant analysis
5. Discriminant analysis We continue from Bayes s rule presented in Section 3 on p. 85 (5.1) where c i is a class, x isap-dimensional vector (data case) and we use class conditional probability (density
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationParametric Techniques Lecture 3
Parametric Techniques Lecture 3 Jason Corso SUNY at Buffalo 22 January 2009 J. Corso (SUNY at Buffalo) Parametric Techniques Lecture 3 22 January 2009 1 / 39 Introduction In Lecture 2, we learned how to
More informationFisher Information Matrix-based Nonlinear System Conversion for State Estimation
Fisher Information Matrix-based Nonlinear System Conversion for State Estimation Ming Lei Christophe Baehr and Pierre Del Moral Abstract In practical target tracing a number of improved measurement conversion
More informationThe Kalman Filter. Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience. Sarah Dance
The Kalman Filter Data Assimilation & Inverse Problems from Weather Forecasting to Neuroscience Sarah Dance School of Mathematical and Physical Sciences, University of Reading s.l.dance@reading.ac.uk July
More informationEIE6207: Estimation Theory
EIE6207: Estimation Theory Man-Wai MAK Dept. of Electronic and Information Engineering, The Hong Kong Polytechnic University enmwmak@polyu.edu.hk http://www.eie.polyu.edu.hk/ mwmak References: Steven M.
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More informationClassical Estimation Topics
Classical Estimation Topics Namrata Vaswani, Iowa State University February 25, 2014 This note fills in the gaps in the notes already provided (l0.pdf, l1.pdf, l2.pdf, l3.pdf, LeastSquares.pdf). 1 Min
More informationOn the convergence of the iterative solution of the likelihood equations
On the convergence of the iterative solution of the likelihood equations R. Moddemeijer University of Groningen, Department of Computing Science, P.O. Box 800, NL-9700 AV Groningen, The Netherlands, e-mail:
More informationA Practitioner s Guide to Generalized Linear Models
A Practitioners Guide to Generalized Linear Models Background The classical linear models and most of the minimum bias procedures are special cases of generalized linear models (GLMs). GLMs are more technically
More informationMachine Learning. Lecture 4: Regularization and Bayesian Statistics. Feng Li. https://funglee.github.io
Machine Learning Lecture 4: Regularization and Bayesian Statistics Feng Li fli@sdu.edu.cn https://funglee.github.io School of Computer Science and Technology Shandong University Fall 207 Overfitting Problem
More informationParametric Techniques
Parametric Techniques Jason J. Corso SUNY at Buffalo J. Corso (SUNY at Buffalo) Parametric Techniques 1 / 39 Introduction When covering Bayesian Decision Theory, we assumed the full probabilistic structure
More informationLECTURE NOTES FYS 4550/FYS EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO
LECTURE NOTES FYS 4550/FYS9550 - EXPERIMENTAL HIGH ENERGY PHYSICS AUTUMN 2013 PART I PROBABILITY AND STATISTICS A. STRANDLIE GJØVIK UNIVERSITY COLLEGE AND UNIVERSITY OF OSLO Before embarking on the concept
More informationMethods of evaluating estimators and best unbiased estimators Hamid R. Rabiee
Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance
More informationBayesian Decision Theory
Bayesian Decision Theory Selim Aksoy Department of Computer Engineering Bilkent University saksoy@cs.bilkent.edu.tr CS 551, Fall 2017 CS 551, Fall 2017 c 2017, Selim Aksoy (Bilkent University) 1 / 46 Bayesian
More information10. Linear Models and Maximum Likelihood Estimation
10. Linear Models and Maximum Likelihood Estimation ECE 830, Spring 2017 Rebecca Willett 1 / 34 Primary Goal General problem statement: We observe y i iid pθ, θ Θ and the goal is to determine the θ that
More informationPART I INTRODUCTION The meaning of probability Basic definitions for frequentist statistics and Bayesian inference Bayesian inference Combinatorics
Table of Preface page xi PART I INTRODUCTION 1 1 The meaning of probability 3 1.1 Classical definition of probability 3 1.2 Statistical definition of probability 9 1.3 Bayesian understanding of probability
More informationLecture 2: CDF and EDF
STAT 425: Introduction to Nonparametric Statistics Winter 2018 Instructor: Yen-Chi Chen Lecture 2: CDF and EDF 2.1 CDF: Cumulative Distribution Function For a random variable X, its CDF F () contains all
More informationComputer Vision Group Prof. Daniel Cremers. 3. Regression
Prof. Daniel Cremers 3. Regression Categories of Learning (Rep.) Learnin g Unsupervise d Learning Clustering, density estimation Supervised Learning learning from a training data set, inference on the
More informationMACHINE LEARNING ADVANCED MACHINE LEARNING
MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 22 MACHINE LEARNING Discrete Probabilities Consider two variables and y taking discrete
More informationInferring from data. Theory of estimators
Inferring from data Theory of estimators 1 Estimators Estimator is any function of the data e(x) used to provide an estimate ( a measurement ) of an unknown parameter. Because estimators are functions
More informationMachine Learning Lecture 2
Announcements Machine Learning Lecture 2 Eceptional number of lecture participants this year Current count: 449 participants This is very nice, but it stretches our resources to their limits Probability
More information2. A Basic Statistical Toolbox
. A Basic Statistical Toolbo Statistics is a mathematical science pertaining to the collection, analysis, interpretation, and presentation of data. Wikipedia definition Mathematical statistics: concerned
More informationThe Finite Sample Properties of the Least Squares Estimator / Basic Hypothesis Testing
1 The Finite Sample Properties of the Least Squares Estimator / Basic Hypothesis Testing Greene Ch 4, Kennedy Ch. R script mod1s3 To assess the quality and appropriateness of econometric estimators, we
More informationNotes on Discriminant Functions and Optimal Classification
Notes on Discriminant Functions and Optimal Classification Padhraic Smyth, Department of Computer Science University of California, Irvine c 2017 1 Discriminant Functions Consider a classification problem
More informationMACHINE LEARNING ADVANCED MACHINE LEARNING
MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,
More informationModel structure. Lecture Note #3 (Chap.6) Identification of time series model. ARMAX Models and Difference Equations
System Modeling and Identification Lecture ote #3 (Chap.6) CHBE 70 Korea University Prof. Dae Ryoo Yang Model structure ime series Multivariable time series x [ ] x x xm Multidimensional time series (temporal+spatial)
More informationCross entropy-based importance sampling using Gaussian densities revisited
Cross entropy-based importance sampling using Gaussian densities revisited Sebastian Geyer a,, Iason Papaioannou a, Daniel Straub a a Engineering Ris Analysis Group, Technische Universität München, Arcisstraße
More informationDS-GA 1002 Lecture notes 11 Fall Bayesian statistics
DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian
More informationMinimum Message Length Analysis of the Behrens Fisher Problem
Analysis of the Behrens Fisher Problem Enes Makalic and Daniel F Schmidt Centre for MEGA Epidemiology The University of Melbourne Solomonoff 85th Memorial Conference, 2011 Outline Introduction 1 Introduction
More informationCOMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION. Aditya K. Jagannatham and Bhaskar D.
COMPLEX CONSTRAINED CRB AND ITS APPLICATION TO SEMI-BLIND MIMO AND OFDM CHANNEL ESTIMATION Aditya K Jagannatham and Bhaskar D Rao University of California, SanDiego 9500 Gilman Drive, La Jolla, CA 92093-0407
More information