It's Only Fitting. Fitting model to data parameterizing model estimating unknown parameters in the model

Similar documents
Chances are good Modeling stochastic forces in ecology

Mathematical statistics

Introduction to Maximum Likelihood Estimation

Chapter 7, continued: MANOVA

Review. December 4 th, Review

Economics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,

Confidence Intervals, Testing and ANOVA Summary

Allee effects in stochastic populations

SIMULTANEOUS CONFIDENCE BANDS FOR THE PTH PERCENTILE AND THE MEAN LIFETIME IN EXPONENTIAL AND WEIBULL REGRESSION MODELS. Ping Sa and S.J.

Math 494: Mathematical Statistics

Modern Methods of Data Analysis - WS 07/08

Introduction to Statistical Analysis

Wald s theorem and the Asimov data set

Multivariate Regression

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Simple and Multiple Linear Regression

QUALIFYING EXAMINATION II IN MATHEMATICAL STATISTICS SATURDAY, MAY 9, Examiners: Drs. K. M. Ramachandran and G. S. Ladde

Théorie Analytique des Probabilités

The Chi-Square Distributions

" M A #M B. Standard deviation of the population (Greek lowercase letter sigma) σ 2

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015


Random and mixed effects models

Asymptotic Statistics-VI. Changliang Zou

F & B Approaches to a simple model

Review of Statistics 101

The Chi-Square Distributions

Psychology 282 Lecture #4 Outline Inferences in SLR

236 Chapter 4 Applications of Derivatives

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Mathematical statistics

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Homework 2: Simple Linear Regression

Statistical machine learning and kernel methods. Primary references: John Shawe-Taylor and Nello Cristianini, Kernel Methods for Pattern Analysis

Econ 583 Homework 7 Suggested Solutions: Wald, LM and LR based on GMM and MLE

Inferences about central values (.)

STAT 501 EXAM I NAME Spring 1999

7. Random variables as observables. Definition 7. A random variable (function) X on a probability measure space is called an observable

y ˆ i = ˆ " T u i ( i th fitted value or i th fit)

Two hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45

So far our focus has been on estimation of the parameter vector β in the. y = Xβ + u

STT 843 Key to Homework 1 Spring 2018

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Brandon C. Kelly (Harvard Smithsonian Center for Astrophysics)

Categorical Data Analysis Chapter 3

Mathematical statistics

Maximum Likelihood (ML) Estimation

Bayesian inference for sample surveys. Roderick Little Module 2: Bayesian models for simple random samples

Primer on statistics:

Signal detection theory

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Central Limit Theorem ( 5.3)

Institute of Actuaries of India

Stat 427/527: Advanced Data Analysis I

Stat 5101 Lecture Notes

DEPARTMENT OF MATHEMATICS AND STATISTICS QUALIFYING EXAMINATION II IN MATHEMATICAL STATISTICS SATURDAY, SEPTEMBER 24, 2016

Financial Econometrics and Quantitative Risk Managenent Return Properties

Estimators as Random Variables

Example: Baseball winning % for NL West pennant winner

Linear models and their mathematical foundations: Simple linear regression

Fundamental Probability and Statistics

ECON 4160, Autumn term Lecture 1

Interval estimation. October 3, Basic ideas CLT and CI CI for a population mean CI for a population proportion CI for a Normal mean

f(x θ)dx with respect to θ. Assuming certain smoothness conditions concern differentiating under the integral the integral sign, we first obtain

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

Probability and Stochastic Processes

PRINCIPLES OF STATISTICAL INFERENCE

ST495: Survival Analysis: Hypothesis testing and confidence intervals

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

DATA IN SERIES AND TIME I. Several different techniques depending on data and what one wants to do

n =10,220 observations. Smaller samples analyzed here to illustrate sample size effect.

1; (f) H 0 : = 55 db, H 1 : < 55.

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

You can compute the maximum likelihood estimate for the correlation

BTRY 4090: Spring 2009 Theory of Statistics

Using R in Undergraduate and Graduate Probability and Mathematical Statistics Courses*

Economics 583: Econometric Theory I A Primer on Asymptotics

Negative binomial distribution and multiplicities in p p( p) collisions

Some New Aspects of Dose-Response Models with Applications to Multistage Models Having Parameters on the Boundary

Stochastic logistic model with environmental noise. Brian Dennis Dept Fish and Wildlife Resources and Dept Statistical Science University of Idaho

Stat 643 Problems. a) Show that V is a sub 5-algebra of T. b) Let \ be an integrable random variable. Find EÐ\lVÑÞ

Multiple Regression Analysis

Table 1: Fish Biomass data set on 26 streams

An R # Statistic for Fixed Effects in the Linear Mixed Model and Extension to the GLMM

[y i α βx i ] 2 (2) Q = i=1

Interval Estimation III: Fisher's Information & Bootstrapping

Correlation and Regression

Empirical Likelihood

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY

STAT Financial Time Series

SIMULATION - PROBLEM SET 1

SPRING 2007 EXAM C SOLUTIONS

Class 24. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Applied Regression Modeling: A Business Approach Chapter 2: Simple Linear Regression Sections

Home Page. Title Page. Page 1 of c. Go Back. Full Screen. Close. Quit

Hypothesis Testing, Power, Sample Size and Confidence Intervals (Part 2)

Multivariate Statistics

Recall the Basics of Hypothesis Testing

Transcription:

It's Only Fitting Fitting model to data parameterizing model estimating unknown parameters in the model Likelihood: an example Cohort of 8! individuals observe survivors at times >œ 1, 2, 3,..., : 8", 8,..., 8 Suppose we want to fit the demographic survival model. The probability of this particular outcome under the model is the joint probability mass function: Pr ÒR" œ 8", R œ 8,..., R œ 8 R! œ 8! Ó œ :8 ˆ 8 :8 ˆ 8 â:8 ˆ 8 "! "."

œ 8>." Š : Ð".:Ñ 8 > 8 8.8 > >." > For instance, suppose the numbers of survivors, starting at >œ0, are: 100, 93, 80, 69, 61, 56, 52, 45, 43, 40, 36. Then PrÒR" œ 93, R œ 80,..., R œ 36 R œ 100Ó œ! ˆ 100 ˆ 93 93 : Ð".:Ñ 80 : Ð".:Ñ... ˆ 40 : Ð".:Ñ. 93 100. 93 80 93. 80 36 œp: a b 36 40 36 R. A. Fisher (1922): use as as estimate of : the value that mazimizes P: a b.

P: a b: likelihood function (joint pmf or pdf, evaluated at the data) s: : maximum likelihood (ML) estimate of : Likelihoods are usually products, so in maximizing it is usually easier to work with log-likelihoods: ln PÐ:Ñ œ! 8>." ln Š 3! 8 ln : 8 > > 3! Ð8.8 Ñ ln : >." > Set. ln P a: bî.: œ 0 and solve for :: s:œ! 8! 8 > >." For the example data, s:œ0.8998

Hypothesis tests and confidence intervals Suppose of : :! is the true (but unknown) value log-likelihood ratio quantity: K œ.òln PÐ: Ñ. ln PÐ:ÑÓ s Math-stat theorem (Wilks 1938): K converges in distribution to a chi-square distribution with 1 df (same type of convergence as CLT)!

Result allows for a statistical hypothesis test: H:! :œ:! :Á: H: "! Calculate K, reject H! in favor of H " if K exceeds the "!!Ð".! Ñth percentile,! Ð"Ñ, of a chi-square (1 df) distribution, for an approximate size! test. Result also allows an approximate 100(1.! Ñ % confidence interval for : (set of all values of :! for which H! would not be rejected): Ö: À PÐ:Ñ s! ln. ln PÐ:! Ñ Ÿ! Ð"ÑÎ The 95th percentile of a chi-square(1) distribution is about!þ!& Ð"Ñ œ 3.843. Finding the endpoinds of the confidence interval usually requires numerical rootfinding. For the data, the 95% CI for : is (0.8750, 0.9215)

Likelihoods for other models NLAR model: \ œ 2Ð\ Ñ 3 I > >." > Recall that the one-step conditional pdf is :B ˆ B œ > >.". " a B.2B a 5 bb > >." a5 1b exp. Suppose B!, B",..., B are the (transformed) observations of population abundances. ): vector of unknown parameters in 2aB : variance of noise, I 5 > Likelihood function: PÐ)5, Ñ œ :ÐB l B Ñ > >." >." b

Log-likelihood: ln PÐ, Ñ œ " ln :ÐB l B Ñ )5 > >." œ. ÐÎÑ ln Ð 1Ñ. ÐÎÑ ln Ð5 Ñ. c"îð Ñd! cb. 2ÐB Ñd 5 > >." ML estimates s), 5 s are the values of ) and 5 that maximize P or ln P. ex. environmental stochastic Ricker model (\ œ ln R ): > > \ \ œ \ 3 +.,/ >." > >." 3 I> Data: 8!, 8",..., 8 Transformed data: B0 œ ln 80, B1 œ ln 81,..., B œ ln 8 ) œ c+,, d w

ln PÐ+,,, Ñ œ. ÐÎÑ ln Ð 1Ñ. ÐÎÑ ln Ð5 Ñ 5. c"îð 5 Ñd! ab. B. + 3,/ B >." b > >." Multivariate NLAR model (such as LPA): \ œ 2Ð\ Ñ 3 I > >." > :ÐB> l B>." Ñ " 5. œ kdk. a 1b w." exp c.ðb.2 ÑD ÐB.2 ÑÎ d > >." > >." 5 : number of state variables in \ > 2 œ 2aB b >." >." Lilelihood is a product of multivariate normal pdf's log-likelihood is: ln PÐ)D, Ñ œ. Ð5ÎÑ ln Ð 1Ñ. ÐÎÑ ln kdk. Ð"ÎÑ! w." ÐB. 2 Ñ D ÐB. 2 Ñ > >." > >."

Properties of ML estimates: asymptotically unbiased E Š s) p ) as p_ consistent PrŠ ¹ s ). ) ¹ K % p 1 as p_ asymptotically efficient VŠ )s p theoretical minimum as p_ asymptotically normal: distribution of s ) converges to a normal distribution as p_

Conditional least squares estimation Conditional expected value: E ˆ \ \ œ B œ 2aB b > >." >." >." ) Conditional sum of squares: UÐ) Ñ œ! cb. 2ÐB ) Ñd > >." Conditional least squares (CLS) estimates of the parameters in ) jointly minimize UÐ) Ñ

Properties of CLS estimates: asymptotically unbiased consistent asymptotically robust to mis-specification of distribution of I > (not effficient)

Estimating D (ML or CLS) Fitted model yields series of residuals (or residual vectors): / œ B.2ÐB s > > >." Ñ s >." for >œ 1, 2,...,, where 2ÐB Ñis the skeleton evaluated at the ML or CLS estimate of ) Let Hœ c/, /,..., / d ( 5 ) then " D s w œhhî

Confidence intervals, hypothesis tests Profile likelihood (ML estimates) = ) œ / = is a parameter vector of interest / vector of nuisance parameters is a Pa=/D,, b: likelihood function PŠ =/, s, Ds : likelihood function = = maximized for fixed = ( profile likelihood function of = ) PŠ =/D sss,, : maximized likelihood function (unrestricted)

Approx 100(1.! Ñ% confidence region for = is the set of all values of = for which K œ. ln P Š =/, s, D s. ln PŠ =/D s, s, s where = = Ÿ! a( b ( is the number of parameters in = Bootstrapping (ML or CLS estimates) /", /,..., /: residuals (or residual vectors) /", /,..., /: random sample with replacement from the centered residuals (i.e. subtract the average from each of /", /,..., / to center them around zero)

Generate a bootstrapped data set B, B, B,..., B (simulated data set from estimated model) using B œ 2ÐB s Ñ 3 / > >." >! " Fit the model to B!, B", B,..., B, obtaining bootstrapped parameter estimates ), D Repeat 2000 or so times (using residuals & estimates from original fitted model), obtaining 2000 sets of bootstrapped parameter estimates (or test statistics, or functions of parameter estimates such as eigenvalues or Lyaponov exponents)

Resulting 2000 values constitute a good (statistically consistent) estimate of the variability of the estimates e.g., 2.5th and 97.5th percentiles of the bootstrapped estimates of a parameter are a valid approximate 95% confidence interval for the parameter

Model evaluation Goodness of fit If the model fits well, then the residuals / ", /,..., / should behave approximately like independent, identically distributed normal (or multivariate normal) variates Evaluating model fit revolves around the residuals e.g. autocorrelation tests, residual plots to look for unchanging variance, normal probability plots, etc. R-square : proportion of variability in data accounted for by fitted model Validation (predictive ability of the model) Ideally, collect new data, independently from the data used to fit the model (possibly with experimental, model-suggested perturbations!)