REGRESSION WITH SPATIALLY MISALIGNED DATA. Lisa Madsen Oregon State University David Ruppert Cornell University

Similar documents
Point-Referenced Data Models

Properties of the least squares estimates

Problem Selected Scores

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Gauge Plots. Gauge Plots JAPANESE BEETLE DATA MAXIMUM LIKELIHOOD FOR SPATIALLY CORRELATED DISCRETE DATA JAPANESE BEETLE DATA

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

1 One-way analysis of variance

Time Series Analysis

Probability and Statistics Notes

Outline of GLMs. Definitions

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Weighted Least Squares

Applied Regression. Applied Regression. Chapter 2 Simple Linear Regression. Hongcheng Li. April, 6, 2013

University of Oxford. Statistical Methods Autocorrelation. Identification and Estimation

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Spatial statistics, addition to Part I. Parameter estimation and kriging for Gaussian random fields

Part IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015

Multiple Linear Regression

Generalized Linear Models. Kurt Hornik

Econ 583 Final Exam Fall 2008

Association studies and regression

[y i α βx i ] 2 (2) Q = i=1

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

Maximum Likelihood Estimation

Simple Regression Model Setup Estimation Inference Prediction. Model Diagnostic. Multiple Regression. Model Setup and Estimation.

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Weighted Least Squares

Integrated Likelihood Estimation in Semiparametric Regression Models. Thomas A. Severini Department of Statistics Northwestern University

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Covariance function estimation in Gaussian process regression

Statistics for analyzing and modeling precipitation isotope ratios in IsoMAP

Likelihood-Based Methods

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Lecture 6 Multiple Linear Regression, cont.

Ch 2: Simple Linear Regression

Kriging Luc Anselin, All Rights Reserved

ESTIMATING THE MEAN LEVEL OF FINE PARTICULATE MATTER: AN APPLICATION OF SPATIAL STATISTICS

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

STAT5044: Regression and Anova

ECE531 Lecture 10b: Maximum Likelihood Estimation

Summer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.

Stat 579: Generalized Linear Models and Extensions

Fractional Imputation in Survey Sampling: A Comparative Review

Estimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.

Kriging models with Gaussian processes - covariance function estimation and impact of spatial sampling

Linear Methods for Prediction

First Year Examination Department of Statistics, University of Florida

Chapter 1. Linear Regression with One Predictor Variable

Regression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Math 423/533: The Main Theoretical Topics

Linear Models in Machine Learning

MEI Exam Review. June 7, 2002

Linear models and their mathematical foundations: Simple linear regression

Linear Regression Models P8111

Spatio-temporal prediction of site index based on forest inventories and climate change scenarios

Parameter estimation: ACVF of AR processes

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Model Selection, Estimation, and Bootstrap Smoothing. Bradley Efron Stanford University

BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation

MAT3379 (Winter 2016)

STAT 3A03 Applied Regression With SAS Fall 2017

Stat 579: Generalized Linear Models and Extensions

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff

Statistics 910, #5 1. Regression Methods

Spatial Lasso with Application to GIS Model Selection. F. Jay Breidt Colorado State University

1 Mixed effect models and longitudinal data analysis

Stat 5102 Final Exam May 14, 2015

Max. Likelihood Estimation. Outline. Econometrics II. Ricardo Mora. Notes. Notes

STAT5044: Regression and Anova. Inyoung Kim

MAT 3379 (Winter 2016) FINAL EXAM (PRACTICE)

MATH 829: Introduction to Data Mining and Analysis Linear Regression: statistical tests

Regression Estimation Least Squares and Maximum Likelihood

LDA, QDA, Naive Bayes

Brief Review on Estimation Theory

K. Model Diagnostics. residuals ˆɛ ij = Y ij ˆµ i N = Y ij Ȳ i semi-studentized residuals ω ij = ˆɛ ij. studentized deleted residuals ɛ ij =

Model Selection for Geostatistical Models

For more information about how to cite these materials visit

1 Introduction to Generalized Least Squares

Introduction The framework Bias and variance Approximate computation of leverage Empirical evaluation Discussion of sampling approach in big data

where x and ȳ are the sample means of x 1,, x n

Statistics & Data Sciences: First Year Prelim Exam May 2018

Introduction to Estimation Methods for Time Series models Lecture 2

Models for spatial data (cont d) Types of spatial data. Types of spatial data (cont d) Hierarchical models for spatial data

F & B Approaches to a simple model

Statistics - Lecture Three. Linear Models. Charlotte Wickham 1.

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

Ch 3: Multiple Linear Regression

Ph.D. Qualifying Exam Monday Tuesday, January 4 5, 2016

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

Linear Regression (9/11/13)

If we want to analyze experimental or simulated data we might encounter the following tasks:

Chapter 4: Asymptotic Properties of the MLE (Part 2)

Econometrics I, Estimation

Regression, Ridge Regression, Lasso

The Slow Convergence of OLS Estimators of α, β and Portfolio. β and Portfolio Weights under Long Memory Stochastic Volatility

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Regression Review. Statistics 149. Spring Copyright c 2006 by Mark E. Irwin

Transcription:

REGRESSION ITH SPATIALL MISALIGNED DATA Lisa Madsen Oregon State University David Ruppert Cornell University

SPATIALL MISALIGNED DATA 10 X X X X X X X X 5 X X X X X 0 X 0 5 10

OUTLINE 1. Introduction 2. The Krige and Regress (KR) Estimator 3. The Maximum Likelihood (ML) Estimator 4. Simulation Results 5. Conclusions

Notation = vector of responses observed at locations s 1,..., s n X = vector of unobserved predictors at locations s 1,..., s n = vector of observations from predictor process at locations t 1,..., t m ɛ= error vector at locations s 1,..., s n

1. = β 0 1 n 1 + β 1 X + ɛ. Assumptions 2. X,, and ɛ are generated by spatially autocorrelated stationary Gaussian processes. 3. X and are generated by the same spatial process. 4. X and ɛ are independent of each other. 5. Spatial autocorrelations are each given by a parametric model.

Model = β 0 1 n 1 + β 1 X + ɛ ( X ɛ N n (0, Σ ɛ ) ) [ ΣX Σ X ]) N n+m (µ X 1 (n+m) 1, Σ X Σ ) Σ X, Σ, Σ X depend on θ X = (θ X,1 θ X,2 θ X,3 ) Σ ɛ depends on θ ɛ = (θ ɛ,1 θ ɛ,2 θ ɛ,3

KRIGE AND REGRESS Kriging ˆX = ˆµ X 1 n 1 + Σ X Σ 1 ( ˆµ X1 m 1 ) where ˆµ X is the best linear unbiased estimator of µ X : so that ˆµ X = 1 1 mσ 1 1 1 m Σ 1 1 m 1 ˆX = Λ where Λ depends on Σ and Σ X.

KRIGE AND REGRESS Regression where e can write ˆβ KR = ( ˆX Σ 1 ɛ ˆX = ˆX) 1 ˆX Σ 1 ɛ [ ] 1 n 1 ˆX. ˆβ 1,KR = ( Λ MΛ ) 1 Λ M where M = 1 1 n Σ 1 ɛ 1 n 1 Σ 1 ɛ Σ 1 ɛ 1 n n Σ 1 ɛ.

Variance of ˆβ KR Starting with the identity we get var( ˆβ 1,KR ) = E[var( ˆβ KR )] + var[e( ˆβ KR )], var( ˆβ 1,KR ) = β 2 1[E(Q 3 Q 2 2 )+E(Q 2 1Q 2 2 ) (E(Q 1 Q 1 2 )) 2 ]+E(Q 1 2 ) where Q 1 = Λ Σ 1 ɛ Σ X Σ 1 Q 2 = Λ Σ 1 ɛ Λ + Λ Σ 1 ɛ Q 3 = Λ Σ 1 ɛ (Σ X Σ X Σ 1 Σ X )Σ 1 ɛ (1 1 n Σ X Σ 1 1 1 m)µ X Λ.

Estimating θ X from e are assuming N (µ X 1 m 1, Σ ) and, with h ij =distance between sampling locations of i and j, (Σ ) ij = { θx,1, h ij = 0 θ X,2 exp( θ X,3 h ij ), h ij > 0 Estimate the parameters θ = (θ X,1, θ X,2, θ X,3 ) by Restricted Maximum Likelihood (REML).

Estimating θ ɛ Given N (Xβ, Σ ɛ ), can t use REML because X not known. May work to estimate θ ɛ from approximate residuals ˆɛ = ˆX ˆβ where ˆβ is the unweighted KR estimate of β. e take Σ ɛ = σ ɛ identity where σ ɛ is scalar.

Approximate Estimators ˆX = ˆµ X 1 n 1 + ˆΣ ˆΣ 1 X ( ˆµ X1 m 1 ) ˆβ KR = ( ˆX ˆX) 1 ˆX var( ˆβ 1,KR ) = ˆβ 1,KR[E( 2 ˆQ ˆQ 2 3 2 ) + E( ˆQ 2 1 1 + E( ˆQ 2 ) ˆQ 2 2 ) (E( ˆQ 1 ˆQ 1 2 )) 2 ] where ˆQ 1, ˆQ2, and ˆQ 3 are as before except with Σ and Σ X estimated.

Point Estimates ˆβ 1,KR KR Estimates of β 1 40 35 30 25 20 15 10 5 0 5 Known Estimated Covariance Parameters

30 Variance Estimates of ˆβ 1,KR 0.5*log(Est. Var./True Var.) 25 20 15 10 5 0 Known Estimated Covariance Parameters

Naive Variance Estimates of ˆβ 1,KR : var naïve = ( ˆX Σ 1 ɛ ˆX) 1 1.6 Naive Variance Estimates 1.4 1.2 1 0.8 0.6 0.4 0.2 Known Covariance Parameters

Naive Variance Estimates of ˆβ 1,KR Log(Naive Var. Est.) 6 4 2 0 2 4 Unknown Covariance Parameters

Nominal 95% Confidence Intervals ˆβ KR ± 1.96 var Covariance Parameters Unknown % Coverage 89.46 Average idth 1.33e+10

e will show Consistency of ˆβ KR N( ˆβKR β) D N(0, Σ L ). hy bother? If we find the maximum likelihood estimates β ML by a Newton-Raphson maximization of the likelihood function with consistent parameter estimates as starting values, then N( ˆβML β) D N(0, E[I] 1 ), where E[I] is the information matrix.

Consistency of ˆβ KR Notation Suppose we have N iid observations [ 1 1 ] [ N N ] where each i is n 1 and each i is m 1. 0 5 10 0 5 10... 0 5 10 0 5 10

Consistency of ˆβ KR Assumptions Assume [Xi] ( [ ]) ΣX Σ X i iid N µ X 1 (n+m) 1, Σ X Σ, ɛ i iid N(0, σ 2 I), and i = X i β + ɛ i, where X i = [1 n 1 X i ] and β = [ β0 β 1 ]

Consistency of ˆβ KR More assumptions ˆµ X = 1 [ N N i=1 (1 1 1 m ˆΣ 1 m 1) 1 1 ˆΣ 1 1 m i] θ X is estimated by REML E( ˆX 1 ˆX 1 ) is invertible Σ X 0 then N( ˆβ KR β) D N(0, Σ L ).

N-consistency of ˆβKR The asymptotic covariance matrix of ˆβ KR is Σ L = covariance when only β unknown + loss of efficiency for estimating µ X + loss of efficiency for estimating θ X

Maximum Likelihood [ ], ith N iid observations i i the negative log-likelihood is 2l = N log Σ + N i=1 V i Σ 1 V i. where Σ is a block-diagonal matrix with copies of [ ] β 2 Σ 1 = 1 Σ X + Σ ɛ β 1 Σ X β 1 Σ X Σ along the diagonal and [ i (β 0 + β 1 µ X )1 n 1 V i = i µ X 1 m 1 ].

Let φ = Efficiency of the Maximum Likelihood Estimator ] [β 0 β 1 µ X θ X θ ɛ. If we compute φ ML by a Newton-Raphson minimization of 2l with consistent parameter estimates as initial values, then N( ˆφML φ) D N(0, E[I(φ)] 1 ), where ( 2 l {E[I(φ)]} ij = E φ i φ j provided some regularity conditions are met. ).

A Variance Estimator for ˆβ ML e can use an information-based variance estimator: or var 1 ( ˆφ ML ) = I 1 ( ˆφ ML ) var 2 ( ˆφ ML ) = [E(I)] 1 φ= ˆφML. var 1 often fails to be positive definite in simulations.

Simulation Study 16 treatments with 75 replications each Fixed parameters: β 0 = β 1 = µ X = 1, θ X,1 = θ ɛ,1 = 0 Remaining parameters varied to define treatments: θ ɛ2 {0.1, 1} θ ɛ3 {0.5, 3} θ X2 {0.25, 1} θ X3 {0.5, 1.5} n = m = 50, but N = 1

Simulation Sampling Locations 0 5 10 0 5 10 Locations

How θ 3 Affects Covariance C(h) = exp( θ 3 h) 1 0.8 Exponential Covariogram θ 3 =0.5 θ 3 =1 θ 3 =2.5 θ 3 =5 Covariance 0.6 0.4 0.2 0 0 2 4 6 8 10 Distance=h

Simulation Study Treatments TR: 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 θ ɛ,2 + + + + + + + + θ ɛ,3 + + + + + + + + θ X,2 + + + + + + + + θ X,3 + + + + + + + + + indicates the larger value. indicates the smaller value.

Simulation Study Variance Estimates 40 30 Sign(Est. Var.)*Log Est. Var. 20 10 0 10 20 30 40 log(monte Carlo MSE) sgn(est. Var)log Est. Var. Est. θ X,2 /Est. θ X,3 <0.075 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 Experiment

Simulation Study KR vs. ML 5 4 3 MSE(ML)=6.816 MSE(KR)=35.98 Bias(ML)= 0.075 Bias(KR)=0.173 2 ML 1 0 1 2 2 1 0 1 2 3 4 5 KR

Conclusions e have no good variance estimator for ˆβ KR. Information-based variance estimator for ˆβ ML appears useful, even when N = 1, provided no numerical problems. ML yields more precise point estimates than KR.