Modelling Non-linear and Non-stationary Time Series

Size: px
Start display at page:

Download "Modelling Non-linear and Non-stationary Time Series"

Transcription

1 Modelling Non-linear and Non-stationary Time Series Chapter 2: Non-parametric methods Henrik Madsen Advanced Time Series Analysis September 206 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September 206 / 45

2 Kernel estimators in time series analysis Non-parametric estimation Non-parametric estimation Free of parameters and model structure (apart from a smoothing constant). Useful if no prior information about the structure is available. The regression curve is a direct function of the data. Important methods: Kernel estimation, splines, and nearest neighbor. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

3 Introduction Kernel estimators in time series analysis Non-parametric estimation Assume that Y has PDF f(y), then f(y) = lim P(Y h < Y Y + h) h 0 2h. Assume further we have N observations. An estimator of f(y) is: ˆf(y) = [Number of observations in (y-h;y+h)] 2hN ˆf(x) = 2hN n χ ]x h,x+h] (x i ) where χ is the characteristic function: { if x i ]x h, x+h], χ [x h,x+h] (x i ) = 0 if x i ]x h, x+h] i= Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

4 Introduction Kernel estimators in time series analysis Non-parametric estimation Define then w(u) = ˆf(y) = N The general kernel estimator: ˆf(y) = Nh { 2 if u < 0 otherwise N ( ) h w y Yt h t= k(u) du = N ( ) y Yt k h t= Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

5 Kernel estimators in time series analysis The kernel estimator The kernel estimator Assume {Y t },{Z t } are strongly (or strictly) stationary processes Madsen (2008). The basic quantity estimated is then Robinson (983): E[g(Z t ) Y t = y] f Yt (y) (*) where f Yt (y) is the PDF of Y t. The solution to (*) is given by N ( ) [G(Z t ); y] = (N y Yt h) G(Z t )k h t= Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

6 Kernel estimators in time series analysis The kernel estimator - examples The kernel estimator Example Example 2 ˆfY (y) = [; y] = Nh N ( ) y Yt k h t= Notice the special case leads to estimation of Ê[G(Z t ) Y t = y] = [G(Z t); y] [; y] = g(z t ) = Y t+ E[Y t+ Y t = y] N G(Zt )k( y Yt N k( y Y t h ) h ) Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

7 Kernel estimators in time series analysis The kernel estimator The kernel estimator Assume that the theoretical relation is Y = g(x (),...,X (q) )+ǫ where ǫ is white noise, and g is an unknown continuous function. Given N observations O = {(X (),...,X(q), Y ),...,(X () N,...,X(q) N, Y N)} the goal is now to estimate the function g. If q = the kernel estimator for g given the observations is ĝ(x) = N N s= Y sk{h (x X s )} N N s= k{h (x X s )} Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

8 Kernel estimators in time series analysis The kernel estimator The kernel estimator It is shown in Robinson (983) that under reasonable assumptions about the stochastic process {X t } and the kernel, the kernel estimator converges in probability to the conditional expectation of Y: More precisely: ĝ(x) p E[Y X = x] N, h 0, Nh Then at every point of continuity of g(x): ĝ(x) p g(x) Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

9 Kernel functions Kernel estimators in time series analysis Kernel functions The function, k is the kernel, and h is called the bandwidth of the kernel. A kernel is a real and bounded function which satisfies k(u) du =. The kernel with bandwidth, h, is given by K h (u) = h k(u/h) Commonly used kernels are Gaussian, Epanechnikov kernel, rectangular, and triangular kernels. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

10 Kernel estimators in time series analysis Kernel functions Important kernels The Epanechnikov kernel with bandwidth h is given by : K Epa h (u) = 3 u2 ( 4h h 2)I { u h} The Gaussian kernel with bandwidth h is given by: Kh Gau (u) = ) ( h 2π exp u2 2h 2 Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

11 Kernel estimators in time series analysis Kernel functions - illustration Kernel functions K(u) Triangular Epanechnikov uniform Gauss u Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September 206 / 45

12 Kernel estimators in time series analysis Kernel functions Nearest Neighbor (NN) estimation Previously: h was fixed. Another (and often better) alternative is * k-nn (k Nearest Neighbor) (Note: here, k is a natural number - not the kernel). Since it is reasonable to link the value of k to N we often consider a fraction (or percentage) of the data. Often the symbol, ħ, is used in this case. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

13 Kernel estimators in time series analysis Multivariate kernels Kernel functions If we allow q > in () a kernel of q th order have to be used, i.e. a real bounded function k q (u,...,u q ), where k q (u)du =. By a generalization of the bandwidth h to the quadratic matrix h with the dimension q q the more general kernel estimator for g(x (),...,X (q) ) is ĝ(x) = N N s= Y sk q [(x X s )h ] N N s= k q[(x X s )h ] where x = (x,...,x q ) and X s = (X s (),...,X s (q) ). This is called the Nadaraya-Watson estimator. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

14 Kernel estimators in time series analysis Multivariate kernels - example Kernel functions If k q is parameterized using the normal distribution, N q (0, I), it is seen that h k q ( (x z)h ) is the value in the point, x of the q-dimensional normal density function with mean z and covariance h h T. Hence it is possible to take into account correlations in data. Y N(0, I), Y = (X z)h X N(z, hh T ) Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

15 Product kernel Kernel estimators in time series analysis Kernel functions In practice product kernels are used, i.e. k q (xh ) = q k(x i /h i ) where k is a kernel of order. Assuming that h i = h, i =,...,q the following generalization of the one dimensional case in () is obtained i= ĝ(x) = N N s= Y q s i= k{h (x i X s (i) )} q i= k{h (x i X s (i) )} N N s= In case of a Gaussian kernel, this corresponds to a diagonal covariance matrix. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

16 Kernel estimators in time series analysis Non-parametric LS Non-parametric LS If we define the weight w s (x) = q i= k{h (x i X s (i) )} q i= k{h (x i X s (i) )} N N s= then it is seen that the estimator corresponds to the local average: ĝ(x) = N N w s (x)y s s= Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

17 Kernel estimators in time series analysis Non-parametric LS - continued Non-parametric LS The estimate is actually a non-parametric least squares estimate at the point x. This is recognized from the fact that the solution to the least squares problem is given by arg min θ N N w s (x)(y s θ) 2 s= ˆθ LS (x) = N s= w s(x)y s N s= w s(x) Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

18 Kernel estimators in time series analysis Non-parametric LS - continued Non-parametric LS This shows that at each x, ĝ is essentially a weighted LS location estimate, i.e. ĝ(x) = ˆθ LS (x) N w s (x) = N ˆθ LS (x) The kernel estimate is equivalent to a local weighted least squares estimate. s= Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

19 Kernel estimators in time series analysis Non-parametric LS Variance of the non-parametric estimates To assess the variance of the curve estimate at point x Härdle (990) proposes the point-wise estimator given by ˆσ 2 (x) = N N w s (x)(y s ĝ(x s )) 2 s= where the weights are given as shown previously by w s (x) = q i= k{h (x i X s (i) )} q i= k{h (x i X s (i) )} N N s= Remembering the WLS interpretation this estimate seems reasonable. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

20 The bandwidth Kernel estimators in time series analysis Selection of bandwidth In analogy with smoothing in spectrum analysis the bandwidth h determines the smoothness of ĝ: If h is large the variance is small but the bias is large. If h is small the variance is large but the bias is small. In the limits: As h it is seen that ĝ(x) = Ȳ. As h 0 it is seen that ĝ(x) = Y i for x = (X () i,...,x (q) i ) and otherwise 0. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

21 Kernel estimators in time series analysis Selection of bandwidth Selection of bandwidth - Cross Validation The ultimate goal for the selection of h is to minimize (see Härdle (990)) MSE(h) = N [ĝ(x i ) g(x i )] 2 π(x i ) N i= Where π(...) is a weighting function which screens off some of the extreme observations. However g is unknown. An easy solution is the plug-in method, where Y i is used as an estimate of g(x i ). Hence, the criterion is MSE2(h) = N N [ĝ(x i ) Y i ] 2 π(x i ) i= but it is clear that MSE2(h) 0 for h 0. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

22 Kernel estimators in time series analysis The Leave one out method Selection of bandwidth The idea is to avoid that (X i ) is used in the estimate for Y i. For every data (X i, Y i ) we define an estimator ĝ (i) for Y i based on all data except (X i ). The N estimators ĝ (),...,ĝ (N) (called the leave one out estimators) are written ĝ (j) (x) = N N s j Y s s j q i= k{h (x i X s (i) )} q i= k{h (x i X s (i) )} Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

23 Kernel estimators in time series analysis The Leave one out method Selection of bandwidth Now the Cross-Validation criterion using the leave one out estimates ĝ (),...,ĝ (N) is CV(h) = N N [Y i ĝ (i) (X i )] 2 π(x i ) i= and the optimal value of h is then found by minimizing CV(h). It can be shown that under weak assumptions the estimate of the bandwidth ĥ that is obtained when minimizing the Cross-Validation criterion is asymptotic optimal, i.e. it minimizes () see Härdle (990). Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

24 Applications Non-parametric estimation of the conditional mean and variance Applications of kernel estimators Assume that a realization X,...,X N of a stochastic process {X t } is available. The goal is now to use the realization to estimate the functions g( ) and h( ) in the model X t = g(x t,...,x t p )+h(x t,...,x t q )ǫ t As in Tjöstheim (994) we shall use the notation M(x i,...,x id ) = E[X t X t i = x i,...,x t id = x id ] V(x i,...,x id ) = V[X t X t i = x i,...,x t id = x id ] One solution is to use the kernel estimator (other possibilities are splines, nearest neighbor or neural network estimates). Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

25 Applications Non-parametric estimation of the conditional mean and variance Applications of kernel estimators - continued Using a product kernel we obtain ˆM(x i,...,x id ) = N N i d s=i d + X d s j= k{h (x ij X s ij )} N+i d N i d +i s=i d + j= k{h (x ij X s ij )} Assuming that E(X t ) = 0 the estimator for V( ) is ˆV(x i,...,x id ) = N d N i d s=i d + X2 s j= k{h (x ij X s ij )} N+i d N i d +i s=i d + j= k{h (x ij X s ij )} Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

26 Applications Non-parametric estimation of the conditional mean and variance Applications of kernel estimators - continued If E[X t ] 0 it is clear that the above estimator is changed to ˆV(x i,...,x id ) = N d N i d s=i d + X2 s j= k{h (x ij X s ij )} N+i d N i d +i s=i d + j= k{h (x ij X s ij )} ˆM 2 (x i,...,x id ) Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

27 Applications Identification of functional relationship Non-parametric estimation of the conditional mean and variance Estimation of conditional mean and variance We have simulated 0 stochastic independent time series for both of the following non-linear models: A : X t = B : X t = { 0.8Xt +ǫ t, X t 0 0.8X t +.5+ǫ t, X t < 0 +Xt 2 ǫ t What kind of models are model A and B? Gaussian kernels with bandwidths, h = 0.6, are used. Simulation of 0 independent time series. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

28 Applications Identification of functional relationship - Estimated M(x) for model A Non-parametric estimation of the conditional mean and variance M(x) x The theoretical conditional mean and variance are shown with +. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

29 Applications Identification of functional relationship - Estimated M(x) for model B Non-parametric estimation of the conditional mean and variance M(x) x Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

30 Applications Identification of functional relationship - Estimated V(x) for model A Non-parametric estimation of the conditional mean and variance V(x) x Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

31 Applications Applications of kernel estimators - Estimated V(x) for model B Non-parametric estimation of the conditional mean and variance V(x) x Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

32 Applications Example: non-parametric estimation of the PDF Example: non-parametric estimation of the PDF An estimate of the probability density function (PDF) is Robinson (983) : ˆf(xi,...,x id ) = N i d + i N+i s=i d + j= d h k{h (x ij X s ij )} Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

33 Applications Non-parametric estimation of non-stationarity Example: Non-parametric estimation of non-stationarity Non-parametric methods can be applied to an identification of the structure of the existing relationships leading to proposals for a parametric model. An example of identifying the diurnal variation in a non-stationary time series is given in the following. Measurements of heat supply from 6 terrace houses in Kulladal, a suburb of Malmö in Sweden, are used to estimate the heat load as a function of the time of the day. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

34 Applications Heat supply vs. time of day Example data - Heat supply versus measurement number [kj/s] 80 Heat Supply No. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

35 Applications Heat supply vs. time of day Example data heat supply versus time of day [kj/s] Heat Supply Hour The observations are equidistantly sampled on the interval [0, 24] with 0.25 hour between the sample points. The transport delay between the consumers and the measuring instruments is not more than a couple of minutes. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

36 Applications Heat supply vs. time of day Dependence between data points Assumptions: The heat supply can be related to the time of day. There is no difference between the days for the considered period. Hence, the regression curve is a function only of the time of day, and it is this functional relationship, which will be considered. Thus, the assumption of independence between measurements is not fulfilled. The regression function contained in data is minor compared to a set of data with the same number of independent observations. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

37 Applications Heat supply vs. time of day Smoothing with different bandwidths Smoothing of the diurnal power load curve using an Epanechnikov kernel and bandwidths 0.25, 0.625,..., [kj/s] Nadaraya-Watson Estimator Epanechnikov Kernel Hour The characteristics of the non-smoothed averages gradually disappear, when the bandwidth is increased. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

38 Applications CV - approach for finding h Heat supply vs. time of day [(kj/s) 2 ] 9.75 CV-function Epanechnikov Kernel Bandwidth [hour] Obviously the minimum of the CV-function is obtained for a bandwidth close to.3. This value is most likely somewhat bigger than the visually chosen. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

39 Applications Heat supply vs. time of day Cross-Validation - Comments The drop at 5.5 in the morning will clearly disappear in a curve estimate with this bandwidth. These results indicate the need of non-parametric curve estimates in which the bandwidth is not only a function of the density of the predictor variables, but for instance also of the gradient of the regression curve. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

40 Cross-Validation Gaussian kernels Applications Heat supply vs. time of day Smoothing of the diurnal power load curve using a Gaussian kernel and bandwidths 0.025, 0.25,..., [kj/s] Nadaraya-Watson Estimator Gaussian Kernel Hour Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

41 Cross-validation Gaussian kernels Applications Heat supply vs. time of day Clearly it is difficult visually to observe any difference between the curve estimates obtained with the Epanechnikov and Gaussian kernel. A general conclusion : The choice of the kernel (window) is not important as long as it is reasonably smooth. It is the choice of the bandwidth that matters. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

42 Applications Estimation of dependence on external variables Example: Non-parametric estimation of (static) dependence on external variables The dependent variable is the heat load in a district heating system, and the independent variables are ambient and supply temperature. Data are collected in the district heating system in Esbjerg, Denmark, during the period August 4 to December 0, 989. Heat load p(t) versus ambient air temperature a(t) and supply temperature s(t). Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

43 Applications Applying the Epanechnikov kernel Estimation of dependence on external variables For the estimation of the regression surface the kernel method is applied using the Epanechnikov kernel (). A product kernel is used, i.e. the power load estimate, p, at time t, at ambient air temperature a and for supply temperature s is estimated as ˆp ht,h a,h s (t, a, s) = 3843 i= p i K ht (t t i )K ha (a a i )K hs (s s i ) K ht (t t i )K ha (a a i )K hs (s s i ) i=009 where K h (u) = h k(u/h). The bandwidths were chosen using a combination of Cross-Validation and visual inspection. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

44 Applications The Epanechnikov kernel estimate Estimation of dependence on external variables Kernel estimate of the dependence on time of day, t d, ambient air temperature, a and supply temperature, s, shown at t d = 7. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

45 Applications The Epanechnikov kernel estimate Estimate of the standard deviation Estimation of dependence on external variables Point-wise estimate of standard deviation of the surface estimates. Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

46 Applications Estimation of dependence on external variables Härdle, W. (990), Applied nonparametric regression, Cambridge University Press, New York. Madsen, H. (2008), Time Series Analysis,. edn, Chapman & Hall/CRC. Robinson, P. (983), Nonparametric estimators for time series, Journal of time series analysis 4(3), Tjöstheim, D. (994), Non-linear time series: A selective review, Scandinavian Journal of Statistics 2, Henrik Madsen (02427 Adv. TS Analysis) Lecture Notes September / 45

Nonparametric Methods

Nonparametric Methods Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis

More information

Econ 582 Nonparametric Regression

Econ 582 Nonparametric Regression Econ 582 Nonparametric Regression Eric Zivot May 28, 2013 Nonparametric Regression Sofarwehaveonlyconsideredlinearregressionmodels = x 0 β + [ x ]=0 [ x = x] =x 0 β = [ x = x] [ x = x] x = β The assume

More information

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β

Introduction. Linear Regression. coefficient estimates for the wage equation: E(Y X) = X 1 β X d β d = X β Introduction - Introduction -2 Introduction Linear Regression E(Y X) = X β +...+X d β d = X β Example: Wage equation Y = log wages, X = schooling (measured in years), labor market experience (measured

More information

41903: Introduction to Nonparametrics

41903: Introduction to Nonparametrics 41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific

More information

12 - Nonparametric Density Estimation

12 - Nonparametric Density Estimation ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6

More information

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao

Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics. Jiti Gao Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics Jiti Gao Department of Statistics School of Mathematics and Statistics The University of Western Australia Crawley

More information

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas

Density estimation Nonparametric conditional mean estimation Semiparametric conditional mean estimation. Nonparametrics. Gabriel Montes-Rojas 0 0 5 Motivation: Regression discontinuity (Angrist&Pischke) Outcome.5 1 1.5 A. Linear E[Y 0i X i] 0.2.4.6.8 1 X Outcome.5 1 1.5 B. Nonlinear E[Y 0i X i] i 0.2.4.6.8 1 X utcome.5 1 1.5 C. Nonlinearity

More information

Nonparametric Density Estimation (Multidimension)

Nonparametric Density Estimation (Multidimension) Nonparametric Density Estimation (Multidimension) Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann February 19, 2007 Setup One-dimensional

More information

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction

Nonparametric Regression Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Härdle, Müller, Sperlich, Werwarz, 1995, Nonparametric and Semiparametric Models, An Introduction Tine Buch-Kromann Univariate Kernel Regression The relationship between two variables, X and Y where m(

More information

AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET. Questions AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET

AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET. Questions AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET The Problem Identification of Linear and onlinear Dynamical Systems Theme : Curve Fitting Division of Automatic Control Linköping University Sweden Data from Gripen Questions How do the control surface

More information

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd

ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation. Petra E. Todd ECON 721: Lecture Notes on Nonparametric Density and Regression Estimation Petra E. Todd Fall, 2014 2 Contents 1 Review of Stochastic Order Symbols 1 2 Nonparametric Density Estimation 3 2.1 Histogram

More information

Local Polynomial Regression

Local Polynomial Regression VI Local Polynomial Regression (1) Global polynomial regression We observe random pairs (X 1, Y 1 ),, (X n, Y n ) where (X 1, Y 1 ),, (X n, Y n ) iid (X, Y ). We want to estimate m(x) = E(Y X = x) based

More information

Error distribution function for parametrically truncated and censored data

Error distribution function for parametrically truncated and censored data Error distribution function for parametrically truncated and censored data Géraldine LAURENT Jointly with Cédric HEUCHENNE QuantOM, HEC-ULg Management School - University of Liège Friday, 14 September

More information

Econometrics I. Lecture 10: Nonparametric Estimation with Kernels. Paul T. Scott NYU Stern. Fall 2018

Econometrics I. Lecture 10: Nonparametric Estimation with Kernels. Paul T. Scott NYU Stern. Fall 2018 Econometrics I Lecture 10: Nonparametric Estimation with Kernels Paul T. Scott NYU Stern Fall 2018 Paul T. Scott NYU Stern Econometrics I Fall 2018 1 / 12 Nonparametric Regression: Intuition Let s get

More information

Nonparametric Regression. Changliang Zou

Nonparametric Regression. Changliang Zou Nonparametric Regression Institute of Statistics, Nankai University Email: nk.chlzou@gmail.com Smoothing parameter selection An overall measure of how well m h (x) performs in estimating m(x) over x (0,

More information

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i

NADARAYA WATSON ESTIMATE JAN 10, 2006: version 2. Y ik ( x i NADARAYA WATSON ESTIMATE JAN 0, 2006: version 2 DATA: (x i, Y i, i =,..., n. ESTIMATE E(Y x = m(x by n i= ˆm (x = Y ik ( x i x n i= K ( x i x EXAMPLES OF K: K(u = I{ u c} (uniform or box kernel K(u = u

More information

Integral approximation by kernel smoothing

Integral approximation by kernel smoothing Integral approximation by kernel smoothing François Portier Université catholique de Louvain - ISBA August, 29 2014 In collaboration with Bernard Delyon Topic of the talk: Given ϕ : R d R, estimation of

More information

Time Series Analysis

Time Series Analysis Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Regression based methods, 1st part: Introduction (Sec.

More information

Introduction to Regression

Introduction to Regression Introduction to Regression p. 1/97 Introduction to Regression Chad Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/97 Acknowledgement Larry Wasserman, All of Nonparametric

More information

Local regression I. Patrick Breheny. November 1. Kernel weighted averages Local linear regression

Local regression I. Patrick Breheny. November 1. Kernel weighted averages Local linear regression Local regression I Patrick Breheny November 1 Patrick Breheny STA 621: Nonparametric Statistics 1/27 Simple local models Kernel weighted averages The Nadaraya-Watson estimator Expected loss and prediction

More information

Goodness-of-fit tests for the cure rate in a mixture cure model

Goodness-of-fit tests for the cure rate in a mixture cure model Biometrika (217), 13, 1, pp. 1 7 Printed in Great Britain Advance Access publication on 31 July 216 Goodness-of-fit tests for the cure rate in a mixture cure model BY U.U. MÜLLER Department of Statistics,

More information

Nonparametric Econometrics

Nonparametric Econometrics Applied Microeconometrics with Stata Nonparametric Econometrics Spring Term 2011 1 / 37 Contents Introduction The histogram estimator The kernel density estimator Nonparametric regression estimators Semi-

More information

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Outline in High Dimensions Using the Rodeo Han Liu 1,2 John Lafferty 2,3 Larry Wasserman 1,2 1 Statistics Department, 2 Machine Learning Department, 3 Computer Science Department, Carnegie Mellon University

More information

3 Nonparametric Density Estimation

3 Nonparametric Density Estimation 3 Nonparametric Density Estimation Example: Income distribution Source: U.K. Family Expenditure Survey (FES) 1968-1995 Approximately 7000 British Households per year For each household many different variables

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University JSM, 2015 E. Christou, M. G. Akritas (PSU) SIQR JSM, 2015

More information

Gaussian Process Regression

Gaussian Process Regression Gaussian Process Regression 4F1 Pattern Recognition, 21 Carl Edward Rasmussen Department of Engineering, University of Cambridge November 11th - 16th, 21 Rasmussen (Engineering, Cambridge) Gaussian Process

More information

A Note on Data-Adaptive Bandwidth Selection for Sequential Kernel Smoothers

A Note on Data-Adaptive Bandwidth Selection for Sequential Kernel Smoothers 6th St.Petersburg Workshop on Simulation (2009) 1-3 A Note on Data-Adaptive Bandwidth Selection for Sequential Kernel Smoothers Ansgar Steland 1 Abstract Sequential kernel smoothers form a class of procedures

More information

Introduction to Probability and Stocastic Processes - Part I

Introduction to Probability and Stocastic Processes - Part I Introduction to Probability and Stocastic Processes - Part I Lecture 2 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark

More information

5 Operations on Multiple Random Variables

5 Operations on Multiple Random Variables EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y

More information

Introduction Dual Representations Kernel Design RBF Linear Reg. GP Regression GP Classification Summary. Kernel Methods. Henrik I Christensen

Introduction Dual Representations Kernel Design RBF Linear Reg. GP Regression GP Classification Summary. Kernel Methods. Henrik I Christensen Kernel Methods Henrik I Christensen Robotics & Intelligent Machines @ GT Georgia Institute of Technology, Atlanta, GA 30332-0280 hic@cc.gatech.edu Henrik I Christensen (RIM@GT) Kernel Methods 1 / 37 Outline

More information

Chapter 3: Maximum Likelihood Theory

Chapter 3: Maximum Likelihood Theory Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood

More information

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones

Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones Introduction to machine learning and pattern recognition Lecture 2 Coryn Bailer-Jones http://www.mpia.de/homes/calj/mlpr_mpia2008.html 1 1 Last week... supervised and unsupervised methods need adaptive

More information

Additive Isotonic Regression

Additive Isotonic Regression Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive

More information

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators

Estimation theory. Parametric estimation. Properties of estimators. Minimum variance estimator. Cramer-Rao bound. Maximum likelihood estimators Estimation theory Parametric estimation Properties of estimators Minimum variance estimator Cramer-Rao bound Maximum likelihood estimators Confidence intervals Bayesian estimation 1 Random Variables Let

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update 3. Juni 2013) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend

More information

ECE 636: Systems identification

ECE 636: Systems identification ECE 636: Systems identification Lectures 3 4 Random variables/signals (continued) Random/stochastic vectors Random signals and linear systems Random signals in the frequency domain υ ε x S z + y Experimental

More information

Introduction to Regression

Introduction to Regression Introduction to Regression Chad M. Schafer May 20, 2015 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures Cross Validation Local Polynomial Regression

More information

PREWHITENING-BASED ESTIMATION IN PARTIAL LINEAR REGRESSION MODELS: A COMPARATIVE STUDY

PREWHITENING-BASED ESTIMATION IN PARTIAL LINEAR REGRESSION MODELS: A COMPARATIVE STUDY REVSTAT Statistical Journal Volume 7, Number 1, April 2009, 37 54 PREWHITENING-BASED ESTIMATION IN PARTIAL LINEAR REGRESSION MODELS: A COMPARATIVE STUDY Authors: Germán Aneiros-Pérez Departamento de Matemáticas,

More information

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II)

Contents Lecture 4. Lecture 4 Linear Discriminant Analysis. Summary of Lecture 3 (II/II) Summary of Lecture 3 (I/II) Contents Lecture Lecture Linear Discriminant Analysis Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University Email: fredriklindsten@ituuse Summary of lecture

More information

Introduction to Regression

Introduction to Regression Introduction to Regression Chad M. Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/100 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Nonparametric Regression

Nonparametric Regression Nonparametric Regression Econ 674 Purdue University April 8, 2009 Justin L. Tobias (Purdue) Nonparametric Regression April 8, 2009 1 / 31 Consider the univariate nonparametric regression model: where y

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2016 F, FRI Uppsala University, Information Technology 13 April 2016 SI-2016 K. Pelckmans

More information

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity

Local linear multiple regression with variable. bandwidth in the presence of heteroscedasticity Local linear multiple regression with variable bandwidth in the presence of heteroscedasticity Azhong Ye 1 Rob J Hyndman 2 Zinai Li 3 23 January 2006 Abstract: We present local linear estimator with variable

More information

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING

EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING EXAM IN STATISTICAL MACHINE LEARNING STATISTISK MASKININLÄRNING DATE AND TIME: August 30, 2018, 14.00 19.00 RESPONSIBLE TEACHER: Niklas Wahlström NUMBER OF PROBLEMS: 5 AIDING MATERIAL: Calculator, mathematical

More information

Unsupervised Kernel Dimension Reduction Supplemental Material

Unsupervised Kernel Dimension Reduction Supplemental Material Unsupervised Kernel Dimension Reduction Supplemental Material Meihong Wang Dept. of Computer Science U. of Southern California Los Angeles, CA meihongw@usc.edu Fei Sha Dept. of Computer Science U. of Southern

More information

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model.

Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model. Minimax Rate of Convergence for an Estimator of the Functional Component in a Semiparametric Multivariate Partially Linear Model By Michael Levine Purdue University Technical Report #14-03 Department of

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Test for Discontinuities in Nonparametric Regression

Test for Discontinuities in Nonparametric Regression Communications of the Korean Statistical Society Vol. 15, No. 5, 2008, pp. 709 717 Test for Discontinuities in Nonparametric Regression Dongryeon Park 1) Abstract The difference of two one-sided kernel

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

Sliced Inverse Regression

Sliced Inverse Regression Sliced Inverse Regression Ge Zhao gzz13@psu.edu Department of Statistics The Pennsylvania State University Outline Background of Sliced Inverse Regression (SIR) Dimension Reduction Definition of SIR Inversed

More information

Nonparameteric Regression:

Nonparameteric Regression: Nonparameteric Regression: Nadaraya-Watson Kernel Regression & Gaussian Process Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro,

More information

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.

Statement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:. MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss

More information

Nonparametric Regression. Badr Missaoui

Nonparametric Regression. Badr Missaoui Badr Missaoui Outline Kernel and local polynomial regression. Penalized regression. We are given n pairs of observations (X 1, Y 1 ),...,(X n, Y n ) where Y i = r(x i ) + ε i, i = 1,..., n and r(x) = E(Y

More information

Local Polynomial Modelling and Its Applications

Local Polynomial Modelling and Its Applications Local Polynomial Modelling and Its Applications J. Fan Department of Statistics University of North Carolina Chapel Hill, USA and I. Gijbels Institute of Statistics Catholic University oflouvain Louvain-la-Neuve,

More information

AGEC 661 Note Eleven Ximing Wu. Exponential regression model: m (x, θ) = exp (xθ) for y 0

AGEC 661 Note Eleven Ximing Wu. Exponential regression model: m (x, θ) = exp (xθ) for y 0 AGEC 661 ote Eleven Ximing Wu M-estimator So far we ve focused on linear models, where the estimators have a closed form solution. If the population model is nonlinear, the estimators often do not have

More information

On the Robust Modal Local Polynomial Regression

On the Robust Modal Local Polynomial Regression International Journal of Statistical Sciences ISSN 683 5603 Vol. 9(Special Issue), 2009, pp 27-23 c 2009 Dept. of Statistics, Univ. of Rajshahi, Bangladesh On the Robust Modal Local Polynomial Regression

More information

OR MSc Maths Revision Course

OR MSc Maths Revision Course OR MSc Maths Revision Course Tom Byrne School of Mathematics University of Edinburgh t.m.byrne@sms.ed.ac.uk 15 September 2017 General Information Today JCMB Lecture Theatre A, 09:30-12:30 Mathematics revision

More information

Lecture 11. Multivariate Normal theory

Lecture 11. Multivariate Normal theory 10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances

More information

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models

More information

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego

Model-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships

More information

CS 556: Computer Vision. Lecture 21

CS 556: Computer Vision. Lecture 21 CS 556: Computer Vision Lecture 21 Prof. Sinisa Todorovic sinisa@eecs.oregonstate.edu 1 Meanshift 2 Meanshift Clustering Assumption: There is an underlying pdf governing data properties in R Clustering

More information

12. Cholesky factorization

12. Cholesky factorization L. Vandenberghe ECE133A (Winter 2018) 12. Cholesky factorization positive definite matrices examples Cholesky factorization complex positive definite matrices kernel methods 12-1 Definitions a symmetric

More information

Introduction to Regression

Introduction to Regression Introduction to Regression David E Jones (slides mostly by Chad M Schafer) June 1, 2016 1 / 102 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures

More information

DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA

DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA Statistica Sinica 18(2008), 515-534 DESIGN-ADAPTIVE MINIMAX LOCAL LINEAR REGRESSION FOR LONGITUDINAL/CLUSTERED DATA Kani Chen 1, Jianqing Fan 2 and Zhezhen Jin 3 1 Hong Kong University of Science and Technology,

More information

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model

Minimum Hellinger Distance Estimation in a. Semiparametric Mixture Model Minimum Hellinger Distance Estimation in a Semiparametric Mixture Model Sijia Xiang 1, Weixin Yao 1, and Jingjing Wu 2 1 Department of Statistics, Kansas State University, Manhattan, Kansas, USA 66506-0802.

More information

System Identification, Lecture 4

System Identification, Lecture 4 System Identification, Lecture 4 Kristiaan Pelckmans (IT/UU, 2338) Course code: 1RT880, Report code: 61800 - Spring 2012 F, FRI Uppsala University, Information Technology 30 Januari 2012 SI-2012 K. Pelckmans

More information

Machine Learning Basics: Maximum Likelihood Estimation

Machine Learning Basics: Maximum Likelihood Estimation Machine Learning Basics: Maximum Likelihood Estimation Sargur N. srihari@cedar.buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics 1. Learning

More information

Nonparametric Modal Regression

Nonparametric Modal Regression Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric

More information

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 4 Spatial Point Patterns Definition Set of point locations with recorded events" within study

More information

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O

O Combining cross-validation and plug-in methods - for kernel density bandwidth selection O O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem

More information

Acceleration of some empirical means. Application to semiparametric regression

Acceleration of some empirical means. Application to semiparametric regression Acceleration of some empirical means. Application to semiparametric regression François Portier Université catholique de Louvain - ISBA November, 8 2013 In collaboration with Bernard Delyon Regression

More information

Transformation and Smoothing in Sample Survey Data

Transformation and Smoothing in Sample Survey Data Scandinavian Journal of Statistics, Vol. 37: 496 513, 2010 doi: 10.1111/j.1467-9469.2010.00691.x Published by Blackwell Publishing Ltd. Transformation and Smoothing in Sample Survey Data YANYUAN MA Department

More information

1 Empirical Likelihood

1 Empirical Likelihood Empirical Likelihood February 3, 2016 Debdeep Pati 1 Empirical Likelihood Empirical likelihood a nonparametric method without having to assume the form of the underlying distribution. It retains some of

More information

Statistical Machine Learning from Data

Statistical Machine Learning from Data January 17, 2006 Samy Bengio Statistical Machine Learning from Data 1 Statistical Machine Learning from Data Other Artificial Neural Networks Samy Bengio IDIAP Research Institute, Martigny, Switzerland,

More information

Lecture Notes 15 Prediction Chapters 13, 22, 20.4.

Lecture Notes 15 Prediction Chapters 13, 22, 20.4. Lecture Notes 15 Prediction Chapters 13, 22, 20.4. 1 Introduction Prediction is covered in detail in 36-707, 36-701, 36-715, 10/36-702. Here, we will just give an introduction. We observe training data

More information

Lecture 4 Discriminant Analysis, k-nearest Neighbors

Lecture 4 Discriminant Analysis, k-nearest Neighbors Lecture 4 Discriminant Analysis, k-nearest Neighbors Fredrik Lindsten Division of Systems and Control Department of Information Technology Uppsala University. Email: fredrik.lindsten@it.uu.se fredrik.lindsten@it.uu.se

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

26. Filtering. ECE 830, Spring 2014

26. Filtering. ECE 830, Spring 2014 26. Filtering ECE 830, Spring 2014 1 / 26 Wiener Filtering Wiener filtering is the application of LMMSE estimation to recovery of a signal in additive noise under wide sense sationarity assumptions. Problem

More information

System Identification: From Data to Model

System Identification: From Data to Model 1 : From Data to Model With Applications to Aircraft Modeling Division of Automatic Control Linköping University Sweden Dynamic Systems 2 A Dynamic system has an output response y that depends on (all)

More information

4 Nonparametric Regression

4 Nonparametric Regression 4 Nonparametric Regression 4.1 Univariate Kernel Regression An important question in many fields of science is the relation between two variables, say X and Y. Regression analysis is concerned with the

More information

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity.

LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. LECTURES 2-3 : Stochastic Processes, Autocorrelation function. Stationarity. Important points of Lecture 1: A time series {X t } is a series of observations taken sequentially over time: x t is an observation

More information

Gaussian with mean ( µ ) and standard deviation ( σ)

Gaussian with mean ( µ ) and standard deviation ( σ) Slide from Pieter Abbeel Gaussian with mean ( µ ) and standard deviation ( σ) 10/6/16 CSE-571: Robotics X ~ N( µ, σ ) Y ~ N( aµ + b, a σ ) Y = ax + b + + + + 1 1 1 1 1 1 1 1 1 1, ~ ) ( ) ( ), ( ~ ), (

More information

Model Selection for Gaussian Processes

Model Selection for Gaussian Processes Institute for Adaptive and Neural Computation School of Informatics,, UK December 26 Outline GP basics Model selection: covariance functions and parameterizations Criteria for model selection Marginal

More information

Nonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors

Nonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors Nonparametric Estimation of Regression Functions In the Presence of Irrelevant Regressors Peter Hall, Qi Li, Jeff Racine 1 Introduction Nonparametric techniques robust to functional form specification.

More information

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes

COMP 551 Applied Machine Learning Lecture 20: Gaussian processes COMP 55 Applied Machine Learning Lecture 2: Gaussian processes Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: (herke.vanhoof@mcgill.ca) Class web page: www.cs.mcgill.ca/~hvanho2/comp55

More information

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017

COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 COMS 4721: Machine Learning for Data Science Lecture 10, 2/21/2017 Prof. John Paisley Department of Electrical Engineering & Data Science Institute Columbia University FEATURE EXPANSIONS FEATURE EXPANSIONS

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

A New Test in Parametric Linear Models with Nonparametric Autoregressive Errors

A New Test in Parametric Linear Models with Nonparametric Autoregressive Errors A New Test in Parametric Linear Models with Nonparametric Autoregressive Errors By Jiti Gao 1 and Maxwell King The University of Western Australia and Monash University Abstract: This paper considers a

More information

14 - Gaussian Stochastic Processes

14 - Gaussian Stochastic Processes 14-1 Gaussian Stochastic Processes S. Lall, Stanford 211.2.24.1 14 - Gaussian Stochastic Processes Linear systems driven by IID noise Evolution of mean and covariance Example: mass-spring system Steady-state

More information

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands

Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Asymptotic Multivariate Kriging Using Estimated Parameters with Bayesian Prediction Methods for Non-linear Predictands Elizabeth C. Mannshardt-Shamseldin Advisor: Richard L. Smith Duke University Department

More information

Titolo Smooth Backfitting with R

Titolo Smooth Backfitting with R Rapporto n. 176 Titolo Smooth Backfitting with R Autori Alberto Arcagni, Luca Bagnato ottobre 2009 Dipartimento di Metodi Quantitativi per le Scienze Economiche ed Aziendali Università degli Studi di Milano

More information

Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University

Mean-Shift Tracker Computer Vision (Kris Kitani) Carnegie Mellon University Mean-Shift Tracker 16-385 Computer Vision (Kris Kitani) Carnegie Mellon University Mean Shift Algorithm A mode seeking algorithm Fukunaga & Hostetler (1975) Mean Shift Algorithm A mode seeking algorithm

More information

Probabilistic & Unsupervised Learning

Probabilistic & Unsupervised Learning Probabilistic & Unsupervised Learning Gaussian Processes Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc ML/CSML, Dept Computer Science University College London

More information

Nonparametric Density Estimation

Nonparametric Density Estimation Nonparametric Density Estimation Econ 690 Purdue University Justin L. Tobias (Purdue) Nonparametric Density Estimation 1 / 29 Density Estimation Suppose that you had some data, say on wages, and you wanted

More information

Bickel Rosenblatt test

Bickel Rosenblatt test University of Latvia 28.05.2011. A classical Let X 1,..., X n be i.i.d. random variables with a continuous probability density function f. Consider a simple hypothesis H 0 : f = f 0 with a significance

More information

Malaya J. Mat. 2(1)(2014) 35 42

Malaya J. Mat. 2(1)(2014) 35 42 Malaya J. Mat. 2)204) 35 42 Note on nonparametric M-estimation for spatial regression A. Gheriballah a and R. Rouane b, a,b Laboratory of Statistical and stochastic processes, Univ. Djillali Liabs, BP

More information