Curve alignment and functional PCA

Size: px
Start display at page:

Download "Curve alignment and functional PCA"

Transcription

1 Curve alignment and functional PCA Juhyun Par* Department of Mathematics and Statistics, Lancaster University, Lancaster, U.K. Abstract When dealing with multiple curves as functional data, it is a common practice to apply functional PCA to summarize and characterize random variation in finite dimension. Often functional data however exhibits additional variability that distorts the assumed common structure. This is recognized as the problem of curve registration. While the registration step is routinely employed, this is considered as a preprocessing step prior to any serious analysis. Consequently, the effect of alignment is mostly ignored in subsequent analyses and is not well understood. We revisit the issue by particularly focusing on the effect of variability on the FPCA and illustrate the phenomena from a borrowed perturbation viewpoint. This allows us to quantify bias in estimating eigenvalues and eigenfunctions when curves are not aligned. Some discussion on statistical implication is given. Keywords functional data analysis; curve alignment; registration; functional principal component analysis. 1 Introduction Repeated measurements in the form of curves are increasingly common in various scientific applications including biomedicine, physical sciences and econometrics. Usually the sample of curves is assumed to have some homogeneous structure in functional shape, while allowed for individual variability. Summarising and characterising variability in parsimonious manner is one of the aims of functional data analysis (Ramsay and Silverman, 2002, 2005). It is much

2 desirable that a few common components are able to extract most variability and are easy to interpret (Par et al. 2007). Functional PCA utilises the well-nown Karhunen-Loève expansion to provide an optimal representation of the function with a small number of common components. This is based on the assumption that the underlying random function shares the common mean and covariance function. Functional PCA also provides a useful representation when functional linear model is employed with functional explanatory variables. Figure 1 about here. Often functional data exhibits additional variability. Figure 1 shows a typical example of functional data, taen from the famous Zürich longitudinal growth studies. (Gasser et al., 1984). The left hand side plot shows height measurements of boys taen over years since childhood, with measurements points mared by circles. An observation that these are monotonically increasing does not seem so informative but simply confirms that children grow over! The plot on the right shows estimated velocity curves. Now distinctive features are apparent, which is often the case for functional data that consideration of the 1st or 2nd derivatives allows more meaningful comparison. In particular with children growth processes, it is well nown that each individual reaches its maximum growth at different, while the overall process shares common features between children. The presence of bumps appearing in the velocity curves was considered as an important characteristic of growth processes and not surprisingly the occurrence of such bumps varies between child to child. It is natural to allow variability but the consequence may sound severe in that a simple averaging would not produce a proper mean and most common statistics cannot be pursued in a usual manner. In the longitudinal data framewor, this phenomenon may be described by mixed effect models but there seems no notion of phase variability itself. Alternative approach would be to tae shape invariant models, which is a semi-parametric approach where a parametric form is assumed for phase variability. If no parametric form is assumed, the model is equivalent to those for curve alignments in functional data. This is treated as registration problem in functional data analysis, for which several methods have been developed. Basically when the functions exhibit identifiable features, curves can be aligned to match those features or landmars (Gasser and Kneip, 1995). This wors well as long as features are correctly identified. Several other methods have been developed to automate the procedure when the features are less prominent, see Ramsay and Silverman (2005) and references therein. Although the issue of phase variability has been rightly acnowledged, most analyses treat registration as a preprocessing step and thus do not consider any carry-on effects on later analyses. We point out that in practice, lie any other statistical procedure, the

3 registration step requires careful human intervention and experiences as well as additional estimation and smoothing step to mae it right. What happens then if registration was not carried out or was done improperly? This question is also relevant when, in modern applied sciences with abundant data, often PCA or a type of factor analysis is carried out to find a low-dimensional representation, ignoring phase variability. Most theoretical analyses view the goal of registration as finding a correct mean function and do not go beyond it. It may sound absurd to worry about the second moment property when the first moment is already in trouble. But as will be explained later, this is not necessarily a linear process and there is a systematic relationship between processes with and without phase variability. For example, Figure 2 demonstrates the effect of additional variability on the number of significant components in function PCA. The top row in Figure 2 shows an example of data with (right) and without (left) variability and the bottom row is the result of functional PCA. When additional variability is present, PCA produces many more uncorrelated principal components scores than anticipated (bottom right). Some issues with interpretability in PCA may be attributed to an improper registration. Figure 2 about here. A recent wor of Kneip and Ramsay (2007) addresses a similar issue (and more) and propose a new procedure to combine registration to fit functional PCA models, extending the convex averaging idea of registration (Liu and Müller, 2004). We also revisit the issue but our focus lies on understanding and quantifying the effect of imperfect registration on functional PCA. This is illustrated based on perturbation theory and numerical examples. Some discussion is given for a simple diagnostic chec for registration methods and a correction for functional PCA. 2 fpca with variability 2.1 Notation for fpca To fix the idea, consider a stochastic process X L 2 (T ) with compact support T = [0, T ], with the mean function µ(t) = E[X(t)] and the covariance function γ(s, t) = Cov(X(s), X(t)). Assume that T E[X(t)2 ] <. Let λ 1 λ 2... be the ordered eigenvalues of the covariance operator defined through γ with the corresponding eigenfunctions φ 1, φ 2,.... We assume that λ <. Then X(t) = µ(t) + ξ φ (t), (1) =1

4 where E[ξ] = 0 and E[ξ j ξ ] = λ j I(j = ). Thus in our analysis we will automatically assume that X admits the representation of (1). With a sample of curves available, these quantities are replaced by their estimates and a finite number of components {φ } are usually considered sufficient to extract significant observed variation. Theoretical properties of estimators are studied in Dauxois et al. (1982), Rice and Silverman (1991), Kenip (1994) and Hall et al. (2006). 2.2 Time variability and warping function When variability is present, we observe a sample from Y ( ) = X(η( )), where η( ) is called a warping function to describe transformation on axis. The warping function η( ) is assumed to satisfy the following conditions: (A1) η is a smooth monotone function; (A2) E[η(t)] = t for all t T ; (A3) T Var[η(t)] <. (A4) η is independent of amplitude variability (ξs). These are standard assumptions when dealing with registration (Gasser and Kneip, 1995). 2.3 Covariance function estimation without registration Suppose that {Y 1,, Y n } are independent random functions from Y. An empirical covariance function estimator is ˆγ Y (s, t) = 1 n {Y i (s) n Ȳ (s)}{y (t) Ȳ (t)}, i=1 where Ȳ (t) = 1 n n i=1 Y i(s). To study statistical properties of the estimator, this may be written as ˆγ Y = γ Y + n 1/2 {n 1/2 (ˆγ Y γ Y )}, (2) where γ Y (s, t) = Cov(Y (s), Y (t)). The usual asymptotic properties of functional PCA rely on the convergence of the latter term {n 1/2 (ˆγ Y γ Y )}, see Hall and Hosseini-Nasab (2006). However, closeness to γ Y is not so much of interest when additional variability is present. Instead we view the estimator as an approximation to γ X and thus decompose the estimator as ˆγ Y = γ X + (γ Y γ X ) + (ˆγ Y γ Y ).

5 The term ˆγ Y γ Y follows the usual asymptotics discussed with (2) and is negligible compared to the bias term γ Y γ X. Hence, from now on we will focus on the bias term only. 2.4 Approximation to covariance function We introduce the following notations to describe the observed process Y. µ(s) = E[Y (s)] φ(s) = φ(η(s)) Also let us define γ η (s, t) = E[{η(s) s}{η(t) t}]. Replacing t by η(t) in (1), we obtain a representation of Y. This is not any longer orthonomal decomposition but provides a useful starting point to study ˆγ Y. Combined with independence between η and ξ in (A4) we obtain E[ˆγ Y (s, t)] = E[{ µ(s) µ η (s)}{ µ(s) µ η (t)}] + n 1 λ E[φ (η i (s)) n φ (s) φ (t)]. Under further smoothness assumptions, the expression can be simplified further, which we summarise in the following proposition. Proposition 1. Assume that γ(s, t) = λ φ (s)φ (t) and has continuous second derivatives. Then, we have γ Y (s, t) = γ(s, t) + γ η v(s, t) + o( γ η ) where v(s, t) = γ η(s, t) γ η µ (s)µ (t) + γ η(s, t) γ η + γ η(s, s) γ η λ φ (s)φ (t) λ 2 φ (s)φ (t) + γ η(t, t) γ η λ 2 φ (s)φ (t). Proof of Proposition 1 To simplify the expression, observe that from (A2) any smooth function g of η can be approximated using Taylor expansion as g(η(s)) = g(s) + g (s)(η(s) s) + g (s) 2 (η(s) s)2 + o p (1).

6 Applying this approximation to µ and φ leads to γ Y (s, t) = γ(s, t) + γ η (s, t)µ (s)µ (t) +γ η (s, t) + γ η(t, t) 2 λ φ (s)φ (t) + γ η(s, s) 2 λ φ (s)φ (t) + o( γ η ) λ φ (s)φ (t) where γ η = sup γ η (s, t). Note that ˆγ Y is not an unbiased estimator and γ Y = E[ n 1 ˆγ n Y ]. Though not neglible itself, phase variability is still assumed to be small relative to amplitude variability.. This opens up the possibility of viewing the bias as additive perturbation. If we assume that the process is well approximated by a finite number of basis functions, it is equivalent to wor with a finite dimensional operator. Then the problem is reduced to finding eigenvalues and eigenfunctions for a perturbed covariance operator. 2.5 Perturbation theory of eigenvalue problem Consider the covariance function of the form γ ε (s, t) = γ(s, t) + εv(s, t) for some function v. Define the corresponding operator by Γ ε φ = γ ε φ for φ L 2. We write the corresponding covariance operator by Γ ε = Γ + εv. The perturbation theory formulates the problem of finding eigenvalues and eigenfunctions as finding approximation terms in order of ε. where Γφ = λ φ = 0, 1, 2, Γ ε φ (ε) = λ (ε)φ (ε) λ (ε) = λ + ελ (1) + ε 2 λ (2) + φ (ε) = φ + εφ (1) + ε 2 φ (2) +

7 Basically, since the {φ l }s are orthonomal bases, any function can be written in terms of φ l. If φ (r) = l c lφ l and the constants can be determined inductively. If we require that φ (ε) be normalised, then 1 =< φ (ε), φ (ε) >. For illustration, below are given first and second order approximation, which quantify the bias. The expansion is not limited to any order but we only list first two order approximation results for simplification. As a reference, general rth order approximation is provided in Appendix. Here the eigenfunction expansions assume that eigenvalues are distinct. First order approximation: λ (1) = < V φ, φ > φ (1) = l < V φ, φ l > φ l λ λ l Therefore we may write λ (ε) = λ + ε < φ, V φ > +O(ε 2 ) φ (ε) = φ + ε < φ, V φ l > φ l + O(ε 2 ). λ λ l l Second order approximation: λ (2) = < V φ, φ (1) >= l φ (2) = c φ + m c m φ m < V φ, φ l > 2 λ λ l c = 1 2 < φ(1), φ(1) >= 1 < V φ, φ l > 2 2 (λ λ l ) 2 l c m = l < V φ, φ l >< V φ m, φ l > (λ λ l )(λ λ m ) < V φ, φ >< V φ, φ m > (λ λ m ) 2 Note that this framewor is not limited to the case of registration error. Additional smoothing bias, sampling error or discrete approximation error can be formulated in a similar

8 manner. For example Silverman (1996) used a similar argument to show benefit of smoothing in functional PCA. In the case of sampling error, as in (2), the expression is given with ε = n 1/2, which was the basis of asymptotic expansions of eigenvalues and eigenfunctions in Hall and Hosseini-Nasab (2006). However note that unlie their results, bias here does not diminish as sample size grows. 2.6 Numerical example To illustrate the approximation numerically, we revisit one component model, shown in Figure 2 and defined by x i (s) = a i f(s) a i i.i.d. Normal(µ a, σ 2 a) y i (s) = a i f(s β i ) β i i.i.d. Normal(0, σ 2 β). It follows that µ(s) = µ a f(s), µ(η i (s)) = µ a f(s β i ). For this model, we have only one significant component with λ = σ 2 a f 2, φ(s) = f(s) f. Although it is possible to derive the exact covariance function, we find it does not help much gaining insight into the problem. Instead we use the approximation to covariance function as where γ η (s, t) = σβ 2 γ y (s, t) = γ x (s, t) + σβv(s, 2 t) + o(σβ) 2, v(s, t) = (µ 2 a + σ 2 a)f (s)f (t) + σ2 a 2 {f (s)f(t) + f(s)f (t)} The first order approximation results gives λ ε λ + σβ 2 < f, V f >. f 2

9 For a concrete example, we consider a special case with f(s) = 1 2π exp ( ) s2 2. With some elementary manipulation, it can be shown that v(s, t) = σ2 a 2 f(s)f(t){(s + t)2 2} λ ε = σ2 a 2 π ( σ2 βσa ) + o(σ 2 β ) 2 = σa σβσ 2 a 2 + o(σβ) 2. Thus, without registration, we would underestimate the eigenvalue. Depending on the magnitude of σ β and σ α, the error term may be significant. Figure 3 about here. For numerical comparison, we simulate 50 curves truncated on [-6,6] with σ a = 2, σ β = 0.5 and estimate the eigenvalues and eigenfunctions of the empirical covariance function estimator. A landmar registration is used to align the curves based on the pea. Results are based on 100 simulations. Figure 3 shows box plots for 4 leading eigenvalues estimated based on x i (0), y i (1), ˆx i (2) respectively and the first order correction (4) suggested in Section 3.1. As expected, the eigenvalues for unregistered curves show spurious variation. Effect of registration is reflected in (2). The bias corrected ones are more variable but can improve the estimation. However there is no guarantee that the ordering will be preserved. We also find that improvements in corrected eigenfunction estimations are rather disappointing. But we have not tried to improve estimation of eigenfunctions, as we consider this beyond the scope of this paper. In addition, for our theoretical derivation and bias approximation, we could ignore all other factors such as smoothing, discrete approximation, different registration methods and possibly higher order approximation errors. In contrast, numerical results will be affected by all these additional factors beyond control. This is particularly relevant to estimating eigenfunctions. 3 Discussion and statistical considerations We focus on the first order approximation. As Γ and Γ ε are exchangeable, we may rewrite the approximation results in terms of λ and φ as λ = λ (ε) ε < φ (ε), V φ (ε) > +O(ε 2 ) = < φ (ε), Γφ (ε) > +O(ε 2 ) φ = φ (ε) < φ (ε), Γφ l (ε) > φ l (ε) + O(ε 2 ). λ (ε) λ l (ε) l

10 Note that even if λ may have ties, it is unliely that the estimates of λ (ε) have ties. To some extent, this formulation suggests a procedure to recover the true values of λ and φ but this requires the full nowledge of Γ which is unnown. Nevertheless, these have two usage. Firstly combined with our imperfect registration, we can devise a simple bias correction procedure. 3.1 Bias correction based on registration 1. Functional PCA with the original data {y i } produces (ˆΓ ε, ˆλ (ε), ˆφ (ε)); 2. Applying registration method gives (ˆx i, ˆη i ) and thus (ˆΓ, ˆλ, ˆφ ); 3. Bias corrected estimators; ˆλ,2 = < ˆφ (ε), ˆΓ ˆφ (ε) > ˆφ,2 = ˆφ (ε) l < ˆφ (ε), ˆΓ ˆφ l (ε) > ˆλ (ε) ˆλ l (ε). Apparently, when the registration is immaculate, there will be no need of bias correction step and it is best to use (ˆλ, ˆφ ) for further inference. To chec if bias correction is necessary, we can apply a simple diagnostic chec as explined below. 3.2 Diagnostics for registration There is no standard diagnostic for checing the registration step, other than heuristic inspection. Most registration methods however produce not only the registered curves but also the warping functions η. The warping functions are relatively simple functions but we have shown that these are directly related to the performance of the estimators. So the first step would be to examine these functions. Recall that ε = sup γ η. If registration is properly done, ε should be small. Moreover, if curves are already aligned, an additional registration should not alter the curves too much. Thus, we may employ the following diagnostic procedure: 1. Register {y i } and estimate (ˆx i, ˆη i ); 2. Calculate ˆγ η. sup ˆγ η is the order of magnitude in error; 3. Register {ˆx i } and estimate (ˆx (2) i, ˆφ (2) i );

11 4. Estimate ˆγ η (2). If it is not negligible, then correct bias. However, It may not be practical to correct all eigenfunctions, as these are much more subtle and the recovery relies on the full eigenspace. Nevertheless, one way of avoiding the pitfall with selection of eigenfunctions, as demonstrated in Figure 2, is at least to correct eigenvalues and determine the importance of eigenfunctions in relation to the corresponding eigenvalues. In our discussion, we have not specified any particular registration method and these are suggested as a general principle. Comparison of γ η may be useful when different registration methods are employed. Alternatively, the moment based registration by James (2008) incorporates this idea by directly penalizing the deviation of η in estimation. More generally, the method developed by Kneip and Ramsay (2007) is based on a template matching and can be viewed as an elaborate implementation of our proposal, by iteratively updating registration based on functional PCA. Appendix General rth order approximation: Aφ (0) Assume that A is real symmetric and satisfies = λ (0) φ(0) = 0, 1,,. For identifiability we ssume that (φ, φ ) = 1 and (φ j, φ ) = 0 for j. Consider where (A + εb)φ (ε) = λ (ε)φ (ε) λ (ε) = λ (0) + ελ (1) + ε 2 λ (2) + φ (ε) = φ (0) + εφ (1) + ε 2 φ (2) +. Here λ = λ (0) and φ = φ (0). The equation may be expressed as (A + εb)(φ + εφ (1) + ε 2 φ (2) + ) = (λ + ελ (1) + ε 2 λ (2) + )(φ + εφ (1) + ε 2 φ (2) + ). Assuming the existence of infinite sum series, we have for r 1 Aφ (r) λ φ (r) = λ (r) φ + λ (r 1) φ (1) Denote the right hand side term by f so that f = λ (r) φ + λ (r 1) φ (1) + + λ (1) + + λ (1) φ(r 1) φ(r 1) Bφ (r 1). Bφ (r 1). (3)

12 and write Aφ (r) Observe that the lefthand side term Aφ (r) < φ l, Aφ (r) In other words, we have λ φ (r) λ φ (r) = f satisifies the following: λ φ (r) > = < (A λ )φ l, φ (r) > { 0 if l = = (λ l λ ) < φ l, φ (r) > if l < φ, f >= 0 < φ l, f >= (λ l λ ) < φ l, φ (r) > l. (4) Equating this with f in place generates λ (r) for r = 1 leads to results stated in Section 2.5. approximation, let us define α (r),l =< φ(r), φ l > so that φ (r) first order approximation results can be summarised as λ (1) = < Bφ, φ > α (1), = 0, α (1),l = < Bφ, φ l > λ λ l l. and φ (r) sequentially. In particular solving Before proceeding this for higher order = α (r), φ + l α(r),l φ l. Then the Consider r 2. For eigenvalue approximation, from (4), the rth order term λ (r) obtained from λ (r) = l α (r 1),l < Bφ, φ l > λ (r 1) α (1), λ(1) α(r 1),. can be For eigenfunction approximation, we need to derive coefficients α (r),l. It can be seen that the coefficient α (r), is related to the normalising condition as 2α (2), = j {α (1),j }2 2α (r), = j α (1),j α(r 1),j j α (2),j α(r 2),j j α (r 1),j α (1),j r 2. The remaining coefficients α (r),l are obtained from (λ λ l )α (r),l = j α (r 1),l < Bφ j, φ l > λ (r 1) α (1),l λ(1) α(r 1),l. To complete calculation, it is necessary to iterate between λ (r) λ (r 1) α (r 1),l ( or φ (r 1) ) λ (r) α (r),l ( or φ(r) ). and φ (r) sequentially, that is,

13 References [1] Dauxois, J. Pousse, A. and Romain, Y. (1982) Asymptotic theory for the principal component analysis of a vector random function: some applications to statistical inference. Journal of Multivariate Analysis, 12, [2] Liu, X. and Müller, H. G. (2004) Functional convex averaging and synchronization for -warped random curves. Journal of the American Statistical Association, 99, [3] Gasser, T. and Kneip, A. (1995) Searching for structure in curve samples. Journal of the American Statistical Association, 90, [4] Gasser, T. Müller, H. G., Köhler, W., Molinari, L. and Prader, A. (1984), Nonparametric regression analysis of growth curves, Annals of Statistics, 12, [5] Hall, P, Müller, H. G. and Wang, J. L. (2006). Properties of principal componenet methods for functional and longitudinal data analysis. Annals of Statistics, 34, [6] Kenip, A. (1994) Nonparametric estimation of common regressors for similar curve data. Annals of Statistics, 22, [7] Kneip, A. and Ramsay, J. O. (2007) Combining registration and fitting for functional models. technical report. [8] Par, J. Gasser, T. and Rousson, V. (2007) Structural components in functional data. technical report. [9] Ramsay, J. O. and Silverman, B. W. (2002) Applied functional data analysis, New Yor: Springer. [10] Ramsay, J. O. and Silverman, B. W. (2005) Functional data analysis, New Yor: Springer. [11] Rice, J. W. and Silverman, B. W. (1991) Estimating the mean and the covariance structure nonparametrically when the data are curves. Journal of Royal Statistical Society, B, 53, [12] Silverman, B. W. (1996) Smoothed functional principal components analysis by choice of norm. Annals of Statistics, 24, 1-24.

14 height velocity age (years) age (years) Figure 1: Example of height measurements (left) with estimated velocity curves (right) from Zürich growth studies. ]

15 pc 1 pc 2 pc 1 pc pc pc 4 pc 3 pc Figure 2: Functional PCA with (right) and without (left) variability. Additional variation produces many more components than anticipated.

16 la(1) la(2) la(3) la(4) Figure 3: Comparison of eigenvalues.

FUNCTIONAL DATA ANALYSIS. Contribution to the. International Handbook (Encyclopedia) of Statistical Sciences. July 28, Hans-Georg Müller 1

FUNCTIONAL DATA ANALYSIS. Contribution to the. International Handbook (Encyclopedia) of Statistical Sciences. July 28, Hans-Georg Müller 1 FUNCTIONAL DATA ANALYSIS Contribution to the International Handbook (Encyclopedia) of Statistical Sciences July 28, 2009 Hans-Georg Müller 1 Department of Statistics University of California, Davis One

More information

FUNCTIONAL DATA ANALYSIS

FUNCTIONAL DATA ANALYSIS FUNCTIONAL DATA ANALYSIS Hans-Georg Müller Department of Statistics University of California, Davis One Shields Ave., Davis, CA 95616, USA. e-mail: mueller@wald.ucdavis.edu KEY WORDS: Autocovariance Operator,

More information

Dynamic Relations for Sparsely Sampled Gaussian Processes

Dynamic Relations for Sparsely Sampled Gaussian Processes TEST manuscript No. (will be inserted by the editor) Dynamic Relations for Sparsely Sampled Gaussian Processes Hans-Georg Müller Wenjing Yang Received: date / Accepted: date Abstract In longitudinal studies,

More information

Smooth Common Principal Component Analysis

Smooth Common Principal Component Analysis 1 Smooth Common Principal Component Analysis Michal Benko Wolfgang Härdle Center for Applied Statistics and Economics benko@wiwi.hu-berlin.de Humboldt-Universität zu Berlin Motivation 1-1 Volatility Surface

More information

Modeling Repeated Functional Observations

Modeling Repeated Functional Observations Modeling Repeated Functional Observations Kehui Chen Department of Statistics, University of Pittsburgh, Hans-Georg Müller Department of Statistics, University of California, Davis Supplemental Material

More information

Functional modeling of longitudinal data

Functional modeling of longitudinal data CHAPTER 1 Functional modeling of longitudinal data 1.1 Introduction Hans-Georg Müller Longitudinal studies are characterized by data records containing repeated measurements per subject, measured at various

More information

Second-Order Inference for Gaussian Random Curves

Second-Order Inference for Gaussian Random Curves Second-Order Inference for Gaussian Random Curves With Application to DNA Minicircles Victor Panaretos David Kraus John Maddocks Ecole Polytechnique Fédérale de Lausanne Panaretos, Kraus, Maddocks (EPFL)

More information

AN INTRODUCTION TO THEORETICAL PROPERTIES OF FUNCTIONAL PRINCIPAL COMPONENT ANALYSIS. Ngoc Mai Tran Supervisor: Professor Peter G.

AN INTRODUCTION TO THEORETICAL PROPERTIES OF FUNCTIONAL PRINCIPAL COMPONENT ANALYSIS. Ngoc Mai Tran Supervisor: Professor Peter G. AN INTRODUCTION TO THEORETICAL PROPERTIES OF FUNCTIONAL PRINCIPAL COMPONENT ANALYSIS Ngoc Mai Tran Supervisor: Professor Peter G. Hall Department of Mathematics and Statistics, The University of Melbourne.

More information

Empirical Dynamics for Longitudinal Data

Empirical Dynamics for Longitudinal Data Empirical Dynamics for Longitudinal Data December 2009 Short title: Empirical Dynamics Hans-Georg Müller 1 Department of Statistics University of California, Davis Davis, CA 95616 U.S.A. Email: mueller@wald.ucdavis.edu

More information

A comparison of stratified simple random sampling and sampling with probability proportional to size

A comparison of stratified simple random sampling and sampling with probability proportional to size A comparison of stratified simple random sampling and sampling with probability proportional to size Edgar Bueno Dan Hedlin Per Gösta Andersson 1 Introduction When planning the sampling strategy (i.e.

More information

Shrinkage Estimation for Functional Principal Component Scores, with Application to the Population Kinetics of Plasma Folate

Shrinkage Estimation for Functional Principal Component Scores, with Application to the Population Kinetics of Plasma Folate Shrinkage Estimation for Functional Principal Component Scores, with Application to the Population Kinetics of Plasma Folate Fang Yao, Hans-Georg Müller,, Andrew J. Clifford, Steven R. Dueker, Jennifer

More information

CONCEPT OF DENSITY FOR FUNCTIONAL DATA

CONCEPT OF DENSITY FOR FUNCTIONAL DATA CONCEPT OF DENSITY FOR FUNCTIONAL DATA AURORE DELAIGLE U MELBOURNE & U BRISTOL PETER HALL U MELBOURNE & UC DAVIS 1 CONCEPT OF DENSITY IN FUNCTIONAL DATA ANALYSIS The notion of probability density for a

More information

Supplement to A Generalized Least Squares Matrix Decomposition. 1 GPMF & Smoothness: Ω-norm Penalty & Functional Data

Supplement to A Generalized Least Squares Matrix Decomposition. 1 GPMF & Smoothness: Ω-norm Penalty & Functional Data Supplement to A Generalized Least Squares Matrix Decomposition Genevera I. Allen 1, Logan Grosenic 2, & Jonathan Taylor 3 1 Department of Statistics and Electrical and Computer Engineering, Rice University

More information

Tests for separability in nonparametric covariance operators of random surfaces

Tests for separability in nonparametric covariance operators of random surfaces Tests for separability in nonparametric covariance operators of random surfaces Shahin Tavakoli (joint with John Aston and Davide Pigoli) April 19, 2016 Analysis of Multidimensional Functional Data Shahin

More information

Functional principal component analysis of aircraft trajectories

Functional principal component analysis of aircraft trajectories Functional principal component analysis of aircraft trajectories Florence Nicol To cite this version: Florence Nicol. Functional principal component analysis of aircraft trajectories. ISIATM 0, nd International

More information

Weakly dependent functional data. Piotr Kokoszka. Utah State University. Siegfried Hörmann. University of Utah

Weakly dependent functional data. Piotr Kokoszka. Utah State University. Siegfried Hörmann. University of Utah Weakly dependent functional data Piotr Kokoszka Utah State University Joint work with Siegfried Hörmann University of Utah Outline Examples of functional time series L 4 m approximability Convergence of

More information

Functional Latent Feature Models. With Single-Index Interaction

Functional Latent Feature Models. With Single-Index Interaction Generalized With Single-Index Interaction Department of Statistics Center for Statistical Bioinformatics Institute for Applied Mathematics and Computational Science Texas A&M University Naisyin Wang and

More information

A Stickiness Coefficient for Longitudinal Data

A Stickiness Coefficient for Longitudinal Data A Stickiness Coefficient for Longitudinal Data June 2011 Andrea Gottlieb 1 Graduate Group in Biostatistics University of California, Davis 1 Shields Avenue Davis, CA 95616 U.S.A. Phone: 1 (530) 752-2361

More information

ON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS

ON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS ON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS Olivier Scaillet a * This draft: July 2016. Abstract This note shows that adding monotonicity or convexity

More information

Derivative Principal Component Analysis for Representing the Time Dynamics of Longitudinal and Functional Data 1

Derivative Principal Component Analysis for Representing the Time Dynamics of Longitudinal and Functional Data 1 Derivative Principal Component Analysis for Representing the Time Dynamics of Longitudinal and Functional Data 1 Short title: Derivative Principal Component Analysis Xiongtao Dai, Hans-Georg Müller Department

More information

A Stickiness Coefficient for Longitudinal Data

A Stickiness Coefficient for Longitudinal Data A Stickiness Coefficient for Longitudinal Data Revised Version November 2011 Andrea Gottlieb 1 Department of Statistics University of California, Davis 1 Shields Avenue Davis, CA 95616 U.S.A. Phone: 1

More information

Modeling Multi-Way Functional Data With Weak Separability

Modeling Multi-Way Functional Data With Weak Separability Modeling Multi-Way Functional Data With Weak Separability Kehui Chen Department of Statistics University of Pittsburgh, USA @CMStatistics, Seville, Spain December 09, 2016 Outline Introduction. Multi-way

More information

ftsa: An R Package for Analyzing Functional Time Series by Han Lin Shang

ftsa: An R Package for Analyzing Functional Time Series by Han Lin Shang CONTRIBUTED RESEARCH ARTICLES 64 ftsa: An R Package for Analyzing Functional Time Series by Han Lin Shang Abstract Recent advances in computer recording and storing technology have tremendously increased

More information

Option 1: Landmark Registration We can try to align specific points. The Registration Problem. Landmark registration. Time-Warping Functions

Option 1: Landmark Registration We can try to align specific points. The Registration Problem. Landmark registration. Time-Warping Functions The Registration Problem Most analyzes only account for variation in amplitude. Frequently, observed data exhibit features that vary in time. Option 1: Landmark Registration We can try to align specific

More information

Regularization via Spectral Filtering

Regularization via Spectral Filtering Regularization via Spectral Filtering Lorenzo Rosasco MIT, 9.520 Class 7 About this class Goal To discuss how a class of regularization methods originally designed for solving ill-posed inverse problems,

More information

The Mahalanobis distance for functional data with applications to classification

The Mahalanobis distance for functional data with applications to classification arxiv:1304.4786v1 [math.st] 17 Apr 2013 The Mahalanobis distance for functional data with applications to classification Esdras Joseph, Pedro Galeano and Rosa E. Lillo Departamento de Estadística Universidad

More information

Mixture of Gaussian Processes and its Applications

Mixture of Gaussian Processes and its Applications Mixture of Gaussian Processes and its Applications Mian Huang, Runze Li, Hansheng Wang, and Weixin Yao The Pennsylvania State University Technical Report Series #10-102 College of Health and Human Development

More information

Course topics (tentative) The role of random effects

Course topics (tentative) The role of random effects Course topics (tentative) random effects linear mixed models analysis of variance frequentist likelihood-based inference (MLE and REML) prediction Bayesian inference The role of random effects Rasmus Waagepetersen

More information

Fractal functional regression for classification of gene expression data by wavelets

Fractal functional regression for classification of gene expression data by wavelets Fractal functional regression for classification of gene expression data by wavelets Margarita María Rincón 1 and María Dolores Ruiz-Medina 2 1 University of Granada Campus Fuente Nueva 18071 Granada,

More information

Spectral Regularization

Spectral Regularization Spectral Regularization Lorenzo Rosasco 9.520 Class 07 February 27, 2008 About this class Goal To discuss how a class of regularization methods originally designed for solving ill-posed inverse problems,

More information

The loss function and estimating equations

The loss function and estimating equations Chapter 6 he loss function and estimating equations 6 Loss functions Up until now our main focus has been on parameter estimating via the maximum likelihood However, the negative maximum likelihood is

More information

FUNCTIONAL DATA ANALYSIS FOR VOLATILITY PROCESS

FUNCTIONAL DATA ANALYSIS FOR VOLATILITY PROCESS FUNCTIONAL DATA ANALYSIS FOR VOLATILITY PROCESS Rituparna Sen Monday, July 31 10:45am-12:30pm Classroom 228 St-C5 Financial Models Joint work with Hans-Georg Müller and Ulrich Stadtmüller 1. INTRODUCTION

More information

Nonresponse weighting adjustment using estimated response probability

Nonresponse weighting adjustment using estimated response probability Nonresponse weighting adjustment using estimated response probability Jae-kwang Kim Yonsei University, Seoul, Korea December 26, 2006 Introduction Nonresponse Unit nonresponse Item nonresponse Basic strategy

More information

Intelligent Data Analysis. Principal Component Analysis. School of Computer Science University of Birmingham

Intelligent Data Analysis. Principal Component Analysis. School of Computer Science University of Birmingham Intelligent Data Analysis Principal Component Analysis Peter Tiňo School of Computer Science University of Birmingham Discovering low-dimensional spatial layout in higher dimensional spaces - 1-D/3-D example

More information

MIXTURE INNER PRODUCT SPACES AND THEIR APPLICATION TO FUNCTIONAL DATA ANALYSIS

MIXTURE INNER PRODUCT SPACES AND THEIR APPLICATION TO FUNCTIONAL DATA ANALYSIS MIXTURE INNER PRODUCT SPACES AND THEIR APPLICATION TO FUNCTIONAL DATA ANALYSIS Zhenhua Lin 1, Hans-Georg Müller 2 and Fang Yao 1,3 Abstract We introduce the concept of mixture inner product spaces associated

More information

Degradation Modeling and Monitoring of Truncated Degradation Signals. Rensheng Zhou, Nagi Gebraeel, and Nicoleta Serban

Degradation Modeling and Monitoring of Truncated Degradation Signals. Rensheng Zhou, Nagi Gebraeel, and Nicoleta Serban Degradation Modeling and Monitoring of Truncated Degradation Signals Rensheng Zhou, Nagi Gebraeel, and Nicoleta Serban School of Industrial and Systems Engineering, Georgia Institute of Technology Abstract:

More information

Exchangeable Bernoulli Random Variables And Bayes Postulate

Exchangeable Bernoulli Random Variables And Bayes Postulate Exchangeable Bernoulli Random Variables And Bayes Postulate Moulinath Banerjee and Thomas Richardson September 30, 20 Abstract We discuss the implications of Bayes postulate in the setting of exchangeable

More information

BIAS-ROBUSTNESS AND EFFICIENCY OF MODEL-BASED INFERENCE IN SURVEY SAMPLING

BIAS-ROBUSTNESS AND EFFICIENCY OF MODEL-BASED INFERENCE IN SURVEY SAMPLING Statistica Sinica 22 (2012), 777-794 doi:http://dx.doi.org/10.5705/ss.2010.238 BIAS-ROBUSTNESS AND EFFICIENCY OF MODEL-BASED INFERENCE IN SURVEY SAMPLING Desislava Nedyalova and Yves Tillé University of

More information

Diagnostics for functional regression via residual processes

Diagnostics for functional regression via residual processes Diagnostics for functional regression via residual processes Jeng-Min Chiou a, Hans-Georg Müller b, a Academia Sinica, 128 Academia Road Sec.2, Taipei 11529, Taiwan b University of California, Davis, One

More information

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data

An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data Jae-Kwang Kim 1 Iowa State University June 28, 2012 1 Joint work with Dr. Ming Zhou (when he was a PhD student at ISU)

More information

Analysis of AneuRisk65 data: warped logistic discrimination

Analysis of AneuRisk65 data: warped logistic discrimination Electronic Journal of Statistics Vol. () ISSN: 935-7524 DOI:./ Analysis of AneuRisk65 data: warped logistic discrimination Daniel Gervini Department of Mathematical Sciences University of Wisconsin Milwaukee

More information

TIME-WARPED GROWTH PROCESSES, WITH APPLICATIONS TO THE MODELING OF BOOM-BUST CYCLES IN HOUSE PRICES

TIME-WARPED GROWTH PROCESSES, WITH APPLICATIONS TO THE MODELING OF BOOM-BUST CYCLES IN HOUSE PRICES Submitted to the Annals of Applied Statistics arxiv: arxiv:0000.0000 TIME-WARPED GROWTH PROCESSES, WITH APPLICATIONS TO THE MODELING OF BOOM-BUST CYCLES IN HOUSE PRICES By Jie Peng,, Debashis Paul, and

More information

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data.

Structure in Data. A major objective in data analysis is to identify interesting features or structure in the data. Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two

More information

sparse and low-rank tensor recovery Cubic-Sketching

sparse and low-rank tensor recovery Cubic-Sketching Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru

More information

ECON 3150/4150, Spring term Lecture 6

ECON 3150/4150, Spring term Lecture 6 ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture

More information

arxiv: v1 [stat.me] 4 Oct 2016

arxiv: v1 [stat.me] 4 Oct 2016 Hybrid Regularisation of Functional Linear Models Anirvan Chakraborty and Victor M. Panaretos Institut de Mathématiques, École Polytechnique Fédérale de Lausanne e-mail: anirvan.chakraborty@epfl.ch; victor.panaretos@epfl.ch

More information

Nonparametric Inference In Functional Data

Nonparametric Inference In Functional Data Nonparametric Inference In Functional Data Zuofeng Shang Purdue University Joint work with Guang Cheng from Purdue Univ. An Example Consider the functional linear model: Y = α + where 1 0 X(t)β(t)dt +

More information

Diagnostics for functional regression via residual processes

Diagnostics for functional regression via residual processes Diagnostics for functional regression via residual processes Jeng-Min Chiou Academia Sinica, Taiwan E-mail: jmchiou@stat.sinica.edu.tw Hans-Georg Müller University of California, Davis, USA E-mail: mueller@wald.ucdavis.edu

More information

Networked Control Systems, Event-Triggering, Small-Gain Theorem, Nonlinear

Networked Control Systems, Event-Triggering, Small-Gain Theorem, Nonlinear EVENT-TRIGGERING OF LARGE-SCALE SYSTEMS WITHOUT ZENO BEHAVIOR C. DE PERSIS, R. SAILER, AND F. WIRTH Abstract. We present a Lyapunov based approach to event-triggering for large-scale systems using a small

More information

7 Day 3: Time Varying Parameter Models

7 Day 3: Time Varying Parameter Models 7 Day 3: Time Varying Parameter Models References: 1. Durbin, J. and S.-J. Koopman (2001). Time Series Analysis by State Space Methods. Oxford University Press, Oxford 2. Koopman, S.-J., N. Shephard, and

More information

Optimal Auxiliary Variable Assisted Two-Phase Sampling Designs

Optimal Auxiliary Variable Assisted Two-Phase Sampling Designs MASTER S THESIS Optimal Auxiliary Variable Assisted Two-Phase Sampling Designs HENRIK IMBERG Department of Mathematical Sciences Division of Mathematical Statistics CHALMERS UNIVERSITY OF TECHNOLOGY UNIVERSITY

More information

Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions

Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions March 6, 2013 Contents 1 Wea second variation 2 1.1 Formulas for variation........................

More information

Convergence of Eigenspaces in Kernel Principal Component Analysis

Convergence of Eigenspaces in Kernel Principal Component Analysis Convergence of Eigenspaces in Kernel Principal Component Analysis Shixin Wang Advanced machine learning April 19, 2016 Shixin Wang Convergence of Eigenspaces April 19, 2016 1 / 18 Outline 1 Motivation

More information

Bindel, Fall 2011 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Jan 9

Bindel, Fall 2011 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Jan 9 Problem du jour Week 3: Wednesday, Jan 9 1. As a function of matrix dimension, what is the asymptotic complexity of computing a determinant using the Laplace expansion (cofactor expansion) that you probably

More information

Diagnostics for Linear Models With Functional Responses

Diagnostics for Linear Models With Functional Responses Diagnostics for Linear Models With Functional Responses Qing Shen Edmunds.com Inc. 2401 Colorado Ave., Suite 250 Santa Monica, CA 90404 (shenqing26@hotmail.com) Hongquan Xu Department of Statistics University

More information

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation

Statistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation Statistics 62: L p spaces, metrics on spaces of probabilites, and connections to estimation Moulinath Banerjee December 6, 2006 L p spaces and Hilbert spaces We first formally define L p spaces. Consider

More information

Flexible Estimation of Treatment Effect Parameters

Flexible Estimation of Treatment Effect Parameters Flexible Estimation of Treatment Effect Parameters Thomas MaCurdy a and Xiaohong Chen b and Han Hong c Introduction Many empirical studies of program evaluations are complicated by the presence of both

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 6: Bias and variance (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 49 Our plan today We saw in last lecture that model scoring methods seem to be trading off two different

More information

Principal Component Analysis

Principal Component Analysis Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used

More information

FERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA. Ning He, Lei Xie, Shu-qing Wang, Jian-ming Zhang

FERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA. Ning He, Lei Xie, Shu-qing Wang, Jian-ming Zhang FERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA Ning He Lei Xie Shu-qing Wang ian-ming Zhang National ey Laboratory of Industrial Control Technology Zhejiang University Hangzhou 3007

More information

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions

Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract

More information

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF

More information

A Note on Hilbertian Elliptically Contoured Distributions

A Note on Hilbertian Elliptically Contoured Distributions A Note on Hilbertian Elliptically Contoured Distributions Yehua Li Department of Statistics, University of Georgia, Athens, GA 30602, USA Abstract. In this paper, we discuss elliptically contoured distribution

More information

Regularized principal components analysis

Regularized principal components analysis 9 Regularized principal components analysis 9.1 Introduction In this chapter, we discuss the application of smoothing to functional principal components analysis. In Chapter 5 we have already seen that

More information

An asymptotic ratio characterization of input-to-state stability

An asymptotic ratio characterization of input-to-state stability 1 An asymptotic ratio characterization of input-to-state stability Daniel Liberzon and Hyungbo Shim Abstract For continuous-time nonlinear systems with inputs, we introduce the notion of an asymptotic

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V

More information

Functional quasi-likelihood regression models with smooth random effects

Functional quasi-likelihood regression models with smooth random effects J. R. Statist. Soc. B (2003) 65, Part 2, pp. 405 423 Functional quasi-likelihood regression models with smooth random effects Jeng-Min Chiou National Health Research Institutes, Taipei, Taiwan and Hans-Georg

More information

Independent component analysis for functional data

Independent component analysis for functional data Independent component analysis for functional data Hannu Oja Department of Mathematics and Statistics University of Turku Version 12.8.216 August 216 Oja (UTU) FICA Date bottom 1 / 38 Outline 1 Probability

More information

A Novel Nonparametric Density Estimator

A Novel Nonparametric Density Estimator A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with

More information

Function-on-Scalar Regression with the refund Package

Function-on-Scalar Regression with the refund Package University of Haifa From the SelectedWorks of Philip T. Reiss July 30, 2012 Function-on-Scalar Regression with the refund Package Philip T. Reiss, New York University Available at: https://works.bepress.com/phil_reiss/28/

More information

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations

Ch. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations Ch 15 Forecasting Having considered in Chapter 14 some of the properties of ARMA models, we now show how they may be used to forecast future values of an observed time series For the present we proceed

More information

Chapter 2: simple regression model

Chapter 2: simple regression model Chapter 2: simple regression model Goal: understand how to estimate and more importantly interpret the simple regression Reading: chapter 2 of the textbook Advice: this chapter is foundation of econometrics.

More information

Matrix Factorizations

Matrix Factorizations 1 Stat 540, Matrix Factorizations Matrix Factorizations LU Factorization Definition... Given a square k k matrix S, the LU factorization (or decomposition) represents S as the product of two triangular

More information

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and

This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and education use, including for instruction at the authors institution

More information

SMOOTHED FUNCTIONAL PRINCIPAL COMPONENTS ANALYSIS BY CHOICE OF NORM 1. BY BERNARD W. SILVERMAN University of Bristol

SMOOTHED FUNCTIONAL PRINCIPAL COMPONENTS ANALYSIS BY CHOICE OF NORM 1. BY BERNARD W. SILVERMAN University of Bristol The Annals of Statistics 1996, Vol. 4, No. 1, 14 SMOOTHED FUNCTIONAL PRINCIPAL COMPONENTS ANALYSIS BY CHOICE OF NORM 1 BY BERNARD W. SILVERMAN University of Bristol The principal components analysis of

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Suppose again we have n sample points x,..., x n R p. The data-point x i R p can be thought of as the i-th row X i of an n p-dimensional

More information

Function Approximation

Function Approximation 1 Function Approximation This is page i Printer: Opaque this 1.1 Introduction In this chapter we discuss approximating functional forms. Both in econometric and in numerical problems, the need for an approximating

More information

Introduction to Functional Data Analysis A CSCU Workshop. Giles Hooker Biological Statistics and Computational Biology

Introduction to Functional Data Analysis A CSCU Workshop. Giles Hooker Biological Statistics and Computational Biology Introduction to Functional Data Analysis A CSCU Workshop Giles Hooker Biological Statistics and Computational Biology gjh27@cornell.edu www.bscb.cornell.edu/ hooker/fdaworkshop 1 / 26 Agenda What is Functional

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models

Economics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe

More information

Online Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω)

Online Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω) Online Appendix Proof of Lemma A.. he proof uses similar arguments as in Dunsmuir 979), but allowing for weak identification and selecting a subset of frequencies using W ω). It consists of two steps.

More information

Curves clustering with approximation of the density of functional random variables

Curves clustering with approximation of the density of functional random variables Curves clustering with approximation of the density of functional random variables Julien Jacques and Cristian Preda Laboratoire Paul Painlevé, UMR CNRS 8524, University Lille I, Lille, France INRIA Lille-Nord

More information

DEGRADATION MODELING AND MONITORING OF ENGINEERING SYSTEMS USING FUNCTIONAL DATA ANALYSIS

DEGRADATION MODELING AND MONITORING OF ENGINEERING SYSTEMS USING FUNCTIONAL DATA ANALYSIS DEGRADATION MODELING AND MONITORING OF ENGINEERING SYSTEMS USING FUNCTIONAL DATA ANALYSIS A Thesis Presented to The Academic Faculty by Rensheng Zhou In Partial Fulfillment of the Requirements for the

More information

Notes on Measure Theory and Markov Processes

Notes on Measure Theory and Markov Processes Notes on Measure Theory and Markov Processes Diego Daruich March 28, 2014 1 Preliminaries 1.1 Motivation The objective of these notes will be to develop tools from measure theory and probability to allow

More information

Estimating Moving Average Processes with an improved version of Durbin s Method

Estimating Moving Average Processes with an improved version of Durbin s Method Estimating Moving Average Processes with an improved version of Durbin s Method Maximilian Ludwig this version: June 7, 4, initial version: April, 3 arxiv:347956v [statme] 6 Jun 4 Abstract This paper provides

More information

Interpretable Functional Principal Component Analysis

Interpretable Functional Principal Component Analysis Biometrics 000, 000 000 DOI: 000 000 0000 Interpretable Functional Principal Component Analysis Zhenhua Lin 1,, Liangliang Wang 2,, and Jiguo Cao 2, 1 Department of Statistical Sciences, University of

More information

arxiv: v1 [stat.me] 18 Jul 2015

arxiv: v1 [stat.me] 18 Jul 2015 Review of Functional Data Analysis Jane-Ling Wang, 1 Jeng-Min Chiou, 2 and Hans-Georg Müller 1 1 Department of Statistics, University of California, Davis, USA, 95616 2 Institute of Statistical Science,

More information

Principal component analysis

Principal component analysis Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and

More information

Fundamental Issues in Bayesian Functional Data Analysis. Dennis D. Cox Rice University

Fundamental Issues in Bayesian Functional Data Analysis. Dennis D. Cox Rice University Fundamental Issues in Bayesian Functional Data Analysis Dennis D. Cox Rice University 1 Introduction Question: What are functional data? Answer: Data that are functions of a continuous variable.... say

More information

Optimal spectral shrinkage and PCA with heteroscedastic noise

Optimal spectral shrinkage and PCA with heteroscedastic noise Optimal spectral shrinage and PCA with heteroscedastic noise William Leeb and Elad Romanov Abstract This paper studies the related problems of denoising, covariance estimation, and principal component

More information

Convex Functions and Optimization

Convex Functions and Optimization Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized

More information

Latent Variable Models and EM algorithm

Latent Variable Models and EM algorithm Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic

More information

Nonparametric Modal Regression

Nonparametric Modal Regression Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

arxiv: v1 [stat.me] 3 Sep 2016

arxiv: v1 [stat.me] 3 Sep 2016 Functional Data Analysis by Matrix Completion Marie-Hélène Descary and Victor M. Panaretos Institut de Mathématiques Ecole Polytechnique Fédérale de Lausanne e-mail: marie-helene.descary@epfl.ch, victor.panaretos@epfl.ch

More information

Time Series and Forecasting Lecture 4 NonLinear Time Series

Time Series and Forecasting Lecture 4 NonLinear Time Series Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations

More information

Asymptotic inference for a nonstationary double ar(1) model

Asymptotic inference for a nonstationary double ar(1) model Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk

More information

Focusing on structural assumptions in regression on functional variable.

Focusing on structural assumptions in regression on functional variable. Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session IPS043) p.798 Focusing on structural assumptions in regression on functional variable. DELSOL, Laurent Université d

More information

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis

Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis Sílvia Gonçalves and Benoit Perron Département de sciences économiques,

More information