Curve alignment and functional PCA
|
|
- Elwin Murphy
- 6 years ago
- Views:
Transcription
1 Curve alignment and functional PCA Juhyun Par* Department of Mathematics and Statistics, Lancaster University, Lancaster, U.K. Abstract When dealing with multiple curves as functional data, it is a common practice to apply functional PCA to summarize and characterize random variation in finite dimension. Often functional data however exhibits additional variability that distorts the assumed common structure. This is recognized as the problem of curve registration. While the registration step is routinely employed, this is considered as a preprocessing step prior to any serious analysis. Consequently, the effect of alignment is mostly ignored in subsequent analyses and is not well understood. We revisit the issue by particularly focusing on the effect of variability on the FPCA and illustrate the phenomena from a borrowed perturbation viewpoint. This allows us to quantify bias in estimating eigenvalues and eigenfunctions when curves are not aligned. Some discussion on statistical implication is given. Keywords functional data analysis; curve alignment; registration; functional principal component analysis. 1 Introduction Repeated measurements in the form of curves are increasingly common in various scientific applications including biomedicine, physical sciences and econometrics. Usually the sample of curves is assumed to have some homogeneous structure in functional shape, while allowed for individual variability. Summarising and characterising variability in parsimonious manner is one of the aims of functional data analysis (Ramsay and Silverman, 2002, 2005). It is much
2 desirable that a few common components are able to extract most variability and are easy to interpret (Par et al. 2007). Functional PCA utilises the well-nown Karhunen-Loève expansion to provide an optimal representation of the function with a small number of common components. This is based on the assumption that the underlying random function shares the common mean and covariance function. Functional PCA also provides a useful representation when functional linear model is employed with functional explanatory variables. Figure 1 about here. Often functional data exhibits additional variability. Figure 1 shows a typical example of functional data, taen from the famous Zürich longitudinal growth studies. (Gasser et al., 1984). The left hand side plot shows height measurements of boys taen over years since childhood, with measurements points mared by circles. An observation that these are monotonically increasing does not seem so informative but simply confirms that children grow over! The plot on the right shows estimated velocity curves. Now distinctive features are apparent, which is often the case for functional data that consideration of the 1st or 2nd derivatives allows more meaningful comparison. In particular with children growth processes, it is well nown that each individual reaches its maximum growth at different, while the overall process shares common features between children. The presence of bumps appearing in the velocity curves was considered as an important characteristic of growth processes and not surprisingly the occurrence of such bumps varies between child to child. It is natural to allow variability but the consequence may sound severe in that a simple averaging would not produce a proper mean and most common statistics cannot be pursued in a usual manner. In the longitudinal data framewor, this phenomenon may be described by mixed effect models but there seems no notion of phase variability itself. Alternative approach would be to tae shape invariant models, which is a semi-parametric approach where a parametric form is assumed for phase variability. If no parametric form is assumed, the model is equivalent to those for curve alignments in functional data. This is treated as registration problem in functional data analysis, for which several methods have been developed. Basically when the functions exhibit identifiable features, curves can be aligned to match those features or landmars (Gasser and Kneip, 1995). This wors well as long as features are correctly identified. Several other methods have been developed to automate the procedure when the features are less prominent, see Ramsay and Silverman (2005) and references therein. Although the issue of phase variability has been rightly acnowledged, most analyses treat registration as a preprocessing step and thus do not consider any carry-on effects on later analyses. We point out that in practice, lie any other statistical procedure, the
3 registration step requires careful human intervention and experiences as well as additional estimation and smoothing step to mae it right. What happens then if registration was not carried out or was done improperly? This question is also relevant when, in modern applied sciences with abundant data, often PCA or a type of factor analysis is carried out to find a low-dimensional representation, ignoring phase variability. Most theoretical analyses view the goal of registration as finding a correct mean function and do not go beyond it. It may sound absurd to worry about the second moment property when the first moment is already in trouble. But as will be explained later, this is not necessarily a linear process and there is a systematic relationship between processes with and without phase variability. For example, Figure 2 demonstrates the effect of additional variability on the number of significant components in function PCA. The top row in Figure 2 shows an example of data with (right) and without (left) variability and the bottom row is the result of functional PCA. When additional variability is present, PCA produces many more uncorrelated principal components scores than anticipated (bottom right). Some issues with interpretability in PCA may be attributed to an improper registration. Figure 2 about here. A recent wor of Kneip and Ramsay (2007) addresses a similar issue (and more) and propose a new procedure to combine registration to fit functional PCA models, extending the convex averaging idea of registration (Liu and Müller, 2004). We also revisit the issue but our focus lies on understanding and quantifying the effect of imperfect registration on functional PCA. This is illustrated based on perturbation theory and numerical examples. Some discussion is given for a simple diagnostic chec for registration methods and a correction for functional PCA. 2 fpca with variability 2.1 Notation for fpca To fix the idea, consider a stochastic process X L 2 (T ) with compact support T = [0, T ], with the mean function µ(t) = E[X(t)] and the covariance function γ(s, t) = Cov(X(s), X(t)). Assume that T E[X(t)2 ] <. Let λ 1 λ 2... be the ordered eigenvalues of the covariance operator defined through γ with the corresponding eigenfunctions φ 1, φ 2,.... We assume that λ <. Then X(t) = µ(t) + ξ φ (t), (1) =1
4 where E[ξ] = 0 and E[ξ j ξ ] = λ j I(j = ). Thus in our analysis we will automatically assume that X admits the representation of (1). With a sample of curves available, these quantities are replaced by their estimates and a finite number of components {φ } are usually considered sufficient to extract significant observed variation. Theoretical properties of estimators are studied in Dauxois et al. (1982), Rice and Silverman (1991), Kenip (1994) and Hall et al. (2006). 2.2 Time variability and warping function When variability is present, we observe a sample from Y ( ) = X(η( )), where η( ) is called a warping function to describe transformation on axis. The warping function η( ) is assumed to satisfy the following conditions: (A1) η is a smooth monotone function; (A2) E[η(t)] = t for all t T ; (A3) T Var[η(t)] <. (A4) η is independent of amplitude variability (ξs). These are standard assumptions when dealing with registration (Gasser and Kneip, 1995). 2.3 Covariance function estimation without registration Suppose that {Y 1,, Y n } are independent random functions from Y. An empirical covariance function estimator is ˆγ Y (s, t) = 1 n {Y i (s) n Ȳ (s)}{y (t) Ȳ (t)}, i=1 where Ȳ (t) = 1 n n i=1 Y i(s). To study statistical properties of the estimator, this may be written as ˆγ Y = γ Y + n 1/2 {n 1/2 (ˆγ Y γ Y )}, (2) where γ Y (s, t) = Cov(Y (s), Y (t)). The usual asymptotic properties of functional PCA rely on the convergence of the latter term {n 1/2 (ˆγ Y γ Y )}, see Hall and Hosseini-Nasab (2006). However, closeness to γ Y is not so much of interest when additional variability is present. Instead we view the estimator as an approximation to γ X and thus decompose the estimator as ˆγ Y = γ X + (γ Y γ X ) + (ˆγ Y γ Y ).
5 The term ˆγ Y γ Y follows the usual asymptotics discussed with (2) and is negligible compared to the bias term γ Y γ X. Hence, from now on we will focus on the bias term only. 2.4 Approximation to covariance function We introduce the following notations to describe the observed process Y. µ(s) = E[Y (s)] φ(s) = φ(η(s)) Also let us define γ η (s, t) = E[{η(s) s}{η(t) t}]. Replacing t by η(t) in (1), we obtain a representation of Y. This is not any longer orthonomal decomposition but provides a useful starting point to study ˆγ Y. Combined with independence between η and ξ in (A4) we obtain E[ˆγ Y (s, t)] = E[{ µ(s) µ η (s)}{ µ(s) µ η (t)}] + n 1 λ E[φ (η i (s)) n φ (s) φ (t)]. Under further smoothness assumptions, the expression can be simplified further, which we summarise in the following proposition. Proposition 1. Assume that γ(s, t) = λ φ (s)φ (t) and has continuous second derivatives. Then, we have γ Y (s, t) = γ(s, t) + γ η v(s, t) + o( γ η ) where v(s, t) = γ η(s, t) γ η µ (s)µ (t) + γ η(s, t) γ η + γ η(s, s) γ η λ φ (s)φ (t) λ 2 φ (s)φ (t) + γ η(t, t) γ η λ 2 φ (s)φ (t). Proof of Proposition 1 To simplify the expression, observe that from (A2) any smooth function g of η can be approximated using Taylor expansion as g(η(s)) = g(s) + g (s)(η(s) s) + g (s) 2 (η(s) s)2 + o p (1).
6 Applying this approximation to µ and φ leads to γ Y (s, t) = γ(s, t) + γ η (s, t)µ (s)µ (t) +γ η (s, t) + γ η(t, t) 2 λ φ (s)φ (t) + γ η(s, s) 2 λ φ (s)φ (t) + o( γ η ) λ φ (s)φ (t) where γ η = sup γ η (s, t). Note that ˆγ Y is not an unbiased estimator and γ Y = E[ n 1 ˆγ n Y ]. Though not neglible itself, phase variability is still assumed to be small relative to amplitude variability.. This opens up the possibility of viewing the bias as additive perturbation. If we assume that the process is well approximated by a finite number of basis functions, it is equivalent to wor with a finite dimensional operator. Then the problem is reduced to finding eigenvalues and eigenfunctions for a perturbed covariance operator. 2.5 Perturbation theory of eigenvalue problem Consider the covariance function of the form γ ε (s, t) = γ(s, t) + εv(s, t) for some function v. Define the corresponding operator by Γ ε φ = γ ε φ for φ L 2. We write the corresponding covariance operator by Γ ε = Γ + εv. The perturbation theory formulates the problem of finding eigenvalues and eigenfunctions as finding approximation terms in order of ε. where Γφ = λ φ = 0, 1, 2, Γ ε φ (ε) = λ (ε)φ (ε) λ (ε) = λ + ελ (1) + ε 2 λ (2) + φ (ε) = φ + εφ (1) + ε 2 φ (2) +
7 Basically, since the {φ l }s are orthonomal bases, any function can be written in terms of φ l. If φ (r) = l c lφ l and the constants can be determined inductively. If we require that φ (ε) be normalised, then 1 =< φ (ε), φ (ε) >. For illustration, below are given first and second order approximation, which quantify the bias. The expansion is not limited to any order but we only list first two order approximation results for simplification. As a reference, general rth order approximation is provided in Appendix. Here the eigenfunction expansions assume that eigenvalues are distinct. First order approximation: λ (1) = < V φ, φ > φ (1) = l < V φ, φ l > φ l λ λ l Therefore we may write λ (ε) = λ + ε < φ, V φ > +O(ε 2 ) φ (ε) = φ + ε < φ, V φ l > φ l + O(ε 2 ). λ λ l l Second order approximation: λ (2) = < V φ, φ (1) >= l φ (2) = c φ + m c m φ m < V φ, φ l > 2 λ λ l c = 1 2 < φ(1), φ(1) >= 1 < V φ, φ l > 2 2 (λ λ l ) 2 l c m = l < V φ, φ l >< V φ m, φ l > (λ λ l )(λ λ m ) < V φ, φ >< V φ, φ m > (λ λ m ) 2 Note that this framewor is not limited to the case of registration error. Additional smoothing bias, sampling error or discrete approximation error can be formulated in a similar
8 manner. For example Silverman (1996) used a similar argument to show benefit of smoothing in functional PCA. In the case of sampling error, as in (2), the expression is given with ε = n 1/2, which was the basis of asymptotic expansions of eigenvalues and eigenfunctions in Hall and Hosseini-Nasab (2006). However note that unlie their results, bias here does not diminish as sample size grows. 2.6 Numerical example To illustrate the approximation numerically, we revisit one component model, shown in Figure 2 and defined by x i (s) = a i f(s) a i i.i.d. Normal(µ a, σ 2 a) y i (s) = a i f(s β i ) β i i.i.d. Normal(0, σ 2 β). It follows that µ(s) = µ a f(s), µ(η i (s)) = µ a f(s β i ). For this model, we have only one significant component with λ = σ 2 a f 2, φ(s) = f(s) f. Although it is possible to derive the exact covariance function, we find it does not help much gaining insight into the problem. Instead we use the approximation to covariance function as where γ η (s, t) = σβ 2 γ y (s, t) = γ x (s, t) + σβv(s, 2 t) + o(σβ) 2, v(s, t) = (µ 2 a + σ 2 a)f (s)f (t) + σ2 a 2 {f (s)f(t) + f(s)f (t)} The first order approximation results gives λ ε λ + σβ 2 < f, V f >. f 2
9 For a concrete example, we consider a special case with f(s) = 1 2π exp ( ) s2 2. With some elementary manipulation, it can be shown that v(s, t) = σ2 a 2 f(s)f(t){(s + t)2 2} λ ε = σ2 a 2 π ( σ2 βσa ) + o(σ 2 β ) 2 = σa σβσ 2 a 2 + o(σβ) 2. Thus, without registration, we would underestimate the eigenvalue. Depending on the magnitude of σ β and σ α, the error term may be significant. Figure 3 about here. For numerical comparison, we simulate 50 curves truncated on [-6,6] with σ a = 2, σ β = 0.5 and estimate the eigenvalues and eigenfunctions of the empirical covariance function estimator. A landmar registration is used to align the curves based on the pea. Results are based on 100 simulations. Figure 3 shows box plots for 4 leading eigenvalues estimated based on x i (0), y i (1), ˆx i (2) respectively and the first order correction (4) suggested in Section 3.1. As expected, the eigenvalues for unregistered curves show spurious variation. Effect of registration is reflected in (2). The bias corrected ones are more variable but can improve the estimation. However there is no guarantee that the ordering will be preserved. We also find that improvements in corrected eigenfunction estimations are rather disappointing. But we have not tried to improve estimation of eigenfunctions, as we consider this beyond the scope of this paper. In addition, for our theoretical derivation and bias approximation, we could ignore all other factors such as smoothing, discrete approximation, different registration methods and possibly higher order approximation errors. In contrast, numerical results will be affected by all these additional factors beyond control. This is particularly relevant to estimating eigenfunctions. 3 Discussion and statistical considerations We focus on the first order approximation. As Γ and Γ ε are exchangeable, we may rewrite the approximation results in terms of λ and φ as λ = λ (ε) ε < φ (ε), V φ (ε) > +O(ε 2 ) = < φ (ε), Γφ (ε) > +O(ε 2 ) φ = φ (ε) < φ (ε), Γφ l (ε) > φ l (ε) + O(ε 2 ). λ (ε) λ l (ε) l
10 Note that even if λ may have ties, it is unliely that the estimates of λ (ε) have ties. To some extent, this formulation suggests a procedure to recover the true values of λ and φ but this requires the full nowledge of Γ which is unnown. Nevertheless, these have two usage. Firstly combined with our imperfect registration, we can devise a simple bias correction procedure. 3.1 Bias correction based on registration 1. Functional PCA with the original data {y i } produces (ˆΓ ε, ˆλ (ε), ˆφ (ε)); 2. Applying registration method gives (ˆx i, ˆη i ) and thus (ˆΓ, ˆλ, ˆφ ); 3. Bias corrected estimators; ˆλ,2 = < ˆφ (ε), ˆΓ ˆφ (ε) > ˆφ,2 = ˆφ (ε) l < ˆφ (ε), ˆΓ ˆφ l (ε) > ˆλ (ε) ˆλ l (ε). Apparently, when the registration is immaculate, there will be no need of bias correction step and it is best to use (ˆλ, ˆφ ) for further inference. To chec if bias correction is necessary, we can apply a simple diagnostic chec as explined below. 3.2 Diagnostics for registration There is no standard diagnostic for checing the registration step, other than heuristic inspection. Most registration methods however produce not only the registered curves but also the warping functions η. The warping functions are relatively simple functions but we have shown that these are directly related to the performance of the estimators. So the first step would be to examine these functions. Recall that ε = sup γ η. If registration is properly done, ε should be small. Moreover, if curves are already aligned, an additional registration should not alter the curves too much. Thus, we may employ the following diagnostic procedure: 1. Register {y i } and estimate (ˆx i, ˆη i ); 2. Calculate ˆγ η. sup ˆγ η is the order of magnitude in error; 3. Register {ˆx i } and estimate (ˆx (2) i, ˆφ (2) i );
11 4. Estimate ˆγ η (2). If it is not negligible, then correct bias. However, It may not be practical to correct all eigenfunctions, as these are much more subtle and the recovery relies on the full eigenspace. Nevertheless, one way of avoiding the pitfall with selection of eigenfunctions, as demonstrated in Figure 2, is at least to correct eigenvalues and determine the importance of eigenfunctions in relation to the corresponding eigenvalues. In our discussion, we have not specified any particular registration method and these are suggested as a general principle. Comparison of γ η may be useful when different registration methods are employed. Alternatively, the moment based registration by James (2008) incorporates this idea by directly penalizing the deviation of η in estimation. More generally, the method developed by Kneip and Ramsay (2007) is based on a template matching and can be viewed as an elaborate implementation of our proposal, by iteratively updating registration based on functional PCA. Appendix General rth order approximation: Aφ (0) Assume that A is real symmetric and satisfies = λ (0) φ(0) = 0, 1,,. For identifiability we ssume that (φ, φ ) = 1 and (φ j, φ ) = 0 for j. Consider where (A + εb)φ (ε) = λ (ε)φ (ε) λ (ε) = λ (0) + ελ (1) + ε 2 λ (2) + φ (ε) = φ (0) + εφ (1) + ε 2 φ (2) +. Here λ = λ (0) and φ = φ (0). The equation may be expressed as (A + εb)(φ + εφ (1) + ε 2 φ (2) + ) = (λ + ελ (1) + ε 2 λ (2) + )(φ + εφ (1) + ε 2 φ (2) + ). Assuming the existence of infinite sum series, we have for r 1 Aφ (r) λ φ (r) = λ (r) φ + λ (r 1) φ (1) Denote the right hand side term by f so that f = λ (r) φ + λ (r 1) φ (1) + + λ (1) + + λ (1) φ(r 1) φ(r 1) Bφ (r 1). Bφ (r 1). (3)
12 and write Aφ (r) Observe that the lefthand side term Aφ (r) < φ l, Aφ (r) In other words, we have λ φ (r) λ φ (r) = f satisifies the following: λ φ (r) > = < (A λ )φ l, φ (r) > { 0 if l = = (λ l λ ) < φ l, φ (r) > if l < φ, f >= 0 < φ l, f >= (λ l λ ) < φ l, φ (r) > l. (4) Equating this with f in place generates λ (r) for r = 1 leads to results stated in Section 2.5. approximation, let us define α (r),l =< φ(r), φ l > so that φ (r) first order approximation results can be summarised as λ (1) = < Bφ, φ > α (1), = 0, α (1),l = < Bφ, φ l > λ λ l l. and φ (r) sequentially. In particular solving Before proceeding this for higher order = α (r), φ + l α(r),l φ l. Then the Consider r 2. For eigenvalue approximation, from (4), the rth order term λ (r) obtained from λ (r) = l α (r 1),l < Bφ, φ l > λ (r 1) α (1), λ(1) α(r 1),. can be For eigenfunction approximation, we need to derive coefficients α (r),l. It can be seen that the coefficient α (r), is related to the normalising condition as 2α (2), = j {α (1),j }2 2α (r), = j α (1),j α(r 1),j j α (2),j α(r 2),j j α (r 1),j α (1),j r 2. The remaining coefficients α (r),l are obtained from (λ λ l )α (r),l = j α (r 1),l < Bφ j, φ l > λ (r 1) α (1),l λ(1) α(r 1),l. To complete calculation, it is necessary to iterate between λ (r) λ (r 1) α (r 1),l ( or φ (r 1) ) λ (r) α (r),l ( or φ(r) ). and φ (r) sequentially, that is,
13 References [1] Dauxois, J. Pousse, A. and Romain, Y. (1982) Asymptotic theory for the principal component analysis of a vector random function: some applications to statistical inference. Journal of Multivariate Analysis, 12, [2] Liu, X. and Müller, H. G. (2004) Functional convex averaging and synchronization for -warped random curves. Journal of the American Statistical Association, 99, [3] Gasser, T. and Kneip, A. (1995) Searching for structure in curve samples. Journal of the American Statistical Association, 90, [4] Gasser, T. Müller, H. G., Köhler, W., Molinari, L. and Prader, A. (1984), Nonparametric regression analysis of growth curves, Annals of Statistics, 12, [5] Hall, P, Müller, H. G. and Wang, J. L. (2006). Properties of principal componenet methods for functional and longitudinal data analysis. Annals of Statistics, 34, [6] Kenip, A. (1994) Nonparametric estimation of common regressors for similar curve data. Annals of Statistics, 22, [7] Kneip, A. and Ramsay, J. O. (2007) Combining registration and fitting for functional models. technical report. [8] Par, J. Gasser, T. and Rousson, V. (2007) Structural components in functional data. technical report. [9] Ramsay, J. O. and Silverman, B. W. (2002) Applied functional data analysis, New Yor: Springer. [10] Ramsay, J. O. and Silverman, B. W. (2005) Functional data analysis, New Yor: Springer. [11] Rice, J. W. and Silverman, B. W. (1991) Estimating the mean and the covariance structure nonparametrically when the data are curves. Journal of Royal Statistical Society, B, 53, [12] Silverman, B. W. (1996) Smoothed functional principal components analysis by choice of norm. Annals of Statistics, 24, 1-24.
14 height velocity age (years) age (years) Figure 1: Example of height measurements (left) with estimated velocity curves (right) from Zürich growth studies. ]
15 pc 1 pc 2 pc 1 pc pc pc 4 pc 3 pc Figure 2: Functional PCA with (right) and without (left) variability. Additional variation produces many more components than anticipated.
16 la(1) la(2) la(3) la(4) Figure 3: Comparison of eigenvalues.
FUNCTIONAL DATA ANALYSIS. Contribution to the. International Handbook (Encyclopedia) of Statistical Sciences. July 28, Hans-Georg Müller 1
FUNCTIONAL DATA ANALYSIS Contribution to the International Handbook (Encyclopedia) of Statistical Sciences July 28, 2009 Hans-Georg Müller 1 Department of Statistics University of California, Davis One
More informationFUNCTIONAL DATA ANALYSIS
FUNCTIONAL DATA ANALYSIS Hans-Georg Müller Department of Statistics University of California, Davis One Shields Ave., Davis, CA 95616, USA. e-mail: mueller@wald.ucdavis.edu KEY WORDS: Autocovariance Operator,
More informationDynamic Relations for Sparsely Sampled Gaussian Processes
TEST manuscript No. (will be inserted by the editor) Dynamic Relations for Sparsely Sampled Gaussian Processes Hans-Georg Müller Wenjing Yang Received: date / Accepted: date Abstract In longitudinal studies,
More informationSmooth Common Principal Component Analysis
1 Smooth Common Principal Component Analysis Michal Benko Wolfgang Härdle Center for Applied Statistics and Economics benko@wiwi.hu-berlin.de Humboldt-Universität zu Berlin Motivation 1-1 Volatility Surface
More informationModeling Repeated Functional Observations
Modeling Repeated Functional Observations Kehui Chen Department of Statistics, University of Pittsburgh, Hans-Georg Müller Department of Statistics, University of California, Davis Supplemental Material
More informationFunctional modeling of longitudinal data
CHAPTER 1 Functional modeling of longitudinal data 1.1 Introduction Hans-Georg Müller Longitudinal studies are characterized by data records containing repeated measurements per subject, measured at various
More informationSecond-Order Inference for Gaussian Random Curves
Second-Order Inference for Gaussian Random Curves With Application to DNA Minicircles Victor Panaretos David Kraus John Maddocks Ecole Polytechnique Fédérale de Lausanne Panaretos, Kraus, Maddocks (EPFL)
More informationAN INTRODUCTION TO THEORETICAL PROPERTIES OF FUNCTIONAL PRINCIPAL COMPONENT ANALYSIS. Ngoc Mai Tran Supervisor: Professor Peter G.
AN INTRODUCTION TO THEORETICAL PROPERTIES OF FUNCTIONAL PRINCIPAL COMPONENT ANALYSIS Ngoc Mai Tran Supervisor: Professor Peter G. Hall Department of Mathematics and Statistics, The University of Melbourne.
More informationEmpirical Dynamics for Longitudinal Data
Empirical Dynamics for Longitudinal Data December 2009 Short title: Empirical Dynamics Hans-Georg Müller 1 Department of Statistics University of California, Davis Davis, CA 95616 U.S.A. Email: mueller@wald.ucdavis.edu
More informationA comparison of stratified simple random sampling and sampling with probability proportional to size
A comparison of stratified simple random sampling and sampling with probability proportional to size Edgar Bueno Dan Hedlin Per Gösta Andersson 1 Introduction When planning the sampling strategy (i.e.
More informationShrinkage Estimation for Functional Principal Component Scores, with Application to the Population Kinetics of Plasma Folate
Shrinkage Estimation for Functional Principal Component Scores, with Application to the Population Kinetics of Plasma Folate Fang Yao, Hans-Georg Müller,, Andrew J. Clifford, Steven R. Dueker, Jennifer
More informationCONCEPT OF DENSITY FOR FUNCTIONAL DATA
CONCEPT OF DENSITY FOR FUNCTIONAL DATA AURORE DELAIGLE U MELBOURNE & U BRISTOL PETER HALL U MELBOURNE & UC DAVIS 1 CONCEPT OF DENSITY IN FUNCTIONAL DATA ANALYSIS The notion of probability density for a
More informationSupplement to A Generalized Least Squares Matrix Decomposition. 1 GPMF & Smoothness: Ω-norm Penalty & Functional Data
Supplement to A Generalized Least Squares Matrix Decomposition Genevera I. Allen 1, Logan Grosenic 2, & Jonathan Taylor 3 1 Department of Statistics and Electrical and Computer Engineering, Rice University
More informationTests for separability in nonparametric covariance operators of random surfaces
Tests for separability in nonparametric covariance operators of random surfaces Shahin Tavakoli (joint with John Aston and Davide Pigoli) April 19, 2016 Analysis of Multidimensional Functional Data Shahin
More informationFunctional principal component analysis of aircraft trajectories
Functional principal component analysis of aircraft trajectories Florence Nicol To cite this version: Florence Nicol. Functional principal component analysis of aircraft trajectories. ISIATM 0, nd International
More informationWeakly dependent functional data. Piotr Kokoszka. Utah State University. Siegfried Hörmann. University of Utah
Weakly dependent functional data Piotr Kokoszka Utah State University Joint work with Siegfried Hörmann University of Utah Outline Examples of functional time series L 4 m approximability Convergence of
More informationFunctional Latent Feature Models. With Single-Index Interaction
Generalized With Single-Index Interaction Department of Statistics Center for Statistical Bioinformatics Institute for Applied Mathematics and Computational Science Texas A&M University Naisyin Wang and
More informationA Stickiness Coefficient for Longitudinal Data
A Stickiness Coefficient for Longitudinal Data June 2011 Andrea Gottlieb 1 Graduate Group in Biostatistics University of California, Davis 1 Shields Avenue Davis, CA 95616 U.S.A. Phone: 1 (530) 752-2361
More informationON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS
ON ILL-POSEDNESS OF NONPARAMETRIC INSTRUMENTAL VARIABLE REGRESSION WITH CONVEXITY CONSTRAINTS Olivier Scaillet a * This draft: July 2016. Abstract This note shows that adding monotonicity or convexity
More informationDerivative Principal Component Analysis for Representing the Time Dynamics of Longitudinal and Functional Data 1
Derivative Principal Component Analysis for Representing the Time Dynamics of Longitudinal and Functional Data 1 Short title: Derivative Principal Component Analysis Xiongtao Dai, Hans-Georg Müller Department
More informationA Stickiness Coefficient for Longitudinal Data
A Stickiness Coefficient for Longitudinal Data Revised Version November 2011 Andrea Gottlieb 1 Department of Statistics University of California, Davis 1 Shields Avenue Davis, CA 95616 U.S.A. Phone: 1
More informationModeling Multi-Way Functional Data With Weak Separability
Modeling Multi-Way Functional Data With Weak Separability Kehui Chen Department of Statistics University of Pittsburgh, USA @CMStatistics, Seville, Spain December 09, 2016 Outline Introduction. Multi-way
More informationftsa: An R Package for Analyzing Functional Time Series by Han Lin Shang
CONTRIBUTED RESEARCH ARTICLES 64 ftsa: An R Package for Analyzing Functional Time Series by Han Lin Shang Abstract Recent advances in computer recording and storing technology have tremendously increased
More informationOption 1: Landmark Registration We can try to align specific points. The Registration Problem. Landmark registration. Time-Warping Functions
The Registration Problem Most analyzes only account for variation in amplitude. Frequently, observed data exhibit features that vary in time. Option 1: Landmark Registration We can try to align specific
More informationRegularization via Spectral Filtering
Regularization via Spectral Filtering Lorenzo Rosasco MIT, 9.520 Class 7 About this class Goal To discuss how a class of regularization methods originally designed for solving ill-posed inverse problems,
More informationThe Mahalanobis distance for functional data with applications to classification
arxiv:1304.4786v1 [math.st] 17 Apr 2013 The Mahalanobis distance for functional data with applications to classification Esdras Joseph, Pedro Galeano and Rosa E. Lillo Departamento de Estadística Universidad
More informationMixture of Gaussian Processes and its Applications
Mixture of Gaussian Processes and its Applications Mian Huang, Runze Li, Hansheng Wang, and Weixin Yao The Pennsylvania State University Technical Report Series #10-102 College of Health and Human Development
More informationCourse topics (tentative) The role of random effects
Course topics (tentative) random effects linear mixed models analysis of variance frequentist likelihood-based inference (MLE and REML) prediction Bayesian inference The role of random effects Rasmus Waagepetersen
More informationFractal functional regression for classification of gene expression data by wavelets
Fractal functional regression for classification of gene expression data by wavelets Margarita María Rincón 1 and María Dolores Ruiz-Medina 2 1 University of Granada Campus Fuente Nueva 18071 Granada,
More informationSpectral Regularization
Spectral Regularization Lorenzo Rosasco 9.520 Class 07 February 27, 2008 About this class Goal To discuss how a class of regularization methods originally designed for solving ill-posed inverse problems,
More informationThe loss function and estimating equations
Chapter 6 he loss function and estimating equations 6 Loss functions Up until now our main focus has been on parameter estimating via the maximum likelihood However, the negative maximum likelihood is
More informationFUNCTIONAL DATA ANALYSIS FOR VOLATILITY PROCESS
FUNCTIONAL DATA ANALYSIS FOR VOLATILITY PROCESS Rituparna Sen Monday, July 31 10:45am-12:30pm Classroom 228 St-C5 Financial Models Joint work with Hans-Georg Müller and Ulrich Stadtmüller 1. INTRODUCTION
More informationNonresponse weighting adjustment using estimated response probability
Nonresponse weighting adjustment using estimated response probability Jae-kwang Kim Yonsei University, Seoul, Korea December 26, 2006 Introduction Nonresponse Unit nonresponse Item nonresponse Basic strategy
More informationIntelligent Data Analysis. Principal Component Analysis. School of Computer Science University of Birmingham
Intelligent Data Analysis Principal Component Analysis Peter Tiňo School of Computer Science University of Birmingham Discovering low-dimensional spatial layout in higher dimensional spaces - 1-D/3-D example
More informationMIXTURE INNER PRODUCT SPACES AND THEIR APPLICATION TO FUNCTIONAL DATA ANALYSIS
MIXTURE INNER PRODUCT SPACES AND THEIR APPLICATION TO FUNCTIONAL DATA ANALYSIS Zhenhua Lin 1, Hans-Georg Müller 2 and Fang Yao 1,3 Abstract We introduce the concept of mixture inner product spaces associated
More informationDegradation Modeling and Monitoring of Truncated Degradation Signals. Rensheng Zhou, Nagi Gebraeel, and Nicoleta Serban
Degradation Modeling and Monitoring of Truncated Degradation Signals Rensheng Zhou, Nagi Gebraeel, and Nicoleta Serban School of Industrial and Systems Engineering, Georgia Institute of Technology Abstract:
More informationExchangeable Bernoulli Random Variables And Bayes Postulate
Exchangeable Bernoulli Random Variables And Bayes Postulate Moulinath Banerjee and Thomas Richardson September 30, 20 Abstract We discuss the implications of Bayes postulate in the setting of exchangeable
More informationBIAS-ROBUSTNESS AND EFFICIENCY OF MODEL-BASED INFERENCE IN SURVEY SAMPLING
Statistica Sinica 22 (2012), 777-794 doi:http://dx.doi.org/10.5705/ss.2010.238 BIAS-ROBUSTNESS AND EFFICIENCY OF MODEL-BASED INFERENCE IN SURVEY SAMPLING Desislava Nedyalova and Yves Tillé University of
More informationDiagnostics for functional regression via residual processes
Diagnostics for functional regression via residual processes Jeng-Min Chiou a, Hans-Georg Müller b, a Academia Sinica, 128 Academia Road Sec.2, Taipei 11529, Taiwan b University of California, Davis, One
More informationAn Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data
An Efficient Estimation Method for Longitudinal Surveys with Monotone Missing Data Jae-Kwang Kim 1 Iowa State University June 28, 2012 1 Joint work with Dr. Ming Zhou (when he was a PhD student at ISU)
More informationAnalysis of AneuRisk65 data: warped logistic discrimination
Electronic Journal of Statistics Vol. () ISSN: 935-7524 DOI:./ Analysis of AneuRisk65 data: warped logistic discrimination Daniel Gervini Department of Mathematical Sciences University of Wisconsin Milwaukee
More informationTIME-WARPED GROWTH PROCESSES, WITH APPLICATIONS TO THE MODELING OF BOOM-BUST CYCLES IN HOUSE PRICES
Submitted to the Annals of Applied Statistics arxiv: arxiv:0000.0000 TIME-WARPED GROWTH PROCESSES, WITH APPLICATIONS TO THE MODELING OF BOOM-BUST CYCLES IN HOUSE PRICES By Jie Peng,, Debashis Paul, and
More informationStructure in Data. A major objective in data analysis is to identify interesting features or structure in the data.
Structure in Data A major objective in data analysis is to identify interesting features or structure in the data. The graphical methods are very useful in discovering structure. There are basically two
More informationsparse and low-rank tensor recovery Cubic-Sketching
Sparse and Low-Ran Tensor Recovery via Cubic-Setching Guang Cheng Department of Statistics Purdue University www.science.purdue.edu/bigdata CCAM@Purdue Math Oct. 27, 2017 Joint wor with Botao Hao and Anru
More informationECON 3150/4150, Spring term Lecture 6
ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture
More informationarxiv: v1 [stat.me] 4 Oct 2016
Hybrid Regularisation of Functional Linear Models Anirvan Chakraborty and Victor M. Panaretos Institut de Mathématiques, École Polytechnique Fédérale de Lausanne e-mail: anirvan.chakraborty@epfl.ch; victor.panaretos@epfl.ch
More informationNonparametric Inference In Functional Data
Nonparametric Inference In Functional Data Zuofeng Shang Purdue University Joint work with Guang Cheng from Purdue Univ. An Example Consider the functional linear model: Y = α + where 1 0 X(t)β(t)dt +
More informationDiagnostics for functional regression via residual processes
Diagnostics for functional regression via residual processes Jeng-Min Chiou Academia Sinica, Taiwan E-mail: jmchiou@stat.sinica.edu.tw Hans-Georg Müller University of California, Davis, USA E-mail: mueller@wald.ucdavis.edu
More informationNetworked Control Systems, Event-Triggering, Small-Gain Theorem, Nonlinear
EVENT-TRIGGERING OF LARGE-SCALE SYSTEMS WITHOUT ZENO BEHAVIOR C. DE PERSIS, R. SAILER, AND F. WIRTH Abstract. We present a Lyapunov based approach to event-triggering for large-scale systems using a small
More information7 Day 3: Time Varying Parameter Models
7 Day 3: Time Varying Parameter Models References: 1. Durbin, J. and S.-J. Koopman (2001). Time Series Analysis by State Space Methods. Oxford University Press, Oxford 2. Koopman, S.-J., N. Shephard, and
More informationOptimal Auxiliary Variable Assisted Two-Phase Sampling Designs
MASTER S THESIS Optimal Auxiliary Variable Assisted Two-Phase Sampling Designs HENRIK IMBERG Department of Mathematical Sciences Division of Mathematical Statistics CHALMERS UNIVERSITY OF TECHNOLOGY UNIVERSITY
More informationMultiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions
Multiple integrals: Sufficient conditions for a local minimum, Jacobi and Weierstrass-type conditions March 6, 2013 Contents 1 Wea second variation 2 1.1 Formulas for variation........................
More informationConvergence of Eigenspaces in Kernel Principal Component Analysis
Convergence of Eigenspaces in Kernel Principal Component Analysis Shixin Wang Advanced machine learning April 19, 2016 Shixin Wang Convergence of Eigenspaces April 19, 2016 1 / 18 Outline 1 Motivation
More informationBindel, Fall 2011 Intro to Scientific Computing (CS 3220) Week 3: Wednesday, Jan 9
Problem du jour Week 3: Wednesday, Jan 9 1. As a function of matrix dimension, what is the asymptotic complexity of computing a determinant using the Laplace expansion (cofactor expansion) that you probably
More informationDiagnostics for Linear Models With Functional Responses
Diagnostics for Linear Models With Functional Responses Qing Shen Edmunds.com Inc. 2401 Colorado Ave., Suite 250 Santa Monica, CA 90404 (shenqing26@hotmail.com) Hongquan Xu Department of Statistics University
More informationStatistics 612: L p spaces, metrics on spaces of probabilites, and connections to estimation
Statistics 62: L p spaces, metrics on spaces of probabilites, and connections to estimation Moulinath Banerjee December 6, 2006 L p spaces and Hilbert spaces We first formally define L p spaces. Consider
More informationFlexible Estimation of Treatment Effect Parameters
Flexible Estimation of Treatment Effect Parameters Thomas MaCurdy a and Xiaohong Chen b and Han Hong c Introduction Many empirical studies of program evaluations are complicated by the presence of both
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 6: Bias and variance (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 49 Our plan today We saw in last lecture that model scoring methods seem to be trading off two different
More informationPrincipal Component Analysis
Machine Learning Michaelmas 2017 James Worrell Principal Component Analysis 1 Introduction 1.1 Goals of PCA Principal components analysis (PCA) is a dimensionality reduction technique that can be used
More informationFERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA. Ning He, Lei Xie, Shu-qing Wang, Jian-ming Zhang
FERMENTATION BATCH PROCESS MONITORING BY STEP-BY-STEP ADAPTIVE MPCA Ning He Lei Xie Shu-qing Wang ian-ming Zhang National ey Laboratory of Industrial Control Technology Zhejiang University Hangzhou 3007
More informationSupplement to Quantile-Based Nonparametric Inference for First-Price Auctions
Supplement to Quantile-Based Nonparametric Inference for First-Price Auctions Vadim Marmer University of British Columbia Artyom Shneyerov CIRANO, CIREQ, and Concordia University August 30, 2010 Abstract
More informationBusiness Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM
Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF
More informationA Note on Hilbertian Elliptically Contoured Distributions
A Note on Hilbertian Elliptically Contoured Distributions Yehua Li Department of Statistics, University of Georgia, Athens, GA 30602, USA Abstract. In this paper, we discuss elliptically contoured distribution
More informationRegularized principal components analysis
9 Regularized principal components analysis 9.1 Introduction In this chapter, we discuss the application of smoothing to functional principal components analysis. In Chapter 5 we have already seen that
More informationAn asymptotic ratio characterization of input-to-state stability
1 An asymptotic ratio characterization of input-to-state stability Daniel Liberzon and Hyungbo Shim Abstract For continuous-time nonlinear systems with inputs, we introduce the notion of an asymptotic
More informationCheng Soon Ong & Christian Walder. Canberra February June 2018
Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 (Many figures from C. M. Bishop, "Pattern Recognition and ") 1of 254 Part V
More informationFunctional quasi-likelihood regression models with smooth random effects
J. R. Statist. Soc. B (2003) 65, Part 2, pp. 405 423 Functional quasi-likelihood regression models with smooth random effects Jeng-Min Chiou National Health Research Institutes, Taipei, Taiwan and Hans-Georg
More informationIndependent component analysis for functional data
Independent component analysis for functional data Hannu Oja Department of Mathematics and Statistics University of Turku Version 12.8.216 August 216 Oja (UTU) FICA Date bottom 1 / 38 Outline 1 Probability
More informationA Novel Nonparametric Density Estimator
A Novel Nonparametric Density Estimator Z. I. Botev The University of Queensland Australia Abstract We present a novel nonparametric density estimator and a new data-driven bandwidth selection method with
More informationFunction-on-Scalar Regression with the refund Package
University of Haifa From the SelectedWorks of Philip T. Reiss July 30, 2012 Function-on-Scalar Regression with the refund Package Philip T. Reiss, New York University Available at: https://works.bepress.com/phil_reiss/28/
More informationCh. 15 Forecasting. 1.1 Forecasts Based on Conditional Expectations
Ch 15 Forecasting Having considered in Chapter 14 some of the properties of ARMA models, we now show how they may be used to forecast future values of an observed time series For the present we proceed
More informationChapter 2: simple regression model
Chapter 2: simple regression model Goal: understand how to estimate and more importantly interpret the simple regression Reading: chapter 2 of the textbook Advice: this chapter is foundation of econometrics.
More informationMatrix Factorizations
1 Stat 540, Matrix Factorizations Matrix Factorizations LU Factorization Definition... Given a square k k matrix S, the LU factorization (or decomposition) represents S as the product of two triangular
More informationThis article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and
This article appeared in a journal published by Elsevier. The attached copy is furnished to the author for internal non-commercial research and education use, including for instruction at the authors institution
More informationSMOOTHED FUNCTIONAL PRINCIPAL COMPONENTS ANALYSIS BY CHOICE OF NORM 1. BY BERNARD W. SILVERMAN University of Bristol
The Annals of Statistics 1996, Vol. 4, No. 1, 14 SMOOTHED FUNCTIONAL PRINCIPAL COMPONENTS ANALYSIS BY CHOICE OF NORM 1 BY BERNARD W. SILVERMAN University of Bristol The principal components analysis of
More informationStochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions
International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.
More informationMultivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma
Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Suppose again we have n sample points x,..., x n R p. The data-point x i R p can be thought of as the i-th row X i of an n p-dimensional
More informationFunction Approximation
1 Function Approximation This is page i Printer: Opaque this 1.1 Introduction In this chapter we discuss approximating functional forms. Both in econometric and in numerical problems, the need for an approximating
More informationIntroduction to Functional Data Analysis A CSCU Workshop. Giles Hooker Biological Statistics and Computational Biology
Introduction to Functional Data Analysis A CSCU Workshop Giles Hooker Biological Statistics and Computational Biology gjh27@cornell.edu www.bscb.cornell.edu/ hooker/fdaworkshop 1 / 26 Agenda What is Functional
More informationPart 6: Multivariate Normal and Linear Models
Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of
More informationEconomics 536 Lecture 7. Introduction to Specification Testing in Dynamic Econometric Models
University of Illinois Fall 2016 Department of Economics Roger Koenker Economics 536 Lecture 7 Introduction to Specification Testing in Dynamic Econometric Models In this lecture I want to briefly describe
More informationOnline Appendix. j=1. φ T (ω j ) vec (EI T (ω j ) f θ0 (ω j )). vec (EI T (ω) f θ0 (ω)) = O T β+1/2) = o(1), M 1. M T (s) exp ( isω)
Online Appendix Proof of Lemma A.. he proof uses similar arguments as in Dunsmuir 979), but allowing for weak identification and selecting a subset of frequencies using W ω). It consists of two steps.
More informationCurves clustering with approximation of the density of functional random variables
Curves clustering with approximation of the density of functional random variables Julien Jacques and Cristian Preda Laboratoire Paul Painlevé, UMR CNRS 8524, University Lille I, Lille, France INRIA Lille-Nord
More informationDEGRADATION MODELING AND MONITORING OF ENGINEERING SYSTEMS USING FUNCTIONAL DATA ANALYSIS
DEGRADATION MODELING AND MONITORING OF ENGINEERING SYSTEMS USING FUNCTIONAL DATA ANALYSIS A Thesis Presented to The Academic Faculty by Rensheng Zhou In Partial Fulfillment of the Requirements for the
More informationNotes on Measure Theory and Markov Processes
Notes on Measure Theory and Markov Processes Diego Daruich March 28, 2014 1 Preliminaries 1.1 Motivation The objective of these notes will be to develop tools from measure theory and probability to allow
More informationEstimating Moving Average Processes with an improved version of Durbin s Method
Estimating Moving Average Processes with an improved version of Durbin s Method Maximilian Ludwig this version: June 7, 4, initial version: April, 3 arxiv:347956v [statme] 6 Jun 4 Abstract This paper provides
More informationInterpretable Functional Principal Component Analysis
Biometrics 000, 000 000 DOI: 000 000 0000 Interpretable Functional Principal Component Analysis Zhenhua Lin 1,, Liangliang Wang 2,, and Jiguo Cao 2, 1 Department of Statistical Sciences, University of
More informationarxiv: v1 [stat.me] 18 Jul 2015
Review of Functional Data Analysis Jane-Ling Wang, 1 Jeng-Min Chiou, 2 and Hans-Georg Müller 1 1 Department of Statistics, University of California, Davis, USA, 95616 2 Institute of Statistical Science,
More informationPrincipal component analysis
Principal component analysis Angela Montanari 1 Introduction Principal component analysis (PCA) is one of the most popular multivariate statistical methods. It was first introduced by Pearson (1901) and
More informationFundamental Issues in Bayesian Functional Data Analysis. Dennis D. Cox Rice University
Fundamental Issues in Bayesian Functional Data Analysis Dennis D. Cox Rice University 1 Introduction Question: What are functional data? Answer: Data that are functions of a continuous variable.... say
More informationOptimal spectral shrinkage and PCA with heteroscedastic noise
Optimal spectral shrinage and PCA with heteroscedastic noise William Leeb and Elad Romanov Abstract This paper studies the related problems of denoising, covariance estimation, and principal component
More informationConvex Functions and Optimization
Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized
More informationLatent Variable Models and EM algorithm
Latent Variable Models and EM algorithm SC4/SM4 Data Mining and Machine Learning, Hilary Term 2017 Dino Sejdinovic 3.1 Clustering and Mixture Modelling K-means and hierarchical clustering are non-probabilistic
More informationNonparametric Modal Regression
Nonparametric Modal Regression Summary In this article, we propose a new nonparametric modal regression model, which aims to estimate the mode of the conditional density of Y given predictors X. The nonparametric
More informationStatistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach
Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score
More informationarxiv: v1 [stat.me] 3 Sep 2016
Functional Data Analysis by Matrix Completion Marie-Hélène Descary and Victor M. Panaretos Institut de Mathématiques Ecole Polytechnique Fédérale de Lausanne e-mail: marie-helene.descary@epfl.ch, victor.panaretos@epfl.ch
More informationTime Series and Forecasting Lecture 4 NonLinear Time Series
Time Series and Forecasting Lecture 4 NonLinear Time Series Bruce E. Hansen Summer School in Economics and Econometrics University of Crete July 23-27, 2012 Bruce Hansen (University of Wisconsin) Foundations
More informationAsymptotic inference for a nonstationary double ar(1) model
Asymptotic inference for a nonstationary double ar() model By SHIQING LING and DONG LI Department of Mathematics, Hong Kong University of Science and Technology, Hong Kong maling@ust.hk malidong@ust.hk
More informationFocusing on structural assumptions in regression on functional variable.
Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session IPS043) p.798 Focusing on structural assumptions in regression on functional variable. DELSOL, Laurent Université d
More informationDiscussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis
Discussion of Bootstrap prediction intervals for linear, nonlinear, and nonparametric autoregressions, by Li Pan and Dimitris Politis Sílvia Gonçalves and Benoit Perron Département de sciences économiques,
More information