Multidimensional Density Smoothing with P-splines

Size: px
Start display at page:

Download "Multidimensional Density Smoothing with P-splines"

Transcription

1 Multidimensional Density Smoothing with P-splines Paul H.C. Eilers, Brian D. Marx 1 Department of Medical Statistics, Leiden University Medical Center, 300 RC, Leiden, The Netherlands (p.eilers@lumc.nl) Department of Experimental Statistics, Louisiana State University, Baton Rouge, Louisiana USA (bmarx@lsu.edu) Abstract: We propose a simple and effective multidimensional density estimator. Our approach is essentially penalized Poisson regression using a rich tensor product B-spline basis, where the Poisson counts come from processing the data into a multidimensional histogram, often consisting of thousands of bins. The penalty enforces smoothness of the B-spline coefficients, specifically within the rows, columns, layers depending on the dimension. In this paper, we focus on how a one-dimensional P-spline density estimator can be extended to two-dimensions, and beyond. In higher dimensions we provide a hint on how efficient grid algorithms can be implemented using array regression. Our approach optimizes the penalty weight parameter(s) using information criteria, specifically AIC. Two examples illustrate our method in two-dimensions. Keywords: AIC; Effective Dimension; Tensor Product. 1 Introduction Density estimation is a core activity in data exploration and applied statistics. Surprisingly, many statisticians approach it in a quite unsophisticated way: they are happy with off-the-shelf kernel smoothers, as can be seen in many papers on computational Bayesian modeling. An attractive alternative is (penalized) likelihood smoothing, modeling the logarithm of the density with splines. Kooperberg and Stone (1991) use adaptive-knot B- splines, while Eilers and Marx (1996) combine fixed-knot B-splines with a roughness penalty (P-splines), simplifying the scheme of O Sullivan (1988). In the P-spline approach, the raw data are processed and reduced to a histogram, then density estimation is achieved through (penalized) Poisson regression. This is attractive for high-volume one-dimensional data, because one deals with, e.g., 100 or 00 histogram bins instead of many thousands of raw observations. In two or more dimensions the histogram approach is even more attractive, when combined with a recently developed algorithm for fast weighted smoothing on grids (Eilers et al., 006).

2 Multidimensional P-spline Density A critical issue is optimization of the amount of smoothing. We use Akaike s Information Criterion (AIC). It is easy to compute, because an effective model dimension can be defined, and it can be computed with relatively little effort. P-spline overview: one-dimensional densities Let (y i, u i ) denote the Poisson count and bin midpoint pairs from a (narrowly binned) histogram, i = 1,..., n. The vector of counts is denoted y. We model the expected values of the counts as µ i = E(y i ) = exp( c b j (u i )θ j ), (1) or in matrix terms µ = Bθ, where B = [b j (u i )] is a (n c) B-spline basis built along the indexing axis u of the density. A rich basis is used (c sufficiently large) and the knots for the basis are equally-spaced. Apart from a constant, the Poisson log-likelihood is l = n i=1 log(µ yi i e µi ) = j=1 n (y i log µ i µ i ). () A penalty on the d-th differences of θ is subtracted from the log-likelihood, to tune smoothness. The number of basis functions in B is chosen large, to give ample flexibility. The penalized log-likelihood, l, then is l = l λ c j=d+1 i=1 Setting the gradient of (3) equal to zero gives ( d θ j ). (3) B (y µ) = λd Dθ, (4) where D is a matrix of contrasts such that Dθ = d θ. Linearization of (4) leads to (B W B + λd D)θ = B W z, () where z = η + W 1 (y µ) is the working variable and η = Bθ. The matrix W = diag(µ) and θ, µ are approximations to the solution of (), iterated until convergence. One easily recognizes the familiar equations for fitting a GLM, modified by the term λd D that stems from the penalty. At convergence, one can interpret () as linear smoothing of the working variable z. We have ẑ = Bˆθ = B(B Ŵ B + λd D) 1 B Ŵ z = Hz, (6)

3 P.H.C. Eilers and B.D. Marx 3 where H is the smoothing or hat matrix. The effective dimension of the smoother is approximated by race(h). We use it to define Akaike s information criterion (AIC) as AIC = n y i log(y i /ˆµ i ) + trace(h). (7) i=1 Essentially this is all old hat. A more detailed description of density smoothing with P-splines is given by Eilers and Marx (1996). We use the presentation above as an impetus to move density smoothing with use tensor product P-splines in two or more dimensions. 3 Two dimensional P-spline densities Let Y ih be counts in a two-dimensional (narrowly binned) n 1 n histogram, forming the array Y of Poisson counts with expectation M. The bin midpoints are now indexed by the n 1 n pairs (u 1 1 n, 1 n1 u ), where u 1 (u ) are the midpoint locations along the first (second) dimension. We model the expected values, now using tensor product B-splines on a rich c 1 c rectangular grid (equally-spaced along each dimension), as µ ih = E(Y ih ) = exp( c 1 c j=1 k=1 b j (u 1i )b k (u h )θ jk ). (8) The n 1 n -vectorized form of the Poisson counts and corresponding expected values are denoted by y = vec(y ) and µ = vec(m), respectively. Ignoring penalties for the moment, (8) is also a GLM, and can be rewritten as µ = E(y) = B B 1 θ = Bθ. (9) We now have the c 1 c array of coefficients Θ, with vectorized form θ = vec(θ). The matrix B, with n 1 n rows and c 1 c columns, collects the tensor product basis functions. As outlined in Durbán et al. (004), the penalty can be written as λ 1 D 1 Θ F + λ ΘD F = θ (λ 1 I c D 1D 1 + λ D D I c1 )θ = θ P θ, where F indicates the quadratic Frobenius norm, the sum of the squares of all elements of a matrix, and D 1 (D ) is the proper contrast matrix for the rows (columns) of Θ. Such a penalty enforces smoothness or tensor product B-spline coefficients within any one row or column, but breaks linkage in penalization across rows or columns. As seen in (??), the weight and order of the penalty is the same from row to row, the same from column to column, but can vary from rows to columns. In this scheme, with η = Bθ, the equivalent of () becomes (B W B + P )θ = B W ( η + W 1 (y µ)). (10)

4 4 Multidimensional P-spline Density 3.1 Optimizing the penalty To optimize the amount of smoothing, AIC can be computed similarly to (7). Upon convergence AIC is now computed as n 1 n AIC = y ih log(y ih /ˆµ ih ) + trace{b Ŵ B(B Ŵ B + P ) 1 }, (11) i=1 h=1 borrowing the result that trace is invariant under cyclical permutation of matrices. One varies the logarithms of λ 1 and λ on a grid, searching for the minimum. In many cases, or for first exploration, isotropic smoothing (λ 1 = λ = λ) will be sufficient and the grid search can be one-dimensional. 3. Toward higher dimensions Although this scheme will work, it can be wasteful with memory and computation time, especially for higher dimensions. For example, in three dimensions (with common n and c), the basis would have n 3 rows and c 3 columns and most practical applications would not fit in the memory of the average PC. To avoid this bottleneck, we use the grid algorithms of Eilers et al. (006). An example is the computation of the mean µ = vec(m) in two dimensions. Instead of computing it as a vector, as in (9), we can use the following equivalence µ = E(y) = B B 1 θ = Bθ M = E(Y ) = B 1 ΘB. (1) This avoids the construction of the large Kronecker product basis, saving space and time. This scheme can be extended to 3 dimensions or more, by recursive re-arrangement of multi-dimensional arrays as matrices and pre-multiplication by the one-dimensional bases. In a similar way the multi-dimensional inner products B W B can be obtained efficiently. The key observation is that in a 4-dimensional representation the elements of B W B can be written f jkj k = w hi b ij b hk b ij b hk, (13) h i which can be rearranged as f jkj k = h b hk b hk w hi b ij b ij. (14) By clever use of row-wise tensor products and switching between four- and two-dimensional representations of matrices and arrays, one can completely avoid the construction of the large tensor-product basis B. In addition to saving memory, computations are sped up by orders of magnitude. There is no room for details here, see Eilers et al. (006) and Currie et al. (006). i

5 P.H.C. Eilers and B.D. Marx 0. AIC Old Faithful, optimal smooth, anisotropic log 10 (λ r ) Waiting time (min) log (λ ) 10 c 40 Duration (min) FIGURE 1. Non-isotropic smoothing of Old Faithful data. Left: contours of AIC; right: contours of smooth density estimate. The density was normalized so that the largest value is equal to 1. The contour levels are 0.9, 0.7, 0., 0.3, 0., 0.1, 0.07, 0.0, 0.03, 0.0, Isotropic smoothing; λ = Old Faithful, optimal smooth AIC 1000 Waiting time (min) log 10 (λ c ) 40 Duration (min) FIGURE. Isotropic smoothing of Old Faithful data. Left: AIC profile; right: contours of smooth density estimate. The density was normalized so that the largest value is equal to 1. The contour levels are 0.9, 0.7, 0., 0.3, 0., 0.1, 0.07, 0.0, 0.03, 0.0, Illustrative examples Figure 1 shows two-dimensional density contours for the well-known Old Faithful data. The number of Poisson observations is on a his-

6 6 Multidimensional P-spline Density 6 Microarray data; # selected points: Isotropic smoothing; λ = 0.00 Array (log Array 1 (log 10 AIC log 10 (λ c ) 6 Density estimate; λ = Square root of density estimate; λ = 0.00 Array (log Array (log Array 1 (log 10 1 Array 1 (log 10 FIGURE 3. Presentation of microarray data scatterplot as a density. Upper: scatterplot of selection of raw data (left); optimal isotropic penalty parameter using AIC (right). lower: image of optimal density (left); square root density (for improved visualization) (right). togram grid, and a grid of tensor products of cubic B-splines were used to construct the two-dimensional density. The order of the difference penalty is 3. We see that AIC is minimized with relatively light smoothing (λ 1.003) along the rows (waiting times) and with heavier smoothing (λ 1000) along the columns (duration). Figure shows the result of isotropic smoothing, yielding a compromise in the single λ 0.0. The large difference between the two λs for anisotropic smoothing might look surprising. The reason is that the conditional distributions of waiting time given (the previous) duration is always unimodal, not much different from a normal distribution. This allows a large weight for a third order penalty without destroying a proper fit to the data. On the other hand the conditional distribution of the (previous) duration given the waiting time is bimodal or rather skewed on part of the data domain. The weight of horizontal penalty cannot therefore be large. As a second example, Figure 3 displays as scatterplot of gene expressions (on a log 10 scale) as measured by two microarrays. The data pairs are plotted using a simple dot symbol. Only a random selection of 1000 observa-

7 P.H.C. Eilers and B.D. Marx 7 tions is shown, to prevent a large part of the cloud of pints to become completely black. As discussed by Eilers and Goeman (004), scatterplots with many dots are difficult to judge, because symbols at the boundaries attract too much attention from the observer, making the cloud of points look wider than it really is. Eilers and Goeman presented a very fast but simple smoother, with a goal of an attractive display of dense scatterplots. The density smoother proposed here can be used as an improvement, especially when λ is optimized by AIC. The results presented were also obtained with a histogram grid and tensor products of cubic B-splines. The optimal penalty weight value for isotropic smoothing is λ The result of the smoother is presented as a gray-scale image of the height of the density. Depending on the medium used (computer screen, projector or printer) the effective dynamic range can be quite small. Therefore we also present a gray scale plot of the square root of the density. Discussion We have shown that tensor products of P-splines and Poisson regression of histogram data leads to an effective density estimator. As discussed in Eilers and Marx (1996), the position and number of bins (i.e. the location and width of the histogram grid) makes little difference to the final fit, provided that the grid is chosen to be sufficiently fine. There is no room here to illustrate that the same holds for two-dimensional smoothing. It is remarkable that a very sparse two-dimensional histogram leads to a very good density estimate. The Old Faithful data set has 7 observations. The average count per -D histogram bin is thus less than 0.03! As a practical guideline in two dimensions, we recommend that users (who are not concerned with computation time) start with histogram bins, 0 0 B-splines, and a second or third order penalty. Some nice features of one-dimensional P-spline density estimation also carry over to higher dimensions. For one, there is a conservation of moments, for any λ, within each row, column, layer and so forth (depending on the dimension of the histogram). For example within any row, a penalty of order d = 1 yields equality of the sum of the observed and the sum of fitted counts. When d =, e.g. within any row, the previous condition holds and the mean of the observed data equals that of the fitted counts. Further, when d = 3, the previous two conditions hold with an additional preservation of variance between observed and fitted data. The P-spline density estimator is also not affected by unpleasant boundary effects, so familiar to kernel smoothers. In fact sharp specialized boundaries are encouraged with the P-spline approach. Especially for isotropic smoothing, optimization of AIC, using a simple grid search, is efficient. Computation, for one value of λ (in Matlab), takes less than a second on a 1000 MHz PIII computer. The time to compute the histogram is negligible, even for many observations.

8 8 Multidimensional P-spline Density Minimization of AIC is not the only possible approach to optimal smoothing. One can interpret models with penalties as mixed models and attempt to estimate the variance of the mixing distribution (Ruppert, Wand and Carroll, 003). Durbán et al. (006) investigated this approach to multidimensional density estimation. Alternatively, a purely Bayesian approach is also possible. The penalty is seen as the logarithm of a prior density of differences of the coefficients of the B-spline tensor products. Efficient simulation, using an optimized Langevin-Hastings algorithm, is possible. See Lambert and Eilers (006), extending their 00 work. References Currie, I.D., Durbán, M. and Eilers, P.H.C. (006). Generalized linear array models with applications to multidimensional smoothing. Journal of the Royal Statistical Society, 68, 1-. Durbán, M., Currie, I.D. and Eilers, P.H.C. (004). Smoothing and forecasting mortality rates. Statistical Modelling, 4, Durbán, M., Currie, I.D. and Eilers, P.H.C. (006). Multi-dimensional density smoothing with P-splines: a mixed model approach. Proceeding of the International Workshop on Statistical Modelling, Galway. Eilers, P.H.C., Currie, I.D. and and Durbán, M. (006). Fast and compact smoothing on large multidimensional grids. Computational Statistics and Data Analysis, 0, Eilers, P.H.C and Goeman, J. (004). Enhancing scatterplots with smoothed densities. Bioinformatics, 0, Eilers, P.H.C. and Marx, B.D. (1996). Flexible smoothing using B-splines and penalized likelihood (with Comments and Rejoinder). Statistical Science, 11(), Kooperberg. C. and Stone, C.J. (1991). A study of logspline density estimation. Computational Statistics and Data Analysis, 1, Lambert, P. and Eilers, P.H.C. (00). Bayesian proportional hazards model with time-varying coefficients: a penalized Poisson regression approach. Statistics in Medicine 4, Lambert, P. and Eilers, P.H.C. (006). Bayesian multidimensional density smoothing. Proceeding of the International Workshop on Statistical Modelling, Galway. O Sullivan, F. (1988). Fast computation of fully automated log-density and log-hazard estimators. SIAM J. Sci. Stat. Comput., 9, Ruppert, D., Wand, M. and Caroll, R. (003). Semiparametric Regression. New York: Cambridge.

Currie, Iain Heriot-Watt University, Department of Actuarial Mathematics & Statistics Edinburgh EH14 4AS, UK

Currie, Iain Heriot-Watt University, Department of Actuarial Mathematics & Statistics Edinburgh EH14 4AS, UK An Introduction to Generalized Linear Array Models Currie, Iain Heriot-Watt University, Department of Actuarial Mathematics & Statistics Edinburgh EH14 4AS, UK E-mail: I.D.Currie@hw.ac.uk 1 Motivating

More information

GLAM An Introduction to Array Methods in Statistics

GLAM An Introduction to Array Methods in Statistics GLAM An Introduction to Array Methods in Statistics Iain Currie Heriot Watt University GLAM A Generalized Linear Array Model is a low-storage, high-speed, method for multidimensional smoothing, when data

More information

A Hierarchical Perspective on Lee-Carter Models

A Hierarchical Perspective on Lee-Carter Models A Hierarchical Perspective on Lee-Carter Models Paul Eilers Leiden University Medical Centre L-C Workshop, Edinburgh 24 The vantage point Previous presentation: Iain Currie led you upward From Glen Gumbel

More information

P -spline ANOVA-type interaction models for spatio-temporal smoothing

P -spline ANOVA-type interaction models for spatio-temporal smoothing P -spline ANOVA-type interaction models for spatio-temporal smoothing Dae-Jin Lee 1 and María Durbán 1 1 Department of Statistics, Universidad Carlos III de Madrid, SPAIN. e-mail: dae-jin.lee@uc3m.es and

More information

Space-time modelling of air pollution with array methods

Space-time modelling of air pollution with array methods Space-time modelling of air pollution with array methods Dae-Jin Lee Royal Statistical Society Conference Edinburgh 2009 D.-J. Lee (Uc3m) GLAM: Array methods in Statistics RSS 09 - Edinburgh # 1 Motivation

More information

Using P-splines to smooth two-dimensional Poisson data

Using P-splines to smooth two-dimensional Poisson data 1 Using P-splines to smooth two-dimensional Poisson data Maria Durbán 1, Iain Currie 2, Paul Eilers 3 17th IWSM, July 2002. 1 Dept. Statistics and Econometrics, Universidad Carlos III de Madrid, Spain.

More information

Flexible Spatio-temporal smoothing with array methods

Flexible Spatio-temporal smoothing with array methods Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session IPS046) p.849 Flexible Spatio-temporal smoothing with array methods Dae-Jin Lee CSIRO, Mathematics, Informatics and

More information

Recovering Indirect Information in Demographic Applications

Recovering Indirect Information in Demographic Applications Recovering Indirect Information in Demographic Applications Jutta Gampe Abstract In many demographic applications the information of interest can only be estimated indirectly. Modelling events and rates

More information

MULTIDIMENSIONAL COVARIATE EFFECTS IN SPATIAL AND JOINT EXTREMES

MULTIDIMENSIONAL COVARIATE EFFECTS IN SPATIAL AND JOINT EXTREMES MULTIDIMENSIONAL COVARIATE EFFECTS IN SPATIAL AND JOINT EXTREMES Philip Jonathan, Kevin Ewans, David Randell, Yanyun Wu philip.jonathan@shell.com www.lancs.ac.uk/ jonathan Wave Hindcasting & Forecasting

More information

Estimation of cumulative distribution function with spline functions

Estimation of cumulative distribution function with spline functions INTERNATIONAL JOURNAL OF ECONOMICS AND STATISTICS Volume 5, 017 Estimation of cumulative distribution function with functions Akhlitdin Nizamitdinov, Aladdin Shamilov Abstract The estimation of the cumulative

More information

Nonparametric Small Area Estimation Using Penalized Spline Regression

Nonparametric Small Area Estimation Using Penalized Spline Regression Nonparametric Small Area Estimation Using Penalized Spline Regression 0verview Spline-based nonparametric regression Nonparametric small area estimation Prediction mean squared error Bootstrapping small

More information

Statistics 203: Introduction to Regression and Analysis of Variance Course review

Statistics 203: Introduction to Regression and Analysis of Variance Course review Statistics 203: Introduction to Regression and Analysis of Variance Course review Jonathan Taylor - p. 1/?? Today Review / overview of what we learned. - p. 2/?? General themes in regression models Specifying

More information

Bayesian covariate models in extreme value analysis

Bayesian covariate models in extreme value analysis Bayesian covariate models in extreme value analysis David Randell, Philip Jonathan, Kathryn Turnbull, Mathew Jones EVA 2015 Ann Arbor Copyright 2015 Shell Global Solutions (UK) EVA 2015 Ann Arbor June

More information

Bayesian density estimation from grouped continuous data

Bayesian density estimation from grouped continuous data Bayesian density estimation from grouped continuous data Philippe Lambert,a, Paul H.C. Eilers b,c a Université de Liège, Institut des sciences humaines et sociales, Méthodes quantitatives en sciences sociales,

More information

Penalized Splines, Mixed Models, and Recent Large-Sample Results

Penalized Splines, Mixed Models, and Recent Large-Sample Results Penalized Splines, Mixed Models, and Recent Large-Sample Results David Ruppert Operations Research & Information Engineering, Cornell University Feb 4, 2011 Collaborators Matt Wand, University of Wollongong

More information

Regularization in Cox Frailty Models

Regularization in Cox Frailty Models Regularization in Cox Frailty Models Andreas Groll 1, Trevor Hastie 2, Gerhard Tutz 3 1 Ludwig-Maximilians-Universität Munich, Department of Mathematics, Theresienstraße 39, 80333 Munich, Germany 2 University

More information

Chapter 3: Regression Methods for Trends

Chapter 3: Regression Methods for Trends Chapter 3: Regression Methods for Trends Time series exhibiting trends over time have a mean function that is some simple function (not necessarily constant) of time. The example random walk graph from

More information

Application of density estimation methods to quantal analysis

Application of density estimation methods to quantal analysis Application of density estimation methods to quantal analysis Koichi Yoshioka Tokyo Medical and Dental University Summary There has been controversy for the quantal nature of neurotransmission of mammalian

More information

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization /

Coordinate descent. Geoff Gordon & Ryan Tibshirani Optimization / Coordinate descent Geoff Gordon & Ryan Tibshirani Optimization 10-725 / 36-725 1 Adding to the toolbox, with stats and ML in mind We ve seen several general and useful minimization tools First-order methods

More information

STA216: Generalized Linear Models. Lecture 1. Review and Introduction

STA216: Generalized Linear Models. Lecture 1. Review and Introduction STA216: Generalized Linear Models Lecture 1. Review and Introduction Let y 1,..., y n denote n independent observations on a response Treat y i as a realization of a random variable Y i In the general

More information

Discussion of Maximization by Parts in Likelihood Inference

Discussion of Maximization by Parts in Likelihood Inference Discussion of Maximization by Parts in Likelihood Inference David Ruppert School of Operations Research & Industrial Engineering, 225 Rhodes Hall, Cornell University, Ithaca, NY 4853 email: dr24@cornell.edu

More information

Towards a Regression using Tensors

Towards a Regression using Tensors February 27, 2014 Outline Background 1 Background Linear Regression Tensorial Data Analysis 2 Definition Tensor Operation Tensor Decomposition 3 Model Attention Deficit Hyperactivity Disorder Data Analysis

More information

An Introduction to GAMs based on penalized regression splines. Simon Wood Mathematical Sciences, University of Bath, U.K.

An Introduction to GAMs based on penalized regression splines. Simon Wood Mathematical Sciences, University of Bath, U.K. An Introduction to GAMs based on penalied regression splines Simon Wood Mathematical Sciences, University of Bath, U.K. Generalied Additive Models (GAM) A GAM has a form something like: g{e(y i )} = η

More information

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis

Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Intensity Analysis of Spatial Point Patterns Geog 210C Introduction to Spatial Data Analysis Chris Funk Lecture 4 Spatial Point Patterns Definition Set of point locations with recorded events" within study

More information

Small Area Estimation for Skewed Georeferenced Data

Small Area Estimation for Skewed Georeferenced Data Small Area Estimation for Skewed Georeferenced Data E. Dreassi - A. Petrucci - E. Rocco Department of Statistics, Informatics, Applications "G. Parenti" University of Florence THE FIRST ASIAN ISI SATELLITE

More information

Multivariate Calibration with Robust Signal Regression

Multivariate Calibration with Robust Signal Regression Multivariate Calibration with Robust Signal Regression Bin Li and Brian Marx from Louisiana State University Somsubhra Chakraborty from Indian Institute of Technology Kharagpur David C Weindorf from Texas

More information

Variable Selection and Model Choice in Survival Models with Time-Varying Effects

Variable Selection and Model Choice in Survival Models with Time-Varying Effects Variable Selection and Model Choice in Survival Models with Time-Varying Effects Boosting Survival Models Benjamin Hofner 1 Department of Medical Informatics, Biometry and Epidemiology (IMBE) Friedrich-Alexander-Universität

More information

Estimating prediction error in mixed models

Estimating prediction error in mixed models Estimating prediction error in mixed models benjamin saefken, thomas kneib georg-august university goettingen sonja greven ludwig-maximilians-university munich 1 / 12 GLMM - Generalized linear mixed models

More information

Generalized Linear Models in R

Generalized Linear Models in R Generalized Linear Models in R NO ORDER Kenneth K. Lopiano, Garvesh Raskutti, Dan Yang last modified 28 4 2013 1 Outline 1. Background and preliminaries 2. Data manipulation and exercises 3. Data structures

More information

Log-Density Estimation with Application to Approximate Likelihood Inference

Log-Density Estimation with Application to Approximate Likelihood Inference Log-Density Estimation with Application to Approximate Likelihood Inference Martin Hazelton 1 Institute of Fundamental Sciences Massey University 19 November 2015 1 Email: m.hazelton@massey.ac.nz WWPMS,

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Advanced Methods for Data Analysis (36-402/36-608 Spring 2014 1 Generalized linear models 1.1 Introduction: two regressions So far we ve seen two canonical settings for regression.

More information

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives

Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives TR-No. 14-06, Hiroshima Statistical Research Group, 1 11 Illustration of the Varying Coefficient Model for Analyses the Tree Growth from the Age and Space Perspectives Mariko Yamamura 1, Keisuke Fukui

More information

Hani Mehrpouyan, California State University, Bakersfield. Signals and Systems

Hani Mehrpouyan, California State University, Bakersfield. Signals and Systems Hani Mehrpouyan, Department of Electrical and Computer Engineering, Lecture 26 (LU Factorization) May 30 th, 2013 The material in these lectures is partly taken from the books: Elementary Numerical Analysis,

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Lecture 3. Hypothesis testing. Goodness of Fit. Model diagnostics GLM (Spring, 2018) Lecture 3 1 / 34 Models Let M(X r ) be a model with design matrix X r (with r columns) r n

More information

ASYMPTOTICS FOR PENALIZED SPLINES IN ADDITIVE MODELS

ASYMPTOTICS FOR PENALIZED SPLINES IN ADDITIVE MODELS Mem. Gra. Sci. Eng. Shimane Univ. Series B: Mathematics 47 (2014), pp. 63 71 ASYMPTOTICS FOR PENALIZED SPLINES IN ADDITIVE MODELS TAKUMA YOSHIDA Communicated by Kanta Naito (Received: December 19, 2013)

More information

The OSCAR for Generalized Linear Models

The OSCAR for Generalized Linear Models Sebastian Petry & Gerhard Tutz The OSCAR for Generalized Linear Models Technical Report Number 112, 2011 Department of Statistics University of Munich http://www.stat.uni-muenchen.de The OSCAR for Generalized

More information

Functional SVD for Big Data

Functional SVD for Big Data Functional SVD for Big Data Pan Chao April 23, 2014 Pan Chao Functional SVD for Big Data April 23, 2014 1 / 24 Outline 1 One-Way Functional SVD a) Interpretation b) Robustness c) CV/GCV 2 Two-Way Problem

More information

Curve Fitting Re-visited, Bishop1.2.5

Curve Fitting Re-visited, Bishop1.2.5 Curve Fitting Re-visited, Bishop1.2.5 Maximum Likelihood Bishop 1.2.5 Model Likelihood differentiation p(t x, w, β) = Maximum Likelihood N N ( t n y(x n, w), β 1). (1.61) n=1 As we did in the case of the

More information

Unsupervised machine learning

Unsupervised machine learning Chapter 9 Unsupervised machine learning Unsupervised machine learning (a.k.a. cluster analysis) is a set of methods to assign objects into clusters under a predefined distance measure when class labels

More information

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty

Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Journal of Data Science 9(2011), 549-564 Selection of Smoothing Parameter for One-Step Sparse Estimates with L q Penalty Masaru Kanba and Kanta Naito Shimane University Abstract: This paper discusses the

More information

CSC2515 Assignment #2

CSC2515 Assignment #2 CSC2515 Assignment #2 Due: Nov.4, 2pm at the START of class Worth: 18% Late assignments not accepted. 1 Pseudo-Bayesian Linear Regression (3%) In this question you will dabble in Bayesian statistics and

More information

Smooth semi- and nonparametric Bayesian estimation of bivariate densities from bivariate histogram data

Smooth semi- and nonparametric Bayesian estimation of bivariate densities from bivariate histogram data Smooth semi- and nonparametric Bayesian estimation of bivariate densities from bivariate histogram data Philippe Lambert a,b a Institut des sciences humaines et sociales, Méthodes quantitatives en sciences

More information

Density Estimation: New Spline Approaches and a Partial Review

Density Estimation: New Spline Approaches and a Partial Review Density Estimation: New Spline Approaches and a Partial Review Martin B. Mächler Seminar für Statistik, ETH Zentrum CH-8092 Zürich, Switzerland e-mail: maechler@stat.math.ethz.ch CompStat, August 1996

More information

Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"

Kneib, Fahrmeir: Supplement to Structured additive regression for categorical space-time data: A mixed model approach Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach" Sonderforschungsbereich 386, Paper 43 (25) Online unter: http://epub.ub.uni-muenchen.de/

More information

Lattice Data. Tonglin Zhang. Spatial Statistics for Point and Lattice Data (Part III)

Lattice Data. Tonglin Zhang. Spatial Statistics for Point and Lattice Data (Part III) Title: Spatial Statistics for Point Processes and Lattice Data (Part III) Lattice Data Tonglin Zhang Outline Description Research Problems Global Clustering and Local Clusters Permutation Test Spatial

More information

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Ruppert A. EMPIRICAL ESTIMATE OF THE KERNEL MIXTURE Here we

More information

COS 424: Interacting with Data

COS 424: Interacting with Data COS 424: Interacting with Data Lecturer: Rob Schapire Lecture #14 Scribe: Zia Khan April 3, 2007 Recall from previous lecture that in regression we are trying to predict a real value given our data. Specically,

More information

Functional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods

Functional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods Université Catholique de Louvain Institut de Statistique, Biostatistique et Sciences Actuarielles Functional Estimation in Systems Defined by Differential Equation using Bayesian Smoothing Methods 19th

More information

Odds ratio estimation in Bernoulli smoothing spline analysis-ofvariance

Odds ratio estimation in Bernoulli smoothing spline analysis-ofvariance The Statistician (1997) 46, No. 1, pp. 49 56 Odds ratio estimation in Bernoulli smoothing spline analysis-ofvariance models By YUEDONG WANG{ University of Michigan, Ann Arbor, USA [Received June 1995.

More information

gamboostlss: boosting generalized additive models for location, scale and shape

gamboostlss: boosting generalized additive models for location, scale and shape gamboostlss: boosting generalized additive models for location, scale and shape Benjamin Hofner Joint work with Andreas Mayr, Nora Fenske, Thomas Kneib and Matthias Schmid Department of Medical Informatics,

More information

Simultaneous Confidence Bands for the Coefficient Function in Functional Regression

Simultaneous Confidence Bands for the Coefficient Function in Functional Regression University of Haifa From the SelectedWorks of Philip T. Reiss August 7, 2008 Simultaneous Confidence Bands for the Coefficient Function in Functional Regression Philip T. Reiss, New York University Available

More information

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown. Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)

More information

MAT300/500 Programming Project Spring 2019

MAT300/500 Programming Project Spring 2019 MAT300/500 Programming Project Spring 2019 Please submit all project parts on the Moodle page for MAT300 or MAT500. Due dates are listed on the syllabus and the Moodle site. You should include all neccessary

More information

Nonparametric Regression. Badr Missaoui

Nonparametric Regression. Badr Missaoui Badr Missaoui Outline Kernel and local polynomial regression. Penalized regression. We are given n pairs of observations (X 1, Y 1 ),...,(X n, Y n ) where Y i = r(x i ) + ε i, i = 1,..., n and r(x) = E(Y

More information

12 - Nonparametric Density Estimation

12 - Nonparametric Density Estimation ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6

More information

Bias-corrected AIC for selecting variables in Poisson regression models

Bias-corrected AIC for selecting variables in Poisson regression models Bias-corrected AIC for selecting variables in Poisson regression models Ken-ichi Kamo (a), Hirokazu Yanagihara (b) and Kenichi Satoh (c) (a) Corresponding author: Department of Liberal Arts and Sciences,

More information

Modelling trends in digit preference patterns

Modelling trends in digit preference patterns Appl. Statist. (2017) 66, Part 5, pp. 893 918 Modelling trends in digit preference patterns Carlo G. Camarda, Institut National d Études Démographiques, Paris, France Paul H. C. Eilers Erasmus University

More information

Array methods in statistics with applications to the modelling and forecasting of mortality. James Gavin Kirkby

Array methods in statistics with applications to the modelling and forecasting of mortality. James Gavin Kirkby Array methods in statistics with applications to the modelling and forecasting of mortality James Gavin Kirkby Submitted for the degree of Doctor of Philosophy Heriot-Watt University School of Mathematical

More information

BLIND SOURCE SEPARATION IN POST NONLINEAR MIXTURES

BLIND SOURCE SEPARATION IN POST NONLINEAR MIXTURES BLIND SURCE SEPARATIN IN PST NNLINEAR MIXTURES Sophie Achard, Dinh Tuan Pham Univ of Grenoble Laboratory of Modeling Computation, IMAG, CNRS BP 53X, 384 Grenoble Cedex, France SophieAchard@imagfr, DinhTuanPham@imagfr

More information

Recap. HW due Thursday by 5 pm Next HW coming on Thursday Logistic regression: Pr(G = k X) linear on the logit scale Linear discriminant analysis:

Recap. HW due Thursday by 5 pm Next HW coming on Thursday Logistic regression: Pr(G = k X) linear on the logit scale Linear discriminant analysis: 1 / 23 Recap HW due Thursday by 5 pm Next HW coming on Thursday Logistic regression: Pr(G = k X) linear on the logit scale Linear discriminant analysis: Pr(G = k X) Pr(X G = k)pr(g = k) Theory: LDA more

More information

Selection of Variables and Functional Forms in Multivariable Analysis: Current Issues and Future Directions

Selection of Variables and Functional Forms in Multivariable Analysis: Current Issues and Future Directions in Multivariable Analysis: Current Issues and Future Directions Frank E Harrell Jr Department of Biostatistics Vanderbilt University School of Medicine STRATOS Banff Alberta 2016-07-04 Fractional polynomials,

More information

mgcv: GAMs in R Simon Wood Mathematical Sciences, University of Bath, U.K.

mgcv: GAMs in R Simon Wood Mathematical Sciences, University of Bath, U.K. mgcv: GAMs in R Simon Wood Mathematical Sciences, University of Bath, U.K. mgcv, gamm4 mgcv is a package supplied with R for generalized additive modelling, including generalized additive mixed models.

More information

Exact Likelihood Ratio Tests for Penalized Splines

Exact Likelihood Ratio Tests for Penalized Splines Exact Likelihood Ratio Tests for Penalized Splines By CIPRIAN CRAINICEANU, DAVID RUPPERT, GERDA CLAESKENS, M.P. WAND Department of Biostatistics, Johns Hopkins University, 615 N. Wolfe Street, Baltimore,

More information

A short introduction to INLA and R-INLA

A short introduction to INLA and R-INLA A short introduction to INLA and R-INLA Integrated Nested Laplace Approximation Thomas Opitz, BioSP, INRA Avignon Workshop: Theory and practice of INLA and SPDE November 7, 2018 2/21 Plan for this talk

More information

Smoothing Age-Period-Cohort models with P -splines: a mixed model approach

Smoothing Age-Period-Cohort models with P -splines: a mixed model approach Smoothing Age-Period-Cohort models with P -splines: a mixed model approach Running headline: Smooth Age-Period-Cohort models I D Currie, Department of Actuarial Mathematics and Statistics, and the Maxwell

More information

Direct Learning: Linear Classification. Donglin Zeng, Department of Biostatistics, University of North Carolina

Direct Learning: Linear Classification. Donglin Zeng, Department of Biostatistics, University of North Carolina Direct Learning: Linear Classification Logistic regression models for classification problem We consider two class problem: Y {0, 1}. The Bayes rule for the classification is I(P(Y = 1 X = x) > 1/2) so

More information

LISA Short Course Series Generalized Linear Models (GLMs) & Categorical Data Analysis (CDA) in R. Liang (Sally) Shan Nov. 4, 2014

LISA Short Course Series Generalized Linear Models (GLMs) & Categorical Data Analysis (CDA) in R. Liang (Sally) Shan Nov. 4, 2014 LISA Short Course Series Generalized Linear Models (GLMs) & Categorical Data Analysis (CDA) in R Liang (Sally) Shan Nov. 4, 2014 L Laboratory for Interdisciplinary Statistical Analysis LISA helps VT researchers

More information

Computational Methods CMSC/AMSC/MAPL 460. Vectors, Matrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. Vectors, Matrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science Computational ethods CSC/ASC/APL 460 Vectors, atrices, Linear Systems, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science Class Outline uch of scientific computation involves solution of linear

More information

An Introduction to Functional Data Analysis

An Introduction to Functional Data Analysis An Introduction to Functional Data Analysis Chongzhi Di Fred Hutchinson Cancer Research Center cdi@fredhutch.org Biotat 578A: Special Topics in (Genetic) Epidemiology November 10, 2015 Textbook Ramsay

More information

Model comparison and selection

Model comparison and selection BS2 Statistical Inference, Lectures 9 and 10, Hilary Term 2008 March 2, 2008 Hypothesis testing Consider two alternative models M 1 = {f (x; θ), θ Θ 1 } and M 2 = {f (x; θ), θ Θ 2 } for a sample (X = x)

More information

Effective Computation for Odds Ratio Estimation in Nonparametric Logistic Regression

Effective Computation for Odds Ratio Estimation in Nonparametric Logistic Regression Communications of the Korean Statistical Society 2009, Vol. 16, No. 4, 713 722 Effective Computation for Odds Ratio Estimation in Nonparametric Logistic Regression Young-Ju Kim 1,a a Department of Information

More information

Outline. Mixed models in R using the lme4 package Part 5: Generalized linear mixed models. Parts of LMMs carried over to GLMMs

Outline. Mixed models in R using the lme4 package Part 5: Generalized linear mixed models. Parts of LMMs carried over to GLMMs Outline Mixed models in R using the lme4 package Part 5: Generalized linear mixed models Douglas Bates University of Wisconsin - Madison and R Development Core Team UseR!2009,

More information

Generalized Linear Models. Last time: Background & motivation for moving beyond linear

Generalized Linear Models. Last time: Background & motivation for moving beyond linear Generalized Linear Models Last time: Background & motivation for moving beyond linear regression - non-normal/non-linear cases, binary, categorical data Today s class: 1. Examples of count and ordered

More information

Regularization Methods for Additive Models

Regularization Methods for Additive Models Regularization Methods for Additive Models Marta Avalos, Yves Grandvalet, and Christophe Ambroise HEUDIASYC Laboratory UMR CNRS 6599 Compiègne University of Technology BP 20529 / 60205 Compiègne, France

More information

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models

Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Jonathan Taylor - p. 1/15 Today s class Bias-Variance tradeoff. Penalized regression. Cross-validation. - p. 2/15 Bias-variance

More information

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS Parametric Distributions Basic building blocks: Need to determine given Representation: or? Recall Curve Fitting Binary Variables

More information

Reparametrization of COM-Poisson Regression Models with Applications in the Analysis of Experimental Count Data

Reparametrization of COM-Poisson Regression Models with Applications in the Analysis of Experimental Count Data Reparametrization of COM-Poisson Regression Models with Applications in the Analysis of Experimental Count Data Eduardo Elias Ribeiro Junior 1 2 Walmes Marques Zeviani 1 Wagner Hugo Bonat 1 Clarice Garcia

More information

Analysis Methods for Supersaturated Design: Some Comparisons

Analysis Methods for Supersaturated Design: Some Comparisons Journal of Data Science 1(2003), 249-260 Analysis Methods for Supersaturated Design: Some Comparisons Runze Li 1 and Dennis K. J. Lin 2 The Pennsylvania State University Abstract: Supersaturated designs

More information

Aggregated cancer incidence data: spatial models

Aggregated cancer incidence data: spatial models Aggregated cancer incidence data: spatial models 5 ième Forum du Cancéropôle Grand-est - November 2, 2011 Erik A. Sauleau Department of Biostatistics - Faculty of Medicine University of Strasbourg ea.sauleau@unistra.fr

More information

Finite Difference Methods (FDMs) 1

Finite Difference Methods (FDMs) 1 Finite Difference Methods (FDMs) 1 1 st - order Approxima9on Recall Taylor series expansion: Forward difference: Backward difference: Central difference: 2 nd - order Approxima9on Forward difference: Backward

More information

Introduction to Statistical modeling: handout for Math 489/583

Introduction to Statistical modeling: handout for Math 489/583 Introduction to Statistical modeling: handout for Math 489/583 Statistical modeling occurs when we are trying to model some data using statistical tools. From the start, we recognize that no model is perfect

More information

Mixed models in R using the lme4 package Part 7: Generalized linear mixed models

Mixed models in R using the lme4 package Part 7: Generalized linear mixed models Mixed models in R using the lme4 package Part 7: Generalized linear mixed models Douglas Bates University of Wisconsin - Madison and R Development Core Team University of

More information

Spatially Adaptive Smoothing Splines

Spatially Adaptive Smoothing Splines Spatially Adaptive Smoothing Splines Paul Speckman University of Missouri-Columbia speckman@statmissouriedu September 11, 23 Banff 9/7/3 Ordinary Simple Spline Smoothing Observe y i = f(t i ) + ε i, =

More information

Multi-resolution models for large data sets

Multi-resolution models for large data sets Multi-resolution models for large data sets Douglas Nychka, National Center for Atmospheric Research National Science Foundation NORDSTAT, Umeå, June, 2012 Credits Steve Sain, NCAR Tia LeRud, UC Davis

More information

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo

Sparse Nonparametric Density Estimation in High Dimensions Using the Rodeo Outline in High Dimensions Using the Rodeo Han Liu 1,2 John Lafferty 2,3 Larry Wasserman 1,2 1 Statistics Department, 2 Machine Learning Department, 3 Computer Science Department, Carnegie Mellon University

More information

Matrices. Math 240 Calculus III. Wednesday, July 10, Summer 2013, Session II. Matrices. Math 240. Definitions and Notation.

Matrices. Math 240 Calculus III. Wednesday, July 10, Summer 2013, Session II. Matrices. Math 240. Definitions and Notation. function Matrices Calculus III Summer 2013, Session II Wednesday, July 10, 2013 Agenda function 1. 2. function function Definition An m n matrix is a rectangular array of numbers arranged in m horizontal

More information

Computational Methods CMSC/AMSC/MAPL 460. Linear Systems, Matrices, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science

Computational Methods CMSC/AMSC/MAPL 460. Linear Systems, Matrices, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science Computational ethods CSC/ASC/APL 460 Linear Systems, atrices, LU Decomposition, Ramani Duraiswami, Dept. of Computer Science Class Outline uch of scientific computation involves solution of linear equations

More information

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference

ECE G: Special Topics in Signal Processing: Sparsity, Structure, and Inference ECE 18-898G: Special Topics in Signal Processing: Sparsity, Structure, and Inference Sparse Recovery using L1 minimization - algorithms Yuejie Chi Department of Electrical and Computer Engineering Spring

More information

Spatial Process Estimates as Smoothers: A Review

Spatial Process Estimates as Smoothers: A Review Spatial Process Estimates as Smoothers: A Review Soutir Bandyopadhyay 1 Basic Model The observational model considered here has the form Y i = f(x i ) + ɛ i, for 1 i n. (1.1) where Y i is the observed

More information

Smoothing Age-Period-Cohort models with P -splines: a mixed model approach

Smoothing Age-Period-Cohort models with P -splines: a mixed model approach Smoothing Age-Period-Cohort models with P -splines: a mixed model approach Running headline: Smooth Age-Period-Cohort models I D Currie, Department of Actuarial Mathematics and Statistics, and the Maxwell

More information

Math 533 Extra Hour Material

Math 533 Extra Hour Material Math 533 Extra Hour Material A Justification for Regression The Justification for Regression It is well-known that if we want to predict a random quantity Y using some quantity m according to a mean-squared

More information

Frailty models. November 30, Programme

Frailty models. November 30, Programme Frailty models November 30, 2004 9.30h-10.30h Programme Veronique Rondeau (INSERM, Bordeaux) Title: A three-level nested frailty model for clustered survival data using a penalized likelihood 10.30h-11.00h

More information

Machine Learning: Assignment 1

Machine Learning: Assignment 1 10-701 Machine Learning: Assignment 1 Due on Februrary 0, 014 at 1 noon Barnabas Poczos, Aarti Singh Instructions: Failure to follow these directions may result in loss of points. Your solutions for this

More information

Terminology for Statistical Data

Terminology for Statistical Data Terminology for Statistical Data variables - features - attributes observations - cases (consist of multiple values) In a standard data matrix, variables or features correspond to columns observations

More information

Appendix A. Numeric example of Dimick Staiger Estimator and comparison between Dimick-Staiger Estimator and Hierarchical Poisson Estimator

Appendix A. Numeric example of Dimick Staiger Estimator and comparison between Dimick-Staiger Estimator and Hierarchical Poisson Estimator Appendix A. Numeric example of Dimick Staiger Estimator and comparison between Dimick-Staiger Estimator and Hierarchical Poisson Estimator As described in the manuscript, the Dimick-Staiger (DS) estimator

More information

Chapter 7: Model Assessment and Selection

Chapter 7: Model Assessment and Selection Chapter 7: Model Assessment and Selection DD3364 April 20, 2012 Introduction Regression: Review of our problem Have target variable Y to estimate from a vector of inputs X. A prediction model ˆf(X) has

More information

The Conjugate Gradient Method

The Conjugate Gradient Method The Conjugate Gradient Method Classical Iterations We have a problem, We assume that the matrix comes from a discretization of a PDE. The best and most popular model problem is, The matrix will be as large

More information

Calibrating Environmental Engineering Models and Uncertainty Analysis

Calibrating Environmental Engineering Models and Uncertainty Analysis Models and Cornell University Oct 14, 2008 Project Team Christine Shoemaker, co-pi, Professor of Civil and works in applied optimization, co-pi Nikolai Blizniouk, PhD student in Operations Research now

More information

LASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape

LASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape LASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape Nikolaus Umlauf https://eeecon.uibk.ac.at/~umlauf/ Overview Joint work with Andreas Groll, Julien Hambuckers

More information