POLYNOMIAL SINGLE COMPONENT ANALYSIS USING MULTIVARIATE MOMENTS

Size: px
Start display at page:

Download "POLYNOMIAL SINGLE COMPONENT ANALYSIS USING MULTIVARIATE MOMENTS"

Transcription

1 POLYNOMIAL SINGLE COMPONENT ANALYSIS USING MULTIVARIATE MOMENTS JAN DE LEEUW Abstract. Polynomial single component analysis is defined. Algorithms, based on multidimensional array approimation methods, to fit a model with a single common factor to observed multivariate moments are described and applied to a psychometric eample. Software in R is included. Date: Tuesday 0 th July, 0 7h 7min Typeset in Lucida Bright. 000 Mathematics Subject Classification. 6H. Key words and phrases. Component Analysis, Polynomials, Array Decmposition.

2 JAN DE LEEUW. Model Suppose y j are random variables, with j =,, m. We use the convention of underlining random variables [Hemelrijk, 966]. In polynomial single component analysis we assume there eist a random variable such that () y j = P j (), where the P j are univariate polynomials of degree s. Thus s () P j () = β jp p. p=0 Closely related previous work on the polynomial factor analysis model, which differs from () because it includes unique factors, is in Mooijaart and Bentler [986], who also discuss earlier work by Gibson and McDonald. Mooijaart and Bentler [986] use higher order product moments, specifically third order moments, and the special case of one-dimensional polynomial factor analysis is worked out and applied to a real eample. Since Mooijaart and Bentler [986] the emphasis has shifted away from higher order statistics to using instrumental variables and other more specialized techniques, and from component and factor analysis to general nonlinear structural equation models. Good historical overviews of these more recent developments are in Yalcin and Amemiya [00] and Mooijaart and Bentler [00]. In this paper we go back to the simple single component model, we use higher-order statistics, specifically multivariate moments, and we use decomposition methods for multidimensional arrays to fit the models to sample moments. Thus, compared to Mooijaart and Bentler [986], we develop the case of non-normal factors, provide algorithms and software to deal with moments of arbitrary order, and develop algorithms and software based on fitting multilinear

3 POLYNOMIAL SINGLE COMPONENT ANALYSIS array decompositions. All software is written in R [R Development Core Team, 0]. In a companion paper [De Leeuw, 0] we have worked out the equations in terms of cumulants, instead of moments. Cumulants are useful for laying the groundwork for generalizations to approimations with multiple components, and to factor analysis models. In the special case of single-factor component analysis, however, working with moments is much simpler.. Product Moments From the multilinearity property of moments for any r variables, not necessarily distinct, we have s s () E(P j () P jr ()) = β j p β jr p r c p p r, p =0 p r =0 where () c p p r = µ p + +p r, and where we write µ t for the t th raw moment of. The product moments of order r of the polynomials are a supersymmetric multilinear function of the elements of B = {β jp }, for a given kernel C = {c p p r }. Supper-symmetry means that the array elements are invariant under all permutations of the subscript. Moreover the kernel must have the moment structure (). The s r elements of the kernel array are simple linear functions of the first sr moments of. In fact we can write () C = sr t= µ t I t, where the I t are super-symmetric orthogonal binary arrays with {I t } p p r = only if p + + p r = t.

4 JAN DE LEEUW Of course not all vectors are possible vectors of the first moments of a random variable. Defining all possible vectors of moments is the truncated version of the classical Moment Problem, which requires that the Hankel matri formed from the moments is positive semi-definite. See, for eample, Curto and Fialkow [99] for the details. Instead of dealing with the complicated moment inequalities we can use the parametrization µ t = t df() and think of the kernel as a function of the distribution function F. By Mulholland and Rogers [98] we can limit ourselves, without loss of generality, to step functions with at most sr + steps. When dealing with multivariate moments it is convenient to introduce an additional variable y 0 which is a.s. equal to one and a corresponding polynomial P 0 which has β 0p = 0 for p =,, s. If we include P 0 in the epectations (), then the moment array includes all moments up to order r, i.e. also the moments of order,, r.. Loss Function Given the result () it becomes straightforward to generalize the methods for linear component analysis in De Leeuw and Kukuyeva [0] to the polynomial case. We first compute an array of moments A (r ) = {a (r ) j j r } of order up to r from a given data matri. This can be done with the code in segment. Insert Code Segment about here The data is a super-symmetric moment array of order (m + ) r. The kernel is a super-symmetric array of order s r, and B is an (m + ) s matri of loadings. The first row of B is zero, ecept for its first element.

5 POLYNOMIAL SINGLE COMPONENT ANALYSIS We then define the least squares loss function by (6) σ (B, C) = j J j r J r w j j r (a j j r s p =0 and minimize it over both B and C. s p r =0 β j p β jr p r c p p r ). The w j j r are given non-negative weights. They can be used to give lower or higher weights to the moments of various orders, or even to incorporate information about the standard errors of the moments. Note that the loss function can be written in matri form as (7) σ (B, C) = {vec(a) (B } {{ B} )vec(c)} V {vec(a) (B } {{ B} )vec(c))}, r times r times where V = diag(vec(w )). In actual computation and reporting we normalize the loss function to its root-mean-square form σ (B, C) (8) σ (B, C) = tr(v ). For the fitting of C three different cases are of interest... Fied Kernel. The problem is relatively simple if the kernel C is completely known. This requires, for eample, the additional assumption that has a standard normal distribution. The code in segment can be used to compute moments up to order r, including µ 0 =, of the standard normal distribution. Insert Code Segment about here The code in can be used to put moments into the kernel array, using the formula (). Insert Code Segment about here

6 6 JAN DE LEEUW In the case of a fied kernel the loss function only needs to be minimized over B... Free Kernel. In the second special case we ignore the fact that the kernel array C = {c p p r } is, according to (), an array formed of moments of the same variable. Thus we ignore the structure imposed by (). We merely imposing super-symmetry on C. We now minimize the loss function over both B and super-symmetric C. Ignoring the structure of the kernel array is convenient, but it does imply we will not be able to distinguish a linear model with s + correlated components from a polynomial model of degree s. This is a major disadvantage of the method. Nevertheless, if using higher order moments forces identification, then we will be able to estimate factor loadings (or polynomial coefficients) consistently, up to a linear transformation... Structured Kernel. This is the more complicated case, and in several important ways the most interesting one. We minimize over B and over all C with the structure (). We minimize loss over B and over these vectors of moments. There are two variations. In the first we minimize over all vectors µ, without imposing the condition that the vector consists of moments of a single variable. In the second we minimize over all distributions F, which determine the moments. In fact, it usually makes more sense to minimize over a finite-dimensional conve set of distribution functions, for eample the set of all conve combinations of B-splines of a given order with a given set of knots. In this case we can write B + (9) µ t = π b t df b (). b= Collecting the moments of the basis distributions in a matri M allows us to write µ = Mπ.

7 POLYNOMIAL SINGLE COMPONENT ANALYSIS 7 The code in segment computes the basis I t of indicators for the kernel used in formula (). It is wasteful to compute and store the indicators, but for small eamples it will be both feasible and computationally efficient. Insert Code Segment about here The code in segment uses Neuman [98, Proposition.] to compute the moments of B-splines for any knot sequence to any order. Insert Code Segment about here. Algorithms Various segments of code needed for our minimizations, written in R, to minimize (6) are given in De Leeuw [008, 0]; De Leeuw and Kukuyeva [0]. We adopt and etend them to deal with our form of component analysis... Initial Solution. Because of the clear possibility of local minima we want to start with a good initial configuration. For the kernel we always use the one based on moments of the standard normal. For the loadings we approimate the first moments y and second momentss by an equation the form y 0 c b. y S b B c C 0 B The product on the right works out to b + c B b + Bc bb + Bcb + bc B + BCB Now choose b = y Bc, so that we fit the means eactly. Then the fit becomes y y yy + B(C cc )B,

8 8 JAN DE LEEUW and we find B by using unweighted least squares to approimate the covariance S yy by B(C cc )B. This is a simple eigenvalueeigenvector problem... Fied Kernel. To minimize over B we use cyclic coordinate descent (CCD), i.e. we change one coordinate of B at a time while keeping all other fied at their current values. One iteration to update B is a cycle in which all coordinates are changed. CCD is discussed, for eample, in Zangwill [969, Page -], and in a similar contet it has been used in ALSCAL [Takane et al., 977]. It is currently popular in machine learning. See for some eamples and references. If we deal with cumulants of order four, then changing one coordinate makes the loss function a polynomial of order eight of that coordinate. If the cumulants are of order three, the polynomial is of order si, and so on. The polynomial is minimized by using the optimize() function in R. There is no guarantee CCD will find the global minimum. As in ALSCAL, it is possible, at least in principle, to use global optimization methods for multivariate (non-negative) polynomials [?]. We have not tried this, and for larger problems the computational requirements are quite daunting... Free Kernel. Fitting a free kernel is a simple modification of the basic algorithm. After a CCD cycle we update the kernel for given coordinates, and this is a simple linear least squares problem. The resulting Alternating Least Squares (ALS) algorithm, which alternates CCD cycles and kernel updates, is a block relaation algorithm in the sense of De Leeuw [99].

9 POLYNOMIAL SINGLE COMPONENT ANALYSIS 9 Note that from (7) we see, using Moore-Penrose inverses, that the optimal kernel for fied B is (0) vec(c) = (B } + {{ B + })vec(a). r times.. Structured Kernel. From () () (B } {{ B} )vec(c) = r times sr t= µ t (B } {{ B} )vec(i t ) r times Collecting the vectors (B } {{ B} )vec(i t ) in the columns of a matri r times H allows us to write the loss function as () σ (B, µ) = (vec(a) Hµ) V (vec(a) Hµ), while using (9) gives () σ (B, π) = (vec(a) HMπ) V (vec(a) HMπ). Thus the two variations for a structured kernel are to minimize () over all µ and to minimize () over all π in the unit simple. Both are linear least squares problems, in the second case with simple linear inequality and equality constraints. We solve the quadratic programming problem with the Frank-Wolfe linearization algorithm using the optimal step-size. The code is in segment. Insert Code Segment about here. Real Eample We use the YouthGratitude data from the R package psychotools [Zeileis et al., 0]. They are described in detail in the article by Froh et al. [0]. The si seven-point Likert scale variables of the GQ-6 scale are used, with responses from 0 students aged 0-9 years. In Table A of Froh et al. [0] we see that a linear factor analysis gives loadings between.8 and.9 for items and, between.6 and.7 for items,, and, and between. and. for

10 0 JAN DE LEEUW item 6. On the basis of these results the authors eliminate item 6 from further analysis, and we will do the same thing. The data are generate by the following code. library (psychotools) data ("YouthGratitude") gq6 <- as.matri (YouthGratitude[,:8]) We then load the required packages and source the required files. require (abind) require (apl) require (polynom) source ("tuckerccd.r") 6 source ("kernelinitial.r") 7 source ("kernelupdate.r") 8 source ("tuckermore.r") 9 source ("tuckerfinal.r") 0 source ("tuckerkernel.r") source ("tuckerau.r") source ("tuckerbasis.r") source ("tuckerplot.r") source ("utilities.r") source ("simplels.r") For the runs we set the iteration precision, the decrease of successive loss function values, to and we set the maimum number of outer iterations to 00. The number of inner iterations, which are the coordinate descent cycles to fit B, is set at. If we use a miture of B-spline densities for the moments of the distribution of we use equally spaced knots at the positive and negative integers from -6 to +6 and choose order (i.e. we use piecewise quadratic polynomials differentiable at the knots).

11 POLYNOMIAL SINGLE COMPONENT ANALYSIS The code below does 8 analyses, for each of the four kernel types, using moments of order r =, and polynomial degrees s =,,. loss <- array (0, c(,, )) itel <- array (0, c(,, )) R <- : S <- : 6 K <- : 7 8 for (r in R) { 9 for (s in S) { 0 for (k in K) { kertype <- switch (k, "fied", "free", " moment", "cdf") vname <- paste (kertype, r, s, sep = "") pname <- paste (vname, ".pdf", sep = "") assign (paste (kertype, r, s, sep = ""), tuckerccd (gq6, r, s, kertype = kertype )) llist <- eval (as.name (vname)) 6 pdf (pname) 7 plotpol (llist $ b, vname) 8 dev.off () 9 loss [s, r -, k] <- llist $ tloss 0 itel [s, r -, k] <- llist $ otel } } } dump (c("loss", "itel"), file = "results.r" ).. Results for Fied Normal Kernel.

12 JAN DE LEEUW 6. Etensions and Elaborations 6.. Weights and Efficiency. Adding weights W = {w j j r } to (6) is straightforward. It is more complicated to use the standard errors of the multivariate cumulants to improve (asymptotically distribution free) efficiency of the estimates. Again we will rely on (??). 6.. Constraints on Loadings. Adding linear constraints (some elements are zero, some elements are equal) on B is again straightforward. This makes it possible to use different polynomial degrees for different variables. We can also impose ordinal constraints, for instance that polynomials are monotonic. 6.. Consistency. Sample cumulants converge in probability to population cumulants. Minimizers of the loss function (or stationary points of the algorithm) are, under regularity conditions, continuous functions of the sample cumulants. Thus, under additional identifiability conditions, the estimates computed by the algorithm are consistent estimates of the population parameters. 6.. Multiple Common Factors. Additional work is needed to adapt this approach to structures with more than one common factor. The common part of each variable will be a multivariate polynomial of the common factors, i.e. a symmetric multilinear form. It is consequently still true that the multivariate polynomial is a linear combination of powers and products of powers, and thus the basic results (??) and (??) continue to apply. 6.. Non-additive Unique Factors. The additive combination rule in () is not as compelling as in the linear case. This is also the reason we have not paid much attention to factor analysis, as opposed to component analysis, in this paper. The more general decomposition y j = P j (, u j ) becomes interesting, but this is more naturally handled by the etension to mutiple common factors. For

13 POLYNOMIAL SINGLE COMPONENT ANALYSIS this case (??) no longer applies, but we can continue to rely on (??). This more general description, together with work on more than one factor, will become a topic for further research Convergence Accelleration. In further work we plan to use the methods discussed in De Leeuw [008] to speed up convergence. It must be emphasized, however, that there are many more parts of the algorithm that can be accelerated in various ways.

14 JAN DE LEEUW References R.E. Curto and L.A. Fialkow. Recursiveness, Positivity, and Truncated Moment Problems. Houston Journal of Mathematics, 7: 60 6, 99. J. De Leeuw. Block Relaation Algorithms in Statistics. In H.H. Bock, W. Lenski, and M.M. Richter, editors, Information Systems and Data Analysis, pages 08, Berlin, 99. Springer Verlag. URL chapters/deleeuw_c_9c.pdf. J. De Leeuw. Polynomial etrapolation to accelerate fied point algorithms. Preprint Series, UCLA Department of Statistics, Los Angeles, CA, 008. URL ~deleeuw/janspubs/008/reports/deleeuw_r_08i.pdf. J. De Leeuw. The Multiway Package. Preprint Series, UCLA Department of Statistics, Los Angeles, CA, 008. URL reports/deleeuw_r_08j.pdf. J. De Leeuw. Multivariate Cumulants in R. Unpublished, 0. URL notes/deleeuw_u_a.pdf. J. De Leeuw. Single Factor Polynomial Component Analysis. Unpublished, 0. URL janspubs/0/notes/deleeuw_u_a.pdf. J. De Leeuw and I. Kukuyeva. Component Analysis Using Multivariate Cumulants. Unpublished, 0. URL notes/deleeuw_kukuyeva_u_.pdf. J.J. Froh, J. Fan, R.A. Emmons, G. Bono, E. S. Huebner, and P. Watkins. Measuring Gratitude in Youth: Assessing the Psychometric Properties of Adult Gratitude Scales in Children and Adolescents. Psychological Assessment, :, 0. J. Hemelrijk. Underlining Random Variables. Statistica Neerlandica, 0: 7, 966.

15 POLYNOMIAL SINGLE COMPONENT ANALYSIS A. Mooijaart and P. Bentler. Random Polynomial Factor Analysis. In E. Diday et al., editor, Data Analysis and Informatics, volume IV, pages 0, 986. A. Mooijaart and P.M. Bentler. An Alternative Approach for Nonlinear Latent Variable Models. Structural Equation Modeling, 7: 7 7, 00. H.P. Mulholland and C.A. Rogers. Representation Theorems for Distribution Functions. Proceedings of the London Mathematical Society, 8:77, 98. E. Neuman. Moments and Fourier Transforms of B-splines. Journal of Computational and Applied Mathematics, 7: 6, 98. R Development Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, 0. URL ISBN Y. Takane, F.W. Young, and J. De Leeuw. Nonmetric individual differences in multidimensional scaling: An alternating least squares method with optimal scaling features. Psychometrika, :7 67, 977. URL janspubs/977/articles/takane_young_deleeuw_a_77. pdf. I. Yalcin and Y. Amemiya. Nonlinear Factor Analysis as a Statistical Method. Statistical Science, 6:7 9, 00. W. I. Zangwill. Nonlinear Programming: a Unified Approach. Prentice-Hall, Englewood-Cliffs, N.J., 969. A. Zeileis, C. Strobl, and F. Wickelmaier. psychotools: Infrastructure for Psychometric Modeling, 0. URL R-project.org/package=psychotools. R package version 0.-.

16 6 JAN DE LEEUW Appendi A. Code Code Segment Raw Moments up to Order p rawmomentsuptop <- function (, p = ) { n <- nrow () m <- ncol () if (p == ) { return (c (, apply (,, mean))) 6 } 7 y <- array (0, rep (m +, p)) 8 for (i in : n) { 9 i <- c (, [i, ]) 0 z <- i for (s in :p) { z <- outer (z, i) } y <- y + z } 6 return (y / n) 7 }

17 POLYNOMIAL SINGLE COMPONENT ANALYSIS 7 Code Segment Moments of Standard Normal normalevenmoment <- function (n) { if ((n %% ) == ) { stop ("even moments only") } return (doublefactorial (n - )) 6 } 7 8 doublefactorial <- function (n) { 9 if ((n %% ) == 0) { 0 stop ("odd argument only") } k <- (n + ) / return (prod ( * ( : k) - )) } 6 normalmoments <- function (n) { 7 m <- rep (0, n) 8 ind <- which((( : n) %% ) == 0) 9 m [ind] <- sapply (ind, normalevenmoment) 0 return (c (, m)) }

18 8 JAN DE LEEUW Code Segment Kernel from Moments require("apl") makemomentkernel <- function (s, r, mu) { if (length (mu)!= ((s * r) + )) { stop ("vector of moments wrong length") 6 } 7 c <- array(0, rep (s +, r)) 8 sr <- (s + ) ^ r 9 for (i in : sr) { 0 l <- aplencode (i, rep (s +, r)) - k <- sum (l) c [i] <- mu [k + ] } return (c) }

19 POLYNOMIAL SINGLE COMPONENT ANALYSIS 9 Code Segment Kernel Indicators makeindicatorbasismatri <- function (s, r) { ii <- replist (array (0, rep (s +, r)), (s * r) + ) sr <- (s + ) ^ r for (i in ( : sr)) { l <- aplencode (i, rep (s +, r)) - 6 k <- sum (l) 7 ii [[k + ]] [i] <- 8 } 9 return (ii) 0 }

20 0 JAN DE LEEUW Code Segment Moments of B-splines makebsplinemoments <- function (lup, t) { kup <- length (t) - m <- matri (0, kup, lup + ) for (l in (0 : lup)) { s <- (t [] ^ (l + )) - (t [] ^ (l + )) 6 s <- (l + ) * (t [] - t []) 7 m [, l + ] <- s / s 8 for (k in ( : kup)) { 9 ss <- 0 0 for (j in (0 : l)) { s <- t [k + ] ^ (l - j) s <- choose (k + j -, j) s <- m [k -, j + ] ss <- ss + s * s * s } 6 m [k, l + ] <- ss / choose (k + l, k) 7 } 8 } 9 return (m [kup, ]) 0 } makebsplinemomentmatri <- function (lup, t, order) { k <- length (t) - order m <- matri (0, k, lup + ) for (i in ( : k)) { 6 m[i, ] <- makebsplinemoments (lup, t [i : (i + order)]) 7 } 8 return (m) 9 }

21 POLYNOMIAL SINGLE COMPONENT ANALYSIS Code Segment 6 CCD Algorithm for Fied/Free Kernel tuckerccd <- function (data, r, s, w = array (, dim (a)), kertype = "fied", order =, knots = -6:6, inma =, ineps = e-, outma = 00, outeps = e-, verbose = ) { a <- rawmomentsuptop (data, r) vassign (i, m, p, q, k, values = kernelinitial (s, r, k, kertype, knots, order)) vassign (b, otel, values = list (tuckerinitial (a, k), )) oloss <- tuckervalue (a, k, w, b) 6 repeat { 7 vassign (itel, bold, values = list (, b)) 8 repeat { 9 bnew <- oneccdcycle (a, k, w, bold) 0 chng <- ma (abs (bnew - bold)) if (verbose > ) cat ("Ineration: ", formatc (itel,, ), "IneChange: ", formatc (chng, 6, 0, " f"), "\n") if ( (chng < ineps) (itel == inma)) break () 6 vassign (bold, itel, values = list (bnew, itel + )) 7 } 8 b <- bnew 9 loss <- tuckervalue (a, k, w, b) 0 vassign (k, p, q, values = kernelupdate (a, k, i, m, p, w, b, kertype)) nloss <- tuckervalue (a, k, w, b) if (verbose >= ) cat ("Iteration:", formatc (otel,, ), "OLoss: ", formatc (oloss, 6, 0, "f"), "BLoss: ", formatc (loss, 6, 0, "f"), 6 "Kloss: ", formatc (nloss, 6, 0, "f"), "\ n") 7 if (((oloss - nloss) < outeps) (otel == outma)) 8 break () 9 vassign (oloss, otel, values = list (nloss, otel + )) 0 } return (list(b = b, otel = otel, tloss = nloss, k = k, p = p, q = q)) }

22 JAN DE LEEUW Code Segment 7 Tucker Subroutines tuckervalue <- function (a, k, w, b) { r <- length (dim (a)) res <- as.vector (a) - arrkronecker (replist (b, r)) % *% as.vector (k) return (sqrt (sum (as.vector (w) * res * res) / sum (w) } 6 )) 7 tuckerinitial <- function (a, k) { 8 r <- length (dim (a)) 9 s <- dim (a)[] 0 t <- dim (k) [] h <- asub (a, c (replist ( : s, ), replist (, r - ) )) m <- (h - outer (h [, ], h[, ])) [-, -] e <- eigen (m) v <- asub (k, c (replist ( : t, ), replist (, r - ) )) p <- nrow (v) - 6 u <- (v - outer (v [, ], v[, ])) [-, -] 7 if (p == ) { 8 <- as.matri (e $ vectors [, ]) %*% sqrt (e $ 9 } else { values []) 0 <- e $ vectors [, : p] %*% diag (sqrt (e $ } values [ : p])) bb <- %*% solve (t (chol (u))) bv <- h [-, ] - bb %*% v [-, ] return (rbind (c (, rep (0, t - )), cbind (bv, bb))) }

23 POLYNOMIAL SINGLE COMPONENT ANALYSIS Code Segment 8 More Tucker Subroutines oneccdcycle <- function (a, k, w, b) { p <- dim (k) [] m <- dim (a) [] f <- function (, j, s) tuckervalue (a, k, w, addelement (b, j, s, )) loss <- tuckervalue (a, k, w, b) 6 for (j in : m) { 7 for (s in : p) { 8 oo <- optimize (f, c(-, ), j, s) 9 ls <- oo $ objective 0 ip <- oo $ minimum if (ls > loss) { net () } loss <- ls b <- addelement (b, j, s, ip) 6 } 7 } 8 return (b) 9 } 0 addelement <- function (, i, j, s) { [i, j] <- [i, j] + s return () }

24 JAN DE LEEUW Code Segment 9 Tucker Auilaries arrkronecker <- function (, fun=" * ") { nmat <- length () if (nmat == 0) { stop("empty argument in arrkronecker") } 6 res <- [[]] 7 if (length () == ) { 8 return (res) 9 } 0 for (i in : nmat) { res <- kronecker (res, [[i]], fun) } return (res) } 6 replist <- function (, n) { 7 z <- list() 8 if (n == 0) { 9 return (z) 0 } for (i in : n) z <- c (z, list ()) return (z) } 6 vassign <- function(..., values, envir=parent.frame()) { 7 vars <- as.character(substitute(...())) 8 values <- rep(values, length.out=length(vars)) 9 for(i in seq_along(vars)) { 0 assign(vars[[i]], values[[i]], envir) } }

25 POLYNOMIAL SINGLE COMPONENT ANALYSIS Code Segment 0 Plotting Polynomials plotpol <- function (, title) { m <- nrow () s <- seq (-.,., length = 00) h <- seq (-.,., length = 0) v <- matri (0, m -, 00) 6 u <- matri (0, m -, 0) 7 cols <- rainbow (m - ) 8 for (i in : (m - )) { 9 v[i, ] <- predict (polynomial ([i +, ]), s) 0 u[i, ] <- predict (polynomial ([i +, ]), h) } yma <- 7 ymin <- plot (0, lim = c (-.,.), ylim = c (ymin, yma), main = title, lab = "", ylab = "", type = "n") for (i in : (m - )) { 6 lines (s, v[i, ]) 7 points (cbind (h, u[i, ]), pch = as.character (i), col = cols[i]) 8 } 9 }

26 6 JAN DE LEEUW Code Segment Frank-Wolfe Algorithm qpfrankwolfe <- function (, y, w = rep (, length (y)), b = rep (, ncol ()) / ncol (), itma = 00, eps = e-6, verbose = FALSE) { w <- crossprod (, w * ) fold <- Inf itel <- repeat { 6 res <- c (y - %*% b) 7 fnew<- sum (w * res ^ ) / 8 g <- - colsums (w * res * ) 9 mg <- min (g) 0 ig <- which.min (g) bnew <- rep (0, length (b)) bnew [ig] <- d <- bnew - b tau <- sum (d * crossprod (, w * res)) / sum (w * outer (d, d)) tau <- min (, ma (0, tau)) 6 b <- b + tau * d 7 if (verbose) { 8 cat ("itel: ", formatc (itel,, ), 9 "fold: ", formatc (fold, 0,, "f"), 0 "fnew: ", formatc (fnew, 0,, "f"), } "\n") if (((fold - fnew) < eps) (itel == itma)) { break () } itel <- itel + 6 fold <- fnew 7 } 8 return (list (b = b, f = fnew)) 9 }

27 POLYNOMIAL SINGLE COMPONENT ANALYSIS 7

28 8 JAN DE LEEUW Appendi B. Figures fied (a) fied (b) fied (c) fied (d) fied (e) fied (f) fied (g) fied (h) fied (i) fied (j) fied (k) fied (l) Figure. Twelve solutions with a fied normal kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and, (l) sr.

29 POLYNOMIAL SINGLE COMPONENT ANALYSIS free (a) free (b) free (c) free (d) free (e) free (f) free (g) free (h) free (i) free (j) free (k) free (l) Figure. Twelve solutions with a free kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and, (l) sr.

30 0 JAN DE LEEUW moment (a) moment (b) moment (c) moment (d) moment (e) moment (f) moment (g) moment (h) moment (i) moment (j) moment (k) moment (l) Figure. Twelve solutions with a moment kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and,?? sr.

31 POLYNOMIAL SINGLE COMPONENT ANALYSIS cdf cdf cdf (a) (b) (c) cdf cdf cdf (d) (e) (f) cdf cdf cdf (g) (h) (i) cdf cdf cdf (j) (k) (l) Figure. Twelve solutions with a CDF kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and, (l) sr.

32 JAN DE LEEUW Appendi C. Tables fied r = r = r = s = s = s = s = free r = r = r = s = s = s = s = moment r = r = r = s = s = s = s = cdf r = r = r = s = s = s = s = Table. Minimum loss function values for different kernel types, polynomial degrees s and moment orders r.

33 POLYNOMIAL SINGLE COMPONENT ANALYSIS fied r = r = r = s = 6 s = s = s = 0 00 free r = r = r = s = 00 s = s = 0 00 s = moment r = r = r = s = 00 s = s = s = cdf r = r = r = s = s = s = 8 00 s = 00 Table. Number of iterations for different kernel types, polynomial degrees s and moment orders r. Department of Statistics, University of California, Los Angeles, CA address, Jan de Leeuw: deleeuw@stat.ucla.edu URL, Jan de Leeuw:

MULTIVARIATE CUMULANTS IN R. 1. Introduction

MULTIVARIATE CUMULANTS IN R. 1. Introduction MULTIVARIATE CUMULANTS IN R JAN DE LEEUW Abstract. We give algorithms and R code to compute all multivariate cumulants up to order p of m variables. 1. Introduction Multivariate cumulants are of increasing

More information

HOMOGENEITY ANALYSIS USING EUCLIDEAN MINIMUM SPANNING TREES 1. INTRODUCTION

HOMOGENEITY ANALYSIS USING EUCLIDEAN MINIMUM SPANNING TREES 1. INTRODUCTION HOMOGENEITY ANALYSIS USING EUCLIDEAN MINIMUM SPANNING TREES JAN DE LEEUW 1. INTRODUCTION In homogeneity analysis the data are a system of m subsets of a finite set of n objects. The purpose of the technique

More information

LEAST SQUARES ORTHOGONAL POLYNOMIAL COMPONENT ANALYSIS IN R. 1. Introduction

LEAST SQUARES ORTHOGONAL POLYNOMIAL COMPONENT ANALYSIS IN R. 1. Introduction LEAST SQUARES ORTHOGONAL POLYNOMIAL COMPONENT ANALYSIS IN R JAN DE LEEUW AND KEKONA KAI SORENSON Abstract. We present a simple algorithm (with code in R) to fit multivariate polynomial components or factors

More information

COMPONENT ANALYSIS USING MULTIVARIATE CUMULANTS

COMPONENT ANALYSIS USING MULTIVARIATE CUMULANTS COMPONENT ANALYSIS USING MULTIVARIATE CUMULANTS JAN DE LEEUW AND IRINA KUKUYEVA Abstract. This paper discusses several different algorithms to fit linear component analysis approximations to multivariate

More information

ALGORITHM CONSTRUCTION BY DECOMPOSITION 1. INTRODUCTION. The following theorem is so simple it s almost embarassing. Nevertheless

ALGORITHM CONSTRUCTION BY DECOMPOSITION 1. INTRODUCTION. The following theorem is so simple it s almost embarassing. Nevertheless ALGORITHM CONSTRUCTION BY DECOMPOSITION JAN DE LEEUW ABSTRACT. We discuss and illustrate a general technique, related to augmentation, in which a complicated optimization problem is replaced by a simpler

More information

LEAST SQUARES METHODS FOR FACTOR ANALYSIS. 1. Introduction

LEAST SQUARES METHODS FOR FACTOR ANALYSIS. 1. Introduction LEAST SQUARES METHODS FOR FACTOR ANALYSIS JAN DE LEEUW AND JIA CHEN Abstract. Meet the abstract. This is the abstract. 1. Introduction Suppose we have n measurements on each of m variables. Collect these

More information

MAJORIZATION OF DIFFERENT ORDER

MAJORIZATION OF DIFFERENT ORDER MAJORIZATION OF DIFFERENT ORDER JAN DE LEEUW 1. INTRODUCTION The majorization method [De Leeuw, 1994; Heiser, 1995; Lange et al., 2000] for minimization of real valued loss functions has become very popular

More information

over the parameters θ. In both cases, consequently, we select the minimizing

over the parameters θ. In both cases, consequently, we select the minimizing MONOTONE REGRESSION JAN DE LEEUW Abstract. This is an entry for The Encyclopedia of Statistics in Behavioral Science, to be published by Wiley in 200. In linear regression we fit a linear function y =

More information

MAXIMUM LIKELIHOOD IN GENERALIZED FIXED SCORE FACTOR ANALYSIS 1. INTRODUCTION

MAXIMUM LIKELIHOOD IN GENERALIZED FIXED SCORE FACTOR ANALYSIS 1. INTRODUCTION MAXIMUM LIKELIHOOD IN GENERALIZED FIXED SCORE FACTOR ANALYSIS JAN DE LEEUW ABSTRACT. We study the weighted least squares fixed rank approximation problem in which the weight matrices depend on unknown

More information

Nonlinear Support Vector Machines through Iterative Majorization and I-Splines

Nonlinear Support Vector Machines through Iterative Majorization and I-Splines Nonlinear Support Vector Machines through Iterative Majorization and I-Splines P.J.F. Groenen G. Nalbantov J.C. Bioch July 9, 26 Econometric Institute Report EI 26-25 Abstract To minimize the primal support

More information

QUADRATIC MAJORIZATION 1. INTRODUCTION

QUADRATIC MAJORIZATION 1. INTRODUCTION QUADRATIC MAJORIZATION JAN DE LEEUW 1. INTRODUCTION Majorization methods are used extensively to solve complicated multivariate optimizaton problems. We refer to de Leeuw [1994]; Heiser [1995]; Lange et

More information

Bracketing an Optima in Univariate Optimization

Bracketing an Optima in Univariate Optimization Bracketing an Optima in Univariate Optimization Pritibhushan Sinha Quantitative Methods & Operations Management Area Indian Institute of Management Kozhikode Kozhikode 673570 Kerala India Email: pritibhushan.sinha@iimk.ac.in

More information

9. Interpretations, Lifting, SOS and Moments

9. Interpretations, Lifting, SOS and Moments 9-1 Interpretations, Lifting, SOS and Moments P. Parrilo and S. Lall, CDC 2003 2003.12.07.04 9. Interpretations, Lifting, SOS and Moments Polynomial nonnegativity Sum of squares (SOS) decomposition Eample

More information

FACTOR ANALYSIS AS MATRIX DECOMPOSITION 1. INTRODUCTION

FACTOR ANALYSIS AS MATRIX DECOMPOSITION 1. INTRODUCTION FACTOR ANALYSIS AS MATRIX DECOMPOSITION JAN DE LEEUW ABSTRACT. Meet the abstract. This is the abstract. 1. INTRODUCTION Suppose we have n measurements on each of taking m variables. Collect these measurements

More information

Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables

Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables Achim Zeileis, Edgar C. Merkle, Ting Wang http://eeecon.uibk.ac.at/~zeileis/ Overview Motivation Framework Score-based

More information

Tail Approximation of the Skew-Normal by the Skew-Normal Laplace: Application to Owen s T Function and the Bivariate Normal Distribution

Tail Approximation of the Skew-Normal by the Skew-Normal Laplace: Application to Owen s T Function and the Bivariate Normal Distribution Journal of Statistical and Econometric ethods vol. no. 3 - ISS: 5-557 print version 5-565online Scienpress Ltd 3 Tail Approimation of the Skew-ormal by the Skew-ormal Laplace: Application to Owen s T Function

More information

Simultaneous Diagonalization of Positive Semi-definite Matrices

Simultaneous Diagonalization of Positive Semi-definite Matrices Simultaneous Diagonalization of Positive Semi-definite Matrices Jan de Leeuw Version 21, May 21, 2017 Abstract We give necessary and sufficient conditions for solvability of A j = XW j X, with the A j

More information

Least squares multidimensional scaling with transformed distances

Least squares multidimensional scaling with transformed distances Least squares multidimensional scaling with transformed distances Patrick J.F. Groenen 1, Jan de Leeuw 2 and Rudolf Mathar 3 1 Department of Data Theory, University of Leiden P.O. Box 9555, 2300 RB Leiden,

More information

Power EP. Thomas Minka Microsoft Research Ltd., Cambridge, UK MSR-TR , October 4, Abstract

Power EP. Thomas Minka Microsoft Research Ltd., Cambridge, UK MSR-TR , October 4, Abstract Power EP Thomas Minka Microsoft Research Ltd., Cambridge, UK MSR-TR-2004-149, October 4, 2004 Abstract This note describes power EP, an etension of Epectation Propagation (EP) that makes the computations

More information

1. Data Overview: North Direction. Total observations for each intersection: 96,432 Missing Percentages:

1. Data Overview: North Direction. Total observations for each intersection: 96,432 Missing Percentages: Preliminary Report LOOP IMPUTATION OF HOURLY DATA ON A SECTION OF THE I-5 (BETWEEN ROUTES 14 AND 99) VIA FUNCTIONAL PRINCIPAL COMPONENT ANALYSIS JAN DE LEEUW, IRINA KUKUYEVA Abstract. We present the results

More information

Statistical Geometry Processing Winter Semester 2011/2012

Statistical Geometry Processing Winter Semester 2011/2012 Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian

More information

UCLA Department of Statistics Papers

UCLA Department of Statistics Papers UCLA Department of Statistics Papers Title Rate of Convergence of the Arithmetic-Geometric Mean Process Permalink https://escholarship.org/uc/item/5kx0j9nm Author de Leeuw, Jan Publication Date 008-0-09

More information

CENTERING IN MULTILEVEL MODELS. Consider the situation in which we have m groups of individuals, where

CENTERING IN MULTILEVEL MODELS. Consider the situation in which we have m groups of individuals, where CENTERING IN MULTILEVEL MODELS JAN DE LEEUW ABSTRACT. This is an entry for The Encyclopedia of Statistics in Behavioral Science, to be published by Wiley in 2005. Consider the situation in which we have

More information

FIXED-RANK APPROXIMATION WITH CONSTRAINTS

FIXED-RANK APPROXIMATION WITH CONSTRAINTS FIXED-RANK APPROXIMATION WITH CONSTRAINTS JAN DE LEEUW AND IRINA KUKUYEVA Abstract. We discuss fixed-rank weighted least squares approximation of a rectangular matrix with possibly singular Kronecker weights

More information

Number of cases (objects) Number of variables Number of dimensions. n-vector with categorical observations

Number of cases (objects) Number of variables Number of dimensions. n-vector with categorical observations PRINCALS Notation The PRINCALS algorithm was first described in Van Rickevorsel and De Leeuw (1979) and De Leeuw and Van Rickevorsel (1980); also see Gifi (1981, 1985). Characteristic features of PRINCALS

More information

Approximate inference, Sampling & Variational inference Fall Cours 9 November 25

Approximate inference, Sampling & Variational inference Fall Cours 9 November 25 Approimate inference, Sampling & Variational inference Fall 2015 Cours 9 November 25 Enseignant: Guillaume Obozinski Scribe: Basile Clément, Nathan de Lara 9.1 Approimate inference with MCMC 9.1.1 Gibbs

More information

NON-NEGATIVE MATRIX FACTORIZATION FOR PARAMETER ESTIMATION IN HIDDEN MARKOV MODELS. Balaji Lakshminarayanan and Raviv Raich

NON-NEGATIVE MATRIX FACTORIZATION FOR PARAMETER ESTIMATION IN HIDDEN MARKOV MODELS. Balaji Lakshminarayanan and Raviv Raich NON-NEGATIVE MATRIX FACTORIZATION FOR PARAMETER ESTIMATION IN HIDDEN MARKOV MODELS Balaji Lakshminarayanan and Raviv Raich School of EECS, Oregon State University, Corvallis, OR 97331-551 {lakshmba,raich@eecs.oregonstate.edu}

More information

A GENERALIZATION OF TAKANE'S ALGORITHM FOR DEDICOM. Yosmo TAKANE

A GENERALIZATION OF TAKANE'S ALGORITHM FOR DEDICOM. Yosmo TAKANE PSYCHOMETR1KA--VOL. 55, NO. 1, 151--158 MARCH 1990 A GENERALIZATION OF TAKANE'S ALGORITHM FOR DEDICOM HENK A. L. KIERS AND JOS M. F. TEN BERGE UNIVERSITY OF GRONINGEN Yosmo TAKANE MCGILL UNIVERSITY JAN

More information

Lecture 3: Latent Variables Models and Learning with the EM Algorithm. Sam Roweis. Tuesday July25, 2006 Machine Learning Summer School, Taiwan

Lecture 3: Latent Variables Models and Learning with the EM Algorithm. Sam Roweis. Tuesday July25, 2006 Machine Learning Summer School, Taiwan Lecture 3: Latent Variables Models and Learning with the EM Algorithm Sam Roweis Tuesday July25, 2006 Machine Learning Summer School, Taiwan Latent Variable Models What to do when a variable z is always

More information

CSCI-6971 Lecture Notes: Probability theory

CSCI-6971 Lecture Notes: Probability theory CSCI-6971 Lecture Notes: Probability theory Kristopher R. Beevers Department of Computer Science Rensselaer Polytechnic Institute beevek@cs.rpi.edu January 31, 2006 1 Properties of probabilities Let, A,

More information

Lecture 9: MATH 329: Introduction to Scientific Computing

Lecture 9: MATH 329: Introduction to Scientific Computing Lecture 9: MATH 329: Introduction to Scientific Computing Sanjeena Dang Department of Mathematical Sciences Vectors and Matrices in R Numeric vector " and matrix objects in R are a close match to mathematical

More information

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric

More information

Equivalence of Minimal l 0 and l p Norm Solutions of Linear Equalities, Inequalities and Linear Programs for Sufficiently Small p

Equivalence of Minimal l 0 and l p Norm Solutions of Linear Equalities, Inequalities and Linear Programs for Sufficiently Small p Equivalence of Minimal l 0 and l p Norm Solutions of Linear Equalities, Inequalities and Linear Programs for Sufficiently Small p G. M. FUNG glenn.fung@siemens.com R&D Clinical Systems Siemens Medical

More information

M.S. Project Report. Efficient Failure Rate Prediction for SRAM Cells via Gibbs Sampling. Yamei Feng 12/15/2011

M.S. Project Report. Efficient Failure Rate Prediction for SRAM Cells via Gibbs Sampling. Yamei Feng 12/15/2011 .S. Project Report Efficient Failure Rate Prediction for SRA Cells via Gibbs Sampling Yamei Feng /5/ Committee embers: Prof. Xin Li Prof. Ken ai Table of Contents CHAPTER INTRODUCTION...3 CHAPTER BACKGROUND...5

More information

Review of Optimization Basics

Review of Optimization Basics Review of Optimization Basics. Introduction Electricity markets throughout the US are said to have a two-settlement structure. The reason for this is that the structure includes two different markets:

More information

AN ITERATIVE MAXIMUM EIGENVALUE METHOD FOR SQUARED DISTANCE SCALING 1. INTRODUCTION

AN ITERATIVE MAXIMUM EIGENVALUE METHOD FOR SQUARED DISTANCE SCALING 1. INTRODUCTION AN ITERATIVE MAXIMUM EIGENVALUE METHOD FOR SQUARED DISTANCE SCALING JAN DE LEEUW ABSTRACT. Meet the abstract. This is the abstract. 1. INTRODUCTION The proble we consider in this paper is a etric ultidiensional

More information

MACHINE LEARNING ADVANCED MACHINE LEARNING

MACHINE LEARNING ADVANCED MACHINE LEARNING MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,

More information

Section - 9 GRAPHS. (a) y f x (b) y f x. (c) y f x (d) y f x. (e) y f x (f) y f x k. (g) y f x k (h) y kf x. (i) y f kx. [a] y f x to y f x

Section - 9 GRAPHS. (a) y f x (b) y f x. (c) y f x (d) y f x. (e) y f x (f) y f x k. (g) y f x k (h) y kf x. (i) y f kx. [a] y f x to y f x 44 Section - 9 GRAPHS In this section, we will discuss graphs and graph-plotting in more detail. Detailed graph plotting also requires a knowledge of derivatives. Here, we will be discussing some general

More information

On Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process

On Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process On Simulations form the Two-Parameter arxiv:1209.5359v1 [stat.co] 24 Sep 2012 Poisson-Dirichlet Process and the Normalized Inverse-Gaussian Process Luai Al Labadi and Mahmoud Zarepour May 8, 2018 ABSTRACT

More information

MULTIVARIATE ANALYSIS OF ORDINAL DATA USING PROBITS

MULTIVARIATE ANALYSIS OF ORDINAL DATA USING PROBITS MULTIVARIATE ANALYSIS OF ORDINAL DATA USING PROBITS JAN DE LEEUW ABSTRACT. We discuss algorithms to fit component and factor representations to multivariate multicategory ordinal data, using a probit threshold

More information

Introduction. x 7. Partial fraction decomposition x 7 of. x 7. x 3 1. x 2. Decomposition of N x /D x into Partial Fractions. N x. D x polynomial N 1 x

Introduction. x 7. Partial fraction decomposition x 7 of. x 7. x 3 1. x 2. Decomposition of N x /D x into Partial Fractions. N x. D x polynomial N 1 x 333202_0704.qd 2/5/05 9:43 M Page 533 Section 7.4 Partial Fractions 533 7.4 Partial Fractions What you should learn Recognize partial fraction decompositions of rational epressions. Find partial fraction

More information

Approximate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem

Approximate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem Approimate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem ALEXEI UTESHEV St.Petersburg State University Department of Applied Mathematics Universitetskij pr. 35, 198504

More information

Least squares contour alignment

Least squares contour alignment Least squares contour alignment Ivan Markovsky and Sasan Mahmoodi School of Electronics and Computer Science University of Southampton, SO7 BJ, UK Emails: {im,sm3}@ecssotonacuk Abstract The contour alignment

More information

Lecture 3: Pattern Classification. Pattern classification

Lecture 3: Pattern Classification. Pattern classification EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and

More information

Precalculus Notes: Functions. Skill: Solve problems using the algebra of functions.

Precalculus Notes: Functions. Skill: Solve problems using the algebra of functions. Skill: Solve problems using the algebra of functions. Modeling a Function: Numerical (data table) Algebraic (equation) Graphical Using Numerical Values: Look for a common difference. If the first difference

More information

2 Statistical Estimation: Basic Concepts

2 Statistical Estimation: Basic Concepts Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:

More information

Multilevel Analysis, with Extensions

Multilevel Analysis, with Extensions May 26, 2010 We start by reviewing the research on multilevel analysis that has been done in psychometrics and educational statistics, roughly since 1985. The canonical reference (at least I hope so) is

More information

Gaussians. Hiroshi Shimodaira. January-March Continuous random variables

Gaussians. Hiroshi Shimodaira. January-March Continuous random variables Cumulative Distribution Function Gaussians Hiroshi Shimodaira January-March 9 In this chapter we introduce the basics of how to build probabilistic models of continuous-valued data, including the most

More information

18.303: Introduction to Green s functions and operator inverses

18.303: Introduction to Green s functions and operator inverses 8.33: Introduction to Green s functions and operator inverses S. G. Johnson October 9, 2 Abstract In analogy with the inverse A of a matri A, we try to construct an analogous inverse  of differential

More information

Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables

Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables Achim Zeileis, Edgar C. Merkle, Ting Wang http://eeecon.uibk.ac.at/~zeileis/ Overview Motivation Framework Score-based

More information

9 - Markov processes and Burt & Allison 1963 AGEC

9 - Markov processes and Burt & Allison 1963 AGEC This document was generated at 8:37 PM on Saturday, March 17, 2018 Copyright 2018 Richard T. Woodward 9 - Markov processes and Burt & Allison 1963 AGEC 642-2018 I. What is a Markov Chain? A Markov chain

More information

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.

Dependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline. Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,

More information

Math 2412 Activity 1(Due by EOC Sep. 17)

Math 2412 Activity 1(Due by EOC Sep. 17) Math 4 Activity (Due by EOC Sep. 7) Determine whether each relation is a unction.(indicate why or why not.) Find the domain and range o each relation.. 4,5, 6,7, 8,8. 5,6, 5,7, 6,6, 6,7 Determine whether

More information

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 4

2.29 Numerical Fluid Mechanics Spring 2015 Lecture 4 2.29 Spring 2015 Lecture 4 Review Lecture 3 Truncation Errors, Taylor Series and Error Analysis Taylor series: 2 3 n n i1 i i i i i n f( ) f( ) f '( ) f ''( ) f '''( )... f ( ) R 2! 3! n! n1 ( n1) Rn f

More information

CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM

CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM Ralf L.M. Peeters Bernard Hanzon Martine Olivi Dept. Mathematics, Universiteit Maastricht, P.O. Box 616, 6200 MD Maastricht,

More information

Exact and Approximate Numbers:

Exact and Approximate Numbers: Eact and Approimate Numbers: The numbers that arise in technical applications are better described as eact numbers because there is not the sort of uncertainty in their values that was described above.

More information

The stability of limit cycles in nonlinear systems

The stability of limit cycles in nonlinear systems Nonlinear Dyn 009) 56: 69 75 DOI 10.1007/s11071-008-9398-3 O R I G I NA L PA P E R The stability of limit cycles in nonlinear systems Ali Ghaffari Masayoshi Tomizuka Reza A. Soltan Received: 1 March 007

More information

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks

SPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks Topics in Machine Learning-EE 5359 Neural Networks 1 The Perceptron Output: A perceptron is a function that maps D-dimensional vectors to real numbers. For notational convenience, we add a zero-th dimension

More information

Machine Learning Lecture 3

Machine Learning Lecture 3 Announcements Machine Learning Lecture 3 Eam dates We re in the process of fiing the first eam date Probability Density Estimation II 9.0.207 Eercises The first eercise sheet is available on L2P now First

More information

Inexact Solution of NLP Subproblems in MINLP

Inexact Solution of NLP Subproblems in MINLP Ineact Solution of NLP Subproblems in MINLP M. Li L. N. Vicente April 4, 2011 Abstract In the contet of conve mied-integer nonlinear programming (MINLP, we investigate how the outer approimation method

More information

Factor Analysis. Qian-Li Xue

Factor Analysis. Qian-Li Xue Factor Analysis Qian-Li Xue Biostatistics Program Harvard Catalyst The Harvard Clinical & Translational Science Center Short course, October 7, 06 Well-used latent variable models Latent variable scale

More information

Mobile Robot Localization

Mobile Robot Localization Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations

More information

Package depth.plot. December 20, 2015

Package depth.plot. December 20, 2015 Package depth.plot December 20, 2015 Type Package Title Multivariate Analogy of Quantiles Version 0.1 Date 2015-12-19 Maintainer Could be used to obtain spatial depths, spatial ranks and outliers of multivariate

More information

The incomplete gamma functions. Notes by G.J.O. Jameson. These notes incorporate the Math. Gazette article [Jam1], with some extra material.

The incomplete gamma functions. Notes by G.J.O. Jameson. These notes incorporate the Math. Gazette article [Jam1], with some extra material. The incomplete gamma functions Notes by G.J.O. Jameson These notes incorporate the Math. Gazette article [Jam], with some etra material. Definitions and elementary properties functions: Recall the integral

More information

Perturbation Theory for Variational Inference

Perturbation Theory for Variational Inference Perturbation heor for Variational Inference Manfred Opper U Berlin Marco Fraccaro echnical Universit of Denmark Ulrich Paquet Apple Ale Susemihl U Berlin Ole Winther echnical Universit of Denmark Abstract

More information

MACHINE LEARNING ADVANCED MACHINE LEARNING

MACHINE LEARNING ADVANCED MACHINE LEARNING MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 22 MACHINE LEARNING Discrete Probabilities Consider two variables and y taking discrete

More information

ON MODEL SELECTION FOR STATE ESTIMATION FOR NONLINEAR SYSTEMS. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof

ON MODEL SELECTION FOR STATE ESTIMATION FOR NONLINEAR SYSTEMS. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof ON MODEL SELECTION FOR STATE ESTIMATION FOR NONLINEAR SYSTEMS Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof Delft Center for Systems and Control, Delft University of Technology, Mekelweg 2, 2628 CD

More information

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form: 3.3 Gradient Vector and Jacobian Matri 3 3.3 Gradient Vector and Jacobian Matri Overview: Differentiable functions have a local linear approimation. Near a given point, local changes are determined by

More information

Bayesian Approach 2. CSC412 Probabilistic Learning & Reasoning

Bayesian Approach 2. CSC412 Probabilistic Learning & Reasoning CSC412 Probabilistic Learning & Reasoning Lecture 12: Bayesian Parameter Estimation February 27, 2006 Sam Roweis Bayesian Approach 2 The Bayesian programme (after Rev. Thomas Bayes) treats all unnown quantities

More information

5.8 Chebyshev Approximation

5.8 Chebyshev Approximation 84 Chapter 5. Evaluation of Functions appropriate setting is ld=, and the value of the derivative is the accumulated sum divided by the sampling interval h. CITED REFERENCES AND FURTHER READING: Dennis,

More information

M 340L CS Homework Set 1

M 340L CS Homework Set 1 M 340L CS Homework Set 1 Solve each system in Problems 1 6 by using elementary row operations on the equations or on the augmented matri. Follow the systematic elimination procedure described in Lay, Section

More information

A Rectangular Trust Region Dogleg Approach for Unconstrained and Bound Constrained Nonlinear Optimization

A Rectangular Trust Region Dogleg Approach for Unconstrained and Bound Constrained Nonlinear Optimization A Rectangular Trust Region Dogleg Approach for Unconstrained and Bound onstrained onlinear Optimization. VOGLIS, I.E. LAGARIS Department of omputer Science University of Ioannina GREEE f g Abstract: -

More information

LECTURE NOTES ON QUANTUM COMPUTATION. Cornell University, Physics , CS 483; Spring, 2005 c 2006, N. David Mermin

LECTURE NOTES ON QUANTUM COMPUTATION. Cornell University, Physics , CS 483; Spring, 2005 c 2006, N. David Mermin LECTURE NOTES ON QUANTUM COMPUTATION Cornell University, Physics 481-681, CS 483; Spring, 2005 c 2006, N. David Mermin IV. Searching with a Quantum Computer Last revised 3/30/06 Suppose you know that eactly

More information

Partitioning variation in multilevel models.

Partitioning variation in multilevel models. Partitioning variation in multilevel models. by Harvey Goldstein, William Browne and Jon Rasbash Institute of Education, London, UK. Summary. In multilevel modelling, the residual variation in a response

More information

Outline. Random Variables. Examples. Random Variable

Outline. Random Variables. Examples. Random Variable Outline Random Variables M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno Random variables. CDF and pdf. Joint random variables. Correlated, independent, orthogonal. Correlation,

More information

REGRESSION, DISCRIMINANT ANALYSIS, AND CANONICAL CORRELATION ANALYSIS WITH HOMALS 1. MORALS

REGRESSION, DISCRIMINANT ANALYSIS, AND CANONICAL CORRELATION ANALYSIS WITH HOMALS 1. MORALS REGRESSION, DISCRIMINANT ANALYSIS, AND CANONICAL CORRELATION ANALYSIS WITH HOMALS JAN DE LEEUW ABSTRACT. It is shown that the homals package in R can be used for multiple regression, multi-group discriminant

More information

Application of modified Leja sequences to polynomial interpolation

Application of modified Leja sequences to polynomial interpolation Special Issue for the years of the Padua points, Volume 8 5 Pages 66 74 Application of modified Leja sequences to polynomial interpolation L. P. Bos a M. Caliari a Abstract We discuss several applications

More information

Gaussian Process Vine Copulas for Multivariate Dependence

Gaussian Process Vine Copulas for Multivariate Dependence Gaussian Process Vine Copulas for Multivariate Dependence José Miguel Hernández-Lobato 1,2 joint work with David López-Paz 2,3 and Zoubin Ghahramani 1 1 Department of Engineering, Cambridge University,

More information

Lecture 2. Tensor Unfoldings. Charles F. Van Loan

Lecture 2. Tensor Unfoldings. Charles F. Van Loan From Matrix to Tensor: The Transition to Numerical Multilinear Algebra Lecture 2. Tensor Unfoldings Charles F. Van Loan Cornell University The Gene Golub SIAM Summer School 2010 Selva di Fasano, Brindisi,

More information

Numerical Methods. Root Finding

Numerical Methods. Root Finding Numerical Methods Solving Non Linear 1-Dimensional Equations Root Finding Given a real valued function f of one variable (say ), the idea is to find an such that: f() 0 1 Root Finding Eamples Find real

More information

Filter-Generating Systems

Filter-Generating Systems 24 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO. 3, MARCH 2000 Filter-Generating Systems Saed Samadi, Akinori Nishihara, and Hiroshi Iwakura Abstract

More information

Projection, Inference, and Consistency

Projection, Inference, and Consistency Projection, Inference, and Consistency John Hooker Carnegie Mellon University IJCAI 2016, New York City A high-level presentation. Don t worry about the details. 2 Projection as a Unifying Concept Projection

More information

A New Lower Bound Technique for Quantum Circuits without Ancillæ

A New Lower Bound Technique for Quantum Circuits without Ancillæ A New Lower Bound Technique for Quantum Circuits without Ancillæ Debajyoti Bera Abstract We present a technique to derive depth lower bounds for quantum circuits. The technique is based on the observation

More information

Quadratic Programming Relaxations for Metric Labeling and Markov Random Field MAP Estimation

Quadratic Programming Relaxations for Metric Labeling and Markov Random Field MAP Estimation Quadratic Programming Relaations for Metric Labeling and Markov Random Field MAP Estimation Pradeep Ravikumar John Lafferty School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213,

More information

Multidimensional partitions of unity and Gaussian terrains

Multidimensional partitions of unity and Gaussian terrains and Gaussian terrains Richard A. Bale, Jeff P. Grossman, Gary F. Margrave, and Michael P. Lamoureu ABSTRACT Partitions of unity play an important rôle as amplitude-preserving windows for nonstationary

More information

Lecture 4b. Bessel functions. Introduction. Generalized factorial function. 4b.1. Using integration by parts it is easy to show that

Lecture 4b. Bessel functions. Introduction. Generalized factorial function. 4b.1. Using integration by parts it is easy to show that 4b. Lecture 4b Using integration by parts it is easy to show that Bessel functions Introduction In the previous lecture the separation of variables method led to Bessel's equation y' ' y ' 2 y= () 2 Here

More information

Taylor Series and Asymptotic Expansions

Taylor Series and Asymptotic Expansions Taylor Series and Asymptotic Epansions The importance of power series as a convenient representation, as an approimation tool, as a tool for solving differential equations and so on, is pretty obvious.

More information

Particle Methods as Message Passing

Particle Methods as Message Passing Particle Methods as Message Passing Justin Dauwels RIKEN Brain Science Institute Hirosawa,2-1,Wako-shi,Saitama,Japan Email: justin@dauwels.com Sascha Korl Phonak AG CH-8712 Staefa, Switzerland Email: sascha.korl@phonak.ch

More information

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010 Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X

More information

Lecture 11. Linear Soft Margin Support Vector Machines

Lecture 11. Linear Soft Margin Support Vector Machines CS142: Machine Learning Spring 2017 Lecture 11 Instructor: Pedro Felzenszwalb Scribes: Dan Xiang, Tyler Dae Devlin Linear Soft Margin Support Vector Machines We continue our discussion of linear soft margin

More information

Counting, symbols, positions, powers, choice, arithmetic, binary, translation, hex, addresses, and gates.

Counting, symbols, positions, powers, choice, arithmetic, binary, translation, hex, addresses, and gates. Counting, symbols, positions, powers, choice, arithmetic, binary, translation, he, addresses, and gates. Counting. Suppose the concern is a collection of objects. As an eample, let the objects be denoted

More information

6. Vector Random Variables

6. Vector Random Variables 6. Vector Random Variables In the previous chapter we presented methods for dealing with two random variables. In this chapter we etend these methods to the case of n random variables in the following

More information

A Comparison of Particle Filters for Personal Positioning

A Comparison of Particle Filters for Personal Positioning VI Hotine-Marussi Symposium of Theoretical and Computational Geodesy May 9-June 6. A Comparison of Particle Filters for Personal Positioning D. Petrovich and R. Piché Institute of Mathematics Tampere University

More information

Lecture 7: Weak Duality

Lecture 7: Weak Duality EE 227A: Conve Optimization and Applications February 7, 2012 Lecture 7: Weak Duality Lecturer: Laurent El Ghaoui 7.1 Lagrange Dual problem 7.1.1 Primal problem In this section, we consider a possibly

More information

1 Convexity, Convex Relaxations, and Global Optimization

1 Convexity, Convex Relaxations, and Global Optimization 1 Conveity, Conve Relaations, and Global Optimization Algorithms 1 1.1 Minima Consider the nonlinear optimization problem in a Euclidean space: where the feasible region. min f(), R n Definition (global

More information

Lecture 2: CDF and EDF

Lecture 2: CDF and EDF STAT 425: Introduction to Nonparametric Statistics Winter 2018 Instructor: Yen-Chi Chen Lecture 2: CDF and EDF 2.1 CDF: Cumulative Distribution Function For a random variable X, its CDF F () contains all

More information

On the projection onto a finitely generated cone

On the projection onto a finitely generated cone Acta Cybernetica 00 (0000) 1 15. On the projection onto a finitely generated cone Miklós Ujvári Abstract In the paper we study the properties of the projection onto a finitely generated cone. We show for

More information

Basic Linear Inverse Method Theory - DRAFT NOTES

Basic Linear Inverse Method Theory - DRAFT NOTES Basic Linear Inverse Method Theory - DRAFT NOTES Peter P. Jones 1 1 Centre for Compleity Science, University of Warwick, Coventry CV4 7AL, UK (Dated: 21st June 2012) BASIC LINEAR INVERSE METHOD THEORY

More information

Consistency as Projection

Consistency as Projection Consistency as Projection John Hooker Carnegie Mellon University INFORMS 2015, Philadelphia USA Consistency as Projection Reconceive consistency in constraint programming as a form of projection. For eample,

More information

Numerical Integration (Quadrature) Another application for our interpolation tools!

Numerical Integration (Quadrature) Another application for our interpolation tools! Numerical Integration (Quadrature) Another application for our interpolation tools! Integration: Area under a curve Curve = data or function Integrating data Finite number of data points spacing specified

More information