POLYNOMIAL SINGLE COMPONENT ANALYSIS USING MULTIVARIATE MOMENTS
|
|
- Ambrose Weaver
- 5 years ago
- Views:
Transcription
1 POLYNOMIAL SINGLE COMPONENT ANALYSIS USING MULTIVARIATE MOMENTS JAN DE LEEUW Abstract. Polynomial single component analysis is defined. Algorithms, based on multidimensional array approimation methods, to fit a model with a single common factor to observed multivariate moments are described and applied to a psychometric eample. Software in R is included. Date: Tuesday 0 th July, 0 7h 7min Typeset in Lucida Bright. 000 Mathematics Subject Classification. 6H. Key words and phrases. Component Analysis, Polynomials, Array Decmposition.
2 JAN DE LEEUW. Model Suppose y j are random variables, with j =,, m. We use the convention of underlining random variables [Hemelrijk, 966]. In polynomial single component analysis we assume there eist a random variable such that () y j = P j (), where the P j are univariate polynomials of degree s. Thus s () P j () = β jp p. p=0 Closely related previous work on the polynomial factor analysis model, which differs from () because it includes unique factors, is in Mooijaart and Bentler [986], who also discuss earlier work by Gibson and McDonald. Mooijaart and Bentler [986] use higher order product moments, specifically third order moments, and the special case of one-dimensional polynomial factor analysis is worked out and applied to a real eample. Since Mooijaart and Bentler [986] the emphasis has shifted away from higher order statistics to using instrumental variables and other more specialized techniques, and from component and factor analysis to general nonlinear structural equation models. Good historical overviews of these more recent developments are in Yalcin and Amemiya [00] and Mooijaart and Bentler [00]. In this paper we go back to the simple single component model, we use higher-order statistics, specifically multivariate moments, and we use decomposition methods for multidimensional arrays to fit the models to sample moments. Thus, compared to Mooijaart and Bentler [986], we develop the case of non-normal factors, provide algorithms and software to deal with moments of arbitrary order, and develop algorithms and software based on fitting multilinear
3 POLYNOMIAL SINGLE COMPONENT ANALYSIS array decompositions. All software is written in R [R Development Core Team, 0]. In a companion paper [De Leeuw, 0] we have worked out the equations in terms of cumulants, instead of moments. Cumulants are useful for laying the groundwork for generalizations to approimations with multiple components, and to factor analysis models. In the special case of single-factor component analysis, however, working with moments is much simpler.. Product Moments From the multilinearity property of moments for any r variables, not necessarily distinct, we have s s () E(P j () P jr ()) = β j p β jr p r c p p r, p =0 p r =0 where () c p p r = µ p + +p r, and where we write µ t for the t th raw moment of. The product moments of order r of the polynomials are a supersymmetric multilinear function of the elements of B = {β jp }, for a given kernel C = {c p p r }. Supper-symmetry means that the array elements are invariant under all permutations of the subscript. Moreover the kernel must have the moment structure (). The s r elements of the kernel array are simple linear functions of the first sr moments of. In fact we can write () C = sr t= µ t I t, where the I t are super-symmetric orthogonal binary arrays with {I t } p p r = only if p + + p r = t.
4 JAN DE LEEUW Of course not all vectors are possible vectors of the first moments of a random variable. Defining all possible vectors of moments is the truncated version of the classical Moment Problem, which requires that the Hankel matri formed from the moments is positive semi-definite. See, for eample, Curto and Fialkow [99] for the details. Instead of dealing with the complicated moment inequalities we can use the parametrization µ t = t df() and think of the kernel as a function of the distribution function F. By Mulholland and Rogers [98] we can limit ourselves, without loss of generality, to step functions with at most sr + steps. When dealing with multivariate moments it is convenient to introduce an additional variable y 0 which is a.s. equal to one and a corresponding polynomial P 0 which has β 0p = 0 for p =,, s. If we include P 0 in the epectations (), then the moment array includes all moments up to order r, i.e. also the moments of order,, r.. Loss Function Given the result () it becomes straightforward to generalize the methods for linear component analysis in De Leeuw and Kukuyeva [0] to the polynomial case. We first compute an array of moments A (r ) = {a (r ) j j r } of order up to r from a given data matri. This can be done with the code in segment. Insert Code Segment about here The data is a super-symmetric moment array of order (m + ) r. The kernel is a super-symmetric array of order s r, and B is an (m + ) s matri of loadings. The first row of B is zero, ecept for its first element.
5 POLYNOMIAL SINGLE COMPONENT ANALYSIS We then define the least squares loss function by (6) σ (B, C) = j J j r J r w j j r (a j j r s p =0 and minimize it over both B and C. s p r =0 β j p β jr p r c p p r ). The w j j r are given non-negative weights. They can be used to give lower or higher weights to the moments of various orders, or even to incorporate information about the standard errors of the moments. Note that the loss function can be written in matri form as (7) σ (B, C) = {vec(a) (B } {{ B} )vec(c)} V {vec(a) (B } {{ B} )vec(c))}, r times r times where V = diag(vec(w )). In actual computation and reporting we normalize the loss function to its root-mean-square form σ (B, C) (8) σ (B, C) = tr(v ). For the fitting of C three different cases are of interest... Fied Kernel. The problem is relatively simple if the kernel C is completely known. This requires, for eample, the additional assumption that has a standard normal distribution. The code in segment can be used to compute moments up to order r, including µ 0 =, of the standard normal distribution. Insert Code Segment about here The code in can be used to put moments into the kernel array, using the formula (). Insert Code Segment about here
6 6 JAN DE LEEUW In the case of a fied kernel the loss function only needs to be minimized over B... Free Kernel. In the second special case we ignore the fact that the kernel array C = {c p p r } is, according to (), an array formed of moments of the same variable. Thus we ignore the structure imposed by (). We merely imposing super-symmetry on C. We now minimize the loss function over both B and super-symmetric C. Ignoring the structure of the kernel array is convenient, but it does imply we will not be able to distinguish a linear model with s + correlated components from a polynomial model of degree s. This is a major disadvantage of the method. Nevertheless, if using higher order moments forces identification, then we will be able to estimate factor loadings (or polynomial coefficients) consistently, up to a linear transformation... Structured Kernel. This is the more complicated case, and in several important ways the most interesting one. We minimize over B and over all C with the structure (). We minimize loss over B and over these vectors of moments. There are two variations. In the first we minimize over all vectors µ, without imposing the condition that the vector consists of moments of a single variable. In the second we minimize over all distributions F, which determine the moments. In fact, it usually makes more sense to minimize over a finite-dimensional conve set of distribution functions, for eample the set of all conve combinations of B-splines of a given order with a given set of knots. In this case we can write B + (9) µ t = π b t df b (). b= Collecting the moments of the basis distributions in a matri M allows us to write µ = Mπ.
7 POLYNOMIAL SINGLE COMPONENT ANALYSIS 7 The code in segment computes the basis I t of indicators for the kernel used in formula (). It is wasteful to compute and store the indicators, but for small eamples it will be both feasible and computationally efficient. Insert Code Segment about here The code in segment uses Neuman [98, Proposition.] to compute the moments of B-splines for any knot sequence to any order. Insert Code Segment about here. Algorithms Various segments of code needed for our minimizations, written in R, to minimize (6) are given in De Leeuw [008, 0]; De Leeuw and Kukuyeva [0]. We adopt and etend them to deal with our form of component analysis... Initial Solution. Because of the clear possibility of local minima we want to start with a good initial configuration. For the kernel we always use the one based on moments of the standard normal. For the loadings we approimate the first moments y and second momentss by an equation the form y 0 c b. y S b B c C 0 B The product on the right works out to b + c B b + Bc bb + Bcb + bc B + BCB Now choose b = y Bc, so that we fit the means eactly. Then the fit becomes y y yy + B(C cc )B,
8 8 JAN DE LEEUW and we find B by using unweighted least squares to approimate the covariance S yy by B(C cc )B. This is a simple eigenvalueeigenvector problem... Fied Kernel. To minimize over B we use cyclic coordinate descent (CCD), i.e. we change one coordinate of B at a time while keeping all other fied at their current values. One iteration to update B is a cycle in which all coordinates are changed. CCD is discussed, for eample, in Zangwill [969, Page -], and in a similar contet it has been used in ALSCAL [Takane et al., 977]. It is currently popular in machine learning. See for some eamples and references. If we deal with cumulants of order four, then changing one coordinate makes the loss function a polynomial of order eight of that coordinate. If the cumulants are of order three, the polynomial is of order si, and so on. The polynomial is minimized by using the optimize() function in R. There is no guarantee CCD will find the global minimum. As in ALSCAL, it is possible, at least in principle, to use global optimization methods for multivariate (non-negative) polynomials [?]. We have not tried this, and for larger problems the computational requirements are quite daunting... Free Kernel. Fitting a free kernel is a simple modification of the basic algorithm. After a CCD cycle we update the kernel for given coordinates, and this is a simple linear least squares problem. The resulting Alternating Least Squares (ALS) algorithm, which alternates CCD cycles and kernel updates, is a block relaation algorithm in the sense of De Leeuw [99].
9 POLYNOMIAL SINGLE COMPONENT ANALYSIS 9 Note that from (7) we see, using Moore-Penrose inverses, that the optimal kernel for fied B is (0) vec(c) = (B } + {{ B + })vec(a). r times.. Structured Kernel. From () () (B } {{ B} )vec(c) = r times sr t= µ t (B } {{ B} )vec(i t ) r times Collecting the vectors (B } {{ B} )vec(i t ) in the columns of a matri r times H allows us to write the loss function as () σ (B, µ) = (vec(a) Hµ) V (vec(a) Hµ), while using (9) gives () σ (B, π) = (vec(a) HMπ) V (vec(a) HMπ). Thus the two variations for a structured kernel are to minimize () over all µ and to minimize () over all π in the unit simple. Both are linear least squares problems, in the second case with simple linear inequality and equality constraints. We solve the quadratic programming problem with the Frank-Wolfe linearization algorithm using the optimal step-size. The code is in segment. Insert Code Segment about here. Real Eample We use the YouthGratitude data from the R package psychotools [Zeileis et al., 0]. They are described in detail in the article by Froh et al. [0]. The si seven-point Likert scale variables of the GQ-6 scale are used, with responses from 0 students aged 0-9 years. In Table A of Froh et al. [0] we see that a linear factor analysis gives loadings between.8 and.9 for items and, between.6 and.7 for items,, and, and between. and. for
10 0 JAN DE LEEUW item 6. On the basis of these results the authors eliminate item 6 from further analysis, and we will do the same thing. The data are generate by the following code. library (psychotools) data ("YouthGratitude") gq6 <- as.matri (YouthGratitude[,:8]) We then load the required packages and source the required files. require (abind) require (apl) require (polynom) source ("tuckerccd.r") 6 source ("kernelinitial.r") 7 source ("kernelupdate.r") 8 source ("tuckermore.r") 9 source ("tuckerfinal.r") 0 source ("tuckerkernel.r") source ("tuckerau.r") source ("tuckerbasis.r") source ("tuckerplot.r") source ("utilities.r") source ("simplels.r") For the runs we set the iteration precision, the decrease of successive loss function values, to and we set the maimum number of outer iterations to 00. The number of inner iterations, which are the coordinate descent cycles to fit B, is set at. If we use a miture of B-spline densities for the moments of the distribution of we use equally spaced knots at the positive and negative integers from -6 to +6 and choose order (i.e. we use piecewise quadratic polynomials differentiable at the knots).
11 POLYNOMIAL SINGLE COMPONENT ANALYSIS The code below does 8 analyses, for each of the four kernel types, using moments of order r =, and polynomial degrees s =,,. loss <- array (0, c(,, )) itel <- array (0, c(,, )) R <- : S <- : 6 K <- : 7 8 for (r in R) { 9 for (s in S) { 0 for (k in K) { kertype <- switch (k, "fied", "free", " moment", "cdf") vname <- paste (kertype, r, s, sep = "") pname <- paste (vname, ".pdf", sep = "") assign (paste (kertype, r, s, sep = ""), tuckerccd (gq6, r, s, kertype = kertype )) llist <- eval (as.name (vname)) 6 pdf (pname) 7 plotpol (llist $ b, vname) 8 dev.off () 9 loss [s, r -, k] <- llist $ tloss 0 itel [s, r -, k] <- llist $ otel } } } dump (c("loss", "itel"), file = "results.r" ).. Results for Fied Normal Kernel.
12 JAN DE LEEUW 6. Etensions and Elaborations 6.. Weights and Efficiency. Adding weights W = {w j j r } to (6) is straightforward. It is more complicated to use the standard errors of the multivariate cumulants to improve (asymptotically distribution free) efficiency of the estimates. Again we will rely on (??). 6.. Constraints on Loadings. Adding linear constraints (some elements are zero, some elements are equal) on B is again straightforward. This makes it possible to use different polynomial degrees for different variables. We can also impose ordinal constraints, for instance that polynomials are monotonic. 6.. Consistency. Sample cumulants converge in probability to population cumulants. Minimizers of the loss function (or stationary points of the algorithm) are, under regularity conditions, continuous functions of the sample cumulants. Thus, under additional identifiability conditions, the estimates computed by the algorithm are consistent estimates of the population parameters. 6.. Multiple Common Factors. Additional work is needed to adapt this approach to structures with more than one common factor. The common part of each variable will be a multivariate polynomial of the common factors, i.e. a symmetric multilinear form. It is consequently still true that the multivariate polynomial is a linear combination of powers and products of powers, and thus the basic results (??) and (??) continue to apply. 6.. Non-additive Unique Factors. The additive combination rule in () is not as compelling as in the linear case. This is also the reason we have not paid much attention to factor analysis, as opposed to component analysis, in this paper. The more general decomposition y j = P j (, u j ) becomes interesting, but this is more naturally handled by the etension to mutiple common factors. For
13 POLYNOMIAL SINGLE COMPONENT ANALYSIS this case (??) no longer applies, but we can continue to rely on (??). This more general description, together with work on more than one factor, will become a topic for further research Convergence Accelleration. In further work we plan to use the methods discussed in De Leeuw [008] to speed up convergence. It must be emphasized, however, that there are many more parts of the algorithm that can be accelerated in various ways.
14 JAN DE LEEUW References R.E. Curto and L.A. Fialkow. Recursiveness, Positivity, and Truncated Moment Problems. Houston Journal of Mathematics, 7: 60 6, 99. J. De Leeuw. Block Relaation Algorithms in Statistics. In H.H. Bock, W. Lenski, and M.M. Richter, editors, Information Systems and Data Analysis, pages 08, Berlin, 99. Springer Verlag. URL chapters/deleeuw_c_9c.pdf. J. De Leeuw. Polynomial etrapolation to accelerate fied point algorithms. Preprint Series, UCLA Department of Statistics, Los Angeles, CA, 008. URL ~deleeuw/janspubs/008/reports/deleeuw_r_08i.pdf. J. De Leeuw. The Multiway Package. Preprint Series, UCLA Department of Statistics, Los Angeles, CA, 008. URL reports/deleeuw_r_08j.pdf. J. De Leeuw. Multivariate Cumulants in R. Unpublished, 0. URL notes/deleeuw_u_a.pdf. J. De Leeuw. Single Factor Polynomial Component Analysis. Unpublished, 0. URL janspubs/0/notes/deleeuw_u_a.pdf. J. De Leeuw and I. Kukuyeva. Component Analysis Using Multivariate Cumulants. Unpublished, 0. URL notes/deleeuw_kukuyeva_u_.pdf. J.J. Froh, J. Fan, R.A. Emmons, G. Bono, E. S. Huebner, and P. Watkins. Measuring Gratitude in Youth: Assessing the Psychometric Properties of Adult Gratitude Scales in Children and Adolescents. Psychological Assessment, :, 0. J. Hemelrijk. Underlining Random Variables. Statistica Neerlandica, 0: 7, 966.
15 POLYNOMIAL SINGLE COMPONENT ANALYSIS A. Mooijaart and P. Bentler. Random Polynomial Factor Analysis. In E. Diday et al., editor, Data Analysis and Informatics, volume IV, pages 0, 986. A. Mooijaart and P.M. Bentler. An Alternative Approach for Nonlinear Latent Variable Models. Structural Equation Modeling, 7: 7 7, 00. H.P. Mulholland and C.A. Rogers. Representation Theorems for Distribution Functions. Proceedings of the London Mathematical Society, 8:77, 98. E. Neuman. Moments and Fourier Transforms of B-splines. Journal of Computational and Applied Mathematics, 7: 6, 98. R Development Core Team. R: A Language and Environment for Statistical Computing. R Foundation for Statistical Computing, Vienna, Austria, 0. URL ISBN Y. Takane, F.W. Young, and J. De Leeuw. Nonmetric individual differences in multidimensional scaling: An alternating least squares method with optimal scaling features. Psychometrika, :7 67, 977. URL janspubs/977/articles/takane_young_deleeuw_a_77. pdf. I. Yalcin and Y. Amemiya. Nonlinear Factor Analysis as a Statistical Method. Statistical Science, 6:7 9, 00. W. I. Zangwill. Nonlinear Programming: a Unified Approach. Prentice-Hall, Englewood-Cliffs, N.J., 969. A. Zeileis, C. Strobl, and F. Wickelmaier. psychotools: Infrastructure for Psychometric Modeling, 0. URL R-project.org/package=psychotools. R package version 0.-.
16 6 JAN DE LEEUW Appendi A. Code Code Segment Raw Moments up to Order p rawmomentsuptop <- function (, p = ) { n <- nrow () m <- ncol () if (p == ) { return (c (, apply (,, mean))) 6 } 7 y <- array (0, rep (m +, p)) 8 for (i in : n) { 9 i <- c (, [i, ]) 0 z <- i for (s in :p) { z <- outer (z, i) } y <- y + z } 6 return (y / n) 7 }
17 POLYNOMIAL SINGLE COMPONENT ANALYSIS 7 Code Segment Moments of Standard Normal normalevenmoment <- function (n) { if ((n %% ) == ) { stop ("even moments only") } return (doublefactorial (n - )) 6 } 7 8 doublefactorial <- function (n) { 9 if ((n %% ) == 0) { 0 stop ("odd argument only") } k <- (n + ) / return (prod ( * ( : k) - )) } 6 normalmoments <- function (n) { 7 m <- rep (0, n) 8 ind <- which((( : n) %% ) == 0) 9 m [ind] <- sapply (ind, normalevenmoment) 0 return (c (, m)) }
18 8 JAN DE LEEUW Code Segment Kernel from Moments require("apl") makemomentkernel <- function (s, r, mu) { if (length (mu)!= ((s * r) + )) { stop ("vector of moments wrong length") 6 } 7 c <- array(0, rep (s +, r)) 8 sr <- (s + ) ^ r 9 for (i in : sr) { 0 l <- aplencode (i, rep (s +, r)) - k <- sum (l) c [i] <- mu [k + ] } return (c) }
19 POLYNOMIAL SINGLE COMPONENT ANALYSIS 9 Code Segment Kernel Indicators makeindicatorbasismatri <- function (s, r) { ii <- replist (array (0, rep (s +, r)), (s * r) + ) sr <- (s + ) ^ r for (i in ( : sr)) { l <- aplencode (i, rep (s +, r)) - 6 k <- sum (l) 7 ii [[k + ]] [i] <- 8 } 9 return (ii) 0 }
20 0 JAN DE LEEUW Code Segment Moments of B-splines makebsplinemoments <- function (lup, t) { kup <- length (t) - m <- matri (0, kup, lup + ) for (l in (0 : lup)) { s <- (t [] ^ (l + )) - (t [] ^ (l + )) 6 s <- (l + ) * (t [] - t []) 7 m [, l + ] <- s / s 8 for (k in ( : kup)) { 9 ss <- 0 0 for (j in (0 : l)) { s <- t [k + ] ^ (l - j) s <- choose (k + j -, j) s <- m [k -, j + ] ss <- ss + s * s * s } 6 m [k, l + ] <- ss / choose (k + l, k) 7 } 8 } 9 return (m [kup, ]) 0 } makebsplinemomentmatri <- function (lup, t, order) { k <- length (t) - order m <- matri (0, k, lup + ) for (i in ( : k)) { 6 m[i, ] <- makebsplinemoments (lup, t [i : (i + order)]) 7 } 8 return (m) 9 }
21 POLYNOMIAL SINGLE COMPONENT ANALYSIS Code Segment 6 CCD Algorithm for Fied/Free Kernel tuckerccd <- function (data, r, s, w = array (, dim (a)), kertype = "fied", order =, knots = -6:6, inma =, ineps = e-, outma = 00, outeps = e-, verbose = ) { a <- rawmomentsuptop (data, r) vassign (i, m, p, q, k, values = kernelinitial (s, r, k, kertype, knots, order)) vassign (b, otel, values = list (tuckerinitial (a, k), )) oloss <- tuckervalue (a, k, w, b) 6 repeat { 7 vassign (itel, bold, values = list (, b)) 8 repeat { 9 bnew <- oneccdcycle (a, k, w, bold) 0 chng <- ma (abs (bnew - bold)) if (verbose > ) cat ("Ineration: ", formatc (itel,, ), "IneChange: ", formatc (chng, 6, 0, " f"), "\n") if ( (chng < ineps) (itel == inma)) break () 6 vassign (bold, itel, values = list (bnew, itel + )) 7 } 8 b <- bnew 9 loss <- tuckervalue (a, k, w, b) 0 vassign (k, p, q, values = kernelupdate (a, k, i, m, p, w, b, kertype)) nloss <- tuckervalue (a, k, w, b) if (verbose >= ) cat ("Iteration:", formatc (otel,, ), "OLoss: ", formatc (oloss, 6, 0, "f"), "BLoss: ", formatc (loss, 6, 0, "f"), 6 "Kloss: ", formatc (nloss, 6, 0, "f"), "\ n") 7 if (((oloss - nloss) < outeps) (otel == outma)) 8 break () 9 vassign (oloss, otel, values = list (nloss, otel + )) 0 } return (list(b = b, otel = otel, tloss = nloss, k = k, p = p, q = q)) }
22 JAN DE LEEUW Code Segment 7 Tucker Subroutines tuckervalue <- function (a, k, w, b) { r <- length (dim (a)) res <- as.vector (a) - arrkronecker (replist (b, r)) % *% as.vector (k) return (sqrt (sum (as.vector (w) * res * res) / sum (w) } 6 )) 7 tuckerinitial <- function (a, k) { 8 r <- length (dim (a)) 9 s <- dim (a)[] 0 t <- dim (k) [] h <- asub (a, c (replist ( : s, ), replist (, r - ) )) m <- (h - outer (h [, ], h[, ])) [-, -] e <- eigen (m) v <- asub (k, c (replist ( : t, ), replist (, r - ) )) p <- nrow (v) - 6 u <- (v - outer (v [, ], v[, ])) [-, -] 7 if (p == ) { 8 <- as.matri (e $ vectors [, ]) %*% sqrt (e $ 9 } else { values []) 0 <- e $ vectors [, : p] %*% diag (sqrt (e $ } values [ : p])) bb <- %*% solve (t (chol (u))) bv <- h [-, ] - bb %*% v [-, ] return (rbind (c (, rep (0, t - )), cbind (bv, bb))) }
23 POLYNOMIAL SINGLE COMPONENT ANALYSIS Code Segment 8 More Tucker Subroutines oneccdcycle <- function (a, k, w, b) { p <- dim (k) [] m <- dim (a) [] f <- function (, j, s) tuckervalue (a, k, w, addelement (b, j, s, )) loss <- tuckervalue (a, k, w, b) 6 for (j in : m) { 7 for (s in : p) { 8 oo <- optimize (f, c(-, ), j, s) 9 ls <- oo $ objective 0 ip <- oo $ minimum if (ls > loss) { net () } loss <- ls b <- addelement (b, j, s, ip) 6 } 7 } 8 return (b) 9 } 0 addelement <- function (, i, j, s) { [i, j] <- [i, j] + s return () }
24 JAN DE LEEUW Code Segment 9 Tucker Auilaries arrkronecker <- function (, fun=" * ") { nmat <- length () if (nmat == 0) { stop("empty argument in arrkronecker") } 6 res <- [[]] 7 if (length () == ) { 8 return (res) 9 } 0 for (i in : nmat) { res <- kronecker (res, [[i]], fun) } return (res) } 6 replist <- function (, n) { 7 z <- list() 8 if (n == 0) { 9 return (z) 0 } for (i in : n) z <- c (z, list ()) return (z) } 6 vassign <- function(..., values, envir=parent.frame()) { 7 vars <- as.character(substitute(...())) 8 values <- rep(values, length.out=length(vars)) 9 for(i in seq_along(vars)) { 0 assign(vars[[i]], values[[i]], envir) } }
25 POLYNOMIAL SINGLE COMPONENT ANALYSIS Code Segment 0 Plotting Polynomials plotpol <- function (, title) { m <- nrow () s <- seq (-.,., length = 00) h <- seq (-.,., length = 0) v <- matri (0, m -, 00) 6 u <- matri (0, m -, 0) 7 cols <- rainbow (m - ) 8 for (i in : (m - )) { 9 v[i, ] <- predict (polynomial ([i +, ]), s) 0 u[i, ] <- predict (polynomial ([i +, ]), h) } yma <- 7 ymin <- plot (0, lim = c (-.,.), ylim = c (ymin, yma), main = title, lab = "", ylab = "", type = "n") for (i in : (m - )) { 6 lines (s, v[i, ]) 7 points (cbind (h, u[i, ]), pch = as.character (i), col = cols[i]) 8 } 9 }
26 6 JAN DE LEEUW Code Segment Frank-Wolfe Algorithm qpfrankwolfe <- function (, y, w = rep (, length (y)), b = rep (, ncol ()) / ncol (), itma = 00, eps = e-6, verbose = FALSE) { w <- crossprod (, w * ) fold <- Inf itel <- repeat { 6 res <- c (y - %*% b) 7 fnew<- sum (w * res ^ ) / 8 g <- - colsums (w * res * ) 9 mg <- min (g) 0 ig <- which.min (g) bnew <- rep (0, length (b)) bnew [ig] <- d <- bnew - b tau <- sum (d * crossprod (, w * res)) / sum (w * outer (d, d)) tau <- min (, ma (0, tau)) 6 b <- b + tau * d 7 if (verbose) { 8 cat ("itel: ", formatc (itel,, ), 9 "fold: ", formatc (fold, 0,, "f"), 0 "fnew: ", formatc (fnew, 0,, "f"), } "\n") if (((fold - fnew) < eps) (itel == itma)) { break () } itel <- itel + 6 fold <- fnew 7 } 8 return (list (b = b, f = fnew)) 9 }
27 POLYNOMIAL SINGLE COMPONENT ANALYSIS 7
28 8 JAN DE LEEUW Appendi B. Figures fied (a) fied (b) fied (c) fied (d) fied (e) fied (f) fied (g) fied (h) fied (i) fied (j) fied (k) fied (l) Figure. Twelve solutions with a fied normal kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and, (l) sr.
29 POLYNOMIAL SINGLE COMPONENT ANALYSIS free (a) free (b) free (c) free (d) free (e) free (f) free (g) free (h) free (i) free (j) free (k) free (l) Figure. Twelve solutions with a free kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and, (l) sr.
30 0 JAN DE LEEUW moment (a) moment (b) moment (c) moment (d) moment (e) moment (f) moment (g) moment (h) moment (i) moment (j) moment (k) moment (l) Figure. Twelve solutions with a moment kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and,?? sr.
31 POLYNOMIAL SINGLE COMPONENT ANALYSIS cdf cdf cdf (a) (b) (c) cdf cdf cdf (d) (e) (f) cdf cdf cdf (g) (h) (i) cdf cdf cdf (j) (k) (l) Figure. Twelve solutions with a CDF kernel. (a) sr; (b) sr; (c) sr; (d) sr; (e) sr; (f) sr; (g) sr; (h) sr; (i) sr; (j) sr; (k) sr; and, (l) sr.
32 JAN DE LEEUW Appendi C. Tables fied r = r = r = s = s = s = s = free r = r = r = s = s = s = s = moment r = r = r = s = s = s = s = cdf r = r = r = s = s = s = s = Table. Minimum loss function values for different kernel types, polynomial degrees s and moment orders r.
33 POLYNOMIAL SINGLE COMPONENT ANALYSIS fied r = r = r = s = 6 s = s = s = 0 00 free r = r = r = s = 00 s = s = 0 00 s = moment r = r = r = s = 00 s = s = s = cdf r = r = r = s = s = s = 8 00 s = 00 Table. Number of iterations for different kernel types, polynomial degrees s and moment orders r. Department of Statistics, University of California, Los Angeles, CA address, Jan de Leeuw: deleeuw@stat.ucla.edu URL, Jan de Leeuw:
MULTIVARIATE CUMULANTS IN R. 1. Introduction
MULTIVARIATE CUMULANTS IN R JAN DE LEEUW Abstract. We give algorithms and R code to compute all multivariate cumulants up to order p of m variables. 1. Introduction Multivariate cumulants are of increasing
More informationHOMOGENEITY ANALYSIS USING EUCLIDEAN MINIMUM SPANNING TREES 1. INTRODUCTION
HOMOGENEITY ANALYSIS USING EUCLIDEAN MINIMUM SPANNING TREES JAN DE LEEUW 1. INTRODUCTION In homogeneity analysis the data are a system of m subsets of a finite set of n objects. The purpose of the technique
More informationLEAST SQUARES ORTHOGONAL POLYNOMIAL COMPONENT ANALYSIS IN R. 1. Introduction
LEAST SQUARES ORTHOGONAL POLYNOMIAL COMPONENT ANALYSIS IN R JAN DE LEEUW AND KEKONA KAI SORENSON Abstract. We present a simple algorithm (with code in R) to fit multivariate polynomial components or factors
More informationCOMPONENT ANALYSIS USING MULTIVARIATE CUMULANTS
COMPONENT ANALYSIS USING MULTIVARIATE CUMULANTS JAN DE LEEUW AND IRINA KUKUYEVA Abstract. This paper discusses several different algorithms to fit linear component analysis approximations to multivariate
More informationALGORITHM CONSTRUCTION BY DECOMPOSITION 1. INTRODUCTION. The following theorem is so simple it s almost embarassing. Nevertheless
ALGORITHM CONSTRUCTION BY DECOMPOSITION JAN DE LEEUW ABSTRACT. We discuss and illustrate a general technique, related to augmentation, in which a complicated optimization problem is replaced by a simpler
More informationLEAST SQUARES METHODS FOR FACTOR ANALYSIS. 1. Introduction
LEAST SQUARES METHODS FOR FACTOR ANALYSIS JAN DE LEEUW AND JIA CHEN Abstract. Meet the abstract. This is the abstract. 1. Introduction Suppose we have n measurements on each of m variables. Collect these
More informationMAJORIZATION OF DIFFERENT ORDER
MAJORIZATION OF DIFFERENT ORDER JAN DE LEEUW 1. INTRODUCTION The majorization method [De Leeuw, 1994; Heiser, 1995; Lange et al., 2000] for minimization of real valued loss functions has become very popular
More informationover the parameters θ. In both cases, consequently, we select the minimizing
MONOTONE REGRESSION JAN DE LEEUW Abstract. This is an entry for The Encyclopedia of Statistics in Behavioral Science, to be published by Wiley in 200. In linear regression we fit a linear function y =
More informationMAXIMUM LIKELIHOOD IN GENERALIZED FIXED SCORE FACTOR ANALYSIS 1. INTRODUCTION
MAXIMUM LIKELIHOOD IN GENERALIZED FIXED SCORE FACTOR ANALYSIS JAN DE LEEUW ABSTRACT. We study the weighted least squares fixed rank approximation problem in which the weight matrices depend on unknown
More informationNonlinear Support Vector Machines through Iterative Majorization and I-Splines
Nonlinear Support Vector Machines through Iterative Majorization and I-Splines P.J.F. Groenen G. Nalbantov J.C. Bioch July 9, 26 Econometric Institute Report EI 26-25 Abstract To minimize the primal support
More informationQUADRATIC MAJORIZATION 1. INTRODUCTION
QUADRATIC MAJORIZATION JAN DE LEEUW 1. INTRODUCTION Majorization methods are used extensively to solve complicated multivariate optimizaton problems. We refer to de Leeuw [1994]; Heiser [1995]; Lange et
More informationBracketing an Optima in Univariate Optimization
Bracketing an Optima in Univariate Optimization Pritibhushan Sinha Quantitative Methods & Operations Management Area Indian Institute of Management Kozhikode Kozhikode 673570 Kerala India Email: pritibhushan.sinha@iimk.ac.in
More information9. Interpretations, Lifting, SOS and Moments
9-1 Interpretations, Lifting, SOS and Moments P. Parrilo and S. Lall, CDC 2003 2003.12.07.04 9. Interpretations, Lifting, SOS and Moments Polynomial nonnegativity Sum of squares (SOS) decomposition Eample
More informationFACTOR ANALYSIS AS MATRIX DECOMPOSITION 1. INTRODUCTION
FACTOR ANALYSIS AS MATRIX DECOMPOSITION JAN DE LEEUW ABSTRACT. Meet the abstract. This is the abstract. 1. INTRODUCTION Suppose we have n measurements on each of taking m variables. Collect these measurements
More informationScore-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables
Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables Achim Zeileis, Edgar C. Merkle, Ting Wang http://eeecon.uibk.ac.at/~zeileis/ Overview Motivation Framework Score-based
More informationTail Approximation of the Skew-Normal by the Skew-Normal Laplace: Application to Owen s T Function and the Bivariate Normal Distribution
Journal of Statistical and Econometric ethods vol. no. 3 - ISS: 5-557 print version 5-565online Scienpress Ltd 3 Tail Approimation of the Skew-ormal by the Skew-ormal Laplace: Application to Owen s T Function
More informationSimultaneous Diagonalization of Positive Semi-definite Matrices
Simultaneous Diagonalization of Positive Semi-definite Matrices Jan de Leeuw Version 21, May 21, 2017 Abstract We give necessary and sufficient conditions for solvability of A j = XW j X, with the A j
More informationLeast squares multidimensional scaling with transformed distances
Least squares multidimensional scaling with transformed distances Patrick J.F. Groenen 1, Jan de Leeuw 2 and Rudolf Mathar 3 1 Department of Data Theory, University of Leiden P.O. Box 9555, 2300 RB Leiden,
More informationPower EP. Thomas Minka Microsoft Research Ltd., Cambridge, UK MSR-TR , October 4, Abstract
Power EP Thomas Minka Microsoft Research Ltd., Cambridge, UK MSR-TR-2004-149, October 4, 2004 Abstract This note describes power EP, an etension of Epectation Propagation (EP) that makes the computations
More information1. Data Overview: North Direction. Total observations for each intersection: 96,432 Missing Percentages:
Preliminary Report LOOP IMPUTATION OF HOURLY DATA ON A SECTION OF THE I-5 (BETWEEN ROUTES 14 AND 99) VIA FUNCTIONAL PRINCIPAL COMPONENT ANALYSIS JAN DE LEEUW, IRINA KUKUYEVA Abstract. We present the results
More informationStatistical Geometry Processing Winter Semester 2011/2012
Statistical Geometry Processing Winter Semester 2011/2012 Linear Algebra, Function Spaces & Inverse Problems Vector and Function Spaces 3 Vectors vectors are arrows in space classically: 2 or 3 dim. Euclidian
More informationUCLA Department of Statistics Papers
UCLA Department of Statistics Papers Title Rate of Convergence of the Arithmetic-Geometric Mean Process Permalink https://escholarship.org/uc/item/5kx0j9nm Author de Leeuw, Jan Publication Date 008-0-09
More informationCENTERING IN MULTILEVEL MODELS. Consider the situation in which we have m groups of individuals, where
CENTERING IN MULTILEVEL MODELS JAN DE LEEUW ABSTRACT. This is an entry for The Encyclopedia of Statistics in Behavioral Science, to be published by Wiley in 2005. Consider the situation in which we have
More informationFIXED-RANK APPROXIMATION WITH CONSTRAINTS
FIXED-RANK APPROXIMATION WITH CONSTRAINTS JAN DE LEEUW AND IRINA KUKUYEVA Abstract. We discuss fixed-rank weighted least squares approximation of a rectangular matrix with possibly singular Kronecker weights
More informationNumber of cases (objects) Number of variables Number of dimensions. n-vector with categorical observations
PRINCALS Notation The PRINCALS algorithm was first described in Van Rickevorsel and De Leeuw (1979) and De Leeuw and Van Rickevorsel (1980); also see Gifi (1981, 1985). Characteristic features of PRINCALS
More informationApproximate inference, Sampling & Variational inference Fall Cours 9 November 25
Approimate inference, Sampling & Variational inference Fall 2015 Cours 9 November 25 Enseignant: Guillaume Obozinski Scribe: Basile Clément, Nathan de Lara 9.1 Approimate inference with MCMC 9.1.1 Gibbs
More informationNON-NEGATIVE MATRIX FACTORIZATION FOR PARAMETER ESTIMATION IN HIDDEN MARKOV MODELS. Balaji Lakshminarayanan and Raviv Raich
NON-NEGATIVE MATRIX FACTORIZATION FOR PARAMETER ESTIMATION IN HIDDEN MARKOV MODELS Balaji Lakshminarayanan and Raviv Raich School of EECS, Oregon State University, Corvallis, OR 97331-551 {lakshmba,raich@eecs.oregonstate.edu}
More informationA GENERALIZATION OF TAKANE'S ALGORITHM FOR DEDICOM. Yosmo TAKANE
PSYCHOMETR1KA--VOL. 55, NO. 1, 151--158 MARCH 1990 A GENERALIZATION OF TAKANE'S ALGORITHM FOR DEDICOM HENK A. L. KIERS AND JOS M. F. TEN BERGE UNIVERSITY OF GRONINGEN Yosmo TAKANE MCGILL UNIVERSITY JAN
More informationLecture 3: Latent Variables Models and Learning with the EM Algorithm. Sam Roweis. Tuesday July25, 2006 Machine Learning Summer School, Taiwan
Lecture 3: Latent Variables Models and Learning with the EM Algorithm Sam Roweis Tuesday July25, 2006 Machine Learning Summer School, Taiwan Latent Variable Models What to do when a variable z is always
More informationCSCI-6971 Lecture Notes: Probability theory
CSCI-6971 Lecture Notes: Probability theory Kristopher R. Beevers Department of Computer Science Rensselaer Polytechnic Institute beevek@cs.rpi.edu January 31, 2006 1 Properties of probabilities Let, A,
More informationLecture 9: MATH 329: Introduction to Scientific Computing
Lecture 9: MATH 329: Introduction to Scientific Computing Sanjeena Dang Department of Mathematical Sciences Vectors and Matrices in R Numeric vector " and matrix objects in R are a close match to mathematical
More informationGeometric Modeling Summer Semester 2010 Mathematical Tools (1)
Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric
More informationEquivalence of Minimal l 0 and l p Norm Solutions of Linear Equalities, Inequalities and Linear Programs for Sufficiently Small p
Equivalence of Minimal l 0 and l p Norm Solutions of Linear Equalities, Inequalities and Linear Programs for Sufficiently Small p G. M. FUNG glenn.fung@siemens.com R&D Clinical Systems Siemens Medical
More informationM.S. Project Report. Efficient Failure Rate Prediction for SRAM Cells via Gibbs Sampling. Yamei Feng 12/15/2011
.S. Project Report Efficient Failure Rate Prediction for SRA Cells via Gibbs Sampling Yamei Feng /5/ Committee embers: Prof. Xin Li Prof. Ken ai Table of Contents CHAPTER INTRODUCTION...3 CHAPTER BACKGROUND...5
More informationReview of Optimization Basics
Review of Optimization Basics. Introduction Electricity markets throughout the US are said to have a two-settlement structure. The reason for this is that the structure includes two different markets:
More informationAN ITERATIVE MAXIMUM EIGENVALUE METHOD FOR SQUARED DISTANCE SCALING 1. INTRODUCTION
AN ITERATIVE MAXIMUM EIGENVALUE METHOD FOR SQUARED DISTANCE SCALING JAN DE LEEUW ABSTRACT. Meet the abstract. This is the abstract. 1. INTRODUCTION The proble we consider in this paper is a etric ultidiensional
More informationMACHINE LEARNING ADVANCED MACHINE LEARNING
MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 2 2 MACHINE LEARNING Overview Definition pdf Definition joint, condition, marginal,
More informationSection - 9 GRAPHS. (a) y f x (b) y f x. (c) y f x (d) y f x. (e) y f x (f) y f x k. (g) y f x k (h) y kf x. (i) y f kx. [a] y f x to y f x
44 Section - 9 GRAPHS In this section, we will discuss graphs and graph-plotting in more detail. Detailed graph plotting also requires a knowledge of derivatives. Here, we will be discussing some general
More informationOn Simulations form the Two-Parameter. Poisson-Dirichlet Process and the Normalized. Inverse-Gaussian Process
On Simulations form the Two-Parameter arxiv:1209.5359v1 [stat.co] 24 Sep 2012 Poisson-Dirichlet Process and the Normalized Inverse-Gaussian Process Luai Al Labadi and Mahmoud Zarepour May 8, 2018 ABSTRACT
More informationMULTIVARIATE ANALYSIS OF ORDINAL DATA USING PROBITS
MULTIVARIATE ANALYSIS OF ORDINAL DATA USING PROBITS JAN DE LEEUW ABSTRACT. We discuss algorithms to fit component and factor representations to multivariate multicategory ordinal data, using a probit threshold
More informationIntroduction. x 7. Partial fraction decomposition x 7 of. x 7. x 3 1. x 2. Decomposition of N x /D x into Partial Fractions. N x. D x polynomial N 1 x
333202_0704.qd 2/5/05 9:43 M Page 533 Section 7.4 Partial Fractions 533 7.4 Partial Fractions What you should learn Recognize partial fraction decompositions of rational epressions. Find partial fraction
More informationApproximate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem
Approimate formulas for the Point-to-Ellipse and for the Point-to-Ellipsoid Distance Problem ALEXEI UTESHEV St.Petersburg State University Department of Applied Mathematics Universitetskij pr. 35, 198504
More informationLeast squares contour alignment
Least squares contour alignment Ivan Markovsky and Sasan Mahmoodi School of Electronics and Computer Science University of Southampton, SO7 BJ, UK Emails: {im,sm3}@ecssotonacuk Abstract The contour alignment
More informationLecture 3: Pattern Classification. Pattern classification
EE E68: Speech & Audio Processing & Recognition Lecture 3: Pattern Classification 3 4 5 The problem of classification Linear and nonlinear classifiers Probabilistic classification Gaussians, mitures and
More informationPrecalculus Notes: Functions. Skill: Solve problems using the algebra of functions.
Skill: Solve problems using the algebra of functions. Modeling a Function: Numerical (data table) Algebraic (equation) Graphical Using Numerical Values: Look for a common difference. If the first difference
More information2 Statistical Estimation: Basic Concepts
Technion Israel Institute of Technology, Department of Electrical Engineering Estimation and Identification in Dynamical Systems (048825) Lecture Notes, Fall 2009, Prof. N. Shimkin 2 Statistical Estimation:
More informationMultilevel Analysis, with Extensions
May 26, 2010 We start by reviewing the research on multilevel analysis that has been done in psychometrics and educational statistics, roughly since 1985. The canonical reference (at least I hope so) is
More informationGaussians. Hiroshi Shimodaira. January-March Continuous random variables
Cumulative Distribution Function Gaussians Hiroshi Shimodaira January-March 9 In this chapter we introduce the basics of how to build probabilistic models of continuous-valued data, including the most
More information18.303: Introduction to Green s functions and operator inverses
8.33: Introduction to Green s functions and operator inverses S. G. Johnson October 9, 2 Abstract In analogy with the inverse A of a matri A, we try to construct an analogous inverse  of differential
More informationScore-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables
Score-Based Tests of Measurement Invariance with Respect to Continuous and Ordinal Variables Achim Zeileis, Edgar C. Merkle, Ting Wang http://eeecon.uibk.ac.at/~zeileis/ Overview Motivation Framework Score-based
More information9 - Markov processes and Burt & Allison 1963 AGEC
This document was generated at 8:37 PM on Saturday, March 17, 2018 Copyright 2018 Richard T. Woodward 9 - Markov processes and Burt & Allison 1963 AGEC 642-2018 I. What is a Markov Chain? A Markov chain
More informationDependence. Practitioner Course: Portfolio Optimization. John Dodson. September 10, Dependence. John Dodson. Outline.
Practitioner Course: Portfolio Optimization September 10, 2008 Before we define dependence, it is useful to define Random variables X and Y are independent iff For all x, y. In particular, F (X,Y ) (x,
More informationMath 2412 Activity 1(Due by EOC Sep. 17)
Math 4 Activity (Due by EOC Sep. 7) Determine whether each relation is a unction.(indicate why or why not.) Find the domain and range o each relation.. 4,5, 6,7, 8,8. 5,6, 5,7, 6,6, 6,7 Determine whether
More information2.29 Numerical Fluid Mechanics Spring 2015 Lecture 4
2.29 Spring 2015 Lecture 4 Review Lecture 3 Truncation Errors, Taylor Series and Error Analysis Taylor series: 2 3 n n i1 i i i i i n f( ) f( ) f '( ) f ''( ) f '''( )... f ( ) R 2! 3! n! n1 ( n1) Rn f
More informationCANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM
CANONICAL LOSSLESS STATE-SPACE SYSTEMS: STAIRCASE FORMS AND THE SCHUR ALGORITHM Ralf L.M. Peeters Bernard Hanzon Martine Olivi Dept. Mathematics, Universiteit Maastricht, P.O. Box 616, 6200 MD Maastricht,
More informationExact and Approximate Numbers:
Eact and Approimate Numbers: The numbers that arise in technical applications are better described as eact numbers because there is not the sort of uncertainty in their values that was described above.
More informationThe stability of limit cycles in nonlinear systems
Nonlinear Dyn 009) 56: 69 75 DOI 10.1007/s11071-008-9398-3 O R I G I NA L PA P E R The stability of limit cycles in nonlinear systems Ali Ghaffari Masayoshi Tomizuka Reza A. Soltan Received: 1 March 007
More informationSPSS, University of Texas at Arlington. Topics in Machine Learning-EE 5359 Neural Networks
Topics in Machine Learning-EE 5359 Neural Networks 1 The Perceptron Output: A perceptron is a function that maps D-dimensional vectors to real numbers. For notational convenience, we add a zero-th dimension
More informationMachine Learning Lecture 3
Announcements Machine Learning Lecture 3 Eam dates We re in the process of fiing the first eam date Probability Density Estimation II 9.0.207 Eercises The first eercise sheet is available on L2P now First
More informationInexact Solution of NLP Subproblems in MINLP
Ineact Solution of NLP Subproblems in MINLP M. Li L. N. Vicente April 4, 2011 Abstract In the contet of conve mied-integer nonlinear programming (MINLP, we investigate how the outer approimation method
More informationFactor Analysis. Qian-Li Xue
Factor Analysis Qian-Li Xue Biostatistics Program Harvard Catalyst The Harvard Clinical & Translational Science Center Short course, October 7, 06 Well-used latent variable models Latent variable scale
More informationMobile Robot Localization
Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations
More informationPackage depth.plot. December 20, 2015
Package depth.plot December 20, 2015 Type Package Title Multivariate Analogy of Quantiles Version 0.1 Date 2015-12-19 Maintainer Could be used to obtain spatial depths, spatial ranks and outliers of multivariate
More informationThe incomplete gamma functions. Notes by G.J.O. Jameson. These notes incorporate the Math. Gazette article [Jam1], with some extra material.
The incomplete gamma functions Notes by G.J.O. Jameson These notes incorporate the Math. Gazette article [Jam], with some etra material. Definitions and elementary properties functions: Recall the integral
More informationPerturbation Theory for Variational Inference
Perturbation heor for Variational Inference Manfred Opper U Berlin Marco Fraccaro echnical Universit of Denmark Ulrich Paquet Apple Ale Susemihl U Berlin Ole Winther echnical Universit of Denmark Abstract
More informationMACHINE LEARNING ADVANCED MACHINE LEARNING
MACHINE LEARNING ADVANCED MACHINE LEARNING Recap of Important Notions on Estimation of Probability Density Functions 22 MACHINE LEARNING Discrete Probabilities Consider two variables and y taking discrete
More informationON MODEL SELECTION FOR STATE ESTIMATION FOR NONLINEAR SYSTEMS. Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof
ON MODEL SELECTION FOR STATE ESTIMATION FOR NONLINEAR SYSTEMS Robert Bos,1 Xavier Bombois Paul M. J. Van den Hof Delft Center for Systems and Control, Delft University of Technology, Mekelweg 2, 2628 CD
More information3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:
3.3 Gradient Vector and Jacobian Matri 3 3.3 Gradient Vector and Jacobian Matri Overview: Differentiable functions have a local linear approimation. Near a given point, local changes are determined by
More informationBayesian Approach 2. CSC412 Probabilistic Learning & Reasoning
CSC412 Probabilistic Learning & Reasoning Lecture 12: Bayesian Parameter Estimation February 27, 2006 Sam Roweis Bayesian Approach 2 The Bayesian programme (after Rev. Thomas Bayes) treats all unnown quantities
More information5.8 Chebyshev Approximation
84 Chapter 5. Evaluation of Functions appropriate setting is ld=, and the value of the derivative is the accumulated sum divided by the sampling interval h. CITED REFERENCES AND FURTHER READING: Dennis,
More informationM 340L CS Homework Set 1
M 340L CS Homework Set 1 Solve each system in Problems 1 6 by using elementary row operations on the equations or on the augmented matri. Follow the systematic elimination procedure described in Lay, Section
More informationA Rectangular Trust Region Dogleg Approach for Unconstrained and Bound Constrained Nonlinear Optimization
A Rectangular Trust Region Dogleg Approach for Unconstrained and Bound onstrained onlinear Optimization. VOGLIS, I.E. LAGARIS Department of omputer Science University of Ioannina GREEE f g Abstract: -
More informationLECTURE NOTES ON QUANTUM COMPUTATION. Cornell University, Physics , CS 483; Spring, 2005 c 2006, N. David Mermin
LECTURE NOTES ON QUANTUM COMPUTATION Cornell University, Physics 481-681, CS 483; Spring, 2005 c 2006, N. David Mermin IV. Searching with a Quantum Computer Last revised 3/30/06 Suppose you know that eactly
More informationPartitioning variation in multilevel models.
Partitioning variation in multilevel models. by Harvey Goldstein, William Browne and Jon Rasbash Institute of Education, London, UK. Summary. In multilevel modelling, the residual variation in a response
More informationOutline. Random Variables. Examples. Random Variable
Outline Random Variables M. Sami Fadali Professor of Electrical Engineering University of Nevada, Reno Random variables. CDF and pdf. Joint random variables. Correlated, independent, orthogonal. Correlation,
More informationREGRESSION, DISCRIMINANT ANALYSIS, AND CANONICAL CORRELATION ANALYSIS WITH HOMALS 1. MORALS
REGRESSION, DISCRIMINANT ANALYSIS, AND CANONICAL CORRELATION ANALYSIS WITH HOMALS JAN DE LEEUW ABSTRACT. It is shown that the homals package in R can be used for multiple regression, multi-group discriminant
More informationApplication of modified Leja sequences to polynomial interpolation
Special Issue for the years of the Padua points, Volume 8 5 Pages 66 74 Application of modified Leja sequences to polynomial interpolation L. P. Bos a M. Caliari a Abstract We discuss several applications
More informationGaussian Process Vine Copulas for Multivariate Dependence
Gaussian Process Vine Copulas for Multivariate Dependence José Miguel Hernández-Lobato 1,2 joint work with David López-Paz 2,3 and Zoubin Ghahramani 1 1 Department of Engineering, Cambridge University,
More informationLecture 2. Tensor Unfoldings. Charles F. Van Loan
From Matrix to Tensor: The Transition to Numerical Multilinear Algebra Lecture 2. Tensor Unfoldings Charles F. Van Loan Cornell University The Gene Golub SIAM Summer School 2010 Selva di Fasano, Brindisi,
More informationNumerical Methods. Root Finding
Numerical Methods Solving Non Linear 1-Dimensional Equations Root Finding Given a real valued function f of one variable (say ), the idea is to find an such that: f() 0 1 Root Finding Eamples Find real
More informationFilter-Generating Systems
24 IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II: ANALOG AND DIGITAL SIGNAL PROCESSING, VOL. 47, NO. 3, MARCH 2000 Filter-Generating Systems Saed Samadi, Akinori Nishihara, and Hiroshi Iwakura Abstract
More informationProjection, Inference, and Consistency
Projection, Inference, and Consistency John Hooker Carnegie Mellon University IJCAI 2016, New York City A high-level presentation. Don t worry about the details. 2 Projection as a Unifying Concept Projection
More informationA New Lower Bound Technique for Quantum Circuits without Ancillæ
A New Lower Bound Technique for Quantum Circuits without Ancillæ Debajyoti Bera Abstract We present a technique to derive depth lower bounds for quantum circuits. The technique is based on the observation
More informationQuadratic Programming Relaxations for Metric Labeling and Markov Random Field MAP Estimation
Quadratic Programming Relaations for Metric Labeling and Markov Random Field MAP Estimation Pradeep Ravikumar John Lafferty School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213,
More informationMultidimensional partitions of unity and Gaussian terrains
and Gaussian terrains Richard A. Bale, Jeff P. Grossman, Gary F. Margrave, and Michael P. Lamoureu ABSTRACT Partitions of unity play an important rôle as amplitude-preserving windows for nonstationary
More informationLecture 4b. Bessel functions. Introduction. Generalized factorial function. 4b.1. Using integration by parts it is easy to show that
4b. Lecture 4b Using integration by parts it is easy to show that Bessel functions Introduction In the previous lecture the separation of variables method led to Bessel's equation y' ' y ' 2 y= () 2 Here
More informationTaylor Series and Asymptotic Expansions
Taylor Series and Asymptotic Epansions The importance of power series as a convenient representation, as an approimation tool, as a tool for solving differential equations and so on, is pretty obvious.
More informationParticle Methods as Message Passing
Particle Methods as Message Passing Justin Dauwels RIKEN Brain Science Institute Hirosawa,2-1,Wako-shi,Saitama,Japan Email: justin@dauwels.com Sascha Korl Phonak AG CH-8712 Staefa, Switzerland Email: sascha.korl@phonak.ch
More informationLinear Regression. Aarti Singh. Machine Learning / Sept 27, 2010
Linear Regression Aarti Singh Machine Learning 10-701/15-781 Sept 27, 2010 Discrete to Continuous Labels Classification Sports Science News Anemic cell Healthy cell Regression X = Document Y = Topic X
More informationLecture 11. Linear Soft Margin Support Vector Machines
CS142: Machine Learning Spring 2017 Lecture 11 Instructor: Pedro Felzenszwalb Scribes: Dan Xiang, Tyler Dae Devlin Linear Soft Margin Support Vector Machines We continue our discussion of linear soft margin
More informationCounting, symbols, positions, powers, choice, arithmetic, binary, translation, hex, addresses, and gates.
Counting, symbols, positions, powers, choice, arithmetic, binary, translation, he, addresses, and gates. Counting. Suppose the concern is a collection of objects. As an eample, let the objects be denoted
More information6. Vector Random Variables
6. Vector Random Variables In the previous chapter we presented methods for dealing with two random variables. In this chapter we etend these methods to the case of n random variables in the following
More informationA Comparison of Particle Filters for Personal Positioning
VI Hotine-Marussi Symposium of Theoretical and Computational Geodesy May 9-June 6. A Comparison of Particle Filters for Personal Positioning D. Petrovich and R. Piché Institute of Mathematics Tampere University
More informationLecture 7: Weak Duality
EE 227A: Conve Optimization and Applications February 7, 2012 Lecture 7: Weak Duality Lecturer: Laurent El Ghaoui 7.1 Lagrange Dual problem 7.1.1 Primal problem In this section, we consider a possibly
More information1 Convexity, Convex Relaxations, and Global Optimization
1 Conveity, Conve Relaations, and Global Optimization Algorithms 1 1.1 Minima Consider the nonlinear optimization problem in a Euclidean space: where the feasible region. min f(), R n Definition (global
More informationLecture 2: CDF and EDF
STAT 425: Introduction to Nonparametric Statistics Winter 2018 Instructor: Yen-Chi Chen Lecture 2: CDF and EDF 2.1 CDF: Cumulative Distribution Function For a random variable X, its CDF F () contains all
More informationOn the projection onto a finitely generated cone
Acta Cybernetica 00 (0000) 1 15. On the projection onto a finitely generated cone Miklós Ujvári Abstract In the paper we study the properties of the projection onto a finitely generated cone. We show for
More informationBasic Linear Inverse Method Theory - DRAFT NOTES
Basic Linear Inverse Method Theory - DRAFT NOTES Peter P. Jones 1 1 Centre for Compleity Science, University of Warwick, Coventry CV4 7AL, UK (Dated: 21st June 2012) BASIC LINEAR INVERSE METHOD THEORY
More informationConsistency as Projection
Consistency as Projection John Hooker Carnegie Mellon University INFORMS 2015, Philadelphia USA Consistency as Projection Reconceive consistency in constraint programming as a form of projection. For eample,
More informationNumerical Integration (Quadrature) Another application for our interpolation tools!
Numerical Integration (Quadrature) Another application for our interpolation tools! Integration: Area under a curve Curve = data or function Integrating data Finite number of data points spacing specified
More information