Supplemental Material for TKDE

Size: px
Start display at page:

Download "Supplemental Material for TKDE"

Transcription

1 Supplemental Material for TKDE Kijung Sh, Lee Sael, U Kang PROPOSED METHODS Proofs of Update Rules In this section, we present the proofs of the update rules Section 35 of the ma paper Specifically, we prove the CDTF update rule for L regularization Theorem 7, the CDTF update rule for the nonnegativity constrat Theorem, the SALS update rule for coupled tensor factorization Theorem, and the update rule for bias terms commonly used by CDTF and SALS for the bias model Theorem 0 Lemma Partial Derivative CDTF: For a parameter i, let ˆr i i N x ii N N s k l al, g i,,i N ˆr Ω n ii N l n al k, and d i,,i N Ω n i,,i N Ω i,,i N Ω i,,i N Ω n i,,i N Ω n g l n al k, as the ma paper x i i N K s l x i i N K x i i N x i i N K s l s l ˆr i i N a l a l K s l a l l a l k a l g a l k l n Theorem 7: Correctness of CDTF with L - regularization The update rule the ma paper mimizes the loss function with respect to the updated parameter For an updated parameter, let ˆr i i N x ii N s k N l al, g i,,i N Ω n d i,,i N Ω n paper arg m k ˆr ii N l n al k, and l n al k, as the ma λ g/d if g > λ L LassoA,, A N λ g/d if g < -λ 0 otherwise By Lemma, L LassoA,, A N i,,i N Ω g i d λ an x i i N K i Case : If g > λ > 0, s l { g i d λ g i d λ L Lasso k 0, should be negative for L Lasso k < 0, L Lasso k < 0 makes L Lasso k g a l λ N A l l d λ, and an if an if an > 0 < 0 > d Sce d to be zero If λ g/d 0 Sce L Lasso d 0, k λg/d mimizes L LassoA,, A N with respect to Case : Likewise, if g < λ < 0, L Lasso k Sce d 0, be zero If L Lasso k < should be positive for L Lasso k > 0, L Lasso g k λ g/d > 0 makes L Lasso k d 0, to d λ, and 0 Sce λ g/d mimizes L Lasso A,, A N with respect to i Case 3: On the other hand, if λ g λ, L Lasso { g i d λ an i d 0 g i d λ an i d 0 if an if an i > 0 i < 0 That is, L Lasso A,, A N decreases if i < 0 and creases if > 0, given other parameters Thus, L Lasso A,, A N, which is a contuous

2 function, is mimized with respect to arg m k if an 0 λ g/d if g > λ L LassoA,, A N λ g/d if g < -λ 0 otherwise N l al, Theorem : Correctness of CDTF with the nonnegativity constrat The update rule 4 the ma paper mimizes the loss function with respect to the updated parameter under the nonnegativity constrat For an updated parameter i, let ˆr i i N x ii N s k g i,,i N ˆr Ω n ii N l n al k, and d i,,i N Ω n arg m k 0 l n al k, as the ma paper LA,, A N g max d λ, 0 By Lemma, LA,, A N i,,i N Ω g d λ Thus, L x i i N K s l a l λ N A l F l > 0 if > g/d λ 0 if g/d λ < 0 otherwise Case : If g/dλ 0, sce L k dλ 0, LA,, A N is mimized if g/dλ with respect to i Case : On the other hand, if g/d λ < 0, under the constrat that i 0, LA,, A N is mimized with respect to if an 0 This is because LA,, A N, which is a contuous function, monotonically creases ie, 0 > g/d λ arg m k 0 L k LA,, A N g max d λ, 0 > 0 if Ny n y Ny n y y ii Ny K k i C c c Likewise, let x Ω and y Ω be the sets of dices of the observable entries X and Y, respectively arg m [a i k,,a i k C ] T L Coupled xa,, xa Nx, ya,, ya Ny xb i yb i λi C xc i yc i, where x B i and y B i are C by C matrices whose entries are xb i c c x c x c, n n yb i c c xc i and y c i xc i c yc i c i,,i Nx xω i i,,i Ny yω i y c n y c n, c, c, are length C vectors whose entries are xˆr i i N x, i,,i Nx xω i i,,i Ny yω i y ˆr i i N and I C is the C by C identity matrix L xa,, xa Nx a i k c i,,i Nx xω λ Nx xa n F n xa i k c i,,i Nx xω i n c y c n LxA,, xa Nx xa x i i Nx K x i i Nx K i,,i Nx xω i λa i k c Likewise, xa i k c x i i Nx x s n xa C i k c N x s n i k c x s n K N x s n x x x λ xa i k c, c x s xˆr i i Nx x c n Theorem : Correctness of SALS Coupled Tensor Factorization The update rule 6 the ma paper mimizes 5 with respect to the updated parameters Let xr and y R be the residual tensors for X and Y, respectively That is, for C updated parameters a i k,, a i k C, xˆr ii Nx x ii Nx K Nx k n x i C Nx c n x c and yˆr ii Ny L ya,, ya Ny a i k c C i,,i Ny yω i λa i k c N y y s n s y ˆr i i Ny y c n

3 From these, L Coupled a i k c i,,i Nx xω i 0, c, c C i,,i Ny yω i C N x s n C x s xˆr i i Nx x c n N y y s n λa i k c 0, c C xa i k s i,,i Nx xω i i,,i Ny yω i λa i k c i,,i Nx xω i i,,i Ny yω i s s y ˆr i i Ny x s n C ya i k s s xˆr i i Nx y ˆr i i Ny y s n x c n y c n y c n x n, c c y n c xb i yb i λi C[a i k,, a i k C ] T xc i yc i arg m [a i k,,a i k C ] T L Coupled xa,, xa Nx, ya,, ya Ny xb i yb i λi C xc i yc i Theorem 0: Correctness of the Update Rule for Bias Terms The update rule the ma paper mimizes 7 with respect to the updated parameter For an updated parameter b n, let r ii N x ii N K N k l al k l n bl µ, as the ma paper arg m b n L BiasA,, A N, b,, b N i,,i N Ω n L BiasA,, A N, b,, b N i,,i N Ω λ A l b n x i i N K a l s l b n N A l F b n r i i N /λ b Ω n λ b l N l N b l F b n b l µ i,,i N Ω n x i i N K i,,i N Ω n x i i N a l s l λ b Ω n b n Sce L Bias b n N K s l l a l N b l µ l b n r i i N b n λ b b n i,,i N Ω n r i i N b l µ λ b l 3 N b l F b n λ b Ω n 0, L Bias A,, A N, b,, b N is mimized with respect to b n if b n i,,i N r Ω n ii N /λ b Ω n 0, which entails L Bias b n arg m b n L BiasA,, A N, b,, b N Pseudocodes i,,i N Ω n r i i N /λ b Ω n We present the pseudocodes of the SALS variants described Section 35 of the ma paper SALS for Coupled Tensor Factorization Algorithm 6 describes SALS for coupled tensor factorization, where two tensors, denoted by X and Y, share their first mode without loss of generality We denote the residual tensors for X and Y by x R and yr, respectively The lengths of the n-th modes of X and Y are denoted by x I n and y I n, respectively Algorithm 6: SALS for Coupled Tensor Factorization Input : X, Y, K, λ Output: A, xa n for n N x, ya n for n N y itialize xr, yr, A, xa n and ya n for all n for outer iter T out do 3 for split iter K C do 4 choose k,, k C from columns not updated yet 5 compute x ˆR and y ˆR usg 6 for ner iter T do 7 for i I xi yi do update a i k,, a i k C usg for n N x do for xi n do update x,, x usg C for n N y do for yi n do update y,, y usg C update xr and yr usg 0

4 4 Algorithm 7: SALS for Bias Model Input : X, K, λ A, λ b Output: A n for all n, b n for all n, µ compute µ the mean of the observable entries of X itialize R, A n for all n, and b n for all n 3 for outer iter T out do 4 for split iter K C do 5 choose k,, k C from columns not updated yet 6 compute ˆR usg 7 for ner iter T do for n N do for I n do 0 update,, C usg update R usg 0 for n N do for I n do update b n usg update R usg SALS for Bias Model SALS for the bias model is described Algorithm 7, where each i,, i N th entry of R is r i x ii N µ N n bn K N k n an, as explaed Section 355 of the ma paper OPTIMIZATION ON MAPREDUCE In this section, we present the details of the optimization techniques described Section 4 of the ma paper Local Disk Cachg As explaed Section 4 of the ma paper, our MAPREDUCE implementation of CDTF and SALS with local disk cachg, X entries are distributed across maches and cached their local disk durg the map and reduce stages Algorithm gives the details of the map and reduce stages The rest of CDTF and SALS runs the close stage cleanup stage Hadoop usg the cached data Direct Communication In the ma paper, we troduce direct communication between reducers usg distributed file system to overcome the rigidity of MAPREDUCE model Algorithm describes the implementation of m k broadcast CDTF le 0 of Algorithm 3 the ma paper based on this communication method 3 Greedy Row Assignment Our MAPREDUCE implementation of SALS and CDTF uses the greedy row assignment, explaed Section 343 of the ma paper In this section, we expla our MAPREDUCE implementation of the greedy row assignment We assume that X is stored Algorithm : Data distribution CDTF and SALS with local disk cachg Input : X, ms n for all m and n Output: mω n entries of R X for all m and n MapKey k, Value v beg 3 i,, i N, x i i N v 4 for n,,n do 5 fd m where ms n 6 emit < m, n, i,, i N, x i i N > 7 end PartitionerKey k, Value v beg 0 m, n k assign < k, v > to mache m end 3 ReduceKey k, Value v[ v ] 4 beg 5 m, n k 6 create a file on the local disk to cache mω n entries of R 7 foreach i,, i N, x i i N v do write i,, i N, x i i N to the file end Algorithm : m k broadcast CDTF Input : m k parameters to broadcast Output: k parameters received from others beg create a data file ma on the distributed file system DFS 3 write m k on the datafile 4 create a dummy file md on DFS 5 while not all data files are read do 6 get the list of dummy files from DFS 7 foreach m D the list do if m A are not read then read m k from m A 0 end on the distributed file system At the first stage, Ω n for all n and is computed Specifically, mappers output < n,, > for all n for each entry x ii N, and reducers output < n,, Ω n > for all n and by countg the number of values for each key At the second stage, the outputs are aggregated to a sgle reducer which runs the rest of Algorithm 5 the ma paper 3 EXPERIMENTS In this section, we design and conduct additional experiments to answer the followg questions: How do different numbers of ner iterations T affect the convergence of SALS? How do different numbers of columns updated at a time C affect the runng time of SALS? 3 Experimental Settgs We ran experiments on a 0-node Hadoop cluster Each node had an Intel Xeon E GHz CPU

5 5 a Netflix 3 C 0 b Netflix 3 C 0 c Yahoo-music 4 C 0 d Yahoo-music 4 C 40 Fig 4: Effects of T ie, ner iterations on the convergence of SALS when C ie, the number of columns updated at a time has large enough values The effects of T on convergence speed and the quality of converged solutions are margal a Netflix 3 3 Effects of the Number of Inner Iterations ie, T on the Convergence of SALS We compared the convergence properties of SALS with different T values Especially, we focused on cases where C ie, the number of columns updated at a time has large enough values The effect of T when C is set to one and thus SALS is equivalent to CDTF can be found the ma paper As seen Figure 4, the effects of T on convergence speed and the quality of converged solutions are neither distct nor consistent When C is set to one, however, high T values are preferred see Section 57 of the ma paper for detailed experimental results b Yahoo-music 4 Fig 5: Effects of the number of columns updated at a time C on the runng time of SALS Runng time per iteration decreased until C 0, then started to crease The maximum heap size per reducer was set to GB Other experimental settgs, cludg datasets and parameter values λ and K, were the same as those the ma paper The number of reducers was set to 0 We used the root mean square error RMSE on a held-out test set, which is commonly used recommender systems, to measure the accuracy, as the ma paper 33 Effects of the Number of Columns Updated at a Time ie, C on Runng Time of SALS We measured the runng times per iteration SALS, as we creased C from to K As seen Figure 5, runng time per iteration decreased until C 0, then started to crease As C creases, the amount of disk I/O decles sce it depends on the number of times that the entries of R or ˆR are streamed from disk, which is versely proportional to C Conversely, computational cost creases quadratically with regard to C At small C values, the decrease the amount of disk I/O was greater and leaded to a downward trend of runng time per iteration The opposite happened at large C values

RECOMMENDATION problems can be viewed as

RECOMMENDATION problems can be viewed as IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING 1 Fully Scalable Methods for Distributed Tensor Factorization Kijung Sh, Lee Sael, and U Kang Abstract Given a high-order large-scale tensor, how can

More information

Distributed Methods for High-dimensional and Large-scale Tensor Factorization

Distributed Methods for High-dimensional and Large-scale Tensor Factorization Distributed Methods for High-dimensional and Large-scale Tensor Factorization Kijung Shin Dept. of Computer Science and Engineering Seoul National University Seoul, Republic of Korea koreaskj@snu.ac.kr

More information

Parallel Matrix Factorization for Recommender Systems

Parallel Matrix Factorization for Recommender Systems Under consideration for publication in Knowledge and Information Systems Parallel Matrix Factorization for Recommender Systems Hsiang-Fu Yu, Cho-Jui Hsieh, Si Si, and Inderjit S. Dhillon Department of

More information

Scalable Coordinate Descent Approaches to Parallel Matrix Factorization for Recommender Systems

Scalable Coordinate Descent Approaches to Parallel Matrix Factorization for Recommender Systems Scalable Coordinate Descent Approaches to Parallel Matrix Factorization for Recommender Systems Hsiang-Fu Yu, Cho-Jui Hsieh, Si Si, and Inderjit Dhillon Department of Computer Science University of Texas

More information

Large-Scale Behavioral Targeting

Large-Scale Behavioral Targeting Large-Scale Behavioral Targeting Ye Chen, Dmitry Pavlov, John Canny ebay, Yandex, UC Berkeley (This work was conducted at Yahoo! Labs.) June 30, 2009 Chen et al. (KDD 09) Large-Scale Behavioral Targeting

More information

CS425: Algorithms for Web Scale Data

CS425: Algorithms for Web Scale Data CS425: Algorithms for Web Scale Data Most of the slides are from the Mining of Massive Datasets book. These slides have been modified for CS425. The original slides can be accessed at: www.mmds.org Challenges

More information

Large-scale Matrix Factorization. Kijung Shin Ph.D. Student, CSD

Large-scale Matrix Factorization. Kijung Shin Ph.D. Student, CSD Large-scale Matrix Factorization Kijung Shin Ph.D. Student, CSD Roadmap Matrix Factorization (review) Algorithms Distributed SGD: DSGD Alternating Least Square: ALS Cyclic Coordinate Descent: CCD++ Experiments

More information

Linear Algebra (Review) Volker Tresp 2018

Linear Algebra (Review) Volker Tresp 2018 Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c

More information

Block AIR Methods. For Multicore and GPU. Per Christian Hansen Hans Henrik B. Sørensen. Technical University of Denmark

Block AIR Methods. For Multicore and GPU. Per Christian Hansen Hans Henrik B. Sørensen. Technical University of Denmark Block AIR Methods For Multicore and GPU Per Christian Hansen Hans Henrik B. Sørensen Technical University of Denmark Model Problem and Notation Parallel-beam 3D tomography exact solution exact data noise

More information

Fast and Scalable Distributed Boolean Tensor Factorization

Fast and Scalable Distributed Boolean Tensor Factorization Fast and Scalable Distributed Boolean Tensor Factorization Namyong Park Seoul National University Email: namyong.park@snu.ac.kr Sejoon Oh Seoul National University Email: ohhenrie@snu.ac.kr U Kang Seoul

More information

GIVEN a tensor that is too large to fit in memory, how can

GIVEN a tensor that is too large to fit in memory, how can Out-of-Core and Distributed Algorithms for Dense Subtensor Mining Kijung Shin, Bryan Hooi, Jisu Kim, and Christos Faloutsos arxiv:802.0065v2 [cs.db] 6 Feb 208 Abstract How can we detect fraudulent lockstep

More information

CHAPTER 2 EXTRACTION OF THE QUADRATICS FROM REAL ALGEBRAIC POLYNOMIAL

CHAPTER 2 EXTRACTION OF THE QUADRATICS FROM REAL ALGEBRAIC POLYNOMIAL 24 CHAPTER 2 EXTRACTION OF THE QUADRATICS FROM REAL ALGEBRAIC POLYNOMIAL 2.1 INTRODUCTION Polynomial factorization is a mathematical problem, which is often encountered in applied sciences and many of

More information

Graph Analysis Using Map/Reduce

Graph Analysis Using Map/Reduce Seminar: Massive-Scale Graph Analysis Summer Semester 2015 Graph Analysis Using Map/Reduce Laurent Linden s9lalind@stud.uni-saarland.de May 14, 2015 Introduction Peta-Scale Graph + Graph mining + MapReduce

More information

Linear Algebra (Review) Volker Tresp 2017

Linear Algebra (Review) Volker Tresp 2017 Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.

More information

Neural Networks. ICS 273A UC Irvine Instructor: Max Welling

Neural Networks. ICS 273A UC Irvine Instructor: Max Welling Neural Networks ICS 273A UC Irve Instructor: Max Wellg Neurons 1 b Neurons communicate by receivg signals on their dendrites. Addg these signals and firg off a new signal along the axon if the total put

More information

Tall-and-skinny! QRs and SVDs in MapReduce

Tall-and-skinny! QRs and SVDs in MapReduce A 1 A 2 A 3 Tall-and-skinny! QRs and SVDs in MapReduce Yangyang Hou " Purdue, CS Austin Benson " Stanford University Paul G. Constantine Col. School. Mines " Joe Nichols U. of Minn James Demmel " UC Berkeley

More information

For more information visit here:

For more information visit here: The length or the magnitude of the vector = (a, b, c) is defined by w = a 2 +b 2 +c 2 A vector may be divided by its own length to convert it into a unit vector, i.e.? = u / u. (The vectors have been denoted

More information

Multiplication of Polynomials

Multiplication of Polynomials Summary 391 Chapter 5 SUMMARY Section 5.1 A polynomial in x is defined by a finite sum of terms of the form ax n, where a is a real number and n is a whole number. a is the coefficient of the term. n is

More information

Regression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)

Regression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Linear Regression Regression Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Example: Height, Gender, Weight Shoe Size Audio features

More information

Regression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)

Regression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Linear Regression Regression Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Example: Height, Gender, Weight Shoe Size Audio features

More information

EEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture 11

EEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture 11 EEC 686/785 Modeling & Performance Evaluation of Computer Systems Lecture Department of Electrical and Computer Engineering Cleveland State University wenbing@ieee.org (based on Dr. Raj Jain s lecture

More information

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices

A robust multilevel approximate inverse preconditioner for symmetric positive definite matrices DICEA DEPARTMENT OF CIVIL, ENVIRONMENTAL AND ARCHITECTURAL ENGINEERING PhD SCHOOL CIVIL AND ENVIRONMENTAL ENGINEERING SCIENCES XXX CYCLE A robust multilevel approximate inverse preconditioner for symmetric

More information

Using R for Iterative and Incremental Processing

Using R for Iterative and Incremental Processing Using R for Iterative and Incremental Processing Shivaram Venkataraman, Indrajit Roy, Alvin AuYoung, Robert Schreiber UC Berkeley and HP Labs UC BERKELEY Big Data, Complex Algorithms PageRank (Dominant

More information

Discrete Multi-material Topology Optimization under Total Mass Constraint

Discrete Multi-material Topology Optimization under Total Mass Constraint Discrete Multi-material Topology Optimization under Total Mass Constraint Xingtong Yang Ming Li State Key Laboratory of CAD&CG, Zhejiang University Solid and Physical Modeling Bilbao, June 12, 2018 Problem

More information

Lecture notes 1: ECEN 489

Lecture notes 1: ECEN 489 Lecture notes : ECEN 489 Power Management Circuits and Systems Department of Electrical & Computer Engeerg Texas A&M University Jose Silva-Martez January 207 Copyright Texas A&M University. All rights

More information

Incomplete Cholesky preconditioners that exploit the low-rank property

Incomplete Cholesky preconditioners that exploit the low-rank property anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université

More information

k-variates++: more pluses in the k-means++

k-variates++: more pluses in the k-means++ (formerly NICTA) k-variates++: more pluses in the k-means++ m p 7 3 n o M, 9 Poster #2 Richard Nock, Raphaël Canyasse, Roksana Boreli, Frank Nielsen DATA61 ANU TECHNION ECOLE POLYTECHNIQUE UNSW SONY CS

More information

StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory

StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory S.V. N. (vishy) Vishwanathan Purdue University and Microsoft vishy@purdue.edu October 9, 2012 S.V. N. Vishwanathan (Purdue,

More information

l 1 and l 2 Regularization

l 1 and l 2 Regularization David Rosenberg New York University February 5, 2015 David Rosenberg (New York University) DS-GA 1003 February 5, 2015 1 / 32 Tikhonov and Ivanov Regularization Hypothesis Spaces We ve spoken vaguely about

More information

Math Models of OR: Branch-and-Bound

Math Models of OR: Branch-and-Bound Math Models of OR: Branch-and-Bound John E. Mitchell Department of Mathematical Sciences RPI, Troy, NY 12180 USA November 2018 Mitchell Branch-and-Bound 1 / 15 Branch-and-Bound Outline 1 Branch-and-Bound

More information

Chapter 7 Iterative Techniques in Matrix Algebra

Chapter 7 Iterative Techniques in Matrix Algebra Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition

More information

Incorporating uncertainty in the design of water distribution systems

Incorporating uncertainty in the design of water distribution systems European Water 58: 449-456, 2017. 2017 E.W. Publications Incorporatg uncertaty the design of water distribution systems M. Spiliotis 1* and G. Tsakiris 2 1 Division of Hydraulic Engeerg, Department of

More information

SOLUTIONS TO ADDITIONAL EXERCISES FOR II.1 AND II.2

SOLUTIONS TO ADDITIONAL EXERCISES FOR II.1 AND II.2 SOLUTIONS TO ADDITIONAL EXERCISES FOR II.1 AND II.2 Here are the solutions to the additional exercises in betsepexercises.pdf. B1. Let y and z be distinct points of L; we claim that x, y and z are not

More information

THE solution of the absolute value equation (AVE) of

THE solution of the absolute value equation (AVE) of The nonlinear HSS-like iterative method for absolute value equations Mu-Zheng Zhu Member, IAENG, and Ya-E Qi arxiv:1403.7013v4 [math.na] 2 Jan 2018 Abstract Salkuyeh proposed the Picard-HSS iteration method

More information

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics

SPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics SPARSE SOLVERS FOR THE POISSON EQUATION Margreet Nool CWI, Multiscale Dynamics November 9, 2015 OUTLINE OF THIS TALK 1 FISHPACK, LAPACK, PARDISO 2 SYSTEM OVERVIEW OF CARTESIUS 3 POISSON EQUATION 4 SOLVERS

More information

Thermodynamics [ENGR 251] [Lyes KADEM 2007]

Thermodynamics [ENGR 251] [Lyes KADEM 2007] CHAPTER V The first law of thermodynamics is a representation of the conservation of energy. It is a necessary, but not a sufficient, condition for a process to occur. Indeed, no restriction is imposed

More information

Behavioral Data Mining. Lecture 7 Linear and Logistic Regression

Behavioral Data Mining. Lecture 7 Linear and Logistic Regression Behavioral Data Mining Lecture 7 Linear and Logistic Regression Outline Linear Regression Regularization Logistic Regression Stochastic Gradient Fast Stochastic Methods Performance tips Linear Regression

More information

A Hybrid Method for the Wave Equation. beilina

A Hybrid Method for the Wave Equation.   beilina A Hybrid Method for the Wave Equation http://www.math.unibas.ch/ beilina 1 The mathematical model The model problem is the wave equation 2 u t 2 = (a 2 u) + f, x Ω R 3, t > 0, (1) u(x, 0) = 0, x Ω, (2)

More information

QR Decomposition in a Multicore Environment

QR Decomposition in a Multicore Environment QR Decomposition in a Multicore Environment Omar Ahsan University of Maryland-College Park Advised by Professor Howard Elman College Park, MD oha@cs.umd.edu ABSTRACT In this study we examine performance

More information

Nesterov-based Alternating Optimization for Nonnegative Tensor Completion: Algorithm and Parallel Implementation

Nesterov-based Alternating Optimization for Nonnegative Tensor Completion: Algorithm and Parallel Implementation Nesterov-based Alternating Optimization for Nonnegative Tensor Completion: Algorithm and Parallel Implementation Georgios Lourakis and Athanasios P. Liavas School of Electrical and Computer Engineering,

More information

Parallelized Variational EM for Latent Dirichlet Allocation: An Experimental Evaluation of Speed and Scalability

Parallelized Variational EM for Latent Dirichlet Allocation: An Experimental Evaluation of Speed and Scalability Parallelized Variational EM for Latent Dirichlet Allocation: An Experimental Evaluation of Speed and Scalability Ramesh Nallapati, William Cohen and John Lafferty Machine Learning Department Carnegie Mellon

More information

Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent

Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent KDD 2011 Rainer Gemulla, Peter J. Haas, Erik Nijkamp and Yannis Sismanis Presenter: Jiawen Yao Dept. CSE, UT Arlington 1 1

More information

Lecture XI. Approximating the Invariant Distribution

Lecture XI. Approximating the Invariant Distribution Lecture XI Approximating the Invariant Distribution Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Invariant Distribution p. 1 /24 SS Equilibrium in the Aiyagari model G.

More information

CS246 Final Exam, Winter 2011

CS246 Final Exam, Winter 2011 CS246 Final Exam, Winter 2011 1. Your name and student ID. Name:... Student ID:... 2. I agree to comply with Stanford Honor Code. Signature:... 3. There should be 17 numbered pages in this exam (including

More information

Scheduling with AND/OR Precedence Constraints

Scheduling with AND/OR Precedence Constraints Scheduling with AND/OR Precedence Constraints Seminar Mathematische Optimierung - SS 2007 23th April 2007 Synthesis synthesis: transfer from the behavioral domain (e. g. system specifications, algorithms)

More information

Math 304 (Spring 2010) - Lecture 2

Math 304 (Spring 2010) - Lecture 2 Math 304 (Spring 010) - Lecture Emre Mengi Department of Mathematics Koç University emengi@ku.edu.tr Lecture - Floating Point Operation Count p.1/10 Efficiency of an algorithm is determined by the total

More information

LINEAR COMPARTMENTAL MODELS: INPUT-OUTPUT EQUATIONS AND OPERATIONS THAT PRESERVE IDENTIFIABILITY. 1. Introduction

LINEAR COMPARTMENTAL MODELS: INPUT-OUTPUT EQUATIONS AND OPERATIONS THAT PRESERVE IDENTIFIABILITY. 1. Introduction LINEAR COMPARTMENTAL MODELS: INPUT-OUTPUT EQUATIONS AND OPERATIONS THAT PRESERVE IDENTIFIABILITY ELIZABETH GROSS, HEATHER HARRINGTON, NICOLETTE MESHKAT, AND ANNE SHIU Abstract. This work focuses on the

More information

A Proof of Theorem 1. Dual Weighted Generalized Nuclear Norm. C Proof of Theorem 2

A Proof of Theorem 1. Dual Weighted Generalized Nuclear Norm. C Proof of Theorem 2 A Proof of Theorem Proof or all Z = P i c i i with kak A = P i c i, where i = Au i v T i BT, we can construct the i-th column of W and H as Clearly, we have Z = WH T and w i = p c i Au i and h i = p c

More information

SOLUTIONS TO EXERCISES FOR. MATHEMATICS 133 Part 2. I. Topics from linear algebra

SOLUTIONS TO EXERCISES FOR. MATHEMATICS 133 Part 2. I. Topics from linear algebra SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part Fall 013 NOTE ON ILLUSTRATIONS. Drawings for several of the solutions in this file are available in the following document: http://math.ucr.edu/ res/math133/math133solutionsa.figures.f13.pdf

More information

High Performance Parallel Tucker Decomposition of Sparse Tensors

High Performance Parallel Tucker Decomposition of Sparse Tensors High Performance Parallel Tucker Decomposition of Sparse Tensors Oguz Kaya INRIA and LIP, ENS Lyon, France SIAM PP 16, April 14, 2016, Paris, France Joint work with: Bora Uçar, CNRS and LIP, ENS Lyon,

More information

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts

Some definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts Some definitions Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization M. M. Sussman sussmanm@math.pitt.edu Office Hours: MW 1:45PM-2:45PM, Thack 622 A matrix A is SPD (Symmetric

More information

Cache Oblivious Stencil Computations

Cache Oblivious Stencil Computations Cache Oblivious Stencil Computations S. HUNOLD J. L. TRÄFF F. VERSACI Lectures on High Performance Computing 13 April 2015 F. Versaci (TU Wien) Cache Oblivious Stencil Computations 13 April 2015 1 / 19

More information

Matrix Factorization Techniques for Recommender Systems

Matrix Factorization Techniques for Recommender Systems Matrix Factorization Techniques for Recommender Systems Patrick Seemann, December 16 th, 2014 16.12.2014 Fachbereich Informatik Recommender Systems Seminar Patrick Seemann Topics Intro New-User / New-Item

More information

Solving Regularized Total Least Squares Problems

Solving Regularized Total Least Squares Problems Solving Regularized Total Least Squares Problems Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Numerical Simulation Joint work with Jörg Lampe TUHH Heinrich Voss Total

More information

Linear Regression (continued)

Linear Regression (continued) Linear Regression (continued) Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 6, 2017 1 / 39 Outline 1 Administration 2 Review of last lecture 3 Linear regression

More information

arxiv: v3 [stat.ml] 14 Apr 2016

arxiv: v3 [stat.ml] 14 Apr 2016 arxiv:1307.0048v3 [stat.ml] 14 Apr 2016 Simple one-pass algorithm for penalized linear regression with cross-validation on MapReduce Kun Yang April 15, 2016 Abstract In this paper, we propose a one-pass

More information

LFTF: A Framework for Efficient Tensor Analytics at Scale

LFTF: A Framework for Efficient Tensor Analytics at Scale : A Framework for Efficient Tensor Analytics at Scale Fan Yang Fanhua Shang Yuzhen Huang James Cheng Jinfeng Li Yunjian Zhao Ruihao Zhao Department of Computer Science and Engineering The Chinese University

More information

ECE521 W17 Tutorial 1. Renjie Liao & Min Bai

ECE521 W17 Tutorial 1. Renjie Liao & Min Bai ECE521 W17 Tutorial 1 Renjie Liao & Min Bai Schedule Linear Algebra Review Matrices, vectors Basic operations Introduction to TensorFlow NumPy Computational Graphs Basic Examples Linear Algebra Review

More information

Efficient Primal- Dual Graph Algorithms for Map Reduce

Efficient Primal- Dual Graph Algorithms for Map Reduce Efficient Primal- Dual Graph Algorithms for Map Reduce Joint work with Bahman Bahmani Ashish Goel, Stanford Kamesh Munagala Duke University Modern Data Models Over the past decade, many commodity distributed

More information

Knowledge Discovery and Data Mining 1 (VO) ( )

Knowledge Discovery and Data Mining 1 (VO) ( ) Knowledge Discovery and Data Mining 1 (VO) (707.003) Map-Reduce Denis Helic KTI, TU Graz Oct 24, 2013 Denis Helic (KTI, TU Graz) KDDM1 Oct 24, 2013 1 / 82 Big picture: KDDM Probability Theory Linear Algebra

More information

Environment (Parallelizing Query Optimization)

Environment (Parallelizing Query Optimization) Advanced d Query Optimization i i Techniques in a Parallel Computing Environment (Parallelizing Query Optimization) Wook-Shin Han*, Wooseong Kwak, Jinsoo Lee Guy M. Lohman, Volker Markl Kyungpook National

More information

CS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions

CS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions CS246: Mining Massive Data Sets Winter 2017 Problem Set 2 Due 11:59pm February 9, 2017 Only one late period is allowed for this homework (11:59pm 2/14). General Instructions Submission instructions: These

More information

An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion

An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion Robert M. Freund, MIT joint with Paul Grigas (UC Berkeley) and Rahul Mazumder (MIT) CDC, December 2016 1 Outline of Topics

More information

MapReduce in Spark. Krzysztof Dembczyński. Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland

MapReduce in Spark. Krzysztof Dembczyński. Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland MapReduce in Spark Krzysztof Dembczyński Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland Software Development Technologies Master studies, second semester

More information

Network Newton. Aryan Mokhtari, Qing Ling and Alejandro Ribeiro. University of Pennsylvania, University of Science and Technology (China)

Network Newton. Aryan Mokhtari, Qing Ling and Alejandro Ribeiro. University of Pennsylvania, University of Science and Technology (China) Network Newton Aryan Mokhtari, Qing Ling and Alejandro Ribeiro University of Pennsylvania, University of Science and Technology (China) aryanm@seas.upenn.edu, qingling@mail.ustc.edu.cn, aribeiro@seas.upenn.edu

More information

MapReduce in Spark. Krzysztof Dembczyński. Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland

MapReduce in Spark. Krzysztof Dembczyński. Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland MapReduce in Spark Krzysztof Dembczyński Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland Software Development Technologies Master studies, first semester

More information

Notes on Mathematics Groups

Notes on Mathematics Groups EPGY Singapore Quantum Mechanics: 2007 Notes on Mathematics Groups A group, G, is defined is a set of elements G and a binary operation on G; one of the elements of G has particularly special properties

More information

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18

CSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18 CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$

More information

2.6 Large Signal Operation

2.6 Large Signal Operation 3/4/2011 section 2_6 Large signal operation 1/2 2.6 Large Signal Operation Readg Assignment: 94-98 Recall that real amplifiers are only approximately lear! If the put signal becomes too large, and/or the

More information

Introduction to Compressed Sensing

Introduction to Compressed Sensing Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral

More information

SIMPLEX LIKE (aka REDUCED GRADIENT) METHODS. REDUCED GRADIENT METHOD (Wolfe)

SIMPLEX LIKE (aka REDUCED GRADIENT) METHODS. REDUCED GRADIENT METHOD (Wolfe) 19 SIMPLEX LIKE (aka REDUCED GRADIENT) METHODS The REDUCED GRADIENT algorithm and its variants such as the CONVEX SIMPLEX METHOD (CSM) and the GENERALIZED REDUCED GRADIENT (GRG) algorithm are approximation

More information

Continuous methods for numerical linear algebra problems

Continuous methods for numerical linear algebra problems Continuous methods for numerical linear algebra problems Li-Zhi Liao (http://www.math.hkbu.edu.hk/ liliao) Department of Mathematics Hong Kong Baptist University The First International Summer School on

More information

Multi-Approximate-Keyword Routing Query

Multi-Approximate-Keyword Routing Query Bin Yao 1, Mingwang Tang 2, Feifei Li 2 1 Department of Computer Science and Engineering Shanghai Jiao Tong University, P. R. China 2 School of Computing University of Utah, USA Outline 1 Introduction

More information

a b = a T b = a i b i (1) i=1 (Geometric definition) The dot product of two Euclidean vectors a and b is defined by a b = a b cos(θ a,b ) (2)

a b = a T b = a i b i (1) i=1 (Geometric definition) The dot product of two Euclidean vectors a and b is defined by a b = a b cos(θ a,b ) (2) This is my preperation notes for teaching in sections during the winter 2018 quarter for course CSE 446. Useful for myself to review the concepts as well. More Linear Algebra Definition 1.1 (Dot Product).

More information

Bidirectional Representation and Backpropagation Learning

Bidirectional Representation and Backpropagation Learning Int'l Conf on Advances in Big Data Analytics ABDA'6 3 Bidirectional Representation and Bacpropagation Learning Olaoluwa Adigun and Bart Koso Department of Electrical Engineering Signal and Image Processing

More information

Lecture 9: September 28

Lecture 9: September 28 0-725/36-725: Convex Optimization Fall 206 Lecturer: Ryan Tibshirani Lecture 9: September 28 Scribes: Yiming Wu, Ye Yuan, Zhihao Li Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These

More information

Chapter Two Elements of Linear Algebra

Chapter Two Elements of Linear Algebra Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to

More information

Learning Multiple Tasks with a Sparse Matrix-Normal Penalty

Learning Multiple Tasks with a Sparse Matrix-Normal Penalty Learning Multiple Tasks with a Sparse Matrix-Normal Penalty Yi Zhang and Jeff Schneider NIPS 2010 Presented by Esther Salazar Duke University March 25, 2011 E. Salazar (Reading group) March 25, 2011 1

More information

FPGA Implementation of a Predictive Controller

FPGA Implementation of a Predictive Controller FPGA Implementation of a Predictive Controller SIAM Conference on Optimization 2011, Darmstadt, Germany Minisymposium on embedded optimization Juan L. Jerez, George A. Constantinides and Eric C. Kerrigan

More information

CSE546 Machine Learning, Autumn 2016: Homework 1

CSE546 Machine Learning, Autumn 2016: Homework 1 CSE546 Machine Learning, Autumn 2016: Homework 1 Due: Friday, October 14 th, 5pm 0 Policies Writing: Please submit your HW as a typed pdf document (not handwritten). It is encouraged you latex all your

More information

Web Appendix for The Value of Reputation in an Online Freelance Marketplace

Web Appendix for The Value of Reputation in an Online Freelance Marketplace Web Appendix for The Value of Reputation in an Online Freelance Marketplace Hema Yoganarasimhan 1 Details of Second-Step Estimation All the estimates from state transitions (θ p, θ g, θ t ), η i s, and

More information

MULTIPLE PRODUCTS OBJECTIVES. If a i j,b j k,c i k, = + = + = + then a. ( b c) ) 8 ) 6 3) 4 5). If a = 3i j+ k and b 3i j k = = +, then a. ( a b) = ) 0 ) 3) 3 4) not defined { } 3. The scalar a. ( b c)

More information

Definitions and Properties of R N

Definitions and Properties of R N Definitions and Properties of R N R N as a set As a set R n is simply the set of all ordered n-tuples (x 1,, x N ), called vectors. We usually denote the vector (x 1,, x N ), (y 1,, y N ), by x, y, or

More information

CS224W: Methods of Parallelized Kronecker Graph Generation

CS224W: Methods of Parallelized Kronecker Graph Generation CS224W: Methods of Parallelized Kronecker Graph Generation Sean Choi, Group 35 December 10th, 2012 1 Introduction The question of generating realistic graphs has always been a topic of huge interests.

More information

Recommendation Systems

Recommendation Systems Recommendation Systems Popularity Recommendation Systems Predicting user responses to options Offering news articles based on users interests Offering suggestions on what the user might like to buy/consume

More information

Transmission Line Input Impedance

Transmission Line Input Impedance 1/22/23 Transmission e Input Impedance.doc 1/9 Transmission e Input Impedance Consider a lossless le, length, termated with a load. I(z) I + V (z) -, β + V - z z What is the put impedance of this le? Q:

More information

Scalable Nonnegative Matrix Factorization with Block-wise Updates

Scalable Nonnegative Matrix Factorization with Block-wise Updates Scalable Nonnegative Matrix Factorization with Block-wise Updates Jiangtao Yin 1, Lixin Gao 1, and Zhongfei (Mark) Zhang 2 1 University of Massachusetts Amherst, Amherst, MA 01003, USA 2 Binghamton University,

More information

Additional Mathematics Lines and circles

Additional Mathematics Lines and circles Additional Mathematics Lines and circles Topic assessment 1 The points A and B have coordinates ( ) and (4 respectively. Calculate (i) The gradient of the line AB [1] The length of the line AB [] (iii)

More information

Math 215 HW #11 Solutions

Math 215 HW #11 Solutions Math 215 HW #11 Solutions 1 Problem 556 Find the lengths and the inner product of 2 x and y [ 2 + ] Answer: First, x 2 x H x [2 + ] 2 (4 + 16) + 16 36, so x 6 Likewise, so y 6 Finally, x, y x H y [2 +

More information

CHAPTER 5. Basic Iterative Methods

CHAPTER 5. Basic Iterative Methods Basic Iterative Methods CHAPTER 5 Solve Ax = f where A is large and sparse (and nonsingular. Let A be split as A = M N in which M is nonsingular, and solving systems of the form Mz = r is much easier than

More information

Machine Learning. Regularization and Feature Selection. Fabio Vandin November 14, 2017

Machine Learning. Regularization and Feature Selection. Fabio Vandin November 14, 2017 Machine Learning Regularization and Feature Selection Fabio Vandin November 14, 2017 1 Regularized Loss Minimization Assume h is defined by a vector w = (w 1,..., w d ) T R d (e.g., linear models) Regularization

More information

On Spectral Factorization and Riccati Equations for Time-Varying Systems in Discrete Time

On Spectral Factorization and Riccati Equations for Time-Varying Systems in Discrete Time On Spectral Factorization and Riccati Equations for Time-Varying Systems in Discrete Time Alle-Jan van der Veen and Michel Verhaegen Delft University of Technology Department of Electrical Engineering

More information

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota

Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work

More information

Cryptanalysis of a hash function, and the modular subset sum problem

Cryptanalysis of a hash function, and the modular subset sum problem Cryptanalysis of a hash function, and the modular subset sum problem Chris Monico Department of Mathematics and Statistics Texas Tech University January 17, 2018 Abstract Recently, Shpilrain and Sosnovski

More information

Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM

Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM Lee, Ching-pei University of Illinois at Urbana-Champaign Joint work with Dan Roth ICML 2015 Outline Introduction Algorithm Experiments

More information

AMS526: Numerical Analysis I (Numerical Linear Algebra)

AMS526: Numerical Analysis I (Numerical Linear Algebra) AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 7: More on Householder Reflectors; Least Squares Problems Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 15 Outline

More information

( ) = ( ) ( ) = ( ) = + = = = ( ) Therefore: , where t. Note: If we start with the condition BM = tab, we will have BM = ( x + 2, y + 3, z 5)

( ) = ( ) ( ) = ( ) = + = = = ( ) Therefore: , where t. Note: If we start with the condition BM = tab, we will have BM = ( x + 2, y + 3, z 5) Chapter Exercise a) AB OB OA ( xb xa, yb ya, zb za),,, 0, b) AB OB OA ( xb xa, yb ya, zb za) ( ), ( ),, 0, c) AB OB OA x x, y y, z z (, ( ), ) (,, ) ( ) B A B A B A ( ) d) AB OB OA ( xb xa, yb ya, zb za)

More information

Linear Regression. Robot Image Credit: Viktoriya Sukhanova 123RF.com

Linear Regression. Robot Image Credit: Viktoriya Sukhanova 123RF.com Linear Regression These slides were assembled by Eric Eaton, with grateful acknowledgement of the many others who made their course materials freely available online. Feel free to reuse or adapt these

More information

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT

MLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT MLCC 2018 Variable Selection and Sparsity Lorenzo Rosasco UNIGE-MIT-IIT Outline Variable Selection Subset Selection Greedy Methods: (Orthogonal) Matching Pursuit Convex Relaxation: LASSO & Elastic Net

More information

Online Nonnegative Matrix Factorization with General Divergences

Online Nonnegative Matrix Factorization with General Divergences Online Nonnegative Matrix Factorization with General Divergences Vincent Y. F. Tan (ECE, Mathematics, NUS) Joint work with Renbo Zhao (NUS) and Huan Xu (GeorgiaTech) IWCT, Shanghai Jiaotong University

More information