Supplemental Material for TKDE
|
|
- Josephine Garrett
- 5 years ago
- Views:
Transcription
1 Supplemental Material for TKDE Kijung Sh, Lee Sael, U Kang PROPOSED METHODS Proofs of Update Rules In this section, we present the proofs of the update rules Section 35 of the ma paper Specifically, we prove the CDTF update rule for L regularization Theorem 7, the CDTF update rule for the nonnegativity constrat Theorem, the SALS update rule for coupled tensor factorization Theorem, and the update rule for bias terms commonly used by CDTF and SALS for the bias model Theorem 0 Lemma Partial Derivative CDTF: For a parameter i, let ˆr i i N x ii N N s k l al, g i,,i N ˆr Ω n ii N l n al k, and d i,,i N Ω n i,,i N Ω i,,i N Ω i,,i N Ω n i,,i N Ω n g l n al k, as the ma paper x i i N K s l x i i N K x i i N x i i N K s l s l ˆr i i N a l a l K s l a l l a l k a l g a l k l n Theorem 7: Correctness of CDTF with L - regularization The update rule the ma paper mimizes the loss function with respect to the updated parameter For an updated parameter, let ˆr i i N x ii N s k N l al, g i,,i N Ω n d i,,i N Ω n paper arg m k ˆr ii N l n al k, and l n al k, as the ma λ g/d if g > λ L LassoA,, A N λ g/d if g < -λ 0 otherwise By Lemma, L LassoA,, A N i,,i N Ω g i d λ an x i i N K i Case : If g > λ > 0, s l { g i d λ g i d λ L Lasso k 0, should be negative for L Lasso k < 0, L Lasso k < 0 makes L Lasso k g a l λ N A l l d λ, and an if an if an > 0 < 0 > d Sce d to be zero If λ g/d 0 Sce L Lasso d 0, k λg/d mimizes L LassoA,, A N with respect to Case : Likewise, if g < λ < 0, L Lasso k Sce d 0, be zero If L Lasso k < should be positive for L Lasso k > 0, L Lasso g k λ g/d > 0 makes L Lasso k d 0, to d λ, and 0 Sce λ g/d mimizes L Lasso A,, A N with respect to i Case 3: On the other hand, if λ g λ, L Lasso { g i d λ an i d 0 g i d λ an i d 0 if an if an i > 0 i < 0 That is, L Lasso A,, A N decreases if i < 0 and creases if > 0, given other parameters Thus, L Lasso A,, A N, which is a contuous
2 function, is mimized with respect to arg m k if an 0 λ g/d if g > λ L LassoA,, A N λ g/d if g < -λ 0 otherwise N l al, Theorem : Correctness of CDTF with the nonnegativity constrat The update rule 4 the ma paper mimizes the loss function with respect to the updated parameter under the nonnegativity constrat For an updated parameter i, let ˆr i i N x ii N s k g i,,i N ˆr Ω n ii N l n al k, and d i,,i N Ω n arg m k 0 l n al k, as the ma paper LA,, A N g max d λ, 0 By Lemma, LA,, A N i,,i N Ω g d λ Thus, L x i i N K s l a l λ N A l F l > 0 if > g/d λ 0 if g/d λ < 0 otherwise Case : If g/dλ 0, sce L k dλ 0, LA,, A N is mimized if g/dλ with respect to i Case : On the other hand, if g/d λ < 0, under the constrat that i 0, LA,, A N is mimized with respect to if an 0 This is because LA,, A N, which is a contuous function, monotonically creases ie, 0 > g/d λ arg m k 0 L k LA,, A N g max d λ, 0 > 0 if Ny n y Ny n y y ii Ny K k i C c c Likewise, let x Ω and y Ω be the sets of dices of the observable entries X and Y, respectively arg m [a i k,,a i k C ] T L Coupled xa,, xa Nx, ya,, ya Ny xb i yb i λi C xc i yc i, where x B i and y B i are C by C matrices whose entries are xb i c c x c x c, n n yb i c c xc i and y c i xc i c yc i c i,,i Nx xω i i,,i Ny yω i y c n y c n, c, c, are length C vectors whose entries are xˆr i i N x, i,,i Nx xω i i,,i Ny yω i y ˆr i i N and I C is the C by C identity matrix L xa,, xa Nx a i k c i,,i Nx xω λ Nx xa n F n xa i k c i,,i Nx xω i n c y c n LxA,, xa Nx xa x i i Nx K x i i Nx K i,,i Nx xω i λa i k c Likewise, xa i k c x i i Nx x s n xa C i k c N x s n i k c x s n K N x s n x x x λ xa i k c, c x s xˆr i i Nx x c n Theorem : Correctness of SALS Coupled Tensor Factorization The update rule 6 the ma paper mimizes 5 with respect to the updated parameters Let xr and y R be the residual tensors for X and Y, respectively That is, for C updated parameters a i k,, a i k C, xˆr ii Nx x ii Nx K Nx k n x i C Nx c n x c and yˆr ii Ny L ya,, ya Ny a i k c C i,,i Ny yω i λa i k c N y y s n s y ˆr i i Ny y c n
3 From these, L Coupled a i k c i,,i Nx xω i 0, c, c C i,,i Ny yω i C N x s n C x s xˆr i i Nx x c n N y y s n λa i k c 0, c C xa i k s i,,i Nx xω i i,,i Ny yω i λa i k c i,,i Nx xω i i,,i Ny yω i s s y ˆr i i Ny x s n C ya i k s s xˆr i i Nx y ˆr i i Ny y s n x c n y c n y c n x n, c c y n c xb i yb i λi C[a i k,, a i k C ] T xc i yc i arg m [a i k,,a i k C ] T L Coupled xa,, xa Nx, ya,, ya Ny xb i yb i λi C xc i yc i Theorem 0: Correctness of the Update Rule for Bias Terms The update rule the ma paper mimizes 7 with respect to the updated parameter For an updated parameter b n, let r ii N x ii N K N k l al k l n bl µ, as the ma paper arg m b n L BiasA,, A N, b,, b N i,,i N Ω n L BiasA,, A N, b,, b N i,,i N Ω λ A l b n x i i N K a l s l b n N A l F b n r i i N /λ b Ω n λ b l N l N b l F b n b l µ i,,i N Ω n x i i N K i,,i N Ω n x i i N a l s l λ b Ω n b n Sce L Bias b n N K s l l a l N b l µ l b n r i i N b n λ b b n i,,i N Ω n r i i N b l µ λ b l 3 N b l F b n λ b Ω n 0, L Bias A,, A N, b,, b N is mimized with respect to b n if b n i,,i N r Ω n ii N /λ b Ω n 0, which entails L Bias b n arg m b n L BiasA,, A N, b,, b N Pseudocodes i,,i N Ω n r i i N /λ b Ω n We present the pseudocodes of the SALS variants described Section 35 of the ma paper SALS for Coupled Tensor Factorization Algorithm 6 describes SALS for coupled tensor factorization, where two tensors, denoted by X and Y, share their first mode without loss of generality We denote the residual tensors for X and Y by x R and yr, respectively The lengths of the n-th modes of X and Y are denoted by x I n and y I n, respectively Algorithm 6: SALS for Coupled Tensor Factorization Input : X, Y, K, λ Output: A, xa n for n N x, ya n for n N y itialize xr, yr, A, xa n and ya n for all n for outer iter T out do 3 for split iter K C do 4 choose k,, k C from columns not updated yet 5 compute x ˆR and y ˆR usg 6 for ner iter T do 7 for i I xi yi do update a i k,, a i k C usg for n N x do for xi n do update x,, x usg C for n N y do for yi n do update y,, y usg C update xr and yr usg 0
4 4 Algorithm 7: SALS for Bias Model Input : X, K, λ A, λ b Output: A n for all n, b n for all n, µ compute µ the mean of the observable entries of X itialize R, A n for all n, and b n for all n 3 for outer iter T out do 4 for split iter K C do 5 choose k,, k C from columns not updated yet 6 compute ˆR usg 7 for ner iter T do for n N do for I n do 0 update,, C usg update R usg 0 for n N do for I n do update b n usg update R usg SALS for Bias Model SALS for the bias model is described Algorithm 7, where each i,, i N th entry of R is r i x ii N µ N n bn K N k n an, as explaed Section 355 of the ma paper OPTIMIZATION ON MAPREDUCE In this section, we present the details of the optimization techniques described Section 4 of the ma paper Local Disk Cachg As explaed Section 4 of the ma paper, our MAPREDUCE implementation of CDTF and SALS with local disk cachg, X entries are distributed across maches and cached their local disk durg the map and reduce stages Algorithm gives the details of the map and reduce stages The rest of CDTF and SALS runs the close stage cleanup stage Hadoop usg the cached data Direct Communication In the ma paper, we troduce direct communication between reducers usg distributed file system to overcome the rigidity of MAPREDUCE model Algorithm describes the implementation of m k broadcast CDTF le 0 of Algorithm 3 the ma paper based on this communication method 3 Greedy Row Assignment Our MAPREDUCE implementation of SALS and CDTF uses the greedy row assignment, explaed Section 343 of the ma paper In this section, we expla our MAPREDUCE implementation of the greedy row assignment We assume that X is stored Algorithm : Data distribution CDTF and SALS with local disk cachg Input : X, ms n for all m and n Output: mω n entries of R X for all m and n MapKey k, Value v beg 3 i,, i N, x i i N v 4 for n,,n do 5 fd m where ms n 6 emit < m, n, i,, i N, x i i N > 7 end PartitionerKey k, Value v beg 0 m, n k assign < k, v > to mache m end 3 ReduceKey k, Value v[ v ] 4 beg 5 m, n k 6 create a file on the local disk to cache mω n entries of R 7 foreach i,, i N, x i i N v do write i,, i N, x i i N to the file end Algorithm : m k broadcast CDTF Input : m k parameters to broadcast Output: k parameters received from others beg create a data file ma on the distributed file system DFS 3 write m k on the datafile 4 create a dummy file md on DFS 5 while not all data files are read do 6 get the list of dummy files from DFS 7 foreach m D the list do if m A are not read then read m k from m A 0 end on the distributed file system At the first stage, Ω n for all n and is computed Specifically, mappers output < n,, > for all n for each entry x ii N, and reducers output < n,, Ω n > for all n and by countg the number of values for each key At the second stage, the outputs are aggregated to a sgle reducer which runs the rest of Algorithm 5 the ma paper 3 EXPERIMENTS In this section, we design and conduct additional experiments to answer the followg questions: How do different numbers of ner iterations T affect the convergence of SALS? How do different numbers of columns updated at a time C affect the runng time of SALS? 3 Experimental Settgs We ran experiments on a 0-node Hadoop cluster Each node had an Intel Xeon E GHz CPU
5 5 a Netflix 3 C 0 b Netflix 3 C 0 c Yahoo-music 4 C 0 d Yahoo-music 4 C 40 Fig 4: Effects of T ie, ner iterations on the convergence of SALS when C ie, the number of columns updated at a time has large enough values The effects of T on convergence speed and the quality of converged solutions are margal a Netflix 3 3 Effects of the Number of Inner Iterations ie, T on the Convergence of SALS We compared the convergence properties of SALS with different T values Especially, we focused on cases where C ie, the number of columns updated at a time has large enough values The effect of T when C is set to one and thus SALS is equivalent to CDTF can be found the ma paper As seen Figure 4, the effects of T on convergence speed and the quality of converged solutions are neither distct nor consistent When C is set to one, however, high T values are preferred see Section 57 of the ma paper for detailed experimental results b Yahoo-music 4 Fig 5: Effects of the number of columns updated at a time C on the runng time of SALS Runng time per iteration decreased until C 0, then started to crease The maximum heap size per reducer was set to GB Other experimental settgs, cludg datasets and parameter values λ and K, were the same as those the ma paper The number of reducers was set to 0 We used the root mean square error RMSE on a held-out test set, which is commonly used recommender systems, to measure the accuracy, as the ma paper 33 Effects of the Number of Columns Updated at a Time ie, C on Runng Time of SALS We measured the runng times per iteration SALS, as we creased C from to K As seen Figure 5, runng time per iteration decreased until C 0, then started to crease As C creases, the amount of disk I/O decles sce it depends on the number of times that the entries of R or ˆR are streamed from disk, which is versely proportional to C Conversely, computational cost creases quadratically with regard to C At small C values, the decrease the amount of disk I/O was greater and leaded to a downward trend of runng time per iteration The opposite happened at large C values
RECOMMENDATION problems can be viewed as
IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING 1 Fully Scalable Methods for Distributed Tensor Factorization Kijung Sh, Lee Sael, and U Kang Abstract Given a high-order large-scale tensor, how can
More informationDistributed Methods for High-dimensional and Large-scale Tensor Factorization
Distributed Methods for High-dimensional and Large-scale Tensor Factorization Kijung Shin Dept. of Computer Science and Engineering Seoul National University Seoul, Republic of Korea koreaskj@snu.ac.kr
More informationParallel Matrix Factorization for Recommender Systems
Under consideration for publication in Knowledge and Information Systems Parallel Matrix Factorization for Recommender Systems Hsiang-Fu Yu, Cho-Jui Hsieh, Si Si, and Inderjit S. Dhillon Department of
More informationScalable Coordinate Descent Approaches to Parallel Matrix Factorization for Recommender Systems
Scalable Coordinate Descent Approaches to Parallel Matrix Factorization for Recommender Systems Hsiang-Fu Yu, Cho-Jui Hsieh, Si Si, and Inderjit Dhillon Department of Computer Science University of Texas
More informationLarge-Scale Behavioral Targeting
Large-Scale Behavioral Targeting Ye Chen, Dmitry Pavlov, John Canny ebay, Yandex, UC Berkeley (This work was conducted at Yahoo! Labs.) June 30, 2009 Chen et al. (KDD 09) Large-Scale Behavioral Targeting
More informationCS425: Algorithms for Web Scale Data
CS425: Algorithms for Web Scale Data Most of the slides are from the Mining of Massive Datasets book. These slides have been modified for CS425. The original slides can be accessed at: www.mmds.org Challenges
More informationLarge-scale Matrix Factorization. Kijung Shin Ph.D. Student, CSD
Large-scale Matrix Factorization Kijung Shin Ph.D. Student, CSD Roadmap Matrix Factorization (review) Algorithms Distributed SGD: DSGD Alternating Least Square: ALS Cyclic Coordinate Descent: CCD++ Experiments
More informationLinear Algebra (Review) Volker Tresp 2018
Linear Algebra (Review) Volker Tresp 2018 1 Vectors k, M, N are scalars A one-dimensional array c is a column vector. Thus in two dimensions, ( ) c1 c = c 2 c i is the i-th component of c c T = (c 1, c
More informationBlock AIR Methods. For Multicore and GPU. Per Christian Hansen Hans Henrik B. Sørensen. Technical University of Denmark
Block AIR Methods For Multicore and GPU Per Christian Hansen Hans Henrik B. Sørensen Technical University of Denmark Model Problem and Notation Parallel-beam 3D tomography exact solution exact data noise
More informationFast and Scalable Distributed Boolean Tensor Factorization
Fast and Scalable Distributed Boolean Tensor Factorization Namyong Park Seoul National University Email: namyong.park@snu.ac.kr Sejoon Oh Seoul National University Email: ohhenrie@snu.ac.kr U Kang Seoul
More informationGIVEN a tensor that is too large to fit in memory, how can
Out-of-Core and Distributed Algorithms for Dense Subtensor Mining Kijung Shin, Bryan Hooi, Jisu Kim, and Christos Faloutsos arxiv:802.0065v2 [cs.db] 6 Feb 208 Abstract How can we detect fraudulent lockstep
More informationCHAPTER 2 EXTRACTION OF THE QUADRATICS FROM REAL ALGEBRAIC POLYNOMIAL
24 CHAPTER 2 EXTRACTION OF THE QUADRATICS FROM REAL ALGEBRAIC POLYNOMIAL 2.1 INTRODUCTION Polynomial factorization is a mathematical problem, which is often encountered in applied sciences and many of
More informationGraph Analysis Using Map/Reduce
Seminar: Massive-Scale Graph Analysis Summer Semester 2015 Graph Analysis Using Map/Reduce Laurent Linden s9lalind@stud.uni-saarland.de May 14, 2015 Introduction Peta-Scale Graph + Graph mining + MapReduce
More informationLinear Algebra (Review) Volker Tresp 2017
Linear Algebra (Review) Volker Tresp 2017 1 Vectors k is a scalar (a number) c is a column vector. Thus in two dimensions, c = ( c1 c 2 ) (Advanced: More precisely, a vector is defined in a vector space.
More informationNeural Networks. ICS 273A UC Irvine Instructor: Max Welling
Neural Networks ICS 273A UC Irve Instructor: Max Wellg Neurons 1 b Neurons communicate by receivg signals on their dendrites. Addg these signals and firg off a new signal along the axon if the total put
More informationTall-and-skinny! QRs and SVDs in MapReduce
A 1 A 2 A 3 Tall-and-skinny! QRs and SVDs in MapReduce Yangyang Hou " Purdue, CS Austin Benson " Stanford University Paul G. Constantine Col. School. Mines " Joe Nichols U. of Minn James Demmel " UC Berkeley
More informationFor more information visit here:
The length or the magnitude of the vector = (a, b, c) is defined by w = a 2 +b 2 +c 2 A vector may be divided by its own length to convert it into a unit vector, i.e.? = u / u. (The vectors have been denoted
More informationMultiplication of Polynomials
Summary 391 Chapter 5 SUMMARY Section 5.1 A polynomial in x is defined by a finite sum of terms of the form ax n, where a is a real number and n is a whole number. a is the coefficient of the term. n is
More informationRegression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)
Linear Regression Regression Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Example: Height, Gender, Weight Shoe Size Audio features
More informationRegression. Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning)
Linear Regression Regression Goal: Learn a mapping from observations (features) to continuous labels given a training set (supervised learning) Example: Height, Gender, Weight Shoe Size Audio features
More informationEEC 686/785 Modeling & Performance Evaluation of Computer Systems. Lecture 11
EEC 686/785 Modeling & Performance Evaluation of Computer Systems Lecture Department of Electrical and Computer Engineering Cleveland State University wenbing@ieee.org (based on Dr. Raj Jain s lecture
More informationA robust multilevel approximate inverse preconditioner for symmetric positive definite matrices
DICEA DEPARTMENT OF CIVIL, ENVIRONMENTAL AND ARCHITECTURAL ENGINEERING PhD SCHOOL CIVIL AND ENVIRONMENTAL ENGINEERING SCIENCES XXX CYCLE A robust multilevel approximate inverse preconditioner for symmetric
More informationUsing R for Iterative and Incremental Processing
Using R for Iterative and Incremental Processing Shivaram Venkataraman, Indrajit Roy, Alvin AuYoung, Robert Schreiber UC Berkeley and HP Labs UC BERKELEY Big Data, Complex Algorithms PageRank (Dominant
More informationDiscrete Multi-material Topology Optimization under Total Mass Constraint
Discrete Multi-material Topology Optimization under Total Mass Constraint Xingtong Yang Ming Li State Key Laboratory of CAD&CG, Zhejiang University Solid and Physical Modeling Bilbao, June 12, 2018 Problem
More informationLecture notes 1: ECEN 489
Lecture notes : ECEN 489 Power Management Circuits and Systems Department of Electrical & Computer Engeerg Texas A&M University Jose Silva-Martez January 207 Copyright Texas A&M University. All rights
More informationIncomplete Cholesky preconditioners that exploit the low-rank property
anapov@ulb.ac.be ; http://homepages.ulb.ac.be/ anapov/ 1 / 35 Incomplete Cholesky preconditioners that exploit the low-rank property (theory and practice) Artem Napov Service de Métrologie Nucléaire, Université
More informationk-variates++: more pluses in the k-means++
(formerly NICTA) k-variates++: more pluses in the k-means++ m p 7 3 n o M, 9 Poster #2 Richard Nock, Raphaël Canyasse, Roksana Boreli, Frank Nielsen DATA61 ANU TECHNION ECOLE POLYTECHNIQUE UNSW SONY CS
More informationStreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory
StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory S.V. N. (vishy) Vishwanathan Purdue University and Microsoft vishy@purdue.edu October 9, 2012 S.V. N. Vishwanathan (Purdue,
More informationl 1 and l 2 Regularization
David Rosenberg New York University February 5, 2015 David Rosenberg (New York University) DS-GA 1003 February 5, 2015 1 / 32 Tikhonov and Ivanov Regularization Hypothesis Spaces We ve spoken vaguely about
More informationMath Models of OR: Branch-and-Bound
Math Models of OR: Branch-and-Bound John E. Mitchell Department of Mathematical Sciences RPI, Troy, NY 12180 USA November 2018 Mitchell Branch-and-Bound 1 / 15 Branch-and-Bound Outline 1 Branch-and-Bound
More informationChapter 7 Iterative Techniques in Matrix Algebra
Chapter 7 Iterative Techniques in Matrix Algebra Per-Olof Persson persson@berkeley.edu Department of Mathematics University of California, Berkeley Math 128B Numerical Analysis Vector Norms Definition
More informationIncorporating uncertainty in the design of water distribution systems
European Water 58: 449-456, 2017. 2017 E.W. Publications Incorporatg uncertaty the design of water distribution systems M. Spiliotis 1* and G. Tsakiris 2 1 Division of Hydraulic Engeerg, Department of
More informationSOLUTIONS TO ADDITIONAL EXERCISES FOR II.1 AND II.2
SOLUTIONS TO ADDITIONAL EXERCISES FOR II.1 AND II.2 Here are the solutions to the additional exercises in betsepexercises.pdf. B1. Let y and z be distinct points of L; we claim that x, y and z are not
More informationTHE solution of the absolute value equation (AVE) of
The nonlinear HSS-like iterative method for absolute value equations Mu-Zheng Zhu Member, IAENG, and Ya-E Qi arxiv:1403.7013v4 [math.na] 2 Jan 2018 Abstract Salkuyeh proposed the Picard-HSS iteration method
More informationSPARSE SOLVERS POISSON EQUATION. Margreet Nool. November 9, 2015 FOR THE. CWI, Multiscale Dynamics
SPARSE SOLVERS FOR THE POISSON EQUATION Margreet Nool CWI, Multiscale Dynamics November 9, 2015 OUTLINE OF THIS TALK 1 FISHPACK, LAPACK, PARDISO 2 SYSTEM OVERVIEW OF CARTESIUS 3 POISSON EQUATION 4 SOLVERS
More informationThermodynamics [ENGR 251] [Lyes KADEM 2007]
CHAPTER V The first law of thermodynamics is a representation of the conservation of energy. It is a necessary, but not a sufficient, condition for a process to occur. Indeed, no restriction is imposed
More informationBehavioral Data Mining. Lecture 7 Linear and Logistic Regression
Behavioral Data Mining Lecture 7 Linear and Logistic Regression Outline Linear Regression Regularization Logistic Regression Stochastic Gradient Fast Stochastic Methods Performance tips Linear Regression
More informationA Hybrid Method for the Wave Equation. beilina
A Hybrid Method for the Wave Equation http://www.math.unibas.ch/ beilina 1 The mathematical model The model problem is the wave equation 2 u t 2 = (a 2 u) + f, x Ω R 3, t > 0, (1) u(x, 0) = 0, x Ω, (2)
More informationQR Decomposition in a Multicore Environment
QR Decomposition in a Multicore Environment Omar Ahsan University of Maryland-College Park Advised by Professor Howard Elman College Park, MD oha@cs.umd.edu ABSTRACT In this study we examine performance
More informationNesterov-based Alternating Optimization for Nonnegative Tensor Completion: Algorithm and Parallel Implementation
Nesterov-based Alternating Optimization for Nonnegative Tensor Completion: Algorithm and Parallel Implementation Georgios Lourakis and Athanasios P. Liavas School of Electrical and Computer Engineering,
More informationParallelized Variational EM for Latent Dirichlet Allocation: An Experimental Evaluation of Speed and Scalability
Parallelized Variational EM for Latent Dirichlet Allocation: An Experimental Evaluation of Speed and Scalability Ramesh Nallapati, William Cohen and John Lafferty Machine Learning Department Carnegie Mellon
More informationLarge-Scale Matrix Factorization with Distributed Stochastic Gradient Descent
Large-Scale Matrix Factorization with Distributed Stochastic Gradient Descent KDD 2011 Rainer Gemulla, Peter J. Haas, Erik Nijkamp and Yannis Sismanis Presenter: Jiawen Yao Dept. CSE, UT Arlington 1 1
More informationLecture XI. Approximating the Invariant Distribution
Lecture XI Approximating the Invariant Distribution Gianluca Violante New York University Quantitative Macroeconomics G. Violante, Invariant Distribution p. 1 /24 SS Equilibrium in the Aiyagari model G.
More informationCS246 Final Exam, Winter 2011
CS246 Final Exam, Winter 2011 1. Your name and student ID. Name:... Student ID:... 2. I agree to comply with Stanford Honor Code. Signature:... 3. There should be 17 numbered pages in this exam (including
More informationScheduling with AND/OR Precedence Constraints
Scheduling with AND/OR Precedence Constraints Seminar Mathematische Optimierung - SS 2007 23th April 2007 Synthesis synthesis: transfer from the behavioral domain (e. g. system specifications, algorithms)
More informationMath 304 (Spring 2010) - Lecture 2
Math 304 (Spring 010) - Lecture Emre Mengi Department of Mathematics Koç University emengi@ku.edu.tr Lecture - Floating Point Operation Count p.1/10 Efficiency of an algorithm is determined by the total
More informationLINEAR COMPARTMENTAL MODELS: INPUT-OUTPUT EQUATIONS AND OPERATIONS THAT PRESERVE IDENTIFIABILITY. 1. Introduction
LINEAR COMPARTMENTAL MODELS: INPUT-OUTPUT EQUATIONS AND OPERATIONS THAT PRESERVE IDENTIFIABILITY ELIZABETH GROSS, HEATHER HARRINGTON, NICOLETTE MESHKAT, AND ANNE SHIU Abstract. This work focuses on the
More informationA Proof of Theorem 1. Dual Weighted Generalized Nuclear Norm. C Proof of Theorem 2
A Proof of Theorem Proof or all Z = P i c i i with kak A = P i c i, where i = Au i v T i BT, we can construct the i-th column of W and H as Clearly, we have Z = WH T and w i = p c i Au i and h i = p c
More informationSOLUTIONS TO EXERCISES FOR. MATHEMATICS 133 Part 2. I. Topics from linear algebra
SOLUTIONS TO EXERCISES FOR MATHEMATICS 133 Part Fall 013 NOTE ON ILLUSTRATIONS. Drawings for several of the solutions in this file are available in the following document: http://math.ucr.edu/ res/math133/math133solutionsa.figures.f13.pdf
More informationHigh Performance Parallel Tucker Decomposition of Sparse Tensors
High Performance Parallel Tucker Decomposition of Sparse Tensors Oguz Kaya INRIA and LIP, ENS Lyon, France SIAM PP 16, April 14, 2016, Paris, France Joint work with: Bora Uçar, CNRS and LIP, ENS Lyon,
More informationSome definitions. Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization. A-inner product. Important facts
Some definitions Math 1080: Numerical Linear Algebra Chapter 5, Solving Ax = b by Optimization M. M. Sussman sussmanm@math.pitt.edu Office Hours: MW 1:45PM-2:45PM, Thack 622 A matrix A is SPD (Symmetric
More informationCache Oblivious Stencil Computations
Cache Oblivious Stencil Computations S. HUNOLD J. L. TRÄFF F. VERSACI Lectures on High Performance Computing 13 April 2015 F. Versaci (TU Wien) Cache Oblivious Stencil Computations 13 April 2015 1 / 19
More informationMatrix Factorization Techniques for Recommender Systems
Matrix Factorization Techniques for Recommender Systems Patrick Seemann, December 16 th, 2014 16.12.2014 Fachbereich Informatik Recommender Systems Seminar Patrick Seemann Topics Intro New-User / New-Item
More informationSolving Regularized Total Least Squares Problems
Solving Regularized Total Least Squares Problems Heinrich Voss voss@tu-harburg.de Hamburg University of Technology Institute of Numerical Simulation Joint work with Jörg Lampe TUHH Heinrich Voss Total
More informationLinear Regression (continued)
Linear Regression (continued) Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 6, 2017 1 / 39 Outline 1 Administration 2 Review of last lecture 3 Linear regression
More informationarxiv: v3 [stat.ml] 14 Apr 2016
arxiv:1307.0048v3 [stat.ml] 14 Apr 2016 Simple one-pass algorithm for penalized linear regression with cross-validation on MapReduce Kun Yang April 15, 2016 Abstract In this paper, we propose a one-pass
More informationLFTF: A Framework for Efficient Tensor Analytics at Scale
: A Framework for Efficient Tensor Analytics at Scale Fan Yang Fanhua Shang Yuzhen Huang James Cheng Jinfeng Li Yunjian Zhao Ruihao Zhao Department of Computer Science and Engineering The Chinese University
More informationECE521 W17 Tutorial 1. Renjie Liao & Min Bai
ECE521 W17 Tutorial 1 Renjie Liao & Min Bai Schedule Linear Algebra Review Matrices, vectors Basic operations Introduction to TensorFlow NumPy Computational Graphs Basic Examples Linear Algebra Review
More informationEfficient Primal- Dual Graph Algorithms for Map Reduce
Efficient Primal- Dual Graph Algorithms for Map Reduce Joint work with Bahman Bahmani Ashish Goel, Stanford Kamesh Munagala Duke University Modern Data Models Over the past decade, many commodity distributed
More informationKnowledge Discovery and Data Mining 1 (VO) ( )
Knowledge Discovery and Data Mining 1 (VO) (707.003) Map-Reduce Denis Helic KTI, TU Graz Oct 24, 2013 Denis Helic (KTI, TU Graz) KDDM1 Oct 24, 2013 1 / 82 Big picture: KDDM Probability Theory Linear Algebra
More informationEnvironment (Parallelizing Query Optimization)
Advanced d Query Optimization i i Techniques in a Parallel Computing Environment (Parallelizing Query Optimization) Wook-Shin Han*, Wooseong Kwak, Jinsoo Lee Guy M. Lohman, Volker Markl Kyungpook National
More informationCS246: Mining Massive Data Sets Winter Only one late period is allowed for this homework (11:59pm 2/14). General Instructions
CS246: Mining Massive Data Sets Winter 2017 Problem Set 2 Due 11:59pm February 9, 2017 Only one late period is allowed for this homework (11:59pm 2/14). General Instructions Submission instructions: These
More informationAn Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion
An Extended Frank-Wolfe Method, with Application to Low-Rank Matrix Completion Robert M. Freund, MIT joint with Paul Grigas (UC Berkeley) and Rahul Mazumder (MIT) CDC, December 2016 1 Outline of Topics
More informationMapReduce in Spark. Krzysztof Dembczyński. Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland
MapReduce in Spark Krzysztof Dembczyński Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland Software Development Technologies Master studies, second semester
More informationNetwork Newton. Aryan Mokhtari, Qing Ling and Alejandro Ribeiro. University of Pennsylvania, University of Science and Technology (China)
Network Newton Aryan Mokhtari, Qing Ling and Alejandro Ribeiro University of Pennsylvania, University of Science and Technology (China) aryanm@seas.upenn.edu, qingling@mail.ustc.edu.cn, aribeiro@seas.upenn.edu
More informationMapReduce in Spark. Krzysztof Dembczyński. Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland
MapReduce in Spark Krzysztof Dembczyński Intelligent Decision Support Systems Laboratory (IDSS) Poznań University of Technology, Poland Software Development Technologies Master studies, first semester
More informationNotes on Mathematics Groups
EPGY Singapore Quantum Mechanics: 2007 Notes on Mathematics Groups A group, G, is defined is a set of elements G and a binary operation on G; one of the elements of G has particularly special properties
More informationCSE 417T: Introduction to Machine Learning. Final Review. Henry Chai 12/4/18
CSE 417T: Introduction to Machine Learning Final Review Henry Chai 12/4/18 Overfitting Overfitting is fitting the training data more than is warranted Fitting noise rather than signal 2 Estimating! "#$
More information2.6 Large Signal Operation
3/4/2011 section 2_6 Large signal operation 1/2 2.6 Large Signal Operation Readg Assignment: 94-98 Recall that real amplifiers are only approximately lear! If the put signal becomes too large, and/or the
More informationIntroduction to Compressed Sensing
Introduction to Compressed Sensing Alejandro Parada, Gonzalo Arce University of Delaware August 25, 2016 Motivation: Classical Sampling 1 Motivation: Classical Sampling Issues Some applications Radar Spectral
More informationSIMPLEX LIKE (aka REDUCED GRADIENT) METHODS. REDUCED GRADIENT METHOD (Wolfe)
19 SIMPLEX LIKE (aka REDUCED GRADIENT) METHODS The REDUCED GRADIENT algorithm and its variants such as the CONVEX SIMPLEX METHOD (CSM) and the GENERALIZED REDUCED GRADIENT (GRG) algorithm are approximation
More informationContinuous methods for numerical linear algebra problems
Continuous methods for numerical linear algebra problems Li-Zhi Liao (http://www.math.hkbu.edu.hk/ liliao) Department of Mathematics Hong Kong Baptist University The First International Summer School on
More informationMulti-Approximate-Keyword Routing Query
Bin Yao 1, Mingwang Tang 2, Feifei Li 2 1 Department of Computer Science and Engineering Shanghai Jiao Tong University, P. R. China 2 School of Computing University of Utah, USA Outline 1 Introduction
More informationa b = a T b = a i b i (1) i=1 (Geometric definition) The dot product of two Euclidean vectors a and b is defined by a b = a b cos(θ a,b ) (2)
This is my preperation notes for teaching in sections during the winter 2018 quarter for course CSE 446. Useful for myself to review the concepts as well. More Linear Algebra Definition 1.1 (Dot Product).
More informationBidirectional Representation and Backpropagation Learning
Int'l Conf on Advances in Big Data Analytics ABDA'6 3 Bidirectional Representation and Bacpropagation Learning Olaoluwa Adigun and Bart Koso Department of Electrical Engineering Signal and Image Processing
More informationLecture 9: September 28
0-725/36-725: Convex Optimization Fall 206 Lecturer: Ryan Tibshirani Lecture 9: September 28 Scribes: Yiming Wu, Ye Yuan, Zhihao Li Note: LaTeX template courtesy of UC Berkeley EECS dept. Disclaimer: These
More informationChapter Two Elements of Linear Algebra
Chapter Two Elements of Linear Algebra Previously, in chapter one, we have considered single first order differential equations involving a single unknown function. In the next chapter we will begin to
More informationLearning Multiple Tasks with a Sparse Matrix-Normal Penalty
Learning Multiple Tasks with a Sparse Matrix-Normal Penalty Yi Zhang and Jeff Schneider NIPS 2010 Presented by Esther Salazar Duke University March 25, 2011 E. Salazar (Reading group) March 25, 2011 1
More informationFPGA Implementation of a Predictive Controller
FPGA Implementation of a Predictive Controller SIAM Conference on Optimization 2011, Darmstadt, Germany Minisymposium on embedded optimization Juan L. Jerez, George A. Constantinides and Eric C. Kerrigan
More informationCSE546 Machine Learning, Autumn 2016: Homework 1
CSE546 Machine Learning, Autumn 2016: Homework 1 Due: Friday, October 14 th, 5pm 0 Policies Writing: Please submit your HW as a typed pdf document (not handwritten). It is encouraged you latex all your
More informationWeb Appendix for The Value of Reputation in an Online Freelance Marketplace
Web Appendix for The Value of Reputation in an Online Freelance Marketplace Hema Yoganarasimhan 1 Details of Second-Step Estimation All the estimates from state transitions (θ p, θ g, θ t ), η i s, and
More informationMULTIPLE PRODUCTS OBJECTIVES. If a i j,b j k,c i k, = + = + = + then a. ( b c) ) 8 ) 6 3) 4 5). If a = 3i j+ k and b 3i j k = = +, then a. ( a b) = ) 0 ) 3) 3 4) not defined { } 3. The scalar a. ( b c)
More informationDefinitions and Properties of R N
Definitions and Properties of R N R N as a set As a set R n is simply the set of all ordered n-tuples (x 1,, x N ), called vectors. We usually denote the vector (x 1,, x N ), (y 1,, y N ), by x, y, or
More informationCS224W: Methods of Parallelized Kronecker Graph Generation
CS224W: Methods of Parallelized Kronecker Graph Generation Sean Choi, Group 35 December 10th, 2012 1 Introduction The question of generating realistic graphs has always been a topic of huge interests.
More informationRecommendation Systems
Recommendation Systems Popularity Recommendation Systems Predicting user responses to options Offering news articles based on users interests Offering suggestions on what the user might like to buy/consume
More informationTransmission Line Input Impedance
1/22/23 Transmission e Input Impedance.doc 1/9 Transmission e Input Impedance Consider a lossless le, length, termated with a load. I(z) I + V (z) -, β + V - z z What is the put impedance of this le? Q:
More informationScalable Nonnegative Matrix Factorization with Block-wise Updates
Scalable Nonnegative Matrix Factorization with Block-wise Updates Jiangtao Yin 1, Lixin Gao 1, and Zhongfei (Mark) Zhang 2 1 University of Massachusetts Amherst, Amherst, MA 01003, USA 2 Binghamton University,
More informationAdditional Mathematics Lines and circles
Additional Mathematics Lines and circles Topic assessment 1 The points A and B have coordinates ( ) and (4 respectively. Calculate (i) The gradient of the line AB [1] The length of the line AB [] (iii)
More informationMath 215 HW #11 Solutions
Math 215 HW #11 Solutions 1 Problem 556 Find the lengths and the inner product of 2 x and y [ 2 + ] Answer: First, x 2 x H x [2 + ] 2 (4 + 16) + 16 36, so x 6 Likewise, so y 6 Finally, x, y x H y [2 +
More informationCHAPTER 5. Basic Iterative Methods
Basic Iterative Methods CHAPTER 5 Solve Ax = f where A is large and sparse (and nonsingular. Let A be split as A = M N in which M is nonsingular, and solving systems of the form Mz = r is much easier than
More informationMachine Learning. Regularization and Feature Selection. Fabio Vandin November 14, 2017
Machine Learning Regularization and Feature Selection Fabio Vandin November 14, 2017 1 Regularized Loss Minimization Assume h is defined by a vector w = (w 1,..., w d ) T R d (e.g., linear models) Regularization
More informationOn Spectral Factorization and Riccati Equations for Time-Varying Systems in Discrete Time
On Spectral Factorization and Riccati Equations for Time-Varying Systems in Discrete Time Alle-Jan van der Veen and Michel Verhaegen Delft University of Technology Department of Electrical Engineering
More informationMultilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota
Multilevel low-rank approximation preconditioners Yousef Saad Department of Computer Science and Engineering University of Minnesota SIAM CSE Boston - March 1, 2013 First: Joint work with Ruipeng Li Work
More informationCryptanalysis of a hash function, and the modular subset sum problem
Cryptanalysis of a hash function, and the modular subset sum problem Chris Monico Department of Mathematics and Statistics Texas Tech University January 17, 2018 Abstract Recently, Shpilrain and Sosnovski
More informationDistributed Box-Constrained Quadratic Optimization for Dual Linear SVM
Distributed Box-Constrained Quadratic Optimization for Dual Linear SVM Lee, Ching-pei University of Illinois at Urbana-Champaign Joint work with Dan Roth ICML 2015 Outline Introduction Algorithm Experiments
More informationAMS526: Numerical Analysis I (Numerical Linear Algebra)
AMS526: Numerical Analysis I (Numerical Linear Algebra) Lecture 7: More on Householder Reflectors; Least Squares Problems Xiangmin Jiao SUNY Stony Brook Xiangmin Jiao Numerical Analysis I 1 / 15 Outline
More information( ) = ( ) ( ) = ( ) = + = = = ( ) Therefore: , where t. Note: If we start with the condition BM = tab, we will have BM = ( x + 2, y + 3, z 5)
Chapter Exercise a) AB OB OA ( xb xa, yb ya, zb za),,, 0, b) AB OB OA ( xb xa, yb ya, zb za) ( ), ( ),, 0, c) AB OB OA x x, y y, z z (, ( ), ) (,, ) ( ) B A B A B A ( ) d) AB OB OA ( xb xa, yb ya, zb za)
More informationLinear Regression. Robot Image Credit: Viktoriya Sukhanova 123RF.com
Linear Regression These slides were assembled by Eric Eaton, with grateful acknowledgement of the many others who made their course materials freely available online. Feel free to reuse or adapt these
More informationMLCC 2018 Variable Selection and Sparsity. Lorenzo Rosasco UNIGE-MIT-IIT
MLCC 2018 Variable Selection and Sparsity Lorenzo Rosasco UNIGE-MIT-IIT Outline Variable Selection Subset Selection Greedy Methods: (Orthogonal) Matching Pursuit Convex Relaxation: LASSO & Elastic Net
More informationOnline Nonnegative Matrix Factorization with General Divergences
Online Nonnegative Matrix Factorization with General Divergences Vincent Y. F. Tan (ECE, Mathematics, NUS) Joint work with Renbo Zhao (NUS) and Huan Xu (GeorgiaTech) IWCT, Shanghai Jiaotong University
More information