Regularized NNLS Algorithms for Nonnegative Matrix Factorization with Application to Text Document Clustering

Size: px
Start display at page:

Download "Regularized NNLS Algorithms for Nonnegative Matrix Factorization with Application to Text Document Clustering"

Transcription

1 Regularized NNLS Algorithms for Nonnegative Matrix Factorization with Application to Text Document Clustering Rafal Zdunek Abstract. Nonnegative Matrix Factorization (NMF) has recently received much attention both in an algorithmic aspect as well as in applications. Text document clustering and supervised classification are important applications of NMF. Various types of numerical optimization algorithms have been proposed for NMF, which includes multiplicative, projected gradient descent, alternating least squares and active-set ones. In this paper, we discuss the selected Non-Negatively constrained Least Squares (NNLS) algorithms (a family of the NNLS algorithm proposed by Lawson and Hanson) that belong to a class of active-set methods. We noticed that applying the NNLS algorithm to the Tikhonov regularized LS objective function with a regularization parameter exponentially decreasing considerably increases the accuracy of data clustering as well as it reduces the risk of getting stuck into unfavorable local minima. Moreover, the experiments demonstrate that the regularized NNLS algorithm is superior to many well-known NMF algorithms used for text document clustering. 1 Introduction NMF decomposes an input matrix into lower-rank factors that have nonnegative values and usually some physical meaning or interpretation. The reduced dimensionality is very useful in data analysis due to many advantages such as denoising, computational efficiency, greater interpretability and easier visualization. Hence, NMF has already found diverse applications [2, 4, 7, 11, 12, 18, 20, 21, 22, 23, 24] in multivariate data analysis, machine learning, and signal/image processing. Text document clustering can be regarded as the unsupervised classification of documents into groups (clusters) that have similar semantic features. The Rafal Zdunek Institute of Telecommunications, Teleinformatics and Acoustics, Wroclaw University of Technology, Wybrzeze Wyspianskiego 27, Wroclaw, Poland rafal.zdunek@pwr.wroc.pl R. Burduk et al. (Eds.): Computer Recognition Systems 4, AISC 95, pp springerlink.com Springer-Verlag Berlin Heidelberg 2011

2 758 R. Zdunek grouping can be basically obtained with hierarchical (tree-like structure) or partitional (flat clustering) techniques [13]. Xu et al [24] showed that NMF can perform text document partitioning by capturing semantic features in a document collection and grouping the documents according to similarity of these features. Different from Latent Semantic Indexing (LSI), NMF provides nonnegative semantic vectors that do not have to be orthogonal or even independent. Moreover, the coefficients of linear combinations of the semantic vectors are also nonnegative, which means that only additive combinations are permitted. The semantic vectors in NMF have also easier interpretation - namely each of them represents a base topic. Thus, each document can be represented by an additive combination of the base topics. Partitional clustering can be readily performed in the space of linear combination coefficients. An intensive development of NMF-based techniques for document clustering brought in a variety of nonnegativity constrained optimization algorithms. Xu et al [24] applied the standard and weighted Lee-Seung multiplicative algorithms to document clustering, and demonstrated that NMF outperforms SVD-based clustering methods. Shahnaz et al [22] proposed the GD-CLS algorithm that combines the standard multiplicative algorithm for updating the semantic vectors and the regularized projected LS algorithm with a constant regularization parameter for updating the other factor. Several NMF algorithms (tri-nmf, semi-orthogonal and convex NMF) for partitional clustering have been developed by Ding et al [8, 9, 10, 19]. Another group of NMF algorithms is based on so-called Graph NMF (GNMF) [6] that incorporates a priori information on a local structure of clustered data. Locality Preserving NMF (LPNMF), which has been proposed by Cai et al [5] for document clustering, is an extension of GNMF. The above-mentioned algorithms mostly perform multiplicative updates that assure a monotonic convergence but their convergence rate is usually very small. To accelerate the slow convergence, several additive update algorithms have been proposed, including Projected Gradient (PG) descent, Alternating Least Squares (ALS) and active-set algorithms. A survey of PG and ALS algorithms for NMF can be found in [7, 25]. Since the estimated factors in NMF-based document clustering are expected to be large and very sparse, a good choice seems to be active-set algorithms. In this paper, we discuss the selected active-set algorithms that are inspired by the Non-Negative Least Squares (NNLS) algorithm, proposed by Lawson and Hanson [17] in Bro and de Jong [3] considerably accelerate the NNLS algorithm by rearranging computations for cross-product matrices. The solution given by the NNLS algorithms is proved to be optimal according to the Karush-Kuhn-Tucker (KKT) conditions. Unfortunately, these algorithms are not very efficient for solving nonnegativity constrained linear systems with multiple Right-Hand Side (RHS) vectors since they compute a complete pseudoinverse once for each RHS. To tackle this problem, Benthem and Keenan [1] devised the Fast Combinatorial NNLS (FC- NNLS) algorithm and experimentally demonstrated it works efficiently for energydispersive X-ray spectroscopy data. The FC-NNLS algorithm works also efficiently for some NMF problems. Kim and Park [16] applied the FC-NNLS algorithm to

3 Regularized NNLS Algorithms for Text Document Clustering 759 the l 1 - and l 2 -norm regularized LS problems in NMF, and showed that the regularized NNLS algorithms work very efficiently for gene expression microarrays. Their approach assumes constant regularization parameters to enforce the desired degree of sparsity. Here we apply the FC-NNLS algorithm to text document clustering, assuming the regularization parameter decrease exponentially with iterations to trigger a given character of iterative updates. The paper is organized in the following way. The next section discusses the concept of NMF for document clustering. Section 3 is concerned with the NNLS algorithms. The experiments for text document clustering are presented in Section 4. Finally, the conclusions are given in Section 5. 2 NMF for Document Clustering According to the model of document representation introduced in [24], first the collection of documents to be clustered is subject to stop-words removal and words stemming, and then the whole processed collection is represented by the termdocument matrix Y = [y it ] R I T +, where T is the number of documents in the document collection, I is the number of words after preprocessing of the documents, and R + is the nonnegative orthant of the Euclidean space R. Each entry of Y is modeled by: ( ) T y it = t it log, (1) d i where t it denotes the frequency of the i-th word in the t-th document, and d i is the number of documents containing the i-th word. Additionally, each column vector of Y is normalized to the unit l 2 -norm. The matrix Y is very sparse, since each document contains only a small portion of the words that occur in the whole collection of documents. The aim of NMF is to find such lower-rank nonnegative matrices A R I J and X R J T that Y = AX R I T, given the matrix Y, the lower rank J (the number of topics), and possibly a prior knowledge on the matrices A and X. Assuming each column vector of Y = [y 1,...,y T ] represents a single observation (a datum point in R I ), and J is a priori known number of clusters (topics), we can interpret the column vectors of A = [a 1,...,a J ] as the semantic feature vectors or centroids (indicating the directions of central points of clusters in R I ) and the entry x in X = [x ] as an indicator (probability) how the t-th document is related to the j-th topic. To estimate the matrices A and X from Y, we assume the squared Euclidean distance: D(Y AX) = 1 2 Y AX 2 F. (2) The gradient matrices of (2) with respect to A and X are expressed by: G A = [g (A) i j ] = A D(Y AX) = (AX Y)X T R I J, (3)

4 760 R. Zdunek G X = [g (X) ] = X D(Y AX) = A T (AX Y) R J T. (4) From the stationarity of the objective function, we have G A 0 and G X 0, which leads to the ALS algorithm: A YX T (XX T ) 1, X (A T A) 1 A T Y. (5) The update rules (5) do not guarantee nonnegative solutions which are essential for NMF. Some operations are necessary to enforce nonnegative entries. The simplest approach is to replace negative entries with zero-values, which leads to the projected ALS algorithm: where A P ΩA [ YX T (XX T ) 1], X P ΩX [ (A T A) 1 A T Y ], (6) Ω A = A R I J : a i j 0 }, Ω X = X R J T : x 0 }. (7) Unfortunately, the projected ALS algorithm does not assure that the limit point is optimal in the sense it satisfies the KKT optimality conditions: g (A) i j = 0, if a i j > 0, and g (A) i j > 0, if a i j = 0. (8) g (X) = 0, if x > 0, and g (X) > 0, if x = 0, (9) and the complementary slackness: g (A) i j a i j = 0, g (X) x = 0. (10) 3 NNLS Algorithms Several update rules have been proposed in the literature that satisfy the KKT conditions. Here we restrict our considerations to the NNLS algorithms for updating the matrix X, given the matrices A and Y. The discussed algorithms can also update the matrix A by applying them to the transposed system: X T A T = Y T. 3.1 NNLS Algorithm for Single RHS The NNLS algorithm was originally proposed by Lawson and Hanson [17], and currently it is used by the command lsqnonneg(.) in Matlab. This algorithm (denoted by LH-NNLS) iteratively partitions the unknown variables into the passive set P that contains basic variables and the active set R that contains active constraints, and updates only the basic variables until the complementary slackness condition in (10) is met. Let P = j : x > 0 } and R = 1,...,J}\P, and the partition: t : x t = [x (P) t ;x (R) t ] T R J, and g t = xt D(y t Ax t ) = [g (P) t ;g (R) t ] T R J.

5 Regularized NNLS Algorithms for Text Document Clustering 761 The columns of A can be also partitioned in the similar way: A = [A P A R ], where A P = [a,p ] and A R = [a,r ]. The basic variables can be estimated by solving the unconstrained LS problem: min A P x (P) x t (P) t y t 2 }, (11) where A P has full column rank. The nonbasic variables in the KKT optimality point should be equal zero. In contrary to the projected ALS, the NNLS algorithm does not replace negative entries with zero-values, which is equivalent to determine nonbasic variables from unconstrained LS updates, but it starts from all nonbasic variables and tries to iteratively update a set of basic variables. The LH-NNLS algorithm is given by Algorithm 1. Algorithm 1: LH-NNLS Algorithm Input : A R I J, y t = R I Output: x t 0 such that x t = argmin xt Ax t y t 2 1 Initialization: P = /0, R = 1,...,J}, x t = 0, g t = A T y t ; 2 repeat 3 k k+ 1 ; 4 m = argmin j R g }; // the constraint to add 5 if g mt < 0 then 6 P P m, and R R\m; 7 else 8 stop with x t as an optimal solution x t (P) = ( (A P ) T ) 1 9 A P (AP ) T y t where A P = [a,p ] R I P ; 10 while x (P) 0 for j = 1,...,J do x (P) α = min j P x (P) x (P) x (P) ; // the step length 11 0 [ x t x t (P) + α( x t (P) x t (P) T 12 ); 0] ; N = j : x = 0 } 13 ; // the constraint to drop 14 P P\N, and R R N ; x t (P) = ( (A P ) T ) 1 15 A P (AP ) T y t where A P = [a,p ] R I P ; [ x t x t (P) T 16 ; 0] ; 17 until R = /0 or max j R g } < tol ; Bro and de Jong [3] considerably speed up this algorithm for I >> J by precomputing the normal matrix A T A and the vector A T y t, and then replace the steps 9 and 15 in Algorithm 1 with:

6 762 R. Zdunek x (P) t = ( (A T A) P,P ) 1(A T y t ) P. (12) Unfortunately, the inverse of (A T A) P,P must be computed for each t, which is very expensive if the number of RHS is very large. 3.2 Regularized NNLS Algorithm for Multiple RHS Van Benthem and Keenan [1] tackled the problem of a high computational cost for multiple RHS. They noticed that for a sparse solution with multiple column vectors, a probability of finding columns vectors that have the same layout of the zero-entries (active constraints) is high. Hence, after detecting such vectors in X, their passive entries can be updated computing the inverse of (A T A) P,P only once. The NNLS algorithm proposed by Van Benthem and Keenan is referred to as FC-NNLS. The CSSLS algorithm (Algorithm 3) is a part of the FC-NNLS algorithm. We applied the FC-NNLS algorithm to the penalized Euclidean function: Algorithm 2: Regularized FC-NNLS Algorithm Input : A R I J, Y = R I T, λ R + Output: X 0 such that X = argmin X AX Y F + λ X F 1 Initialization: M = 1,...,T }, N = 1,...,J}; 2 Precompute: B = [b i j ] = A T A+λI J and C = [c it ] = A T Y ; 3 X = B 1 C ; // unconstrained minimizer 1 if x P = [p ], where p = > 0, ; 4 0 otherwise // passive entries 5 F = t M : j p I} ; // set of columns to be optimized x x if p = 1, 0 otherwise ; 6 7 while F /0 do 8 P F = [p,f ] R J F, C F = [c,f ] R J F ; 9 [x,f ] =cssls(b,c F,P F ) ; // Solved with the CSSLS algorithm 10 H = t F : min j N x } < 0} ; // Columns with negative vars. 11 while H /0 do 12 s H, select the variables to move out of the passive set P; 13 P H = [p,h ] R J H, C H = [c,h ] R J H ; 14 [x,h ] =cssls(b,c H,P H ); 15 H = t F : min j N x } < 0} ; // Columns with negative vars W = [w ] = C F BX F, where X F = [x,f ] ; // negative gradient Z = t F : j w (1 P F ) = 0} ; // set of optimized columns F F\Z ; // set of columns to be optimized 1 if j = argmax p = j w (1 P F ), t F }, ; // updating p otherwise

7 Regularized NNLS Algorithms for Text Document Clustering 763 Algorithm 3: CSSLS Algorithm Input : B R J J, C = R J K, P R J K Output: X 1 M = 1,...,K}, N = 1,...,J}, P = [p 1,...,p K ] ; 2 Find the set of L unique columns in P: U = [u 1,...,u L ] = uniquep} ; d j = } 3 t M : p t = u j ; // columns with identical passive sets 4 for j = 1,...,L do 5 [X] u j,d j = ( [B] u j,u j ) 1 [C]u j,d j D(Y AX) = 1 2 Y AX 2 F + λ X 2 F, (13) where λ is a regularization parameter. In consequence, we have the Regularized FC-NNLS algorithm. The aim of using regularization for the discussed problems is rather to trigger the character of iterations than to stabilize ill-posed problems since the matrix A is not expected to be ill-conditioned. 3.3 Regularized NNLS-NMF Algorithm The Regularized NNLS-NMF algorithm is given by Algorithm 4. Motivated by Algorithm 4: Regularized NNLS-NMF Algorithm Input : Y R I T, J - lower rank, λ 0 - initial regularization parameter, τ - decay rate, λ - minimal value of regularization parameter, Output: Factors A and X 1 Initialize (randomly) A and X; 2 repeat 3 k k+ 1; 4 λ = λ + λ 0 exp τk} ; // Regularization parameter schedule 5 X fcnnls(a,y,λ) ; // Update for X d (X) j = t=1 T x2, X diag (d (X) j ) 1} } X, A Adiag d (X) 6 j ; 7 Ā fcnnls(x T,Y T,λ), A = Ā T ; // Update for A } d (A) j = I i=1 a2 i j, X diag d (A) j X, A Adiag (d (A) j ) 1} 8 ; 9 until Stop criterion is satisfied ; [26], we propose to gradually decrease the regularization parameter λ with alternating iterations for NMF, starting from a large value λ 0. Considering updates for X and A in terms of Singular Value Decomposition (SVD) of these matrices, it is obvious that if λ is large, only the right singular vectors that correspond to the largest singular values take part in the updating process. When the number of iterations is

8 764 R. Zdunek large (in practice about 100), λ λ, where we take λ = 10 12, and the singular vectors corresponding to the smallest singular values participate in the updates. 4 Experiments The experiments are carried out for two databases of text documents. The first database (D1) is created from the Reuters database by randomly selecting 500 documents from the following topics: acq, coffee, crude, eran, gold, interest, money-fx, ship, sugar, trade. The database D1 has 5960 distinct words; thus Y R and J = 10. The other database (D2) comes from the TopicPlanet document collection. We selected 178 documents classified into 6 topics: air-travel, broadband, cruises, domain-names, investments, technologies, which gives 8054 words. Thus Y R and J = 6. For clustering, we used the following algorithms: standard Lee-Seung NMF for the Euclidean distance [18], GD-CLS [22], Projected ALS [2, 7], Uni-orth NMF [10] separately for the matrix A and X, Bi-orth NMF [10], Convex-NMF [8], Newton NMF [7], k-means (from Matlab2008) with a cosine measure, LSI, and Regularized FC-NNLS (RFC-NNLS). All the algorithms are terminated after 20 iterations. For the RFC-NNLS algorithm, we set the following parameter: λ = 10 12, λ 0 = 10 8 and τ = 0.2. All the algorithms are intensively tested under the software developed by M. Jankowiak [14]. Table 1 Mean-accuracy, standard deviations, and averaged elapsed time (in seconds) over 10 MC runs of the tested NMF algorithms. Algorithm TopicPlanet Reuters Accuracy Std. Time [sec] Accuracy Std. Time [sec] Lee-Seung NMF GD-CLS Projected ALS Uni-orth (A) Uni-orth (X) Bi-orth Convex-NMF Newton-NMF k-means LSI RFC-NNLS Each tested algorithm (except for LSI) is run for 10 Monte Carlo (MC) trials with a random initialization. The algorithms are evaluated (Table 1) in terms of

9 Regularized NNLS Algorithms for Text Document Clustering 765 the averaged accuracy of clustering, standard deviations, and the averaged elapsed time over 10 trials. The accuracy is expressed with a ratio of correctly classified documents to the total number of documents. 5 Conclusions We proposed the Tikhonov regularized version of the FC-NNLS (RFC-NNLS) algorithm and efficiently applied it to text document clustering. The experiments demonstrate that the RFC-NNLS algorithm outperforms all the tested algorithms in terms of the accuracy and resistance to initialization. The computational complexity of the RFC-NNLS is comparable to fast algorithms such as the standard Lee-Seung NMF, ALS, Newton-NMF, and LSI. Acknowledgment. This work was supported by the habilitation grant N N ( ) from the Ministry of Science and Higher Education, Poland. References [1] Benthem, M.H.V., Keenan, M.R.: J. Chemometr. 18, (2004) [2] Berry, M., Browne, M., Langville, A., Pauca, P., Plemmons, R.: Comput. Stat. Data An. 52, (2007) [3] Bro, R., Jong, S.D.: J. Chemometr. 11, (1997) [4] Buciu, I., Pitas, I.: Application of non-negative and local nonnegative matrix factorization to facial expression recognition. In: Proc. Intl. Conf. Pattern Recognition (ICPR), pp (2004) [5] Cai, D., He, X., Wu, X., Bao, H., Han, J.: Locality preserving nonnegative matrix factorization. In: Proc. IJCAI 2009, pp (2009) [6] Cai, D., He, X., Wu, X., Han, J.: Nonnegative matrix factorization on manifold. In: Proc. 8th IEEE Intl. Conf. Data Mining (ICDM), pp (2008) [7] Cichocki, A., Zdunek, R., Phan, A.H., Amari, S.I.: Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multi-way Data Analysis and Blind Source Separation. Wiley and Sons, Chichester (2009) [8] Ding, C., Li, T., Jordan, M.I.: IEEE T. Pattern. Anal. 32, (2010) [9] Ding, C., Li, T., Peng, W.: Nonnegative matrix factorization and probabilistic latent semantic indexing: Equivalence, chi-square statistic, and a hybrid method. In: Proc. AAAI National Conf. Artificial Intelligence (AAAI 2006) (2006) [10] Ding, C., Li, T., Peng, W., Park, H.: Orthogonal nonnegative matrix tri-factorizations for clustering. In: Proc 12th ACM SIGKDD Intl. Conf. Knowledge Discovery and Data Mining, pp ACM Press, New York (2006) [11] Du, Q., Kopriva, I.: Neurocomputing 72, (2009) [12] Heiler, M., Schnoerr, C.: J. Mach. Learn. Res. 7, (2006) [13] Jain, A.K., Murty, M.N., Flynn, P.J.: ACM Comput. Surv. 31, (1999) [14] Jankowiak, M.: Application of nonnegative matrix factorization for text document classification. MSc thesis (supervised by Dr. R. Zdunek), Wroclaw University of Technology, Poland (2010) (in Polish) [15] Kim, H., Park, H.: Bioinformatics 23, (2007)

10 766 R. Zdunek [16] Kim, H., Park, H.: SIAM J. Matrix Anal. A 30, (2008) [17] Lawson, C.L., Hanson, R.J.: Solving Least Squares Problems. Prentice-Hall, Englewood Cliffs (1974) [18] Lee, D.D., Seung, H.S.: Nature 401, (1999) [19] Li, T., Ding, C.: The relationships among various nonnegative matrix factorization methods for clustering. In: Proc. 6th Intl. Conf. Data Mining (ICDM 2006), pp IEEE Computer Society, Washington DC, USA (2006) [20] O Grady, P., Pearlmutte, B.: Neurocomputing 72, (2008) [21] Sajda, P., Du, S., Brown, T.R., Stoyanova, R., Shungu, D.C., Mao, X., Parra, L.C.: IEEE T. Med. Imaging 23, (2004) [22] Shahnaz, F., Berry, M., Pauca, P., Plemmons, R.: Inform. Process. Manag. 42, (2006) [23] Sra, S., Dhillon, I.S.: Nonnegative matrix approximation: Algorithms and Applications. UTCS Technical Report TR-06-27, Austin, USA (2006), [24] Xu, W., Liu, X., Gong, Y.: Document clustering based on non-negative matrix factorization. In: SIGIR 2003: Proc 26th Annual Intl ACM SIGIR Conf. Research and Development in Informaion Retrieval, pp ACM Press, New York (2003) [25] Zdunek, R., Cichocki, A.: Comput. Intel. Neurosci. (939567) (2008) [26] Zdunek, R., Phan, A.H., Cichocki, A.: Aust. J. Intel. Inform. Process. Syst. 12, (2010)

Non-Negative Matrix Factorization with Quasi-Newton Optimization

Non-Negative Matrix Factorization with Quasi-Newton Optimization Non-Negative Matrix Factorization with Quasi-Newton Optimization Rafal ZDUNEK, Andrzej CICHOCKI Laboratory for Advanced Brain Signal Processing BSI, RIKEN, Wako-shi, JAPAN Abstract. Non-negative matrix

More information

MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION

MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION Ehsan Hosseini Asl 1, Jacek M. Zurada 1,2 1 Department of Electrical and Computer Engineering University of Louisville, Louisville,

More information

Non-Negative Matrix Factorization

Non-Negative Matrix Factorization Chapter 3 Non-Negative Matrix Factorization Part 2: Variations & applications Geometry of NMF 2 Geometry of NMF 1 NMF factors.9.8 Data points.7.6 Convex cone.5.4 Projections.3.2.1 1.5 1.5 1.5.5 1 3 Sparsity

More information

Orthogonal Nonnegative Matrix Factorization: Multiplicative Updates on Stiefel Manifolds

Orthogonal Nonnegative Matrix Factorization: Multiplicative Updates on Stiefel Manifolds Orthogonal Nonnegative Matrix Factorization: Multiplicative Updates on Stiefel Manifolds Jiho Yoo and Seungjin Choi Department of Computer Science Pohang University of Science and Technology San 31 Hyoja-dong,

More information

Note on Algorithm Differences Between Nonnegative Matrix Factorization And Probabilistic Latent Semantic Indexing

Note on Algorithm Differences Between Nonnegative Matrix Factorization And Probabilistic Latent Semantic Indexing Note on Algorithm Differences Between Nonnegative Matrix Factorization And Probabilistic Latent Semantic Indexing 1 Zhong-Yuan Zhang, 2 Chris Ding, 3 Jie Tang *1, Corresponding Author School of Statistics,

More information

Non-negative Tensor Factorization Based on Alternating Large-scale Non-negativity-constrained Least Squares

Non-negative Tensor Factorization Based on Alternating Large-scale Non-negativity-constrained Least Squares Non-negative Tensor Factorization Based on Alternating Large-scale Non-negativity-constrained Least Squares Hyunsoo Kim, Haesun Park College of Computing Georgia Institute of Technology 66 Ferst Drive,

More information

Regularized Alternating Least Squares Algorithms for Non-negative Matrix/Tensor Factorization

Regularized Alternating Least Squares Algorithms for Non-negative Matrix/Tensor Factorization Regularized Alternating Least Squares Algorithms for Non-negative Matrix/Tensor Factorization Andrzej CICHOCKI and Rafal ZDUNEK Laboratory for Advanced Brain Signal Processing, RIKEN BSI, Wako-shi, Saitama

More information

Dimension Reduction Using Nonnegative Matrix Tri-Factorization in Multi-label Classification

Dimension Reduction Using Nonnegative Matrix Tri-Factorization in Multi-label Classification 250 Int'l Conf. Par. and Dist. Proc. Tech. and Appl. PDPTA'15 Dimension Reduction Using Nonnegative Matrix Tri-Factorization in Multi-label Classification Keigo Kimura, Mineichi Kudo and Lu Sun Graduate

More information

Iterative Laplacian Score for Feature Selection

Iterative Laplacian Score for Feature Selection Iterative Laplacian Score for Feature Selection Linling Zhu, Linsong Miao, and Daoqiang Zhang College of Computer Science and echnology, Nanjing University of Aeronautics and Astronautics, Nanjing 2006,

More information

Automatic Rank Determination in Projective Nonnegative Matrix Factorization

Automatic Rank Determination in Projective Nonnegative Matrix Factorization Automatic Rank Determination in Projective Nonnegative Matrix Factorization Zhirong Yang, Zhanxing Zhu, and Erkki Oja Department of Information and Computer Science Aalto University School of Science and

More information

Simplicial Nonnegative Matrix Tri-Factorization: Fast Guaranteed Parallel Algorithm

Simplicial Nonnegative Matrix Tri-Factorization: Fast Guaranteed Parallel Algorithm Simplicial Nonnegative Matrix Tri-Factorization: Fast Guaranteed Parallel Algorithm Duy-Khuong Nguyen 13, Quoc Tran-Dinh 2, and Tu-Bao Ho 14 1 Japan Advanced Institute of Science and Technology, Japan

More information

arxiv: v2 [math.oc] 6 Oct 2011

arxiv: v2 [math.oc] 6 Oct 2011 Accelerated Multiplicative Updates and Hierarchical ALS Algorithms for Nonnegative Matrix Factorization Nicolas Gillis 1 and François Glineur 2 arxiv:1107.5194v2 [math.oc] 6 Oct 2011 Abstract Nonnegative

More information

Nonnegative Tensor Factorization with Smoothness Constraints

Nonnegative Tensor Factorization with Smoothness Constraints Nonnegative Tensor Factorization with Smoothness Constraints Rafal ZDUNEK 1 and Tomasz M. RUTKOWSKI 2 1 Institute of Telecommunications, Teleinformatics and Acoustics, Wroclaw University of Technology,

More information

Data Mining and Matrices

Data Mining and Matrices Data Mining and Matrices 6 Non-Negative Matrix Factorization Rainer Gemulla, Pauli Miettinen May 23, 23 Non-Negative Datasets Some datasets are intrinsically non-negative: Counters (e.g., no. occurrences

More information

Rank Selection in Low-rank Matrix Approximations: A Study of Cross-Validation for NMFs

Rank Selection in Low-rank Matrix Approximations: A Study of Cross-Validation for NMFs Rank Selection in Low-rank Matrix Approximations: A Study of Cross-Validation for NMFs Bhargav Kanagal Department of Computer Science University of Maryland College Park, MD 277 bhargav@cs.umd.edu Vikas

More information

Cholesky Decomposition Rectification for Non-negative Matrix Factorization

Cholesky Decomposition Rectification for Non-negative Matrix Factorization Cholesky Decomposition Rectification for Non-negative Matrix Factorization Tetsuya Yoshida Graduate School of Information Science and Technology, Hokkaido University N-14 W-9, Sapporo 060-0814, Japan yoshida@meme.hokudai.ac.jp

More information

B-Spline Smoothing of Feature Vectors in Nonnegative Matrix Factorization

B-Spline Smoothing of Feature Vectors in Nonnegative Matrix Factorization B-Spline Smoothing of Feature Vectors in Nonnegative Matrix Factorization Rafa l Zdunek 1, Andrzej Cichocki 2,3,4, and Tatsuya Yokota 2,5 1 Department of Electronics, Wroclaw University of Technology,

More information

NONNEGATIVE matrix factorization (NMF) is a

NONNEGATIVE matrix factorization (NMF) is a Algorithms for Orthogonal Nonnegative Matrix Factorization Seungjin Choi Abstract Nonnegative matrix factorization (NMF) is a widely-used method for multivariate analysis of nonnegative data, the goal

More information

Fast Nonnegative Tensor Factorization with an Active-Set-Like Method

Fast Nonnegative Tensor Factorization with an Active-Set-Like Method Fast Nonnegative Tensor Factorization with an Active-Set-Like Method Jingu Kim and Haesun Park Abstract We introduce an efficient algorithm for computing a low-rank nonnegative CANDECOMP/PARAFAC(NNCP)

More information

Non-negative matrix factorization and term structure of interest rates

Non-negative matrix factorization and term structure of interest rates Non-negative matrix factorization and term structure of interest rates Hellinton H. Takada and Julio M. Stern Citation: AIP Conference Proceedings 1641, 369 (2015); doi: 10.1063/1.4906000 View online:

More information

Non-Negative Matrix Factorization

Non-Negative Matrix Factorization Chapter 3 Non-Negative Matrix Factorization Part : Introduction & computation Motivating NMF Skillicorn chapter 8; Berry et al. (27) DMM, summer 27 2 Reminder A T U Σ V T T Σ, V U 2 Σ 2,2 V 2.8.6.6.3.6.5.3.6.3.6.4.3.6.4.3.3.4.5.3.5.8.3.8.3.3.5

More information

Nonnegative Matrix Factorization Clustering on Multiple Manifolds

Nonnegative Matrix Factorization Clustering on Multiple Manifolds Proceedings of the Twenty-Fourth AAAI Conference on Artificial Intelligence (AAAI-10) Nonnegative Matrix Factorization Clustering on Multiple Manifolds Bin Shen, Luo Si Department of Computer Science,

More information

AN ALTERNATING MINIMIZATION ALGORITHM FOR NON-NEGATIVE MATRIX APPROXIMATION

AN ALTERNATING MINIMIZATION ALGORITHM FOR NON-NEGATIVE MATRIX APPROXIMATION AN ALTERNATING MINIMIZATION ALGORITHM FOR NON-NEGATIVE MATRIX APPROXIMATION JOEL A. TROPP Abstract. Matrix approximation problems with non-negativity constraints arise during the analysis of high-dimensional

More information

Using Matrix Decompositions in Formal Concept Analysis

Using Matrix Decompositions in Formal Concept Analysis Using Matrix Decompositions in Formal Concept Analysis Vaclav Snasel 1, Petr Gajdos 1, Hussam M. Dahwa Abdulla 1, Martin Polovincak 1 1 Dept. of Computer Science, Faculty of Electrical Engineering and

More information

Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning

Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning LIU, LU, GU: GROUP SPARSE NMF FOR MULTI-MANIFOLD LEARNING 1 Group Sparse Non-negative Matrix Factorization for Multi-Manifold Learning Xiangyang Liu 1,2 liuxy@sjtu.edu.cn Hongtao Lu 1 htlu@sjtu.edu.cn

More information

ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering

ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering ONP-MF: An Orthogonal Nonnegative Matrix Factorization Algorithm with Application to Clustering Filippo Pompili 1, Nicolas Gillis 2, P.-A. Absil 2,andFrançois Glineur 2,3 1- University of Perugia, Department

More information

Multi-Task Clustering using Constrained Symmetric Non-Negative Matrix Factorization

Multi-Task Clustering using Constrained Symmetric Non-Negative Matrix Factorization Multi-Task Clustering using Constrained Symmetric Non-Negative Matrix Factorization Samir Al-Stouhi Chandan K. Reddy Abstract Researchers have attempted to improve the quality of clustering solutions through

More information

Novel Alternating Least Squares Algorithm for Nonnegative Matrix and Tensor Factorizations

Novel Alternating Least Squares Algorithm for Nonnegative Matrix and Tensor Factorizations Novel Alternating Least Squares Algorithm for Nonnegative Matrix and Tensor Factorizations Anh Huy Phan 1, Andrzej Cichocki 1,, Rafal Zdunek 1,2,andThanhVuDinh 3 1 Lab for Advanced Brain Signal Processing,

More information

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors Nico Vervliet Joint work with Lieven De Lathauwer SIAM AN17, July 13, 2017 2 Classification of hazardous

More information

1 Non-negative Matrix Factorization (NMF)

1 Non-negative Matrix Factorization (NMF) 2018-06-21 1 Non-negative Matrix Factorization NMF) In the last lecture, we considered low rank approximations to data matrices. We started with the optimal rank k approximation to A R m n via the SVD,

More information

Clustering. SVD and NMF

Clustering. SVD and NMF Clustering with the SVD and NMF Amy Langville Mathematics Department College of Charleston Dagstuhl 2/14/2007 Outline Fielder Method Extended Fielder Method and SVD Clustering with SVD vs. NMF Demos with

More information

CVPR A New Tensor Algebra - Tutorial. July 26, 2017

CVPR A New Tensor Algebra - Tutorial. July 26, 2017 CVPR 2017 A New Tensor Algebra - Tutorial Lior Horesh lhoresh@us.ibm.com Misha Kilmer misha.kilmer@tufts.edu July 26, 2017 Outline Motivation Background and notation New t-product and associated algebraic

More information

Non-negative Matrix Factorization: Algorithms, Extensions and Applications

Non-negative Matrix Factorization: Algorithms, Extensions and Applications Non-negative Matrix Factorization: Algorithms, Extensions and Applications Emmanouil Benetos www.soi.city.ac.uk/ sbbj660/ March 2013 Emmanouil Benetos Non-negative Matrix Factorization March 2013 1 / 25

More information

On the equivalence between Non-negative Matrix Factorization and Probabilistic Latent Semantic Indexing

On the equivalence between Non-negative Matrix Factorization and Probabilistic Latent Semantic Indexing Computational Statistics and Data Analysis 52 (2008) 3913 3927 www.elsevier.com/locate/csda On the equivalence between Non-negative Matrix Factorization and Probabilistic Latent Semantic Indexing Chris

More information

Block Coordinate Descent for Regularized Multi-convex Optimization

Block Coordinate Descent for Regularized Multi-convex Optimization Block Coordinate Descent for Regularized Multi-convex Optimization Yangyang Xu and Wotao Yin CAAM Department, Rice University February 15, 2013 Multi-convex optimization Model definition Applications Outline

More information

Multi-Task Co-clustering via Nonnegative Matrix Factorization

Multi-Task Co-clustering via Nonnegative Matrix Factorization Multi-Task Co-clustering via Nonnegative Matrix Factorization Saining Xie, Hongtao Lu and Yangcheng He Shanghai Jiao Tong University xiesaining@gmail.com, lu-ht@cs.sjtu.edu.cn, yche.sjtu@gmail.com Abstract

More information

Non-negative matrix factorization with fixed row and column sums

Non-negative matrix factorization with fixed row and column sums Available online at www.sciencedirect.com Linear Algebra and its Applications 9 (8) 5 www.elsevier.com/locate/laa Non-negative matrix factorization with fixed row and column sums Ngoc-Diep Ho, Paul Van

More information

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin

Introduction to Machine Learning. PCA and Spectral Clustering. Introduction to Machine Learning, Slides: Eran Halperin 1 Introduction to Machine Learning PCA and Spectral Clustering Introduction to Machine Learning, 2013-14 Slides: Eran Halperin Singular Value Decomposition (SVD) The singular value decomposition (SVD)

More information

Accelerated Anti-lopsided Algorithm for Non-negative Least Squares

Accelerated Anti-lopsided Algorithm for Non-negative Least Squares Noname manuscript No. (will be inserted by the editor) Accelerated Anti-lopsided Algorithm for Non-negative Least Squares Duy Khuong Nguyen Tu Bao Ho Received: date / Accepted: date Abstract Non-negative

More information

Preserving Privacy in Data Mining using Data Distortion Approach

Preserving Privacy in Data Mining using Data Distortion Approach Preserving Privacy in Data Mining using Data Distortion Approach Mrs. Prachi Karandikar #, Prof. Sachin Deshpande * # M.E. Comp,VIT, Wadala, University of Mumbai * VIT Wadala,University of Mumbai 1. prachiv21@yahoo.co.in

More information

Dimension reduction based on centroids and least squares for efficient processing of text data

Dimension reduction based on centroids and least squares for efficient processing of text data Dimension reduction based on centroids and least squares for efficient processing of text data M. Jeon,H.Park, and J.B. Rosen Abstract Dimension reduction in today s vector space based information retrieval

More information

Nonnegative Matrix Factorization and Probabilistic Latent Semantic Indexing: Equivalence, Chi-square Statistic, and a Hybrid Method

Nonnegative Matrix Factorization and Probabilistic Latent Semantic Indexing: Equivalence, Chi-square Statistic, and a Hybrid Method Nonnegative Matrix Factorization and Probabilistic Latent Semantic Indexing: Equivalence, hi-square Statistic, and a Hybrid Method hris Ding a, ao Li b and Wei Peng b a Lawrence Berkeley National Laboratory,

More information

Research Article Relationship Matrix Nonnegative Decomposition for Clustering

Research Article Relationship Matrix Nonnegative Decomposition for Clustering Mathematical Problems in Engineering Volume 2011, Article ID 864540, 15 pages doi:10.1155/2011/864540 Research Article Relationship Matrix Nonnegative Decomposition for Clustering Ji-Yuan Pan and Jiang-She

More information

Nonnegative Matrix Factorization. Data Mining

Nonnegative Matrix Factorization. Data Mining The Nonnegative Matrix Factorization in Data Mining Amy Langville langvillea@cofc.edu Mathematics Department College of Charleston Charleston, SC Yahoo! Research 10/18/2005 Outline Part 1: Historical Developments

More information

c Springer, Reprinted with permission.

c Springer, Reprinted with permission. Zhijian Yuan and Erkki Oja. A FastICA Algorithm for Non-negative Independent Component Analysis. In Puntonet, Carlos G.; Prieto, Alberto (Eds.), Proceedings of the Fifth International Symposium on Independent

More information

L 2,1 Norm and its Applications

L 2,1 Norm and its Applications L 2, Norm and its Applications Yale Chang Introduction According to the structure of the constraints, the sparsity can be obtained from three types of regularizers for different purposes.. Flat Sparsity.

More information

Research Article Comparative Features Extraction Techniques for Electrocardiogram Images Regression

Research Article Comparative Features Extraction Techniques for Electrocardiogram Images Regression Research Journal of Applied Sciences, Engineering and Technology (): 6, 7 DOI:.96/rjaset..6 ISSN: -79; e-issn: -767 7 Maxwell Scientific Organization Corp. Submitted: September 8, 6 Accepted: November,

More information

Nonnegative Matrix Factorization with Toeplitz Penalty

Nonnegative Matrix Factorization with Toeplitz Penalty Journal of Informatics and Mathematical Sciences Vol. 10, Nos. 1 & 2, pp. 201 215, 2018 ISSN 0975-5748 (online); 0974-875X (print) Published by RGN Publications http://dx.doi.org/10.26713/jims.v10i1-2.851

More information

Fast Nonnegative Matrix Factorization with Rank-one ADMM

Fast Nonnegative Matrix Factorization with Rank-one ADMM Fast Nonnegative Matrix Factorization with Rank-one Dongjin Song, David A. Meyer, Martin Renqiang Min, Department of ECE, UCSD, La Jolla, CA, 9093-0409 dosong@ucsd.edu Department of Mathematics, UCSD,

More information

Supervised Nonnegative Matrix Factorization to Predict ICU Mortality Risk

Supervised Nonnegative Matrix Factorization to Predict ICU Mortality Risk Supervised Nonnegative Matrix Factorization to Predict ICU Mortality Risk Guoqing Chao guoqingchao10@gmail.com Chengsheng Mao chengsheng.mao@northwestern.edu Fei Wang Weill Cornell Medicine Cornell University

More information

Single-channel source separation using non-negative matrix factorization

Single-channel source separation using non-negative matrix factorization Single-channel source separation using non-negative matrix factorization Mikkel N. Schmidt Technical University of Denmark mns@imm.dtu.dk www.mikkelschmidt.dk DTU Informatics Department of Informatics

More information

AN ENHANCED INITIALIZATION METHOD FOR NON-NEGATIVE MATRIX FACTORIZATION. Liyun Gong 1, Asoke K. Nandi 2,3 L69 3BX, UK; 3PH, UK;

AN ENHANCED INITIALIZATION METHOD FOR NON-NEGATIVE MATRIX FACTORIZATION. Liyun Gong 1, Asoke K. Nandi 2,3 L69 3BX, UK; 3PH, UK; 213 IEEE INERNAIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEP. 22 25, 213, SOUHAMPON, UK AN ENHANCED INIIALIZAION MEHOD FOR NON-NEGAIVE MARIX FACORIZAION Liyun Gong 1, Asoke K. Nandi 2,3

More information

A Purely Geometric Approach to Non-Negative Matrix Factorization

A Purely Geometric Approach to Non-Negative Matrix Factorization A Purely Geometric Approach to Non-Negative Matrix Factorization Christian Bauckhage B-IT, University of Bonn, Bonn, Germany Fraunhofer IAIS, Sankt Augustin, Germany http://mmprec.iais.fraunhofer.de/bauckhage.html

More information

Non-Negative Factorization for Clustering of Microarray Data

Non-Negative Factorization for Clustering of Microarray Data INT J COMPUT COMMUN, ISSN 1841-9836 9(1):16-23, February, 2014. Non-Negative Factorization for Clustering of Microarray Data L. Morgos Lucian Morgos Dept. of Electronics and Telecommunications Faculty

More information

Matrix Factorization & Latent Semantic Analysis Review. Yize Li, Lanbo Zhang

Matrix Factorization & Latent Semantic Analysis Review. Yize Li, Lanbo Zhang Matrix Factorization & Latent Semantic Analysis Review Yize Li, Lanbo Zhang Overview SVD in Latent Semantic Indexing Non-negative Matrix Factorization Probabilistic Latent Semantic Indexing Vector Space

More information

An interior-point gradient method for large-scale totally nonnegative least squares problems

An interior-point gradient method for large-scale totally nonnegative least squares problems An interior-point gradient method for large-scale totally nonnegative least squares problems Michael Merritt and Yin Zhang Technical Report TR04-08 Department of Computational and Applied Mathematics Rice

More information

Matrix Factorization with Applications to Clustering Problems: Formulation, Algorithms and Performance

Matrix Factorization with Applications to Clustering Problems: Formulation, Algorithms and Performance Date: Mar. 3rd, 2017 Matrix Factorization with Applications to Clustering Problems: Formulation, Algorithms and Performance Presenter: Songtao Lu Department of Electrical and Computer Engineering Iowa

More information

One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017

One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 One Picture and a Thousand Words Using Matrix Approximtions October 2017 Oak Ridge National Lab Dianne P. O Leary c 2017 1 One Picture and a Thousand Words Using Matrix Approximations Dianne P. O Leary

More information

c 2010 by Florian Markus Lorenz. All rights reserved.

c 2010 by Florian Markus Lorenz. All rights reserved. c 2010 by Florian Markus Lorenz. All rights reserved. SEQUENTIALLY-FIT ALTERNATING LEAST SQUARES ALGORITHMS IN NONNEGATIVE MATRIX FACTORIZATION BY FLORIAN MARKUS LORENZ THESIS Submitted in partial fulfillment

More information

Chapter 2. Optimization. Gradients, convexity, and ALS

Chapter 2. Optimization. Gradients, convexity, and ALS Chapter 2 Optimization Gradients, convexity, and ALS Contents Background Gradient descent Stochastic gradient descent Newton s method Alternating least squares KKT conditions 2 Motivation We can solve

More information

EUSIPCO

EUSIPCO EUSIPCO 2013 1569741067 CLUSERING BY NON-NEGAIVE MARIX FACORIZAION WIH INDEPENDEN PRINCIPAL COMPONEN INIIALIZAION Liyun Gong 1, Asoke K. Nandi 2,3 1 Department of Electrical Engineering and Electronics,

More information

Algorithms for Nonnegative Tensor Factorization

Algorithms for Nonnegative Tensor Factorization Algorithms for Nonnegative Tensor Factorization Markus Flatz Technical Report 2013-05 November 2013 Department of Computer Sciences Jakob-Haringer-Straße 2 5020 Salzburg Austria www.cosy.sbg.ac.at Technical

More information

Nonnegative Tensor Factorization with Smoothness Constraints

Nonnegative Tensor Factorization with Smoothness Constraints Nonnegative Tensor Factorization with Smoothness Constraints Rafal Zdunek 1 and Tomasz M. Rutkowski 2 1 Institute of Telecommunications, Teleinformatics and Acoustics, Wroclaw University of Technology,

More information

2 GU, ZHOU: NEIGHBORHOOD PRESERVING NONNEGATIVE MATRIX FACTORIZATION graph regularized NMF (GNMF), which assumes that the nearby data points are likel

2 GU, ZHOU: NEIGHBORHOOD PRESERVING NONNEGATIVE MATRIX FACTORIZATION graph regularized NMF (GNMF), which assumes that the nearby data points are likel GU, ZHOU: NEIGHBORHOOD PRESERVING NONNEGATIVE MATRIX FACTORIZATION 1 Neighborhood Preserving Nonnegative Matrix Factorization Quanquan Gu gqq03@mails.tsinghua.edu.cn Jie Zhou jzhou@tsinghua.edu.cn State

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

Nonnegative Matrix Factorization with Orthogonality Constraints

Nonnegative Matrix Factorization with Orthogonality Constraints Nonnegative Matrix Factorization with Orthogonality Constraints Jiho Yoo and Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology (POSTECH) Pohang, Republic

More information

PROJECTIVE NON-NEGATIVE MATRIX FACTORIZATION WITH APPLICATIONS TO FACIAL IMAGE PROCESSING

PROJECTIVE NON-NEGATIVE MATRIX FACTORIZATION WITH APPLICATIONS TO FACIAL IMAGE PROCESSING st Reading October 0, 200 8:4 WSPC/-IJPRAI SPI-J068 008 International Journal of Pattern Recognition and Artificial Intelligence Vol. 2, No. 8 (200) 0 c World Scientific Publishing Company PROJECTIVE NON-NEGATIVE

More information

Probabilistic Latent Variable Models as Non-Negative Factorizations

Probabilistic Latent Variable Models as Non-Negative Factorizations 1 Probabilistic Latent Variable Models as Non-Negative Factorizations Madhusudana Shashanka, Bhiksha Raj, Paris Smaragdis Abstract In this paper, we present a family of probabilistic latent variable models

More information

Recursive Generalized Eigendecomposition for Independent Component Analysis

Recursive Generalized Eigendecomposition for Independent Component Analysis Recursive Generalized Eigendecomposition for Independent Component Analysis Umut Ozertem 1, Deniz Erdogmus 1,, ian Lan 1 CSEE Department, OGI, Oregon Health & Science University, Portland, OR, USA. {ozertemu,deniz}@csee.ogi.edu

More information

On Spectral Basis Selection for Single Channel Polyphonic Music Separation

On Spectral Basis Selection for Single Channel Polyphonic Music Separation On Spectral Basis Selection for Single Channel Polyphonic Music Separation Minje Kim and Seungjin Choi Department of Computer Science Pohang University of Science and Technology San 31 Hyoja-dong, Nam-gu

More information

Slice Oriented Tensor Decomposition of EEG Data for Feature Extraction in Space, Frequency and Time Domains

Slice Oriented Tensor Decomposition of EEG Data for Feature Extraction in Space, Frequency and Time Domains Slice Oriented Tensor Decomposition of EEG Data for Feature Extraction in Space, and Domains Qibin Zhao, Cesar F. Caiafa, Andrzej Cichocki, and Liqing Zhang 2 Laboratory for Advanced Brain Signal Processing,

More information

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization

Proximal Newton Method. Zico Kolter (notes by Ryan Tibshirani) Convex Optimization Proximal Newton Method Zico Kolter (notes by Ryan Tibshirani) Convex Optimization 10-725 Consider the problem Last time: quasi-newton methods min x f(x) with f convex, twice differentiable, dom(f) = R

More information

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems

Numerical optimization. Numerical optimization. Longest Shortest where Maximal Minimal. Fastest. Largest. Optimization problems 1 Numerical optimization Alexander & Michael Bronstein, 2006-2009 Michael Bronstein, 2010 tosca.cs.technion.ac.il/book Numerical optimization 048921 Advanced topics in vision Processing and Analysis of

More information

Local Feature Extraction Models from Incomplete Data in Face Recognition Based on Nonnegative Matrix Factorization

Local Feature Extraction Models from Incomplete Data in Face Recognition Based on Nonnegative Matrix Factorization American Journal of Software Engineering and Applications 2015; 4(3): 50-55 Published online May 12, 2015 (http://www.sciencepublishinggroup.com/j/ajsea) doi: 10.11648/j.ajsea.20150403.12 ISSN: 2327-2473

More information

USING SINGULAR VALUE DECOMPOSITION (SVD) AS A SOLUTION FOR SEARCH RESULT CLUSTERING

USING SINGULAR VALUE DECOMPOSITION (SVD) AS A SOLUTION FOR SEARCH RESULT CLUSTERING POZNAN UNIVE RSIY OF E CHNOLOGY ACADE MIC JOURNALS No. 80 Electrical Engineering 2014 Hussam D. ABDULLA* Abdella S. ABDELRAHMAN* Vaclav SNASEL* USING SINGULAR VALUE DECOMPOSIION (SVD) AS A SOLUION FOR

More information

Fast Coordinate Descent methods for Non-Negative Matrix Factorization

Fast Coordinate Descent methods for Non-Negative Matrix Factorization Fast Coordinate Descent methods for Non-Negative Matrix Factorization Inderjit S. Dhillon University of Texas at Austin SIAM Conference on Applied Linear Algebra Valencia, Spain June 19, 2012 Joint work

More information

Machine Learning (BSMC-GA 4439) Wenke Liu

Machine Learning (BSMC-GA 4439) Wenke Liu Machine Learning (BSMC-GA 4439) Wenke Liu 02-01-2018 Biomedical data are usually high-dimensional Number of samples (n) is relatively small whereas number of features (p) can be large Sometimes p>>n Problems

More information

NMF WITH SPECTRAL AND TEMPORAL CONTINUITY CRITERIA FOR MONAURAL SOUND SOURCE SEPARATION. Julian M. Becker, Christian Sohn and Christian Rohlfing

NMF WITH SPECTRAL AND TEMPORAL CONTINUITY CRITERIA FOR MONAURAL SOUND SOURCE SEPARATION. Julian M. Becker, Christian Sohn and Christian Rohlfing NMF WITH SPECTRAL AND TEMPORAL CONTINUITY CRITERIA FOR MONAURAL SOUND SOURCE SEPARATION Julian M. ecker, Christian Sohn Christian Rohlfing Institut für Nachrichtentechnik RWTH Aachen University D-52056

More information

On Tolerant Fuzzy c-means Clustering with L 1 -Regularization

On Tolerant Fuzzy c-means Clustering with L 1 -Regularization On Tolerant Fuzzy c-means Clustering with L 1 -Regularization HAMASUNA Yuhiro 1 ENDO Yasunori 2 MIYAMOTO Sadaa 2 1. Doctoral Program in Risk Engineering, University of Tsukuba 1-1-1 Tennodai, Tsukuba,

More information

Novel Multi-layer Non-negative Tensor Factorization with Sparsity Constraints

Novel Multi-layer Non-negative Tensor Factorization with Sparsity Constraints Novel Multi-layer Non-negative Tensor Factorization with Sparsity Constraints Andrzej CICHOCKI 1, Rafal ZDUNEK 1, Seungjin CHOI, Robert PLEMMONS 2, and Shun-ichi AMARI 1 1 RIKEN Brain Science Institute,

More information

Numerical optimization

Numerical optimization Numerical optimization Lecture 4 Alexander & Michael Bronstein tosca.cs.technion.ac.il/book Numerical geometry of non-rigid shapes Stanford University, Winter 2009 2 Longest Slowest Shortest Minimal Maximal

More information

Machine learning for pervasive systems Classification in high-dimensional spaces

Machine learning for pervasive systems Classification in high-dimensional spaces Machine learning for pervasive systems Classification in high-dimensional spaces Department of Communications and Networking Aalto University, School of Electrical Engineering stephan.sigg@aalto.fi Version

More information

CS47300: Web Information Search and Management

CS47300: Web Information Search and Management CS47300: Web Information Search and Management Prof. Chris Clifton 6 September 2017 Material adapted from course created by Dr. Luo Si, now leading Alibaba research group 1 Vector Space Model Disadvantages:

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning Christoph Lampert Spring Semester 2015/2016 // Lecture 12 1 / 36 Unsupervised Learning Dimensionality Reduction 2 / 36 Dimensionality Reduction Given: data X = {x 1,..., x

More information

Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY

Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY Matrix Decomposition in Privacy-Preserving Data Mining JUN ZHANG DEPARTMENT OF COMPUTER SCIENCE UNIVERSITY OF KENTUCKY OUTLINE Why We Need Matrix Decomposition SVD (Singular Value Decomposition) NMF (Nonnegative

More information

Lecture Notes 10: Matrix Factorization

Lecture Notes 10: Matrix Factorization Optimization-based data analysis Fall 207 Lecture Notes 0: Matrix Factorization Low-rank models. Rank- model Consider the problem of modeling a quantity y[i, j] that depends on two indices i and j. To

More information

Probabilistic & Unsupervised Learning

Probabilistic & Unsupervised Learning Probabilistic & Unsupervised Learning Week 2: Latent Variable Models Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc ML/CSML, Dept Computer Science University College

More information

Examination of Initialization Techniques for Nonnegative Matrix Factorization

Examination of Initialization Techniques for Nonnegative Matrix Factorization Georgia State University ScholarWorks @ Georgia State University Mathematics Theses Department of Mathematics and Statistics 11-21-2008 Examination of Initialization Techniques for Nonnegative Matrix Factorization

More information

Matrix Derivatives and Descent Optimization Methods

Matrix Derivatives and Descent Optimization Methods Matrix Derivatives and Descent Optimization Methods 1 Qiang Ning Department of Electrical and Computer Engineering Beckman Institute for Advanced Science and Techonology University of Illinois at Urbana-Champaign

More information

Pose estimation from point and line correspondences

Pose estimation from point and line correspondences Pose estimation from point and line correspondences Giorgio Panin October 17, 008 1 Problem formulation Estimate (in a LSE sense) the pose of an object from N correspondences between known object points

More information

component risk analysis

component risk analysis 273: Urban Systems Modeling Lec. 3 component risk analysis instructor: Matteo Pozzi 273: Urban Systems Modeling Lec. 3 component reliability outline risk analysis for components uncertain demand and uncertain

More information

Data Mining and Matrices

Data Mining and Matrices Data Mining and Matrices 05 Semi-Discrete Decomposition Rainer Gemulla, Pauli Miettinen May 16, 2013 Outline 1 Hunting the Bump 2 Semi-Discrete Decomposition 3 The Algorithm 4 Applications SDD alone SVD

More information

N-mode Analysis (Tensor Framework) Behrouz Saghafi

N-mode Analysis (Tensor Framework) Behrouz Saghafi N-mode Analysis (Tensor Framework) Behrouz Saghafi N-mode Analysis (Tensor Framework) Drawback of 1-mode analysis (e.g. PCA): Captures the variance among just a single factor Our training set contains

More information

Convex and Semi-Nonnegative Matrix Factorizations

Convex and Semi-Nonnegative Matrix Factorizations CONVEX AND SEMI-NONNEGATIVE MATRIX FACTORIZATIONS: DING, LI AND JORDAN 1 Convex and Semi-Nonnegative Matrix Factorizations Chris Ding, Tao Li, and Michael I. Jordan Chris Ding is with the Department of

More information

arxiv: v4 [math.na] 10 Nov 2014

arxiv: v4 [math.na] 10 Nov 2014 NEWTON-BASED OPTIMIZATION FOR KULLBACK-LEIBLER NONNEGATIVE TENSOR FACTORIZATIONS SAMANTHA HANSEN, TODD PLANTENGA, TAMARA G. KOLDA arxiv:134.4964v4 [math.na] 1 Nov 214 Abstract. Tensor factorizations with

More information

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44

Convex Optimization. Newton s method. ENSAE: Optimisation 1/44 Convex Optimization Newton s method ENSAE: Optimisation 1/44 Unconstrained minimization minimize f(x) f convex, twice continuously differentiable (hence dom f open) we assume optimal value p = inf x f(x)

More information

Preconditioning. Noisy, Ill-Conditioned Linear Systems

Preconditioning. Noisy, Ill-Conditioned Linear Systems Preconditioning Noisy, Ill-Conditioned Linear Systems James G. Nagy Emory University Atlanta, GA Outline 1. The Basic Problem 2. Regularization / Iterative Methods 3. Preconditioning 4. Example: Image

More information

SVD based initialization: A head start for nonnegative matrix factorization

SVD based initialization: A head start for nonnegative matrix factorization SVD based initialization: A head start for nonnegative matrix factorization C. Boutsidis a, E. Gallopoulos b a Computer Science Department, Rensselaer Polytechnic Institute, Troy, NY 12180, USA b Computer

More information

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan

Clustering. CSL465/603 - Fall 2016 Narayanan C Krishnan Clustering CSL465/603 - Fall 2016 Narayanan C Krishnan ckn@iitrpr.ac.in Supervised vs Unsupervised Learning Supervised learning Given x ", y " "%& ', learn a function f: X Y Categorical output classification

More information

Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy

Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy Caroline Chaux Joint work with X. Vu, N. Thirion-Moreau and S. Maire (LSIS, Toulon) Aix-Marseille

More information