Sparseness Constraints on Nonnegative Tensor Decomposition

Similar documents
Non-Negative Tensor Factorisation for Sound Source Separation

Recovering Tensor Data from Incomplete Measurement via Compressive Sampling

Non-negative Matrix Factor Deconvolution; Extraction of Multiple Sound Sources from Monophonic Inputs

Nonnegative Matrix Factor 2-D Deconvolution for Blind Single Channel Source Separation

On Spectral Basis Selection for Single Channel Polyphonic Music Separation

Analysis of polyphonic audio using source-filter model and non-negative matrix factorization

CONVOLUTIVE NON-NEGATIVE MATRIX FACTORISATION WITH SPARSENESS CONSTRAINT

MATRIX COMPLETION AND TENSOR RANK

Fast Nonnegative Tensor Factorization with an Active-Set-Like Method

NMF WITH SPECTRAL AND TEMPORAL CONTINUITY CRITERIA FOR MONAURAL SOUND SOURCE SEPARATION. Julian M. Becker, Christian Sohn and Christian Rohlfing

The multiple-vector tensor-vector product

Introduction to Tensors. 8 May 2014

Single Channel Music Sound Separation Based on Spectrogram Decomposition and Note Classification

SINGLE CHANNEL SPEECH MUSIC SEPARATION USING NONNEGATIVE MATRIX FACTORIZATION AND SPECTRAL MASKS. Emad M. Grais and Hakan Erdogan

/16/$ IEEE 1728

A randomized block sampling approach to the canonical polyadic decomposition of large-scale tensors

Computing nonnegative tensor factorizations

Non-Negative Matrix Factorization And Its Application to Audio. Tuomas Virtanen Tampere University of Technology

c Springer, Reprinted with permission.

Third-Order Tensor Decompositions and Their Application in Quantum Chemistry

ENGG5781 Matrix Analysis and Computations Lecture 10: Non-Negative Matrix Factorization and Tensor Decomposition

MULTI-RESOLUTION SIGNAL DECOMPOSITION WITH TIME-DOMAIN SPECTROGRAM FACTORIZATION. Hirokazu Kameoka

LOW MULTILINEAR RANK TENSOR APPROXIMATION VIA SEMIDEFINITE PROGRAMMING

Fast Nonnegative Matrix Factorization with Rank-one ADMM

Non-negative Matrix Factorization: Algorithms, Extensions and Applications

A FLEXIBLE MODELING FRAMEWORK FOR COUPLED MATRIX AND TENSOR FACTORIZATIONS

NON-NEGATIVE SPARSE CODING

Fitting a Tensor Decomposition is a Nonlinear Optimization Problem

Tensor Decompositions and Applications

LOW RANK TENSOR DECONVOLUTION

Fundamentals of Multilinear Subspace Learning

Tensor-Based Dictionary Learning for Multidimensional Sparse Recovery. Florian Römer and Giovanni Del Galdo

Non-negative Tensor Factorization Based on Alternating Large-scale Non-negativity-constrained Least Squares

Nonnegative Matrix Factorization with Markov-Chained Bases for Modeling Time-Varying Patterns in Music Spectrograms

ORTHOGONALITY-REGULARIZED MASKED NMF FOR LEARNING ON WEAKLY LABELED AUDIO DATA. Iwona Sobieraj, Lucas Rencker, Mark D. Plumbley

Computational Linear Algebra

The Canonical Tensor Decomposition and Its Applications to Social Network Analysis

arxiv: v4 [math.na] 10 Nov 2014

A BLIND SPARSE APPROACH FOR ESTIMATING CONSTRAINT MATRICES IN PARALIND DATA MODELS

Discovering Convolutive Speech Phones using Sparseness and Non-Negativity Constraints

Author manuscript, published in "7th International Symposium on Computer Music Modeling and Retrieval (CMMR 2010), Málaga : Spain (2010)"

Preserving Privacy in Data Mining using Data Distortion Approach

t Tao Group Limited, Reading, RG6 IAZ, U.K. wenwu.wang(ieee.org t Samsung Electronics Research Institute, Staines, TW1 8 4QE, U.K.

Factor Analysis (FA) Non-negative Matrix Factorization (NMF) CSE Artificial Intelligence Grad Project Dr. Debasis Mitra

IN real-world audio signals, several sound sources are usually

Non-Negative Tensor Factorization with Applications to Statistics and Computer Vision

On the convergence of higher-order orthogonality iteration and its extension

Robust Speaker Modeling Based on Constrained Nonnegative Tensor Factorization

CVPR A New Tensor Algebra - Tutorial. July 26, 2017

NONNEGATIVE matrix factorization (NMF) is a

Scalable Tensor Factorizations with Incomplete Data

Convolutive Non-Negative Matrix Factorization for CQT Transform using Itakura-Saito Divergence

of Orthogonal Matching Pursuit

BOOLEAN TENSOR FACTORIZATIONS. Pauli Miettinen 14 December 2011

Constrained Nonnegative Matrix Factorization with Applications to Music Transcription

c 2008 Society for Industrial and Applied Mathematics

Multiplicative and Additive Perturbation Effects on the Recovery of Sparse Signals on the Sphere using Compressed Sensing

Noisy Signal Recovery via Iterative Reweighted L1-Minimization

Non-Negative Matrix Factorization with Quasi-Newton Optimization

Faloutsos, Tong ICDE, 2009

MULTIPLICATIVE ALGORITHM FOR CORRENTROPY-BASED NONNEGATIVE MATRIX FACTORIZATION

New algorithms for tensor decomposition based on a reduced functional

Nonnegative Tensor Factorization with Smoothness Constraints

SOUND SOURCE SEPARATION BASED ON NON-NEGATIVE TENSOR FACTORIZATION INCORPORATING SPATIAL CUE AS PRIOR KNOWLEDGE

A new method on deterministic construction of the measurement matrix in compressed sensing

An Effective Tensor Completion Method Based on Multi-linear Tensor Ring Decomposition

Two- and three-way component models for LR fuzzy data in a possibilistic framework

CP DECOMPOSITION AND ITS APPLICATION IN NOISE REDUCTION AND MULTIPLE SOURCES IDENTIFICATION

Single Channel Signal Separation Using MAP-based Subspace Decomposition

A Simpler Approach to Low-Rank Tensor Canonical Polyadic Decomposition

Non-negative matrix factorization and term structure of interest rates


Optimization for Compressed Sensing

Convention Paper Presented at the 133rd Convention 2012 October San Francisco, USA

Nonnegative Tensor Factorization using a proximal algorithm: application to 3D fluorescence spectroscopy

Machine Learning for Signal Processing Non-negative Matrix Factorization

GREEDY SIGNAL RECOVERY REVIEW

On Kruskal s uniqueness condition for the Candecomp/Parafac decomposition

Compressive Sensing of Sparse Tensor

Automated Unmixing of Comprehensive Two-Dimensional Chemical Separations with Mass Spectrometry. 1 Introduction. 2 System modeling

c 2008 Society for Industrial and Applied Mathematics

Strengthened Sobolev inequalities for a random subspace of functions

Generalized Orthogonal Matching Pursuit- A Review and Some

Robust PCA. CS5240 Theoretical Foundations in Multimedia. Leow Wee Kheng

GROUP-SPARSE SUBSPACE CLUSTERING WITH MISSING DATA

ADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA. Mark Plumbley

Minimizing the Difference of L 1 and L 2 Norms with Applications

REVIEW OF SINGLE CHANNEL SOURCE SEPARATION TECHNIQUES

Lecture 18 Nov 3rd, 2015

New Coherence and RIP Analysis for Weak. Orthogonal Matching Pursuit

Sensing systems limited by constraints: physical size, time, cost, energy

STRUCTURE-AWARE DICTIONARY LEARNING WITH HARMONIC ATOMS

Two-View Segmentation of Dynamic Scenes from the Multibody Fundamental Matrix

Near Optimal Signal Recovery from Random Projections

COMPRESSED SENSING IN PYTHON

arxiv: v1 [math.na] 26 Nov 2009

ACCOUNTING FOR PHASE CANCELLATIONS IN NON-NEGATIVE MATRIX FACTORIZATION USING WEIGHTED DISTANCES. Sebastian Ewert Mark D. Plumbley Mark Sandler

ESTIMATING TRAFFIC NOISE LEVELS USING ACOUSTIC MONITORING: A PRELIMINARY STUDY

1 Non-negative Matrix Factorization (NMF)

SIGNALS with sparse representations can be recovered

Transcription:

Sparseness Constraints on Nonnegative Tensor Decomposition Na Li nali@clarksonedu Carmeliza Navasca cnavasca@clarksonedu Department of Mathematics Clarkson University Potsdam, New York 3699, USA Department of Mathematics Clarkson University Potsdam, New York 3699, USA November 24, 2 Abstract The Nonnegative Tensor Factorization (NTF) method, has been shown to separate the mixture of several sound sources reasonably well Based on the sparsity of power spectrogram of signals, we propose to add sparseness constraints to one factor matrix, which contains frequency basis, to obtain a sparse representation of this nonnegative factor Two methods are presented for the sparse and nonnegative factor matrix The algorithms are tested to separate the instrument sources from a two channel mixtures containing a clarinet, a piano and a steel drum Some comparison results of these two NTFbased methods are then presented Introduction The purpose of this paper is to use a new numerical technique for tensor decomposition Given a tensor T R I J K, we find the decomposition of T into matrix factors A R I R, S R J R and G R K R, such that, T T R = A r S r G r r= where T is an approximation to T Here G r, A r and S r are the rth column of the matrices G, A and S, respectively We use the CANDECOMP/PARAFAC (CP) decomposition that is due to Harshman [] and Carol and Chang [6] In the CP decomposition, the matrices A R I R, S R J R, and G R K R are assumed to exist in which the original tensor T is defined as T ijk = R A ir S jr G kr, () r= for (i, j, k) [, I] [, J] [, K] A specific application of CP in signal processing is in sound source separation Sound source separation refers to the problem of synthesizing source signals given a mixture of those same source signals For example, the original tensor T R I J K contains the power spectrogram of the multi-channel mixed signals [] The unknown factors are the following: the entries of matrix G are the gains of each independent source from R channels and each matrix A r S T r is a power spectrogram of a source signal Due to some properties of the power spectrogram of signals [], the data has an inherent tensor structure and hence, tensor decomposition methods are applied The Alternating Least-Squares (ALS) method introduced in [, 6] is a well-known method for approximating the factors of the original tensor T described in For sound source separation, it is typical to add a nonnegativity constraint on all three factors due to the nature of the data Thus, each of the subproblem of the ALS is replaced by a nonnegative matrix factorization subproblem This method is called the Nonnegative Tensor Factorization (NTF); see [6, 3] for the implementation There are also various methods for sound source separation; see the references [7, 8,, 7, 8] In this paper, a two-channel mixture of the source signals from a clarinet, a piano and a steel drum is considered A tensor data consists of the spectrogram of the input channel signal in one channel Therefore, in the case of two channels, we obtain a tensor of size I J 2 By using the ALS algorithm, we can obtain the factor A, S and G with no constraints imposed on the factors Thus, we propose two hybrid nonnegative tensor decompositions which add sparseness constraint to one of the nonnegative factors Unlike the application of the NTF in [], all three factors are subjected to nonnegativity constraints In our application, the CP decomposition () is used while assuming that the matrix A is sparse since it represents the signals in the frequency basis

The proposed methods are optimization based methods relying on [5, 6] methods and sparsity methods, namely, compressed sensing for recovering sparse signals [5, 4] and projection-based technique to handle sparseness constraints [3] 2 Preliminaries We denote the scalars in R with lower-case letters (a, b, ) and the vectors with bold lower-case letters (a, b, ) The matrices are written as bold upper-case letters (A, B, ) and the symbol for tensors are calligraphic letters (A, B, ) The subscripts represent the following scalars: (A) ijk = a ijk, (A) ij = a ij, (a) i = a i and A r is the rth column of A The superscripts indicate the length of the vector or the size of the matrices For example, b K is a vector with length K and B N K is a N K matrix In addition, the lower-case superscripts on a matrix indicate the mode in which it has been matricized For example, T (n) is the mode-n matricization of the tensor T R I J K for n =, 2, 3 Definition 2 The Kronecker product of matrices A and B is defined as 2 3 a B a 2B 6 A B = a 4 2B a 22B 7 5 Definition 22 The Khatri-Rao product is the matching columnwise Kronecker product Given matrices A R I K and B R J K, their Khatri-Rao product is denoted by A B The result is a matrix of size (IJ K) defined by A B = [A B A 2 B 2 ] If a and b are vectors, then the Khatri-Rao and Kronecker products are identical, ie, a b = a b Definition 23 (Mode-n matricization) Matricization is the process of reordering the elements of an N th order tensor into a matrix The mode-n matricization of a tensor T R I I2 I N is denoted by T (n) and arranges the mode-n fibers, the vectors obtained from fixing every index with the exception of the nth mode, as the columns of the resulting matrix Definition 24 (Rank-one tensor) An N th order tensor T R I I2 I N is a rank-one if it can be written as the outer product of N vectors, ie, T = a () a (2) a (N), where a (r) R Ir, r N The symbol represents the vector outer product This means that each element of the tensor is the product of the corresponding vector elements: t ii 2 i N = a () i a (2) i 2 a (N) i N, for all i n I n 3 Optimization Methods The CP decomposition is the simplest factorization for representing a given tensor T R I J K into sums of rank-one tensors; ie solve for three factor matrices A R I R, S J R and G K R for a fixed R in the optimization problem: min A,S,G R T A r S r G r 2 F (2) r= Through tensor matricization via the Khatri-Rao products, the problem (2) is reformulated into subproblems: A k+ = argmin T I JK () A(G b k S k ) T 2 F, ba R I R S k+ = argmin T J IK (2) S(G b k A k+ ) T 2 F, bs R J R G k+ = argmin T K IJ (3) G(S b k+ A k+ ) T 2 F bg R K R This is the Alternating Least-Squares method This iterative technique requires initial guesses A, S and G For sound separation applications, we require additional constraints on each of the subproblems in the ALS algorithm: S and G are nonnegative matrices while A is a sparse nonnegative matrix To obtain nonnegative matrices S and G, the last two subproblems in (3) is replaced by the optimization problem of finding nonnegative factors W and H from a nonnegative matrix V : minimize V WH 2 F, subject to W, H (4) where denotes the nonnegative entries of W and H This is called Nonnegative Matrix Factorization () Several algorithms have been proposed for, namely, a gradient-based method by Paatero [6] and a multiplicative updating algorithm by Lee and Seung [5] 3 -Sparse vs l - For factor A, we have to deal with two constraints: sparsity and nonnegativity We implement two methods differing in the order of how the constraints are imposed: with sparse constraints [3] (-Sparse) and l - minimization with nonnegative constraints (l -) 3 l - The first subproblem in (3) is reformulated by first vectorizing the equation T () I JK = Â(Gk S k ) T into t = Qâ via column stacking with Q = I I I (G k S k ) R IJK IR, t R IJK and â R IR Then the leastsquares subproblem is replaced by an l -minimization with equality constraints [2]: min â l subject to t = Qâ (5) (3)

Note that if x R n, then x l = n x i The idea i= behind this model (5) is to construct sparse and exact solution vector â which matricizes into a sparse factor matrix  satisfying the CP decomposition The l minimization problem can be recast as linear program [2]: min â subject to â and Qâ = t (6) with both equality and inequality constraints The link between l minimization and linear programs has been known since the 95 s in the paper of [4] Moreover, numerical techniques for solving linear programs have been well studied In our codes, we implement Matlab s linear programming algorithm based on the simplex and interior-point methods To impose the nonnegative constraint on Â, the approximation from the linear program is further refined by a nonnegative least-squares method frame using the discrete Fourier transform (DFT) for each channel (see Figure 2a) Then the two spectrograms are stacked into a tensor format (see Figure 2b) The two-channel mixture of the signals is tensor T of size 24 89 2 5 5 2 3 4 5 6 7 8 9 5 5 left channel right channel 2 3 4 5 6 7 8 9 Figure : The two-channel sound signal mixtures x 4 x 4 32 -Sparse In [3], based on the method proposed in [6] and [5], Hoyer added a sparseness constraint on the algorithm Below we summarize the algorithm found in [3]: for the nonnegative matrix V = T () k Initialize W = A k calculated in a previous ALS iteration and set H = (G k S k ) T 2 W W µ W (WH V)H T 3 Project each column of W to a vector that is nonnegative with same l 2 norm, but the l norm is set to achieve the desired sparseness W ia (WHH T ) ia where µ W = is small positive multiplicative step from [5] In [3], Hoyer provided the projection operator algorithm find the closest (in the Euclidean sense) nonnegative vector s for any vector x constrained to a given the l norm and a given l 2 norm The desired sparsity of the factors, S ω and S h, are defined as where sparseness(w i ) = S ω, sparseness(h i ) = S h, for a vector x sparseness(x) = n x l x l2 n 4 Numerical Experiments A two-channel mixture of the sound source signals from a clarinet, a piano and a steel drum is the time-domain input signal (see Figure ) Each channel is a 936 time sampled signal in 2 seconds and is divided into 89 frames, the power spectrum is calculated within each (a) The frequency of the two channels (b) Tensor Figure 2: Tensor structure: the left frequency is the top slice and the right frequency is the bottom slice, obtaining a rotated tensor T R 24 89 2 According to the previous discussion in Section 3 and 32, below we describe the algorithm allowing different constraints on factor A The algorithm is the following: CP decomposition is applied to T via ALS (3) to obtain A, S and G Choose one from the following three methods for factor A and apply the nonnegative matrix factorization method for factors S and G () For nonnegative A, solve the first subproblem of (3) by (l -) For nonnegative sparse matrix A, use l minimization with nonnegative constraints (5) as the method for the first leastsquares subproblem (-Sparse) For nonnegative sparse A, use with sparseness constraints as a method by using the algorithm in Section (32) Nonnegative factor matrices S and G are solved by

2 3 4 5 6 7 8 9 original clarinet original piano original drum x 4 2 3 4 5 6 7 8 9 x 4 2 3 4 5 6 7 8 9 x 4 2 3 4 5 6 7 8 9 x 4 l 2 3 4 5 6 7 8 9 x 4 l 2 3 4 5 6 7 8 9 x 4 l 2 3 4 5 6 7 8 9 x 4 Sparse 2 3 4 5 6 7 8 9 x 4 Sparse 2 3 4 5 6 7 8 9 x 4 Sparse 2 3 4 5 6 7 8 9 x 4 (a) Separated waveforms for clarinet 2 3 4 5 6 7 8 9 x 4 (b) Separated waveforms for piano 2 3 4 5 6 7 8 9 x 4 (c) Separated waveforms for steel drum Figure 3: Waveforms of original signal,, l nonnegative minimization and with sparseness 2 For each r R, construct the source spectrogram, F r = A(:, r)s(:, r) T 3 The matrix G gives the ratios of each signal of the instrument in two channels let a vector of size R, H = log G(,:) G(2,:) Using k-means cluster method [9], H is divided into 3 clusters, where each is from an instrument So, H r corresponds to F r which should be in the same cluster 7 6 5 4 3 error comparison l sparse 4 Apply phase information [] for the spectrogram obtained in 2 where the corresponding source signal up to the clusters is dominant to F r Invert the spectrogram to obtain the time domain waveforms See Figures 3a 3c 2 2 3 Figure shows the two channel sound signal mixture input formed into a tensor We compared these methods (-Sparse, l -) and to the original waveforms in terms of the time domain waveform plots and frequency plots Figures 3a, 3b and 3c show the comparison results, the waveform (top) in each figure is the original sound signal waveform The rest of the waveforms (second from the top to bottom) in Figures 3a, 3b and 3c are, l - and -Sparse, respectively It can be seen that all three methods can capture the main characteristics of the sources We measure the distance between two signals w and v in a least squared error sense [] by the following equation: Figure 4: Error comparison of the different methods N E = ( w i v i ) 2 (7) i= where N is the length of the signals In Figure 4, set is the comparison of the separated clarinet signals, set 2 and set 3 are the separated piano and steel drum signals, respectively It shows that l - (the green bar) is better than the other two methods in all three sets Although both of l - and -Sparse obtain a sparse and nonnegative factor matrix A, we have seen that l - is better in the least squared error sense Figure 5: Spectrograms of the clarinet for the varying methods, the top left one is the original spectrogram Let us look at the clarinet spectrogram plots of the separated and original signals Figure 5 shows that the

-Sparse method does not capture the spectral density at frequency between to 5 4 and over 4 4 while the higher densities appear in the spectrograms of the and l - methods 5 Conclusion Two methods for hybrid tensor nonnegative decomposition are presented in the application of sound source separation The two methods, the l - and - Sparse, are proposed to implement the hybrid tensor decomposition to obtain a sparse and nonnegative factor due to the sparsity of power spectral The numerical examples show the effectiveness of these techniques Moreover, the error comparison plots show the l nonnegative minimization (l -) is better than The results of l - shows that it performs well with respect to the error and the resulting spectrogram plots of the separated signals which coincide with the actual sound The -Sparse method is the worst according to the error comparison More study is needed to compare the two methods Furthermore, we plan to develop more efficient techniques for sound source separation with better clustering methods 6 Acknowledgements CN and NL are both in part supported by the US National Science Foundation DMS-95 CN and NL are very grateful to Jason Holloway for his assistance NL would like to thank Prof Dr-Ing Gerald Schuller for useful conversations at the 44th IEEE Asilomar Conference on Signals, Systems and Computers References [] D Barry, R Lawlor and E Coyle, Comparison of Signal Reconstruction Methods for the Azimuth Discrimination and Resynthesis Algorithm, 8 th AES Convention Barcelona, Spain, 25 [2] S Boyd and L Vandenberghe Convex Optimization, Cambridge University Press, Cambridge, 24 [3] R Bro and S De Jong, A fast non-negativityconstrained least squares algorithm, Journal of Chemometrics, vol, pp 393-4, 997 [4] E Candès, J Romberg and T Tao, Stable signal recovery from incomplete and inaccurate measurements, Communications of Pure and Applied Mathematics, vol 59, pp 27-223, 25 [5] E Candès and T Tao, Decoding by linear programming, IEEE Trans Inform Theory, vol 52, no2, pp 423-425, 25 [6] J Carrol and J Chang, Analysis of Individual Differences in Multidimensional Scaling via an N-way Generalization of Eckart-Young Decomposition Psychometrika, vol 9, pp 267-283, 97 [7] M A Casey and A Westner, Separation of Mixed Audio Sources By Independent Subspace Analysis, Proc Of ICMC 2, pp 54-6, Berlin, Germany [8] L De Lathauwer, B De Moor and J Vandewalle, An introduction to independent component analysis, J Chemometrics, vol 4, pp 23-49, 2 [9] L Eldén, Matrix Methods in Data Mining and Pattern Recoqnition, Fundamentals of Algorithms 4, SIAM, 27 [] D Fitzgerald, M Cranitch and E Coyle, Non- Negative Tensor Factorization for Sound Source Separation, Dublin, Sept -2, ISSC 25 [] R A Harshman, Foundations of the PARAFAC procedure: Model and Conditions for an Explanatory Mutli-code Factor Analysis, UCLA Working Papers in Phonetics, vol 6, pp -84, 97 [2] JR Holloway and C Navasca Recovering tensor data from incomplete measurement via compressive sampling Proc IEEE Asilomar Conference on Signals, Controls and Systems 29 Pacific Grove, California, November, 29 [3] P O Hoyer, Non-negative Matrix Factorization with Sparseness Constraints, Journal of Machine Learning Research,5, 29, 457-469 [4] A Karmarkar, A new polynomial-time algorithm for linear programming, Combinatorica, vol 4, 984 [5] D Lee, H S Seung, Algorithms for non-negative matrix factorization, In Advances in Neural Information Processing 3 (Proc NIPS 2), MIT Press, 2 [6] P Paatero and U Tapper, Positive matrix factorization: A non-negative factor model with optimal utilization of error estimates of data values, Environmetrics, vol 5, pp -26, 994 [7] P Smaragdis and JC Brown, Non-negative Matrix Factorization for Polyphonic Music Transcription, IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), pp 77-8, October 23 [8] T Virtanen, Sound Source Separation Using Sparse Coding with Temporal Continuity Objective, Proc of International Computer Music Conference (ICMC23), Singapore, 23