CONVEX NONNEGATIVE MATRIX FACTORIZATION WITH MISSING DATA

Size: px
Start display at page:

Download "CONVEX NONNEGATIVE MATRIX FACTORIZATION WITH MISSING DATA"

Transcription

1 2016 IEEE INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING, SEPT , 2016, SALERNO, ITALY CONVEX NONNEGATIVE MATRIX FACTORIZATION WITH MISSING DATA Ronan Hamon, Valentin Emiya Aix Marseille Univ, CNRS, LIF, Marseille, France Cédric Févotte CNRS & IRIT, Toulouse, France ABSTRACT Convex nonnegative matrix factorization (CNMF) is a variant of nonnegative matrix factorization (NMF) in which the comonents are a convex combination of atoms of a known dictionary. In this contribution, we roose to extend CNMF to the case where the data matrix and the dictionary have missing entries. After a formulation of the roblem in this context of missing data, we roose a majorization-minimization algorithm for the solving of the otimization roblem incurred. Exerimental results with synthetic data and audio sectrograms highlight an imrovement of the erformance of reconstruction with resect to standard NMF. The erformance ga is articularly significant when the task of reconstruction becomes arduous, e.g. when the ratio of missing data is high, the noise is stee, or the comlexity of data is high. Index Terms matrix factorization, nonnegativity, lowrankness, matrix comletion, sectrogram inainting 1. INTRODUCTION Convex NMF (CNMF) [1] is a secial case of nonnegative matrix factorization (NMF) [2], in which the matrix of comonents is constrained to be a linear combination of atoms of a known dictionary. The term convex refers to the constraint of the linear combination, where the combination coefficients forming each comonent are nonnegative and sum to 1. Comared to the fully unsuervised NMF setting, the use of known atoms is a source of suervision that may guide learning based on this additional data: in articular, an interesting case of CNMF consists in auto-encoding the data themselves, by defining the atoms as the data matrix. CNMF has been of interest in a number of contexts, such as clustering, data analysis, face recognition, or music transcrition [1, 3]. It is also related to the self-exressive dictionary-based reresentation roosed in [4]. An issue that has not yet been addressed is when the data matrix has missing coefficients. Such an extension of CNMF is worth being considered, as it oens the way to data-reconstruction settings with nonnegative low-rank constraints, which covers several relevant alications. One This work was suorted by ANR JCJC rogram MAD (ANR- 14- CE ). examle concerns the field of image or audio inainting [5, 6, 7, 8], where CNMF may imrove the current reconstruction techniques. In inainting of audio sectrograms for examle, setting u the dictionary to be a comrehensive collection of notes from a secific instrument may guide the factorization toward a realistic and meaningful decomosition, increasing the quality of the reconstruction of the missing data. In this contribution, we also consider the case where the dictionary may have missing coefficients itself. The aer is organized as follows. Section 2 formulates CNMF in the resence of missing entries in the data matrix and in the dictionary. Section 3 describes the roosed majorization-minimization (MM) algorithm. Sections 4 and 5 reort exerimental results with synthetic data and audio sectrograms. 2. CONVEX NONNEGATIVE MATRIX FACTORIZATION WITH MISSING DATA 2.1. Notations and definitions For any integer N, the integer set t1, 2,...,Nu is denoted by rns. The coefficients of a matrix A P R MˆN are denoted by either a mn or ras mn. The element-wise matrix roduct, A matrix division and matrix ower are denoted by A.B, B and A., resectively where A and B are matrices with same dimensions and is a scalar. 0 and 1 denote vectors or matrices comosed of zeros and ones, resectively, with dimensions that can be deduced from the context. The element-wise negation of a binary matrix M is denoted by M fi 1 M NMF and Convex NMF NMF consists in aroximating a data matrix V P R F ` ˆN as the roduct WH of two nonnegative matrices W P R F ` ˆK and H P R KˆN `. Often, K min F, Nq, such that WH is a low-rank aroximation of V. Every samle v n, the n- th column of V, is thus decomosed as a linear combination of K elementary comonents or atterns w 1,...,w K P R F`, the columns of W. The coefficients of the linear combination are given by the n-th column h n of H. In [9] and [10], algorithms have been roosed for the unsuervised estimation of W and H from V, by minimization /16/$31.00 c 2016 IEEE

2 of the cost function D V WHq d v rwhs q, where d x yq is the -divergence defined as: $ 1 1q `x ` 1q y xy 1 & for P Rzt0, 1u d x yq fi x log x y x ` y for 1 (1) x y log x y 1 for 0 When ill-defined, we set by convention d 0 0q 0. CNMF is a variant of NMF in which W SL. S rs 1,...,s P spr F ` ˆP is a nonnegative matrix of atoms and L rl 1,...,l K spr P ` ˆK is the so-called labeling matrix. Each dictionary element w k is thus equal to Sl k, with usually P K, and the data is in the end decomosed as V SLH. The scale indeterminacy between L and H may be lifted by imosing }l k } 1 1, in which case w k is recisely a convex combination of the elements of the subsace S. CNMF can be related to the so-called archetyal analysis [11], but without considering any nonnegativity constraint. The use of known examles in S can then be seen as a source of suervision that guides learning. A secial case of CNMF is obtained by setting S V, thus auto-encoding the data as VLH. This articular case is studied in deth in [1]. In this aer, we consider the general case for S, with or without missing data Convex NMF with missing data We assume that some coefficients in V and S may be missing. Let V ÄrFsˆrNs be a set of airs of indices that locates the observed coefficients in V: f,nq PV iff v is known. Similarly, let S ÄrFsˆrPs be a set of airs of indices that locates the observed coefficients in S. The use of sets V and S may be reformulated equivalently by defining masking matrices M V Pt0, 1u F ˆN and M S Pt0, 1u F ˆP from V and S as # 1 if f,nq PV rm V s PrFsˆrNs (2) 0 otherwise # 1 if f,q PS rm S s f PrF sˆrps (3) 0 otherwise A major goal in this aer is to estimate L, H and the missing entries in S, given the artially observed data matrix V. Denoting by S o the set of observed/known dictionary matrix coefficients, our aim is to minimize the objective function C S, L, Hq fi D M V.V M V.SLHq (4) subject to S P R F ˆP `, L P R P ˆK `, H P R KˆN `, and M S.S M S.S o. The articular case where the dictionary equals the data matrix itself is obtained by setting M S, S o q fi M V, Vq. Algorithm 1 CNMF with missing data Require: V, S o, M V, M S, Initialize S, L, H with random nonnegative values loo Udate S: S M S.S o` (5) M V. SLHq. 2q.V LHq T. q M S.S.. M V. SLHq 1q LHq T Udate L: L L. ST Udate H: M V. SLHq. 2q.V H T S T M V. SLHq. 1q H T H H. SLqT M V. SLHq. 2q.V SLq T. M V. SLHq 1q Rescale L and H: end loo return S, L, H. q. q (6) PrKs, h k }l k } 1 ˆ h k (8) l k 3. PROPOSED ALGORITHM l k }l k } 1 (9) 3.1. General descrition of the algorithm Algorithm 1 extends the algorithm roosed in [9] for comlete CNMF with the -divergence to the case of missing entries in V or S. The algorithm is a block-coordinate descent rocedure in which each block is one the three matrix factors. The udates of each block/factor is obtained via majorizationminimization (MM), a classic rocedure that consists in iteratively minimizing a tight uer bound (called auxiliary function) of the objective function. In the resent setting, the MM rocedure leads to multilicative udates, characteristic of many NMF algorithms, that automatically reserve nonnegativity given ositive initialization Detailed udates We consider the otimization of C S, L, Hq with resect to each of its three arguments individually, using MM. Current udates are denoted with a tilde, i.e., S, r L r and H. r We start by recalling the definition of an auxiliary function: Definition 1 (Auxiliary function). The maing G A A r :

3 R IˆJ ` ˆ R IˆJ ` fiñ R` is an auxiliary function to C Aq iff P R IˆJ `, CAq G A A r P R IˆJ `, CAq G A A r (10). The iterative minimization of G A Aq with resect to A, with relacement of A r at every iteration, monotonically decreases the objective C Aq until convergence. As exlained in detail in [9], the -divergence may be decomosed into the sum of a convex term d q..q, a concave term d u..q and a constant term cst. The first two terms can be majorized using routine Jensen and tangent inequalities, resectively, leading to tractable udates. The auxiliary functions used to derive Algorithm 1 are given by the three following roositions 1 and the monotonicity of the algorithm follows by construction. Proosition 1 (Auxiliary function for S). Let S r P R F ` ˆP be such PrFsˆrNs, ṽ 0, PrFsˆrPs, s f 0, where V r fi SLH. r Then the function G S S S r fi ÿ rm V s qg S r S ` ug S r ı S ` cst where G q S S r fi ÿ ˆ rlhs n rs f q s f d v rv rv rs f S S r fi d u v rv q ` ud 1 v rv q ÿ f rlhs n s f rs f q is an auxiliary function to C S, L, Hq with resect to S and its minimum is given by equation (5). The auxiliary function decoules with resect to the individual coefficients of S and as such, the constraint M S.S M S.S o is directly imosed by only udating the coefficients of S with indices in S. Proosition 2 (Auxiliary function for L). Let L r P R P ` ˆK be such PrF sˆrns, ṽ 0 kq PrPsˆrKs, l k 0, where V r fi SLH. r Then the function G L fi ÿ rm V s qg L r L ` ug L r ı L ` cst where G q fi ÿ s f r lk h kn q l k d v rv rv k r lk fi d u v rv q ` ud 1 v rv q ÿ k s f h kn l k rl k 1 The roof of these roositions are available in the extended version at htts://hal-amu.archives-ouvertes.fr/hal is an auxiliary function to C S, L, Hq with resect to L and its minimum subject to M S.S M S.S o for M S P t0, 1u F ˆP and S o P R F ` ˆP is given by equation (6). Proosition 3 (Auxiliary function for H). Let us define W fi SL and let H r P R KˆN ` be such P rf sˆrns, rv 0 nq P rks ˆ rns, r h kn 0, where V r fi WH. r Then the function G H H H r fi ÿ rm V s qg H r H ` ug H r ı H ` cst where G q H H r fi ÿ w fk r ˆ hkn q h kn d v rv rv k r hkn H H r fi d u v rv q ` ud 1 v rv q ÿ k w fk h kn rh kn is an auxiliary function to C S, L, Hq with resect to H and its minimum is given by equation (7). 4. EXPERIMENT ON SYNTHETIC DATA 4.1. Exerimental setting The objective of this exeriment is to analyze the erformance of CNMF for reconstructing missing data, by comaring it with the regular NMF. We consider a data matrix V of rank K synthesized under the CNMF model V S L H, where the matrix of atoms S and the ground truth factors L and H are generated as the absolute values of Gaussian noise. It is worth noting that V is also consistent with a NMF model by defining W S L. A erturbed data matrix V is obtained by considering a multilicative noise, obtained using a Gamma distribution with mean 1 and variance 1. Hence the arameter controls the imortance of the erturbation. The mask M V of known elements in V is derived by considering missing coefficients randomly and uniformly distributed over the matrix, such that the ratio of missing values is equal to V. Generation of data is reeated 3 times, as well as the generation of the masks. Results are averaged over these reetitions. From a matrix V with missing entries, NMF and CNMF with missing values are alied using K comonents. Only the case where 2 has been considered in this exeriment. In both algorithms, the convergence is reached when the relative difference of the cost function between two iterations is below reetitions are erformed using different random initialization, and the best instance (i.e., the instance which minimizes the cost function) is retained. The reconstructed data matrix is obtained as V r SLH. The reconstruction error is obtained by comuting the - divergence between the noiseless data matrix V, and the reconstructed matrix V; r the error is comuted on and averaged

4 along the missing coefficients only, as e test 1 ij r M V s ij d M V.V, M V r Vq (11) where M V is the mask of unknown elements in V. In the case of CNMF, we consider two choices for S: the data matrix V with missing values, and the ground truth matrix of atoms S, considered here without missing values. The following arameters are fixed: F 100, N 250, P 50, K 10, 2. We articularly investigate the influence of four factors: the number of estimated comonents K Pr2, 14s; the ratio of missing data V Pr0.1, 0.9s in V, i.e., of zeros in M V ; the choice of the matrix S PtS, Vu for the CNMF; the noise level Pr10, 5000s in V ( is inversely roortional to the variance of the noise) Results We first focus on the influence of the number of estimated comonents k for the case where the true dictionary is fully known. Figure 1 dislays the test error with resect to the number of estimated comonents K, for two levels of noise. Performance of reconstruction obtained by NMF and CNMF with S S are lotted, for different values of ratio of missing values in V. = 100 V =0.2 -NMF V =0.2 -CNMF V =0.6 -NMF V =0.6 -CNMF V =0.8 -NMF V =0.8 -CNMF = 3000 for CNMF or NMF, the former still erforming better than the latter. These first results outline the difference between NMF and CNMF, emhasized in the next figures. This comarison is augmented by considering the case where S V, with resect to the ratio V of missing values in V. Figure 2 dislays the erformance of NMF and CNMF as a function of V, for two noise levels. CNMF with the true atoms (S S ) gives the best results on the full range missing data ratio. When there are very few missing data and a low noise, NMF erforms almost as well as CNMF. However, the NMF error increases much faster than the CNMF error as the number of missing data grows, or as the noise in data becomes imortant. This higher sensitivity of NMF to missing data may be exlained by overfitting since the number of free arameters in NMF is higher than in CNMF. In the case of CNMF with S V, the model cannot fit the data as well as CNMF with S S or as NMF. Consequently, the resulting modeling error is observed when there is few missing data, and when comaring S V and S S on all values. However, it erforms better than NMF at high values of V since the constraint S V can be seen as a regularization. We finally investigate the robustness of methods by looking at the influence of the multilicative noise, controlled by the arameter, on the erformance. Figure 3 shows the test error for the NMF and the CNMF with S S with resect to and for some values of V. As exected, the test error decreases according to the variance of the noise, inversely roortional to. If a low value of disruts abrutly the erformance of reconstruction ( ), the test error is slightly decreasing for. When the variance of the noise is close to zero ( 5000), the erformance of NMF and CNMF are almost the same. The erformance of reconstruction differs when the variance of the noise increases, as well as the ratio of missing values in V Number of comonents Number of comonents = 100 = 3000 NMF S = S -CNMF S = V -CNMF Fig. 1. vs. number of comonents. Two levels of noise are dislayed: high noise level ( 100, left) and low noise level ( 3000, right). These results show that the noise level has a high influence on the best number of estimated comonents. As exected, a high noise requires a strong regularization, obtained here by selecting a low value of K. On the contrary, when the noise is low, the best choice of K is closer to the true value K 10. Similarly, when the number of missing data is low ( V 0.2), one should set K to K 10, either for CNMF or for NMF. When it gets higher, the otimal K gets lower in order to limit the effect of overfitting. In this case, the best number of comonents dros down to K 2, either Ratio of missing data in V Ratio of missing data in V Fig. 2. vs. ratio of missing data in V, with K K. Two levels of noise are dislayed: high noise level ( 100, left) and low noise level ( 3000, right).

5 V =0.2 -NMF V =0.2 -CNMF V =0.6 -NMF V =0.6 -CNMF V =0.8 -NMF V =0.8 -CNMF Fig. 3. vs. noise level. Each curve describes the erformance of reconstruction of missing data in V, with K K, according to the method and the ratio of missing data V. 5. APPLICATION TO SPECTROGRAM INPAINTING In order to illustrate the erformance of the roosed algorithm on real data, we consider sectrograms of iano music signals, which are known to be well modeled by NMF methods [12]. Indeed, NMF may rovide a note-level decomosition of a full music iece, each NMF comonent being the estimated sectrum and activation of a single note. This aroximation has roved successful and is also limited in terms of modelling error and of lack of suervision to guide the NMF algorithm. In such condition, we have designed an exeriment with missing data to comare regular NMF against two CNMF variants: in the first one, we set S V; in the second one, S contains examles of all ossible isolated note sectra from another iano. the recording of another iano instrument from the MAPS dataset, from C3 to C Results Figure 4 dislays the test error with resect to the ratio of missing data in V, averaged over the 17 iano ieces. It clearly shows that the CNMF with the secific dictionary S D is much more robust to missing data than the other two systems. When less than 40% of data are missing, NMF erforms slightly better; however, the NMF test error dramatically increases when more data are missing, by a factor 5. when more than 80% data are missing. This must be due to overfitting since NMF has a large number of free arameters to be estimated from very few observations when data are missing. The erformance of the CNMF system with S V robably suffers from modelling error when very few data are missing since the columns of V may not be able to combine into K comonents in a convex way. In the range 50 70%, its erformance is similar to that of NMF. Beyond this range, it seems to be less rone to overfitting than NMF, robably due to less free arameters or to a regularization effect rovided by S V. NMF ConvexNMF - S=D ConvexNMF - S=V 5.1. Exerimental setting We consider 17 iano ieces from the MAPS dataset [13]. For each recording, the magnitude sectrogram is comuted using a 46-ms sine window with 50%-overla and 4096 frequency bins, the samling frequency being 44.1kHz. Matrix V is created from the resulting sectrogram by selecting the F 500 lower-frequency art and the first five seconds, i.e., the N 214 first time frames. Missing data in V are artificially created by removing coefficients uniformly at random. Three systems are comared, based on the test error defined as the -divergence comuted on the estimation of missing data. In all of them, the number of comonent K is set to the true number of notes available from the dataset annotation and we set 1. The first system is the regular NMF, randomly initialized. The second system is the roosed CNMF with M S, S o q fi M V, Vq. The third system is the roosed CNMF with S D set as a secific matrix D of P 61 atoms. Each atom is a single-note sectrum extracted from Ratio of missing data in V Fig. 4. vs. missing data ratio in audio sectrograms. We now investigate the influence of the comlexity of the audio signal on the test error when the ratio of missing data is set to the high value 80%. Figure 5 dislays, for each music recording, the test error with resect to the number of different itches for all notes in the iano iece, which also equals the number of comonent K used by each system. CNMF with S D erforms better than NMF whatever the number of notes and the error increases by a small factor along the reresented range. NMF erforms about five times worse for easy ieces, i.e., ieces comosed of notes with about 6 different itches and it erforms about 10 4 times worse when the number of itches is larger than 25. CNMF 2 The code of the exeriments is available on the webage of the MAD roject htt://mad.lif.univ-mrs.fr/.

6 with S V erforms slightly better than NMF. Since the number of comonents K increases equals the number of note itches, those results confirm that NMF may highly suffer from overtraining while CNMF may not, being robust to missing data even for large values of K. NMF ConvexNMF - S=D ConvexNMF - S=V Number of note itches (or K) Fig. 5. vs. number of different note itches for 80% missing data (each dot reresents one iece of music). 6. CONCLUSION In this aer, we have roosed an extension of convex nonnegative matrix factorization in the case where missing data, as it has been reviously resented for regular NMF. The roosed method can deal with missing values both in the data matrix V and in the dictionary S, which is articularly useful in the case S V where the data is autoencoded. In this framework, an algorithm has been rovided and analyzed using a Majorization-Minimization (MM) scheme to guarantee the convergence to a local minimum. A large set of exeriments on synthetic data showed romising results for this variant of NMF for the task of reconstruction of missing data, and validated the value of this aroach. In many situations, CNMF outerforms NMF, esecially when the ratio of missing values is high and when the matrix data V is noisy. This trend has been confirmed on real audio sectrograms of iano music. In articular, we have shown how the use of a generic set of isolated iano notes as atoms could dramatically enhance the robustness to missing data. This reliminary study indicates that it is worthy of further investigation, beyond the roosed settings where missing values are uniformly distributed over the matrix. Furthermore, the influence of missing values in the dictionary has not been comletely assessed, as only the case where S V has been taken into account. On the alication side, this aroach could give new insight in many roblems dealing with estimation of missing data. 7. REFERENCES [1] C. Ding, Tao Li, and M.I. Jordan, Convex and seminonnegative matrix factorizations, IEEE Trans. Pattern Anal. Mach. Intell., vol. 32, no. 1, , Jan [2] D.D. Lee and H.S. Seung, Learning the arts of objects by non-negative matrix factorization, Nature, vol. 401, no. 6755, , [3] E. Vincent, N. Bertin, and R. Badeau, Adative harmonic sectral decomosition for multile itch estimation, IEEE Trans. Audio, Seech, Language Process., vol. 18, no. 3, , Mar [4] E. Elhamifar and R. Vidal, Sarse subsace clustering: Algorithm, theory, and alications, IEEE Trans. Pattern Anal. Mach. Intell., vol. 35, no. 11, , [5] M. Bertalmio, G. Sairo, V. Caselles, and C. Ballester, Image inainting, in Proc of SIGGRAPH. ACM, 2000, [6] P. Smaragdis, B. Raj, and M. Shashanka, Missing data imutation for sectral audio signals, in Proc. of MLSP, Grenoble, France, Set [7] J. Le Roux, H. Kameoka, N. Ono, A. De Cheveigne, and S. Sagayama, Comutational auditory induction as a missing-data model-fitting roblem with bregman divergence, Seech Communication, vol. 53, no. 5, , [8] D.L. Sun and R. Mazumder, Non-negative matrix comletion for bandwidth extension: A convex otimization aroach, in Proc. of MLSP, Set. 2013, [9] C. Févotte and J. Idier, Algorithms for nonnegative matrix factorization with the -divergence, Neural Comut., vol. 23, no. 9, , [10] M. Nakano, H. Kameoka, J. Le Roux, Y. Kitano, N. Ono, and S. Sagayama, Convergence-guaranteed multilicative algorithms for nonnegative matrix factorization with -divergence, Proc. of MLSP, vol. 10,. 1, [11] A. Cutler and L. Breiman, Archetyal analysis, Technometrics, vol. 36, no. 4, , [12] C. Févotte, N. Bertin, and J-L Durrieu, Nonnegative matrix factorization with the Itakura-Saito divergence: With alication to music analysis, Neural Comut., vol. 21, no. 3, , [13] V. Emiya, R. Badeau, and B. David, Multiitch estimation of iano sounds using a new robabilistic sectral smoothness rincile, IEEE Trans. Audio, Seech, Language Process., vol. 18, no. 6, , 2010.

7 A. PROOFS We detail the roofs of Proositions 1 and 2. The roof of Proosition 3 is straithforward using the same methodology. We first recall reliminary elementss from [9]. The -divergence d x yq can be decomosed as a sum of a convex term, a concave term and a constant term with resect to its second variable y as d x yq q d x yq` ud x yq`d xq (12) This decomosition is not unique. We will use decomosition given in [9, Table 1], for which we have the following derivatives w.r.t. variable y: $ & xy 2 if 1 qd 1 x yq fi y 2 y xq if 1 2 (13) y 1 if 2 $ & y 1 if 1 ud 1 x yq fi 0 if 1 2 (14) xy 2 if 2 A.1. Udate of S (Proof of Proosition 1) We rove Proosition 1 by first constructing the auxiliary function (Proosition 4 below) and then focussing on its minimum (Proosition 5 below). Due to searability, the udate of S P R F ` ˆP relies on the udate of each of its columns. Hence, we only derive the udate for a vector s P R P`. Definition 2 (Objective function C S sq). For v P R Ǹ, L P R P ` ˆK define C S sq fi ÿ n, H P R KˆN `, m Pt0, 1u N, m o Pt0, 1u P, s P R P`, let us m qcn n sq` uc ı n sq ` C (15) 0where qc n sq fi q d `v n s T LH n, u Cn sq fi u d `v n s T LH n and C fi d m.vq` d m o.s o q (16) Proosition 4 (Auxiliary function G S s rsq for C S sq). Let rs P R P` be such rv n 0 rs 0, where rv fi s T LH T. Then the function GS s rsq defined by G S s rsq fi ÿ n m qgn n s rsq` ug ı n s rsq ` C (17) where qg n s rsq fi ÿ rlhs n rs d q rv n ˆ s v n rv n and G rs u n s rsq fi d u v n rv n q` ud 1 v n rv n q ÿ rlhs n s rs q. (18) is an auxiliary function for C S sq. Proof. We trivially have G S s sq C S sq. We use the searability in n and in in order to uer bound each convex term qc n sq and each concave term u C n sq. Convex term q C n sq. Let us rove that q G n s rsq q C n sq. Let P be the set of indices such that rlhs n 0 and define, for P P, r n fi rlhs n rs rlhs n rs rv n 1 PP rlhs 1 n rs. (19) 1

8 We have PP r n 1 and qg n s rsq ÿ PP r n q d v n rlhs n s r n q d v n ÿ PP rlhs r n s n d r q n Pÿ v n rlhs n s C q n sq (20) 1 Concave term C u n sq. We have G u n s rsq C u n sq since C u n sq is concave and s fiñ G u n s rsq is a tangent lane to C u n sq in rs: ug n s rsq d u v n rv n q`ÿ ud 1 v n ÿ 1 rlhs 1 n rs rlhs n s rs q C u n rsq` r C u n rsq, s rs (21) 1 Proosition 5 (Minimum of G S s rsq). The minimum of s fiñ G S s rsq subject to the constraint m o.s m o.s o is reached at s MM with $ ˆ 2 q s MM n mn rv n v nrlhs rs n fi 1 if m o n mn rv n rlhs 0 n (22) s o if m o 1. Proof. Since variable s is fixed for such that m o 1, we only consider variables s for such that m o 0. The related enalty term in G S s rsq vanishes when m o 0. Using (13) and (14), the minimum is obtained by cancelling the gradient r s G S s rsq ÿ n m n LH n qd ˆv 1 s n rv n ` rs ud 1 v n rv n q and by considering that the Hessian matrix is diagonal with nonnegative entries since q d x yq is convex: r 2 s G S s rsq ÿ n (23) ˆ LH n m n rv n d q2 s v n rv n 0. (24) rs rs A.2. Udate of L (Proof of Proosition 2) We roove Proosition 2 by first constructing the auxiliary function (Proosition 6 below) and then focussing on its minimum (Proosition 7 below). As oosed to the udate of S, no searability is considered here. Definition 3 (Objective function C L lq). For v P R F`, S P R F ` ˆP ı C L Lq fi ÿ m qc Lq` uc Lq, H P R KˆN `, M Pt0, 1u F ˆN, L P R P ` ˆK, let us define ` C (25) where qc Lq fi q d v rslhs, u C Lq fi u d v rslhs and C fi d M.Vq. (26) Proosition 6 (Auxiliary function G L for C L Lq). Let L r P R P ` ˆK be such V r 0 k, L r k 0, where V r fi SLH. r Then the function G L defined by G L fi ÿ m qg L r L ` ug L r ı L ` C (27)

9 where qg fi ÿ s f r lk h kn rv k q d v rv l k r lk «fi ud v rv q` ud 1 v rv q ÿ s f h kn l k k rl k (28) (29) is an auxiliary function for C L Lq. Proof. We trivially have G L L Lq C L Lq. In order to rove that G L C L Lq, we use the searability in f and n and we uer bound the convex terms C q Lq and the concave terms C u Lq. Convex term C q Lq. Let us rove that G q C q Lq. Let P be the set of indices such that s f 0, K be the set of indices such that h kn 0 and define, for, kq PP ˆ K, We have,kqppˆk r fkn 1 and qg q d ÿ,kqppˆk v r fkn fi s f r l k h kn rv r fkn q d v s fl k h kn r fkn ÿ,kqppˆk r fkn s f l k h kn r fkn s f r lk h kn 1,k 1 qppˆk s r f 1 l1 k 1h k 1 n q d v Pÿ 1 k 1. (30) (31) Kÿ s f l k h kn C q Lq (32) Concave term C u Lq. We have G u C u Lq since C u Lq is concave and L fiñ G u L L r is a tangent lane to uc Lq in r L: d u v rv q ` ÿ u C rl ` k ud 1 v ÿ 1 k 1 s f 1r l1 k 1h k 1 n DrC u rl, L rl E s f h kn l k rl k (33) (34) Proosition 7 (Minimum of G L L L r ). The minimum of L fiñ G L L L r is reached at L MM k, l MM,k fi $ ˆ & l,k l,k otherwise. s fm rv 2 v h kn s fm rv 1 h kn q if M 0 (35) Proof. Using (13) and (14), the minimum is obtained by cancelling the gradient r lk G L ÿ «m s f h kn qd v 1 l k rv ` r ud 1 v rv q (36) lk and by considering that the Hessian matrix is diagonal with nonnegative entries since d q x yq is convex: r 2 l k G L ÿ s f h kn m rv d q2 l k v r rv 0. (37) lk r hk

Convex Optimization methods for Computing Channel Capacity

Convex Optimization methods for Computing Channel Capacity Convex Otimization methods for Comuting Channel Caacity Abhishek Sinha Laboratory for Information and Decision Systems (LIDS), MIT sinhaa@mit.edu May 15, 2014 We consider a classical comutational roblem

More information

L P -NORM NON-NEGATIVE MATRIX FACTORIZATION AND ITS APPLICATION TO SINGING VOICE ENHANCEMENT. Tomohiko Nakamura and Hirokazu Kameoka,

L P -NORM NON-NEGATIVE MATRIX FACTORIZATION AND ITS APPLICATION TO SINGING VOICE ENHANCEMENT. Tomohiko Nakamura and Hirokazu Kameoka, L P -NORM NON-NEGATIVE MATRIX FACTORIZATION AND ITS APPLICATION TO SINGING VOICE ENHANCEMENT Tomohio Naamura and Hiroazu Kameoa, Graduate School of Information Science and Technology, The University of

More information

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning

Uncorrelated Multilinear Principal Component Analysis for Unsupervised Multilinear Subspace Learning TNN-2009-P-1186.R2 1 Uncorrelated Multilinear Princial Comonent Analysis for Unsuervised Multilinear Subsace Learning Haiing Lu, K. N. Plataniotis and A. N. Venetsanooulos The Edward S. Rogers Sr. Deartment

More information

Estimation of the large covariance matrix with two-step monotone missing data

Estimation of the large covariance matrix with two-step monotone missing data Estimation of the large covariance matrix with two-ste monotone missing data Masashi Hyodo, Nobumichi Shutoh 2, Takashi Seo, and Tatjana Pavlenko 3 Deartment of Mathematical Information Science, Tokyo

More information

Radial Basis Function Networks: Algorithms

Radial Basis Function Networks: Algorithms Radial Basis Function Networks: Algorithms Introduction to Neural Networks : Lecture 13 John A. Bullinaria, 2004 1. The RBF Maing 2. The RBF Network Architecture 3. Comutational Power of RBF Networks 4.

More information

Positive decomposition of transfer functions with multiple poles

Positive decomposition of transfer functions with multiple poles Positive decomosition of transfer functions with multile oles Béla Nagy 1, Máté Matolcsi 2, and Márta Szilvási 1 Deartment of Analysis, Technical University of Budaest (BME), H-1111, Budaest, Egry J. u.

More information

Approximating min-max k-clustering

Approximating min-max k-clustering Aroximating min-max k-clustering Asaf Levin July 24, 2007 Abstract We consider the roblems of set artitioning into k clusters with minimum total cost and minimum of the maximum cost of a cluster. The cost

More information

Topic: Lower Bounds on Randomized Algorithms Date: September 22, 2004 Scribe: Srinath Sridhar

Topic: Lower Bounds on Randomized Algorithms Date: September 22, 2004 Scribe: Srinath Sridhar 15-859(M): Randomized Algorithms Lecturer: Anuam Guta Toic: Lower Bounds on Randomized Algorithms Date: Setember 22, 2004 Scribe: Srinath Sridhar 4.1 Introduction In this lecture, we will first consider

More information

Numerical Linear Algebra

Numerical Linear Algebra Numerical Linear Algebra Numerous alications in statistics, articularly in the fitting of linear models. Notation and conventions: Elements of a matrix A are denoted by a ij, where i indexes the rows and

More information

A Recursive Block Incomplete Factorization. Preconditioner for Adaptive Filtering Problem

A Recursive Block Incomplete Factorization. Preconditioner for Adaptive Filtering Problem Alied Mathematical Sciences, Vol. 7, 03, no. 63, 3-3 HIKARI Ltd, www.m-hiari.com A Recursive Bloc Incomlete Factorization Preconditioner for Adative Filtering Problem Shazia Javed School of Mathematical

More information

4. Score normalization technical details We now discuss the technical details of the score normalization method.

4. Score normalization technical details We now discuss the technical details of the score normalization method. SMT SCORING SYSTEM This document describes the scoring system for the Stanford Math Tournament We begin by giving an overview of the changes to scoring and a non-technical descrition of the scoring rules

More information

EXACTLY PERIODIC SUBSPACE DECOMPOSITION BASED APPROACH FOR IDENTIFYING TANDEM REPEATS IN DNA SEQUENCES

EXACTLY PERIODIC SUBSPACE DECOMPOSITION BASED APPROACH FOR IDENTIFYING TANDEM REPEATS IN DNA SEQUENCES EXACTLY ERIODIC SUBSACE DECOMOSITION BASED AROACH FOR IDENTIFYING TANDEM REEATS IN DNA SEUENCES Ravi Guta, Divya Sarthi, Ankush Mittal, and Kuldi Singh Deartment of Electronics & Comuter Engineering, Indian

More information

An Analysis of Reliable Classifiers through ROC Isometrics

An Analysis of Reliable Classifiers through ROC Isometrics An Analysis of Reliable Classifiers through ROC Isometrics Stijn Vanderlooy s.vanderlooy@cs.unimaas.nl Ida G. Srinkhuizen-Kuyer kuyer@cs.unimaas.nl Evgueni N. Smirnov smirnov@cs.unimaas.nl MICC-IKAT, Universiteit

More information

On the capacity of the general trapdoor channel with feedback

On the capacity of the general trapdoor channel with feedback On the caacity of the general tradoor channel with feedback Jui Wu and Achilleas Anastasooulos Electrical Engineering and Comuter Science Deartment University of Michigan Ann Arbor, MI, 48109-1 email:

More information

ON OPTIMIZATION OF THE MEASUREMENT MATRIX FOR COMPRESSIVE SENSING

ON OPTIMIZATION OF THE MEASUREMENT MATRIX FOR COMPRESSIVE SENSING 8th Euroean Signal Processing Conference (EUSIPCO-2) Aalborg, Denmark, August 23-27, 2 ON OPTIMIZATION OF THE MEASUREMENT MATRIX FOR COMPRESSIVE SENSING Vahid Abolghasemi, Saideh Ferdowsi, Bahador Makkiabadi,2,

More information

A multiplicative algorithm for convolutive non-negative matrix factorization based on squared euclidean distance

A multiplicative algorithm for convolutive non-negative matrix factorization based on squared euclidean distance Loughborough University Institutional Reository A multilicative algorithm for convolutive non-negative matrix factorization based on squared euclidean distance This item was submitted to Loughborough University's

More information

GOOD MODELS FOR CUBIC SURFACES. 1. Introduction

GOOD MODELS FOR CUBIC SURFACES. 1. Introduction GOOD MODELS FOR CUBIC SURFACES ANDREAS-STEPHAN ELSENHANS Abstract. This article describes an algorithm for finding a model of a hyersurface with small coefficients. It is shown that the aroach works in

More information

16.2. Infinite Series. Introduction. Prerequisites. Learning Outcomes

16.2. Infinite Series. Introduction. Prerequisites. Learning Outcomes Infinite Series 6.2 Introduction We extend the concet of a finite series, met in Section 6., to the situation in which the number of terms increase without bound. We define what is meant by an infinite

More information

Optimal array pattern synthesis with desired magnitude response

Optimal array pattern synthesis with desired magnitude response Otimal array attern synthesis with desired magnitude resonse A.M. Pasqual a, J.R. Arruda a and P. erzog b a Universidade Estadual de Caminas, Rua Mendeleiev, 00, Cidade Universitária Zeferino Vaz, 13083-970

More information

AI*IA 2003 Fusion of Multiple Pattern Classifiers PART III

AI*IA 2003 Fusion of Multiple Pattern Classifiers PART III AI*IA 23 Fusion of Multile Pattern Classifiers PART III AI*IA 23 Tutorial on Fusion of Multile Pattern Classifiers by F. Roli 49 Methods for fusing multile classifiers Methods for fusing multile classifiers

More information

Finding a sparse vector in a subspace: linear sparsity using alternating directions

Finding a sparse vector in a subspace: linear sparsity using alternating directions IEEE TRANSACTION ON INFORMATION THEORY VOL XX NO XX 06 Finding a sarse vector in a subsace: linear sarsity using alternating directions Qing Qu Student Member IEEE Ju Sun Student Member IEEE and John Wright

More information

Information collection on a graph

Information collection on a graph Information collection on a grah Ilya O. Ryzhov Warren Powell February 10, 2010 Abstract We derive a knowledge gradient olicy for an otimal learning roblem on a grah, in which we use sequential measurements

More information

ON THE LEAST SIGNIFICANT p ADIC DIGITS OF CERTAIN LUCAS NUMBERS

ON THE LEAST SIGNIFICANT p ADIC DIGITS OF CERTAIN LUCAS NUMBERS #A13 INTEGERS 14 (014) ON THE LEAST SIGNIFICANT ADIC DIGITS OF CERTAIN LUCAS NUMBERS Tamás Lengyel Deartment of Mathematics, Occidental College, Los Angeles, California lengyel@oxy.edu Received: 6/13/13,

More information

Elementary Analysis in Q p

Elementary Analysis in Q p Elementary Analysis in Q Hannah Hutter, May Szedlák, Phili Wirth November 17, 2011 This reort follows very closely the book of Svetlana Katok 1. 1 Sequences and Series In this section we will see some

More information

ENHANCING TIMBRE MODEL USING MFCC AND ITS TIME DERIVATIVES FOR MUSIC SIMILARITY ESTIMATION

ENHANCING TIMBRE MODEL USING MFCC AND ITS TIME DERIVATIVES FOR MUSIC SIMILARITY ESTIMATION th Euroean Signal Processing Conference (EUSIPCO ) Bucharest, Romania, August 7-3, ENHANCING TIMBRE MODEL USING AND ITS TIME DERIVATIVES FOR MUSIC SIMILARITY ESTIMATION Franz de Leon, Kirk Martinez Electronics

More information

The analysis and representation of random signals

The analysis and representation of random signals The analysis and reresentation of random signals Bruno TOÉSNI Bruno.Torresani@cmi.univ-mrs.fr B. Torrésani LTP Université de Provence.1/30 Outline 1. andom signals Introduction The Karhunen-Loève Basis

More information

Information collection on a graph

Information collection on a graph Information collection on a grah Ilya O. Ryzhov Warren Powell October 25, 2009 Abstract We derive a knowledge gradient olicy for an otimal learning roblem on a grah, in which we use sequential measurements

More information

A Note on Guaranteed Sparse Recovery via l 1 -Minimization

A Note on Guaranteed Sparse Recovery via l 1 -Minimization A Note on Guaranteed Sarse Recovery via l -Minimization Simon Foucart, Université Pierre et Marie Curie Abstract It is roved that every s-sarse vector x C N can be recovered from the measurement vector

More information

Sampling and Distortion Tradeoffs for Bandlimited Periodic Signals

Sampling and Distortion Tradeoffs for Bandlimited Periodic Signals Samling and Distortion radeoffs for Bandlimited Periodic Signals Elaheh ohammadi and Farokh arvasti Advanced Communications Research Institute ACRI Deartment of Electrical Engineering Sharif University

More information

arxiv: v1 [physics.data-an] 26 Oct 2012

arxiv: v1 [physics.data-an] 26 Oct 2012 Constraints on Yield Parameters in Extended Maximum Likelihood Fits Till Moritz Karbach a, Maximilian Schlu b a TU Dortmund, Germany, moritz.karbach@cern.ch b TU Dortmund, Germany, maximilian.schlu@cern.ch

More information

Semi-Orthogonal Multilinear PCA with Relaxed Start

Semi-Orthogonal Multilinear PCA with Relaxed Start Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI 2015) Semi-Orthogonal Multilinear PCA with Relaxed Start Qiuan Shi and Haiing Lu Deartment of Comuter Science

More information

arxiv: v2 [stat.ml] 7 May 2015

arxiv: v2 [stat.ml] 7 May 2015 Semi-Orthogonal Multilinear PCA with Relaxed Start Qiuan Shi and Haiing Lu Deartment of Comuter Science Hong Kong Batist University, Hong Kong, China csshi@com.hkbu.edu.hk, haiing@hkbu.edu.hk arxiv:1504.08142v2

More information

Hotelling s Two- Sample T 2

Hotelling s Two- Sample T 2 Chater 600 Hotelling s Two- Samle T Introduction This module calculates ower for the Hotelling s two-grou, T-squared (T) test statistic. Hotelling s T is an extension of the univariate two-samle t-test

More information

t 0 Xt sup X t p c p inf t 0

t 0 Xt sup X t p c p inf t 0 SHARP MAXIMAL L -ESTIMATES FOR MARTINGALES RODRIGO BAÑUELOS AND ADAM OSȨKOWSKI ABSTRACT. Let X be a suermartingale starting from 0 which has only nonnegative jums. For each 0 < < we determine the best

More information

A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement

A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement A Variance Modeling Framework Based on Variational Autoencoders for Speech Enhancement Simon Leglaive 1 Laurent Girin 1,2 Radu Horaud 1 1: Inria Grenoble Rhône-Alpes 2: Univ. Grenoble Alpes, Grenoble INP,

More information

On split sample and randomized confidence intervals for binomial proportions

On split sample and randomized confidence intervals for binomial proportions On slit samle and randomized confidence intervals for binomial roortions Måns Thulin Deartment of Mathematics, Usala University arxiv:1402.6536v1 [stat.me] 26 Feb 2014 Abstract Slit samle methods have

More information

On Line Parameter Estimation of Electric Systems using the Bacterial Foraging Algorithm

On Line Parameter Estimation of Electric Systems using the Bacterial Foraging Algorithm On Line Parameter Estimation of Electric Systems using the Bacterial Foraging Algorithm Gabriel Noriega, José Restreo, Víctor Guzmán, Maribel Giménez and José Aller Universidad Simón Bolívar Valle de Sartenejas,

More information

The Noise Power Ratio - Theory and ADC Testing

The Noise Power Ratio - Theory and ADC Testing The Noise Power Ratio - Theory and ADC Testing FH Irons, KJ Riley, and DM Hummels Abstract This aer develos theory behind the noise ower ratio (NPR) testing of ADCs. A mid-riser formulation is used for

More information

Using the Divergence Information Criterion for the Determination of the Order of an Autoregressive Process

Using the Divergence Information Criterion for the Determination of the Order of an Autoregressive Process Using the Divergence Information Criterion for the Determination of the Order of an Autoregressive Process P. Mantalos a1, K. Mattheou b, A. Karagrigoriou b a.deartment of Statistics University of Lund

More information

Various Proofs for the Decrease Monotonicity of the Schatten s Power Norm, Various Families of R n Norms and Some Open Problems

Various Proofs for the Decrease Monotonicity of the Schatten s Power Norm, Various Families of R n Norms and Some Open Problems Int. J. Oen Problems Comt. Math., Vol. 3, No. 2, June 2010 ISSN 1998-6262; Coyright c ICSRS Publication, 2010 www.i-csrs.org Various Proofs for the Decrease Monotonicity of the Schatten s Power Norm, Various

More information

Approximate Dynamic Programming for Dynamic Capacity Allocation with Multiple Priority Levels

Approximate Dynamic Programming for Dynamic Capacity Allocation with Multiple Priority Levels Aroximate Dynamic Programming for Dynamic Caacity Allocation with Multile Priority Levels Alexander Erdelyi School of Oerations Research and Information Engineering, Cornell University, Ithaca, NY 14853,

More information

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression

A Comparison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Journal of Modern Alied Statistical Methods Volume Issue Article 7 --03 A Comarison between Biased and Unbiased Estimators in Ordinary Least Squares Regression Ghadban Khalaf King Khalid University, Saudi

More information

Supplementary Materials for Robust Estimation of the False Discovery Rate

Supplementary Materials for Robust Estimation of the False Discovery Rate Sulementary Materials for Robust Estimation of the False Discovery Rate Stan Pounds and Cheng Cheng This sulemental contains roofs regarding theoretical roerties of the roosed method (Section S1), rovides

More information

Improved Capacity Bounds for the Binary Energy Harvesting Channel

Improved Capacity Bounds for the Binary Energy Harvesting Channel Imroved Caacity Bounds for the Binary Energy Harvesting Channel Kaya Tutuncuoglu 1, Omur Ozel 2, Aylin Yener 1, and Sennur Ulukus 2 1 Deartment of Electrical Engineering, The Pennsylvania State University,

More information

MODELING THE RELIABILITY OF C4ISR SYSTEMS HARDWARE/SOFTWARE COMPONENTS USING AN IMPROVED MARKOV MODEL

MODELING THE RELIABILITY OF C4ISR SYSTEMS HARDWARE/SOFTWARE COMPONENTS USING AN IMPROVED MARKOV MODEL Technical Sciences and Alied Mathematics MODELING THE RELIABILITY OF CISR SYSTEMS HARDWARE/SOFTWARE COMPONENTS USING AN IMPROVED MARKOV MODEL Cezar VASILESCU Regional Deartment of Defense Resources Management

More information

System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests

System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests 009 American Control Conference Hyatt Regency Riverfront, St. Louis, MO, USA June 0-, 009 FrB4. System Reliability Estimation and Confidence Regions from Subsystem and Full System Tests James C. Sall Abstract

More information

16.2. Infinite Series. Introduction. Prerequisites. Learning Outcomes

16.2. Infinite Series. Introduction. Prerequisites. Learning Outcomes Infinite Series 6. Introduction We extend the concet of a finite series, met in section, to the situation in which the number of terms increase without bound. We define what is meant by an infinite series

More information

AKRON: An Algorithm for Approximating Sparse Kernel Reconstruction

AKRON: An Algorithm for Approximating Sparse Kernel Reconstruction : An Algorithm for Aroximating Sarse Kernel Reconstruction Gregory Ditzler Det. of Electrical and Comuter Engineering The University of Arizona Tucson, AZ 8572 USA ditzler@email.arizona.edu Nidhal Carla

More information

Probability Estimates for Multi-class Classification by Pairwise Coupling

Probability Estimates for Multi-class Classification by Pairwise Coupling Probability Estimates for Multi-class Classification by Pairwise Couling Ting-Fan Wu Chih-Jen Lin Deartment of Comuter Science National Taiwan University Taiei 06, Taiwan Ruby C. Weng Deartment of Statistics

More information

ESTIMATION OF THE RECIPROCAL OF THE MEAN OF THE INVERSE GAUSSIAN DISTRIBUTION WITH PRIOR INFORMATION

ESTIMATION OF THE RECIPROCAL OF THE MEAN OF THE INVERSE GAUSSIAN DISTRIBUTION WITH PRIOR INFORMATION STATISTICA, anno LXVIII, n., 008 ESTIMATION OF THE RECIPROCAL OF THE MEAN OF THE INVERSE GAUSSIAN DISTRIBUTION WITH PRIOR INFORMATION 1. INTRODUCTION The Inverse Gaussian distribution was first introduced

More information

Nonnegative Matrix Factorization with Markov-Chained Bases for Modeling Time-Varying Patterns in Music Spectrograms

Nonnegative Matrix Factorization with Markov-Chained Bases for Modeling Time-Varying Patterns in Music Spectrograms Nonnegative Matrix Factorization with Markov-Chained Bases for Modeling Time-Varying Patterns in Music Spectrograms Masahiro Nakano 1, Jonathan Le Roux 2, Hirokazu Kameoka 2,YuKitano 1, Nobutaka Ono 1,

More information

q-ary Symmetric Channel for Large q

q-ary Symmetric Channel for Large q List-Message Passing Achieves Caacity on the q-ary Symmetric Channel for Large q Fan Zhang and Henry D Pfister Deartment of Electrical and Comuter Engineering, Texas A&M University {fanzhang,hfister}@tamuedu

More information

Recursive Estimation of the Preisach Density function for a Smart Actuator

Recursive Estimation of the Preisach Density function for a Smart Actuator Recursive Estimation of the Preisach Density function for a Smart Actuator Ram V. Iyer Deartment of Mathematics and Statistics, Texas Tech University, Lubbock, TX 7949-142. ABSTRACT The Preisach oerator

More information

Distributed Rule-Based Inference in the Presence of Redundant Information

Distributed Rule-Based Inference in the Presence of Redundant Information istribution Statement : roved for ublic release; distribution is unlimited. istributed Rule-ased Inference in the Presence of Redundant Information June 8, 004 William J. Farrell III Lockheed Martin dvanced

More information

Robustness of classifiers to uniform l p and Gaussian noise Supplementary material

Robustness of classifiers to uniform l p and Gaussian noise Supplementary material Robustness of classifiers to uniform l and Gaussian noise Sulementary material Jean-Yves Franceschi Ecole Normale Suérieure de Lyon LIP UMR 5668 Omar Fawzi Ecole Normale Suérieure de Lyon LIP UMR 5668

More information

Combining Logistic Regression with Kriging for Mapping the Risk of Occurrence of Unexploded Ordnance (UXO)

Combining Logistic Regression with Kriging for Mapping the Risk of Occurrence of Unexploded Ordnance (UXO) Combining Logistic Regression with Kriging for Maing the Risk of Occurrence of Unexloded Ordnance (UXO) H. Saito (), P. Goovaerts (), S. A. McKenna (2) Environmental and Water Resources Engineering, Deartment

More information

Uncorrelated Multilinear Discriminant Analysis with Regularization and Aggregation for Tensor Object Recognition

Uncorrelated Multilinear Discriminant Analysis with Regularization and Aggregation for Tensor Object Recognition TNN-2007-P-0332.R1 1 Uncorrelated Multilinear Discriminant Analysis with Regularization and Aggregation for Tensor Object Recognition Haiing Lu, K.N. Plataniotis and A.N. Venetsanooulos The Edward S. Rogers

More information

Th P10 13 Alternative Misfit Functions for FWI Applied to Surface Waves

Th P10 13 Alternative Misfit Functions for FWI Applied to Surface Waves Th P0 3 Alternative Misfit Functions for FWI Alied to Surface Waves I. Masoni* Total E&P, Joseh Fourier University), R. Brossier Joseh Fourier University), J. Virieux Joseh Fourier University) & J.L. Boelle

More information

PER-PATCH METRIC LEARNING FOR ROBUST IMAGE MATCHING. Sezer Karaoglu, Ivo Everts, Jan C. van Gemert, and Theo Gevers

PER-PATCH METRIC LEARNING FOR ROBUST IMAGE MATCHING. Sezer Karaoglu, Ivo Everts, Jan C. van Gemert, and Theo Gevers PER-PATCH METRIC LEARNING FOR ROBUST IMAGE MATCHING Sezer Karaoglu, Ivo Everts, Jan C. van Gemert, and Theo Gevers Intelligent Systems Lab, Amsterdam, University of Amsterdam, 1098 XH Amsterdam, The Netherlands

More information

DETC2003/DAC AN EFFICIENT ALGORITHM FOR CONSTRUCTING OPTIMAL DESIGN OF COMPUTER EXPERIMENTS

DETC2003/DAC AN EFFICIENT ALGORITHM FOR CONSTRUCTING OPTIMAL DESIGN OF COMPUTER EXPERIMENTS Proceedings of DETC 03 ASME 003 Design Engineering Technical Conferences and Comuters and Information in Engineering Conference Chicago, Illinois USA, Setember -6, 003 DETC003/DAC-48760 AN EFFICIENT ALGORITHM

More information

Compressed Sensing Based Video Multicast

Compressed Sensing Based Video Multicast Comressed Sensing Based Video Multicast Markus B. Schenkel a,b, Chong Luo a, Pascal Frossard b and Feng Wu a a Microsoft Research Asia, Beijing, China b Signal Processing Laboratory (LTS4), EPFL, Lausanne,

More information

GENERATING FUZZY RULES FOR PROTEIN CLASSIFICATION E. G. MANSOORI, M. J. ZOLGHADRI, S. D. KATEBI, H. MOHABATKAR, R. BOOSTANI AND M. H.

GENERATING FUZZY RULES FOR PROTEIN CLASSIFICATION E. G. MANSOORI, M. J. ZOLGHADRI, S. D. KATEBI, H. MOHABATKAR, R. BOOSTANI AND M. H. Iranian Journal of Fuzzy Systems Vol. 5, No. 2, (2008). 21-33 GENERATING FUZZY RULES FOR PROTEIN CLASSIFICATION E. G. MANSOORI, M. J. ZOLGHADRI, S. D. KATEBI, H. MOHABATKAR, R. BOOSTANI AND M. H. SADREDDINI

More information

IMPROVED BOUNDS IN THE SCALED ENFLO TYPE INEQUALITY FOR BANACH SPACES

IMPROVED BOUNDS IN THE SCALED ENFLO TYPE INEQUALITY FOR BANACH SPACES IMPROVED BOUNDS IN THE SCALED ENFLO TYPE INEQUALITY FOR BANACH SPACES OHAD GILADI AND ASSAF NAOR Abstract. It is shown that if (, ) is a Banach sace with Rademacher tye 1 then for every n N there exists

More information

Solved Problems. (a) (b) (c) Figure P4.1 Simple Classification Problems First we draw a line between each set of dark and light data points.

Solved Problems. (a) (b) (c) Figure P4.1 Simple Classification Problems First we draw a line between each set of dark and light data points. Solved Problems Solved Problems P Solve the three simle classification roblems shown in Figure P by drawing a decision boundary Find weight and bias values that result in single-neuron ercetrons with the

More information

General Linear Model Introduction, Classes of Linear models and Estimation

General Linear Model Introduction, Classes of Linear models and Estimation Stat 740 General Linear Model Introduction, Classes of Linear models and Estimation An aim of scientific enquiry: To describe or to discover relationshis among events (variables) in the controlled (laboratory)

More information

CHAPTER-II Control Charts for Fraction Nonconforming using m-of-m Runs Rules

CHAPTER-II Control Charts for Fraction Nonconforming using m-of-m Runs Rules CHAPTER-II Control Charts for Fraction Nonconforming using m-of-m Runs Rules. Introduction: The is widely used in industry to monitor the number of fraction nonconforming units. A nonconforming unit is

More information

Deriving Indicator Direct and Cross Variograms from a Normal Scores Variogram Model (bigaus-full) David F. Machuca Mory and Clayton V.

Deriving Indicator Direct and Cross Variograms from a Normal Scores Variogram Model (bigaus-full) David F. Machuca Mory and Clayton V. Deriving ndicator Direct and Cross Variograms from a Normal Scores Variogram Model (bigaus-full) David F. Machuca Mory and Clayton V. Deutsch Centre for Comutational Geostatistics Deartment of Civil &

More information

Understanding Big Data Spectral Clustering

Understanding Big Data Spectral Clustering Understanding Big Data Sectral Clustering Romain Couillet, Florent Benaych-Georges To cite this version: Romain Couillet, Florent Benaych-Georges. Understanding Big Data Sectral Clustering. IEEE 6th International

More information

MATH 361: NUMBER THEORY EIGHTH LECTURE

MATH 361: NUMBER THEORY EIGHTH LECTURE MATH 361: NUMBER THEORY EIGHTH LECTURE 1. Quadratic Recirocity: Introduction Quadratic recirocity is the first result of modern number theory. Lagrange conjectured it in the late 1700 s, but it was first

More information

MATH 2710: NOTES FOR ANALYSIS

MATH 2710: NOTES FOR ANALYSIS MATH 270: NOTES FOR ANALYSIS The main ideas we will learn from analysis center around the idea of a limit. Limits occurs in several settings. We will start with finite limits of sequences, then cover infinite

More information

PETER J. GRABNER AND ARNOLD KNOPFMACHER

PETER J. GRABNER AND ARNOLD KNOPFMACHER ARITHMETIC AND METRIC PROPERTIES OF -ADIC ENGEL SERIES EXPANSIONS PETER J. GRABNER AND ARNOLD KNOPFMACHER Abstract. We derive a characterization of rational numbers in terms of their unique -adic Engel

More information

An Ant Colony Optimization Approach to the Probabilistic Traveling Salesman Problem

An Ant Colony Optimization Approach to the Probabilistic Traveling Salesman Problem An Ant Colony Otimization Aroach to the Probabilistic Traveling Salesman Problem Leonora Bianchi 1, Luca Maria Gambardella 1, and Marco Dorigo 2 1 IDSIA, Strada Cantonale Galleria 2, CH-6928 Manno, Switzerland

More information

Uncorrelated Multilinear Discriminant Analysis with Regularization and Aggregation for Tensor Object Recognition

Uncorrelated Multilinear Discriminant Analysis with Regularization and Aggregation for Tensor Object Recognition Uncorrelated Multilinear Discriminant Analysis with Regularization and Aggregation for Tensor Object Recognition Haiing Lu, K.N. Plataniotis and A.N. Venetsanooulos The Edward S. Rogers Sr. Deartment of

More information

CONVOLUTIVE NON-NEGATIVE MATRIX FACTORISATION WITH SPARSENESS CONSTRAINT

CONVOLUTIVE NON-NEGATIVE MATRIX FACTORISATION WITH SPARSENESS CONSTRAINT CONOLUTIE NON-NEGATIE MATRIX FACTORISATION WITH SPARSENESS CONSTRAINT Paul D. O Grady Barak A. Pearlmutter Hamilton Institute National University of Ireland, Maynooth Co. Kildare, Ireland. ABSTRACT Discovering

More information

ON THE DEVELOPMENT OF PARAMETER-ROBUST PRECONDITIONERS AND COMMUTATOR ARGUMENTS FOR SOLVING STOKES CONTROL PROBLEMS

ON THE DEVELOPMENT OF PARAMETER-ROBUST PRECONDITIONERS AND COMMUTATOR ARGUMENTS FOR SOLVING STOKES CONTROL PROBLEMS Electronic Transactions on Numerical Analysis. Volume 44,. 53 72, 25. Coyright c 25,. ISSN 68 963. ETNA ON THE DEVELOPMENT OF PARAMETER-ROBUST PRECONDITIONERS AND COMMUTATOR ARGUMENTS FOR SOLVING STOKES

More information

A New Perspective on Learning Linear Separators with Large L q L p Margins

A New Perspective on Learning Linear Separators with Large L q L p Margins A New Persective on Learning Linear Searators with Large L q L Margins Maria-Florina Balcan Georgia Institute of Technology Christoher Berlind Georgia Institute of Technology Abstract We give theoretical

More information

arxiv: v1 [stat.ml] 10 Mar 2016

arxiv: v1 [stat.ml] 10 Mar 2016 Global and Local Uncertainty Princiles for Signals on Grahs Nathanael Perraudin, Benjamin Ricaud, David I Shuman, and Pierre Vandergheynst March, 206 arxiv:603.03030v [stat.ml] 0 Mar 206 Abstract Uncertainty

More information

PARTIAL FACE RECOGNITION: A SPARSE REPRESENTATION-BASED APPROACH. Luoluo Liu, Trac D. Tran, and Sang Peter Chin

PARTIAL FACE RECOGNITION: A SPARSE REPRESENTATION-BASED APPROACH. Luoluo Liu, Trac D. Tran, and Sang Peter Chin PARTIAL FACE RECOGNITION: A SPARSE REPRESENTATION-BASED APPROACH Luoluo Liu, Trac D. Tran, and Sang Peter Chin Det. of Electrical and Comuter Engineering, Johns Hokins Univ., Baltimore, MD 21218, USA {lliu69,trac,schin11}@jhu.edu

More information

Positive Definite Uncertain Homogeneous Matrix Polynomials: Analysis and Application

Positive Definite Uncertain Homogeneous Matrix Polynomials: Analysis and Application BULGARIA ACADEMY OF SCIECES CYBEREICS AD IFORMAIO ECHOLOGIES Volume 9 o 3 Sofia 009 Positive Definite Uncertain Homogeneous Matrix Polynomials: Analysis and Alication Svetoslav Savov Institute of Information

More information

On Code Design for Simultaneous Energy and Information Transfer

On Code Design for Simultaneous Energy and Information Transfer On Code Design for Simultaneous Energy and Information Transfer Anshoo Tandon Electrical and Comuter Engineering National University of Singaore Email: anshoo@nus.edu.sg Mehul Motani Electrical and Comuter

More information

2 K. ENTACHER 2 Generalized Haar function systems In the following we x an arbitrary integer base b 2. For the notations and denitions of generalized

2 K. ENTACHER 2 Generalized Haar function systems In the following we x an arbitrary integer base b 2. For the notations and denitions of generalized BIT 38 :2 (998), 283{292. QUASI-MONTE CARLO METHODS FOR NUMERICAL INTEGRATION OF MULTIVARIATE HAAR SERIES II KARL ENTACHER y Deartment of Mathematics, University of Salzburg, Hellbrunnerstr. 34 A-52 Salzburg,

More information

Metrics Performance Evaluation: Application to Face Recognition

Metrics Performance Evaluation: Application to Face Recognition Metrics Performance Evaluation: Alication to Face Recognition Naser Zaeri, Abeer AlSadeq, and Abdallah Cherri Electrical Engineering Det., Kuwait University, P.O. Box 5969, Safat 6, Kuwait {zaery, abeer,

More information

Unsupervised Hyperspectral Image Analysis Using Independent Component Analysis (ICA)

Unsupervised Hyperspectral Image Analysis Using Independent Component Analysis (ICA) Unsuervised Hyersectral Image Analysis Using Indeendent Comonent Analysis (ICA) Shao-Shan Chiang Chein-I Chang Irving W. Ginsberg Remote Sensing Signal and Image Processing Laboratory Deartment of Comuter

More information

Robust Beamforming via Matrix Completion

Robust Beamforming via Matrix Completion Robust Beamforming via Matrix Comletion Shunqiao Sun and Athina P. Petroulu Deartment of Electrical and Comuter Engineering Rutgers, the State University of Ne Jersey Piscataay, Ne Jersey 8854-858 Email:

More information

2-D Analysis for Iterative Learning Controller for Discrete-Time Systems With Variable Initial Conditions Yong FANG 1, and Tommy W. S.

2-D Analysis for Iterative Learning Controller for Discrete-Time Systems With Variable Initial Conditions Yong FANG 1, and Tommy W. S. -D Analysis for Iterative Learning Controller for Discrete-ime Systems With Variable Initial Conditions Yong FANG, and ommy W. S. Chow Abstract In this aer, an iterative learning controller alying to linear

More information

Some Unitary Space Time Codes From Sphere Packing Theory With Optimal Diversity Product of Code Size

Some Unitary Space Time Codes From Sphere Packing Theory With Optimal Diversity Product of Code Size IEEE TRANSACTIONS ON INFORMATION THEORY, VOL. 5, NO., DECEMBER 4 336 Some Unitary Sace Time Codes From Shere Packing Theory With Otimal Diversity Product of Code Size Haiquan Wang, Genyuan Wang, and Xiang-Gen

More information

Solution sheet ξi ξ < ξ i+1 0 otherwise ξ ξ i N i,p 1 (ξ) + where 0 0

Solution sheet ξi ξ < ξ i+1 0 otherwise ξ ξ i N i,p 1 (ξ) + where 0 0 Advanced Finite Elements MA5337 - WS7/8 Solution sheet This exercise sheets deals with B-slines and NURBS, which are the basis of isogeometric analysis as they will later relace the olynomial ansatz-functions

More information

Coding Along Hermite Polynomials for Gaussian Noise Channels

Coding Along Hermite Polynomials for Gaussian Noise Channels Coding Along Hermite olynomials for Gaussian Noise Channels Emmanuel A. Abbe IG, EFL Lausanne, 1015 CH Email: emmanuel.abbe@efl.ch Lizhong Zheng LIDS, MIT Cambridge, MA 0139 Email: lizhong@mit.edu Abstract

More information

State Estimation with ARMarkov Models

State Estimation with ARMarkov Models Deartment of Mechanical and Aerosace Engineering Technical Reort No. 3046, October 1998. Princeton University, Princeton, NJ. State Estimation with ARMarkov Models Ryoung K. Lim 1 Columbia University,

More information

where x i is the ith coordinate of x R N. 1. Show that the following upper bound holds for the growth function of H:

where x i is the ith coordinate of x R N. 1. Show that the following upper bound holds for the growth function of H: Mehryar Mohri Foundations of Machine Learning Courant Institute of Mathematical Sciences Homework assignment 2 October 25, 2017 Due: November 08, 2017 A. Growth function Growth function of stum functions.

More information

Feedback-error control

Feedback-error control Chater 4 Feedback-error control 4.1 Introduction This chater exlains the feedback-error (FBE) control scheme originally described by Kawato [, 87, 8]. FBE is a widely used neural network based controller

More information

Introduction to Probability and Statistics

Introduction to Probability and Statistics Introduction to Probability and Statistics Chater 8 Ammar M. Sarhan, asarhan@mathstat.dal.ca Deartment of Mathematics and Statistics, Dalhousie University Fall Semester 28 Chater 8 Tests of Hyotheses Based

More information

Provided by the author(s) and University College Dublin Library in accordance with ublisher olicies. Please cite the ublished version when available. Title Low Comlexity Stochastic Otimization-Based Model

More information

Fig. 4. Example of Predicted Workers. Fig. 3. A Framework for Tackling the MQA Problem.

Fig. 4. Example of Predicted Workers. Fig. 3. A Framework for Tackling the MQA Problem. 217 IEEE 33rd International Conference on Data Engineering Prediction-Based Task Assignment in Satial Crowdsourcing Peng Cheng #, Xiang Lian, Lei Chen #, Cyrus Shahabi # Hong Kong University of Science

More information

ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM

ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM ANALYTIC NUMBER THEORY AND DIRICHLET S THEOREM JOHN BINDER Abstract. In this aer, we rove Dirichlet s theorem that, given any air h, k with h, k) =, there are infinitely many rime numbers congruent to

More information

On the Chvatál-Complexity of Knapsack Problems

On the Chvatál-Complexity of Knapsack Problems R u t c o r Research R e o r t On the Chvatál-Comlexity of Knasack Problems Gergely Kovács a Béla Vizvári b RRR 5-08, October 008 RUTCOR Rutgers Center for Oerations Research Rutgers University 640 Bartholomew

More information

Proximal methods for the latent group lasso penalty

Proximal methods for the latent group lasso penalty Proximal methods for the latent grou lasso enalty The MIT Faculty has made this article oenly available. Please share how this access benefits you. Your story matters. Citation As Published Publisher Villa,

More information

MULTI-CHANNEL PARAMETRIC ESTIMATOR FAST BLOCK MATRIX INVERSES

MULTI-CHANNEL PARAMETRIC ESTIMATOR FAST BLOCK MATRIX INVERSES MULTI-CANNEL ARAMETRIC ESTIMATOR FAST BLOCK MATRIX INVERSES S Lawrence Marle Jr School of Electrical Engineering and Comuter Science Oregon State University Corvallis, OR 97331 Marle@eecsoregonstateedu

More information

Spectral Clustering based on the graph p-laplacian

Spectral Clustering based on the graph p-laplacian Sectral Clustering based on the grah -Lalacian Thomas Bühler tb@cs.uni-sb.de Matthias Hein hein@cs.uni-sb.de Saarland University Comuter Science Deartment Camus E 663 Saarbrücken Germany Abstract We resent

More information

Tensor-Based Sparsity Order Estimation for Big Data Applications

Tensor-Based Sparsity Order Estimation for Big Data Applications Tensor-Based Sarsity Order Estimation for Big Data Alications Kefei Liu, Florian Roemer, João Paulo C. L. da Costa, Jie Xiong, Yi-Sheng Yan, Wen-Qin Wang and Giovanni Del Galdo Indiana University School

More information