A Review of Independent Component Analysis Techniques
|
|
- Cynthia Booth
- 5 years ago
- Views:
Transcription
1 A Revie of Independent Component Analysis Techniques Bogdan Matei Electrical and Computer Engineering Department Rutgers University, Piscataay, NJ, , USA 1 Introduction The Blind Signal Separation (BSS) consists in the recovery of unknon independent signals from observed linear combinations of them. To set up the notations, let be the unknon signals and the "! matrix # signal $ % & ' ' ')( to * be the unknon mixing matrix relating the observed $ #,+ (1) here the explicit time-dependence on time as dropped. Without loss of generality it is assumed that the source signals are zero mean -.0/, 21. The model (1) can be modified in order to accommodate the presence of additive Gaussian noise $ #4365 (2) An illustration of a BSS application is the cocktail party problem in hich the speech of several people is received from several microphones present in the room. The task is to recover the speech of the individual speakers from the overlapped talk 14]. The column 7 / of the matrix # is called directional vector associated to the 8 -th source /. When 7 / are knon the spatial filters 9 / can be estimated using the minimum variance distortionless response (MVDR) 9 /: <;>=? 7 /0+ ;? - $ $"@ A+ (3) or the linear constrained minimum variance estimator (LCMV) in the presence of zero-mean hite noise ith noise variance B and independent sources 4] 9 /: C A;?ED B GF IH 7 /J (4)
2 the pseudo-inverse. Blind deconvolution (BD) is a particular case of BSS in hich the mixing matrix # is assumed to have a Toeplitz triangular structure. When # is not triangular the filter is non-causal and hen Throughout the paper K denotes the hermitian operator and L is not Toeplitz the filter is non-stationary 9]. In many applications, hoever, either the mixing matrix # is not available, or the errors in # due to multipath propagation may render the estimates 9 / unusable. In these situations it is better not to make any assumptions about # other than the Toeplitz/triangular structure for BD. Independent component analysis (ICA) is a class of techniques of performing BSS 14], though many references use ICA and BSS interchangeably 22, 6, 9]. We ill also consider that ICA and BSS are synonymous. There are to distinct approaches toards computing the ICA. One employs high order cumulants and is found mainly in the statistical signal processing literature 9, 6, 22] and the other uses the gradient-descent of non-linear activation functions in neuron-like devices and is mainly developed in the neural netorks community 1, 18, 19]. Each of the above approaches has advantages and shortcomings: the computation of high order cumulants is very sensitive to outliers and lack of sufficient support in the data especially for signals having a long-tailed probability density function (p.d.f.), hile the neural-netorks algorithms may become unstable, converge sloly and most often require some extra knoledge about the p.d.f. of the source signals in order to choose the non-linearities in the neurons. There are both batch and adaptive versions of ICA algorithms depending on the particular application. The ICA requires that the number of received signals! be at least equal to the number of independent sources. A close examination of the model (1) reveals the folloing ambiguities 1. The variances - MN/*M can not be determined since $ # diag POQ+R+ OST U V diag AO = +W+ O = G0X #ZY Y J (5) To fix this scale indeterminacy it is assumed that - M / M found only up to a sign., \, hoever 7 / and / can be 2. The order of the components can not be specified. If ] is a permutation matrix, i.e., it has a single 1 on every ro and column, then similarly to (5) $ ^ #_] ` ] = 2 ) " #ZY Y J (6)
3 B o The blind decomposition into # as it ill be shon in Section 2. and can be found only if at most one of the sources / is Gaussian, The ICA techniques ere applied in antenna array processing for the estimation of radiating sources 11], in the separation of the evoked potentials from EEG 19], in the recovery of fetal EKG 8], in the separation of reflections from images 12], etc. In Section 2 the differences beteen the Principal Component Analyis (PCA) and ICA are investigated and it shon that ICA is a generalization of PCA. The cumulant based approach for doing ICA is introduced in Section 4. In Section 5 several gradient based techniques for ICA are presented. Section 6 contains simulations using real data extracted from images. 2 Principal Component Analysis vs. Independent Component Analysis Principal Component Analysis is a standard statistical tool used to find the orthogonal directions corresponding to the highest variance. It is equivalent to a decorrelation of the data using secondorder information. If a? - is the covariance of $, (since E($ ) = b ), let the eigenvalue decomposition of ;? be and assume that the elements of the diagonal matrix d ;? (7) are sorted in decreasing order. In the absence of noise and assuming!hgi, the covariance matrix ;? has rank equal to, therefore the last! D eigenvalues jlk C1)+GmW 3in++!. When independent and identically distributed (i.i.d.) noise ith zero-mean and noise variance B jlk, m" 3in++!. Define the hitening transform o diag U j D B and let the hitened data vector be affect the measurements, the smallest eigenvalues satisfy U=qprs+t j D B o It can be easily seen that the covariance matrix of U=uprs++ j D B G=qp rv+ 1)++ 1s (8) $ (9) is ;Zx " - o diag U0++n+ 1)++ (10) 3
4 F B o ( F In practice ;{z must be estimated from the available data. If available An estimate of the noise variance and When! o diag G j D B samples of $, E }0+l+ are ;? + ~ $ U + $ A s +l+ $ AJ (11) B is given by! D G=qp rv+w j D B k " * j k +!h ˆ (12) G=qp rv++ j D B U=uprs+`1)+l+`1y $ (13), the noise variance can not be estimated and ICA proceeds similarly to the noiseless situation. A degradation in the performance of ICA algorithms is therefore expected depending on the signal to noise (SNR) ratio. Since the last! D components of are identically zero, e ill consider in the folloing that e ork only in the signal subspace and that is -dimensional. The hitening transform o can be determined only up to an orthogonal transform. Indeed, since the covariance of q ;ŒR $"$"@` - (14) The ICA attempts to go further one step then PCA, by finding the orthogonal matrix hich transforms the hitened data into Š having `+ ++ statistically independent. The difference beteen PCA and ICA is illustrated in Figure 1. The independent signals Ž, % 0++1n101 having uniform, respectively Gaussian distribution are transformed into $ using the mixing matrix The PCA hitens the data into " # \ 1)J 0 0 n 1)J 0 n 0 1)J 0 ) 1)J 0 0 (15) $ as shon in Figure 1(e),(f). The hitened uniform data is distributed inside a square, hoever the components are still not independent. The ICA techniques allo the determination of the orthogonal matrix hich aligns the sides of the square ith the coordinate axes. The independent components n/ are recovered up to a scale, sign and order. On the other hand the hitened normal data has circular symmetry, therefore the matrix can not be found. The ICA is thus more general than PCA in trying not only to decorrelate the data, but also to find a decomposition transforming the input into independent components. 4
5 0.4 Uniform 2.5 Gaussian (a) (b) (c) (d) (e) (f) (g) (h) Figure 1: Whitening uniform (left column) and Gaussian data (right column). (a), (b) original data; (c), (d) rotated data using c ; (e), (f) scaled data; (g) The independent components are recovered by finding the corresponding matrix in the case of uniformly distributed data hich aligns the axis of the square ith the ' and axes; (h) the orthogonal matrix can not be found for Gaussian data since the hitened data has circular symmetry. 5
6 o The transform e seek is therefore If an orthogonal matrix Š " še$ š # * % œ Ž (16) can be found hich transforms the mixed signals $ into Š ith independent components, then assuming that at most one independent source / is normally distributed, then Š 2œ_ ith œ being a non-mixing matrix (i.e. it has exactly one non-zero component on each ro and column) 9]. The ICA algorithms attempt to find the matrix š Section 3 several criteria performing ICA are presented. 3 Contrast Functions hich ensures that Š are independent. In Contrast functions (or contrasts) ž serve as objective criteria for ICA. They are real functions of a probability distribution and are designed such that 6, 9] ž Š Ÿ ž š $ Ÿ ž œ_n gž n (17) Most often contrast functions are derived using concepts from information theory like the entropy and the Kullback distance beteen to probability densities 10]. The entropy of a random variable having a p.d.f. is defined as S De 0 0 The Kullback distance beteen the probability densities and is M 0 0 j (19) The Kullback distance doesn t satisfy the distance axioms, since M «ª M, hoever, M g 1 ith equality if and only if. Assuming that the p.d.f. of the independent signals / are knon e have R v " / j (18) R A/ (20) The Maximum Likelihood criterion seeks to find the matrix š such that the distribution of Š še$ is closest to the distribution of. A suitable contrast is then yielded by the Kullback distance beteen these to distributions ž Š Š M.* U T± Š M²n (21) 6
7 F Bell and Sejnoski proposed in 1] the infomax contrast function žÿ³a in the context of a neural netork approach. Cardoso proved that žs³a is in fact identical to ž :µ 7]. When no assumption about the distribution of the sources can be made, the minimization of Š M²s should be done not only ith respect to š but also ith respect to the distribution of. Let Š Y be a random vector hose components / Y are independent and have a p.d.f. equal to the marginal densities of /. Then the Kullback distance beteen Š š and is 6] Š M 0 Š M Š Y *3 Š Y M²n (22) Since Š Y M²n is minimum (zero) hen the distribution of is equal to the distribution of Š Y, the minimization of Š M²n is equivalent to the minimization of Š M Š Y, hich is nothing else then the mutual information beteen Š and Š Y. The mutual information contrast function ž ³ as proposed by Comon 9] and measures the deviation from independence of a random vector ž ³ Š Ÿ Š M Š Y SJ (23) From the definition of the Kullback distance (19), it follos that ž¹ ³ Š g 1 ith equality iff the Š Š Š / J entries of are independent. Since is hite /v3 const (24) and ž ³ Š can be brought to a simpler form, the minimum marginal entropy 2] ž º ³ Š / /Ÿ+ (25) here the» superscript denotes the fact that the minimization of the contrast function ž º ³ Š is done under the hiteness constraint - S. The minimum marginal entropy (25) is related to the negentropy maximization criterion 16] ¼ Š¹½ D Š (26) hich measures the difference in entropy beteen a normally distributed vector Š¹½, having the same covariance matrix as Š, and the vector Š. The normal distribution has the largest entropy among all the distributions ith the same first to moments 10], therefore the negentropy is a strictly positive quantity. From (24) and using the constancy of Š*½ the maximization of the negentropy (26) is equivalent to (25). The contrast function ž¹ ³ is the most general in the sense that it makes no assumption about the distribution of the independent sources. On the other hand hen some knoledge about the p.d.f. of the sources is available ž :µ should be used 2, 6]. 7
8 k É Ã k k à / 4 Cumulant Based Approaches The contrast functions (21, 23, 25) can be expressed using high order statistics. The most convenient ay of expressing higher orders statistics is through cumulants 21] because: 1. The computations are simpler ith cumulants than ith moments 2. The cumulants of a sum are the sum of the cumulants 3. The Edgeorth approximations of a distribution can be conveniently expressed using cumulants 4. The higher order cumulants of a normal vector are zero, contrary to the moments The moments and cumulants can be obtained using moment, respectively cumulant generating functions. The moment generating function of a random vector, ²Q+ {+"t is ¾\ ÀS " - Á' * P (27) and the cumulant generating function is defined as  ÀS " f n S ¾\ (28) The cumulant generating function has a Taylor expansion  ÀS \ )Æ klå k 3 keãnä kèçéaç / à væ É Å kèç É 3i (29) kèç ÉTà ÃnÄ / Å kèçéaç /W3 Ã Ä *Æ kêçéaç / É ËÅ kèçéaç / Ç Ë 3i Ç ËGÃ Ã Ä Ã Ä The cumulants Å are obtained from (30) by identifying the corresponding terms. For example the order to cumulants are Å kèç É, the order three cumulants are Å kêç ÉAÇ /, etc. Another ay of defining the cumulants is presented belo 2] in terms of the centered random vectors ÌÍ D - 0. The first order cumulant is hile the second order cumulants are Å k - ² k Ÿ+ (30) Å kèçé - Ì k Ì É Ä SJ (31) 8
9 V Å V Å V Å X k 3 3 V Å 3 Å Å V Å V V Å Å X Thus the second order cumulants are given by the components of the covariance matrix. Similarly the third and fourth order cumulants also called skeness, respectively kurtosis are Å kèçéaç /4 For a normally distributed vector function is - Ì k Ì É Ì / Ä Å kèçépç / Ç Ë - Ì k Ì É / Ì Ì Ä Ë (32) Ä having mean Î x and covariance œzx, the cumulant generating  xs À¹ " Î x 3 therefore the cumulants ith order greater than to vanish. 4.1 Contrast Function Approximation Using Cumulants œzxqà (33) The approximations of the contrast functions introduced in Section 3 in terms of up to order four cumulants, assuming that distributions having odd moments equal to zero are obtained in terms of Edgeorth expansions 2, 16]. Assuming that Ï and Ð are to random vectors hose p.d.f. function Ï, respectively Ð are close to a normal distribution and have odd moments zero, the Kullback distance beteen the to distributions are given in terms of second and fourth order cumulants by Ï M Ð SÑ kèçé Å)ÒkÈÇ É D Å)ÓkÈÇ É s kèçépç / Ç Ë ÅŽÒkÊÇÉAÇ / Ç Ë D Å)ÓkÈÇÉAÇ / Ç Ë Using the approximation (34) and the independence of the components / corresponding to the sources it follos that ž¹ µ Š Ñ kêçé kêç É D kèçé`ô k É X s kèçépç / Ç Ë kêçéaç / Ç Ë D kèçéaç / Ç ËÕÔ k É / Ë X here the Kronecker product Ô is one only hen all the indexes are identical. Taking into account that after hitening e have Å ž º :µ kèçé Š Ö Ô k É after some algebra (35) can be further simplified kèçéaç / Ç Ë D kèçéaç / Ç Ë Ô k É / Ë X D l kêçéaç / Ç Ë kêçkèçkèçk - M k Ì M Ø )3 const (34) (35) (36) The mutual information contrast ž ³ is approximated by 2] ž¹ ³ Š Ñ kèç ÉÙ ykèç k kèçé X s kèçépç / Ç ËÊÙ ykèçkèçkèçk kèçépç / Ç Ë X (37) 9
10 Ä F é V Å Ä ä V Å é é Ä k D Ä V Å and the minimum marginal entropy contrast ž º Ú³ ž º ³ Š SÑ s kèçéaç / kèçépç / Ç Ë X Ç ËÈÙ skèçkèçkèçk D kèçkèçkêçk X (38) A contrast function similar to (38) hich can be minimized using the joint diagonalization of matrices 3] as proposed by Cardoso 4] žsûqü)ý Þ Š S kèçépç / Ç ËÊÙ ykèçkèç / Ç Ë kèçéaç / Ç Ë X The JADE contrast function žsûqü)ý Þ and the orthogonal maximum likelihood function ž º µ ill be used in the folloing in performing ICA. 4.2 Computation of Fourth Order Cumulants In this section several computational algorithms for determining the fourth order cumulants Å of hitened data Š Ul `++ Š. Define the ß matrix à } Š Gl `+ Š A s ++ Š A. A convenient ay of storing Å kèçépç / Ç Ë is in a square matrix á 4, 22]. The Kronecker product 13] beteen the \Œ! matrix d and _ matrix â is defined as the Í _Œ!ã matrix. â ( ë ì â ì dåä â ßæç çè é... ê â ( â The vectorization of the matrix à Define the and matrices Note that for hitened data ;ò expressed as (39) kèçépç / Ç Ë í (40) is obtained by stacking its columns one belo the other vec à " î Š Gl ï Š q ð à ß î Š Ul ä Š U + Š A s Š A s +ñ Š ä Š ð à ß î Š Ul ä Š Gl `+ Š A y ä Š P y `+ñ Š ä Š ð á ;ò ;{ô" à à (41) (42). Similarly to 22] the fourth order cumulant matrix á can be à D vec a vec D a ä a 10 ;Œô Js; + (43)
11 ö = = ( Å þ F = here J has the same significance ith MATLAB. A different approach for the computation of the fourth moments using cumulant matrices is presented in 4, 2]. Given a ß_ matrix õ and the hitened random vectors Š, the cumulant õ has the mè -th component matrix ö ö õ A k É A complete set of cumulant matrices is given by úsö / Ç Ë Å kêçéaç / Ç Ë ø / Ë (44) - õv D tr õ D õ D ð õ\k ûmm ü ý here õÿþ + n++ (45) and kèçéaçþ Ç î ö ð The Ø fourth order cumulants are therefore determined by computing the using the definition (44). Since ö Another choice is given by the base õþ ö þ@ + if = 3 þ@ + if < = + if { < it follos that ö õ2þ k É (46) cumulant matrices + C0++ (47), 21 for Ö, only 3l cumulant matrices need to be computed 2]. A maximal base containing only cumulant matrices is obtained by performing an eigenvalue decomposition of the matrix á computed in (43) á ^c diag j + j ++ j r + j <j ij r (48) ith here the Œ õk jlkêk m (49) matrix k is obtained by reshaping the Wm -th column of c. 11
12 k Ä o $ 4.3 JADE Algorithm In this section e sho that the contrast function žÿûüvý Þ, hich is an approximation to the minimum marginal entropy contrast (38) can be minimized using the joint diagonalization of the cumulant matrices ö õ k. The orthogonal matrix introduced in Section 2 is found as the minimizer of 4] žsûüvý Þ õòkp (50) here õ\k is a maximal set of cumulant matrices and off # denotes the off diagonal elements of the matrix #. The matrix can be found be a Jacobi technique using Givens rotations 3] as here each of the Givens rotations ; k k k É É k ÉÉ ;Œ ;_ `;{/ (51) is an identity matrix ith the exception of D + M M 3^MN*M & (52) At each step and are found from the minimization of (51). See 3] for the proof and their expression. The JADE algorithm is summarized next. 1. Compute the hitening transform o using (13) 2. Compute the maximal set of cumulant matrices ö õ using (44) or (49) 3. Find the orthogonal matrix hich minimizes (50) 4. Estimate the mixing matrix as #! o H 5. Estimate the independent sources 4.4 MaxKurt/MinKurt Algorithm When the kurtoses of the sources are knon the contrast function ž º µ (37) should be minimized. Under the assumption of equal kurtosis ith knon signs (37) should be maximized for positive kurtosis (a random variable ith positive kurtosis is also knon as super-gaussian or super-kurtic). On the other hand for variables having negative kurtoses (for example the uniform distribution) (37) should be maximized. 12
13 š F F F F F The MaxKurt/ MinKurt algorithm minimizes / maximizes (37) using also the joint diagonalization of matrices. This approach has the advantage of operating directly on the data à, ithout the need of computing explicitly the cumulant matrices ö and yields closed-form solutions for the Givens angles and (52). The disadvantage is that ith bad assumptions about the sign of the kurtosis, the ICA may fail as illustrated in Figure 2 A similar approach as used by Farid and Adelson 12] to separate the reflections and the lighting from images, hoever they have assumed only sources ith positive kurtosis. 5 Gradient Techniques for ICA The approaches presented in Section 4 are batch oriented and can be used only off-line, after the hole data is acquired. In this Section several gradient based techniques are described hich can be used in both on-line and off-line applications. Most of the adaptive approaches ere developed by researchers from the neural netorks community 1, 19, 20, 16]. Assuming that the input data at moment is $ and the separation matrix is š of ICA is, see also (16), the output Š š $ (53) The update rule is of the form 5] š 3<l " D#" %$ U A š (54) here $ñ U is a vector to matrix function dependent on the specific algorithm used and the learning factor " can be constant or time dependent. 5.1 The EASI Adaptive Algorithm An equivariant adaptive separation via independence (EASI) algorithm as proposed in 5]. Based on the notion of relative gradient the folloing update rule can be derived š 3<l " ÖV D#" tî Š D 3'&t Š G D Š (&% Š ð X š (55) ith Š defined in (53). A more stable rule is the normalized version of (55) 3<l W *) D#" Š D &W Š U D Š +&" Š «3 " Š W3 " M Š &W Š U M -, š (56) 13
14 Uniform Exponential ( (a) (b) (c) (d) (e) (f) Figure 2: MaxKurt/MinKurt fails hen the kurtosis signs are rong. Calling the function ith positive sign for the uniform data, or alternatively ith negative sign for the exponential (hich is super-gaussian) causes ICA to find directions rotated ith m from the directions corresponding to independence. (a), (b) The mixed sources; (c), (d) The mixed sources after hitening; (e), (f) The output of ICA using MaxKurt. 14
15 F o F 8 The nonlinearity & can be chosen as &t Š G Š Š (57) though other expressions may be employed. Several observations should be made at this point: (i) The data as assumed centered; (ii) the. hitening is adaptively performed by the term Š D 5.2 Neural Approaches Based on Information Maximization Bell and Sejnoski 1] introduced a class of neural netorks based on information maximization. As mentioned in Section 4 the information maximization criterion is in fact equivalent to the maximum likelihood. The learning rule proposed for super-kurtic signals is š q3i % V 3 " î š D 0/214365q Š U ðgx š (58) See 1] for other learning rules for sub-kurtic signal. The algorithm is very simple, hoever it is rather slo as pointed out in 16] and requires the hitening of $. 5.3 The FastICA Algorithm The FastICA algorithm as proposed in 15] as a fast, batch type algorithm for performing ICA. Assuming hite data the orthogonal matrix (16) the FastICA can be summarized as follos 15, 14]. 1. Center the data, by subtracting the mean 2. Whiten the data using " $ $ D $ Ì $ Ì 3. Choose a random orthogonal matrix. $, ith o defined in (13). 7 $ (59) 4. Compute Š (60) 15
16 8 P 8 7 Y F F 5. Update as ith 6. Orthogonalize :9 3; diag D=< 3?>Z + (61) k 7 k BAŸ k G (62) ; diag ) k DDC A k U, U= (64) 7. If not finished goto Step 4 The non-linearity A can be chosen as 14] AŸ A'S " '6E AŸ P'ã F/1G365q A'S AŸ A'S 'IHKJML DON (65) The FastICA can be applied ithout modification to super and sub kurtic sources and has very fast convergence, since is a Neton type algorithm. In the Section 6 e ill sho some results using the FastIca implementation available for donload from 6 Experiments The ability of several idely used ICA algorithms to recover the independent components from unknon mixtures is analyzed next. Simulations ith real data extracted from images ere performed using Jade, MaxKurt and FastICA algorithms. The pixels from several images play the role of the one-dimensional independent components. For instance assume that and are to images having the same dimensions. Then the independent data is P î vec F vec F ð (66) A random matrix ith positive entries plays the role of the mixing matrix # Then, the mixed signals are obtained as ~ # in the ICA model. (67) 16
17 P The data ~ is analyzed subsequently ith ICA. In Figure 3 the independent and mixed signals obtained from to independent images containing human faces are displayed. Since there is a scale uncertainty, e have scaled the ~ in order to cover all 256 gray levels 12]. The sign uncertainty in the recovered signals! as solved by using the a priori information about the positive elements of the mixing matrix #. The Jade and MaxKurt algorithms ere implemented by the author using the guidelines from the sources available from cardoso/guidesepsou.html. Cardoso s eb site contains also numerous links to papers and people involved in research in ICA. The FastICA algorithm developed by the Independent Component Analysis research group from Helsinki University of Technology as also tested. The FastICA used the nonlinearity AŸ A'S " i' E (see Section 5.3). The EASI adaptive algorithm as also tested, hoever the convergence on the real data used as extremely slo and the performance rather poor. The results of performing ICA on real data are presented in Figures We have noticed that both Jade and FastICA have very close results, ith FastICA being usually somehat faster. MaxKurt also gave good results hen the exact sign of the kurtosis of the sources is knon. A histogram in the images can reveal if the pdf has long tails or short tail and this visual information can help in taking a decision about the sign of the kurtosis. The importance of the correct sign can be observed from Figure 6. An interesting conclusion is illustrated in Figure 9. When the data doesn t cooperate all three algorithms failed to recover perfectly the independent signal. For good data all three give fairly good estimates. An experiment containing three independent sources is presented in Figure 11. The Maxkurt failed to recover the original images for both positive and negative signs for the kurtosis. It can be seen that Jade does a slightly better job of recovering the independent components. 7 Conclusion We have investigated several idely used approaches for performing ICA. We have recovered successfully the original images from mixtures of to and three images. 17
18 (a) (b) (c) (d) Figure 3: Images containing human faces. (a), (b) originals; (c), (d) mixed images. 18
19 (a) (b) (c) (d) Figure 4: The result of ICA on the images from Figure 3. MaxKurt/MinKurt assuming positive kurtosis; (e), (f) FastICA (e) (f) (a), (b) Jade Algorithm; (c), (d) 19
20 (a) (b) (c) (d) Figure 5: Natural Images. (a), (b) originals; (c), (d) mixed 20
21 (a) (b) (c) (d) (e) (f) Figure 6: The result of ICA on the images from Figure 5. (a), (b) Jade Algorithm; (c), (d) MaxKurt/MinKurt assuming positive kurtosis; (e), (f) MaxKurt/MinKurt assuming negative kurtosis. 21
22 (a) (b) (c) (d) Figure 7: Interion images (PUMA set). (a), (b) originals; (c), (d) mixed images. 22
23 (a) (b) (c) (d) (e) (f) Figure 8: The result of ICA on the images from Figure 7. (a), (b) Jade Algorithm; (c), (d) MaxKurt/MinKurt assuming positive kurtosis; (e), (f) FastICA. 23
24 (a) (b) (c) (d) Figure 9: Baboon and Human images. (a), (b) originals; (c), (d) mixed images. 24
25 (a) (b) (c) (d) (e) (f) Figure 10: The result of ICA on the images from Figure 9. (a), (b) Jade Algorithm; (c), (d) MaxKurt/MinKurt assuming positive kurtosis; (e), (f) FastICA algorithm 25
26 (a) (b) (c) Figure 11: The result of ICA using 3 overlapped images. (a) the mixed images; (b) Output of the Jade algorithm; (c) Output of FastICA algorithm. 26
27 References 1] Anthony J. Bell and Terrence J. Sejnoski, An Information-Maximization Approach to Blind Separation and Blind Deconvolution, Neural Computation, Vol. 7, No. 6, pp , ] Jean-François Cardoso, High-order contrast for independent analysis, Neural Computation, Vol. 11, No. 1, pp , ] Jean-François Cardoso and Antoine Souloumiac, Jacobi angles for simultaneous diagonalization, SIAM Journal of Matrix Analysis and Applications, Vol. 17, No. 1, ] Jean-François Cardoso and Antoine Souloumiac, Blind beamforming for non Gaussian signals, IEE Proceedings-F, Vol. 140(6), pp , ] Jean-François Cardoso and Beate Laheld, Equivariant adaptive source separation, IEEE Trans. on Signal Processing, Vol. 44, No. 12, pp , ] Jean-François Cardoso, Blind signal separation: statistical principles, Proceedings of the IEEE, Special issue on blind identification and estimation, R.-W. Liu andl. Tong editors. Vol. 90, No. 8, pp , ] Jean-François Cardoso, Infomax and maximum likelihood for source separation, IEEE Letters on Signal Processing, Vol. 4, No. 4, pp , ] Jean-François Cardoso, Multidimensional independent component analysis, IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP 98), Seattle, ] Pierre Comon, Independent component analysis. A ne concept?, Signal Processing, Special issue on High-Order Statistics, Vol. 36, No. 3, pp , ] T. M. Cover and J. A. Thomas, Elements of information theory, Wiley series in telecommunications, John Wiley, ] G. Desodt and D. Muller, Complex Independent Component Analysis applied to the Separation of Radar Signals, Proc. EUSIPCO, Barcelona, pp ,
28 12] Hany Farid and Edard H. Adelson, Separating Reflections from Images Using Independent Component Analysis, Journal of the Optical Society of America, Vol. 16, No. 9, pp , ] Alexander Graham, Kronecker Products and Matrix Calculus ith Applications, Ellis Horood series in mathematics and its applications, ] Aapo Hyvärinen and Erkki Oja, Independent Component Analysis: A Tutorial, Available at 15] Aapo Hyvärinen, A fast fixed-point algorithm for independent component analysis, Neural Computation, Vol. 9, No. 7, pp , ] Aapo Hyvärinen, Survey on Independent Component Analysis, Neural Computing Surveys Vol. 2, pp , ] M.C. Jones and Robin Sibson, What is Projection Pursuit, Journal of the Royal Statistical Society A, Vol. 150, Part 1, pp. 1 36, ] Juha Karhunen, Neural Approaches to Independent Component Analysis and Source Separation, In Proc. 4th European Symposium on Artificial Neural Netorks (ESANN 96), Bruges, Belgium, April 24-26, pp , ] Juha Karhunen et al., Applications of Neural Blind Separation to Signal and Image Processing, Proc. IEEE 1997 Int. Conf. on Acoustics, Speech, and Signal Processing (ICASSP 97), April 21-24, Munich, Germany, pp , ] Petteri Pajunen and Juha Karhunen, Least-Squares Methods for Blind Source Separation Based on Nonlinear PCA, International Journal of Neural Systems, Vol. 8, No. 5 and 6, pp , ] Peter McCullagh, Tensor Methods in Statistics, Monographs on Statistics and Applied Probability, Chapman and Hall, ] Jacob Sheinvald, On Blind Beamforming for Multiple Non-Gaussian Signals and the Constant-Modulus Algorithm, IEEE Transactions on Signal Processing, Vol. 46, No. 7, pp ,
Independent Component Analysis. Contents
Contents Preface xvii 1 Introduction 1 1.1 Linear representation of multivariate data 1 1.1.1 The general statistical setting 1 1.1.2 Dimension reduction methods 2 1.1.3 Independence as a guiding principle
More informationAn Improved Cumulant Based Method for Independent Component Analysis
An Improved Cumulant Based Method for Independent Component Analysis Tobias Blaschke and Laurenz Wiskott Institute for Theoretical Biology Humboldt University Berlin Invalidenstraße 43 D - 0 5 Berlin Germany
More informationAdaptive Noise Cancellation
Adaptive Noise Cancellation P. Comon and V. Zarzoso January 5, 2010 1 Introduction In numerous application areas, including biomedical engineering, radar, sonar and digital communications, the goal is
More informationIntroduction to Independent Component Analysis. Jingmei Lu and Xixi Lu. Abstract
Final Project 2//25 Introduction to Independent Component Analysis Abstract Independent Component Analysis (ICA) can be used to solve blind signal separation problem. In this article, we introduce definition
More informationFundamentals of Principal Component Analysis (PCA), Independent Component Analysis (ICA), and Independent Vector Analysis (IVA)
Fundamentals of Principal Component Analysis (PCA),, and Independent Vector Analysis (IVA) Dr Mohsen Naqvi Lecturer in Signal and Information Processing, School of Electrical and Electronic Engineering,
More information1 Introduction Independent component analysis (ICA) [10] is a statistical technique whose main applications are blind source separation, blind deconvo
The Fixed-Point Algorithm and Maximum Likelihood Estimation for Independent Component Analysis Aapo Hyvarinen Helsinki University of Technology Laboratory of Computer and Information Science P.O.Box 5400,
More informationIndependent Component Analysis and Its Applications. By Qing Xue, 10/15/2004
Independent Component Analysis and Its Applications By Qing Xue, 10/15/2004 Outline Motivation of ICA Applications of ICA Principles of ICA estimation Algorithms for ICA Extensions of basic ICA framework
More informationPROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata
' / PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE Noboru Murata Waseda University Department of Electrical Electronics and Computer Engineering 3--
More informationIndependent Component Analysis
1 Independent Component Analysis Background paper: http://www-stat.stanford.edu/ hastie/papers/ica.pdf 2 ICA Problem X = AS where X is a random p-vector representing multivariate input measurements. S
More informationMULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES
MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES S. Visuri 1 H. Oja V. Koivunen 1 1 Signal Processing Lab. Dept. of Statistics Tampere Univ. of Technology University of Jyväskylä P.O.
More informationBlind separation of instantaneous mixtures of dependent sources
Blind separation of instantaneous mixtures of dependent sources Marc Castella and Pierre Comon GET/INT, UMR-CNRS 7, 9 rue Charles Fourier, 9 Évry Cedex, France marc.castella@int-evry.fr, CNRS, I3S, UMR
More informationICA. Independent Component Analysis. Zakariás Mátyás
ICA Independent Component Analysis Zakariás Mátyás Contents Definitions Introduction History Algorithms Code Uses of ICA Definitions ICA Miture Separation Signals typical signals Multivariate statistics
More informationEstimation of linear non-gaussian acyclic models for latent factors
Estimation of linear non-gaussian acyclic models for latent factors Shohei Shimizu a Patrik O. Hoyer b Aapo Hyvärinen b,c a The Institute of Scientific and Industrial Research, Osaka University Mihogaoka
More informationwhere A 2 IR m n is the mixing matrix, s(t) is the n-dimensional source vector (n» m), and v(t) is additive white noise that is statistically independ
BLIND SEPARATION OF NONSTATIONARY AND TEMPORALLY CORRELATED SOURCES FROM NOISY MIXTURES Seungjin CHOI x and Andrzej CICHOCKI y x Department of Electrical Engineering Chungbuk National University, KOREA
More informationIndependent component analysis: algorithms and applications
PERGAMON Neural Networks 13 (2000) 411 430 Invited article Independent component analysis: algorithms and applications A. Hyvärinen, E. Oja* Neural Networks Research Centre, Helsinki University of Technology,
More informationModèles stochastiques II
Modèles stochastiques II INFO 154 Gianluca Bontempi Département d Informatique Boulevard de Triomphe - CP 1 http://ulbacbe/di Modéles stochastiques II p1/50 The basics of statistics Statistics starts ith
More informationIndependent Component Analysis and Blind Source Separation
Independent Component Analysis and Blind Source Separation Aapo Hyvärinen University of Helsinki and Helsinki Institute of Information Technology 1 Blind source separation Four source signals : 1.5 2 3
More informationOne-unit Learning Rules for Independent Component Analysis
One-unit Learning Rules for Independent Component Analysis Aapo Hyvarinen and Erkki Oja Helsinki University of Technology Laboratory of Computer and Information Science Rakentajanaukio 2 C, FIN-02150 Espoo,
More informationLecture'12:' SSMs;'Independent'Component'Analysis;' Canonical'Correla;on'Analysis'
Lecture'12:' SSMs;'Independent'Component'Analysis;' Canonical'Correla;on'Analysis' Lester'Mackey' May'7,'2014' ' Stats'306B:'Unsupervised'Learning' Beyond'linearity'in'state'space'modeling' Credit:'Alex'Simma'
More informationBLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES
BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES Dinh-Tuan Pham Laboratoire de Modélisation et Calcul URA 397, CNRS/UJF/INPG BP 53X, 38041 Grenoble cédex, France Dinh-Tuan.Pham@imag.fr
More informationCIFAR Lectures: Non-Gaussian statistics and natural images
CIFAR Lectures: Non-Gaussian statistics and natural images Dept of Computer Science University of Helsinki, Finland Outline Part I: Theory of ICA Definition and difference to PCA Importance of non-gaussianity
More informationIndependent Component Analysis. PhD Seminar Jörgen Ungh
Independent Component Analysis PhD Seminar Jörgen Ungh Agenda Background a motivater Independence ICA vs. PCA Gaussian data ICA theory Examples Background & motivation The cocktail party problem Bla bla
More informationIndependent Component Analysis
Independent Component Analysis James V. Stone November 4, 24 Sheffield University, Sheffield, UK Keywords: independent component analysis, independence, blind source separation, projection pursuit, complexity
More informationA ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT
Progress In Electromagnetics Research Letters, Vol. 16, 53 60, 2010 A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT Y. P. Liu and Q. Wan School of Electronic Engineering University of Electronic
More informationHST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007
MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationIndependent Component Analysis
A Short Introduction to Independent Component Analysis Aapo Hyvärinen Helsinki Institute for Information Technology and Depts of Computer Science and Psychology University of Helsinki Problem of blind
More informationGatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II
Gatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II Gatsby Unit University College London 27 Feb 2017 Outline Part I: Theory of ICA Definition and difference
More informationBLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES
BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES Dinh-Tuan Pham Laboratoire de Modélisation et Calcul URA 397, CNRS/UJF/INPG BP 53X, 38041 Grenoble cédex, France Dinh-Tuan.Pham@imag.fr
More informationIndependent Component Analysis
000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 1 Introduction Indepent
More informationMassoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39
Blind Source Separation (BSS) and Independent Componen Analysis (ICA) Massoud BABAIE-ZADEH Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39 Outline Part I Part II Introduction
More informationFASTGEO - A HISTOGRAM BASED APPROACH TO LINEAR GEOMETRIC ICA. Andreas Jung, Fabian J. Theis, Carlos G. Puntonet, Elmar W. Lang
FASTGEO - A HISTOGRAM BASED APPROACH TO LINEAR GEOMETRIC ICA Andreas Jung, Fabian J Theis, Carlos G Puntonet, Elmar W Lang Institute for Theoretical Physics, University of Regensburg Institute of Biophysics,
More informationBlind Machine Separation Te-Won Lee
Blind Machine Separation Te-Won Lee University of California, San Diego Institute for Neural Computation Blind Machine Separation Problem we want to solve: Single microphone blind source separation & deconvolution
More informationBlind Signal Separation: Statistical Principles
Blind Signal Separation: Statistical Principles JEAN-FRANÇOIS CARDOSO, MEMBER, IEEE Invited Paper Blind signal separation (BSS) and independent component analysis (ICA) are emerging techniques of array
More informationRecursive Generalized Eigendecomposition for Independent Component Analysis
Recursive Generalized Eigendecomposition for Independent Component Analysis Umut Ozertem 1, Deniz Erdogmus 1,, ian Lan 1 CSEE Department, OGI, Oregon Health & Science University, Portland, OR, USA. {ozertemu,deniz}@csee.ogi.edu
More informationHST.582J/6.555J/16.456J
Blind Source Separation: PCA & ICA HST.582J/6.555J/16.456J Gari D. Clifford gari [at] mit. edu http://www.mit.edu/~gari G. D. Clifford 2005-2009 What is BSS? Assume an observation (signal) is a linear
More informationTWO METHODS FOR ESTIMATING OVERCOMPLETE INDEPENDENT COMPONENT BASES. Mika Inki and Aapo Hyvärinen
TWO METHODS FOR ESTIMATING OVERCOMPLETE INDEPENDENT COMPONENT BASES Mika Inki and Aapo Hyvärinen Neural Networks Research Centre Helsinki University of Technology P.O. Box 54, FIN-215 HUT, Finland ABSTRACT
More informationBLIND SEPARATION USING ABSOLUTE MOMENTS BASED ADAPTIVE ESTIMATING FUNCTION. Juha Karvanen and Visa Koivunen
BLIND SEPARATION USING ABSOLUTE MOMENTS BASED ADAPTIVE ESTIMATING UNCTION Juha Karvanen and Visa Koivunen Signal Processing Laboratory Helsinki University of Technology P.O. Box 3, IN-215 HUT, inland Tel.
More informationNon-orthogonal Support-Width ICA
ESANN'6 proceedings - European Symposium on Artificial Neural Networks Bruges (Belgium), 6-8 April 6, d-side publi., ISBN -9337-6-4. Non-orthogonal Support-Width ICA John A. Lee, Frédéric Vrins and Michel
More informationIndependent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego
Independent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego Email: brao@ucsdedu References 1 Hyvarinen, A, Karhunen, J, & Oja, E (2004) Independent component analysis (Vol 46)
More informationParametric independent component analysis for stable distributions
ORIGIAL RESEARCH Parametric independent component analysis for stable distributions ohammad Reza Ameri, 2, ona Shokripour 3, Adel ohammadpour 2, Vahid assiri 2, 4. Department of Computer Science and Software
More informationMTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen
MTTS1 Dimensionality Reduction and Visualization Spring 2014 Jaakko Peltonen Lecture 3: Linear feature extraction Feature extraction feature extraction: (more general) transform the original to (k < d).
More informationLeast square joint diagonalization of matrices under an intrinsic scale constraint
Least square joint diagonalization of matrices under an intrinsic scale constraint Dinh-Tuan Pham 1 and Marco Congedo 2 1 Laboratory Jean Kuntzmann, cnrs - Grenoble INP - UJF (Grenoble, France) Dinh-Tuan.Pham@imag.fr
More informationAutoregressive Independent Process Analysis with Missing Observations
Autoregressive Independent Process Analysis with Missing Observations Zoltán Szabó Eötvös Loránd University - Department of Software Technology and Methodology Pázmány P. sétány 1/C, Budapest, H-1117 -
More informationIndependent Component Analysis (ICA)
Independent Component Analysis (ICA) Université catholique de Louvain (Belgium) Machine Learning Group http://www.dice.ucl ucl.ac.be/.ac.be/mlg/ 1 Overview Uncorrelation vs Independence Blind source separation
More informationPrincipal Component Analysis
Principal Component Analysis Introduction Consider a zero mean random vector R n with autocorrelation matri R = E( T ). R has eigenvectors q(1),,q(n) and associated eigenvalues λ(1) λ(n). Let Q = [ q(1)
More informationNon-Euclidean Independent Component Analysis and Oja's Learning
Non-Euclidean Independent Component Analysis and Oja's Learning M. Lange 1, M. Biehl 2, and T. Villmann 1 1- University of Appl. Sciences Mittweida - Dept. of Mathematics Mittweida, Saxonia - Germany 2-
More informationAdvanced Introduction to Machine Learning CMU-10715
Advanced Introduction to Machine Learning CMU-10715 Independent Component Analysis Barnabás Póczos Independent Component Analysis 2 Independent Component Analysis Model original signals Observations (Mixtures)
More informationBlind Source Separation with a Time-Varying Mixing Matrix
Blind Source Separation with a Time-Varying Mixing Matrix Marcus R DeYoung and Brian L Evans Department of Electrical and Computer Engineering The University of Texas at Austin 1 University Station, Austin,
More informationACENTRAL problem in neural-network research, as well
626 IEEE TRANSACTIONS ON NEURAL NETWORKS, VOL. 10, NO. 3, MAY 1999 Fast and Robust Fixed-Point Algorithms for Independent Component Analysis Aapo Hyvärinen Abstract Independent component analysis (ICA)
More informationOn Information Maximization and Blind Signal Deconvolution
On Information Maximization and Blind Signal Deconvolution A Röbel Technical University of Berlin, Institute of Communication Sciences email: roebel@kgwtu-berlinde Abstract: In the following paper we investigate
More informationAn Introduction to Independent Components Analysis (ICA)
An Introduction to Independent Components Analysis (ICA) Anish R. Shah, CFA Northfield Information Services Anish@northinfo.com Newport Jun 6, 2008 1 Overview of Talk Review principal components Introduce
More informationDistinguishing Causes from Effects using Nonlinear Acyclic Causal Models
JMLR Workshop and Conference Proceedings 6:17 164 NIPS 28 workshop on causality Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Kun Zhang Dept of Computer Science and HIIT University
More informationAnalytical solution of the blind source separation problem using derivatives
Analytical solution of the blind source separation problem using derivatives Sebastien Lagrange 1,2, Luc Jaulin 2, Vincent Vigneron 1, and Christian Jutten 1 1 Laboratoire Images et Signaux, Institut National
More informationTutorial on Blind Source Separation and Independent Component Analysis
Tutorial on Blind Source Separation and Independent Component Analysis Lucas Parra Adaptive Image & Signal Processing Group Sarnoff Corporation February 09, 2002 Linear Mixtures... problem statement...
More informationBlind signal processing algorithms
12th Int. Workshop on Systems, Signals & Image Processing, 22-24 September 2005, Chalkida, Greece 105 Blind signal processing algorithms Athanasios Margaris and Efthimios Kotsialos Department of Applied
More informationBlind separation of sources that have spatiotemporal variance dependencies
Blind separation of sources that have spatiotemporal variance dependencies Aapo Hyvärinen a b Jarmo Hurri a a Neural Networks Research Centre, Helsinki University of Technology, Finland b Helsinki Institute
More informationADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA. Mark Plumbley
Submitteed to the International Conference on Independent Component Analysis and Blind Signal Separation (ICA2) ADAPTIVE LATERAL INHIBITION FOR NON-NEGATIVE ICA Mark Plumbley Audio & Music Lab Department
More informationIndependent Component Analysis
Independent Component Analysis Seungjin Choi Department of Computer Science Pohang University of Science and Technology, Korea seungjin@postech.ac.kr March 4, 2009 1 / 78 Outline Theory and Preliminaries
More informationy 2 a 12 a 1n a 11 s 2 w 11 f n a 21 s n f n y 1 a n1 x n x 2 x 1
Maximum Likelihood Blind Source Separation: A Context-Sensitive Generalization of ICA Barak A. Pearlmutter Computer Science Dept, FEC 33 University of Ne Mexico Albuquerque, NM 873 bap@cs.unm.edu Lucas
More information! " # $! % & '! , ) ( + - (. ) ( ) * + / 0 1 2 3 0 / 4 5 / 6 0 ; 8 7 < = 7 > 8 7 8 9 : Œ Š ž P P h ˆ Š ˆ Œ ˆ Š ˆ Ž Ž Ý Ü Ý Ü Ý Ž Ý ê ç è ± ¹ ¼ ¹ ä ± ¹ w ç ¹ è ¼ è Œ ¹ ± ¹ è ¹ è ä ç w ¹ ã ¼ ¹ ä ¹ ¼ ¹ ±
More informationRobust extraction of specific signals with temporal structure
Robust extraction of specific signals with temporal structure Zhi-Lin Zhang, Zhang Yi Computational Intelligence Laboratory, School of Computer Science and Engineering, University of Electronic Science
More informationPCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani
PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given
More informationUsing Kernel PCA for Initialisation of Variational Bayesian Nonlinear Blind Source Separation Method
Using Kernel PCA for Initialisation of Variational Bayesian Nonlinear Blind Source Separation Method Antti Honkela 1, Stefan Harmeling 2, Leo Lundqvist 1, and Harri Valpola 1 1 Helsinki University of Technology,
More informationMachine Learning: Logistic Regression. Lecture 04
Machine Learning: Logistic Regression Razvan C. Bunescu School of Electrical Engineering and Computer Science bunescu@ohio.edu Supervised Learning Task = learn an (unkon function t : X T that maps input
More informationAcoustic Source Separation with Microphone Arrays CCNY
Acoustic Source Separation with Microphone Arrays Lucas C. Parra Biomedical Engineering Department City College of New York CCNY Craig Fancourt Clay Spence Chris Alvino Montreal Workshop, Nov 6, 2004 Blind
More informationUnsupervised learning: beyond simple clustering and PCA
Unsupervised learning: beyond simple clustering and PCA Liza Rebrova Self organizing maps (SOM) Goal: approximate data points in R p by a low-dimensional manifold Unlike PCA, the manifold does not have
More informationA more efficient second order blind identification method for separation of uncorrelated stationary time series
A more efficient second order blind identification method for separation of uncorrelated stationary time series Sara Taskinen 1a, Jari Miettinen a, Klaus Nordhausen b,c a Department of Mathematics and
More informationSemi-Blind approaches to source separation: introduction to the special session
Semi-Blind approaches to source separation: introduction to the special session Massoud BABAIE-ZADEH 1 Christian JUTTEN 2 1- Sharif University of Technology, Tehran, IRAN 2- Laboratory of Images and Signals
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationSimultaneous Diagonalization in the Frequency Domain (SDIF) for Source Separation
Simultaneous Diagonalization in the Frequency Domain (SDIF) for Source Separation Hsiao-Chun Wu and Jose C. Principe Computational Neuro-Engineering Laboratory Department of Electrical and Computer Engineering
More informationComparative Analysis of ICA Based Features
International Journal of Emerging Engineering Research and Technology Volume 2, Issue 7, October 2014, PP 267-273 ISSN 2349-4395 (Print) & ISSN 2349-4409 (Online) Comparative Analysis of ICA Based Features
More informationNatural Gradient Learning for Over- and Under-Complete Bases in ICA
NOTE Communicated by Jean-François Cardoso Natural Gradient Learning for Over- and Under-Complete Bases in ICA Shun-ichi Amari RIKEN Brain Science Institute, Wako-shi, Hirosawa, Saitama 351-01, Japan Independent
More informationSeparation of Different Voices in Speech using Fast Ica Algorithm
Volume-6, Issue-6, November-December 2016 International Journal of Engineering and Management Research Page Number: 364-368 Separation of Different Voices in Speech using Fast Ica Algorithm Dr. T.V.P Sundararajan
More informationRobust Subspace DOA Estimation for Wireless Communications
Robust Subspace DOA Estimation for Wireless Communications Samuli Visuri Hannu Oja ¾ Visa Koivunen Laboratory of Signal Processing Computer Technology Helsinki Univ. of Technology P.O. Box 3, FIN-25 HUT
More informationComparison of Fast ICA and Gradient Algorithms of Independent Component Analysis for Separation of Speech Signals
K. Mohanaprasad et.al / International Journal of Engineering and echnolog (IJE) Comparison of Fast ICA and Gradient Algorithms of Independent Component Analsis for Separation of Speech Signals K. Mohanaprasad
More informationBlind Instantaneous Noisy Mixture Separation with Best Interference-plus-noise Rejection
Blind Instantaneous Noisy Mixture Separation with Best Interference-plus-noise Rejection Zbyněk Koldovský 1,2 and Petr Tichavský 1 1 Institute of Information Theory and Automation, Pod vodárenskou věží
More informationc Springer, Reprinted with permission.
Zhijian Yuan and Erkki Oja. A FastICA Algorithm for Non-negative Independent Component Analysis. In Puntonet, Carlos G.; Prieto, Alberto (Eds.), Proceedings of the Fifth International Symposium on Independent
More informationIndependent Component Analysis of Incomplete Data
Independent Component Analysis of Incomplete Data Max Welling Markus Weber California Institute of Technology 136-93 Pasadena, CA 91125 fwelling,rmwg@vision.caltech.edu Keywords: EM, Missing Data, ICA
More information1 Introduction Blind source separation (BSS) is a fundamental problem which is encountered in a variety of signal processing problems where multiple s
Blind Separation of Nonstationary Sources in Noisy Mixtures Seungjin CHOI x1 and Andrzej CICHOCKI y x Department of Electrical Engineering Chungbuk National University 48 Kaeshin-dong, Cheongju Chungbuk
More informationDistinguishing Causes from Effects using Nonlinear Acyclic Causal Models
Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Kun Zhang Dept of Computer Science and HIIT University of Helsinki 14 Helsinki, Finland kun.zhang@cs.helsinki.fi Aapo Hyvärinen
More informationORIENTED PCA AND BLIND SIGNAL SEPARATION
ORIENTED PCA AND BLIND SIGNAL SEPARATION K. I. Diamantaras Department of Informatics TEI of Thessaloniki Sindos 54101, Greece kdiamant@it.teithe.gr Th. Papadimitriou Department of Int. Economic Relat.
More informationASYMPTOTIC PERFORMANCE OF MIMO WIRELESS CHANNELS WITH LIMITED FEEDBACK
ASYMPTTIC PERFRMANCE F MIM WIRELESS CHANNELS WITH LIMITED FEEDBACK Wiroonsak Santipach and Michael L. Honig Department of Electrical and Computer Engineering Northestern University 2145 Sheridan Road Evanston,
More information3 - Vector Spaces Definition vector space linear space u, v,
3 - Vector Spaces Vectors in R and R 3 are essentially matrices. They can be vieed either as column vectors (matrices of size and 3, respectively) or ro vectors ( and 3 matrices). The addition and scalar
More informationy(x) = x w + ε(x), (1)
Linear regression We are ready to consider our first machine-learning problem: linear regression. Suppose that e are interested in the values of a function y(x): R d R, here x is a d-dimensional vector-valued
More informationSpeed and Accuracy Enhancement of Linear ICA Techniques Using Rational Nonlinear Functions
Speed and Accuracy Enhancement of Linear ICA Techniques Using Rational Nonlinear Functions Petr Tichavský 1, Zbyněk Koldovský 1,2, and Erkki Oja 3 1 Institute of Information Theory and Automation, Pod
More informationIndependent Component Analysis
A Short Introduction to Independent Component Analysis with Some Recent Advances Aapo Hyvärinen Dept of Computer Science Dept of Mathematics and Statistics University of Helsinki Problem of blind source
More informationArtificial Intelligence Module 2. Feature Selection. Andrea Torsello
Artificial Intelligence Module 2 Feature Selection Andrea Torsello We have seen that high dimensional data is hard to classify (curse of dimensionality) Often however, the data does not fill all the space
More informationCOS 424: Interacting with Data. Lecturer: Rob Schapire Lecture #15 Scribe: Haipeng Zheng April 5, 2007
COS 424: Interacting ith Data Lecturer: Rob Schapire Lecture #15 Scribe: Haipeng Zheng April 5, 2007 Recapitulation of Last Lecture In linear regression, e need to avoid adding too much richness to the
More informationICA Using Kernel Entropy Estimation with NlogN Complexity
ICA Using Kernel Entropy Estimation with NlogN Complexity Sarit Shwartz, Michael Zibulevsky, and Yoav Y. Schechner Department of Electrical Engineering Technion - Israel Institute of Technology, Haifa
More informationPrincipal Component Analysis vs. Independent Component Analysis for Damage Detection
6th European Workshop on Structural Health Monitoring - Fr..D.4 Principal Component Analysis vs. Independent Component Analysis for Damage Detection D. A. TIBADUIZA, L. E. MUJICA, M. ANAYA, J. RODELLAR
More informationA Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag
A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationCHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum
CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum 1997 19 CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION 3.0. Introduction
More informationA Canonical Genetic Algorithm for Blind Inversion of Linear Channels
A Canonical Genetic Algorithm for Blind Inversion of Linear Channels Fernando Rojas, Jordi Solé-Casals, Enric Monte-Moreno 3, Carlos G. Puntonet and Alberto Prieto Computer Architecture and Technology
More informationIndependent Component Analysis
Independent Component Analysis Philippe B. Laval KSU Fall 2017 Philippe B. Laval (KSU) ICA Fall 2017 1 / 18 Introduction Independent Component Analysis (ICA) falls under the broader topic of Blind Source
More informationAPPLICATION OF INDEPENDENT COMPONENT ANALYSIS TO CHEMICAL REACTIONS. S.Triadaphillou, A. J. Morris and E. B. Martin
APPLICAION OF INDEPENDEN COMPONEN ANALYSIS O CHEMICAL REACIONS S.riadaphillou, A. J. Morris and E. B. Martin Centre for Process Analytics and Control echnology School of Chemical Engineering and Advanced
More informationComparative Performance Analysis of Three Algorithms for Principal Component Analysis
84 R. LANDQVIST, A. MOHAMMED, COMPARATIVE PERFORMANCE ANALYSIS OF THR ALGORITHMS Comparative Performance Analysis of Three Algorithms for Principal Component Analysis Ronnie LANDQVIST, Abbas MOHAMMED Dept.
More informationGENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS. Mitsuru Kawamoto 1,2 and Yujiro Inouye 1
GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS Mitsuru Kawamoto,2 and Yuiro Inouye. Dept. of Electronic and Control Systems Engineering, Shimane University,
More informationLecture VIII Dim. Reduction (I)
Lecture VIII Dim. Reduction (I) Contents: Subset Selection & Shrinkage Ridge regression, Lasso PCA, PCR, PLS Lecture VIII: MLSC - Dr. Sethu Viayakumar Data From Human Movement Measure arm movement and
More information