BLIND source separation is a set of techniques that is used

Size: px
Start display at page:

Download "BLIND source separation is a set of techniques that is used"

Transcription

1 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK A Novel Approach to Blind Separation of Delayed Sources in Linear Mixtures Jakob Ashtar, Kristoffer Møller Jørgensen, Jacob Lindvig, Lars Juul Mikkelsen, Jakob Birkedal Nielsen, Lars Sommer Søndergaard, and Dario Farina, Supervisor, Aalborg University Abstract This paper addresses the problem of blind estimation of time-delayed sources from a set of linear mixtures. The mixtures are analyzed and approximated by a linear, instantaneous signal model and the sources are recovered using second order statistical methods. An algorithm is implemented and test results are demonstrated and discussed. The proposed algorithm is referred to as SOBIDS Second Order Blind Identification of Dealyed Sources). SOBIDS is compared to another algorithm using second order statistics and test results show that SOBIDS performs better for delayed sources. I. INTRODUCTION BLIND source separation is a set of techniques that is used in many fields of engineering such as speech recognition, telecommunications, and biomedical signal processing. One of the first papers describing the problem of recovering sources was published in 985 by Herault et al. [5] These authors addressed the problem by using and extending principal component analysis. Various approaches have been proposed since then. Many of them are based on mathematical disciplines such as principal component analysis, independent component analysis, or singular value decomposition [] [2]. The classical cocktail party problem where two persons communicate in a room with loud music and background noise is an example of a situation where the human brain utilizes blind source separation. In this case the human brain focuses on a speaker, distinguishes one sound from the others, and ignores the background noise. Simulating this process with machines is, however, a very difficult task. The wide range and nature of applications further complicates the task. In some cases the sources can be assumed stationary. In other cases, e.g. speech signals, the sources are non-stationary. Signal propagation delays might also be relevant to consider. Also, the mixtures of sources may be either instantaneous and linear or convolutive. The aim of this paper is to propose a new method SOBIDS) for the separation of sources mixed with unknown, small delays in a linear, instantaneous way. The paper is organized as follows. In section II and III we provide a description of a blind identification algorithm previously proposed to separate sources from their linear, instantaneous mixtures using second order statistics []. This algorithm is referred to as SOBI Second Order Blind Identification). Section IV introduces a convolutive signal model where the sources are delayed. To solve the blind source separation problem for a convolutive model we approximate the model with a linear, instantaneous signal model. In section V we show that it is possible to blindly estimate the sources for small delays. In section VI we compare test results for SOBI and SOBIDS and discuss the performance of the new algorithm. II. LINEAR INSTANTANEOUS MIXTURES Assuming linear, instantaneous mixtures we have the following signal model. A. Signal Model Let a set of n unknown sources s t),s 2 t),...,s j t),...,s n t) be linearly combined to form m known observations y t),y 2 t),...,y i t),...,y m t). Each observation y i t) is contaminated by unknown additive noise p i t). Expressed mathematically we get the following relation: y t) = a s t)+a 2 s 2 t)+...+ a n s n t)+ p t) y 2 t) = a 2 s t)+a 22 s 2 t)+...+ a 2n s n t)+ p 2 t). y m t) = a m s t)+a m2 s 2 t)+...+ a mn s n t)+ p m t) ) The above system of linear equations can also be represented in matrix notation as a transformation T : R n R m of the source vector st): yt) = Tst)) = xt)+pt) = Ast)+pt) 2) where the observation vector yt), the signal vector xt), and the noise vector pt) are m column vectors and the source vector st) is an n column vector. The unknown m n matrix A is called the mixing matrix. B. Assumptions It is assumed that the sources are real, ergodic, zero mean stochastic processes. Since the power of the sources cannot be determined [], it is assumed without loss of generality that the sources have unit variance. Furthermore, the sources are assumed to be mutually uncorrelated spatially white). It is also assumed that the sources are uncorrelated with the additive noise. The noise process is a stationary, temporally and spatially white, zero-mean, and real random process. The noise variance is the same for all observations. The mixing matrix has full rank. Furthermore, it is assumed that there are at least as many observations as sources, that is m n.

2 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 2 C. Derivations Based on the signal model and the above assumptions the n n covariance matrix of the sources R s τ) is given by: R s τ) = E [ st)s T t + τ) ] = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 3) where R s j s j τ) = E [ s j t)s j t +τ) ]. Since the sources are mutually uncorrelated and assumed to have unit variance, we have: R s 0) = I 4) As the noise is temporally and spatially white the m m covariance matrix R p τ) of the noise is 0 at time lags τ 0. At τ = 0 the covariance matrix R p τ) is the identity matrix scaled by the noise variance σ 2 p : R p τ) = E [ pt)p T t + τ) ] = σ 2 pδτ)i 5) where δτ) is the Kronecker delta. The m m covariance matrix R y τ) of the observations is given by: R y τ) = E [ yt)y T t + τ) ] = E [ Ast)+pt))Ast + τ)+pt + τ)) T] = E [ Ast)s T t + τ)a T + pt)p T t + τ) ] = AR s τ)a T + σ 2 pδτ)i 6) = R x τ)+σ 2 p δτ)i where R x τ) is the covariance matrix of the signal vector. III. THE SOBI ALGORITHM The method for blind separation described in [] is based on two steps: whitening and rotation. The whitening step requires an estimation of the noise variance from the observations. A. Estimation of the Noise Variance The first step is to estimate the noise variance σ 2 p. This can be done by applying 6) for τ = 0 which is the only time instant where the noise variance is a part of the equation: R y 0) = AR s 0)A T + σ 2 p δ0)i = AIA T + σ 2 pi = AA T + σ 2 p I 7) Since AA T is a symmetric matrix it is also orthogonally diagonalizable [3, pp. 445]. If we orthogonally diagonalize AA T we can rewrite 7) as: R y 0) = UDU T + σ 2 pi 8) where the column vectors of U are the orthonormal eigenvectors of the positive definite matrix AA T. The corresponding positive and sorted eigenvalues [3, pp. 456]: are in the diagonal matrix λ λ n λ m 0 9) D = diag[λ,λ 2,...,λ n,λ n+,...,λ m ] 0) Understanding that the right side of 8) is, in fact, an orthogonal diagonalization of R y 0) becomes more easy when the equation is written in a different way: R y 0) = UDU T + Uσ 2 p IUT = UD+σ 2 pi)u T = UVU T ) From ) we observe that the eigenvalues v ii in the diagonal of V contain the noise variance: v ii = λ i + σ 2 p 2) From our assumptions the mixing matrix A has full rank equal to n). Since the column vectors of AA T are linear combinations of the column vectors of A we can derive that: λ i = 0, i > n 3) This means that the last m n eigenvalues in V are: v ii = σ 2 p, i > n 4) Thus the noise variance can be estimated by calculating an average of the last m n eigenvalues v ii in V, that is: B. Whitening σ 2 p = m n m i=n+ v ii 5) From the definition of the transform in 2) we see that the n-dimensional source vector is mapped from R n to an m- dimensional signal vector in R m. Disregarding the noise vector we notice that the signal vector is a linear combination of the n linearly independent m-dimensional column vectors of A. Since there are only n linearly independent column vectors they only span R n and not all of R m. As a consequence the representation of the source vector will contain redundant information when m > n. Another problem is that the source vector is mapped from an orthogonal coordinate system to a coordinate system which is not guaranteed to be orthogonal. To eliminate the redundant information and to ensure that the source vector is mapped into an orthogonal coordinate system we introduce the whitening matrix W. The n m whitening matrix W is defined in the following way: WR x 0)W T = I WAR s 0)A T W T = I WAIA T W T = I WAA T W T = I 6) such that the transformed sources are uncorrelated at time lag τ = 0. Thus, the whitening matrix has the following property: WA)WA) T = WAA T W T = I 7) If we observe what happens when the whitening matrix is applied to the observation vector: Wyt) = Wxt)+Wpt) = WAst)+Wpt) 8)

3 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 3 we see that the source vector is now mapped from R n to R n no redundant information) and that the transformation of the signal vector is orthogonal since WA is orthogonal). To derive an expression for the whitening matrix we continue by substituting AA T in 7) with UDU T : WUDU T W T = I 9) The last m n eigenvalues in the matrix D are zero. Thus we can reduce the size of D and U without affecting the diagonalization of AA T, that is: AA T = U r D r U T r 20) where the n n reduced eigenvalue matrix D r is: D r = diag[λ,λ 2,...,λ n ] 2) Column vector j in the m n matrix U r is the eigenvector corresponding to the eigenvalue λ j. The matrix V is reduced in a similar way, such that: We substitute UDU T in 9) with U r D r U T r and replace D r with V r σ 2 pi: V r = D r + σ 2 p I 22) and get: WU r D r U T r W T = I 23) WU r Vr σ 2 p I) U T r WT = I 24) Rearranging the equation such that it reads: WU r Vr σ 2 pi ) 2 V r σ 2 pi ) 2 U T r W T = I 25) [WU r Vr σ 2p I) 2 ][ WU r Vr σ 2 p I) 2 ] T = I 26) we see that it is possible to obtain a closed analytical expression for the whitening matrix by solving the following equation: WU r Vr σ 2 p I) 2 = I 27) under the assumption that an estimate of the noise variance is known. The matrices U r and V r can be obtained by orthogonally diagonalizing R y 0). The expression V r σ 2 pi ) 2 is calculated by taking the square root of each entry in the diagonal matrix V r σ 2 pi. As a conclusion we derive the following expression for the whitening matrix: C. Estimation of the Mixing Matrix W = V r σ 2 p I) 2 U T r 28) The final step in recovering the sources is to estimate the mixing matrix A. To get an estimate of the mixing matrix we choose to investigate the whitened covariance matrix R z τ) at different time lags τ 0: R z τ) = WR y τ)w T = WAR s τ)a T W T = WAR s τ)wa) T = QR s τ)q T, τ 0 29) and realize that the right side of the equation is an orthogonal diagonalization of R z τ). Consequently, the key idea here, since W and R y τ) can be estimated, is to solve 29) with respect to Q at K different time lags τ K > τ K > τ K 2 > τ 2 > τ > 0. The set of solutions {Q,Q 2,...,Q K,Q K } will be the basis for calculating an estimate Q of Q and hence also an estimate of the mixing matrix since: Q = ŴÂ 30) We rearrange the equation to obtain an expression for the estimate of the mixing matrix: Â = Ŵ Q 3) where Ŵ is the pseudo-inverse of Ŵ. Since it is beyond the scope of this paper to delve further into how an estimate is obtained, we refer to [] which in greater detail explains how Q is estimated using a technique called joint diagonalization. IV. MIXTURES WITH DELAYED SOURCES When propagation delays are introduced the signal model changes from a model with linear, instantaneous mixtures to a convolutive model. In the following sections we will approximate this convolutive model with a linear, instantaneous model in which the sources can be estimated using second order statistics. A. Convolutive Signal Model Let a set of n unknown sources s t),s 2 t),...,s j t),...,s n t) be linearly combined with unknown delays t i j to form m observations y t),y 2 t),...,y i t),...,y m t). Each observation y i t) is contaminated by unknown additive noise p i t). This model can be expressed mathematically in the following way: y t) = y 2 t) =. y m t) = a s t t )+...+ a n s n t t n )+ p t) a 2 s t t 2 )+...+ a 2n s n t t 2 )+ p 2 t) a m s t t m )+...+ a mn s n t t mn )+ p m t) B. Approximation of a Delayed Source 32) From the definition of the derivative: st + ε) st) ṡt) = lim 33) ε 0 ε we arrive at the following approximation when substituting ε with a small delay t i j : which shows that: t i j ṡt) st t i j ) st) 34) st t i j ) st) t i j ṡt) 35) This approximation corresponds to the Taylor expansion truncated at the first order and is used in [2] where a delay is assumed small if: t i j 36) 2 π fmax where f max denotes the maximum frequency in st).

4 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 4 C. An Approximation of the Convolutive Signal Model From 35) it is possible to define a linear, instantaneous signal model, which is an approximation of the convolutive signal model. The approximated signal model is defined as: y i t) n j= or equivalently in matrix notation as: a i j s j t) t i j ṡ j t))+ p i t) 37) yt) Ast)+pt) 38) where the m 2n mixing matrix A is defined as: a a n a t σṡ a n t n σṡn a 2 a 2n a 2 t 2 σṡ a 2n t 2n σṡn A = a m a mn a m t m σṡ a mn t mn σṡn 39) The 2n source vector st) contains the sources and their normalized derivatives and is defined as: st) = [s t),s 2 t),...,s n t), ṡt), ṡ2t),..., ṡnt) ] T σṡ σṡ2 σṡn = [s t),s 2 t),...,s n t), s t), s 2 t),..., s n t)] T 40) where s j t) is the derivative of s j t) normalized with respect to the standard deviation σṡ j of ṡ j t). D. Assumptions The sources and noise have the same statistical properties as defined in II-B. Hence the derivatives are zero mean, second order stationary, ergodic, stochastic processes. Moreover, the derivatives are uncorrelated with the noise and also spatially white. Finally it is assumed that there are at least twice as many observations as sources, that is m 2n. The mixing matrix A has full rank equal to 2n. E. Derivations Based on the approximative signal model and the above assumptions the 2n 2n covariance matrix R s τ) is given by: [ ] Rss τ) R R s τ) = s s τ) 4) R ss τ) R s s τ) where the n n quadrant matrices R ss τ),r s s τ),r ss τ) and R s s τ) are defined as: R ss τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 42) R s s τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 43) R ss τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 44) R s s τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 45) Considering that the sources are mutually uncorrelated we have that: R ss 0) = I 46) Considering that the derivatives of the sources are mutually uncorrelated we have that: R s s 0) = I 47) From [4, pp. 65] and [4, pp. 44] we know that: and R s j ṡ j = Ṙ s j s j R s j s j τ) = Ṙs j s j τ) σṡ j R s j s j τ) = R s j s j τ) = Ṙs j s j τ) σṡ j 48) Since σ ṡ j Ṙ s j s j τ) is symmetric around τ = 0 we have that R s j s j τ) = σ ṡ j Ṙ s j s j τ) = σ ṡ j Ṙ s j s j τ) = R s j s j τ). Furthermore, we know that R s j s j τ) has a maximum for τ = 0. Based on this we derive that Ṙ s j s j 0) = 0. Combining the two findings we conclude that the covariance matrix R s τ) at τ = 0 is: R s 0) = [ ] I 0 = I 49) 0 I Due to the fact that the noise is temporally and spatially white the m m covariance matrix R p τ) of the noise is 0 at time lags τ 0. At τ = 0 the covariance matrix R p τ) is the identity matrix scaled by the noise variance σ 2 p: R p τ) = E [ pt)p T t + τ) ] = σ 2 pδτ)i 50) where δτ) is the Kronecker delta. The m m covariance matrix R y τ) of the observations is given by: R y τ) = E [ yt)y T t + τ) ] = E [ Ast)+pt))Ast + τ)+pt + τ)) T] = E [ Ast)s T t + τ)a T + pt)p T t + τ) ] = AR s τ)a T + σ 2 pδτ)i 5) = R x τ)+σ 2 pδτ)i where R x τ) is the covariance matrix of the signal vector. V. METHOD In the following text we will show that it is possible to estimate the sources using second order statistics. A. Estimation of the Whitening Matrix and the Noise Variance For τ = 0 we have: R y 0) = AR s 0)A T + σ 2 pi = AA T + σ 2 p I 52) Since 52) is similar to 7) we can estimate the whitening matrix and the noise variance using the method described in section III-A and III-B.

5 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 5 B. Estimation of the Mixing Matrix For τ 0 the whitened covariance matrix of the observations is: WR y τ)w T = WAR s τ)a T W T 53) The covariance matrix of the sources R s τ) is in general non-diagonal. We note that the entries in 43) and 44) are the off-diagonal entries in R s τ). These entries have been shown to be related to the derivative of the autocorrelation function of the sources. The size of the off-diagonal entries compared to the size of the main diagonal entries depends on the autocorrelation functions of the sources and, thus, on the bandwidth and frequency content of each source. To estimate the mixing matrix we use the same method as described in section III-C. Due to the fact that R s τ) has to be diagonal or close to diagonal in order to use this method the performance of the SOBIDS algorithm will depend on the frequency characteristics of the sources. Normalized Amplitude Original Source Estimated Source Original Source Estimated Source Samples C. Implementation Based on our analysis we introduce a second order blind identification algorithm for delayed sources SOBIDS). SO- BIDS is implemented in the following way: ) Estimate the sample covariance R y 0) from L data samples. Denote by λ,λ 2,...,λ 2n the 2n largest eigenvalues and u,u 2,...,u 2n the corresponding eigenvectors of R y 0). 2) Calculate an estimate of the noise variance σ 2 p as the average of the m 2n smallest eigenvalues of R y 0). Calculate an estimate of the whitening matrix as: Ŵ = [ λ σ 2 p) 2 u,...,λ 2n σ 2 p) 2 ] T u 2n 3) Calculate the sample covariance matrix R z τ) of Ŵyt) for a fixed set of time lags τ {τ j j =,...,K}. 4) Calculate an estimate Q of the orthogonal matrix that is a joint diagonalizer of the set { R z τ j ) j =,...,K}. 5) Calculate an estimate of the sources and their normalized derivatives as ŝt) = Q T Ŵyt). 6) Calculate an estimate of the mixing matrix as  = Ŵ# Q. D. Delay Estimation In the following we describe a method for estimating the relative delays using the estimated source vector ŝt) and the estimated mixing matrix Â. We calculate the derivatives of the sources in ŝt) and then the variance of these derivatives to get an estimate of σṡi. Using these estimates and  it is possible to estimate the delays. Note that we have no information on the order of the entries of ŝt) or the arrangement of the columns in Â. Since ŝt) is calculated using the invers of  we do know that the order of the entries in ŝt) corresponds to the order of the columns of Â. By calculating the derivative of each entry in ŝt) we state that it is possible to do a matching of each estimated source ŝ i t) with the corresponding estimated normalized derivative s i t). This provides sufficient information to calculate the relative delay of each source. Fig. : An example of the goodness factor for two source estimates. The goodness factor is 0.93 and 0.72 for top and bottom respectively. VI. RESULTS In this section the results of our performed test of the SOBIDS and SOBI algorithms will be presented. The intention of this paper is to describe an algorithm based on principles from the SOBI algorithm but with increased performance when estimating sources with small delays. Therefore the algorithms will be tested with delayed sources and the results will be compared. These tests will be performed with noiseless and noise contaminated observations. A. Performance Index To identify and compare the performance of the two algorithms a performance index has been used. This index is based on the cross-correlation of each generated source and the corresponding estimated source and is defined as: G = n n j= max{ R s j ŝ j } 54) where max{ R s j ŝ j } is the maximum value of the normalized and absolute cross-correlation between the original source sequence s j and the estimated source sequence ŝ j. Since there are n sources we also have n maximum values and the performance parameter is therefore an average of these n maximum values. The idea is to study the behavior of G which indicates how well the algorithm performs under different circumstances, e.g. various delay values. We refer to G as the goodness factor. In fig. two different matchings of signals are shown with goodness factors of 0.72 and The maximum value of G is which corresponds to a full correlation of the matched signals.

6 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 6 B. Simulation Environment The sources used for the simulations are generated by filtering n stationary white noise sources with second order filters with partly overlapping passbands. The observations are generated as linear mixtures of delayed sources. Noise is added for each observation. In section IV-A this mixture is shown. Preliminary tests have been made to identify the parameters which influence the goodness of the algorithms. The tests show that the goodness is almost independent of a number of parameters which therefore were decided to be kept fixed for all simulations. The tests were thus performed with: 3 sources, 4 matrices to be jointly diagonalized, and an overlap ratio of the passband filters of 0% of the filter bandwidth. The filters were created randomly with equal bandwidth. With 3 sources and 0% overlap we will produce filters with a maximum bandwidth of 0.37 f s/2. The number of observations used in the simulation differs from SOBI to SOBIDS. For SOBI the number of observations is n+2 = 5 and for SOBIDS this value is 2n+2 = 8. Tests have shown that this difference is irrelevant for the performance. The parameters essential for the performance of the algorithms are the SNR of the observations and the size of the relative delay of each source. The delay values are randomly generated and the values will be uniformly distributed in an interval from 0 to a defined maximum value. We state that the single parameter which provides the best information on how a single source is delayed in the observations is the standard deviation of the delays for the source. We therefore define the following parameter σ d as the mean of the standard deviations of the delays for each source: σ d = n n j= Vart j,...,t m j ) 55) We refer to σ d as the mean standard deviation of delay. C. Testing the Algorithms with Noiseless Observations To create an ideal environment for testing the performance when using delayed sources no noise is added to the observations when simulating. Fig. 2 shows the result of a simulation with a variation of the mean standard deviation of delay from 0 to 3. The graph for the SOBI algorithm shows that at σ d = 0 the goodness is 0.96 decreasing significantly to G 0.76 at σ d = 0.5. From this point the performance is falling slightly to a goodness of 0.7 at σ d = 3. We note that at zero delay the goodness of the SOBIDS algorithm is 0.97 decreasing to G 0.94 at σ d = 0.. At 0. σ d 0.5 the performance is approximately constant with G From that point the performance is decreasing approximately linearly to G 0.83 at σ d = 3. Comparing the SOBI and the SOBIDS algorithms we notice that with no delay the performance of SOBIDS is slightly better than SOBI. For mean standard deviation of delays of 0.2 the SOBIDS is performing approximately 5% higher than Goodness, G SOBIDS algorithm SOBI algorithm Mean standard deviation of delay, σ d Fig. 2: Result of simulation with noiseless observations with σ d 3. Goodness, G SOBIDS algorithm SOBI algorithm Mean standard deviation of delay, σ d Fig. 3: Result of simulation with noiseless observations for σ d 30. SOBI and for σ d = 0.5 the performance is approximately 22% higher. At a mean standard deviation of delay of 3 the performance of the SOBIDS algorithm is 7% higher than the SOBI. Simulations with σ d up to 30 show that the performance decreases exponentially fig. 3). D. Testing the Algorithms with Noise Contaminated Observations The performance of the SOBI and the SOBIDS is tested with noise contaminated observations to examine the performance of the algorithms in a more demanding applications. The observations have a SNR of 0dB. Fig. 4 shows the result of a simulation with a variation of the mean standard deviation of delay from 0 to 3. The performance of SOBI with noisy observations at σ d = 0 is approximately 0.82 which is 5% lower than with noiseless

7 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 7 Goodness, G SOBIDS algorithm SOBI algorithm Mean standard deviation of delay, σ d Fig. 4: Simulation result of observations with 0dB SNR for σ d 3. Goodness, G SOBIDS algorithm SOBI algorithm E. Discussion The tests show that the proposed SOBIDS algorithm outperforms SOBI for linear mixtures of delayed sources. Adding noise has an influence on the overall performance characteristics. Since the full effect of adding noise is far from tested in details this is an interesting subject for further study concerning the robustness and effeciency of the SOBIDS algorithm in real-world cases. The performance of the algorithms with noiseless observations decreases as the relative delay increases. This can be explained from section IV-B concerning Taylor series expansion. 35) is an approximation which is only valid for small delays. The SOBIDS uses the first order Taylor series expansion and SOBI uses a zero order Taylor series expansion which explains why the sources recovered from SOBI become inaccurate at smaller delays than the SOBIDS algorithm. As expected, the performance of both algorithms decreases when noise is added. The larger the delay the smaller the difference. For 0.3 σ d 3 the performance decrease of the simulations made with noisy sources is for the SOBI algorithm between 4% and 8% and for SOBIDS between 7% and 3%. This indicates that SOBIDS is slightly less noise robust than the SOBI algorithm. The first order Taylor series expansion has a positive effect on the performance for relatively small delays. For relatively larger delays it would be interesting to recalculate the expressions for the Taylor series and expanding it to higher orders. This might result in a more constant overall performance. As a concequence more observations would be needed and there is still an unsolved problem of whitening as the second derivative of a source is not uncorrelated with the source even for time lag zero. This is left for further studies Mean standard deviation of delay, σ d Fig. 5: Simulation result of observations with 0dB SNR for σ d 30. observations. As delays become larger the difference in goodness decreases to 0.03 at σ d = 3 which corresponds to a drop in performance of 4%. Also for the SOBIDS the performance decreased when noise was added to the observations. The difference of the goodness between tests performed with noisy observations and with noiseless observations is 0. at zero delay corresponding to a performance decrease of approximately 9%. The goodness decreases to 0.79 at σ d = 0.5 which is a 9% drop in performance. At σ d = 3 the performance is 8% lower than the performance of the tests with noiseless observations. We note that there is a large decrease in performance in the interval 0 σ d 0.3. Simulations with σ d up to 30 show that the performance decreases exponentially shown in fig. 5. The goodness of the SOBIDS algorithm is approximately 0.05 higher than SOBI for σ d > 5. VII. CONCLUSION In this paper we present a novel method for blindly separating mutually uncorrelated sources from a convolutive linear mixture contaminated with white noise. The paper covers a signal model introducing an approximation of sources with small delays as a first order Taylor series expansion. The sources are estimated mainly in two steps, a whitening process in order to make the observations spatially white followed by a rotation of the whitened observations making the principal axes the estimated sources. Tests revealed that the proposed SOBIDS algorithm outperforms SOBI for linear mixtures of delayed sources. An analytical method for estimating the delays has been proposed. REFERENCES [] A. Belouchrani, K. Abed-Meraim, J. F. Cardoso and E. Moulines, "A Blind Source Separation Technique Using Second-Order Statistics", IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 45, NO. 2, FEBRUARY 997 [2] J. Barrère, G. Chabriel, "A Compact Sensor Array for Blind Separation of Sources", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMSŮI: FUNDAMENTAL THEORY AND APPLICATIONS, VOL. 49, NO. 5, MAY 2002

8 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 8 [3] D. Lay, "Linear Algebra and Its Applications", 2nd ed., Addison-Wesley Publishing Company, April [4] K. Sam Shanmugan, A. M. Breipohl, "Random Signals - Detection, Estimation and Data Analysis", John Wiley & Sons, 988. [5] J. Herault, C. Jutten and B Ans., "D etection de grandeurs primitives dans un message composite par une architecture de calcul neuromim etique en apprentissage non supervis e.", Proc. Gretsi pp , 985. Jakob Ashtar was born in 972. He is a student at the 7th semester Signal Processing at Aalborg University. Kristoffer Møller Jørgensen was born in 979. He is a student at the 7th semester Signal Processing at Aalborg University. Jacob Lindvig was born in 973. He is a student at the 7th semester Signal Processing at Aalborg University. Lars Juul Mikkelsen was born in 98. He is a student at the 7th semester Signal Processing at Aalborg University. Jakob Birkedal Nielsen was born in 979. He is a student at the 7th semester Signal Processing at Aalborg University. Lars Sommer Søndergaard was born in 980. He is a student at the 7th semester Signal Processing at Aalborg University. Dario Farina Supervisor) was born in 973. He Graduated from Politecnico di Torino in 998. In 200 he obtained his PhD in Electronics Engineering and Communications at Politecnico di Torino and at the Ecole Centrale de Nantes, Nantes France).

where A 2 IR m n is the mixing matrix, s(t) is the n-dimensional source vector (n» m), and v(t) is additive white noise that is statistically independ

where A 2 IR m n is the mixing matrix, s(t) is the n-dimensional source vector (n» m), and v(t) is additive white noise that is statistically independ BLIND SEPARATION OF NONSTATIONARY AND TEMPORALLY CORRELATED SOURCES FROM NOISY MIXTURES Seungjin CHOI x and Andrzej CICHOCKI y x Department of Electrical Engineering Chungbuk National University, KOREA

More information

A UNIFIED PRESENTATION OF BLIND SEPARATION METHODS FOR CONVOLUTIVE MIXTURES USING BLOCK-DIAGONALIZATION

A UNIFIED PRESENTATION OF BLIND SEPARATION METHODS FOR CONVOLUTIVE MIXTURES USING BLOCK-DIAGONALIZATION A UNIFIED PRESENTATION OF BLIND SEPARATION METHODS FOR CONVOLUTIVE MIXTURES USING BLOCK-DIAGONALIZATION Cédric Févotte and Christian Doncarli Institut de Recherche en Communications et Cybernétique de

More information

1 Introduction Blind source separation (BSS) is a fundamental problem which is encountered in a variety of signal processing problems where multiple s

1 Introduction Blind source separation (BSS) is a fundamental problem which is encountered in a variety of signal processing problems where multiple s Blind Separation of Nonstationary Sources in Noisy Mixtures Seungjin CHOI x1 and Andrzej CICHOCKI y x Department of Electrical Engineering Chungbuk National University 48 Kaeshin-dong, Cheongju Chungbuk

More information

Analytical solution of the blind source separation problem using derivatives

Analytical solution of the blind source separation problem using derivatives Analytical solution of the blind source separation problem using derivatives Sebastien Lagrange 1,2, Luc Jaulin 2, Vincent Vigneron 1, and Christian Jutten 1 1 Laboratoire Images et Signaux, Institut National

More information

Wavelet de-noising for blind source separation in noisy mixtures.

Wavelet de-noising for blind source separation in noisy mixtures. Wavelet for blind source separation in noisy mixtures. Bertrand Rivet 1, Vincent Vigneron 1, Anisoara Paraschiv-Ionescu 2 and Christian Jutten 1 1 Institut National Polytechnique de Grenoble. Laboratoire

More information

Robust extraction of specific signals with temporal structure

Robust extraction of specific signals with temporal structure Robust extraction of specific signals with temporal structure Zhi-Lin Zhang, Zhang Yi Computational Intelligence Laboratory, School of Computer Science and Engineering, University of Electronic Science

More information

ORIENTED PCA AND BLIND SIGNAL SEPARATION

ORIENTED PCA AND BLIND SIGNAL SEPARATION ORIENTED PCA AND BLIND SIGNAL SEPARATION K. I. Diamantaras Department of Informatics TEI of Thessaloniki Sindos 54101, Greece kdiamant@it.teithe.gr Th. Papadimitriou Department of Int. Economic Relat.

More information

Independent Component Analysis. Contents

Independent Component Analysis. Contents Contents Preface xvii 1 Introduction 1 1.1 Linear representation of multivariate data 1 1.1.1 The general statistical setting 1 1.1.2 Dimension reduction methods 2 1.1.3 Independence as a guiding principle

More information

MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES

MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES S. Visuri 1 H. Oja V. Koivunen 1 1 Signal Processing Lab. Dept. of Statistics Tampere Univ. of Technology University of Jyväskylä P.O.

More information

Independent Component Analysis and Its Application on Accelerator Physics

Independent Component Analysis and Its Application on Accelerator Physics Independent Component Analysis and Its Application on Accelerator Physics Xiaoying Pang LA-UR-12-20069 ICA and PCA Similarities: Blind source separation method (BSS) no model Observed signals are linear

More information

NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group

NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION M. Schwab, P. Noll, and T. Sikora Technical University Berlin, Germany Communication System Group Einsteinufer 17, 1557 Berlin (Germany) {schwab noll

More information

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given

More information

Massoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39

Massoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39 Blind Source Separation (BSS) and Independent Componen Analysis (ICA) Massoud BABAIE-ZADEH Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39 Outline Part I Part II Introduction

More information

Statistical signal processing

Statistical signal processing Statistical signal processing Short overview of the fundamentals Outline Random variables Random processes Stationarity Ergodicity Spectral analysis Random variable and processes Intuition: A random variable

More information

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)

The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will

More information

Semi-Blind approaches to source separation: introduction to the special session

Semi-Blind approaches to source separation: introduction to the special session Semi-Blind approaches to source separation: introduction to the special session Massoud BABAIE-ZADEH 1 Christian JUTTEN 2 1- Sharif University of Technology, Tehran, IRAN 2- Laboratory of Images and Signals

More information

THEORETICAL CONCEPTS & APPLICATIONS OF INDEPENDENT COMPONENT ANALYSIS

THEORETICAL CONCEPTS & APPLICATIONS OF INDEPENDENT COMPONENT ANALYSIS THEORETICAL CONCEPTS & APPLICATIONS OF INDEPENDENT COMPONENT ANALYSIS SONALI MISHRA 1, NITISH BHARDWAJ 2, DR. RITA JAIN 3 1,2 Student (B.E.- EC), LNCT, Bhopal, M.P. India. 3 HOD (EC) LNCT, Bhopal, M.P.

More information

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag

A Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data

More information

Least square joint diagonalization of matrices under an intrinsic scale constraint

Least square joint diagonalization of matrices under an intrinsic scale constraint Least square joint diagonalization of matrices under an intrinsic scale constraint Dinh-Tuan Pham 1 and Marco Congedo 2 1 Laboratory Jean Kuntzmann, cnrs - Grenoble INP - UJF (Grenoble, France) Dinh-Tuan.Pham@imag.fr

More information

7. Symmetric Matrices and Quadratic Forms

7. Symmetric Matrices and Quadratic Forms Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value

More information

Blind Source Separation with a Time-Varying Mixing Matrix

Blind Source Separation with a Time-Varying Mixing Matrix Blind Source Separation with a Time-Varying Mixing Matrix Marcus R DeYoung and Brian L Evans Department of Electrical and Computer Engineering The University of Texas at Austin 1 University Station, Austin,

More information

BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES

BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES Dinh-Tuan Pham Laboratoire de Modélisation et Calcul URA 397, CNRS/UJF/INPG BP 53X, 38041 Grenoble cédex, France Dinh-Tuan.Pham@imag.fr

More information

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE

3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3.0 INTRODUCTION The purpose of this chapter is to introduce estimators shortly. More elaborated courses on System Identification, which are given

More information

Concentration Ellipsoids

Concentration Ellipsoids Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE

More information

BLIND SOURCE SEPARATION TECHNIQUES ANOTHER WAY OF DOING OPERATIONAL MODAL ANALYSIS

BLIND SOURCE SEPARATION TECHNIQUES ANOTHER WAY OF DOING OPERATIONAL MODAL ANALYSIS BLIND SOURCE SEPARATION TECHNIQUES ANOTHER WAY OF DOING OPERATIONAL MODAL ANALYSIS F. Poncelet, Aerospace and Mech. Eng. Dept., University of Liege, Belgium G. Kerschen, Aerospace and Mech. Eng. Dept.,

More information

Natural Gradient Learning for Over- and Under-Complete Bases in ICA

Natural Gradient Learning for Over- and Under-Complete Bases in ICA NOTE Communicated by Jean-François Cardoso Natural Gradient Learning for Over- and Under-Complete Bases in ICA Shun-ichi Amari RIKEN Brain Science Institute, Wako-shi, Hirosawa, Saitama 351-01, Japan Independent

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.011: Introduction to Communication, Control and Signal Processing QUIZ, April 1, 010 QUESTION BOOKLET Your

More information

ON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS

ON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS Yugoslav Journal of Operations Research 5 (25), Number, 79-95 ON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS Slavica TODOROVIĆ-ZARKULA EI Professional Electronics, Niš, bssmtod@eunet.yu Branimir TODOROVIĆ,

More information

Tutorial on Blind Source Separation and Independent Component Analysis

Tutorial on Blind Source Separation and Independent Component Analysis Tutorial on Blind Source Separation and Independent Component Analysis Lucas Parra Adaptive Image & Signal Processing Group Sarnoff Corporation February 09, 2002 Linear Mixtures... problem statement...

More information

Blind channel deconvolution of real world signals using source separation techniques

Blind channel deconvolution of real world signals using source separation techniques Blind channel deconvolution of real world signals using source separation techniques Jordi Solé-Casals 1, Enric Monte-Moreno 2 1 Signal Processing Group, University of Vic, Sagrada Família 7, 08500, Vic

More information

Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides

Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering Stochastic Processes and Linear Algebra Recap Slides Stochastic processes and variables XX tt 0 = XX xx nn (tt) xx 2 (tt) XX tt XX

More information

Blind Source Separation Using Second-Order Cyclostationary Statistics

Blind Source Separation Using Second-Order Cyclostationary Statistics 694 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 4, APRIL 2001 Blind Source Separation Using Second-Order Cyclostationary Statistics Karim Abed-Meraim, Yong Xiang, Jonathan H. Manton, Yingbo Hua,

More information

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012

Machine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012 Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component

More information

Single-User MIMO systems: Introduction, capacity results, and MIMO beamforming

Single-User MIMO systems: Introduction, capacity results, and MIMO beamforming Single-User MIMO systems: Introduction, capacity results, and MIMO beamforming Master Universitario en Ingeniería de Telecomunicación I. Santamaría Universidad de Cantabria Contents Introduction Multiplexing,

More information

Chapter 7: Symmetric Matrices and Quadratic Forms

Chapter 7: Symmetric Matrices and Quadratic Forms Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved

More information

Singular Value Decomposition

Singular Value Decomposition Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =

More information

Independent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego

Independent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego Independent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego Email: brao@ucsdedu References 1 Hyvarinen, A, Karhunen, J, & Oja, E (2004) Independent component analysis (Vol 46)

More information

ADAPTIVE ANTENNAS. SPATIAL BF

ADAPTIVE ANTENNAS. SPATIAL BF ADAPTIVE ANTENNAS SPATIAL BF 1 1-Spatial reference BF -Spatial reference beamforming may not use of embedded training sequences. Instead, the directions of arrival (DoA) of the impinging waves are used

More information

Acoustic Source Separation with Microphone Arrays CCNY

Acoustic Source Separation with Microphone Arrays CCNY Acoustic Source Separation with Microphone Arrays Lucas C. Parra Biomedical Engineering Department City College of New York CCNY Craig Fancourt Clay Spence Chris Alvino Montreal Workshop, Nov 6, 2004 Blind

More information

Blind Identification of FIR Systems and Deconvolution of White Input Sequences

Blind Identification of FIR Systems and Deconvolution of White Input Sequences Blind Identification of FIR Systems and Deconvolution of White Input Sequences U. SOVERINI, P. CASTALDI, R. DIVERSI and R. GUIDORZI Dipartimento di Elettronica, Informatica e Sistemistica Università di

More information

HST.582J/6.555J/16.456J

HST.582J/6.555J/16.456J Blind Source Separation: PCA & ICA HST.582J/6.555J/16.456J Gari D. Clifford gari [at] mit. edu http://www.mit.edu/~gari G. D. Clifford 2005-2009 What is BSS? Assume an observation (signal) is a linear

More information

A Canonical Genetic Algorithm for Blind Inversion of Linear Channels

A Canonical Genetic Algorithm for Blind Inversion of Linear Channels A Canonical Genetic Algorithm for Blind Inversion of Linear Channels Fernando Rojas, Jordi Solé-Casals, Enric Monte-Moreno 3, Carlos G. Puntonet and Alberto Prieto Computer Architecture and Technology

More information

A more efficient second order blind identification method for separation of uncorrelated stationary time series

A more efficient second order blind identification method for separation of uncorrelated stationary time series A more efficient second order blind identification method for separation of uncorrelated stationary time series Sara Taskinen 1a, Jari Miettinen a, Klaus Nordhausen b,c a Department of Mathematics and

More information

Real-time Blind Source Separation

Real-time Blind Source Separation Real-time Blind Source Separation A Feasibility Study using CUDA th semester project, AAU, Applied Signal Processing and Implementation Spring 9 Group 4 Søren Reinholt Søndergaard Martin Brinch Sørensen

More information

Introduction to Machine Learning

Introduction to Machine Learning 10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what

More information

Econ Slides from Lecture 7

Econ Slides from Lecture 7 Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for

More information

DOA Estimation using MUSIC and Root MUSIC Methods

DOA Estimation using MUSIC and Root MUSIC Methods DOA Estimation using MUSIC and Root MUSIC Methods EE602 Statistical signal Processing 4/13/2009 Presented By: Chhavipreet Singh(Y515) Siddharth Sahoo(Y5827447) 2 Table of Contents 1 Introduction... 3 2

More information

Computational Methods. Eigenvalues and Singular Values

Computational Methods. Eigenvalues and Singular Values Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations

More information

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random

More information

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007 MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions

YORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 3. M Test # Solutions. (8 pts) For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For this

More information

Performance Analysis for Strong Interference Remove of Fast Moving Target in Linear Array Antenna

Performance Analysis for Strong Interference Remove of Fast Moving Target in Linear Array Antenna Performance Analysis for Strong Interference Remove of Fast Moving Target in Linear Array Antenna Kwan Hyeong Lee Dept. Electriacal Electronic & Communicaton, Daejin University, 1007 Ho Guk ro, Pochen,Gyeonggi,

More information

Independent Component Analysis and Its Applications. By Qing Xue, 10/15/2004

Independent Component Analysis and Its Applications. By Qing Xue, 10/15/2004 Independent Component Analysis and Its Applications By Qing Xue, 10/15/2004 Outline Motivation of ICA Applications of ICA Principles of ICA estimation Algorithms for ICA Extensions of basic ICA framework

More information

A METHOD OF ICA IN TIME-FREQUENCY DOMAIN

A METHOD OF ICA IN TIME-FREQUENCY DOMAIN A METHOD OF ICA IN TIME-FREQUENCY DOMAIN Shiro Ikeda PRESTO, JST Hirosawa 2-, Wako, 35-98, Japan Shiro.Ikeda@brain.riken.go.jp Noboru Murata RIKEN BSI Hirosawa 2-, Wako, 35-98, Japan Noboru.Murata@brain.riken.go.jp

More information

Singular Value Decomposition and Principal Component Analysis (PCA) I

Singular Value Decomposition and Principal Component Analysis (PCA) I Singular Value Decomposition and Principal Component Analysis (PCA) I Prof Ned Wingreen MOL 40/50 Microarray review Data per array: 0000 genes, I (green) i,i (red) i 000 000+ data points! The expression

More information

BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS

BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS M. Kawamoto 1,2, Y. Inouye 1, A. Mansour 2, and R.-W. Liu 3 1. Department of Electronic and Control Systems Engineering,

More information

Name of the Student: Problems on Discrete & Continuous R.Vs

Name of the Student: Problems on Discrete & Continuous R.Vs Engineering Mathematics 05 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA6 MATERIAL NAME : University Questions MATERIAL CODE : JM08AM004 REGULATION : R008 UPDATED ON : Nov-Dec 04 (Scan

More information

926 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 3, MARCH Monica Nicoli, Member, IEEE, and Umberto Spagnolini, Senior Member, IEEE (1)

926 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 3, MARCH Monica Nicoli, Member, IEEE, and Umberto Spagnolini, Senior Member, IEEE (1) 926 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 3, MARCH 2005 Reduced-Rank Channel Estimation for Time-Slotted Mobile Communication Systems Monica Nicoli, Member, IEEE, and Umberto Spagnolini,

More information

Blind Source Separation via Generalized Eigenvalue Decomposition

Blind Source Separation via Generalized Eigenvalue Decomposition Journal of Machine Learning Research 4 (2003) 1261-1269 Submitted 10/02; Published 12/03 Blind Source Separation via Generalized Eigenvalue Decomposition Lucas Parra Department of Biomedical Engineering

More information

Numerical Methods I Singular Value Decomposition

Numerical Methods I Singular Value Decomposition Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Principal

More information

Lecture 19 Observability and state estimation

Lecture 19 Observability and state estimation EE263 Autumn 2007-08 Stephen Boyd Lecture 19 Observability and state estimation state estimation discrete-time observability observability controllability duality observers for noiseless case continuous-time

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,

More information

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:

VAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where: VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector

More information

Lecture 3: Review of Linear Algebra

Lecture 3: Review of Linear Algebra ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak, scribe: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters,

More information

Advanced Introduction to Machine Learning CMU-10715

Advanced Introduction to Machine Learning CMU-10715 Advanced Introduction to Machine Learning CMU-10715 Independent Component Analysis Barnabás Póczos Independent Component Analysis 2 Independent Component Analysis Model original signals Observations (Mixtures)

More information

EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4]

EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: Au = λu λ C : eigenvalue u C n : eigenvector Example:

More information

Large Scale Data Analysis Using Deep Learning

Large Scale Data Analysis Using Deep Learning Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset

More information

Second-order statistics based blind source separation using a bank of subband filters

Second-order statistics based blind source separation using a bank of subband filters Digital Signal Processing 13 (2003) 252 274 www.elsevier.com/locate/dsp Second-order statistics based blind source separation using a bank of subband filters R.R. Gharieb a,b, and A. Cichocki a,c a Laboratory

More information

Estimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition

Estimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition Estimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition Seema Sud 1 1 The Aerospace Corporation, 4851 Stonecroft Blvd. Chantilly, VA 20151 Abstract

More information

Vector and Matrix Norms. Vector and Matrix Norms

Vector and Matrix Norms. Vector and Matrix Norms Vector and Matrix Norms Vector Space Algebra Matrix Algebra: We let x x and A A, where, if x is an element of an abstract vector space n, and A = A: n m, then x is a complex column vector of length n whose

More information

A New High-Resolution and Stable MV-SVD Algorithm for Coherent Signals Detection

A New High-Resolution and Stable MV-SVD Algorithm for Coherent Signals Detection Progress In Electromagnetics Research M, Vol. 35, 163 171, 2014 A New High-Resolution and Stable MV-SVD Algorithm for Coherent Signals Detection Basma Eldosouky, Amr H. Hussein *, and Salah Khamis Abstract

More information

1254 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 4, APRIL On the Virtual Array Concept for Higher Order Array Processing

1254 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 4, APRIL On the Virtual Array Concept for Higher Order Array Processing 1254 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 4, APRIL 2005 On the Virtual Array Concept for Higher Order Array Processing Pascal Chevalier, Laurent Albera, Anne Ferréol, and Pierre Comon,

More information

Recipes for the Linear Analysis of EEG and applications

Recipes for the Linear Analysis of EEG and applications Recipes for the Linear Analysis of EEG and applications Paul Sajda Department of Biomedical Engineering Columbia University Can we read the brain non-invasively and in real-time? decoder 1001110 if YES

More information

Master Thesis Separation of Non-Stationary Sources

Master Thesis Separation of Non-Stationary Sources Master Thesis Separation of Non-Stationary Sources Lisbeth Andersson Perth, March 21 Australian Telecommunications Research Institute Curtin University of Technology Australia i Preface This master thesis

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Laurenz Wiskott Institute for Theoretical Biology Humboldt-University Berlin Invalidenstraße 43 D-10115 Berlin, Germany 11 March 2004 1 Intuition Problem Statement Experimental

More information

BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES

BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES Dinh-Tuan Pham Laboratoire de Modélisation et Calcul URA 397, CNRS/UJF/INPG BP 53X, 38041 Grenoble cédex, France Dinh-Tuan.Pham@imag.fr

More information

GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS. Mitsuru Kawamoto 1,2 and Yujiro Inouye 1

GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS. Mitsuru Kawamoto 1,2 and Yujiro Inouye 1 GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS Mitsuru Kawamoto,2 and Yuiro Inouye. Dept. of Electronic and Control Systems Engineering, Shimane University,

More information

Overview of Beamforming

Overview of Beamforming Overview of Beamforming Arye Nehorai Preston M. Green Department of Electrical and Systems Engineering Washington University in St. Louis March 14, 2012 CSSIP Lab 1 Outline Introduction Spatial and temporal

More information

Tensor approach for blind FIR channel identification using 4th-order cumulants

Tensor approach for blind FIR channel identification using 4th-order cumulants Tensor approach for blind FIR channel identification using 4th-order cumulants Carlos Estêvão R Fernandes Gérard Favier and João Cesar M Mota contact: cfernand@i3s.unice.fr June 8, 2006 Outline 1. HOS

More information

/16/$ IEEE 1728

/16/$ IEEE 1728 Extension of the Semi-Algebraic Framework for Approximate CP Decompositions via Simultaneous Matrix Diagonalization to the Efficient Calculation of Coupled CP Decompositions Kristina Naskovska and Martin

More information

A Root-MUSIC-Like Direction Finding Method for Cyclostationary Signals

A Root-MUSIC-Like Direction Finding Method for Cyclostationary Signals EURASIP Journal on Applied Signal Processing 25:1, 69 73 c 25 Hindawi Publishing Corporation A Root-MUSIC-Like Direction Finding Method for Cyclostationary Signals Pascal Chargé LESIA, DGEI, INSA Toulouse,

More information

1 Principal component analysis and dimensional reduction

1 Principal component analysis and dimensional reduction Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an

More information

Principal Component Analysis

Principal Component Analysis Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature

More information

Data Mining Lecture 4: Covariance, EVD, PCA & SVD

Data Mining Lecture 4: Covariance, EVD, PCA & SVD Data Mining Lecture 4: Covariance, EVD, PCA & SVD Jo Houghton ECS Southampton February 25, 2019 1 / 28 Variance and Covariance - Expectation A random variable takes on different values due to chance The

More information

Lecture: Face Recognition and Feature Reduction

Lecture: Face Recognition and Feature Reduction Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the

More information

Linear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg

Linear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg Linear Algebra, part 3 Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2010 Going back to least squares (Sections 1.7 and 2.3 from Strang). We know from before: The vector

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas

Dimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx

More information

Let A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1

Let A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1 Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Au = λu Example: ( ) 2 0 A = 2 1 λ 1 = 1 with eigenvector

More information

A Subspace Approach to Estimation of. Measurements 1. Carlos E. Davila. Electrical Engineering Department, Southern Methodist University

A Subspace Approach to Estimation of. Measurements 1. Carlos E. Davila. Electrical Engineering Department, Southern Methodist University EDICS category SP 1 A Subspace Approach to Estimation of Autoregressive Parameters From Noisy Measurements 1 Carlos E Davila Electrical Engineering Department, Southern Methodist University Dallas, Texas

More information

IV. Matrix Approximation using Least-Squares

IV. Matrix Approximation using Least-Squares IV. Matrix Approximation using Least-Squares The SVD and Matrix Approximation We begin with the following fundamental question. Let A be an M N matrix with rank R. What is the closest matrix to A that

More information

MIMO instantaneous blind identification based on second-order temporal structure

MIMO instantaneous blind identification based on second-order temporal structure MIMO instantaneous blind identification based on second-order temporal structure Citation for published version (APA): Laar, van de, J, Moonen, M, & Sommen, P C W (2008) MIMO instantaneous blind identification

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Empirical Macroeconomics

Empirical Macroeconomics Empirical Macroeconomics Francesco Franco Nova SBE April 5, 2016 Francesco Franco Empirical Macroeconomics 1/39 Growth and Fluctuations Supply and Demand Figure : US dynamics Francesco Franco Empirical

More information

VARIANCE COMPUTATION OF MODAL PARAMETER ES- TIMATES FROM UPC SUBSPACE IDENTIFICATION

VARIANCE COMPUTATION OF MODAL PARAMETER ES- TIMATES FROM UPC SUBSPACE IDENTIFICATION VARIANCE COMPUTATION OF MODAL PARAMETER ES- TIMATES FROM UPC SUBSPACE IDENTIFICATION Michael Döhler 1, Palle Andersen 2, Laurent Mevel 1 1 Inria/IFSTTAR, I4S, Rennes, France, {michaeldoehler, laurentmevel}@inriafr

More information

Math 307 Learning Goals. March 23, 2010

Math 307 Learning Goals. March 23, 2010 Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent

More information

Growing Window Recursive Quadratic Optimization with Variable Regularization

Growing Window Recursive Quadratic Optimization with Variable Regularization 49th IEEE Conference on Decision and Control December 15-17, Hilton Atlanta Hotel, Atlanta, GA, USA Growing Window Recursive Quadratic Optimization with Variable Regularization Asad A. Ali 1, Jesse B.

More information

POLYNOMIAL SINGULAR VALUES FOR NUMBER OF WIDEBAND SOURCES ESTIMATION AND PRINCIPAL COMPONENT ANALYSIS

POLYNOMIAL SINGULAR VALUES FOR NUMBER OF WIDEBAND SOURCES ESTIMATION AND PRINCIPAL COMPONENT ANALYSIS POLYNOMIAL SINGULAR VALUES FOR NUMBER OF WIDEBAND SOURCES ESTIMATION AND PRINCIPAL COMPONENT ANALYSIS Russell H. Lambert RF and Advanced Mixed Signal Unit Broadcom Pasadena, CA USA russ@broadcom.com Marcel

More information

Blind Machine Separation Te-Won Lee

Blind Machine Separation Te-Won Lee Blind Machine Separation Te-Won Lee University of California, San Diego Institute for Neural Computation Blind Machine Separation Problem we want to solve: Single microphone blind source separation & deconvolution

More information