BLIND source separation is a set of techniques that is used
|
|
- Rodney Cameron
- 5 years ago
- Views:
Transcription
1 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK A Novel Approach to Blind Separation of Delayed Sources in Linear Mixtures Jakob Ashtar, Kristoffer Møller Jørgensen, Jacob Lindvig, Lars Juul Mikkelsen, Jakob Birkedal Nielsen, Lars Sommer Søndergaard, and Dario Farina, Supervisor, Aalborg University Abstract This paper addresses the problem of blind estimation of time-delayed sources from a set of linear mixtures. The mixtures are analyzed and approximated by a linear, instantaneous signal model and the sources are recovered using second order statistical methods. An algorithm is implemented and test results are demonstrated and discussed. The proposed algorithm is referred to as SOBIDS Second Order Blind Identification of Dealyed Sources). SOBIDS is compared to another algorithm using second order statistics and test results show that SOBIDS performs better for delayed sources. I. INTRODUCTION BLIND source separation is a set of techniques that is used in many fields of engineering such as speech recognition, telecommunications, and biomedical signal processing. One of the first papers describing the problem of recovering sources was published in 985 by Herault et al. [5] These authors addressed the problem by using and extending principal component analysis. Various approaches have been proposed since then. Many of them are based on mathematical disciplines such as principal component analysis, independent component analysis, or singular value decomposition [] [2]. The classical cocktail party problem where two persons communicate in a room with loud music and background noise is an example of a situation where the human brain utilizes blind source separation. In this case the human brain focuses on a speaker, distinguishes one sound from the others, and ignores the background noise. Simulating this process with machines is, however, a very difficult task. The wide range and nature of applications further complicates the task. In some cases the sources can be assumed stationary. In other cases, e.g. speech signals, the sources are non-stationary. Signal propagation delays might also be relevant to consider. Also, the mixtures of sources may be either instantaneous and linear or convolutive. The aim of this paper is to propose a new method SOBIDS) for the separation of sources mixed with unknown, small delays in a linear, instantaneous way. The paper is organized as follows. In section II and III we provide a description of a blind identification algorithm previously proposed to separate sources from their linear, instantaneous mixtures using second order statistics []. This algorithm is referred to as SOBI Second Order Blind Identification). Section IV introduces a convolutive signal model where the sources are delayed. To solve the blind source separation problem for a convolutive model we approximate the model with a linear, instantaneous signal model. In section V we show that it is possible to blindly estimate the sources for small delays. In section VI we compare test results for SOBI and SOBIDS and discuss the performance of the new algorithm. II. LINEAR INSTANTANEOUS MIXTURES Assuming linear, instantaneous mixtures we have the following signal model. A. Signal Model Let a set of n unknown sources s t),s 2 t),...,s j t),...,s n t) be linearly combined to form m known observations y t),y 2 t),...,y i t),...,y m t). Each observation y i t) is contaminated by unknown additive noise p i t). Expressed mathematically we get the following relation: y t) = a s t)+a 2 s 2 t)+...+ a n s n t)+ p t) y 2 t) = a 2 s t)+a 22 s 2 t)+...+ a 2n s n t)+ p 2 t). y m t) = a m s t)+a m2 s 2 t)+...+ a mn s n t)+ p m t) ) The above system of linear equations can also be represented in matrix notation as a transformation T : R n R m of the source vector st): yt) = Tst)) = xt)+pt) = Ast)+pt) 2) where the observation vector yt), the signal vector xt), and the noise vector pt) are m column vectors and the source vector st) is an n column vector. The unknown m n matrix A is called the mixing matrix. B. Assumptions It is assumed that the sources are real, ergodic, zero mean stochastic processes. Since the power of the sources cannot be determined [], it is assumed without loss of generality that the sources have unit variance. Furthermore, the sources are assumed to be mutually uncorrelated spatially white). It is also assumed that the sources are uncorrelated with the additive noise. The noise process is a stationary, temporally and spatially white, zero-mean, and real random process. The noise variance is the same for all observations. The mixing matrix has full rank. Furthermore, it is assumed that there are at least as many observations as sources, that is m n.
2 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 2 C. Derivations Based on the signal model and the above assumptions the n n covariance matrix of the sources R s τ) is given by: R s τ) = E [ st)s T t + τ) ] = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 3) where R s j s j τ) = E [ s j t)s j t +τ) ]. Since the sources are mutually uncorrelated and assumed to have unit variance, we have: R s 0) = I 4) As the noise is temporally and spatially white the m m covariance matrix R p τ) of the noise is 0 at time lags τ 0. At τ = 0 the covariance matrix R p τ) is the identity matrix scaled by the noise variance σ 2 p : R p τ) = E [ pt)p T t + τ) ] = σ 2 pδτ)i 5) where δτ) is the Kronecker delta. The m m covariance matrix R y τ) of the observations is given by: R y τ) = E [ yt)y T t + τ) ] = E [ Ast)+pt))Ast + τ)+pt + τ)) T] = E [ Ast)s T t + τ)a T + pt)p T t + τ) ] = AR s τ)a T + σ 2 pδτ)i 6) = R x τ)+σ 2 p δτ)i where R x τ) is the covariance matrix of the signal vector. III. THE SOBI ALGORITHM The method for blind separation described in [] is based on two steps: whitening and rotation. The whitening step requires an estimation of the noise variance from the observations. A. Estimation of the Noise Variance The first step is to estimate the noise variance σ 2 p. This can be done by applying 6) for τ = 0 which is the only time instant where the noise variance is a part of the equation: R y 0) = AR s 0)A T + σ 2 p δ0)i = AIA T + σ 2 pi = AA T + σ 2 p I 7) Since AA T is a symmetric matrix it is also orthogonally diagonalizable [3, pp. 445]. If we orthogonally diagonalize AA T we can rewrite 7) as: R y 0) = UDU T + σ 2 pi 8) where the column vectors of U are the orthonormal eigenvectors of the positive definite matrix AA T. The corresponding positive and sorted eigenvalues [3, pp. 456]: are in the diagonal matrix λ λ n λ m 0 9) D = diag[λ,λ 2,...,λ n,λ n+,...,λ m ] 0) Understanding that the right side of 8) is, in fact, an orthogonal diagonalization of R y 0) becomes more easy when the equation is written in a different way: R y 0) = UDU T + Uσ 2 p IUT = UD+σ 2 pi)u T = UVU T ) From ) we observe that the eigenvalues v ii in the diagonal of V contain the noise variance: v ii = λ i + σ 2 p 2) From our assumptions the mixing matrix A has full rank equal to n). Since the column vectors of AA T are linear combinations of the column vectors of A we can derive that: λ i = 0, i > n 3) This means that the last m n eigenvalues in V are: v ii = σ 2 p, i > n 4) Thus the noise variance can be estimated by calculating an average of the last m n eigenvalues v ii in V, that is: B. Whitening σ 2 p = m n m i=n+ v ii 5) From the definition of the transform in 2) we see that the n-dimensional source vector is mapped from R n to an m- dimensional signal vector in R m. Disregarding the noise vector we notice that the signal vector is a linear combination of the n linearly independent m-dimensional column vectors of A. Since there are only n linearly independent column vectors they only span R n and not all of R m. As a consequence the representation of the source vector will contain redundant information when m > n. Another problem is that the source vector is mapped from an orthogonal coordinate system to a coordinate system which is not guaranteed to be orthogonal. To eliminate the redundant information and to ensure that the source vector is mapped into an orthogonal coordinate system we introduce the whitening matrix W. The n m whitening matrix W is defined in the following way: WR x 0)W T = I WAR s 0)A T W T = I WAIA T W T = I WAA T W T = I 6) such that the transformed sources are uncorrelated at time lag τ = 0. Thus, the whitening matrix has the following property: WA)WA) T = WAA T W T = I 7) If we observe what happens when the whitening matrix is applied to the observation vector: Wyt) = Wxt)+Wpt) = WAst)+Wpt) 8)
3 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 3 we see that the source vector is now mapped from R n to R n no redundant information) and that the transformation of the signal vector is orthogonal since WA is orthogonal). To derive an expression for the whitening matrix we continue by substituting AA T in 7) with UDU T : WUDU T W T = I 9) The last m n eigenvalues in the matrix D are zero. Thus we can reduce the size of D and U without affecting the diagonalization of AA T, that is: AA T = U r D r U T r 20) where the n n reduced eigenvalue matrix D r is: D r = diag[λ,λ 2,...,λ n ] 2) Column vector j in the m n matrix U r is the eigenvector corresponding to the eigenvalue λ j. The matrix V is reduced in a similar way, such that: We substitute UDU T in 9) with U r D r U T r and replace D r with V r σ 2 pi: V r = D r + σ 2 p I 22) and get: WU r D r U T r W T = I 23) WU r Vr σ 2 p I) U T r WT = I 24) Rearranging the equation such that it reads: WU r Vr σ 2 pi ) 2 V r σ 2 pi ) 2 U T r W T = I 25) [WU r Vr σ 2p I) 2 ][ WU r Vr σ 2 p I) 2 ] T = I 26) we see that it is possible to obtain a closed analytical expression for the whitening matrix by solving the following equation: WU r Vr σ 2 p I) 2 = I 27) under the assumption that an estimate of the noise variance is known. The matrices U r and V r can be obtained by orthogonally diagonalizing R y 0). The expression V r σ 2 pi ) 2 is calculated by taking the square root of each entry in the diagonal matrix V r σ 2 pi. As a conclusion we derive the following expression for the whitening matrix: C. Estimation of the Mixing Matrix W = V r σ 2 p I) 2 U T r 28) The final step in recovering the sources is to estimate the mixing matrix A. To get an estimate of the mixing matrix we choose to investigate the whitened covariance matrix R z τ) at different time lags τ 0: R z τ) = WR y τ)w T = WAR s τ)a T W T = WAR s τ)wa) T = QR s τ)q T, τ 0 29) and realize that the right side of the equation is an orthogonal diagonalization of R z τ). Consequently, the key idea here, since W and R y τ) can be estimated, is to solve 29) with respect to Q at K different time lags τ K > τ K > τ K 2 > τ 2 > τ > 0. The set of solutions {Q,Q 2,...,Q K,Q K } will be the basis for calculating an estimate Q of Q and hence also an estimate of the mixing matrix since: Q = ŴÂ 30) We rearrange the equation to obtain an expression for the estimate of the mixing matrix: Â = Ŵ Q 3) where Ŵ is the pseudo-inverse of Ŵ. Since it is beyond the scope of this paper to delve further into how an estimate is obtained, we refer to [] which in greater detail explains how Q is estimated using a technique called joint diagonalization. IV. MIXTURES WITH DELAYED SOURCES When propagation delays are introduced the signal model changes from a model with linear, instantaneous mixtures to a convolutive model. In the following sections we will approximate this convolutive model with a linear, instantaneous model in which the sources can be estimated using second order statistics. A. Convolutive Signal Model Let a set of n unknown sources s t),s 2 t),...,s j t),...,s n t) be linearly combined with unknown delays t i j to form m observations y t),y 2 t),...,y i t),...,y m t). Each observation y i t) is contaminated by unknown additive noise p i t). This model can be expressed mathematically in the following way: y t) = y 2 t) =. y m t) = a s t t )+...+ a n s n t t n )+ p t) a 2 s t t 2 )+...+ a 2n s n t t 2 )+ p 2 t) a m s t t m )+...+ a mn s n t t mn )+ p m t) B. Approximation of a Delayed Source 32) From the definition of the derivative: st + ε) st) ṡt) = lim 33) ε 0 ε we arrive at the following approximation when substituting ε with a small delay t i j : which shows that: t i j ṡt) st t i j ) st) 34) st t i j ) st) t i j ṡt) 35) This approximation corresponds to the Taylor expansion truncated at the first order and is used in [2] where a delay is assumed small if: t i j 36) 2 π fmax where f max denotes the maximum frequency in st).
4 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 4 C. An Approximation of the Convolutive Signal Model From 35) it is possible to define a linear, instantaneous signal model, which is an approximation of the convolutive signal model. The approximated signal model is defined as: y i t) n j= or equivalently in matrix notation as: a i j s j t) t i j ṡ j t))+ p i t) 37) yt) Ast)+pt) 38) where the m 2n mixing matrix A is defined as: a a n a t σṡ a n t n σṡn a 2 a 2n a 2 t 2 σṡ a 2n t 2n σṡn A = a m a mn a m t m σṡ a mn t mn σṡn 39) The 2n source vector st) contains the sources and their normalized derivatives and is defined as: st) = [s t),s 2 t),...,s n t), ṡt), ṡ2t),..., ṡnt) ] T σṡ σṡ2 σṡn = [s t),s 2 t),...,s n t), s t), s 2 t),..., s n t)] T 40) where s j t) is the derivative of s j t) normalized with respect to the standard deviation σṡ j of ṡ j t). D. Assumptions The sources and noise have the same statistical properties as defined in II-B. Hence the derivatives are zero mean, second order stationary, ergodic, stochastic processes. Moreover, the derivatives are uncorrelated with the noise and also spatially white. Finally it is assumed that there are at least twice as many observations as sources, that is m 2n. The mixing matrix A has full rank equal to 2n. E. Derivations Based on the approximative signal model and the above assumptions the 2n 2n covariance matrix R s τ) is given by: [ ] Rss τ) R R s τ) = s s τ) 4) R ss τ) R s s τ) where the n n quadrant matrices R ss τ),r s s τ),r ss τ) and R s s τ) are defined as: R ss τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 42) R s s τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 43) R ss τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 44) R s s τ) = diag [ R s s τ),r s2 s 2 τ),...,r sn s n τ) ] 45) Considering that the sources are mutually uncorrelated we have that: R ss 0) = I 46) Considering that the derivatives of the sources are mutually uncorrelated we have that: R s s 0) = I 47) From [4, pp. 65] and [4, pp. 44] we know that: and R s j ṡ j = Ṙ s j s j R s j s j τ) = Ṙs j s j τ) σṡ j R s j s j τ) = R s j s j τ) = Ṙs j s j τ) σṡ j 48) Since σ ṡ j Ṙ s j s j τ) is symmetric around τ = 0 we have that R s j s j τ) = σ ṡ j Ṙ s j s j τ) = σ ṡ j Ṙ s j s j τ) = R s j s j τ). Furthermore, we know that R s j s j τ) has a maximum for τ = 0. Based on this we derive that Ṙ s j s j 0) = 0. Combining the two findings we conclude that the covariance matrix R s τ) at τ = 0 is: R s 0) = [ ] I 0 = I 49) 0 I Due to the fact that the noise is temporally and spatially white the m m covariance matrix R p τ) of the noise is 0 at time lags τ 0. At τ = 0 the covariance matrix R p τ) is the identity matrix scaled by the noise variance σ 2 p: R p τ) = E [ pt)p T t + τ) ] = σ 2 pδτ)i 50) where δτ) is the Kronecker delta. The m m covariance matrix R y τ) of the observations is given by: R y τ) = E [ yt)y T t + τ) ] = E [ Ast)+pt))Ast + τ)+pt + τ)) T] = E [ Ast)s T t + τ)a T + pt)p T t + τ) ] = AR s τ)a T + σ 2 pδτ)i 5) = R x τ)+σ 2 pδτ)i where R x τ) is the covariance matrix of the signal vector. V. METHOD In the following text we will show that it is possible to estimate the sources using second order statistics. A. Estimation of the Whitening Matrix and the Noise Variance For τ = 0 we have: R y 0) = AR s 0)A T + σ 2 pi = AA T + σ 2 p I 52) Since 52) is similar to 7) we can estimate the whitening matrix and the noise variance using the method described in section III-A and III-B.
5 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 5 B. Estimation of the Mixing Matrix For τ 0 the whitened covariance matrix of the observations is: WR y τ)w T = WAR s τ)a T W T 53) The covariance matrix of the sources R s τ) is in general non-diagonal. We note that the entries in 43) and 44) are the off-diagonal entries in R s τ). These entries have been shown to be related to the derivative of the autocorrelation function of the sources. The size of the off-diagonal entries compared to the size of the main diagonal entries depends on the autocorrelation functions of the sources and, thus, on the bandwidth and frequency content of each source. To estimate the mixing matrix we use the same method as described in section III-C. Due to the fact that R s τ) has to be diagonal or close to diagonal in order to use this method the performance of the SOBIDS algorithm will depend on the frequency characteristics of the sources. Normalized Amplitude Original Source Estimated Source Original Source Estimated Source Samples C. Implementation Based on our analysis we introduce a second order blind identification algorithm for delayed sources SOBIDS). SO- BIDS is implemented in the following way: ) Estimate the sample covariance R y 0) from L data samples. Denote by λ,λ 2,...,λ 2n the 2n largest eigenvalues and u,u 2,...,u 2n the corresponding eigenvectors of R y 0). 2) Calculate an estimate of the noise variance σ 2 p as the average of the m 2n smallest eigenvalues of R y 0). Calculate an estimate of the whitening matrix as: Ŵ = [ λ σ 2 p) 2 u,...,λ 2n σ 2 p) 2 ] T u 2n 3) Calculate the sample covariance matrix R z τ) of Ŵyt) for a fixed set of time lags τ {τ j j =,...,K}. 4) Calculate an estimate Q of the orthogonal matrix that is a joint diagonalizer of the set { R z τ j ) j =,...,K}. 5) Calculate an estimate of the sources and their normalized derivatives as ŝt) = Q T Ŵyt). 6) Calculate an estimate of the mixing matrix as  = Ŵ# Q. D. Delay Estimation In the following we describe a method for estimating the relative delays using the estimated source vector ŝt) and the estimated mixing matrix Â. We calculate the derivatives of the sources in ŝt) and then the variance of these derivatives to get an estimate of σṡi. Using these estimates and  it is possible to estimate the delays. Note that we have no information on the order of the entries of ŝt) or the arrangement of the columns in Â. Since ŝt) is calculated using the invers of  we do know that the order of the entries in ŝt) corresponds to the order of the columns of Â. By calculating the derivative of each entry in ŝt) we state that it is possible to do a matching of each estimated source ŝ i t) with the corresponding estimated normalized derivative s i t). This provides sufficient information to calculate the relative delay of each source. Fig. : An example of the goodness factor for two source estimates. The goodness factor is 0.93 and 0.72 for top and bottom respectively. VI. RESULTS In this section the results of our performed test of the SOBIDS and SOBI algorithms will be presented. The intention of this paper is to describe an algorithm based on principles from the SOBI algorithm but with increased performance when estimating sources with small delays. Therefore the algorithms will be tested with delayed sources and the results will be compared. These tests will be performed with noiseless and noise contaminated observations. A. Performance Index To identify and compare the performance of the two algorithms a performance index has been used. This index is based on the cross-correlation of each generated source and the corresponding estimated source and is defined as: G = n n j= max{ R s j ŝ j } 54) where max{ R s j ŝ j } is the maximum value of the normalized and absolute cross-correlation between the original source sequence s j and the estimated source sequence ŝ j. Since there are n sources we also have n maximum values and the performance parameter is therefore an average of these n maximum values. The idea is to study the behavior of G which indicates how well the algorithm performs under different circumstances, e.g. various delay values. We refer to G as the goodness factor. In fig. two different matchings of signals are shown with goodness factors of 0.72 and The maximum value of G is which corresponds to a full correlation of the matched signals.
6 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 6 B. Simulation Environment The sources used for the simulations are generated by filtering n stationary white noise sources with second order filters with partly overlapping passbands. The observations are generated as linear mixtures of delayed sources. Noise is added for each observation. In section IV-A this mixture is shown. Preliminary tests have been made to identify the parameters which influence the goodness of the algorithms. The tests show that the goodness is almost independent of a number of parameters which therefore were decided to be kept fixed for all simulations. The tests were thus performed with: 3 sources, 4 matrices to be jointly diagonalized, and an overlap ratio of the passband filters of 0% of the filter bandwidth. The filters were created randomly with equal bandwidth. With 3 sources and 0% overlap we will produce filters with a maximum bandwidth of 0.37 f s/2. The number of observations used in the simulation differs from SOBI to SOBIDS. For SOBI the number of observations is n+2 = 5 and for SOBIDS this value is 2n+2 = 8. Tests have shown that this difference is irrelevant for the performance. The parameters essential for the performance of the algorithms are the SNR of the observations and the size of the relative delay of each source. The delay values are randomly generated and the values will be uniformly distributed in an interval from 0 to a defined maximum value. We state that the single parameter which provides the best information on how a single source is delayed in the observations is the standard deviation of the delays for the source. We therefore define the following parameter σ d as the mean of the standard deviations of the delays for each source: σ d = n n j= Vart j,...,t m j ) 55) We refer to σ d as the mean standard deviation of delay. C. Testing the Algorithms with Noiseless Observations To create an ideal environment for testing the performance when using delayed sources no noise is added to the observations when simulating. Fig. 2 shows the result of a simulation with a variation of the mean standard deviation of delay from 0 to 3. The graph for the SOBI algorithm shows that at σ d = 0 the goodness is 0.96 decreasing significantly to G 0.76 at σ d = 0.5. From this point the performance is falling slightly to a goodness of 0.7 at σ d = 3. We note that at zero delay the goodness of the SOBIDS algorithm is 0.97 decreasing to G 0.94 at σ d = 0.. At 0. σ d 0.5 the performance is approximately constant with G From that point the performance is decreasing approximately linearly to G 0.83 at σ d = 3. Comparing the SOBI and the SOBIDS algorithms we notice that with no delay the performance of SOBIDS is slightly better than SOBI. For mean standard deviation of delays of 0.2 the SOBIDS is performing approximately 5% higher than Goodness, G SOBIDS algorithm SOBI algorithm Mean standard deviation of delay, σ d Fig. 2: Result of simulation with noiseless observations with σ d 3. Goodness, G SOBIDS algorithm SOBI algorithm Mean standard deviation of delay, σ d Fig. 3: Result of simulation with noiseless observations for σ d 30. SOBI and for σ d = 0.5 the performance is approximately 22% higher. At a mean standard deviation of delay of 3 the performance of the SOBIDS algorithm is 7% higher than the SOBI. Simulations with σ d up to 30 show that the performance decreases exponentially fig. 3). D. Testing the Algorithms with Noise Contaminated Observations The performance of the SOBI and the SOBIDS is tested with noise contaminated observations to examine the performance of the algorithms in a more demanding applications. The observations have a SNR of 0dB. Fig. 4 shows the result of a simulation with a variation of the mean standard deviation of delay from 0 to 3. The performance of SOBI with noisy observations at σ d = 0 is approximately 0.82 which is 5% lower than with noiseless
7 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 7 Goodness, G SOBIDS algorithm SOBI algorithm Mean standard deviation of delay, σ d Fig. 4: Simulation result of observations with 0dB SNR for σ d 3. Goodness, G SOBIDS algorithm SOBI algorithm E. Discussion The tests show that the proposed SOBIDS algorithm outperforms SOBI for linear mixtures of delayed sources. Adding noise has an influence on the overall performance characteristics. Since the full effect of adding noise is far from tested in details this is an interesting subject for further study concerning the robustness and effeciency of the SOBIDS algorithm in real-world cases. The performance of the algorithms with noiseless observations decreases as the relative delay increases. This can be explained from section IV-B concerning Taylor series expansion. 35) is an approximation which is only valid for small delays. The SOBIDS uses the first order Taylor series expansion and SOBI uses a zero order Taylor series expansion which explains why the sources recovered from SOBI become inaccurate at smaller delays than the SOBIDS algorithm. As expected, the performance of both algorithms decreases when noise is added. The larger the delay the smaller the difference. For 0.3 σ d 3 the performance decrease of the simulations made with noisy sources is for the SOBI algorithm between 4% and 8% and for SOBIDS between 7% and 3%. This indicates that SOBIDS is slightly less noise robust than the SOBI algorithm. The first order Taylor series expansion has a positive effect on the performance for relatively small delays. For relatively larger delays it would be interesting to recalculate the expressions for the Taylor series and expanding it to higher orders. This might result in a more constant overall performance. As a concequence more observations would be needed and there is still an unsolved problem of whitening as the second derivative of a source is not uncorrelated with the source even for time lag zero. This is left for further studies Mean standard deviation of delay, σ d Fig. 5: Simulation result of observations with 0dB SNR for σ d 30. observations. As delays become larger the difference in goodness decreases to 0.03 at σ d = 3 which corresponds to a drop in performance of 4%. Also for the SOBIDS the performance decreased when noise was added to the observations. The difference of the goodness between tests performed with noisy observations and with noiseless observations is 0. at zero delay corresponding to a performance decrease of approximately 9%. The goodness decreases to 0.79 at σ d = 0.5 which is a 9% drop in performance. At σ d = 3 the performance is 8% lower than the performance of the tests with noiseless observations. We note that there is a large decrease in performance in the interval 0 σ d 0.3. Simulations with σ d up to 30 show that the performance decreases exponentially shown in fig. 5. The goodness of the SOBIDS algorithm is approximately 0.05 higher than SOBI for σ d > 5. VII. CONCLUSION In this paper we present a novel method for blindly separating mutually uncorrelated sources from a convolutive linear mixture contaminated with white noise. The paper covers a signal model introducing an approximation of sources with small delays as a first order Taylor series expansion. The sources are estimated mainly in two steps, a whitening process in order to make the observations spatially white followed by a rotation of the whitened observations making the principal axes the estimated sources. Tests revealed that the proposed SOBIDS algorithm outperforms SOBI for linear mixtures of delayed sources. An analytical method for estimating the delays has been proposed. REFERENCES [] A. Belouchrani, K. Abed-Meraim, J. F. Cardoso and E. Moulines, "A Blind Source Separation Technique Using Second-Order Statistics", IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 45, NO. 2, FEBRUARY 997 [2] J. Barrère, G. Chabriel, "A Compact Sensor Array for Blind Separation of Sources", IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMSŮI: FUNDAMENTAL THEORY AND APPLICATIONS, VOL. 49, NO. 5, MAY 2002
8 GROUP 743, 7 TH SEMESTER SIGNAL PROCESSING, AUTUMN 2004, AALBORG UNIVERSITY, FREDRIK BAJERS VEJ 7A, 9220 AALBORG ØST, DENMARK 8 [3] D. Lay, "Linear Algebra and Its Applications", 2nd ed., Addison-Wesley Publishing Company, April [4] K. Sam Shanmugan, A. M. Breipohl, "Random Signals - Detection, Estimation and Data Analysis", John Wiley & Sons, 988. [5] J. Herault, C. Jutten and B Ans., "D etection de grandeurs primitives dans un message composite par une architecture de calcul neuromim etique en apprentissage non supervis e.", Proc. Gretsi pp , 985. Jakob Ashtar was born in 972. He is a student at the 7th semester Signal Processing at Aalborg University. Kristoffer Møller Jørgensen was born in 979. He is a student at the 7th semester Signal Processing at Aalborg University. Jacob Lindvig was born in 973. He is a student at the 7th semester Signal Processing at Aalborg University. Lars Juul Mikkelsen was born in 98. He is a student at the 7th semester Signal Processing at Aalborg University. Jakob Birkedal Nielsen was born in 979. He is a student at the 7th semester Signal Processing at Aalborg University. Lars Sommer Søndergaard was born in 980. He is a student at the 7th semester Signal Processing at Aalborg University. Dario Farina Supervisor) was born in 973. He Graduated from Politecnico di Torino in 998. In 200 he obtained his PhD in Electronics Engineering and Communications at Politecnico di Torino and at the Ecole Centrale de Nantes, Nantes France).
where A 2 IR m n is the mixing matrix, s(t) is the n-dimensional source vector (n» m), and v(t) is additive white noise that is statistically independ
BLIND SEPARATION OF NONSTATIONARY AND TEMPORALLY CORRELATED SOURCES FROM NOISY MIXTURES Seungjin CHOI x and Andrzej CICHOCKI y x Department of Electrical Engineering Chungbuk National University, KOREA
More informationA UNIFIED PRESENTATION OF BLIND SEPARATION METHODS FOR CONVOLUTIVE MIXTURES USING BLOCK-DIAGONALIZATION
A UNIFIED PRESENTATION OF BLIND SEPARATION METHODS FOR CONVOLUTIVE MIXTURES USING BLOCK-DIAGONALIZATION Cédric Févotte and Christian Doncarli Institut de Recherche en Communications et Cybernétique de
More information1 Introduction Blind source separation (BSS) is a fundamental problem which is encountered in a variety of signal processing problems where multiple s
Blind Separation of Nonstationary Sources in Noisy Mixtures Seungjin CHOI x1 and Andrzej CICHOCKI y x Department of Electrical Engineering Chungbuk National University 48 Kaeshin-dong, Cheongju Chungbuk
More informationAnalytical solution of the blind source separation problem using derivatives
Analytical solution of the blind source separation problem using derivatives Sebastien Lagrange 1,2, Luc Jaulin 2, Vincent Vigneron 1, and Christian Jutten 1 1 Laboratoire Images et Signaux, Institut National
More informationWavelet de-noising for blind source separation in noisy mixtures.
Wavelet for blind source separation in noisy mixtures. Bertrand Rivet 1, Vincent Vigneron 1, Anisoara Paraschiv-Ionescu 2 and Christian Jutten 1 1 Institut National Polytechnique de Grenoble. Laboratoire
More informationRobust extraction of specific signals with temporal structure
Robust extraction of specific signals with temporal structure Zhi-Lin Zhang, Zhang Yi Computational Intelligence Laboratory, School of Computer Science and Engineering, University of Electronic Science
More informationORIENTED PCA AND BLIND SIGNAL SEPARATION
ORIENTED PCA AND BLIND SIGNAL SEPARATION K. I. Diamantaras Department of Informatics TEI of Thessaloniki Sindos 54101, Greece kdiamant@it.teithe.gr Th. Papadimitriou Department of Int. Economic Relat.
More informationIndependent Component Analysis. Contents
Contents Preface xvii 1 Introduction 1 1.1 Linear representation of multivariate data 1 1.1.1 The general statistical setting 1 1.1.2 Dimension reduction methods 2 1.1.3 Independence as a guiding principle
More informationMULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES
MULTICHANNEL SIGNAL PROCESSING USING SPATIAL RANK COVARIANCE MATRICES S. Visuri 1 H. Oja V. Koivunen 1 1 Signal Processing Lab. Dept. of Statistics Tampere Univ. of Technology University of Jyväskylä P.O.
More informationIndependent Component Analysis and Its Application on Accelerator Physics
Independent Component Analysis and Its Application on Accelerator Physics Xiaoying Pang LA-UR-12-20069 ICA and PCA Similarities: Blind source separation method (BSS) no model Observed signals are linear
More informationNOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION. M. Schwab, P. Noll, and T. Sikora. Technical University Berlin, Germany Communication System Group
NOISE ROBUST RELATIVE TRANSFER FUNCTION ESTIMATION M. Schwab, P. Noll, and T. Sikora Technical University Berlin, Germany Communication System Group Einsteinufer 17, 1557 Berlin (Germany) {schwab noll
More informationPCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani
PCA & ICA CE-717: Machine Learning Sharif University of Technology Spring 2015 Soleymani Dimensionality Reduction: Feature Selection vs. Feature Extraction Feature selection Select a subset of a given
More informationMassoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39
Blind Source Separation (BSS) and Independent Componen Analysis (ICA) Massoud BABAIE-ZADEH Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39 Outline Part I Part II Introduction
More informationStatistical signal processing
Statistical signal processing Short overview of the fundamentals Outline Random variables Random processes Stationarity Ergodicity Spectral analysis Random variable and processes Intuition: A random variable
More informationThe Singular Value Decomposition (SVD) and Principal Component Analysis (PCA)
Chapter 5 The Singular Value Decomposition (SVD) and Principal Component Analysis (PCA) 5.1 Basics of SVD 5.1.1 Review of Key Concepts We review some key definitions and results about matrices that will
More informationSemi-Blind approaches to source separation: introduction to the special session
Semi-Blind approaches to source separation: introduction to the special session Massoud BABAIE-ZADEH 1 Christian JUTTEN 2 1- Sharif University of Technology, Tehran, IRAN 2- Laboratory of Images and Signals
More informationTHEORETICAL CONCEPTS & APPLICATIONS OF INDEPENDENT COMPONENT ANALYSIS
THEORETICAL CONCEPTS & APPLICATIONS OF INDEPENDENT COMPONENT ANALYSIS SONALI MISHRA 1, NITISH BHARDWAJ 2, DR. RITA JAIN 3 1,2 Student (B.E.- EC), LNCT, Bhopal, M.P. India. 3 HOD (EC) LNCT, Bhopal, M.P.
More informationA Tutorial on Data Reduction. Principal Component Analysis Theoretical Discussion. By Shireen Elhabian and Aly Farag
A Tutorial on Data Reduction Principal Component Analysis Theoretical Discussion By Shireen Elhabian and Aly Farag University of Louisville, CVIP Lab November 2008 PCA PCA is A backbone of modern data
More informationLeast square joint diagonalization of matrices under an intrinsic scale constraint
Least square joint diagonalization of matrices under an intrinsic scale constraint Dinh-Tuan Pham 1 and Marco Congedo 2 1 Laboratory Jean Kuntzmann, cnrs - Grenoble INP - UJF (Grenoble, France) Dinh-Tuan.Pham@imag.fr
More information7. Symmetric Matrices and Quadratic Forms
Linear Algebra 7. Symmetric Matrices and Quadratic Forms CSIE NCU 1 7. Symmetric Matrices and Quadratic Forms 7.1 Diagonalization of symmetric matrices 2 7.2 Quadratic forms.. 9 7.4 The singular value
More informationBlind Source Separation with a Time-Varying Mixing Matrix
Blind Source Separation with a Time-Varying Mixing Matrix Marcus R DeYoung and Brian L Evans Department of Electrical and Computer Engineering The University of Texas at Austin 1 University Station, Austin,
More informationBLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES
BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES Dinh-Tuan Pham Laboratoire de Modélisation et Calcul URA 397, CNRS/UJF/INPG BP 53X, 38041 Grenoble cédex, France Dinh-Tuan.Pham@imag.fr
More information3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE
3. ESTIMATION OF SIGNALS USING A LEAST SQUARES TECHNIQUE 3.0 INTRODUCTION The purpose of this chapter is to introduce estimators shortly. More elaborated courses on System Identification, which are given
More informationConcentration Ellipsoids
Concentration Ellipsoids ECE275A Lecture Supplement Fall 2008 Kenneth Kreutz Delgado Electrical and Computer Engineering Jacobs School of Engineering University of California, San Diego VERSION LSECE275CE
More informationBLIND SOURCE SEPARATION TECHNIQUES ANOTHER WAY OF DOING OPERATIONAL MODAL ANALYSIS
BLIND SOURCE SEPARATION TECHNIQUES ANOTHER WAY OF DOING OPERATIONAL MODAL ANALYSIS F. Poncelet, Aerospace and Mech. Eng. Dept., University of Liege, Belgium G. Kerschen, Aerospace and Mech. Eng. Dept.,
More informationNatural Gradient Learning for Over- and Under-Complete Bases in ICA
NOTE Communicated by Jean-François Cardoso Natural Gradient Learning for Over- and Under-Complete Bases in ICA Shun-ichi Amari RIKEN Brain Science Institute, Wako-shi, Hirosawa, Saitama 351-01, Japan Independent
More informationMassachusetts Institute of Technology
Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.011: Introduction to Communication, Control and Signal Processing QUIZ, April 1, 010 QUESTION BOOKLET Your
More informationON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS
Yugoslav Journal of Operations Research 5 (25), Number, 79-95 ON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS Slavica TODOROVIĆ-ZARKULA EI Professional Electronics, Niš, bssmtod@eunet.yu Branimir TODOROVIĆ,
More informationTutorial on Blind Source Separation and Independent Component Analysis
Tutorial on Blind Source Separation and Independent Component Analysis Lucas Parra Adaptive Image & Signal Processing Group Sarnoff Corporation February 09, 2002 Linear Mixtures... problem statement...
More informationBlind channel deconvolution of real world signals using source separation techniques
Blind channel deconvolution of real world signals using source separation techniques Jordi Solé-Casals 1, Enric Monte-Moreno 2 1 Signal Processing Group, University of Vic, Sagrada Família 7, 08500, Vic
More informationProf. Dr.-Ing. Armin Dekorsy Department of Communications Engineering. Stochastic Processes and Linear Algebra Recap Slides
Prof. Dr.-Ing. Armin Dekorsy Department of Communications Engineering Stochastic Processes and Linear Algebra Recap Slides Stochastic processes and variables XX tt 0 = XX xx nn (tt) xx 2 (tt) XX tt XX
More informationBlind Source Separation Using Second-Order Cyclostationary Statistics
694 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 49, NO. 4, APRIL 2001 Blind Source Separation Using Second-Order Cyclostationary Statistics Karim Abed-Meraim, Yong Xiang, Jonathan H. Manton, Yingbo Hua,
More informationMachine Learning. Principal Components Analysis. Le Song. CSE6740/CS7641/ISYE6740, Fall 2012
Machine Learning CSE6740/CS7641/ISYE6740, Fall 2012 Principal Components Analysis Le Song Lecture 22, Nov 13, 2012 Based on slides from Eric Xing, CMU Reading: Chap 12.1, CB book 1 2 Factor or Component
More informationSingle-User MIMO systems: Introduction, capacity results, and MIMO beamforming
Single-User MIMO systems: Introduction, capacity results, and MIMO beamforming Master Universitario en Ingeniería de Telecomunicación I. Santamaría Universidad de Cantabria Contents Introduction Multiplexing,
More informationChapter 7: Symmetric Matrices and Quadratic Forms
Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved
More informationSingular Value Decomposition
Singular Value Decomposition Motivatation The diagonalization theorem play a part in many interesting applications. Unfortunately not all matrices can be factored as A = PDP However a factorization A =
More informationIndependent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego
Independent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego Email: brao@ucsdedu References 1 Hyvarinen, A, Karhunen, J, & Oja, E (2004) Independent component analysis (Vol 46)
More informationADAPTIVE ANTENNAS. SPATIAL BF
ADAPTIVE ANTENNAS SPATIAL BF 1 1-Spatial reference BF -Spatial reference beamforming may not use of embedded training sequences. Instead, the directions of arrival (DoA) of the impinging waves are used
More informationAcoustic Source Separation with Microphone Arrays CCNY
Acoustic Source Separation with Microphone Arrays Lucas C. Parra Biomedical Engineering Department City College of New York CCNY Craig Fancourt Clay Spence Chris Alvino Montreal Workshop, Nov 6, 2004 Blind
More informationBlind Identification of FIR Systems and Deconvolution of White Input Sequences
Blind Identification of FIR Systems and Deconvolution of White Input Sequences U. SOVERINI, P. CASTALDI, R. DIVERSI and R. GUIDORZI Dipartimento di Elettronica, Informatica e Sistemistica Università di
More informationHST.582J/6.555J/16.456J
Blind Source Separation: PCA & ICA HST.582J/6.555J/16.456J Gari D. Clifford gari [at] mit. edu http://www.mit.edu/~gari G. D. Clifford 2005-2009 What is BSS? Assume an observation (signal) is a linear
More informationA Canonical Genetic Algorithm for Blind Inversion of Linear Channels
A Canonical Genetic Algorithm for Blind Inversion of Linear Channels Fernando Rojas, Jordi Solé-Casals, Enric Monte-Moreno 3, Carlos G. Puntonet and Alberto Prieto Computer Architecture and Technology
More informationA more efficient second order blind identification method for separation of uncorrelated stationary time series
A more efficient second order blind identification method for separation of uncorrelated stationary time series Sara Taskinen 1a, Jari Miettinen a, Klaus Nordhausen b,c a Department of Mathematics and
More informationReal-time Blind Source Separation
Real-time Blind Source Separation A Feasibility Study using CUDA th semester project, AAU, Applied Signal Processing and Implementation Spring 9 Group 4 Søren Reinholt Søndergaard Martin Brinch Sørensen
More informationIntroduction to Machine Learning
10-701 Introduction to Machine Learning PCA Slides based on 18-661 Fall 2018 PCA Raw data can be Complex, High-dimensional To understand a phenomenon we measure various related quantities If we knew what
More informationEcon Slides from Lecture 7
Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for
More informationDOA Estimation using MUSIC and Root MUSIC Methods
DOA Estimation using MUSIC and Root MUSIC Methods EE602 Statistical signal Processing 4/13/2009 Presented By: Chhavipreet Singh(Y515) Siddharth Sahoo(Y5827447) 2 Table of Contents 1 Introduction... 3 2
More informationComputational Methods. Eigenvalues and Singular Values
Computational Methods Eigenvalues and Singular Values Manfred Huber 2010 1 Eigenvalues and Singular Values Eigenvalues and singular values describe important aspects of transformations and of data relations
More informationELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random
More informationHST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007
MIT OpenCourseWare http://ocw.mit.edu HST.582J / 6.555J / 16.456J Biomedical Signal and Image Processing Spring 2007 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationYORK UNIVERSITY. Faculty of Science Department of Mathematics and Statistics MATH M Test #2 Solutions
YORK UNIVERSITY Faculty of Science Department of Mathematics and Statistics MATH 3. M Test # Solutions. (8 pts) For each statement indicate whether it is always TRUE or sometimes FALSE. Note: For this
More informationPerformance Analysis for Strong Interference Remove of Fast Moving Target in Linear Array Antenna
Performance Analysis for Strong Interference Remove of Fast Moving Target in Linear Array Antenna Kwan Hyeong Lee Dept. Electriacal Electronic & Communicaton, Daejin University, 1007 Ho Guk ro, Pochen,Gyeonggi,
More informationIndependent Component Analysis and Its Applications. By Qing Xue, 10/15/2004
Independent Component Analysis and Its Applications By Qing Xue, 10/15/2004 Outline Motivation of ICA Applications of ICA Principles of ICA estimation Algorithms for ICA Extensions of basic ICA framework
More informationA METHOD OF ICA IN TIME-FREQUENCY DOMAIN
A METHOD OF ICA IN TIME-FREQUENCY DOMAIN Shiro Ikeda PRESTO, JST Hirosawa 2-, Wako, 35-98, Japan Shiro.Ikeda@brain.riken.go.jp Noboru Murata RIKEN BSI Hirosawa 2-, Wako, 35-98, Japan Noboru.Murata@brain.riken.go.jp
More informationSingular Value Decomposition and Principal Component Analysis (PCA) I
Singular Value Decomposition and Principal Component Analysis (PCA) I Prof Ned Wingreen MOL 40/50 Microarray review Data per array: 0000 genes, I (green) i,i (red) i 000 000+ data points! The expression
More informationBLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS
BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS M. Kawamoto 1,2, Y. Inouye 1, A. Mansour 2, and R.-W. Liu 3 1. Department of Electronic and Control Systems Engineering,
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 05 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA6 MATERIAL NAME : University Questions MATERIAL CODE : JM08AM004 REGULATION : R008 UPDATED ON : Nov-Dec 04 (Scan
More information926 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 3, MARCH Monica Nicoli, Member, IEEE, and Umberto Spagnolini, Senior Member, IEEE (1)
926 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 3, MARCH 2005 Reduced-Rank Channel Estimation for Time-Slotted Mobile Communication Systems Monica Nicoli, Member, IEEE, and Umberto Spagnolini,
More informationBlind Source Separation via Generalized Eigenvalue Decomposition
Journal of Machine Learning Research 4 (2003) 1261-1269 Submitted 10/02; Published 12/03 Blind Source Separation via Generalized Eigenvalue Decomposition Lucas Parra Department of Biomedical Engineering
More informationNumerical Methods I Singular Value Decomposition
Numerical Methods I Singular Value Decomposition Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 9th, 2014 A. Donev (Courant Institute)
More informationPrincipal Component Analysis
Principal Component Analysis CS5240 Theoretical Foundations in Multimedia Leow Wee Kheng Department of Computer Science School of Computing National University of Singapore Leow Wee Kheng (NUS) Principal
More informationLecture 19 Observability and state estimation
EE263 Autumn 2007-08 Stephen Boyd Lecture 19 Observability and state estimation state estimation discrete-time observability observability controllability duality observers for noiseless case continuous-time
More informationLecture 3: Review of Linear Algebra
ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters, transforms,
More informationVAR Model. (k-variate) VAR(p) model (in the Reduced Form): Y t-2. Y t-1 = A + B 1. Y t + B 2. Y t-p. + ε t. + + B p. where:
VAR Model (k-variate VAR(p model (in the Reduced Form: where: Y t = A + B 1 Y t-1 + B 2 Y t-2 + + B p Y t-p + ε t Y t = (y 1t, y 2t,, y kt : a (k x 1 vector of time series variables A: a (k x 1 vector
More informationLecture 3: Review of Linear Algebra
ECE 83 Fall 2 Statistical Signal Processing instructor: R Nowak, scribe: R Nowak Lecture 3: Review of Linear Algebra Very often in this course we will represent signals as vectors and operators (eg, filters,
More informationAdvanced Introduction to Machine Learning CMU-10715
Advanced Introduction to Machine Learning CMU-10715 Independent Component Analysis Barnabás Póczos Independent Component Analysis 2 Independent Component Analysis Model original signals Observations (Mixtures)
More informationEIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4]
EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: Au = λu λ C : eigenvalue u C n : eigenvector Example:
More informationLarge Scale Data Analysis Using Deep Learning
Large Scale Data Analysis Using Deep Learning Linear Algebra U Kang Seoul National University U Kang 1 In This Lecture Overview of linear algebra (but, not a comprehensive survey) Focused on the subset
More informationSecond-order statistics based blind source separation using a bank of subband filters
Digital Signal Processing 13 (2003) 252 274 www.elsevier.com/locate/dsp Second-order statistics based blind source separation using a bank of subband filters R.R. Gharieb a,b, and A. Cichocki a,c a Laboratory
More informationEstimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition
Estimation of the Optimum Rotational Parameter for the Fractional Fourier Transform Using Domain Decomposition Seema Sud 1 1 The Aerospace Corporation, 4851 Stonecroft Blvd. Chantilly, VA 20151 Abstract
More informationVector and Matrix Norms. Vector and Matrix Norms
Vector and Matrix Norms Vector Space Algebra Matrix Algebra: We let x x and A A, where, if x is an element of an abstract vector space n, and A = A: n m, then x is a complex column vector of length n whose
More informationA New High-Resolution and Stable MV-SVD Algorithm for Coherent Signals Detection
Progress In Electromagnetics Research M, Vol. 35, 163 171, 2014 A New High-Resolution and Stable MV-SVD Algorithm for Coherent Signals Detection Basma Eldosouky, Amr H. Hussein *, and Salah Khamis Abstract
More information1254 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 4, APRIL On the Virtual Array Concept for Higher Order Array Processing
1254 IEEE TRANSACTIONS ON SIGNAL PROCESSING, VOL. 53, NO. 4, APRIL 2005 On the Virtual Array Concept for Higher Order Array Processing Pascal Chevalier, Laurent Albera, Anne Ferréol, and Pierre Comon,
More informationRecipes for the Linear Analysis of EEG and applications
Recipes for the Linear Analysis of EEG and applications Paul Sajda Department of Biomedical Engineering Columbia University Can we read the brain non-invasively and in real-time? decoder 1001110 if YES
More informationMaster Thesis Separation of Non-Stationary Sources
Master Thesis Separation of Non-Stationary Sources Lisbeth Andersson Perth, March 21 Australian Telecommunications Research Institute Curtin University of Technology Australia i Preface This master thesis
More informationPrincipal Component Analysis
Principal Component Analysis Laurenz Wiskott Institute for Theoretical Biology Humboldt-University Berlin Invalidenstraße 43 D-10115 Berlin, Germany 11 March 2004 1 Intuition Problem Statement Experimental
More informationBLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES
BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES Dinh-Tuan Pham Laboratoire de Modélisation et Calcul URA 397, CNRS/UJF/INPG BP 53X, 38041 Grenoble cédex, France Dinh-Tuan.Pham@imag.fr
More informationGENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS. Mitsuru Kawamoto 1,2 and Yujiro Inouye 1
GENERALIZED DEFLATION ALGORITHMS FOR THE BLIND SOURCE-FACTOR SEPARATION OF MIMO-FIR CHANNELS Mitsuru Kawamoto,2 and Yuiro Inouye. Dept. of Electronic and Control Systems Engineering, Shimane University,
More informationOverview of Beamforming
Overview of Beamforming Arye Nehorai Preston M. Green Department of Electrical and Systems Engineering Washington University in St. Louis March 14, 2012 CSSIP Lab 1 Outline Introduction Spatial and temporal
More informationTensor approach for blind FIR channel identification using 4th-order cumulants
Tensor approach for blind FIR channel identification using 4th-order cumulants Carlos Estêvão R Fernandes Gérard Favier and João Cesar M Mota contact: cfernand@i3s.unice.fr June 8, 2006 Outline 1. HOS
More information/16/$ IEEE 1728
Extension of the Semi-Algebraic Framework for Approximate CP Decompositions via Simultaneous Matrix Diagonalization to the Efficient Calculation of Coupled CP Decompositions Kristina Naskovska and Martin
More informationA Root-MUSIC-Like Direction Finding Method for Cyclostationary Signals
EURASIP Journal on Applied Signal Processing 25:1, 69 73 c 25 Hindawi Publishing Corporation A Root-MUSIC-Like Direction Finding Method for Cyclostationary Signals Pascal Chargé LESIA, DGEI, INSA Toulouse,
More information1 Principal component analysis and dimensional reduction
Linear Algebra Working Group :: Day 3 Note: All vector spaces will be finite-dimensional vector spaces over the field R. 1 Principal component analysis and dimensional reduction Definition 1.1. Given an
More informationPrincipal Component Analysis
Principal Component Analysis Anders Øland David Christiansen 1 Introduction Principal Component Analysis, or PCA, is a commonly used multi-purpose technique in data analysis. It can be used for feature
More informationData Mining Lecture 4: Covariance, EVD, PCA & SVD
Data Mining Lecture 4: Covariance, EVD, PCA & SVD Jo Houghton ECS Southampton February 25, 2019 1 / 28 Variance and Covariance - Expectation A random variable takes on different values due to chance The
More informationLecture: Face Recognition and Feature Reduction
Lecture: Face Recognition and Feature Reduction Juan Carlos Niebles and Ranjay Krishna Stanford Vision and Learning Lab 1 Recap - Curse of dimensionality Assume 5000 points uniformly distributed in the
More informationLinear Algebra, part 3. Going back to least squares. Mathematical Models, Analysis and Simulation = 0. a T 1 e. a T n e. Anna-Karin Tornberg
Linear Algebra, part 3 Anna-Karin Tornberg Mathematical Models, Analysis and Simulation Fall semester, 2010 Going back to least squares (Sections 1.7 and 2.3 from Strang). We know from before: The vector
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More informationDimensionality Reduction: PCA. Nicholas Ruozzi University of Texas at Dallas
Dimensionality Reduction: PCA Nicholas Ruozzi University of Texas at Dallas Eigenvalues λ is an eigenvalue of a matrix A R n n if the linear system Ax = λx has at least one non-zero solution If Ax = λx
More informationLet A an n n real nonsymmetric matrix. The eigenvalue problem: λ 1 = 1 with eigenvector u 1 = ( ) λ 2 = 2 with eigenvector u 2 = ( 1
Eigenvalue Problems. Introduction Let A an n n real nonsymmetric matrix. The eigenvalue problem: EIGENVALE PROBLEMS AND THE SVD. [5.1 TO 5.3 & 7.4] Au = λu Example: ( ) 2 0 A = 2 1 λ 1 = 1 with eigenvector
More informationA Subspace Approach to Estimation of. Measurements 1. Carlos E. Davila. Electrical Engineering Department, Southern Methodist University
EDICS category SP 1 A Subspace Approach to Estimation of Autoregressive Parameters From Noisy Measurements 1 Carlos E Davila Electrical Engineering Department, Southern Methodist University Dallas, Texas
More informationIV. Matrix Approximation using Least-Squares
IV. Matrix Approximation using Least-Squares The SVD and Matrix Approximation We begin with the following fundamental question. Let A be an M N matrix with rank R. What is the closest matrix to A that
More informationMIMO instantaneous blind identification based on second-order temporal structure
MIMO instantaneous blind identification based on second-order temporal structure Citation for published version (APA): Laar, van de, J, Moonen, M, & Sommen, P C W (2008) MIMO instantaneous blind identification
More informationLecture Notes 1: Vector spaces
Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector
More informationEmpirical Macroeconomics
Empirical Macroeconomics Francesco Franco Nova SBE April 5, 2016 Francesco Franco Empirical Macroeconomics 1/39 Growth and Fluctuations Supply and Demand Figure : US dynamics Francesco Franco Empirical
More informationVARIANCE COMPUTATION OF MODAL PARAMETER ES- TIMATES FROM UPC SUBSPACE IDENTIFICATION
VARIANCE COMPUTATION OF MODAL PARAMETER ES- TIMATES FROM UPC SUBSPACE IDENTIFICATION Michael Döhler 1, Palle Andersen 2, Laurent Mevel 1 1 Inria/IFSTTAR, I4S, Rennes, France, {michaeldoehler, laurentmevel}@inriafr
More informationMath 307 Learning Goals. March 23, 2010
Math 307 Learning Goals March 23, 2010 Course Description The course presents core concepts of linear algebra by focusing on applications in Science and Engineering. Examples of applications from recent
More informationGrowing Window Recursive Quadratic Optimization with Variable Regularization
49th IEEE Conference on Decision and Control December 15-17, Hilton Atlanta Hotel, Atlanta, GA, USA Growing Window Recursive Quadratic Optimization with Variable Regularization Asad A. Ali 1, Jesse B.
More informationPOLYNOMIAL SINGULAR VALUES FOR NUMBER OF WIDEBAND SOURCES ESTIMATION AND PRINCIPAL COMPONENT ANALYSIS
POLYNOMIAL SINGULAR VALUES FOR NUMBER OF WIDEBAND SOURCES ESTIMATION AND PRINCIPAL COMPONENT ANALYSIS Russell H. Lambert RF and Advanced Mixed Signal Unit Broadcom Pasadena, CA USA russ@broadcom.com Marcel
More informationBlind Machine Separation Te-Won Lee
Blind Machine Separation Te-Won Lee University of California, San Diego Institute for Neural Computation Blind Machine Separation Problem we want to solve: Single microphone blind source separation & deconvolution
More information