y 2 a 12 a 1n a 11 s 2 w 11 f n a 21 s n f n y 1 a n1 x n x 2 x 1
|
|
- Harvey Boone
- 6 years ago
- Views:
Transcription
1 Maximum Likelihood Blind Source Separation: A Context-Sensitive Generalization of ICA Barak A. Pearlmutter Computer Science Dept, FEC 33 University of Ne Mexico Albuquerque, NM 873 bap@cs.unm.edu Lucas C. Parra Siemens Corporate Research 755 College Road East Princeton, NJ lucas@scr.siemens.com Abstract In the square linear blind source separation problem, one must nd a linear unmixing operator hich can detangle the result x i (t) of mixing n unknon independent sources s i (t) through an unknon n n mixing matrix A(t) of causal linear lters: x i = P a i s. We cast the problem as one of maximum likelihood density estimation, and in that frameork introduce an algorithm that searches for independent components using both temporal and spatial cues. We call the resulting algorithm \Contextual ICA," after the (Bell and Senoski 995) Infomax algorithm, hich e sho to be a special case of cica. Because cica can make use of the temporal structure of its input, it is able separate in a number of situations here standard methods cannot, including sources ith lo kurtosis, colored Gaussian sources, and sources hich have Gaussian histograms. The Blind Source Separation Problem Consider a set of n indepent sources s(t) ::: s n (t). We P are given n linearly distorted sensor reading hich combine these sources, x i = a is, here a i is a lter beteen source and sensor i, asshon in gure a. This can be expressed as X X X x i (t) = a i ()s (t ; ) = = a i s
2 f( s(t) s ( t ),... ) f 2 s s 2 a a 2 a 2 a n f( y (t) y ( t ),...; () ) f 2 y y n f n s n f n y n a n n x x 2 x n x x 2 x n Figure : The left diagram shos a generative model of data production for blind source separation problem. The cica algorithm ts the reparametrized generative model on the right to the data. Since (unless the mixing process is singular) both diagrams give linear maps beteen the sources and the sensors, they are mathematically equivalent. Hoever, (a) makes the transformation from s to x explicit, hile (b) makes the transformation from x to y, the estimated sources, explicit. P or, in matrix notation, x(t) = = A()s(t ; ) =A s: The square linear blind source separation problem is to recover s from x. There is an inherent ambiguity in this, for if e dene a ne set of sources s by s i = b i s i here b i () is some invertable lter, then the various s i are independent, P and constitute ust as good a solution to the problem as the true s i,sincex i = (a i b ; ) s : Similarly the sources could be arbitrarily permuted. Surprisingly, up to permutation of the sources and linear ltering of the individual sources, the problem is ell posed assuming that the sources s are not Gaussian. The reason for this is that only ith a correct separation are the recovered sources truly statistically independent, and this fact serves as a sucient constraint. Under the assumptions e have made, and further assuming that P the linear transformation A is invertible, e ill speak of recovering y i (t) = i x here these y i are a ltered and permuted version of the original unknon s i. For clarity of exposition, ill often refer to \the" solution and refer to the y i as \the" recovered sources, rather than refering to an point inthe manifold of solutions and a set of consistent recovered sources. 2 Maximum likelihood density estimation Folloing Pham, Garrat, and Jutten (992) and Belouchrani and Cardoso (995), e cast the BSS problem as one of maximum likelihood density estimation. In the MLE frameork, one begins ith a probabilistic model of the data production process. This probabilistic model is parametrized by a vector of modiable parameters, and it therefore assigns a -dependent probability density p(x x ::: ) toa each possible dataset x x :::. The task is then to nd a hich maximizes this probability. There are a number of approaches to performing this maximization. Here e apply Without these assumptions, for instance in the presence of noise, even a linear mixing process leads to an optimal unmixing process that is highly nonlinear.
3 the stochastic gradient method, in hich a single stochastic sample x is chosen from the dataset and ;dlogp(x P )=d is used as a stochastic estimate of the gradient of the negative likelihood t ;dlogp(x(t) )=d. 2. The likelihood of the data The model of data production e consider is shon in gure a. In that model, the sensor readings x are an explicit linear function of the underlying sources s. In this model of the data production, there are to stages. In the rst stage, the sources independently produce signals. These signals are time-dependent, and the probability density of source i producing value s (t) at time t is f (s (t)s (t ; ) s (t ; 2) :::). Although this source model could be of almost any dierentiable form, e used a generalized autoregressive model described in appendix A. For expository purposes, e can consider using a simple AR model, so e models (t) = b ()s (t ; ) + b (2)s (t ; 2) + + b (T )s (t ; T )+r, here r is an iid random variable, perhaps ith a complicated density. It is important to distinguish to dierent, although related, linear lters. When the source models are simple AR models, there are to types of linear convolutions being performed. The rst is in the ay each source produces its signal: as a linear function of its recent history plus a hite driving term, hich could be expressed as a moving average model, a convolution ith a hite driving term, s = b r. The second is in the ay the sources are mixed: linear functions of the output of each source are added, x i = P a i s = P (a i b ) r. Thus, ith AR sources, the source convolution could be folded into the convolutions of the linear mixing process. If e ere to estimate values for the free parameters of this model, i.e. to estimate the lters, then the task of recovering the estimated sources from the sensor output ould require inverting the linear A =(a i ), as ell as some technique to guarantee its non-singularity. Such a model is shon in gure a. Instead, e parameterize the model by W = A ;, an estimated unmixing matrix, as shon in gure b. In this indirect representation, s is an explicit linear function of x, and therefore x is only an implicit linear function of s. This parameterization of the model is equally convenient for assigning probabilities to samples x, and is therefore suitable for MLE. Its advantage is that because the transformation from sensors to sources is estimated explicitly, the sources can be recovered directly from the data and the estimated model, ithout invertion. Note that in this inverse parameterization, the estimated mixture process is stored in inverse form. The source-specic models f i are kept in forard form. Each source-specic model i has a vector of parameters, hich e denote (i). We are no in a position to calculate the likelihood of the data. For simplicityeuse a matrix W of real numbers rather than FIR lters. Generalizing this derivation to a matrix of lters is straightforard, folloing the same techniques used bylambert (996), Torkkola (996), A. Bell (997), but space precludes a derivation here. The individual generative source models give p(y(t)y(t ; ) y(t ; 2) :::)= Y i f i (y i (t)y i (t ; ) y i (t ; 2) :::) ()
4 here the probability densities f i are each parameterized by vectors (i). Using these equations, e ould like to express the likelihood of x(t) in closed form, given the history x(t ; ) x(t ; 2) :::. Since the history is knon, e therefore also kno the history of the recovered sources, y(t ; ) y(t ; 2) :::. This means that e can calculate the density p(y(t)y(t ; ) :::). Using this, e can express the density ofx(t) and expand G b =log^p(x ) =logw + P log f (y (t)y (t ; ) y (t ; 2) ::: () ) There are to sorts of parameters hich emust take the derivative ith respect to: the matrix W and the source parameters (). The source parameters do not inuence our recovered sources, and therefore have a simple form d G b = ; df (y )=d d f (y ) Hoever, a change to the matrix W changes y, hichintroduces a fe extra terms. Note that d log W =dw = W ;T, the transpose inverse. Next, since y = Wx, e see that dy =dw = (x) T, a matrix of zeros except for the vector x in ro. No e note that df ()=dw term has to logical components: the rst from the eectofchanging W upon y (t), and the second from the eect of changing W upon y (t ; ) y (t ; 2) :::. (This second is called the \recurrent term", and such terms are frequently dropped for convenience. As shon in gure 3, dropping this term here is not a reasonable approximation.) df (y (t)y (t ; ) ::: ) dw dy (t) (t) dw dy (t ; (t ; ) dw Note that the expression dy(t;) is the only matrix, and it is zero except for the dw th ro, hich isx(t ; ). The =@y (t) e shall denote f (), and (t ; ) e shall denote f () (). We thenhave d b G dw = ;W;T ; f () f () X x(t) T ; = f () () f ()! x(t ; ) T (2) here (expr()) denotes the column vector hose elements are expr() ::: expr(n). 2.2 The natural gradient Folloing Amari, Cichocki, and Yang (996), e follo a pseudogradient. Instead of using equation 2, e post-multiply this quantity by W T W. Since this is a positivedenite matrix, it does not aect the stochastic gradient convergence criteria, and the resulting quantity simplies in a fashion that neatly eliminates the costly matrix inversion otherise required. Convergence is also accelerated. 3 Experiments We conducted a number of experiments to test the ecacy of the cica algorithm. The rst, shon in gure 2, as a toy problem involving a set of processed deliberately constructed to be dicult for conventional source separation algorithms. In the second experiment, shon in gure 3, ten real sources ere digitally mixed ith an instantaneous matrix and separation performance as measured as a funciton of varying model complexity parameters. These sources have are available for benchmarking purposes in
5 5 Filtered and Mixed Uniform Distributions Recovered Source Values Estimated Density in Unximed Source Space estimated density source source Figure 2: cica using a history of one time step and a mixture of ve logistic densities for each source as applied to 5, samples of a mixture of to one-dimensional uniform distributions each ltered by convolution ith a decaying exponential of time constant of99.5. Shon is a scatterplot of the data input to the algorithm, along ith the true source axes (left), the estimated residual probability density (center), and a scatterplot of the residuals of the data transformed into the estimated source space coordinates (right). The product of the true mixing matrix and the estimated unmixing matrix deviates from a scaling and permutation matrix by about 3%. Truncated Gradient Full Gradient Noise Model S/N Ratio S/N Ratio S/N Ratio number of AR filter taps number of AR filter taps 2 number of logistics Figure 3: The performance of cica as a function of model complexity and gradient accuracy. In all simulations, ten ve-second clips taken digitally from ten audio CD ere digitally mixed through a random ten-by-ten instantanious mixing matrix. The signal to noise ratio of each original source as expressed in the recovered sources is plotted. In (a) and (b), AR source models ith a logistic noise term ere used, and the number of taps of the AR model as varied. (This reduces to Bell-Senoski infomax hen the number of taps is zero.) Is (a), the recurrent term of the gradient as left out, hile in (b) the recurrent term as included. Clearly the recurrent term is important. In (c), a degenerate AR model ith zero taps as used, but the noise term as a mixture of logistics, and the number of logistics as varied. 4 Discussion The Infomax algorithm (Baram and Roth 994) used for source separation (Bell and Senoski 995) is a special case of the above algorithm in hich (a) the mixing is not convolutional, so W () = W (2) = ::: =, and (b) the sources are assumed to be iid, and therefore the distributions f i (y(t)) are not history sensitive. Further, the form of the f i is restricted to avery special distribution: the logistic density,
6 the derivative of the sigmoidal function =( + exp ;). Although ICA has enoyed avariety of applications (Makeig et al. 996 Bell and Senoski 996b Baram and Roth 995 Bell and Senoski 996a), there are a number of sources hich it cannot separate. These include all sources ith Gaussian histograms (e.g. colored gaussian sources, or even speech to run through the right sort of slight nonlinearity), and sources ith lo kurtosis. As shon in the experiments above, these are of more than theoretical interest. If e simplify our model to use ordinary AR models for the sources, ith gaussian noise terms of xed variance, it is possible to derive a closed-form expression for W (Hagai Attias, personal communication). It may be that for many sources of practical interest, trading aay this model accuracy for speed ill be fruitful. 4. Weakened assumptions It seems clear that, in general, separating hen there are feer microphones than sources requires a strong bayesian prior, and even given perfect knoledge of the mixture process and perfect source models, inverting the mixing process ill be computationally burdensome. Hoever, hen there are more microphones than sources, there is an opportunity to improve the performance of the system in the presence of noise. This seems straightforard to integrate into our frameork. Similarly, fast-timescale microphone nonlinearities are easily incorporated into this maximum likelihood approach. The structure of this problem ould seem to lend itself to EM. Certainly the individual source models can be easily optimized using EM, assuming that they themselves are of suitable form. References A. Bell, T.-W. L. (997). Blind separation of delayed and convolved sources. In Advances in Neural Information Processing Systems 9. MIT Press. In this volume. Amari, S., Cichocki, A., and Yang, H. H. (996). A ne learning algorithm for blind signal separation. In Advances in Neural Information Processing Systems 8. MIT Press. Baram, Y. and Roth, Z. (994). Density Shaping by Neural Netorks ith Application to Classication, Estimation and Forecasting. Tech. rep. CIS-94-2, Center for Intelligent Systems, Technion, Israel Institute for Technology, Haifa. Baram, Y. and Roth, Z. (995). Forecasting by Density Shaping Using Neural Netorks. In Computational Intelligence for Financial Engineering Ne York City. IEEE Press. Bell, A. J. and Senoski, T. J. (995). An Information-Maximization Approach to Blind Separation and Blind Deconvolution. Neural Computation, 7 (6), 29{59. Bell, A. J. and Senoski, T. J. (996a). The Independent Components of Natural Scenes. Vision Research. Submitted.
7 Bell, A. J. and Senoski, T. J. (996b). Learning the higher-order structure of a natural sound. Netork: Computation in Neural Systems. In press. Belouchrani, A. and Cardoso, J.-F. (995). Maximum likelihood source separation by the expectation-maximization technique: Deterministic and stochastic implementation. In Proceedings of 995 International Symposium on Non-Linear Theory and Applications, pp. 49{53 Las Vegas, NV. In press. Lambert, R. H. (996). Multichannel Blind Deconvolution: FIR Matrix Algebra and Separation of Multipath Mixtures. Ph.D. thesis, USC. Makeig, S., Anllo-Vento, L., Jung, T.-P., Bell, A. J., Senoski, T. J., and Hillyard, S. A. (996). Independent component analysis of event-related potentials during selective attention. Society for Neuroscience Abstracts, 22. Pearlmutter, B. A. and Parra, L. C. (996). A Context-Sensitive Generalization of ICA. In International Conference on Neural Information Processing Hong Kong. Springer-Verlag. Url ftp://ftp.cnl.salk.edu/pub/bap/iconip-96- cica.ps.gz. Pham, D., Garrat, P., and Jutten, C. (992). Separation of a mixture of independent sources through a maximum likelihood approach. In European Signal Processing Conference, pp. 77{774. Torkkola, K. (996). Blind separation of convolved sources based on information maximization. In Neural Netorks for Signal Processing VI Kyoto, Japan. IEEE Press. In press. A Fixed mixture AR models The f (u )e used ere a mixture AR processes driven by logistic noise terms, as in Pearlmutter and Parra (996). Each source model as f (u (t)u (t ; ) u (t ; 2) ::: )= X k m k h((u (t) ; u k )= k )= k (3) here k is a scale parameter for logistic density k of source andisanelement Pof, and the mixing coecients m k are elements of and are constrained by m k k =. The component means u k are taken to be linear functions of the recent values of that source, u k = X = a k () u (t ; )+b k (4) here the linear prediction coecients a k () and bias b k are elements of. The derivatives of these are straightforard see Pearlmutter and Parra (996) for details. One complication is to note that, after P each eight update, the mixing coecients must be normalized, m k m k = m k k :
Maximum Likelihood Blind Source Separation: A Context-Sensitive Generalization of ICA
Maximum Likelihood Blind Source Separation: A Context-Sensitive Generalization of ICA Barak A. Pearlmutter Computer Science Dept, FEC 313 University of New Mexico Albuquerque, NM 87131 bap@cs.unm.edu Lucas
More informationBlind Machine Separation Te-Won Lee
Blind Machine Separation Te-Won Lee University of California, San Diego Institute for Neural Computation Blind Machine Separation Problem we want to solve: Single microphone blind source separation & deconvolution
More informationLecture 3a: The Origin of Variational Bayes
CSC535: 013 Advanced Machine Learning Lecture 3a: The Origin of Variational Bayes Geoffrey Hinton The origin of variational Bayes In variational Bayes, e approximate the true posterior across parameters
More informationOn Information Maximization and Blind Signal Deconvolution
On Information Maximization and Blind Signal Deconvolution A Röbel Technical University of Berlin, Institute of Communication Sciences email: roebel@kgwtu-berlinde Abstract: In the following paper we investigate
More informationy(x) = x w + ε(x), (1)
Linear regression We are ready to consider our first machine-learning problem: linear regression. Suppose that e are interested in the values of a function y(x): R d R, here x is a d-dimensional vector-valued
More informationIndependent Component Analysis
Independent Component Analysis James V. Stone November 4, 24 Sheffield University, Sheffield, UK Keywords: independent component analysis, independence, blind source separation, projection pursuit, complexity
More information1 Introduction Independent component analysis (ICA) [10] is a statistical technique whose main applications are blind source separation, blind deconvo
The Fixed-Point Algorithm and Maximum Likelihood Estimation for Independent Component Analysis Aapo Hyvarinen Helsinki University of Technology Laboratory of Computer and Information Science P.O.Box 5400,
More informationA Constrained EM Algorithm for Independent Component Analysis
LETTER Communicated by Hagai Attias A Constrained EM Algorithm for Independent Component Analysis Max Welling Markus Weber California Institute of Technology, Pasadena, CA 91125, U.S.A. We introduce a
More informationNatural Gradient Learning for Over- and Under-Complete Bases in ICA
NOTE Communicated by Jean-François Cardoso Natural Gradient Learning for Over- and Under-Complete Bases in ICA Shun-ichi Amari RIKEN Brain Science Institute, Wako-shi, Hirosawa, Saitama 351-01, Japan Independent
More informationx 1 (t) Spectrogram t s
A METHOD OF ICA IN TIME-FREQUENCY DOMAIN Shiro Ikeda PRESTO, JST Hirosawa 2-, Wako, 35-98, Japan Shiro.Ikeda@brain.riken.go.jp Noboru Murata RIKEN BSI Hirosawa 2-, Wako, 35-98, Japan Noboru.Murata@brain.riken.go.jp
More informationArtificial Neural Networks. Part 2
Artificial Neural Netorks Part Artificial Neuron Model Folloing simplified model of real neurons is also knon as a Threshold Logic Unit x McCullouch-Pitts neuron (943) x x n n Body of neuron f out Biological
More informationMULTICHANNEL BLIND SEPARATION AND. Scott C. Douglas 1, Andrzej Cichocki 2, and Shun-ichi Amari 2
MULTICHANNEL BLIND SEPARATION AND DECONVOLUTION OF SOURCES WITH ARBITRARY DISTRIBUTIONS Scott C. Douglas 1, Andrzej Cichoci, and Shun-ichi Amari 1 Department of Electrical Engineering, University of Utah
More informationConvolutive Blind Source Separation based on Multiple Decorrelation. Lucas Parra, Clay Spence, Bert De Vries Sarno Corporation, CN-5300, Princeton, NJ
Convolutive Blind Source Separation based on Multiple Decorrelation. Lucas Parra, Clay Spence, Bert De Vries Sarno Corporation, CN-5300, Princeton, NJ 08543 lparra j cspence j bdevries @ sarno.com Abstract
More informationLecture VIII Dim. Reduction (I)
Lecture VIII Dim. Reduction (I) Contents: Subset Selection & Shrinkage Ridge regression, Lasso PCA, PCR, PLS Lecture VIII: MLSC - Dr. Sethu Viayakumar Data From Human Movement Measure arm movement and
More informationAn Improved Cumulant Based Method for Independent Component Analysis
An Improved Cumulant Based Method for Independent Component Analysis Tobias Blaschke and Laurenz Wiskott Institute for Theoretical Biology Humboldt University Berlin Invalidenstraße 43 D - 0 5 Berlin Germany
More informationNon-Euclidean Independent Component Analysis and Oja's Learning
Non-Euclidean Independent Component Analysis and Oja's Learning M. Lange 1, M. Biehl 2, and T. Villmann 1 1- University of Appl. Sciences Mittweida - Dept. of Mathematics Mittweida, Saxonia - Germany 2-
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationIndependent Component Analysis
A Short Introduction to Independent Component Analysis Aapo Hyvärinen Helsinki Institute for Information Technology and Depts of Computer Science and Psychology University of Helsinki Problem of blind
More informationMMSE Equalizer Design
MMSE Equalizer Design Phil Schniter March 6, 2008 [k] a[m] P a [k] g[k] m[k] h[k] + ṽ[k] q[k] y [k] P y[m] For a trivial channel (i.e., h[k] = δ[k]), e kno that the use of square-root raisedcosine (SRRC)
More informationAnalytical solution of the blind source separation problem using derivatives
Analytical solution of the blind source separation problem using derivatives Sebastien Lagrange 1,2, Luc Jaulin 2, Vincent Vigneron 1, and Christian Jutten 1 1 Laboratoire Images et Signaux, Institut National
More informationFREQUENCY-DOMAIN IMPLEMENTATION OF BLOCK ADAPTIVE FILTERS FOR ICA-BASED MULTICHANNEL BLIND DECONVOLUTION
FREQUENCY-DOMAIN IMPLEMENTATION OF BLOCK ADAPTIVE FILTERS FOR ICA-BASED MULTICHANNEL BLIND DECONVOLUTION Kyungmin Na, Sang-Chul Kang, Kyung Jin Lee, and Soo-Ik Chae School of Electrical Engineering Seoul
More informationGatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II
Gatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II Gatsby Unit University College London 27 Feb 2017 Outline Part I: Theory of ICA Definition and difference
More informationLinear Regression Linear Regression with Shrinkage
Linear Regression Linear Regression ith Shrinkage Introduction Regression means predicting a continuous (usually scalar) output y from a vector of continuous inputs (features) x. Example: Predicting vehicle
More informationRemaining energy on log scale Number of linear PCA components
NONLINEAR INDEPENDENT COMPONENT ANALYSIS USING ENSEMBLE LEARNING: EXPERIMENTS AND DISCUSSION Harri Lappalainen, Xavier Giannakopoulos, Antti Honkela, and Juha Karhunen Helsinki University of Technology,
More informationIndependent Component Analysis of Incomplete Data
Independent Component Analysis of Incomplete Data Max Welling Markus Weber California Institute of Technology 136-93 Pasadena, CA 91125 fwelling,rmwg@vision.caltech.edu Keywords: EM, Missing Data, ICA
More informationFundamentals of Principal Component Analysis (PCA), Independent Component Analysis (ICA), and Independent Vector Analysis (IVA)
Fundamentals of Principal Component Analysis (PCA),, and Independent Vector Analysis (IVA) Dr Mohsen Naqvi Lecturer in Signal and Information Processing, School of Electrical and Electronic Engineering,
More informationBlind Source Separation with a Time-Varying Mixing Matrix
Blind Source Separation with a Time-Varying Mixing Matrix Marcus R DeYoung and Brian L Evans Department of Electrical and Computer Engineering The University of Texas at Austin 1 University Station, Austin,
More informationLearning with Ensembles: How. over-tting can be useful. Anders Krogh Copenhagen, Denmark. Abstract
Published in: Advances in Neural Information Processing Systems 8, D S Touretzky, M C Mozer, and M E Hasselmo (eds.), MIT Press, Cambridge, MA, pages 190-196, 1996. Learning with Ensembles: How over-tting
More informationAdaptive Noise Cancellation
Adaptive Noise Cancellation P. Comon and V. Zarzoso January 5, 2010 1 Introduction In numerous application areas, including biomedical engineering, radar, sonar and digital communications, the goal is
More informationA Canonical Genetic Algorithm for Blind Inversion of Linear Channels
A Canonical Genetic Algorithm for Blind Inversion of Linear Channels Fernando Rojas, Jordi Solé-Casals, Enric Monte-Moreno 3, Carlos G. Puntonet and Alberto Prieto Computer Architecture and Technology
More informationRecursive Generalized Eigendecomposition for Independent Component Analysis
Recursive Generalized Eigendecomposition for Independent Component Analysis Umut Ozertem 1, Deniz Erdogmus 1,, ian Lan 1 CSEE Department, OGI, Oregon Health & Science University, Portland, OR, USA. {ozertemu,deniz}@csee.ogi.edu
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More informationCIFAR Lectures: Non-Gaussian statistics and natural images
CIFAR Lectures: Non-Gaussian statistics and natural images Dept of Computer Science University of Helsinki, Finland Outline Part I: Theory of ICA Definition and difference to PCA Importance of non-gaussianity
More informationICA [6] ICA) [7, 8] ICA ICA ICA [9, 10] J-F. Cardoso. [13] Matlab ICA. Comon[3], Amari & Cardoso[4] ICA ICA
16 1 (Independent Component Analysis: ICA) 198 9 ICA ICA ICA 1 ICA 198 Jutten Herault Comon[3], Amari & Cardoso[4] ICA Comon (PCA) projection persuit projection persuit ICA ICA ICA 1 [1] [] ICA ICA EEG
More informationLogistic Regression. Machine Learning Fall 2018
Logistic Regression Machine Learning Fall 2018 1 Where are e? We have seen the folloing ideas Linear models Learning as loss minimization Bayesian learning criteria (MAP and MLE estimation) The Naïve Bayes
More informationPROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata
' / PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE Noboru Murata Waseda University Department of Electrical Electronics and Computer Engineering 3--
More informationoutput dimension input dimension Gaussian evidence Gaussian Gaussian evidence evidence from t +1 inputs and outputs at time t x t+2 x t-1 x t+1
To appear in M. S. Kearns, S. A. Solla, D. A. Cohn, (eds.) Advances in Neural Information Processing Systems. Cambridge, MA: MIT Press, 999. Learning Nonlinear Dynamical Systems using an EM Algorithm Zoubin
More informationNonlinear Projection Trick in kernel methods: An alternative to the Kernel Trick
Nonlinear Projection Trick in kernel methods: An alternative to the Kernel Trick Nojun Kak, Member, IEEE, Abstract In kernel methods such as kernel PCA and support vector machines, the so called kernel
More informationIndependent Components Analysis
CS229 Lecture notes Andrew Ng Part XII Independent Components Analysis Our next topic is Independent Components Analysis (ICA). Similar to PCA, this will find a new basis in which to represent our data.
More informationA ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT
Progress In Electromagnetics Research Letters, Vol. 16, 53 60, 2010 A ROBUST BEAMFORMER BASED ON WEIGHTED SPARSE CONSTRAINT Y. P. Liu and Q. Wan School of Electronic Engineering University of Electronic
More informationIndependent Component Analysis
A Short Introduction to Independent Component Analysis with Some Recent Advances Aapo Hyvärinen Dept of Computer Science Dept of Mathematics and Statistics University of Helsinki Problem of blind source
More informationUndercomplete Independent Component. Analysis for Signal Separation and. Dimension Reduction. Category: Algorithms and Architectures.
Undercomplete Independent Component Analysis for Signal Separation and Dimension Reduction John Porrill and James V Stone Psychology Department, Sheeld University, Sheeld, S10 2UR, England. Tel: 0114 222
More informationIndependent Component Analysis. Contents
Contents Preface xvii 1 Introduction 1 1.1 Linear representation of multivariate data 1 1.1.1 The general statistical setting 1 1.1.2 Dimension reduction methods 2 1.1.3 Independence as a guiding principle
More informationElec4621 Advanced Digital Signal Processing Chapter 11: Time-Frequency Analysis
Elec461 Advanced Digital Signal Processing Chapter 11: Time-Frequency Analysis Dr. D. S. Taubman May 3, 011 In this last chapter of your notes, we are interested in the problem of nding the instantaneous
More informationNeural networks and support vector machines
Neural netorks and support vector machines Perceptron Input x 1 Weights 1 x 2 x 3... x D 2 3 D Output: sgn( x + b) Can incorporate bias as component of the eight vector by alays including a feature ith
More informationComplete Blind Subspace Deconvolution
Complete Blind Subspace Deconvolution Zoltán Szabó Department of Information Systems, Eötvös Loránd University, Pázmány P. sétány 1/C, Budapest H-1117, Hungary szzoli@cs.elte.hu http://nipg.inf.elte.hu
More informationCHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION. From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum
CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION From Exploratory Factor Analysis Ledyard R Tucker and Robert C. MacCallum 1997 19 CHAPTER 3 THE COMMON FACTOR MODEL IN THE POPULATION 3.0. Introduction
More informationTutorial on Blind Source Separation and Independent Component Analysis
Tutorial on Blind Source Separation and Independent Component Analysis Lucas Parra Adaptive Image & Signal Processing Group Sarnoff Corporation February 09, 2002 Linear Mixtures... problem statement...
More informationPerformance Comparison of Two Implementations of the Leaky. LMS Adaptive Filter. Scott C. Douglas. University of Utah. Salt Lake City, Utah 84112
Performance Comparison of Two Implementations of the Leaky LMS Adaptive Filter Scott C. Douglas Department of Electrical Engineering University of Utah Salt Lake City, Utah 8411 Abstract{ The leaky LMS
More informationDETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION. Alexandre Iline, Harri Valpola and Erkki Oja
DETECTING PROCESS STATE CHANGES BY NONLINEAR BLIND SOURCE SEPARATION Alexandre Iline, Harri Valpola and Erkki Oja Laboratory of Computer and Information Science Helsinki University of Technology P.O.Box
More informationMassoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39
Blind Source Separation (BSS) and Independent Componen Analysis (ICA) Massoud BABAIE-ZADEH Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39 Outline Part I Part II Introduction
More informationsection 3 we will reformulate the problem of ICA, that is the recovery of source signals from the linear instantaneous mix, as a Maximum Likelihood pr
Temporal Models in Blind Source Separation Lucas C. Parra Sarno Corporation, CN-5300, Princeton, NJ 08543 1 Introduction Independent component analysis (ICA) aims to nd statistical independent signals
More information1 Introduction The Separation of Independent Sources (SIS) assumes that some unknown but independent temporal signals propagate through a mixing and/o
Appeared in IEEE Trans. on Circuits and Systems, vol. 42, no. 11, pp. 748-751, November 95 c Implementation and Test Results of a Chip for The Separation of Mixed Signals Ammar B. A. Gharbi and Fathi M.
More informationDistinguishing Causes from Effects using Nonlinear Acyclic Causal Models
JMLR Workshop and Conference Proceedings 6:17 164 NIPS 28 workshop on causality Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Kun Zhang Dept of Computer Science and HIIT University
More informationDistinguishing Causes from Effects using Nonlinear Acyclic Causal Models
Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models Kun Zhang Dept of Computer Science and HIIT University of Helsinki 14 Helsinki, Finland kun.zhang@cs.helsinki.fi Aapo Hyvärinen
More informationAn Iterative Blind Source Separation Method for Convolutive Mixtures of Images
An Iterative Blind Source Separation Method for Convolutive Mixtures of Images Marc Castella and Jean-Christophe Pesquet Université de Marne-la-Vallée / UMR-CNRS 8049 5 bd Descartes, Champs-sur-Marne 77454
More informationwhere A 2 IR m n is the mixing matrix, s(t) is the n-dimensional source vector (n» m), and v(t) is additive white noise that is statistically independ
BLIND SEPARATION OF NONSTATIONARY AND TEMPORALLY CORRELATED SOURCES FROM NOISY MIXTURES Seungjin CHOI x and Andrzej CICHOCKI y x Department of Electrical Engineering Chungbuk National University, KOREA
More informationUndercomplete Blind Subspace Deconvolution via Linear Prediction
Undercomplete Blind Subspace Deconvolution via Linear Prediction Zoltán Szabó, Barnabás Póczos, and András L rincz Department of Information Systems, Eötvös Loránd University, Pázmány P. sétány 1/C, Budapest
More informationA METHOD OF ICA IN TIME-FREQUENCY DOMAIN
A METHOD OF ICA IN TIME-FREQUENCY DOMAIN Shiro Ikeda PRESTO, JST Hirosawa 2-, Wako, 35-98, Japan Shiro.Ikeda@brain.riken.go.jp Noboru Murata RIKEN BSI Hirosawa 2-, Wako, 35-98, Japan Noboru.Murata@brain.riken.go.jp
More informationON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS
Yugoslav Journal of Operations Research 5 (25), Number, 79-95 ON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS Slavica TODOROVIĆ-ZARKULA EI Professional Electronics, Niš, bssmtod@eunet.yu Branimir TODOROVIĆ,
More informationBLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES
BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES Dinh-Tuan Pham Laboratoire de Modélisation et Calcul URA 397, CNRS/UJF/INPG BP 53X, 38041 Grenoble cédex, France Dinh-Tuan.Pham@imag.fr
More informationAcoustic Source Separation with Microphone Arrays CCNY
Acoustic Source Separation with Microphone Arrays Lucas C. Parra Biomedical Engineering Department City College of New York CCNY Craig Fancourt Clay Spence Chris Alvino Montreal Workshop, Nov 6, 2004 Blind
More informationAcoustic MIMO Signal Processing
Yiteng Huang Jacob Benesty Jingdong Chen Acoustic MIMO Signal Processing With 71 Figures Ö Springer Contents 1 Introduction 1 1.1 Acoustic MIMO Signal Processing 1 1.2 Organization of the Book 4 Part I
More informationAN ADAPTIVE NONLINEAR FUNCTION CONTROLLED BY ESTIMATED OUTPUT PDF FOR BLIND SOURCE SEPARATION. Kenji Nakayama Akihiro Hirano Takayuki Sakai
AN ADAPTIVE NONLINEAR FUNCTION CONTROLLED BY ESTIMATED OUTPUT PDF FOR BLIND SOURCE SEPARATION Kenji Nakaama Akihiro Hirano Takauki Sakai Dept. of Information Sstems Eng., Facult of Eng., Kanazaa Univ.
More informationBoxlets: a Fast Convolution Algorithm for. Signal Processing and Neural Networks. Patrice Y. Simard, Leon Bottou, Patrick Haner and Yann LeCun
Boxlets: a Fast Convolution Algorithm for Signal Processing and Neural Networks Patrice Y. Simard, Leon Bottou, Patrick Haner and Yann LeCun AT&T Labs-Research 100 Schultz Drive, Red Bank, NJ 07701-7033
More informationCHARACTERIZATION OF ULTRASONIC IMMERSION TRANSDUCERS
CHARACTERIZATION OF ULTRASONIC IMMERSION TRANSDUCERS INTRODUCTION David D. Bennink, Center for NDE Anna L. Pate, Engineering Science and Mechanics Ioa State University Ames, Ioa 50011 In any ultrasonic
More informationIndependent Component Analysis on the Basis of Helmholtz Machine
Independent Component Analysis on the Basis of Helmholtz Machine Masashi OHATA *1 ohatama@bmc.riken.go.jp Toshiharu MUKAI *1 tosh@bmc.riken.go.jp Kiyotoshi MATSUOKA *2 matsuoka@brain.kyutech.ac.jp *1 Biologically
More informationBayesian Properties of Normalized Maximum Likelihood and its Fast Computation
Bayesian Properties of Normalized Maximum Likelihood and its Fast Computation Andre Barron Department of Statistics Yale University Email: andre.barron@yale.edu Teemu Roos HIIT & Dept of Computer Science
More informationTo appear in Proceedings of the ICA'99, Aussois, France, A 2 R mn is an unknown mixture matrix of full rank, v(t) is the vector of noises. The
To appear in Proceedings of the ICA'99, Aussois, France, 1999 1 NATURAL GRADIENT APPROACH TO BLIND SEPARATION OF OVER- AND UNDER-COMPLETE MIXTURES L.-Q. Zhang, S. Amari and A. Cichocki Brain-style Information
More informationPreliminaries. Definition: The Euclidean dot product between two vectors is the expression. i=1
90 8 80 7 70 6 60 0 8/7/ Preliminaries Preliminaries Linear models and the perceptron algorithm Chapters, T x + b < 0 T x + b > 0 Definition: The Euclidean dot product beteen to vectors is the expression
More informationForecasting Time Series by SOFNN with Reinforcement Learning
Forecasting Time Series by SOFNN ith einforcement Learning Takashi Kuremoto, Masanao Obayashi, and Kunikazu Kobayashi Abstract A self-organized fuzzy neural netork (SOFNN) ith a reinforcement learning
More informationBusiness Cycles: The Classical Approach
San Francisco State University ECON 302 Business Cycles: The Classical Approach Introduction Michael Bar Recall from the introduction that the output per capita in the U.S. is groing steady, but there
More informationLECTURE :ICA. Rita Osadchy. Based on Lecture Notes by A. Ng
LECURE :ICA Rita Osadchy Based on Lecture Notes by A. Ng Cocktail Party Person 1 2 s 1 Mike 2 s 3 Person 3 1 Mike 1 s 2 Person 2 3 Mike 3 microphone signals are mied speech signals 1 2 3 ( t) ( t) ( t)
More informationROBUSTNESS OF PARAMETRIC SOURCE DEMIXING IN ECHOIC ENVIRONMENTS. Radu Balan, Justinian Rosca, Scott Rickard
ROBUSTNESS OF PARAMETRIC SOURCE DEMIXING IN ECHOIC ENVIRONMENTS Radu Balan, Justinian Rosca, Scott Rickard Siemens Corporate Research Multimedia and Video Technology Princeton, NJ 5 fradu.balan,justinian.rosca,scott.rickardg@scr.siemens.com
More informationRegression coefficients may even have a different sign from the expected.
Multicolinearity Diagnostics : Some of the diagnostics e have just discussed are sensitive to multicolinearity. For example, e kno that ith multicolinearity, additions and deletions of data cause shifts
More informationDefinition of a new Parameter for use in Active Structural Acoustic Control
Definition of a ne Parameter for use in Active Structural Acoustic Control Brigham Young University Abstract-- A ne parameter as recently developed by Jeffery M. Fisher (M.S.) for use in Active Structural
More informationManaging Uncertainty
Managing Uncertainty Bayesian Linear Regression and Kalman Filter December 4, 2017 Objectives The goal of this lab is multiple: 1. First it is a reminder of some central elementary notions of Bayesian
More informationCOS 424: Interacting with Data. Lecturer: Rob Schapire Lecture #15 Scribe: Haipeng Zheng April 5, 2007
COS 424: Interacting ith Data Lecturer: Rob Schapire Lecture #15 Scribe: Haipeng Zheng April 5, 2007 Recapitulation of Last Lecture In linear regression, e need to avoid adding too much richness to the
More informationSemi-Blind approaches to source separation: introduction to the special session
Semi-Blind approaches to source separation: introduction to the special session Massoud BABAIE-ZADEH 1 Christian JUTTEN 2 1- Sharif University of Technology, Tehran, IRAN 2- Laboratory of Images and Signals
More informationMODULE -4 BAYEIAN LEARNING
MODULE -4 BAYEIAN LEARNING CONTENT Introduction Bayes theorem Bayes theorem and concept learning Maximum likelihood and Least Squared Error Hypothesis Maximum likelihood Hypotheses for predicting probabilities
More informationChapter 7 Interconnected Systems and Feedback: Well-Posedness, Stability, and Performance 7. Introduction Feedback control is a powerful approach to o
Lectures on Dynamic Systems and Control Mohammed Dahleh Munther A. Dahleh George Verghese Department of Electrical Engineering and Computer Science Massachuasetts Institute of Technology c Chapter 7 Interconnected
More informationTIME DOMAIN ACOUSTIC CONTRAST CONTROL IMPLEMENTATION OF SOUND ZONES FOR LOW-FREQUENCY INPUT SIGNALS
TIME DOMAIN ACOUSTIC CONTRAST CONTROL IMPLEMENTATION OF SOUND ZONES FOR LOW-FREQUENCY INPUT SIGNALS Daan H. M. Schellekens 12, Martin B. Møller 13, and Martin Olsen 4 1 Bang & Olufsen A/S, Struer, Denmark
More informationFile: ica tutorial2.tex. James V Stone and John Porrill, Psychology Department, Sheeld University, Tel: Fax:
File: ica tutorial2.tex Independent Component Analysis and Projection Pursuit: A Tutorial Introduction James V Stone and John Porrill, Psychology Department, Sheeld University, Sheeld, S 2UR, England.
More informationEnhancing Generalization Capability of SVM Classifiers with Feature Weight Adjustment
Enhancing Generalization Capability of SVM Classifiers ith Feature Weight Adjustment Xizhao Wang and Qiang He College of Mathematics and Computer Science, Hebei University, Baoding 07002, Hebei, China
More informationTRINICON: A Versatile Framework for Multichannel Blind Signal Processing
TRINICON: A Versatile Framework for Multichannel Blind Signal Processing Herbert Buchner, Robert Aichner, Walter Kellermann {buchner,aichner,wk}@lnt.de Telecommunications Laboratory University of Erlangen-Nuremberg
More informationFCOMBI Algorithm for Fetal ECG Extraction
I J C A, 8(5), 05, pp. 997-003 International Science Press FCOMBI Algorithm for Fetal ECG Extraction Breesha S.R.*, X. Felix Joseph** and L. Padma Suresh*** Abstract: Fetal ECG extraction has an important
More informationA NEW VIEW OF ICA. G.E. Hinton, M. Welling, Y.W. Teh. S. K. Osindero
( ( A NEW VIEW OF ICA G.E. Hinton, M. Welling, Y.W. Teh Department of Computer Science University of Toronto 0 Kings College Road, Toronto Canada M5S 3G4 S. K. Osindero Gatsby Computational Neuroscience
More informationA Generalization of a result of Catlin: 2-factors in line graphs
AUSTRALASIAN JOURNAL OF COMBINATORICS Volume 72(2) (2018), Pages 164 184 A Generalization of a result of Catlin: 2-factors in line graphs Ronald J. Gould Emory University Atlanta, Georgia U.S.A. rg@mathcs.emory.edu
More informationMATHEMATICAL MODEL OF IMAGE DEGRADATION. = s
MATHEMATICAL MODEL OF IMAGE DEGRADATION H s u v G s u v F s u v ^ F u v G u v H s u v Gaussian Kernel Source: C. Rasmussen Gaussian filters pixel 5 pixels 0 pixels 30 pixels Gaussian filter Removes high-frequency
More information1.1. Contributions. The most important feature of problem (1.1) is that A is
FAST AND STABLE ALGORITHMS FOR BANDED PLUS SEMISEPARABLE SYSTEMS OF LINEAR EQUATIONS S. HANDRASEKARAN AND M. GU y Abstract. We present fast and numerically stable algorithms for the solution of linear
More informationBlind channel deconvolution of real world signals using source separation techniques
Blind channel deconvolution of real world signals using source separation techniques Jordi Solé-Casals 1, Enric Monte-Moreno 2 1 Signal Processing Group, University of Vic, Sagrada Família 7, 08500, Vic
More information3 - Vector Spaces Definition vector space linear space u, v,
3 - Vector Spaces Vectors in R and R 3 are essentially matrices. They can be vieed either as column vectors (matrices of size and 3, respectively) or ro vectors ( and 3 matrices). The addition and scalar
More informationMax-Margin Ratio Machine
JMLR: Workshop and Conference Proceedings 25:1 13, 2012 Asian Conference on Machine Learning Max-Margin Ratio Machine Suicheng Gu and Yuhong Guo Department of Computer and Information Sciences Temple University,
More informationUnsupervised Feature Extraction by Time-Contrastive Learning and Nonlinear ICA
Unsupervised Feature Extraction by Time-Contrastive Learning and Nonlinear ICA with Hiroshi Morioka Dept of Computer Science University of Helsinki, Finland Facebook AI Summit, 13th June 2016 Abstract
More informationBLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS
BLIND DECONVOLUTION ALGORITHMS FOR MIMO-FIR SYSTEMS DRIVEN BY FOURTH-ORDER COLORED SIGNALS M. Kawamoto 1,2, Y. Inouye 1, A. Mansour 2, and R.-W. Liu 3 1. Department of Electronic and Control Systems Engineering,
More informationNetworks of McCulloch-Pitts Neurons
s Lecture 4 Netorks of McCulloch-Pitts Neurons The McCulloch and Pitts (M_P) Neuron x x sgn x n Netorks of M-P Neurons One neuron can t do much on its on, but a net of these neurons x i x i i sgn i ij
More informationA Modified ν-sv Method For Simplified Regression A. Shilton M. Palaniswami
A Modified ν-sv Method For Simplified Regression A Shilton apsh@eemuozau), M Palanisami sami@eemuozau) Department of Electrical and Electronic Engineering, University of Melbourne, Victoria - 3 Australia
More informationOn the Slow Convergence of EM and VBEM in Low-Noise Linear Models
NOTE Communicated by Zoubin Ghahramani On the Slow Convergence of EM and VBEM in Low-Noise Linear Models Kaare Brandt Petersen kbp@imm.dtu.dk Ole Winther owi@imm.dtu.dk Lars Kai Hansen lkhansen@imm.dtu.dk
More informationSimultaneous Diagonalization in the Frequency Domain (SDIF) for Source Separation
Simultaneous Diagonalization in the Frequency Domain (SDIF) for Source Separation Hsiao-Chun Wu and Jose C. Principe Computational Neuro-Engineering Laboratory Department of Electrical and Computer Engineering
More information