BLIND SOURCE SEPARATION IN POST NONLINEAR MIXTURES

Similar documents
FAST ALGORITHM FOR ESTIMATING MUTUAL INFORMATION, ENTROPIES AND SCORE FUNCTIONS. Dinh Tuan Pham

BLIND SEPARATION OF TEMPORALLY CORRELATED SOURCES USING A QUASI MAXIMUM LIKELIHOOD APPROACH

MINIMIZATION-PROJECTION (MP) APPROACH FOR BLIND SOURCE SEPARATION IN DIFFERENT MIXING MODELS

Massoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39

Analytical solution of the blind source separation problem using derivatives

Markovian Source Separation

Semi-Blind approaches to source separation: introduction to the special session

BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES

A Canonical Genetic Algorithm for Blind Inversion of Linear Channels

Blind Source Separation in Nonlinear Mixture for Colored Sources Using Signal Derivatives

By choosing to view this document, you agree to all provisions of the copyright laws protecting it.

Blind compensation of polynomial mixtures of Gaussian signals with application in nonlinear blind source separation

Blind Signal Separation: Statistical Principles

BLIND SEPARATION OF INSTANTANEOUS MIXTURES OF NON STATIONARY SOURCES

Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models

Distinguishing Causes from Effects using Nonlinear Acyclic Causal Models

Blind separation of instantaneous mixtures of dependent sources

One-unit Learning Rules for Independent Component Analysis

Entropy Manipulation of Arbitrary Non I inear Map pings

Least square joint diagonalization of matrices under an intrinsic scale constraint

Blind channel deconvolution of real world signals using source separation techniques

Gatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II

1 Introduction Independent component analysis (ICA) [10] is a statistical technique whose main applications are blind source separation, blind deconvo

Natural Gradient Learning for Over- and Under-Complete Bases in ICA

CIFAR Lectures: Non-Gaussian statistics and natural images

On Information Maximization and Blind Signal Deconvolution

Variational Methods in Bayesian Deconvolution

Linear Regression. Aarti Singh. Machine Learning / Sept 27, 2010

An Iterative Blind Source Separation Method for Convolutive Mixtures of Images

Statistical Geometry Processing Winter Semester 2011/2012

BLIND INVERSION OF WIENER SYSTEM USING A MINIMIZATION-PROJECTION (MP) APPROACH

An Improved Cumulant Based Method for Independent Component Analysis

PROPERTIES OF THE EMPIRICAL CHARACTERISTIC FUNCTION AND ITS APPLICATION TO TESTING FOR INDEPENDENCE. Noboru Murata

Recursive Generalized Eigendecomposition for Independent Component Analysis

Linear Regression and Its Applications

Independent Component Analysis

Blind Source Separation with a Time-Varying Mixing Matrix

BLIND source separation (BSS) aims at recovering a vector

Wavelet de-noising for blind source separation in noisy mixtures.

Chapter 9. Non-Parametric Density Function Estimation

Lecture 16: Small Sample Size Problems (Covariance Estimation) Many thanks to Carlos Thomaz who authored the original version of these slides

Getting Started with Communications Engineering

Using Kernel PCA for Initialisation of Variational Bayesian Nonlinear Blind Source Separation Method

BLIND SEPARATION OF SOURCES: A NON-LINEAR NEURAL ALGORITHM

Fundamentals of Principal Component Analysis (PCA), Independent Component Analysis (ICA), and Independent Vector Analysis (IVA)

Blind Machine Separation Te-Won Lee

Cheng Soon Ong & Christian Walder. Canberra February June 2018

MEASUREMENTS that are telemetered to the control

Remaining energy on log scale Number of linear PCA components

To appear in Proceedings of the ICA'99, Aussois, France, A 2 R mn is an unknown mixture matrix of full rank, v(t) is the vector of noises. The

SPARSE signal representations have gained popularity in recent

CPSC 340: Machine Learning and Data Mining. Sparse Matrix Factorization Fall 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Chapter 9. Non-Parametric Density Function Estimation

IDENTIFIABILITY AND SEPARABILITY OF LINEAR ICA MODELS REVISITED. Jan Eriksson and Visa Koivunen

Lecture 3 September 1

Nonparametric Bayesian Methods (Gaussian Processes)

On Application Of Modified Gradient To Extended Conjugate Gradient Method Algorithm For Solving Optimal Control Problems

Trust Regions. Charles J. Geyer. March 27, 2013

Scan Time Optimization for Post-injection PET Scans

Bayesian Inference of Noise Levels in Regression

ON-LINE BLIND SEPARATION OF NON-STATIONARY SIGNALS

ASIGNIFICANT research effort has been devoted to the. Optimal State Estimation for Stochastic Systems: An Information Theoretic Approach

Topic 5: Non-Linear Relationships and Non-Linear Least Squares

CPSC 340: Machine Learning and Data Mining. More PCA Fall 2017

MACHINE LEARNING. Methods for feature extraction and reduction of dimensionality: Probabilistic PCA and kernel PCA

Neural Network Training

Connections between score matching, contrastive divergence, and pseudolikelihood for continuous-valued variables. Revised submission to IEEE TNN

On the convergence of the iterative solution of the likelihood equations

Designing Information Devices and Systems I Spring 2018 Lecture Notes Note 3

Lecture 5: Logistic Regression. Neural Networks

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

THE SPIKE PROCESS: A SIMPLE TEST CASE FOR INDEPENDENT OR SPARSE COMPONENT ANALYSIS. Naoki Saito and Bertrand Bénichou

5601 Notes: The Sandwich Estimator

Nonlinear Support Vector Machines through Iterative Majorization and I-Splines

TWO METHODS FOR ESTIMATING OVERCOMPLETE INDEPENDENT COMPONENT BASES. Mika Inki and Aapo Hyvärinen

Exploring the energy landscape

2 Tikhonov Regularization and ERM

Stability of Switched Linear Hyperbolic Systems by Lyapunov Techniques

SOURCE SEPARATION OF BASEBAND SIGNALS IN POST-NONLINEAR MIXTURES

Nonlinear reverse-correlation with synthesized naturalistic noise

ON SOME EXTENSIONS OF THE NATURAL GRADIENT ALGORITHM. Brain Science Institute, RIKEN, Wako-shi, Saitama , Japan

INDEPENDENT COMPONENT ANALYSIS (ICA) is a

Implicitly and Explicitly Constrained Optimization Problems for Training of Recurrent Neural Networks

Statistics: Learning models from data

Efficient Line Search Methods for Riemannian Optimization Under Unitary Matrix Constraint

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Analysis of Interest Rate Curves Clustering Using Self-Organising Maps

Chapter 0 of Calculus ++, Differential calculus with several variables

On-Line Hardware Implementation for Complex Exponential and Logarithm

Morphological Diversity and Source Separation

Upon successful completion of MATH 220, the student will be able to:

ROBERTO BATTITI, MAURO BRUNATO. The LION Way: Machine Learning plus Intelligent Optimization. LIONlab, University of Trento, Italy, Apr 2015

Machine Learning. Neural Networks. (slides from Domingos, Pardo, others)

On the convergence of the iterative solution of the likelihood equations

arxiv: v1 [cs.sc] 17 Apr 2013

Hamiltonian Dynamics In The Theory of Abstraction

Non-Euclidean Independent Component Analysis and Oja's Learning

HST.582J/6.555J/16.456J

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Transcription:

BLIND SURCE SEPARATIN IN PST NNLINEAR MIXTURES Sophie Achard, Dinh Tuan Pham Univ of Grenoble Laboratory of Modeling Computation, IMAG, CNRS BP 53X, 384 Grenoble Cedex, France SophieAchard@imagfr, DinhTuanPham@imagfr Christian Jutten Laboratory of Images Signals, INPG 46 avenue Félix Viallet, 383 Grenoble Cedex, France ChristianJutten@inpgfr ABSTRACT This work implements alternative algorithms to that of Taleb Jutten for blind source separation in post nonlinear mixtures We use the same mutual information criterion as them, but we exploit its invariance with respect to translation to express its relative gradient in terms of the derivatives of the nonlinear transformations Then we develop algorithms based on these derivatives In a semiparametric approach, the latter are parametrized by piecewise constant functions All algorithms require only the estimation of the score functions of the reconstructed sources A new method for score function estimation is presented for this purpose INTRDUCTIN Blind source separation has been well studied in the case of linear mixtures [, 2, 3, 4], but only begins to attract attention in the case of nonlinear mixtures Yet in many situations, it is more realistic to assume a nonlinear rather than a linear mixture The main difficulty in using this class of mixtures is that they are too broad thus lead to the problem of identifiability Indeed Darmois [5] has shown that there exist nonlinear mixtures of sources which preserve their independence To avoid this difficulty, in this paper we shall follow Taleb Jutten [6] by restricting the mixture to the smaller class of post nonlinear mixtures In such mixture, the observed channels,, are related to the sources through the relations where denotes the general element of the mixing matrix are nonlinear functions It is assumed here that there are the same number of sources sensors the matrix is invertible the functions are Thanks to the European Bliss Project for funding monotonous so that the sources can be recovered from the observations if one knew The goal of blind source separation is to recover the sources without any particular knowledge on their distributions the mixing mecanism The separation will be based solely on the mutual independence assumption of the sources Specifically, one tries to find a matrix applications so that the rom variables "# $ % % where $ & which represent the reconstructed sources, are as independent as it is possible Using the mutual information as the natural measure of independence, Taleb Jutten [6] has obtained the criterion )(* () )+, +, $ /2354768:9 4 These authors have derived a separation algorithm based on the relative gradient of the criterion with respect to,, ur approach stems from the remark that the above criterion, due to its invariance with respect to translation, is actually a function of the derivatives of not of these functions themselves Thus we shall compute the gradient of the criterion with respect to these derivatives obtain new algorithms by the gradient descent method They are described in sections 2 3 Section 4 introduces a new method for the score function estimation ur algorithms have the nice property that they can be expessed entirely in terms of the estimated score functions of the sources, hence these estimations play a crucial role Section 5 discusses the numerical implementation of the algorithms some simulation results are presented in section 6 Finally, some comments discussions about our algorithms are made in section 7 (2) 295

U U G 2 RELATIVE GRADIENT We begin by computing the relative gradient of the criterion with respect to the derivatives of The general result can then be specialized to the case where these functions are parametrized 2 Nonparametric approach To obtain the linear functional of the relative gradient of, we express the first order Taylor expansion of with respect to around We then deduce the relative gradient functional " " " / $5 " $#&%( *),+ where is the score function of (ie /23#(&, #( being the density of ) l is the indicator function of (ie l / if /2, else) By setting this gradient to zero, we deduce the following estimating equations: : / $5 3 5476 894 =< #&%( 54>24 ne can verify that these equations are indeed satisfied if the variables,, " are independent 22 Semiparametric approach The above approach allows arbitrary values for the functions thus could give them too much degree of freedom, ignoring the fact that they are generally smooth functions It is thus of interest to consider a semiparametric approach which restricts the degree of freedom of these functions but still allows a sufficiently rich nonlinear mixture model ur approach consists of representing the functions by continuous piecewise linear functions The main reason, besides simplicity, is that the space of piecewise linear functions is stable with respect to composition This will simplify considerably the calculation of the relative gradient, as the increments in subsection 2 will be also piecewise linear As we will show shortly, our general formula can be specialized to this case with ease Note that the function cannot (need not) be defined outside the support? @? @ A9 of the distribution of Although? @? @ A could theoretically be, in practice, one can be satisfied with them being the minimum the maximum of the observed values of, since there is no data outside? @ :? @ A hence it is not necessary to estimate outside this interval In this interval, let us denote by? @ B? @ ADC the change points in slope of the function, arranged in increasing order As is linear in? @ E? @ E7, F,, G, it is clear that its relative increment in subsection 2 is also linear in H:@ E H:@ E7, F,, G, where H:@ E? @ E5, F Therefore, the derivatives of can be repesented as ADC E +@ E / I J K L @ J K L&MN I to be in being increasing We will see in section 52 how this condition can be easily checked in the formula Without loss of generality, we take creasing, which is equivalent to the H @ E By replacing the above expression for for relative gradient, we obtain, Relative gradient of according to : " " " Relative gradient of according to +@ : +@ ADC : +@ +@ ADC ADC J K L&MN P E J K L l $5 RQD+ 5S H:@ E 4 $9TH:@ E7,+@ E 54>4 3 GRADIENT DESCENT METHD The minimization of the criterion is achieved through the gradient descent method The method consists of making successive small relative changes in in the opposite direction of the relative gradient so as to decrease the criterion Explicitly, one adjusts by the following iteration, V&W YXZ [4>4 where V X are two tuning parameters W Z are obtained from the relative gradient 3 Nonparametric approach Here the matrix W has general element W 3 3 5476 894% else 296

U As for Z, 54>4%, there are two possibilities according to the adopted metric: with the normal metric (of the Hilbert space of square summable functions with respect to the Lebesgue measure): Z / $5 " $#&%( with the probabilistic metric (of the Hilbert space of square summable functions with respect to the probability measure): Z #&%( / $5 " 32 Semiparametric approach The matrix W is given by the same formula as above while Z is given by ADC Z +@ E E / I J K L J K L&MN I where +@ E, F G, >, are given in two different ways according to the adopted metric: with the Lebesgue metric (associated to the scalar product +@ + @ +@ E + @ E H:@ E7 H:@ E5 ): +@ E E7 E H YH J K L&MN J K L : / $5 S H:@ E 4 $9TH:@ E7 =<9+&Q with the probabilistic metric (associated to the scalar product +@ + @ +@ E5+ @ E S H:@ E 4 $ 4 H:@ E7 ): +@ E S H:@ E 4 $9TH:@ E7 J K L&MN : J K L l $5 4 SCRE FUNCTIN ESTIMATIN =< + In all our algorithms, only the score functions of the reconstructed sources " are involved Different methods for score function estimation thus provide different algorithms Note that unlike the linear mixture case where the score functions need not to be accurately estimated, since the estimation equations are still satisfied even with a wrong score function, in the present case, the accuracy of the score function estimation is crucial, since the estimating equations are not satisfied when the score functions are wrong Several score function estimation methods are available: the kernel method (used by Taleb Jutten [6]), the spline method [7, 8] We present here a new method derived from an entropy estimator It is suggested by the following expansion of the entropy +, # +, terms of higher order in (3) where is a rom variable, is a small rom increment denotes the score function of Let be a sample from an estimator of +, based on this sample The relation (3) suggests estimating at the sampling points by 54D4 It is desirable that be invariant with respect to translation equivariant with respect to scaling, that is /23 4 4 for any real number, since the entropy functional satisfies these properties By differentiating the above equalities with respect to then putting, one gets (4) (5) The above interesting properties of mimic that of the true score function:, which can be easily obtained by integration by part As an estimator of the entropy, one may take a discretization of its defining integral: /23 # " X#$&% # " X# ($&% % /23 where # is an estimate of the density of, X) are the sample mean stard deviation of, % represents the discretization step Taking #* to be a kernel estimator of the density of with the smoothing parameter proportional to its sample stard deviation, one can show that the estimator possesses the desired translation invariance scale equivariance properties ne can also use an empirical estimator of the mean, leading to /23 # 297

E P F B B This estimator again possesses the desired translation invariance scale equivariance properties, if one takes # to be a kernel estimator of the density of with a smoothing parameter proportional to its sample stard deviation This second form of the entropy estimator, although appears simpler than the first, can actually be costlier computationally The reason is that the first form involves only the estimated density at a regular grid points computation of the kernel density over such a grid can be efficiently done through the binning technique Further, the number of the grid points can usually be much less than the sample size without incurring any appreciable loss of accuracy Finally, the probability density function #R%( the probabilities S H:@ E 4 $T[H:@ E7, which are needed in the algorithm, can also be estimated via the estimated score function Indeed, noting that the density # of satisfies # 5 /, we propose to estimate it by / This expression can be seen to be constant in in each interval E E7, F, where 4 4 are the order statistics of the sample By integration, one obtains the estimator S E 4 4 E E7 5 NUMERICAL IMPLEMENTATIN 5 Algorithms Let / @,,, denote the observations recorded by the > th sensor put @ / @ The algorithm developed by Taleb Jutten [9] changes at each iteration the matrix the @ Based on the previous expression of the relative gradient, the proposed method adapts the matrix the differences * B C instead, where * E / E / 4 4/ are the order statistics of The values of the @, hence of the @ can be recovered by centering, for example The proposed algorithm, in the nonparametric approach with probabilistic metric, is described in the boxed text below, in which S#@ E is an estimation of E E7 S#@ E S / 4 9T / S E 4 E7 $9T & is the permutation defined by * E @= X V are two small tuning parameters The algotrihms with the normal metric in the semiparametric approach are quite similar, so we don t detail them E Initialisations : loop : $ an estimation of, # E7 E E7 E L&MN C L X X K L &E7 >*, Normalization of $ where V W W 35 $ Normalization of until convergence 52 Normalisations @ 3 @ @= sinon 5476 894% It should be noted that there is a redundancy in the post nonlinear mixture model: the nonlinear transformations,, can be multiplied by arbitrary constant factors the corresponding columns of the matrix divided by the same constants, without changing the mixtures To avoid this ambiguity, which affects the convergence of the algorithm, one should normalize the functions,, (or equivalently the rows of ) Many different possibilities exist, the first which may come to mind is to require that the sample variance of $ >, be ( ) But we find more convenient to require that preserve the entropy, that is /2354 / 4# /,+ /, >, ( ), since our algorithm works with the derivatives hence the last integral can be easily estimated This still leaves the sign of undetermined For definiteness, we require that E7 be positive, E which can be easily checked by looking at In blind source separation, it is well known that one can only recover the sources up to a scaling a permutation However, instead of normalizing " at each iteration step, we shall modify our algorithm to make it invariant with respect to scale change, by taking as W 3 53: : 3 var &3 &3 " var &3 < 298

The denominators in the above formula are actually the diagonal terms of the Hessian of the criterion with respect to the linear transformations 6 EXPERIMENTAL RESULTS 32 iterations =/ / B / 9 / Figure presents some comparison of the score function estimation by three different methods ne can see that the method with the splines with the entropy discretization are rather similar But the method with kernels is rather bad near the end points of the interval This may be explained by the fact that the first two method possess the properties (4) (5) while the kernel method does not nly the result of one simulation is presented here to save space thers simulations performed have shown that the performances of the algorithms depend on the mixture components the estimation of the score function The two metrics also yield different effects on the convergence of the algorithms The coefficients X V which control the gradient descent are fixed empirically 3 2 7 DISCUSSIN 2 3 3 2 2 3 4 3 2 2 3 Estimation of score function of a gaussian variable 4 2 5 5 5 5 2 Estimation of score function of an uniform variable Fig Score function estimations with different methods: dotted line: kernel method, dashed line: spline method, solid line: method with discretization of entropy Next, we present in figure 2, the results of a simulation of the semiparametric method with: a grid with points among the observations the probabilistic metric the score function estimation with the method of the entropy discretization X, V A first comment regards the values of Z at the end points of the interval We observ that these values are often quite large in comparison with the others This may be attributed to the fact that there are too few data to achieve a good estimation of the score function near these points To avoid instability of the algorithm caused by an unduly large value of Z in usually a quite narrow region, we transform non linearly Z into 9 Z When Z is small this doesn t have any effect since then 9 Z Z, but when Z is large, the hyperbolic tangente function reduces it to We find that this device can improve drastically the convergence of the algorithm Another comment regards the coefficients V X which control the gradient descent It seems that keeping them fixed is not a good strategy When one approaches the solution, these coefficents appear to be far too large, causing an oscilatory behaviour of the algorithm may destroy its convergence To avoid that, they must be drastically reduced with the consequence that the algorithm becomes extremely slow A good strategy is to adapt V X such that the criterion is decreased at each iteration Since the theoretical criterion is unknown, an estimator of it must be used But then, in order to ensure that can be decreased, the relative gradient involved in the algorithm must be the gradient of not an estimated relative gradient of Naturally, one would estimate +,& by the same formula (2) but with +, $ replaced by their estimators @ @ :@ :@ Then it can be shown that the relative gradient of is given by the same formula that for the relative gradient of in section 2, except that the score functions %( are replaced by the score estimators of of $ as defined in section 4 that the expectation operator is replaced by a sample average Simulations show that with this method, the algorithms are much more stable 299

8 CNCLUSIN In this paper, we provide alternative methods for blind source separation in post nonlinear mixtures Although we use the same mutual information as Taleb Jutten, our algorithms differ in that they work with the derivatives of the nonlinear transformations are based on the relative gradient of the criterion with respect to these derivatives This approach can be extended easily to a semiparametric setting in which the nonlinear transformations are represented by continuous piecewise linear finctions The method has the nice property that it involves only certain estimated score functions with an adequate choice of the latter, it amounts to minimizing some empirical criterion This can be exploited to better control the convergence of the gradient descent All the algorithms were implemented tested in different situations 5 5 5 5 2 5 5 5 5 2 5 5 Sources 5 9 REFERENCES [] A Bell T Sejnowski, An informationmaximization approach to blind separation blind deconvolution, Neural Computation, vol 7, no 6, 995 [2] DT Pham, Blind separation of instantaneous mixture of sources via an independent component analysis, IEEE Transactions on Signal Processing, vol 44, no, pp 2768 2779, Nov 996 [3] P Comon, Independent component analysis, a new concept?, Signal Processing, vol 3, no 36, pp 287 34, Apr 994 [4] JF Cardoso, Blind signal separation : Statistical principles, Proceedings IEEE, vol, no 86, pp 29 225, ct 998 [5] G Darmois, Analyse gnrale des liaisons stochastiques, Rev Inst Internat Stat, vol 2, pp 2 8, 953 [6] A Taleb C Jutten, Sources separation in postnonlinear mixtures, IEEE Transactions on Signal Processing, vol, no 47, pp 287 282, ct 999 [7] D D Cox, A penalty method for nonparametric estimation of the logarithmic derivative of density function, Ann Inst Statist Math, vol 37, no Part A, pp 27 288, 985 [8] Pin T Ng, Smoothing spline score estimation, SIAM J Sci Compt, vol 5, no 5, pp 3 25, Sept 994 [9] A Taleb, Séparation de Sources dans les Mélanges Non Linéaires, PhD thesis, INPG Laboratoire LIS, 999 5 5 5 5 5 5 5 5 bservations 5 5 5 5 5 5 5 5 5 Distribution of % at convergence 2 2 5 5 5 5 2 5 5 5 Distribution of at convergence g o f 5 4 2 2 4 5 5 5 g2 o f2 5 4 2 2 4 Composition of non linear funtions at convergence Fig 2 Simulation 3