Slide11 Haykin Chapter 10: Information-Theoretic Models

Similar documents
CIFAR Lectures: Non-Gaussian statistics and natural images

Independent Component Analysis and Its Applications. By Qing Xue, 10/15/2004

Gatsby Theoretical Neuroscience Lectures: Non-Gaussian statistics and natural images Parts I-II

Information Theory. Coding and Information Theory. Information Theory Textbooks. Entropy

Efficient Coding. Odelia Schwartz 2017

Independent Component Analysis

Neural coding Ecological approach to sensory coding: efficient adaptation to the natural environment

ECE 4400:693 - Information Theory

Information Theory. Mark van Rossum. January 24, School of Informatics, University of Edinburgh 1 / 35

Introduction to Information Theory. Uncertainty. Entropy. Surprisal. Joint entropy. Conditional entropy. Mutual information.

Dept. of Linguistics, Indiana University Fall 2015

Ch. 8 Math Preliminaries for Lossy Coding. 8.4 Info Theory Revisited

Lecture 11: Continuous-valued signals and differential entropy

Lecture 1: Introduction, Entropy and ML estimation

An Introduction to Independent Components Analysis (ICA)

Expectation Maximization

EE/Stat 376B Handout #5 Network Information Theory October, 14, Homework Set #2 Solutions

Advanced Introduction to Machine Learning CMU-10715

Information maximization in a network of linear neurons

Chapter I: Fundamental Information Theory

Fundamentals of Principal Component Analysis (PCA), Independent Component Analysis (ICA), and Independent Vector Analysis (IVA)

Introduction to Machine Learning

Independent Component Analysis

Lecture 8: Channel Capacity, Continuous Random Variables

Independent Component Analysis

Ch. 8 Math Preliminaries for Lossy Coding. 8.5 Rate-Distortion Theory

Independent Component Analysis. Contents

Lecture 5 Channel Coding over Continuous Channels

Information Theory Primer:

Independent Component Analysis. PhD Seminar Jörgen Ungh

Principles of Communications

Block 2: Introduction to Information Theory

Different Estimation Methods for the Basic Independent Component Analysis Model

3. If a choice is broken down into two successive choices, the original H should be the weighted sum of the individual values of H.

HST.582J/6.555J/16.456J

Independent component analysis: algorithms and applications

Continuous Random Variables

Independent Component Analysis (ICA)

Communication Theory II

EE/Stats 376A: Homework 7 Solutions Due on Friday March 17, 5 pm

Higher Order Statistics

PCA & ICA. CE-717: Machine Learning Sharif University of Technology Spring Soleymani

CPSC 340: Machine Learning and Data Mining. Sparse Matrix Factorization Fall 2018

Lecture 6: Gaussian Channels. Copyright G. Caire (Sample Lectures) 157

Exercises with solutions (Set D)

Dimensionality Reduction. CS57300 Data Mining Fall Instructor: Bruno Ribeiro

Lecture 14 February 28

Unsupervised learning: beyond simple clustering and PCA

Independent Component Analysis (ICA) Bhaskar D Rao University of California, San Diego

Machine learning - HT Maximum Likelihood

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning

Machine Learning (BSMC-GA 4439) Wenke Liu

Natural Image Statistics and Neural Representations

CS 630 Basic Probability and Information Theory. Tim Campbell

Efficient coding of natural images with a population of noisy Linear-Nonlinear neurons

x log x, which is strictly convex, and use Jensen s Inequality:

Random Variables. P(x) = P[X(e)] = P(e). (1)

Massoud BABAIE-ZADEH. Blind Source Separation (BSS) and Independent Componen Analysis (ICA) p.1/39

Lecture 17: Differential Entropy

HST.582J / 6.555J / J Biomedical Signal and Image Processing Spring 2007

Statistical Pattern Recognition

How Behavioral Constraints May Determine Optimal Sensory Representations

Noisy channel communication

Introduction to Independent Component Analysis. Jingmei Lu and Xixi Lu. Abstract

FuncICA for time series pattern discovery

Machine Learning Srihari. Information Theory. Sargur N. Srihari

Lecture 5 - Information theory

Chapter 9 Fundamental Limits in Information Theory

LECTURE 15. Last time: Feedback channel: setting up the problem. Lecture outline. Joint source and channel coding theorem

ELEC546 Review of Information Theory

Revision of Lecture 5

C.M. Liu Perceptual Signal Processing Lab College of Computer Science National Chiao-Tung University

Information Theory in Intelligent Decision Making

9 Classification. 9.1 Linear Classifiers

STATS 306B: Unsupervised Learning Spring Lecture 12 May 7

Lecture 22: Final Review

conditional cdf, conditional pdf, total probability theorem?

Comparative Analysis of ICA Based Features

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

ON SOME EXTENSIONS OF THE NATURAL GRADIENT ALGORITHM. Brain Science Institute, RIKEN, Wako-shi, Saitama , Japan

Lecture 2: Repetition of probability theory and statistics

Statistical signal processing

Independent Component Analysis and Unsupervised Learning. Jen-Tzung Chien

Independent Component Analysis (ICA)

Computational Systems Biology: Biology X

Entropies & Information Theory

Review of Probability Theory

Natural Image Statistics

Nonlinear reverse-correlation with synthesized naturalistic noise

Lecture 5: GPs and Streaming regression

ENGG2430A-Homework 2

Independent Component Analysis and Unsupervised Learning

Lecture 2: August 31

Slide05 Haykin Chapter 5: Radial-Basis Function Networks

EE4601 Communication Systems

Joint Distributions. (a) Scalar multiplication: k = c d. (b) Product of two matrices: c d. (c) The transpose of a matrix:

A review of probability theory

Latent Variable Models and EM algorithm

[POLS 8500] Review of Linear Algebra, Probability and Information Theory

3F1: Signals and Systems INFORMATION THEORY Examples Paper Solutions

Transcription:

Slide11 Haykin Chapter 10: Information-Theoretic Models CPSC 636-600 Instructor: Yoonsuck Choe Spring 2015 ICA section is heavily derived from Aapo Hyvärinen s ICA tutorial: http://www.cis.hut.fi/aapo/papers/ijcnn99_tutorialweb/. Shannon s Information Theory Originally developed to help design communication systems that are efficient and reliable (Shannon, 1948. It is a deep mathematical theory concerned with the essence of the communication process. Provides a framework for: efficiency of information representation, limitations in reliable transmission of information over a communication channel. Gives bounds on optimum representation and transmission of signals. 1 2 Motivation Information-theoretic models that lead to self-organization in a principled manner. Maximum mutual information principle (Linsker 1988: Synaptic connections of a multilayered neural network develop in such a way as to maximize the amount of information preserved when signals are transformed at each processing stage of the network, subject to certain constraints. Information Theory Review Topics to be covered: Entropy Mutual information Relative entropy Differential entropy of continuous random variables Redundancy reduction (Attneave 1954: Major function of perceptual machinary is to strip away some of the redundancy of stimulation, to describe or encode information in a form more economical than that in which it impinges on the receptors. In other words, redundancy reduction = feature extraction. 3 4

Random Variables Notations: X random variable, x value of random variable. If X can take continuous values, theoretically it can carry infinite amount of information. However, this it is meaningless to think of infinite-precision measurement, in most cases values of X can be quantized into a finite number of discrete levels. X = {x k k = 0, ±1,..., ±K} Let event X = x k occur with probability p k = P (X = x k with the requirement 0 p k 1, K k= K 5 p k = 1 Uncertainty, Surprise, Information, and Entropy If p k is 1 (i.e., probability of event X = x k is 1, when X = x k is observed, there is no surprise. You are also pretty sure about the next outcome (X = x k, so you are more certain (i.e., less uncertain. High probability events are less surprising. High probability events are less uncertain. Thus, surprisal/uncertainty of an event are related to the inverse of the probability of that event. You gain information when you go from a high-uncertainty state to a low-uncertainty state. 6 Entropy Uncertainty measure for event X = x k (log assumes log 2 : I(x k = log ( 1 p k = log p k. I(x k = 0 when p k = 1 (no uncertainty, no surprisal. I(x k 0 for 0 p k 1: no negative uncertainty. I(x k > I(x i for p k < p i : more uncertain for less probable events. Average uncertainty = Entropy of a random variable: H(X = E[I(x k ] = K k= K p ki(x k = K k= K p k log p k 7 Properties of Entropy The higher the H(X, the higher the potential information you can gain through observation/measurement. Bounds on the entropy: 0 H(X log(2k + 1 H(X = 0 when p k = 1 and p j = 0 for j k: No uncertainty. H(X = log(2k + 1 when p k = 1/(2K + 1 for all k: Maximum uncertainty, when all events are equiprobable. 8

Properties of Entropy (cont d Max entropy when p k = 1/(2K + 1 for all k follows from ( pk p k log 0 q k k for two probability distributions {p k } and {q k }, with the equality holding when p k = q k for all k. (Multiply both sides with -1. Kullback-Leibler divergence (relative entropy: D p q = x X ( px (x p X (x log q X (x measures how different two probability distributions are (note that it is not symmetric, i.e., D p q D q p. 9 Differential Entropy of Cont. Rand. Variables Differential entropy: h(x = f X (x log f X (xdx = E[log f X (x] Note that H(X, in the limit, does not equal h(x: H(X = lim δx 0 f X (x k δx log(f X (xδx k= }{{} }{{} p k p k = lim δx 0 [ k= f X (x k log(f X (xδx + log(δx f X (x k δx k= = f X (x k log(f X (xdx lim δx 0 log δx f X (xδx = h(x lim δx 0 log δx 10 ] Diff. Entropy of Uniform Distribution Uniform distribution within interval [0, 1]: f X (x = 1 for 0 x 1 and 0 otherwise h(x = 1 log 1dx = 1 0dx = 0. (1 Properties of Differential Entropy h(x + c = h(x h(ax = h(x + log a h(y = f Y (y = 1 a f Y E[log f Y (y] ( y a [ ( ( 1 = E log a f y Y a] [ ( = E log f y Y a] + log a. Plugging in Y = ax to the above, we get the desired result. For vector random variable X, h(ax = h(x + log det(a. 11 12

Maximum Entropy Principle One Dimensional Gaussian Dist. When choosing a probability model given a set of known states of a stochastic system and constraints, there could be potentially an infinite number of choices. Which one to choose? Jaynes (1957 proposed the maximum entropy principle: Pick the probability distribution that maximizes the entropy, subject to constraints on the distribution. Stating the problem in an constrained optimization framework, we can get interesting general results. For a given variance σ 2, the Gaussian random variable has the largest differential entropy attainable by any random variable. The entropy of a Gaussian random variable X is uniquely determined by the variance of X. 13 14 Mutual Information Conditional entropy: What is the entropy in X after observing Y? How much uncertainty remains in X after observing Y? H(X Y = H(X, Y H(Y where the joint-entropy is defined as H(X, Y = x X y Y p(x, y log p(x, y Mutual information: How much uncertainty is reduced in X when we observe Y? The amount of reduced uncertainty is equal to the amount of information we gained! I(X; Y = H(X H(X Y = 15 x X y Y p(x, y p(x, y log p(xp(y Mutual Information for Continuous Random Variables In analogy with the discrete case: I(X; Y = And it has the same property = h(x + h(y h(x, Y ( fx (x y f X,Y (x, y log f X (x I(X; Y = h(x h(x Y 16 = h(y h(y X dxdy

Summary Properties of KL Divergence It is always positive or zero. Zero, when there is a perfect match between the two distributions. It is invariant w.r.t. Permutation of the order in which the components of the vector random variable x are arranged. Amplitude scaling. Monotonic nonlinear transformation. Various relationships among entropy, conditional entropy, joint entropy, and mutual information can be summarized as shown above. It is related to mutual information: I(X; Y = DfX,Y kfx fy 17 18 Application of Information Theory to Neural Network Mutual Information as an Objective Function Learning (a Maximize mutual info between input vector X and output vector Y. We can use mutual information as an objective function to be (b Maximize mutual info between Ya and Yb driven by near-by input vectors Xa and Xb from a single image. optimized when developing learning rules for neural networks. 19 20

Mutual Info. as an Objective Function (cont d Maximum Mutual Information Principle Infomax principle by Linsker (1987, 1988, 1989: Maximize I(Y; X for input vector X and output vector Y. Appealing as the basis for statistical signal processing. Infomax provides a mathematical framework for self-organization. Relation to channel capacity, which defines the Shannon limit on the rate of information transmission through a communication channel. (c Minimize information between Y a and Y b driven by input vectors from different images. (d Minimize statistical dependence between Y i s. 21 22 Example: Single Neuron + Output Noise Single neuron with additive output noise: Y = ( m w i X i + N, where Y is the output, w i the weight, X i the input, and N the processing noise. Assumptions: Output Y is a Gaussian r.v. with variance σ 2 Y. Noise N is also a Gaussian r.v. with µ = 0 and variance σ 2 N. Input and noise are uncorrelated: E[X i N] = 0 for all i. Ex.: Single Neuron + Output Noise (cont d Mutual information between input and output: I(Y ; X = h(y h(y X. Since P (Y X = c + P (N, where c is a constant, h(y X = h(n. Given X, what remains in Y is just noise N. So, we get I(Y ; X = h(y h(n. 23 24

Ex.: Single Neuron + Output Noise (cont d Since both Y and N are Gaussian, So, finally we get: The ratio σ 2 Y /σ2 N h(y = 1 2 [1 + log(2πσ2 Y ] h(n = 1 2 [1 + log(2πσ2 N ] I(Y ; X = 1 2 log ( σ 2 Y σ 2 N can be viewed as a signal-to-noise ratio. If noise variance σn 2 be maximized simply by maximizing the output variance σy 2! is fixed, the mutual information I(Y ; X can 25. Example: Single Neuron + Input Noise Single neuron, with noise on each input line: Y = m We can decompose the above to Y = m w i (X i + N i. w i X i + m w i N i }{{} call this N N is also a Gaussian distribution, with variance: σ 2 N = m 26 w 2 i σ2 N. Example: Single Neuron + Input Noise Lessons Learned As before: h(y X = h(n = 1 2 (1+2πσ2 N = 1 2 Again, we can get the mutual information as: I(Y ; X = h(y h(n = 1 2 log ( [ 1 + 2πσ 2 N σ 2 Y σ 2 N m w2 i m w 2 i ]. Application of Infomax principle is problem-dependent. When m w2 i = 1, then the two additive noise models behave similarly. Assumptions such as Gaussianity need to be justified (it s hard to calculate mutual information without such tricks. Adpoting a Gaussian noise model, we can invoke a surrogate mutual information computed relatively easily. Now, with fixed σn 2, information is maximized by maximizing the ratio σy 2 / m w2 i, where σ2 Y is a function of w i. 27 28

Noiseless Network Noiseless network that transforms a random vector X of arbitrary distribution to a new random vector Y of different distribution: Y = WX. Mutual information in this case is: I(Y; X = H(Y H(Y X. With noiseless mapping, H(Y X attains the lowest value (. However, we can consider the gradient instead: Infomax and Redundancy Reduction In Shannon s framework, Order and structure = Redundancy. Increase in the above reduces uncertainty. More redundancy in the signal implies less information conveyed. More information conveyed means less redundancy. Thus, Infomax principle leads to reduced reduncancy in output Y compared to input X. I(Y; X W = H(Y W. When noise is present: Input noise: add redundancy in input to combat noise. Since H(Y X is independent of W, it drops out. Maximizing mutual information between input and output is equivalent ot maximing entropy in the output, both with respect to the weight matrix W (Bell and Sejnowski 1995. 29 Output noise: add more output components to combat noise. High level of noise favors redundancy of representation. Low level of noise favors diversity of representation. 30 Modeling of a Perceptual System Importance of redundancy in sensory messages: Attneave (1954, Barlow (1959. Redundancy provides knowledge that enables the brain to build cognitive maps or working models of the environment (Barlow 1989. Reduncany reduction: specific form of Barlow s hypothesis early processing is to turn highly redundant sensory input into more efficient factorial code. Outputs become statistically independent. Atick and Redlich (1990: principle of minumum redundancy. Principle of Minimum Redundancy Sensory signal S, Noisy input X, Recoding system A, noisy output Y. X = S + N 1 Y = AX + N 2 Retinal input includes redundant information. Purpose of retinal coding is to reduce/eliminate the redundant bits of data due to correlations and noise, before sending the signal along the optic nerve. Redundancy measure (with channel capacity C( : I(Y; S R = 1 C(Y 31 32

Principle of Minimum Redundancy (cont d Spatially Coherent Features Objective: find recoder matrix A such that R = 1 I(Y; S C(Y is minimized, subject to the no information loss constaraint: I(Y; X = I(X; X ɛ. When S and Y have the same dimensionality and there is no noise, principle of minimum redundancy is equivalent to the Infomax principle. Thus, Infomax on input/output lead to reduncancy reduction. 33 Infomax for unsupervised processing of the image of natural scenes (Becker and Hinton, 1992. Goal: design a self-organizing system that is capable of learning to encode complex scene information in a simpler form. Objective: extract higher-order features that exhibit simple coherence across space so that representation for one spatial region can be used to produce that of representation of neighboring regions. 34 Spatially Coherent Features (cont d Let S denote a signal component common to both Y a and Y b. We can then express the outputs in terms of S and some noise: Y a = S + N a Y b = S + N b and further assume that N a and N b are independent and zero-mean Gaussian. Also assume S is Gaussian. The mutual information then becomes I(Y a ; Y b = h(y a + h(y b h(y a, Y b. Spatially Coherent Features (cont d With I(Y a ; Y b = h(y a + h(y b h(y a, Y b and [ ( ] 1 + log 2πσ 2 a we get h(y a = 1 2 h(y b = 1 2 [ ( ] 1 + log 2πσ 2 b h(y a, Y b = 1 + log(2π + 1 log det(σ, 2 Σ = [ σ 2 a ρ ab σ a σ b ρ ab σ a σ b σ 2 b ρ ab = E[(Y a E[Y a ](Y b E[Y b ]] σ a σ b ] I(Y a ; Y b = 1 2 log (1 ρ 2 ab (covariance matrix (correlation. 35 36

Spatially Coherent Features (cont d The final results was: I(Y a ; Y b = 1 2 log ( 1 ρ 2 ab. That is, maximizing information is equivalent to maximizing correlation between Y a and Y b, which is intuitively appealing. Relation to canonical correlation in statistics: Given random input vectors X a and X b, find two weight vectors w a and w b so that Spatially Coherent Features When the inputs come from two separate regions, we want to minimize the mutual information between the two outputs (Ukrainec and Haykin, 1992, 1996. Applications include when input sources such as different polarizations of the signal are imaged: mutual information between outputs driven by two orthogonal polarizations should be minimized. Y a = w T a X a and Y b = w T b X b have maximum correlation between them (Anderson 1984. Applications: stereo disparity extraction (Becker and Hinton, 1992. 37 38 Independent Components Analysis (ICA ICA (cont d Unknown random source vector U(n: U = [U 1, U 2,..., U m ] T, where the m components are supplied by a set of independent sources. Note that we need a series of source vectors. U is transformed by an unknown mixing matrix A: X = AU, where X = [X 1, X 2,..., X m ] T. 39 A = [ 2 3 2 1 Left: u 1 on x-axis, u 2 on y-axis (source Right: x 1 on x-axis, x 2 on y-axis (observation Thoughts: how would PCA transform this? Examples from Aapo Hyvarinen s ICA tutorial: http://www.cis.hut.fi/aapo/papers/ijcnn99_tutorialweb/. 40 ].

ICA (cont d ICA (cont d In X = AU, both A and U are unknown. Task: find an estimate of the inverse of the mixing matrix (the demixing matrix W Y = WX. The hope is to recover the unknown source U. (A good example is the cocktail party problem. This is known as the blind source separation problem. Examples from AApo Hyvarinen s ICA tutorial: http://www.cis.hut.fi/aapo/papers/ijcnn99_tutorialweb/. 41 Solution: It is actually feasible, but certain ambiguities cannot be resolved: sign, permutation, scaling (variance. Solution can be obtained by enforcing independence among components of Y while adjusting W, thus the name independent components analysis. 42 ICA: Ambiguities Consider X = AU, and Y = WX. Permutation: X = AP 1 PU, where P is a permutation matrix. Permuting U and A in the same way will give the same X. Sign: the model is unaffected by multiplication of one of the sources by -1. Scaling (variance: estimate scaling up U and scaling down A will give the same X. Source vector u 1 u 2 A ICA: Neural Network View u m x m x m x 1 x 1 y 1 x 2 Observation vector The mixer on the left is an unknown physical process. The demixer on the right could be seen as a neural network. Observation vector x 2 W y y 2 m Output vector 43 44

ICA: Independence Two random variables X and Y are statistically independent when f X,Y (x, y = f X (xf Y (y, where f( is the probability density function. A weaker form of independence is uncorrelatedness (zero covariance, which is E[(X µ X (Y µ Y ] = E[XY ] E[X]E[Y ] = 0, i.e., E[XY ] = E[X]E[Y ]. Gaussians are bad: When the unknown source is Gaussian, any orthogonal transformation A results in the same Gaussian distribution. 45 Statistical Aside: Central Limit Theorem 130 250 x1.dat x2.dat 120 200 110 150 100 100 90 50 80 70 0 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 250 350 x3.dat x10.dat 300 200 250 150 200 150 100 100 50 50 0 0 0 0.5 1 1.5 2 2.5 3 1 2 3 4 5 6 7 8 9 When i.i.d. random variables X 1, X 2,... are added to get another random variable X, X tends to a normal distribution. So, Gaussians are prevalent and hard to avoid in statistics. 46 ICA: Non-Gaussianity Non-Gaussianity can be used as a measure of independence. The intuition is as follows: X = AU, Y = WX Consider one component of Y: Y i = [W i1, W i2,..., W im ]X ICA: Measures of Non-Gaussianity There are several measures of non-gaussianity Kurtosis Negentropy etc. Y i = [W i1, W i2,..., W im ]A U } {{ } call this Z T So, Y i is a linear combination of random variables U k (Y i = m j=1 Z iu i, so it is more Gaussian than any individual U k s. The Gaussianity is minimized when Y i equals one of U k s (one Z p is 1 and all the rest 0. 47 48

ICA: Kurtosis ICA: Negentropy Kurtosis is the fourth-order cumulant. Kurtosis(Y = E[Y 4 ] 3 ( E [ Y 2] 2. Gaussian distributions have kurtosis = 0. More peaked distributions have kurtosis > 0. More flatter distributions have kurtosis < 0. Learning: Start with random W. Adjust W and measure change in kurtosis. We can also use gradient-based methods. Drawback: Kurtosis is sensitive to outliers, and thus not robust. Negentropy J is defined as J(Y = H(Y gauss H(Y where Y gauss is a Gaussian random variable that has the same covariance matrix as Y. Negentropy is always non-negative, and it is zero iff Y is Gaussian. Thus, maximizing negentropy is to maximize non-gaussianity. Problem is that estimating negentropy is difficult, and requires the knowledge of the pdfs. 49 50 Classical method: ICA: Approximation of Negentropy J(Y 1 2 E[Y 3 ] 2 + 1 Kurtosis(Y 2 48 but it is not robust due to the involvement of the kurtoris. Another variant: J(Y p k=1 k i (E[G i (Y ] E[G i (N] 2 where k i s are coefficients, G i ( s are nonquadratic functions, and N is ICA: Minimizing Mutual Information We can also aim to minimize mutual information between Y i s. This turns out to be equivalent to maximizing negentropy (when Y i s have unit variance. I(Y 1 ; Y 2 ;...; Y m = C J(Y i i where C is a constant that does not depend on the weight matrix W. a zero-mean, unit-variance Gaussian r.v. This can be further simplified by J(Y (E[G(Y ] E[G(N] 2 G 1 (Y = 1 a 1 log cosh a 1 Y, G 2 (Y = exp( Y 2 /2. 51 52

ICA: Achieving Independence Given output vector Y, we want Y i and Y j to be statistically independent. This can achieved when I(Y i ; Y j = 0. Another alternative is to make the probability density f Y (y, W parameterized by the matrix W to approach the factorial distribution: f Y (y, W = m f Yi (y i, W, where fyi (y i, W is the marginal probability density of Y i. This can be measured by D (W. f f ICA: KL Divergence with Factorial Dist The KL divergence can be shown to be: m D (W = h(y + h(yi. f f Next, we need to calculate the output entropy: h(y = h(wx = h(x + log det(w. Finally, we need to calculate the marginal entropy h(yi, which gets tricky. This calculation involves a polynomial activation function ϕ(y i. See the textbook for details. 53 54 ICA: Learning W Learning objective is to minimize the KL divergence D f f. We can do gradient descent: ICA Examples Visit the url http://www.cis.hut.fi/projects/ compneuro/whatisica.html for interesting results. w ik = η w ik D f f = η ( (W T ik ϕ(y i x k. The final learning rule, in matrix form, is: W(n+1 = W(n+η(n [ I ϕ(y(ny T (n ] W T (n. 55 56