EFFICIENT GEOMETRIC METHODS FOR KERNEL DENSITY ESTIMATION BASED INDEPENDENT COMPONENT ANALYSIS

Size: px
Start display at page:

Download "EFFICIENT GEOMETRIC METHODS FOR KERNEL DENSITY ESTIMATION BASED INDEPENDENT COMPONENT ANALYSIS"

Transcription

1 15th European Signal Processing Conference (EUSIPCO 2007), Poznan, Poland, Septeber 3-7, 2007, copyright by EURASIP EFFICIENT GEOMETRIC METHODS FOR KERNEL DENSITY ESTIMATION BASED INDEPENDENT COMPONENT ANALYSIS Hao Shen, Martin Kleinsteuber and Knut Hüper National ICT Australia, Canberra Research Laboratory Locked Bag 8001, Canberra ACT 2601, Australia and Departent of Inforation Engineering, RSISE, The Australian National University, Canberra ACT 0200, Australia ABSTRACT The perforance of Independent Coponent Analysis (ICA) ethods significantly depends on the choice of the contrast function and the optiisation algorith used in obtaining the deixing atrix. It has been shown that nonparaetric ICA approaches are ore robust than its paraetric counterparts. One basic nonparaetric ICA contrast was developed by approxiating utual inforation using kernel density estiations. In this work we study the kernel density estiation based linear ICA proble fro an optiisation point of view. Two geoetric ethods are proposed to optiise the kernel density estiation based linear ICA contrast function, a Jacobi-type ethod and an approxiate Newton-like ethod. Rigorous analysis shows that both geoetric ethods converge locally quadratically fast to the correct deixing. The perforance of the proposed algoriths is investigated by nuerical experients. 1. INTRODUCTION In the past decade, Independent Coponent Analysis (ICA) has becoe a standard statistical tool for solving the Blind Source Separation (BSS) proble. It has received considerable attentions in various counities. Generally, the perforance of an ICA ethod depends significantly on the choice of the contrast function easuring statistical independence of signals and on the appropriate optiisation technique as well. Approaches to designing ICA contrasts can be classified into two categories: paraetric and nonparaetric. Paraetric approaches construct the ICA criterion according to certain hypotheses of distributions of the source signals by soe paraeterised failies of density functions. Since the distributions of sources are generally unknown, incorrect hypothesis on these distributions can lead to poor separation perforance, and ay even cause paraetric ICA ethods to fail copletely in any real applications. For these reasons, there has been an increasing interest in developing nonparaetric ICA ethods. One of such nonparaetric ICA approaches is to iniise the epirical utual inforation between recovered signals by eploying the Kernel Density Estiation (KDE) technique to deal with the unknown distributions of the sources, such as 14, 20, 2. There also exist other sophisticated nonparaetric ICA approaches, which do not work with the probability density, refer to 6, 1, 9 and references therein. In this work we only focus on the nonparaetric linear ICA proble based on the kernel density estiation proposed in 2 fro an optiisation point of view. Fro an optiisation point of view, developent of efficient geoetric ICA ethods has been very active ever since the early work of Coon 5. Jacobi-type ethods are one of the proinent ethods, including Joint Approxiate Diagonalisation of Eigenatrices (JADE) 3, Extended Maxiu Likelihood (EML) 21, MaxKurt 3, Jacobi optiised Epirical Characteristic Function ICA (JECFICA) 6, RAD- ICAL 14, and so on. Although the efficiency of these Jacobi-type ICA ethods has been verified by nuerical experients, the convergence properties of ost Jacobi-type ICA algoriths are still theoretically unclear. Recently with a coplete understanding of the FastICA algorith 12, 18, the so-called approxiate Newton-like ethod on anifolds, which is locally quadratically convergent, has been successfully applied to the probles of parallel paraetric linear ICA 17 and one-unit KDE based linear ICA 19. The focus of this work is on the developent of efficient geoetric ethods, naely a Jacobi-type ethod and an approxiate Newton-like ethod, for solving the KDE based linear ICA proble. Rigorous analysis shows that both ethods are locally quadratically convergent to a correct deixing atrix under reasonable assuptions. The paper is organised as follows. Section 2 introduces the KDE based linear ICA proble and gives a critical point analysis of the contrast function followed by a study of the Hessian. In Section 3, we present a Jacobi-type ethod and an approxiate Newton-like ethod for optiising the KDE based linear ICA contrast. The local convergence properties of both ethods are studied as well. Finally in Section 4, several nuerical experients are provided to investigate the perforance of the proposed algoriths. 2. THE KDE BASED LINEAR ICA PROBLEM 2.1 Linear ICA Model and the KDE Based Contrast The standard instantaneous noiseless linear ICA odel is stated as the following relation, see 11 for details, Z = AS, (1) where S = s 1,...,s n R n is a data atrix of n saples of sources ( n), which are utually statistically independent and drawn independently and identically, A R is the ixing atrix of full rank, and Z = z 1,...,z n R n 2007 EURASIP 1711

2 15th European Signal Processing Conference (EUSIPCO 2007), Poznan, Poland, Septeber 3-7, 2007, copyright by EURASIP represents the observed ixtures. Note that, the utual statistical independence can only be ensured if the saple size n tends to infinity. Nevertheless, for our theoretical analysis in Section 2.2, we assue that the coplete independence holds even if the saple size is finite. The task of the linear ICA proble is to recover the source signals S by estiating the ixing atrix A based only on the observed ixtures Z. After the so-called whitening process of the ixtures, the corresponding deixing odel of (1) can be stated as follows Y = X W, (2) where W = w 1,...,w n = V Z R n is the whitened observation (V R is the whitening atrix), X R is an orthogonal atrix being the deixing atrix, and Y = y 1,...,y n R n contains the reconstructed signals. It has been shown in 4, that the whitening process is statistically { n-consistent. Let SO() := X R X X = 1,det(X) = 1 } denote the special orthogonal group of order. In this work, we study the linear ICA proble of finding an X = x 1,...,x SO() which reconstructs the sources via the odel (2) based only on the whitened observations. Miniisation of the utual inforation between the recovered signal is a coon criterion for ICA. In absence of knowing the true distribution of the sources, an epirical utual inforation between the extracted signals can be coputed by eploying the kernel density estiation technique, refer to 2 for ore details, f : SO() R, X k=1 ( ( 1 E i log h E j φ x k w i j h )), (3) where w i j := w i w j R represents the difference between the i-th and j-th saple, φ : R R is an appropriate kernel function, and h R + is the kernel bandwidth. In this work we specify φ as the Gaussian kernel φ(t) = exp( t 2 /2). 2.2 Analysis of the KDE based linear ICA contrast Before analysing the KDE based linear ICA contrast defined in (3), we recall the tangent space of SO() at X as T X SO() := { Ξ R Ξ = XΩ, Ω so() }, (4) where so() := {Ω R Ω = Ω } denote the set of all skew-syetric atrices. Let Ω = ω 1,...,ω = (ω kl ) k,l=1 so() and define y i j (x k ) := (xk w i j)/h. By the chain rule, the first derivative of the contrast f in direction XΩ T X SO() is calculated as where D f (X)(XΩ) = 1 k<l ω kl (u kl (X) u lk (X)), (5) φ u kl (X) := E (y i j (x k ))y i j (x l ) i E j φ(y i j (x k )) R. (6) It can be shown that if X SO() is a correct deixing atrix, by the whitening properties of the sources, the ter u kl (X ) is equal to zero for all k l. Thus it follows that the first derivative of f vanishes at X, i.e., D f (X )(X Ω) = 0. Therefore a correct deixing atrix X is indeed a critical point of f. Now coputing the second derivative of f, one gets the following quadratic for f (Xe tω ) dt 2 = ω k H(k) (X)ω k k=1 1 k<l ω 2 kl (u kk(x) + u ll (X)), where H (k) (X) = (h (k) pq (X)) p,q=1 R with the pq-th entry h (k) φ pq (X) =E (y i j (x k ))y i j (x p )y i j (x q ) i E j φ(y i j (x k )) φ E (y i j (x k ))y i j (x p )E j φ (y i j (x k ))y i j (x q ) i. (E j φ(y i j (x k ))) 2 Let X = X. Using the properties of the whitened signals, tedious but straightforward coputations lead to f (X e tω ) dt 2 =D 2 f (X )(X Ω,X Ω) where = 1 k<l û kk (X) := 2 φ E (y i j (x k )) h 2 i E j φ(y i j (x k )) E i φ (y i j (x k ))y i j (x k ) E j φ(y i j (x k )) ω 2 kl (û kk(x ) + û ll (X )), 1 h 2 E i ( φ (y i j (x k )) E j φ(y i j (x k )). ) 2 (7) (8) (9) (10) Consequently, the Hessian of f at X is indeed diagonal. It is iportant to notice that the properties in (9) hold true only if the coplete statistical independence can be ensured for the sources. It is known already that the global iniu of the actual utual inforation between reconstructed signals gives the correct separation. However there is still no proof showing that the global iniu of the nonparaetric epirical utual inforation defined in (3) yields the best estiation of the sources. Nevertheless, in this work, we assue that the best estiation of the sources is attained at the global iniu of the KDE based contrast defined in (3). In other words, we assue that any correct deixing atrix X is a nondegenerate critical point of the contrast f, i.e., for a given set of sources, the ter û kk (X ) defined in (10) is nonzero. 3. GEOMETRIC METHODS FOR THE KDE BASED LINEAR ICA PROBLEM 3.1 Jacobi-type ICA Method Jacobi-type ethods are well-known tools for solving atrix eigenvalue probles or singular value probles 7. They can also be considered as optiisation procedures. The local convergence properties of Jacobi-type ethods have been recently discussed in 10, 13. In this section, we develop a Jacobi-type ethod to optiise the ICA constrast (3). A Jacobi-type ethod consists of iterating so-called sweeps. Within a sweep, the Jacobi-type ethod uses predeterined directions, which reduce the coputational burden. Once fixing a predeterined direction, a step size is chosen to yield either a local or a global optiu of the restricted 2007 EURASIP 1712

3 15th European Signal Processing Conference (EUSIPCO 2007), Poznan, Poland, Septeber 3-7, 2007, copyright by EURASIP cost function. However, this is often unfeasible for general cost functions. In this work, we follow a different approach that is based on a one diensional Newton optiisation step. Siilar approxiations of the optial step size have already been used in 15, 8. Now let us denote a basis atrix of so() by Ω kl so() with ω kl = ω lk = 1 and zeros anywhere else. We define a set of directions in so() as := {Ω kl so() 1 k < l }. Recall the forula for a geodesic in SO() eanating fro a point X in direction XΩ T X SO() γ X : R SO(), t X exp(tω). (11) That is γ X (0) = X and γ X (0) = XΩ. By fixing a direction Ω so(), we constrain the ICA contrast function f defined in (3) to a geodesic, such that in each step, a one diensional optiisation proble f γ X : R R, t f (X exp(tω)) (12) has to be solved. Having chosen a predeterined direction Ω kl, we copute the first and second derivatives of f γ X as follows d dt f (X exp(tω kl)) = u kl (X) u lk (X), dt 2 f (X exp(tω kl )) = ĥkl(x) + ĥlk(x), (13) where ĥkl(x) := h (k) ll (X) u kk (X). Hence, a Newton step λ kl R for a fixed direction Ω kl can be coputed by λ kl := λ kl (X) := u kl(x) u lk (X) ĥ kl (X)+ĥlk(X). (14) Now by the special structure of Ω kl, the atrix Φ = (φ pq ) p,q=1 := exp(λ klω kl ) is coputed as being a Jacobi rotation, i.e., it is equal to the identity atrix except for the kk, ll, kl and lk entries with φ kk = φ ll = cos(λ kl ), φ kl = φ lk = sin(λ kl ). (15) To ensure the one diensional Newton direction to point downhill, one can force the denoinator defined in (14) to be positive by taking the absolute value of the Hessian of f γ X. A Jacobi-type ethod for optiising the KDE based linear ICA contrast function can be suarised as follows: Jacobi-type KDE based linear ICA ethod Step 1: Given an initial guess X SO() and a set of predeterined directions so(); Step 2: Let X old = X. For 1 k < l ; (i) Calculate a Newton stepsize λ kl R λ kl = u kl(x) u lk (X) ĥkl(x)+ĥlk(x) ; (ii) Update X X exp(λ kl Ω kl ); Step 3: If X old X is sall enough, Stop. Otherwise, goto Step 2. To show the local convergence properties of the Jacobitype KDE based linear ICA ethod, we first state the following theore regarding the local quadratic convergence of a Jacobi-type ethod, which eploys a Newton step rotation, for optiising a general cost function on SO(). This result can be proven by siilar arguents explored already in 10, 13. Due to the page liits, the proof will be oitted here. Theore 1 Let f : SO() R be a sooth cost function, X SO() be a local iniu of f and {Ω 1,...,Ω N } be a basis of so(). If the Hessian H f (X ) is nondegenerated and if X Ω i,x Ω j T X SO() are orthogonal with respect to H f (X ) for all 1 i < j N, then the Jacobi-type ethod using a Newton step converges locally quadratically fast. The local convergence properties of the Jacobi-type KDE based linear ICA ethod can be stated as follows Corollary 1 Let X SO() be a correct deixing atrix of a linear ICA proble. Then the Jacobi-type KDE based linear ICA ethod is locally quadratically convergent to X. Sketch of proof. Let be a standard basis of so(). It can be shown that D 2 f (X )(X Ω kl,x Ω pq ) = 0 for any Ω kl Ω pq. Then by the assuption that X is a nondegenerate critical point, the result follows. 3.2 Approxiate Newton-like ICA Method For a Newton type ethod, coputing the Hessian is generally too expensive to evaluate at each iteration. Therefore we propose a cheap approxiation of the true Hessian according to the explicit structure of the Hessian at a correct deixing atrix. As suggested in (9), the Hessian of f at a correct deixing atrix X is indeed diagonal. We therefore ight approxiate the Hessian of f considered now as the linear operator T X SO() T X SO() by approxiating dt 2 f (Xe tω ) using the expression ωkl 2 (û kk(x) + û ll (X)). (16) 1 k<l It is easily seen that the approxiation (16) gives the true Hessian of f as in (9) if X = X. Hence, for an approxiate Newton ethod, the approxiate Newton direction Ω = (ω kl ) k,l=1 so() can be cheaply coputed by ω kl = u kl(x) u lk (X) û kk (X)+û ll (X). (17) Now, an approxiate Newton iteration can be closed by projecting the Newton direction Ω back onto SO() using the atrix exponential. However, in contrast to the Jacobi-type update, the atrix exponential of an arbitrary Ω so() requires an eigenvalue decoposition being an unnecessarily expensive iterative process 16. However, one can overcoe this issue by using a first order approxiation of the atrix exponential via a QR decoposition, which is still orthogonal as follows. Let I + Ω = Q(Ω)R(Ω), (18) be the unique QR decoposition of the atrix I + Ω with Ω so(). Note that, the deterinant of I + Ω is always positive. Therefore the QR decoposition is indeed unique with diagonal entries of R(Ω) being positive and detq(ω) = 1. To suarise, we construct the Approxiate Newtonlike KDE based linear ICA ethod as follows: 2007 EURASIP 1713

4 15th European Signal Processing Conference (EUSIPCO 2007), Poznan, Poland, Septeber 3-7, 2007, copyright by EURASIP Approxiate Newton-like KDE based linear ICA ethod Step 1: Given an initial guess X SO(); Step 2: Let X old = X. Copute the direction Ω = (ω kl ) k,l=1 so() by ω kl = u kl(x) u lk (X) û kk (X)+û ll (X) ; Step 3: Update X XQ(Ω); Step 4: If X old X is sall enough, Stop. Otherwise, goto Step 2. Finally, the local convergence properties of the approxiate Newton-like KDE based linear ICA ethod are suarised in the following result. Theore 2 Let X SO() be a correct deixing atrix of a linear ICA proble. Then the approxiate Newton-like KDE based linear ICA ethod is locally quadratically convergent to X. Sketch of proof. Since the approxiate Newton direction defined in (17) is locally a sooth ap on SO() around X. Each iteration of the approxiate Newton-like KDE based linear ICA ethod is then locally a sooth ap on SO() around X as well. It can be shown that the first derivative of the algorith considered now as locally a sooth ap around X vanishes at X. The result follows by a Taylor series arguent. 4. NUMERICAL EXPERIMENTS In this section, we copare our new ethods with an existing parallelised approxiate Newton linear ICA algorith, PANNICA proposed in 19, which can be considered as an nonparaetric version of FastICA. Firstly, the local convergence properties of all algoriths are investigated by applying the to an ideal dataset and a siple speech dataset. The perforance are further copared in ters of both separation quality and convergence speed. Without fine tuning the kernel bandwidth h, we set h = Convergence Properties The convergence is easured by the distance of the accuulation point X to the current iterate X k, i.e., by the Frobenius nor X k X F. All three ethods are initialised by the sae deixing atrix. It is worthwhile to notice that, by the nature of the linear ICA proble, all algoriths can converge to different correct deixing atrices, where one is a colun-wise perutation of the other. Here we only study the cases where they converge to the sae deixing atrix. First, we construct an ideal dataset where the properties of the Hessian given in (9) hold true. It consists of two siple signals, one being a square wave and the other a sine wave. The nuerical results shown in Figure 1 verify the local quadratic convergence properties of both ethods presented in Corollary 1 and Theore 2. It also indicates that PAN- NICA is locally quadratically convergent as well. Based on our ipleentations, all algoriths have coparable coputational burden. Now, let us study the local convergence properties of these ethods on a real speech dataset, which consists of three signals with n = 5,000 saples. The results in Figure 2 suggest that all three algoriths appear to converge only linearly fast. This is ost likely caused by the fact that, for X(k) X * F X(k) X * F Jacobi ICA ANL ICA PANNICA Iteration/Sweep (k) Figure 1: Convergence on an ideal dataset. Jacobi ICA ANL ICA PANNICA Iteration/Sweep (k) Figure 2: Convergence on a siple speech dataset. this speech dataset, the ideal properties of the Hessian shown in (9) are not fulfilled. Nevertheless fro this experient, one can still conclude that the Jacobi-type ethod converges uch faster than the other two approxiate Newton based algoriths within a real environent. 4.2 Coparison of perforance In this experient, all ethods are tested on the benchark speech signal dataset fro the Brain Science Institute, RIKEN, see The dataset consists of 200 natural speech signals sapled at 4 khz. The task of this experient is to separate = 3 signals which are randoly chosen out of 200 sources, with a fixed saple size n = 1,000. The separation perforance is easured by the average Signal-to-Interference-Ratio (SIR) index, refer to 2 for the definition. Usually, a large SIR index indicates a better separation. The convergence perforance is copared by the nuber of iterations/sweeps required by each algorith to reach the sae level of error, e.g., X k X < Each ethod is initialised by the sae randoly generated deixing atrix. By replicating the experient 100 ties, Figure 3 and Figure 4 present the quartile based boxplots of the SIR index and the nuber of iterations/sweeps for each ethod to converge respectively. In Figure 3, it is shown that both the Jacobi-type ethod and the approxiate Newton-like ethod reach alost identical accuracy, while the parallelised approxiate Newton ethod perfors slightly worse than the others. By coparing the convergence perforance 2007 EURASIP 1714

5 15th European Signal Processing Conference (EUSIPCO 2007), Poznan, Poland, Septeber 3-7, 2007, copyright by EURASIP SIR (db) Nuber of Iterations/Sweeps Jacobi ICA ANL ICA PANNICA Figure 3: Coparison of separation quality Jacobi ICA ANL ICA PANNICA Figure 4: Coparison of convergence speed. in Figure 4, the Jacobi-type ethod outperfors the other two ethods. In average, it converges about 3 ties faster than the others. 5. ACKNOWLEDGMENT National ICT Australia is funded by the Australian Governent s Departent of Counications, Inforation Technology and the Arts and the Australian Research Council through Backing Australia s Ability and the ICT Research Centre of Excellence progras. REFERENCES 1 F. R. Bach and M. I. Jordan. Kernel independent coponent analysis. Journal of Machine Learning Research, 3:1 48, R. Boscolo, H. Pan, and V. P. Roychowdhury. Independent coponent analysis based on nonparaetric density estiation. IEEE Transactions on Neural Networks, 15(1):55 65, J.-F. Cardoso. High-order contrasts for independent coponent analysis. Neural Coputation, 11(1): , A. Chen and P. J. Bickel. Consistent independent coponent analysis and prewhitening. IEEE Transactions on Signal Processing, 53(10): , P. Coon. Independent coponent analysis, a new concept? Signal Processing, 36(3): , J. Eriksson and K. Koivunen. Characteristic-function based independent coponent analysis. Signal Processing, 83(10): , G. Golub and C. F. van Loan. Matrix Coputations. The Johns Hopkins University Press, Baltiore, 2nd edition, J. Götze, S. Paul, and M. Sauer. An efficient Jacobi-like algorith for parallel eigenvalue coputation. IEEE Transactions on Coputers, 42(9): , A. Gretton, R. Herbrich, A. J. Sola, O. Bousquet, and B. Schölkopf. Kernel ethods for easuring independence. Journal of Machine Learning Research, 6: , K. Hüper. A Calculus Approach to Matrix Eigenvalue Algoriths. Habilitation Dissertation, Departent of Matheatics, University of Würzburg, Gerany, July A. Hyvärinen, J. Karhunen, and E. Oja. Independent Coponent Analysis. Wiley, New York, A. Hyvärinen and E. Oja. A fast fixed-point algorith for independent coponent analysis. Neural Coputation, 9(7): , M. Kleinsteuber. Jacobi-Type Methods on Seisiple Lie Algebras A Lie Algebraic Approach to the Syetric Eigenvalue Proble. PhD thesis, Bayerische Julius-Maxiilians-Universität Würzburg, E. G. Learned-Miller and J. W. Fisher III. ICA using spacings estiates of entropy. Journal of Machine Learning Research, 4: , J. J. Modi and J. D. Pryce. Efficient ipleentation of Jacobi s diagonalization ethod on the DAP. Nuerische Matheatik, 46(3): , C. Moler and C. van Loan. Nineteen dubious ways to copute the exponential of a atrix, twenty-five years later. SIAM Review, 45(1):3 49, H. Shen and K. Hüper. Newton-like ethods for parallel independent coponent analysis. In Proceedings of the 16 th IEEE International Workshop on Machine Learning for Signal Processing (MLSP 2006), pages , Maynooth, Ireland, H. Shen, K. Hüper, and A.-K. Seghouane. Geoetric optiisation and FastICA algoriths. In Proceedings of the 17 th International Syposiu of Matheatical Theory of Networks and Systes (MTNS 2006), pages , Kyoto, Japan, H. Shen, K. Hüper, and A. J. Sola. Newton-like ethods for nonparaetric independent coponent analysis. In Proceedings of the 13 th International Conference on Neural Inforation Processing, Part I, (ICONIP 2006), Hong Kong, China, October 3-6, 2006, volue 4232 of Lecture Notes in Coputer Science, pages , Berlin/Heidelberg, Springer-Verlag. 20 H. Stögbauer, A. Kraskov, S. Astakhov, and P. Grassberger. Least dependent coponent analysis based on utual inforation. Phys. Rev. E, 70(6):066123, V. Zarzoso and A. K. Nandi. Blind separation of independent sources for virtually any source probability density function. IEEE Transactions on Signal Processing, 47(9): , EURASIP 1715

Geometric Analysis of Hilbert-Schmidt Independence Criterion Based ICA Contrast Function

Geometric Analysis of Hilbert-Schmidt Independence Criterion Based ICA Contrast Function Geometric Analysis of Hilbert-Schmidt Independence Criterion Based ICA Contrast Function NICTA Technical Report: PA006080 ISSN 1833-9646 Hao Shen 1, Stefanie Jegelka 2 and Arthur Gretton 2 1 Systems Engineering

More information

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS ISSN 1440-771X AUSTRALIA DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS An Iproved Method for Bandwidth Selection When Estiating ROC Curves Peter G Hall and Rob J Hyndan Working Paper 11/00 An iproved

More information

DISSIMILARITY MEASURES FOR ICA-BASED SOURCE NUMBER ESTIMATION. Seungchul Lee 2 2. University of Michigan. Ann Arbor, MI, USA.

DISSIMILARITY MEASURES FOR ICA-BASED SOURCE NUMBER ESTIMATION. Seungchul Lee 2 2. University of Michigan. Ann Arbor, MI, USA. Proceedings of the ASME International Manufacturing Science and Engineering Conference MSEC June -8,, Notre Dae, Indiana, USA MSEC-7 DISSIMILARIY MEASURES FOR ICA-BASED SOURCE NUMBER ESIMAION Wei Cheng,

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Lecture 21. Interior Point Methods Setup and Algorithm

Lecture 21. Interior Point Methods Setup and Algorithm Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

A Note on the Applied Use of MDL Approximations

A Note on the Applied Use of MDL Approximations A Note on the Applied Use of MDL Approxiations Daniel J. Navarro Departent of Psychology Ohio State University Abstract An applied proble is discussed in which two nested psychological odels of retention

More information

PAC-Bayes Analysis Of Maximum Entropy Learning

PAC-Bayes Analysis Of Maximum Entropy Learning PAC-Bayes Analysis Of Maxiu Entropy Learning John Shawe-Taylor and David R. Hardoon Centre for Coputational Statistics and Machine Learning Departent of Coputer Science University College London, UK, WC1E

More information

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

Ensemble Based on Data Envelopment Analysis

Ensemble Based on Data Envelopment Analysis Enseble Based on Data Envelopent Analysis So Young Sohn & Hong Choi Departent of Coputer Science & Industrial Systes Engineering, Yonsei University, Seoul, Korea Tel) 82-2-223-404, Fax) 82-2- 364-7807

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

Supplementary Material for Fast and Provable Algorithms for Spectrally Sparse Signal Reconstruction via Low-Rank Hankel Matrix Completion

Supplementary Material for Fast and Provable Algorithms for Spectrally Sparse Signal Reconstruction via Low-Rank Hankel Matrix Completion Suppleentary Material for Fast and Provable Algoriths for Spectrally Sparse Signal Reconstruction via Low-Ran Hanel Matrix Copletion Jian-Feng Cai Tianing Wang Ke Wei March 1, 017 Abstract We establish

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search

Quantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search Quantu algoriths (CO 781, Winter 2008) Prof Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search ow we begin to discuss applications of quantu walks to search algoriths

More information

The linear sampling method and the MUSIC algorithm

The linear sampling method and the MUSIC algorithm INSTITUTE OF PHYSICS PUBLISHING INVERSE PROBLEMS Inverse Probles 17 (2001) 591 595 www.iop.org/journals/ip PII: S0266-5611(01)16989-3 The linear sapling ethod and the MUSIC algorith Margaret Cheney Departent

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

arxiv: v1 [stat.ml] 31 Jan 2018

arxiv: v1 [stat.ml] 31 Jan 2018 Increental kernel PCA and the Nyströ ethod arxiv:802.00043v [stat.ml] 3 Jan 208 Fredrik Hallgren Departent of Statistical Science University College London London WCE 6BT, United Kingdo fredrik.hallgren@ucl.ac.uk

More information

Introduction to Machine Learning. Recitation 11

Introduction to Machine Learning. Recitation 11 Introduction to Machine Learning Lecturer: Regev Schweiger Recitation Fall Seester Scribe: Regev Schweiger. Kernel Ridge Regression We now take on the task of kernel-izing ridge regression. Let x,...,

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

Lower Bounds for Quantized Matrix Completion

Lower Bounds for Quantized Matrix Completion Lower Bounds for Quantized Matrix Copletion Mary Wootters and Yaniv Plan Departent of Matheatics University of Michigan Ann Arbor, MI Eail: wootters, yplan}@uich.edu Mark A. Davenport School of Elec. &

More information

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

arxiv: v1 [math.na] 10 Oct 2016

arxiv: v1 [math.na] 10 Oct 2016 GREEDY GAUSS-NEWTON ALGORITHM FOR FINDING SPARSE SOLUTIONS TO NONLINEAR UNDERDETERMINED SYSTEMS OF EQUATIONS MÅRTEN GULLIKSSON AND ANNA OLEYNIK arxiv:6.395v [ath.na] Oct 26 Abstract. We consider the proble

More information

Bayes Decision Rule and Naïve Bayes Classifier

Bayes Decision Rule and Naïve Bayes Classifier Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili,

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili, Australian Journal of Basic and Applied Sciences, 5(3): 35-358, 20 ISSN 99-878 Generalized AOR Method for Solving Syste of Linear Equations Davod Khojasteh Salkuyeh Departent of Matheatics, University

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science

Qualitative Modelling of Time Series Using Self-Organizing Maps: Application to Animal Science Proceedings of the 6th WSEAS International Conference on Applied Coputer Science, Tenerife, Canary Islands, Spain, Deceber 16-18, 2006 183 Qualitative Modelling of Tie Series Using Self-Organizing Maps:

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a

More information

IN modern society that various systems have become more

IN modern society that various systems have become more Developent of Reliability Function in -Coponent Standby Redundant Syste with Priority Based on Maxiu Entropy Principle Ryosuke Hirata, Ikuo Arizono, Ryosuke Toohiro, Satoshi Oigawa, and Yasuhiko Takeoto

More information

Boosting with log-loss

Boosting with log-loss Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

Testing equality of variances for multiple univariate normal populations

Testing equality of variances for multiple univariate normal populations University of Wollongong Research Online Centre for Statistical & Survey Methodology Working Paper Series Faculty of Engineering and Inforation Sciences 0 esting equality of variances for ultiple univariate

More information

Lecture 13 Eigenvalue Problems

Lecture 13 Eigenvalue Problems Lecture 13 Eigenvalue Probles MIT 18.335J / 6.337J Introduction to Nuerical Methods Per-Olof Persson October 24, 2006 1 The Eigenvalue Decoposition Eigenvalue proble for atrix A: Ax = λx with eigenvalues

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

3.3 Variational Characterization of Singular Values

3.3 Variational Characterization of Singular Values 3.3. Variational Characterization of Singular Values 61 3.3 Variational Characterization of Singular Values Since the singular values are square roots of the eigenvalues of the Heritian atrices A A and

More information

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics

ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS A Thesis Presented to The Faculty of the Departent of Matheatics San Jose State University In Partial Fulfillent of the Requireents

More information

Handwriting Detection Model Based on Four-Dimensional Vector Space Model

Handwriting Detection Model Based on Four-Dimensional Vector Space Model Journal of Matheatics Research; Vol. 10, No. 4; August 2018 ISSN 1916-9795 E-ISSN 1916-9809 Published by Canadian Center of Science and Education Handwriting Detection Model Based on Four-Diensional Vector

More information

Generalized Sampling Theorem for Bandpass Signals

Generalized Sampling Theorem for Bandpass Signals Hindawi Publishing Corporation EURASIP Journal on Applied Signal Processing Volue 26, Article ID 59587, Pages 6 DOI 55/ASP/26/59587 Generalized Sapling Theore for Bandpass Signals Ales Prokes Departent

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

ON THE TWO-LEVEL PRECONDITIONING IN LEAST SQUARES METHOD

ON THE TWO-LEVEL PRECONDITIONING IN LEAST SQUARES METHOD PROCEEDINGS OF THE YEREVAN STATE UNIVERSITY Physical and Matheatical Sciences 04,, p. 7 5 ON THE TWO-LEVEL PRECONDITIONING IN LEAST SQUARES METHOD M a t h e a t i c s Yu. A. HAKOPIAN, R. Z. HOVHANNISYAN

More information

A note on the realignment criterion

A note on the realignment criterion A note on the realignent criterion Chi-Kwong Li 1, Yiu-Tung Poon and Nung-Sing Sze 3 1 Departent of Matheatics, College of Willia & Mary, Williasburg, VA 3185, USA Departent of Matheatics, Iowa State University,

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3

More information

Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links

Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Tie-Varying Jaing Links Jun Kurihara KDDI R&D Laboratories, Inc 2 5 Ohara, Fujiino, Saitaa, 356 8502 Japan Eail: kurihara@kddilabsjp

More information

RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS

RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS BIT Nuerical Matheatics 43: 459 466, 2003. 2003 Kluwer Acadeic Publishers. Printed in The Netherlands 459 RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS V. SIMONCINI Dipartiento di

More information

lecture 36: Linear Multistep Mehods: Zero Stability

lecture 36: Linear Multistep Mehods: Zero Stability 95 lecture 36: Linear Multistep Mehods: Zero Stability 5.6 Linear ultistep ethods: zero stability Does consistency iply convergence for linear ultistep ethods? This is always the case for one-step ethods,

More information

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical IEEE TRANSACTIONS ON INFORMATION THEORY Large Alphabet Source Coding using Independent Coponent Analysis Aichai Painsky, Meber, IEEE, Saharon Rosset and Meir Feder, Fellow, IEEE arxiv:67.7v [cs.it] Jul

More information

Scale Invariant Conditional Dependence Measures

Scale Invariant Conditional Dependence Measures Sashank J. Reddi sjakkar@cs.cu.edu Machine Learning Departent, School of Coputer Science, Carnegie Mellon University Barnabás Póczos bapoczos@cs.cu.edu Machine Learning Departent, School of Coputer Science,

More information

The Methods of Solution for Constrained Nonlinear Programming

The Methods of Solution for Constrained Nonlinear Programming Research Inventy: International Journal Of Engineering And Science Vol.4, Issue 3(March 2014), PP 01-06 Issn (e): 2278-4721, Issn (p):2319-6483, www.researchinventy.co The Methods of Solution for Constrained

More information

Feedforward Networks

Feedforward Networks Feedforward Neural Networks - Backpropagation Feedforward Networks Gradient Descent Learning and Backpropagation CPSC 533 Fall 2003 Christian Jacob Dept.of Coputer Science,University of Calgary Feedforward

More information

Chaotic Coupled Map Lattices

Chaotic Coupled Map Lattices Chaotic Coupled Map Lattices Author: Dustin Keys Advisors: Dr. Robert Indik, Dr. Kevin Lin 1 Introduction When a syste of chaotic aps is coupled in a way that allows the to share inforation about each

More information

Antenna Saturation Effects on MIMO Capacity

Antenna Saturation Effects on MIMO Capacity Antenna Saturation Effects on MIMO Capacity T S Pollock, T D Abhayapala, and R A Kennedy National ICT Australia Research School of Inforation Sciences and Engineering The Australian National University,

More information

A remark on a success rate model for DPA and CPA

A remark on a success rate model for DPA and CPA A reark on a success rate odel for DPA and CPA A. Wieers, BSI Version 0.5 andreas.wieers@bsi.bund.de Septeber 5, 2018 Abstract The success rate is the ost coon evaluation etric for easuring the perforance

More information

An improved self-adaptive harmony search algorithm for joint replenishment problems

An improved self-adaptive harmony search algorithm for joint replenishment problems An iproved self-adaptive harony search algorith for joint replenishent probles Lin Wang School of Manageent, Huazhong University of Science & Technology zhoulearner@gail.co Xiaojian Zhou School of Manageent,

More information

RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS MEMBRANE

RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS MEMBRANE Proceedings of ICIPE rd International Conference on Inverse Probles in Engineering: Theory and Practice June -8, 999, Port Ludlow, Washington, USA : RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS

More information

Stochastic Subgradient Methods

Stochastic Subgradient Methods Stochastic Subgradient Methods Lingjie Weng Yutian Chen Bren School of Inforation and Coputer Science University of California, Irvine {wengl, yutianc}@ics.uci.edu Abstract Stochastic subgradient ethods

More information

An l 1 Regularized Method for Numerical Differentiation Using Empirical Eigenfunctions

An l 1 Regularized Method for Numerical Differentiation Using Empirical Eigenfunctions Journal of Matheatical Research with Applications Jul., 207, Vol. 37, No. 4, pp. 496 504 DOI:0.3770/j.issn:2095-265.207.04.0 Http://jre.dlut.edu.cn An l Regularized Method for Nuerical Differentiation

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

Fairness via priority scheduling

Fairness via priority scheduling Fairness via priority scheduling Veeraruna Kavitha, N Heachandra and Debayan Das IEOR, IIT Bobay, Mubai, 400076, India vavitha,nh,debayan}@iitbacin Abstract In the context of ulti-agent resource allocation

More information

The Transactional Nature of Quantum Information

The Transactional Nature of Quantum Information The Transactional Nature of Quantu Inforation Subhash Kak Departent of Coputer Science Oklahoa State University Stillwater, OK 7478 ABSTRACT Inforation, in its counications sense, is a transactional property.

More information

Feedforward Networks

Feedforward Networks Feedforward Networks Gradient Descent Learning and Backpropagation Christian Jacob CPSC 433 Christian Jacob Dept.of Coputer Science,University of Calgary CPSC 433 - Feedforward Networks 2 Adaptive "Prograing"

More information

Kernel-Based Nonparametric Anomaly Detection

Kernel-Based Nonparametric Anomaly Detection Kernel-Based Nonparaetric Anoaly Detection Shaofeng Zou Dept of EECS Syracuse University Eail: szou@syr.edu Yingbin Liang Dept of EECS Syracuse University Eail: yliang6@syr.edu H. Vincent Poor Dept of

More information

Probability Distributions

Probability Distributions Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples

More information

Estimation of the Mean of the Exponential Distribution Using Maximum Ranked Set Sampling with Unequal Samples

Estimation of the Mean of the Exponential Distribution Using Maximum Ranked Set Sampling with Unequal Samples Open Journal of Statistics, 4, 4, 64-649 Published Online Septeber 4 in SciRes http//wwwscirporg/ournal/os http//ddoiorg/436/os4486 Estiation of the Mean of the Eponential Distribution Using Maiu Ranked

More information

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40 On Poset Merging Peter Chen Guoli Ding Steve Seiden Abstract We consider the follow poset erging proble: Let X and Y be two subsets of a partially ordered set S. Given coplete inforation about the ordering

More information

Asynchronous Gossip Algorithms for Stochastic Optimization

Asynchronous Gossip Algorithms for Stochastic Optimization Asynchronous Gossip Algoriths for Stochastic Optiization S. Sundhar Ra ECE Dept. University of Illinois Urbana, IL 680 ssrini@illinois.edu A. Nedić IESE Dept. University of Illinois Urbana, IL 680 angelia@illinois.edu

More information

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy

Nonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy Storage Capacity and Dynaics of Nononotonic Networks Bruno Crespi a and Ignazio Lazzizzera b a. IRST, I-38050 Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I-38050 Povo (Trento) Italy INFN Gruppo

More information

Page 1 Lab 1 Elementary Matrix and Linear Algebra Spring 2011

Page 1 Lab 1 Elementary Matrix and Linear Algebra Spring 2011 Page Lab Eleentary Matri and Linear Algebra Spring 0 Nae Due /03/0 Score /5 Probles through 4 are each worth 4 points.. Go to the Linear Algebra oolkit site ransforing a atri to reduced row echelon for

More information

Feedforward Networks. Gradient Descent Learning and Backpropagation. Christian Jacob. CPSC 533 Winter 2004

Feedforward Networks. Gradient Descent Learning and Backpropagation. Christian Jacob. CPSC 533 Winter 2004 Feedforward Networks Gradient Descent Learning and Backpropagation Christian Jacob CPSC 533 Winter 2004 Christian Jacob Dept.of Coputer Science,University of Calgary 2 05-2-Backprop-print.nb Adaptive "Prograing"

More information

Solving initial value problems by residual power series method

Solving initial value problems by residual power series method Theoretical Matheatics & Applications, vol.3, no.1, 13, 199-1 ISSN: 179-9687 (print), 179-979 (online) Scienpress Ltd, 13 Solving initial value probles by residual power series ethod Mohaed H. Al-Sadi

More information

Explicit solution of the polynomial least-squares approximation problem on Chebyshev extrema nodes

Explicit solution of the polynomial least-squares approximation problem on Chebyshev extrema nodes Explicit solution of the polynoial least-squares approxiation proble on Chebyshev extrea nodes Alfredo Eisinberg, Giuseppe Fedele Dipartiento di Elettronica Inforatica e Sisteistica, Università degli Studi

More information

Ph 20.3 Numerical Solution of Ordinary Differential Equations

Ph 20.3 Numerical Solution of Ordinary Differential Equations Ph 20.3 Nuerical Solution of Ordinary Differential Equations Due: Week 5 -v20170314- This Assignent So far, your assignents have tried to failiarize you with the hardware and software in the Physics Coputing

More information

Combining Classifiers

Combining Classifiers Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/

More information

1 Proof of learning bounds

1 Proof of learning bounds COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #4 Scribe: Akshay Mittal February 13, 2013 1 Proof of learning bounds For intuition of the following theore, suppose there exists a

More information

Weighted- 1 minimization with multiple weighting sets

Weighted- 1 minimization with multiple weighting sets Weighted- 1 iniization with ultiple weighting sets Hassan Mansour a,b and Özgür Yılaza a Matheatics Departent, University of British Colubia, Vancouver - BC, Canada; b Coputer Science Departent, University

More information

Keywords: Estimator, Bias, Mean-squared error, normality, generalized Pareto distribution

Keywords: Estimator, Bias, Mean-squared error, normality, generalized Pareto distribution Testing approxiate norality of an estiator using the estiated MSE and bias with an application to the shape paraeter of the generalized Pareto distribution J. Martin van Zyl Abstract In this work the norality

More information

On Constant Power Water-filling

On Constant Power Water-filling On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives

More information

MAXIMUM LIKELIHOOD BASED TECHNIQUES FOR BLIND SOURCE SEPARATION AND APPROXIMATE JOINT DIAGONALIZATION

MAXIMUM LIKELIHOOD BASED TECHNIQUES FOR BLIND SOURCE SEPARATION AND APPROXIMATE JOINT DIAGONALIZATION BEN-GURION UNIVERSITY OF TE NEGEV FACULTY OF ENGINEERING SCIENCE DEPARTENT OF ELECTRICAL AND COPUTER ENGINEERING AXIU LIKELIOOD BASED TECNIQUES FOR BLIND SOURCE SEPARATION AND APPROXIATE JOINT DIAGONALIZATION

More information

An Improved Particle Filter with Applications in Ballistic Target Tracking

An Improved Particle Filter with Applications in Ballistic Target Tracking Sensors & ransducers Vol. 72 Issue 6 June 204 pp. 96-20 Sensors & ransducers 204 by IFSA Publishing S. L. http://www.sensorsportal.co An Iproved Particle Filter with Applications in Ballistic arget racing

More information

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Proc. of the IEEE/OES Seventh Working Conference on Current Measureent Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Belinda Lipa Codar Ocean Sensors 15 La Sandra Way, Portola Valley, CA 98 blipa@pogo.co

More information

Ufuk Demirci* and Feza Kerestecioglu**

Ufuk Demirci* and Feza Kerestecioglu** 1 INDIRECT ADAPTIVE CONTROL OF MISSILES Ufuk Deirci* and Feza Kerestecioglu** *Turkish Navy Guided Missile Test Station, Beykoz, Istanbul, TURKEY **Departent of Electrical and Electronics Engineering,

More information

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm Acta Polytechnica Hungarica Vol., No., 04 Sybolic Analysis as Universal Tool for Deriving Properties of Non-linear Algoriths Case study of EM Algorith Vladiir Mladenović, Miroslav Lutovac, Dana Porrat

More information

A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION

A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION A eshsize boosting algorith in kernel density estiation A MESHSIZE BOOSTING ALGORITHM IN KERNEL DENSITY ESTIMATION C.C. Ishiekwene, S.M. Ogbonwan and J.E. Osewenkhae Departent of Matheatics, University

More information

2 Q 10. Likewise, in case of multiple particles, the corresponding density in 2 must be averaged over all

2 Q 10. Likewise, in case of multiple particles, the corresponding density in 2 must be averaged over all Lecture 6 Introduction to kinetic theory of plasa waves Introduction to kinetic theory So far we have been odeling plasa dynaics using fluid equations. The assuption has been that the pressure can be either

More information

On Conditions for Linearity of Optimal Estimation

On Conditions for Linearity of Optimal Estimation On Conditions for Linearity of Optial Estiation Erah Akyol, Kuar Viswanatha and Kenneth Rose {eakyol, kuar, rose}@ece.ucsb.edu Departent of Electrical and Coputer Engineering University of California at

More information

C na (1) a=l. c = CO + Clm + CZ TWO-STAGE SAMPLE DESIGN WITH SMALL CLUSTERS. 1. Introduction

C na (1) a=l. c = CO + Clm + CZ TWO-STAGE SAMPLE DESIGN WITH SMALL CLUSTERS. 1. Introduction TWO-STGE SMPLE DESIGN WITH SMLL CLUSTERS Robert G. Clark and David G. Steel School of Matheatics and pplied Statistics, University of Wollongong, NSW 5 ustralia. (robert.clark@abs.gov.au) Key Words: saple

More information

Geometrical intuition behind the dual problem

Geometrical intuition behind the dual problem Based on: Geoetrical intuition behind the dual proble KP Bennett, EJ Bredensteiner, Duality and Geoetry in SVM Classifiers, Proceedings of the International Conference on Machine Learning, 2000 1 Geoetrical

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter

More information

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition

Pattern Recognition and Machine Learning. Learning and Evaluation for Pattern Recognition Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lesson 1 4 October 2017 Outline Learning and Evaluation for Pattern Recognition Notation...2 1. The Pattern Recognition

More information

Fixed-to-Variable Length Distribution Matching

Fixed-to-Variable Length Distribution Matching Fixed-to-Variable Length Distribution Matching Rana Ali Ajad and Georg Böcherer Institute for Counications Engineering Technische Universität München, Gerany Eail: raa2463@gail.co,georg.boecherer@tu.de

More information

In this chapter, we consider several graph-theoretic and probabilistic models

In this chapter, we consider several graph-theoretic and probabilistic models THREE ONE GRAPH-THEORETIC AND STATISTICAL MODELS 3.1 INTRODUCTION In this chapter, we consider several graph-theoretic and probabilistic odels for a social network, which we do under different assuptions

More information

1 Rademacher Complexity Bounds

1 Rademacher Complexity Bounds COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #10 Scribe: Max Goer March 07, 2013 1 Radeacher Coplexity Bounds Recall the following theore fro last lecture: Theore 1. With probability

More information

1 Generalization bounds based on Rademacher complexity

1 Generalization bounds based on Rademacher complexity COS 5: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #0 Scribe: Suqi Liu March 07, 08 Last tie we started proving this very general result about how quickly the epirical average converges

More information