A New Class of APEX-Like PCA Algorithms
|
|
- Alvin Blake
- 5 years ago
- Views:
Transcription
1 Reprinted fro Proceedings of ISCAS-98, IEEE Int. Syposiu on Circuit and Systes, Monterey (USA), June 1998 A New Class of APEX-Like PCA Algoriths Sione Fiori, Aurelio Uncini, Francesco Piazza Dipartiento di Elettronica e Autoatica - Università di Ancona Via Brecce Bianche, Ancona-Italy. Fax:+39 (071) eail: aurel@ieee.org Internet:
2 A NEW CLASS OF APEX LIKE PCA ALGORITHMS Sione Fiori, Aurelio Uncini and Francesco Piazza Dept. Electronics and Autoatics University of Ancona (Italy) E-ail : sione@eealab. unian. it ABSTRACT One of the ost coonly known algorith to perfor neural Principal Coponent Analysis of real-valued rando signals is the Kung-Diaantaras Adaptive Principal coponent EXtractor (APEX) for a laterally-connected neural architecture. In this paper we present a new approach to obtain an APEX-like PCA procedure as a special case of a ore general class of learning rules, by eans of an optiization theory specialized for the laterally-connected topology. Through siulations we show the new algoriths can be faster than the original one. 1. INTRODUCTION Principal Coponent Analysis (PCA) of ultivariate rando signals is a well-known statistical data analysis technique [1, 7]. It is possible to show that a linear transforation z = W t x of a given ultiple rando signal x into a new rando signal z with less coponents than x, such that: Hebbian Algorith, GHA, [8]), Kung-Diaantaras (APEX, [5, 6, 7]). All of these ethods are characterized by different architectural coplexity, convergence speed properties and nuerical precision at the equilibriu. In this paper we deal with one of these: the Adaptive Principal coponent EXtractor (APEX, [6, 7]) based on a laterally-connected neural architecture. It has a wide relevance in the field of analog ipleentations because it is characterized by a very low coplexity. Here we derive a new class of PCA algoriths based on the laterallyconnected neural architecture, arising fro a siple optiization theory specialized for this topology. Such a class contains, as a special case, an APEX-like algorith, but it contains also a subclass of algoriths that show a saller architectural coplexity and interesting convergence features when copared with the original one. NOTATION. In the following E[.] returns the atheatical expectation of the arguent; operator SUT[A] returns the strictly upper triangular part of the square atrix A; the i:th entry of the generic vector v is denoted with vi. the transfored signal power is axiized under suitable constraints [4]; the transfored scalar signals are statistically decorrelated [4]; the signal x is optially represented by z (in the ean squared reconstruction error sense) [4]; a proper easure of the uncertain [9], y of z is axiized can be obtained by assuing W = F, where (F,.) is a PCA of x. (The foral definition of PCA in ters of atrix pairs (F, D) can be found in [3].) Matrix F contains eigenvectors (noralized to unitary nors) of the covariance atrix of the analyzed signal, while D contains the powers of the Principal Coponents arranged in a descending order. In the literature several algoriths are known that allow the extraction of the (unique) PCA of a signal fro itself. The ost coonly used are those by Sanger (Generalized This researchwassupportedby the Italian MURST. 2. THE LATERALLY-CONNECTED NEURAL ARCHITECTURE Kung and Diaantaras realized a Principal Coponent analyzer using a linear neural network described by the following neural schee: y= Wtx+Lty, (1) with a proper unsupervised learning rule. The input vector x E 7?P, the output vector y E R (with < p, arbitrarily fixed), the direct-connection p x weight-atrix W and the lateral-connection x weight-atrix L are intended to be evaluated at the sae teporal instant. The coluns of W and L are naed in the following way: W = [W, w,.. w~], L= [ ~]. Notice that, being Lt a strictly lower-triangular square atrix (i.e. L~k = O if i z k), this neural network is hierarchical not recurrent. The original learning rule for the weight-atrix W was: AW = q[xy WY2], (2)
3 and the learning rule for the weight-atrix L was: AL= qsut[yy] qly2, (3) where q is a positive learning rate, X is a p x atrix, Y and Y are x atrices, defined by: Y = diag(yl, ys,....y~). Kung and Diaantaras were able to prove the convergence of the above algorith under soe conditions. In particular, we can restate their result saying that: (Theore.) Let x be a p-coponents real rando signal, zero-ean, with a finite covariance endowed with non-null distinct eigenvalues, and let (F, D) the unique PCA of x. Let DKN be the neural-net described by (1) trained by eans of the learningpair (2)-(3). If the rate q is chosen so sall that the behavior of the algorith is asyptotically stable and the initial entries of W are sall rando nubers and L(0) = O, then in the ean it holds true that: li L(t) = O, t-+ #~1 W(t) = F, JliI E[y(t)y (t)] = D In other words, under the above conditions the DKNasyptotically becoes, in the ean, a principal coponent ana- Iyzez rl Strictly speaking, they proved that if q is sufficiently sall and suitable initial conditions are assued, then in the ean (W, E[yyt]) ~ (F, D), where (F, D) is a PCA of the signal x. We call the above issue the Kung-Diaantaras Result (KDR). 3. THE +APEX CLASS In the following Subsection a new class of APEX like algoriths is presented. Later, differences and siilarities between our new algoriths and other ones will be discussed APEX-like algoriths based on an optiization forulation A PCA transforation is such that the transfored signals (with the above sybology, z = Wtx) are characterized by axiu variance. Furtherore, fro the foral definition of PCA it is known that, at the equilibriu, any unique PCA vector w~ ust be orthogonal with respect to each other and with an unitary nor. These targets can be thought as separated objectives to be attained by eans of laterally-connected More forally we can state the following: neural topology. (Proposition.) It is possible to dejine a pair (J, C) of objective functions whose extreization process yields a class of PCA algoriths containing, as a special case, an APEX-like. Functions J and C can be properly fixed by exaining the structure of a generic output signal yi fro(1) squared. Direct calculations show: Y? = (W:x)z + (UY)2 + 2(W: X)(UY). (4) The first ter at the right hand contains in the ean the power of the transfored signal.q = w~x, while the second ter at the right hand of the above equation contains in the ean a linear cobination of the output signals crosscorrelation, in fact it holds true that 11[(1~y) 2] = 1:E[yyt ]1~. By definition of PCA, the first one has to be axiized under the constraint w~w~ = 1 ([4]), while the second one ust be zeroed. Here we propose to use the direct-connection adaptation to axiize the powers of the transfored signal by axiizing the following objective function: J(W, L) ~f ~E[y~] + ; ~(W:Wi 1)/ui, (5) i=l %=1 with respect to W only. In the above equation pi are socalled Lagrange ultipliers to be deterined by iposing the constraints w: Wi = 1 in the equilibriu conditions ~ = O. It is iportant to notice that by definition of L, a scalar product (l: y) does not depend on w,, but only on W2 for j < i, then fro equations (4) and (5) we obtain: 8J 8W% = 2E[(W:X)X] + 2E[(l:y)x] + p~w~ = 2E[yix] + p~w~, therefore the optiu w~ satisfies:, (9J 2E[yi(w:x)] + Pi = 0, i dw~ and optiu pi is pi = 213[yi.zi]. If the Gradient Steepest Ascent (GSA) ethod is used to adapt each wi, that eans Awi = + ~q ~, the stochastic learning rule for W reads: AW = q[xy WYZ], (6) and true gradient of J has been replaced by its stochastic instantaneous approxiation. Finally, we choose to adapt the lateral-connection weightatrix L only, in order to iniize a cost function defined as: C(W, L) :f ~ E[y;] + ~(l;li)tii, (7) i=l i=l
4 where a set of Lagrange has been introduced for the constraints llli//2 = O(that have to be reached attheequilibriu to preserve KDR)and toaddtothesystern a nuber of degree of freedo. Besides, it is interesting to recognize that those constraints also ebed a regularization property on the global criterion [2]. As it can be directly proved by using standard Kuhn-Tucker theory [2], under these constraints there are no theoretical reasons to force functions ~i to assue anyparticularshape. This second objective function C can be iniized, with respect to the variable atrix L, by eans of a Gradient Steepest Descent (GSD) ethod Ali = ~ q%. Fro equations (4) and (7) it follows: (%3 (Xi = 2E[yiy[~l] + 2@ili, where y[~l=[ylyz...yl l ]tfor2~i~, andyi1l=[o O... O O]t. By rewriting GSD equations in atrix notation, ignoring again expectation operator, the new stochastic learning rule for L reads: AL = +.UT[YY] ql*, (8) with being defined Rule (8) provides iniization of the cross-correlation between the network s output signals. Now we have all the eleents to propose the following definition, relative to the class of algorith represented by the above new neural learning rules: (Definition.) The faily of learning rules described by equations (6) and (8) is called the + APEX Principal Coponent analyzer class. The special eleent in this fareily with $ = Y2 is called Y2 APEX. Notice that Y2 APEX is not the s~e alg~rith as the original APEX, but as L -+ O also Z ~ Y, thus these algoriths asyptotically behave in the sae way, and we call it APEX-like. It is also iportant to reark that, apart fro firther stability considerations, the choice of the ultiplying fictions ii(t) is free. h fact, we can adopt any suitable arbitrarily chosen function that guarantees the asyptotic stability of the global learning process and good perforances of the Principal Coponent analyzing algorith Discussion In practice, in our experients we have exained the following three cases: 1. all are chosen null; 2. vi(t) are arbitrarily chosen non-null constant values tii ; 3. the vi (t) are assued as particular non-constant fhnctions of the unique variables yi (t). Roughly speaking, we can identie the special PC s extractor obtained by vanishing free functions ~i (t) as the & APEX algorith, whose descriptive equations are: AW = ~[XY WYZ], (9) AL = qsut[yy], (lo) In a coputational-coplexity point of view this algorith is the ost interesting one, since it requires a saller aount of operations than the original APEX, as shown in Table 1. The above rule recalls the linearized Rubner-Tavan odel that the O-APEX asyptotically behaves like. (For knowing details about Rubner-Tavan approach readers please refer to [7].) We observed that the ter y? in each of the (3) is too uch large and can also lead the algorith very far fro the right solution. Thus, when non-constant non-null fictions $$ are used, we found useful they satisfy this constraint: Each +~ (t)should be a positive fiction that grows less than t2 at least for large It 1. For instance, we found good results with ~; = Iyi 1. Other suitable choices are of course possible. Algorith Coplexity (Operations) GHA 2p+ ~(2 +)(p+ 1) APEX 3p+ ~2 ~ O-APEX Table 1: Coplexity 3P + 2 coparison. Table 1 provides estiates of the architectural coplexity of the neural networks in ters of the nuber of eleentary operations required by the corresponding learning rules with respect to the network diensions. We define an operation as a product eventually followed by a su. 4. EXPERIMENTAL RESULTS To assess our theoretical analysis and copare algoriths perforances, we perfored siulations by using Sanger s GHA, standard APEX and our new algoriths belonging to the O APEX class. Such PCA algoriths have been run with a network input signal x = Qs, where Q is a p x p orthonoral atrix (Qt Q = I) randoly generated, and s contains p utually uncorrelated zero-ean rando signals Si with dif- 2 = ~[5~]. signals Si are placed in s so ferent powers Ui that their powers are decreasingly ordered, i.e. u? > C! if i < j. This iplies that the first Principal Coponents of x (with < p) are the first colun-vectors of Q.
5 :L C Figure 1: Convergence speed coparison. Figure 2: Coparison of APEX algoriths. Each algorith starts fro the sae initial conditions, that are rando for W and null for L. In order to copare the convergence speed of the new algoriths with those of the GHA and APEX, a suitable easure of convergence b is used. This easure is defined as 6(W) = /IW Q IIF, where Q is that atrix whose coluns are the first of Q, and II. IIF denotes the Frobenius nor. Note that the quantity S ay converge to different values since recovering the coluns of Q is sign-blind. Siulation presented in Figure 1 concerns GHA, APEX and Iyl APEX (that = Iy~1) algoriths. The above results are obtained with a learning stepsize q = 0.01, network s diension p = 10 and = 5. Powers u; were assued fro the exponential law u? = 22 i (where i ranges fro 1 to p) in order to keep a good eigenvalue spread. New Iyl APEX perfors well: its convergence toward KDR looks faster and its precision sees to be coparable at all with that of the other algoriths. Figure 2 shows typical courses of lyl- APEX and Y2 APEX copared together for a: = 0.1 (p i + 1). Here input signals have sall powers one close to another. In this case all algoriths after few steps behave alost identically, therefore the O-APEX is the ost convenient one. 5. CONCLUSION In [7] a wide generalization of the standard APEX has been presented, but to our knowledge special attention has not been paid to particularization nor tests have been perfored in order to discover their features, hence we believe this paper points out soe new issue and contains new contributions. Extensions of the new ethod to the coplex-valued case is currently under investigation. 6. REFERENCES [2] [3] [4] [5] [6] [7] [8] [9] ral networks: A survey, IEEE Trans. on Neural works, Vol. 6, No. 4, pp , July 1995 Net- A. CICHOCKI AND R. UNBEHAUEN, Neural networks for optiization andsignalprocessing, J. Wiley Ltd., 1993 P. COMON, Independent Coponent Analysis, a new concept?, Signal Processing, Vol. 36, pp , 1994 J. KARHUNEN, Optiization criteria and nonlinear PCA neural networks, Proc. of International Joint Conference on Neural Networks (IJCNN), pp ,1994 S.Y. KUNG, Constrained Principal Coponent Analysis via an orthogonal learning network, Proc. of International Syposiu on Circuits and Systes (IS- CAS), pp , 1990 S.Y. KUNG AND K.I. DIAMANTARAS, A network learning algorith for adaptive principal coponent extraction, Proc. of International Conference on Acoustics, Speech and Signal Processing (ICASSP), 1990, pp S.Y. KUNG AND K.I. DIAMANTARAS, Principal Coponent Neural Networks: Theory and Applications, J. Wiley, 1996 r.d. SANGER, Optial unsupervised learning in a single-layer neural network, Neural Networks, Vol. 2, pp , 1989 L. XU, Theories for unsupervised learning: PCA and its nonlinear extension, Proc. of International Joint Conference on Neural Networks (IJCNN), pp ,1994 [1] P.F. BALDI AND K. HORNIK, Learning in linear neu-
Feature Extraction Techniques
Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that
More informationIntroduction to Machine Learning. Recitation 11
Introduction to Machine Learning Lecturer: Regev Schweiger Recitation Fall Seester Scribe: Regev Schweiger. Kernel Ridge Regression We now take on the task of kernel-izing ridge regression. Let x,...,
More informationCh 12: Variations on Backpropagation
Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith
More informationBlock designs and statistics
Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent
More informationExperimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis
City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October
More informationBoosting with log-loss
Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the
More informationPrincipal Components Analysis
Principal Coponents Analysis Cheng Li, Bingyu Wang Noveber 3, 204 What s PCA Principal coponent analysis (PCA) is a statistical procedure that uses an orthogonal transforation to convert a set of observations
More informationInteractive Markov Models of Evolutionary Algorithms
Cleveland State University EngagedScholarship@CSU Electrical Engineering & Coputer Science Faculty Publications Electrical Engineering & Coputer Science Departent 2015 Interactive Markov Models of Evolutionary
More information3.3 Variational Characterization of Singular Values
3.3. Variational Characterization of Singular Values 61 3.3 Variational Characterization of Singular Values Since the singular values are square roots of the eigenvalues of the Heritian atrices A A and
More informationSharp Time Data Tradeoffs for Linear Inverse Problems
Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used
More informationChapter 6 1-D Continuous Groups
Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:
More informationSupport Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization
Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering
More informationA Simple Regression Problem
A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where
More informationModel Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon
Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential
More informationA Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair
Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Notes for EE7C (Spring 018: Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee7c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee7c@berkeley.edu October 15,
More informationKernel Methods and Support Vector Machines
Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic
More informationThe proofs of Theorem 1-3 are along the lines of Wied and Galeano (2013).
A Appendix: Proofs The proofs of Theore 1-3 are along the lines of Wied and Galeano (2013) Proof of Theore 1 Let D[d 1, d 2 ] be the space of càdlàg functions on the interval [d 1, d 2 ] equipped with
More informationThe Transactional Nature of Quantum Information
The Transactional Nature of Quantu Inforation Subhash Kak Departent of Coputer Science Oklahoa State University Stillwater, OK 7478 ABSTRACT Inforation, in its counications sense, is a transactional property.
More informationIntelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines
Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes
More informationSlide10. Haykin Chapter 8: Principal Components Analysis. Motivation. Principal Component Analysis: Variance Probe
Slide10 Motivation Haykin Chapter 8: Principal Coponents Analysis 1.6 1.4 1.2 1 0.8 cloud.dat 0.6 CPSC 636-600 Instructor: Yoonsuck Choe Spring 2015 0.4 0.2 0 0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 How can we
More information13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices
CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay
More informationNon-Parametric Non-Line-of-Sight Identification 1
Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,
More informationRandomized Recovery for Boolean Compressed Sensing
Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch
More informationCOS 424: Interacting with Data. Written Exercises
COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well
More informationOptimal Jamming Over Additive Noise: Vector Source-Channel Case
Fifty-first Annual Allerton Conference Allerton House, UIUC, Illinois, USA October 2-3, 2013 Optial Jaing Over Additive Noise: Vector Source-Channel Case Erah Akyol and Kenneth Rose Abstract This paper
More informationLeast Squares Fitting of Data
Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a
More informationOn Constant Power Water-filling
On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives
More informationFast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials
Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter
More informationExplicit solution of the polynomial least-squares approximation problem on Chebyshev extrema nodes
Explicit solution of the polynoial least-squares approxiation proble on Chebyshev extrea nodes Alfredo Eisinberg, Giuseppe Fedele Dipartiento di Elettronica Inforatica e Sisteistica, Università degli Studi
More informationA Low-Complexity Congestion Control and Scheduling Algorithm for Multihop Wireless Networks with Order-Optimal Per-Flow Delay
A Low-Coplexity Congestion Control and Scheduling Algorith for Multihop Wireless Networks with Order-Optial Per-Flow Delay Po-Kai Huang, Xiaojun Lin, and Chih-Chun Wang School of Electrical and Coputer
More informationA remark on a success rate model for DPA and CPA
A reark on a success rate odel for DPA and CPA A. Wieers, BSI Version 0.5 andreas.wieers@bsi.bund.de Septeber 5, 2018 Abstract The success rate is the ost coon evaluation etric for easuring the perforance
More informationQuantum algorithms (CO 781, Winter 2008) Prof. Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search
Quantu algoriths (CO 781, Winter 2008) Prof Andrew Childs, University of Waterloo LECTURE 15: Unstructured search and spatial search ow we begin to discuss applications of quantu walks to search algoriths
More informationMulti-Dimensional Hegselmann-Krause Dynamics
Multi-Diensional Hegselann-Krause Dynaics A. Nedić Industrial and Enterprise Systes Engineering Dept. University of Illinois Urbana, IL 680 angelia@illinois.edu B. Touri Coordinated Science Laboratory
More informationESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS. A Thesis. Presented to. The Faculty of the Department of Mathematics
ESTIMATING AND FORMING CONFIDENCE INTERVALS FOR EXTREMA OF RANDOM POLYNOMIALS A Thesis Presented to The Faculty of the Departent of Matheatics San Jose State University In Partial Fulfillent of the Requireents
More informationTopic 5a Introduction to Curve Fitting & Linear Regression
/7/08 Course Instructor Dr. Rayond C. Rup Oice: A 337 Phone: (95) 747 6958 E ail: rcrup@utep.edu opic 5a Introduction to Curve Fitting & Linear Regression EE 4386/530 Coputational ethods in EE Outline
More informationDISSIMILARITY MEASURES FOR ICA-BASED SOURCE NUMBER ESTIMATION. Seungchul Lee 2 2. University of Michigan. Ann Arbor, MI, USA.
Proceedings of the ASME International Manufacturing Science and Engineering Conference MSEC June -8,, Notre Dae, Indiana, USA MSEC-7 DISSIMILARIY MEASURES FOR ICA-BASED SOURCE NUMBER ESIMAION Wei Cheng,
More informationConstrained Consensus and Optimization in Multi-Agent Networks arxiv: v2 [math.oc] 17 Dec 2008
LIDS Report 2779 1 Constrained Consensus and Optiization in Multi-Agent Networks arxiv:0802.3922v2 [ath.oc] 17 Dec 2008 Angelia Nedić, Asuan Ozdaglar, and Pablo A. Parrilo February 15, 2013 Abstract We
More informationarxiv: v1 [cs.ds] 3 Feb 2014
arxiv:40.043v [cs.ds] 3 Feb 04 A Bound on the Expected Optiality of Rando Feasible Solutions to Cobinatorial Optiization Probles Evan A. Sultani The Johns Hopins University APL evan@sultani.co http://www.sultani.co/
More informationA note on the multiplication of sparse matrices
Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani
More informationA method to determine relative stroke detection efficiencies from multiplicity distributions
A ethod to deterine relative stroke detection eiciencies ro ultiplicity distributions Schulz W. and Cuins K. 2. Austrian Lightning Detection and Inoration Syste (ALDIS), Kahlenberger Str.2A, 90 Vienna,
More informationA Note on the Applied Use of MDL Approximations
A Note on the Applied Use of MDL Approxiations Daniel J. Navarro Departent of Psychology Ohio State University Abstract An applied proble is discussed in which two nested psychological odels of retention
More informationRecursive Algebraic Frisch Scheme: a Particle-Based Approach
Recursive Algebraic Frisch Schee: a Particle-Based Approach Stefano Massaroli Renato Myagusuku Federico Califano Claudio Melchiorri Atsushi Yaashita Hajie Asaa Departent of Precision Engineering, The University
More informationHESSIAN MATRICES OF PENALTY FUNCTIONS FOR SOLVING CONSTRAINED-OPTIMIZATION PROBLEMS
R 702 Philips Res. Repts 24, 322-330, 1969 HESSIAN MATRICES OF PENALTY FUNCTIONS FOR SOLVING CONSTRAINED-OPTIMIZATION PROBLEMS by F. A. LOOTSMA Abstract This paper deals with the Hessian atrices of penalty
More informationMachine Learning Basics: Estimators, Bias and Variance
Machine Learning Basics: Estiators, Bias and Variance Sargur N. srihari@cedar.buffalo.edu This is part of lecture slides on Deep Learning: http://www.cedar.buffalo.edu/~srihari/cse676 1 Topics in Basics
More informationHybrid System Identification: An SDP Approach
49th IEEE Conference on Decision and Control Deceber 15-17, 2010 Hilton Atlanta Hotel, Atlanta, GA, USA Hybrid Syste Identification: An SDP Approach C Feng, C M Lagoa, N Ozay and M Sznaier Abstract The
More informationASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical
IEEE TRANSACTIONS ON INFORMATION THEORY Large Alphabet Source Coding using Independent Coponent Analysis Aichai Painsky, Meber, IEEE, Saharon Rosset and Meir Feder, Fellow, IEEE arxiv:67.7v [cs.it] Jul
More informationare equal to zero, where, q = p 1. For each gene j, the pairwise null and alternative hypotheses are,
Page of 8 Suppleentary Materials: A ultiple testing procedure for ulti-diensional pairwise coparisons with application to gene expression studies Anjana Grandhi, Wenge Guo, Shyaal D. Peddada S Notations
More information. The univariate situation. It is well-known for a long tie that denoinators of Pade approxiants can be considered as orthogonal polynoials with respe
PROPERTIES OF MULTIVARIATE HOMOGENEOUS ORTHOGONAL POLYNOMIALS Brahi Benouahane y Annie Cuyt? Keywords Abstract It is well-known that the denoinators of Pade approxiants can be considered as orthogonal
More informationTesting equality of variances for multiple univariate normal populations
University of Wollongong Research Online Centre for Statistical & Survey Methodology Working Paper Series Faculty of Engineering and Inforation Sciences 0 esting equality of variances for ultiple univariate
More informationUnsupervised Learning: Dimension Reduction
Unsupervised Learning: Diension Reduction by Prof. Seungchul Lee isystes Design Lab http://isystes.unist.ac.kr/ UNIST Table of Contents I.. Principal Coponent Analysis (PCA) II. 2. PCA Algorith I. 2..
More informationFixed-to-Variable Length Distribution Matching
Fixed-to-Variable Length Distribution Matching Rana Ali Ajad and Georg Böcherer Institute for Counications Engineering Technische Universität München, Gerany Eail: raa2463@gail.co,georg.boecherer@tu.de
More informationIntelligent Systems: Reasoning and Recognition. Artificial Neural Networks
Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial
More informationPattern Recognition and Machine Learning. Artificial Neural networks
Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial
More informationComplex Quadratic Optimization and Semidefinite Programming
Coplex Quadratic Optiization and Seidefinite Prograing Shuzhong Zhang Yongwei Huang August 4 Abstract In this paper we study the approxiation algoriths for a class of discrete quadratic optiization probles
More informationStochastic Subgradient Methods
Stochastic Subgradient Methods Lingjie Weng Yutian Chen Bren School of Inforation and Coputer Science University of California, Irvine {wengl, yutianc}@ics.uci.edu Abstract Stochastic subgradient ethods
More informationRecovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup)
Recovering Data fro Underdeterined Quadratic Measureents (CS 229a Project: Final Writeup) Mahdi Soltanolkotabi Deceber 16, 2011 1 Introduction Data that arises fro engineering applications often contains
More informationPattern Recognition and Machine Learning. Artificial Neural networks
Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial
More informationE0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis
E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds
More informationA new type of lower bound for the largest eigenvalue of a symmetric matrix
Linear Algebra and its Applications 47 7 9 9 www.elsevier.co/locate/laa A new type of lower bound for the largest eigenvalue of a syetric atrix Piet Van Mieghe Delft University of Technology, P.O. Box
More informationBipartite subgraphs and the smallest eigenvalue
Bipartite subgraphs and the sallest eigenvalue Noga Alon Benny Sudaov Abstract Two results dealing with the relation between the sallest eigenvalue of a graph and its bipartite subgraphs are obtained.
More informationLower Bounds for Quantized Matrix Completion
Lower Bounds for Quantized Matrix Copletion Mary Wootters and Yaniv Plan Departent of Matheatics University of Michigan Ann Arbor, MI Eail: wootters, yplan}@uich.edu Mark A. Davenport School of Elec. &
More informationExact tensor completion with sum-of-squares
Proceedings of Machine Learning Research vol 65:1 54, 2017 30th Annual Conference on Learning Theory Exact tensor copletion with su-of-squares Aaron Potechin Institute for Advanced Study, Princeton David
More informationSupplementary Information for Design of Bending Multi-Layer Electroactive Polymer Actuators
Suppleentary Inforation for Design of Bending Multi-Layer Electroactive Polyer Actuators Bavani Balakrisnan, Alek Nacev, and Elisabeth Sela University of Maryland, College Park, Maryland 074 1 Analytical
More informationUse of PSO in Parameter Estimation of Robot Dynamics; Part One: No Need for Parameterization
Use of PSO in Paraeter Estiation of Robot Dynaics; Part One: No Need for Paraeterization Hossein Jahandideh, Mehrzad Navar Abstract Offline procedures for estiating paraeters of robot dynaics are practically
More informationUsing EM To Estimate A Probablity Density With A Mixture Of Gaussians
Using EM To Estiate A Probablity Density With A Mixture Of Gaussians Aaron A. D Souza adsouza@usc.edu Introduction The proble we are trying to address in this note is siple. Given a set of data points
More informationMSEC MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL SOLUTION FOR MAINTENANCE AND PERFORMANCE
Proceeding of the ASME 9 International Manufacturing Science and Engineering Conference MSEC9 October 4-7, 9, West Lafayette, Indiana, USA MSEC9-8466 MODELING OF DEGRADATION PROCESSES TO OBTAIN AN OPTIMAL
More informationThis model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.
CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when
More informationThe Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters
journal of ultivariate analysis 58, 96106 (1996) article no. 0041 The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Paraeters H. S. Steyn
More informationDesign of Spatially Coupled LDPC Codes over GF(q) for Windowed Decoding
IEEE TRANSACTIONS ON INFORMATION THEORY (SUBMITTED PAPER) 1 Design of Spatially Coupled LDPC Codes over GF(q) for Windowed Decoding Lai Wei, Student Meber, IEEE, David G. M. Mitchell, Meber, IEEE, Thoas
More informationFigure 1: Equivalent electric (RC) circuit of a neurons membrane
Exercise: Leaky integrate and fire odel of neural spike generation This exercise investigates a siplified odel of how neurons spike in response to current inputs, one of the ost fundaental properties of
More informationExtension of CSRSM for the Parametric Study of the Face Stability of Pressurized Tunnels
Extension of CSRSM for the Paraetric Study of the Face Stability of Pressurized Tunnels Guilhe Mollon 1, Daniel Dias 2, and Abdul-Haid Soubra 3, M.ASCE 1 LGCIE, INSA Lyon, Université de Lyon, Doaine scientifique
More informatione-companion ONLY AVAILABLE IN ELECTRONIC FORM
OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer
More informationDERIVING PROPER UNIFORM PRIORS FOR REGRESSION COEFFICIENTS
DERIVING PROPER UNIFORM PRIORS FOR REGRESSION COEFFICIENTS N. van Erp and P. van Gelder Structural Hydraulic and Probabilistic Design, TU Delft Delft, The Netherlands Abstract. In probles of odel coparison
More informationLeonardo R. Bachega*, Student Member, IEEE, Srikanth Hariharan, Student Member, IEEE Charles A. Bouman, Fellow, IEEE, and Ness Shroff, Fellow, IEEE
Distributed Signal Decorrelation and Detection in Sensor Networks Using the Vector Sparse Matrix Transfor Leonardo R Bachega*, Student Meber, I, Srikanth Hariharan, Student Meber, I Charles A Bouan, Fellow,
More informationOPTIMIZATION in multi-agent networks has attracted
Distributed constrained optiization and consensus in uncertain networks via proxial iniization Kostas Margellos, Alessandro Falsone, Sione Garatti and Maria Prandini arxiv:603.039v3 [ath.oc] 3 May 07 Abstract
More informationMultivariate Methods. Matlab Example. Principal Components Analysis -- PCA
Multivariate Methos Xiaoun Qi Principal Coponents Analysis -- PCA he PCA etho generates a new set of variables, calle principal coponents Each principal coponent is a linear cobination of the original
More informationFairness via priority scheduling
Fairness via priority scheduling Veeraruna Kavitha, N Heachandra and Debayan Das IEOR, IIT Bobay, Mubai, 400076, India vavitha,nh,debayan}@iitbacin Abstract In the context of ulti-agent resource allocation
More informationLecture 21. Interior Point Methods Setup and Algorithm
Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and
More informationAn Improved Particle Filter with Applications in Ballistic Target Tracking
Sensors & ransducers Vol. 72 Issue 6 June 204 pp. 96-20 Sensors & ransducers 204 by IFSA Publishing S. L. http://www.sensorsportal.co An Iproved Particle Filter with Applications in Ballistic arget racing
More informationRESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS
BIT Nuerical Matheatics 43: 459 466, 2003. 2003 Kluwer Acadeic Publishers. Printed in The Netherlands 459 RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS V. SIMONCINI Dipartiento di
More information1 Bounding the Margin
COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost
More informationWeighted- 1 minimization with multiple weighting sets
Weighted- 1 iniization with ultiple weighting sets Hassan Mansour a,b and Özgür Yılaza a Matheatics Departent, University of British Colubia, Vancouver - BC, Canada; b Coputer Science Departent, University
More informationProbability Distributions
Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples
More informationProc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES
Proc. of the IEEE/OES Seventh Working Conference on Current Measureent Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Belinda Lipa Codar Ocean Sensors 15 La Sandra Way, Portola Valley, CA 98 blipa@pogo.co
More informationREDUCTION OF FINITE ELEMENT MODELS BY PARAMETER IDENTIFICATION
ISSN 139 14X INFORMATION TECHNOLOGY AND CONTROL, 008, Vol.37, No.3 REDUCTION OF FINITE ELEMENT MODELS BY PARAMETER IDENTIFICATION Riantas Barauskas, Vidantas Riavičius Departent of Syste Analysis, Kaunas
More informationKeywords: Estimator, Bias, Mean-squared error, normality, generalized Pareto distribution
Testing approxiate norality of an estiator using the estiated MSE and bias with an application to the shape paraeter of the generalized Pareto distribution J. Martin van Zyl Abstract In this work the norality
More informationUpper and Lower Bounds on the Capacity of Wireless Optical Intensity Channels
ISIT7, Nice, France, June 4 June 9, 7 Upper and Lower Bounds on the Capacity of Wireless Optical Intensity Channels Ahed A. Farid and Steve Hranilovic Dept. Electrical and Coputer Engineering McMaster
More informationNonmonotonic Networks. a. IRST, I Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I Povo (Trento) Italy
Storage Capacity and Dynaics of Nononotonic Networks Bruno Crespi a and Ignazio Lazzizzera b a. IRST, I-38050 Povo (Trento) Italy, b. Univ. of Trento, Physics Dept., I-38050 Povo (Trento) Italy INFN Gruppo
More informationCompression and Predictive Distributions for Large Alphabet i.i.d and Markov models
2014 IEEE International Syposiu on Inforation Theory Copression and Predictive Distributions for Large Alphabet i.i.d and Markov odels Xiao Yang Departent of Statistics Yale University New Haven, CT, 06511
More informationVulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links
Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Tie-Varying Jaing Links Jun Kurihara KDDI R&D Laboratories, Inc 2 5 Ohara, Fujiino, Saitaa, 356 8502 Japan Eail: kurihara@kddilabsjp
More informationOBJECTIVES INTRODUCTION
M7 Chapter 3 Section 1 OBJECTIVES Suarize data using easures of central tendency, such as the ean, edian, ode, and idrange. Describe data using the easures of variation, such as the range, variance, and
More informationReed-Muller Codes. m r inductive definition. Later, we shall explain how to construct Reed-Muller codes using the Kronecker product.
Coding Theory Massoud Malek Reed-Muller Codes An iportant class of linear block codes rich in algebraic and geoetric structure is the class of Reed-Muller codes, which includes the Extended Haing code.
More informationarxiv: v1 [stat.ot] 7 Jul 2010
Hotelling s test for highly correlated data P. Bubeliny e-ail: bubeliny@karlin.ff.cuni.cz Charles University, Faculty of Matheatics and Physics, KPMS, Sokolovska 83, Prague, Czech Republic, 8675. arxiv:007.094v
More informationCONTROL SYSTEMS, ROBOTICS, AND AUTOMATION Vol. IX Uncertainty Models For Robustness Analysis - A. Garulli, A. Tesi and A. Vicino
UNCERTAINTY MODELS FOR ROBUSTNESS ANALYSIS A. Garulli Dipartiento di Ingegneria dell Inforazione, Università di Siena, Italy A. Tesi Dipartiento di Sistei e Inforatica, Università di Firenze, Italy A.
More informationVariations on Backpropagation
2 Variations on Backpropagation 2 Variations Heuristic Modifications Moentu Variable Learning Rate Standard Nuerical Optiization Conjugate Gradient Newton s Method (Levenberg-Marquardt) 2 2 Perforance
More informationAn Adaptive UKF Algorithm for the State and Parameter Estimations of a Mobile Robot
Vol. 34, No. 1 ACTA AUTOMATICA SINICA January, 2008 An Adaptive UKF Algorith for the State and Paraeter Estiations of a Mobile Robot SONG Qi 1, 2 HAN Jian-Da 1 Abstract For iproving the estiation accuracy
More informationarxiv: v1 [math.na] 10 Oct 2016
GREEDY GAUSS-NEWTON ALGORITHM FOR FINDING SPARSE SOLUTIONS TO NONLINEAR UNDERDETERMINED SYSTEMS OF EQUATIONS MÅRTEN GULLIKSSON AND ANNA OLEYNIK arxiv:6.395v [ath.na] Oct 26 Abstract. We consider the proble
More informationMAXIMUM LIKELIHOOD BASED TECHNIQUES FOR BLIND SOURCE SEPARATION AND APPROXIMATE JOINT DIAGONALIZATION
BEN-GURION UNIVERSITY OF TE NEGEV FACULTY OF ENGINEERING SCIENCE DEPARTENT OF ELECTRICAL AND COPUTER ENGINEERING AXIU LIKELIOOD BASED TECNIQUES FOR BLIND SOURCE SEPARATION AND APPROXIATE JOINT DIAGONALIZATION
More information