HESSIAN MATRICES OF PENALTY FUNCTIONS FOR SOLVING CONSTRAINED-OPTIMIZATION PROBLEMS

Size: px
Start display at page:

Download "HESSIAN MATRICES OF PENALTY FUNCTIONS FOR SOLVING CONSTRAINED-OPTIMIZATION PROBLEMS"

Transcription

1 R 702 Philips Res. Repts 24, , 1969 HESSIAN MATRICES OF PENALTY FUNCTIONS FOR SOLVING CONSTRAINED-OPTIMIZATION PROBLEMS by F. A. LOOTSMA Abstract This paper deals with the Hessian atrices of penalty functions, evaluated at their iniizing point. It is concerned with the condition nuber of these atrices for sall values of a controlling paraeter. At the end of the paper a coparison is ade between different types of penalty functions on the grounds of the results obtained. 1. Classification of penalty-function techniques Throughout this paper we shall be concerned with the proble iniize f(x) subject t~ } g,(x) ~ 0; 1 = I,...,, (1.1) where x denotes an eleent of the n-diensional vector space En. We shall be assuing that the functions J, -gl>..., -g are convex and twice differentiable with continuous second-order partial derivatives on an open, convex subset V of En. The constraint set R={xlg,(x);;:;::O; i=i,...,} (1.2) is a bounded subset of V. The interior Ro of R is non-epty. We consider a nuber of penalty-function techniques for solving proble (1.1). One can distinguish two classes, both of which have been referred to by expressive naes. The interior-point ethods operate in the interior Ro of R. The penalty function is given by Br(x) =f(x) - r ~ cp[g,(x)], '=1 (1.3) where cp is a concave function of one variable, say y. Its derivative tp' cp'(y) = y-v reads (l.4) with a positive, integer '11. A point x(r) iniizing (1.3) over Ro 'exists then for "any r > 0, Any convergent sequence {x(rk)}' where {rk} is a onotonie, decreasing null sequence as k --* 00, converges to a solution of (1.1). The exterior-point ethods or outside-in ethods present an approach to a iniu solution fro outside the constraint set. The' general for of the

2 HESSIAN MATRICES OF PENALTY FUNCTIONS 323 penalty function is given by Lr(x) = f(x) - r- 1 ~ lp[g,(x)], '=1 (1.5) where 11' is a concave function of one variable y, such that for y < 0, 1p(y) = co(y) for y::::;; 0. The derivative co' of co is given by { co'(y) = (-y)v. (1.6) Let z(r) denote a point iniizing (1.5) over En. Any convergent sequence {zerk)}, where {rk} denotes again a onotonie, decreasing null sequence, converges to a solution of (1.1). It will be convenient to extend the terinology that we have been using in previous papers. Following Murray S) we shall refer to interior-point penalty functions of the type (1.3) as barrier functions. The exterior-point penalty functions (1.5) will briefly be indicated as loss functions, a nae which has also been used by Fiacco and McCorick 1). Furtherore, we introduce a classification based on the behaviour of the functions tp' and co' in a neighbourhood of y = O. A barrier function is said to be of order 'JI if the function tp' has a pole of order 'JI at y = o. Siilarly, a loss function is of order 'JI if the function co' has a zero of order 'JI at y = o. 2. Conditioning An intriguing point is the choice of a penalty function for nuerical purposes. We shall not repeat here all the arguents supporting the choice of the firstorder penalty functions for coputational purposes. Our concern is an arguent which has been introduced only by Murray S), naely the question of "conditioning". This is a qualification referring to the Hessian atrix of a penalty function. The otivation for such a study is the idea that failures of (second-order) unconstrained-iniization techniques ay be due to illconditioning of the Hessian atrix at soe iteration points. Throughout this paper it is tacitly assued that penalty functions are strictly convex so that they have a unique iniizing point in their definition area. We shall priarily be concerned with the Hessian atrix of penalty functions at the iniizing point. In what follows we shall refer to it as the principal Hessian atrix. The reason will be clear. In a neighbourhood ofthe iniizing point a useful approxiation of a penalty function is given by a quadratic function, with the principal Hessian atrix as the coefficient atrix of the quadratic ter. It is therefore reasonable to assue that unconstrained ini-

3 324 F. A. LOOTSMA izatiori ay be obstructed by ill-conditioning of the principal Hessian atrix. Conditioning of a atrix is easured by the condition nuber, which for syetric, positive definite atrices is defined as the ratio of the greatest to the sallest eigenvalue. We are particularly interested in variations of the condition nuber of the principal Hessian atrix as a function of r in the case where r decreases to O. The condition nuber is also affected by the order of the penalty function. These are, roughly, the points to be discussed in the present paper. 3. The principal Hessian atrix of penalty functions The following analysis will be carried out under the uniqueness conditions forulated in a previous paper 4). They are sufficient for proble (1.1) to have a unique iniu solution x with unique Lagrangian ultipliers Ul>..., Urn. We take \If and \l2f to represent the gradient and the Hessian atrix of J, etc. It will be convenient to define D(x,u) = \l2f(x) - 2: UI\l2gl(X). We think of the constraints as arranged in such a way that gt(.x') = 0; i = 1,..., oe, gt(.x') > 0; i = oe+ 1,...,. Hence, oestands for the nuber of active constraints at the iniu solution x. Let us, first, consider the barrier functions of the type (1.3). The gradient of (1.3) vanishes at its iniizing point x(r) whence \I f[x(r)] - r ~ \lgl[x(r)].f ; gt"[x(r)] = o. (3.1) Let u(r) denote the vector with coponents ul(r), i = 1,...,, defined by We have shown 4) that li ul(r) = UI; r.j.o (3.2) i = 1,...,. (3.3) It.is readily verified that the Hessian atrix of (1.3) evaluated at x(r) can be written as D[x(r), u(r)] + r- 1tv G 1 [x(r), u(r)], (3.4)

4 HESSIAN MATRICES OF PENALTY FUNCTIONS where RI G 1 (X,u) = 11 ~ U /v \1gl(X) \1gl(x)T. The gradients in the above expression are thought of as colun vectors, whereas the index sybol T is used for transposition. The second ter in (3.4) represents a su of atrices of rank 1. A convenient approxiation of (3.4) for sall, positive values of r is given by (3.5) where and F 1 (x,ü) = (~ " G 1 [x(r),u(r)]) dr 1/v,=0 (3.6) Tbe atrices D(x,ü) and G 1 (x,ü) are syetric and positive sei-definite. The uniqueness conditions 4) iposed on proble (1.1) iply that either D(x,ü) or G 1 (x,ü) has rank n. Finally, it can be shown that F 1 (x,ü) is syetric and for any y such!hat G 1 (x,ü) y = O. Fro tbese arguents we obtain that the approxiation (3.5) of the principal Hessian atrix (3.4) is positive definite and syetric for sufficiently sall, positive values of r. The loss function (1.5) can be treated in a siilar way. The gradient of (1.5) vanishes at the iniizing point z(r). For sall values of r this leads to 4) \1f[z(r)] _r- 1 " ~ {-gl[z(r)]}" \1gl[z(r)] = O. (3.7) Taking tl(r) to denote the vector with coponents we have vl(r) = r=: {-gl[z(r)]}"; i = 1,..., a, li vtcr) = ÜI; 'to i = 1,..., ri.. A useful approxiation of the principal Hessian atrix is then given by H 2 = D(x,ü) + F 2 (x,ü) + r- 1/v Gix,ü), (3.8)

5 326 F. A. LOOTSMA where the atrix Fix,ü) is defined in a siilar way as Fl(x,ü), and ex G 2 (x,ü) = v ~ (Ü,)l-l/v \1g,(x) \1g,(x)T., (3.9) There is clearly a striking siilarity in the for of HI defined by (3.5) and H 2 of (3.8). 4. Eigenvalues and condition nuber of the principal Hessian atrix It will be convenient to confine our attention in this section to a atrix of the for 1 D+F+-G, (4.1) P where D is a positive definite atrix of rank n, G a positive sei-definite atrix of rank (X < n, and F such that yt F y = 0 for all y satisfying G y = O. The paraeter p is a positive controlling paraeter. The eigenvalues of (4.1) are not affected by a coordinate transforation. Let us therefore transfor D, F and G into atrices D', F' and G', respectively, in such a way that G' is a diagonal atrix. Then (4.1) reduces to (4.2) Here Gu' is a diagonal atrix with (X rows and (X coluns, and positive diagonal eleents ru',..., rex.." The partitioning of D' and F' is siilar to that of G' so that DIl' has (X rows and (X coluns, etc. For sall values of p the eigenvalues of (4.2), and consequently the eigenvalues of (4.1), are given by Át ~p-lrtl'; Át ~ f-l,'; i = 1,..., oe, i = (X + 1,..., n, where f-lex+/,..., f-ln' denote the eigenvalues of D 22 '. The proof rests on the theores of Gerschgorin (see Wilkinson 7), p. 71 ff.). Suppose that and For sall values of p the condition nuber of (4.1), here the ratio of the greatest to the sallest eigenvalue, is given by -1 ru' p =rf-ln

6 HESSIAN MATRICES OF PENALTY FUNCTIONS 327 SO that the condition nuber is proportional to p-l. Finally, the deterinant of (4.1) varies with p-a. Substituting p = r+!" one obtains the results of the next section. We ay note that the condition nuber converges to the finite value Yll' (as PtO) in the particular case that oe= n (if, for exaple, the proble is one of linear prograing). 5. Coparison of first-order and higher-order penalty functions The penalty functions of order 'JI have a principal Hessian atrix with a condition nuber proportional to r- 1/v for sall values of r. This iplies that the first-order penalty functions (the logarithic barrier function and the quadratic loss function) tend to becoe ill-conditioned ore rapidly than any of the higher-order penalty functions. The inverse of the principal Hessian atrix, which plays a doinant part in the successful algorith of Davidon, Fletcher and Powell 3) as well as in the variable-etric or quasi-newton ethods 6) derived fro it, tends to becoe singular as r t O. For sall values of r its deterinant is proportional to rt: Here again, the first-order penalty functions have a disadvantage with respect to higher-order penalty functions. On the other hand, first-order penalty-function techniques converge faster than the higher-order ones. Generally, we have by the rule of De I'Hêpital that. f [x(r)] - f(x) Ii = 1. (5.1) r~o a ~ ÜI gl[x(r)] Using (3.2) and (3.3) for the barrier functions we are led to Working with a loss function we obtain in a siilar way that a f[z(r)]-f(x) ~ _rl/v~ (ÜI)I+1/V. Hence, a variation of the condition nuber proportional to r- 1 / v is copensated by a rate of convergence varying with r'!". In other words, it is doubtful whether a higher-order penalty-function technique will give a substantial iproveent in coparison with a first-order technique.

7 328 F. A. LOOTSMA 6. Balancing the constraints The condition nuber of a principal Hessian atrix depends alost entirely on the greatest eigenvalue of (3.6), if a barrier function is eployed, or (3.9), if we are dealing with a loss function. It would apparently be a nuisance if this greatest eigenvalue differs considerably fro the reaining eigenvalues of these atrices. The question arises whether we can balance the (active) constraints in such a way that the positive eigenvalues of (3.6) or (3.9) are of the sae order of agnitude. We confine ourselves to first-order penalty functions. The atrices (3.6) and (3.9) are then reduced to and Cl ~ (ÜI)2 V'gl(x) 'Vgl(x)T (6.1) Cl ~ V'gl(x) V'gl(x)T (6.2) respectively. Let us first consider the case that the gradients V'gl(X), i = 1,..., ()(,are orthogonal. This is not the general case, but the orthogonality hypothesis leads to soe understanding of the causes of ibalance. It follows easily that (6.1) has a positive eigenvalues given by (ü,liv'g,(x)11)2; i = 1,..., a, (6.3) Using the Kuhn-Tucker relations and the orthogonality relations one can readily show that _ {J111V'j(x)11 UI = ; i = 1,..., ()(, IIV'g,(x) II where (JI is the cosine of the angle between 'Vj(x) and V'gl(x). The positive eigenvalues of( 6.1) are accordingly given by ({JIIIV'j(x)IIY; i = 1,..., a, (6.4) which deonstrates that ibalance is only due to the angles between the gradients of the objective function and the active constraints. In a siilar way we obtain that the positive eigenvalues of (6.2) can be written as II'Vgl(x)W; i = 1,..., ()(. (6.5) Ibalance of the active constraints is here only produced by differences in the lengths of gradients. Hence, a proble ay be unbalanced with respect to the logarithic barrier function, but balanced with respect to the quadratic loss function, and vice versa..

8 HESSIAN MATRICES OF PENALTY FUNCTIONS 329 The balance of the active constraints function is taken to be is restored if the logarithic barrier I(x) - r ~ PI In g,(x), (6.6) and if the weight factors P, are chosen by (Ü,JJ\7g,(X)JI)-2; i = 1,...,0(, PI= {. o or 1; l = 0( + 1,..., n. (6.7) Then (6.1) is reduced to a atrix of the for '" ~ ala?, (6.8) where the vectors al>..., a", are orthogonal and noralized to length 1. The quantities appearing in the right-hand side of (6.7) are not available at the beginning of the coputational process for solving (1.1). Convenient approxiations, however, are gradually obtained at successive r-inia. One could start with the barrier function (6.6) and with P, = 1 for all i = 1,..., n. Ibalance of the constraints is then indicated by differences in the quantities (u,(r) JJ\7g,[x(r)]JI)2; 1,..., 0(, (6.9) for sall values of r. Using (6.9) to odify the weight factors one can restart the coputations with an iproved equilibriu. A siilar device can of course be eployed in the case that the quadratic loss function is used. Let us finally depart fro the orthogonality hypothesis, and consider what happens if the weight factors (6.7) are eployed in the case that the gradients \7g,(x), i = 1,..., et, are independent but not necessarily orthogonal. Ibalance is still governed by a atrix ofthe for (6.8) with noralized vectors al'..., a", of length 1. This atrix can also be written as AAT where A is the atrix with coluns al'..., a",. The positive eigenvalues of AAT are given by the eigenvalues of the atrix ATA, which has, as a consequence of the noralization, diagonal eleents equal to 1. Hence, the positive eigenvalues of AA T su up «ae. Orthogonality of the vectors al'..., a", iplies that the eigenvalues are toll qual to 1. The continuity of the eigenvalues ensures that sall perturbations of the orthogonality will only lead to sall deviations of the eigenvalues fro unity. The above device of choosing weight factors can therefore frequently be used in order to balance the constraints, even in the case that the gradients of the active constraints are not orthogonal.

9 330 F. A. LOOTSMA Acknowledgeent The author wishes to thank Prof. Dr J. F. Benders (Technological University, Eindhoven) and Dr J. D. Pearson (Inforation Systes and Autoation) for the inspiring discussions and criticiss. REFERENCES Eindhoven, May ) A. V. Fiacco and G. P. McCorick, Nonlinearprograing, Sequential unconstrained iniization techniques, Wiley, New York, ) R. Fletcher and A. P. McCann, Acceleration techniques for non-linear prograing, Paper presented at the conference on optiization, Keele, ) R. Fletcher and M. J. D. Powell, The Coputer Journal 6, , ) F. A. Lootsa, Philips Res. Repts. 23, , ) W. Murr ay, Ill-conditioning in barrier and penalty functions arising in constrained nonlinear prograing, Paper presented at the sixth int. syp. on ath. prograing, Princeton, N.J., ) J. D. Pearson, The Coputer Journal12, , ) J. H. Wilkinson, The algebraic eigenvalue proble, Clarendon Press, Oxford, ) W. I. Zangwill, Manageent Science 13, , 1967.

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

The Methods of Solution for Constrained Nonlinear Programming

The Methods of Solution for Constrained Nonlinear Programming Research Inventy: International Journal Of Engineering And Science Vol.4, Issue 3(March 2014), PP 01-06 Issn (e): 2278-4721, Issn (p):2319-6483, www.researchinventy.co The Methods of Solution for Constrained

More information

Page 1 Lab 1 Elementary Matrix and Linear Algebra Spring 2011

Page 1 Lab 1 Elementary Matrix and Linear Algebra Spring 2011 Page Lab Eleentary Matri and Linear Algebra Spring 0 Nae Due /03/0 Score /5 Probles through 4 are each worth 4 points.. Go to the Linear Algebra oolkit site ransforing a atri to reduced row echelon for

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fitting of Data David Eberly, Geoetric Tools, Redond WA 98052 https://www.geoetrictools.co/ This work is licensed under the Creative Coons Attribution 4.0 International License. To view a

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

ON THE TWO-LEVEL PRECONDITIONING IN LEAST SQUARES METHOD

ON THE TWO-LEVEL PRECONDITIONING IN LEAST SQUARES METHOD PROCEEDINGS OF THE YEREVAN STATE UNIVERSITY Physical and Matheatical Sciences 04,, p. 7 5 ON THE TWO-LEVEL PRECONDITIONING IN LEAST SQUARES METHOD M a t h e a t i c s Yu. A. HAKOPIAN, R. Z. HOVHANNISYAN

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

A new type of lower bound for the largest eigenvalue of a symmetric matrix

A new type of lower bound for the largest eigenvalue of a symmetric matrix Linear Algebra and its Applications 47 7 9 9 www.elsevier.co/locate/laa A new type of lower bound for the largest eigenvalue of a syetric atrix Piet Van Mieghe Delft University of Technology, P.O. Box

More information

Introduction to Optimization Techniques. Nonlinear Programming

Introduction to Optimization Techniques. Nonlinear Programming Introduction to Optiization echniques Nonlinear Prograing Optial Solutions Consider the optiization proble in f ( x) where F R n xf Definition : x F is optial (global iniu) for this proble, if f( x ) f(

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

INNER CONSTRAINTS FOR A 3-D SURVEY NETWORK

INNER CONSTRAINTS FOR A 3-D SURVEY NETWORK eospatial Science INNER CONSRAINS FOR A 3-D SURVEY NEWORK hese notes follow closely the developent of inner constraint equations by Dr Willie an, Departent of Building, School of Design and Environent,

More information

Polygonal Designs: Existence and Construction

Polygonal Designs: Existence and Construction Polygonal Designs: Existence and Construction John Hegean Departent of Matheatics, Stanford University, Stanford, CA 9405 Jeff Langford Departent of Matheatics, Drake University, Des Moines, IA 5011 G

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

arxiv: v1 [math.na] 10 Oct 2016

arxiv: v1 [math.na] 10 Oct 2016 GREEDY GAUSS-NEWTON ALGORITHM FOR FINDING SPARSE SOLUTIONS TO NONLINEAR UNDERDETERMINED SYSTEMS OF EQUATIONS MÅRTEN GULLIKSSON AND ANNA OLEYNIK arxiv:6.395v [ath.na] Oct 26 Abstract. We consider the proble

More information

Hybrid System Identification: An SDP Approach

Hybrid System Identification: An SDP Approach 49th IEEE Conference on Decision and Control Deceber 15-17, 2010 Hilton Atlanta Hotel, Atlanta, GA, USA Hybrid Syste Identification: An SDP Approach C Feng, C M Lagoa, N Ozay and M Sznaier Abstract The

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

An Algorithm for Posynomial Geometric Programming, Based on Generalized Linear Programming

An Algorithm for Posynomial Geometric Programming, Based on Generalized Linear Programming An Algorith for Posynoial Geoetric Prograing, Based on Generalized Linear Prograing Jayant Rajgopal Departent of Industrial Engineering University of Pittsburgh, Pittsburgh, PA 526 Dennis L. Bricer Departent

More information

CHAPTER 8 CONSTRAINED OPTIMIZATION 2: SEQUENTIAL QUADRATIC PROGRAMMING, INTERIOR POINT AND GENERALIZED REDUCED GRADIENT METHODS

CHAPTER 8 CONSTRAINED OPTIMIZATION 2: SEQUENTIAL QUADRATIC PROGRAMMING, INTERIOR POINT AND GENERALIZED REDUCED GRADIENT METHODS CHAPER 8 CONSRAINED OPIMIZAION : SEQUENIAL QUADRAIC PROGRAMMING, INERIOR POIN AND GENERALIZED REDUCED GRADIEN MEHODS 8. Introduction In the previous chapter we eained the necessary and sufficient conditions

More information

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters

The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Parameters journal of ultivariate analysis 58, 96106 (1996) article no. 0041 The Distribution of the Covariance Matrix for a Subset of Elliptical Distributions with Extension to Two Kurtosis Paraeters H. S. Steyn

More information

Lecture 21. Interior Point Methods Setup and Algorithm

Lecture 21. Interior Point Methods Setup and Algorithm Lecture 21 Interior Point Methods In 1984, Kararkar introduced a new weakly polynoial tie algorith for solving LPs [Kar84a], [Kar84b]. His algorith was theoretically faster than the ellipsoid ethod and

More information

Topic 5a Introduction to Curve Fitting & Linear Regression

Topic 5a Introduction to Curve Fitting & Linear Regression /7/08 Course Instructor Dr. Rayond C. Rup Oice: A 337 Phone: (95) 747 6958 E ail: rcrup@utep.edu opic 5a Introduction to Curve Fitting & Linear Regression EE 4386/530 Coputational ethods in EE Outline

More information

Physics 215 Winter The Density Matrix

Physics 215 Winter The Density Matrix Physics 215 Winter 2018 The Density Matrix The quantu space of states is a Hilbert space H. Any state vector ψ H is a pure state. Since any linear cobination of eleents of H are also an eleent of H, it

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup)

Recovering Data from Underdetermined Quadratic Measurements (CS 229a Project: Final Writeup) Recovering Data fro Underdeterined Quadratic Measureents (CS 229a Project: Final Writeup) Mahdi Soltanolkotabi Deceber 16, 2011 1 Introduction Data that arises fro engineering applications often contains

More information

RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS

RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS BIT Nuerical Matheatics 43: 459 466, 2003. 2003 Kluwer Acadeic Publishers. Printed in The Netherlands 459 RESTARTED FULL ORTHOGONALIZATION METHOD FOR SHIFTED LINEAR SYSTEMS V. SIMONCINI Dipartiento di

More information

Interactive Markov Models of Evolutionary Algorithms

Interactive Markov Models of Evolutionary Algorithms Cleveland State University EngagedScholarship@CSU Electrical Engineering & Coputer Science Faculty Publications Electrical Engineering & Coputer Science Departent 2015 Interactive Markov Models of Evolutionary

More information

paper prepared for the 1996 PTRC Conference, September 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL

paper prepared for the 1996 PTRC Conference, September 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL paper prepared for the 1996 PTRC Conference, Septeber 2-6, Brunel University, UK ON THE CALIBRATION OF THE GRAVITY MODEL Nanne J. van der Zijpp 1 Transportation and Traffic Engineering Section Delft University

More information

Anisotropic reference media and the possible linearized approximations for phase velocities of qs waves in weakly anisotropic media

Anisotropic reference media and the possible linearized approximations for phase velocities of qs waves in weakly anisotropic media INSTITUTE OF PHYSICS PUBLISHING JOURNAL OF PHYSICS D: APPLIED PHYSICS J. Phys. D: Appl. Phys. 5 00 007 04 PII: S00-770867-6 Anisotropic reference edia and the possible linearized approxiations for phase

More information

Chaotic Coupled Map Lattices

Chaotic Coupled Map Lattices Chaotic Coupled Map Lattices Author: Dustin Keys Advisors: Dr. Robert Indik, Dr. Kevin Lin 1 Introduction When a syste of chaotic aps is coupled in a way that allows the to share inforation about each

More information

Supporting Information for Supression of Auger Processes in Confined Structures

Supporting Information for Supression of Auger Processes in Confined Structures Supporting Inforation for Supression of Auger Processes in Confined Structures George E. Cragg and Alexander. Efros Naval Research aboratory, Washington, DC 20375, USA 1 Solution of the Coupled, Two-band

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

1 Bounding the Margin

1 Bounding the Margin COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #12 Scribe: Jian Min Si March 14, 2013 1 Bounding the Margin We are continuing the proof of a bound on the generalization error of AdaBoost

More information

RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS MEMBRANE

RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS MEMBRANE Proceedings of ICIPE rd International Conference on Inverse Probles in Engineering: Theory and Practice June -8, 999, Port Ludlow, Washington, USA : RECOVERY OF A DENSITY FROM THE EIGENVALUES OF A NONHOMOGENEOUS

More information

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving

More information

A1. Find all ordered pairs (a, b) of positive integers for which 1 a + 1 b = 3

A1. Find all ordered pairs (a, b) of positive integers for which 1 a + 1 b = 3 A. Find all ordered pairs a, b) of positive integers for which a + b = 3 08. Answer. The six ordered pairs are 009, 08), 08, 009), 009 337, 674) = 35043, 674), 009 346, 673) = 3584, 673), 674, 009 337)

More information

Variations on Backpropagation

Variations on Backpropagation 2 Variations on Backpropagation 2 Variations Heuristic Modifications Moentu Variable Learning Rate Standard Nuerical Optiization Conjugate Gradient Newton s Method (Levenberg-Marquardt) 2 2 Perforance

More information

On Nonlinear Controllability of Homogeneous Systems Linear in Control

On Nonlinear Controllability of Homogeneous Systems Linear in Control IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 48, NO. 1, JANUARY 2003 139 On Nonlinear Controllability of Hoogeneous Systes Linear in Control Jaes Melody, Taer Başar, and Francesco Bullo Abstract This work

More information

which together show that the Lax-Milgram lemma can be applied. (c) We have the basic Galerkin orthogonality

which together show that the Lax-Milgram lemma can be applied. (c) We have the basic Galerkin orthogonality UPPSALA UNIVERSITY Departent of Inforation Technology Division of Scientific Coputing Solutions to exa in Finite eleent ethods II 14-6-5 Stefan Engblo, Daniel Elfverson Question 1 Note: a inus sign in

More information

On Conditions for Linearity of Optimal Estimation

On Conditions for Linearity of Optimal Estimation On Conditions for Linearity of Optial Estiation Erah Akyol, Kuar Viswanatha and Kenneth Rose {eakyol, kuar, rose}@ece.ucsb.edu Departent of Electrical and Coputer Engineering University of California at

More information

Support Vector Machines MIT Course Notes Cynthia Rudin

Support Vector Machines MIT Course Notes Cynthia Rudin Support Vector Machines MIT 5.097 Course Notes Cynthia Rudin Credit: Ng, Hastie, Tibshirani, Friedan Thanks: Şeyda Ertekin Let s start with soe intuition about argins. The argin of an exaple x i = distance

More information

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS

DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS ISSN 1440-771X AUSTRALIA DEPARTMENT OF ECONOMETRICS AND BUSINESS STATISTICS An Iproved Method for Bandwidth Selection When Estiating ROC Curves Peter G Hall and Rob J Hyndan Working Paper 11/00 An iproved

More information

Boosting with log-loss

Boosting with log-loss Boosting with log-loss Marco Cusuano-Towner Septeber 2, 202 The proble Suppose we have data exaples {x i, y i ) i =... } for a two-class proble with y i {, }. Let F x) be the predictor function with the

More information

COMPONENT MODE SYNTHESIS, FIXED-INTERFACE MODEL Revision A

COMPONENT MODE SYNTHESIS, FIXED-INTERFACE MODEL Revision A COMPONEN MODE SYNHESS, FXED-NERFACE MODEL Revision A By o rvine Eail: toirvine@aol.co February, ntroduction Coponent ode synthesis is a ethod for analyzing the dynaic behavior of a syste consisting of

More information

Deflation of the I-O Series Some Technical Aspects. Giorgio Rampa University of Genoa April 2007

Deflation of the I-O Series Some Technical Aspects. Giorgio Rampa University of Genoa April 2007 Deflation of the I-O Series 1959-2. Soe Technical Aspects Giorgio Rapa University of Genoa g.rapa@unige.it April 27 1. Introduction The nuber of sectors is 42 for the period 1965-2 and 38 for the initial

More information

Neural Network Learning as an Inverse Problem

Neural Network Learning as an Inverse Problem Neural Network Learning as an Inverse Proble VĚRA KU RKOVÁ, Institute of Coputer Science, Acadey of Sciences of the Czech Republic, Pod Vodárenskou věží 2, 182 07 Prague 8, Czech Republic. Eail: vera@cs.cas.cz

More information

Solutions 1. Introduction to Coding Theory - Spring 2010 Solutions 1. Exercise 1.1. See Examples 1.2 and 1.11 in the course notes.

Solutions 1. Introduction to Coding Theory - Spring 2010 Solutions 1. Exercise 1.1. See Examples 1.2 and 1.11 in the course notes. Solutions 1 Exercise 1.1. See Exaples 1.2 and 1.11 in the course notes. Exercise 1.2. Observe that the Haing distance of two vectors is the iniu nuber of bit flips required to transfor one into the other.

More information

A model reduction approach to numerical inversion for a parabolic partial differential equation

A model reduction approach to numerical inversion for a parabolic partial differential equation Inverse Probles Inverse Probles 30 (204) 250 (33pp) doi:0.088/0266-56/30/2/250 A odel reduction approach to nuerical inversion for a parabolic partial differential equation Liliana Borcea, Vladiir Drusin

More information

Convolutional Codes. Lecture Notes 8: Trellis Codes. Example: K=3,M=2, rate 1/2 code. Figure 95: Convolutional Encoder

Convolutional Codes. Lecture Notes 8: Trellis Codes. Example: K=3,M=2, rate 1/2 code. Figure 95: Convolutional Encoder Convolutional Codes Lecture Notes 8: Trellis Codes In this lecture we discuss construction of signals via a trellis. That is, signals are constructed by labeling the branches of an infinite trellis with

More information

On Constant Power Water-filling

On Constant Power Water-filling On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

On the Use of A Priori Information for Sparse Signal Approximations

On the Use of A Priori Information for Sparse Signal Approximations ITS TECHNICAL REPORT NO. 3/4 On the Use of A Priori Inforation for Sparse Signal Approxiations Oscar Divorra Escoda, Lorenzo Granai and Pierre Vandergheynst Signal Processing Institute ITS) Ecole Polytechnique

More information

Saddle Points in Random Matrices: Analysis of Knuth Search Algorithms

Saddle Points in Random Matrices: Analysis of Knuth Search Algorithms Saddle Points in Rando Matrices: Analysis of Knuth Search Algoriths Micha Hofri Philippe Jacquet Dept. of Coputer Science INRIA, Doaine de Voluceau - Rice University Rocquencourt - B.P. 05 Houston TX 77005

More information

Asynchronous Gossip Algorithms for Stochastic Optimization

Asynchronous Gossip Algorithms for Stochastic Optimization Asynchronous Gossip Algoriths for Stochastic Optiization S. Sundhar Ra ECE Dept. University of Illinois Urbana, IL 680 ssrini@illinois.edu A. Nedić IESE Dept. University of Illinois Urbana, IL 680 angelia@illinois.edu

More information

Multivariate Methods. Matlab Example. Principal Components Analysis -- PCA

Multivariate Methods. Matlab Example. Principal Components Analysis -- PCA Multivariate Methos Xiaoun Qi Principal Coponents Analysis -- PCA he PCA etho generates a new set of variables, calle principal coponents Each principal coponent is a linear cobination of the original

More information

THE WEIGHTING METHOD AND MULTIOBJECTIVE PROGRAMMING UNDER NEW CONCEPTS OF GENERALIZED (, )-INVEXITY

THE WEIGHTING METHOD AND MULTIOBJECTIVE PROGRAMMING UNDER NEW CONCEPTS OF GENERALIZED (, )-INVEXITY U.P.B. Sci. Bull., Series A, Vol. 80, Iss. 2, 2018 ISSN 1223-7027 THE WEIGHTING METHOD AND MULTIOBJECTIVE PROGRAMMING UNDER NEW CONCEPTS OF GENERALIZED (, )-INVEXITY Tadeusz ANTCZAK 1, Manuel ARANA-JIMÉNEZ

More information

Kinetic Theory of Gases: Elementary Ideas

Kinetic Theory of Gases: Elementary Ideas Kinetic Theory of Gases: Eleentary Ideas 17th February 2010 1 Kinetic Theory: A Discussion Based on a Siplified iew of the Motion of Gases 1.1 Pressure: Consul Engel and Reid Ch. 33.1) for a discussion

More information

Lecture 9 November 23, 2015

Lecture 9 November 23, 2015 CSC244: Discrepancy Theory in Coputer Science Fall 25 Aleksandar Nikolov Lecture 9 Noveber 23, 25 Scribe: Nick Spooner Properties of γ 2 Recall that γ 2 (A) is defined for A R n as follows: γ 2 (A) = in{r(u)

More information

Introduction to Robotics (CS223A) (Winter 2006/2007) Homework #5 solutions

Introduction to Robotics (CS223A) (Winter 2006/2007) Homework #5 solutions Introduction to Robotics (CS3A) Handout (Winter 6/7) Hoework #5 solutions. (a) Derive a forula that transfors an inertia tensor given in soe frae {C} into a new frae {A}. The frae {A} can differ fro frae

More information

OPTIMIZATION in multi-agent networks has attracted

OPTIMIZATION in multi-agent networks has attracted Distributed constrained optiization and consensus in uncertain networks via proxial iniization Kostas Margellos, Alessandro Falsone, Sione Garatti and Maria Prandini arxiv:603.039v3 [ath.oc] 3 May 07 Abstract

More information

Solving initial value problems by residual power series method

Solving initial value problems by residual power series method Theoretical Matheatics & Applications, vol.3, no.1, 13, 199-1 ISSN: 179-9687 (print), 179-979 (online) Scienpress Ltd, 13 Solving initial value probles by residual power series ethod Mohaed H. Al-Sadi

More information

The linear sampling method and the MUSIC algorithm

The linear sampling method and the MUSIC algorithm INSTITUTE OF PHYSICS PUBLISHING INVERSE PROBLEMS Inverse Probles 17 (2001) 591 595 www.iop.org/journals/ip PII: S0266-5611(01)16989-3 The linear sapling ethod and the MUSIC algorith Margaret Cheney Departent

More information

Uniform Approximation and Bernstein Polynomials with Coefficients in the Unit Interval

Uniform Approximation and Bernstein Polynomials with Coefficients in the Unit Interval Unifor Approxiation and Bernstein Polynoials with Coefficients in the Unit Interval Weiang Qian and Marc D. Riedel Electrical and Coputer Engineering, University of Minnesota 200 Union St. S.E. Minneapolis,

More information

Exact tensor completion with sum-of-squares

Exact tensor completion with sum-of-squares Proceedings of Machine Learning Research vol 65:1 54, 2017 30th Annual Conference on Learning Theory Exact tensor copletion with su-of-squares Aaron Potechin Institute for Advanced Study, Princeton David

More information

A New Class of APEX-Like PCA Algorithms

A New Class of APEX-Like PCA Algorithms Reprinted fro Proceedings of ISCAS-98, IEEE Int. Syposiu on Circuit and Systes, Monterey (USA), June 1998 A New Class of APEX-Like PCA Algoriths Sione Fiori, Aurelio Uncini, Francesco Piazza Dipartiento

More information

by F. A. LOOTSMA Throughout this paper we shall be dealing with a number of methods for solving the constrained-minimization problem:

by F. A. LOOTSMA Throughout this paper we shall be dealing with a number of methods for solving the constrained-minimization problem: R 676 Philips Res. Repts 23, 408-423, 1968 CONSTRAINED VIA PENALTY OPTIMIZATION FUNCTIONS by F. A. LOOTSMA Abstract 1. Introduetion This paper is concerned with a generalized combination ofthe interiorpoint

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Curious Bounds for Floor Function Sums

Curious Bounds for Floor Function Sums 1 47 6 11 Journal of Integer Sequences, Vol. 1 (018), Article 18.1.8 Curious Bounds for Floor Function Sus Thotsaporn Thanatipanonda and Elaine Wong 1 Science Division Mahidol University International

More information

Numerical issues in the implementation of high order polynomial multidomain penalty spectral Galerkin methods for hyperbolic conservation laws

Numerical issues in the implementation of high order polynomial multidomain penalty spectral Galerkin methods for hyperbolic conservation laws Nuerical issues in the ipleentation of high order polynoial ultidoain penalty spectral Galerkin ethods for hyperbolic conservation laws Sigal Gottlieb 1 and Jae-Hun Jung 1, 1 Departent of Matheatics, University

More information

Kinetic Theory of Gases: Elementary Ideas

Kinetic Theory of Gases: Elementary Ideas Kinetic Theory of Gases: Eleentary Ideas 9th February 011 1 Kinetic Theory: A Discussion Based on a Siplified iew of the Motion of Gases 1.1 Pressure: Consul Engel and Reid Ch. 33.1) for a discussion of

More information

Tight Bounds for Maximal Identifiability of Failure Nodes in Boolean Network Tomography

Tight Bounds for Maximal Identifiability of Failure Nodes in Boolean Network Tomography Tight Bounds for axial Identifiability of Failure Nodes in Boolean Network Toography Nicola Galesi Sapienza Università di Roa nicola.galesi@uniroa1.it Fariba Ranjbar Sapienza Università di Roa fariba.ranjbar@uniroa1.it

More information

Lecture 13 Eigenvalue Problems

Lecture 13 Eigenvalue Problems Lecture 13 Eigenvalue Probles MIT 18.335J / 6.337J Introduction to Nuerical Methods Per-Olof Persson October 24, 2006 1 The Eigenvalue Decoposition Eigenvalue proble for atrix A: Ax = λx with eigenvalues

More information

Comparison of Stability of Selected Numerical Methods for Solving Stiff Semi- Linear Differential Equations

Comparison of Stability of Selected Numerical Methods for Solving Stiff Semi- Linear Differential Equations International Journal of Applied Science and Technology Vol. 7, No. 3, Septeber 217 Coparison of Stability of Selected Nuerical Methods for Solving Stiff Sei- Linear Differential Equations Kwaku Darkwah

More information

Optimal quantum detectors for unambiguous detection of mixed states

Optimal quantum detectors for unambiguous detection of mixed states PHYSICAL REVIEW A 69, 06318 (004) Optial quantu detectors for unabiguous detection of ixed states Yonina C. Eldar* Departent of Electrical Engineering, Technion Israel Institute of Technology, Haifa 3000,

More information

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science A Better Algorith For an Ancient Scheduling Proble David R. Karger Steven J. Phillips Eric Torng Departent of Coputer Science Stanford University Stanford, CA 9435-4 Abstract One of the oldest and siplest

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

Decentralized Adaptive Control of Nonlinear Systems Using Radial Basis Neural Networks

Decentralized Adaptive Control of Nonlinear Systems Using Radial Basis Neural Networks 050 IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 44, NO., NOVEMBER 999 Decentralized Adaptive Control of Nonlinear Systes Using Radial Basis Neural Networks Jeffrey T. Spooner and Kevin M. Passino Abstract

More information

PROXSCAL. Notation. W n n matrix with weights for source k. E n s matrix with raw independent variables F n p matrix with fixed coordinates

PROXSCAL. Notation. W n n matrix with weights for source k. E n s matrix with raw independent variables F n p matrix with fixed coordinates PROXSCAL PROXSCAL perfors ultidiensional scaling of proxiity data to find a leastsquares representation of the obects in a low-diensional space. Individual differences odels can be specified for ultiple

More information

Explicit solution of the polynomial least-squares approximation problem on Chebyshev extrema nodes

Explicit solution of the polynomial least-squares approximation problem on Chebyshev extrema nodes Explicit solution of the polynoial least-squares approxiation proble on Chebyshev extrea nodes Alfredo Eisinberg, Giuseppe Fedele Dipartiento di Elettronica Inforatica e Sisteistica, Università degli Studi

More information

(t, m, s)-nets and Maximized Minimum Distance, Part II

(t, m, s)-nets and Maximized Minimum Distance, Part II (t,, s)-nets and Maxiized Miniu Distance, Part II Leonhard Grünschloß and Alexander Keller Abstract The quality paraeter t of (t,, s)-nets controls extensive stratification properties of the generated

More information

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x)

Ştefan ŞTEFĂNESCU * is the minimum global value for the function h (x) 7Applying Nelder Mead s Optiization Algorith APPLYING NELDER MEAD S OPTIMIZATION ALGORITHM FOR MULTIPLE GLOBAL MINIMA Abstract Ştefan ŞTEFĂNESCU * The iterative deterinistic optiization ethod could not

More information

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm Acta Polytechnica Hungarica Vol., No., 04 Sybolic Analysis as Universal Tool for Deriving Properties of Non-linear Algoriths Case study of EM Algorith Vladiir Mladenović, Miroslav Lutovac, Dana Porrat

More information

Principal Components Analysis

Principal Components Analysis Principal Coponents Analysis Cheng Li, Bingyu Wang Noveber 3, 204 What s PCA Principal coponent analysis (PCA) is a statistical procedure that uses an orthogonal transforation to convert a set of observations

More information

Genetic Quantum Algorithm and its Application to Combinatorial Optimization Problem

Genetic Quantum Algorithm and its Application to Combinatorial Optimization Problem Genetic Quantu Algorith and its Application to Cobinatorial Optiization Proble Kuk-Hyun Han Dept. of Electrical Engineering, KAIST, 373-, Kusong-dong Yusong-gu Taejon, 305-70, Republic of Korea khhan@vivaldi.kaist.ac.kr

More information

Constrained Consensus and Optimization in Multi-Agent Networks arxiv: v2 [math.oc] 17 Dec 2008

Constrained Consensus and Optimization in Multi-Agent Networks arxiv: v2 [math.oc] 17 Dec 2008 LIDS Report 2779 1 Constrained Consensus and Optiization in Multi-Agent Networks arxiv:0802.3922v2 [ath.oc] 17 Dec 2008 Angelia Nedić, Asuan Ozdaglar, and Pablo A. Parrilo February 15, 2013 Abstract We

More information

OBJECTIVES INTRODUCTION

OBJECTIVES INTRODUCTION M7 Chapter 3 Section 1 OBJECTIVES Suarize data using easures of central tendency, such as the ean, edian, ode, and idrange. Describe data using the easures of variation, such as the range, variance, and

More information

THE KALMAN FILTER: A LOOK BEHIND THE SCENE

THE KALMAN FILTER: A LOOK BEHIND THE SCENE HE KALMA FILER: A LOOK BEHID HE SCEE R.E. Deain School of Matheatical and Geospatial Sciences, RMI University eail: rod.deain@rit.edu.au Presented at the Victorian Regional Survey Conference, Mildura,

More information

Tail estimates for norms of sums of log-concave random vectors

Tail estimates for norms of sums of log-concave random vectors Tail estiates for nors of sus of log-concave rando vectors Rados law Adaczak Rafa l Lata la Alexander E. Litvak Alain Pajor Nicole Toczak-Jaegerann Abstract We establish new tail estiates for order statistics

More information

Ph 20.3 Numerical Solution of Ordinary Differential Equations

Ph 20.3 Numerical Solution of Ordinary Differential Equations Ph 20.3 Nuerical Solution of Ordinary Differential Equations Due: Week 5 -v20170314- This Assignent So far, your assignents have tried to failiarize you with the hardware and software in the Physics Coputing

More information

Multi-Dimensional Hegselmann-Krause Dynamics

Multi-Dimensional Hegselmann-Krause Dynamics Multi-Diensional Hegselann-Krause Dynaics A. Nedić Industrial and Enterprise Systes Engineering Dept. University of Illinois Urbana, IL 680 angelia@illinois.edu B. Touri Coordinated Science Laboratory

More information

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili,

Generalized AOR Method for Solving System of Linear Equations. Davod Khojasteh Salkuyeh. Department of Mathematics, University of Mohaghegh Ardabili, Australian Journal of Basic and Applied Sciences, 5(3): 35-358, 20 ISSN 99-878 Generalized AOR Method for Solving Syste of Linear Equations Davod Khojasteh Salkuyeh Departent of Matheatics, University

More information

Lecture 20 November 7, 2013

Lecture 20 November 7, 2013 CS 229r: Algoriths for Big Data Fall 2013 Prof. Jelani Nelson Lecture 20 Noveber 7, 2013 Scribe: Yun Willia Yu 1 Introduction Today we re going to go through the analysis of atrix copletion. First though,

More information

Necessity of low effective dimension

Necessity of low effective dimension Necessity of low effective diension Art B. Owen Stanford University October 2002, Orig: July 2002 Abstract Practitioners have long noticed that quasi-monte Carlo ethods work very well on functions that

More information

SOLUTIONS. PROBLEM 1. The Hamiltonian of the particle in the gravitational field can be written as, x 0, + U(x), U(x) =

SOLUTIONS. PROBLEM 1. The Hamiltonian of the particle in the gravitational field can be written as, x 0, + U(x), U(x) = SOLUTIONS PROBLEM 1. The Hailtonian of the particle in the gravitational field can be written as { Ĥ = ˆp2, x 0, + U(x), U(x) = (1) 2 gx, x > 0. The siplest estiate coes fro the uncertainty relation. If

More information

Lower Bounds for Quantized Matrix Completion

Lower Bounds for Quantized Matrix Completion Lower Bounds for Quantized Matrix Copletion Mary Wootters and Yaniv Plan Departent of Matheatics University of Michigan Ann Arbor, MI Eail: wootters, yplan}@uich.edu Mark A. Davenport School of Elec. &

More information

Moving Least-Squares: A Numerical Differentiation Method for Irregularly Spaced Calculation Points

Moving Least-Squares: A Numerical Differentiation Method for Irregularly Spaced Calculation Points SANDIA EPOT SAND001-1669 Unliited elease Printed June 001 Moving Least-Squares: A Nuerical Differentiation Method for Irregularly Spaced Calculation Points Albert Gossler Prepared by Sandia National Laboratories

More information

DESIGN OF THE DIE PROFILE FOR THE INCREMENTAL RADIAL FORGING PROCESS *

DESIGN OF THE DIE PROFILE FOR THE INCREMENTAL RADIAL FORGING PROCESS * IJST, Transactions of Mechanical Engineering, Vol. 39, No. M1, pp 89-100 Printed in The Islaic Republic of Iran, 2015 Shira University DESIGN OF THE DIE PROFILE FOR THE INCREMENTAL RADIAL FORGING PROCESS

More information

ma x = -bv x + F rod.

ma x = -bv x + F rod. Notes on Dynaical Systes Dynaics is the study of change. The priary ingredients of a dynaical syste are its state and its rule of change (also soeties called the dynaic). Dynaical systes can be continuous

More information