Near-Isometry by Relaxation: Supplement
|
|
- Gilbert Nigel Short
- 5 years ago
- Views:
Transcription
1 Near-Isometry by Relaxation: Supplement Proof of Proposition,3. We first prove the following Lemma. Anonymous Author(s) Affiliation Address Proposition. If f : S R D R is convex, non-negative and f exists for all x int S, then f (x) is convex. Proof ( f ) = f f; ( f ) = f f + f f which is positive definite whenever f f is.. Using the above Lemma, and the fact that H k I n is non-negative and infinitely differentiable almost everywhere, we obtain the desired result. Proof of Proposition sup u R s u Gu G G0+εI s = sup u 0 u G 0 u + ɛ u () u T Gu u T G 0 u + ɛ u = sup u 0 v G ɛ For (), we first prove the following fact v v with G ɛ = (G 0 + ɛi) / and v = G ɛ u () GG ɛ = G with G = (G 0 + ɛi) / G(G 0 + ɛi) / (3) = sup u Null G u T Gu u T G 0u+ɛ u if Null(G) = Null(G 0 ) max α +β = β λ (G)+α λ max(g)+αβθ max(g,g 0) β ɛ+α (λ min (G0)+ɛ) if Null(G) Null(G 0 ) (4) where λ max (G) is the spectral radius of G, Θ(G, G 0 ) = sup u = v =,v Null G,u Null G0 u Gv is the cosine of the principal angle between Null G and Null G 0, and λ min (G 0) is the smallest nonzero eigenvalue of G 0. u Denote for simplicity g) = T Gu u T G 0u+ɛ u. () If Null(G) = Null(G 0 ) then for u Null G the value is 0, which cannot be the sup. Let u = v u 0 with u 0 Null G, v Null G. Then u T G 0 u + ɛ u = v T G 0 v + ɛ v + ɛ u 0 > v T G 0 v + ɛ v. Hence, the u which attains the supremum must be in Null G. Now note that, if Null G Null G 0, R s = Null G 0 Null G 0, and Null G 0 = (Null G 0 Null G) V, with V the orthogonal complement of Null G 0 Null G in Null G 0 and the supremum of g) = is attained on U = V Null G 0 (as adding any component along the orthogonal complement of this space only adds a positive value to the denominator, increasing g)). Any u U can be written as u = αu 0 βv 0 with u 0 Null G 0 and v 0 V unit vectors. By upper bounding every term in
2 the numerator and lower bounding u 0G 0 u 0 we obtain the result. Note that for ɛ small enough, the expression in 4 is close to ɛ λ (G). For (), let v V and compute g(v) as above, with α = 0. It follows that g(v) = v Gv ɛ v and by taking the supremum over v V we obtain that sup V g(v) = ɛ λ (G) < r, from which the result follows. For (3), it is obvious that when ɛ 0, g(v) on V, but remains finite for u V. More precisely, G G0 = iff Null G 0 G. To verify that G0 is a norm, we must verify the triangle inequality, since the other two properties obviously hold. If A G0 = or B G0 =, triangle inequality holds trivially. Assume then that A G0, B G0 <. Since A G0+εI s + B G0+εI s A + B G0+εI s for every ɛ > 0, then in the limit we will have that A G0 + B G0 A + B G0. The norm for comparing Riemannian metric The norm of a bilinear functional f : R R R is defined as sup u = v = f, v), or since for a fixed orthonormal base of R s f, v) = u Av, f = sup u = v = u Av. If A is hermitian, then f = max λ(a) λ i where λ(a) denotes the spectrum of A. One can define the norm with respect to any metric G 0 on R s where G 0 is a symmetric, positive definite matrix by f G0 = sup u G0 = v G0 = u Av = sup ũ = ṽ = ũ G / 0 AG / 0 ṽ = max / λ(g λ 0 AG / i In other words, the appropriate operator norm we seek can be expressed as a (generalized) matrix spectral norm. In our cases G 0 = I d 0 and A = H k I d 3 Proof of Propositions 3 Note that we can write the loss as: n Π ky L k YΠ k Π k U k U k Π k k= Where Π k = (U k U k + (ε orth ) k I s ) /. We take the Π k matrices to be fixed and don t depend on the data points Y (in practice they do, however, after taking a gradient step we update the Π k in an E-M style algorithm). Since U k U k and Π k are the identity matrix (the latter multiplied by /( + ε orth )) when s = d we can compute the derivative when s > d without loss of generality. 3. Proof of Derivative Since the derivative is a linear operator it s sufficient to show that the derivative of a single loss function is of the form: l k Y = ( λ k )sgn(λ k)l k YΠ k u k u kπ k To compute the derivative we will make use of the chain rule. First define the function l k as a composition of functions: l k (Y) ρ(p k (H k (Y)) C k ) With C k = Π k U k U k Π k and ρ(u) = (max λ k (U) ) k P k (H) = Π khπ k H k (Y) = Y L k Y Where U, H are both symmetric. Here we note that the matrix spectral norm reduces to the spectral radius if U is symmetric. Since H k (Y) is defined to be symmetric and C k is symmetric this is the case. By the chain rule: Taking these from left to right: Dl k (Y) = Dρ(P k (H k (Y)) C k )DP k (H k (Y))DH k (Y)
3 Dρ Since ρ is defined to be the largest (in absolute value) eigenvalue of U (squared) the derivative is the kronecker product between the corresponding eigenvector and itself multiplied by the sign of the eigenvalue: D ρ(u) = sgn(λ k) k u k) Where λ k = ρ(u) and Uu k = λ k u k Then since we square the spectral radius we add the factor of ( λ k ) so that: D(ρ(U) = ( λ k )sgn(λ k) k u k) 3.. DP k DPk(H) = (Π k Π k) P k (H) = Π khπ k dp k (H) = Π kdhπ k vec(dp k (H)) = vec(π kdhπ k ) = (Π k Π k)dvec(h) 3..3 DH k DHk(Y) = Ns(Is Y Lk) Where N s = I s + K ss for K ss the commutation matrix defined in Magnus & Neudecker ch H k (Y) = Y L k Y dh k (Y) = [(dy) L k Y + Y L k dy] vec(dh k (Y)) = [(Y L k I s )dvec(y ) + (I s Y L k )dvec(y)] 3..4 Dc k Putting it all together = [(Y L k I s )K ns dvec(y) + (I s Y L k )dvec(y)] = [K ss(i s Y L k)dvec(y) + (I s Y L k )dvec(y)] = [(K ss + I s )(I s Y L k )dvec(y)] L k is symmetric = [N s(i s Y L k )dvec(y)] = N s (I s Y L k )dvec(y) Dc k (Y) = ( λ k )sgn(λ k) k u k)(π k Π k)n s (I s Y L k ) = vec ( ) ck Y see Matrix Differential Calculus With Applications in Statistics And Economics by Magnus & Neudecker ch. 9 for proof 3
4 We can simplify this to get the claim: c k Y = ( λ k )sgn(λ k)l k YΠ k u k u kπ k Dc k (Y) = ( λ k )sgn(λ k) k u k)(π k Π k)n s (I s Y L k ) Then note that: = ( λ k )sgn(λ k) k u k)(π k Π k) (K ss + I s )(I s Y L k ) = ( λ k )sgn(λ k) kπ k u kπ k)(k ss + I s )(I s Y L k ) = ( λ k )sgn(λ k) [ k Π k u kπ k)k ss (I s Y L k ) + kπ k u kπ k)(i s Y L k ) ] = ( λ k )sgn(λ k) [ k Π k u kπ k)(y L k I s )K ns + kπ k u kπ k)(i s Y L k ) ] = ( λ k )sgn(λ k) [ k Π ky L k u kπ k)k ns + kπ k u kπ ky L k ) ] = ( λ k )sgn(λ k) [ K kπ k u kπ ky L k ) + kπ k u kπ ky L k ) ] = ( λ k )sgn(λ k) kπ k u kπ ky L k ) K = = ( λ k )sgn(λ k)(π k u k L k YΠ k u k ) vec(( λ k )sgn(λ k)l k YΠ k u k u kπ k) = ( λ k )sgn(λ k)vec([l k YΠ k u k ][][u kπ k]) = ( λ k )sgn(λ k)(π k u k L k YΠ k u k )vec() = ( λ k )sgn(λ k)(π k u k L k YΠ k u k ) = (Dc k (Y) So that c k Y = ( λ k )sgn(λ k)l k YΠ k u k u kπ k The proposition then follows by removing the absolute value and multiplication by the sign. 4
5 Figure : The average number of neighbors m(r) vs the neighborhood radius r, on a log-log scale, for the SDSS spectra data, computed on the whole sample of 675,000 galaxies. The blue regression line, is fitted to the graph points in the shown r range, and has slope.87. The absence of a linear region on this graph suggests that the data dimension varies with the scale. The analysis and visualization in this paper corresponds to the largest meaningful scale. 5
SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS
SPECTRAL PROPERTIES OF THE LAPLACIAN ON BOUNDED DOMAINS TSOGTGEREL GANTUMUR Abstract. After establishing discrete spectra for a large class of elliptic operators, we present some fundamental spectral properties
More informationLecture 8 : Eigenvalues and Eigenvectors
CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with
More informationNearly Isometric Embedding by Relaxation
Nearly Isometric Embedding by Relaxation Anonymous Author(s) Affiliation Address email Abstract 1 2 3 4 5 6 7 8 9 Many manifold learning algorithms aim to create embeddings with low or no distortion (i.e.
More information08a. Operators on Hilbert spaces. 1. Boundedness, continuity, operator norms
(February 24, 2017) 08a. Operators on Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/real/notes 2016-17/08a-ops
More informationNORMS ON SPACE OF MATRICES
NORMS ON SPACE OF MATRICES. Operator Norms on Space of linear maps Let A be an n n real matrix and x 0 be a vector in R n. We would like to use the Picard iteration method to solve for the following system
More informationTutorials in Optimization. Richard Socher
Tutorials in Optimization Richard Socher July 20, 2008 CONTENTS 1 Contents 1 Linear Algebra: Bilinear Form - A Simple Optimization Problem 2 1.1 Definitions........................................ 2 1.2
More informationBoolean Inner-Product Spaces and Boolean Matrices
Boolean Inner-Product Spaces and Boolean Matrices Stan Gudder Department of Mathematics, University of Denver, Denver CO 80208 Frédéric Latrémolière Department of Mathematics, University of Denver, Denver
More informationEE/ACM Applications of Convex Optimization in Signal Processing and Communications Lecture 2
EE/ACM 150 - Applications of Convex Optimization in Signal Processing and Communications Lecture 2 Andre Tkacenko Signal Processing Research Group Jet Propulsion Laboratory April 5, 2012 Andre Tkacenko
More informationFunctional Analysis. Franck Sueur Metric spaces Definitions Completeness Compactness Separability...
Functional Analysis Franck Sueur 2018-2019 Contents 1 Metric spaces 1 1.1 Definitions........................................ 1 1.2 Completeness...................................... 3 1.3 Compactness......................................
More informationCourse 212: Academic Year Section 1: Metric Spaces
Course 212: Academic Year 1991-2 Section 1: Metric Spaces D. R. Wilkins Contents 1 Metric Spaces 3 1.1 Distance Functions and Metric Spaces............. 3 1.2 Convergence and Continuity in Metric Spaces.........
More informationMATH 51H Section 4. October 16, Recall what it means for a function between metric spaces to be continuous:
MATH 51H Section 4 October 16, 2015 1 Continuity Recall what it means for a function between metric spaces to be continuous: Definition. Let (X, d X ), (Y, d Y ) be metric spaces. A function f : X Y is
More informationNonlinear Programming Algorithms Handout
Nonlinear Programming Algorithms Handout Michael C. Ferris Computer Sciences Department University of Wisconsin Madison, Wisconsin 5376 September 9 1 Eigenvalues The eigenvalues of a matrix A C n n are
More information5 Compact linear operators
5 Compact linear operators One of the most important results of Linear Algebra is that for every selfadjoint linear map A on a finite-dimensional space, there exists a basis consisting of eigenvectors.
More informationThe following definition is fundamental.
1. Some Basics from Linear Algebra With these notes, I will try and clarify certain topics that I only quickly mention in class. First and foremost, I will assume that you are familiar with many basic
More informationCS 246 Review of Linear Algebra 01/17/19
1 Linear algebra In this section we will discuss vectors and matrices. We denote the (i, j)th entry of a matrix A as A ij, and the ith entry of a vector as v i. 1.1 Vectors and vector operations A vector
More informationFunctional Analysis HW #5
Functional Analysis HW #5 Sangchul Lee October 29, 2015 Contents 1 Solutions........................................ 1 1 Solutions Exercise 3.4. Show that C([0, 1]) is not a Hilbert space, that is, there
More informationSet, functions and Euclidean space. Seungjin Han
Set, functions and Euclidean space Seungjin Han September, 2018 1 Some Basics LOGIC A is necessary for B : If B holds, then A holds. B A A B is the contraposition of B A. A is sufficient for B: If A holds,
More informationLinear Algebra Massoud Malek
CSUEB Linear Algebra Massoud Malek Inner Product and Normed Space In all that follows, the n n identity matrix is denoted by I n, the n n zero matrix by Z n, and the zero vector by θ n An inner product
More information6 Inner Product Spaces
Lectures 16,17,18 6 Inner Product Spaces 6.1 Basic Definition Parallelogram law, the ability to measure angle between two vectors and in particular, the concept of perpendicularity make the euclidean space
More informationRichard S. Palais Department of Mathematics Brandeis University Waltham, MA The Magic of Iteration
Richard S. Palais Department of Mathematics Brandeis University Waltham, MA 02254-9110 The Magic of Iteration Section 1 The subject of these notes is one of my favorites in all mathematics, and it s not
More informationSPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS
SPECTRAL THEOREM FOR COMPACT SELF-ADJOINT OPERATORS G. RAMESH Contents Introduction 1 1. Bounded Operators 1 1.3. Examples 3 2. Compact Operators 5 2.1. Properties 6 3. The Spectral Theorem 9 3.3. Self-adjoint
More informationExtreme points of compact convex sets
Extreme points of compact convex sets In this chapter, we are going to show that compact convex sets are determined by a proper subset, the set of its extreme points. Let us start with the main definition.
More informationTypical Problem: Compute.
Math 2040 Chapter 6 Orhtogonality and Least Squares 6.1 and some of 6.7: Inner Product, Length and Orthogonality. Definition: If x, y R n, then x y = x 1 y 1 +... + x n y n is the dot product of x and
More informationKernel Method: Data Analysis with Positive Definite Kernels
Kernel Method: Data Analysis with Positive Definite Kernels 2. Positive Definite Kernel and Reproducing Kernel Hilbert Space Kenji Fukumizu The Institute of Statistical Mathematics. Graduate University
More informationSpectral Theory, with an Introduction to Operator Means. William L. Green
Spectral Theory, with an Introduction to Operator Means William L. Green January 30, 2008 Contents Introduction............................... 1 Hilbert Space.............................. 4 Linear Maps
More informationLecture 4 Lebesgue spaces and inequalities
Lecture 4: Lebesgue spaces and inequalities 1 of 10 Course: Theory of Probability I Term: Fall 2013 Instructor: Gordan Zitkovic Lecture 4 Lebesgue spaces and inequalities Lebesgue spaces We have seen how
More informationPCA with random noise. Van Ha Vu. Department of Mathematics Yale University
PCA with random noise Van Ha Vu Department of Mathematics Yale University An important problem that appears in various areas of applied mathematics (in particular statistics, computer science and numerical
More informationThroughout these notes we assume V, W are finite dimensional inner product spaces over C.
Math 342 - Linear Algebra II Notes Throughout these notes we assume V, W are finite dimensional inner product spaces over C 1 Upper Triangular Representation Proposition: Let T L(V ) There exists an orthonormal
More informationFunctional Analysis Review
Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all
More informationU.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018
U.C. Berkeley Better-than-Worst-Case Analysis Handout 3 Luca Trevisan May 24, 2018 Lecture 3 In which we show how to find a planted clique in a random graph. 1 Finding a Planted Clique We will analyze
More informationTHE INVERSE FUNCTION THEOREM FOR LIPSCHITZ MAPS
THE INVERSE FUNCTION THEOREM FOR LIPSCHITZ MAPS RALPH HOWARD DEPARTMENT OF MATHEMATICS UNIVERSITY OF SOUTH CAROLINA COLUMBIA, S.C. 29208, USA HOWARD@MATH.SC.EDU Abstract. This is an edited version of a
More informationReal Variables # 10 : Hilbert Spaces II
randon ehring Real Variables # 0 : Hilbert Spaces II Exercise 20 For any sequence {f n } in H with f n = for all n, there exists f H and a subsequence {f nk } such that for all g H, one has lim (f n k,
More information1 Geometry of R Conic Sections Parametric Equations More Parametric Equations Polar Coordinates...
Contents 1 Geometry of R 1.1 Conic Sections............................................ 1. Parametric Equations........................................ 3 1.3 More Parametric Equations.....................................
More informationSection 3.9. Matrix Norm
3.9. Matrix Norm 1 Section 3.9. Matrix Norm Note. We define several matrix norms, some similar to vector norms and some reflecting how multiplication by a matrix affects the norm of a vector. We use matrix
More informationTHE INVERSE FUNCTION THEOREM
THE INVERSE FUNCTION THEOREM W. PATRICK HOOPER The implicit function theorem is the following result: Theorem 1. Let f be a C 1 function from a neighborhood of a point a R n into R n. Suppose A = Df(a)
More informationPart IB Geometry. Theorems. Based on lectures by A. G. Kovalev Notes taken by Dexter Chua. Lent 2016
Part IB Geometry Theorems Based on lectures by A. G. Kovalev Notes taken by Dexter Chua Lent 2016 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationj=1 [We will show that the triangle inequality holds for each p-norm in Chapter 3 Section 6.] The 1-norm is A F = tr(a H A).
Math 344 Lecture #19 3.5 Normed Linear Spaces Definition 3.5.1. A seminorm on a vector space V over F is a map : V R that for all x, y V and for all α F satisfies (i) x 0 (positivity), (ii) αx = α x (scale
More informationHomework set 4 - Solutions
Homework set 4 - Solutions Math 407 Renato Feres 1. Exercise 4.1, page 49 of notes. Let W := T0 m V and denote by GLW the general linear group of W, defined as the group of all linear isomorphisms of W
More informationOctober 25, 2013 INNER PRODUCT SPACES
October 25, 2013 INNER PRODUCT SPACES RODICA D. COSTIN Contents 1. Inner product 2 1.1. Inner product 2 1.2. Inner product spaces 4 2. Orthogonal bases 5 2.1. Existence of an orthogonal basis 7 2.2. Orthogonal
More informationMetric Spaces and Topology
Chapter 2 Metric Spaces and Topology From an engineering perspective, the most important way to construct a topology on a set is to define the topology in terms of a metric on the set. This approach underlies
More informationMath 577 Assignment 7
Math 577 Assignment 7 Thanks for Yu Cao 1. Solution. The linear system being solved is Ax = 0, where A is a (n 1 (n 1 matrix such that 2 1 1 2 1 A =......... 1 2 1 1 2 and x = (U 1, U 2,, U n 1. By the
More informationComputational math: Assignment 1
Computational math: Assignment 1 Thanks Ting Gao for her Latex file 11 Let B be a 4 4 matrix to which we apply the following operations: 1double column 1, halve row 3, 3add row 3 to row 1, 4interchange
More informationCharacterization of half-radial matrices
Characterization of half-radial matrices Iveta Hnětynková, Petr Tichý Faculty of Mathematics and Physics, Charles University, Sokolovská 83, Prague 8, Czech Republic Abstract Numerical radius r(a) is the
More informationAngles Between Infinite Dimensional Subspaces with Applications to the Rayleigh-Ritz and Alternating Projectors Methods
Angles Between Infinite Dimensional Subspaces with Applications to the Rayleigh-Ritz and Alternating Projectors Methods Andrew Knyazev a,,1, Abram Jujunashvili a, Merico Argentati a a Department of Mathematical
More informationStrongly Regular Graphs, part 1
Spectral Graph Theory Lecture 23 Strongly Regular Graphs, part 1 Daniel A. Spielman November 18, 2009 23.1 Introduction In this and the next lecture, I will discuss strongly regular graphs. Strongly regular
More informationFinite-dimensional spaces. C n is the space of n-tuples x = (x 1,..., x n ) of complex numbers. It is a Hilbert space with the inner product
Chapter 4 Hilbert Spaces 4.1 Inner Product Spaces Inner Product Space. A complex vector space E is called an inner product space (or a pre-hilbert space, or a unitary space) if there is a mapping (, )
More informationCHAPTER 2 -idempotent matrices
CHAPTER 2 -idempotent matrices A -idempotent matrix is defined and some of its basic characterizations are derived (see [33]) in this chapter. It is shown that if is a -idempotent matrix then it is quadripotent
More informationFrom now on, we will represent a metric space with (X, d). Here are some examples: i=1 (x i y i ) p ) 1 p, p 1.
Chapter 1 Metric spaces 1.1 Metric and convergence We will begin with some basic concepts. Definition 1.1. (Metric space) Metric space is a set X, with a metric satisfying: 1. d(x, y) 0, d(x, y) = 0 x
More informationLINEAR MAPS PRESERVING NUMERICAL RADIUS OF TENSOR PRODUCTS OF MATRICES
LINEAR MAPS PRESERVING NUMERICAL RADIUS OF TENSOR PRODUCTS OF MATRICES AJDA FOŠNER, ZEJUN HUANG, CHI-KWONG LI, AND NUNG-SING SZE Abstract. Let m, n 2 be positive integers. Denote by M m the set of m m
More informationCOMP 558 lecture 18 Nov. 15, 2010
Least squares We have seen several least squares problems thus far, and we will see more in the upcoming lectures. For this reason it is good to have a more general picture of these problems and how to
More informationMATH36001 Perron Frobenius Theory 2015
MATH361 Perron Frobenius Theory 215 In addition to saying something useful, the Perron Frobenius theory is elegant. It is a testament to the fact that beautiful mathematics eventually tends to be useful,
More information1 9/5 Matrices, vectors, and their applications
1 9/5 Matrices, vectors, and their applications Algebra: study of objects and operations on them. Linear algebra: object: matrices and vectors. operations: addition, multiplication etc. Algorithms/Geometric
More informationMath 414: Linear Algebra II, Fall 2015 Final Exam
Math 414: Linear Algebra II, Fall 2015 Final Exam December 17, 2015 Name: This is a closed book single author exam. Use of books, notes, or other aids is not permissible, nor is collaboration with any
More informationMath 350 Fall 2011 Notes about inner product spaces. In this notes we state and prove some important properties of inner product spaces.
Math 350 Fall 2011 Notes about inner product spaces In this notes we state and prove some important properties of inner product spaces. First, recall the dot product on R n : if x, y R n, say x = (x 1,...,
More informationChapter 2 Spectra of Finite Graphs
Chapter 2 Spectra of Finite Graphs 2.1 Characteristic Polynomials Let G = (V, E) be a finite graph on n = V vertices. Numbering the vertices, we write down its adjacency matrix in an explicit form of n
More informationChapter 6: Orthogonality
Chapter 6: Orthogonality (Last Updated: November 7, 7) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved around.. Inner products
More informationLINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM
LINEAR ALGEBRA BOOT CAMP WEEK 4: THE SPECTRAL THEOREM Unless otherwise stated, all vector spaces in this worksheet are finite dimensional and the scalar field F is R or C. Definition 1. A linear operator
More informationL p Spaces and Convexity
L p Spaces and Convexity These notes largely follow the treatments in Royden, Real Analysis, and Rudin, Real & Complex Analysis. 1. Convex functions Let I R be an interval. For I open, we say a function
More informationChapter 4 Euclid Space
Chapter 4 Euclid Space Inner Product Spaces Definition.. Let V be a real vector space over IR. A real inner product on V is a real valued function on V V, denoted by (, ), which satisfies () (x, y) = (y,
More informationProblem Set 2: Solutions Math 201A: Fall 2016
Problem Set 2: s Math 201A: Fall 2016 Problem 1. (a) Prove that a closed subset of a complete metric space is complete. (b) Prove that a closed subset of a compact metric space is compact. (c) Prove that
More informationMath Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88
Math Camp 2010 Lecture 4: Linear Algebra Xiao Yu Wang MIT Aug 2010 Xiao Yu Wang (MIT) Math Camp 2010 08/10 1 / 88 Linear Algebra Game Plan Vector Spaces Linear Transformations and Matrices Determinant
More informationReal symmetric matrices/1. 1 Eigenvalues and eigenvectors
Real symmetric matrices 1 Eigenvalues and eigenvectors We use the convention that vectors are row vectors and matrices act on the right. Let A be a square matrix with entries in a field F; suppose that
More informationMath 328 Course Notes
Math 328 Course Notes Ian Robertson March 3, 2006 3 Properties of C[0, 1]: Sup-norm and Completeness In this chapter we are going to examine the vector space of all continuous functions defined on the
More informationDS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.
DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1
More informationMathematics for Economists
Mathematics for Economists Victor Filipe Sao Paulo School of Economics FGV Metric Spaces: Basic Definitions Victor Filipe (EESP/FGV) Mathematics for Economists Jan.-Feb. 2017 1 / 34 Definitions and Examples
More informationProblem set 5, Real Analysis I, Spring, otherwise. (a) Verify that f is integrable. Solution: Compute since f is even, 1 x (log 1/ x ) 2 dx 1
Problem set 5, Real Analysis I, Spring, 25. (5) Consider the function on R defined by f(x) { x (log / x ) 2 if x /2, otherwise. (a) Verify that f is integrable. Solution: Compute since f is even, R f /2
More informationConvexity in R n. The following lemma will be needed in a while. Lemma 1 Let x E, u R n. If τ I(x, u), τ 0, define. f(x + τu) f(x). τ.
Convexity in R n Let E be a convex subset of R n. A function f : E (, ] is convex iff f(tx + (1 t)y) (1 t)f(x) + tf(y) x, y E, t [0, 1]. A similar definition holds in any vector space. A topology is needed
More informationLecture notes: Applied linear algebra Part 1. Version 2
Lecture notes: Applied linear algebra Part 1. Version 2 Michael Karow Berlin University of Technology karow@math.tu-berlin.de October 2, 2008 1 Notation, basic notions and facts 1.1 Subspaces, range and
More informationTHE ALTERNATIVE DUNFORD-PETTIS PROPERTY FOR SUBSPACES OF THE COMPACT OPERATORS
THE ALTERNATIVE DUNFORD-PETTIS PROPERTY FOR SUBSPACES OF THE COMPACT OPERATORS MARÍA D. ACOSTA AND ANTONIO M. PERALTA Abstract. A Banach space X has the alternative Dunford-Pettis property if for every
More informationLecture 5 - Hausdorff and Gromov-Hausdorff Distance
Lecture 5 - Hausdorff and Gromov-Hausdorff Distance August 1, 2011 1 Definition and Basic Properties Given a metric space X, the set of closed sets of X supports a metric, the Hausdorff metric. If A is
More informationNumerical Sequences and Series
Numerical Sequences and Series Written by Men-Gen Tsai email: b89902089@ntu.edu.tw. Prove that the convergence of {s n } implies convergence of { s n }. Is the converse true? Solution: Since {s n } is
More informationMath 320-2: Midterm 2 Practice Solutions Northwestern University, Winter 2015
Math 30-: Midterm Practice Solutions Northwestern University, Winter 015 1. Give an example of each of the following. No justification is needed. (a) A metric on R with respect to which R is bounded. (b)
More informationElements of Convex Optimization Theory
Elements of Convex Optimization Theory Costis Skiadas August 2015 This is a revised and extended version of Appendix A of Skiadas (2009), providing a self-contained overview of elements of convex optimization
More informationClass notes: Approximation
Class notes: Approximation Introduction Vector spaces, linear independence, subspace The goal of Numerical Analysis is to compute approximations We want to approximate eg numbers in R or C vectors in R
More informationIterative Reweighted Minimization Methods for l p Regularized Unconstrained Nonlinear Programming
Iterative Reweighted Minimization Methods for l p Regularized Unconstrained Nonlinear Programming Zhaosong Lu October 5, 2012 (Revised: June 3, 2013; September 17, 2013) Abstract In this paper we study
More informationMatrices and Vectors. Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A =
30 MATHEMATICS REVIEW G A.1.1 Matrices and Vectors Definition of Matrix. An MxN matrix A is a two-dimensional array of numbers A = a 11 a 12... a 1N a 21 a 22... a 2N...... a M1 a M2... a MN A matrix can
More informationShiqian Ma, MAT-258A: Numerical Optimization 1. Chapter 4. Subgradient
Shiqian Ma, MAT-258A: Numerical Optimization 1 Chapter 4 Subgradient Shiqian Ma, MAT-258A: Numerical Optimization 2 4.1. Subgradients definition subgradient calculus duality and optimality conditions Shiqian
More informationFunctional Analysis Exercise Class
Functional Analysis Exercise Class Week 9 November 13 November Deadline to hand in the homeworks: your exercise class on week 16 November 20 November Exercises (1) Show that if T B(X, Y ) and S B(Y, Z)
More informationMATH 140B - HW 5 SOLUTIONS
MATH 140B - HW 5 SOLUTIONS Problem 1 (WR Ch 7 #8). If I (x) = { 0 (x 0), 1 (x > 0), if {x n } is a sequence of distinct points of (a,b), and if c n converges, prove that the series f (x) = c n I (x x n
More informationAn LMI description for the cone of Lorentz-positive maps II
An LMI description for the cone of Lorentz-positive maps II Roland Hildebrand October 6, 2008 Abstract Let L n be the n-dimensional second order cone. A linear map from R m to R n is called positive if
More informationLecture 4: Purifications and fidelity
CS 766/QIC 820 Theory of Quantum Information (Fall 2011) Lecture 4: Purifications and fidelity Throughout this lecture we will be discussing pairs of registers of the form (X, Y), and the relationships
More informationAnalysis II: The Implicit and Inverse Function Theorems
Analysis II: The Implicit and Inverse Function Theorems Jesse Ratzkin November 17, 2009 Let f : R n R m be C 1. When is the zero set Z = {x R n : f(x) = 0} the graph of another function? When is Z nicely
More information(x, y) = d(x, y) = x y.
1 Euclidean geometry 1.1 Euclidean space Our story begins with a geometry which will be familiar to all readers, namely the geometry of Euclidean space. In this first chapter we study the Euclidean distance
More information2 (Bonus). Let A X consist of points (x, y) such that either x or y is a rational number. Is A measurable? What is its Lebesgue measure?
MA 645-4A (Real Analysis), Dr. Chernov Homework assignment 1 (Due 9/5). Prove that every countable set A is measurable and µ(a) = 0. 2 (Bonus). Let A consist of points (x, y) such that either x or y is
More information2 Garrett: `A Good Spectral Theorem' 1. von Neumann algebras, density theorem The commutant of a subring S of a ring R is S 0 = fr 2 R : rs = sr; 8s 2
1 A Good Spectral Theorem c1996, Paul Garrett, garrett@math.umn.edu version February 12, 1996 1 Measurable Hilbert bundles Measurable Banach bundles Direct integrals of Hilbert spaces Trivializing Hilbert
More informationEXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018
EXPOSITORY NOTES ON DISTRIBUTION THEORY, FALL 2018 While these notes are under construction, I expect there will be many typos. The main reference for this is volume 1 of Hörmander, The analysis of liner
More informationHilbert spaces. 1. Cauchy-Schwarz-Bunyakowsky inequality
(October 29, 2016) Hilbert spaces Paul Garrett garrett@math.umn.edu http://www.math.umn.edu/ garrett/ [This document is http://www.math.umn.edu/ garrett/m/fun/notes 2016-17/03 hsp.pdf] Hilbert spaces are
More informationAlgorithms for Picture Analysis. Lecture 07: Metrics. Axioms of a Metric
Axioms of a Metric Picture analysis always assumes that pictures are defined in coordinates, and we apply the Euclidean metric as the golden standard for distance (or derived, such as area) measurements.
More informationFRAMES, GRAPHS AND ERASURES
FRAMES, GRAPHS AND ERASURES BERNHARD G. BODMANN AND VERN I. PAULSEN Abstract. Two-uniform frames and their use for the coding of vectors are the main subject of this paper. These frames are known to be
More informationBasic Calculus Review
Basic Calculus Review Lorenzo Rosasco ISML Mod. 2 - Machine Learning Vector Spaces Functionals and Operators (Matrices) Vector Space A vector space is a set V with binary operations +: V V V and : R V
More informationMetric-based classifiers. Nuno Vasconcelos UCSD
Metric-based classifiers Nuno Vasconcelos UCSD Statistical learning goal: given a function f. y f and a collection of eample data-points, learn what the function f. is. this is called training. two major
More information(convex combination!). Use convexity of f and multiply by the common denominator to get. Interchanging the role of x and y, we obtain that f is ( 2M ε
1. Continuity of convex functions in normed spaces In this chapter, we consider continuity properties of real-valued convex functions defined on open convex sets in normed spaces. Recall that every infinitedimensional
More informationElementary linear algebra
Chapter 1 Elementary linear algebra 1.1 Vector spaces Vector spaces owe their importance to the fact that so many models arising in the solutions of specific problems turn out to be vector spaces. The
More informationChapter 3 Transformations
Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases
More informationMaximizing the numerical radii of matrices by permuting their entries
Maximizing the numerical radii of matrices by permuting their entries Wai-Shun Cheung and Chi-Kwong Li Dedicated to Professor Pei Yuan Wu. Abstract Let A be an n n complex matrix such that every row and
More informationDEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix to upper-triangular
form) Given: matrix C = (c i,j ) n,m i,j=1 ODE and num math: Linear algebra (N) [lectures] c phabala 2016 DEN: Linear algebra numerical view (GEM: Gauss elimination method for reducing a full rank matrix
More informationFoundations of Matrix Analysis
1 Foundations of Matrix Analysis In this chapter we recall the basic elements of linear algebra which will be employed in the remainder of the text For most of the proofs as well as for the details, the
More informationMath 117: Topology of the Real Numbers
Math 117: Topology of the Real Numbers John Douglas Moore November 10, 2008 The goal of these notes is to highlight the most important topics presented in Chapter 3 of the text [1] and to provide a few
More informationChapter 7: Symmetric Matrices and Quadratic Forms
Chapter 7: Symmetric Matrices and Quadratic Forms (Last Updated: December, 06) These notes are derived primarily from Linear Algebra and its applications by David Lay (4ed). A few theorems have been moved
More informationMathematical foundations - linear algebra
Mathematical foundations - linear algebra Andrea Passerini passerini@disi.unitn.it Machine Learning Vector space Definition (over reals) A set X is called a vector space over IR if addition and scalar
More information