UNCONSTRAINED OPTIMIZATION PAUL SCHRIMPF OCTOBER 24, 2013

Size: px
Start display at page:

Download "UNCONSTRAINED OPTIMIZATION PAUL SCHRIMPF OCTOBER 24, 2013"

Transcription

1 PAUL SCHRIMPF OCTOBER 24, 213 UNIVERSITY OF BRITISH COLUMBIA ECONOMICS 26 Today s lecture is about unconstrained optimization. If you re following along in the syllabus, you ll notice that we ve skipped the fourth topic, eigenvalues and definite matrices. We will cover these things as part of our study of optimization. 1. NOTATION AND DEFINITIONS An optimization problem refers to finding the maimum or minimum of a function, perhaps subject to some constraints. In economics, the most common optimization problems are utility maimization and profit maimization. Because of this, we will state most of our definitions and results for maimization problems. Of course, we could just as well state each definition and result for a minimization problem by reversing the sign of most inequalities. In this lecture we will be interested in unconstrained optimization problems such as ma U F where U R n and F : U R. If F = ma U F, we mean that F F for all U and F = F for some U. Definition 1.1. F = ma U F is the maimum of F on U if F F for all U and F = F for some U There may be more than one such. We denote the set of all such that F = F by arg ma U F and might write arg ma U F, or, if we know there is only one such,, we sometimes write = arg ma U F. Definition 1.2. U is a maimizer of F on U if F = ma U F. The set of all maimizers is denoted arg ma U F. Definition 1.3. U is a strict maimizer of F on U if F > F for all U with =. Recall the definition of a local maimum from lecture 8. Definition 1.4. F has a local maimum at if δ > such that Fy F for all y N δ U. Each such is called a local maimizer of F. If Fy < F for all y =, y N δ U, then we say F has a strict local maimum at. When we want to be eplicit about the distinction between local maimum and the maimum in definition 1.1, we refer to the later as the global maimum. 1

2 Eample 1.1. Here are some eamples of functions from R R and their maima and minima. 1 F = 2 is minimized at = with minimum. 2 F = c has minimum and maimum c. Any is a maimizer. 3 F = cos has maimum 1 and minimum 1. 2πn for n Z is a maimizer. 4 F = cos + /2 has no global maimizer or minimizer, but has many local ones. 2. FIRST ORDER CONDITIONS In lecture 8, we proved that if F has a local maimum at, then DF =. We restate that theorem and proof here. Theorem 2.1. Let U R n, F : U R, and suppose F has a local maimum or minimum at, F is differentiable at, and interioru. Then DF =. Proof. We will write the proof for when F has a local maimum at. The eact same reasoning works when F has a local minimum. Since is in the interior of U, we can choose δ > such that N δ U. Since is a local maimum we can also choose such that F Fy for all y N δ. Since F is differentiable, we can write F + h F h = DF h + r, h h =. Let h = tv for some v R n with v = 1 and t R. If r,h where lim h h DF v >, then for t > small enough, we would have F+tv F t = DF v + r,tv > t DF v/2 > and F + tv > F in contradiction to being a local maimum. Similarly, if DF v v < then for t < and small, we would have F+tv F t = DF v + r,tv t > D f v/2 > and F + tv > F. Thus, it must be that DF v = for all v, i.e. DF =. The first order condition is the fact that DF = is a necessary condition for to be a local maimizer or minimizer of F. Definition 2.1. Any point such that DF = is call a critical point of F. If F is differentiable, F cannot have local minima or maima =local etrema at noncritical points. F might have a local etrema its critical points, but it does not have to. Consider F = 3 F =, but is not a local maimizer or minimizer of F. Similarly, if F : R 2 R, F = , then DF =, but is not a local minimum or maimum of F. 3. SECOND ORDER CONDITIONS To determine whether a given critical point is a local minimum or maimum or neither we can look at the second derivative of the function. Let F : R n R and suppose is a critical point. Then DF =. To see if is a local maimum, we need to look at 2

3 F + h for small h. If F is twice continuously differentiable, we can take a second order Taylor epansion of F around. F + h = F + DF h ht D 2 F h + r, h where D 2 F is the matri of F s second derivatives. It is called the Hessian of F. 2 F 2 F 2 D n F = F 1 n 2 2 F n Since is a critical, point DF =, so We can see that is a local maimum if F + h F = 1 2 ht D 2 F h + r, h. 1 2 ht D 2 F h + r, h for all h =, h < δ for some δ >. We know that r, h is small so, we epect that the above inequality will be true if h T D 2 F h for all h =. The Hessian, D 2 F is just some symmetric n by n matri, and h T D 2 F h is a quadratic form in h. This motivates the following definition. Definition 3.1. Let A be a symmetric matri, then A is Negative definite if T A < for all = Negative semi-definite if T A for all = Positive definite if T A > for all = Positive semi-definite if T A for all = Indefinite if 1 s.t. T 1 A 1 > and some other 2 such that T 2 A 2 <. In the net section we will derive some conditions on A that ensure it is negative semi- definite. For now, just observe that if D 2 F is negative semi-definite, then must be a local maimum. If D 2 F is negative definite, then is a strict local maimum. The following theorem restates the results of this discussion. Theorem 3.1. Let F : U R be twice continuously differentiable on U and let be a critical point in the interior of U. If 1 The Hessian, D 2 F is negative definite, then is a strict local maimizer. 2 The Hessian, D 2 F is positive definite, then is a strict local minimizer. 3 The Hessian, D 2 F is indefinite, is neither a local min nor a local ma. 4 The Hessian is positive or negative semi-definite, then could be a local maimum, minimum, or neither. Proof. We will only prove the first case. The second and third cases can be proven similarly. We will go over some eamples of the fourth case. 3

4 The main idea of the proof is contained in the discussion at the start of this section. The only tricky part is carefully showing that r, h is small enough to ignore. As in the discussion preceding the theorem, is a local minimizer if We can rewrite this as Factoring, we have F + h F = 1 2 ht D 2 F h + r, h. h T D 2 F h + r, h =h T D 2 F h + h T h r, h h 2 h T D 2 F h + h T h r, h h 2 =h T D 2 F + r, h h 2 h. r From our theorem on Taylor series, we know lim,h h h 2 δ > such that if h < δ, then r,h < ϵ. So, h 2 Let t = h and q = h T D 2 F + r, h h 2 h h T D 2 F + ϵ h h T D 2 F h + ϵ h 2 h, so h = tq. Then we have h h T D 2 F + r, h h 2 h t 2 q T D 2 F q + ϵt 2 =. Then for any ϵ > We can pick ϵ < inf q =1 q T D 2 F q. The set {q : q = 1} is compact so there is some q that achieves this minimum, and q T D 2 F q <, so ϵ >. 1 Then for all t < δ and q = 1, t 2 q T D 2 F q + ϵt 2 < t 2 q T D 2 F q + q T D 2 F q t 2 < where the last inequality follows from q T D 2 F q <. Thus, F + h F < for all h < δ. When the Hessian is not positive definite, negative definite, or indefinite, the result of this theorem is ambiguous. Let s go over some eamples of this case. Eample 3.1. F : R R, F = 4. The first order condition is 4 3 =, so = is the only critical point. The Hessian is F = 12 2 = at. However, 4 has a strict local minimum at. 1 If we didn t have compactness, we might have ended up with ϵ =, which is not allowed. 4

5 Eample 3.2. F : R 2 R, F 1, 2 = 1 2. The first order condition is DF = 2 1, =, so the 1 =, 2 R are all critical points. The Hessian is D 2 2 F = This is negative semi-definite because h T D 2 F h = 2h 2 1. Also, graphing the function would make it clear that 1 =, 2 R are all non-strict local maima. Eample 3.3. F : R 2 R, F 1, 2 = The first order condition is DF = 2 1, 42 3 =, so the =, is a critical point. The Hessian is D 2 2 F = This is negative semi-definite at because h T D 2 F h = 2h 2 1. However, is not a local maimum because F, 2 > F, for any 2 =. is also not a local minimum because F 1, < F, for all 1 =. In each of these eamples, the second order condition is inconclusive because h T D 2 F h = for some h. In these cases we could determine whether is a local maimum, local minimum, or neither by either looking at higher derivatives of F at, or look at D 2 F for all in a neighborhood of. We will not often encounter cases where the second order condition is inconclusive, so we will not study these possibilities in detail. The converse of theorem 3.1 is nearly true. If is a local maimizer, then D 2 F must be negative semi-definite. Theorem 3.2. Let F : U R be twice continuously differentiable on U and let in the interior of U be a local maimizer or minimizer of F. Then DF = and D 2 F is negative or positive semi-definite. Proof. Using the same notation and setup as in the proof of theorem 3.1, > F + tq F =t qt D 2 F q + r, tq >t qt D 2 F q + r, tq t 2 r,tq Because lim t =, for any ϵ > δ > such that if t < δ, then t 2 1 >t 2 2 qt D 2 r, tq 1 F q + t 2 t 2 2 qt D 2 F q ϵ t 2 ϵ > 1 2 t2 q T D 2 F q This is only possible for all ϵ > if q T D 2 F q. I do not epect you to remember the proofs of the last two theorems. However, it is important to remember the second order condition, and know how to check it. To check the second order condition, we need some practical way to tell whether a matri is positive or negative semi-definite.

6 4. DEFINITE MATRICES The second order condition says that a critical point is a local maimum if the Hessian is negative definite. In this section we will develop some conditions for whether a matri is negative definite. Let s start with the simplest case where A is a one by one matri. Then T A = 2 a is negative definite if and only if a <. A is negative semi-definite if a. What if A is a two by two symmetric matri? Then, T T 1 a b 1 A = b c Completing the square we get T A =a 1 + 2b 2 =a 1 + b a 2 2 =a b c 2 2 a b2 2 + a ac b2 2 2 a b2 a c2 2 Thus T A < for all = if a < and ac b2 a <, i.e. ac b 2 >. Notice that ac b 2 = deta. So A is negative definite if a < and deta >. If we wanted A to be positive definite, we would need a > and deta >. For semi-definite we get the same thing with weak instead of strict inequalities. It turns out that in general this sort of pattern continues. We can determine whether A is negative definite by looking at the determinants of certain submatrices of A. Definition 4.1. Let A by an n by n matri. The k by k submatri a 11 a 1k A k =.. a k1 a kk is the kth leading principal submatri of A. The determinant of A k is the kth order leading principal minor of A. Theorem 4.1. Let A be an n by n symmetric matri. Then 1 A is positive definite if and only if all n of its leading principal minors are strictly positive. 2 A is positive semi-definite if and only if all n of its leading principal minors are weakly positive. 3 A is negative definite if and only if all n of its leading principal minors alternate in sign as follows: deta 1 <, deta 2 >, deta 3 <, etc. 4 A is negative semi-definite if and only if all n of its leading principal minors weakly alternate in sign as follows: deta 1, deta 2, deta 3, etc A is indefinite if and only if none of the five above cases hold, and deta k = for at least one k. The proof of this theorem is a bit tedious, so we will skip it. There is a proof in chapter 16.4 of Simon and Blume. Figure 1 shows each of the five types of definiteness. 6

7 FIGURE 1. Definite quadratic forms Positive definite 2 + y 2 Negative definite 2 y y - - y Positive semi-definite 2 Negative semi-definite y y - - y Indefinite 2 y y Eigenvectors and eigenvalues. Another way to check whether a matri is negative definite is by looking at the matri s eigenvalues. Eigenvalues are of interest in their own right as well because eigenvalues have many other uses, such as determining the stability of systems of difference and differential equations. 7

8 Definition 4.2. If A is an n by n matri, λ is a scalar, v R n with v = 1, and Av = λv then λ is a eigenvalue of A and v is an eigenvector. If v is an eigenvector of A with eigenvalue λ, then Atv = λtv for any t =. Thus, the requirement that v = 1 is just a normalization to pin down v. There are a few equivalent ways of defining eigenvalues, some of which are given by the following lemma. Lemma 4.1. Let A be an n by n matri and λ a scalar. Each of the following are equivalent. 1 λ is a eigenvalue of A. 2 A λi is singular. 3 deta λi =. Proof. We know that 2 and 3 from our results on systems of linear equations and matrices. Also, if A λi is singular, then the null space of A λi contains non-zero vectors. Choose v N A λi such that v =. Then A λiv =, so Av/ v = λv/ v as in the definition of eigenvalues. The function χ A : R R defined by χ A = deta I is called the characteristic polynomial of A. It will be a polynomial of order n. You may know from some other math course that polynomials of degree n have n roots, some of which might be not be real and some of which might not be distinct. Suppose that A has m n distinct real eigenvalues, λ 1,..., λ m. Associated with each eigenvalue is a linear subspace of eigenvalues, {v R n : Av = λ i v} = N A λ i I. By definition of eigenvalues each of these subspaces has dimension at least 1. Some of them may be of larger dimension, but they can be of dimension at most n. For each N A λ i I there an orthonormal basis of eigenvectors, i.e. v 1 i,..., vk i i and v j i T vi l = for j = l. such that Av j i = λ iv j v i, j = 1 Lemma 4.2. Let A be an n by n matri with m distinct real eigenvalues, and let v 1 i,..., vk i be an orthonormal basis for N A λ i I. Then {v 1 1,..., vk 1 1,..., v1 m,..., v k m m } are linearly independent. Proof. Suppose the eigenvectors are not linearly independent. Then we could write 2 v 1 1 = c2 1 v ck mv k m. i 2 Possibly we cannot do this for v1, but we can for some v i. Without loss of generality, assume that we can for v 1. 8

9 with at least one c j i =. By the definition of eigenvalues and eigenvectors, we have Av 1 =λ 1 v 1 Ac 2 1 v ck mv k m =λ 1 c 2 1 v ck mv k m c 2 1 λ 1v ck mλ m v k m =λ 1 c 2 1 v c k mv k m c 1 2 λ 2 λ 1 v ck mλ m λ 1 v m = If c 1 2 = = ck m, then the original equation becomes v 1 1 = c2 1 v ck 1 1 vk 1 1 with one of these c s non-zero. That would contradict the way v 1 1,...vk 1 1 were constructed to be linearly independent. Hence, c j i = for at least one i > 1. Since the λ i are distinct, λ k λ 1 = for any k, so this says that v 1 2,..., vk m m are linearly dependent as well. We can repeat this argument to show that v 1 3,..., vk m m are linearly dependent, and then repeat it again and again and eventually show that v m 1 and v m are linearly dependent. Then v m 1 = bv m for some b =, and and Av m 1 = λ m 1 v m 1 = λ m 1 bv m Av m 1 = Abv m = bλ m v m. This implies that λ m = λ m 1, contrary to the eigenvalues being distinct. Therefore, the collection of eigenvectors must be linearly independent. Since the eigenvectors are linearly independent, in R n there can be at most n of them. However, there are some matrices that have strictly fewer eigenvectors. To say much more about eigenvalues we must impose the etra condition that A is symmetric. Fortunately, Hessian matrices are symmetric, so for the current purposes, imposing symmetry is not a problem. Since A is symmetric, A = A T. Also, by the definition of transpose, A, y =, A T y, or specializing to R n, A T y = A T y = Ay, where the last equality comes from symmetry of A. By the fundamental theorem of algebra which says that polynomials have possibly comple roots, at least one λ 1 and v 1 such that Av 1 = λ 1 v 1. A short argument using comple numbers which we have not covered and the symmetry of A shows that λ 1 must be real. If we then consider A as a symmetric linear operator on the n 1 dimensional space span{v 1 }, we can repeat this argument to show that there is a second not necessarily distinct eigenvalue λ 2 and associated eigenvector v 2 with v 2 v 1. If we repeat this n times, we will construct n eigenvalues some of which may be the same and n orthogonal eigenvectors. Moreover, we can rescale the eigenvectors so that v i = 1 for each i. These n eigenvectors form an orthonormal basis for R n. You can imagine them as some rotation of the usual aes. 9

10 Net, we can write A in terms of its eigenvectors and eigenvalues. If we make Λ a diagonal matri consisting of the eigenvalues of A and V a matri whose columns are the eigenvectors of A, then AV = VΛ. Moreover, V is an orthogonal matri, so V 1 = V T. This relationship is called the eigendecomposition of A. Theorem 4.2 Eigendecomposition. Let A be an n by n symmetric matri, then A has n not necessarily distinct eigenvalues and A = VΛV T where Λ is the diagonal matri consisting of the eigenvalues of A and the columns of V are the associated eigenvectors of A, and V is an orthonormal matri. Comment 4.1. There are non-symmetric matrices that can not be decomposed into eigenvalues and eigenvectors, for eample There are other non-symmetric matrices that can be eigendecomposed, for eample can be eigendecomposed. Any square matri with A T A = AA T can be eigendecomposed. 1 1 Using the eigendecomposition, we can relate eigenvalues to the definiteness of a matri. Theorem 4.3. If A is an n by n symmetric non-singular matri with eigenvalues λ 1,..., λ n, then 1 λ i > for all i, iff A is positive definite, 2 λ i for all i, iff A is positive semi-definite, 3 λ i < for all i, iff A is negative definite, 4 λ i for all i, iff A is negative semi-definite, if some λ i > and some λ j <, then A is indefinite. Proof. Let A = VΛV T be the eigendecomposition of A. Let = R n. Then T V = z T = because V is nonsingular. Also, Since Λ is diagonal we have T A = T VΛV T = z T Λz. z T Λz = z 2 1 λ z 2 nλ n Thus, T A = z T Λz > for all iff λ i > for each i. The other parts of the theorem follow similarly Other facts about eigendecomposition. The results in this subsection are not essential for the rest of this lecture, but they are occasionally useful. If a matri is singular, then = such that A =. By definition, then / is an eigenvector and is an eigenvalue of A. Conversely, if is an eigenvalue of A, then with = 1 such that A =. Then A is singular. 1

11 Lemma 4.3 Eigenvalues and invertibility. A is singular iff is an eigenvalue of A. Recall that we defined the norm of a linear transformation as A BL = sup A. : =1 When A is symmetric so that its eigenvalues eist, this norm is equal to the largest eigenvalue of A. Lemma 4.4 Eigenvalues and operator norm. Suppose the n n matri A has n eigenvalues, λ 1,..., λ n, then A BL = ma 1 i n λ i. We saw that definiteness is related to both determinant and eigenvalues. It should come as no suprise that determinants and eigenvalues are directly related. The determinant of a matri is equal to the product of its eigenvalues. Lemma 4. Eigenvalues and determinants. If A has eigenvalues λ 1,..., λ n, such that then deta = n i=1 λ i. deta ti = n i=1 t λ i. GLOBAL MAXIMUM AND MINIMUM The second order condition 3.1 along with the first order condition gives a nice way of finding local maima of a function, but what about the global maimum? In general, you could use the first and second order conditions to find each of the local maima of F in the interior of U. You then must compare each of these local maima with each other and the value of F on the boundary of U to find the global maimum of F on U. If there are many local maima or the boundary of U is complicated, this procedure can be quite complicated. It would be nice if there was some simpler necessary and sufficient for a global maimum. Unfortunately, there is no such general necessary and sufficient condition. There is, however, a sufficient condition that is sometimes useful. Definition.1. Let f : U R. f is conve if for all, y U with l, y U we have f t + 1 ty t f + 1 t f y for all t [, 1]. Equivalently, F is conve if the set {y, : U, y f } U R is conve. This set is called the epigraph of f. If you draw the graph of the function, the epigraph is the set of points above the function. Definition.2. Let f : U R. f is concave if for all, y U with l, y U we have f t + 1 ty t f + 1 t f y for all t [, 1]. Equivalently, F is concave if the set {y, : U, y f } is conve. Theorem.1. Let F : U R be twice continuously differentiable and U R n be conve. Then, 1 For maimization: a The following three conditions are equivalent: 11

12 i F is a concave function on U ii Fy F DF y for all, y U, iii D 2 F is negative semi-definite for all, y U b If F is a concave function on U and DF = for some U, then is the global maimizer of F on U. 2 For minimization: a The following three conditions are equivalent: i F is a conve function on U ii Fy F DF y for all, y U, iii D 2 F is positive semi-definite for all, y U b If F is a conve function on U and DF = for some U, then is the global minimizer of F on U. This theorem is often summarized by saying that conve minimization problems i.e. the set U and function F are both conve have unique solutions. We will not prove this theorem until after covering conveity in more detail, which we may or may not have time for. An important fact about conve optimization problems is that their solutions can be efficiently computed. A general optimization problem is what is known as NPhard, which in practice means that you can write down what looks like a reasonable not too large optimization problem whose solution takes prohibitively long to compute. In contrast conve optimization problems can be solved in polynomial time. In particular, if f is conve, is n dimensional, the steps needed to compute f is M, and you want to find ˆ such that f ˆ f < ϵ, then there are algorithms that can find ˆ in O nn 3 + M log1/ϵ steps. In fact, there are algorithms for special types of conve problems that can solve them nearly as fast as a least squares problem or a linear program. In practice, this means that most conve problems can be solved quickly with up to about a one thousand variables, and some conve problems can be solved very quickly even with tens or hundreds of thousands of variables. The discovery that conve optimization problems can be solved efficiently was fairly recent. The main theoretical breakthroughs began in the late eighties and early nineties, and it remains an active area of research Profit maimization. 6. APPLICATIONS Competitive multi-product firm. Suppose a firm has produces k goods using n inputs with production function f : R n R k. The prices of the goods are p, and the prices of the inputs are w, so that the firm s profits are The firm chooses to maimize profits. Π = p T f w T. The first order condition is ma p T f w T p T D f w =. 12

13 or without using matrices, k f j p j = w i j=1 i for i = 1,..., n. The second order condition is that k j=1 p 2 f j j 2 k j=1 p 2 f j j D[p T 1 1 n D f ] =.. k j=1 p 2 f j j 1 n k j=1 p 2 f j j must be negative semidefinite Multi-product monopolist. Consider the same setup as before, but now with a monopolist who recognizes that prices depend on output. Let the inverse demand function be Pq where P : R k R k. Now the firm s problem is The first order condition is or without matrices, k j=1 ma P f T f w T 2 n D f T DPT f f + P f D f w = k P j f f l f q l=1 l l + P j f f j i i = w i We can get something a bit more interpretable by writing this in terms of elasticities. Recall that the elasticity of demand for good l with respect to the price of good j is ϵ l j 1 = P j q l q l P l. Then, k P j f j=1 k P j f ϵ l f j 1 l + f j =w l=1 i i i [ ] fj 1 + ϵ j j 1 + i ϵ l f j 1 l =w i l =j i k j=1 There is a way to compare the price and quantity produced by the monopolist to the competitive firm. There are also things that can be said about the comparison between a single product monopolist k = 1 vs a multi-product monopolist. It might be interesting to derive some of these results. To begin with let k = 1 and compare the monopolist to the competitive firm. Under some reasonable assumptions on f and P, you can show that m < c where m is for the monopolist and c is for the competitive firm. You can also show that p m > p c and that ϵ1 1 < 1 for the monopolist. Although I left it out of the notation above, ϵ depends on. 13

Paul Schrimpf. October 18, UBC Economics 526. Unconstrained optimization. Paul Schrimpf. Notation and definitions. First order conditions

Paul Schrimpf. October 18, UBC Economics 526. Unconstrained optimization. Paul Schrimpf. Notation and definitions. First order conditions Unconstrained UBC Economics 526 October 18, 2013 .1.2.3.4.5 Section 1 Unconstrained problem x U R n F : U R. max F (x) x U Definition F = max x U F (x) is the maximum of F on U if F (x) F for all x U and

More information

Math (P)refresher Lecture 8: Unconstrained Optimization

Math (P)refresher Lecture 8: Unconstrained Optimization Math (P)refresher Lecture 8: Unconstrained Optimization September 2006 Today s Topics : Quadratic Forms Definiteness of Quadratic Forms Maxima and Minima in R n First Order Conditions Second Order Conditions

More information

Economics 205 Exercises

Economics 205 Exercises Economics 05 Eercises Prof. Watson, Fall 006 (Includes eaminations through Fall 003) Part 1: Basic Analysis 1. Using ε and δ, write in formal terms the meaning of lim a f() = c, where f : R R.. Write the

More information

Optimization Paul Schrimpf September 12, 2018

Optimization Paul Schrimpf September 12, 2018 Optimization Paul Schrimpf September 1, 018 University of British Columbia Economics 56 cba1 Today s lecture is about optimization Useful references are chapters 1-5 of Diit (1990), chapters 16-19 of Simon

More information

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018

MATH 5720: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 2018 MATH 57: Unconstrained Optimization Hung Phan, UMass Lowell September 13, 18 1 Global and Local Optima Let a function f : S R be defined on a set S R n Definition 1 (minimizers and maximizers) (i) x S

More information

1 Kernel methods & optimization

1 Kernel methods & optimization Machine Learning Class Notes 9-26-13 Prof. David Sontag 1 Kernel methods & optimization One eample of a kernel that is frequently used in practice and which allows for highly non-linear discriminant functions

More information

3 (Maths) Linear Algebra

3 (Maths) Linear Algebra 3 (Maths) Linear Algebra References: Simon and Blume, chapters 6 to 11, 16 and 23; Pemberton and Rau, chapters 11 to 13 and 25; Sundaram, sections 1.3 and 1.5. The methods and concepts of linear algebra

More information

September Math Course: First Order Derivative

September Math Course: First Order Derivative September Math Course: First Order Derivative Arina Nikandrova Functions Function y = f (x), where x is either be a scalar or a vector of several variables (x,..., x n ), can be thought of as a rule which

More information

Eigenvalues and diagonalization

Eigenvalues and diagonalization Eigenvalues and diagonalization Patrick Breheny November 15 Patrick Breheny BST 764: Applied Statistical Modeling 1/20 Introduction The next topic in our course, principal components analysis, revolves

More information

Eigenvalues, Eigenvectors, and Diagonalization

Eigenvalues, Eigenvectors, and Diagonalization Math 240 TA: Shuyi Weng Winter 207 February 23, 207 Eigenvalues, Eigenvectors, and Diagonalization The concepts of eigenvalues, eigenvectors, and diagonalization are best studied with examples. We will

More information

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Remark By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero. Sec 6 Eigenvalues and Eigenvectors Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called an eigenvalue of A if there is a nontrivial

More information

Chap 3. Linear Algebra

Chap 3. Linear Algebra Chap 3. Linear Algebra Outlines 1. Introduction 2. Basis, Representation, and Orthonormalization 3. Linear Algebraic Equations 4. Similarity Transformation 5. Diagonal Form and Jordan Form 6. Functions

More information

EC5555 Economics Masters Refresher Course in Mathematics September 2013

EC5555 Economics Masters Refresher Course in Mathematics September 2013 EC5555 Economics Masters Reresher Course in Mathematics September 3 Lecture 5 Unconstraine Optimization an Quaratic Forms Francesco Feri We consier the unconstraine optimization or the case o unctions

More information

Chapter 3 Transformations

Chapter 3 Transformations Chapter 3 Transformations An Introduction to Optimization Spring, 2014 Wei-Ta Chu 1 Linear Transformations A function is called a linear transformation if 1. for every and 2. for every If we fix the bases

More information

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors

Chapter 7. Canonical Forms. 7.1 Eigenvalues and Eigenvectors Chapter 7 Canonical Forms 7.1 Eigenvalues and Eigenvectors Definition 7.1.1. Let V be a vector space over the field F and let T be a linear operator on V. An eigenvalue of T is a scalar λ F such that there

More information

CS 143 Linear Algebra Review

CS 143 Linear Algebra Review CS 143 Linear Algebra Review Stefan Roth September 29, 2003 Introductory Remarks This review does not aim at mathematical rigor very much, but instead at ease of understanding and conciseness. Please see

More information

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra. DS-GA 1002 Lecture notes 0 Fall 2016 Linear Algebra These notes provide a review of basic concepts in linear algebra. 1 Vector spaces You are no doubt familiar with vectors in R 2 or R 3, i.e. [ ] 1.1

More information

Linear Algebra - Part II

Linear Algebra - Part II Linear Algebra - Part II Projection, Eigendecomposition, SVD (Adapted from Sargur Srihari s slides) Brief Review from Part 1 Symmetric Matrix: A = A T Orthogonal Matrix: A T A = AA T = I and A 1 = A T

More information

1 Last time: least-squares problems

1 Last time: least-squares problems MATH Linear algebra (Fall 07) Lecture Last time: least-squares problems Definition. If A is an m n matrix and b R m, then a least-squares solution to the linear system Ax = b is a vector x R n such that

More information

MATH 583A REVIEW SESSION #1

MATH 583A REVIEW SESSION #1 MATH 583A REVIEW SESSION #1 BOJAN DURICKOVIC 1. Vector Spaces Very quick review of the basic linear algebra concepts (see any linear algebra textbook): (finite dimensional) vector space (or linear space),

More information

Chiang/Wainwright: Fundamental Methods of Mathematical Economics

Chiang/Wainwright: Fundamental Methods of Mathematical Economics Chiang/Wainwright: Fundamental Methods of Mathematical Economics CHAPTER 9 EXERCISE 9.. Find the stationary values of the following (check whether they are relative maima or minima or inflection points),

More information

f'(x) = x 4 (2)(x - 6)(1) + (x - 6) 2 (4x 3 ) f'(x) = (x - 2) -1/3 = x 2 ; domain of f: (-, ) f'(x) = (x2 + 1)4x! 2x 2 (2x) 4x f'(x) =

f'(x) = x 4 (2)(x - 6)(1) + (x - 6) 2 (4x 3 ) f'(x) = (x - 2) -1/3 = x 2 ; domain of f: (-, ) f'(x) = (x2 + 1)4x! 2x 2 (2x) 4x f'(x) = 85. f() = 4 ( - 6) 2 f'() = 4 (2)( - 6)(1) + ( - 6) 2 (4 3 ) = 2 3 ( - 6)[ + 2( - 6)] = 2 3 ( - 6)(3-12) = 6 3 ( - 4)( - 6) Thus, the critical values are = 0, = 4, and = 6. Now we construct the sign chart

More information

ARE211, Fall 2005 CONTENTS. 5. Characteristics of Functions Surjective, Injective and Bijective functions. 5.2.

ARE211, Fall 2005 CONTENTS. 5. Characteristics of Functions Surjective, Injective and Bijective functions. 5.2. ARE211, Fall 2005 LECTURE #18: THU, NOV 3, 2005 PRINT DATE: NOVEMBER 22, 2005 (COMPSTAT2) CONTENTS 5. Characteristics of Functions. 1 5.1. Surjective, Injective and Bijective functions 1 5.2. Homotheticity

More information

0.1. Linear transformations

0.1. Linear transformations Suggestions for midterm review #3 The repetitoria are usually not complete; I am merely bringing up the points that many people didn t now on the recitations Linear transformations The following mostly

More information

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016

1 Overview. 2 A Characterization of Convex Functions. 2.1 First-order Taylor approximation. AM 221: Advanced Optimization Spring 2016 AM 221: Advanced Optimization Spring 2016 Prof. Yaron Singer Lecture 8 February 22nd 1 Overview In the previous lecture we saw characterizations of optimality in linear optimization, and we reviewed the

More information

8 Extremal Values & Higher Derivatives

8 Extremal Values & Higher Derivatives 8 Extremal Values & Higher Derivatives Not covered in 2016\17 81 Higher Partial Derivatives For functions of one variable there is a well-known test for the nature of a critical point given by the sign

More information

Matrices and Linear transformations

Matrices and Linear transformations Matrices and Linear transformations We have been thinking of matrices in connection with solutions to linear systems of equations like Ax = b. It is time to broaden our horizons a bit and start thinking

More information

4.3 - How Derivatives Affect the Shape of a Graph

4.3 - How Derivatives Affect the Shape of a Graph 4.3 - How Derivatives Affect the Shape of a Graph 1. Increasing and Decreasing Functions Definition: A function f is (strictly) increasing on an interval I if for every 1, in I with 1, f 1 f. A function

More information

5 More on Linear Algebra

5 More on Linear Algebra 14.102, Math for Economists Fall 2004 Lecture Notes, 9/23/2004 These notes are primarily based on those written by George Marios Angeletos for the Harvard Math Camp in 1999 and 2000, and updated by Stavros

More information

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero.

Remark 1 By definition, an eigenvector must be a nonzero vector, but eigenvalue could be zero. Sec 5 Eigenvectors and Eigenvalues In this chapter, vector means column vector Definition An eigenvector of an n n matrix A is a nonzero vector x such that A x λ x for some scalar λ A scalar λ is called

More information

Basic Calculus Review

Basic Calculus Review Basic Calculus Review Lorenzo Rosasco ISML Mod. 2 - Machine Learning Vector Spaces Functionals and Operators (Matrices) Vector Space A vector space is a set V with binary operations +: V V V and : R V

More information

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag.

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag. San Francisco State University Math Review Notes Michael Bar Sets A set is any collection of elements Eamples: a A {,,4,6,8,} - the set of even numbers between zero and b B { red, white, bule} - the set

More information

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP)

MATH 20F: LINEAR ALGEBRA LECTURE B00 (T. KEMP) MATH 20F: LINEAR ALGEBRA LECTURE B00 (T KEMP) Definition 01 If T (x) = Ax is a linear transformation from R n to R m then Nul (T ) = {x R n : T (x) = 0} = Nul (A) Ran (T ) = {Ax R m : x R n } = {b R m

More information

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review

10.34 Numerical Methods Applied to Chemical Engineering Fall Quiz #1 Review 10.34 Numerical Methods Applied to Chemical Engineering Fall 2015 Quiz #1 Review Study guide based on notes developed by J.A. Paulson, modified by K. Severson Linear Algebra We ve covered three major topics

More information

MATH 221, Spring Homework 10 Solutions

MATH 221, Spring Homework 10 Solutions MATH 22, Spring 28 - Homework Solutions Due Tuesday, May Section 52 Page 279, Problem 2: 4 λ A λi = and the characteristic polynomial is det(a λi) = ( 4 λ)( λ) ( )(6) = λ 6 λ 2 +λ+2 The solutions to the

More information

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION

MATH 1120 (LINEAR ALGEBRA 1), FINAL EXAM FALL 2011 SOLUTIONS TO PRACTICE VERSION MATH (LINEAR ALGEBRA ) FINAL EXAM FALL SOLUTIONS TO PRACTICE VERSION Problem (a) For each matrix below (i) find a basis for its column space (ii) find a basis for its row space (iii) determine whether

More information

Lecture 7: Positive Semidefinite Matrices

Lecture 7: Positive Semidefinite Matrices Lecture 7: Positive Semidefinite Matrices Rajat Mittal IIT Kanpur The main aim of this lecture note is to prepare your background for semidefinite programming. We have already seen some linear algebra.

More information

Functional Analysis Review

Functional Analysis Review Outline 9.520: Statistical Learning Theory and Applications February 8, 2010 Outline 1 2 3 4 Vector Space Outline A vector space is a set V with binary operations +: V V V and : R V V such that for all

More information

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by

MATH SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER Given vector spaces V and W, V W is the vector space given by MATH 110 - SOLUTIONS TO PRACTICE MIDTERM LECTURE 1, SUMMER 2009 GSI: SANTIAGO CAÑEZ 1. Given vector spaces V and W, V W is the vector space given by V W = {(v, w) v V and w W }, with addition and scalar

More information

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each

More information

Convex Functions and Optimization

Convex Functions and Optimization Chapter 5 Convex Functions and Optimization 5.1 Convex Functions Our next topic is that of convex functions. Again, we will concentrate on the context of a map f : R n R although the situation can be generalized

More information

Econ Slides from Lecture 7

Econ Slides from Lecture 7 Econ 205 Sobel Econ 205 - Slides from Lecture 7 Joel Sobel August 31, 2010 Linear Algebra: Main Theory A linear combination of a collection of vectors {x 1,..., x k } is a vector of the form k λ ix i for

More information

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r

Then x 1,..., x n is a basis as desired. Indeed, it suffices to verify that it spans V, since n = dim(v ). We may write any v V as r Practice final solutions. I did not include definitions which you can find in Axler or in the course notes. These solutions are on the terse side, but would be acceptable in the final. However, if you

More information

4 Matrix Diagonalization and Eigensystems

4 Matrix Diagonalization and Eigensystems 14.102, Math for Economists Fall 2004 Lecture Notes, 9/21/2004 These notes are primarily based on those written by George Marios Angeletos for the Harvard Math Camp in 1999 and 2000, and updated by Stavros

More information

Recall the convention that, for us, all vectors are column vectors.

Recall the convention that, for us, all vectors are column vectors. Some linear algebra Recall the convention that, for us, all vectors are column vectors. 1. Symmetric matrices Let A be a real matrix. Recall that a complex number λ is an eigenvalue of A if there exists

More information

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY

STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY STATIC LECTURE 4: CONSTRAINED OPTIMIZATION II - KUHN TUCKER THEORY UNIVERSITY OF MARYLAND: ECON 600 1. Some Eamples 1 A general problem that arises countless times in economics takes the form: (Verbally):

More information

Multivariate Statistical Analysis

Multivariate Statistical Analysis Multivariate Statistical Analysis Fall 2011 C. L. Williams, Ph.D. Lecture 4 for Applied Multivariate Analysis Outline 1 Eigen values and eigen vectors Characteristic equation Some properties of eigendecompositions

More information

Spectral Theorem for Self-adjoint Linear Operators

Spectral Theorem for Self-adjoint Linear Operators Notes for the undergraduate lecture by David Adams. (These are the notes I would write if I was teaching a course on this topic. I have included more material than I will cover in the 45 minute lecture;

More information

Matrix Vector Products

Matrix Vector Products We covered these notes in the tutorial sessions I strongly recommend that you further read the presented materials in classical books on linear algebra Please make sure that you understand the proofs and

More information

21 Symmetric and skew-symmetric matrices

21 Symmetric and skew-symmetric matrices 21 Symmetric and skew-symmetric matrices 21.1 Decomposition of a square matrix into symmetric and skewsymmetric matrices Let C n n be a square matrix. We can write C = (1/2)(C + C t ) + (1/2)(C C t ) =

More information

Calculating determinants for larger matrices

Calculating determinants for larger matrices Day 26 Calculating determinants for larger matrices We now proceed to define det A for n n matrices A As before, we are looking for a function of A that satisfies the product formula det(ab) = det A det

More information

1 Invariant subspaces

1 Invariant subspaces MATH 2040 Linear Algebra II Lecture Notes by Martin Li Lecture 8 Eigenvalues, eigenvectors and invariant subspaces 1 In previous lectures we have studied linear maps T : V W from a vector space V to another

More information

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET

IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET IMPORTANT DEFINITIONS AND THEOREMS REFERENCE SHEET This is a (not quite comprehensive) list of definitions and theorems given in Math 1553. Pay particular attention to the ones in red. Study Tip For each

More information

Transpose & Dot Product

Transpose & Dot Product Transpose & Dot Product Def: The transpose of an m n matrix A is the n m matrix A T whose columns are the rows of A. So: The columns of A T are the rows of A. The rows of A T are the columns of A. Example:

More information

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm

University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm University of Colorado at Denver Mathematics Department Applied Linear Algebra Preliminary Exam With Solutions 16 January 2009, 10:00 am 2:00 pm Name: The proctor will let you read the following conditions

More information

Functions of Several Variables

Functions of Several Variables Functions of Several Variables The Unconstrained Minimization Problem where In n dimensions the unconstrained problem is stated as f() x variables. minimize f()x x, is a scalar objective function of vector

More information

The Kuhn-Tucker and Envelope Theorems

The Kuhn-Tucker and Envelope Theorems The Kuhn-Tucker and Envelope Theorems Peter Ireland ECON 77200 - Math for Economists Boston College, Department of Economics Fall 207 The Kuhn-Tucker and envelope theorems can be used to characterize the

More information

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2

EC /11. Math for Microeconomics September Course, Part II Problem Set 1 with Solutions. a11 a 12. x 2 LONDON SCHOOL OF ECONOMICS Professor Leonardo Felli Department of Economics S.478; x7525 EC400 2010/11 Math for Microeconomics September Course, Part II Problem Set 1 with Solutions 1. Show that the general

More information

1. In this problem, if the statement is always true, circle T; otherwise, circle F.

1. In this problem, if the statement is always true, circle T; otherwise, circle F. Math 1553, Extra Practice for Midterm 3 (sections 45-65) Solutions 1 In this problem, if the statement is always true, circle T; otherwise, circle F a) T F If A is a square matrix and the homogeneous equation

More information

Stat 159/259: Linear Algebra Notes

Stat 159/259: Linear Algebra Notes Stat 159/259: Linear Algebra Notes Jarrod Millman November 16, 2015 Abstract These notes assume you ve taken a semester of undergraduate linear algebra. In particular, I assume you are familiar with the

More information

18.303: Introduction to Green s functions and operator inverses

18.303: Introduction to Green s functions and operator inverses 8.33: Introduction to Green s functions and operator inverses S. G. Johnson October 9, 2 Abstract In analogy with the inverse A of a matri A, we try to construct an analogous inverse  of differential

More information

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.]

[Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty.] Math 43 Review Notes [Disclaimer: This is not a complete list of everything you need to know, just some of the topics that gave people difficulty Dot Product If v (v, v, v 3 and w (w, w, w 3, then the

More information

Transpose & Dot Product

Transpose & Dot Product Transpose & Dot Product Def: The transpose of an m n matrix A is the n m matrix A T whose columns are the rows of A. So: The columns of A T are the rows of A. The rows of A T are the columns of A. Example:

More information

MATH 1553 PRACTICE MIDTERM 3 (VERSION B)

MATH 1553 PRACTICE MIDTERM 3 (VERSION B) MATH 1553 PRACTICE MIDTERM 3 (VERSION B) Name Section 1 2 3 4 5 Total Please read all instructions carefully before beginning. Each problem is worth 10 points. The maximum score on this exam is 50 points.

More information

Foundations of Computer Vision

Foundations of Computer Vision Foundations of Computer Vision Wesley. E. Snyder North Carolina State University Hairong Qi University of Tennessee, Knoxville Last Edited February 8, 2017 1 3.2. A BRIEF REVIEW OF LINEAR ALGEBRA Apply

More information

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is,

A = 3 1. We conclude that the algebraic multiplicity of the eigenvalues are both one, that is, 65 Diagonalizable Matrices It is useful to introduce few more concepts, that are common in the literature Definition 65 The characteristic polynomial of an n n matrix A is the function p(λ) det(a λi) Example

More information

Eigenspaces and Diagonalizable Transformations

Eigenspaces and Diagonalizable Transformations Chapter 2 Eigenspaces and Diagonalizable Transformations As we explored how heat states evolve under the action of a diffusion transformation E, we found that some heat states will only change in amplitude.

More information

In economics, the amount of a good x demanded is a function of the price of that good. In other words,

In economics, the amount of a good x demanded is a function of the price of that good. In other words, I. UNIVARIATE CALCULUS Given two sets X and Y, a function is a rule that associates each member of X with eactly one member of Y. That is, some goes in, and some y comes out. These notations are used to

More information

Intro to Nonlinear Optimization

Intro to Nonlinear Optimization Intro to Nonlinear Optimization We now rela the proportionality and additivity assumptions of LP What are the challenges of nonlinear programs NLP s? Objectives and constraints can use any function: ma

More information

Lecture 8 : Eigenvalues and Eigenvectors

Lecture 8 : Eigenvalues and Eigenvectors CPS290: Algorithmic Foundations of Data Science February 24, 2017 Lecture 8 : Eigenvalues and Eigenvectors Lecturer: Kamesh Munagala Scribe: Kamesh Munagala Hermitian Matrices It is simpler to begin with

More information

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators.

MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. MATH 423 Linear Algebra II Lecture 33: Diagonalization of normal operators. Adjoint operator and adjoint matrix Given a linear operator L on an inner product space V, the adjoint of L is a transformation

More information

Nonlinear Programming (NLP)

Nonlinear Programming (NLP) Natalia Lazzati Mathematics for Economics (Part I) Note 6: Nonlinear Programming - Unconstrained Optimization Note 6 is based on de la Fuente (2000, Ch. 7), Madden (1986, Ch. 3 and 5) and Simon and Blume

More information

Preliminary draft only: please check for final version

Preliminary draft only: please check for final version ARE211, Fall2012 CALCULUS4: THU, OCT 11, 2012 PRINTED: AUGUST 22, 2012 (LEC# 15) Contents 3. Univariate and Multivariate Differentiation (cont) 1 3.6. Taylor s Theorem (cont) 2 3.7. Applying Taylor theory:

More information

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as

Here each term has degree 2 (the sum of exponents is 2 for all summands). A quadratic form of three variables looks as Reading [SB], Ch. 16.1-16.3, p. 375-393 1 Quadratic Forms A quadratic function f : R R has the form f(x) = a x. Generalization of this notion to two variables is the quadratic form Q(x 1, x ) = a 11 x

More information

The Kuhn-Tucker and Envelope Theorems

The Kuhn-Tucker and Envelope Theorems The Kuhn-Tucker and Envelope Theorems Peter Ireland EC720.01 - Math for Economists Boston College, Department of Economics Fall 2010 The Kuhn-Tucker and envelope theorems can be used to characterize the

More information

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015

Solutions to Final Practice Problems Written by Victoria Kala Last updated 12/5/2015 Solutions to Final Practice Problems Written by Victoria Kala vtkala@math.ucsb.edu Last updated /5/05 Answers This page contains answers only. See the following pages for detailed solutions. (. (a x. See

More information

Symmetric Matrices and Eigendecomposition

Symmetric Matrices and Eigendecomposition Symmetric Matrices and Eigendecomposition Robert M. Freund January, 2014 c 2014 Massachusetts Institute of Technology. All rights reserved. 1 2 1 Symmetric Matrices and Convexity of Quadratic Functions

More information

k is a product of elementary matrices.

k is a product of elementary matrices. Mathematics, Spring Lecture (Wilson) Final Eam May, ANSWERS Problem (5 points) (a) There are three kinds of elementary row operations and associated elementary matrices. Describe what each kind of operation

More information

Eigenvalues for Triangular Matrices. ENGI 7825: Linear Algebra Review Finding Eigenvalues and Diagonalization

Eigenvalues for Triangular Matrices. ENGI 7825: Linear Algebra Review Finding Eigenvalues and Diagonalization Eigenvalues for Triangular Matrices ENGI 78: Linear Algebra Review Finding Eigenvalues and Diagonalization Adapted from Notes Developed by Martin Scharlemann June 7, 04 The eigenvalues for a triangular

More information

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes

22.3. Repeated Eigenvalues and Symmetric Matrices. Introduction. Prerequisites. Learning Outcomes Repeated Eigenvalues and Symmetric Matrices. Introduction In this Section we further develop the theory of eigenvalues and eigenvectors in two distinct directions. Firstly we look at matrices where one

More information

Geometric Modeling Summer Semester 2010 Mathematical Tools (1)

Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Geometric Modeling Summer Semester 2010 Mathematical Tools (1) Recap: Linear Algebra Today... Topics: Mathematical Background Linear algebra Analysis & differential geometry Numerical techniques Geometric

More information

A Basic Course in Real Analysis Prof. P. D. Srivastava Department of Mathematics Indian Institute of Technology, Kharagpur

A Basic Course in Real Analysis Prof. P. D. Srivastava Department of Mathematics Indian Institute of Technology, Kharagpur A Basic Course in Real Analysis Prof. P. D. Srivastava Department of Mathematics Indian Institute of Technology, Kharagpur Lecture - 36 Application of MVT, Darbou Theorem, L Hospital Rule (Refer Slide

More information

Preliminary draft only: please check for final version

Preliminary draft only: please check for final version ARE, Fall CALCULUS: THU, OCT 4, PRINTED: AUGUST, LEC# 3) Contents 3. Univariate and Multivariate Differentiation cont) 3.3. Partial, Directional and Total derivatives 3.3.. Directional derivatives 3.3..

More information

MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS

MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS MATH 221: SOLUTIONS TO SELECTED HOMEWORK PROBLEMS 1. HW 1: Due September 4 1.1.21. Suppose v, w R n and c is a scalar. Prove that Span(v + cw, w) = Span(v, w). We must prove two things: that every element

More information

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then

Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then Lecture 2 - Unconstrained Optimization Definition[Global Minimum and Maximum]Let f : S R be defined on a set S R n. Then 1. x S is a global minimum point of f over S if f (x) f (x ) for any x S. 2. x S

More information

Study Notes on Matrices & Determinants for GATE 2017

Study Notes on Matrices & Determinants for GATE 2017 Study Notes on Matrices & Determinants for GATE 2017 Matrices and Determinates are undoubtedly one of the most scoring and high yielding topics in GATE. At least 3-4 questions are always anticipated from

More information

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix

DIAGONALIZATION. In order to see the implications of this definition, let us consider the following example Example 1. Consider the matrix DIAGONALIZATION Definition We say that a matrix A of size n n is diagonalizable if there is a basis of R n consisting of eigenvectors of A ie if there are n linearly independent vectors v v n such that

More information

Lecture 8: Linear Algebra Background

Lecture 8: Linear Algebra Background CSE 521: Design and Analysis of Algorithms I Winter 2017 Lecture 8: Linear Algebra Background Lecturer: Shayan Oveis Gharan 2/1/2017 Scribe: Swati Padmanabhan Disclaimer: These notes have not been subjected

More information

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017

Math 4A Notes. Written by Victoria Kala Last updated June 11, 2017 Math 4A Notes Written by Victoria Kala vtkala@math.ucsb.edu Last updated June 11, 2017 Systems of Linear Equations A linear equation is an equation that can be written in the form a 1 x 1 + a 2 x 2 +...

More information

A Brief Outline of Math 355

A Brief Outline of Math 355 A Brief Outline of Math 355 Lecture 1 The geometry of linear equations; elimination with matrices A system of m linear equations with n unknowns can be thought of geometrically as m hyperplanes intersecting

More information

DM554 Linear and Integer Programming. Lecture 9. Diagonalization. Marco Chiarandini

DM554 Linear and Integer Programming. Lecture 9. Diagonalization. Marco Chiarandini DM554 Linear and Integer Programming Lecture 9 Marco Chiarandini Department of Mathematics & Computer Science University of Southern Denmark Outline 1. More on 2. 3. 2 Resume Linear transformations and

More information

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS

LINEAR ALGEBRA 1, 2012-I PARTIAL EXAM 3 SOLUTIONS TO PRACTICE PROBLEMS LINEAR ALGEBRA, -I PARTIAL EXAM SOLUTIONS TO PRACTICE PROBLEMS Problem (a) For each of the two matrices below, (i) determine whether it is diagonalizable, (ii) determine whether it is orthogonally diagonalizable,

More information

Eigenvectors and Hermitian Operators

Eigenvectors and Hermitian Operators 7 71 Eigenvalues and Eigenvectors Basic Definitions Let L be a linear operator on some given vector space V A scalar λ and a nonzero vector v are referred to, respectively, as an eigenvalue and corresponding

More information

Math 3191 Applied Linear Algebra

Math 3191 Applied Linear Algebra Math 9 Applied Linear Algebra Lecture 9: Diagonalization Stephen Billups University of Colorado at Denver Math 9Applied Linear Algebra p./9 Section. Diagonalization The goal here is to develop a useful

More information

B553 Lecture 5: Matrix Algebra Review

B553 Lecture 5: Matrix Algebra Review B553 Lecture 5: Matrix Algebra Review Kris Hauser January 19, 2012 We have seen in prior lectures how vectors represent points in R n and gradients of functions. Matrices represent linear transformations

More information

Vector Spaces. Addition : R n R n R n Scalar multiplication : R R n R n.

Vector Spaces. Addition : R n R n R n Scalar multiplication : R R n R n. Vector Spaces Definition: The usual addition and scalar multiplication of n-tuples x = (x 1,..., x n ) R n (also called vectors) are the addition and scalar multiplication operations defined component-wise:

More information

Fredholm Theory. April 25, 2018

Fredholm Theory. April 25, 2018 Fredholm Theory April 25, 208 Roughly speaking, Fredholm theory consists of the study of operators of the form I + A where A is compact. From this point on, we will also refer to I + A as Fredholm operators.

More information

LU Factorization. A m x n matrix A admits an LU factorization if it can be written in the form of A = LU

LU Factorization. A m x n matrix A admits an LU factorization if it can be written in the form of A = LU LU Factorization A m n matri A admits an LU factorization if it can be written in the form of Where, A = LU L : is a m m lower triangular matri with s on the diagonal. The matri L is invertible and is

More information

Eigenvalues, Eigenvectors, and Diagonalization

Eigenvalues, Eigenvectors, and Diagonalization Week12 Eigenvalues, Eigenvectors, and Diagonalization 12.1 Opening Remarks 12.1.1 Predicting the Weather, Again View at edx Let us revisit the example from Week 4, in which we had a simple model for predicting

More information

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form:

3.3.1 Linear functions yet again and dot product In 2D, a homogenous linear scalar function takes the general form: 3.3 Gradient Vector and Jacobian Matri 3 3.3 Gradient Vector and Jacobian Matri Overview: Differentiable functions have a local linear approimation. Near a given point, local changes are determined by

More information