Optimization. Nuno Vasconcelos ECE Department, UCSD

Size: px
Start display at page:

Download "Optimization. Nuno Vasconcelos ECE Department, UCSD"

Transcription

1 Optmzaton Nuno Vasconcelos ECE Department, UCSD

2 Optmzaton many engneerng problems bol on to optmzaton goal: n mamum or mnmum o a uncton Denton: gven unctons, g,,...,k an h,,...m ene on some oman Ω R n mn subject to, Ω g, h, : cost; h equalty, g nequalty: constrants or compactness e rte g nstea o g,. Smlarly h note that no nee or

3 Optmzaton note: mamzng s the same as mnmzng, ths enton also orks or mamzaton the easble regon s the regon here. s ene an all constrants hol R { Ω g, h } * s a global mnmum o *, Ω * s a local mnmum o ε > s.t. * < ε * local global 3

4 he graent the graent o a uncton at z s h th t t n z z z,, L heorem: the graent ponts n the recton o mamum groth proo: proo: rom aylor seres epanson α α α O ervatve along,.cos. lm α α α * 4 s mamum hen s n the recton o the graent

5 he graent note that there s no recton o groth also -, an there s no recton o ecrease e are ether at a local mnmum or mamum or sale pont conversely, at local mn or ma or sale pont no recton o groth or ecrease ths shos that e have a crtcal pont an only to etermne hch type e nee secon orer contons ma mn sale 5

6 ma he Hessan, by aylor seres α an 3 α α α α O 3 an α α α O mn pck α such that Oα <<, mamum at an only, mnmum at an only sale mnmum at an only, sale otherse ths proves the ollong theorems 6 ths proves the ollong theorems

7 Mnma contons unconstrane let be contnuously erentable * s a local mnmum o an only s a local mnmum o an only has zero graent at * * an the Hessan o at * s postve ente n t R * here R, n M L 7 n n L

8 Mama contons unconstrane let be contnuously erentable * s a local mamum o an only s a local mamum o an only has zero graent at * * an the Hessan o at * s negatve ente n t R * here R, n M L 8 n n L

9 Eample conser the unctons g g the graents are has no mnma or mama, g g has a crtcal pont at the orgn, snce Hessan s postve ente, ths s a mnmum g 9

10 Eample makes sense because s a plane, graent s constant so-contours o -

11 Eample makes sense because g g s a quaratc, postve everyhere but the orgn note ho graent ponts toars largest ncrease g p g g h h h g h

12 Conve unctons Denton: s conve,u Ωan λ [,] λ λ u λ λ u heorem: s conve an only ts Hessan s postve ente t or all t, Ω proo: λ-λv u requres some ntermeate results that e ll not cover λ-λv u e ll skp t λ-λv

13 Concave unctons Denton: s concave,u Ωan λ [,] λ λ u λ λ u heorem: s concave an only ts Hessan s negatve ente t or all t, Ω proo: - s conve by prevous theorem, Hessan s negatve ente Hessan o s postve ente 3

14 Conve unctons heorem: s conve any local mnmum * s also a global mnmum Proo: e nee to sho that, or any u, * u or any u: *-[λ*-λu -λ *-u an, makng λ arbtrarly close to, e can make *-[λ*-λu ε, or any ε > snce * s local mnmum, t ollos that * λ*-λu an, by convety, that * λ*-λu λu or *-λ u-λ an * u 4

15 Constrane optmzaton n summary: e kno hat are contons or unconstrane ma an mn e lke conve unctons n a mnma, t ll be global mnmum hat about optmzaton th constrants? a e entons to start th nequalty g : s actve g, otherse nactve nequaltes can be epresse as equaltes by ntroucton o slack varables g g ξ, an ξ 5

16 Conve optmzaton Denton: a set Ω s conve,u Ωan λ [,] then λ-λu Ω a lne beteen any to ponts n Ω s also n Ω conve not conve Denton: an optmzaton problem here the set Ω, the cost an all constrants g an h are conve s sa to be conve note: lnear constrants g Ab are alays conve zero Hessan 6

17 Constrane optmzaton e ll conser general not only conve constrane optmzaton problems, start by case th only equaltes heorem: conser the problem * arg mn subject to h here the constrant graents h * are lnearly nepenent. hen * s a soluton an only there ets a unque vector λ, such that m λ * λ h * y * m λ h * y, y s.t. h * y 7

18 Alternatve ormulaton state the contons through the Lagrangan L, λ m λ h the theorem can be compactly rtten as * L *, λ y λ L *, * λ * L *, λ y, y s.t. h * y the entres o λ are reerre to as Lagrange multplers 8

19 Graent revste recall rom * that ervatve o along s lm α α α..cos, ths means thatt greatest ncrease hen no ncrease hen snce there s no ncrease hen s tangent to so-contour k the graent s perpencular to the tangent o the so-contour ths suggests a geometrc proo no ncrease 9

20 Lagrangan optmzaton geometrc nterpretaton: snce h s a so-contour o h, h* s perpencular to the so-contour says that * span{h *}.e. to tangent space o the constrant surace ntutve recton o largest ncrease o s to constrant surace the graent s zero along the constrant no ay to gve an nntesmal graent step, thout enng up volatng t t s mpossble to ncrease an stll satsy the constrant h span{h*} tg plane

21 Eample conser the problem mn subject to t leas to the ollong pcture h so-contours o h -

22 Eample conser the problem mn subject to to the so-contours o k h h -

23 Eample conser the problem mn subject to h to the so-contour o h - h h - 3

24 Eample recall that ervatve along s α lm..cos, α α crtcal pont crtcal pont - movng along the tangent s escent as long as cos tg, < -.e. π/ < angle,tg < 3π/ - can alays n such unless tg - crtcal pont hen h - to n hch type e nee n orer as beore 4

25 Alternatve ve conser the tangent space to the so-contour h ths s the subspace o rst orer easble varatons { h * } V *, space o or hch satses the constrant up to rst orer appromaton V* easble varatons h * h* 5

26 Feasble varatons multplyng our rst Lagrangan conton by * λ h * t ollos that m *, V * ths s a generalzaton o * n unconstrane case mples that * V* an thereore * h* note: Hessan constrant only ene or y n V* makes sense: e cannot move anyhere else, oes not really matter hat Hessan s outse V* 6

27 In summary or a constrane optmzaton problem, th equalty constrants heorem: conser the problem * arg mn subject to h here the constrant graents h * are lnearly nepenent. hen * s a soluton an only there ets a unque vector λ, such that m λ * λ h * y * m λ h * y, y s.t. h * y 7

28 Alternatve ormulaton state the contons through the Lagrangan L, λ m λ h the theorem can be compactly rtten as * L *, λ y λ L *, * λ * L *, λ y, y s.t. h * y the entres o λ are reerre to as Lagrange multplers 8

29 9

The Karush-Kuhn-Tucker. Nuno Vasconcelos ECE Department, UCSD

The Karush-Kuhn-Tucker. Nuno Vasconcelos ECE Department, UCSD e Karus-Kun-ucker condtons and dualt Nuno Vasconcelos ECE Department, UCSD Optmzaton goal: nd mamum or mnmum o a uncton Denton: gven unctons, g, 1,...,k and, 1,...m dened on some doman Ω R n mn w, w Ω

More information

Mathematical Economics MEMF e ME. Filomena Garcia. Topic 2 Calculus

Mathematical Economics MEMF e ME. Filomena Garcia. Topic 2 Calculus Mathematcal Economcs MEMF e ME Flomena Garca Topc 2 Calculus Mathematcal Economcs - www.seg.utl.pt/~garca/economa_matematca . Unvarate Calculus Calculus Functons : X Y y ( gves or each element X one element

More information

Chapter 7: Conservation of Energy

Chapter 7: Conservation of Energy Lecture 7: Conservaton o nergy Chapter 7: Conservaton o nergy Introucton I the quantty o a subject oes not change wth tme, t means that the quantty s conserve. The quantty o that subject remans constant

More information

Which Separator? Spring 1

Which Separator? Spring 1 Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal

More information

Support Vector Machines CS434

Support Vector Machines CS434 Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We

More information

OPTIMISATION. Introduction Single Variable Unconstrained Optimisation Multivariable Unconstrained Optimisation Linear Programming

OPTIMISATION. Introduction Single Variable Unconstrained Optimisation Multivariable Unconstrained Optimisation Linear Programming OPTIMIATION Introducton ngle Varable Unconstraned Optmsaton Multvarable Unconstraned Optmsaton Lnear Programmng Chapter Optmsaton /. Introducton In an engneerng analss, sometmes etremtes, ether mnmum or

More information

Pattern Classification

Pattern Classification Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher

More information

Shuai Dong. Isaac Newton. Gottfried Leibniz

Shuai Dong. Isaac Newton. Gottfried Leibniz Computatonal pyscs Sua Dong Isaac Newton Gottred Lebnz Numercal calculus poston dervatve ntegral v velocty dervatve ntegral a acceleraton Numercal calculus Numercal derentaton Numercal ntegraton Roots

More information

Linear discriminants. Nuno Vasconcelos ECE Department, UCSD

Linear discriminants. Nuno Vasconcelos ECE Department, UCSD Lnear dscrmnants Nuno Vasconcelos ECE Department UCSD Classfcaton a classfcaton problem as to tpes of varables e.g. X - vector of observatons features n te orld Y - state class of te orld X R 2 fever blood

More information

PHYS 705: Classical Mechanics. Calculus of Variations II

PHYS 705: Classical Mechanics. Calculus of Variations II 1 PHYS 705: Classcal Mechancs Calculus of Varatons II 2 Calculus of Varatons: Generalzaton (no constrant yet) Suppose now that F depends on several dependent varables : We need to fnd such that has a statonary

More information

Lagrange Multipliers Kernel Trick

Lagrange Multipliers Kernel Trick Lagrange Multplers Kernel Trck Ncholas Ruozz Unversty of Texas at Dallas Based roughly on the sldes of Davd Sontag General Optmzaton A mathematcal detour, we ll come back to SVMs soon! subject to: f x

More information

15-381: Artificial Intelligence. Regression and cross validation

15-381: Artificial Intelligence. Regression and cross validation 15-381: Artfcal Intellgence Regresson and cross valdaton Where e are Inputs Densty Estmator Probablty Inputs Classfer Predct category Inputs Regressor Predct real no. Today Lnear regresson Gven an nput

More information

T-Forward Method: A Closed-Form Solution and Polynomial Time Approach for Convex Nonlinear Programming

T-Forward Method: A Closed-Form Solution and Polynomial Time Approach for Convex Nonlinear Programming Algorthms Research 4, 3: -5 DOI:.593/j.algorthms.43. T-Forwar etho: A Close-Form Soluton an Polynomal Tme Approach or Conve onlnear Programmng Gang Lu Technology Research Department, acroronter, Elmhurst,

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that e have state of the orld X observatons decson functon L[,y] loss of predctn y th Bayes decson rule s the rule

More information

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017

U.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017 U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that

More information

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg

princeton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there

More information

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution.

Solutions HW #2. minimize. Ax = b. Give the dual problem, and make the implicit equality constraints explicit. Solution. Solutons HW #2 Dual of general LP. Fnd the dual functon of the LP mnmze subject to c T x Gx h Ax = b. Gve the dual problem, and make the mplct equalty constrants explct. Soluton. 1. The Lagrangan s L(x,

More information

COS 521: Advanced Algorithms Game Theory and Linear Programming

COS 521: Advanced Algorithms Game Theory and Linear Programming COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton

More information

Practical Newton s Method

Practical Newton s Method Practcal Newton s Method Lecture- n Newton s Method n Pure Newton s method converges radly once t s close to. It may not converge rom the remote startng ont he search drecton to be a descent drecton rue

More information

CHAPTER 7 CONSTRAINED OPTIMIZATION 2: SQP AND GRG

CHAPTER 7 CONSTRAINED OPTIMIZATION 2: SQP AND GRG Chapter 7: Constraned Optmzaton CHAPER 7 CONSRAINED OPIMIZAION : SQP AND GRG Introducton In the prevous chapter we eamned the necessary and suffcent condtons for a constraned optmum. We dd not, however,

More information

Ch. 7 Lagrangian and Hamiltonian dynamics Homework Problems 7-3, 7-7, 7-15, 7-16, 7-17, 7-18, 7-34, 7-37, where y'(x) dy dx Δ Δ Δ. f x.

Ch. 7 Lagrangian and Hamiltonian dynamics Homework Problems 7-3, 7-7, 7-15, 7-16, 7-17, 7-18, 7-34, 7-37, where y'(x) dy dx Δ Δ Δ. f x. Ch. 7 Laranan an Hamltonan namcs Homewor Problems 7-3 7-7 7-5 7-6 7-7 7-8 7-34 7-37 7-40 A. revew o calculus o varatons (Chapter 6. basc problem or J { ( '(; } where '( For e en ponts an ntereste n the

More information

MA209 Variational Principles

MA209 Variational Principles MA209 Varatonal Prncples June 3, 203 The course covers the bascs of the calculus of varatons, an erves the Euler-Lagrange equatons for mnmsng functonals of the type Iy) = fx, y, y )x. It then gves examples

More information

Complex Variables. Chapter 18 Integration in the Complex Plane. March 12, 2013 Lecturer: Shih-Yuan Chen

Complex Variables. Chapter 18 Integration in the Complex Plane. March 12, 2013 Lecturer: Shih-Yuan Chen omplex Varables hapter 8 Integraton n the omplex Plane March, Lecturer: Shh-Yuan hen Except where otherwse noted, content s lcensed under a BY-N-SA. TW Lcense. ontents ontour ntegrals auchy-goursat theorem

More information

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD

The Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s

More information

CHAPTER 4d. ROOTS OF EQUATIONS

CHAPTER 4d. ROOTS OF EQUATIONS CHAPTER 4d. ROOTS OF EQUATIONS A. J. Clark School o Engneerng Department o Cvl and Envronmental Engneerng by Dr. Ibrahm A. Assakka Sprng 00 ENCE 03 - Computaton Methods n Cvl Engneerng II Department o

More information

Lecture 10 Support Vector Machines. Oct

Lecture 10 Support Vector Machines. Oct Lecture 10 Support Vector Machnes Oct - 20-2008 Lnear Separators Whch of the lnear separators s optmal? Concept of Margn Recall that n Perceptron, we learned that the convergence rate of the Perceptron

More information

Maximal Margin Classifier

Maximal Margin Classifier CS81B/Stat41B: Advanced Topcs n Learnng & Decson Makng Mamal Margn Classfer Lecturer: Mchael Jordan Scrbes: Jana van Greunen Corrected verson - /1/004 1 References/Recommended Readng 1.1 Webstes www.kernel-machnes.org

More information

Convex Optimization. Optimality conditions. (EE227BT: UC Berkeley) Lecture 9 (Optimality; Conic duality) 9/25/14. Laurent El Ghaoui.

Convex Optimization. Optimality conditions. (EE227BT: UC Berkeley) Lecture 9 (Optimality; Conic duality) 9/25/14. Laurent El Ghaoui. Convex Optmzaton (EE227BT: UC Berkeley) Lecture 9 (Optmalty; Conc dualty) 9/25/14 Laurent El Ghaou Organsatonal Mdterm: 10/7/14 (1.5 hours, n class, double-sded cheat sheet allowed) Project: Intal proposal

More information

Lecture 2 Solution of Nonlinear Equations ( Root Finding Problems )

Lecture 2 Solution of Nonlinear Equations ( Root Finding Problems ) Lecture Soluton o Nonlnear Equatons Root Fndng Problems Dentons Classcaton o Methods Analytcal Solutons Graphcal Methods Numercal Methods Bracketng Methods Open Methods Convergence Notatons Root Fndng

More information

Local Approximation of Pareto Surface

Local Approximation of Pareto Surface Proceedngs o the World Congress on Engneerng 007 Vol II Local Approxmaton o Pareto Surace S.V. Utyuzhnkov, J. Magnot, and M.D. Guenov Abstract In the desgn process o complex systems, the desgner s solvng

More information

Solutions to exam in SF1811 Optimization, Jan 14, 2015

Solutions to exam in SF1811 Optimization, Jan 14, 2015 Solutons to exam n SF8 Optmzaton, Jan 4, 25 3 3 O------O -4 \ / \ / The network: \/ where all lnks go from left to rght. /\ / \ / \ 6 O------O -5 2 4.(a) Let x = ( x 3, x 4, x 23, x 24 ) T, where the varable

More information

MMA and GCMMA two methods for nonlinear optimization

MMA and GCMMA two methods for nonlinear optimization MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons

More information

Chapter 3 Differentiation and Integration

Chapter 3 Differentiation and Integration MEE07 Computer Modelng Technques n Engneerng Chapter Derentaton and Integraton Reerence: An Introducton to Numercal Computatons, nd edton, S. yakowtz and F. zdarovsky, Mawell/Macmllan, 990. Derentaton

More information

Field and Wave Electromagnetic. Chapter.4

Field and Wave Electromagnetic. Chapter.4 Fel an Wave Electromagnetc Chapter.4 Soluton of electrostatc Problems Posson s s an Laplace s Equatons D = ρ E = E = V D = ε E : Two funamental equatons for electrostatc problem Where, V s scalar electrc

More information

CHAPTER 7 CONSTRAINED OPTIMIZATION 1: THE KARUSH-KUHN-TUCKER CONDITIONS

CHAPTER 7 CONSTRAINED OPTIMIZATION 1: THE KARUSH-KUHN-TUCKER CONDITIONS CHAPER 7 CONSRAINED OPIMIZAION : HE KARUSH-KUHN-UCKER CONDIIONS 7. Introducton We now begn our dscusson of gradent-based constraned optzaton. Recall that n Chapter 3 we looked at gradent-based unconstraned

More information

p(z) = 1 a e z/a 1(z 0) yi a i x (1/a) exp y i a i x a i=1 n i=1 (y i a i x) inf 1 (y Ax) inf Ax y (1 ν) y if A (1 ν) = 0 otherwise

p(z) = 1 a e z/a 1(z 0) yi a i x (1/a) exp y i a i x a i=1 n i=1 (y i a i x) inf 1 (y Ax) inf Ax y (1 ν) y if A (1 ν) = 0 otherwise Dustn Lennon Math 582 Convex Optmzaton Problems from Boy, Chapter 7 Problem 7.1 Solve the MLE problem when the nose s exponentally strbute wth ensty p(z = 1 a e z/a 1(z 0 The MLE s gven by the followng:

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

CHAPTER 6 CONSTRAINED OPTIMIZATION 1: K-T CONDITIONS

CHAPTER 6 CONSTRAINED OPTIMIZATION 1: K-T CONDITIONS Chapter 6: Constraned Optzaton CHAPER 6 CONSRAINED OPIMIZAION : K- CONDIIONS Introducton We now begn our dscusson of gradent-based constraned optzaton. Recall that n Chapter 3 we looked at gradent-based

More information

LECTURE NOTE III Chapter 4 Linear Programming

LECTURE NOTE III Chapter 4 Linear Programming Department of Chemcal and Bologcal Engneerng LECURE NOE III Chapter 4 Lnear Programmng Ref: Luenberger, D. G., "Lnear and Nonlnear Programmng,: 2nd Ed., Addson-Wesley, 1984 - Obectve functon and constrants

More information

The General Nonlinear Constrained Optimization Problem

The General Nonlinear Constrained Optimization Problem St back, relax, and enjoy the rde of your lfe as we explore the condtons that enable us to clmb to the top of a concave functon or descend to the bottom of a convex functon whle constraned wthn a closed

More information

Hard Problems from Advanced Partial Differential Equations (18.306)

Hard Problems from Advanced Partial Differential Equations (18.306) Har Problems from Avance Partal Dfferental Equatons (18.306) Kenny Kamrn June 27, 2004 1. We are gven the PDE 2 Ψ = Ψ xx + Ψ yy = 0. We must fn solutons of the form Ψ = x γ f (ξ), where ξ x/y. We also

More information

Polynomial Barrier Method for Solving Linear Programming Problems

Polynomial Barrier Method for Solving Linear Programming Problems Internatonal Journal o Engneerng & echnology IJE-IJENS Vol: No: 45 Polynoal Barrer Method or Solvng Lnear Prograng Probles Parwad Moengn, Meber, IAENG Abstract In ths wor, we study a class o polynoal ordereven

More information

Hopfield Training Rules 1 N

Hopfield Training Rules 1 N Hopfeld Tranng Rules To memorse a sngle pattern Suppose e set the eghts thus - = p p here, s the eght beteen nodes & s the number of nodes n the netor p s the value requred for the -th node What ll the

More information

Summary with Examples for Root finding Methods -Bisection -Newton Raphson -Secant

Summary with Examples for Root finding Methods -Bisection -Newton Raphson -Secant Summary wth Eamples or Root ndng Methods -Bsecton -Newton Raphson -Secant Nonlnear Equaton Solvers Bracketng Graphcal Open Methods Bsecton False Poston (Regula-Fals) Newton Raphson Secant All Iteratve

More information

Spring Force and Power

Spring Force and Power Lecture 13 Chapter 9 Sprng Force and Power Yeah, energy s better than orces. What s net? Course webste: http://aculty.uml.edu/andry_danylov/teachng/physcsi IN THIS CHAPTER, you wll learn how to solve problems

More information

Duality in linear programming

Duality in linear programming MPRA Munch Personal RePEc Archve Dualty n lnear programmng Mhaela Albc and Dela Teselos and Raluca Prundeanu and Ionela Popa Unversty Constantn Brancoveanu Ramncu Valcea 7 January 00 Onlne at http://mpraubun-muenchende/986/

More information

ρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to

ρ some λ THE INVERSE POWER METHOD (or INVERSE ITERATION) , for , or (more usually) to THE INVERSE POWER METHOD (or INVERSE ITERATION) -- applcaton of the Power method to A some fxed constant ρ (whch s called a shft), x λ ρ If the egenpars of A are { ( λ, x ) } ( ), or (more usually) to,

More information

,, MRTS is the marginal rate of technical substitution

,, MRTS is the marginal rate of technical substitution Mscellaneous Notes on roducton Economcs ompled by eter F Orazem September 9, 00 I Implcatons of conve soquants Two nput case, along an soquant 0 along an soquant Slope of the soquant,, MRTS s the margnal

More information

Global Optimization of Bilinear Generalized Disjunctive Programs

Global Optimization of Bilinear Generalized Disjunctive Programs Global Optmzaton o Blnear Generalzed Dsunctve Programs Juan Pablo Ruz Ignaco E. Grossmann Department o Chemcal Engneerng Center or Advanced Process Decson-mang Unversty Pttsburgh, PA 15213 1 Non-Convex

More information

PHZ 6607 Lecture Notes

PHZ 6607 Lecture Notes NOTE PHZ 6607 Lecture Notes 1. Lecture 2 1.1. Defntons Books: ( Tensor Analyss on Manfols ( The mathematcal theory of black holes ( Carroll (v Schutz Vector: ( In an N-Dmensonal space, a vector s efne

More information

e i is a random error

e i is a random error Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where + β + β e for,..., and are observable varables e s a random error How can an estmaton rule be constructed for the unknown

More information

EE 330 Lecture 24. Small Signal Analysis Small Signal Analysis of BJT Amplifier

EE 330 Lecture 24. Small Signal Analysis Small Signal Analysis of BJT Amplifier EE 0 Lecture 4 Small Sgnal Analss Small Sgnal Analss o BJT Ampler Eam Frda March 9 Eam Frda Aprl Revew Sesson or Eam : 6:00 p.m. on Thursda March 8 n Room Sweene 6 Revew rom Last Lecture Comparson o Gans

More information

Chapter 2 Transformations and Expectations. , and define f

Chapter 2 Transformations and Expectations. , and define f Revew for the prevous lecture Defnton: support set of a ranom varable, the monotone functon; Theorem: How to obtan a cf, pf (or pmf) of functons of a ranom varable; Eamples: several eamples Chapter Transformatons

More information

Scatter Plot x

Scatter Plot x Construct a scatter plot usng excel for the gven data. Determne whether there s a postve lnear correlaton, negatve lnear correlaton, or no lnear correlaton. Complete the table and fnd the correlaton coeffcent

More information

Proseminar Optimierung II. Victor A. Kovtunenko SS 2012/2013: LV

Proseminar Optimierung II. Victor A. Kovtunenko SS 2012/2013: LV Prosemnar Optmerung II Vctor A. Kovtunenko Insttute for Mathematcs and Scentfc Computng, Karl-Franzens Unversty of Graz, Henrchstr. 36, 8010 Graz, Austra; Lavrent ev Insttute of Hydrodynamcs, Sberan Dvson

More information

b ), which stands for uniform distribution on the interval a x< b. = 0 elsewhere

b ), which stands for uniform distribution on the interval a x< b. = 0 elsewhere Fall Analyss of Epermental Measurements B. Esensten/rev. S. Errede Some mportant probablty dstrbutons: Unform Bnomal Posson Gaussan/ormal The Unform dstrbuton s often called U( a, b ), hch stands for unform

More information

1.050 Content overview Engineering Mechanics I Content overview. Outline and goals. Lecture 28

1.050 Content overview Engineering Mechanics I Content overview. Outline and goals. Lecture 28 .5 Content overvew.5 Engneerng Mechancs I Lecture 8 Introucton: Energy bouns n lnear elastcty (cont I. Dmensonal analyss. On monsters, mce an mushrooms Lectures -. Smlarty relatons: Important engneerng

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.

Chapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise. Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the

More information

Competitive Experimentation and Private Information

Competitive Experimentation and Private Information Compettve Expermentaton an Prvate Informaton Guseppe Moscarn an Francesco Squntan Omtte Analyss not Submtte for Publcaton Dervatons for te Gamma-Exponental Moel Dervaton of expecte azar rates. By Bayes

More information

Support Vector Machines CS434

Support Vector Machines CS434 Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? + + + + + + + + + Intuton of Margn Consder ponts

More information

14 Lagrange Multipliers

14 Lagrange Multipliers Lagrange Multplers 14 Lagrange Multplers The Method of Lagrange Multplers s a powerful technque for constraned optmzaton. Whle t has applcatons far beyond machne learnng t was orgnally developed to solve

More information

Affine transformations and convexity

Affine transformations and convexity Affne transformatons and convexty The purpose of ths document s to prove some basc propertes of affne transformatons nvolvng convex sets. Here are a few onlne references for background nformaton: http://math.ucr.edu/

More information

SINGULARLY PERTURBED BOUNDARY VALUE. We consider the following singularly perturbed boundary value problem

SINGULARLY PERTURBED BOUNDARY VALUE. We consider the following singularly perturbed boundary value problem Cater PARAETRIC QUITIC SPLIE SOLUTIO OR SIGULARLY PERTURBED BOUDARY VALUE PROBLES Introcton We conser te ollong snglarly ertrbe bonary vale roblem " L r r > an ere s a small ostve arameter st

More information

( ) [ ( k) ( k) ( x) ( ) ( ) ( ) [ ] ξ [ ] [ ] [ ] ( )( ) i ( ) ( )( ) 2! ( ) = ( ) 3 Interpolation. Polynomial Approximation.

( ) [ ( k) ( k) ( x) ( ) ( ) ( ) [ ] ξ [ ] [ ] [ ] ( )( ) i ( ) ( )( ) 2! ( ) = ( ) 3 Interpolation. Polynomial Approximation. 3 Interpolaton {( y } Gven:,,,,,, [ ] Fnd: y for some Mn, Ma Polynomal Appromaton Theorem (Weerstrass Appromaton Theorem --- estence ε [ ab] f( P( , then there ests a polynomal

More information

COMPLEX NUMBERS AND QUADRATIC EQUATIONS

COMPLEX NUMBERS AND QUADRATIC EQUATIONS COMPLEX NUMBERS AND QUADRATIC EQUATIONS INTRODUCTION We know that x 0 for all x R e the square of a real number (whether postve, negatve or ero) s non-negatve Hence the equatons x, x, x + 7 0 etc are not

More information

a b a In case b 0, a being divisible by b is the same as to say that

a b a In case b 0, a being divisible by b is the same as to say that Secton 6.2 Dvsblty among the ntegers An nteger a ε s dvsble by b ε f there s an nteger c ε such that a = bc. Note that s dvsble by any nteger b, snce = b. On the other hand, a s dvsble by only f a = :

More information

CS 331 DESIGN AND ANALYSIS OF ALGORITHMS DYNAMIC PROGRAMMING. Dr. Daisy Tang

CS 331 DESIGN AND ANALYSIS OF ALGORITHMS DYNAMIC PROGRAMMING. Dr. Daisy Tang CS DESIGN ND NLYSIS OF LGORITHMS DYNMIC PROGRMMING Dr. Dasy Tang Dynamc Programmng Idea: Problems can be dvded nto stages Soluton s a sequence o decsons and the decson at the current stage s based on the

More information

18-660: Numerical Methods for Engineering Design and Optimization

18-660: Numerical Methods for Engineering Design and Optimization 8-66: Numercal Methods for Engneerng Desgn and Optmzaton n L Department of EE arnege Mellon Unversty Pttsburgh, PA 53 Slde Overve lassfcaton Support vector machne Regularzaton Slde lassfcaton Predct categorcal

More information

Mechanics Physics 151

Mechanics Physics 151 Mechancs Physcs 5 Lecture 3 Contnuous Systems an Fels (Chapter 3) Where Are We Now? We ve fnshe all the essentals Fnal wll cover Lectures through Last two lectures: Classcal Fel Theory Start wth wave equatons

More information

6) Derivatives, gradients and Hessian matrices

6) Derivatives, gradients and Hessian matrices 30C00300 Mathematcal Methods for Economsts (6 cr) 6) Dervatves, gradents and Hessan matrces Smon & Blume chapters: 14, 15 Sldes by: Tmo Kuosmanen 1 Outlne Defnton of dervatve functon Dervatve notatons

More information

Machine Learning. What is a good Decision Boundary? Support Vector Machines

Machine Learning. What is a good Decision Boundary? Support Vector Machines Machne Learnng 0-70/5 70/5-78 78 Sprng 200 Support Vector Machnes Erc Xng Lecture 7 March 5 200 Readng: Chap. 6&7 C.B book and lsted papers Erc Xng @ CMU 2006-200 What s a good Decson Boundar? Consder

More information

f(x,y) = (4(x 2 4)x,2y) = 0 H(x,y) =

f(x,y) = (4(x 2 4)x,2y) = 0 H(x,y) = Problem Set 3: Unconstraned mzaton n R N. () Fnd all crtcal ponts of f(x,y) (x 4) +y and show whch are ma and whch are mnma. () Fnd all crtcal ponts of f(x,y) (y x ) x and show whch are ma and whch are

More information

Some modelling aspects for the Matlab implementation of MMA

Some modelling aspects for the Matlab implementation of MMA Some modellng aspects for the Matlab mplementaton of MMA Krster Svanberg krlle@math.kth.se Optmzaton and Systems Theory Department of Mathematcs KTH, SE 10044 Stockholm September 2004 1. Consdered optmzaton

More information

15 Lagrange Multipliers

15 Lagrange Multipliers 15 The Method of s a powerful technque for constraned optmzaton. Whle t has applcatons far beyond machne learnng t was orgnally developed to solve physcs equatons), t s used for several ey dervatons n

More information

INTERMEDIATE FLUID MECHANICS

INTERMEDIATE FLUID MECHANICS INTERMEDITE FLUID MEHNIS enot shman-rosn Thaer School of Engneerng Dartmoth ollege See: Kn et al. Secton 3.4 pages 76-8 Lectre : Stran Vortct rclaton an Stress The ector eloct fel has 3 components each

More information

1 GSW Iterative Techniques for y = Ax

1 GSW Iterative Techniques for y = Ax 1 for y = A I m gong to cheat here. here are a lot of teratve technques that can be used to solve the general case of a set of smultaneous equatons (wrtten n the matr form as y = A), but ths chapter sn

More information

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD

Bayesian decision theory. Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world observatons decson functon L[,y] loss of predctn y wth the epected value of the

More information

Work is the change in energy of a system (neglecting heat transfer). To examine what could

Work is the change in energy of a system (neglecting heat transfer). To examine what could Work Work s the change n energy o a system (neglectng heat transer). To eamne what could cause work, let s look at the dmensons o energy: L ML E M L F L so T T dmensonally energy s equal to a orce tmes

More information

Mathematics Intersection of Lines

Mathematics Intersection of Lines a place of mnd F A C U L T Y O F E D U C A T I O N Department of Currculum and Pedagog Mathematcs Intersecton of Lnes Scence and Mathematcs Educaton Research Group Supported b UBC Teachng and Learnng Enhancement

More information

If there are k binding constraints at x then re-label these constraints so that they are the first k constraints.

If there are k binding constraints at x then re-label these constraints so that they are the first k constraints. Mathematcal Foundatons -1- Constaned Optmzaton Constaned Optmzaton Ma{ f ( ) X} whee X {, h ( ), 1,, m} Necessay condtons fo to be a soluton to ths mamzaton poblem Mathematcally, f ag Ma{ f ( ) X}, then

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

Discriminative classifier: Logistic Regression. CS534-Machine Learning

Discriminative classifier: Logistic Regression. CS534-Machine Learning Dscrmnatve classfer: Logstc Regresson CS534-Machne Learnng 2 Logstc Regresson Gven tranng set D stc regresson learns the condtonal dstrbuton We ll assume onl to classes and a parametrc form for here s

More information

Discriminative classifier: Logistic Regression. CS534-Machine Learning

Discriminative classifier: Logistic Regression. CS534-Machine Learning Dscrmnatve classfer: Logstc Regresson CS534-Machne Learnng robablstc Classfer Gven an nstance, hat does a probablstc classfer do dfferentl compared to, sa, perceptron? It does not drectl predct Instead,

More information

Solutions to selected problems from homework 1.

Solutions to selected problems from homework 1. Jan Hagemejer 1 Soltons to selected problems from homeork 1. Qeston 1 Let be a tlty fncton hch generates demand fncton xp, ) and ndrect tlty fncton vp, ). Let F : R R be a strctly ncreasng fncton. If the

More information

ENGI9496 Lecture Notes Multiport Models in Mechanics

ENGI9496 Lecture Notes Multiport Models in Mechanics ENGI9496 Moellng an Smulaton of Dynamc Systems Mechancs an Mechansms ENGI9496 Lecture Notes Multport Moels n Mechancs (New text Secton 4..3; Secton 9.1 generalzes to 3D moton) Defntons Generalze coornates

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?

More information

CHAPTER-5 INFORMATION MEASURE OF FUZZY MATRIX AND FUZZY BINARY RELATION

CHAPTER-5 INFORMATION MEASURE OF FUZZY MATRIX AND FUZZY BINARY RELATION CAPTER- INFORMATION MEASURE OF FUZZY MATRI AN FUZZY BINARY RELATION Introducton The basc concept of the fuzz matr theor s ver smple and can be appled to socal and natural stuatons A branch of fuzz matr

More information

Single Variable Optimization

Single Variable Optimization 8/4/07 Course Instructor Dr. Raymond C. Rump Oce: A 337 Phone: (95) 747 6958 E Mal: rcrump@utep.edu Topc 8b Sngle Varable Optmzaton EE 4386/530 Computatonal Methods n EE Outlne Mathematcal Prelmnares Sngle

More information

Goal Programming Approach to Solve Multi- Objective Intuitionistic Fuzzy Non- Linear Programming Models

Goal Programming Approach to Solve Multi- Objective Intuitionistic Fuzzy Non- Linear Programming Models Internatonal Journal o Mathematcs rends and echnoloy IJM Volume Number 7 - January 8 Goal Prorammn Approach to Solve Mult- Objectve Intutonstc Fuzzy Non- Lnear Prorammn Models S.Rukman #, R.Sopha Porchelv

More information

Chapter 8. Potential Energy and Conservation of Energy

Chapter 8. Potential Energy and Conservation of Energy Chapter 8 Potental Energy and Conservaton of Energy In ths chapter we wll ntroduce the followng concepts: Potental Energy Conservatve and non-conservatve forces Mechancal Energy Conservaton of Mechancal

More information

Perfect Competition and the Nash Bargaining Solution

Perfect Competition and the Nash Bargaining Solution Perfect Competton and the Nash Barganng Soluton Renhard John Department of Economcs Unversty of Bonn Adenauerallee 24-42 53113 Bonn, Germany emal: rohn@un-bonn.de May 2005 Abstract For a lnear exchange

More information

The Karush-Kuhn-Tucker Test of Convexity of Univariate Observations and Certain Economic Applications

The Karush-Kuhn-Tucker Test of Convexity of Univariate Observations and Certain Economic Applications Proceedngs of the World Congress on Engneerng 007 Vol II WCE 007 July - 4 007 London U.K. he Karush-Kuhn-ucker est of Convety of Unvarate Observatons and Certan Economc Applcatons Sofa A. Georgadou and

More information

6.854J / J Advanced Algorithms Fall 2008

6.854J / J Advanced Algorithms Fall 2008 MIT OpenCourseWare http://ocw.mt.edu 6.854J / 18.415J Advanced Algorthms Fall 2008 For nformaton about ctng these materals or our Terms of Use, vst: http://ocw.mt.edu/terms. 18.415/6.854 Advanced Algorthms

More information

Unit 5: Quadratic Equations & Functions

Unit 5: Quadratic Equations & Functions Date Perod Unt 5: Quadratc Equatons & Functons DAY TOPIC 1 Modelng Data wth Quadratc Functons Factorng Quadratc Epressons 3 Solvng Quadratc Equatons 4 Comple Numbers Smplfcaton, Addton/Subtracton & Multplcaton

More information

: Numerical Analysis Topic 2: Solution of Nonlinear Equations Lectures 5-11:

: Numerical Analysis Topic 2: Solution of Nonlinear Equations Lectures 5-11: 764: Numercal Analyss Topc : Soluton o Nonlnear Equatons Lectures 5-: UIN Malang Read Chapters 5 and 6 o the tetbook 764_Topc Lecture 5 Soluton o Nonlnear Equatons Root Fndng Problems Dentons Classcaton

More information

Parameter estimation class 5

Parameter estimation class 5 Parameter estmaton class 5 Multple Ve Geometr Comp 9-89 Marc Pollefes Content Background: Projectve geometr (D, 3D), Parameter estmaton, Algortm evaluaton. Sngle Ve: Camera model, Calbraton, Sngle Ve Geometr.

More information

Exercises of Chapter 2

Exercises of Chapter 2 Exercses of Chapter Chuang-Cheh Ln Department of Computer Scence and Informaton Engneerng, Natonal Chung Cheng Unversty, Mng-Hsung, Chay 61, Tawan. Exercse.6. Suppose that we ndependently roll two standard

More information

SELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d.

SELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d. SELECTED SOLUTIONS, SECTION 4.3 1. Weak dualty Prove that the prmal and dual values p and d defned by equatons 4.3. and 4.3.3 satsfy p d. We consder an optmzaton problem of the form The Lagrangan for ths

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information