Machine Learning 4771
|
|
- Douglas Wade
- 6 years ago
- Views:
Transcription
1 Machne Leanng 4771 Instucto: Tony Jebaa
2 Topc 6 Revew: Suppot Vecto Machnes Pmal & Dual Soluton Non-sepaable SVMs Kenels SVM Demo
3 Revew: SVM Suppot vecto machnes ae (n the smplest case) lnea classfes that do stuctual sk mnmzaton (SRM) Dectly maxmze magn to educe guaanteed sk J(θ) Assume fst the 2-class data s lnealy sepaable: have ( x 1,y 1 ),, x N,y N { } { } whee x D and y 1,1 = sgn w T x +b f x;θ Decson bounday o hypeplane gven by w T x +b = 0 Note: can scale w & b whle keepng same bounday Many solutons exst whch have empcal eo R emp (θ)=0 Want wdest o thckest one (max magn), also t s unque!
4 Suppot Vecto Machnes Defne: w T x +b = 0 H + =postve magn hypeplane H - =negatve magn hypeplane q =dstance fom decson plane to ogn q = mn x x 0 subject to w T x +b = 0 x 0 2 λ w T x +b mn x 1 2 1) gad x 1 = 0 2 xt x λ w T x +b x λw = 0 x = λw 3) Sol n ˆx = ( b )w w T w 4) dstance q = ˆx 0 = b w = w T w 5) Defne wthout loss of genealty snce can scale b & w 2) plug nto constant b w T w w T w = b w H w T x +b = 0 H + w T x +b = +1 H w T x +b = 1 w T x +b = 0 w T ( λw) +b = 0 λ = b w T w
5 Suppot Vecto Machnes d +d H The constants on the SVM fo R emp (θ)=0 ae thus: w T x +b +1 y = +1 w T x +b 1 y = 1 y ( w T x +b) 1 0 O moe smply: The magn of the SVM s: H + H H + w T x +b = +1 H w T x +b = 1 m = d + + d Dstance to ogn: Theefoe: d + = d = 1 w H q = b and magn Want to max magn, o equvalently mnmze: SVM Poblem: mn 1 w 2 subject to y 2 ( w T x +b) 1 0 Ths s a quadatc pogam! Can plug ths nto a matlab functon called qp(), done! w H + q + = b 1 m = w 2 w H q = 1 b w o w 2 w
6 SVM n Dual Fom We can also solve the poblem va convex dualty Pmal SVM poblem L P : Ths s a quadatc pogam, quadatc cost functon wth multple lnea nequaltes (these cave out a convex hull) Subtact fom cost each nequalty tmes an α Lagange multple, take devatves of w & b: Plug back n, dual: Also have constants: mn 1 w 2 subject to y 2 ( w T x +b) L P = mn w,b max w 2 α 0 α 2 ( y ( w T x +b) 1) L = w α w P y x = 0 w = α y x L = α b P y = 0 L D = α 1 α 2 α j y y j x T j x j α y = 0 & α 0 Above L D must be maxmzed! convex dualty also qp()
7 SVM Dual Soluton Popetes We have dual convex pogam: 1 2 α α α j y y j x T x j subject to α y = 0 & α 0,j Solve fo N alphas (one pe data pont) nstead of D w s Stll convex (qp) so unque soluton, gves alphas Alphas can be used to get w: w = α y x Suppot Vectos: have non-zeo alphas shown wth thcke ccles, all lve on the magn: w T x +b = ±1 Soluton s spase, most alphas=0 these ae non-suppot vectos SVM gnoes them f they move (wthout cossng magn) o f they ae deleted, SVM doesn t change (stays obust)
8 SVM Dual Soluton Popetes Pmal & Dual Illustaton: w T x +b ±1 Recall we could get w fom alphas: O could use as s: f x w = α y x +b Kaush-Kuhn-Tucke Condtons (KKT): solve value of b on magn (fo nonzeo alphas) have: w T x +b = y usng known w, compute b fo each suppot vecto then Spasty (few nonzeo alphas) s useful fo seveal easons Means SVM only uses some of tanng data to lean Should help mpove ts ablty to genealze to test data Computatonally faste when usng fnal leaned classfe = sgn ( x T w +b) = sgn ( α y x T x ) b = y w T x : α > 0 b = aveage b
9 Non-Sepaable SVMs What happens when non-sepaable? Thee s no soluton and convex hull shnks to nothng Not all constants can be esolved, the alphas go to Instead of pefectly classfyng each pont: y ( w T x +b) 1 we Relax the poblem wth (postve) slack vaables x s allow data to (sometmes) fall on wong sde, fo example: w T x +b 0.03 f y = +1 w T x +b +1 ξ f y = +1 whee ξ 0 New constants: w T x +b 1 + ξ f y = 1 whee ξ 0 But too much x s means too much slack, so penalze them L P : mn 1 w 2 +C ξ 2 subject to y ( w T x +b) 1 + ξ 0
10 Non-Sepaable SVMs Ths new poblem s stll convex, stll qp()! Use chooses scala C (o coss-valdates) whch contols how much slack x to use (how non-sepaable) and how obust to outles o bad ponts on the wong sde Lage magn Low slack On ght sde L P : mn 1 w 2 +C ξ 2 α ( y ( w T x +b) 1 + ξ ) β ξ L and L asbefoe... L b P w P ξ P = C α β = 0 Can now wte dual poblem (to maxmze): L D : max α 1 2 α = C β but... α & β 0 0 α C Fo x postvty α α j y y j x T x,j j subject to α y = 0 and α 0,C Same dual as befoe but alphas can t gow beyond C
11 Non-Sepaable SVMs As we ty to enfoce a classfcaton fo a data pont ts Lagange multple alpha keeps gowng endlessly Clampng alpha to stop gowng at C makes SVM gve up on those non-sepaable ponts The dual pogam s now: Solve as befoe wth exta constants that alphas postve AND less than C gves alphas fom alphas get w = α y x Kaush-Kuhn-Tucke Condtons (KKT): solve value of b on magn fo not=zeo alphas AND not=c alphas fo all othes have suppot vectos, assume and use fomula y w T x + ξ = 0 ( b ) 1 + ξ = 0 to get b and b = aveage ( b ) Mechancal analogy: suppot vecto foces & toques
12 Nonlnea SVMs What f the poblem s not lnea?
13 Nonlnea SVMs What f the poblem s not lnea? We can use ou old tck Map d-dmensonal x data fom L-space to hgh dmensonal H (Hlbet) featue-space va bass functons Φ(x) Fo example, quadatc classfe: x Φ( x ) va Φ( x ) = Call ph s featue vectos computed fom ognal x nputs Replace all x s n the SVM equatons wth ph s Now solve the followng leanng poblem: L D : max α 1 2 Whch gves a nonlnea classfe n ognal space: f x x vec x x T T φ( x j ) α α j y y j φ x,j s.t. α 0,C, α y = 0 +b = sgn α y φ( x) T φ( x ) L Η
14 Kenels (see One mpotant aspect of SVMs: all math nvolves only the nne poducts between the ph featues! = sgn α y φ( x) T φ( x ) f x +b Replace all nne poducts wth a geneal kenel functon Mece kenel: accepts 2 nputs and outputs a scala va: k ( x, x ) = φ( x),φ( x ) = Example: quadatc polynomal t T φ( x ) fo fnte φ φ x φ( x,t)φ( x,t )dt φ x othewse 2 = x 1 = φ( x) T φ( x ) k x, x 2x 1 x 2 x 2 2 T = x 1 2 x x 1 x 1 x 2 x 2 + x 2 2 x 2 2 = ( x 1 x 1 + x 2 x 2 ) 2
15 Kenels Sometmes, many Φ(x) wll poduce the same k(x,x ) Sometmes k(x,x ) computable but featues huge o nfnte! Example: polynomals If explct polynomal mappng, featue space Φ(x) s huge d-dmensonal data, p-th ode polynomal, mages of sze 16x16 wth p=4 have dm(h)=183mllon but can equvalently just use kenel: = ( x T x ) p = ( x x ) p k x, x φ( x)φ x k x,y dm Η = ( x T y) p = d + p 1 Multnomal Theoem p! 1! 2! 3! ( p 1 2 )! x 1 x x d d x 1 1 x 2 2 x d d ( w x 1 1 x 2 2 x d )( d w x 1 1 x 2 2 x d ) w=weght on tem d Equvalent! p
16 Kenels Replace each x T x j k x,x j, fo example: P-th Ode Polynomal Kenel: RBF Kenel (nfnte!): Sgmod (hypebolc tan) Kenel: Usng kenels we get genealzed nne poduct SVM: L D : max α 1 α 2 α j y y j k x,x,j j s.t. α 0,C, α y = 0 f ( x) = sgn ( α y k ( x,x) +b) Stll qp solve, just use Gam matx K (postve defnte) K = k ( x 1,x 2 ) k ( x 1,x 3 ) k ( x 2,x 2 ) k ( x 2,x 3 ) k ( x 2,x 3 ) k ( x 3,x 3 ) k x 1,x 1 k x 1,x 2 k x 1,x 3 = ( x T x + 1) p = exp 1 2σ x x 2 2 = tanh κx T x δ k x, x k x, x k x, x K j = k x,x j
17 Kenelzed SVMs Polynomal kenel: Radal bass functon kenel: Polynomal Kenel = ( x T x j + 1) p k x,x j = exp 12σ x x 2 j k x,x j RBF kenel 2 Least-squaes, logstc-egesson, pecepton ae also kenelzable
18 SVM Demo SVM Demo by Steve Gunn: In svc.m eplace [alpha lambda how] = qp( ); wth [alpha lambda how] = quadpog(h,c,[],[],a,b,vlb,vub,x0); Ths eplaces the old Matlab command qp (quadatc pogammng) wth the new one fo moe ecent vesons
If there are k binding constraints at x then re-label these constraints so that they are the first k constraints.
Mathematcal Foundatons -1- Constaned Optmzaton Constaned Optmzaton Ma{ f ( ) X} whee X {, h ( ), 1,, m} Necessay condtons fo to be a soluton to ths mamzaton poblem Mathematcally, f ag Ma{ f ( ) X}, then
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationLecture 3: Dual problems and Kernels
Lecture 3: Dual problems and Kernels C4B Machne Learnng Hlary 211 A. Zsserman Prmal and dual forms Lnear separablty revsted Feature mappng Kernels for SVMs Kernel trck requrements radal bass functons SVM
More informationPHYS 705: Classical Mechanics. Derivation of Lagrange Equations from D Alembert s Principle
1 PHYS 705: Classcal Mechancs Devaton of Lagange Equatons fom D Alembet s Pncple 2 D Alembet s Pncple Followng a smla agument fo the vtual dsplacement to be consstent wth constants,.e, (no vtual wok fo
More informationSet of square-integrable function 2 L : function space F
Set of squae-ntegable functon L : functon space F Motvaton: In ou pevous dscussons we have seen that fo fee patcles wave equatons (Helmholt o Schödnge) can be expessed n tems of egenvalue equatons. H E,
More informationEngineering Mechanics. Force resultants, Torques, Scalar Products, Equivalent Force systems
Engneeng echancs oce esultants, Toques, Scala oducts, Equvalent oce sstems Tata cgaw-hll Companes, 008 Resultant of Two oces foce: acton of one bod on anothe; chaacteed b ts pont of applcaton, magntude,
More informationMachine Learning. Spectral Clustering. Lecture 23, April 14, Reading: Eric Xing 1
Machne Leanng -7/5 7/5-78, 78, Spng 8 Spectal Clusteng Ec Xng Lectue 3, pl 4, 8 Readng: Ec Xng Data Clusteng wo dffeent ctea Compactness, e.g., k-means, mxtue models Connectvty, e.g., spectal clusteng
More information8 Baire Category Theorem and Uniform Boundedness
8 Bae Categoy Theoem and Unfom Boundedness Pncple 8.1 Bae s Categoy Theoem Valdty of many esults n analyss depends on the completeness popety. Ths popety addesses the nadequacy of the system of atonal
More informationChapter Fifiteen. Surfaces Revisited
Chapte Ffteen ufaces Revsted 15.1 Vecto Descpton of ufaces We look now at the vey specal case of functons : D R 3, whee D R s a nce subset of the plane. We suppose s a nce functon. As the pont ( s, t)
More informationOptimization Methods: Linear Programming- Revised Simplex Method. Module 3 Lecture Notes 5. Revised Simplex Method, Duality and Sensitivity analysis
Optmzaton Meods: Lnea Pogammng- Revsed Smple Meod Module Lectue Notes Revsed Smple Meod, Dualty and Senstvty analyss Intoducton In e pevous class, e smple meod was dscussed whee e smple tableau at each
More informationRigid Bodies: Equivalent Systems of Forces
Engneeng Statcs, ENGR 2301 Chapte 3 Rgd Bodes: Equvalent Sstems of oces Intoducton Teatment of a bod as a sngle patcle s not alwas possble. In geneal, the se of the bod and the specfc ponts of applcaton
More informationIntegral Vector Operations and Related Theorems Applications in Mechanics and E&M
Dola Bagayoko (0) Integal Vecto Opeatons and elated Theoems Applcatons n Mechancs and E&M Ι Basc Defnton Please efe to you calculus evewed below. Ι, ΙΙ, andιιι notes and textbooks fo detals on the concepts
More informationChapter 6 Support vector machine. Séparateurs à vaste marge
Chapter 6 Support vector machne Séparateurs à vaste marge Méthode de classfcaton bnare par apprentssage Introdute par Vladmr Vapnk en 1995 Repose sur l exstence d un classfcateur lnéare Apprentssage supervsé
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationUNIT10 PLANE OF REGRESSION
UIT0 PLAE OF REGRESSIO Plane of Regesson Stuctue 0. Intoducton Ojectves 0. Yule s otaton 0. Plane of Regesson fo thee Vaales 0.4 Popetes of Resduals 0.5 Vaance of the Resduals 0.6 Summay 0.7 Solutons /
More informationSupport Vector Machines
Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x n class
More informationKristin P. Bennett. Rensselaer Polytechnic Institute
Support Vector Machnes and Other Kernel Methods Krstn P. Bennett Mathematcal Scences Department Rensselaer Polytechnc Insttute Support Vector Machnes (SVM) A methodology for nference based on Statstcal
More informationSupport Vector Machines
/14/018 Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x
More informationScalars and Vectors Scalar
Scalas and ectos Scala A phscal quantt that s completel chaacteed b a eal numbe (o b ts numecal value) s called a scala. In othe wods a scala possesses onl a magntude. Mass denst volume tempeatue tme eneg
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationAPPENDIX A Some Linear Algebra
APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,
More informationSolutions to exam in SF1811 Optimization, Jan 14, 2015
Solutons to exam n SF8 Optmzaton, Jan 4, 25 3 3 O------O -4 \ / \ / The network: \/ where all lnks go from left to rght. /\ / \ / \ 6 O------O -5 2 4.(a) Let x = ( x 3, x 4, x 23, x 24 ) T, where the varable
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More information3. A Review of Some Existing AW (BT, CT) Algorithms
3. A Revew of Some Exstng AW (BT, CT) Algothms In ths secton, some typcal ant-wndp algothms wll be descbed. As the soltons fo bmpless and condtoned tansfe ae smla to those fo ant-wndp, the pesented algothms
More informationSupport Vector Machines
CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at
More informationPart V: Velocity and Acceleration Analysis of Mechanisms
Pat V: Velocty an Acceleaton Analyss of Mechansms Ths secton wll evew the most common an cuently pactce methos fo completng the knematcs analyss of mechansms; escbng moton though velocty an acceleaton.
More informationPHY126 Summer Session I, 2008
PHY6 Summe Sesson I, 8 Most of nfomaton s avalable at: http://nngoup.phscs.sunsb.edu/~chak/phy6-8 ncludng the sllabus and lectue sldes. Read sllabus and watch fo mpotant announcements. Homewok assgnment
More informationMultilayer Perceptron (MLP)
Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne
More informationPhysics 202, Lecture 2. Announcements
Physcs 202, Lectue 2 Today s Topcs Announcements Electc Felds Moe on the Electc Foce (Coulomb s Law The Electc Feld Moton of Chaged Patcles n an Electc Feld Announcements Homewok Assgnment #1: WebAssgn
More informationA. Thicknesses and Densities
10 Lab0 The Eath s Shells A. Thcknesses and Denstes Any theoy of the nteo of the Eath must be consstent wth the fact that ts aggegate densty s 5.5 g/cm (ecall we calculated ths densty last tme). In othe
More informationExact Simplification of Support Vector Solutions
Jounal of Machne Leanng Reseach 2 (200) 293-297 Submtted 3/0; Publshed 2/0 Exact Smplfcaton of Suppot Vecto Solutons Tom Downs TD@ITEE.UQ.EDU.AU School of Infomaton Technology and Electcal Engneeng Unvesty
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017
U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that
More informationLinear Classification, SVMs and Nearest Neighbors
1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush
More informationMMA and GCMMA two methods for nonlinear optimization
MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons
More informationEnergy in Closed Systems
Enegy n Closed Systems Anamta Palt palt.anamta@gmal.com Abstact The wtng ndcates a beakdown of the classcal laws. We consde consevaton of enegy wth a many body system n elaton to the nvese squae law and
More informationINTRODUCTION. consider the statements : I there exists x X. f x, such that. II there exists y Y. such that g y
INRODUCION hs dssetaton s the eadng of efeences [1], [] and [3]. Faas lemma s one of the theoems of the altenatve. hese theoems chaacteze the optmalt condtons of seveal mnmzaton poblems. It s nown that
More informationLecture 6: Support Vector Machines
Lecture 6: Support Vector Machnes Marna Melă mmp@stat.washngton.edu Department of Statstcs Unversty of Washngton November, 2018 Lnear SVM s The margn and the expected classfcaton error Maxmum Margn Lnear
More informationPhysics 11b Lecture #2. Electric Field Electric Flux Gauss s Law
Physcs 11b Lectue # Electc Feld Electc Flux Gauss s Law What We Dd Last Tme Electc chage = How object esponds to electc foce Comes n postve and negatve flavos Conseved Electc foce Coulomb s Law F Same
More informationSupport Vector Machines
Support Vector Machnes Konstantn Tretyakov (kt@ut.ee) MTAT.03.227 Machne Learnng So far Supervsed machne learnng Lnear models Least squares regresson Fsher s dscrmnant, Perceptron, Logstc model Non-lnear
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? + + + + + + + + + Intuton of Margn Consder ponts
More informationTian Zheng Department of Statistics Columbia University
Haplotype Tansmsson Assocaton (HTA) An "Impotance" Measue fo Selectng Genetc Makes Tan Zheng Depatment of Statstcs Columba Unvesty Ths s a jont wok wth Pofesso Shaw-Hwa Lo n the Depatment of Statstcs at
More informationSELECTED SOLUTIONS, SECTION (Weak duality) Prove that the primal and dual values p and d defined by equations (4.3.2) and (4.3.3) satisfy p d.
SELECTED SOLUTIONS, SECTION 4.3 1. Weak dualty Prove that the prmal and dual values p and d defned by equatons 4.3. and 4.3.3 satsfy p d. We consder an optmzaton problem of the form The Lagrangan for ths
More informationCSE 252C: Computer Vision III
CSE 252C: Computer Vson III Lecturer: Serge Belonge Scrbe: Catherne Wah LECTURE 15 Kernel Machnes 15.1. Kernels We wll study two methods based on a specal knd of functon k(x, y) called a kernel: Kernel
More information1. A body will remain in a state of rest, or of uniform motion in a straight line unless it
Pncples of Dnamcs: Newton's Laws of moton. : Foce Analss 1. A bod wll eman n a state of est, o of unfom moton n a staght lne unless t s acted b etenal foces to change ts state.. The ate of change of momentum
More informationFeature Selection: Part 1
CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?
More informationSupport Vector Machines
Support Vector Machnes Konstantn Tretyakov (kt@ut.ee) MTAT.03.227 Machne Learnng So far So far Supervsed machne learnng Lnear models Non-lnear models Unsupervsed machne learnng Generc scaffoldng So far
More informationChapter 13 - Universal Gravitation
Chapte 3 - Unesal Gataton In Chapte 5 we studed Newton s thee laws of moton. In addton to these laws, Newton fomulated the law of unesal gataton. Ths law states that two masses ae attacted by a foce gen
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We
More informationLagrange Multipliers Kernel Trick
Lagrange Multplers Kernel Trck Ncholas Ruozz Unversty of Texas at Dallas Based roughly on the sldes of Davd Sontag General Optmzaton A mathematcal detour, we ll come back to SVMs soon! subject to: f x
More informationDetection and Estimation Theory
ESE 54 Detecton and Etmaton Theoy Joeph A. O Sullvan Samuel C. Sach Pofeo Electonc Sytem and Sgnal Reeach Laboatoy Electcal and Sytem Engneeng Wahngton Unvety 411 Jolley Hall 314-935-4173 (Lnda anwe) jao@wutl.edu
More informationSection 8.3 Polar Form of Complex Numbers
80 Chapter 8 Secton 8 Polar Form of Complex Numbers From prevous classes, you may have encountered magnary numbers the square roots of negatve numbers and, more generally, complex numbers whch are the
More informationn α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0
MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector
More informationRemember: When an object falls due to gravity its potential energy decreases.
Chapte 5: lectc Potental As mentoned seveal tmes dung the uate Newton s law o gavty and Coulomb s law ae dentcal n the mathematcal om. So, most thngs that ae tue o gavty ae also tue o electostatcs! Hee
More informationMATH 417 Homework 3 Instructor: D. Cabrera Due June 30. sin θ v x = v r cos θ v θ r. (b) Then use the Cauchy-Riemann equations in polar coordinates
MATH 417 Homewok 3 Instucto: D. Cabea Due June 30 1. Let a function f(z) = u + iv be diffeentiable at z 0. (a) Use the Chain Rule and the fomulas x = cosθ and y = to show that u x = u cosθ u θ, v x = v
More information1 Matrix representations of canonical matrices
1 Matrx representatons of canoncal matrces 2-d rotaton around the orgn: ( ) cos θ sn θ R 0 = sn θ cos θ 3-d rotaton around the x-axs: R x = 1 0 0 0 cos θ sn θ 0 sn θ cos θ 3-d rotaton around the y-axs:
More informationMULTILAYER PERCEPTRONS
Last updated: Nov 26, 2012 MULTILAYER PERCEPTRONS Outline 2 Combining Linea Classifies Leaning Paametes Outline 3 Combining Linea Classifies Leaning Paametes Implementing Logical Relations 4 AND and OR
More informationInner Product. Euclidean Space. Orthonormal Basis. Orthogonal
Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,
More informationKernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan
Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems
More informationLecture 10 Support Vector Machines. Oct
Lecture 10 Support Vector Machnes Oct - 20-2008 Lnear Separators Whch of the lnear separators s optmal? Concept of Margn Recall that n Perceptron, we learned that the convergence rate of the Perceptron
More informationCorrespondence Analysis & Related Methods
Coespondence Analyss & Related Methods Ineta contbutons n weghted PCA PCA s a method of data vsualzaton whch epesents the tue postons of ponts n a map whch comes closest to all the ponts, closest n sense
More informationNatural Language Processing and Information Retrieval
Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support
More informationMechanics Physics 151
Mechancs Physcs 151 Lectue 18 Hamltonan Equatons of Moton (Chapte 8) What s Ahead We ae statng Hamltonan fomalsm Hamltonan equaton Today and 11/6 Canoncal tansfomaton 1/3, 1/5, 1/10 Close lnk to non-elatvstc
More informationprinceton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg
prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there
More information24-2: Electric Potential Energy. 24-1: What is physics
D. Iyad SAADEDDIN Chapte 4: Electc Potental Electc potental Enegy and Electc potental Calculatng the E-potental fom E-feld fo dffeent chage dstbutons Calculatng the E-feld fom E-potental Potental of a
More informationA NOTE ON ELASTICITY ESTIMATION OF CENSORED DEMAND
Octobe 003 B 003-09 A NOT ON ASTICITY STIATION OF CNSOD DAND Dansheng Dong an Hay. Kase Conell nvesty Depatment of Apple conomcs an anagement College of Agcultue an fe Scences Conell nvesty Ithaca New
More information4 SingularValue Decomposition (SVD)
/6/00 Z:\ jeh\self\boo Kannan\Jan-5-00\4 SVD 4 SngulaValue Decomposton (SVD) Chapte 4 Pat SVD he sngula value decomposton of a matx s the factozaton of nto the poduct of thee matces = UDV whee the columns
More informationAPPLICATIONS OF SEMIGENERALIZED -CLOSED SETS
Intenatonal Jounal of Mathematcal Engneeng Scence ISSN : 22776982 Volume Issue 4 (Apl 202) http://www.mes.com/ https://stes.google.com/ste/mesounal/ APPLICATIONS OF SEMIGENERALIZED CLOSED SETS G.SHANMUGAM,
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationChapter 23: Electric Potential
Chapte 23: Electc Potental Electc Potental Enegy It tuns out (won t show ths) that the tostatc foce, qq 1 2 F ˆ = k, s consevatve. 2 Recall, fo any consevatve foce, t s always possble to wte the wok done
More informationP 365. r r r )...(1 365
SCIENCE WORLD JOURNAL VOL (NO4) 008 www.scecncewoldounal.og ISSN 597-64 SHORT COMMUNICATION ANALYSING THE APPROXIMATION MODEL TO BIRTHDAY PROBLEM *CHOJI, D.N. & DEME, A.C. Depatment of Mathematcs Unvesty
More informationA Novel Ordinal Regression Method with Minimum Class Variance Support Vector Machine
Intenatonal Confeence on Mateals Engneeng and Infomaton echnology Applcatons (MEIA 05) A ovel Odnal Regesson Method wth Mnmum Class Vaance Suppot Vecto Machne Jnong Hu,, a, Xaomng Wang and Zengx Huang
More information= y and Normed Linear Spaces
304-50 LINER SYSTEMS Lectue 8: Solutos to = ad Nomed Lea Spaces 73 Fdg N To fd N, we eed to chaacteze all solutos to = 0 Recall that ow opeatos peseve N, so that = 0 = 0 We ca solve = 0 ecusvel backwads
More informationC4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )
C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z
More information2/24/2014. The point mass. Impulse for a single collision The impulse of a force is a vector. The Center of Mass. System of particles
/4/04 Chapte 7 Lnea oentu Lnea oentu of a Sngle Patcle Lnea oentu: p υ It s a easue of the patcle s oton It s a vecto, sla to the veloct p υ p υ p υ z z p It also depends on the ass of the object, sla
More informationTest 1 phy What mass of a material with density ρ is required to make a hollow spherical shell having inner radius r i and outer radius r o?
Test 1 phy 0 1. a) What s the pupose of measuement? b) Wte all fou condtons, whch must be satsfed by a scala poduct. (Use dffeent symbols to dstngush opeatons on ectos fom opeatons on numbes.) c) What
More informationChapter I Matrices, Vectors, & Vector Calculus 1-1, 1-9, 1-10, 1-11, 1-17, 1-18, 1-25, 1-27, 1-36, 1-37, 1-41.
Chapte I Matces, Vectos, & Vecto Calculus -, -9, -0, -, -7, -8, -5, -7, -36, -37, -4. . Concept of a Scala Consde the aa of patcles shown n the fgue. he mass of the patcle at (,) can be epessed as. M (,
More informationLinear, affine, and convex sets and hulls In the sequel, unless otherwise specified, X will denote a real vector space.
Lnear, affne, and convex sets and hulls In the sequel, unless otherwse specfed, X wll denote a real vector space. Lnes and segments. Gven two ponts x, y X, we defne xy = {x + t(y x) : t R} = {(1 t)x +
More informationKhintchine-Type Inequalities and Their Applications in Optimization
Khntchne-Type Inequaltes and The Applcatons n Optmzaton Anthony Man-Cho So Depatment of Systems Engneeng & Engneeng Management The Chnese Unvesty of Hong Kong ISDS-Kolloquum Unvestaet Wen 29 June 2009
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationALL QUESTIONS ARE WORTH 20 POINTS. WORK OUT FIVE PROBLEMS.
GNRAL PHYSICS PH -3A (D. S. Mov) Test (/3/) key STUDNT NAM: STUDNT d #: -------------------------------------------------------------------------------------------------------------------------------------------
More informationMULTIPOLE FIELDS. Multipoles, 2 l poles. Monopoles, dipoles, quadrupoles, octupoles... Electric Dipole R 1 R 2. P(r,θ,φ) e r
MULTIPOLE FIELDS Mutpoes poes. Monopoes dpoes quadupoes octupoes... 4 8 6 Eectc Dpoe +q O θ e R R P(θφ) -q e The potenta at the fed pont P(θφ) s ( θϕ )= q R R Bo E. Seneus : Now R = ( e) = + cosθ R = (
More informationNonlinear Classifiers II
Nonlnear Classfers II Nonlnear Classfers: Introducton Classfers Supervsed Classfers Lnear Classfers Perceptron Least Squares Methods Lnear Support Vector Machne Nonlnear Classfers Part I: Mult Layer Neural
More informationCollege of Computer & Information Science Fall 2009 Northeastern University 20 October 2009
College of Computer & Informaton Scence Fall 2009 Northeastern Unversty 20 October 2009 CS7880: Algorthmc Power Tools Scrbe: Jan Wen and Laura Poplawsk Lecture Outlne: Prmal-dual schema Network Desgn:
More informationA Brief Guide to Recognizing and Coping With Failures of the Classical Regression Assumptions
A Bef Gude to Recognzng and Copng Wth Falues of the Classcal Regesson Assumptons Model: Y 1 k X 1 X fxed n epeated samples IID 0, I. Specfcaton Poblems A. Unnecessay explanatoy vaables 1. OLS s no longe
More informationNeural networks. Nuno Vasconcelos ECE Department, UCSD
Neural networs Nuno Vasconcelos ECE Department, UCSD Classfcaton a classfcaton problem has two types of varables e.g. X - vector of observatons (features) n the world Y - state (class) of the world x X
More informationMultistage Median Ranked Set Sampling for Estimating the Population Median
Jounal of Mathematcs and Statstcs 3 (: 58-64 007 ISSN 549-3644 007 Scence Publcatons Multstage Medan Ranked Set Samplng fo Estmatng the Populaton Medan Abdul Azz Jeman Ame Al-Oma and Kamaulzaman Ibahm
More informationImage classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?
Image classfcaton Gven te bag-of-features representatons of mages from dfferent classes ow do we learn a model for dstngusng tem? Classfers Learn a decson rule assgnng bag-offeatures representatons of
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More information12. The Hamilton-Jacobi Equation Michael Fowler
1. The Hamlton-Jacob Equaton Mchael Fowler Back to Confguraton Space We ve establshed that the acton, regarded as a functon of ts coordnate endponts and tme, satsfes ( ) ( ) S q, t / t+ H qpt,, = 0, and
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationSOME NEW SELF-DUAL [96, 48, 16] CODES WITH AN AUTOMORPHISM OF ORDER 15. KEYWORDS: automorphisms, construction, self-dual codes
Факултет по математика и информатика, том ХVІ С, 014 SOME NEW SELF-DUAL [96, 48, 16] CODES WITH AN AUTOMORPHISM OF ORDER 15 NIKOLAY I. YANKOV ABSTRACT: A new method fo constuctng bnay self-dual codes wth
More information1 Convex Optimization
Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,
More informationTHE CHINESE REMAINDER THEOREM. We should thank the Chinese for their wonderful remainder theorem. Glenn Stevens
THE CHINESE REMAINDER THEOREM KEITH CONRAD We should thank the Chnese for ther wonderful remander theorem. Glenn Stevens 1. Introducton The Chnese remander theorem says we can unquely solve any par of
More informationV. Principles of Irreversible Thermodynamics. s = S - S 0 (7.3) s = = - g i, k. "Flux": = da i. "Force": = -Â g a ik k = X i. Â J i X i (7.
Themodynamcs and Knetcs of Solds 71 V. Pncples of Ievesble Themodynamcs 5. Onsage s Teatment s = S - S 0 = s( a 1, a 2,...) a n = A g - A n (7.6) Equlbum themodynamcs detemnes the paametes of an equlbum
More informationAssortment Optimization under MNL
Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.
More informationFoundations of Arithmetic
Foundatons of Arthmetc Notaton We shall denote the sum and product of numbers n the usual notaton as a 2 + a 2 + a 3 + + a = a, a 1 a 2 a 3 a = a The notaton a b means a dvdes b,.e. ac = b where c s an
More information1. Estimation, Approximation and Errors Percentages Polynomials and Formulas Identities and Factorization 52
ontents ommonly Used Formulas. Estmaton, pproxmaton and Errors. Percentages. Polynomals and Formulas 8. Identtes and Factorzaton. Equatons and Inequaltes 66 6. Rate and Rato 8 7. Laws of Integral Indces
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More information