Slide03 Historical Overview Haykin Chapter 3 (Chap 1, 3, 3rd Ed): Single-Layer Perceptrons Multiple Faces of a Single Neuron Part I: Adaptive Filter
|
|
- Mildred Hodge
- 6 years ago
- Views:
Transcription
1 Slide3 Haykin Chaper 3 (Chap, 3, 3rd Ed): Single-Layer Perceprons CPSC Insrucor: Yoonsuck Choe Hisorical Overview McCulloch and Pis (943): neural neworks as compuing machines. Hebb (949): posulaed he firs rule for self-organizing learning. Rosenbla (958): percepron as a firs model of supervised learning. Widrow and Hoff (96): adapive filers using leas-mean-square (LMS) algorihm (dela rule). 2 Muliple Faces of a Single Neuron Wha a single neuron does can be viewed from differen perspecives: Adapive filer: as in signal processing Classifier: as in percepron The wo aspecs will be reviewed, in he above order. Par I: Adapive Filer 3 4
2 Adapive Filering Problem Consider an unknown dynamical sysem, ha akes m inpus and generaes one oupu. Behavior of he sysem described as is inpu/oupu pair: T : {x(i), d(i); i =, 2,..., n,...} where x(i) = [x (i), x 2 (i),..., x m (i)] T is he inpu and d(i) he desired response (or arge signal). Inpu vecor can be eiher a spaial snapsho or a emporal sequence uniformly spaced in ime. There are wo imporan processes in adapive filering: Filering process: generaion of oupu based on he inpu: y(i) = x T (i)w(i). Adapaive process: auomaic adjusmen of weighs o reduce error: e(i) = d(i) y(i). 5 Unconsrained Opimizaion Techniques How can we adjus w(i) o gradually minimize e(i)? Noe ha e(i) = d(i) y(i) = d(i) x T (i)w(i). Since d(i) and x(i) are fixed, only he change in w(i) can change e(i). In oher words, we wan o minimize he cos funcion E(w) wih respec o he weigh vecor w: Find he opimal soluion w. The necessary condiion for opimaliy is E(w ) =, where he gradien operaor is defined as [ ] T =,,... w w 2 w m Wih his, we ge E(w ) = [ E w, 6 ] E E T,.... w 2 w m Seepes Descen We wan he ieraive updae algorihm o have he following propery: E(w(n )) < E(w(n)). Define he gradien vecor E(w) as g. The ieraive weigh updae rule hen becomes: w(n ) = w(n) ηg(n) where η is a small learning-rae parameer. So we can say, w(n) = w(n ) w(n) = ηg(n) Seepes Descen (con d) We now check if E(w(n )) < E(w(n)). Using firs-order Taylor expansion of E( ) near w(n), E(w(n )) E(w(n)) g T (n) w(n) and w(n) = ηg(n), we ge E(w(n )) E(w(n)) ηg T (n)g(n) So, i is indeed (for small η): = E(w(n)) η g(n) 2. }{{} Posiive! E(w(n )) < E(w(n)). 7 Taylor series: f(x) = f(a) f (a)(x a) f (a)(xa) 2 2!... 8
3 Seepes Descen: Example Seepes Descen: Anoher Example x*xy*y y 7 Gradien of x*xy*y Convergence o opimal w is very slow x Small η: overdamped, smooh rajecory Large η: underdamped, jagged rajecory η oo large: algorihm becomes unsable 9 For f(x) = f(x, y) = x 2 y 2, [ ] f(x, y) = f x, f T y = [2x, 2y] T. Noe ha () he gradien vecors are poining upward, away from he origin, (2) lengh of he vecors are shorer near he origin. If you follow f(x, y), you will end up a he origin. We can see ha he gradien vecors are perpendicular o he level curves. * The vecor lenghs were scaled down by a facor of o avoid cluer. Newon s Mehod Newon s mehod is an exension of seepes descen, where he second-order erm in he Taylor series expansion is used. I is generally faser and shows a less erraic meandering compared o he seepes descen mehod. There are cerain condiions o be me hough, such as he Hessian marix 2 E(w) being posiive definie (for an arbiarry x, x T Hx > ). Gauss-Newon Mehod Applicable for cos-funcions expressed as sum of error squares: E(w) = 2 n e i (w) 2, i= where e i (w) is he error in he i-h rial, wih he weigh w. Recalling he Taylor series f(x) = f(a) f (a)(x a)..., we can express e i (w) evaluaed near e i (w k ) as [ ] T ei e i (w) = e i (w k ) (w w k ). w w=w k In marix noaion, we ge: e(w) = e(w k ) J e (w k )(w w k ). * We will use a slighly differen noaion han he exbook, for clariy. 2
4 Gauss-Newon Mehod (con d) Quick Example: Jacobian Marix J e (w) is he Jacobian marix, where each row is he gradien of e i (w): J e (w) = e w e w 2... e 2 w e 2 w 2... e wn e 2 wn : : : : : : en w en w 2... en wn = ( e (w)) T ( e 2 (w)) T : : ( e n (w)) T Given e(x, y) = e (x, y) e 2 (x, y) = The Jacobian of e(x, y) becomes J e (x, y) = [ e (x,y) x e 2 (x,y) x e (x,y) y e 2 (x,y) y x 2 y 2 cos(x) sin(y) ] = [ 2x sin(x), 2y cos(y) ]. We can hen evaluae J e (w k ) by plugging in acual values of w k ino he Jabobian marix above. For (x, y) = (.5π, π), we ge [ π J e (.5π, π) = sin(.5π) 2π cos(π) ] = [ π 2π ]. 3 4 Gauss-Newon Mehod (con d) Linear Leas-Square Filer Again, saring wih e(w) = e(w k ) J e (w k )(w w k ), Given m inpu and oupu funcion y(i) = φ(x T i w i) where φ(x) = x, i.e., i is linear, and a se of raining samples {x i, d i } n i=, we can define he error vecor for an arbirary weigh w as wha we wan is o se w so ha he error approaches. Tha is, we wan o minimize he norm of e(w): e(w) 2 = e(w k ) 2 2e(w k ) T J e (w k )(w w k ) e(w) = d [x, x 2,..., x n ] T w. where d = [d, d 2,..., d n ] T. Seing X = [x, x 2,..., x n ] T, we ge: e(w) = d Xw. (w w k ) T J T e (w k)j e (w k )(w w k ). Differeniaing he above wr w and seing he resul o, we ge Differeniaing he above wr w, we ge e(w) = X T. So, he Jacobian becomes J e (w) = ( e(w)) T = X. Plugging his in o he Gauss-Newon equaion, we finally ge: J T e (w k)e(w k )J T e (w k)j e (w k )(ww k ) =, from which we ge w = w k (J T e (w k)j e (w k )) J T e (w k)e(w k ). * J T e (w k)j e (w k ) needs o be nonsingular (inverse is needed). 5 w = w k (X T X) X T (d Xw k ) = w k (X T X) X T d (X T X) X T Xw k }{{} = (X T X) X T d. 6 This is Iw k = w k.
5 Poins worh noing: Linear Leas-Square Filer (con d) X does no need o be a square marix! We ge w = (X T X) X T d off he ba parly because he oupu is linear (oherwise, he formula would be more complex). The Jacobian of he error funcion only depends on he inpu, and is invarian wr he weigh w. The facor (X T X) X T (le s call i X ) is like an inverse. Muliply X o boh sides of d = Xw Linear Leas-Square Filer: Example See src/pseudoinv.m. X = ceil(rand(4,2)*), wrue = rand(2,)*, d=x*wrue, w = inv(x *X)*X *d X = wrue = d = hen we ge: w = X d = X X w. }{{} =I 7 w = Leas-Mean-Square Algorihm Cos funcion is based on insananeous values. E(w) = 2 e2 (w) Differeniaing he above wr w, we ge E(w) w Pluggin in e(w) = d x T w, e(w) w e(w) = e(w) w. E(w) = x, and hence w = xe(w). Using his in he seepes descen rule, we ge he LMS algorihm: ŵ n = ŵ n ηx n e n. Noe ha his weigh updae is done wih only one (x i, d i ) pair! 9 Leas-Mean-Square Algorihm: Evaluaion LMS algorihm behaves like a low-pass filer. LMS algorihm is simple, model-independen, and hus robus. LMS does no follow he direcion of seepes descen: Insead, i follows i sochasically (sochasic gradien descen). Slow convergence is an issue. LMS is sensiive o he inpu correlaion marix s condiion number (raio beween larges vs. smalles eigenvalue of he correl. marix). LMS can be shown o converge if he learning rae has he following propery: < η < 2 λ max where λ max is he larges eigenvalue of he correl. marix. 2
6 Improving Convergence in LMS Search-Then-Converge in LMS The main problem arises because of he fixed η. One soluion: Use a ime-varying learning rae: η(n) = c/n, as in sochasic opimizaion heory. A beer alernaive: use a hybrid mehod called search-hen-converge. η(n) = η (n/τ) When n < τ, performance is similar o sandard LMS. When n > τ, i behaves like sochasic opimizaion. η(n) = η n vs. η(n) = η (n/τ) 2 22 The Percepron Model Par II: Percepron Percepron uses a non-linear neuron model (McCulloch-Pis model). v = m if v > w i x i b, y = φ(v) = if v i= Goal: classify inpu vecors ino wo classes
7 Boolean Logic Gaes wih Percepron Unis Wha Perceprons Can Represen =.5 = W2= =.5 =.5 AND = W2= OR = NOT Russel & Norvig w w Oupu = Slope = W Perceprons can represen basic boolean funcions. Oupu=fs Thus, a nework of percepron unis can compue any Boolean funcion. Wha abou XOR or EQUIV? Perceprons can only represen linearly separable funcions. Oupu of he percepron: W I W I >, hen oupu is W I W I, hen oupu is Geomeric Inerpreaion w w Oupu=fs Oupu = Slope = W w w = The Role of he Bias = Slope = W Rearranging W I W I >, hen oupu is, we ge (if W > ) I > W W I W, where poins above he line, he oupu is, and for hose below he line. Compare wih y = W x. W 27 W Wihou he bias ( = ), learning is limied o adjusmen of he slope of he separaing line passing hrough he origin. Three example lines wih differen weighs are shown. 28
8 Limiaion of Perceprons w w Oupu=fs Oupu = Slope = W x Generalizing o n-dimensions z (x,y,z) n = [a b c] T (x,y,z ) y x y z a b c d hp://mahworld.wolfram.com/plane.hml Only funcions where he poins and poins are clearly linearly separable can be represened by perceprons. The geomeric inerpreaion is generalizable o funcions of n argumens, i.e. percepron wih n inpus plus one hreshold (or bias) uni. 29 n = (a, b, c), x = (x, y, z), x = (x, y, z ). Equaion of a plane: n ( x x ) = In shor, ax by cz d =, where a, b, c can serve as he weigh, and d = n x as he bias. For n-d inpu space, he decision boundary becomes a (n )-D hyperplane (-D less han he inpu space). 3 Linear Separabiliy Linear Separabiliy (con d) Linearlyseparable No Linearlyseparable No Linearlyseparable For funcions ha ake ineger or real values as argumens and oupu eiher or. Lef: linearly separable (i.e., can draw a sraigh line beween he classes). Righ: no linearly separable (i.e., perceprons canno represen such a funcion) AND OR XOR Perceprons canno represen XOR! Minsky and Paper (969)? 3 32
9 XOR in Deail Perceprons: A Differen Perspecive # I I XOR 2 w w Oupu = Slope = W x i w 3 4 Oupu=fs θ d W I W I >, hen oupu is : 2 W > W > 3 W > W > 4 W W W W 2 < W W < (from 2, 3, and 4), bu (from ), a conradicion. 33 w T x > b hen, oupu is w T x = w x cos θ > b hen, oupu is x cos θ > b w So, if d = x cos θ in he figure above is greaer han hen, oupu is b, hen oupu =. w Adjusing w changes he il of he decision boundary, and adjusing he bias b (and w ) moves he decision boundary closer or away from he origin. 34 Percepron Learning Rule Percepron Learning Rule (con d) Given a linearly separable se of inpus ha can belong o class C or C 2, The goal of percepron learning is o have w T x > for all inpu in class C w T x for all inpu in class C 2 If all inpus are correcly classified wih he curren weighs w(n), For misclassified inpus (η(n) is he learning rae): w(n ) = w(n) η(n)x(n) if w T x > and x C 2. w(n ) = w(n) η(n)x(n) if w T x and x C. Or, simply x(n ) = w(n) η(n)e(n)x(n), where e(n) = d(n) y(n) (he error). w(n) T x >, for all inpu in class C, and w(n) T x, for all inpu in class C 2, hen w(n ) = w(n) (no change). Oherwise, adjus he weighs
10 Learning in Percepron: Anoher Look Percepron Convergence Theorem x xw w x wx wx w Given a se of linearly separable inpus, Wihou loss of generaliy, assume η =, w() =. Assume he firs n examples C are all misclassified. Then, using w(n ) = w(n) x(n), we ge w(n ) = x() x(2)... x(n). () When a posiive example (C ) is misclassified, w(n ) = w(n) η(n)x(n). When a negaive example (C 2 ) is misclassified, w(n ) = w(n) η(n)x(n). Noe he il in he weigh vecor, and observe how i would change he decision boundary. 37 Since he inpu se is linearly separable, here is a leas on soluion w such ha w T x(n) > for all inpus in C. Define α = min x(n) C w T x(n) >. Muliply boh sides in eq. wih w, we ge: w T w(n) = wt x()wt x(2)...wt x(n). (2) From he wo seps above, we ge: w T w(n ) > nα (3) 38 Percepron Convergence Theorem (con d) Percepron Convergence Theorem (con d) Using Cauchy-Schwarz inequaliy ] w 2 w(n ) 2 [w T 2 w(n ) From he above and w T w(n ) > nα, So, finally, we ge w 2 w(n ) 2 n 2 α 2 w(n ) 2 n2 α 2 w 2 }{{} Firs main resul 39 (4) Taking he Euclidean norm of w(k ) = w(k) x(k), w(k ) 2 = w(k) 2 2w T (k)x(k) x(k) 2 Since all n inpus in C are misclassified, w T (k)x(k) for k =, 2,..., n, w(k ) 2 w(k) 2 x(k) 2 = 2w T (k)x(k), w(k ) 2 w(k) 2 x(k) 2 w(k ) 2 w(k) 2 x(k) 2 Summing up he inequaliies for all k =, 2,..., n, and w() =, we ge w(k ) 2 n x(k) 2 nβ, (5) k= where β = max x (k) C x(k) 2. 4
11 Percepron Convergence Theorem (con d) From eq. 4 and eq. 5, n 2 α 2 w 2 w(n ) 2 nβ Here, α is a consan, depending on he fixed inpu se and he fixed soluion w (so, w is also a consan), and β is also a consan since i depends only on he fixed inpu se. In his case, if n grows o a large value, he above inequaliy will becomes invalid (n is a posiive ineger). Fixed-Incremen Convergence Theorem Le he subses of raining vecors C and C 2 be linearly separable. Le he inpus presened o percepron originae from hese wo subses. The percepron converges afer some n ieraions, in he sense ha w(n ) = w(n ) = w(n 2) =... is a soluion vecor for n n max. Thus, n canno grow beyond a cerain n max, where n 2 max α2 w 2 = n maxβ n max = β w 2 α 2, and when n = n max, all inpus will be correcly classified 4 42 Summary Adapive filer using he LMS algorihm and perceprons are closely relaed (he learning rule is almos idenical). LMS and perceprons are differen, however, since one uses linear acivaion and he oher hard limiers. LMS is used in coninuous learning, while perceprons are rained for only a finie number of seps. Single-neuron or single-layer has severe limis: How can muliple layers help? 43 44
12 XOR wih Mulilayer Perceprons XOR AND Noe: he bias unis are no shown in he nework on he righ, bu hey are needed. Only hree percepron unis are needed o implemen XOR. However, you need wo layers o achieve his. 45
The Rosenblatt s LMS algorithm for Perceptron (1958) is built around a linear neuron (a neuron with a linear
In The name of God Lecure4: Percepron and AALIE r. Majid MjidGhoshunih Inroducion The Rosenbla s LMS algorihm for Percepron 958 is buil around a linear neuron a neuron ih a linear acivaion funcion. Hoever,
More informationCourse Notes for EE227C (Spring 2018): Convex Optimization and Approximation
Course Noes for EE7C Spring 018: Convex Opimizaion and Approximaion Insrucor: Moriz Hard Email: hard+ee7c@berkeley.edu Graduae Insrucor: Max Simchowiz Email: msimchow+ee7c@berkeley.edu Ocober 15, 018 3
More informationt is a basis for the solution space to this system, then the matrix having these solutions as columns, t x 1 t, x 2 t,... x n t x 2 t...
Mah 228- Fri Mar 24 5.6 Marix exponenials and linear sysems: The analogy beween firs order sysems of linear differenial equaions (Chaper 5) and scalar linear differenial equaions (Chaper ) is much sronger
More information15. Vector Valued Functions
1. Vecor Valued Funcions Up o his poin, we have presened vecors wih consan componens, for example, 1, and,,4. However, we can allow he componens of a vecor o be funcions of a common variable. For example,
More information2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes
Some common engineering funcions 2.7 Inroducion This secion provides a caalogue of some common funcions ofen used in Science and Engineering. These include polynomials, raional funcions, he modulus funcion
More informationSection 3.5 Nonhomogeneous Equations; Method of Undetermined Coefficients
Secion 3.5 Nonhomogeneous Equaions; Mehod of Undeermined Coefficiens Key Terms/Ideas: Linear Differenial operaor Nonlinear operaor Second order homogeneous DE Second order nonhomogeneous DE Soluion o homogeneous
More informationLecture 9: September 25
0-725: Opimizaion Fall 202 Lecure 9: Sepember 25 Lecurer: Geoff Gordon/Ryan Tibshirani Scribes: Xuezhi Wang, Subhodeep Moira, Abhimanu Kumar Noe: LaTeX emplae couresy of UC Berkeley EECS dep. Disclaimer:
More informationLecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.
Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in
More information1 1 + x 2 dx. tan 1 (2) = ] ] x 3. Solution: Recall that the given integral is improper because. x 3. 1 x 3. dx = lim dx.
. Use Simpson s rule wih n 4 o esimae an () +. Soluion: Since we are using 4 seps, 4 Thus we have [ ( ) f() + 4f + f() + 4f 3 [ + 4 4 6 5 + + 4 4 3 + ] 5 [ + 6 6 5 + + 6 3 + ]. 5. Our funcion is f() +.
More informationMath Week 14 April 16-20: sections first order systems of linear differential equations; 7.4 mass-spring systems.
Mah 2250-004 Week 4 April 6-20 secions 7.-7.3 firs order sysems of linear differenial equaions; 7.4 mass-spring sysems. Mon Apr 6 7.-7.2 Sysems of differenial equaions (7.), and he vecor Calculus we need
More informationCHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK
175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he
More informationINTRODUCTION TO MACHINE LEARNING 3RD EDITION
ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class
More informationDimitri Solomatine. D.P. Solomatine. Data-driven modelling (part 2). 2
Daa-driven modelling. Par. Daa-driven Arificial di Neural modelling. Newors Par Dimiri Solomaine Arificial neural newors D.P. Solomaine. Daa-driven modelling par. 1 Arificial neural newors ANN: main pes
More information23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes
Represening Periodic Funcions by Fourier Series 3. Inroducion In his Secion we show how a periodic funcion can be expressed as a series of sines and cosines. We begin by obaining some sandard inegrals
More information1 Review of Zero-Sum Games
COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any
More informationMATH 5720: Gradient Methods Hung Phan, UMass Lowell October 4, 2018
MATH 5720: Gradien Mehods Hung Phan, UMass Lowell Ocober 4, 208 Descen Direcion Mehods Consider he problem min { f(x) x R n}. The general descen direcions mehod is x k+ = x k + k d k where x k is he curren
More informationChapter 6. Systems of First Order Linear Differential Equations
Chaper 6 Sysems of Firs Order Linear Differenial Equaions We will only discuss firs order sysems However higher order sysems may be made ino firs order sysems by a rick shown below We will have a sligh
More informationTwo Coupled Oscillators / Normal Modes
Lecure 3 Phys 3750 Two Coupled Oscillaors / Normal Modes Overview and Moivaion: Today we ake a small, bu significan, sep owards wave moion. We will no ye observe waves, bu his sep is imporan in is own
More informationKINEMATICS IN ONE DIMENSION
KINEMATICS IN ONE DIMENSION PREVIEW Kinemaics is he sudy of how hings move how far (disance and displacemen), how fas (speed and velociy), and how fas ha how fas changes (acceleraion). We say ha an objec
More informationChapter 3 Boundary Value Problem
Chaper 3 Boundary Value Problem A boundary value problem (BVP) is a problem, ypically an ODE or a PDE, which has values assigned on he physical boundary of he domain in which he problem is specified. Le
More informationLet us start with a two dimensional case. We consider a vector ( x,
Roaion marices We consider now roaion marices in wo and hree dimensions. We sar wih wo dimensions since wo dimensions are easier han hree o undersand, and one dimension is a lile oo simple. However, our
More informationGround Rules. PC1221 Fundamentals of Physics I. Kinematics. Position. Lectures 3 and 4 Motion in One Dimension. A/Prof Tay Seng Chuan
Ground Rules PC11 Fundamenals of Physics I Lecures 3 and 4 Moion in One Dimension A/Prof Tay Seng Chuan 1 Swich off your handphone and pager Swich off your lapop compuer and keep i No alking while lecure
More informationLecture 20: Riccati Equations and Least Squares Feedback Control
34-5 LINEAR SYSTEMS Lecure : Riccai Equaions and Leas Squares Feedback Conrol 5.6.4 Sae Feedback via Riccai Equaions A recursive approach in generaing he marix-valued funcion W ( ) equaion for i for he
More informationVehicle Arrival Models : Headway
Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where
More informationCHAPTER 12 DIRECT CURRENT CIRCUITS
CHAPTER 12 DIRECT CURRENT CIUITS DIRECT CURRENT CIUITS 257 12.1 RESISTORS IN SERIES AND IN PARALLEL When wo resisors are conneced ogeher as shown in Figure 12.1 we said ha hey are conneced in series. As
More informationLearning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power
Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.
More informationArticle from. Predictive Analytics and Futurism. July 2016 Issue 13
Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning
More informationEnsamble methods: Boosting
Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room
More informationNeural Networks. Understanding the Brain
Neural Neworks Threshold unis Neural Neworks Gradien descen Mulilayer neworks Backpropagaion Hidden layer represenaions Example: Face Recogniion Advanced opics And, more Neworks of processing unis (neurons)
More informationEnsamble methods: Bagging and Boosting
Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par
More informationPENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD
PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.
More informationSOLUTIONS TO ECE 3084
SOLUTIONS TO ECE 384 PROBLEM 2.. For each sysem below, specify wheher or no i is: (i) memoryless; (ii) causal; (iii) inverible; (iv) linear; (v) ime invarian; Explain your reasoning. If he propery is no
More informationLearning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power
Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.
More information3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon
3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of
More informationEECE 301 Signals & Systems Prof. Mark Fowler
EECE 31 Signals & Sysems Prof. Mark Fowler Noe Se #1 C-T Sysems: Convoluion Represenaion Reading Assignmen: Secion 2.6 of Kamen and Heck 1/11 Course Flow Diagram The arrows here show concepual flow beween
More informationGMM - Generalized Method of Moments
GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................
More informationMatlab and Python programming: how to get started
Malab and Pyhon programming: how o ge sared Equipping readers he skills o wrie programs o explore complex sysems and discover ineresing paerns from big daa is one of he main goals of his book. In his chaper,
More informationFinal Spring 2007
.615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o
More information6.003 Homework #9 Solutions
6.00 Homework #9 Soluions Problems. Fourier varieies a. Deermine he Fourier series coefficiens of he following signal, which is periodic in 0. x () 0 0 a 0 5 a k sin πk 5 sin πk 5 πk for k 0 a k 0 πk j
More informationThen. 1 The eigenvalues of A are inside R = n i=1 R i. 2 Union of any k circles not intersecting the other (n k)
Ger sgorin Circle Chaper 9 Approimaing Eigenvalues Per-Olof Persson persson@berkeley.edu Deparmen of Mahemaics Universiy of California, Berkeley Mah 128B Numerical Analysis (Ger sgorin Circle) Le A be
More informationSome Basic Information about M-S-D Systems
Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,
More informationPhysics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle
Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,
More informationSimulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010
Simulaion-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Week Descripion Reading Maerial 2 Compuer Simulaion of Dynamic Models Finie Difference, coninuous saes, discree ime Simple Mehods Euler Trapezoid
More informationL07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms
L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS NA568 Mobile Roboics: Mehods & Algorihms Today s Topic Quick review on (Linear) Kalman Filer Kalman Filering for Non-Linear Sysems Exended Kalman Filer (EKF)
More informationNotes on online convex optimization
Noes on online convex opimizaion Karl Sraos Online convex opimizaion (OCO) is a principled framework for online learning: OnlineConvexOpimizaion Inpu: convex se S, number of seps T For =, 2,..., T : Selec
More informationHamilton- J acobi Equation: Explicit Formulas In this lecture we try to apply the method of characteristics to the Hamilton-Jacobi equation: u t
M ah 5 2 7 Fall 2 0 0 9 L ecure 1 0 O c. 7, 2 0 0 9 Hamilon- J acobi Equaion: Explici Formulas In his lecure we ry o apply he mehod of characerisics o he Hamilon-Jacobi equaion: u + H D u, x = 0 in R n
More information6.003 Homework #9 Solutions
6.003 Homework #9 Soluions Problems. Fourier varieies a. Deermine he Fourier series coefficiens of he following signal, which is periodic in 0. x () 0 3 0 a 0 5 a k a k 0 πk j3 e 0 e j πk 0 jπk πk e 0
More informationA Dynamic Model of Economic Fluctuations
CHAPTER 15 A Dynamic Model of Economic Flucuaions Modified for ECON 2204 by Bob Murphy 2016 Worh Publishers, all righs reserved IN THIS CHAPTER, OU WILL LEARN: how o incorporae dynamics ino he AD-AS model
More informationLie Derivatives operator vector field flow push back Lie derivative of
Lie Derivaives The Lie derivaive is a mehod of compuing he direcional derivaive of a vecor field wih respec o anoher vecor field We already know how o make sense of a direcional derivaive of real valued
More informationEchocardiography Project and Finite Fourier Series
Echocardiography Projec and Finie Fourier Series 1 U M An echocardiagram is a plo of how a porion of he hear moves as he funcion of ime over he one or more hearbea cycles If he hearbea repeas iself every
More informationOnline Convex Optimization Example And Follow-The-Leader
CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion
More informationTraveling Waves. Chapter Introduction
Chaper 4 Traveling Waves 4.1 Inroducion To dae, we have considered oscillaions, i.e., periodic, ofen harmonic, variaions of a physical characerisic of a sysem. The sysem a one ime is indisinguishable from
More informationEE363 homework 1 solutions
EE363 Prof. S. Boyd EE363 homework 1 soluions 1. LQR for a riple accumulaor. We consider he sysem x +1 = Ax + Bu, y = Cx, wih 1 1 A = 1 1, B =, C = [ 1 ]. 1 1 This sysem has ransfer funcion H(z) = (z 1)
More informationChapter 2. First Order Scalar Equations
Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.
More informationMath 10B: Mock Mid II. April 13, 2016
Name: Soluions Mah 10B: Mock Mid II April 13, 016 1. ( poins) Sae, wih jusificaion, wheher he following saemens are rue or false. (a) If a 3 3 marix A saisfies A 3 A = 0, hen i canno be inverible. True.
More informationCSE/NB 528 Lecture 14: From Supervised to Reinforcement Learning (Chapter 9) R. Rao, 528: Lecture 14
CSE/NB 58 Lecure 14: From Supervised o Reinforcemen Learning Chaper 9 1 Recall from las ime: Sigmoid Neworks Oupu v T g w u g wiui w Inpu nodes u = u 1 u u 3 T i Sigmoid oupu funcion: 1 g a 1 a e 1 ga
More informationCHAPTER 2 Signals And Spectra
CHAPER Signals And Specra Properies of Signals and Noise In communicaion sysems he received waveform is usually caegorized ino he desired par conaining he informaion, and he undesired par. he desired par
More informationA Primal-Dual Type Algorithm with the O(1/t) Convergence Rate for Large Scale Constrained Convex Programs
PROC. IEEE CONFERENCE ON DECISION AND CONTROL, 06 A Primal-Dual Type Algorihm wih he O(/) Convergence Rae for Large Scale Consrained Convex Programs Hao Yu and Michael J. Neely Absrac This paper considers
More informationFinish reading Chapter 2 of Spivak, rereading earlier sections as necessary. handout and fill in some missing details!
MAT 257, Handou 6: Ocober 7-2, 20. I. Assignmen. Finish reading Chaper 2 of Spiva, rereading earlier secions as necessary. handou and fill in some missing deails! II. Higher derivaives. Also, read his
More information10. State Space Methods
. Sae Space Mehods. Inroducion Sae space modelling was briefly inroduced in chaper. Here more coverage is provided of sae space mehods before some of heir uses in conrol sysem design are covered in he
More information6.003 Homework #8 Solutions
6.003 Homework #8 Soluions Problems. Fourier Series Deermine he Fourier series coefficiens a k for x () shown below. x ()= x ( + 0) 0 a 0 = 0 a k = e /0 sin(/0) for k 0 a k = π x()e k d = 0 0 π e 0 k d
More informationGuest Lectures for Dr. MacFarlane s EE3350 Part Deux
Gues Lecures for Dr. MacFarlane s EE3350 Par Deux Michael Plane Mon., 08-30-2010 Wrie name in corner. Poin ou his is a review, so I will go faser. Remind hem o go lisen o online lecure abou geing an A
More information556: MATHEMATICAL STATISTICS I
556: MATHEMATICAL STATISTICS I INEQUALITIES 5.1 Concenraion and Tail Probabiliy Inequaliies Lemma (CHEBYCHEV S LEMMA) c > 0, If X is a random variable, hen for non-negaive funcion h, and P X [h(x) c] E
More informationDeep Learning: Theory, Techniques & Applications - Recurrent Neural Networks -
Deep Learning: Theory, Techniques & Applicaions - Recurren Neural Neworks - Prof. Maeo Maeucci maeo.maeucci@polimi.i Deparmen of Elecronics, Informaion and Bioengineering Arificial Inelligence and Roboics
More informationTHE 2-BODY PROBLEM. FIGURE 1. A pair of ellipses sharing a common focus. (c,b) c+a ROBERT J. VANDERBEI
THE 2-BODY PROBLEM ROBERT J. VANDERBEI ABSTRACT. In his shor noe, we show ha a pair of ellipses wih a common focus is a soluion o he 2-body problem. INTRODUCTION. Solving he 2-body problem from scrach
More informationT L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB
Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal
More informationParticle Swarm Optimization
Paricle Swarm Opimizaion Speaker: Jeng-Shyang Pan Deparmen of Elecronic Engineering, Kaohsiung Universiy of Applied Science, Taiwan Email: jspan@cc.kuas.edu.w 7/26/2004 ppso 1 Wha is he Paricle Swarm Opimizaion
More information!!"#"$%&#'()!"#&'(*%)+,&',-)./0)1-*23)
"#"$%&#'()"#&'(*%)+,&',-)./)1-*) #$%&'()*+,&',-.%,/)*+,-&1*#$)()5*6$+$%*,7&*-'-&1*(,-&*6&,7.$%$+*&%'(*8$&',-,%'-&1*(,-&*6&,79*(&,%: ;..,*&1$&$.$%&'()*1$$.,'&',-9*(&,%)?%*,('&5
More informationMath 2142 Exam 1 Review Problems. x 2 + f (0) 3! for the 3rd Taylor polynomial at x = 0. To calculate the various quantities:
Mah 4 Eam Review Problems Problem. Calculae he 3rd Taylor polynomial for arcsin a =. Soluion. Le f() = arcsin. For his problem, we use he formula f() + f () + f ()! + f () 3! for he 3rd Taylor polynomial
More informationInventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions
Muli-Period Sochasic Models: Opimali of (s, S) Polic for -Convex Objecive Funcions Consider a seing similar o he N-sage newsvendor problem excep ha now here is a fixed re-ordering cos (> 0) for each (re-)order.
More informationIMPLICIT AND INVERSE FUNCTION THEOREMS PAUL SCHRIMPF 1 OCTOBER 25, 2013
IMPLICI AND INVERSE FUNCION HEOREMS PAUL SCHRIMPF 1 OCOBER 25, 213 UNIVERSIY OF BRIISH COLUMBIA ECONOMICS 526 We have exensively sudied how o solve sysems of linear equaions. We know how o check wheher
More informationSupplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence
Supplemen for Sochasic Convex Opimizaion: Faser Local Growh Implies Faser Global Convergence Yi Xu Qihang Lin ianbao Yang Proof of heorem heorem Suppose Assumpion holds and F (w) obeys he LGC (6) Given
More informationGeorey E. Hinton. University oftoronto. Technical Report CRG-TR February 22, Abstract
Parameer Esimaion for Linear Dynamical Sysems Zoubin Ghahramani Georey E. Hinon Deparmen of Compuer Science Universiy oftorono 6 King's College Road Torono, Canada M5S A4 Email: zoubin@cs.orono.edu Technical
More informationSolutions from Chapter 9.1 and 9.2
Soluions from Chaper 9 and 92 Secion 9 Problem # This basically boils down o an exercise in he chain rule from calculus We are looking for soluions of he form: u( x) = f( k x c) where k x R 3 and k is
More informationRobust estimation based on the first- and third-moment restrictions of the power transformation model
h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,
More information1 Solutions to selected problems
1 Soluions o seleced problems 1. Le A B R n. Show ha in A in B bu in general bd A bd B. Soluion. Le x in A. Then here is ɛ > 0 such ha B ɛ (x) A B. This shows x in B. If A = [0, 1] and B = [0, 2], hen
More informationOnline Appendix to Solution Methods for Models with Rare Disasters
Online Appendix o Soluion Mehods for Models wih Rare Disasers Jesús Fernández-Villaverde and Oren Levinal In his Online Appendix, we presen he Euler condiions of he model, we develop he pricing Calvo block,
More informationMathcad Lecture #8 In-class Worksheet Curve Fitting and Interpolation
Mahcad Lecure #8 In-class Workshee Curve Fiing and Inerpolaion A he end of his lecure, you will be able o: explain he difference beween curve fiing and inerpolaion decide wheher curve fiing or inerpolaion
More informationEXERCISES FOR SECTION 1.5
1.5 Exisence and Uniqueness of Soluions 43 20. 1 v c 21. 1 v c 1 2 4 6 8 10 1 2 2 4 6 8 10 Graph of approximae soluion obained using Euler s mehod wih = 0.1. Graph of approximae soluion obained using Euler
More informationClass Meeting # 10: Introduction to the Wave Equation
MATH 8.5 COURSE NOTES - CLASS MEETING # 0 8.5 Inroducion o PDEs, Fall 0 Professor: Jared Speck Class Meeing # 0: Inroducion o he Wave Equaion. Wha is he wave equaion? The sandard wave equaion for a funcion
More informationMulti-scale 2D acoustic full waveform inversion with high frequency impulsive source
Muli-scale D acousic full waveform inversion wih high frequency impulsive source Vladimir N Zubov*, Universiy of Calgary, Calgary AB vzubov@ucalgaryca and Michael P Lamoureux, Universiy of Calgary, Calgary
More informationEcon107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8)
I. Definiions and Problems A. Perfec Mulicollineariy Econ7 Applied Economerics Topic 7: Mulicollineariy (Sudenmund, Chaper 8) Definiion: Perfec mulicollineariy exiss in a following K-variable regression
More informationAppendix to Online l 1 -Dictionary Learning with Application to Novel Document Detection
Appendix o Online l -Dicionary Learning wih Applicaion o Novel Documen Deecion Shiva Prasad Kasiviswanahan Huahua Wang Arindam Banerjee Prem Melville A Background abou ADMM In his secion, we give a brief
More informationConcourse Math Spring 2012 Worked Examples: Matrix Methods for Solving Systems of 1st Order Linear Differential Equations
Concourse Mah 80 Spring 0 Worked Examples: Marix Mehods for Solving Sysems of s Order Linear Differenial Equaions The Main Idea: Given a sysem of s order linear differenial equaions d x d Ax wih iniial
More informationDiebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles
Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance
More informationSolutions Problem Set 3 Macro II (14.452)
Soluions Problem Se 3 Macro II (14.452) Francisco A. Gallego 04/27/2005 1 Q heory of invesmen in coninuous ime and no uncerainy Consider he in nie horizon model of a rm facing adjusmen coss o invesmen.
More informationLecture Notes 2. The Hilbert Space Approach to Time Series
Time Series Seven N. Durlauf Universiy of Wisconsin. Basic ideas Lecure Noes. The Hilber Space Approach o Time Series The Hilber space framework provides a very powerful language for discussing he relaionship
More informationHOMEWORK # 2: MATH 211, SPRING Note: This is the last solution set where I will describe the MATLAB I used to make my pictures.
HOMEWORK # 2: MATH 2, SPRING 25 TJ HITCHMAN Noe: This is he las soluion se where I will describe he MATLAB I used o make my picures.. Exercises from he ex.. Chaper 2.. Problem 6. We are o show ha y() =
More information5. Stochastic processes (1)
Lec05.pp S-38.45 - Inroducion o Teleraffic Theory Spring 2005 Conens Basic conceps Poisson process 2 Sochasic processes () Consider some quaniy in a eleraffic (or any) sysem I ypically evolves in ime randomly
More information= ( ) ) or a system of differential equations with continuous parametrization (T = R
XIII. DIFFERENCE AND DIFFERENTIAL EQUATIONS Ofen funcions, or a sysem of funcion, are paramerized in erms of some variable, usually denoed as and inerpreed as ime. The variable is wrien as a funcion of
More informationCE 395 Special Topics in Machine Learning
CE 395 Special Topics in Machine Learning Assoc. Prof. Dr. Yuriy Mishchenko Fall 2017 DIGITAL FILTERS AND FILTERING Why filers? Digial filering is he workhorse of digial signal processing Filering is a
More information( ) a system of differential equations with continuous parametrization ( T = R + These look like, respectively:
XIII. DIFFERENCE AND DIFFERENTIAL EQUATIONS Ofen funcions, or a sysem of funcion, are paramerized in erms of some variable, usually denoed as and inerpreed as ime. The variable is wrien as a funcion of
More informationPMR5406 Redes Neurais e Lógica Fuzzy Aula 3 Single Layer Percetron
PMR5406 Redes Neurais e Aula 3 Single Layer Percetron Baseado em: Neural Networks, Simon Haykin, Prentice-Hall, 2 nd edition Slides do curso por Elena Marchiori, Vrije Unviersity Architecture We consider
More informationWEEK-3 Recitation PHYS 131. of the projectile s velocity remains constant throughout the motion, since the acceleration a x
WEEK-3 Reciaion PHYS 131 Ch. 3: FOC 1, 3, 4, 6, 14. Problems 9, 37, 41 & 71 and Ch. 4: FOC 1, 3, 5, 8. Problems 3, 5 & 16. Feb 8, 018 Ch. 3: FOC 1, 3, 4, 6, 14. 1. (a) The horizonal componen of he projecile
More informationPhysics 180A Fall 2008 Test points. Provide the best answer to the following questions and problems. Watch your sig figs.
Physics 180A Fall 2008 Tes 1-120 poins Name Provide he bes answer o he following quesions and problems. Wach your sig figs. 1) The number of meaningful digis in a number is called he number of. When numbers
More informationExplaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015
Explaining Toal Facor Produciviy Ulrich Kohli Universiy of Geneva December 2015 Needed: A Theory of Toal Facor Produciviy Edward C. Presco (1998) 2 1. Inroducion Toal Facor Produciviy (TFP) has become
More informationIB Physics Kinematics Worksheet
IB Physics Kinemaics Workshee Wrie full soluions and noes for muliple choice answers. Do no use a calculaor for muliple choice answers. 1. Which of he following is a correc definiion of average acceleraion?
More informationAnnouncements: Warm-up Exercise:
Fri Apr 13 7.1 Sysems of differenial equaions - o model muli-componen sysems via comparmenal analysis hp//en.wikipedia.org/wiki/muli-comparmen_model Announcemens Warm-up Exercise Here's a relaively simple
More informationEE 315 Notes. Gürdal Arslan CLASS 1. (Sections ) What is a signal?
EE 35 Noes Gürdal Arslan CLASS (Secions.-.2) Wha is a signal? In his class, a signal is some funcion of ime and i represens how some physical quaniy changes over some window of ime. Examples: velociy of
More information