Neural Networks. Class 22: MLSP, Fall 2016 Instructor: Bhiksha Raj
|
|
- Derick Ramsey
- 5 years ago
- Views:
Transcription
1 Neural Networs Class 22: MLSP, Fall 2016 Instructor: Bhsha Raj
2 IMPORTANT ADMINSTRIVIA Fnal wee. Project presentatons on 6th 18797/
3 Neural Networs are tang over! Neural networs have become one of the major thrust areas recently n varous pattern recognton, predcton, and analyss problems In many problems they have establshed the state of the art Often exceedng prevous benchmars by large margns 18797/
4 Recent success wth neural networs Some recent successes wth neural networs A bt of hyperbole, but stll /
5 Recent success wth neural networs Some recent successes wth neural networs A bt of hyperbole, but stll /
6 Recent success wth neural networs Some recent successes wth neural networs A bt of hyperbole, but stll /
7 Recent success wth neural networs Some recent successes wth neural networs A bt of hyperbole, but 18797/11755 stll.. 7
8 Recent success wth neural networs Captons generated entrely by a neural networ 18797/
9 Successes wth neural networs And a varety of other problems: Image recognton Sgnal enhancement Even predctng stoc marets! 18797/
10 So what are neural networs?? Voce sgnal N.Net Transcrpton Image N.Net Text capton Game State N.Net Next move What are these boxes? 18797/
11 So what are neural networs?? It began wth ths.. Humans are very good at the tass we just saw Can we model the human bran/ human ntellgence? An old queston datng 18797/11755 bac to Plato and Arstotle.. 11
12 Observaton: The Bran Md 1800s: The bran s a mass of nterconnected neurons 18797/
13 Bran: Interconnected Neurons Many neurons connect n to each neuron Each neuron connects out to many neurons 18797/
14 The bran s a connectonst machne The human bran s a connectonst machne Ban, A. (1873). Mnd and body. The theores of ther relaton. London: Henry Kng. Ferrer, D. (1876). The Functons of the Bran. London: Smth, Elder and Co Neurons connect to other neurons. The processng/capacty of the bran s a functon of these connectons Connectonst machnes emulate ths structure 18797/
15 Connectonst Machnes Neural networs are connectonst machnes As opposed to Von Neumann Machnes Von Neumann Machne Neural Networ PROCESSOR PROGRAM DATA NETWORK Processng unt Memory The machne has many processng unts The program s the connectons between these unts Connectons may also defne memory 18797/
16 Modellng the bran What are the unts? A neuron: Soma Dendrtes Axon Sgnals come n through the dendrtes nto the Soma A sgnal goes out va the axon to other neurons Only one axon per neuron Factod that may only nterest me: Neurons do not undergo cell dvson 18797/
17 McCullough and Ptts The Doctor and the Hobo.. Warren McCulloch: Neurophyscan Walter Ptts: Homeless wannabe logcan who arrved at hs door 18797/
18 The McCulloch and Ptts model A sngle neuron A mathematcal model of a neuron McCulloch, W.S. & Ptts, W.H. (1943). A Logcal Calculus of the Ideas Immanent n Nervous Actvty, Bulletn of Mathematcal Bophyscs, 5: , 1943 Threshold Logc Note: McCullough and Ptts orgnal model was actually slghtly dfferent ths model s actually due to Rosenblatt 18
19 The soluton to everythng Fran Rosenblatt Psychologst, Logcan Inventor of the soluton to everythng, aa the Perceptron (1958) A mathematcal model of the neuron that could solve everythng!!! 19
20 Smplfed mathematcal model Number of nputs combne lnearly Threshold logc: Fre f combned nput exceeds threshold Y = 1 f w x + b > 0 0 else 18797/
21 Smplfed mathematcal model A mathematcal model Orgnally assumed could represent any Boolean crcut Rosenblatt, 1958 : the embryo of an electronc computer that [the Navy] expects wll be able to wal, tal, see, wrte, reproduce tself and be conscous of ts exstence 18797/
22 Perceptron X X Y X -1 0 X Y X X Y Y Boolean Gates But 18797/
23 Perceptron No soluton for XOR! Not unversal! X??? X Y Y Mnsy and Papert, /
24 A sngle neuron s not enough Indvdual elements are wea computatonal elements Marvn Mnsy and Seymour Papert, 1969, Perceptrons: An Introducton to Computatonal Geometry Networed elements are requred 18797/
25 Mult-layer Perceptron! X X Y X Y XOR Y -1-1 Hdden Layer X Y The frst layer s a hdden layer 18797/
26 Mult-Layer Perceptron ( A&X&Z A&Y )&( X & Y X&Z ) X Y Z A Even more complex Boolean functons can be composed usng layered networs of perceptrons Two hdden layers n above model 1 21 In fact can buld any Boolean functon over any number of nputs 18797/
27 Mult-layer perceptrons are unversal Boolean functons A mult-layer perceptron s a unversal Boolean functon In fact, an MLP wth only one hdden layer s a unversal Boolean functon! 18797/
28 Constructng Boolean functons wth only one hdden layer x 1 x 2 y 1 y 2 x 1 x 2 y 1 y 2 + x 1 x 2 y 1 y 2 + x 1 x 2 y 1 y terms 2 n unts 2n nputs Any Boolean formula can be expressed by an MLP wth one hdden layer Any Boolean formula can be expressed n Conjunctve Normal Form The one hdden layer can be exponentally wde But the same formula can be obtaned wth a much smaller networ f we have multple hdden layers 18797/
29 Neural Networs: Mult-layer Perceptrons Voce sgnal N.Net Transcrpton Image N.Net Text capton In realty the nput to these systems s not Boolean Inputs are contnuous valued Sgnals may be contnuous valued Image features are contnuous valued Pxels are mult-valued (e.g ) 29
30 MLP on contnuous nputs The nputs are contnuous valued Threshold logc: Fre f combned nput exceeds threshold Y = 1 f w x + b > 0 0 else 18797/
31 A Perceptron on Reals w 1,w 2 x 2 y = 1 f w x T 0 else x 1 A perceptron operates on real-valued vectors Ths s just a lnear classfer x 2 x /
32 Booleans over the reals x 2 Can now be composed nto networs to compute arbtrary classfcaton boundares x 1 The networ must fre f the nput s n the coloured area 18797/
33 Booleans over the reals x 2 x 1 x 2 x 1 The networ must fre f the nput s n the coloured area 18797/
34 Booleans over the reals x 2 x 1 x 2 x 1 The networ must fre f the nput s n the coloured area 18797/
35 Booleans over the reals x 2 x 1 x 2 x 1 The networ must fre f the nput s n the coloured area 18797/
36 Booleans over the reals x 2 x 1 x 2 x 1 The networ must fre f the nput s n the coloured area 18797/
37 Booleans over the reals x 2 x 1 x 2 x 1 The networ must fre f the nput s n the coloured area 18797/
38 Booleans over the reals 3 N x x =1 AND y N? y 1 y 2 y 3 y 4 y x 2 x 1 The networ must fre f the nput s n the coloured area 18797/
39 Booleans over the reals OR AND AND x 2 OR two polygons A thrd layer s requred x 1 x 1 x /
40 How Complex Can t Get An arbtrarly complex decson boundary Bascally any Boolean functon over the basc lnear boundares Even wth a sngle hdden layer! 18797/
41 Composng a polygon N y =1 AND N? y 1 y 2 y 3 y 4 y 5 x 2 x The polygon net Increasng the number of sdes shrns the area outsde the polygon that have sum close to N 18797/
42 Composng a crcle No nonlnearty appled! + The crcle net Very large number of neurons Crcle can be of arbtrary dameter, at any locaton Acheved wthout usng a thresholdng functon!! 42
43 Addng crcles No nonlnearty appled! + The sum of two crcles sub nets s exactly a net wth output 1 f the nput falls wthn ether crcle 18797/
44 Composng an arbtrary fgure + Just ft n an arbtrary number of crcles More accurate approxmaton wth greater number of smaller crcles A lesson here that we wll refer to agan shortly /
45 Story so far.. Mult-layer perceptrons are Boolean networs They represent Boolean functons over lnear boundares They can approxmate any boundary Usng a suffcently large number of lnear unts Complex Boolean functons are better modeled wth more layers Complex boundares are more compactly represented usng more layers 18797/
46 Lets loo at the weghts x 1 y = 1 f w x T 0 else x 2 x 3 x N y = 1 f x T w T 0 else What do the weghts tell us? The neuron fres f the nner product between the weghts and the nputs exceeds a threshold 18797/
47 The weght as a template x 1 x 2 X T W > T cos θ > T X w x 3 x N θ < cos 1 T X The perceptron fres f the nput s wthn a specfed angle of the weght Represents a convex regon on the surface of the sphere! The networ s a Boolean functon over these regons. The overall decson regon can be arbtrarly nonconvex Neuron fres f the nput vector s close enough to the weght vector. If the nput pattern matches the weght pattern closely enough 18797/
48 The weght as a template W X X y = 1 f w x T 0 else Correlaton = 0.57 Correlaton = 0.82 If the correlaton between the weght pattern and the nputs exceeds a threshold, fre The perceptron s a correlaton flter! 18797/
49 The MLP as a Boolean functon over feature detectors DIGIT OR NOT? The nput layer comprses feature detectors Detect f certan patterns have occurred n the nput The networ s a Boolean functon over the feature detectors I.e. t s mportant for the frst layer to capture relevant patterns 49
50 The MLP as a cascade of feature detectors DIGIT OR NOT? The networ s a cascade of feature detectors Hgher level neurons compose complex templates from features represented by lower-level neurons Rs n ths perspectve: Upper level neurons may be performng OR Loong for a choce of compound patterns 50
51 Story so far MLPs are Boolean machnes They represent Boolean functons over lnear boundares They can represent arbtrary boundares Perceptrons are correlaton flters They detect patterns n the nput MLPs are Boolean formulae over patterns detected by perceptrons Hgher-level perceptrons may also be vewed as feature detectors Extra: MLP n classfcaton The networ wll fre f the combnaton of the detected basc features matches an acceptable pattern for a desred class of sgnal E.g. Approprate combnatons of (Nose, Eyes, Eyebrows, Chee, Chn) Face 18797/
52 MLP as a contnuous-valued regresson h 2 h 1 h n x 1 1 T 1 T 1 T 2 T f(x) T 1 T 2 x h 1 + h 2 x h n MLPs can actually compose arbtrary functons to arbtrary precson Not just classfcaton/boolean functons 1D example Left: A net wth a par of unts can create a pulse of any wdth at any locaton Rght: A networ of N such pars approxmates the functon wth N scaled pulses 52
53 MLP as a contnuous-valued regresson + h 1 h 2 h 1 h 2 h n h n MLPs can actually compose arbtrary functons Even wth only one layer To arbtrary precson The MLP s a unversal approxmator! 18797/
54 Mult-layer perceptrons are unversal functon approxmators X, Y N.Net Z, Colour A mult-layer perceptron s a unversal functon approxmator Horn, Stnchcombe and Whte 1989, several others 18797/
55 What s nsde these boxes? Voce sgnal N.Net Transcrpton Image N.Net Text capton Game State N.Net Next move Each of these tass s performed by a net le ths one Functons that tae the gven nput and produce the requred output 18797/
56 Story so far MLPs are Boolean machnes They represent arbtrary Boolean functons over arbtrary lnear boundares MLPs perform classfcaton MLPs can compute arbtrary real-valued functons of arbtrary real-valued nputs To arbtrary precson They are unversal approxmators 18797/
57 Buldng a networ for a tas 18797/
58 These tass are functons Voce sgnal N.Net Transcrpton Image N.Net Text capton Game State N.Net Next move Each of these boxes s actually a functon E.g f: Image Capton 18797/
59 These tass are functons Voce sgnal Transcrpton Image Text capton Game State Next move Each box s actually a functon E.g f: Image Capton It can be approxmated by a neural networ 59
60 The networ as a functon Input Output Inputs are numerc vectors Numerc representaton of nput, e.g. audo, mage, game state, etc. Outputs are numerc scalars or vectors Numerc encodng of output from whch actual output can be derved E.g. a score, whch can be compared to a threshold to decde f the nput s a face or not Output may be mult-dmensonal, f tas requres t 18797/
61 The networ s a functon Gven an nput, t computes the functon layer wse to predct an output 18797/
62 A note on actvatons x 1 x 2 x 3 x N sgmod tanh Prevous explanatons assumed networ unts used a threshold functon for actvaton In realty, we use a number of other dfferentable functons Mostly, but not always, squashng functons 62
63 The entre networ Inputs (D-dmensonal): X 1,, X D Weghts from th node n l th layer to j th node n l+1 th l layer: W,j Complete set of weghts {W l,j ; l = 0.. L, = 1 N l, j = 1 N l+1 } Outputs (M-dmensonal): O 1,, O M 18797/
64 Mang a predcton (Aa forward pass) Illustraton wth a networ wth X N-1 hdden layers f N O 18797/
65 Forward Computaton X I 1 f 1 O 1 I N-1 O N-1 IN O N f 1 f 1 f N f N f 1 I W 1 j 1 j X f N 1 For the jth neuron of the 1 st hdden layer: Input to the actvaton functons n the frst layer 18797/
66 Forward Computaton X I 1 f 1 O 1 I N-1 O N-1 IN O N f 1 f 1 f N f N f 1 I W 1 j 1 j X f N 1 For the jth neuron of the 1 st hdden layer: Input to the actvaton functons n the frst layer O 1 j f ( 1 1 I j ) Output of the jth neuron of the 1 st hdden layer: 18797/
67 Forward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N f N f f N 1 For the jth neuron of the th hdden layer: Input to the actvaton functons n the th layer I j W j O /
68 Forward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N f N f f N 1 For the jth neuron of the th hdden layer: Input to the actvaton functons n the th layer I j W j O 1 O j Output of the jth neuron of the th hdden layer: f ( I j 18797/ )
69 Forward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N f N I W 1 j 1 j X f f N 1 ITERATE FOR = 1:N for j = 1:layer-wdth Output O N s the output of the networ O I j j W f ( I j j ) O /
70 The problem of tranng the networ The networ must be composed correctly to produce the desred outputs Ths s the problem of tranng the networ 18797/
71 The Problem of Tranng a Networ f(x) X Y = f(x) X Y = f(x) A neural networ s effectvely just a functon What maes t an NNet s the structure of the functon Taes an nput Generally mult-dmensonal Produces an output May be un- or mult-dmensonal Challenge: How to mae t produce the desred output for any gven nput 18797/
72 Buldng by desgn Y X 2 X 2 X 1 X 1 Soluton 1: Just hand desgn t What wll the networ for the above functon be? We can hand draw ths one /
73 Tranng from examples Most functons are too complex to hand-desgn Partcularly n hgh-dmensonal spaces Instead, we wll try to learn t from tranng examples Input-output pars In realty the tranng data wll not be even remotely as dense as shown n the fgure It wll be several orders of magntude sparser 73
74 General approach to tranng Blue lnes: error when functon s below desred output Blac lnes: error when functon s above desred output Defne an error between the actual networ output for any parameter value and the desred output Error can be defned n a number of ways Many Dvergence functons have been defned Typcally defned as the sum of the error over ndvdual tranng nstances 18797/
75 Recall: A note on actvatons x 1 x 2 x 3 x N sgmod tanh Composng/learnng networs usng threshold actvatons s a combnatoral problem Exponentally hard to fnd the rght soluton The smoother dfferentable actvaton functons enable learnng through optmzaton 75
76 ERROR Mnmzng Error Problem: Fnd the parameters at whch ths functon acheves a mnmum Subject to any constrants we may pose Typcally, ths cannot be found usng a closed-form formula 18797/
77 The Approach of Gradent Descent Iteratve soluton: Start at some pont Fnd drecton n whch to shft ths pont to decrease error Ths can be found from the slope of the functon A postve slope movng left decreases error A negatve slope movng rght decreases error Shft pont n ths drecton The sze of the shft depends 18797/11755 on the slope 77
78 The Approach of Gradent Descent Mult-dmensonal functon: Slope replaced by vector gradent From current pont shft n drecton opposte to gradent Sze of shft depends on magntude of gradent y y x f x y x f f ), ( ), ( 18797/
79 The Approach of Gradent Descent Steps are smaller when the slope s smaller, because the optmum value s generally at a locaton of near-zero slope The gradent descent algorthm W W η W E X; W Untl error E(X;W) converges η s the step sze 18797/
80 The Approach of Gradent Descent Steps are smaller when the slope s smaller, because the optmum value s generally at a locaton of near-zero slope Needs computaton of gradent of error w.r.t. networ parameters The gradent descent algorthm W W η W E X; W Untl error E(X;W) converges η s the step sze 18797/
81 Gradents: Algebrac Formulaton W N f N O X W 1 f 1 f 1 W f f N 1 Weghts matrx Actvaton functon The networ s a nested functon o = f N (W N f N 1 W N 1 f N 2 W N 2 f (W f 1 ( f 1 W 1 X ) )) E = Dv y, o = y o /
82 A note on algebra Y f N O Dv(O,Y) E W N W X W 1 f 1 f 1 I f f N 1 I W f 1 W E E f T I 1 Usng f to represent the output of the actvatons rather than O for easer nterpretaton Usng subscrpts rather than superscrpts to represent layer 18797/
83 A note on algebra E E I f I f X f 1 W 1 f f N 1 W N f N O Y E Dv(O,Y) I ', ',2 ', L f f f f I E f f f E f f I f ',1 ',1 ', ', ',1 ' L f f f E E f I f ' 18797/
84 Gradents: Algebrac Formulaton Y f N O Dv(O,Y) E W N X W 1 f 1 Generc rule ' T W E f E f f 1 T f 1 s the vector of outputs of the -1 th layer f E s the gradent of the error w.r.t. the vector of outputs of the th layer ' f f 1 W f s the vector of dervatves of the actvatons of the th layer w.r.t ther nputs W f N 1 E 1 ' f1 f 1 E X taen at current nput W f /
85 Gradents: Algebrac Formulaton Y f N O Dv(O,Y) E W N X W 1 f 1 f 1 W f f N 1 ' T W E f E f f 1 What s ths? 18797/
86 A note on algebra Y f N O Dv(O,Y) E W N W X W 1 f 1 f 1 I f f N 1 I W f 1 f 1 I W T f 1 f f 1 I I f W T ' f 0 0 0,1 f 0 ',2 0 0 ', L 18797/ f 0 0 0
87 Gradents: Algebrac Formulaton Y f N O Dv(O,Y) E W N X W 1 f 1 f 1 W f W +1 f +1 ' T W E f E f f 1 f E ( T ' ) f f E W 1 f 1 1 f 1 f 1 E T ' f E W f E f 18797/
88 Gradents: Algebrac Formulaton ' T W E f E f f 1 T ' f E W f 1 1 f1 E f N W N O Y Dv(O,Y) E X W 1 f 1 f N E O f 1 W Output layer f Dv( O, Y ) f N 1 Dv( O, Y O1 Dv( O, Y O N ) ) 18797/
89 Gradents: Algebrac Formulaton Y f N O Dv(O,Y) E W N X W 1 f 1 f 1 W f f N 1 ' T ' T ' T ' T W E f W W WN N ODv O Y 1 f 1 2 f 2... f (, ) f /
90 Gradents: Bac Propagaton Y f N O Dv(O,Y) E W N X W 1 f 1 f 1 W f f N 1 Forward pass: Intalze f X 0 For = 1 to N: f f W f ) ( 1 Output O f N 18797/
91 Gradents: Bac Propagaton Y f N O Dv(O,Y) E W N X W 1 f 1 f 1 W f f N 1 Error dvergence: For = N downto 1: f N E Dv O ( O, Y ) ' T W E f E f f 1 (actual recurson) 1 T ' f E W f f E 91
92 BP: Local Formulaton Y f N O Dv(O,Y) E W N X W 1 f 1 f 1 W f f N 1 The networ agan 18797/
93 Gradents: Local Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N f f N 1 Redrawn Separately label nput and output of each node 18797/
94 Forward Computaton X I 1 f 1 O 1 I N-1 O N-1 IN O N f 1 f 1 f N f N f 1 I W 1 j 1 j X f N /
95 Forward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N f N I W 1 j 1 j X f f N 1 I j W j O /
96 Forward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N f N I W 1 j 1 j X f f N 1 I j W j O 1 O j f ( I j 18797/ )
97 Forward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N f N I W 1 j 1 j X f f N 1 ITERATE FOR = 1:N for j = 1:layer-wdth O I j j W f ( I j j ) O /
98 Gradents: Bacward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N f f N /
99 Gradents: Bacward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N E Dv( O, Y ) O O N N f f N /
100 Gradents: Bacward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N E Dv( O, Y ) O O N N f f N 1 E I N do di N N E O N f ' N ( I N ) E O N 18797/
101 Gradents: Bacward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N f f N 1 E Dv( O, Y ) O O E I N N f ' N ( I N N E ) O N E N O 1 j I O N j N 1 E I N j j W E I N N N 1 N Note: j I O W other N j j j 18797/ stuff
102 Gradents: Bacward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N E I f f ' ( I ) E O f N 1 E Dv( O, Y ) O O E I N N E N O 1 f ' N ( I N N E ) O N N E Wj I j N j 18797/
103 Gradents: Bacward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N f f N 1 E Dv( O, Y ) O O E I N f ' ( I N E ) O E O 1 j I O j 1 E I j j W E I j j 18797/
104 Gradents: Bacward Computaton I -1 O -1 W j I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N E W j I W j j f E I j O 1 f N 1 E I j E Dv( O, Y ) O O E I N E O 1 f ' ( I N E ) O E Wj I j j 18797/
105 Gradents: Bacward Computaton I -1 O -1 I O I N-1 O N-1 IN O N f N Dv(O,Y) E f N Intalze: Gradent w.r.t networ output E Dv( O, Y ) O O N N For = N..1 For = 1:layer-wdth E I E O E 1 f ' ( I E Wj I j E ) O E /11755 Wj 105 I j O
106 Bacpropagaton: Multplcatve Networs Some types of networs have multplcatve combnaton I -1 O -1 Z W Forward: 1 1 j O O z Bacward: j j z E O z E O z O E j z E O O E /
107 Overall Approach For each data nstance Forward pass: Pass nstance forward through the net. Store all ntermedate outputs of all computaton Bacward pass: Sweep bacward through the net, teratvely compute all dervatves w.r.t weghts Actual Error s the sum of the error over all tranng nstances E = Dv y(x), o(x) = E x x Actual gradent s the sum or average of the dervatves computed for each tranng nstance x W E = x W E(x) W W η W E 18797/
108 Issues and Challenges What does t learn? Speed Wll not address ths Does t do what we want t to do? Next class: Varatons of nnets MLP, Convoluton, recurrence Nnets for varous tass Image recognton, speech recognton, sgnal enhancement, modellng language /
Admin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester
0/25/6 Admn Assgnment 7 Class /22 Schedule for the rest of the semester NEURAL NETWORKS Davd Kauchak CS58 Fall 206 Perceptron learnng algorthm Our Nervous System repeat untl convergence (or for some #
More informationMultilayer Perceptron (MLP)
Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationEEE 241: Linear Systems
EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they
More informationNeural networks. Nuno Vasconcelos ECE Department, UCSD
Neural networs Nuno Vasconcelos ECE Department, UCSD Classfcaton a classfcaton problem has two types of varables e.g. X - vector of observatons (features) n the world Y - state (class) of the world x X
More informationMultilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata
Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,
More informationMulti-layer neural networks
Lecture 0 Mult-layer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Lnear regresson w Lnear unts f () Logstc regresson T T = w = p( y =, w) = g( w ) w z f () = p ( y = ) w d w d Gradent
More informationIntroduction to the Introduction to Artificial Neural Network
Introducton to the Introducton to Artfcal Neural Netork Vuong Le th Hao Tang s sldes Part of the content of the sldes are from the Internet (possbly th modfcatons). The lecturer does not clam any onershp
More informationEvaluation of classifiers MLPs
Lecture Evaluaton of classfers MLPs Mlos Hausrecht mlos@cs.ptt.edu 539 Sennott Square Evaluaton For any data set e use to test the model e can buld a confuson matrx: Counts of examples th: class label
More informationMultilayer neural networks
Lecture Multlayer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Mdterm exam Mdterm Monday, March 2, 205 In-class (75 mnutes) closed book materal covered by February 25, 205 Multlayer
More informationLecture 10 Support Vector Machines. Oct
Lecture 10 Support Vector Machnes Oct - 20-2008 Lnear Separators Whch of the lnear separators s optmal? Concept of Margn Recall that n Perceptron, we learned that the convergence rate of the Perceptron
More information1 Convex Optimization
Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationCHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD
CHALMERS, GÖTEBORGS UNIVERSITET SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 35, FIM 72 GU, PhD Tme: Place: Teachers: Allowed materal: Not allowed: January 2, 28, at 8 3 2 3 SB
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationNeural Networks. Perceptrons and Backpropagation. Silke Bussen-Heyen. 5th of Novemeber Universität Bremen Fachbereich 3. Neural Networks 1 / 17
Neural Networks Perceptrons and Backpropagaton Slke Bussen-Heyen Unverstät Bremen Fachberech 3 5th of Novemeber 2012 Neural Networks 1 / 17 Contents 1 Introducton 2 Unts 3 Network structure 4 Snglelayer
More informationMATH 567: Mathematical Techniques in Data Science Lab 8
1/14 MATH 567: Mathematcal Technques n Data Scence Lab 8 Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 11, 2017 Recall We have: a (2) 1 = f(w (1) 11 x 1 + W (1) 12 x 2 + W
More informationModel of Neurons. CS 416 Artificial Intelligence. Early History of Neural Nets. Cybernetics. McCulloch-Pitts Neurons. Hebbian Modification.
Page 1 Model of Neurons CS 416 Artfcal Intellgence Lecture 18 Neural Nets Chapter 20 Multple nputs/dendrtes (~10,000!!!) Cell body/soma performs computaton Sngle output/axon Computaton s typcally modeled
More informationKernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan
Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationFeature Selection: Part 1
CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationNeural Networks & Learning
Neural Netorks & Learnng. Introducton The basc prelmnares nvolved n the Artfcal Neural Netorks (ANN) are descrbed n secton. An Artfcal Neural Netorks (ANN) s an nformaton-processng paradgm that nspred
More informationLecture 23: Artificial neural networks
Lecture 23: Artfcal neural networks Broad feld that has developed over the past 20 to 30 years Confluence of statstcal mechancs, appled math, bology and computers Orgnal motvaton: mathematcal modelng of
More informationEnsemble Methods: Boosting
Ensemble Methods: Boostng Ncholas Ruozz Unversty of Texas at Dallas Based on the sldes of Vbhav Gogate and Rob Schapre Last Tme Varance reducton va baggng Generate new tranng data sets by samplng wth replacement
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationNeural Networks. Neural Network Motivation. Why Neural Networks? Comments on Blue Gene. More Comments on Blue Gene
Motvaton for non-lnear Classfers Neural Networs CPS 27 Ron Parr Lnear methods are wea Mae strong assumptons Can only express relatvely smple functons of nputs Comng up wth good features can be hard Why
More informationNonlinear Classifiers II
Nonlnear Classfers II Nonlnear Classfers: Introducton Classfers Supervsed Classfers Lnear Classfers Perceptron Least Squares Methods Lnear Support Vector Machne Nonlnear Classfers Part I: Mult Layer Neural
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationPattern Classification
Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher
More information1 Matrix representations of canonical matrices
1 Matrx representatons of canoncal matrces 2-d rotaton around the orgn: ( ) cos θ sn θ R 0 = sn θ cos θ 3-d rotaton around the x-axs: R x = 1 0 0 0 cos θ sn θ 0 sn θ cos θ 3-d rotaton around the y-axs:
More informationStructure and Drive Paul A. Jensen Copyright July 20, 2003
Structure and Drve Paul A. Jensen Copyrght July 20, 2003 A system s made up of several operatons wth flow passng between them. The structure of the system descrbes the flow paths from nputs to outputs.
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationSupporting Information
Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to
More informationWe present the algorithm first, then derive it later. Assume access to a dataset {(x i, y i )} n i=1, where x i R d and y i { 1, 1}.
CS 189 Introducton to Machne Learnng Sprng 2018 Note 26 1 Boostng We have seen that n the case of random forests, combnng many mperfect models can produce a snglodel that works very well. Ths s the dea
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationIV. Performance Optimization
IV. Performance Optmzaton A. Steepest descent algorthm defnton how to set up bounds on learnng rate mnmzaton n a lne (varyng learnng rate) momentum learnng examples B. Newton s method defnton Gauss-Newton
More informationADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING
1 ADVANCED ACHINE LEARNING ADVANCED ACHINE LEARNING Non-lnear regresson technques 2 ADVANCED ACHINE LEARNING Regresson: Prncple N ap N-dm. nput x to a contnuous output y. Learn a functon of the type: N
More informationOnline Classification: Perceptron and Winnow
E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng
More informationInternet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks
Internet Engneerng Jacek Mazurkewcz, PhD Softcomputng Part 3: Recurrent Artfcal Neural Networks Self-Organsng Artfcal Neural Networks Recurrent Artfcal Neural Networks Feedback sgnals between neurons Dynamc
More informationVQ widely used in coding speech, image, and video
at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationLecture 3: Dual problems and Kernels
Lecture 3: Dual problems and Kernels C4B Machne Learnng Hlary 211 A. Zsserman Prmal and dual forms Lnear separablty revsted Feature mappng Kernels for SVMs Kernel trck requrements radal bass functons SVM
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017
U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that
More informationModule 3: Element Properties Lecture 1: Natural Coordinates
Module 3: Element Propertes Lecture : Natural Coordnates Natural coordnate system s bascally a local coordnate system whch allows the specfcaton of a pont wthn the element by a set of dmensonless numbers
More informationLecture 12: Discrete Laplacian
Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly
More informationLinear Feature Engineering 11
Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19
More informationCSC 411 / CSC D11 / CSC C11
18 Boostng s a general strategy for learnng classfers by combnng smpler ones. The dea of boostng s to take a weak classfer that s, any classfer that wll do at least slghtly better than chance and use t
More informationINF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018
INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton
More informationErrors for Linear Systems
Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationProblem Set 9 Solutions
Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem
More informationDiscriminative classifier: Logistic Regression. CS534-Machine Learning
Dscrmnatve classfer: Logstc Regresson CS534-Machne Learnng robablstc Classfer Gven an nstance, hat does a probablstc classfer do dfferentl compared to, sa, perceptron? It does not drectl predct Instead,
More informationChapter Newton s Method
Chapter 9. Newton s Method After readng ths chapter, you should be able to:. Understand how Newton s method s dfferent from the Golden Secton Search method. Understand how Newton s method works 3. Solve
More information1 Input-Output Mappings. 2 Hebbian Failure. 3 Delta Rule Success.
Task Learnng 1 / 27 1 Input-Output Mappngs. 2 Hebban Falure. 3 Delta Rule Success. Input-Output Mappngs 2 / 27 0 1 2 3 4 5 6 7 8 9 Output 3 8 2 7 Input 5 6 0 9 1 4 Make approprate: Response gven stmulus.
More informationGrover s Algorithm + Quantum Zeno Effect + Vaidman
Grover s Algorthm + Quantum Zeno Effect + Vadman CS 294-2 Bomb 10/12/04 Fall 2004 Lecture 11 Grover s algorthm Recall that Grover s algorthm for searchng over a space of sze wors as follows: consder the
More informationCSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography
CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve
More informationSection 8.3 Polar Form of Complex Numbers
80 Chapter 8 Secton 8 Polar Form of Complex Numbers From prevous classes, you may have encountered magnary numbers the square roots of negatve numbers and, more generally, complex numbers whch are the
More informationC4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )
C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z
More informationRBF Neural Network Model Training by Unscented Kalman Filter and Its Application in Mechanical Fault Diagnosis
Appled Mechancs and Materals Submtted: 24-6-2 ISSN: 662-7482, Vols. 62-65, pp 2383-2386 Accepted: 24-6- do:.428/www.scentfc.net/amm.62-65.2383 Onlne: 24-8- 24 rans ech Publcatons, Swtzerland RBF Neural
More informationWE extend the familiar unidirectional backpropagation
To appear n the IEEE Transactons on Systems, Man, and Cybernetcs Bdrectonal Bacpropagaton Olaoluwa Adgun, Member, IEEE, and Bart Koso, Fellow, IEEE Abstract We extend bacpropagaton learnng from ordnary
More informationPHYS 705: Classical Mechanics. Calculus of Variations II
1 PHYS 705: Classcal Mechancs Calculus of Varatons II 2 Calculus of Varatons: Generalzaton (no constrant yet) Suppose now that F depends on several dependent varables : We need to fnd such that has a statonary
More informationMACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression
11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING
More informationSolving Nonlinear Differential Equations by a Neural Network Method
Solvng Nonlnear Dfferental Equatons by a Neural Network Method Luce P. Aarts and Peter Van der Veer Delft Unversty of Technology, Faculty of Cvlengneerng and Geoscences, Secton of Cvlengneerng Informatcs,
More informationFundamentals of Computational Neuroscience 2e
Fundamentals of Computatonal Neuroscence e Thomas Trappenberg February 7, 9 Chapter 6: Feed-forward mappng networks Dgtal representaton of letter A 3 3 4 5 3 33 4 5 34 35
More informationSupport Vector Machines
Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x n class
More informationGaussian Mixture Models
Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous
More informationMultigradient for Neural Networks for Equalizers 1
Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT
More informationSupport Vector Machines
CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at
More information18-660: Numerical Methods for Engineering Design and Optimization
8-66: Numercal Methods for Engneerng Desgn and Optmzaton n L Department of EE arnege Mellon Unversty Pttsburgh, PA 53 Slde Overve lassfcaton Support vector machne Regularzaton Slde lassfcaton Predct categorcal
More informationSociété de Calcul Mathématique SA
Socété de Calcul Mathématque SA Outls d'ade à la décson Tools for decson help Probablstc Studes: Normalzng the Hstograms Bernard Beauzamy December, 202 I. General constructon of the hstogram Any probablstc
More informationBoostrapaggregating (Bagging)
Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod
More informationUnsupervised Learning
Unsupervsed Learnng Kevn Swngler What s Unsupervsed Learnng? Most smply, t can be thought of as learnng to recognse and recall thngs Recognton I ve seen that before Recall I ve seen that before and I can
More informationChapter 9: Statistical Inference and the Relationship between Two Variables
Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,
More informationOutline and Reading. Dynamic Programming. Dynamic Programming revealed. Computing Fibonacci. The General Dynamic Programming Technique
Outlne and Readng Dynamc Programmng The General Technque ( 5.3.2) -1 Knapsac Problem ( 5.3.3) Matrx Chan-Product ( 5.3.1) Dynamc Programmng verson 1.4 1 Dynamc Programmng verson 1.4 2 Dynamc Programmng
More information1 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations
Physcs 171/271 -Davd Klenfeld - Fall 2005 (revsed Wnter 2011) 1 Dervaton of Rate Equatons from Sngle-Cell Conductance (Hodgkn-Huxley-lke) Equatons We consder a network of many neurons, each of whch obeys
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? + + + + + + + + + Intuton of Margn Consder ponts
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We
More informationChapter 12. Ordinary Differential Equation Boundary Value (BV) Problems
Chapter. Ordnar Dfferental Equaton Boundar Value (BV) Problems In ths chapter we wll learn how to solve ODE boundar value problem. BV ODE s usuall gven wth x beng the ndependent space varable. p( x) q(
More informationHopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen
Hopfeld networks and Boltzmann machnes Geoffrey Hnton et al. Presented by Tambet Matsen 18.11.2014 Hopfeld network Bnary unts Symmetrcal connectons http://www.nnwj.de/hopfeld-net.html Energy functon The
More informationAssortment Optimization under MNL
Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.
More informationMaximal Margin Classifier
CS81B/Stat41B: Advanced Topcs n Learnng & Decson Makng Mamal Margn Classfer Lecturer: Mchael Jordan Scrbes: Jana van Greunen Corrected verson - /1/004 1 References/Recommended Readng 1.1 Webstes www.kernel-machnes.org
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationGenerative classification models
CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn
More informationSDMML HT MSc Problem Sheet 4
SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be
More informationLinear Classification, SVMs and Nearest Neighbors
1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush
More information1 GSW Iterative Techniques for y = Ax
1 for y = A I m gong to cheat here. here are a lot of teratve technques that can be used to solve the general case of a set of smultaneous equatons (wrtten n the matr form as y = A), but ths chapter sn
More informationReport on Image warping
Report on Image warpng Xuan Ne, Dec. 20, 2004 Ths document summarzed the algorthms of our mage warpng soluton for further study, and there s a detaled descrpton about the mplementaton of these algorthms.
More informationChapter 2 - The Simple Linear Regression Model S =0. e i is a random error. S β2 β. This is a minimization problem. Solution is a calculus exercise.
Chapter - The Smple Lnear Regresson Model The lnear regresson equaton s: where y + = β + β e for =,..., y and are observable varables e s a random error How can an estmaton rule be constructed for the
More informationThe exam is closed book, closed notes except your one-page cheat sheet.
CS 89 Fall 206 Introducton to Machne Learnng Fnal Do not open the exam before you are nstructed to do so The exam s closed book, closed notes except your one-page cheat sheet Usage of electronc devces
More informationCIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M
CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute
More informationAGC Introduction
. Introducton AGC 3 The prmary controller response to a load/generaton mbalance results n generaton adjustment so as to mantan load/generaton balance. However, due to droop, t also results n a non-zero
More informationRadial-Basis Function Networks
Radal-Bass uncton Networs v.0 March 00 Mchel Verleysen Radal-Bass uncton Networs - Radal-Bass uncton Networs p Orgn: Cover s theorem p Interpolaton problem p Regularzaton theory p Generalzed RBN p Unversal
More informationSupport Vector Machines
Support Vector Machnes Konstantn Tretyakov (kt@ut.ee) MTAT.03.227 Machne Learnng So far Supervsed machne learnng Lnear models Least squares regresson Fsher s dscrmnant, Perceptron, Logstc model Non-lnear
More informationMachine Learning & Data Mining CS/CNS/EE 155. Lecture 4: Regularization, Sparsity & Lasso
Machne Learnng Data Mnng CS/CS/EE 155 Lecture 4: Regularzaton, Sparsty Lasso 1 Recap: Complete Ppelne S = {(x, y )} Tranng Data f (x, b) = T x b Model Class(es) L(a, b) = (a b) 2 Loss Functon,b L( y, f
More informationBézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0
Bézer curves Mchael S. Floater September 1, 215 These notes provde an ntroducton to Bézer curves. 1 Bernsten polynomals Recall that a real polynomal of a real varable x R, wth degree n, s a functon of
More informationTHE SUMMATION NOTATION Ʃ
Sngle Subscrpt otaton THE SUMMATIO OTATIO Ʃ Most of the calculatons we perform n statstcs are repettve operatons on lsts of numbers. For example, we compute the sum of a set of numbers, or the sum of the
More information2016 Wiley. Study Session 2: Ethical and Professional Standards Application
6 Wley Study Sesson : Ethcal and Professonal Standards Applcaton LESSON : CORRECTION ANALYSIS Readng 9: Correlaton and Regresson LOS 9a: Calculate and nterpret a sample covarance and a sample correlaton
More information