18.7 Artificial Neural Networks
|
|
- Hilda Shaw
- 5 years ago
- Views:
Transcription
1 Artfcl Neurl Networks Neuroscence hs hypotheszed tht mentl ctvty conssts prmrly of electrochemcl ctvty n networks of brn cells clled neurons Ths led McCulloch nd Ptts to devse ther mthemtcl model of the neuron lredy n 1943 Roughly spekng, t fres when lner combnton of ts nputs exceeds some (hrd or soft) threshold Hence, t mplements lner clssfer A neurl network s ust collecton of unts connected together The propertes of the network re determned by ts topology nd the propertes of the neurons Nmes for the reserch feld: connectonsm, prllel dstrbuted processng, neurl computton, nd computtonl neuroscence =1 w 0, Actvton Functon Output Lnks w, g Input Lnks Input Functon Output g n 0 w, 1
2 312 Neurl network structures Neurl networks re composed of nodes or unts connected by drected lnks A lnk from unt to unt serves to propgte the ctvton from to Ech lnk lso hs weght w, ssocted wth t, whch determnes the strength nd sgn of the connecton Ech unt hs dummy nput 0 = 1 wth n ssocted weght w 0, (lner regresson model) Ech unt frst computes weghted sum of ts nputs: n n 0 w, 313 Then the unt pples n ctvton functon g to ths sum to derve the output g( n ) g 0 The ctvton functon g s typclly ether A hrd threshold, n whch cse the unt s clled perceptron, or A logstc functon, n whch cse the term sgmod unt s sometmes used Both of these nonlner ctvton functon ensure the mportnt property tht the entre network of unts cn represent nonlner functon n w, 2
3 314 There re two fundmentlly dstnct wys to connect ndvdul neurons together to form network A feed-forwrd network hs connectons only n one drecton.e., t forms drected cyclc grph Every node receves nput from upstrem nodes nd delvers output to downstrem nodes; there re no loops A feed-forwrd network represents functon of ts current nput It hs no nternl stte other thn the weghts themselves A recurrent network feeds ts output bck nto ts own nputs Ths mens tht the ctvton levels of the network form dynmcl system tht my rech stble stte or exhbt osclltons or even chotc behvor Moreover, the response of the network to gven nput depends on ts ntl stte, whch my depend on prevous nputs Hence, recurrent networks cn support short-term memory Interestng models of the brn, but dffcult to understnd 315 Feed-forwrd networks re usully rrnged n lyers, such tht ech unt receves nput only from unts n the mmedtely precedng lyer Multlyer networks hve one or more lyers of hdden unts tht re not connected to the outputs of the network A network wth ll the nputs connected drectly to the outputs s clled sngle-lyer neurl network, or perceptron network Let us thnk of feed-forwrd neurl network s functon h w (x) prmetrzed by the weghts We cn express the output s functon of the nputs nd the weghts As long s we cn clculte the dervtves of such expressons wth respect to the weghts, we cn use the grdent-descent loss-mnmzton method to trn the network 3
4 316 Becuse the functon represented by network cn be hghly nonlner composed, s t s, of nested nonlner soft threshold functons we cn vew neurl networks s tool for dong nonlner regresson Wth sngle, suffcently lrge hdden lyer, t s possble to represent ny contnuous functon of the nputs wth rbtrry ccurcy Wth two lyers, even dscontnuous functons cn be represented Unfortuntely, for ny prtculr network structure, t s hrd to chrcterze exctly whch functons cn be represented nd whch ones cnnot We cn bck-propgte the error from the output lyer to the hdden lyers The bck-propgton process emerges drectly from dervton of the overll error grdent 317 Lner clssfers Let f(x) be lner functon of ts rgument vector x = (x 1,,x m ) T f ( x) w x b w x b w x b Input x s clssfed postve, f f(x) 0 nd otherwse x s clssfed negtve 1, sgn( f ( x)) 1, The hyperplne determned by the equton w x+b = 0 dvdes the nput spce n two hlfspces If for n exmple (x, y) nd hypothess h t holds tht yh(x) > 0, then the exmple hs been correctly clssfed = yh(x) s the mrgn of the exmple w.r.t. h m 1 f ( x) 0 otherwse 4
5 Support Vector Mchnes Vldmr Vpnk & collegues, 1990 s The SVM frmework s currently the most populr pproch for off-the-shelf supervsed lernng Three dvntges of SVMs: 1.SVMs construct mxmum mrgn seprtor decson boundry wth the lrgest possble dstnce to exmple ponts. Ths helps them to generlze well 2.SVMs crete lner seprtng hyperplne, but they hve the blty to embed the dt nto hgher-dmensonl spce usng the so-clled kernel trck. Often dt tht re not lnerly seprble n the orgnl nput spce re esly seprble n the hgher-dmensonl spce. The hgh-dmensonl lner seprtor s ctully nonlner n the orgnl spce. Ths mens the hypothess spce s gretly expnded over methods tht use strctly lner representtons SVMs re nonprmetrc method they retn trnng exmples nd potentlly need to store them ll. In prctce they often end up retnng only smll frcton of the number of exmples sometmes s few s smll constnt tmes the number of dmensons. SVMs combne the dvntges of nonprmetrc nd prmetrc models: they hve flexblty to represent complex functons, but they re resstnt to overfttng Consder lnerly seprble bnry clssfcton problem. There s n nfnte number of decson boundres consstent wth ll the dt From the pont of vew of 0/1 loss ech would be eqully good Logstc regresson would fnd some seprtng lne, the exct locton of whch depends on ll the exmple ponts The key nsght of SVMs s tht some exmples re more mportnt thn other, nd tht pyng ttenton to them cn led to better generlzton 5
6 320 Insted of mnmzng expected emprcl loss on the trnng dt, SVMs ttempt to mnmze expected generlzton loss Under the probblstc ssumpton tht the s-yet-unseen ponts re drwn from the sme dstrbuton s the prevously seen exmples, there re rguments suggestng tht we mnmze generlzton loss by choosng the seprtor tht s frthest wy from the exmples we hve seen so fr Ths s clled the mxmum mrgn seprtor The mrgn s twce the dstnce from the seprtor to the nerest exmple pont Trdtonlly SVMs use the conventon tht clss lbels re +1 nd 1 The seprtor s defned s the set of ponts {x w x + b = 0} Grdent descent n the spce of w nd b could be used to fnd the prmeters tht mxmze the mrgn whle correctly clssfyng ll the exmples 321 There s nother pproch to solvng the problem There s n lterntve representton clled the dul representton, n whch the optml soluton s found by solvng 1 rgmx k y yk x xk 2, k subect to constrnts 0 nd y = 0 Ths s qudrtc progrmmng optmzton problem, for whch there re good softwre pckges Once we hve found the vector we cn get bck to w wth the equton w = x, or we cn sty n the dul representton The frst mportnt property of the bove dul optmzton problem s tht the expresson s convex: t hs sngle globl mxmum tht cn be found effcently 6
7 322 Second, the dt enter the expresson n the form of dot products of prs of ponts Ths s lso true of the equton for the seprtor tself; once the optml hve been clculted, t s h ( x) sgn y ( x x ) b The thrd mportnt property s tht the weghts ssocted wth ech dt pont re zero except for support vectors the ponts closest to the seprtor Becuse there re usully mny fewer support vectors thn exmples, SVMs gn some of the dvntges of prmetrc models 323 However, we would not usully expect to fnd lner seprtor n the nput spce x, but we cn fnd lner seprtors n hghdmensonl feture spce F(x) smply by replcng x x k n the dul representton optmzton problem wth F(x ) F(x k ) Kernel functon K llows us to fnd lner seprtors n the hgherdmensonl feture spce F(x) smply by replcng x x k wth kernel functon K(x, x k ) Thus we cn lern n the hgher-dmensonl spce, but we compute only kernel functons rther thn the full lst of fetures for ech dt pont As n exmple, F(x ) F(x k ) = (x x k ) 2 when we use three fetures: f 1 = x 12 f 2 = x 2 2 f 3 = 2 x 1 x 2 Kernel trck: Pluggng the kernels nto the optmzton problem optml lner seprtors cn be found effcently n feture spces wth bllons of (or, n some cses, nfntely mny) dmensons 7
An Introduction to Support Vector Machines
An Introducton to Support Vector Mchnes Wht s good Decson Boundry? Consder two-clss, lnerly seprble clssfcton problem Clss How to fnd the lne (or hyperplne n n-dmensons, n>)? Any de? Clss Per Lug Mrtell
More informationMachine Learning Support Vector Machines SVM
Mchne Lernng Support Vector Mchnes SVM Lesson 6 Dt Clssfcton problem rnng set:, D,,, : nput dt smple {,, K}: clss or lbel of nput rget: Construct functon f : X Y f, D Predcton of clss for n unknon nput
More informationCS434a/541a: Pattern Recognition Prof. Olga Veksler. Lecture 9
CS434/541: Pttern Recognton Prof. Olg Veksler Lecture 9 Announcements Fnl project proposl due Nov. 1 1-2 prgrph descrpton Lte Penlt: s 1 pont off for ech d lte Assgnment 3 due November 10 Dt for fnl project
More informationSVMs for regression Multilayer neural networks
Lecture SVMs for regresson Muter neur netors Mos Husrecht mos@cs.ptt.edu 539 Sennott Squre Support vector mchne SVM SVM mmze the mrgn round the seprtng hperpne. he decson functon s fu specfed suset of
More informationSVMs for regression Non-parametric/instance based classification method
S 75 Mchne ernng ecture Mos Huskrecht mos@cs.ptt.edu 539 Sennott Squre SVMs for regresson Non-prmetrc/nstnce sed cssfcton method S 75 Mchne ernng Soft-mrgn SVM Aos some fet on crossng the seprtng hperpne
More informationPrinciple Component Analysis
Prncple Component Anlyss Jng Go SUNY Bufflo Why Dmensonlty Reducton? We hve too mny dmensons o reson bout or obtn nsghts from o vsulze oo much nose n the dt Need to reduce them to smller set of fctors
More informationUNIVERSITY OF IOANNINA DEPARTMENT OF ECONOMICS. M.Sc. in Economics MICROECONOMIC THEORY I. Problem Set II
Mcroeconomc Theory I UNIVERSITY OF IOANNINA DEPARTMENT OF ECONOMICS MSc n Economcs MICROECONOMIC THEORY I Techng: A Lptns (Note: The number of ndctes exercse s dffculty level) ()True or flse? If V( y )
More informationLecture 4: Piecewise Cubic Interpolation
Lecture notes on Vrtonl nd Approxmte Methods n Appled Mthemtcs - A Perce UBC Lecture 4: Pecewse Cubc Interpolton Compled 6 August 7 In ths lecture we consder pecewse cubc nterpolton n whch cubc polynoml
More informationRank One Update And the Google Matrix by Al Bernstein Signal Science, LLC
Introducton Rnk One Updte And the Google Mtrx y Al Bernsten Sgnl Scence, LLC www.sgnlscence.net here re two dfferent wys to perform mtrx multplctons. he frst uses dot product formulton nd the second uses
More informationDennis Bricker, 2001 Dept of Industrial Engineering The University of Iowa. MDP: Taxi page 1
Denns Brcker, 2001 Dept of Industrl Engneerng The Unversty of Iow MDP: Tx pge 1 A tx serves three djcent towns: A, B, nd C. Ech tme the tx dschrges pssenger, the drver must choose from three possble ctons:
More informationPartially Observable Systems. 1 Partially Observable Markov Decision Process (POMDP) Formalism
CS294-40 Lernng for Rootcs nd Control Lecture 10-9/30/2008 Lecturer: Peter Aeel Prtlly Oservle Systems Scre: Dvd Nchum Lecture outlne POMDP formlsm Pont-sed vlue terton Glol methods: polytree, enumerton,
More informationMachine Learning. Support Vector Machines. Le Song. CSE6740/CS7641/ISYE6740, Fall Lecture 8, Sept. 13, 2012 Based on slides from Eric Xing, CMU
Mchne Lernng CSE6740/CS764/ISYE6740 Fll 0 Support Vector Mchnes Le Song Lecture 8 Sept. 3 0 Bsed on sldes fro Erc Xng CMU Redng: Chp. 6&7 C.B ook Outlne Mu rgn clssfcton Constrned optzton Lgrngn dult Kernel
More informationReview of linear algebra. Nuno Vasconcelos UCSD
Revew of lner lgebr Nuno Vsconcelos UCSD Vector spces Defnton: vector spce s set H where ddton nd sclr multplcton re defned nd stsf: ) +( + ) (+ )+ 5) λ H 2) + + H 6) 3) H, + 7) λ(λ ) (λλ ) 4) H, - + 8)
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationSupport vector machines for regression
S 75 Mchne ernng ecture 5 Support vector mchnes for regresson Mos Huskrecht mos@cs.ptt.edu 539 Sennott Squre S 75 Mchne ernng he decson oundr: ˆ he decson: Support vector mchnes ˆ α SV ˆ sgn αˆ SV!!: Decson
More informationMultilayer Perceptron (MLP)
Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne
More informationESCI 342 Atmospheric Dynamics I Lesson 1 Vectors and Vector Calculus
ESI 34 tmospherc Dnmcs I Lesson 1 Vectors nd Vector lculus Reference: Schum s Outlne Seres: Mthemtcl Hndbook of Formuls nd Tbles Suggested Redng: Mrtn Secton 1 OORDINTE SYSTEMS n orthonorml coordnte sstem
More informationNon-Linear Data for Neural Networks Training and Testing
Proceedngs of the 4th WSEAS Int Conf on Informton Securty, Communctons nd Computers, Tenerfe, Spn, December 6-8, 005 (pp466-47) Non-Lner Dt for Neurl Networks Trnng nd Testng ABDEL LATIF ABU-DALHOUM MOHAMMED
More informationChapter Newton-Raphson Method of Solving a Nonlinear Equation
Chpter.4 Newton-Rphson Method of Solvng Nonlner Equton After redng ths chpter, you should be ble to:. derve the Newton-Rphson method formul,. develop the lgorthm of the Newton-Rphson method,. use the Newton-Rphson
More informationDCDM BUSINESS SCHOOL NUMERICAL METHODS (COS 233-8) Solutions to Assignment 3. x f(x)
DCDM BUSINESS SCHOOL NUMEICAL METHODS (COS -8) Solutons to Assgnment Queston Consder the followng dt: 5 f() 8 7 5 () Set up dfference tble through fourth dfferences. (b) Wht s the mnmum degree tht n nterpoltng
More informationIntroduction to Numerical Integration Part II
Introducton to umercl Integrton Prt II CS 75/Mth 75 Brn T. Smth, UM, CS Dept. Sprng, 998 4/9/998 qud_ Intro to Gussn Qudrture s eore, the generl tretment chnges the ntegrton prolem to ndng the ntegrl w
More informationApplied Statistics Qualifier Examination
Appled Sttstcs Qulfer Exmnton Qul_june_8 Fll 8 Instructons: () The exmnton contns 4 Questons. You re to nswer 3 out of 4 of them. () You my use ny books nd clss notes tht you mght fnd helpful n solvng
More information4. Eccentric axial loading, cross-section core
. Eccentrc xl lodng, cross-secton core Introducton We re strtng to consder more generl cse when the xl force nd bxl bendng ct smultneousl n the cross-secton of the br. B vrtue of Snt-Vennt s prncple we
More informationRemember: Project Proposals are due April 11.
Bonformtcs ecture Notes Announcements Remember: Project Proposls re due Aprl. Clss 22 Aprl 4, 2002 A. Hdden Mrov Models. Defntons Emple - Consder the emple we tled bout n clss lst tme wth the cons. However,
More informationChapter Newton-Raphson Method of Solving a Nonlinear Equation
Chpter 0.04 Newton-Rphson Method o Solvng Nonlner Equton Ater redng ths chpter, you should be ble to:. derve the Newton-Rphson method ormul,. develop the lgorthm o the Newton-Rphson method,. use the Newton-Rphson
More informationLeast squares. Václav Hlaváč. Czech Technical University in Prague
Lest squres Václv Hlváč Czech echncl Unversty n Prgue hlvc@fel.cvut.cz http://cmp.felk.cvut.cz/~hlvc Courtesy: Fred Pghn nd J.P. Lews, SIGGRAPH 2007 Course; Outlne 2 Lner regresson Geometry of lest-squres
More informationInternational Journal of Pure and Applied Sciences and Technology
Int. J. Pure Appl. Sc. Technol., () (), pp. 44-49 Interntonl Journl of Pure nd Appled Scences nd Technolog ISSN 9-67 Avlle onlne t www.jopst.n Reserch Pper Numercl Soluton for Non-Lner Fredholm Integrl
More informationQuiz: Experimental Physics Lab-I
Mxmum Mrks: 18 Totl tme llowed: 35 mn Quz: Expermentl Physcs Lb-I Nme: Roll no: Attempt ll questons. 1. In n experment, bll of mss 100 g s dropped from heght of 65 cm nto the snd contner, the mpct s clled
More informationMultiple view geometry
EECS 442 Computer vson Multple vew geometry Perspectve Structure from Moton - Perspectve structure from moton prolem - mgutes - lgerc methods - Fctorzton methods - Bundle djustment - Self-clrton Redng:
More informationDefinition of Tracking
Trckng Defnton of Trckng Trckng: Generte some conclusons bout the moton of the scene, objects, or the cmer, gven sequence of mges. Knowng ths moton, predct where thngs re gong to project n the net mge,
More informationA Family of Multivariate Abel Series Distributions. of Order k
Appled Mthemtcl Scences, Vol. 2, 2008, no. 45, 2239-2246 A Fmly of Multvrte Abel Seres Dstrbutons of Order k Rupk Gupt & Kshore K. Ds 2 Fculty of Scence & Technology, The Icf Unversty, Agrtl, Trpur, Ind
More informationNeural Networks. Perceptrons and Backpropagation. Silke Bussen-Heyen. 5th of Novemeber Universität Bremen Fachbereich 3. Neural Networks 1 / 17
Neural Networks Perceptrons and Backpropagaton Slke Bussen-Heyen Unverstät Bremen Fachberech 3 5th of Novemeber 2012 Neural Networks 1 / 17 Contents 1 Introducton 2 Unts 3 Network structure 4 Snglelayer
More informationMultilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata
Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,
More information6 Roots of Equations: Open Methods
HK Km Slghtly modfed 3//9, /8/6 Frstly wrtten t Mrch 5 6 Roots of Equtons: Open Methods Smple Fed-Pont Iterton Newton-Rphson Secnt Methods MATLAB Functon: fzero Polynomls Cse Study: Ppe Frcton Brcketng
More informationLecture 36. Finite Element Methods
CE 60: Numercl Methods Lecture 36 Fnte Element Methods Course Coordntor: Dr. Suresh A. Krth, Assocte Professor, Deprtment of Cvl Engneerng, IIT Guwht. In the lst clss, we dscussed on the ppromte methods
More informationImproving Anytime Point-Based Value Iteration Using Principled Point Selections
In In Proceedngs of the Twenteth Interntonl Jont Conference on Artfcl Intellgence (IJCAI-7) Improvng Anytme Pont-Bsed Vlue Iterton Usng Prncpled Pont Selectons Mchel R. Jmes, Mchel E. Smples, nd Dmtr A.
More informationThe Schur-Cohn Algorithm
Modelng, Estmton nd Otml Flterng n Sgnl Processng Mohmed Njm Coyrght 8, ISTE Ltd. Aendx F The Schur-Cohn Algorthm In ths endx, our m s to resent the Schur-Cohn lgorthm [] whch s often used s crteron for
More informationBi-level models for OD matrix estimation
TNK084 Trffc Theory seres Vol.4, number. My 2008 B-level models for OD mtrx estmton Hn Zhng, Quyng Meng Abstrct- Ths pper ntroduces two types of O/D mtrx estmton model: ME2 nd Grdent. ME2 s mxmum-entropy
More informationUnsupervised Learning
Unsupervsed Learnng Kevn Swngler What s Unsupervsed Learnng? Most smply, t can be thought of as learnng to recognse and recall thngs Recognton I ve seen that before Recall I ve seen that before and I can
More informationIn this Chapter. Chap. 3 Markov chains and hidden Markov models. Probabilistic Models. Example: CpG Islands
In ths Chpter Chp. 3 Mrov chns nd hdden Mrov models Bontellgence bortory School of Computer Sc. & Eng. Seoul Ntonl Unversty Seoul 5-74, Kore The probblstc model for sequence nlyss HMM (hdden Mrov model)
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationLinear and Nonlinear Optimization
Lner nd Nonlner Optmzton Ynyu Ye Deprtment of Mngement Scence nd Engneerng Stnford Unversty Stnford, CA 9430, U.S.A. http://www.stnford.edu/~yyye http://www.stnford.edu/clss/msnde/ Ynyu Ye, Stnford, MS&E
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationSection 8.3 Polar Form of Complex Numbers
80 Chapter 8 Secton 8 Polar Form of Complex Numbers From prevous classes, you may have encountered magnary numbers the square roots of negatve numbers and, more generally, complex numbers whch are the
More informationINTRODUCTION TO COMPLEX NUMBERS
INTRODUCTION TO COMPLEX NUMBERS The numers -4, -3, -, -1, 0, 1,, 3, 4 represent the negtve nd postve rel numers termed ntegers. As one frst lerns n mddle school they cn e thought of s unt dstnce spced
More information6. Chemical Potential and the Grand Partition Function
6. Chemcl Potentl nd the Grnd Prtton Functon ome Mth Fcts (see ppendx E for detls) If F() s n nlytc functon of stte vrles nd such tht df d pd then t follows: F F p lso snce F p F we cn conclude: p In other
More informationMath 426: Probability Final Exam Practice
Mth 46: Probbility Finl Exm Prctice. Computtionl problems 4. Let T k (n) denote the number of prtitions of the set {,..., n} into k nonempty subsets, where k n. Argue tht T k (n) kt k (n ) + T k (n ) by
More informationMath 497C Sep 17, Curves and Surfaces Fall 2004, PSU
Mth 497C Sep 17, 004 1 Curves nd Surfces Fll 004, PSU Lecture Notes 3 1.8 The generl defnton of curvture; Fox-Mlnor s Theorem Let α: [, b] R n be curve nd P = {t 0,...,t n } be prtton of [, b], then the
More informationChapter 2 Introduction to Algebra. Dr. Chih-Peng Li ( 李 )
Chpter Introducton to Algebr Dr. Chh-Peng L 李 Outlne Groups Felds Bnry Feld Arthetc Constructon of Glos Feld Bsc Propertes of Glos Feld Coputtons Usng Glos Feld Arthetc Vector Spces Groups 3 Let G be set
More informationFall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede. with respect to λ. 1. χ λ χ λ ( ) λ, and thus:
More on χ nd errors : uppose tht we re fttng for sngle -prmeter, mnmzng: If we epnd The vlue χ ( ( ( ; ( wth respect to. χ n Tlor seres n the vcnt of ts mnmum vlue χ ( mn χ χ χ χ + + + mn mnmzes χ, nd
More informationExpected Value and Variance
MATH 38 Expected Value and Varance Dr. Neal, WKU We now shall dscuss how to fnd the average and standard devaton of a random varable X. Expected Value Defnton. The expected value (or average value, or
More informationJens Siebel (University of Applied Sciences Kaiserslautern) An Interactive Introduction to Complex Numbers
Jens Sebel (Unversty of Appled Scences Kserslutern) An Interctve Introducton to Complex Numbers 1. Introducton We know tht some polynoml equtons do not hve ny solutons on R/. Exmple 1.1: Solve x + 1= for
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationprinceton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg
prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there
More informationOnline Appendix to. Mandating Behavioral Conformity in Social Groups with Conformist Members
Onlne Appendx to Mndtng Behvorl Conformty n Socl Groups wth Conformst Members Peter Grzl Andrze Bnk (Correspondng uthor) Deprtment of Economcs, The Wllms School, Wshngton nd Lee Unversty, Lexngton, 4450
More informationEEE 241: Linear Systems
EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they
More informationComputing a complete histogram of an image in Log(n) steps and minimum expected memory requirements using hypercubes
Computng complete hstogrm of n mge n Log(n) steps nd mnmum expected memory requrements usng hypercubes TAREK M. SOBH School of Engneerng, Unversty of Brdgeport, Connectcut, USA. Abstrct Ths work frst revews
More informationUNIFORM CONVERGENCE. Contents 1. Uniform Convergence 1 2. Properties of uniform convergence 3
UNIFORM CONVERGENCE Contents 1. Uniform Convergence 1 2. Properties of uniform convergence 3 Suppose f n : Ω R or f n : Ω C is sequence of rel or complex functions, nd f n f s n in some sense. Furthermore,
More informationLet us look at a linear equation for a one-port network, for example some load with a reflection coefficient s, Figure L6.
ECEN 5004, prng 08 Actve Mcrowve Crcut Zoy Popovc, Unverty of Colordo, Boulder LECURE 5 IGNAL FLOW GRAPH FOR MICROWAVE CIRCUI ANALYI In mny text on mcrowve mplfer (e.g. the clc one by Gonzlez), gnl flow-grph
More informationSection 14.3 Arc Length and Curvature
Section 4.3 Arc Length nd Curvture Clculus on Curves in Spce In this section, we ly the foundtions for describing the movement of n object in spce.. Vector Function Bsics In Clc, formul for rc length in
More informationLimited Dependent Variables
Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages
More information6.6 The Marquardt Algorithm
6.6 The Mqudt Algothm lmttons of the gdent nd Tylo expnson methods ecstng the Tylo expnson n tems of ch-sque devtves ecstng the gdent sech nto n tetve mtx fomlsm Mqudt's lgothm utomtclly combnes the gdent
More informationNeural networks. Nuno Vasconcelos ECE Department, UCSD
Neural networs Nuno Vasconcelos ECE Department, UCSD Classfcaton a classfcaton problem has two types of varables e.g. X - vector of observatons (features) n the world Y - state (class) of the world x X
More information( dg. ) 2 dt. + dt. dt j + dh. + dt. r(t) dt. Comparing this equation with the one listed above for the length of see that
Arc Length of Curves in Three Dimensionl Spce If the vector function r(t) f(t) i + g(t) j + h(t) k trces out the curve C s t vries, we cn mesure distnces long C using formul nerly identicl to one tht we
More informationSparse and Overcomplete Representation: Finding Statistical Orders in Natural Images
Sprse nd Overcomplete Representton: Fndng Sttstcl Orders n Nturl Imges Amr Rez Sffr Azr Insttute for Theoretcl Computer Scence, Grz Unversty of Technology mr@g.tugrz.t Outlne Vsul Cortex. Sprse nd Overcomplete
More informationStructure and Drive Paul A. Jensen Copyright July 20, 2003
Structure and Drve Paul A. Jensen Copyrght July 20, 2003 A system s made up of several operatons wth flow passng between them. The structure of the system descrbes the flow paths from nputs to outputs.
More informationLAPLACE TRANSFORM SOLUTION OF THE PROBLEM OF TIME-FRACTIONAL HEAT CONDUCTION IN A TWO-LAYERED SLAB
Journl of Appled Mthemtcs nd Computtonl Mechncs 5, 4(4), 5-3 www.mcm.pcz.pl p-issn 99-9965 DOI:.75/jmcm.5.4. e-issn 353-588 LAPLACE TRANSFORM SOLUTION OF THE PROBLEM OF TIME-FRACTIONAL HEAT CONDUCTION
More informationTHE COMBINED SHEPARD ABEL GONCHAROV UNIVARIATE OPERATOR
REVUE D ANALYSE NUMÉRIQUE ET DE THÉORIE DE L APPROXIMATION Tome 32, N o 1, 2003, pp 11 20 THE COMBINED SHEPARD ABEL GONCHAROV UNIVARIATE OPERATOR TEODORA CĂTINAŞ Abstrct We extend the Sheprd opertor by
More informationState space systems analysis (continued) Stability. A. Definitions A system is said to be Asymptotically Stable (AS) when it satisfies
Stte spce systems nlysis (continued) Stbility A. Definitions A system is sid to be Asymptoticlly Stble (AS) when it stisfies ut () = 0, t > 0 lim xt () 0. t A system is AS if nd only if the impulse response
More informationAdmin NEURAL NETWORKS. Perceptron learning algorithm. Our Nervous System 10/25/16. Assignment 7. Class 11/22. Schedule for the rest of the semester
0/25/6 Admn Assgnment 7 Class /22 Schedule for the rest of the semester NEURAL NETWORKS Davd Kauchak CS58 Fall 206 Perceptron learnng algorthm Our Nervous System repeat untl convergence (or for some #
More informationp 1 c 2 + p 2 c 2 + p 3 c p m c 2
Where to put a faclty? Gven locatons p 1,..., p m n R n of m houses, want to choose a locaton c n R n for the fre staton. Want c to be as close as possble to all the house. We know how to measure dstance
More informationDemand. Demand and Comparative Statics. Graphically. Marshallian Demand. ECON 370: Microeconomic Theory Summer 2004 Rice University Stanley Gilbert
Demnd Demnd nd Comrtve Sttcs ECON 370: Mcroeconomc Theory Summer 004 Rce Unversty Stnley Glbert Usng the tools we hve develoed u to ths ont, we cn now determne demnd for n ndvdul consumer We seek demnd
More informationModel Fitting and Robust Regression Methods
Dertment o Comuter Engneerng Unverst o Clorn t Snt Cruz Model Fttng nd Robust Regresson Methods CMPE 64: Imge Anlss nd Comuter Vson H o Fttng lnes nd ellses to mge dt Dertment o Comuter Engneerng Unverst
More informationMore metrics on cartesian products
More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of
More informationExercises. 18 Algorithms
18 Algorthms Exercses 0.1. In each of the followng stuatons, ndcate whether f = O(g), or f = Ω(g), or both (n whch case f = Θ(g)). f(n) g(n) (a) n 100 n 200 (b) n 1/2 n 2/3 (c) 100n + log n n + (log n)
More informationM/G/1/GD/ / System. ! Pollaczek-Khinchin (PK) Equation. ! Steady-state probabilities. ! Finding L, W q, W. ! π 0 = 1 ρ
M/G//GD/ / System! Pollcze-Khnchn (PK) Equton L q 2 2 λ σ s 2( + ρ ρ! Stedy-stte probbltes! π 0 ρ! Fndng L, q, ) 2 2 M/M/R/GD/K/K System! Drw the trnston dgrm! Derve the stedy-stte probbltes:! Fnd L,L
More informationThe Regulated and Riemann Integrals
Chpter 1 The Regulted nd Riemnn Integrls 1.1 Introduction We will consider severl different pproches to defining the definite integrl f(x) dx of function f(x). These definitions will ll ssign the sme vlue
More informationSCALED GRADIENT DESCENT LEARNING RATE Reinforcement learning with light-seeking robot
SCALED GRADIET DESCET LEARIG RATE Renforcement lernng wth lght-seekng robot Kry Främlng Helsnk Unversty of Technology, P.O. Box 54, FI-5 HUT, Fnlnd. Eml: Kry.Frmlng@hut.f Keywords: Abstrct: Lner functon
More informationFeature Selection: Part 1
CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?
More informationReview of Calculus, cont d
Jim Lmbers MAT 460 Fll Semester 2009-10 Lecture 3 Notes These notes correspond to Section 1.1 in the text. Review of Clculus, cont d Riemnn Sums nd the Definite Integrl There re mny cses in which some
More informationTheoretical foundations of Gaussian quadrature
Theoreticl foundtions of Gussin qudrture 1 Inner product vector spce Definition 1. A vector spce (or liner spce) is set V = {u, v, w,...} in which the following two opertions re defined: (A) Addition of
More informationDynamic Power Management in a Mobile Multimedia System with Guaranteed Quality-of-Service
Dynmc Power Mngement n Moble Multmed System wth Gurnteed Qulty-of-Servce Qnru Qu, Qng Wu, nd Mssoud Pedrm Dept. of Electrcl Engneerng-Systems Unversty of Southern Clforn Los Angeles CA 90089 Outlne! Introducton
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationVectors and Tensors. R. Shankar Subramanian. R. Aris, Vectors, Tensors, and the Equations of Fluid Mechanics, Prentice Hall (1962).
005 Vectors nd Tensors R. Shnkr Subrmnn Good Sources R. rs, Vectors, Tensors, nd the Equtons of Flud Mechncs, Prentce Hll (96). nd ppendces n () R. B. Brd, W. E. Stewrt, nd E. N. Lghtfoot, Trnsport Phenomen,
More information7.2 The Definite Integral
7.2 The Definite Integrl the definite integrl In the previous section, it ws found tht if function f is continuous nd nonnegtive, then the re under the grph of f on [, b] is given by F (b) F (), where
More informationUsing Predictions in Online Optimization: Looking Forward with an Eye on the Past
Usng Predctons n Onlne Optmzton: Lookng Forwrd wth n Eye on the Pst Nngjun Chen Jont work wth Joshu Comden, Zhenhu Lu, Anshul Gndh, nd Adm Wermn 1 Predctons re crucl for decson mkng 2 Predctons re crucl
More informationMixed Type Duality for Multiobjective Variational Problems
Ž. ournl of Mthemtcl Anlyss nd Applctons 252, 571 586 2000 do:10.1006 m.2000.7000, vlle onlne t http: www.delrry.com on Mxed Type Dulty for Multoectve Vrtonl Prolems R. N. Mukheree nd Ch. Purnchndr Ro
More informationThe Number of Rows which Equal Certain Row
Interntonl Journl of Algebr, Vol 5, 011, no 30, 1481-1488 he Number of Rows whch Equl Certn Row Ahmd Hbl Deprtment of mthemtcs Fcult of Scences Dmscus unverst Dmscus, Sr hblhmd1@gmlcom Abstrct Let be X
More informationLECTURE NOTE #12 PROF. ALAN YUILLE
LECTURE NOTE #12 PROF. ALAN YUILLE 1. Clustering, K-mens, nd EM Tsk: set of unlbeled dt D = {x 1,..., x n } Decompose into clsses w 1,..., w M where M is unknown. Lern clss models p(x w)) Discovery of
More informationCSC 411 / CSC D11 / CSC C11
18 Boostng s a general strategy for learnng classfers by combnng smpler ones. The dea of boostng s to take a weak classfer that s, any classfer that wll do at least slghtly better than chance and use t
More informationTrade-offs in Optimization of GMDH-Type Neural Networks for Modelling of A Complex Process
Proceedngs of the 6th WSEAS Int. Conf. on Systems Theory & Scentfc Computton, Elound, Greece, August -3, 006 (pp48-5) Trde-offs n Optmzton of GDH-Type Neurl Networs for odellng of A Complex Process N.
More information8. INVERSE Z-TRANSFORM
8. INVERSE Z-TRANSFORM The proce by whch Z-trnform of tme ere, nmely X(), returned to the tme domn clled the nvere Z-trnform. The nvere Z-trnform defned by: Computer tudy Z X M-fle trn.m ued to fnd nvere
More informationGAUSS ELIMINATION. Consider the following system of algebraic linear equations
Numercl Anlyss for Engneers Germn Jordnn Unversty GAUSS ELIMINATION Consder the followng system of lgebrc lner equtons To solve the bove system usng clsscl methods, equton () s subtrcted from equton ()
More informationn f(x i ) x. i=1 In section 4.2, we defined the definite integral of f from x = a to x = b as n f(x i ) x; f(x) dx = lim i=1
The Fundmentl Theorem of Clculus As we continue to study the re problem, let s think bck to wht we know bout computing res of regions enclosed by curves. If we wnt to find the re of the region below the
More informationThe exam is closed book, closed notes except your one-page cheat sheet.
CS 89 Fall 206 Introducton to Machne Learnng Fnal Do not open the exam before you are nstructed to do so The exam s closed book, closed notes except your one-page cheat sheet Usage of electronc devces
More informationModule 6: LINEAR TRANSFORMATIONS
Module 6: LINEAR TRANSFORMATIONS. Trnsformtions nd mtrices Trnsformtions re generliztions of functions. A vector x in some set S n is mpped into m nother vector y T( x). A trnsformtion is liner if, for
More informationContinuous Random Variables
STAT/MATH 395 A - PROBABILITY II UW Winter Qurter 217 Néhémy Lim Continuous Rndom Vribles Nottion. The indictor function of set S is rel-vlued function defined by : { 1 if x S 1 S (x) if x S Suppose tht
More informationDifferentiating Gaussian Processes
Dfferentatng Gaussan Processes Andrew McHutchon Aprl 17, 013 1 Frst Order Dervatve of the Posteror Mean The posteror mean of a GP s gven by, f = x, X KX, X 1 y x, X α 1 Only the x, X term depends on the
More information