x i 2. FEEDFORWARD NETWORKS

Size: px
Start display at page:

Download "x i 2. FEEDFORWARD NETWORKS"

Transcription

1 . FEEDFORWARD NEWORKS In retrosectve, t as not untl astonshngly late that methods for tranng ercetrons n several layers, multlayer ercetrons, became knon. Such methods had been roosed by Bryson and Ho 1969, Werbos 1974, Parker 1985, and Le Cun 1985, but the real breakthrough as not eerenced untl the ublcaton of Rumelhart s and McClelland s books 1986 and a aer n Nature 1986, here the back-roagaton or generalzed delta-rule method as roosed. A reason for ths as the fact that gradent-based methods ere not alcable for tranng as long as a Heavsde actvaton functon of the nodes as used. Hofeld - nsred by fndngs n bologcal systems hoever n 198 resented A netork confguraton today called Hofeld netork here the outut sgnalers from the nodes sho a graded resonse on nuts. he sgmod e a.1 that Hofeld ntroduced as to become the most common actvaton functon n neural netorks. Eercse: Dra the sgmod and study ts asymtotc behavor. Mult-layer feedforard netork.1 Central actvaton functons Lnear: ADALINE has already llustrated the behavor of a lnear element th a. Heavsde: he acton of ths actvaton functon can be understood by rertng the functon as N N a a~ 0 +. hus, the bas 0 can adust the locaton of the ste snce 0 acts as a threshold

2 Sgmod: Lke n the Heavsde functon, the bas 0 determnes the locaton of the transton ont of the sgmod see eq..1. If a gan, β, s ncluded n the equaton e β a. t s also ossble to adust the steeness of the curve; β yelds a sgmod aromatng the Heavsde functon, hle small β yeld a ractcally lnear nut-outut relaton. Note! hese adustments of the curve can be carred out by the eghts, snce β can be absorbed by. If the node outut s to vary beteen mn and ma one may use Another nterestng case s obtaned for mn 1 and ma 1. 1 e + a 1+ e 1+ e a tanh a 1 a.4 An advantage of the symmetrcal actvaton functon s that the outut can take on zero th requrng the node to saturate. Eercse: Dra a sgmod netork aromatng f 5 f < 4 f 4 ma + 1+ e a mn mn.3 Eq..3 s a secal case of ths more general formulaton th mn 0 and ma 1. Ram: A lnear ram can be mlemented by restrctng the outut sgnal, e.g., [0,1], by 1, ma0, a mn.5 0 1

3 . Some tranng algorthms In the follong some tranng algorthms for feedforard neural netorks are descrbed. A more detaled summary s gven n, e.g., Hertz et al and Battt Back-roagaton BP hs tranng method s llustrated n the form of the equatons for a netork th one layer of hdden neurons. It s qute easy to generalze the results to M > 1 hdden layers. Consder mnmzaton of the loss functon n 1.10 e t E 1 1 here the teraton nde for the sake of clarty has been omtted or relaced by. All eghts and varables deendent on them, such as net nuts, nternal and eternal oututs, etc., ll n hat follos be 3 assumed to be calculated at ths moment. Introduce the notaton for outut nodes, for hdden nodes and k for nut sgnals.e., nut nodes. Snce and a a σ.6 and and k k k a a σ.7 the netork s oututs are gven by k k k σ σ σ.8 he artal dervatve of the loss functon th resect to s gven by the chan rule as ' th ' 1 a t a t a a e E σ δ δ σ.9

4 If the eght udate s gven by E η η δ.10 Pros and cons: + Comutatonally effcent to calculate gradent and eght udate + Well suted for arallel mlementaton: Uses only local nformaton Slo convergence e obtan the Wdro-Hoff delta-rule ecet the dervatve term ' In an ADALINE σ a a, so ' 1 σ. a σ. a Eercse: Wrte the dervatve of the sgmod, σ a, as a functon of the node s outut only. Eercse: Derve a corresondng term for k. 1 Generally, on can rte for the eghts lm beteen the layers, hose nodes have been labeled m and l, σ σ σ 0.5 lm η δ l m he correcton of the eghts s thus receved by roagatng the error, δ, backard from the outut layer to the nut layer; ths gves the method ts name error back-roagaton a he sgmod and ts dervatve 4 5

5 A large number of modfcatons of the BP have been roosed to enhance the convergence. An early nventon as to nclude an etra momentum term n the eght udate η δ + α old.14 here the second term on the RHS ncludes the revous eght change and, tycally, α Conugate gradents CG Otmzaton methods based on conugate gradents are ell suted for tranng the eghts of neural netorks, snce they buld an estmate of the Hessan matr of the second dervatve terms thout large storage or comutatonal requrements. he CG method ales the udate η k k k.16 One can sho that ths modfcaton of udatng resembles the one that s gven by the method of conugate gradent. he effcency of the search s, hoever, often strongly deendent of the choce of η and α. here the search drecton s gven by g + α k + 1 k + 1 k k.17 Observe that the BP equatons can also be aled ncrementally to udate the eght after every resented attern,.e., the summaton terms over n eqs..10,.1,.13 and.15 can be droed. An advantage of ths s that the search drectons get a stochastc contrbuton by hch local otma may be avoded snce a temoral ncrease n E s alloed. startng from 0 g 0,.e., from the negatve gradent at the startng ont 0. he scalar α s gven by Fletcher and Reeves 1964 α k g k 1 g + k g g k k

6 or Polak and Rbère 1969 α k g k + 1 g k g k + 1 g g k k.19 Pros and Cons: + Converges more radly; may sho quadratc convergence + No ad-hoc arameter values to be set Requres more comutaton effort er teraton than BP. Hgher storage requrements than BP...3 he Levenberg-Marquardt metod LM A method for arameter estmaton n nonlnear systems e.g., n nonlnear regresson as resented by Levenberg 1944 and Marquardt hs trust-regon method aled an nterolaton beteen the drectons of the steeest descent and Neton s method n the search, hch makes t a robust rocedure. Aled on neural netork tranng, here the arameters are the eghts,, the changes n the eghts,, are solved from k k k k k k J J I J e +µ.0 here J s the Jacoban.e., the matr th artal dervatves of e th resect to the eghts, e s the resdual vector th elements e cf. Eq. 1.10, and µ s a arameter that s adusted durng the search. A eakness of all method roosed thus far s that they are only suted to fnd a local otmum and not the globally best soluton. hs roblem can be somehat resolved by startng from several dfferent randomzed ntal eghts, but to tackle the roblem rgorously, stochastc methods see belo are to refer...4 Smulated Annealng SA A method for global otmzaton, that at least artly can avod gettng traed n local mnma, as develoed n the 1950 s by Metrools et al based on an analogy beteen mathematcal mnmzaton of a loss or energy functon and the mnmzaton of the energy of a hyscal system that takes lace hen a materal s annealed. In the latter system, the atoms movement s consderable at hgh temerature, makng ractcally any states occur, hle the system at loer temeratures strves to organze nto a mnmum-energy state. 8 9

7 Introduce an artfcal temerature,, that follos an a ror fed rogram here the value decreases th tme, and a constant c. E 1. Intalze the system: Set k 0 and generate a random startng ont, 0.. Calculate the value of the obectve functon E k E k. 3. Adust the eghts randomly e.g., only one eght,, n turn to k Calculate E k+1 and E k E k+1 E k 5. Accet the ne ont and set k k + 1 th the robablty Search for a global mnmum of a functon th local mnma he dstrbuton of the states s gven by P E E k B e.1 rob e 1 E c k f E k f E < 0 k 0 6. est for convergence; f not, go to.. here PE s the robablty of fndng the system n an energy state E, k B s Boltzmann s constant and s the thermodynamc temerature. Based on ths a search method can be formulated Krkatrck et al. 1983: Some central roblems that have to be tackled n the mlementaton of the algorthm are he temerature rogram: he artfcal temerature should be reduced sloly enough to avod t from beng traed n local mnma

8 It has been shon that should be nversely roortonal to the dervatve of the tme, e.g., 0 k log1 + k.3 but ths requres a large number of teratons. Instead, more ractcal choces are made, e.g.,.3 Imortant ractcal asects Choce of nuts and oututs are the most mortant matters n statstcal model buldng! Often, 70% of the modelng tme may be sent on selecton and rerocessng of nuts and oututs. k k 1 γ ; γ 0.8, Startng ont: It s often motvated to start from several dfferent onts, 0. Search drecton and ste length: he former may be assumed to be of Gaussan dstrbuton k k k P e.5 Another choce s to make the ste length roortonal to the temerature Wasserman he search drecton, n turn, can be generated by randomly selectng a coordnate drecton,, modfyng. Alternatvely, all s may be vsted n turn to guarantee that all eghts be changed. A roer scalng of the nuts and oututs s essental for facltatng a good numercal soluton of the roblem. A scalng rocedure often aled s ~ s or ~ mn ma mn.6a,b Choce of netork structure. Because of a large number of eghts tycally used n neural netork models, Runge effects often occur: As the model comlety e.g., number of hdden nodes gros, the ft on the tranng data becomes better and better, but the model s erformance on novel data not seen ll get orse after some ont, snce the model ll start to ft nose and outlers n the tranng set. 3 33

9 tranng test redcton error and AIC Informaton Crteron A, Akake 1970, hese estmate the model s redcton error on novel data and are, therefore, esecally useful n cases here the number of observatons data ros s lmted. E A ay to tackle the roblem s to use Model comlety Number of teratons Overfttng and overtranng In summary, there s a need for crtera for selecton of netork comlety. Even though rocedures for such statstcal testng have been roosed at an early stage n the lterature Baum and Haussler 1989, Moody 199, Kendall and Hall 199, t, hoever, seems that no comletely general crtera est. It should also be noted that there s not even consensus on such tests for selecton among lnear models, even though certan test crtera are commonly aled, such as the FPE fnal Constructve algorthms that start from a small netork and, f needed, add hdden nodes e.g., Fahlman and Lebere 1990, Frean 1991 Destructve algorthms that start from a large traned netork and rune unnecessary connectons e.g., Mozer and Smolensky 1989, Chauvn 1989, Le Cun et al. 1990, Karnn 1990, Burktt 1991, Fnnoff et al Smultaneous otmzaton of eghts and structure. Even though mathematcal rogrammng can be used to tackle the roblem n theory, the nonlnearty of and the comaratvely large number of nteger varables make the task etremely challengng. Recently, genetc and evolutonary rogrammng has been aled by several nvestgators to tackle the roblem Fogel et al. 1990, Manezzo 1994, Gao et al. 1999, Yen and Lu

10 he crtera aled for decdng hether to ncrease/decrease the netork comlety are often rather ad hoc, even though most have a background n statstcs. Yet another alternatve, hch today has become a routne task n neural netork modelng, s to study the generalzaton erformance durng tranng by follong the model s behavor on a set of novel ndeendent data onts Wegend et al. 1990, Fnnoff et al. 1993, and sto tranng hen the test set errors are at ther mnmum. A tycal evoluton of the errors as schematcally dected n the revous fgure; the ncrease eerenced n the test error s the result of overtranng. It s ossble to sho Lung and Söberg 199 that such tranng nterruton s equvalent to usng regularzaton terms n the loss functon see belo. hs may serve to elan hy clearly overszed netorks commonly reorted n the lterature are stll able to roduce models that generalze ell: As the netorks are traned by neffcent methods such as backroagaton overfttng s avoded, snce a consderable art of the eressonal caacty remans unused. hs s analogous to use a smaller ell-traned netork Wegend et al Choce of loss functon: A desred feature of the loss functon, E, that s mnmzed durng tranng s that t shos a theoretcal mnmum of zero hen the model has learned to roduce all oututs erfectly ell. he squared Eucldean norm square sum of errors, Eq. 1.10, that s often used satsfes ths requrement, but there s usually no theoretcal ustfcaton for ths choce. Sometmes the netork s task s to classfy or categorze data, and n such cases the th outut, should gve the robablty that the nut vector,, corresonds to attrbute. hen, t may be motvated to use a relatve entroy as a measure of the ft Solla et al gven by th e E t t ln e + 1 t 1 t ln here the re-logarthmc terms stems from condtonal robabltes. hs loss functon has some nce roertes: It dverges f the nodal outut s saturated n the rong end of the sgmod and qute small errors for t 0 affect the loss functon consderably. Moreover, art of the loss functon can be shon to be conve, so t s lkely that feer local mnma est

11 Constranng the magntude of eghts: If large eghts should be avoded, a enalty term, E ~, can be added to the loss functon. he roblem can, e.g., be rtten as mn E ~ + E t t + γ.9 or n vector form { e e + γ } mn.30 hs gves an etra term n the dervatve that s roortonal to the value of the eght n queston, so the backroagaton udate becomes k + 1 k η δ 1 η γ k η η γ δ k.31 he term E ~ can, hoever, be consdered to enalze unreasonably much for large but ossbly sgnfcant eghts, so one may nstead use the form ~ E γ 1+.3 hs forces small eghts toards zero because of quadratc enalty, hle larger eghts may stll be alloed. hs formulaton has been clamed to rune unnecessary connectons n the netork, but studes on dfferent decay equatons have not revealed any clear ractcal advantage of.3 over the formulaton n.9. Another nterestng observaton s that the alcaton of a regularzaton such as.9 n the begnnng of the tranng by hgher order methods, here the eght estmates are stll oor, consderably decreases the robablty of gettng stuck n clearly local mnma. he reason s that node aralyss can be avoded,.e., stuatons here nodes saturate n an early stage of the search. he rocedure s often called eght decay cf. Eq

12 .4 Some areas of alcaton.4.1 Functon aromaton An obvous alcaton of feedforard neural netorks s functon aromaton and nonlnear regresson: Netorks th a hdden layer of sgmods and a lnear outut layer have been shon to be able to aromate arbtrary contnuous bounded functons to any accuracy, gven enough caacty,.e., a suffcent number of hdden nodes Cybenko 1988,1989, Funahash In ractce, hoever, the sze of the netork must be strctly controlled to consder features of the data sets studed, e.g., to avod overfttng. he ay n hch the netork acts s qute easy to understand, even though the fne onts of ho t does the ob are more comlcated: In regons here the functon to be aromated shos large changes, the netork allocates nodes, and the steeness of the functon ll be reflected n the magntude of the eghts. An eamle of a ft s gven belo Haykn Eamle of a functon aromated by a neural netork Eercse: Adat a neural netork to y 0.5sn8 + 4 n the nterval 0,1. Ho many hdden nodes are requred to make the ft accurate?.4. Prncal comonent analyss Prncal comonent analyss s a mathematcal method for analyzng multcomonent data that strves to reduce the dmenson of the varable sace th mnmum loss of sgnfcant nformaton. It can be aled to data comresson, data mnng, fault detecton elmnaton of redundant nformaton, rocess montorng descrbes the rocess varables actual dmenson, vsualzaton, etc. Let the measurement nformaton be gven by observatons of an m- dmensonal vector. PCA sets u f <m orthogonal vectors n the m

13 dmensonal sace that descrbe as much as ossble of the varaton n the data see the belo fgure. he covarance matr s gven by C or smly C for he frst rncal comonent s chosen along the drecton of mamum varance, the net one orthogonal to the frst and agan th mamum varance, and so on. hese drectons can be shon to corresond to the drectons of egenvectors, th largest egenvalues, of the covarance matr of the data. λ λ λ K 1 3 λ f.33 normalzed data. In hat follos a bref descrton of the lnear PCA s gven and ts ossble etenson to nonlnear PCA by neural netorks Kramer 1991: If the n observaton vectors are collected ro-se n the n m matr X, the method can be condensed nto E subect to X A B + E mn.34,35 here dma n f, dmb m f and dme n m. hs factorzaton of X yelds columns of B, denoted by b, that are egenvectors corresondng to the largest f egenvalues of C. Assume that B B I, hch hold s B s columns are orthogonal and normalzed,.e., b b δ, here δ s the Kronecker delta. A data vector s net rocessed from the m-dmensonal observaton sace to the f-dmensonal rncal comonent sace feature sace by z B.36 From z the orgnal vector can be reconstructed by Smle roectons z B.37 th the reconstructon error e. 4 43

14 In the nonlnear case Eqs can be rtten as z G and H z.38,.39 here G and H are nonlnear vector functons that each can be descrbed by a neural netork th a layer of hdden nodes. hs gves the archtecture Some noteorthy facts: he nodes n the nut and outut layers, as ell as the mddle layer, are lnear, and those of the remanng to layers are nonlnear e.g., sgmods. hs means that, theoretcally, the eghts of the connectons to and from the mddle layer could be merged, but then e loose the bottleneck,.e., no reducton n rank occurs nformaton about the values of the varables n the reduced dmenson.4.3 Dynamc modelng he easest ay to use neural netorks for modelng dynamc rocesses s to treat the tme dmenson as a satal dmenson. hs leads to an archtecture that s often called tme-delay netorks, th hch temoral sequences can be detected by smultaneously feedng as nuts t, t-1, t-,,t-l. Neural netork for NLPCA me-delay netork 44 45

15 A classcal eamle of an early alcaton of the above s gven by Senosk s and Rosenberg s 1987 NE-talk, here a ndo assng over a tet as fed as nut to a neural netork, that as traned to roduce a honeme code to a seech syntheszer for the centermost letter of the ndo see fgure belo. In such alcatons, neural netorks act as tme seres models, the theory of hch s brefly revsted belo. An autoregressve dscrete tme seres model of k:th order for one varable,, can be rtten Snce the letters ere read from a tet, every letter used 9 nut sgnals; the dmenson of the nut vector as therefore 7 9. hree letters before and after the centermost one ere used because the ronuncaton of a letter s not ndeendent of the contet. he netork as traned on 104 ords and roduced correct results n about 95 % of the honemes. For novel ords, the rate as slghtly belo 80 %. In ste of the fact that t could not comete th a sohstcated conventonal model n terms of accuracy, the eercse shoed that a qute robust neural-netork based model could be develoed usng a fracton of the tme requred for develong the conventonal model. t φ + φ t 1 + φ t + K + φk t k + ε t here ε s usually taken to be a normally dstrbuted uncorrelated random varable hte nose. he equaton can also be rtten comactly f, further, the varable s normalzed to zero mean as φ t ε t here φ 1 φ 1 z 1 φ z φ k z k.41 and z 1 s the backard shft oerator. he etenson of the equatons to vector autoregressve models VAR s obvous. he rncle of NE-talk he advantage of usng feedforard neural netorks for tme-seres modelng les n ther ablty to cature nonlneartes, thout a ror knoledge of the tye of nonlnearty encountered. If φ n eq..41 s relaced by a nonlnear oerator, mlemented by a neural netork, e obtan a NAR nonlnear autoregressve model. ycal roblems tackled 46 47

16 by the technque are found n seech recognton, stock market redctons, rocess dentfcaton, analyss of chaotc tme seres, etc. An eamle from the latter category s gven belo. Study the logstc ma Fegenbaum t + 1 α t t.4 he system shos a fed ont attractor at * α 1/α f α < 0, hle t for α 3.0, shos oscllatons of grong erodcty hch fnally for α > 3.57 change nto chaotc behavor. If α eceeds 4.0, the values of the sequence start to fall outsde 0,1. Recurson of the logstc ma Snce the ne value of s obtaned through recurson, t+1 ft, an l-ste ahead redcton can be rtten l t + l 1 f f t + l... f t t + l f.43 he net fgure llustrates grahcally the recurson, and the fgure belo t a tycal evoluton of the varable for α 3.99 n Eq observatons of t t 1 t 48 49

17 It s easly found that a small netork aromates the tme seres very ell Ydste 1990, Saén he belo fgure shos the quadratc functon and ts aromaton gven by a 1,,1 feedforard netork traned on the frst 500 observatons of the seres. he average redcton error for the follong 500 observatons s of the same magntude, hch shos that the netork has learned the nonlnear ma. Hoever, f the netork s aled for several-ste ahead redcton by feedng ts on outut back as nut, t fals f the redcton horzon s long l > Control On the bass of hat as resented n subsectons.4.1 and.4.3 t s straghtforard to aly neural netorks on control roblems, and the lterature resents numerous such alcatons. In ts smlest form, a netork can act as an nverse model of the rocess system to be controlled. hs model can be dentfed by tranng the netork n arallel th the rocess n queston to roduce the nut,, on the bass of the outut, y see fgure. t e Σ + Process Netork y Identfcaton of the nverse model of the rocess After ths, the netork can be connected before the rocess to calculate he seres and ts aromaton by the neural netork the control sgnal by feedforard control. It should be noted that the aroach requres the nverse of the rocess to est. In the aroach the resultng control qualty s strongly deendent on the qualty of the rocess model, and hether the rocess s tme-ndeendent

18 t Netork B Udated coy of Netork A e Σ Process Netork A Alcaton of the nverse model to rocess control + y he dentfed model can, n turn, form the bass of more sohstcated control strateges, such as otmal e.g., mnmum varance control and model-redctve control MPC, here the model s aled for a redcton horzon over hch a loss functon s mnmzed. It s also ossble to used neural netorks for learnng otmal control traectores. hs aroach can be useful n cases here a more sohstcated, but comutatonally eensve, rocess model s avalable for off-lne studes of the control roblem at hand. Normally, feedback control s aled to tackle the roblems of model naccuracy or tme-deendent rocesses mentoned above. Often, a rocess forard model s used n feedback control. hs can be estmated by tranng a neural netork, ossbly recursvely. Some early aers on alcatons of neural netorks to dentfcaton and control are Narendra and Parthasarathy 1990, Ydste 1990, Zurada 199 and Ydste and Narendra 199. Process Netork Process dentfcaton y t Σ Classfcaton As noted n Chater 1, feedforard neural netorks can be aled to solve classfcaton roblems, here the result s a Boolean varable. Snce all classfcaton roblems th several classes can be rertten nto a set of roblems th to classes, e shall n hat follos only consder a roblem here the nut vector s classfed to belong to ether A + or A. 5 53

19 Neural netorks thus consttute an alternatve technque to tradtonal classfcaton methods, such as nearest neghbor and Bayesan classfers. he classfcaton s based on the caablty of a node to lnearly searate the sace nto to arts; f the nut eghts are large enough, the outut of the node s zero A or one A +. For an n-dmensonal nut vector the node bsects the sace by an n-1- dmensonal hyerlane. Problems that are not lnearly searable can be tackled by combnng the outut from several nodes. A conve regon n an n-dmensonal sace can be descrbed by a logcal conuncton of several lnearly searable sub-regons. For nstance, a trangular regon n a to-dmensonal sace s formed as a logcal conuncton of three lnearly searable regons, as ndcated n the belo fgure. hus, three nodes are requred to form these classfcaton boundares, and a fourth node to combne the oututs of the three nodes. From ths follos that a large number of nodes may be requred to accurately form a nonlnear conve boundary. he number of almost lnear segments that are needed to form the desred shae of the classfcaton boundary gves an ndcaton of the requred number of hdden nodes. Unfortunately, the shae of the boundary s seldom knon a ror. A concave regon can be formed as the unon of several conve regons, hch means that netorks th to hdden layers of nodes ould be requred; the frst layer creates the conve regon logcal conuncton hle the other takes the unon of these logcal dsuncton. An eamle of ths s gven n the belo fgure. Huang and Lmann 1987, Lmann 1988 and Zurada 199 treat classfcaton roblems tackled by neural netorks n more detal. rangle-formed classfcaton regon and classfcaton boundares Concave classfcaton regons 54 55

20 An undesred henomenon that arses n classfcaton of concave regons by feedforard neural netorks s that t s ossble that the netorks create regons outsde the tranng sace th odd classfcaton results. he reason s that the classfcaton boundares ntersect here, formng ne regons here the logcal condtons are satsfed. hs s esecally alarmng n, e.g., fault detecton alcatons of the technque. Kramer and Leonard 1990 resent a ustfed crtcsm of the technque. Eercse: Suggest a netork that solves the classfcaton roblem of dfferng beteen the regons reresented by crcles and astersks th a mnmum number of hdden nodes

Confidence intervals for weighted polynomial calibrations

Confidence intervals for weighted polynomial calibrations Confdence ntervals for weghted olynomal calbratons Sergey Maltsev, Amersand Ltd., Moscow, Russa; ur Kalambet, Amersand Internatonal, Inc., Beachwood, OH e-mal: kalambet@amersand-ntl.com htt://www.chromandsec.com

More information

Pattern Classification

Pattern Classification Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher

More information

Classification (klasifikácia) Feedforward Multi-Layer Perceptron (Dopredná viacvrstvová sieť) 14/11/2016. Perceptron (Frank Rosenblatt, 1957)

Classification (klasifikácia) Feedforward Multi-Layer Perceptron (Dopredná viacvrstvová sieť) 14/11/2016. Perceptron (Frank Rosenblatt, 1957) 4//06 IAI: Lecture 09 Feedforard Mult-Layer Percetron (Doredná vacvrstvová seť) Lubca Benuskova AIMA 3rd ed. Ch. 8.6.4 8.7.5 Classfcaton (klasfkáca) In machne learnng and statstcs, classfcaton s the roblem

More information

Hopfield Training Rules 1 N

Hopfield Training Rules 1 N Hopfeld Tranng Rules To memorse a sngle pattern Suppose e set the eghts thus - = p p here, s the eght beteen nodes & s the number of nodes n the netor p s the value requred for the -th node What ll the

More information

Pattern Recognition. Approximating class densities, Bayesian classifier, Errors in Biometric Systems

Pattern Recognition. Approximating class densities, Bayesian classifier, Errors in Biometric Systems htt://.cubs.buffalo.edu attern Recognton Aromatng class denstes, Bayesan classfer, Errors n Bometrc Systems B. W. Slverman, Densty estmaton for statstcs and data analyss. London: Chaman and Hall, 986.

More information

Multigradient for Neural Networks for Equalizers 1

Multigradient for Neural Networks for Equalizers 1 Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT

More information

Neural Networks & Learning

Neural Networks & Learning Neural Netorks & Learnng. Introducton The basc prelmnares nvolved n the Artfcal Neural Netorks (ANN) are descrbed n secton. An Artfcal Neural Netorks (ANN) s an nformaton-processng paradgm that nspred

More information

Discriminative classifier: Logistic Regression. CS534-Machine Learning

Discriminative classifier: Logistic Regression. CS534-Machine Learning Dscrmnatve classfer: Logstc Regresson CS534-Machne Learnng 2 Logstc Regresson Gven tranng set D stc regresson learns the condtonal dstrbuton We ll assume onl to classes and a parametrc form for here s

More information

Pattern Classification (II) 杜俊

Pattern Classification (II) 杜俊 attern lassfcaton II 杜俊 junu@ustc.eu.cn Revew roalty & Statstcs Bayes theorem Ranom varales: screte vs. contnuous roalty struton: DF an DF Statstcs: mean, varance, moment arameter estmaton: MLE Informaton

More information

A total variation approach

A total variation approach Denosng n dgtal radograhy: A total varaton aroach I. Froso M. Lucchese. A. Borghese htt://as-lab.ds.unm.t / 46 I. Froso, M. Lucchese,. A. Borghese Images are corruted by nose ) When measurement of some

More information

6. Hamilton s Equations

6. Hamilton s Equations 6. Hamlton s Equatons Mchael Fowler A Dynamcal System s Path n Confguraton Sace and n State Sace The story so far: For a mechancal system wth n degrees of freedom, the satal confguraton at some nstant

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Digital PI Controller Equations

Digital PI Controller Equations Ver. 4, 9 th March 7 Dgtal PI Controller Equatons Probably the most common tye of controller n ndustral ower electroncs s the PI (Proortonal - Integral) controller. In feld orented motor control, PI controllers

More information

Naïve Bayes Classifier

Naïve Bayes Classifier 9/8/07 MIST.6060 Busness Intellgence and Data Mnng Naïve Bayes Classfer Termnology Predctors: the attrbutes (varables) whose values are used for redcton and classfcaton. Predctors are also called nut varables,

More information

Hidden Markov Model Cheat Sheet

Hidden Markov Model Cheat Sheet Hdden Markov Model Cheat Sheet (GIT ID: dc2f391536d67ed5847290d5250d4baae103487e) Ths document s a cheat sheet on Hdden Markov Models (HMMs). It resembles lecture notes, excet that t cuts to the chase

More information

Evaluation of classifiers MLPs

Evaluation of classifiers MLPs Lecture Evaluaton of classfers MLPs Mlos Hausrecht mlos@cs.ptt.edu 539 Sennott Square Evaluaton For any data set e use to test the model e can buld a confuson matrx: Counts of examples th: class label

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Exerments-I MODULE III LECTURE - 2 EXPERIMENTAL DESIGN MODELS Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 2 We consder the models

More information

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,

More information

Bayesian Decision Theory

Bayesian Decision Theory No.4 Bayesan Decson Theory Hu Jang Deartment of Electrcal Engneerng and Comuter Scence Lassonde School of Engneerng York Unversty, Toronto, Canada Outlne attern Classfcaton roblems Bayesan Decson Theory

More information

Gradient Compared lp-lms Algorithms for Sparse System Identification

Gradient Compared lp-lms Algorithms for Sparse System Identification Gradent Comared l-lms Algorthms for Sarse System Identfcaton Yong Feng 1,, Jasong Wu, Ru Zeng, Lmn Luo, Huazhong Shu 1. School of Bologcal Scence and Medcal Engneerng, Southeast Unversty, Nanjng 10096,

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

A Tutorial on Data Reduction. Linear Discriminant Analysis (LDA) Shireen Elhabian and Aly A. Farag. University of Louisville, CVIP Lab September 2009

A Tutorial on Data Reduction. Linear Discriminant Analysis (LDA) Shireen Elhabian and Aly A. Farag. University of Louisville, CVIP Lab September 2009 A utoral on Data Reducton Lnear Dscrmnant Analss (LDA) hreen Elhaban and Al A Farag Unverst of Lousvlle, CVIP Lab eptember 009 Outlne LDA objectve Recall PCA No LDA LDA o Classes Counter eample LDA C Classes

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

Fuzzy approach to solve multi-objective capacitated transportation problem

Fuzzy approach to solve multi-objective capacitated transportation problem Internatonal Journal of Bonformatcs Research, ISSN: 0975 087, Volume, Issue, 00, -0-4 Fuzzy aroach to solve mult-objectve caactated transortaton roblem Lohgaonkar M. H. and Bajaj V. H.* * Deartment of

More information

Introduction to the Introduction to Artificial Neural Network

Introduction to the Introduction to Artificial Neural Network Introducton to the Introducton to Artfcal Neural Netork Vuong Le th Hao Tang s sldes Part of the content of the sldes are from the Internet (possbly th modfcatons). The lecturer does not clam any onershp

More information

Practical Newton s Method

Practical Newton s Method Practcal Newton s Method Lecture- n Newton s Method n Pure Newton s method converges radly once t s close to. It may not converge rom the remote startng ont he search drecton to be a descent drecton rue

More information

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU

MIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern

More information

Discriminative classifier: Logistic Regression. CS534-Machine Learning

Discriminative classifier: Logistic Regression. CS534-Machine Learning Dscrmnatve classfer: Logstc Regresson CS534-Machne Learnng robablstc Classfer Gven an nstance, hat does a probablstc classfer do dfferentl compared to, sa, perceptron? It does not drectl predct Instead,

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Neural networks. Nuno Vasconcelos ECE Department, UCSD

Neural networks. Nuno Vasconcelos ECE Department, UCSD Neural networs Nuno Vasconcelos ECE Department, UCSD Classfcaton a classfcaton problem has two types of varables e.g. X - vector of observatons (features) n the world Y - state (class) of the world x X

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

Machine Learning. Classification. Theory of Classification and Nonparametric Classifier. Representing data: Hypothesis (classifier) Eric Xing

Machine Learning. Classification. Theory of Classification and Nonparametric Classifier. Representing data: Hypothesis (classifier) Eric Xing Machne Learnng 0-70/5 70/5-78, 78, Fall 008 Theory of Classfcaton and Nonarametrc Classfer Erc ng Lecture, Setember 0, 008 Readng: Cha.,5 CB and handouts Classfcaton Reresentng data: M K Hyothess classfer

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

b ), which stands for uniform distribution on the interval a x< b. = 0 elsewhere

b ), which stands for uniform distribution on the interval a x< b. = 0 elsewhere Fall Analyss of Epermental Measurements B. Esensten/rev. S. Errede Some mportant probablty dstrbutons: Unform Bnomal Posson Gaussan/ormal The Unform dstrbuton s often called U( a, b ), hch stands for unform

More information

Classification learning II

Classification learning II Lecture 8 Classfcaton learnng II Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Logstc regresson model Defnes a lnear decson boundar Dscrmnant functons: g g g g here g z / e z f, g g - s a logstc functon

More information

15-381: Artificial Intelligence. Regression and cross validation

15-381: Artificial Intelligence. Regression and cross validation 15-381: Artfcal Intellgence Regresson and cross valdaton Where e are Inputs Densty Estmator Probablty Inputs Classfer Predct category Inputs Regressor Predct real no. Today Lnear regresson Gven an nput

More information

CS407 Neural Computation

CS407 Neural Computation CS407 Neural Computaton Lecture 8: Neural Netorks for Constraned Optmzaton. Lecturer: A/Prof. M. Bennamoun Neural Nets for Constraned Optmzaton. Introducton Boltzmann machne Introducton Archtecture and

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Advanced Topics in Optimization. Piecewise Linear Approximation of a Nonlinear Function

Advanced Topics in Optimization. Piecewise Linear Approximation of a Nonlinear Function Advanced Tocs n Otmzaton Pecewse Lnear Aroxmaton of a Nonlnear Functon Otmzaton Methods: M8L Introducton and Objectves Introducton There exsts no general algorthm for nonlnear rogrammng due to ts rregular

More information

Combinational Circuit Design

Combinational Circuit Design Combnatonal Crcut Desgn Part I: Desgn Procedure and Examles Part II : Arthmetc Crcuts Part III : Multlexer, Decoder, Encoder, Hammng Code Combnatonal Crcuts n nuts Combnatonal Crcuts m oututs A combnatonal

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Lecture 23: Artificial neural networks

Lecture 23: Artificial neural networks Lecture 23: Artfcal neural networks Broad feld that has developed over the past 20 to 30 years Confluence of statstcal mechancs, appled math, bology and computers Orgnal motvaton: mathematcal modelng of

More information

1 GSW Iterative Techniques for y = Ax

1 GSW Iterative Techniques for y = Ax 1 for y = A I m gong to cheat here. here are a lot of teratve technques that can be used to solve the general case of a set of smultaneous equatons (wrtten n the matr form as y = A), but ths chapter sn

More information

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018 INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton

More information

Pattern Classification

Pattern Classification attern Classfcaton All materals n these sldes were taken from attern Classfcaton nd ed by R. O. Duda,. E. Hart and D. G. Stork, John Wley & Sons, 000 wth the ermsson of the authors and the ublsher Chater

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

( ) 2 ( ) ( ) Problem Set 4 Suggested Solutions. Problem 1

( ) 2 ( ) ( ) Problem Set 4 Suggested Solutions. Problem 1 Problem Set 4 Suggested Solutons Problem (A) The market demand functon s the soluton to the followng utlty-maxmzaton roblem (UMP): The Lagrangean: ( x, x, x ) = + max U x, x, x x x x st.. x + x + x y x,

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

The Bellman Equation

The Bellman Equation The Bellman Eqaton Reza Shadmehr In ths docment I wll rovde an elanaton of the Bellman eqaton, whch s a method for otmzng a cost fncton and arrvng at a control olcy.. Eamle of a game Sose that or states

More information

Support Vector Machines CS434

Support Vector Machines CS434 Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We

More information

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD CHALMERS, GÖTEBORGS UNIVERSITET SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 35, FIM 72 GU, PhD Tme: Place: Teachers: Allowed materal: Not allowed: January 2, 28, at 8 3 2 3 SB

More information

Multi layer feed-forward NN FFNN. XOR problem. XOR problem. Neural Network for Speech. NETtalk (Sejnowski & Rosenberg, 1987) NETtalk (contd.

Multi layer feed-forward NN FFNN. XOR problem. XOR problem. Neural Network for Speech. NETtalk (Sejnowski & Rosenberg, 1987) NETtalk (contd. NN 3-00 Mult layer feed-forard NN FFNN We consder a more general netor archtecture: beteen the nput and output layers there are hdden layers, as llustrated belo. Hdden nodes do not drectly send outputs

More information

Independent Component Analysis

Independent Component Analysis Indeendent Comonent Analyss Mture Data Data that are mngled from multle sources May not now how many sources May not now the mng mechansm Good Reresentaton Uncorrelated, nformaton-bearng comonents PCA

More information

Mixture of Gaussians Expectation Maximization (EM) Part 2

Mixture of Gaussians Expectation Maximization (EM) Part 2 Mture of Gaussans Eectaton Mamaton EM Part 2 Most of the sldes are due to Chrstoher Bsho BCS Summer School Eeter 2003. The rest of the sldes are based on lecture notes by A. Ng Lmtatons of K-means Hard

More information

4DVAR, according to the name, is a four-dimensional variational method.

4DVAR, according to the name, is a four-dimensional variational method. 4D-Varatonal Data Assmlaton (4D-Var) 4DVAR, accordng to the name, s a four-dmensonal varatonal method. 4D-Var s actually a drect generalzaton of 3D-Var to handle observatons that are dstrbuted n tme. The

More information

Time-Varying Systems and Computations Lecture 6

Time-Varying Systems and Computations Lecture 6 Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy

More information

THERMODYNAMICS. Temperature

THERMODYNAMICS. Temperature HERMODYNMICS hermodynamcs s the henomenologcal scence whch descrbes the behavor of macroscoc objects n terms of a small number of macroscoc arameters. s an examle, to descrbe a gas n terms of volume ressure

More information

Multi-layer neural networks

Multi-layer neural networks Lecture 0 Mult-layer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Lnear regresson w Lnear unts f () Logstc regresson T T = w = p( y =, w) = g( w ) w z f () = p ( y = ) w d w d Gradent

More information

Logistic regression with one predictor. STK4900/ Lecture 7. Program

Logistic regression with one predictor. STK4900/ Lecture 7. Program Logstc regresson wth one redctor STK49/99 - Lecture 7 Program. Logstc regresson wth one redctor 2. Maxmum lkelhood estmaton 3. Logstc regresson wth several redctors 4. Devance and lkelhood rato tests 5.

More information

Ensemble Methods: Boosting

Ensemble Methods: Boosting Ensemble Methods: Boostng Ncholas Ruozz Unversty of Texas at Dallas Based on the sldes of Vbhav Gogate and Rob Schapre Last Tme Varance reducton va baggng Generate new tranng data sets by samplng wth replacement

More information

Generative classification models

Generative classification models CS 675 Intro to Machne Learnng Lecture Generatve classfcaton models Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square Data: D { d, d,.., dn} d, Classfcaton represents a dscrete class value Goal: learn

More information

MATH 567: Mathematical Techniques in Data Science Lab 8

MATH 567: Mathematical Techniques in Data Science Lab 8 1/14 MATH 567: Mathematcal Technques n Data Scence Lab 8 Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 11, 2017 Recall We have: a (2) 1 = f(w (1) 11 x 1 + W (1) 12 x 2 + W

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

Constructing Control Process for Wafer Defects Using Data Mining Technique

Constructing Control Process for Wafer Defects Using Data Mining Technique The Fourth nternatonal Conference on Electronc Busness (CEB004) / Bejng 5 Constructng Control ocess for Wafer Defects Usng Data Mnng Technque Leeng Tong Hsngyn Lee Chfeng Huang Changke Ln Chenhu Yang Department

More information

Primer on High-Order Moment Estimators

Primer on High-Order Moment Estimators Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

An Efficient Least-Squares Trilateration Algorithm for Mobile Robot Localization

An Efficient Least-Squares Trilateration Algorithm for Mobile Robot Localization he IEEE/RSJ Internatonal Conference on Intellgent Robots and Systems October -5, St. Lous, USA An Effcent Least-Squares rlateraton Algorthm for Moble Robot Localzaton Yu Zhou, Member, IEEE Abstract A novel

More information

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen

Hopfield networks and Boltzmann machines. Geoffrey Hinton et al. Presented by Tambet Matiisen Hopfeld networks and Boltzmann machnes Geoffrey Hnton et al. Presented by Tambet Matsen 18.11.2014 Hopfeld network Bnary unts Symmetrcal connectons http://www.nnwj.de/hopfeld-net.html Energy functon The

More information

Supervised Learning NNs

Supervised Learning NNs EE788 Robot Cognton and Plannng, Prof. J.-H. Km Lecture 6 Supervsed Learnng NNs Robot Intellgence Technolog Lab. From Jang, Sun, Mzutan, Ch.9, Neuro-Fuzz and Soft Computng, Prentce Hall Contents. Introducton.

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Algorithms for factoring

Algorithms for factoring CSA E0 235: Crytograhy Arl 9,2015 Instructor: Arta Patra Algorthms for factorng Submtted by: Jay Oza, Nranjan Sngh Introducton Factorsaton of large ntegers has been a wdely studed toc manly because of

More information

Multilayer neural networks

Multilayer neural networks Lecture Multlayer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Mdterm exam Mdterm Monday, March 2, 205 In-class (75 mnutes) closed book materal covered by February 25, 205 Multlayer

More information

A Quadratic Cumulative Production Model for the Material Balance of Abnormally-Pressured Gas Reservoirs F.E. Gonzalez M.S.

A Quadratic Cumulative Production Model for the Material Balance of Abnormally-Pressured Gas Reservoirs F.E. Gonzalez M.S. Natural as Engneerng A Quadratc Cumulatve Producton Model for the Materal Balance of Abnormally-Pressured as Reservors F.E. onale M.S. Thess (2003) T.A. Blasngame, Texas A&M U. Deartment of Petroleum Engneerng

More information

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands

1. Inference on Regression Parameters a. Finding Mean, s.d and covariance amongst estimates. 2. Confidence Intervals and Working Hotelling Bands Content. Inference on Regresson Parameters a. Fndng Mean, s.d and covarance amongst estmates.. Confdence Intervals and Workng Hotellng Bands 3. Cochran s Theorem 4. General Lnear Testng 5. Measures of

More information

IV. Performance Optimization

IV. Performance Optimization IV. Performance Optmzaton A. Steepest descent algorthm defnton how to set up bounds on learnng rate mnmzaton n a lne (varyng learnng rate) momentum learnng examples B. Newton s method defnton Gauss-Newton

More information

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks

Internet Engineering. Jacek Mazurkiewicz, PhD Softcomputing. Part 3: Recurrent Artificial Neural Networks Self-Organising Artificial Neural Networks Internet Engneerng Jacek Mazurkewcz, PhD Softcomputng Part 3: Recurrent Artfcal Neural Networks Self-Organsng Artfcal Neural Networks Recurrent Artfcal Neural Networks Feedback sgnals between neurons Dynamc

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 6. Rdge regresson The OLSE s the best lnear unbased

More information

A Quadratic Cumulative Production Model for the Material Balance of Abnormally-Pressured Gas Reservoirs F.E. Gonzalez M.S.

A Quadratic Cumulative Production Model for the Material Balance of Abnormally-Pressured Gas Reservoirs F.E. Gonzalez M.S. Formaton Evaluaton and the Analyss of Reservor Performance A Quadratc Cumulatve Producton Model for the Materal Balance of Abnormally-Pressured as Reservors F.E. onale M.S. Thess (2003) T.A. Blasngame,

More information

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015 CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research

More information

Why feed-forward networks are in a bad shape

Why feed-forward networks are in a bad shape Why feed-forward networks are n a bad shape Patrck van der Smagt, Gerd Hrznger Insttute of Robotcs and System Dynamcs German Aerospace Center (DLR Oberpfaffenhofen) 82230 Wesslng, GERMANY emal smagt@dlr.de

More information

A General Class of Selection Procedures and Modified Murthy Estimator

A General Class of Selection Procedures and Modified Murthy Estimator ISS 684-8403 Journal of Statstcs Volume 4, 007,. 3-9 A General Class of Selecton Procedures and Modfed Murthy Estmator Abdul Bast and Muhammad Qasar Shahbaz Abstract A new selecton rocedure for unequal

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Polynomial Regression Models

Polynomial Regression Models LINEAR REGRESSION ANALYSIS MODULE XII Lecture - 6 Polynomal Regresson Models Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Test of sgnfcance To test the sgnfcance

More information

A Mathematical Theory of Communication. Claude Shannon s paper presented by Kate Jenkins 2/19/00

A Mathematical Theory of Communication. Claude Shannon s paper presented by Kate Jenkins 2/19/00 A Mathematcal Theory of Communcaton Claude hannon s aer resented by Kate Jenkns 2/19/00 Publshed n two arts, July 1948 and October 1948 n the Bell ystem Techncal Journal Foundng aer of Informaton Theory

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

Classification as a Regression Problem

Classification as a Regression Problem Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class

More information

On the Connectedness of the Solution Set for the Weak Vector Variational Inequality 1

On the Connectedness of the Solution Set for the Weak Vector Variational Inequality 1 Journal of Mathematcal Analyss and Alcatons 260, 15 2001 do:10.1006jmaa.2000.7389, avalable onlne at htt:.dealbrary.com on On the Connectedness of the Soluton Set for the Weak Vector Varatonal Inequalty

More information

Module 9. Lecture 6. Duality in Assignment Problems

Module 9. Lecture 6. Duality in Assignment Problems Module 9 1 Lecture 6 Dualty n Assgnment Problems In ths lecture we attempt to answer few other mportant questons posed n earler lecture for (AP) and see how some of them can be explaned through the concept

More information

Supplementary Material for Spectral Clustering based on the graph p-laplacian

Supplementary Material for Spectral Clustering based on the graph p-laplacian Sulementary Materal for Sectral Clusterng based on the grah -Lalacan Thomas Bühler and Matthas Hen Saarland Unversty, Saarbrücken, Germany {tb,hen}@csun-sbde May 009 Corrected verson, June 00 Abstract

More information

Managing Capacity Through Reward Programs. on-line companion page. Byung-Do Kim Seoul National University College of Business Administration

Managing Capacity Through Reward Programs. on-line companion page. Byung-Do Kim Seoul National University College of Business Administration Managng Caacty Through eward Programs on-lne comanon age Byung-Do Km Seoul Natonal Unversty College of Busness Admnstraton Mengze Sh Unversty of Toronto otman School of Management Toronto ON M5S E6 Canada

More information

Conservative Surrogate Model using Weighted Kriging Variance for Sampling-based RBDO

Conservative Surrogate Model using Weighted Kriging Variance for Sampling-based RBDO 9 th World Congress on Structural and Multdsclnary Otmzaton June 13-17, 011, Shzuoka, Jaan Conservatve Surrogate Model usng Weghted Krgng Varance for Samlng-based RBDO Lang Zhao 1, K.K. Cho, Ikn Lee 3,

More information

Approximation of Optimal Interface Boundary Conditions for Two-Lagrange Multiplier FETI Method

Approximation of Optimal Interface Boundary Conditions for Two-Lagrange Multiplier FETI Method Aroxmaton of Otmal Interface Boundary Condtons for Two-Lagrange Multler FETI Method F.-X. Roux, F. Magoulès, L. Seres, Y. Boubendr ONERA, 29 av. de la Dvson Leclerc, BP72, 92322 Châtllon, France, ,

More information

Topology optimization of plate structures subject to initial excitations for minimum dynamic performance index

Topology optimization of plate structures subject to initial excitations for minimum dynamic performance index th World Congress on Structural and Multdsclnary Otmsaton 7 th -2 th, June 25, Sydney Australa oology otmzaton of late structures subject to ntal exctatons for mnmum dynamc erformance ndex Kun Yan, Gengdong

More information

Priority Queuing with Finite Buffer Size and Randomized Push-out Mechanism

Priority Queuing with Finite Buffer Size and Randomized Push-out Mechanism ICN 00 Prorty Queung wth Fnte Buffer Sze and Randomzed Push-out Mechansm Vladmr Zaborovsy, Oleg Zayats, Vladmr Muluha Polytechncal Unversty, Sant-Petersburg, Russa Arl 4, 00 Content I. Introducton II.

More information

Numerical studies of space filling designs: optimization algorithms and subprojection properties

Numerical studies of space filling designs: optimization algorithms and subprojection properties umercal studes of sace fllng desgns: otmzaton algorthms and subroecton roertes Bertrand Iooss wth Gullaume Dambln & Matheu Coulet CEMRACS 03 July, 30th, 03 Motvatng eamle: Uncertantes management n smulaton

More information

Bayesian networks for scenario analysis of nuclear waste repositories

Bayesian networks for scenario analysis of nuclear waste repositories Bayesan networks for scenaro analyss of nuclear waste reostores Edoardo Toson ab Aht Salo a Enrco Zo bc a. Systems Analyss Laboratory Det of Mathematcs and Systems Analyss - Aalto Unversty b. Laboratory

More information

Not-for-Publication Appendix to Optimal Asymptotic Least Aquares Estimation in a Singular Set-up

Not-for-Publication Appendix to Optimal Asymptotic Least Aquares Estimation in a Singular Set-up Not-for-Publcaton Aendx to Otmal Asymtotc Least Aquares Estmaton n a Sngular Set-u Antono Dez de los Ros Bank of Canada dezbankofcanada.ca December 214 A Proof of Proostons A.1 Proof of Prooston 1 Ts roof

More information