Image Classification Using EM And JE algorithms

Size: px
Start display at page:

Download "Image Classification Using EM And JE algorithms"

Transcription

1 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Image Cassfcaton Usng EM And JE agorthms Xaojn Sh Department of Computer Engneerng, Unversty of Caforna, Santa Cruz, CA, 9564 jennfer@soe.ucsc.edu CS242: Machne Learnng, Fa 22, Dec 4, 22 Abstract Parameter estmaton probem s defntey one of the most mportant probems n the research fed of mage cassfcaton. Snce t s not very possbe to drecty get the parameters that ft data best, peope generay w use teratve way to sove ths probem. EM agorthm s a typca method to sove ths probem, and has been aready eadng ths fed for severa decades. Recenty, Yoram Snger and Manfred K. Warmuth provded a method---je agorthm to sove ths probem. JE agorthm has two versons: batch case and onne case. The purpose of ths project s to compare the performance of the three agorthms: EM, JE batch case, and JE onne case n the mage cassfcaton probem. Keywords: Image cassfcaton, EM agorthm, JE agorthm, onne agorthm, Gaussan Mxture.. Introducton Image object cassfcaton s a very mportant research topc n the pattern recognton fed. The purpose of ths work s: gven an mage, et computer to extract dfferent objects automatcay. That s, we want the computer to have sort of ntegence, so that an mage s no onger just a set of numbers to the computer, but a pcture that contans severa dfferent objects. And the computer can understand the dfference between dfferent objects. Ths work surey has ots of temptng appcatons, but how to acheve ths goa s a bg queston. - -

2 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Generay speakng, the steps of an mage cassfcaton probem can be seen n fgure. From fgure we can see, the quaty of the cassfers s the cruca part for the whoe mage cassfcaton probem. A good cassfer shoud have severa features: frst, t must be abe to ft the tranng data we. Ths s because the test data w be smar to the tranng data, f the cassfer can ft the tranng data we, t s reasonabe to conjecture the cassfer w ft the tranng data we; second, the cassfer must be robust. Ths s because that athough the tranng data s smar wth the test data, they are not dentca, the cassfer shoud be abe to toerate the sma dfference between the two datasets. Ony when a cassfer has both of the quaty can t perform we n the mage cassfcaton probem. Feature Extracton Tranng dataset Tranng Parameter estmaton Cassfer Cassfer 2 Cassfer n Cassfcaton Feature extracton Test dataset Fgure Cassfcaton Because of the nature of probabty, the cassfer based on probabty theory shoud be robust to nose. Ths s the reason why parameter estmaton methods are very popuar n the supervsed earnng probem. The basc dea of parameter - 2 -

3 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe estmaton methods s: we assume every cassfer has an underyng dstrbuton controng the data for ths cass, and we mode the dstrbuton of each cass wth a certan known or we-studed dstrbuton (for exampe, use Gaussan Mxture, the ony unknown factor for the cassfer s the parameters used to defne the dstrbuton (for exampe, the mean and covarance matrx for Gaussan dstrbuton. So the tranng turns to be a parameter estmaton probem. Snce generay t s not possbe to get the best parameters drecty, we often use some teratve agorthm to fnd a resut. If we mode the cass dstrbuton as Gaussan mxtures, EM agorthm s the typca method to sove ths probem, and peope has used ths method for severa decades. Recenty, Yoram Snger and Manfred K. Warmuth provded a parameter estmaton method for Gaussan Mxtures----JE agorthm. The purpose of ths project s to compare the performance of EM and JE agorthm. In ths project, I use two knds of datasets. The frst type s synthetc datasets, that s, I generate severa datasets based on severa known Gaussan mxture dstrbutons, and run the two agorthms over these synthetc datasets. Snce we ve aready known exacty what shoud be the fna resuts for the estmaton, we can use these datasets to compare the performance of the two approaches. There are two performances can be compared, frst s the accuracy, that s, we can estmate the error between the resut and the orgna dstrbuton, and see whether t s accurate enough; second, we can compare the number of teratons to make the agorthm converge, the convergence rate s very mportant because a these agorthms are teraton based, they need to run severa tmes to get the fna resut. The more number of teratons s, the more the computaton tme. If we can make the computaton faster, t w have much broader appcatons. The other knd of datasets I m gong to use s the rea mage datasets. These datasets are the rea datasets we encountered when we do mage cassfcaton, and are provded by professor Roberto Manduch. We have aready abeed the datasets (that s, ths s a supervsed tranng, and we ve aready had the tranng data for each cass. For each cass, we use the Gaussan Mxture to mode ts underyng dstrbuton. After we run the Monte Caro cross vadaton over the abeed data for each cass, we can know the number of custers (modes for each cass, whch means - 3 -

4 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe we ve aready known the number of dfferent Gaussans n the mxture dstrbuton, the ony probem s to estmate the parameters. Athough ths tme, we don t know exacty what s the rea vaue of a parameters, but we can have a genera dea of whether the estmaton s good enough from the fna cassfcaton resuts. If the cassfcaton resuts are very good, we can reasonaby thnk that the parameter estmaton resuts are good. Snce these are the rea data we encounter, the resuts w be meanngfu. It w show the robustness of the two agorthms. Aso t w show whether these agorthms are usefu n the rea appcatons. The foowng sectons are organzed as foows. Secton 2 w descrbe the three dfferent agorthms used n ths project, and made a comparson between them. Secton 3 w descrbe the experment datasets and resuts. Secton 4 w dscuss the experment resut gven n the Secton 3. Secton 5 w provde severa open probems and then provde the future work. Secton 6 gves the concuson. And Secton 7 and 8 are the acknowedgment and reference. 2. Agorthm descrpton There are gong to be three agorthms used n ths project: EM (Expectaton Maxmzaton agorthm, JE (Jont Entropy update agorthm (batch case, whch s provded by Yoram Snger and Manfred K. Warmuth, and the JE agorthm (on-ne case. Foowng w be a bref descrpton of these three agorthms. EM agorthm For the parameter estmaton probem for Mxture Gaussans, EM agorthm s amost a standard. Athough there are many dfferent ways to derve the update formua of EM agorthm, here I want to use the method ntroduced n the machne earnng cass

5 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Let H ndcate the hdden varabes, and V ndcate the vsbe varabes. Our purpose s to fnd the set of parameters Θ, whch can mnmze the negatve og ~ kehood of the vsbe data,.e. Θ = arg mn( n P ( V Θ. Snce mnmzng ths P( H V, Θ ~ quantty s very hard, we mnmze P( H V, Θn ~ + η( n P( V Θ H P( H V, Θ nstead. After set η =, and et the dervatve respect to the parameters equa to, we can get the EM update rue as foows. w µ Σ = = = = = = = p( x, Θ p( x, Θ od x p( x, Θ od = od ( x µ ( x od p( x, Θ µ p( x, Θ od T JE agorthm JE agorthm s an aternatve agorthm used to estmate the parameters for Mxture Gaussan, and s provded by Yoram Snger and Manfred K. Warmuth. The motvaton for ths agorthm probaby s ke ths. The formua to be mnmzed n the EM agorthm can be regard as a dstance between the and od parameter, and pus a oss part. If we use another knd of dstance, we w get a set of dfferent update formua. The dstance used for JE agorthm s the reatve entropy between the od and jont densty of the observed and hdden varabes (ote, n EM agorthm, the dstance s the reatve entropy between the od and densty of the - 5 -

6 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe hdden varabes ony. After a the dervatons, pease see [] for deta. We have the foowng update rues for JE agorthm n batch case. β ( x w µ Σ = m j = P( x w od j Θ od P( x Θ od od η w exp( β ( x = = m od η w j= j exp( β j ( x = od η od = µ + β ( x ( x µ = Σ od = η + = β ( x ( Σ Σ ( x µ ( x µ od od T Σ od JE onne agorthm Another mert of JE agorthm s that t s easy to be modfed to an on-ne agorthm. The update for the onne agorthm can be seen as foows. The ony requrement for the agorthm to converge s 2 η =, < t= t and η t= t. β ( x = w t = m P( x Θ j= t P( x w od w t od j Θ exp( η β ( x t t exp( η β ( x od µ = µ + η β x ( x µ t j ( t t t t od od od Σ = Σ + η β ( x ( Σ Σ ( x µ ( x µ T od t t t t Σ - 6 -

7 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe 3. Experment resuts In order to compare the performances of these three agorthms, I set up four sets of experments, each set s based on severa datasets. Experment and 2 are used to compare the performance of EM agorthm and the batch case JE agorthm; experment 3 and 4 are used to compare the performance of the batch case JE agorthm and the onne case JE agorthm. For a the experments, I use a KMean agorthm to gve nta vaues. Experment The purpose of experment s to compare the parameter estmaton performance of EM and JE agorthm usng synthetc data. Snce for the synthetc data, we know exacty the data dstrbuton, we can compare the accuracy and convergence speed of these two agorthms. I generated severa dfferent datasets, but due to the ack of space, I ony descrbe three of them. The frst dataset s a 3 modes D Mxture Gaussan. The sze of dataset s In order to see the accuracy of the parameter estmaton, I pot the orgna dstrbuton, the estmated dstrbuton provded by EM agorthm, and the estmated dstrbuton provded by JE agorthm n the same fgure. Pease see fgure for resut. From fgure we can see, the resuts of both of the agorthms are very good. But the convergent speeds are dfferent. Fgure2 pot the curve for the agorthms to get converged. From fgure2 we can see, the earnng rate η have very bg effect on the convergence speed. The detaed dscusson of the effect of earnng rate η can be seen n Secton

8 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Fgure Fgure 2 The second dataset s a 2 modes 2D Mxture Gaussan data. The sze s Fgure3 shows the orgna dstrbuton, and fgure4 shows the estmated resut from JE. The thrd dataset s a 3 modes 3D Mxture Gaussan data. The sze s Tabe shows the comparson of the orgna dstrbuton and the experment resut. From ths we can see both agorthms can acheve pretty good accuracy, but JE (n the best case agorthm need ess teratons to get converge (or get the same accuracy as EM agorthm

9 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Fgure 3 Fgure 4 Experment 2 In ths experment, I use the rea mage datasets. Because the rea mages won t foow a certan dstrbuton exacty, the parameter estmaton w be just a knd of approxmaton for the rea mage data, whch means there must be some error n the parameter estmaton. The purpose of ths experment s to test the performances of these two agorthms n a rea word appcaton, and see the robustness. Fgure5 and fgure6 show an exampe of the mages n the two rea mage datasets. These datasets are abeed, whch means ths s a supervsed earnng probem. The exampes of abeed data are show n fgure7 and fgure

10 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Orgna EM JE w [.2.5.3] [.2.5.3] [.3.5.2] µ Σ n 27 2 Tabe - -

11 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Fgure 5 Fgure 6 Fgure 7 Fgure 8 In ths experment, I frst use a the abeed mage data to tran a cassfer, and then used the traned cassfer to cassfy a the mages, and compute the cassfcaton error rate over the tranng data. In ths way, I ony dea wth the parameter estmaton probem (no predcton, and the error rate can party show the resut of the parameter estmaton by the two agorthms. Tabe2 and tabe3 show the number of teratons needed n tranng the cassfer for a partcuar cass, the tota teratons, and the error rate. - -

12 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe EM JE EM JE Cass Cass 2 Cass 3 4 (dverge (converge 9 49 Cass Cass 2 Cass 3 Cass Cass 4 Tota (norma case Error rate (dverge 247 7% 4 (dverge 4 % Cass 5 Cass 6 Tota (norma case Error rate 24% 23% Tab e 2 Tabe 3 Experment 3 The purpose of ths experment s to compare the parameter estmaton performance between the JE batch agorthm and the JE onne agorthm. In ths experment, I test the two agorthms on both synthetc data and rea mage data. Due to the ack of space, I ony show the resut of one mage dataset. Fgure9 and fgure s the resut of the rea mage dataset. And the way of carryng on ths experment s the same as the one descrbed n experment

13 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Fgure 9 Fgure Experment 4 As we mentoned before, a good cassfer shoud be not ony good at estmatng par ameters (.e. to ft the data we, but aso need to be good at predctng mages. The purpose of ths experment s to test the predctng abty of the two cassfers generated by the two agorthms separatey. The experments are run over both rea mage datasets n a eave-one-out way,.e., every tme, eave one mage of the dataset out as the test data, and use a the other mages as the tranng dataset. Ths knd of work w be done over a the mages, and get the average error rate as the fna resut. The resut can be shown n Tabe4. JE batch JE onne Road dataset Rock dataset Tabe 4 From tabe4 we can see, when we try to predct the mage usng the traned cassfers, the JE onne agorthm acheves much better resut than the JE batch agorthm

14 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe 4. Dscusson Experment & 2, why s the performance of JE batch agorthm out-performs EM?. From the update rue of EM and JE batch agorthm we can see, the updates for the covarance matrx are dfferent. In EM agorthm, we aways update the covarance matrx tsef. But n JE batch agorthm, we update the nverse of covarance matrx nstead. Ths means that n JE agorthm, we ony need to compute the nverse of matrx once, but n EM agorthm, we need to compute the nverse n every teraton. Because the nverse operaton s very unstabe, and t s easy to make the matrx become snguar when we have nose data, EM agorthm sometmes w be stuck at certan pont, but JE batch agorthm can avod ths probem. 2. The dfference between EM agorthm and JE agorthm s that the defntons of dstance n these two agorthms are dfferent. In EM agorthm, the dstance s ony consdered over the hdden varabes, but n JE agorthm, the dstance s consdered over both the hdden and vsbe varabe. It seems that ths s a more reasonabe defnton. In Experment & 2, what s the effect of the earnng rate to the JE agorthm? The effect of the earnng rate over the JE agorthm can be seen from the fgure and fgure2. From these two fgures we can see, the earnng rate does have a bg effect over the performance of JE agorthm, especay about the convergence rate. If we can get a good earnng rate, then the convergence s much faster than EM agorthm, but on the contrary, f we got a bad earnng rate, the performance w be much worse than the EM agorthm

15 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe Fgure Fgure 2 E xperment 3, why the performance of JE on-ne agorthm s a tte worse than the JE batch agorthm when we ony deang wth the parameter estmaton probem? I thnk the man probem s we ack a good way to tune the earnng rate. The advantage of an on-ne agorthm s that t can adjust the cassfer adaptvey to the nput data. But when we have a the data together, the advantage cannot show up. The dsadvantage of the JE on-ne agorthm s that we need to tune the earnng rate at every teraton, but for the batch case, we ony need to tune t once. So far, the way to tune the earnng rate s just through experments, so t s very possbe to make some mstakes. And as the tmes need to tune the earnng rate ncrease, t s more and more possbe to use bad earnng rate. When n the parameter estmaton probem ony the dsadvantage of JE onne agorthm can show up, t w work a tte worse. But from the experment we can see JE onne agorthm st works pretty good, the dfference between JE batch agorthm and JE onne agorthm s very sma. Experment 4, why does the JE on-ne agorthm out-performs JE batch agorthm when we use the cassfers n the rea cassfcaton probems (ncudng both the parameter estmaton over the tranng dataset and the predcton over the test data? When we try to predct over the test dataset, we meet some data that s smar but not dentca to the tranng dataset, whch means that we shoud make a tte change of the cassfer accordng to the data. In ths case, JE onne - 5 -

16 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe agorthm w show ts mert. So athough JE onne agorthm doesn t have a good parameter estmaton as the JE batch agorthm over the tranng dataset, when t receve the nput test data, t can change adaptvey to the data, make tsef ft the data better. So JE onne agorthm can get better resut. Possbe mprovement for JE onne agorthm. When we mpement the JE onne agorthm over the rea word mage data, there are severa probem shoud be notced. The frst one s how to dea wth the nose and outer n the tranng dataset. Because the onne agorthm deas wth the nput data one by one, each of them w have some effect over the cassfer. When we have nose or outer, generay t w be qute dfferent from the rea data, and as a resut, t w cause a bg change of the cassfer f we use the update rue, but ths change s not what we want. In my appcaton, I estmate the dfference between the and od parameters at each teraton, f the dfference s too bg, we can assume the data as nose and just dscard t. Snce n the mage cassfcaton probem, we aways have enough data, dscard severa data won t have bg nfuence over the fna resut, so we can mprove the performance of the onne agorthm. The second probem s how to arrange the sequence of data. In order to make the agorthm converge, we need the earnng rate to decrease accordng to tme. Ths means the earer data w have more effect over the fna resut, but the ater data w have ess effect. Ths won t have troube wth the mean and covarance, but w cause some troube of the pror for each mode. Generay, when we dea wth the mage data, we process data accordng to the sequence of mages, that s, we process the frst mage, then the second mage, and so on. But t s very possbe that the data n the same mage w be very smar to each other, whch means they come from the same mode of the mxture Gaussan, say mode. Then f we process a the data beong to mode frst, the pror for mode w be over emphaszed, and w ead to a wrong resut at ast. A more reasonabe way s to nput data accordng to sequence of modes, that s we frst get a data whch cose to mode, then we get a data cose to - 6 -

17 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe mode 2, and so on. In ths ntereavng way, we can avod the over-emphasze probem. Ths project s fnshed n ths way, and acheves very good resut. 5. Future work After ths project, there are st ots of thngs need to be soved n future. Frst of a, regardng to the earnng rate, how to tune the rght vaue s bg probem. From our experments we can see, the earnng rate s very cruca to the fna resut. If we can fnd a good earnng rate, we can get a much better performance than EM, but f not, t s very possbe that the resut w be even worse than EM, sometmes t even cannot converge. So far, my souton to ths probem s to run the JE agorthm over a set of possbe earnng rates, then when some tmes the og kehood of the tranng data becomes decreasng, I termnate the program. But n ths way, the JE agorthm oses part of ts mert: the fast converge rate. So ths s ony a temporary souton. How to tune the earnng rate s an open probem. Second, for the on-ne verson of JE agorthm, snce the earnng rate changes (decreases accordng to tme, whch means the earer data w have more effect over the fna resut, and the ater data w have ess effect over the fna resut, how to arrange the best nput data sequence becomes a probem. If we nput the data mage by mage, t s very possbe that the data correspond to a certan mode w be connected to each other, and f these data occur earer than other data, ths w ead the certan mode w be over-emphaszed n the fna resut (snce a of them have bgger earnng rate. So a more reasonabe souton to ths probem s to ntereave data from dfferent mode. So far, ths work s done manuay. But how to automatcay fnsh ths work s an open probem

18 Machne earnng project report Fa, 2 Xaojn Sh, jennfer@soe 6. Concuson Ths project compares the performance of three dfferent parameter estmaton agorthms: EM, JE batch case, and JE onne case. From the experment resut we can see, f we have a good earnng rate, JE batch case out-performs a ot than the EM agorthm. JE onne agorthm s a tte worse at parameter estmaton probem than the JE batch agorthm, but t acheves a much better performance n the predctng process, so that t can acheve a much better resut than n the rea mage cassfcaton probem. 7. Acknowedgment Thanks a ot to professor Manfred Warmuth, who gves me a ot of hepfu advce durng our dscusson of my project. Thanks a ot to professor Roberto Manduch, who gves me the test dataset, aso heps me a ot n understandng the probem and how to proceed. 8. Reference [] A ew Parameter Estmaton Method for Gaussan Mxtures, by Yoram Snger, Manfred K. Warmuth; [2] Addtve versus exponentated gradent updates for near predcton, by Jyrk Kvnen and Manfred K. Warmuth; [3] A Gente Tutora of the EM Agorthm and ts Appcaton to Parameter Estmaton for Gaussan Mxture and Hdden Markov Modes, by Jeff A. Bmes; [4] The EM Agorthm and Extensons, by Geoffrey J. McLachan, Thryambakam Krshnan; [5] On Bas, Varance, /-Loss, and the Curse-of-Dmensonaty, by Jerome H. Fredman [6] Fa, 22, machne earnng ecture notes

MARKOV CHAIN AND HIDDEN MARKOV MODEL

MARKOV CHAIN AND HIDDEN MARKOV MODEL MARKOV CHAIN AND HIDDEN MARKOV MODEL JIAN ZHANG JIANZHAN@STAT.PURDUE.EDU Markov chan and hdden Markov mode are probaby the smpest modes whch can be used to mode sequenta data,.e. data sampes whch are not

More information

Supplementary Material: Learning Structured Weight Uncertainty in Bayesian Neural Networks

Supplementary Material: Learning Structured Weight Uncertainty in Bayesian Neural Networks Shengyang Sun, Changyou Chen, Lawrence Carn Suppementary Matera: Learnng Structured Weght Uncertanty n Bayesan Neura Networks Shengyang Sun Changyou Chen Lawrence Carn Tsnghua Unversty Duke Unversty Duke

More information

Example: Suppose we want to build a classifier that recognizes WebPages of graduate students.

Example: Suppose we want to build a classifier that recognizes WebPages of graduate students. Exampe: Suppose we want to bud a cassfer that recognzes WebPages of graduate students. How can we fnd tranng data? We can browse the web and coect a sampe of WebPages of graduate students of varous unverstes.

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Supervised Learning. Neural Networks and Back-Propagation Learning. Credit Assignment Problem. Feedforward Network. Adaptive System.

Supervised Learning. Neural Networks and Back-Propagation Learning. Credit Assignment Problem. Feedforward Network. Adaptive System. Part 7: Neura Networ & earnng /2/05 Superved earnng Neura Networ and Bac-Propagaton earnng Produce dered output for tranng nput Generaze reaonaby & appropratey to other nput Good exampe: pattern recognton

More information

Research on Complex Networks Control Based on Fuzzy Integral Sliding Theory

Research on Complex Networks Control Based on Fuzzy Integral Sliding Theory Advanced Scence and Technoogy Letters Vo.83 (ISA 205), pp.60-65 http://dx.do.org/0.4257/ast.205.83.2 Research on Compex etworks Contro Based on Fuzzy Integra Sdng Theory Dongsheng Yang, Bngqng L, 2, He

More information

Neural network-based athletics performance prediction optimization model applied research

Neural network-based athletics performance prediction optimization model applied research Avaabe onne www.jocpr.com Journa of Chemca and Pharmaceutca Research, 04, 6(6):8-5 Research Artce ISSN : 0975-784 CODEN(USA) : JCPRC5 Neura networ-based athetcs performance predcton optmzaton mode apped

More information

Gaussian Mixture Models

Gaussian Mixture Models Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous

More information

Associative Memories

Associative Memories Assocatve Memores We consder now modes for unsupervsed earnng probems, caed auto-assocaton probems. Assocaton s the task of mappng patterns to patterns. In an assocatve memory the stmuus of an ncompete

More information

Multispectral Remote Sensing Image Classification Algorithm Based on Rough Set Theory

Multispectral Remote Sensing Image Classification Algorithm Based on Rough Set Theory Proceedngs of the 2009 IEEE Internatona Conference on Systems Man and Cybernetcs San Antono TX USA - October 2009 Mutspectra Remote Sensng Image Cassfcaton Agorthm Based on Rough Set Theory Yng Wang Xaoyun

More information

Singular Value Decomposition: Theory and Applications

Singular Value Decomposition: Theory and Applications Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real

More information

Delay tomography for large scale networks

Delay tomography for large scale networks Deay tomography for arge scae networks MENG-FU SHIH ALFRED O. HERO III Communcatons and Sgna Processng Laboratory Eectrca Engneerng and Computer Scence Department Unversty of Mchgan, 30 Bea. Ave., Ann

More information

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1 On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

GENERATIVE AND DISCRIMINATIVE CLASSIFIERS: NAIVE BAYES AND LOGISTIC REGRESSION. Machine Learning

GENERATIVE AND DISCRIMINATIVE CLASSIFIERS: NAIVE BAYES AND LOGISTIC REGRESSION. Machine Learning CHAPTER 3 GENERATIVE AND DISCRIMINATIVE CLASSIFIERS: NAIVE BAYES AND LOGISTIC REGRESSION Machne Learnng Copyrght c 205. Tom M. Mtche. A rghts reserved. *DRAFT OF September 23, 207* *PLEASE DO NOT DISTRIBUTE

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

A finite difference method for heat equation in the unbounded domain

A finite difference method for heat equation in the unbounded domain Internatona Conerence on Advanced ectronc Scence and Technoogy (AST 6) A nte derence method or heat equaton n the unbounded doman a Quan Zheng and Xn Zhao Coege o Scence North Chna nversty o Technoogy

More information

Probability Density Function Estimation by different Methods

Probability Density Function Estimation by different Methods EEE 739Q SPRIG 00 COURSE ASSIGMET REPORT Probablty Densty Functon Estmaton by dfferent Methods Vas Chandraant Rayar Abstract The am of the assgnment was to estmate the probablty densty functon (PDF of

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Online Classification: Perceptron and Winnow

Online Classification: Perceptron and Winnow E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng

More information

Differentiating Gaussian Processes

Differentiating Gaussian Processes Dfferentatng Gaussan Processes Andrew McHutchon Aprl 17, 013 1 Frst Order Dervatve of the Posteror Mean The posteror mean of a GP s gven by, f = x, X KX, X 1 y x, X α 1 Only the x, X term depends on the

More information

Supporting Information

Supporting Information Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to

More information

Application of support vector machine in health monitoring of plate structures

Application of support vector machine in health monitoring of plate structures Appcaton of support vector machne n heath montorng of pate structures *Satsh Satpa 1), Yogesh Khandare ), Sauvk Banerjee 3) and Anrban Guha 4) 1), ), 4) Department of Mechanca Engneerng, Indan Insttute

More information

Cyclic Codes BCH Codes

Cyclic Codes BCH Codes Cycc Codes BCH Codes Gaos Feds GF m A Gaos fed of m eements can be obtaned usng the symbos 0,, á, and the eements beng 0,, á, á, á 3 m,... so that fed F* s cosed under mutpcaton wth m eements. The operator

More information

Chapter 6 Hidden Markov Models. Chaochun Wei Spring 2018

Chapter 6 Hidden Markov Models. Chaochun Wei Spring 2018 896 920 987 2006 Chapter 6 Hdden Markov Modes Chaochun We Sprng 208 Contents Readng materas Introducton to Hdden Markov Mode Markov chans Hdden Markov Modes Parameter estmaton for HMMs 2 Readng Rabner,

More information

3. Stress-strain relationships of a composite layer

3. Stress-strain relationships of a composite layer OM PO I O U P U N I V I Y O F W N ompostes ourse 8-9 Unversty of wente ng. &ech... tress-stran reatonshps of a composte ayer - Laurent Warnet & emo Aerman.. tress-stran reatonshps of a composte ayer Introducton

More information

Lecture 4. Instructor: Haipeng Luo

Lecture 4. Instructor: Haipeng Luo Lecture 4 Instructor: Hapeng Luo In the followng lectures, we focus on the expert problem and study more adaptve algorthms. Although Hedge s proven to be worst-case optmal, one may wonder how well t would

More information

Homework Assignment 3 Due in class, Thursday October 15

Homework Assignment 3 Due in class, Thursday October 15 Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.

More information

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements

CS 2750 Machine Learning. Lecture 5. Density estimation. CS 2750 Machine Learning. Announcements CS 750 Machne Learnng Lecture 5 Densty estmaton Mlos Hauskrecht mlos@cs.ptt.edu 539 Sennott Square CS 750 Machne Learnng Announcements Homework Due on Wednesday before the class Reports: hand n before

More information

CHAPTER 7: CLUSTERING

CHAPTER 7: CLUSTERING CHAPTER 7: CLUSTERING Semparamerc Densy Esmaon 3 Paramerc: Assume a snge mode for p ( C ) (Chapers 4 and 5) Semparamerc: p ( C ) s a mure of denses Mupe possbe epanaons/prooypes: Dfferen handwrng syes,

More information

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012

MLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012 MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:

More information

VQ widely used in coding speech, image, and video

VQ widely used in coding speech, image, and video at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng

More information

CS229 Lecture notes. Andrew Ng

CS229 Lecture notes. Andrew Ng CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view

More information

SDMML HT MSc Problem Sheet 4

SDMML HT MSc Problem Sheet 4 SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

COXREG. Estimation (1)

COXREG. Estimation (1) COXREG Cox (972) frst suggested the modes n whch factors reated to fetme have a mutpcatve effect on the hazard functon. These modes are caed proportona hazards (PH) modes. Under the proportona hazards

More information

Nested case-control and case-cohort studies

Nested case-control and case-cohort studies Outne: Nested case-contro and case-cohort studes Ørnuf Borgan Department of Mathematcs Unversty of Oso NORBIS course Unversty of Oso 4-8 December 217 1 Radaton and breast cancer data Nested case contro

More information

This column is a continuation of our previous column

This column is a continuation of our previous column Comparson of Goodness of Ft Statstcs for Lnear Regresson, Part II The authors contnue ther dscusson of the correlaton coeffcent n developng a calbraton for quanttatve analyss. Jerome Workman Jr. and Howard

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #16 Scribe: Yannan Wang April 3, 2014

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #16 Scribe: Yannan Wang April 3, 2014 COS 511: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture #16 Scrbe: Yannan Wang Aprl 3, 014 1 Introducton The goal of our onlne learnng scenaro from last class s C comparng wth best expert and

More information

Relevance Vector Machines Explained

Relevance Vector Machines Explained October 19, 2010 Relevance Vector Machnes Explaned Trstan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introducton Ths document has been wrtten n an attempt to make Tppng s [1] Relevance Vector Machnes

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

Quantum Runge-Lenz Vector and the Hydrogen Atom, the hidden SO(4) symmetry

Quantum Runge-Lenz Vector and the Hydrogen Atom, the hidden SO(4) symmetry Quantum Runge-Lenz ector and the Hydrogen Atom, the hdden SO(4) symmetry Pasca Szrftgser and Edgardo S. Cheb-Terrab () Laboratore PhLAM, UMR CNRS 85, Unversté Le, F-59655, France () Mapesoft Let's consder

More information

Chapter 6. Rotations and Tensors

Chapter 6. Rotations and Tensors Vector Spaces n Physcs 8/6/5 Chapter 6. Rotatons and ensors here s a speca knd of near transformaton whch s used to transforms coordnates from one set of axes to another set of axes (wth the same orgn).

More information

Multilayer Perceptron (MLP)

Multilayer Perceptron (MLP) Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne

More information

A DIMENSION-REDUCTION METHOD FOR STOCHASTIC ANALYSIS SECOND-MOMENT ANALYSIS

A DIMENSION-REDUCTION METHOD FOR STOCHASTIC ANALYSIS SECOND-MOMENT ANALYSIS A DIMESIO-REDUCTIO METHOD FOR STOCHASTIC AALYSIS SECOD-MOMET AALYSIS S. Rahman Department of Mechanca Engneerng and Center for Computer-Aded Desgn The Unversty of Iowa Iowa Cty, IA 52245 June 2003 OUTLIE

More information

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD CHALMERS, GÖTEBORGS UNIVERSITET SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 35, FIM 72 GU, PhD Tme: Place: Teachers: Allowed materal: Not allowed: January 2, 28, at 8 3 2 3 SB

More information

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30

STATS 306B: Unsupervised Learning Spring Lecture 10 April 30 STATS 306B: Unsupervsed Learnng Sprng 2014 Lecture 10 Aprl 30 Lecturer: Lester Mackey Scrbe: Joey Arthur, Rakesh Achanta 10.1 Factor Analyss 10.1.1 Recap Recall the factor analyss (FA) model for lnear

More information

The Entire Solution Path for Support Vector Machine in Positive and Unlabeled Classification 1

The Entire Solution Path for Support Vector Machine in Positive and Unlabeled Classification 1 Abstract The Entre Souton Path for Support Vector Machne n Postve and Unabeed Cassfcaton 1 Yao Lmn, Tang Je, and L Juanz Department of Computer Scence, Tsnghua Unversty 1-308, FIT, Tsnghua Unversty, Bejng,

More information

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve

More information

9 Adaptive Soft K-Nearest-Neighbour Classifiers with Large Margin

9 Adaptive Soft K-Nearest-Neighbour Classifiers with Large Margin 9 Adaptve Soft -Nearest-Neghbour Cassfers wth Large argn Abstract- A nove cassfer s ntroduced to overcome the mtatons of the -NN cassfcaton systems. It estmates the posteror cass probabtes usng a oca Parzen

More information

Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data

Conditional Random Fields: Probabilistic Models for Segmenting and Labeling Sequence Data Condtonal Random Felds: Probablstc Models for Segmentng and Labelng Sequence Data Paper by John Lafferty, Andrew McCallum, and Fernando Perera ICML 2001 Presentaton by Joe Drsh May 9, 2002 Man Goals Present

More information

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,

More information

Ensemble Methods: Boosting

Ensemble Methods: Boosting Ensemble Methods: Boostng Ncholas Ruozz Unversty of Texas at Dallas Based on the sldes of Vbhav Gogate and Rob Schapre Last Tme Varance reducton va baggng Generate new tranng data sets by samplng wth replacement

More information

On the Power Function of the Likelihood Ratio Test for MANOVA

On the Power Function of the Likelihood Ratio Test for MANOVA Journa of Mutvarate Anayss 8, 416 41 (00) do:10.1006/jmva.001.036 On the Power Functon of the Lkehood Rato Test for MANOVA Dua Kumar Bhaumk Unversty of South Aabama and Unversty of Inos at Chcago and Sanat

More information

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression 11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING

More information

Semi-supervised Classification with Active Query Selection

Semi-supervised Classification with Active Query Selection Sem-supervsed Classfcaton wth Actve Query Selecton Jao Wang and Swe Luo School of Computer and Informaton Technology, Beng Jaotong Unversty, Beng 00044, Chna Wangjao088@63.com Abstract. Labeled samples

More information

Note 2. Ling fong Li. 1 Klein Gordon Equation Probablity interpretation Solutions to Klein-Gordon Equation... 2

Note 2. Ling fong Li. 1 Klein Gordon Equation Probablity interpretation Solutions to Klein-Gordon Equation... 2 Note 2 Lng fong L Contents Ken Gordon Equaton. Probabty nterpretaton......................................2 Soutons to Ken-Gordon Equaton............................... 2 2 Drac Equaton 3 2. Probabty nterpretaton.....................................

More information

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities

Supplementary material: Margin based PU Learning. Matrix Concentration Inequalities Supplementary materal: Margn based PU Learnng We gve the complete proofs of Theorem and n Secton We frst ntroduce the well-known concentraton nequalty, so the covarance estmator can be bounded Then we

More information

The Order Relation and Trace Inequalities for. Hermitian Operators

The Order Relation and Trace Inequalities for. Hermitian Operators Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora

princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 3: Large deviations bounds and applications Lecturer: Sanjeev Arora prnceton unv. F 13 cos 521: Advanced Algorthm Desgn Lecture 3: Large devatons bounds and applcatons Lecturer: Sanjeev Arora Scrbe: Today s topc s devaton bounds: what s the probablty that a random varable

More information

Boostrapaggregating (Bagging)

Boostrapaggregating (Bagging) Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod

More information

STAT 3008 Applied Regression Analysis

STAT 3008 Applied Regression Analysis STAT 3008 Appled Regresson Analyss Tutoral : Smple Lnear Regresson LAI Chun He Department of Statstcs, The Chnese Unversty of Hong Kong 1 Model Assumpton To quantfy the relatonshp between two factors,

More information

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z ) C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z

More information

Natural Images, Gaussian Mixtures and Dead Leaves Supplementary Material

Natural Images, Gaussian Mixtures and Dead Leaves Supplementary Material Natural Images, Gaussan Mxtures and Dead Leaves Supplementary Materal Danel Zoran Interdscplnary Center for Neural Computaton Hebrew Unversty of Jerusalem Israel http://www.cs.huj.ac.l/ danez Yar Wess

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

1 Motivation and Introduction

1 Motivation and Introduction Instructor: Dr. Volkan Cevher EXPECTATION PROPAGATION September 30, 2008 Rce Unversty STAT 63 / ELEC 633: Graphcal Models Scrbes: Ahmad Beram Andrew Waters Matthew Nokleby Index terms: Approxmate nference,

More information

A Dissimilarity Measure Based on Singular Value and Its Application in Incremental Discounting

A Dissimilarity Measure Based on Singular Value and Its Application in Incremental Discounting A Dssmarty Measure Based on Snguar Vaue and Its Appcaton n Incrementa Dscountng KE Xaou Department of Automaton, Unversty of Scence and Technoogy of Chna, Hefe, Chna Ema: kxu@ma.ustc.edu.cn MA Lyao Department

More information

Excess Error, Approximation Error, and Estimation Error

Excess Error, Approximation Error, and Estimation Error E0 370 Statstcal Learnng Theory Lecture 10 Sep 15, 011 Excess Error, Approxaton Error, and Estaton Error Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton So far, we have consdered the fnte saple

More information

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture # 15 Scribe: Jieming Mao April 1, 2013

COS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture # 15 Scribe: Jieming Mao April 1, 2013 COS 511: heoretcal Machne Learnng Lecturer: Rob Schapre Lecture # 15 Scrbe: Jemng Mao Aprl 1, 013 1 Bref revew 1.1 Learnng wth expert advce Last tme, we started to talk about learnng wth expert advce.

More information

Short-Term Load Forecasting for Electric Power Systems Using the PSO-SVR and FCM Clustering Techniques

Short-Term Load Forecasting for Electric Power Systems Using the PSO-SVR and FCM Clustering Techniques Energes 20, 4, 73-84; do:0.3390/en40073 Artce OPEN ACCESS energes ISSN 996-073 www.mdp.com/journa/energes Short-Term Load Forecastng for Eectrc Power Systems Usng the PSO-SVR and FCM Custerng Technques

More information

Regularized Discriminant Analysis for Face Recognition

Regularized Discriminant Analysis for Face Recognition 1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths

More information

QUARTERLY OF APPLIED MATHEMATICS

QUARTERLY OF APPLIED MATHEMATICS QUARTERLY OF APPLIED MATHEMATICS Voume XLI October 983 Number 3 DIAKOPTICS OR TEARING-A MATHEMATICAL APPROACH* By P. W. AITCHISON Unversty of Mantoba Abstract. The method of dakoptcs or tearng was ntroduced

More information

22.51 Quantum Theory of Radiation Interactions

22.51 Quantum Theory of Radiation Interactions .51 Quantum Theory of Radaton Interactons Fna Exam - Soutons Tuesday December 15, 009 Probem 1 Harmonc oscator 0 ponts Consder an harmonc oscator descrbed by the Hamtonan H = ω(nˆ + ). Cacuate the evouton

More information

Note on EM-training of IBM-model 1

Note on EM-training of IBM-model 1 Note on EM-tranng of IBM-model INF58 Language Technologcal Applcatons, Fall The sldes on ths subject (nf58 6.pdf) ncludng the example seem nsuffcent to gve a good grasp of what s gong on. Hence here are

More information

Markov Chain Monte Carlo Lecture 6

Markov Chain Monte Carlo Lecture 6 where (x 1,..., x N ) X N, N s called the populaton sze, f(x) f (x) for at least one {1, 2,..., N}, and those dfferent from f(x) are called the tral dstrbutons n terms of mportance samplng. Dfferent ways

More information

Separation of Variables and a Spherical Shell with Surface Charge

Separation of Variables and a Spherical Shell with Surface Charge Separation of Variabes and a Spherica She with Surface Charge In cass we worked out the eectrostatic potentia due to a spherica she of radius R with a surface charge density σθ = σ cos θ. This cacuation

More information

CS : Algorithms and Uncertainty Lecture 14 Date: October 17, 2016

CS : Algorithms and Uncertainty Lecture 14 Date: October 17, 2016 CS 294-128: Algorthms and Uncertanty Lecture 14 Date: October 17, 2016 Instructor: Nkhl Bansal Scrbe: Antares Chen 1 Introducton In ths lecture, we revew results regardng follow the regularzed leader (FTRL.

More information

Prediction Error of the Multivariate Additive Loss Reserving Method for Dependent Lines of Business

Prediction Error of the Multivariate Additive Loss Reserving Method for Dependent Lines of Business Predcton Error of the Mutvarate Addtve Loss Reservng Method for Dependent Lnes of Busness by Mchae Merz and Maro V Wüthrch ABSTRACT Often n non-fe nsurance, cams reserves are the argest poston on the abty

More information

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence

More information

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015 CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research

More information

An Application of Fuzzy Hypotheses Testing in Radar Detection

An Application of Fuzzy Hypotheses Testing in Radar Detection Proceedngs of the th WSES Internatonal Conference on FUZZY SYSEMS n pplcaton of Fuy Hypotheses estng n Radar Detecton.K.ELSHERIF, F.M.BBDY, G.M.BDELHMID Department of Mathematcs Mltary echncal Collage

More information

Deriving the Dual. Prof. Bennett Math of Data Science 1/13/06

Deriving the Dual. Prof. Bennett Math of Data Science 1/13/06 Dervng the Dua Prof. Bennett Math of Data Scence /3/06 Outne Ntty Grtty for SVM Revew Rdge Regresson LS-SVM=KRR Dua Dervaton Bas Issue Summary Ntty Grtty Need Dua of w, b, z w 2 2 mn st. ( x w ) = C z

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

MODEL TUNING WITH THE USE OF HEURISTIC-FREE GMDH (GROUP METHOD OF DATA HANDLING) NETWORKS

MODEL TUNING WITH THE USE OF HEURISTIC-FREE GMDH (GROUP METHOD OF DATA HANDLING) NETWORKS MODEL TUNING WITH THE USE OF HEURISTIC-FREE (GROUP METHOD OF DATA HANDLING) NETWORKS M.C. Schrver (), E.J.H. Kerchoffs (), P.J. Water (), K.D. Saman () () Rswaterstaat Drecte Zeeand () Deft Unversty of

More information

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics

Space of ML Problems. CSE 473: Artificial Intelligence. Parameter Estimation and Bayesian Networks. Learning Topics /7/7 CSE 73: Artfcal Intellgence Bayesan - Learnng Deter Fox Sldes adapted from Dan Weld, Jack Breese, Dan Klen, Daphne Koller, Stuart Russell, Andrew Moore & Luke Zettlemoyer What s Beng Learned? Space

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Outline. Communication. Bellman Ford Algorithm. Bellman Ford Example. Bellman Ford Shortest Path [1]

Outline. Communication. Bellman Ford Algorithm. Bellman Ford Example. Bellman Ford Shortest Path [1] DYNAMIC SHORTEST PATH SEARCH AND SYNCHRONIZED TASK SWITCHING Jay Wagenpfel, Adran Trachte 2 Outlne Shortest Communcaton Path Searchng Bellmann Ford algorthm Algorthm for dynamc case Modfcatons to our algorthm

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

Joint Statistical Meetings - Biopharmaceutical Section

Joint Statistical Meetings - Biopharmaceutical Section Iteratve Ch-Square Test for Equvalence of Multple Treatment Groups Te-Hua Ng*, U.S. Food and Drug Admnstraton 1401 Rockvlle Pke, #200S, HFM-217, Rockvlle, MD 20852-1448 Key Words: Equvalence Testng; Actve

More information

IV. Performance Optimization

IV. Performance Optimization IV. Performance Optmzaton A. Steepest descent algorthm defnton how to set up bounds on learnng rate mnmzaton n a lne (varyng learnng rate) momentum learnng examples B. Newton s method defnton Gauss-Newton

More information

Feature Selection: Part 1

Feature Selection: Part 1 CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?

More information

Hidden Markov Models

Hidden Markov Models Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011

Stanford University CS359G: Graph Partitioning and Expanders Handout 4 Luca Trevisan January 13, 2011 Stanford Unversty CS359G: Graph Parttonng and Expanders Handout 4 Luca Trevsan January 3, 0 Lecture 4 In whch we prove the dffcult drecton of Cheeger s nequalty. As n the past lectures, consder an undrected

More information