Experiments on logistic regression

Size: px
Start display at page:

Download "Experiments on logistic regression"

Transcription

1 Experimens on logisic regression Ning Bao March, 8 Absrac In his repor, several experimens have been conduced on a spam daa se wih Logisic Regression based on Gradien Descen approach. Firs, he overfiing effec is shown wih basic seings (vanilla version). Then Sochasic Gradien Descen and -Norm Regularizaion echniques are boh implemened wih demonsraion of he benefis of hese wo mehods in prevening overfiing. Besides, a new rick of modifying Sigmoid ransfer funcion in he raining sage is described and implemened, which is shown could help reducing overfiing as well. A las, he Label Shrinking echnique is implemened. Ye, he resuls on his daa se are no saisfying. Preliminaries. Logisic Regression The daa is a se of examples (x, y ). The feaures of an example x is ypical binary; and he label y is ypically binary as well. To predic, firs using linear acivaion of individual feaures â = w x well. This is acually a single Neuron. One of our goal is o rain his neuron o make i predicing Then use he Sigmoid funcion as he ransfer funcion o ransfer he linear acivaion ino probabiliy as he predicion ŷ = σ(â ) = And Logisic Loss has he following form + e â = + e w x. Loss(y, ŷ ) = y ln y + ( y ) ln y ŷ ŷ ln( ŷ ) = ln( + e w x ) if y = = ln ŷ = ln( + e w x ) w x if y = = ln( + e w x ) y w x

2 PRELIMINARIES. Gradien Descen Basically speaking, he Gradien Descen I have used is essenially abou solving he below minimizaion problem: inf w Loss(y, σ(w x )) T For he unregularized version of Gradien Descen., we ried o minimize he unregularized average logisic loss wih respec o he vecor w Loss(w) = T Loss(w) = Loss(y, ŷ ) T = (ln( + ew x ) y w x ) T (Theoreically, I should use he oal logisic loss as described in Secion. as he arge funcion o minimize. However, during he experimen I found if I use he oal logisic loss, some of he neuron will become raher larger and lead o considerably huge linear acivaion. This in urn will cause infinie logisic loss in laer raining passes. Thus, I swiched o using average logisic loss over he examples as he arge funcion. ) Consequenly, he gradien of he arge funcion will be w = Loss(w) w = (ŷ y )x T = (σ(w x ) y )x T i, wi = (ŷ y )x,i T = (σ(w x ) y )x,i T If we le w = o minimize he loss, he resuling below equaions are difficul o solve i, σ(w x )x,i = y x,i Thus, we use he G.D. updaing ieraively o minimize he loss, as w p+ = w p η w i, w p+,i = w p,i η wi in which i is he index of feaures and p is he index of ieraions. (I migh ake quie a few

3 PRELIMINARIES 3 ieraions o converge o he approximae minimum.).3 Daa The size of he daa is, in which here are examples (rows) and each example consiss of feaures (columns). The firs column of each row is he real label. All he labels and feaures are binary aking on values of {, }. In he daa, here are 68 posiive examples (wih label value ). The visualizaion of original and permued daa ses are shown in Figure, in which he black dos represens a value or eiher label or feaure poin. (MATLAB : imagesc(-daa); colormap gray) (MATLAB : p=randperm(); daa_perm=daa(p(:),:); imagesc(-daa_perm); colormap gray) The map of original daase (*) The map of permued daase (*) Figure : The map of he original daase.4 Cross Validaion For some experimens conduced, I employed Cross Validaion echnique o pick he bes model. In his conex, he bes model is composed of paricular parameers and he corresponding vecor (or Neuron) since for each differen value of parameers we need o rain he neuron o learn he daa. Using Cross Validaion is he sandard echnique in machine learning for model selecion and resul reporing. I will preven he bias incurred wih cerain daa. Thus, i makes he resul of cerain algorihm more accepable. In he raining sage, using he 3-fold Cross Validaion scheme as below:

4 VANILLA VERSION OF LOGISTIC REGRESSION 4 pariion he raining se ino 3 pars for each of he 3 holdous rain all models on he /3 pars record average logisic loss on he /3 par he bes model is chosen as he one wih hen bes average validaion loss over he 3 handous Figure : The 3-fold Cross Validaion scheme.5 Srucure of his repor I have conduced several experimens wih logisic regression. The ouline of hese resuls is lised below. In Secion, I experimened wih he vanilla version of logisic regression on illusrae he overfiing problem. Then, briefly I demonsraed in Secion 3 ha Sochasic Gradien Descen wih simulaed annealing echnique could help o preven he overfiing effec in logisic regression. Besides, he -Norm regularizaion could help reduce he overfiing effec as well, as shown in Secion 4. Then, I ried he new rick of modifying he Sigmoid ransfer funcion during he raining process. The deail informaion is provided in Secion 5. A las, he echnique of label shrinking is also implemened and described in Secion 6. Vanilla version of Logisic Regression For he Vanilla version of Logisic Regression, I only use he mehod menioned in Secion. and Secion.. Ye, one need o decide when o sop he raining process using Gradien Descen mehod. Usually, he -Norm of gradien is a rule-of-humb sopping crierion. Also, one could use -Norm of gradien as well. The basic idea is he same: sop he raining when he gradiens are small enough which implies ha he model approaches he local minimum of he loss funcion. Noe ha by using -Norm (or -Norm) or gradien as sopping crieria is kind of implici regularizaion - early sopping. Since as one rains he model more on he raining daase, he model will overfi he raining se wih relaively small loss. However, when predicing he model may incur considerable large loss due o he unknown new daa.. -Norm of gradien as sopping crieria For his par, he mehod s ouline is lised in Figure 3. Training he neuron on 3/4 of he whole daase, /4 as he esing se early sopping wih he crierion ha he -Norm of gradien converge o some level repor average logisic loss on raining se & esing se for comparison Figure 3: Mehod s ouline of -Norm of gradien sopping crieria, Vanilla Log. Reg.

5 VANILLA VERSION OF LOGISTIC REGRESSION 5 Noe ha he learning rae η is he parameer, which should be uned. Ye, for simpliciy, I picked η =. as a rule-of-humb value for all he experimens below. The jusificaion here is ha if η is larger, hen each sep of gradien descen is larger which is faser in convergence ye may no be able o approach he local minimum sufficienly enough since he jumping back-and-forh effec in gradien descen. On he oher hand, if η is smaller he convergence speed is exremely slow which is no feasible in pracical implemenaion. The sopping crieria here is (σ(w x ) y )x T i, i [, 3] I ploed he mean logisic loss on he raining se as well as on he esing se in Figure 4..6 Vanilla Log. Reg., Norm sopping crierion, (η =.) ^() ^( ) ^( ) Figure 4: Vanilla version, Early sopping by -Norm of gradien, v.s. es loss We could see from Figure 4, as he model is raining more and more he mean logisic loss on he esing se increases (afer w < ) due o he overfiing effec.. -Norm of gradien as sopping crieria Nex, I experimened wih -Norm of gradien as sopping crieria. Oher seings are he same as in.. The sopping crieria now is

6 VANILLA VERSION OF LOGISTIC REGRESSION 6 And he resul is depiced in Figure 5. (σ(w x ) y )x T i, i [, 4].7 Vanilla Log. Reg., Norm of gradien sopping crierion, (η =.) ^ ^( ) ^( ) ^( 3) ^( 4) The Norm of Gradien a sopping poin Figure 5: Vanilla version, Early sopping by -Norm of gradien, v.s. es loss This resul has wo poins need o discuss: Firs, he doesn blow up subsanially as he -Norm resul. In oher words, he overfiing phenomenon is no so obvious in his case. My explanaion is ha he -Norm value is basically smaller han he -Norm of gradien, hus by adoping his crierion he neuron is no so over-rained as before. Second, he mean logisic losses in his resul are greaer han he -Norm case. The explanaion is similar as above, he neuron is no rained ha much, herefore he average logisic loss is no minimized enough here. For a beer visualizaion, I ploed he -Norm and -Norm cases ogeher as in Figure 6.

7 3 STOCHASTIC GRADIENT DESCENT WITH SIMULATED ANNEALING 7.7 Vanilla Log. Reg., Norm vs Norm of gradien sopping crieria, (η =.).6.5 meanlogisic Loss.4.3, Norm Norm, Norm Norm.. ^() ^( ) ^( ) Figure 6: Early sopping by -Norm vs -Norm of gradien Anoher ineresing observaion I have made is ha using -Norm of gradien as he sopping crierion is exremely slow han using he -Norm. Whereas i seems he -Norm crierion could also provide reasonable loss if se o cerain level. Basically speaking, he wo sopping crieria have similar effecs in he sense of early sopping. I s difficul o judge which one is beer if one only considers he mean logisic loss since he only difference beween hem is he raining ime. If he model is rained sufficienly enough wihou overfiing, hey will provide he exac same resuls. To conclude, by seing a reasonable sopping crierion, one could use early sopping o help prevening he overfiing effec wihou changing oher pars of he gradien descen mehod in logisic regression. 3 Sochasic Gradien Descen wih Simulaed Annealing In his par, I implemened sochasic gradien descen approach. Insead of updaing all he afer seeing all he examples, sochasic gradien descen will updae he weigh vecor example by example, similar as he online seing. Besides, his approach employed a simulaed annealing echnique o sabilize he converging process - he learning rae will decay as he raining process goes which helps prevening overfiing effec.

8 4 -NORM REGULARIZATION 8 The sochasic gradien descen will updae he wih w = w η α i w in which i is he number of raining passes over he raining se and α <. Thus as he raining goes over passes by passes, he learning rae will drop and sabilize ( cool down ) he converging. Consequenly, he will change less and less. This is shown in Figure 7 Sochasic Gradien Descen, (η =., α =.95) Sochasic G.D. (η =., α =.95) Number of passes over raining se Number of passes over he raining se Figure 7: Mean log. loss and of Sochasic Gradien Descen (η =., α =.95) 4 -Norm Regularizaion In his par, I experimened wih he -Norm regularizaion effecs of prevening overfiing. By using -Norm regularizaion, he moivaion of updae would become inf w and he updae formula will become ( λ w + Loss(y, σ(w x ))) T w = w η(λw + (σ(w x ) y )x ) T Figure 8 is a plo of -Norm regularized logisic regression, raining unil w < 6. As we could see, since we have he regularizer in he moivaion i helps preven overfiing by punishing he large. Thus, as we rained more he mean loss on esing se flaens ou. Also, he is conrolled wihin a relaively small range.

9 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA Norm Regularized logisic reg. (η =., λ =.5).5 Weighs variaion of Norm Regularizaion, (η =., λ =.5) ^() ^( ) ^( ) ^( 3) ^( 4) ^( 5) ^( 6) ^() ^( ) ^( ) ^( 3) ^( 4) ^( 5) ^( 6) Norm of gradiens a sopping poins Figure 8: Mean log. loss and of -Norm Regularized G.D. (η =., λ =.5) 5 Modifying predicion funcion in updae formula In his par, I changed he original sigmoid funcion of predicion in he updae formula of during raining, in order o conrol he variaion of. However, when predicing on esing se he sigmoid funcion is lef unchanged. By changing he ransfer funcion, one could conrol he since if he linear acivaion is larger/smaller han cerain heshold he won be updaed. In oher words, he idea could be expressed as once he predicion is good, he say pu. 5. Modified sigmoid funcion Here, I use he following predicion if w x σ heshold ŷ = oherwise +e w x if w x σ heshold which looks like in Figure 9

10 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA Modified Sigmoid func, (σ heshold = 3) Original Sigmoid func Figure 9: Modified Sigmoid funcion as predicion (σ heshold = 3) 5.. Effecs in prevening overfiing Wih his change he will be conrol from growing/dropping oo much. As shown in Figure, he are inenionally bounded compared wih he original sigmoid case. 5 Bounded sigmoid predicion, early sopping by Norm gradien (η =., σ heshold = 3) 5 Original Sigmoid func., Norm of gradien early sopping, (η =.) ^() ^( ) ^( ) ^( 3) gradiens a sopping poins 5 ^() ^( ) ^( ) ^( 3) gradien a sopping poins Figure : Weighs comparison, modified sigmoid vs original Inuiively, his bounding effec on would help prevening overfiing on raining se. As shown in Figure, he mean logisic loss on esing se increase moderaely as raining more (i.e. sopping a smaller -Norm of gradien).

11 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA.6.4 Bounded sigmoid predicion, (η =., σ heshold = 3).6.4 Vanilla Log. Reg., Norm sopping crierion, (η =.) ^() ^( ) ^( ) ^( 3) ^() ^( ) ^( ) Figure : Mean log. loss comparison, modified sigmoid vs original 5.. Varying σ heshold By varying he value of σ heshold, I ploed he curves of mean logisic loss on boh raining and esing ses as shown in Figure. I is observed ha as σ heshold increases he increases as well while he drops..8 Bounded Sigmoid func., varying σ heshold, (η =.) σ heshold Figure : The effecs of differen values of σ heshold This resul accords wih inuiion since as σ heshold increases, he funcion will be more like he

12 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA original sigmoid and he are no inenionally bounded. Therefore, he overfiing effec will show up as he goes up in Figure. Also, in Table σ heshold loss rain. valid. rain. valid. rain. valid. rain. valid. rain. valid. s fold nd fold rd fold avg σ heshold loss rain. valid. rain. valid. rain. valid. rain. valid. rain. valid. s nd rd avg Table : 3-fold Cross Validaion, varying σ heshold, raining/validaion/es log. loss In Table, is repored based on bes se of (wih minimum validaion loss, fixed cerain σ heshold ). As he value of σ heshold increases, he keeps dropping, which accords wih expecaion since he Modified sigmoid funcion is more and more like he original one. Based on he resul, he bes model seems o be σ heshold =.5 plus corresponding rained Side-effec of modified sigmoid One ineresing propery of his modified sigmoid funcion is is insabiliy. The gradiens will jump back-and-forh from ieraion o ieraion. Consider he Figure 3 below.

13 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA 3.35 Norm of gradiens, Bounded Sigmoid func. (η =.).3.5 Norm of gradiens Training ieraions x 4 Figure 3: -Norm of gradien variaion of he modified sigmoid mehod This is due o he arificial modificaion o he original sigmoid funcion. In oher words, since he Modified Sigmoid funcion is no more coninuous he changes of could be dramaic herefore causing he unsable effec. (For he Sep-like ransfer funcion in Secion 5., I didn observe similar effec since he new ransfer funcion is sill coninuous.) I is also observed ha as σ heshold becomes bigger (he modified sigmoid funcion will approximae he original one hen) he -Norm of gradien ends o vary less. 5. Sep-like ransfer funcion Nex, I ried use a sep-like ransfer funcion as which looks as if w x σ heshold w x ŷ = +σ heshold σ oherwise heshold if w x σ heshold

14 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA 4 Sep like ransfer func., (σ heshold = 3) Original Sigmoid func Figure 4: Modified Sigmoid funcion as predicion (σ heshold = 3) 5.. Effecs in prevening overfiing This corresponding resuls of his change are shown in Figure 5 and 6..6 Sep like ransfer func., (η =., σ heshold = 3).6 Vanilla Log. Reg., Norm sopping crierion, (η =.) ^() ^( ) ^( ) ^( 3) gradien a sopping poins ^() ^( ) ^( ) Figure 5: Mean log. loss comparison, sep-like ransfer func. vs original sigmoid

15 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA 5 5 Sep like ransfer func., early sopping w/ Norm gradien 5 Original Sigmoid func., Norm of gradien early sopping, (η =.) ^() ^( ) ^( ) ^( 3) gradiens a sopping poins 5 ^() ^( ) ^( ) ^( 3) gradien a sopping poins Figure 6: Weighs comparison, sep-like ransfer func. vs original sigmoid This mehod helps o conrol o some exen. And consequenly, i also helps o reduce he effec of overfiing. 5.. Compared wih Modified Sigmoid Compared wih he Modified Sigmoid above by raining unil w < 4 yields he plos (Figure 7, 8) below. Bounded sigmoid predicion, (η =., σ heshold = 3) Sep like ransfer func., (η =., σ heshold = 3) ^() ^( ) ^( ) ^( 3) ^( 4) ^() ^( ) ^( ) ^( 3) ^( 4) Figure 7: Mean log. loss comparison 3, Modified sigmoid vs sep-like ransfer func.

16 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA 6 5 Bounded sigmoid predicion, early sopping by Norm gradien (η =., σ heshold = 3) 5 Sep like ransfer func., early sopping w/ Norm gradien ^() ^( ) ^( ) ^( 3) ^( 4) Norm Gradiens a sopping poins 5 ^() ^( ) ^( ) ^( 3) ^( 4) Gradiens a sopping poins Figure 8: Weighs comparison 3, Modified sigmoid vs sep-like ransfer func. As seen in his comparison, Modified Sigmoid performs beer if only rained unil w < 3 and he don diffuse. Ye, if rain more hese mehods performance are geing closer. They boh arificially conrol he by changing he predicion in he updae formula. As shown in he plo, all he will evenually be bounded in he range of [ σ heshold, σ heshold ] Varying σ heshold By varying he value of σ heshold, I ploed he curves of average logisic loss on boh raining and esing ses as shown in Figure 9. I is observed ha as σ heshold increases he increases as well while he drops.

17 5 MODIFYING PREDICTION FUNCTION IN UPDATE FORMULA Sep like ransfer func., varying σ heshold, (η =.) σ heshold Figure 9: The effecs of differen values of σ heshold Also, in Table he relevan daa is recorded. σ heshold loss rain. valid. rain. valid. rain. valid. rain. valid. rain. valid. s fold nd fold rd fold avg σ heshold loss rain. valid. rain. valid. rain. valid. rain. valid. rain. valid. s nd rd avg Table : 3-fold Cross Validaion, varying σ heshold, raining/validaion/es log. loss In Table, is repored based on bes se of (wih minimum validaion loss, fixed cerain σ heshold ). Based on he resul, he bes model seems o be σ heshold =.5 plus corresponding rained.

18 6 LABEL SHRINKING 8 6 Label Shrinking Based on [], I implemened he label shrinking echnique for prevening overfiing. The original idea in [] is more concerned wih overfiing resuled from adding sparse feaures while here I employed his echnique in he conex of logisic regression o observe he resul. 6. Label Shrinking As menioned above, for he spam daa se he labels are binary, y {, }.Shrinking will use a linear ransformaion mapping binary labels o a shrunk label se {a, b} in which a >, b <, as demonsraed below: y := a + (b a)y i.e. a, b. Consider he plos in Figure, in which one could see he effec of shrinking on he corresponding logisic loss..5 σ(ah) a=. loss(., ah) a=. loss(., ah) a=.4 loss(.4, ah) Label Shrinking effecs, a.5 Label Shrinking effecs, b σ(ah) b= loss(.99999, ah) b=.8 loss(.8, ah) b=.4 loss(.4, ah) y, σ(ah), loss(y, ah).5 y, σ(ah), loss(y, ah) ah ah Figure : Effec of Label Shrinking on logisic loss Firs, noe ha by shrinking he logisic loss becomes quadradic-shaped, ha is he loss curve won flaen ou if he predicion ŷ is close o or (equivalenly, he loss curve won flaen ou as he linear predicion â is relaively large or small). In implemenaion, I employed a single parameer ɛ o shrink labels, he label mapping relaion is y := ɛ + ( ɛ)y ɛ i.e.{, } {ɛ, ɛ}

19 6 LABEL SHRINKING Label Shrinking, (ε =.).6.4 Degeneraed Label Shrinking, (ε = ) ^() ^( ) ^( ). ^() ^( ) ^( ) Label Shrinking, (ε =.) ^() ^( ) ^( ) Figure : Mean log. loss comparison 4, Label Shrinking (ɛ =.,.) wih no shrinking I is observed in Figure ha label shrinking doesn help preven overfiing on his daa se, and he mean logisic loss is even worse han he Vanilla version of logisic regression. I proposed wo explanaions o his resul: () The performance of label shrinking echnique may highly relaed wih he characerisic of he daase. Therefore, i s hard o conclude his echnique won help in general case. Besides, he experimen I have done is based on original daase. Permuing he order of he daa may yield differen resul. () Theoreically, raining wih shrunk labels may cause he model minimizing he modified logisic loss funcion (as demonsraed in Figure ) which is differen from he original loss funcion. This may incur a higher loss han raining wihou shrinking. Ye, by ploing ou he change in Figure we could observe ha he vecor is indeed well conrolled as expeced. Moreover, as he ɛ increases he are more consrained.

20 7 CONCLUSION AND FUTURE WORK 3 Label Shrinking, (ε =.) 3 Degeneraed Label Shrinking, (ε = ) 3 3 ^() ^( ) ^( ) gradien a sopping poins 4 ^() ^( ) ^( ) gradien a sopping poins 3 Label Shrinking, (ε =.) 3 ^() ^( ) ^( ) gradien a sopping poins Figure : Weighs comparison 4, Label Shrinking (ɛ =.,.) wih no shrinking. Label shrinking could help conrol, ye he effec in prevening overfiing and benefiing is sill no confirmed. Moreover, due o ime limiaion he Predicion Sreching echnique is no fully esed wih reasonable resuls o presen. Fuure experimens are needed o invesigae hese echniques. 7 Conclusion and Fuure work To conclude, I proposed he poins below: () Overfiing effec is obvious in logisic regression based on gradien descen. Early sopping could help reduce his effec, ye i requires exra compuaion o decide he reasonable sopping poins as well as he sopping crierion o adop. -Norm/-Norm of gradien are he common sopping crieria. There is no essenial difference beween hese wo crieria. () Sochasic Gradien Descen wih Simulaed Annealing approach could help preven overfiing clearly due o he naure of his echnique. Ye, o apply his echnique in pracice, many parameers need o une. -Norm regularizaion is also effecive in prevening overfiing and i s more pracical. (3) Modifying ransfer funcion when raining is observed effecive in prevening overfiing by

21 8 ACKNOWLEDGMENT conrolling he. More exensive experimens are needed o jusify he usabiliy of his echnique in pracical problems. (4) Label Shrinking echnique is useful in conrolling. However, he resul here doesn show his echnique could benefi as in []. Fuure experimens on various daases as well as wih Predicion Sreching echnique should be conduced. 8 Acknowledgmen Thanks Maya Hrisakeva and Nikhila Arkalgud for heir excellen work on homework 3 abou Logisic Regression. I learned a lo from heir work. Thanks Bruno Asuo Arouche Nunes for eaching me he screen deaching rick on SSH clien, which faciliaed my experimens a grea deal. Special hanks o Professor Manfred Warmuh, for his encouragemen, guidance and pleny of help o me which suppors me working hrough his projec. References [] M. Warmuh, Shrink-srech of labels for regularizing logisic regression, 8.

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he

More information

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

Assignment 6. Tyler Shendruk December 6, 2010

Assignment 6. Tyler Shendruk December 6, 2010 Assignmen 6 Tyler Shendruk December 6, 1 1 Harden Problem 1 Le K be he coupling and h he exernal field in a 1D Ising model. From he lecures hese can be ransformed ino effecive coupling and fields K and

More information

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.

More information

The Rosenblatt s LMS algorithm for Perceptron (1958) is built around a linear neuron (a neuron with a linear

The Rosenblatt s LMS algorithm for Perceptron (1958) is built around a linear neuron (a neuron with a linear In The name of God Lecure4: Percepron and AALIE r. Majid MjidGhoshunih Inroducion The Rosenbla s LMS algorihm for Percepron 958 is buil around a linear neuron a neuron ih a linear acivaion funcion. Hoever,

More information

Random Walk with Anti-Correlated Steps

Random Walk with Anti-Correlated Steps Random Walk wih Ani-Correlaed Seps John Noga Dirk Wagner 2 Absrac We conjecure he expeced value of random walks wih ani-correlaed seps o be exacly. We suppor his conjecure wih 2 plausibiliy argumens and

More information

Dimitri Solomatine. D.P. Solomatine. Data-driven modelling (part 2). 2

Dimitri Solomatine. D.P. Solomatine. Data-driven modelling (part 2). 2 Daa-driven modelling. Par. Daa-driven Arificial di Neural modelling. Newors Par Dimiri Solomaine Arificial neural newors D.P. Solomaine. Daa-driven modelling par. 1 Arificial neural newors ANN: main pes

More information

Online Convex Optimization Example And Follow-The-Leader

Online Convex Optimization Example And Follow-The-Leader CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion

More information

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H.

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H. ACE 56 Fall 005 Lecure 5: he Simple Linear Regression Model: Sampling Properies of he Leas Squares Esimaors by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Inference in he Simple

More information

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8)

Econ107 Applied Econometrics Topic 7: Multicollinearity (Studenmund, Chapter 8) I. Definiions and Problems A. Perfec Mulicollineariy Econ7 Applied Economerics Topic 7: Mulicollineariy (Sudenmund, Chaper 8) Definiion: Perfec mulicollineariy exiss in a following K-variable regression

More information

STATE-SPACE MODELLING. A mass balance across the tank gives:

STATE-SPACE MODELLING. A mass balance across the tank gives: B. Lennox and N.F. Thornhill, 9, Sae Space Modelling, IChemE Process Managemen and Conrol Subjec Group Newsleer STE-SPACE MODELLING Inroducion: Over he pas decade or so here has been an ever increasing

More information

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class

More information

Sequential Importance Resampling (SIR) Particle Filter

Sequential Importance Resampling (SIR) Particle Filter Paricle Filers++ Pieer Abbeel UC Berkeley EECS Many slides adaped from Thrun, Burgard and Fox, Probabilisic Roboics 1. Algorihm paricle_filer( S -1, u, z ): 2. Sequenial Imporance Resampling (SIR) Paricle

More information

Vehicle Arrival Models : Headway

Vehicle Arrival Models : Headway Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where

More information

Chapter 7: Solving Trig Equations

Chapter 7: Solving Trig Equations Haberman MTH Secion I: The Trigonomeric Funcions Chaper 7: Solving Trig Equaions Le s sar by solving a couple of equaions ha involve he sine funcion EXAMPLE a: Solve he equaion sin( ) The inverse funcions

More information

Fishing limits and the Logistic Equation. 1

Fishing limits and the Logistic Equation. 1 Fishing limis and he Logisic Equaion. 1 1. The Logisic Equaion. The logisic equaion is an equaion governing populaion growh for populaions in an environmen wih a limied amoun of resources (for insance,

More information

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still. Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in

More information

Inventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions

Inventory Analysis and Management. Multi-Period Stochastic Models: Optimality of (s, S) Policy for K-Convex Objective Functions Muli-Period Sochasic Models: Opimali of (s, S) Polic for -Convex Objecive Funcions Consider a seing similar o he N-sage newsvendor problem excep ha now here is a fixed re-ordering cos (> 0) for each (re-)order.

More information

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal

More information

Let us start with a two dimensional case. We consider a vector ( x,

Let us start with a two dimensional case. We consider a vector ( x, Roaion marices We consider now roaion marices in wo and hree dimensions. We sar wih wo dimensions since wo dimensions are easier han hree o undersand, and one dimension is a lile oo simple. However, our

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Linear Response Theory: The connecion beween QFT and experimens 3.1. Basic conceps and ideas Q: How do we measure he conduciviy of a meal? A: we firs inroduce a weak elecric field E, and

More information

20. Applications of the Genetic-Drift Model

20. Applications of the Genetic-Drift Model 0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0

More information

d 1 = c 1 b 2 - b 1 c 2 d 2 = c 1 b 3 - b 1 c 3

d 1 = c 1 b 2 - b 1 c 2 d 2 = c 1 b 3 - b 1 c 3 and d = c b - b c c d = c b - b c c This process is coninued unil he nh row has been compleed. The complee array of coefficiens is riangular. Noe ha in developing he array an enire row may be divided or

More information

UNC resolution Uncertainty Learning Objectives: measurement interval ( You will turn in two worksheets and

UNC resolution Uncertainty Learning Objectives: measurement interval ( You will turn in two worksheets and UNC Uncerainy revised Augus 30, 017 Learning Objecives: During his lab, you will learn how o 1. esimae he uncerainy in a direcly measured quaniy.. esimae he uncerainy in a quaniy ha is calculaed from quaniies

More information

An introduction to the theory of SDDP algorithm

An introduction to the theory of SDDP algorithm An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Noes for EE7C Spring 018: Convex Opimizaion and Approximaion Insrucor: Moriz Hard Email: hard+ee7c@berkeley.edu Graduae Insrucor: Max Simchowiz Email: msimchow+ee7c@berkeley.edu Ocober 15, 018 3

More information

Learning Objectives: Practice designing and simulating digital circuits including flip flops Experience state machine design procedure

Learning Objectives: Practice designing and simulating digital circuits including flip flops Experience state machine design procedure Lab 4: Synchronous Sae Machine Design Summary: Design and implemen synchronous sae machine circuis and es hem wih simulaions in Cadence Viruoso. Learning Objecives: Pracice designing and simulaing digial

More information

Two Coupled Oscillators / Normal Modes

Two Coupled Oscillators / Normal Modes Lecure 3 Phys 3750 Two Coupled Oscillaors / Normal Modes Overview and Moivaion: Today we ake a small, bu significan, sep owards wave moion. We will no ye observe waves, bu his sep is imporan in is own

More information

Deep Learning: Theory, Techniques & Applications - Recurrent Neural Networks -

Deep Learning: Theory, Techniques & Applications - Recurrent Neural Networks - Deep Learning: Theory, Techniques & Applicaions - Recurren Neural Neworks - Prof. Maeo Maeucci maeo.maeucci@polimi.i Deparmen of Elecronics, Informaion and Bioengineering Arificial Inelligence and Roboics

More information

Lecture 9: September 25

Lecture 9: September 25 0-725: Opimizaion Fall 202 Lecure 9: Sepember 25 Lecurer: Geoff Gordon/Ryan Tibshirani Scribes: Xuezhi Wang, Subhodeep Moira, Abhimanu Kumar Noe: LaTeX emplae couresy of UC Berkeley EECS dep. Disclaimer:

More information

Technical Report Doc ID: TR March-2013 (Last revision: 23-February-2016) On formulating quadratic functions in optimization models.

Technical Report Doc ID: TR March-2013 (Last revision: 23-February-2016) On formulating quadratic functions in optimization models. Technical Repor Doc ID: TR--203 06-March-203 (Las revision: 23-Februar-206) On formulaing quadraic funcions in opimizaion models. Auhor: Erling D. Andersen Convex quadraic consrains quie frequenl appear

More information

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature On Measuring Pro-Poor Growh 1. On Various Ways of Measuring Pro-Poor Growh: A Shor eview of he Lieraure During he pas en years or so here have been various suggesions concerning he way one should check

More information

Lecture 33: November 29

Lecture 33: November 29 36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure

More information

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,

More information

Bias-Variance Error Bounds for Temporal Difference Updates

Bias-Variance Error Bounds for Temporal Difference Updates Bias-Variance Bounds for Temporal Difference Updaes Michael Kearns AT&T Labs mkearns@research.a.com Sainder Singh AT&T Labs baveja@research.a.com Absrac We give he firs rigorous upper bounds on he error

More information

EXERCISES FOR SECTION 1.5

EXERCISES FOR SECTION 1.5 1.5 Exisence and Uniqueness of Soluions 43 20. 1 v c 21. 1 v c 1 2 4 6 8 10 1 2 2 4 6 8 10 Graph of approximae soluion obained using Euler s mehod wih = 0.1. Graph of approximae soluion obained using Euler

More information

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence Supplemen for Sochasic Convex Opimizaion: Faser Local Growh Implies Faser Global Convergence Yi Xu Qihang Lin ianbao Yang Proof of heorem heorem Suppose Assumpion holds and F (w) obeys he LGC (6) Given

More information

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Kriging Models Predicing Arazine Concenraions in Surface Waer Draining Agriculural Waersheds Paul L. Mosquin, Jeremy Aldworh, Wenlin Chen Supplemenal Maerial Number

More information

Západočeská Univerzita v Plzni, Czech Republic and Groupe ESIEE Paris, France

Západočeská Univerzita v Plzni, Czech Republic and Groupe ESIEE Paris, France ADAPTIVE SIGNAL PROCESSING USING MAXIMUM ENTROPY ON THE MEAN METHOD AND MONTE CARLO ANALYSIS Pavla Holejšovsá, Ing. *), Z. Peroua, Ing. **), J.-F. Bercher, Prof. Assis. ***) Západočesá Univerzia v Plzni,

More information

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LDA, logisic

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LTU, decision

More information

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time.

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time. Supplemenary Figure 1 Spike-coun auocorrelaions in ime. Normalized auocorrelaion marices are shown for each area in a daase. The marix shows he mean correlaion of he spike coun in each ime bin wih he spike

More information

Solutions to Odd Number Exercises in Chapter 6

Solutions to Odd Number Exercises in Chapter 6 1 Soluions o Odd Number Exercises in 6.1 R y eˆ 1.7151 y 6.3 From eˆ ( T K) ˆ R 1 1 SST SST SST (1 R ) 55.36(1.7911) we have, ˆ 6.414 T K ( ) 6.5 y ye ye y e 1 1 Consider he erms e and xe b b x e y e b

More information

GMM - Generalized Method of Moments

GMM - Generalized Method of Moments GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................

More information

Longest Common Prefixes

Longest Common Prefixes Longes Common Prefixes The sandard ordering for srings is he lexicographical order. I is induced by an order over he alphabe. We will use he same symbols (,

More information

Chapter 2. First Order Scalar Equations

Chapter 2. First Order Scalar Equations Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.

More information

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes Represening Periodic Funcions by Fourier Series 3. Inroducion In his Secion we show how a periodic funcion can be expressed as a series of sines and cosines. We begin by obaining some sandard inegrals

More information

A variational radial basis function approximation for diffusion processes.

A variational radial basis function approximation for diffusion processes. A variaional radial basis funcion approximaion for diffusion processes. Michail D. Vreas, Dan Cornford and Yuan Shen {vreasm, d.cornford, y.shen}@ason.ac.uk Ason Universiy, Birmingham, UK hp://www.ncrg.ason.ac.uk

More information

A Forward-Backward Splitting Method with Component-wise Lazy Evaluation for Online Structured Convex Optimization

A Forward-Backward Splitting Method with Component-wise Lazy Evaluation for Online Structured Convex Optimization A Forward-Backward Spliing Mehod wih Componen-wise Lazy Evaluaion for Online Srucured Convex Opimizaion Yukihiro Togari and Nobuo Yamashia March 28, 2016 Absrac: We consider large-scale opimizaion problems

More information

Seminar 4: Hotelling 2

Seminar 4: Hotelling 2 Seminar 4: Hoelling 2 November 3, 211 1 Exercise Par 1 Iso-elasic demand A non renewable resource of a known sock S can be exraced a zero cos. Demand for he resource is of he form: D(p ) = p ε ε > A a

More information

Integration Over Manifolds with Variable Coordinate Density

Integration Over Manifolds with Variable Coordinate Density Inegraion Over Manifolds wih Variable Coordinae Densiy Absrac Chrisopher A. Lafore clafore@gmail.com In his paper, he inegraion of a funcion over a curved manifold is examined in he case where he curvaure

More information

CSCE 496/896 Lecture 2: Basic Artificial Neural Networks. Stephen Scott. Introduction. Supervised Learning. Basic Units.

CSCE 496/896 Lecture 2: Basic Artificial Neural Networks. Stephen Scott. Introduction. Supervised Learning. Basic Units. (Adaped from Vinod Variyam, Ehem Alpaydin, Tom Michell, Ian Goodfellow, and Aurélien Géron) learning is mos fundamenal, classic form machine learning par comes from he par labels for examples (insances)

More information

1.1. Example: Polynomial Curve Fitting 4 1. INTRODUCTION

1.1. Example: Polynomial Curve Fitting 4 1. INTRODUCTION 4. INTRODUCTION Figure.2 Plo of a raining daa se of N = poins, shown as blue circles, each comprising an observaion of he inpu variable along wih he corresponding arge variable. The green curve shows he

More information

CSE/NB 528 Lecture 14: From Supervised to Reinforcement Learning (Chapter 9) R. Rao, 528: Lecture 14

CSE/NB 528 Lecture 14: From Supervised to Reinforcement Learning (Chapter 9) R. Rao, 528: Lecture 14 CSE/NB 58 Lecure 14: From Supervised o Reinforcemen Learning Chaper 9 1 Recall from las ime: Sigmoid Neworks Oupu v T g w u g wiui w Inpu nodes u = u 1 u u 3 T i Sigmoid oupu funcion: 1 g a 1 a e 1 ga

More information

On Boundedness of Q-Learning Iterates for Stochastic Shortest Path Problems

On Boundedness of Q-Learning Iterates for Stochastic Shortest Path Problems MATHEMATICS OF OPERATIONS RESEARCH Vol. 38, No. 2, May 2013, pp. 209 227 ISSN 0364-765X (prin) ISSN 1526-5471 (online) hp://dx.doi.org/10.1287/moor.1120.0562 2013 INFORMS On Boundedness of Q-Learning Ieraes

More information

An recursive analytical technique to estimate time dependent physical parameters in the presence of noise processes

An recursive analytical technique to estimate time dependent physical parameters in the presence of noise processes WHAT IS A KALMAN FILTER An recursive analyical echnique o esimae ime dependen physical parameers in he presence of noise processes Example of a ime and frequency applicaion: Offse beween wo clocks PREDICTORS,

More information

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates

Biol. 356 Lab 8. Mortality, Recruitment, and Migration Rates Biol. 356 Lab 8. Moraliy, Recruimen, and Migraion Raes (modified from Cox, 00, General Ecology Lab Manual, McGraw Hill) Las week we esimaed populaion size hrough several mehods. One assumpion of all hese

More information

Notes for Lecture 17-18

Notes for Lecture 17-18 U.C. Berkeley CS278: Compuaional Complexiy Handou N7-8 Professor Luca Trevisan April 3-8, 2008 Noes for Lecure 7-8 In hese wo lecures we prove he firs half of he PCP Theorem, he Amplificaion Lemma, up

More information

NCSS Statistical Software. , contains a periodic (cyclic) component. A natural model of the periodic component would be

NCSS Statistical Software. , contains a periodic (cyclic) component. A natural model of the periodic component would be NCSS Saisical Sofware Chaper 468 Specral Analysis Inroducion This program calculaes and displays he periodogram and specrum of a ime series. This is someimes nown as harmonic analysis or he frequency approach

More information

Final Spring 2007

Final Spring 2007 .615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o

More information

- If one knows that a magnetic field has a symmetry, one may calculate the magnitude of B by use of Ampere s law: The integral of scalar product

- If one knows that a magnetic field has a symmetry, one may calculate the magnitude of B by use of Ampere s law: The integral of scalar product 11.1 APPCATON OF AMPEE S AW N SYMMETC MAGNETC FEDS - f one knows ha a magneic field has a symmery, one may calculae he magniude of by use of Ampere s law: The inegral of scalar produc Closed _ pah * d

More information

Solutions Problem Set 3 Macro II (14.452)

Solutions Problem Set 3 Macro II (14.452) Soluions Problem Se 3 Macro II (14.452) Francisco A. Gallego 04/27/2005 1 Q heory of invesmen in coninuous ime and no uncerainy Consider he in nie horizon model of a rm facing adjusmen coss o invesmen.

More information

Inventory Control of Perishable Items in a Two-Echelon Supply Chain

Inventory Control of Perishable Items in a Two-Echelon Supply Chain Journal of Indusrial Engineering, Universiy of ehran, Special Issue,, PP. 69-77 69 Invenory Conrol of Perishable Iems in a wo-echelon Supply Chain Fariborz Jolai *, Elmira Gheisariha and Farnaz Nojavan

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning

More information

Problem Set 5. Graduate Macro II, Spring 2017 The University of Notre Dame Professor Sims

Problem Set 5. Graduate Macro II, Spring 2017 The University of Notre Dame Professor Sims Problem Se 5 Graduae Macro II, Spring 2017 The Universiy of Nore Dame Professor Sims Insrucions: You may consul wih oher members of he class, bu please make sure o urn in your own work. Where applicable,

More information

CSE/NB 528 Lecture 14: Reinforcement Learning (Chapter 9)

CSE/NB 528 Lecture 14: Reinforcement Learning (Chapter 9) CSE/NB 528 Lecure 14: Reinforcemen Learning Chaper 9 Image from hp://clasdean.la.asu.edu/news/images/ubep2001/neuron3.jpg Lecure figures are from Dayan & Abbo s book hp://people.brandeis.edu/~abbo/book/index.hml

More information

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi Creep in Viscoelasic Subsances Numerical mehods o calculae he coefficiens of he Prony equaion using creep es daa and Herediary Inegrals Mehod Navnee Saini, Mayank Goyal, Vishal Bansal (23); Term Projec

More information

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1

RL Lecture 7: Eligibility Traces. R. S. Sutton and A. G. Barto: Reinforcement Learning: An Introduction 1 RL Lecure 7: Eligibiliy Traces R. S. Suon and A. G. Baro: Reinforcemen Learning: An Inroducion 1 N-sep TD Predicion Idea: Look farher ino he fuure when you do TD backup (1, 2, 3,, n seps) R. S. Suon and

More information

New Boosting Methods of Gaussian Processes for Regression

New Boosting Methods of Gaussian Processes for Regression New Boosing Mehods of Gaussian Processes for Regression Yangqiu Song Sae Key Laboraory of Inelligen echnology and Sysems Deparmen of Auomaion singhua Universiy Beijing 00084, P.R.China E-mail: songyq99@mails.singhua.edu.cn

More information

SOLUTIONS TO ECE 3084

SOLUTIONS TO ECE 3084 SOLUTIONS TO ECE 384 PROBLEM 2.. For each sysem below, specify wheher or no i is: (i) memoryless; (ii) causal; (iii) inverible; (iv) linear; (v) ime invarian; Explain your reasoning. If he propery is no

More information

Energy Storage Benchmark Problems

Energy Storage Benchmark Problems Energy Sorage Benchmark Problems Daniel F. Salas 1,3, Warren B. Powell 2,3 1 Deparmen of Chemical & Biological Engineering 2 Deparmen of Operaions Research & Financial Engineering 3 Princeon Laboraory

More information

ES.1803 Topic 22 Notes Jeremy Orloff

ES.1803 Topic 22 Notes Jeremy Orloff ES.83 Topic Noes Jeremy Orloff Fourier series inroducion: coninued. Goals. Be able o compue he Fourier coefficiens of even or odd periodic funcion using he simplified formulas.. Be able o wrie and graph

More information

Guest Lectures for Dr. MacFarlane s EE3350 Part Deux

Guest Lectures for Dr. MacFarlane s EE3350 Part Deux Gues Lecures for Dr. MacFarlane s EE3350 Par Deux Michael Plane Mon., 08-30-2010 Wrie name in corner. Poin ou his is a review, so I will go faser. Remind hem o go lisen o online lecure abou geing an A

More information

RC, RL and RLC circuits

RC, RL and RLC circuits Name Dae Time o Complee h m Parner Course/ Secion / Grade RC, RL and RLC circuis Inroducion In his experimen we will invesigae he behavior of circuis conaining combinaions of resisors, capaciors, and inducors.

More information

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

Online Appendix to Solution Methods for Models with Rare Disasters

Online Appendix to Solution Methods for Models with Rare Disasters Online Appendix o Soluion Mehods for Models wih Rare Disasers Jesús Fernández-Villaverde and Oren Levinal In his Online Appendix, we presen he Euler condiions of he model, we develop he pricing Calvo block,

More information

FITTING OF A PARTIALLY REPARAMETERIZED GOMPERTZ MODEL TO BROILER DATA

FITTING OF A PARTIALLY REPARAMETERIZED GOMPERTZ MODEL TO BROILER DATA FITTING OF A PARTIALLY REPARAMETERIZED GOMPERTZ MODEL TO BROILER DATA N. Okendro Singh Associae Professor (Ag. Sa.), College of Agriculure, Cenral Agriculural Universiy, Iroisemba 795 004, Imphal, Manipur

More information

A First Course on Kinetics and Reaction Engineering. Class 19 on Unit 18

A First Course on Kinetics and Reaction Engineering. Class 19 on Unit 18 A Firs ourse on Kineics and Reacion Engineering lass 19 on Uni 18 Par I - hemical Reacions Par II - hemical Reacion Kineics Where We re Going Par III - hemical Reacion Engineering A. Ideal Reacors B. Perfecly

More information

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing Applicaion of a Sochasic-Fuzzy Approach o Modeling Opimal Discree Time Dynamical Sysems by Using Large Scale Daa Processing AA WALASZE-BABISZEWSA Deparmen of Compuer Engineering Opole Universiy of Technology

More information

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis Speaker Adapaion Techniques For Coninuous Speech Using Medium and Small Adapaion Daa Ses Consaninos Boulis Ouline of he Presenaion Inroducion o he speaker adapaion problem Maximum Likelihood Sochasic Transformaions

More information

( ) is the stretch factor, and x the

( ) is the stretch factor, and x the (Lecures 7-8) Liddle, Chaper 5 Simple cosmological models (i) Hubble s Law revisied Self-similar srech of he universe All universe models have his characerisic v r ; v = Hr since only his conserves homogeneiy

More information

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits DOI: 0.545/mjis.07.5009 Exponenial Weighed Moving Average (EWMA) Char Under The Assumpion of Moderaeness And Is 3 Conrol Limis KALPESH S TAILOR Assisan Professor, Deparmen of Saisics, M. K. Bhavnagar Universiy,

More information

KINEMATICS IN ONE DIMENSION

KINEMATICS IN ONE DIMENSION KINEMATICS IN ONE DIMENSION PREVIEW Kinemaics is he sudy of how hings move how far (disance and displacemen), how fas (speed and velociy), and how fas ha how fas changes (acceleraion). We say ha an objec

More information

Isolated-word speech recognition using hidden Markov models

Isolated-word speech recognition using hidden Markov models Isolaed-word speech recogniion using hidden Markov models Håkon Sandsmark December 18, 21 1 Inroducion Speech recogniion is a challenging problem on which much work has been done he las decades. Some of

More information

Class Meeting # 10: Introduction to the Wave Equation

Class Meeting # 10: Introduction to the Wave Equation MATH 8.5 COURSE NOTES - CLASS MEETING # 0 8.5 Inroducion o PDEs, Fall 0 Professor: Jared Speck Class Meeing # 0: Inroducion o he Wave Equaion. Wha is he wave equaion? The sandard wave equaion for a funcion

More information

3.1 More on model selection

3.1 More on model selection 3. More on Model selecion 3. Comparing models AIC, BIC, Adjused R squared. 3. Over Fiing problem. 3.3 Sample spliing. 3. More on model selecion crieria Ofen afer model fiing you are lef wih a handful of

More information

CHEAPEST PMT ONLINE TEST SERIES AIIMS/NEET TOPPER PREPARE QUESTIONS

CHEAPEST PMT ONLINE TEST SERIES AIIMS/NEET TOPPER PREPARE QUESTIONS CHEAPEST PMT ONLINE TEST SERIES AIIMS/NEET TOPPER PREPARE QUESTIONS For more deails see las page or conac @aimaiims.in Physics Mock Tes Paper AIIMS/NEET 07 Physics 06 Saurday Augus 0 Uni es : Moion in

More information

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17 EES 16A Designing Informaion Devices and Sysems I Spring 019 Lecure Noes Noe 17 17.1 apaciive ouchscreen In he las noe, we saw ha a capacior consiss of wo pieces on conducive maerial separaed by a nonconducive

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H.

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H. ACE 56 Fall 5 Lecure 8: The Simple Linear Regression Model: R, Reporing he Resuls and Predicion by Professor Sco H. Irwin Required Readings: Griffihs, Hill and Judge. "Explaining Variaion in he Dependen

More information

A Dynamic Model of Economic Fluctuations

A Dynamic Model of Economic Fluctuations CHAPTER 15 A Dynamic Model of Economic Flucuaions Modified for ECON 2204 by Bob Murphy 2016 Worh Publishers, all righs reserved IN THIS CHAPTER, OU WILL LEARN: how o incorporae dynamics ino he AD-AS model

More information

SPH3U: Projectiles. Recorder: Manager: Speaker:

SPH3U: Projectiles. Recorder: Manager: Speaker: SPH3U: Projeciles Now i s ime o use our new skills o analyze he moion of a golf ball ha was ossed hrough he air. Le s find ou wha is special abou he moion of a projecile. Recorder: Manager: Speaker: 0

More information

Rapid Termination Evaluation for Recursive Subdivision of Bezier Curves

Rapid Termination Evaluation for Recursive Subdivision of Bezier Curves Rapid Terminaion Evaluaion for Recursive Subdivision of Bezier Curves Thomas F. Hain School of Compuer and Informaion Sciences, Universiy of Souh Alabama, Mobile, AL, U.S.A. Absrac Bézier curve flaening

More information

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015 Explaining Toal Facor Produciviy Ulrich Kohli Universiy of Geneva December 2015 Needed: A Theory of Toal Facor Produciviy Edward C. Presco (1998) 2 1. Inroducion Toal Facor Produciviy (TFP) has become

More information

Unit Root Time Series. Univariate random walk

Unit Root Time Series. Univariate random walk Uni Roo ime Series Univariae random walk Consider he regression y y where ~ iid N 0, he leas squares esimae of is: ˆ yy y y yy Now wha if = If y y hen le y 0 =0 so ha y j j If ~ iid N 0, hen y ~ N 0, he

More information

Some Basic Information about M-S-D Systems

Some Basic Information about M-S-D Systems Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,

More information