Ensemble Confidence Estimates Posterior Probability

Size: px
Start display at page:

Download "Ensemble Confidence Estimates Posterior Probability"

Transcription

1 Ensemble Esimaes Poserior Probabiliy Michael Muhlbaier, Aposolos Topalis, and Robi Polikar Rowan Universiy, Elecrical and Compuer Engineering, Mullica Hill Rd., Glassboro, NJ 88, USA {muhlba6, Absrac. We have previously inroduced he Learn ++ algorihm ha provides surprisingly promising performance for incremenal learning as well as daa fusion applicaions. In his conribuion we show ha he algorihm can also be used o esimae he poserior probabiliy, or he confidence of is decision on each es insance. On hree increasingly difficul ess ha are specifically designed o compare poserior probabiliy esimaes of he algorihm o ha of he opimal Bayes classifier, we have observed ha esimaed poserior probabiliy approaches o ha of he Bayes classifier as he number of classifiers in he ensemble increase. This saisfying and inuiively expeced oucome shows ha ensemble sysems can also be used o esimae confidence of heir oupu. Inroducion Ensemble / muliple classifier sysems have enjoyed increasing aenion and populariy over he las decade due o heir favorable performances and/or oher advanages over single classifier based sysems. In paricular, ensemble based sysems have been shown, among oher hings, o successfully generae srong classifiers from weak classifiers, resis over-fiing problems [, ], provide an inuiive srucure for daa fusion [-4], as well as incremenal learning problems [5]. One area ha has received somewha less of an aenion, however, is he confidence esimaion poenial of such sysems. Due o heir very characer of generaing muliple classifiers for a given daabase, ensemble sysems provide a naural seing for esimaing he confidence of he classificaion sysem on is generalizaion performance. In his conribuion, we show how our previously inroduced algorihm Learn ++ [5], inspired by AdaBoos bu specifically modified for incremenal learning applicaions, can also be used o deermine is own confidence on any given specific es daa insance. We esimae he poserior probabiliy of he class chosen by he ensemble using a weighed sofmax approach, and use ha esimae as he confidence measure. We empirically show on hree increasingly difficul daases ha as addiional classifiers are added o he ensemble, he poserior probabiliy of he class chosen by he ensemble approaches o ha of he opimal Bayes classifier. I is imporan o noe ha he mehod of ensemble confidence esimaion being proposed is no specific o Learn ++, bu can be applied o any ensemble based sysem. N.C. Oza e al. (Eds.): MCS 5, LNCS 354, pp , 5. Springer-Verlag Berlin Heidelberg 5

2 Ensemble Esimaes Poserior Probabiliy 37 Learn ++ In ensemble approaches using a voing mechanism o combine classifier oupus, he individual classifiers voe on he class hey predic. The final classificaion is hen deermined as he class ha receives he highes oal voe from all classifiers. Learn ++ uses weighed majoriy voing, a raher non-democraic voing scheme, where each classifier receives a voing weigh based on is raining performance. One novely of he Learn ++ algorihm is is abiliy o incremenally learn from newly inroduced daa. For breviy, his feaure of he algorihm is no discussed here and ineresed readers are referred o [4,5]. Insead, we briefly explain he algorihm and discuss how i can be used o deermine is confidence as an esimae of he poserior probabiliy on classifying es daa. For each daase (D k ) ha consecuively becomes available o Learn ++, he inpus o he algorihm are (i) a sequence of m raining daa insances x k,i along wih heir correc labels y i, (ii) a classificaion algorihm BaseClassifier, and (iii) an ineger T k specifying he maximum number of classifiers o be generaed using ha daabase. If he algorihm is seeing is firs daabase (k=), a daa disribuion (D ) from which raining insances will be drawn - is iniialized o be uniform, making he probabiliy of any insance being seleced equal. If k>, hen a disribuion iniializaion sequence, iniializes he daa disribuion. The algorihm hen adds T k classifiers o he ensemble saring a =et k + where et k denoes he number of classifiers ha currenly exis in he ensemble. The pseudocode of he algorihm is given in Figure. For each ieraion, he insance weighs, w, from he previous ieraion are firs normalized (sep ) o creae a weigh disribuion D. A hypohesis, h, is generaed using a subse of D k drawn from D (sep ). The error, ε, of h is calculaed: if ε > ½, he algorihm deems he curren classifier h o be oo weak, discards i, and reurns o sep ; oherwise, calculaes he normalized error β (sep 3). The weighed majoriy voing algorihm is called o obain he composie hypohesis, H, of he ensemble (sep 4). H represens he ensemble decision of he firs hypoheses generaed hus far. The error E of H is hen compued and normalized (sep 5). The insance weighs w are finally updaed according o he performance of H (sep 6), such ha he weighs of insances correcly classified by H are reduced and hose ha are misclassified are effecively increased. This ensures ha he ensemble focus on hose regions of he feaure space ha are ye o be learned. We noe ha H allows Learn ++ o make is disribuion updae based on he ensemble decision, as opposed o AdaBoos which makes is updae based on he curren hypohesis h. 3 as an Esimae of Poserior Probabiliy In applicaions where he daa disribuion is known, an opimal Bayes classifier can be used for which he poserior probabiliy of he chosen class can be calculaed; a quaniy which can hen be inerpreed as a measure of confidence [6]. The poserior probabiliy of class ω j given insance x is classically defined using he Bayes rule as:

3 38 M. Muhlbaier, A. Topalis, and R. Polikar Inpu: For each daase D k k=,,,k Sequence of i=,,m k insances x k,i wih labels y Y {,..., c} i k = Weak learning algorihm BaseClassifier. Ineger T k, specifying he number of ieraions. Do for k=,,,k If k= Iniialize w = D( i) = / m, et = for all i. Else Go o Sep 5 o evaluae he curren ensemble on new daase D k, updae weighs, and recall curren number of classifiers! j= Do for = et k +, et k +,, et k + Tk : m. Se D = w i= w ( i) so ha D is a disribuion. et = k k T j. Call BaseClassifier wih a subse of D k randomly chosen using D. 3. Obain h : X! Y, and calculae is error: ε = D() i ih : ( x ) y i i If ε > ½, discard h and go o sep. Oherwise, compue normalized error as β = ε ε ). ( 4. Call weighed majoriy voing o obain he composie hypohesis H = arg max log ( β) y Y h : ( xi) = y i 5. Compue he error of he composie hypohesis E = D() i ih : ( xi) yi 6. Se B =E /(-E ), <B <, and updae he insance weighs: B, if H( xi) = yi D+ () i = D, oherwise Call weighed majoriy voing o obain he final hypohesis. H final K arg max log y Y k = : h ( xi) = y i = ( β ) Fig.. Learn ++ Algorihm P( x ωj) P( ωj) P( ω j x) = N P( x ω ) ( ) k k P ω = k Since class disribuions are rarely known in pracice, poserior probabiliies mus be esimaed. While here are several echniques for densiy esimaion [7], such echniques are difficul o apply for large dimensional problems. A mehod ha can ()

4 Ensemble Esimaes Poserior Probabiliy 39 esimae he Bayesian poserior probabiliy would herefore prove o be a mos valuable ool in evaluaing classifier performance. Several mehods have been proposed for his purpose [6-9]. One example is he sofmax model [8], commonly used wih classifiers whose oupus are binary encoded, as such oupus can be mapped ino an esimae of he poserior class probabiliy using Aj ( x) e P( ω x) C ( x ) = () j j N Ak ( x) e k= where A j (x) represens he oupu for class j, and N is he number of classes. C j (x) is hen he confidence of he classifier in predicing class ω j for insance x, which is an esimae of he poserior probabiliy P(ω j x). The sofmax funcion essenially akes he exponenial of he oupu and normalizes i o [ ] range by summing over he exponenials of all oupus. This model is generally believed o provide good esimaes if he classifier is well rained using sufficienly dense raining daa. In an effor o generae a measure of confidence for an ensemble of classifiers in general, and for Learn ++ in paricular, we expand he sofmax concep by using he individual classifier weighs in place of a single exper s oupu. The ensemble confidence, esimaing he poserior probabiliy, can herefore be calculaed as: where Fj ( x) e P( ω x) C ( x ) = (3) j j N Fk ( x) e k= ( β ) N log h( x) = ω j Fj ( x ) = = oherwise (4) The confidence, C j (x), associaed wih class ω j for insance x is herefore he exponenial of he sum of classifier weighs ha seleced class ω j, divided by he sum of he aforemenioned exponenials corresponding o each class. The significance of his confidence esimaion scheme is in is consideraion of he diversiy in he classifier decisions: in calculaing he confidence of class ω j, he confidence will increase if he classifiers ha did no choose class ω j have varying decisions as opposed o having a common decision, ha is, if he evidence agains class ω j is no srong. On he oher hand, he confidence will decrease if he classifiers ha did no choose class ω j have a common decision, ha is, here is srong evidence agains class ω j. 4 Simulaion Resuls In order o find ou if and how well he Learn ++ ensemble confidence approximaes he Bayesian poserior probabiliy, he modified sofmax approach was analyzed on hree increasingly difficul problems. In order o calculae he heoreical Bayesian poserior probabiliies, and hence compare he Learn ++ confidences o hose of Bayes-

5 33 M. Muhlbaier, A. Topalis, and R. Polikar ian probabiliies, experimenal daa were generaed from Gaussian disribuion. For raining, random insances were seleced from each class disribuion, using which an ensemble of 3 MLP classifiers were generaed wih Learn ++. The daa and classifier generaion process was hen repeaed and averaged imes wih randomly seleced daa o ensure generaliy. For each simulaion, we also benchmark he resuls by calculaing a mean square error beween Learn ++ and Bayes confidences over he enire grid of he feaure space, wih each added classifier o he ensemble. 4. Experimen A wo feaure, hree class problem, where each class has a known Gaussian disribuion is seen in Fig.. In his experimen class,, and 3 have a variance of.5 and are cenered a [-, ], [, ], and [, -], respecively. Since he disribuion is known (and is Gaussian), he acual poserior probabiliy can be calculaed from Equaion, given he known likelihood P(x ω j ) ha can be calculaed as ( π ) d / / Σ j T ( x µ j) Σ j ( x µ j) P( x ω ) = e (5) j where d is he dimensionaliy, and µ j and Σ j are he mean and he covariance marix of he disribuion from which j h class daa are generaed. Each class was equally likely, hence P(ω j )=/3. For each insance, over he enire grid of he feaure space shown in Fig., we calculaed he poserior probabiliy of he class chosen by he Bayes classifier, and ploed hem as a confidence surface, as shown in Fig.3a. Calculaing he confidences of Learn ++ decisions on he same feaure space provided he plo in Fig 3b, indicaing ha he ensemble confidence surface closely approximaes ha of he Bayes classifier. Densiy y Feaure Feaure Fig.. Daa disribuions used in Experimen

6 Ensemble Esimaes Poserior Probabiliy Feaure Feaure Feaure Feaure Fig. 3. (a) Bayesian and (b) Learn ++ confidence surface for Experimen I is ineresing o noe ha he confidences in boh cases plumme around he decision boundaries and approach away from he decision boundary, an oucome ha makes inuiive sense. To quaniaively deermine how closely he Learn ++ confidence approximaes ha of Bayes classifier, and how his approximaion changes wih each addiional classifier, he mean squared error (MSE) was calculaed beween he ideal Bayesian confidence surface and he Learn ++ confidence over he enire grid of he feaure space - for each addiional classifier added o he ensemble. As seen in Fig.4, MSE beween he wo decreases as new classifiers are added o he ensemble, an expeced, bu neverheless immensely saisfying oucome. Furhermore, he decrease in he error is exponenial and raher monoonic, and does no appear o indicae any over-fiing, a leas for as many as 3 classifiers added o he ensemble. The ensemble confidence was hen compared o ha of a single MLP classifier, where he confidence was calculaed using he MLP s raw oupu values. The mean squared error was calculaed beween he resuling confidence and he Bayesian confidence and has been ploed as a doed line in Fig. 4 in comparison o he Learn ++ confidence. The single MLP differs from classifiers generaed using he Learn ++ algorihm on wo accouns. Firs, he single MLP is rained using all of he raining daa where each classifier in he Learn ++ ensemble is rained on /3 of he raining daa. Also, Learn ++ confidence is based on he discree decision of each classifier. If here were only one classifier in he ensemble, all classifiers would agree resuling in a confidence of. Therefore, confidence of a single MLP can only be calculaed based on he (sofmax normalized) acual oupu values unlike Learn ++ which uses a weighed voe of he discree oupu labels. 4. Experimen To furher characerize he behavior of his confidence esimaion scheme, Experimen was repeaed by increasing he variances of he class disribuions from.5 o.75, resuling in a more overlapping disribuion (Fig. 5) and a ougher classificaion problem. Learn ++ was rained wih daa generaed from his disribuion, is confidence calculaed over he enire grid of he feaure space and ploed in comparison o ha of Bayes classifier in Fig. 6. We noe ha low confidence valleys around he decision boundaries are wider in his case, an expeced oucome of he increased variance.

7 33 M. Muhlbaier, A. Topalis, and R. Polikar.3.3 Mean Squared Error Number of Classifiers Fig. 4. Mean square error as a funcion of number of classifiers - Experimen Densiy.4. Feaure Feaure Fig. 5. Daa disribuions used in Experimen Feaure Feaure Feaure Feaure Fig. 6. (a) Bayesian and (b) Learn ++ confidence surface for Experimen

8 Ensemble Esimaes Poserior Probabiliy Mean Squared Error Number of Classifiers Fig. 7. Mean square error as a funcion of number of classifiers - Experimen Fig.7 shows ha he MSE beween he Bayes and Learn ++ confidences is once again decreasing as new classifiers are added o he ensemble. Fig. 7 also compares Learn ++ performance o a single MLP, shown as he doed line, as described above. 4.3 Experimen 3 Finally, an addiional class was added o he disribuion from Experimen wih a variance of.5 and mean a [ ] (Fig. 8), making i an even more challenging classificaion problem due o addiional overlap beween classes. Similar o he previous wo experimens, an ensemble of 3 classifiers was generaed by Learn ++, and rained wih daa drawn from he above disribuion. The confidence of he ensemble over he enire feaure space was calculaed and ploed in comparison wih he poserior probabiliy based confidence of he Bayes classifier over he same feaure space. Fig. 9 shows hese confidence plos, where he Learn ++ based ensemble confidence (Fig. 9b) closely approximaes ha of Bayes (Fig. 9a). Densiy.6.4. Feaure Feaure Fig. 8. Daa disribuions used in Experimen 3

9 334 M. Muhlbaier, A. Topalis, and R. Polikar Feaure Feaure Feaure Feaure Fig. 9. (a) Bayesian and (b) Learn ++ confidence surface for Experimen 3 Fig. 9 indicaes ha Learn ++ assigns a larger peak confidence o he middle class han he Bayes classifier. Since he Learn ++ confidence is based on he discree decision of each classifier, when a es insance is presened from his porion of he space, mos classifiers agree on he middle class resuling in a high confidence. However, he Bayesian confidence is based on he disribuion of he paricular class and he disribuion overlap of he surrounding classes, hus lowering he confidence. Finally, he MSE beween he Learn ++ confidence and he Bayesian confidence, ploed in Fig., as a funcion of ensemble populaion, shows he now-familiar characerisic of decreasing error wih each new classifier added o he ensemble. For comparison, a single MLP was also rained on he same daa, and is mean squared error wih respec o he Bayesian confidence is shown by a doed line Mean Squared Error Number of Classifiers Fig.. Mean square error as a funcion of number of classifiers - Experimen 3 5 Conclusions and Discussions In his conribuion we have shown ha he confidence of an ensemble based classificaion algorihm in is own decision can easily be calculaed as an exponenially nor-

10 Ensemble Esimaes Poserior Probabiliy 335 malized raio of he weighs. Furhermore, we have shown - on hree experimens of increasingly difficul Gaussian disribuion - ha he confidence calculaed in his way approximaes he poserior probabiliy of he class chosen by he opimal Bayes classifier. In each case, we have observed ha he confidences calculaed by Learn ++ approximaed he Bayes poserior probabiliies raher well. However, in order o quaniaively assess exacly how close he approximaion was, we have also compued he mean square error beween he wo over he enire grid of he feaure space on which he wo classifiers were evaluaed. We have ploed his error as a funcion of he number of classifiers in he ensemble, and noiced ha he error decreased exponenially and monoonically as he number of classifiers increased; an inuiive, ye quie saisfying oucome. No over-fiing effecs were observed afer as many as 3 classifiers, and he final confidences esimaed by Learn ++ was ypically wihin % of he poserior probabiliies calculaed for he Bayes classifier. While hese resuls were obained by using Learn ++ as he ensemble algorihm, hey should generalize well o oher ensemble and/or boosing based algorihms. Acknowledgemen This maerial is based upon work suppored by he Naional Science Foundaion under Gran No. ECS-399, CAREER: An Ensemble of Classifiers Approach for Incremenal Learning. References. Kuncheva L.I. Combining Paern Classifiers, Mehods and Algorihms, Hoboken, NJ: Wiley Inerscience, 4.. Y. Freund and R. Schapire, A decision heoreic generalizaion of on-line learning and an applicaion o boosing, Compuer and Sysem Sci., vol. 57, no., pp. 9-39, Kuncheva L.I., A Theoreical Sudy on Six Classifier Fusion Sraegies, IEEE Trans. Paern Analysis and Machine Inelligence, vol. 4, no., pp. 8 86,. 4. Lewi M. and Polikar R., An ensemble approach for daa fusion wih Learn ++, Proc. 4 h In. Work. on Muliple Classifier Sysems, (Windea T.and Roli F., eds.) LNCS vol. 79, pp , Berlin: Springer, Polikar R., Udpa L., Udpa S., and Honavar V., Learn ++ : An incremenal learning algorihm for supervised neural neworks, IEEE Trans. on Sysem, Man and Cyberneics (C), vol. 3, no. 4, pp ,. 6. Duin R.P., Tax M., Classifier condiional poserior probabiliies, Lecure Noes in Compuer Science, LNCS vol. 45, pp. 6-69, Berlin: Springer, Duda R., Har P., Sork D., In Paern Classificaion /e, Chap. 3 &4, pp. 8-4, New York, NY: Wiley Inerscience,. 8. Alpaydin E. and Jordan M. Local linear perceprons for classificaion. IEEE Transacions on Neural Neworks vol. 7, no. 3, pp , Wilson D., Marinez T., Combining cross-validaion and confidence o measure finess, IEEE Join Conf. on Neural Neworks, vol., pp , 999.

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning

More information

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

Vehicle Arrival Models : Headway

Vehicle Arrival Models : Headway Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where

More information

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LTU, decision

More information

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LDA, logisic

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

An Ensemble Approach for Incremental Learning in Nonstationary Environments

An Ensemble Approach for Incremental Learning in Nonstationary Environments An Ensemble Approach for Incremenal Learning in Nonsaionary Environmens Michael D. Muhlbaier and Robi Polikar * Signal Processing and Paern Recogniion laboraory Elecrical and Compuer Engineering, Rowan

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

Modal identification of structures from roving input data by means of maximum likelihood estimation of the state space model

Modal identification of structures from roving input data by means of maximum likelihood estimation of the state space model Modal idenificaion of srucures from roving inpu daa by means of maximum likelihood esimaion of he sae space model J. Cara, J. Juan, E. Alarcón Absrac The usual way o perform a forced vibraion es is o fix

More information

20. Applications of the Genetic-Drift Model

20. Applications of the Genetic-Drift Model 0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Linear Response Theory: The connecion beween QFT and experimens 3.1. Basic conceps and ideas Q: How do we measure he conduciviy of a meal? A: we firs inroduce a weak elecric field E, and

More information

Sequential Importance Resampling (SIR) Particle Filter

Sequential Importance Resampling (SIR) Particle Filter Paricle Filers++ Pieer Abbeel UC Berkeley EECS Many slides adaped from Thrun, Burgard and Fox, Probabilisic Roboics 1. Algorihm paricle_filer( S -1, u, z ): 2. Sequenial Imporance Resampling (SIR) Paricle

More information

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing Applicaion of a Sochasic-Fuzzy Approach o Modeling Opimal Discree Time Dynamical Sysems by Using Large Scale Daa Processing AA WALASZE-BABISZEWSA Deparmen of Compuer Engineering Opole Universiy of Technology

More information

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he

More information

Comparing Means: t-tests for One Sample & Two Related Samples

Comparing Means: t-tests for One Sample & Two Related Samples Comparing Means: -Tess for One Sample & Two Relaed Samples Using he z-tes: Assumpions -Tess for One Sample & Two Relaed Samples The z-es (of a sample mean agains a populaion mean) is based on he assumpion

More information

STATE-SPACE MODELLING. A mass balance across the tank gives:

STATE-SPACE MODELLING. A mass balance across the tank gives: B. Lennox and N.F. Thornhill, 9, Sae Space Modelling, IChemE Process Managemen and Conrol Subjec Group Newsleer STE-SPACE MODELLING Inroducion: Over he pas decade or so here has been an ever increasing

More information

Air Traffic Forecast Empirical Research Based on the MCMC Method

Air Traffic Forecast Empirical Research Based on the MCMC Method Compuer and Informaion Science; Vol. 5, No. 5; 0 ISSN 93-8989 E-ISSN 93-8997 Published by Canadian Cener of Science and Educaion Air Traffic Forecas Empirical Research Based on he MCMC Mehod Jian-bo Wang,

More information

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence

Supplement for Stochastic Convex Optimization: Faster Local Growth Implies Faster Global Convergence Supplemen for Sochasic Convex Opimizaion: Faser Local Growh Implies Faser Global Convergence Yi Xu Qihang Lin ianbao Yang Proof of heorem heorem Suppose Assumpion holds and F (w) obeys he LGC (6) Given

More information

Testing for a Single Factor Model in the Multivariate State Space Framework

Testing for a Single Factor Model in the Multivariate State Space Framework esing for a Single Facor Model in he Mulivariae Sae Space Framework Chen C.-Y. M. Chiba and M. Kobayashi Inernaional Graduae School of Social Sciences Yokohama Naional Universiy Japan Faculy of Economics

More information

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Simulaion-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Week Descripion Reading Maerial 2 Compuer Simulaion of Dynamic Models Finie Difference, coninuous saes, discree ime Simple Mehods Euler Trapezoid

More information

Random Walk with Anti-Correlated Steps

Random Walk with Anti-Correlated Steps Random Walk wih Ani-Correlaed Seps John Noga Dirk Wagner 2 Absrac We conjecure he expeced value of random walks wih ani-correlaed seps o be exacly. We suppor his conjecure wih 2 plausibiliy argumens and

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Georey E. Hinton. University oftoronto. Technical Report CRG-TR February 22, Abstract

Georey E. Hinton. University oftoronto.   Technical Report CRG-TR February 22, Abstract Parameer Esimaion for Linear Dynamical Sysems Zoubin Ghahramani Georey E. Hinon Deparmen of Compuer Science Universiy oftorono 6 King's College Road Torono, Canada M5S A4 Email: zoubin@cs.orono.edu Technical

More information

Bias-Variance Error Bounds for Temporal Difference Updates

Bias-Variance Error Bounds for Temporal Difference Updates Bias-Variance Bounds for Temporal Difference Updaes Michael Kearns AT&T Labs mkearns@research.a.com Sainder Singh AT&T Labs baveja@research.a.com Absrac We give he firs rigorous upper bounds on he error

More information

Matlab and Python programming: how to get started

Matlab and Python programming: how to get started Malab and Pyhon programming: how o ge sared Equipping readers he skills o wrie programs o explore complex sysems and discover ineresing paerns from big daa is one of he main goals of his book. In his chaper,

More information

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time.

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time. Supplemenary Figure 1 Spike-coun auocorrelaions in ime. Normalized auocorrelaion marices are shown for each area in a daase. The marix shows he mean correlaion of he spike coun in each ime bin wih he spike

More information

A Shooting Method for A Node Generation Algorithm

A Shooting Method for A Node Generation Algorithm A Shooing Mehod for A Node Generaion Algorihm Hiroaki Nishikawa W.M.Keck Foundaion Laboraory for Compuaional Fluid Dynamics Deparmen of Aerospace Engineering, Universiy of Michigan, Ann Arbor, Michigan

More information

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17 EES 16A Designing Informaion Devices and Sysems I Spring 019 Lecure Noes Noe 17 17.1 apaciive ouchscreen In he las noe, we saw ha a capacior consiss of wo pieces on conducive maerial separaed by a nonconducive

More information

Dimitri Solomatine. D.P. Solomatine. Data-driven modelling (part 2). 2

Dimitri Solomatine. D.P. Solomatine. Data-driven modelling (part 2). 2 Daa-driven modelling. Par. Daa-driven Arificial di Neural modelling. Newors Par Dimiri Solomaine Arificial neural newors D.P. Solomaine. Daa-driven modelling par. 1 Arificial neural newors ANN: main pes

More information

New Boosting Methods of Gaussian Processes for Regression

New Boosting Methods of Gaussian Processes for Regression New Boosing Mehods of Gaussian Processes for Regression Yangqiu Song Sae Key Laboraory of Inelligen echnology and Sysems Deparmen of Auomaion singhua Universiy Beijing 00084, P.R.China E-mail: songyq99@mails.singhua.edu.cn

More information

A new flexible Weibull distribution

A new flexible Weibull distribution Communicaions for Saisical Applicaions and Mehods 2016, Vol. 23, No. 5, 399 409 hp://dx.doi.org/10.5351/csam.2016.23.5.399 Prin ISSN 2287-7843 / Online ISSN 2383-4757 A new flexible Weibull disribuion

More information

Asymptotic Equipartition Property - Seminar 3, part 1

Asymptotic Equipartition Property - Seminar 3, part 1 Asympoic Equipariion Propery - Seminar 3, par 1 Ocober 22, 2013 Problem 1 (Calculaion of ypical se) To clarify he noion of a ypical se A (n) ε and he smalles se of high probabiliy B (n), we will calculae

More information

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED 0.1 MAXIMUM LIKELIHOOD ESTIMATIO EXPLAIED Maximum likelihood esimaion is a bes-fi saisical mehod for he esimaion of he values of he parameers of a sysem, based on a se of observaions of a random variable

More information

Online Convex Optimization Example And Follow-The-Leader

Online Convex Optimization Example And Follow-The-Leader CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion

More information

Christos Papadimitriou & Luca Trevisan November 22, 2016

Christos Papadimitriou & Luca Trevisan November 22, 2016 U.C. Bereley CS170: Algorihms Handou LN-11-22 Chrisos Papadimiriou & Luca Trevisan November 22, 2016 Sreaming algorihms In his lecure and he nex one we sudy memory-efficien algorihms ha process a sream

More information

The ROC-Boost Design Algorithm for Asymmetric Classification

The ROC-Boost Design Algorithm for Asymmetric Classification The ROC-Boos Design Algorihm for Asymmeric Classificaion Guido Cesare Deparmen of Mahemaics Universiy of Genova (Ialy guido.cesare@gmail.com Robero Manduchi Deparmen of Compuer Engineering Universiy of

More information

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis Speaker Adapaion Techniques For Coninuous Speech Using Medium and Small Adapaion Daa Ses Consaninos Boulis Ouline of he Presenaion Inroducion o he speaker adapaion problem Maximum Likelihood Sochasic Transformaions

More information

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070

More information

Stability and Bifurcation in a Neural Network Model with Two Delays

Stability and Bifurcation in a Neural Network Model with Two Delays Inernaional Mahemaical Forum, Vol. 6, 11, no. 35, 175-1731 Sabiliy and Bifurcaion in a Neural Nework Model wih Two Delays GuangPing Hu and XiaoLing Li School of Mahemaics and Physics, Nanjing Universiy

More information

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017 Two Popular Bayesian Esimaors: Paricle and Kalman Filers McGill COMP 765 Sep 14 h, 2017 1 1 1, dx x Bel x u x P x z P Recall: Bayes Filers,,,,,,, 1 1 1 1 u z u x P u z u x z P Bayes z = observaion u =

More information

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1 CpS 570 Machine Learning School of EECS Washingon Sae Universiy CpS 570 - Machine Learning 1 Form of underlying disribuions unknown Bu sill wan o perform classificaion and regression Semi-parameric esimaion

More information

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Kriging Models Predicing Arazine Concenraions in Surface Waer Draining Agriculural Waersheds Paul L. Mosquin, Jeremy Aldworh, Wenlin Chen Supplemenal Maerial Number

More information

STRUCTURAL CHANGE IN TIME SERIES OF THE EXCHANGE RATES BETWEEN YEN-DOLLAR AND YEN-EURO IN

STRUCTURAL CHANGE IN TIME SERIES OF THE EXCHANGE RATES BETWEEN YEN-DOLLAR AND YEN-EURO IN Inernaional Journal of Applied Economerics and Quaniaive Sudies. Vol.1-3(004) STRUCTURAL CHANGE IN TIME SERIES OF THE EXCHANGE RATES BETWEEN YEN-DOLLAR AND YEN-EURO IN 001-004 OBARA, Takashi * Absrac The

More information

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi

Navneet Saini, Mayank Goyal, Vishal Bansal (2013); Term Project AML310; Indian Institute of Technology Delhi Creep in Viscoelasic Subsances Numerical mehods o calculae he coefficiens of he Prony equaion using creep es daa and Herediary Inegrals Mehod Navnee Saini, Mayank Goyal, Vishal Bansal (23); Term Projec

More information

Chapter 2. Models, Censoring, and Likelihood for Failure-Time Data

Chapter 2. Models, Censoring, and Likelihood for Failure-Time Data Chaper 2 Models, Censoring, and Likelihood for Failure-Time Daa William Q. Meeker and Luis A. Escobar Iowa Sae Universiy and Louisiana Sae Universiy Copyrigh 1998-2008 W. Q. Meeker and L. A. Escobar. Based

More information

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t

R t. C t P t. + u t. C t = αp t + βr t + v t. + β + w t Exercise 7 C P = α + β R P + u C = αp + βr + v (a) (b) C R = α P R + β + w (c) Assumpions abou he disurbances u, v, w : Classical assumions on he disurbance of one of he equaions, eg. on (b): E(v v s P,

More information

Guest Lectures for Dr. MacFarlane s EE3350 Part Deux

Guest Lectures for Dr. MacFarlane s EE3350 Part Deux Gues Lecures for Dr. MacFarlane s EE3350 Par Deux Michael Plane Mon., 08-30-2010 Wrie name in corner. Poin ou his is a review, so I will go faser. Remind hem o go lisen o online lecure abou geing an A

More information

) were both constant and we brought them from under the integral.

) were both constant and we brought them from under the integral. YIELD-PER-RECRUIT (coninued The yield-per-recrui model applies o a cohor, bu we saw in he Age Disribuions lecure ha he properies of a cohor do no apply in general o a collecion of cohors, which is wha

More information

GMM - Generalized Method of Moments

GMM - Generalized Method of Moments GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................

More information

Lecture 33: November 29

Lecture 33: November 29 36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure

More information

UNIVERSITY OF TRENTO MEASUREMENTS OF TRANSIENT PHENOMENA WITH DIGITAL OSCILLOSCOPES. Antonio Moschitta, Fabrizio Stefani, Dario Petri.

UNIVERSITY OF TRENTO MEASUREMENTS OF TRANSIENT PHENOMENA WITH DIGITAL OSCILLOSCOPES. Antonio Moschitta, Fabrizio Stefani, Dario Petri. UNIVERSIY OF RENO DEPARMEN OF INFORMAION AND COMMUNICAION ECHNOLOGY 385 Povo reno Ialy Via Sommarive 4 hp://www.di.unin.i MEASUREMENS OF RANSIEN PHENOMENA WIH DIGIAL OSCILLOSCOPES Anonio Moschia Fabrizio

More information

The Rosenblatt s LMS algorithm for Perceptron (1958) is built around a linear neuron (a neuron with a linear

The Rosenblatt s LMS algorithm for Perceptron (1958) is built around a linear neuron (a neuron with a linear In The name of God Lecure4: Percepron and AALIE r. Majid MjidGhoshunih Inroducion The Rosenbla s LMS algorihm for Percepron 958 is buil around a linear neuron a neuron ih a linear acivaion funcion. Hoever,

More information

Robotics I. April 11, The kinematics of a 3R spatial robot is specified by the Denavit-Hartenberg parameters in Tab. 1.

Robotics I. April 11, The kinematics of a 3R spatial robot is specified by the Denavit-Hartenberg parameters in Tab. 1. Roboics I April 11, 017 Exercise 1 he kinemaics of a 3R spaial robo is specified by he Denavi-Harenberg parameers in ab 1 i α i d i a i θ i 1 π/ L 1 0 1 0 0 L 3 0 0 L 3 3 able 1: able of DH parameers of

More information

Some Basic Information about M-S-D Systems

Some Basic Information about M-S-D Systems Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,

More information

3.1 More on model selection

3.1 More on model selection 3. More on Model selecion 3. Comparing models AIC, BIC, Adjused R squared. 3. Over Fiing problem. 3.3 Sample spliing. 3. More on model selecion crieria Ofen afer model fiing you are lef wih a handful of

More information

Echocardiography Project and Finite Fourier Series

Echocardiography Project and Finite Fourier Series Echocardiography Projec and Finie Fourier Series 1 U M An echocardiagram is a plo of how a porion of he hear moves as he funcion of ime over he one or more hearbea cycles If he hearbea repeas iself every

More information

RC, RL and RLC circuits

RC, RL and RLC circuits Name Dae Time o Complee h m Parner Course/ Secion / Grade RC, RL and RLC circuis Inroducion In his experimen we will invesigae he behavior of circuis conaining combinaions of resisors, capaciors, and inducors.

More information

Západočeská Univerzita v Plzni, Czech Republic and Groupe ESIEE Paris, France

Západočeská Univerzita v Plzni, Czech Republic and Groupe ESIEE Paris, France ADAPTIVE SIGNAL PROCESSING USING MAXIMUM ENTROPY ON THE MEAN METHOD AND MONTE CARLO ANALYSIS Pavla Holejšovsá, Ing. *), Z. Peroua, Ing. **), J.-F. Bercher, Prof. Assis. ***) Západočesá Univerzia v Plzni,

More information

Lab 10: RC, RL, and RLC Circuits

Lab 10: RC, RL, and RLC Circuits Lab 10: RC, RL, and RLC Circuis In his experimen, we will invesigae he behavior of circuis conaining combinaions of resisors, capaciors, and inducors. We will sudy he way volages and currens change in

More information

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015

Explaining Total Factor Productivity. Ulrich Kohli University of Geneva December 2015 Explaining Toal Facor Produciviy Ulrich Kohli Universiy of Geneva December 2015 Needed: A Theory of Toal Facor Produciviy Edward C. Presco (1998) 2 1. Inroducion Toal Facor Produciviy (TFP) has become

More information

Distributed Deep Learning Parallel Sparse Autoencoder. 2 Serial Sparse Autoencoder. 1 Introduction. 2.1 Stochastic Gradient Descent

Distributed Deep Learning Parallel Sparse Autoencoder. 2 Serial Sparse Autoencoder. 1 Introduction. 2.1 Stochastic Gradient Descent Disribued Deep Learning Parallel Sparse Auoencoder Inroducion Abhik Lahiri Raghav Pasari Bobby Prochnow December 0, 00 Much of he bleeding edge research in he areas of compuer vision, naural language processing,

More information

Particle Swarm Optimization Combining Diversification and Intensification for Nonlinear Integer Programming Problems

Particle Swarm Optimization Combining Diversification and Intensification for Nonlinear Integer Programming Problems Paricle Swarm Opimizaion Combining Diversificaion and Inensificaion for Nonlinear Ineger Programming Problems Takeshi Masui, Masaoshi Sakawa, Kosuke Kao and Koichi Masumoo Hiroshima Universiy 1-4-1, Kagamiyama,

More information

Some Ramsey results for the n-cube

Some Ramsey results for the n-cube Some Ramsey resuls for he n-cube Ron Graham Universiy of California, San Diego Jozsef Solymosi Universiy of Briish Columbia, Vancouver, Canada Absrac In his noe we esablish a Ramsey-ype resul for cerain

More information

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon 3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of

More information

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter Sae-Space Models Iniializaion, Esimaion and Smoohing of he Kalman Filer Iniializaion of he Kalman Filer The Kalman filer shows how o updae pas predicors and he corresponding predicion error variances when

More information

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II Roland Siegwar Margaria Chli Paul Furgale Marco Huer Marin Rufli Davide Scaramuzza ETH Maser Course: 151-0854-00L Auonomous Mobile Robos Localizaion II ACT and SEE For all do, (predicion updae / ACT),

More information

Tasty Coffee example

Tasty Coffee example Lecure Slides for (Binary) Classificaion: Learning a Class from labeled Examples ITRODUCTIO TO Machine Learning ETHEM ALPAYDI The MIT Press, 00 (modified by dph, 0000) CHAPTER : Supervised Learning Things

More information

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature

On Measuring Pro-Poor Growth. 1. On Various Ways of Measuring Pro-Poor Growth: A Short Review of the Literature On Measuring Pro-Poor Growh 1. On Various Ways of Measuring Pro-Poor Growh: A Shor eview of he Lieraure During he pas en years or so here have been various suggesions concerning he way one should check

More information

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.

More information

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H.

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H. ACE 56 Fall 005 Lecure 5: he Simple Linear Regression Model: Sampling Properies of he Leas Squares Esimaors by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Inference in he Simple

More information

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal

More information

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still. Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in

More information

5.2. The Natural Logarithm. Solution

5.2. The Natural Logarithm. Solution 5.2 The Naural Logarihm The number e is an irraional number, similar in naure o π. Is non-erminaing, non-repeaing value is e 2.718 281 828 59. Like π, e also occurs frequenly in naural phenomena. In fac,

More information

Particle Swarm Optimization

Particle Swarm Optimization Paricle Swarm Opimizaion Speaker: Jeng-Shyang Pan Deparmen of Elecronic Engineering, Kaohsiung Universiy of Applied Science, Taiwan Email: jspan@cc.kuas.edu.w 7/26/2004 ppso 1 Wha is he Paricle Swarm Opimizaion

More information

ACE 564 Spring Lecture 7. Extensions of The Multiple Regression Model: Dummy Independent Variables. by Professor Scott H.

ACE 564 Spring Lecture 7. Extensions of The Multiple Regression Model: Dummy Independent Variables. by Professor Scott H. ACE 564 Spring 2006 Lecure 7 Exensions of The Muliple Regression Model: Dumm Independen Variables b Professor Sco H. Irwin Readings: Griffihs, Hill and Judge. "Dumm Variables and Varing Coefficien Models

More information

The Arcsine Distribution

The Arcsine Distribution The Arcsine Disribuion Chris H. Rycrof Ocober 6, 006 A common heme of he class has been ha he saisics of single walker are ofen very differen from hose of an ensemble of walkers. On he firs homework, we

More information

A Specification Test for Linear Dynamic Stochastic General Equilibrium Models

A Specification Test for Linear Dynamic Stochastic General Equilibrium Models Journal of Saisical and Economeric Mehods, vol.1, no.2, 2012, 65-70 ISSN: 2241-0384 (prin), 2241-0376 (online) Scienpress Ld, 2012 A Specificaion Tes for Linear Dynamic Sochasic General Equilibrium Models

More information

A Video Vehicle Detection Algorithm Based on Improved Adaboost Algorithm Weiguang Liu and Qian Zhang*

A Video Vehicle Detection Algorithm Based on Improved Adaboost Algorithm Weiguang Liu and Qian Zhang* A Video Vehicle Deecion Algorihm Based on Improved Adaboos Algorihm Weiguang Liu and Qian Zhang* Zhongyuan Universiy of Technology, Zhengzhou 450000, China lwg66123@163.com, 2817343431@qq.com *The corresponding

More information

Lecture 2 April 04, 2018

Lecture 2 April 04, 2018 Sas 300C: Theory of Saisics Spring 208 Lecure 2 April 04, 208 Prof. Emmanuel Candes Scribe: Paulo Orensein; edied by Sephen Baes, XY Han Ouline Agenda: Global esing. Needle in a Haysack Problem 2. Threshold

More information

Applying Genetic Algorithms for Inventory Lot-Sizing Problem with Supplier Selection under Storage Capacity Constraints

Applying Genetic Algorithms for Inventory Lot-Sizing Problem with Supplier Selection under Storage Capacity Constraints IJCSI Inernaional Journal of Compuer Science Issues, Vol 9, Issue 1, No 1, January 2012 wwwijcsiorg 18 Applying Geneic Algorihms for Invenory Lo-Sizing Problem wih Supplier Selecion under Sorage Capaciy

More information

Robust estimation based on the first- and third-moment restrictions of the power transformation model

Robust estimation based on the first- and third-moment restrictions of the power transformation model h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,

More information

Final Spring 2007

Final Spring 2007 .615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o

More information

Recent Developments In Evolutionary Data Assimilation And Model Uncertainty Estimation For Hydrologic Forecasting Hamid Moradkhani

Recent Developments In Evolutionary Data Assimilation And Model Uncertainty Estimation For Hydrologic Forecasting Hamid Moradkhani Feb 6-8, 208 Recen Developmens In Evoluionary Daa Assimilaion And Model Uncerainy Esimaion For Hydrologic Forecasing Hamid Moradkhani Cener for Complex Hydrosysems Research Deparmen of Civil, Consrucion

More information

(Not) Bounding the True Error

(Not) Bounding the True Error (No) Bounding he True Error John Langford Deparmen of Compuer Science Carnegie-Mellon Universiy Pisburgh, PA 15213 jcl+@cs.cmu.edu Rich Caruana Deparmen of Compuer Science Carnegie-Mellon Universiy Pisburgh,

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Noes for EE7C Spring 018: Convex Opimizaion and Approximaion Insrucor: Moriz Hard Email: hard+ee7c@berkeley.edu Graduae Insrucor: Max Simchowiz Email: msimchow+ee7c@berkeley.edu Ocober 15, 018 3

More information

EKF SLAM vs. FastSLAM A Comparison

EKF SLAM vs. FastSLAM A Comparison vs. A Comparison Michael Calonder, Compuer Vision Lab Swiss Federal Insiue of Technology, Lausanne EPFL) michael.calonder@epfl.ch The wo algorihms are described wih a planar robo applicaion in mind. Generalizaion

More information

Ecological Archives E A1. Meghan A. Duffy, Spencer R. Hall, Carla E. Cáceres, and Anthony R. Ives.

Ecological Archives E A1. Meghan A. Duffy, Spencer R. Hall, Carla E. Cáceres, and Anthony R. Ives. Ecological Archives E9-95-A1 Meghan A. Duffy, pencer R. Hall, Carla E. Cáceres, and Anhony R. ves. 29. Rapid evoluion, seasonaliy, and he erminaion of parasie epidemics. Ecology 9:1441 1448. Appendix A.

More information

Maintenance Models. Prof. Robert C. Leachman IEOR 130, Methods of Manufacturing Improvement Spring, 2011

Maintenance Models. Prof. Robert C. Leachman IEOR 130, Methods of Manufacturing Improvement Spring, 2011 Mainenance Models Prof Rober C Leachman IEOR 3, Mehods of Manufacuring Improvemen Spring, Inroducion The mainenance of complex equipmen ofen accouns for a large porion of he coss associaed wih ha equipmen

More information

The field of mathematics has made tremendous impact on the study of

The field of mathematics has made tremendous impact on the study of A Populaion Firing Rae Model of Reverberaory Aciviy in Neuronal Neworks Zofia Koscielniak Carnegie Mellon Universiy Menor: Dr. G. Bard Ermenrou Universiy of Pisburgh Inroducion: The field of mahemaics

More information

Financial Econometrics Kalman Filter: some applications to Finance University of Evry - Master 2

Financial Econometrics Kalman Filter: some applications to Finance University of Evry - Master 2 Financial Economerics Kalman Filer: some applicaions o Finance Universiy of Evry - Maser 2 Eric Bouyé January 27, 2009 Conens 1 Sae-space models 2 2 The Scalar Kalman Filer 2 21 Presenaion 2 22 Summary

More information

Appendix to Creating Work Breaks From Available Idleness

Appendix to Creating Work Breaks From Available Idleness Appendix o Creaing Work Breaks From Available Idleness Xu Sun and Ward Whi Deparmen of Indusrial Engineering and Operaions Research, Columbia Universiy, New York, NY, 127; {xs2235,ww24}@columbia.edu Sepember

More information

Learning Naive Bayes Classifier from Noisy Data

Learning Naive Bayes Classifier from Noisy Data UCLA Compuer Science Deparmen Technical Repor CSD-TR No 030056 1 Learning Naive Bayes Classifier from Noisy Daa Yirong Yang, Yi Xia, Yun Chi, and Richard R Munz Universiy of California, Los Angeles, CA

More information

Shiva Akhtarian MSc Student, Department of Computer Engineering and Information Technology, Payame Noor University, Iran

Shiva Akhtarian MSc Student, Department of Computer Engineering and Information Technology, Payame Noor University, Iran Curren Trends in Technology and Science ISSN : 79-055 8hSASTech 04 Symposium on Advances in Science & Technology-Commission-IV Mashhad, Iran A New for Sofware Reliabiliy Evaluaion Based on NHPP wih Imperfec

More information

An introduction to the theory of SDDP algorithm

An introduction to the theory of SDDP algorithm An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking

More information

Vectorautoregressive Model and Cointegration Analysis. Time Series Analysis Dr. Sevtap Kestel 1

Vectorautoregressive Model and Cointegration Analysis. Time Series Analysis Dr. Sevtap Kestel 1 Vecorauoregressive Model and Coinegraion Analysis Par V Time Series Analysis Dr. Sevap Kesel 1 Vecorauoregression Vecor auoregression (VAR) is an economeric model used o capure he evoluion and he inerdependencies

More information

Effect of Pruning and Early Stopping on Performance of a Boosting Ensemble

Effect of Pruning and Early Stopping on Performance of a Boosting Ensemble Effec of Pruning and Early Sopping on Performance of a Boosing Ensemble Harris Drucker, Ph.D Monmouh Universiy, Wes ong Branch, NJ 07764 USA drucker@monmouh.edu Absrac: Generaing an archiecure for an ensemble

More information