Density Estimation Trees

Size: px
Start display at page:

Download "Density Estimation Trees"

Transcription

1 Densiy Esimaion Trees Parikshi Ram Georgia Insiue of Technology Alana GA, Alexander G. Gray Georgia Insiue of Technology Alana GA, ABSTRACT In his paper we develop densiy esimaion rees (DETs), he naural analog of classificaion rees and regression rees, for he ask of densiy esimaion. We consider he esimaion of a join probabiliy densiy funcion of a d-dimensional random vecor X and define a piecewise consan esimaor srucured as a decision ree. The inegraed squared error is minimized o learn he ree. We show ha he mehod is nonparameric: under sandard condiions of nonparameric densiy esimaion, DETs are shown o be asympoically consisen. In addiion, being decision rees, DETs perform auomaic feaure selecion. They empirically exhibi he inerpreabiliy, adapabiliy and feaure selecion properies of supervised decision rees while incurring sligh loss in accuracy over oher nonparameric densiy esimaors. Hence hey migh be able o avoid he curse of dimensionaliy if he rue densiy is sparse in dimensions. We believe ha densiy esimaion rees provide a new ool for exploraory daa analysis wih unique capabiliies. Caegories and Subjec Descripors I.5 [Paern Recogniion]: Models; G. [Probabiliy and Saisics]: onparameric saisics General Terms Algorihms, Experimenaion Keywords Decision rees, densiy esimaion, daa analysis. ITRODUCTIO Three mos fundamenal asks of machine learning are classificaion, regression, and densiy esimaion. Classificaion and regression are insances of supervised daa analysis where a raining se of examples is provided. The learning ask is o esimae a funcion using he raining se which also performs well on a es se. The hird ask, densiy esimaion, is an insance of unsupervised learning. This is Permission o make digial or hard copies of all or par of his work for personal or classroom use is graned wihou fee provided ha copies are no made or disribued for profi or commercial advanage and ha copies bear his noice and he full ciaion on he firs page. To copy oherwise, o republish, o pos on servers or o redisribue o liss, requires prior specific permission and/or a fee. KDD, Augus 4,, San Diego, California, USA. Copyrigh ACM //8...$.. generally harder because one does no have any insance of he ground ruh regarding he quaniy being esimaed. Decision rees [] have been widely used in he supervised seing for classificaion and regression. In his paper we inroduce, derive, and explore he naural analog of classificaion rees and regression rees for he unsupervised ask of densiy esimaion. To our knowledge his analogy has never been explored rigorously, hough several oher ideas for densiy esimaion involving hierarchical schemes have been proposed. Our hope is ha he unique advanages of decision rees in he supervised seing will ransfer o he unsupervised seing, o give rise o a new nonparameric densiy esimaion mehod wih ineresing capabiliies. Densiy esimaion. The problem of densiy esimaion can be defined as esimaing an unknown disribuion f on X wih a close approximaion ˆf : X R + given a se of iid (independen and idenically disribued) observaions {X,X,...,X } X drawn from f. Esimaing he probabiliy densiy of he given daa is a fundamenal ask in mulivariae saisics. I ofen appears as a subrouine in oher inference for example, in classificaion, one is required o find ˆP ˆp(X C)P (C) (C X) =, where C is one of ˆp(X) he possible K classes and ˆp(X C) is he class-condiional densiy of he daa X [,, 4]. I is also widely used for conducing exploraory analyses such as oulier or cluser deecion. Decision rees. Decision rees [] were developed for classificaion [5] and regression [6]. They are primarily used for he purpose of supervised learning and have been widely and successfully used in nonparameric classificaion and regression. The piecewise consan esimaors and he simplisic model of univariae binary splis of he daa in decision rees lead o relaively less accurae esimaors wih unknown convergence raes [7]. Techniques such as bagging are used o boos his accuracy. More sophisicaed mehods wih beer accuracies and clearer asympoic properies such as local linear regression [8] and suppor vecor machines [9] exis for regression and classificaion respecively. oneheless, anecdoally, decision rees remain one of he more widely used mehods in pracice [7], possibly even he mos widely used. This is due o he inheren inuiiveness, adapabiliy and inerpreabiliy of he models learned. Moreover, hey are capable of dealing wih mixed caegorical, discree and coninuous variables wihin one unified framework. They also perform feaure selecion and are easy o implemen. These properies come a he cos of accuracy in predicion bu sill make decision rees very desirable and If all he aribues of he daa are coninuous hen X R d 67

2 Table : Characerisics of Mehods for Densiy Esimaion Mehods np Accuracy Inerpreabiliy Adapabiliy Speed COD VI Rules ABD AWD Training Query MoG low fas EM algorihm very fas O(d ) Hisogram low fas O(Md) very fas O(d) KDE (FBW) medium slow O(Hd ) slow O(d) KDE (FBW) medium fas O(Hd) medium O(d log ) KDE (VBW) highes very slow O(H d d)/q slow O(d) KDE (VBW) high slow O(H d d log )/q medium O(d log ) local r-kde very high fas CV sep/q slow O(d) global r-kde medium/high fas CV seps slow O(d) DET medium slow LOO-CV fas O(D T ) pracical. Anoher advanage of decision rees is he efficien es querying once he model (in his case a ree) has been rained. Decision rees rade some accuracy for a simple, inerpreable model. Considering ha densiy esimaion is widely used in exploraory analysis, giving up accuracy (which is no a useful concep in he absence of he ground ruh) for undersanding is accepable.. onparameric Densiy Esimaors Hisograms and kernel densiy esimaors (KDEs) [] are simpler nonparameric (np) echniques, whereas varians of KDEs such as rodeo-kdes [] and wavele based mehods [8] are more complex nonparameric esimaors. A number of nonparameric Bayesian mehods have also been developed for he ask of densiy esimaion []. Mixures of Gaussians (MoGs) are widely used parameric densiy esimaors. Even hough he nonparameric mehods have been shown o be significanly more accurae han he parameric mehods and require fewer assumpions on he daa han parameric mehods, MoGs are widely used because of heir inerpreabiliy as clusers (as well as heir simple implemenaion). Adapabiliy and Speed. Adapabiliy has wo implicaions in densiy esimaion: () adapable beween dimensions (ABD) he esimaion process should rea dimensions differenly depending on heir influence on he densiy (for example, addiion of unimporan dimensions of uniform noise should no affec he esimaor) () adapable wihin a dimension (AWD) deermining regions of ineres for a given dimension and adjusing he esimaor o he local behavior of he densiy in ha dimension (for example, regions wih fas changing densiy vs. regions wih fla densiies). MoGs are no known o adap beween or wihin dimensions. Fixed-bandwidh KDEs (FBW) are accurae nonparameric esimaors bu are no adapable beween or wihin dimensions because of he resricion of using a mulivariae kernel wih he same bandwidh in every dimension (spherical kernel) for every poin. The bandwidh is chosen hrough leave-one-ou cross-validaion (LOO-CV) by picking he one wih he bes CV loss among he H differen values ried. A significanly high value of H is required o obain accurae resuls. LOO-CV for he naïve implemenaion of KDE akes O(Hd ). Spaial pariioning ree daa srucures have been used for fas-approximae compuaion of KDEs (KDE s) [, 4]. For cover-rees [5], his process requires O(Hd) ime [6]. Various KDEs wih adapive bandwidhs have been developed bu are complicaed in pracice. The neares-neighbor KDEs [] are locally adapive by choosing bandwidhs based on heir k h -neares-neighbor disance. However, hey are no adapive over dimensions since hey sill use a spherical mulivariae kernel. Moreover, he esimae does no represen a probabiliy densiy funcion (pdf), and is disconinuous and heavy ailed. A ruly adapive (ABD) KDE would require an opimal bandwidh for each dimension (VBW). This means he kernel used would be ellipical, adaping o he densiy of he daa. This makes he VBW-KDEs adapable beween dimensions. If H bandwidhs are ried in each dimension, he raining ime required for he naïve implemenaion of KDEs would be O(H d d). To make his locally adapive for each dimension(awd), he bandwidh esimaion would be required for each query q. Even he faser mehods (KDE ) would require O(H d d log ) raining ime for each query. This makes he compuaional cos of CV for VBW-KDEs inracable even in medium dimensions. A recen algorihm [] uses rodeo [7] o greedily selec adapive bandwidhs for a KDE in each dimension. Under cerain sparsiy condiions on he unknown densiy funcion, he resuling esimaor is shown o empirically perform well and obain near opimal convergence rae for KDEs. The local version (local r-kde) of his esimaor compues he opimal bandwidh in every dimension for every query poin. The paper demonsraes rodeo s adapabiliy implicily idenifying he dimensions of ineres during he process of bandwidh selecion as well as idenifying regions of ineres wihin a single dimension, selecing smaller bandwidhs where he densiy varies more rapidly. However, his echnique is expensive, requiring a rodeo sep for every single query. The rodeo sep is ieraive and we are no aware of is runime bound. However, i is empirically faser han LOO-CV for VBW-KDEs. The global-rodeo [] (global r- KDE) improves efficiency by using a fixed bandwidh wihin each dimension, and esimaing i by averaging he esimaed bandwidhs for raining poins in every dimension. Hence, he resuling esimaor loses local adapabiliy wihin a dimension (propery of adapabiliy). The raining ime now involves rodeo seps insead of a rodeo sep for each query. The accuracy of he esimae obviously depends on he (larger number raining queries imply more accurae esimaes and slower raining). KDEs can be made adapable, however, a he cos of compuaional complexiy. Moreover, given he esimaed bandwidhs, he ime aken for a single query using he naïve implemenaion of a KDE is O(d). The cover-ree daa srucure provides an approximae esimae in O(d log ) query ime. For O() queries, he query cos can be amor- 68

3 ized over he queries using a ree-daa-srucure over he queries, requiring a oal ime of O(d) [6]. Decision rees are known for heir adapabiliy over he daa (over feaures by implici feaure selecion and wihin dimensions by choosing differen leaf size in differen regions). In his paper, we will demonsrae heir adapabiliy for he ask of densiy esimaion. However, raining decision rees using LOO-CV is an expensive sep involving he growing of he ree, and he subsequen cos-complexiy pruning [, 7]. However his cos is amorized over he muliple empirically efficien queries. The query ime for a decision ree esimaor is O(D T )whered T is he deph of a decision ree T. The wors case upper bound for D T is O(), bu is empirically seen o be much igher in pracice (much more closer o O(log )). Hence decision rees also bring efficiency o nonparameric densiy esimaion a he cos of some accuracy. Hisograms are also fas nonparameric esimaors during raining as well as query ime. They require O(Md) raining ime for rying M differen binwidhs and choosing he one wih he bes CV error. The query ime is a blazing O(d). However, hey lack he adapabiliy of decision rees, and become prohibiively inaccurae as he number of dimensions increases. Inerpreabiliy. We also propose he use of decision rees for densiy esimaion o inroduce inerpreabiliy in he nonparameric seing. The decision rees provide an ineresing overlap beween he accuracy of nonparameric mehods and he simpliciy and inerpreabiliy of parameric mehods. Inerpreabiliy in densiy esimaion can be useful in he providing he following informaion abou he densiy of he daa: () deecing clusers and ouliers (COD) () providing relaive variable/dimension imporance (VI) idenifying dimensions ha significanly affec he densiy () providing simple univariae rules hese simple rules can be used, for example, for direcly specifying some chunk of a huge daa se which sis on a DBMS daabase. In he case of a DET, his chunk migh represen a cluser. MoGs are known o deec clusers bu are incapable of imparing any inerpreabiliy regarding he dimensions. The aforemenioned mehod using rodeo [] implicily does idenify dimensions of ineres. However, i is hard o obain relaive imporance of dimensions by jus looking a he values of he esimaed bandwidhs. Moreover, cluser and oulier deecion is complicaed wih KDEs an exhausive scan of he whole space provides regions of high (clusers) and low densiy (ouliers). So KDEs are no inerpreable densiy esimaors. When accuracy is required wih no concern over compuaion ime, adapive-kdes are he mehod of choice. However, hey migh be oo cosly for exploraory analysis. We aemp o achieve he properies of inerpreabiliy, while reaining adapabiliy, by using he CART-syle univariae splis which gives rise o he uniquely readable rule srucure of decision rees. This key propery is no shared by he several oher mehods for densiy esimaion involving hierarchical schemes which have been proposed over he years. Deermining dimensions of ineres is a direc byproduc of he decision ree framework via he measure of relaive imporance of predicor variables []. While linear models such as linear regression and linear SVMs yield such I is imporan o noe ha he spaial pariioning ree building has a one-ime cos of O(d log ). We will define his measure in he echnical secion of he paper. Figure : Densiy esimae in form of a decision ree and he relaive variable imporance. a capabiliy via examinaion of he linear coefficiens, his capabiliy is harder o find in nonparameric mehods. A sored lis of he leaves of a DET can easily idenify he clusers and he ouliers in he daa. All he aforemenioned properies of hese esimaors are summarized in Table. We furher moivae he desirabiliy of inerpreabiliy of a decision ree in he following example: Moivaing example. Consider he densiy esimaes for he Iris daa se (5 4) 4 as a se of rules depiced in Figure. Each leaf node conains he piecewise consan densiy esimae. More explicily, ˆf(x) =.6 x R (say). Here he leaf R = {x: x 4. &x.5}. We have also lised he class memberships of he poins in he leaf nodes 5. This figure also presens he relaive imporance of he aribues. The ree represenaion les us easily find regions of relaively high (R ) and low densiies (R ). The subree corresponding o he regions R and R conains purely class, represening he well-known linearly separable class in he Iris daa. The char conaining he relaive imporance of he predicor variables implies ha he peal lengh has he highes imporance wih regards o he densiy, whereas, he sepal lengh has no influence on he densiy (example of feaure deleion). Thus we see ha densiy esimaion in he decision ree framework provides much more informaion abou he daa han jus he densiy esimaes. There are several oher cases where i would be useful o ge his kind of informaion abou he daa: ()Asronomy performing densiy esimaion on quasar daa is a common ask in asronomy [8]. Common feaures in he daa are colors. The feaure-imporance analysis corresponding o he densiy will indicae he unimporan color feaures. This informaion can be useful in deciding which filers are useful in fuure analyses, hereby, conserving resources. ()Bioinformaics consider a meabolomics daa se conaining pa- 4 We will coninue o represen he size of daa ses wih poin wih d aribues/dimensions in he form ( d) hroughou his paper. 5 We inend o perform densiy esimaion only on he poins in he unsupervised seing. However, we provide class memberships of he poins in he node o moivae a well known propery of he Iris daa se. 69

4 ien informaion wih over feaures, where he ask is o differeniae cancer paiens from healhy paiens [9]. Densiy esimaion performed over all paiens profiles can reveal which feaures are responsible for he main variaion wihin, say, cancer paiens, or wihin healhy paiens. This is differen from feaure selecion for classificaion which would selec feaures which mos differeniae he paiens wih cancer from he res. In addiion, cluser and oulier deecion performed on his daa se will produce ineresing and possibly crucial informaion. Moreover, a represenaive pariion of he sae space by iself sheds ligh on he underlying srucure of he disribuion. Such informaion is paricularly valuable in high dimensional problems where direc visualizaion of he daa is difficul. Remark. DETs are similar o variable bin-widh hisograms, bu are resriced only o a hierarchical pariioning of he daa (hence possibly having lower accuracy han he bes possible variable bin-widh hisogram, alhough our experimens sugges oherwise).. Overview The focus of his paper is o develop a rigorous decisionree-srucured densiy esimaor and demonsrae he usefulness and inerpreabiliy of he same. In Secion, we discuss exising connecions beween densiy esimaion and decision rees. In he following secion, we define an esimaor based on he decision ree and apply he decision-ree framework for densiy esimaion using his esimaor. We also provide cerain asympoic properies of he decision-ree-srucured esimaor. Secion 4 conains experimenal resuls displaying he performance of DETs wih some comparisons wih hisograms and KDEs. The adapabiliy of DETs is demonsraed by using some synheic daa ses. High dimensional image daa ses are used o demonsrae he DETs inerpreabiliy and heir applicaion o classificaion. Training and querying ime of DETs are compared wih he fases raining and querying mehod for KDEs (KDE s). In he final secion, conclusions are discussed along wih cerain open quesions.. FURTHER RELATED WORK Decision rees have been used alongside densiy esimaion in earlier works. For example, Kohavi, 996 [] and Smyh e al., 995 [] use decision rees for he supervised ask of classificaion, and densiy esimaion (wih naïve Bayes classifiers and KDEs respecively) is used solely for he purpose of obaining a smooher and more accurae classificaion and class-condiional probabiliies in conras o he piecewise consan non-smooh esimaes of he sandard classificaion rees. Decision rees been used for he ask of diagnosing exrapolaion [] by building a classificaion ree differeniaing he given daa se from a random sample from a uniform disribuion in he suppor of he daa se. This ree provides a way o compue a measure for exrapolaion. This measure can be indirecly used o compue he densiy, bu he ree sill performs classificaion. Decision rees have also been used in he supervised seing for esimaing he join probabiliy (of he daa and he label of he raining examples) [] for censored daa by replacing he sandard loss funcion (for example - loss or Gini-index) wih he negaive loglikelihood of he join densiy. The idea of having a nesed hierarchy of pariions of he suppor of he daa has been used for dicreizaion of univariae daa [4]. The se of observaions are pariioned on he basis of he densiy using he loglikelihood loss funcion. Bu he focus is solely on univariae daa. Decision rees have also been used in an unsupervised seing o perform hierarchical clusering [5] bu do no rivially ranslae o densiy esimaion. Siedl e al., 9 [6] propose a novel mehod for indexing he densiy model in he form of a ree (called Bayes ree) for fas access wih desired level of accuracy. This ree is grown boom-up from he esimae of a KDE forhewholedaase. TheBayesreeusesMoGsoindex he densiy a he inermediae levels (increasing he number of Gaussians wih he deph). This ree successfully locaes clusers in he daa. However, being a MoG, i fails o deermine relaive relevance of he dimensions. Pólya rees [7] are hierarchical spaial pariioning rees bu used as a prior for probabiliy measures over he daa space. Densiy esimaion is done eiher by compuing he poserior mean densiy given his prior or learning a fixed ree opology and compuing he piecewise consan esimae condiional on his opology. DETs are relaed in spiri, however, hey will be learned by direcly minimizing he densiy esimaion loss and he esimaes are finally obained direcly from he ree. Moreover, he aforemenioned mehods, hough hierarchical in naure, do no share he inerpreabiliy and feaure selecion properies of decision rees, which are based on univariae splis.. DESITY ESTIMATIO TREE This secion provides he road map o perform densiy esimaion using decision rees. We define an esimaor and he corresponding loss funcion o be minimized during he raining of he decision ree for coninuous and mixed daa. Following ha, we explain he process of learning he opimal decision ree over he given sample and provide an asympoic propery of he DET.. Coninuous Feaures Definiion. The piecewise consan densiy esimae of he decision ree T buil on a se S of observaions in R d is defined as ˆf (x) = l I(x l ) () l T where T is he se of leaves of he decision ree T represening he pariions of he daa space, l is he number of observaions of S in he leaf l, V l is he volume of he leaf l wihin he d dimensional bounding box of S and I( ) is he indicaor funcion. A decision ree T requires a noion of a loss funcion R(T ) which is minimized using a greedy algorihm o consruc a ree on he se of observaions. For he unsupervised ask of densiy esimaion, we consider he Inegraed Squared Error (ISE) loss funcion []. The ISE gives a noion of overall disance beween he esimaed and he rue densiy and is a favored choice in nonparameric densiy esimaion for is inheren robusness in comparison o maximum-likelihoodbased loss funcions [8]. However, oher disance funcions such as he KL-divergence can be used as he loss funcion. We will explore his in he longer version of he paper. 6

5 The ask of learning a DET would involve solving he following opimizaion problem: ( ) ˆf (x) f(x) dx () min ˆf F X where F is he class of esimaors of form in Definiion ha can learned wih any se of observaions. Afer expanding he square and he following Mone-Carlo subsiuion ˆf X (x)f(x)dx ˆf i= (X i)(where{x i} i= is he raining se), he objecive funcion in Eq. is replaced by he following consisen plug-in esimaor of he ISE [] 6 : { min ˆf F X ( ) ˆf (x) dx } ˆf (X i) Using he piecewise consan esimaor from Definiion (which is consan wihin each leaf l ), he objecive funcion in Eq. is replaced wih he following: { ( l l T by subsiuing l ) dx ˆf (x) = l T ( l i= l l ) I(x l ) (since he cross erms in he expansion of ˆf (x) vanishbecause of he indicaor funcion) and simple calculaion shows ha l ˆf (X i)= I(X i l )= l l i= i= l T l T By making he following subsiuion ( ) ( ) ( ) l l l dx = dx = V l, l he esimaor of he ISE for DET has he following form: } l { (5) V l l T Defining Eq. 5 as he error R(T ) of he ree, he greedy surrogae of he error for any node (inernal or oherwise) can be defined as R() =. (6) V The ree is grown in a op down manner by maximizing he reducion in his greedy surrogae of he error over he given observaions.. Mixed Feaures For densiy esimaion over daa wih mixed feaures, we define a novel densiy esimaor and a loss funcion involving ordinal and caegorical daa along wih coninuous daa ha can be used o learn DETs wih mixed daa: 6 The erm X (f(x)) dx is removed from he objecive funcion since i is independen of he esimae and hence doesn affec he opimum. l } () (4) Definiion. Le S R d Z d C d wih d real feaures, d ordinal feaures and d caegorical feaures. The piecewise consan densiy esimaor of he decision ree T buil on S is defined as ˆf (x) = l I(x l ) (7) l T V ld d l j= R l j d l i= M l i where V ld is he volume of he leaf l wihin he d dimensional bounding box of he real par of S, R lj is he range of he ordinal values in he j h of he d l ordinal dimensions presen in l and M li is he number of caegories presen in l for he i h of he d l caegorical dimensions presen in l. Theerroraanode corresponding o he ISE is hen obained as R() = V d d j= R j d i= M i where V d is he volume of he node in he d real dimensions, R j is he range of he ordinal values in he j h of he d ordinal dimensions presen in and M i is he number of caegories presen in for he i h of he d caegorical dimensions presen in.. Tree Consrucion We use he ree learning algorihm presened in Breiman, e al.,984 []. The spliing, pruning and he cross-validaion procedures are modified o work wih his new loss funcion and for he unsupervised ask of densiy esimaion. Spliing. For growing he ree, each node is spli ino wo children. Le S be he se of all univariae splis. Definiion. [] The bes spli s of a node is he spli in he se of splis S which maximally reduces R(T ). This is done by greedily reducing R() for all he erminal nodes of he curren ree. Hence, for any currenly erminal node, we need o find a spli s for a node ino L and R such ha (8) s =argmax{r() R(L) R(R)} (9) s S where = L + R. For coninuous and ordinal dimensions, his opimizaion is performed by rying every possible splis of he daa in each of he dimensions. For caegorical dimensions, he splis are performed in he manner similar o he CART model []. The spliing is sopped when he node size goes below a cerain hreshold (say min ). Pruning. To avoid overfiing, we use he minimal coscomplexiy pruning []. The regularized error of a subree rooed a a node is defined as R α() =R( )+α () where α is a regularizaion parameer (o be esimaed hrough cross-validaion) and is he se of leaves in he subree rooed a. The value of α is gradually increased and a subree rooed a is pruned for he value of α where R α() =R α({}), he regularized error of he pruned subree. Since he size of he iniial ree consruced by he spliing algorihm described previously is finie, he number of possible values of α a which a prune occurs is finie 6

6 and can be calculaed efficienly. Hence we only need o selec he opimal α from a finie se of values (See Secion. in Breiman, e al., 984 [] for complee deails). Cross-validaion. The leave-one-ou cross-validaion (LOO-CV) esimaor of he densiy esimaion loss funcion in Eq. is given by ( ) Ĵ(α) = ˆf α (x) dx ˆf α ( i)(x i) () X where ˆf α is he esimaor wih he decision ree T pruned wih a regularizaion parameer α, and ˆf α ( i) is he esimaor wih he decision ree T ( i) buil wihou he raining example X i pruned wih he regularizaion parameer α. This LOO-CV esimaor is obained from Silverman, 986 [] by swiching he regularizaion parameer (replacing he bandwidh wih α). The bes sized ree is he ree T pruned wih he parameer α such ha: i= α =argmin α Ĵ(α) ().4 Asympoic Properies We show ha he mehod is nonparameric iisconsisen under mild assumpions on he model class of he inpu disribuion f. Consisency is ypically shown [8] for a nonparameric esimaor ˆf obained from a se of observaions by showing ha ( ( ) Pr ˆf (x) f(x)) dx = =. () lim X We prove he consisency of DETs on daa wih coninuous feaures. The proof of consisency of he densiy esimaor proposed in Definiion follows argumens similar o hose used o show he consisency of regression rees []. Theorem. The esimaor ˆf defined in Definiion saisfies Eq.. Proof. Given a fixed posiive ineger d,lebdenoe he collecion of all ses X ha can be described as he soluion se o a sysem of d inequaliies, each of he form b T x c for b R d and c R. ow in a decision ree T,everyleafl T is he soluion se of a sysem of d inequaliies of he form b T x c wih b R d wih jus one enry equal o and he res of he enries equal o. Therefore, T B. Le X n, n, be a random sample from a densiy funcion f on X. For, le ˆF denoe he empirical disribuion of X n, n, definedonase X by ˆF () = I(X n ) = = ˆf (x)dx (4) n= where is he number of random samples in he se {X n} n= and ˆf (x) is he esimaor given in Definiion. According o a general version of he Glivenko-Canelli heorem [9], ( ) Pr lim sup ˆF () f(x)dx = =. (5) B By Eq.4 and 5, we ge ( Pr lim ˆf (x)dx sup B ) f(x)dx = = ( ) Pr lim sup ˆf (x) f(x) dx =. B Assuming ha lim Pr(diameer() ɛ) =, hence Pr ( lim dx =) =, we ge he following wih probabiliy lim sup ˆf (x) f(x) dx B lim ˆf (x ) f(x ) dx for some x = This assumpion is commonly used for he consisency of daa-pariioning esimaors [] and is jusified since as, he diameer of any leaf node would become smaller and smaller since he leaf node can only have a bounded number of poins. Hence ( ) Pr lim sup ˆf (x) f(x) dx = = B ( ( ) Pr lim ˆf (x) f(x)) dx = =. X Hence ˆf saisfies Eq.. 4. EXPERIMETS In his secion, we demonsrae he performance of DETs under differen condiions using synheic and real daa ses. Esimaion accuracies of DETs are compared wih exising nonparameric esimaors. We exhibi he inerpreabiliy of DETs wih wo real daa ses. Furhermore, densiy esimaes of DETs are applied o classificaion and subsequen labelling accuracies are presened. Finally, he speed of raining and querying DETs are presened on several real daa ses and compared wih an exising mehod. We only consider coninuous and ordinal daa in his paper for he lack of space. Experimens wih caegorical and mixed daa will be presened in he longer version of he paper. 4. Esimaion Accuracy: Synheic Examples For he unsupervised ask of densiy esimaion, esimaion accuracy can only be compued on synheic daa ses. We choose he bes-sized ree hrough LOO-CV. To compue esimaion error on synheic daa for densiy queries, we use measures: Roo Mean Squared Error (RMSE) and Hellinger Disance (HD). We compare DETs only wih oher nonparameric densiy esimaors since parameric esimaors would involve choosing he model class (like choosing he number of Gaussians in MoGs). We repor he comparison beween DETs, hisograms (bin-widh seleced by LOO-CV) and KDEs (using he Gaussian kernel) wih he bandwidh seleced by unbiased LOO-CV using he ISE crierion (FBW) and by local rodeo (local r-kde) (VBW) on synheic daa ses. Example. (Srongly skewed disribuion in one dimension) This disribuion is defined as X 7 i= 8 ( ( ( ) i ), ( ) ) i. Figure shows he esimaed densiy funcions by a DET, a hisogram and wo KDEs for a sample size =. Because of he high skewness of he densiy funcion, he KDE (FBW) and he hisogram fail o fi he very smooh 6

7 .4.4 DET True Densiy Hisogram True Densiy Densiy densiy densiy x.5. (a) DET.4.8 Irrelevan dimension Relevan dimension (b) Hisogram x (a) DET (b) Hisogram.4 KDE True Densiy RODEO KDE True Densiy x RMSE HD His KDE(FBW) local r-kde x Irrelevan dim Relevan dim Irrelevan dimension Relevan dimension (c) KDE (FBW) (d) local r-kde Figure : Perspecive plos of he densiy esimaes for a DET, a Hisogram and wo KDEs on he daa se in Example. Figure shows he -dimensional densiy esimaes of a DET, a hisogram and wo KDEs. The DET fis he irrelevan uniform dimension perfecly, while closely approximaing he mixure of Bea disribuions in he relevan dimension. The KDE wih fixed bandwidh and he hisogram esimaor compleely fail o fi he irrelevan dimension. The local r-kde does a beer job of fiing he irrelevan dimension compared o he KDE (FBW), bu sill does no enirely capure he uniform densiy. DET ail. The fixed bandwidh/binwidh resuls in a highly wiggly esimae of he very smooh ail. The DET provides a piecewise consan esimae, bu i adjuss o he differen pars of he densiy funcion by varying he leaf size he ree has small leaves closer o he spike where he densiy funcion changes rapidly, having larger leaves where he densiy funcion does no change rapidly. This demonsraes he adapabiliy of DETs wihin he same dimension. The local r-kde exhibis he same adapabiliy by obaining differen bandwidhs a differen regions, hence capuring he smooh ail of he rue densiy. Table provides he esimaion errors wih increasing sample sizes for he differen mehods. The RMSE values are fairly close and he HD values for he DETs are in fac beer han ha of KDE using unbiased LOO-CV. This can be aribued o he adapive naure of DET wih respec o he leaf sizes. The error values of local r-kde are beer han DET when he daa se size is small (in case of larger daa ses, he local mehod migh be overfiing over his insance of he daa se). The accuracies of he hisogram esimaors and DETs are comparable in his univariae example. Bu we will demonsrae in he nex example ha he performance of hisograms decline wih increasing dimensions. Example. (Two dimensional daa Mixure of Bea disribuions in one dimension, uniform in he oher dimension) We creae a (6 ) daa se by sampling one dimension as a mixure of Bea disribuions and he oher dimension as a uniform disribuion X.5 Table : Esimaion errors for Hisograms, KDEs and DETs wih increasing number of observaions (c) KDE (FBW) (d) local r-kde Figure : The densiy esimaes obained wih poins for Example using DET, Hisogram and KDEs compared o he rue densiy. Type Densiy.6 Densiy densiy densiy.5 4. Inerpreabiliy: Variable Imporance The decision-ree framework defines a measure of relevance for each of he predicor variables [] as following: Definiion 4. The measure of relevance Id (T ) for each aribue Xd (he d h aribue) in a decision ree T is defined as T Id (T ) = ι I(d( ) = d ) (6) = where he summaion is over he T inernal nodes of he ree T, and he aribue d( ) used o pariion node obains ι improvemen in squared error over he consan fi in node. We use his measure o presen he inerpreabiliy of he DETs on wo real daa ses. Example. (Iris daa se) We perform densiy esimaion on his 4 dimensional daa se and compue variable imporance for each of he dimensions. Figure displays he ree and he relaive variable imporance. The inerpreaion has been explained in Secion.. Example 4. (MIST digis image daa) Each image has 8-by-8 pixels providing a 784-dimensional daa se. We perform densiy esimaion on he 585 images of he digi 8 (lef panel of Figure 4). This is also an insance of performing densiy esimaion wih ordinal daa since he pixel values are discree. We compue he variable imporance of each of he dimensions (pixels in his case). The righ panel of Figure 4 displays he resuls obained. The black pixels indicae pixels wih zero variable imporance in densiy B(, ) + B(, ); X U(, ). 6

8 Table 4: Timings (in seconds) for raining(t) and all queries(q) wih KDE sanddets Daa se T(KDE ) T(DET) Q(KDE ) Q(DET) sj Colors Bio Pall psf Speedups in query ime (a) Originals (b) Variable Imporance Figure 4: Relaive imporance of he predicor variables (pixels in his case) in densiy esimaion for images of digi 8. Table : Classificaion Accuracy: r-kde vs. DET Case #Train # Tes r-kde DET v v v v v All esimaion. For he pixels wih non-zero variable imporance, he color ransiions for green (indicaing low relaive variable imporance) o red (indicaing high relaive variable imporance). The marginal densiies of many pixels are close o poin masses. Hence he densiy does no vary a all in hose dimensions, and he relaive variable imporance depiced in Figure 4 indicaes ha he DET is capable of capuring his propery of he daa. 4. Classificaion As menioned earlier, densiy esimaion is a common subrouine in classificaion where he class-condiional densiy ˆp(x C) is o be esimaed for he query poin x using he raining se X. The esimaed class Ĉ(x) = arg max C P (x C) P (C). We compare he accuracy of classificaion beween he densiy esimaes compued by DETs and local r-kdes. Example 5. (Op-digis image daa (797 64) []) We conduc binary as well as muli-class classificaion. For binary classificaion, we consider he following cases: vs. 7, vs. 7, vs. 8, 5 vs. 8, 8 vs. 9. We perform muliclass classificaion using all he classes (-9). Table liss he classificaion accuracies for he differen asks. In mos cases, he accuracies are close for he wo mehods of densiy esimaion, wih he r-kdes being more accurae han he DETs. Bu in he las wo cases, he DETs ouperform he r-kdes. Moreover, he whole experimen (raining and esing) wih he DETs is much faser han he one using he local r-kdes. For example, in he las experimen performing muli-class classificaion required 8 seconds using DETs and 5 seconds using r-kdes Speed We compare DETs o KDE s(fbw) using he fas approximae dual-ree algorihm [] wih kd-rees. We use LOO-CV o esimae he opimal bandwidh for he KDE s 7 The DET is implemened in C++ and he rodeo-kde is implemened using MATLAB on he same machine. Speedup sj colors bio pall7 psf Figure 5: Query Time speedup for he DTE over he KDE. oe ha he KDE is faser han he DTE in case of raining. (FBW). Real world daa ses drawn from he SDSS reposiory [] (SJ (5 ), PSF ( )) and he UCI Machine Learning Reposiory [] (Colors (5 ), dimension reduced version of he Bio daase ( 5), Pall7 ( 7)) are used. The query se is same as he raining se in all he experimens and he KDE sprovide he leave-one-ou esimaes. For he KDE, we use around H bandwidhs and chose he one wih he bes CV error. Usually, a much larger number of bandwidhs are ried and he raining imes scale linearly wih H. For he DETs, we use -fold crossvalidaion since we were no able o find an efficien way o conduc LOO-CV wih decision rees and i would require prohibiively long ime for hese large daa ses. This is one major limiaion of his mehod. Howeover, for sufficienly large daa ses, our experimens indicae ha DETs rained wih LOO-CV are only slighly more accurae han hose wih -fold CV. The absolue iming values (in seconds) for raining(t) and querying(q) are given in Table 4. The speedups of he DETs over he KDE s in es/query ime are shown in Figure 5. As can be seen from he resuls, he raining ime for he decision ree algorihm is significanly larger han he LOO-CV for he KDE s, bu he DETs provide up o.5 orders of magniude speedup in query ime. The KDE s are fas o rain, bu heir work comes a es ime. This experimen demonsraes he efficien querying of decision rees while also indicaing ha he ask of raining decision rees using LOO-CV is quie challenging. 5. DISCUSSIO AD COCLUSIOS This framework of decision-ree-srucured densiy esimaion provides a new way o esimae he densiy of a given se of observaions in he form of a simple se of rules which are inherenly inuiive and inerpreable. This framework has he capabiliy of easily dealing wih caegorical, discree as well as coninuous variables and performs auomaic fea- 64

9 ure selecion. On having hese rules, he densiy of a new es poin can be compued cheaply. All hese feaures sem from he fac ha he densiy esimaion is performed in he form of a decision ree. Along wih ha, he DETs are easily implemenable like supervised decision rees. Alhough hese advanages come wih he loss of accuracy, he DETs are shown o be consisen, and hence are quie reliable in he presence of enough daa. Fuure direcions of improvemen include he reducion of he disconinuiies in he densiy esimae because of he piecewise consan esimaor and he boundary bias since he DETs pu no mass ouside he span of he daa. Boundary bias is a common problem for almos all densiy esimaors. One possible remedy is o use a normalized KDE a each of he leaves insead of using he piecewise consan esimaor (similar o he approach in Smyh e.al. []. An analysis of he convergence rae of he DETs would quanify he amoun of loss of accuracy o accoun for he simpliciy of he esimaor. Being effecively a variable binwidh hisogram, we conjecure ha he convergence rae for he DETs would be o( / ) for univariae daa and beer han hisograms in higher dimensions since we demonsrae ha he DETs can effecively model uninformaive dimensions easily wihou requiring ha exra number of poins as imposed by he curse of dimensionaliy. Moreover, we are acively working on obaining densiy dependen bounds on he deph of he DETs o quanify he runimes for raining and es query. Overall, his mehod for densiy esimaion has immediae applicaion o various fields of daa analysis (for example, oulier and anomaly deecion) 8 and machine learning due o is simple and inerpreable soluion o he fundamenal ask of densiy esimaion. 6. REFERECES [] L. Breiman, J. H. Friedman, R. A. Olshen, and C. J. Sone. Classificaion and Regression Trees. Wadsworh, 984. [] C. Bishop. eural eworks for Paern Recogniion. Oxford Universiy Press, 995. [] C. M. Bishop. Paern recogniion and machine learning. Springer, 6. [4] R.O.DudaandP.E.Har.Paern Classificaion and Scene Analysis. Wiley ew York, 97. [5] J. H. Friedman. A Recursive Pariioning Decision Rule for onparameric Classificaion. Transacions on Compuers, 977. [6] J. H. Friedman. A ree-srucured approach o nonparameric muliple regression. Smoohing Techniques for Curve Esimaion, 979. [7] T. Hasie, R. Tibshirani, and J. Friedman. The Elemens of Saisical Learning: Daa Mining, Inference, and Predicion. Springer,. [8] L. Wasserman. All of onparameric Saisics. Physica-Verlag, 6. [9] C. Cores and V.. Vapnik. Suppor vecor neworks. Machine Learning, 995. [] B. W. Silverman. Densiy Esimaion for Saisics and Daa Analysis. Chapman & Hall/CRC, 986. [] H. Liu, J. Laffery, and L. Wasserman. Sparse onparameric Densiy Esimaion in High Dimensions Using he Rodeo. In AISTATS, 7. [] P. Müller and F.A. Quinana. onparameric Bayesian daa analysis. Saisical Science, 4. [] A. G. Gray and A. W. Moore. onparameric densiy esimaion: Toward compuaional racabiliy. In SIAM Daa Mining,. [4] D. Lee and A. G. Gray. Fas high-dimensional kernel summaions using he mone carlo mulipole mehod. In Advances in IPS, 8. [5] A. Beygelzimer, S. Kakade, and J.C. Langford. Cover Trees for eares eighbor. ICML, 6. [6] P. Ram, D. Lee, W. March, and A. Gray. Linear-ime algorihms for pairwise saisical problems. In Advances in IPS. 9. [7] J. Laffery and L. Wasserman. Rodeo: Sparse onparameric Regression in High Dimensions. Arxiv preprin mah.st/564, 5. [8] R. Riegel, A. G. Gray, and G. Richards. Massive-Scale Kernel Discriminan Analysis: Mining for Quasars. In SIAM Daa Mining (SDM), 8. [9] W. Guan, M. Zhou, C. Y. Hampon, B. B. Benigno, L. D. Walker, A. G. Gray, J. F. McDonald, and F. M. Fernandez. Ovarian Cancer Deecion from Meabolomic Liquid Chromaography/Mass Specromery Daa by Suppor Vecor Machines. BMC Bioinformaics Journal, 9. [] R. Kohavi. Scaling up he accuracy of naive-bayes classifiers: A decision-ree hybrid. In KDD, 996. [] P. Smyh, A. G. Gray, and U. M. Fayyad. Rerofiing decision ree classifiers using kernel densiy esimaion. In ICML, 995. [] G. Hooker. Diagnosing exrapolaion: Tree-based densiy esimaion. In SIGKDD, 4. [] A.M. Molinaro, S. Dudoi, and M.J. Van der Laan. Tree-based mulivariae regression and densiy esimaion wih righ-censored daa. Journal of Mulivariae Analysis, 4. [4] G. Schmidberger and E. Frank. Unsupervised discreizaion using ree-based densiy esimaion. Lecure oes in Compuer Science, 5. [5] J. Basak and R. Krishnapuram. Inerpreable hierarchical clusering by consrucing an unsupervised decision ree. IEEE ransacions on knowledge and daa engineering, 5. [6] T. Seidl, I. Assen, P. Kranen, R. Krieger, and J. Herrmann. Indexing densiy models for incremenal learning and anyime classificaion on daa sreams. In ICEDT: Advances in Daabase Technology, 9. [7] W.H. Wong and L. Ma. Opional Pólya ree and Bayesian inference. The Annals of Saisics,. [8] E. Parzen. On he esimaion of a probabiliy densiy funcion and mode. Annals of Mahemaical Saisics, pages 65 76, 96. [9] V.. Vapnik and A. Y. Chervonenkis. On he uniform convergence of relaive frequencies of evens o heir probabiliies. Theory of Probabiliy and is Applicaions, 97. [] L. Devroye, L. Györfi, and G. Lugosi. A probabilisic heory of paern recogniion. Springer Verlag, 996. [] C.L. Blake and C.J. Merz. UCI reposiory of machine learning daabases [] A. G. Gray and A. W. Moore. n-body problems in saisical learning. In Advances in IPS,. [] R. Lupon, J.E. Gunn, Z. Ivezic, G.R. Knapp, S. Ken, and. Yasuda. The SDSS Imaging Pipelines. Arxiv preprin asro-ph/4,. 8 We will demonsrae hese applicaions in he longer version of he paper. 65

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LTU, decision

More information

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important

Non-parametric techniques. Instance Based Learning. NN Decision Boundaries. Nearest Neighbor Algorithm. Distance metric important on-parameric echniques Insance Based Learning AKA: neares neighbor mehods, non-parameric, lazy, memorybased, or case-based learning Copyrigh 2005 by David Helmbold 1 Do no fi a model (as do LDA, logisic

More information

INTRODUCTION TO MACHINE LEARNING 3RD EDITION

INTRODUCTION TO MACHINE LEARNING 3RD EDITION ETHEM ALPAYDIN The MIT Press, 2014 Lecure Slides for INTRODUCTION TO MACHINE LEARNING 3RD EDITION alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/~ehem/i2ml3e CHAPTER 2: SUPERVISED LEARNING Learning a Class

More information

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1

CptS 570 Machine Learning School of EECS Washington State University. CptS Machine Learning 1 CpS 570 Machine Learning School of EECS Washingon Sae Universiy CpS 570 - Machine Learning 1 Form of underlying disribuions unknown Bu sill wan o perform classificaion and regression Semi-parameric esimaion

More information

Lecture 33: November 29

Lecture 33: November 29 36-705: Inermediae Saisics Fall 2017 Lecurer: Siva Balakrishnan Lecure 33: November 29 Today we will coninue discussing he boosrap, and hen ry o undersand why i works in a simple case. In he las lecure

More information

Ensamble methods: Bagging and Boosting

Ensamble methods: Bagging and Boosting Lecure 21 Ensamble mehods: Bagging and Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Ensemble mehods Mixure of expers Muliple base models (classifiers, regressors), each covers a differen par

More information

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H.

ACE 562 Fall Lecture 5: The Simple Linear Regression Model: Sampling Properties of the Least Squares Estimators. by Professor Scott H. ACE 56 Fall 005 Lecure 5: he Simple Linear Regression Model: Sampling Properies of he Leas Squares Esimaors by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Inference in he Simple

More information

Vehicle Arrival Models : Headway

Vehicle Arrival Models : Headway Chaper 12 Vehicle Arrival Models : Headway 12.1 Inroducion Modelling arrival of vehicle a secion of road is an imporan sep in raffic flow modelling. I has imporan applicaion in raffic flow simulaion where

More information

Chapter 2. Models, Censoring, and Likelihood for Failure-Time Data

Chapter 2. Models, Censoring, and Likelihood for Failure-Time Data Chaper 2 Models, Censoring, and Likelihood for Failure-Time Daa William Q. Meeker and Luis A. Escobar Iowa Sae Universiy and Louisiana Sae Universiy Copyrigh 1998-2008 W. Q. Meeker and L. A. Escobar. Based

More information

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED

0.1 MAXIMUM LIKELIHOOD ESTIMATION EXPLAINED 0.1 MAXIMUM LIKELIHOOD ESTIMATIO EXPLAIED Maximum likelihood esimaion is a bes-fi saisical mehod for he esimaion of he values of he parameers of a sysem, based on a se of observaions of a random variable

More information

Ensamble methods: Boosting

Ensamble methods: Boosting Lecure 21 Ensamble mehods: Boosing Milos Hauskrech milos@cs.pi.edu 5329 Senno Square Schedule Final exam: April 18: 1:00-2:15pm, in-class Term projecs April 23 & April 25: a 1:00-2:30pm in CS seminar room

More information

Robust estimation based on the first- and third-moment restrictions of the power transformation model

Robust estimation based on the first- and third-moment restrictions of the power transformation model h Inernaional Congress on Modelling and Simulaion, Adelaide, Ausralia, 6 December 3 www.mssanz.org.au/modsim3 Robus esimaion based on he firs- and hird-momen resricions of he power ransformaion Nawaa,

More information

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still.

Lecture 2-1 Kinematics in One Dimension Displacement, Velocity and Acceleration Everything in the world is moving. Nothing stays still. Lecure - Kinemaics in One Dimension Displacemen, Velociy and Acceleraion Everyhing in he world is moving. Nohing says sill. Moion occurs a all scales of he universe, saring from he moion of elecrons in

More information

STATE-SPACE MODELLING. A mass balance across the tank gives:

STATE-SPACE MODELLING. A mass balance across the tank gives: B. Lennox and N.F. Thornhill, 9, Sae Space Modelling, IChemE Process Managemen and Conrol Subjec Group Newsleer STE-SPACE MODELLING Inroducion: Over he pas decade or so here has been an ever increasing

More information

Random Walk with Anti-Correlated Steps

Random Walk with Anti-Correlated Steps Random Walk wih Ani-Correlaed Seps John Noga Dirk Wagner 2 Absrac We conjecure he expeced value of random walks wih ani-correlaed seps o be exacly. We suppor his conjecure wih 2 plausibiliy argumens and

More information

Longest Common Prefixes

Longest Common Prefixes Longes Common Prefixes The sandard ordering for srings is he lexicographical order. I is induced by an order over he alphabe. We will use he same symbols (,

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power

Learning a Class from Examples. Training set X. Class C 1. Class C of a family car. Output: Input representation: x 1 : price, x 2 : engine power Alpaydin Chaper, Michell Chaper 7 Alpaydin slides are in urquoise. Ehem Alpaydin, copyrigh: The MIT Press, 010. alpaydin@boun.edu.r hp://www.cmpe.boun.edu.r/ ehem/imle All oher slides are based on Michell.

More information

Retrieval Models. Boolean and Vector Space Retrieval Models. Common Preprocessing Steps. Boolean Model. Boolean Retrieval Model

Retrieval Models. Boolean and Vector Space Retrieval Models. Common Preprocessing Steps. Boolean Model. Boolean Retrieval Model 1 Boolean and Vecor Space Rerieval Models Many slides in his secion are adaped from Prof. Joydeep Ghosh (UT ECE) who in urn adaped hem from Prof. Dik Lee (Univ. of Science and Tech, Hong Kong) Rerieval

More information

Online Convex Optimization Example And Follow-The-Leader

Online Convex Optimization Example And Follow-The-Leader CSE599s, Spring 2014, Online Learning Lecure 2-04/03/2014 Online Convex Opimizaion Example And Follow-The-Leader Lecurer: Brendan McMahan Scribe: Sephen Joe Jonany 1 Review of Online Convex Opimizaion

More information

12: AUTOREGRESSIVE AND MOVING AVERAGE PROCESSES IN DISCRETE TIME. Σ j =

12: AUTOREGRESSIVE AND MOVING AVERAGE PROCESSES IN DISCRETE TIME. Σ j = 1: AUTOREGRESSIVE AND MOVING AVERAGE PROCESSES IN DISCRETE TIME Moving Averages Recall ha a whie noise process is a series { } = having variance σ. The whie noise process has specral densiy f (λ) = of

More information

Christos Papadimitriou & Luca Trevisan November 22, 2016

Christos Papadimitriou & Luca Trevisan November 22, 2016 U.C. Bereley CS170: Algorihms Handou LN-11-22 Chrisos Papadimiriou & Luca Trevisan November 22, 2016 Sreaming algorihms In his lecure and he nex one we sudy memory-efficien algorihms ha process a sream

More information

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis

Speaker Adaptation Techniques For Continuous Speech Using Medium and Small Adaptation Data Sets. Constantinos Boulis Speaker Adapaion Techniques For Coninuous Speech Using Medium and Small Adapaion Daa Ses Consaninos Boulis Ouline of he Presenaion Inroducion o he speaker adapaion problem Maximum Likelihood Sochasic Transformaions

More information

An introduction to the theory of SDDP algorithm

An introduction to the theory of SDDP algorithm An inroducion o he heory of SDDP algorihm V. Leclère (ENPC) Augus 1, 2014 V. Leclère Inroducion o SDDP Augus 1, 2014 1 / 21 Inroducion Large scale sochasic problem are hard o solve. Two ways of aacking

More information

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017

Two Popular Bayesian Estimators: Particle and Kalman Filters. McGill COMP 765 Sept 14 th, 2017 Two Popular Bayesian Esimaors: Paricle and Kalman Filers McGill COMP 765 Sep 14 h, 2017 1 1 1, dx x Bel x u x P x z P Recall: Bayes Filers,,,,,,, 1 1 1 1 u z u x P u z u x z P Bayes z = observaion u =

More information

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB

T L. t=1. Proof of Lemma 1. Using the marginal cost accounting in Equation(4) and standard arguments. t )+Π RB. t )+K 1(Q RB Elecronic Companion EC.1. Proofs of Technical Lemmas and Theorems LEMMA 1. Le C(RB) be he oal cos incurred by he RB policy. Then we have, T L E[C(RB)] 3 E[Z RB ]. (EC.1) Proof of Lemma 1. Using he marginal

More information

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD

PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD PENALIZED LEAST SQUARES AND PENALIZED LIKELIHOOD HAN XIAO 1. Penalized Leas Squares Lasso solves he following opimizaion problem, ˆβ lasso = arg max β R p+1 1 N y i β 0 N x ij β j β j (1.1) for some 0.

More information

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle

Physics 235 Chapter 2. Chapter 2 Newtonian Mechanics Single Particle Chaper 2 Newonian Mechanics Single Paricle In his Chaper we will review wha Newon s laws of mechanics ell us abou he moion of a single paricle. Newon s laws are only valid in suiable reference frames,

More information

Notes on Kalman Filtering

Notes on Kalman Filtering Noes on Kalman Filering Brian Borchers and Rick Aser November 7, Inroducion Daa Assimilaion is he problem of merging model predicions wih acual measuremens of a sysem o produce an opimal esimae of he curren

More information

Tasty Coffee example

Tasty Coffee example Lecure Slides for (Binary) Classificaion: Learning a Class from labeled Examples ITRODUCTIO TO Machine Learning ETHEM ALPAYDI The MIT Press, 00 (modified by dph, 0000) CHAPTER : Supervised Learning Things

More information

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles

Diebold, Chapter 7. Francis X. Diebold, Elements of Forecasting, 4th Edition (Mason, Ohio: Cengage Learning, 2006). Chapter 7. Characterizing Cycles Diebold, Chaper 7 Francis X. Diebold, Elemens of Forecasing, 4h Ediion (Mason, Ohio: Cengage Learning, 006). Chaper 7. Characerizing Cycles Afer compleing his reading you should be able o: Define covariance

More information

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé

Bias in Conditional and Unconditional Fixed Effects Logit Estimation: a Correction * Tom Coupé Bias in Condiional and Uncondiional Fixed Effecs Logi Esimaion: a Correcion * Tom Coupé Economics Educaion and Research Consorium, Naional Universiy of Kyiv Mohyla Academy Address: Vul Voloska 10, 04070

More information

2. Nonlinear Conservation Law Equations

2. Nonlinear Conservation Law Equations . Nonlinear Conservaion Law Equaions One of he clear lessons learned over recen years in sudying nonlinear parial differenial equaions is ha i is generally no wise o ry o aack a general class of nonlinear

More information

Understanding the asymptotic behaviour of empirical Bayes methods

Understanding the asymptotic behaviour of empirical Bayes methods Undersanding he asympoic behaviour of empirical Bayes mehods Boond Szabo, Aad van der Vaar and Harry van Zanen EURANDOM, 11.10.2011. Conens 2/20 Moivaion Nonparameric Bayesian saisics Signal in Whie noise

More information

Chapter 2. First Order Scalar Equations

Chapter 2. First Order Scalar Equations Chaper. Firs Order Scalar Equaions We sar our sudy of differenial equaions in he same way he pioneers in his field did. We show paricular echniques o solve paricular ypes of firs order differenial equaions.

More information

Overview. COMP14112: Artificial Intelligence Fundamentals. Lecture 0 Very Brief Overview. Structure of this course

Overview. COMP14112: Artificial Intelligence Fundamentals. Lecture 0 Very Brief Overview. Structure of this course OMP: Arificial Inelligence Fundamenals Lecure 0 Very Brief Overview Lecurer: Email: Xiao-Jun Zeng x.zeng@mancheser.ac.uk Overview This course will focus mainly on probabilisic mehods in AI We shall presen

More information

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK

CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 175 CHAPTER 10 VALIDATION OF TEST WITH ARTIFICAL NEURAL NETWORK 10.1 INTRODUCTION Amongs he research work performed, he bes resuls of experimenal work are validaed wih Arificial Neural Nework. From he

More information

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II

Zürich. ETH Master Course: L Autonomous Mobile Robots Localization II Roland Siegwar Margaria Chli Paul Furgale Marco Huer Marin Rufli Davide Scaramuzza ETH Maser Course: 151-0854-00L Auonomous Mobile Robos Localizaion II ACT and SEE For all do, (predicion updae / ACT),

More information

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter

State-Space Models. Initialization, Estimation and Smoothing of the Kalman Filter Sae-Space Models Iniializaion, Esimaion and Smoohing of he Kalman Filer Iniializaion of he Kalman Filer The Kalman filer shows how o updae pas predicors and he corresponding predicion error variances when

More information

NCSS Statistical Software. , contains a periodic (cyclic) component. A natural model of the periodic component would be

NCSS Statistical Software. , contains a periodic (cyclic) component. A natural model of the periodic component would be NCSS Saisical Sofware Chaper 468 Specral Analysis Inroducion This program calculaes and displays he periodogram and specrum of a ime series. This is someimes nown as harmonic analysis or he frequency approach

More information

4.1 Other Interpretations of Ridge Regression

4.1 Other Interpretations of Ridge Regression CHAPTER 4 FURTHER RIDGE THEORY 4. Oher Inerpreaions of Ridge Regression In his secion we will presen hree inerpreaions for he use of ridge regression. The firs one is analogous o Hoerl and Kennard reasoning

More information

1 Review of Zero-Sum Games

1 Review of Zero-Sum Games COS 5: heoreical Machine Learning Lecurer: Rob Schapire Lecure #23 Scribe: Eugene Brevdo April 30, 2008 Review of Zero-Sum Games Las ime we inroduced a mahemaical model for wo player zero-sum games. Any

More information

CHERNOFF DISTANCE AND AFFINITY FOR TRUNCATED DISTRIBUTIONS *

CHERNOFF DISTANCE AND AFFINITY FOR TRUNCATED DISTRIBUTIONS * haper 5 HERNOFF DISTANE AND AFFINITY FOR TRUNATED DISTRIBUTIONS * 5. Inroducion In he case of disribuions ha saisfy he regulariy condiions, he ramer- Rao inequaliy holds and he maximum likelihood esimaor

More information

Introduction to Probability and Statistics Slides 4 Chapter 4

Introduction to Probability and Statistics Slides 4 Chapter 4 Inroducion o Probabiliy and Saisics Slides 4 Chaper 4 Ammar M. Sarhan, asarhan@mahsa.dal.ca Deparmen of Mahemaics and Saisics, Dalhousie Universiy Fall Semeser 8 Dr. Ammar Sarhan Chaper 4 Coninuous Random

More information

Sequential Importance Resampling (SIR) Particle Filter

Sequential Importance Resampling (SIR) Particle Filter Paricle Filers++ Pieer Abbeel UC Berkeley EECS Many slides adaped from Thrun, Burgard and Fox, Probabilisic Roboics 1. Algorihm paricle_filer( S -1, u, z ): 2. Sequenial Imporance Resampling (SIR) Paricle

More information

Final Spring 2007

Final Spring 2007 .615 Final Spring 7 Overview The purpose of he final exam is o calculae he MHD β limi in a high-bea oroidal okamak agains he dangerous n = 1 exernal ballooning-kink mode. Effecively, his corresponds o

More information

Georey E. Hinton. University oftoronto. Technical Report CRG-TR February 22, Abstract

Georey E. Hinton. University oftoronto.   Technical Report CRG-TR February 22, Abstract Parameer Esimaion for Linear Dynamical Sysems Zoubin Ghahramani Georey E. Hinon Deparmen of Compuer Science Universiy oftorono 6 King's College Road Torono, Canada M5S A4 Email: zoubin@cs.orono.edu Technical

More information

SUPPLEMENTARY INFORMATION

SUPPLEMENTARY INFORMATION SUPPLEMENTARY INFORMATION DOI: 0.038/NCLIMATE893 Temporal resoluion and DICE * Supplemenal Informaion Alex L. Maren and Sephen C. Newbold Naional Cener for Environmenal Economics, US Environmenal Proecion

More information

ACE 562 Fall Lecture 4: Simple Linear Regression Model: Specification and Estimation. by Professor Scott H. Irwin

ACE 562 Fall Lecture 4: Simple Linear Regression Model: Specification and Estimation. by Professor Scott H. Irwin ACE 56 Fall 005 Lecure 4: Simple Linear Regression Model: Specificaion and Esimaion by Professor Sco H. Irwin Required Reading: Griffihs, Hill and Judge. "Simple Regression: Economic and Saisical Model

More information

1.1. Example: Polynomial Curve Fitting 4 1. INTRODUCTION

1.1. Example: Polynomial Curve Fitting 4 1. INTRODUCTION 4. INTRODUCTION Figure.2 Plo of a raining daa se of N = poins, shown as blue circles, each comprising an observaion of he inpu variable along wih he corresponding arge variable. The green curve shows he

More information

Stability and Bifurcation in a Neural Network Model with Two Delays

Stability and Bifurcation in a Neural Network Model with Two Delays Inernaional Mahemaical Forum, Vol. 6, 11, no. 35, 175-1731 Sabiliy and Bifurcaion in a Neural Nework Model wih Two Delays GuangPing Hu and XiaoLing Li School of Mahemaics and Physics, Nanjing Universiy

More information

Solutions to Odd Number Exercises in Chapter 6

Solutions to Odd Number Exercises in Chapter 6 1 Soluions o Odd Number Exercises in 6.1 R y eˆ 1.7151 y 6.3 From eˆ ( T K) ˆ R 1 1 SST SST SST (1 R ) 55.36(1.7911) we have, ˆ 6.414 T K ( ) 6.5 y ye ye y e 1 1 Consider he erms e and xe b b x e y e b

More information

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds

Kriging Models Predicting Atrazine Concentrations in Surface Water Draining Agricultural Watersheds 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 Kriging Models Predicing Arazine Concenraions in Surface Waer Draining Agriculural Waersheds Paul L. Mosquin, Jeremy Aldworh, Wenlin Chen Supplemenal Maerial Number

More information

Notes for Lecture 17-18

Notes for Lecture 17-18 U.C. Berkeley CS278: Compuaional Complexiy Handou N7-8 Professor Luca Trevisan April 3-8, 2008 Noes for Lecure 7-8 In hese wo lecures we prove he firs half of he PCP Theorem, he Amplificaion Lemma, up

More information

Lecture Notes 2. The Hilbert Space Approach to Time Series

Lecture Notes 2. The Hilbert Space Approach to Time Series Time Series Seven N. Durlauf Universiy of Wisconsin. Basic ideas Lecure Noes. The Hilber Space Approach o Time Series The Hilber space framework provides a very powerful language for discussing he relaionship

More information

3.1 More on model selection

3.1 More on model selection 3. More on Model selecion 3. Comparing models AIC, BIC, Adjused R squared. 3. Over Fiing problem. 3.3 Sample spliing. 3. More on model selecion crieria Ofen afer model fiing you are lef wih a handful of

More information

Numerical Dispersion

Numerical Dispersion eview of Linear Numerical Sabiliy Numerical Dispersion n he previous lecure, we considered he linear numerical sabiliy of boh advecion and diffusion erms when approimaed wih several spaial and emporal

More information

Air Traffic Forecast Empirical Research Based on the MCMC Method

Air Traffic Forecast Empirical Research Based on the MCMC Method Compuer and Informaion Science; Vol. 5, No. 5; 0 ISSN 93-8989 E-ISSN 93-8997 Published by Canadian Cener of Science and Educaion Air Traffic Forecas Empirical Research Based on he MCMC Mehod Jian-bo Wang,

More information

Presentation Overview

Presentation Overview Acion Refinemen in Reinforcemen Learning by Probabiliy Smoohing By Thomas G. Dieerich & Didac Busques Speaer: Kai Xu Presenaion Overview Bacground The Probabiliy Smoohing Mehod Experimenal Sudy of Acion

More information

Article from. Predictive Analytics and Futurism. July 2016 Issue 13

Article from. Predictive Analytics and Futurism. July 2016 Issue 13 Aricle from Predicive Analyics and Fuurism July 6 Issue An Inroducion o Incremenal Learning By Qiang Wu and Dave Snell Machine learning provides useful ools for predicive analyics The ypical machine learning

More information

International Journal of Scientific & Engineering Research, Volume 4, Issue 10, October ISSN

International Journal of Scientific & Engineering Research, Volume 4, Issue 10, October ISSN Inernaional Journal of Scienific & Engineering Research, Volume 4, Issue 10, Ocober-2013 900 FUZZY MEAN RESIDUAL LIFE ORDERING OF FUZZY RANDOM VARIABLES J. EARNEST LAZARUS PIRIYAKUMAR 1, A. YAMUNA 2 1.

More information

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms

L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS. NA568 Mobile Robotics: Methods & Algorithms L07. KALMAN FILTERING FOR NON-LINEAR SYSTEMS NA568 Mobile Roboics: Mehods & Algorihms Today s Topic Quick review on (Linear) Kalman Filer Kalman Filering for Non-Linear Sysems Exended Kalman Filer (EKF)

More information

Approximation Algorithms for Unique Games via Orthogonal Separators

Approximation Algorithms for Unique Games via Orthogonal Separators Approximaion Algorihms for Unique Games via Orhogonal Separaors Lecure noes by Konsanin Makarychev. Lecure noes are based on he papers [CMM06a, CMM06b, LM4]. Unique Games In hese lecure noes, we define

More information

Lecture 2 April 04, 2018

Lecture 2 April 04, 2018 Sas 300C: Theory of Saisics Spring 208 Lecure 2 April 04, 208 Prof. Emmanuel Candes Scribe: Paulo Orensein; edied by Sephen Baes, XY Han Ouline Agenda: Global esing. Needle in a Haysack Problem 2. Threshold

More information

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes

23.2. Representing Periodic Functions by Fourier Series. Introduction. Prerequisites. Learning Outcomes Represening Periodic Funcions by Fourier Series 3. Inroducion In his Secion we show how a periodic funcion can be expressed as a series of sines and cosines. We begin by obaining some sandard inegrals

More information

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17

Designing Information Devices and Systems I Spring 2019 Lecture Notes Note 17 EES 16A Designing Informaion Devices and Sysems I Spring 019 Lecure Noes Noe 17 17.1 apaciive ouchscreen In he las noe, we saw ha a capacior consiss of wo pieces on conducive maerial separaed by a nonconducive

More information

5. Stochastic processes (1)

5. Stochastic processes (1) Lec05.pp S-38.45 - Inroducion o Teleraffic Theory Spring 2005 Conens Basic conceps Poisson process 2 Sochasic processes () Consider some quaniy in a eleraffic (or any) sysem I ypically evolves in ime randomly

More information

HYPERSPECTRAL IMAGE CLASSIFICATIO BASED O A FAST BREGMA SPARSE MULTI OMIAL LOGISTIC REGRESSIO ALGORITHM. Portugal - (jun,

HYPERSPECTRAL IMAGE CLASSIFICATIO BASED O A FAST BREGMA SPARSE MULTI OMIAL LOGISTIC REGRESSIO ALGORITHM. Portugal - (jun, HYPERSPECRAL IMAGE CLASSIFICAIO BASED O A FAS BREGMA SPARSE MULI OMIAL LOGISIC REGRESSIO ALGORIHM J. Li a, *, J. Bioucas-Dias a, Anonio Plaza b a Insiuo de elecomunicações, Insiuo Superior écnico, 049-00,

More information

GMM - Generalized Method of Moments

GMM - Generalized Method of Moments GMM - Generalized Mehod of Momens Conens GMM esimaion, shor inroducion 2 GMM inuiion: Maching momens 2 3 General overview of GMM esimaion. 3 3. Weighing marix...........................................

More information

Unit Root Time Series. Univariate random walk

Unit Root Time Series. Univariate random walk Uni Roo ime Series Univariae random walk Consider he regression y y where ~ iid N 0, he leas squares esimae of is: ˆ yy y y yy Now wha if = If y y hen le y 0 =0 so ha y j j If ~ iid N 0, hen y ~ N 0, he

More information

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits

Exponential Weighted Moving Average (EWMA) Chart Under The Assumption of Moderateness And Its 3 Control Limits DOI: 0.545/mjis.07.5009 Exponenial Weighed Moving Average (EWMA) Char Under The Assumpion of Moderaeness And Is 3 Conrol Limis KALPESH S TAILOR Assisan Professor, Deparmen of Saisics, M. K. Bhavnagar Universiy,

More information

KINEMATICS IN ONE DIMENSION

KINEMATICS IN ONE DIMENSION KINEMATICS IN ONE DIMENSION PREVIEW Kinemaics is he sudy of how hings move how far (disance and displacemen), how fas (speed and velociy), and how fas ha how fas changes (acceleraion). We say ha an objec

More information

Some Ramsey results for the n-cube

Some Ramsey results for the n-cube Some Ramsey resuls for he n-cube Ron Graham Universiy of California, San Diego Jozsef Solymosi Universiy of Briish Columbia, Vancouver, Canada Absrac In his noe we esablish a Ramsey-ype resul for cerain

More information

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010

Simulation-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Simulaion-Solving Dynamic Models ABE 5646 Week 2, Spring 2010 Week Descripion Reading Maerial 2 Compuer Simulaion of Dynamic Models Finie Difference, coninuous saes, discree ime Simple Mehods Euler Trapezoid

More information

Experiments on logistic regression

Experiments on logistic regression Experimens on logisic regression Ning Bao March, 8 Absrac In his repor, several experimens have been conduced on a spam daa se wih Logisic Regression based on Gradien Descen approach. Firs, he overfiing

More information

Hidden Markov Models. Adapted from. Dr Catherine Sweeney-Reed s slides

Hidden Markov Models. Adapted from. Dr Catherine Sweeney-Reed s slides Hidden Markov Models Adaped from Dr Caherine Sweeney-Reed s slides Summary Inroducion Descripion Cenral in HMM modelling Exensions Demonsraion Specificaion of an HMM Descripion N - number of saes Q = {q

More information

References are appeared in the last slide. Last update: (1393/08/19)

References are appeared in the last slide. Last update: (1393/08/19) SYSEM IDEIFICAIO Ali Karimpour Associae Professor Ferdowsi Universi of Mashhad References are appeared in he las slide. Las updae: 0..204 393/08/9 Lecure 5 lecure 5 Parameer Esimaion Mehods opics o be

More information

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H.

ACE 562 Fall Lecture 8: The Simple Linear Regression Model: R 2, Reporting the Results and Prediction. by Professor Scott H. ACE 56 Fall 5 Lecure 8: The Simple Linear Regression Model: R, Reporing he Resuls and Predicion by Professor Sco H. Irwin Required Readings: Griffihs, Hill and Judge. "Explaining Variaion in he Dependen

More information

Lecture 10: The Poincaré Inequality in Euclidean space

Lecture 10: The Poincaré Inequality in Euclidean space Deparmens of Mahemaics Monana Sae Universiy Fall 215 Prof. Kevin Wildrick n inroducion o non-smooh analysis and geomery Lecure 1: The Poincaré Inequaliy in Euclidean space 1. Wha is he Poincaré inequaliy?

More information

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time.

Nature Neuroscience: doi: /nn Supplementary Figure 1. Spike-count autocorrelations in time. Supplemenary Figure 1 Spike-coun auocorrelaions in ime. Normalized auocorrelaion marices are shown for each area in a daase. The marix shows he mean correlaion of he spike coun in each ime bin wih he spike

More information

20. Applications of the Genetic-Drift Model

20. Applications of the Genetic-Drift Model 0. Applicaions of he Geneic-Drif Model 1) Deermining he probabiliy of forming any paricular combinaion of genoypes in he nex generaion: Example: If he parenal allele frequencies are p 0 = 0.35 and q 0

More information

( ) ( ) if t = t. It must satisfy the identity. So, bulkiness of the unit impulse (hyper)function is equal to 1. The defining characteristic is

( ) ( ) if t = t. It must satisfy the identity. So, bulkiness of the unit impulse (hyper)function is equal to 1. The defining characteristic is UNIT IMPULSE RESPONSE, UNIT STEP RESPONSE, STABILITY. Uni impulse funcion (Dirac dela funcion, dela funcion) rigorously defined is no sricly a funcion, bu disribuion (or measure), precise reamen requires

More information

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon

3.1.3 INTRODUCTION TO DYNAMIC OPTIMIZATION: DISCRETE TIME PROBLEMS. A. The Hamiltonian and First-Order Conditions in a Finite Time Horizon 3..3 INRODUCION O DYNAMIC OPIMIZAION: DISCREE IME PROBLEMS A. he Hamilonian and Firs-Order Condiions in a Finie ime Horizon Define a new funcion, he Hamilonian funcion, H. H he change in he oal value of

More information

Lecture 3: Exponential Smoothing

Lecture 3: Exponential Smoothing NATCOR: Forecasing & Predicive Analyics Lecure 3: Exponenial Smoohing John Boylan Lancaser Cenre for Forecasing Deparmen of Managemen Science Mehods and Models Forecasing Mehod A (numerical) procedure

More information

2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes

2.7. Some common engineering functions. Introduction. Prerequisites. Learning Outcomes Some common engineering funcions 2.7 Inroducion This secion provides a caalogue of some common funcions ofen used in Science and Engineering. These include polynomials, raional funcions, he modulus funcion

More information

On Multicomponent System Reliability with Microshocks - Microdamages Type of Components Interaction

On Multicomponent System Reliability with Microshocks - Microdamages Type of Components Interaction On Mulicomponen Sysem Reliabiliy wih Microshocks - Microdamages Type of Componens Ineracion Jerzy K. Filus, and Lidia Z. Filus Absrac Consider a wo componen parallel sysem. The defined new sochasic dependences

More information

Linear Response Theory: The connection between QFT and experiments

Linear Response Theory: The connection between QFT and experiments Phys540.nb 39 3 Linear Response Theory: The connecion beween QFT and experimens 3.1. Basic conceps and ideas Q: How do we measure he conduciviy of a meal? A: we firs inroduce a weak elecric field E, and

More information

Isolated-word speech recognition using hidden Markov models

Isolated-word speech recognition using hidden Markov models Isolaed-word speech recogniion using hidden Markov models Håkon Sandsmark December 18, 21 1 Inroducion Speech recogniion is a challenging problem on which much work has been done he las decades. Some of

More information

Chapter 15: Phenomena. Chapter 15 Chemical Kinetics. Reaction Rates. Reaction Rates R P. Reaction Rates. Rate Laws

Chapter 15: Phenomena. Chapter 15 Chemical Kinetics. Reaction Rates. Reaction Rates R P. Reaction Rates. Rate Laws Chaper 5: Phenomena Phenomena: The reacion (aq) + B(aq) C(aq) was sudied a wo differen emperaures (98 K and 35 K). For each emperaure he reacion was sared by puing differen concenraions of he 3 species

More information

Block Diagram of a DCS in 411

Block Diagram of a DCS in 411 Informaion source Forma A/D From oher sources Pulse modu. Muliplex Bandpass modu. X M h: channel impulse response m i g i s i Digial inpu Digial oupu iming and synchronizaion Digial baseband/ bandpass

More information

Internet Traffic Modeling for Efficient Network Research Management Prof. Zhili Sun, UniS Zhiyong Liu, CATR

Internet Traffic Modeling for Efficient Network Research Management Prof. Zhili Sun, UniS Zhiyong Liu, CATR Inerne Traffic Modeling for Efficien Nework Research Managemen Prof. Zhili Sun, UniS Zhiyong Liu, CATR UK-China Science Bridge Workshop 13-14 December 2011, London Ouline Inroducion Background Classical

More information

WATER LEVEL TRACKING WITH CONDENSATION ALGORITHM

WATER LEVEL TRACKING WITH CONDENSATION ALGORITHM WATER LEVEL TRACKING WITH CONDENSATION ALGORITHM Shinsuke KOBAYASHI, Shogo MURAMATSU, Hisakazu KIKUCHI, Masahiro IWAHASHI Dep. of Elecrical and Elecronic Eng., Niigaa Universiy, 8050 2-no-cho Igarashi,

More information

Modal identification of structures from roving input data by means of maximum likelihood estimation of the state space model

Modal identification of structures from roving input data by means of maximum likelihood estimation of the state space model Modal idenificaion of srucures from roving inpu daa by means of maximum likelihood esimaion of he sae space model J. Cara, J. Juan, E. Alarcón Absrac The usual way o perform a forced vibraion es is o fix

More information

not to be republished NCERT MATHEMATICAL MODELLING Appendix 2 A.2.1 Introduction A.2.2 Why Mathematical Modelling?

not to be republished NCERT MATHEMATICAL MODELLING Appendix 2 A.2.1 Introduction A.2.2 Why Mathematical Modelling? 256 MATHEMATICS A.2.1 Inroducion In class XI, we have learn abou mahemaical modelling as an aemp o sudy some par (or form) of some real-life problems in mahemaical erms, i.e., he conversion of a physical

More information

Chapter 3 Boundary Value Problem

Chapter 3 Boundary Value Problem Chaper 3 Boundary Value Problem A boundary value problem (BVP) is a problem, ypically an ODE or a PDE, which has values assigned on he physical boundary of he domain in which he problem is specified. Le

More information

Some Basic Information about M-S-D Systems

Some Basic Information about M-S-D Systems Some Basic Informaion abou M-S-D Sysems 1 Inroducion We wan o give some summary of he facs concerning unforced (homogeneous) and forced (non-homogeneous) models for linear oscillaors governed by second-order,

More information

) were both constant and we brought them from under the integral.

) were both constant and we brought them from under the integral. YIELD-PER-RECRUIT (coninued The yield-per-recrui model applies o a cohor, bu we saw in he Age Disribuions lecure ha he properies of a cohor do no apply in general o a collecion of cohors, which is wha

More information

10. State Space Methods

10. State Space Methods . Sae Space Mehods. Inroducion Sae space modelling was briefly inroduced in chaper. Here more coverage is provided of sae space mehods before some of heir uses in conrol sysem design are covered in he

More information

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing

Application of a Stochastic-Fuzzy Approach to Modeling Optimal Discrete Time Dynamical Systems by Using Large Scale Data Processing Applicaion of a Sochasic-Fuzzy Approach o Modeling Opimal Discree Time Dynamical Sysems by Using Large Scale Daa Processing AA WALASZE-BABISZEWSA Deparmen of Compuer Engineering Opole Universiy of Technology

More information

A Forward-Backward Splitting Method with Component-wise Lazy Evaluation for Online Structured Convex Optimization

A Forward-Backward Splitting Method with Component-wise Lazy Evaluation for Online Structured Convex Optimization A Forward-Backward Spliing Mehod wih Componen-wise Lazy Evaluaion for Online Srucured Convex Opimizaion Yukihiro Togari and Nobuo Yamashia March 28, 2016 Absrac: We consider large-scale opimizaion problems

More information