Classification with Reject Option in Text Categorisation Systems
|
|
- Eustace Carson
- 5 years ago
- Views:
Transcription
1 Classfcaton wth Reject Opton n Text Categorsaton Systems Gorgo Fumera Ignazo Plla Fabo Rol Dept. of Electrcal and Electronc Eng., Pazza d Arm, 0923 Caglar, Italy {fumera,plla,rol}@dee.unca.t Abstract The am of ths paper s to evaluate the potental usefulness of the reject opton for text categorsaton (TC) tasks. The reject opton s a technque used n statstcal pattern recognton for mprovng classfcaton relablty. Our work s motvated by the fact that, although the reject opton proved to be useful n several pattern recognton problems, t has not yet been consdered for TC tasks. Snce TC tasks dffer from usual pattern recognton problems n the performance measures used and n the fact that documents can belong to more than one category, we developed a specfc rejecton technque for TC problems. The performance mprovement achevable by usng the reject opton was expermentally evaluated on the Reuters dataset, whch s a standard benchmark for TC systems.. Introducton Text categorsaton (TC) problems conssts n assgnng text documents, on the bass of ther content, to one or more predefned thematc categores. Ths s a typcal nformaton retreval (IR) task, and s currently an actve research feld, wth applcatons lke document ndexng and flterng, and herarchcal categorsaton of web pages. Approaches to TC have shfted snce the 90s from knowledge engneerng to the machne learnng approach, ganng n flexblty and savng expert labor power. Usng the machne learnng approach, a classfer s constructed on the bass of a tranng set of labeled documents, by usng nductve learnng methods. Several classfcaton technques have been appled to TC tasks, both symbolc, rule-based methods from the machne learnng feld, and non-symbolc methods (lke neural networks), from statstcal pattern recognton (see [8] for a comprehensve overvew). So far, works n the TC lterature focused on two man topcs: feature selcton/extracton, and comparson between dfferent classfcaton technques [6,9,0,,2]. In ths paper we focus on a technque used n statstcal pattern recognton for mprovng classfcaton relablty, namely, classfcaton wth reject opton. The reject opton conssts n wthholdng the classfcaton of an nput pattern, f t s lkely to be msclassfed. Rejected patterns are then handled n a dfferent way, for nstance, they are manually classfed. The reject opton s useful n pattern recognton tasks n whch the cost of an error s hgher than that of a rejecton. To the best of our knowledge, so far no work consdered the reject opton n TC tasks. evertheless, we beleve that the reject opton can be useful also n TC. Accordngly, the objectve of ths paper s to develop a method for ntroducng the reject opton n TC tasks, and to evaluate ts potental usefulness. To ths am, we consder the usual decomposton of -category problems nto ndependent bnary problems [8]: each bnary problem conssts n decdng f an nput document does or does not belong to one of the predefned categores. We ntroduce the reject opton by allowng the classfer to wthhold assgnng or not an nput document to the categores for whch the decson s consdered not suffcently relable. To ths end, we defne a reject rule based on applyng two threshold values to the scores provded by the classfer for each category. We then evaluate the performance mprovement achevable by usng the reject opton on the Reuters dataset, whch s a standard benchmark for TC systems. The paper s structured as follows. Sectons 2 and 3 provde the theoretcal background about TC and about the reject opton n statstcal pattern recognton. In sectons 4 and 5 we descrbe our method for mplementng the reject opton n TC tasks. The expermental results are reported n secton 6. Conclusons are drawn n secton Text categorsaton Formally, TC can be defned as the task of assgnng a Boolean value T or F to each par (d j,c ) D C, ndcatng whether document d j s to be labeled as belongng to category c, where D s a set of documents, and C s a set of predefned categores. The goal s to fnd a labelng functon (classfer) φ:d C {T,F} whch approxmates at best an unknown target funton Φ, accordng to a gven performance measure [8].
2 The most used performance measures, precson (π) and recall (ρ), are derved from IR. Precson π for the - th category s defned as the probablty that, f a random document d s classfed as belongng to c, ths decson s correct: π =P(Φ(d,c )=T φ(d,c )=T). Recall ρ for the -th class s defned as the probablty that, f d s to be classfed as belongng to c, ths decson s taken: ρ =P(φ(d,c )=T Φ(d,c )=T). For a gven classfer φ, these probabltes are usually estmated from a test set of prelabeled documents, as a functon of the number of true postves (TP ), false postves (FP ), true negatves (T ) and false negatves (F ) documents, defned as shown n Table. The estmates are computed as: TP π =, ρ = TP + FP Table. Contngency table for class Φ(d,c ) T F T TP FP φ(d,c ) F F T TP TP + F. () Global precson and recall are computed ether as mcro- or macro-averaged values (dependng on applcaton requrements). Mcro-averagng conssts n applyng defnton () to the global values of TP, FP, T and F, obtaned by summng over all classes: TP π µ = TP, ρ ( TP + FP ) µ =, (2) ( TP + F ) whle n macro-averagng, the values of π and ρ gven n Eq. () are averaged over the classes: π π M =, ρ ρ M =. (3) Precson and recall (both macro- and mcro-averaged) take on values n the range [0,], and are not ndependent measures: a hgher precson correspond to a lower recall, and vce-versa. The optmal classfer would be the one for whch π=, ρ=. In real tasks, for varyng values of the classfer parameters, precson and recall descrbe a curve n the π-ρ plane lke the one of Fg.. The workng pont on ths curve s applcaton-dependent. π 0 ρ Fg.. A typcal behavour of the π-ρ curve Several effectveness measures whch combne π and ρ have also been defned. A wdely used one s the F measure, defned as the harmonc mean of π and ρ: πρ F = 2. (4) ( π + ρ) In TC systems documents are usually represented wth a weght vector, each weght correspondng to a term (ether a word or a phrase) of the so-called vocabulary, whch s the set of all terms occurrng n the set D (note that n real tasks D s the avalable tranng set). Typcally, only terms obtaned after stop-word removal and stemmng are consdered. Weghts are manly computed as term frequences (tf), n whch they represent the frequency of occurrence of the correspondng terms n a document, and as nverted term frequences (tfdf), defned as: tfdf(t k,d j )=#(t k,d j ) log( D /# D (t k )), where #(t k,d j ) denotes the number of occurrences of term t k n document d j, and # D (t k ) denotes the number of documents n D n whch t k occurs. 3. Classfcaton wth reject opton n statstcal pattern recognton In statstcal pattern recognton, classfcaton wth reject opton has been formalsed under the mnmum rsk theory n [,2,4]. When the costs of correct classfcatons, rejectons, and msclassfcatons are respectvely c C, c R and c E (obvously, c C <c R <c E ), then the expected rsk (.e. the expected value of the cost of classfyng any pattern) s mnmsed by Chow s rule, whch s defned as follows. Consder the class for whch the a posteror probablty of x s maxmum: ω =argmax j P(ω j x); f P(ω x) T, assgn x to ω, otherwse reject x. T s the so-called reject threshold, whose optmal value depends on classfcaton costs: T=(c E -c R )/(c E -c C ) [2]. ote that when the cost of rejectons equals that of msclassfcatons, one obtans T=0: therefore, no pattern s rejected, and Chow s rule reduces to the well-known Bayes rule. We pont out that the above approach to the reject opton can not be mmedately appled to TC tasks, for two man reasons. Frst, Chow s rule apples to classfcaton problems n whch each pattern belongs to only one class, whle ths s not the case of TC problems, as explaned n prevous sectons. Moreover, Chow s rule s amed at mnmsng the expected rsk of a classfer, whle the performance measures n TC tasks are dfferent than the expected rsk. In the next secton we dscuss the way n whch the reject opton can be ntroduced n TC tasks.
3 4. A method for ntroducng the reject opton n text categorsaton As ponted out n [8], -category TC problems are usually consdered as ndependent bnary problems: for each category c, the classfer has to decde whether an nput document d does or does not belong to c, ndependently on the other categores. In ths framework, t s reasonable to ntroduce the reject opton by allowng a classfer to wthhold makng a decson, for categores where the decson s consdered unrelable. A document can then be rejected from any subset of the categores, whle t can be accepted and automatcally categorsed as belongng or not to the other categores. Therefore, each document has to be exceptonally handled (typcally, by manual classfcaton) only for the categores from whch t has been rejected, f any. A text classfer wth reject opton can then be vewed as mplementng functons φ :D C {T,F,R}, one for each category, where R stands for the reject decson. The goal of the reject opton should be that of ncreasng the effectveness measure (Eq. ()-(4)) by turnng as many ncorrect decsons (.e. FP and F ) as possble nto rejectons. Obvously, ths can be acheved at the expense of exceptonally handlng the rejected documents: a trade-off between the achevable performance mprovement and the cost of rejectons must then be found. However, the cost of handlng the rejected documents s clearly applcaton-dependent. In ths work, we evaluate the performance mprovement that can be acheved by the reject opton, wth respect to the rate of rejected decsons. By denotng as r the number of documents rejected from the -th category, the rate r of rejected decsons s defned as: r r D, (5) = ( / ) = where D s the total number of documents. From now on, we wll call r the reject rate. Accordngly, we wll consder the followng goal for the reject opton: maxmse the performance measure for any gven value of the reject rate (5). Let us now dscuss the way n whch the decson of rejectng an nput document from a category can be mplemented. Most classfcaton technques used so far n TC (lke the naïve Bayes and neural networks) are based on computng for any document d a score s [0,] for each category c, representng the evdence of the fact d c. Scores can be obtaned, for nstance, from the output of neural network classfers, or as probabltes computed usng a naïve Bayes classfer. The decson of assgnng or not assgnng d to c s usually made by ntroducng a threshold τ for each category, so that d s labeled as belongng to category c, only f s (d) τ, =,..., [8,2]. In other words, f s (d) τ, then φ(d,c )=T, otherwse φ(d,c )=F. ote that the threshold values are usually computed from a set of valdaton documents. To ntroduce the reject opton, t s reasonable to assume that the relablty of the above decson s hgher for values of s far from τ, than for values near to τ. In other words, the lower s -τ, the more lkely s an erroneous decson for category. Accordngly, we mplement the reject opton by ntroducng two threshold values for each category, denoted as τ H and τ L, wth τ H τ L, such that f s s between these values, the document s rejected from category. Formally, the classfer φ s the defned as follows: f s (d) τ H, φ(d,c )=T, f τ L <s (d)<τ H, φ(d,c )=R, (6) f s (d) τ L, φ(d,c )=F. ote that, f τ H =τ L, one obtans the standard classfer wthout reject opton. We pont out that our approach for ntroducng the reject opton by usng two threshold values for each category, s analogous to that proposed n [3] for two-class pattern recognton problems. In order to expermentally evaluate the performance mprovement achevable by the reject opton, wth respect to the number of rejected decsons, we developed algorthms for maxmsng the performance measure F (4), for any gven value of the reject rate (). These algorthms are presented n secton Algorthms for determnnng threshold values When the reject opton s mplemented n a TC problem as descrbed n secton 4, the optmal values of the 2 threshold {τ H,τ L,...,τ H,τ L } are the ones that maxmse the performance measure for any gven value of the reject rate (). In practce, t s convenent to mplement an algorthm that maxmses the performance measure whle keepng the reject rate r below the gven value. We wll denote such value as r MAX. In ths work, we focused on the F performance measure, both macro- and mcro-averaged. The macro-averaged F s defned as ( / ) ( / ) 2 / = ( = ), (7) F M M M M M M = F π ρ π ρ = + whle the mcro-averaged measure s defned as: m m m m m F = 2 π ρ /( π + ρ ) = 2/ ( 2 + ( FP+ F) / TP), (8) where FP = FP, F = F, TP = TP = = =. To maxmse F M, we chose to ndependently maxmse each F M, enforcng the constrant r r MAX by requrng a maxmum reject rate r for each category equal to r/: r r/. We also exploted the fact that F M s a nonncreasng functon of any τ L (ths property follows from Eqs. (7),(6),(3),()). Snce for each class only two thresholds τ H and τ L must be computed, a smple
4 exhaustve search of ther optmal values can be performed, usng a predefned dscretsaton step τ. To maxmse F m, we exploted the followng propertes (we omt the proof for the sake of brevty): m F s a non-ncreasng functon of any τ L (from (8),(6),(2)); consder the values of the thresholds of category, (τ * L,τ * H ), whch maxmse F m, for any fxed value of the thresholds of the other classes; now, f the thresholds of the other classes are changed so as to ncrease F m, then values of τ L and τ H such that * m τ H <τ H can not further ncrease F (from (8),(6),(2)). On the bass of the above propertes, we developed an algorthm for maxmsng F m under the constrant r r MAX, based on an teratve search n the space of the 2 threshold values. The thresholds are ntally set to zero. The above propertes allow to reduce the search space by consderng at each step only hgher values of the thresholds wth respect to the current ones. To further reduce the search space, we consder at each step only ponts n the search space obtaned by ncrementng the thresholds of only one category at a tme, whle keepng the ones of the other categores at ther current values. Accordngly, at each step, the thresholds of only one category are changed, provded that F m can be ncreased wth r r MAX. If no threshold values satsfyng ths condton are found, the algorthm termnates. The algorthm can be schematsed as follows. ntalse τ H =0, τ H =0, =,..., repeat for each category c : evaluate F m and r for τ H + k τ, for each k n 0,, maxk, and for the maxmum τ L such that r r MAX /, and F m does not decrease wth respect to τ H + k τ and to the ntal value of τ L ; select the par of threshold values (τ L *,τ H * ) that gve the maxmum mprovement wth respect to the current F m, f any, and set them as new values for the correspondng category; untl a hgher F m wth r r MAX s found wth respect to the prevous step. The parameter τ s a predefned dscretsaton step, whle maxk determnes the number of dfferent values of τ H whch are explored at each step. 6. Expermental results The goal of the experments presented n ths secton was to evaluate the potental usefulness of the reject opton n real TC tasks. More precsely, our am was to assess the mprovement of the performance of a TC system achevable by the reject opton, mplemented as descrbed n secton 4, wth respect to the rate of rejected decsons. We conducted our experments on the Reuters dataset, whch s a standard benchmark for TC systems [8,4,,3]. Ths dataset conssts of a set of 2,578 newswre stores (provded n standard text fles) classfed under categores related to economcs. In partcular, we used the ApteMod verson of ths dataset, whch was obtaned by elmnatng unlabelled documents and selectng the categores whch have at least one document n the tranng set and one n the test set. Ths verson conssts of a tranng set of 7,769 documents and a test set of 3,09 documents, belongng to 90 categores. The vocabulary, extracted from the tranng set, conssts of 6,635 terms, after stemmng and stop-word removal. Our experments have been performed wth three knds of classfers, commonly used n the TC lterature [4,9,0,]: k-nearest neghbour (k-), mult-layer perceptron neural network (MLP) and support vector machne (SVM) classfers. Pre-processng of the dataset (stop-word removal, stemmng and feature extracton from the tranng set) was performed wth the software SMART [7]. In partcular, both tf and tfdf weghts have been consdered. Due to the hgh number of features, feature selecton was performed on the tranng set by usng the nformaton gan crteron [0,]. Feature selecton was performed separately for each classfer, and for macro- and mcro-averaged performance measures. Then, the 70% of the documents n the orgnal tranng set was randomly extracted (by keepng the same proporton between the number of documents n each class) to be used for determnng classfer parameters and for tranng the classfers, whle the remanng 30% of documents were used as valdaton set for estmatng threshold values. In partcular, for categores wth less than fve documents, we always kept at least one document both n the tranng and n the valdaton set; for categores wth only one document n the orgnal tranng set, the document was duplcated n the valdaton set. For statstcal sgnfcance, the experments were repeated for ten dfferent randomly generated tranng sets. Reported results refer to the test set performance, averaged over these ten runs. For MLPs, we used one hdden layer wth 64 unts, and 90 output unts, as n []. The tf features were used. The number of nput unts was equal to that of features, whch was set to,000. MLPs were trand wth the standard backpropagaton algorthm, wth a learnng rate equal to 0.0. For k-s, the dstance-weghted verson of Yang [0] was used, wth 2,500 tfdf features. The value of k was 45 for macro- and 65 for mcro-averaged measures. On the bass of experments presented n [4,], we used SVMs wth lnear kernel, traned wth the SVM lght software [5]. 0,000 tfdf features were used. One SVM
5 for each category was traned, and the outputs were scaled to the range [0,] usng a sgmodal functon. Thresholds estmaton was performed on the valdaton set, wth the algorthms descrbed n secton 5. We used a threshold ncrement unt τ equal to 0.00, and a value of maxk for F m equal to 00. Values of the maxmum allowed reject rate r MAX between 0 and 5% were consdered. In Fgs. 2-7 we report the test set macro- and mcroaveraged values of F versus the reject rate, averaged over the ten runs, for the three classfers used. F Macro 0,48 0,47 0,46 0,45 0,44 0,43 0,42 0,4 0,40 0,39 0,38 0,0%,0% 2,0% 3,0% 4,0% Fg. 2. Test set F M vs reject rate (MLP) F mcro 0,94 0,92 0,90 0,88 0,86 0,84 0,82 0,0%,0% 2,0% 3,0% 4,0% 5,0% Fg. 3. Test set F m vs reject rate (MLP) F Macro 0,66 0,64 0,62 0,60 0,58 0,56 0,54 0,0% 0,5%,0%,5% 2,0% 2,5% Fg. 4. Test set F M vs reject rate (k-) F mcro 0,96 0,94 0,92 0,90 0,88 0,86 0,84 0,82 0,0% 0,5%,0%,5% 2,0% 2,5% 3,0% Fg. 5. Test set F m vs reject rate (k-) F Macro 0,58 0,57 0,56 0,55 0,54 0,53 0,52 0,5 0,50 0,0%,0% 2,0% 3,0% Fg. 6. Test set F M vs reject rate (SVM) F mcro 0,96 0,95 0,94 0,93 0,92 0,9 0,90 0,89 0,88 0,87 0,0% 0,5%,0%,5% 2,0% 2,5% 3,0% Fg. 7. Test set F m vs reject rate (SVM) ote that macro-averagng gves much worse performances than mcro-averagng. Ths s due to the fact that Reuters categores have a very dfferent generalty, and macro-averagng s domnated by the performance of the system on rare categores, as noted n [2] and [8]. Let us analyse the performance mprovement acheved by usng the reject opton. We frst pont out that the reject rate acheved on the test set s always much lower than the maxmum requred reject rate, whch was 5%. In partcular, the maxmum value of the reject rate (.e., the rate of rejected decsons) on the test set s about 4% for MLPs (Fgs. 2,3), 3% for k- (Fgs. 4,5), and 3.5% for SVMs (Fgs. 6,7). ote now that F always ncreases for ncreasng values of the reject rate (except for the fnal part of the curve n Fg. 2). Moreover, a sgnfcant ncrement of F was always obtaned by usng the reject opton, especally for macro-averaged F. In partcular, the maxmum ncrement of macro-averaged F was about 20% wth MLPs, 6% wth k-, and 4% wth SVM classfers; the ncrement of mcro-averaged F was 3% wth MLPs, 2% wth k-, and 8% wth SVM classfers. 7. Conclusons In ths work we proposed a method for ntroducng the reject opton n text categorsaton problems, and expermentally evaluated the performance mprovement achevable by usng the reject opton on a real TC task. The expermental results showed that the reject opton can allow to sgnfcantly mprove the performance of a text
6 categorsaton system, at the expense of a reasonably small rate of rejected decsons for each category. As ponted out n secton 4, the cost of handlng rejected documents strongly depends on the partcular applcaton. The approach proposed n ths work was based on the usual decomposton of -category TC problems nto bnary ndependent problems. On the bass of our expermental results, ths approach proved to be useful for applcatons n whch the cost of rejectons depends manly on the rate of rejected decsons. Our future work wll be devoted to analyse the case n whch the cost of rejectons s sgnfcantly affected also by the number of rejected documents. In ths case our approach should be modfed, n order to guarantee a small number of rejected documents. [3] F. Tortorella, An optmal reject rule for bnary classfers, Proc. Int. Workshop on Statstcal Pattern Recognton, Alcante, Span, 2000, pp [4] G. Fumera, F. Rol and G. Gacnto, Reject opton wth multple thresholds, Pattern Recognton 33, 2000, pp References [] C.K. Chow, An Optmum character recognton system usng decson functons, IRE Trans. Electronc Computers 6, 957, pp [2] C.K. Chow, On Optmum Error and Reject Tradeoff, IEEE Trans. on Inf. Theory 6, 970, pp [3] V. Dasg et al., Informaton fuson for text classfcaton an expermental comparson, Pattern Recognton 34, 200, pp [4] T. Joachms, Text categorzaton wth support vector machnes: learnng wth many relevant feature, Proc. 0th European Conf. on Machne Learnng, Chemntz, Germany, 998, pp [5] T. Joachms, Makng large-scale SVM learnng practcal, n Advances n Kernel Methods Support Vector Learnng, MIT Press, 999, pp [6] D.D. Lews, An evaluaton of phrasal and clustered representatons on a text categorzaton task, Proc. 5th ACM Int. Conf. on Research and Development n Inf. Retreval, Kobenhavn, DK, 992, pp [7] G. Salton, Automatc Text Processng: The Transformaton, analyss, and Retreval of Informaton by Computer, Addson- Wesley, Pennsylvana, 989. [8] F. Sebastan, Machne Learnng n Automated Text Categorzaton, ACM Computng Surveys 34, 2002, pp. 47. [9] E. Wener, J.O. Pedersen, and A.S. Wegend, A neural network approache to topc spottng, Proc. 4th Annual Symp. on Doc. Analyss and Inf. Retreval, Las Vegas, USA, 995, pp [0] Y. Yang, and J.O. Pedersen, A comparatve study on feature selecton n text categorzaton, Proc. of 4th Int. Conf. on Machne Learnng, ashvlle, USA, 997, pp [] Y. Yang, and X. Lu, A re-examnaton of text categorzaton methods, Proc. of 22nd ACM Int. Conf. on Research and Development n Inf. Retreval, Berkeley, US, 999. [2] Y. Yang, A Study on Thresholdng Strateges for Text Categorzaton, Proc. of SIGIR-0, Lousana, US, 200.
Evaluation for sets of classes
Evaluaton for Tet Categorzaton Classfcaton accuracy: usual n ML, the proporton of correct decsons, Not approprate f the populaton rate of the class s low Precson, Recall and F 1 Better measures 21 Evaluaton
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationA Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach
A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland
More informationP R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /
Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationMultilayer Perceptron (MLP)
Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationNatural Language Processing and Information Retrieval
Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationPattern Classification
Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationRegularized Discriminant Analysis for Face Recognition
1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths
More informationSupporting Information
Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to
More informationSDMML HT MSc Problem Sheet 4
SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be
More informationStatistics II Final Exam 26/6/18
Statstcs II Fnal Exam 26/6/18 Academc Year 2017/18 Solutons Exam duraton: 2 h 30 mn 1. (3 ponts) A town hall s conductng a study to determne the amount of leftover food produced by the restaurants n the
More informationAppendix B: Resampling Algorithms
407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationMultigradient for Neural Networks for Equalizers 1
Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationVQ widely used in coding speech, image, and video
at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng
More informationMDL-Based Unsupervised Attribute Ranking
MDL-Based Unsupervsed Attrbute Rankng Zdravko Markov Computer Scence Department Central Connectcut State Unversty New Brtan, CT 06050, USA http://www.cs.ccsu.edu/~markov/ markovz@ccsu.edu MDL-Based Unsupervsed
More informationAssortment Optimization under MNL
Assortment Optmzaton under MNL Haotan Song Aprl 30, 2017 1 Introducton The assortment optmzaton problem ams to fnd the revenue-maxmzng assortment of products to offer when the prces of products are fxed.
More informationPop-Click Noise Detection Using Inter-Frame Correlation for Improved Portable Auditory Sensing
Advanced Scence and Technology Letters, pp.164-168 http://dx.do.org/10.14257/astl.2013 Pop-Clc Nose Detecton Usng Inter-Frame Correlaton for Improved Portable Audtory Sensng Dong Yun Lee, Kwang Myung Jeon,
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationEvaluation of classifiers MLPs
Lecture Evaluaton of classfers MLPs Mlos Hausrecht mlos@cs.ptt.edu 539 Sennott Square Evaluaton For any data set e use to test the model e can buld a confuson matrx: Counts of examples th: class label
More information1 Convex Optimization
Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,
More informationCS47300: Web Information Search and Management
CS47300: Web Informaton Search and Management Probablstc Retreval Models Prof. Chrs Clfton 7 September 2018 Materal adapted from course created by Dr. Luo S, now leadng Albaba research group 14 Why probabltes
More informationInductance Calculation for Conductors of Arbitrary Shape
CRYO/02/028 Aprl 5, 2002 Inductance Calculaton for Conductors of Arbtrary Shape L. Bottura Dstrbuton: Internal Summary In ths note we descrbe a method for the numercal calculaton of nductances among conductors
More informationThe Expectation-Maximization Algorithm
The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationSampling Theory MODULE VII LECTURE - 23 VARYING PROBABILITY SAMPLING
Samplng heory MODULE VII LECURE - 3 VARYIG PROBABILIY SAMPLIG DR. SHALABH DEPARME OF MAHEMAICS AD SAISICS IDIA ISIUE OF ECHOLOGY KAPUR he smple random samplng scheme provdes a random sample where every
More informationCOMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
More informationCONTRAST ENHANCEMENT FOR MIMIMUM MEAN BRIGHTNESS ERROR FROM HISTOGRAM PARTITIONING INTRODUCTION
CONTRAST ENHANCEMENT FOR MIMIMUM MEAN BRIGHTNESS ERROR FROM HISTOGRAM PARTITIONING N. Phanthuna 1,2, F. Cheevasuvt 2 and S. Chtwong 2 1 Department of Electrcal Engneerng, Faculty of Engneerng Rajamangala
More informationThe Minimum Universal Cost Flow in an Infeasible Flow Network
Journal of Scences, Islamc Republc of Iran 17(2): 175-180 (2006) Unversty of Tehran, ISSN 1016-1104 http://jscencesutacr The Mnmum Unversal Cost Flow n an Infeasble Flow Network H Saleh Fathabad * M Bagheran
More informationKernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan
Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems
More informationEEE 241: Linear Systems
EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they
More informationTemperature. Chapter Heat Engine
Chapter 3 Temperature In prevous chapters of these notes we ntroduced the Prncple of Maxmum ntropy as a technque for estmatng probablty dstrbutons consstent wth constrants. In Chapter 9 we dscussed the
More informationFUZZY GOAL PROGRAMMING VS ORDINARY FUZZY PROGRAMMING APPROACH FOR MULTI OBJECTIVE PROGRAMMING PROBLEM
Internatonal Conference on Ceramcs, Bkaner, Inda Internatonal Journal of Modern Physcs: Conference Seres Vol. 22 (2013) 757 761 World Scentfc Publshng Company DOI: 10.1142/S2010194513010982 FUZZY GOAL
More informationChapter Newton s Method
Chapter 9. Newton s Method After readng ths chapter, you should be able to:. Understand how Newton s method s dfferent from the Golden Secton Search method. Understand how Newton s method works 3. Solve
More informationSTAT 511 FINAL EXAM NAME Spring 2001
STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte
More informationLinear Regression Analysis: Terminology and Notation
ECON 35* -- Secton : Basc Concepts of Regresson Analyss (Page ) Lnear Regresson Analyss: Termnology and Notaton Consder the generc verson of the smple (two-varable) lnear regresson model. It s represented
More informationImproving the performance of radial basis function classifiers in condition monitoring and fault diagnosis applications where unknown faults may occur
Improvng the performance of radal bass functon classfers n condton montorng and fault dagnoss applcatons where unknown faults may occur Yuhua L, Mchael J. Pont and N. Barre Jones Control & Instrumentaton
More informationCSC 411 / CSC D11 / CSC C11
18 Boostng s a general strategy for learnng classfers by combnng smpler ones. The dea of boostng s to take a weak classfer that s, any classfer that wll do at least slghtly better than chance and use t
More informationCSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography
CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve
More informationOn the Multicriteria Integer Network Flow Problem
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 5, No 2 Sofa 2005 On the Multcrtera Integer Network Flow Problem Vassl Vasslev, Marana Nkolova, Maryana Vassleva Insttute of
More informationProblem Set 9 Solutions
Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem
More informationTransfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system
Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng
More informationLecture 4. Instructor: Haipeng Luo
Lecture 4 Instructor: Hapeng Luo In the followng lectures, we focus on the expert problem and study more adaptve algorthms. Although Hedge s proven to be worst-case optmal, one may wonder how well t would
More informationNegative Binomial Regression
STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...
More informationMAXIMUM A POSTERIORI TRANSDUCTION
MAXIMUM A POSTERIORI TRANSDUCTION LI-WEI WANG, JU-FU FENG School of Mathematcal Scences, Peng Unversty, Bejng, 0087, Chna Center for Informaton Scences, Peng Unversty, Bejng, 0087, Chna E-MIAL: {wanglw,
More informationThe L(2, 1)-Labeling on -Product of Graphs
Annals of Pure and Appled Mathematcs Vol 0, No, 05, 9-39 ISSN: 79-087X (P, 79-0888(onlne Publshed on 7 Aprl 05 wwwresearchmathscorg Annals of The L(, -Labelng on -Product of Graphs P Pradhan and Kamesh
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationThe big picture. Outline
The bg pcture Vncent Claveau IRISA - CNRS, sldes from E. Kjak INSA Rennes Notatons classes: C = {ω = 1,.., C} tranng set S of sze m, composed of m ponts (x, ω ) per class ω representaton space: R d (=
More informationMultilayer neural networks
Lecture Multlayer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Mdterm exam Mdterm Monday, March 2, 205 In-class (75 mnutes) closed book materal covered by February 25, 205 Multlayer
More informationMethods of Combining Multiple Classifiers with Different Features and Their Applications to Text-Independent Speaker Identification
Internatonal Journal of Pattern Recognton and Artfcal Intellgence, Vol. 11, No. 3, pp. 417-445, 1997. World Scentfc Publsher Co. Methods of Combnng Multple Classfers wth Dfferent Features and Ther Applcatons
More informationMulti-layer neural networks
Lecture 0 Mult-layer neural networks Mlos Hauskrecht mlos@cs.ptt.edu 5329 Sennott Square Lnear regresson w Lnear unts f () Logstc regresson T T = w = p( y =, w) = g( w ) w z f () = p ( y = ) w d w d Gradent
More informationBoostrapaggregating (Bagging)
Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.
More informationDepartment of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING
MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/
More informationCourse 395: Machine Learning - Lectures
Course 395: Machne Learnng - Lectures Lecture 1-2: Concept Learnng (M. Pantc Lecture 3-4: Decson Trees & CC Intro (M. Pantc Lecture 5-6: Artfcal Neural Networks (S.Zaferou Lecture 7-8: Instance ased Learnng
More informationOnline Classification: Perceptron and Winnow
E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng
More informationTransient Stability Assessment of Power System Based on Support Vector Machine
ransent Stablty Assessment of Power System Based on Support Vector Machne Shengyong Ye Yongkang Zheng Qngquan Qan School of Electrcal Engneerng, Southwest Jaotong Unversty, Chengdu 610031, P. R. Chna Abstract
More informationComparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method
Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method
More informationThe Study of Teaching-learning-based Optimization Algorithm
Advanced Scence and Technology Letters Vol. (AST 06), pp.05- http://dx.do.org/0.57/astl.06. The Study of Teachng-learnng-based Optmzaton Algorthm u Sun, Yan fu, Lele Kong, Haolang Q,, Helongang Insttute
More informationNonlinear Classifiers II
Nonlnear Classfers II Nonlnear Classfers: Introducton Classfers Supervsed Classfers Lnear Classfers Perceptron Least Squares Methods Lnear Support Vector Machne Nonlnear Classfers Part I: Mult Layer Neural
More informationChapter 13: Multiple Regression
Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to
More informationSemi-supervised Classification with Active Query Selection
Sem-supervsed Classfcaton wth Actve Query Selecton Jao Wang and Swe Luo School of Computer and Informaton Technology, Beng Jaotong Unversty, Beng 00044, Chna Wangjao088@63.com Abstract. Labeled samples
More informationManning & Schuetze, FSNLP (c)1999, 2001
page 589 16.2 Maxmum Entropy Modelng 589 Mannng & Schuetze, FSNLP (c)1999, 2001 a decson tree that detects spam. Fndng the rght features s paramount for ths task, so desgn your feature set carefully. Exercse
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationStatistical Foundations of Pattern Recognition
Statstcal Foundatons of Pattern Recognton Learnng Objectves Bayes Theorem Decson-mang Confdence factors Dscrmnants The connecton to neural nets Statstcal Foundatons of Pattern Recognton NDE measurement
More informationSupport Vector Machines
CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at
More informationx = , so that calculated
Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to
More informationLearning Theory: Lecture Notes
Learnng Theory: Lecture Notes Lecturer: Kamalka Chaudhur Scrbe: Qush Wang October 27, 2012 1 The Agnostc PAC Model Recall that one of the constrants of the PAC model s that the data dstrbuton has to be
More informationWe present the algorithm first, then derive it later. Assume access to a dataset {(x i, y i )} n i=1, where x i R d and y i { 1, 1}.
CS 189 Introducton to Machne Learnng Sprng 2018 Note 26 1 Boostng We have seen that n the case of random forests, combnng many mperfect models can produce a snglodel that works very well. Ths s the dea
More information2 STATISTICALLY OPTIMAL TRAINING DATA 2.1 A CRITERION OF OPTIMALITY We revew the crteron of statstcally optmal tranng data (Fukumzu et al., 1994). We
Advances n Neural Informaton Processng Systems 8 Actve Learnng n Multlayer Perceptrons Kenj Fukumzu Informaton and Communcaton R&D Center, Rcoh Co., Ltd. 3-2-3, Shn-yokohama, Yokohama, 222 Japan E-mal:
More informationLearning from Data 1 Naive Bayes
Learnng from Data 1 Nave Bayes Davd Barber dbarber@anc.ed.ac.uk course page : http://anc.ed.ac.uk/ dbarber/lfd1/lfd1.html c Davd Barber 2001, 2002 1 Learnng from Data 1 : c Davd Barber 2001,2002 2 1 Why
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationCredit Card Pricing and Impact of Adverse Selection
Credt Card Prcng and Impact of Adverse Selecton Bo Huang and Lyn C. Thomas Unversty of Southampton Contents Background Aucton model of credt card solctaton - Errors n probablty of beng Good - Errors n
More informationISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013
ISSN: 2277-375 Constructon of Trend Free Run Orders for Orthogonal rrays Usng Codes bstract: Sometmes when the expermental runs are carred out n a tme order sequence, the response can depend on the run
More informationNotes on Frequency Estimation in Data Streams
Notes on Frequency Estmaton n Data Streams In (one of) the data streamng model(s), the data s a sequence of arrvals a 1, a 2,..., a m of the form a j = (, v) where s the dentty of the tem and belongs to
More informationChapter 9: Statistical Inference and the Relationship between Two Variables
Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,
More information2.3 Nilpotent endomorphisms
s a block dagonal matrx, wth A Mat dm U (C) In fact, we can assume that B = B 1 B k, wth B an ordered bass of U, and that A = [f U ] B, where f U : U U s the restrcton of f to U 40 23 Nlpotent endomorphsms
More informationINF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018
INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton
More informationIntro to Visual Recognition
CS 2770: Computer Vson Intro to Vsual Recognton Prof. Adrana Kovashka Unversty of Pttsburgh February 13, 2018 Plan for today What s recognton? a.k.a. classfcaton, categorzaton Support vector machnes Separable
More informationCHAPTER 17 Amortized Analysis
CHAPTER 7 Amortzed Analyss In an amortzed analyss, the tme requred to perform a sequence of data structure operatons s averaged over all the operatons performed. It can be used to show that the average
More informationfind (x): given element x, return the canonical element of the set containing x;
COS 43 Sprng, 009 Dsjont Set Unon Problem: Mantan a collecton of dsjont sets. Two operatons: fnd the set contanng a gven element; unte two sets nto one (destructvely). Approach: Canoncal element method:
More informationNeryškioji dichotominių testo klausimų ir socialinių rodiklių diferencijavimo savybių klasifikacija
Neryškoj dchotomnų testo klausmų r socalnų rodklų dferencjavmo savybų klasfkacja Aleksandras KRYLOVAS, Natalja KOSAREVA, Julja KARALIŪNAITĖ Technologcal and Economc Development of Economy Receved 9 May
More informationStatistical pattern recognition
Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve
More informationOn the correction of the h-index for career length
1 On the correcton of the h-ndex for career length by L. Egghe Unverstet Hasselt (UHasselt), Campus Depenbeek, Agoralaan, B-3590 Depenbeek, Belgum 1 and Unverstet Antwerpen (UA), IBW, Stadscampus, Venusstraat
More informationRelevance Vector Machines Explained
October 19, 2010 Relevance Vector Machnes Explaned Trstan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introducton Ths document has been wrtten n an attempt to make Tppng s [1] Relevance Vector Machnes
More informationUncertainty in measurements of power and energy on power networks
Uncertanty n measurements of power and energy on power networks E. Manov, N. Kolev Department of Measurement and Instrumentaton, Techncal Unversty Sofa, bul. Klment Ohrdsk No8, bl., 000 Sofa, Bulgara Tel./fax:
More informationTornado and Luby Transform Codes. Ashish Khisti Presentation October 22, 2003
Tornado and Luby Transform Codes Ashsh Khst 6.454 Presentaton October 22, 2003 Background: Erasure Channel Elas[956] studed the Erasure Channel β x x β β x 2 m x 2 k? Capacty of Noseless Erasure Channel
More informationCHAPTER IV RESEARCH FINDING AND DISCUSSIONS
CHAPTER IV RESEARCH FINDING AND DISCUSSIONS A. Descrpton of Research Fndng. The Implementaton of Learnng Havng ganed the whole needed data, the researcher then dd analyss whch refers to the statstcal data
More informationANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)
Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of
More informationarxiv:cs.cv/ Jun 2000
Correlaton over Decomposed Sgnals: A Non-Lnear Approach to Fast and Effectve Sequences Comparson Lucano da Fontoura Costa arxv:cs.cv/0006040 28 Jun 2000 Cybernetc Vson Research Group IFSC Unversty of São
More informationLinear Feature Engineering 11
Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19
More information