Bayesian Network Learning for Rare Events
|
|
- Frederick Adams
- 5 years ago
- Views:
Transcription
1 Internatonal Conference on Comuter Systems and Technologes - ComSysTech 06 Bayesan etwor Learnng for Rare Events Samuel G. Gerssen, Leon J. M. Rothrantz Abstract: Parameter learnng from data n Bayesan networs s a straghtforward tas. The average number of observed occurrences s stored n a condtonal roblty tle, from whch future redctons can be calculated. Ths method reles heavly on the qualty of the data. A data set wth rare events wll not yeld statstcally relle estmates. Bayesan networs allow ror and osteror learnng. In ths aer, new ror assessment technques are ntroduced to obtan stle rors for a condtonal roblty tle. These learnng algorthms are mlemented and tested, and the results wll be resented. Key words: Bayesan networs, Learnng, ave Bayes rors. ITRODUCTIO A Bayesan networ [8, 0] s a drected acyclc grah wth each node reresentng a varle and each arc reresentng a causal relaton between two varles. Varles are characterzed by a roblty dstrbuton for each value. The roblty dstrbuton of each node s nfluenced by the states (for dscrete nodes) or values (for contnuous nodes) of the arent node. The condtonal robltes of a node are stored n a condtonal roblty tle (CPT). The CPT s needed to calculate any condtonal roblty n the model, nference [5]. The sze of the CPT deends on the number of states (s), the number of arents (), and the number of arent states (s) n the followng way: sze ( CPT ) s ( ) () For every ossble combnaton of arent states, there s an entry lsted n the CPT. otce that for a large number of arents the CPT wll eand drastcally. Assume the varles n the Bayesan networ llustrated n Fgure are bnary. s Fgure : Eamle Bayesan networ The condtonal roblty tle of node C wll have eght entres, wth four degrees of freedom. The CPT for node C s gven n Tle. Tle : CPT for node C ( C AB) c c a, b a, b a a, b a a, b b b The research reorted here s art of the Interactve Colloratve Informaton Systems (ICIS) roect, suorted by the Dutch Mnstry of Economc Affars, grant nr: BSIK II.5- -
2 Internatonal Conference on Comuter Systems and Technologes - ComSysTech 06 The number of degrees of freedom of a CPT wth number of states (s), number of arents (), and number of arent states ( s ) s s ( s ). Values for can be obtaned from the data by arameter learnng. s the number of occurrences ( c, ) dvded by the number of occurrences ( ), or ( c, ) + ( c, ). In general, f for a varle X, wth states,...,, s the -th combnaton of arent states, then gven. Let (, ) now be calculated as follows: o be the number of observatons ) o(, ) (2) o(, ) denotes the roblty of, n the data set. can ( Parameter learnng s bascally countng the number of observatons of a secfc event. otce that the accuracy of heavly deends on the number of observed events. A large number of observatons wll result n a more accurate estmate than ust a small samle. Therefore, large data sets usually rovde good, stle models. Unfortunately many realworld data sets are mbalanced; some states of the resonse varle may be dozens to thousands of tmes less lely than other states. Ths s the case n, for eamle, customer banrutces n bans, nternatonal armed conflcts, or edemologcal nfectons. Thus the effect of havng a large data set avalle s canceled by the fact that t contans only a small number of nterestng records. Therefore, a model based on such data may be severely based or hghly unstle. PREVIOUS WORK Rare event roblems occur n generalzed lnear models, such as logstc regresson, but also n models learned from data, such as neural networs and Bayesan networs. Esecally n the case of generalzed lnear models, technques such as ror correcton and weghtng have been develoed to dscard most of the unnterestng art of the data set wthout much erformance loss [4]. Addtonally there are ways of reducng bas and varance [9]. In Bayesan networs, small samle roblems are usually solved by settng a good ror dstrbuton [, 6], based on eert nowledge. However, f ths eert nowledge s not avalle, an accetle ror, based on the data, needs to be set. [7] rooses nosy-or to obtan rors. Ths aroach uses cuttng to shft from ror to osteror estmate. ave Bayes s a very good classfer, as descrbed n [2]. Parameter learnng s elaned n detal by Davd Hecerman n [3]. Inference n a Bayesan networ s the calculaton of condtonal robltes, gven the robltes n a CPT. Inference s based on two rules. The frst one s Bayes rule, whch s defned as: y ) ) ( (3) The second rule s the eanson rule, whch s defned for bnary varles ( ) ) ) as: - II.5-2 -
3 Internatonal Conference on Comuter Systems and Technologes - ComSysTech 06 ) + ) Y Z (4) Usng these two rules, n the eamle n Fgure, ( a c) can now be calculated: a c) a) A a) c a) c) B B c )b) c )) (5) Inference wll be used for the assessment of nave Bayes rors. CPT Calculaton Dervng CPT arameters usng a ror-osteror aroach conssts of three stages:. ror assessment 2. osteror assessment 3. mergng The osteror assessment s equvalent wth CPT arameter learnng, shown n equaton 2, and wll not be dscussed here. Frst, the mergng method s descrbed and then the ror assessment. Mergng The mergng rocess of rors and osterors can be handled n a coule of ways. [7] uses a cut value for the number of observatons, n that aer called smoothng. If, n the data set, a combnaton of arent states occurs less than the cut value, the ror wll be used n the CPT. Otherwse, the osteror wll be used. Let the ror for be, the osteror r, and the cut value c, the smoothng mergng s defned as: Substtuton of equatons 2 and 6 yelds:, f < o(, ) c (6) r, otherwse, f o(, ) < c o(, ) (7), otherwse o(, ) - II.5-3 -
4 Internatonal Conference on Comuter Systems and Technologes - ComSysTech 06 In ths aer, a gentle transton s roosed, by weghtng the rors. The rors wll receve an nteger value, weght w, whch s equvalent to a number of observatons. The resultng CPT entres wll be a weghted average of the rors and the osterors: (w ) + o(, ) (8) w + o(, ) If o(, ) s low, the CPT value wll manly be based on the ror, however, f (, ) s often observed n the data set, the data wll nfluence the CPT value more than the ror. Bascally ths effect s the same as smoothng, only the transton s gentler. In smoothng and weghtng, the values for c and w, resectvely, need to be set. The tests n Secton 4 show that these arameters have the greatest effect f a value between 5 and 80 s chosen, referly between 0 and 40. If the values are too low, the rocedure wll be smlar to normal CPT learnng. If the values are too hgh, the osterors wll have very lttle effect. otce that for a model wth etremely good rors, the weght or cut value should be very hgh. Pror Assessment A CPT contans very detaled nformaton out condtonal robltes for all ossble arent states. As shown n the ntroducton, the number of degrees of freedom eands radly f the number of arents and the number of arent states ncrease. A stle ror should therefore be derved from a low number of degrees of freedom, but as accurate as ossble. [7] roose the usage of nosy-max rors, because t s a good modelng technque for rare events. osy-max maes the assumton that the state of the resonse varle s a logcal combnaton of the states of the nut arameters. In ractce, for a lot of data sets, nosy-max shows hgh erformance, because n many cases, an effect s an addton or multlcaton of the causes. An eamle of a bnary nosy-max (nosy-or) model s gven n Fgure 2. Fgure 2: osy-max networ Asde from the nut varles X...X and resonse varle Y, there s a set of nhbtor nodes Z...Z. X can cause Z to be resent wth a roblty c, but sence of X always mles sence of Z. The CPT of node Y s smlar to a logcal MAX gate. ave Bayes s n many cases a sueror classfer [2]. Just as nosy-max, nave Bayes has a relatvely low number of degrees of freedom. The erformance as a classfer s slghtly better than nosy-max. The networ structure s comletely ooste from a regular CPT networ. It may be aganst ntuton to use rors from an ooste networ. In the ror assessment, the causaltes are less relevant than how y ) s calculated. A nave Bayes networ maes very unrealstc assumtons out causalty. It assumes the redctve varles to be deendent on the resonse varle. Also, t assumes condtonal ndeendence between the redctve varles, meanng that gven the value of Y, X...X are ndeendent. Asde from these assumtons, a nave Bayes model s a - II.5-4 -
5 Internatonal Conference on Comuter Systems and Technologes - ComSysTech 06 very owerful classfer. Inference n a nave Bayes networ s not as straghtforward as n a CPT networ. A networ wth redctve varles X...X and resonse varle Y wth states y...ym wll have the robltes y ) lsted n the CPT. In a nave Bayes networ, these values need to be calculated as follows: y ) ) ) ( y ) y )) (9) The values for y ) and can be obtaned by arameter learnng n the nave Bayes networ. from equaton 8 can be relaced by the rght art of equaton 9. ow the formula for calculaton of the CPT values s comlete. RESULTS The method ove was mlemented and tested on a banrutcy data set from a ban. The banrutces were rare (around 2% of all records). As a erformance measure for scorng the models, the GII nde was used. In a grah where a curve s lotted as the cumulatve banrut ercentage of clents aganst the total ercentage of clents when they are sorted on rs (low rs on the left, hgh rs on the rght), the GII nde s defned as the area between the dagonal and the curve (the Lorentz curve) dvded by the total area under the dagonal. It has a range between - and, or 00% to 00%. Hgh scores ndcate good models. The GII nde s wdely used n socal scences to measure the dscrmnatve ower of a model. Tle 2: Performance of models MODEL GII ave Weghtng W 76.4% ave Weghtng W5 77.0% ave Weghtng W0 ave Weghtng W20 77.% 77.2% ave Weghtng W % ave Weghtng W % osy-max Weghtng W0 7.4% ave Smoothng C0 ave Smoothng C % 76.6% ave Smoothng C40 osy-max Smoothng C % 7.4% More nformaton out the GII nde and ts alcatons can be found n []. A coule of modelng methods were comared, of whch the results are lsted n Tle 2. Aarently, for ths data set, nave Bayes rovdes much better rors than nosy-max. Ths s artly due to the fact that a subotmal learnng algorthm for nosy-max s used. - II.5-5 -
6 Internatonal Conference on Comuter Systems and Technologes - ComSysTech 06 Even wth EM learnng, nosy-max scores wll not be hgher than 76%. Secondly, the weghtng seems to erform better than the smoothng for ths data set. COCLUSIOS AD FUTURE WORK Two mrovements to learnng for rare event data were suggested n ths aer. Frstly, weghted mergng nstead of cuttng, whch allows a more gentle balance between a ror and a osteror. Secondly, nference n a nave Bayes models can rovde ecellent rors for a CPT n a normal Bayesan networ. Comared to estng methods, such as nosy-max rors, nave Bayes rors erform better on the test data set. Unfortunately, the best erformng model on ths data set was nave Bayes wthout any osteror learnng. Therefore, more data sets to test these methods on are requred for a defnte statement. These ntal results are very romsng. REFERECES [] Bec,., G. Kng & L. Zeng. The Problem wth Quanttatve Studes of Internatonal Conflct. htt://web.olmeth.ufl.edu/aers/98/bec98.z. [2] Fredman,., D. Geger & M. Goldszmdt. Bayesan etwor Classfers. Machne Learnng 29(2-3):3-63, 997. [3] Hecerman, D. A Tutoral on Learnng Bayesan etwors. Techncal Reort, Mcrosoft Research, 995. [4] Kng, G. & L. Zeng. Logstc Regresson n Rare Events Data. Poltcal Analyss 9(2):37-63, 200. [5] Laurtzen, S.L. & D.J. Segelhalter. Local comutatons wth robltes on grahcal structures and ther alcaton to eert systems (wth dscusson). Journal of Royal Statstcal Socety, Seres B 50(2):57-224, 988. [6] eal, R.M. Bayesan Learnng for eural etwors. Srnger-Verlag, 996. [7] Onso, A., M.J. Druzdzel & H. Wasylu. Learnng Bayesan networ arameters from small data sets: alcaton of osy-or gates. Internatonal Journal of Aromate Reasonng 27(2):65-82, 200. [8] Pearl, J. Problstc Reasonng n Intellgent Systems: etwors of Plausble Inference. Morgan Kaufmann, 988. [9] Rley, B.D. Pattern Recognton and eural etwors. Cambrdge Unversty Press, 996. [0] Srtes, P., C. Glymour & R. Schenes Causaton, Predcton and Search. Lecture otes n Statstcs 8, Srnger Verlag, 993. [] Xu, K. How has the lterature on Gn s Inde evolved n the ast 80 years? Worng aer, ABOUT THE AUTHOR Assoc. Prof. L. J. M. Rothrantz, Deartment of Man-Machne Interacton, Delft Unversty of Technology, Phone: , -mal: L.J.M.Rothrantz@ew.tudelft.nl - II.5-6 -
Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Exerments-I MODULE III LECTURE - 2 EXPERIMENTAL DESIGN MODELS Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 2 We consder the models
More informationConfidence intervals for weighted polynomial calibrations
Confdence ntervals for weghted olynomal calbratons Sergey Maltsev, Amersand Ltd., Moscow, Russa; ur Kalambet, Amersand Internatonal, Inc., Beachwood, OH e-mal: kalambet@amersand-ntl.com htt://www.chromandsec.com
More informationPattern Classification (II) 杜俊
attern lassfcaton II 杜俊 junu@ustc.eu.cn Revew roalty & Statstcs Bayes theorem Ranom varales: screte vs. contnuous roalty struton: DF an DF Statstcs: mean, varance, moment arameter estmaton: MLE Informaton
More informationClassification Bayesian Classifiers
lassfcaton Bayesan lassfers Jeff Howbert Introducton to Machne Learnng Wnter 2014 1 Bayesan classfcaton A robablstc framework for solvng classfcaton roblems. Used where class assgnment s not determnstc,.e.
More informationNaïve Bayes Classifier
9/8/07 MIST.6060 Busness Intellgence and Data Mnng Naïve Bayes Classfer Termnology Predctors: the attrbutes (varables) whose values are used for redcton and classfcaton. Predctors are also called nut varables,
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Exerments-I MODULE II LECTURE - GENERAL LINEAR HYPOTHESIS AND ANALYSIS OF VARIANCE Dr. Shalabh Deartment of Mathematcs and Statstcs Indan Insttute of Technology Kanur 3.
More informationOutline. Bayesian Networks: Maximum Likelihood Estimation and Tree Structure Learning. Our Model and Data. Outline
Outlne Bayesan Networks: Maxmum Lkelhood Estmaton and Tree Structure Learnng Huzhen Yu janey.yu@cs.helsnk.f Dept. Computer Scence, Unv. of Helsnk Probablstc Models, Sprng, 200 Notces: I corrected a number
More informationMachine Learning. Classification. Theory of Classification and Nonparametric Classifier. Representing data: Hypothesis (classifier) Eric Xing
Machne Learnng 0-70/5 70/5-78, 78, Fall 008 Theory of Classfcaton and Nonarametrc Classfer Erc ng Lecture, Setember 0, 008 Readng: Cha.,5 CB and handouts Classfcaton Reresentng data: M K Hyothess classfer
More informationSimulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests
Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth
More informationSupporting Information
Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to
More informationWeb-Mining Agents Probabilistic Information Retrieval
Web-Mnng Agents Probablstc Informaton etreval Prof. Dr. alf Möller Unverstät zu Lübeck Insttut für Informatonssysteme Karsten Martny Übungen Acknowledgements Sldes taken from: Introducton to Informaton
More informationAn (almost) unbiased estimator for the S-Gini index
An (almost unbased estmator for the S-Gn ndex Thomas Demuynck February 25, 2009 Abstract Ths note provdes an unbased estmator for the absolute S-Gn and an almost unbased estmator for the relatve S-Gn for
More informationWhat Independencies does a Bayes Net Model? Bayesian Networks: Independencies and Inference. Quick proof that independence is symmetric
Bayesan Networks: Indeendences and Inference Scott Daves and ndrew Moore Note to other teachers and users of these sldes. ndrew and Scott would be delghted f you found ths source materal useful n gvng
More informationAssignment 5. Simulation for Logistics. Monti, N.E. Yunita, T.
Assgnment 5 Smulaton for Logstcs Mont, N.E. Yunta, T. November 26, 2007 1. Smulaton Desgn The frst objectve of ths assgnment s to derve a 90% two-sded Confdence Interval (CI) for the average watng tme
More informationBayesian Decision Theory
No.4 Bayesan Decson Theory Hu Jang Deartment of Electrcal Engneerng and Comuter Scence Lassonde School of Engneerng York Unversty, Toronto, Canada Outlne attern Classfcaton roblems Bayesan Decson Theory
More informationManaging Capacity Through Reward Programs. on-line companion page. Byung-Do Kim Seoul National University College of Business Administration
Managng Caacty Through eward Programs on-lne comanon age Byung-Do Km Seoul Natonal Unversty College of Busness Admnstraton Mengze Sh Unversty of Toronto otman School of Management Toronto ON M5S E6 Canada
More informationStatistical Evaluation of WATFLOOD
tatstcal Evaluaton of WATFLD By: Angela MacLean, Dept. of Cvl & Envronmental Engneerng, Unversty of Waterloo, n. ctober, 005 The statstcs program assocated wth WATFLD uses spl.csv fle that s produced wth
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have
More informationComparing two Quantiles: the Burr Type X and Weibull Cases
IOSR Journal of Mathematcs (IOSR-JM) e-issn: 78-578, -ISSN: 39-765X. Volume, Issue 5 Ver. VII (Se. - Oct.06), PP 8-40 www.osrjournals.org Comarng two Quantles: the Burr Tye X and Webull Cases Mohammed
More informationDurban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications
Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department
More informationThis column is a continuation of our previous column
Comparson of Goodness of Ft Statstcs for Lnear Regresson, Part II The authors contnue ther dscusson of the correlaton coeffcent n developng a calbraton for quanttatve analyss. Jerome Workman Jr. and Howard
More informationNumber of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k
ANOVA Model and Matrx Computatons Notaton The followng notaton s used throughout ths chapter unless otherwse stated: N F CN Y Z j w W Number of cases Number of factors Number of covarates Number of levels
More informationChapter 6. Supplemental Text Material
Chapter 6. Supplemental Text Materal S6-. actor Effect Estmates are Least Squares Estmates We have gven heurstc or ntutve explanatons of how the estmates of the factor effects are obtaned n the textboo.
More informationFuzzy approach to solve multi-objective capacitated transportation problem
Internatonal Journal of Bonformatcs Research, ISSN: 0975 087, Volume, Issue, 00, -0-4 Fuzzy aroach to solve mult-objectve caactated transortaton roblem Lohgaonkar M. H. and Bajaj V. H.* * Deartment of
More informationDefinition. Measures of Dispersion. Measures of Dispersion. Definition. The Range. Measures of Dispersion 3/24/2014
Measures of Dsperson Defenton Range Interquartle Range Varance and Standard Devaton Defnton Measures of dsperson are descrptve statstcs that descrbe how smlar a set of scores are to each other The more
More informationAn application of generalized Tsalli s-havrda-charvat entropy in coding theory through a generalization of Kraft inequality
Internatonal Journal of Statstcs and Aled Mathematcs 206; (4): 0-05 ISS: 2456-452 Maths 206; (4): 0-05 206 Stats & Maths wwwmathsjournalcom Receved: 0-09-206 Acceted: 02-0-206 Maharsh Markendeshwar Unversty,
More informationA Quadratic Cumulative Production Model for the Material Balance of Abnormally-Pressured Gas Reservoirs F.E. Gonzalez M.S.
Natural as Engneerng A Quadratc Cumulatve Producton Model for the Materal Balance of Abnormally-Pressured as Reservors F.E. onale M.S. Thess (2003) T.A. Blasngame, Texas A&M U. Deartment of Petroleum Engneerng
More informationDigital PI Controller Equations
Ver. 4, 9 th March 7 Dgtal PI Controller Equatons Probably the most common tye of controller n ndustral ower electroncs s the PI (Proortonal - Integral) controller. In feld orented motor control, PI controllers
More informationA General Class of Selection Procedures and Modified Murthy Estimator
ISS 684-8403 Journal of Statstcs Volume 4, 007,. 3-9 A General Class of Selecton Procedures and Modfed Murthy Estmator Abdul Bast and Muhammad Qasar Shahbaz Abstract A new selecton rocedure for unequal
More informationA Quadratic Cumulative Production Model for the Material Balance of Abnormally-Pressured Gas Reservoirs F.E. Gonzalez M.S.
Formaton Evaluaton and the Analyss of Reservor Performance A Quadratc Cumulatve Producton Model for the Materal Balance of Abnormally-Pressured as Reservors F.E. onale M.S. Thess (2003) T.A. Blasngame,
More information/ n ) are compared. The logic is: if the two
STAT C141, Sprng 2005 Lecture 13 Two sample tests One sample tests: examples of goodness of ft tests, where we are testng whether our data supports predctons. Two sample tests: called as tests of ndependence
More informationChapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems
Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons
More informationCOMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
More informationPriority Queuing with Finite Buffer Size and Randomized Push-out Mechanism
ICN 00 Prorty Queung wth Fnte Buffer Sze and Randomzed Push-out Mechansm Vladmr Zaborovsy, Oleg Zayats, Vladmr Muluha Polytechncal Unversty, Sant-Petersburg, Russa Arl 4, 00 Content I. Introducton II.
More informationSystem in Weibull Distribution
Internatonal Matheatcal Foru 4 9 no. 9 94-95 Relablty Equvalence Factors of a Seres-Parallel Syste n Webull Dstrbuton M. A. El-Dacese Matheatcs Departent Faculty of Scence Tanta Unversty Tanta Egypt eldacese@yahoo.co
More informationComplete Variance Decomposition Methods. Cédric J. Sallaberry
Comlete Varance Decomoston Methods Cédrc J. allaberry enstvty Analyss y y [,,, ] [ y, y,, ] y ny s a vector o uncertan nuts s a vector o results s a comle uncton successon o derent codes, systems o de,
More informationComparison of Outlier Detection Methods in Crossover Design Bioequivalence Studies
Journal of Pharmacy and Nutrton Scences, 01,, 16-170 16 Comarson of Outler Detecton Methods n Crossover Desgn Boequvalence Studes A. Rasheed 1,*, T. Ahmad,# and J.S. Sddq,# 1 Deartment of Research, Dow
More informationRegularized Discriminant Analysis for Face Recognition
1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths
More informationNegative Binomial Regression
STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...
More informationMultinomial logit regression
07/0/6 Multnomal logt regresson Introducton We now turn our attenton to regresson models for the analyss of categorcal dependent varables wth more than two response categores: Y car owned (many possble
More informationA New Evolutionary Computation Based Approach for Learning Bayesian Network
Avalable onlne at www.scencedrect.com Proceda Engneerng 15 (2011) 4026 4030 Advanced n Control Engneerng and Informaton Scence A New Evolutonary Computaton Based Approach for Learnng Bayesan Network Yungang
More informationLecture 3 Stat102, Spring 2007
Lecture 3 Stat0, Sprng 007 Chapter 3. 3.: Introducton to regresson analyss Lnear regresson as a descrptve technque The least-squares equatons Chapter 3.3 Samplng dstrbuton of b 0, b. Contnued n net lecture
More informationGlobal Sensitivity. Tuesday 20 th February, 2018
Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values
More informationBoostrapaggregating (Bagging)
Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationSTATS 306B: Unsupervised Learning Spring Lecture 10 April 30
STATS 306B: Unsupervsed Learnng Sprng 2014 Lecture 10 Aprl 30 Lecturer: Lester Mackey Scrbe: Joey Arthur, Rakesh Achanta 10.1 Factor Analyss 10.1.1 Recap Recall the factor analyss (FA) model for lnear
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More informationDepartment of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING
MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/
More informationComputing MLE Bias Empirically
Computng MLE Bas Emprcally Kar Wa Lm Australan atonal Unversty January 3, 27 Abstract Ths note studes the bas arses from the MLE estmate of the rate parameter and the mean parameter of an exponental dstrbuton.
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationMODELING TRAFFIC LIGHTS IN INTERSECTION USING PETRI NETS
The 3 rd Internatonal Conference on Mathematcs and Statstcs (ICoMS-3) Insttut Pertanan Bogor, Indonesa, 5-6 August 28 MODELING TRAFFIC LIGHTS IN INTERSECTION USING PETRI NETS 1 Deky Adzkya and 2 Subono
More informationChapter 14: Logit and Probit Models for Categorical Response Variables
Chapter 4: Logt and Probt Models for Categorcal Response Varables Sect 4. Models for Dchotomous Data We wll dscuss only ths secton of Chap 4, whch s manly about Logstc Regresson, a specal case of the famly
More informationLogistic regression with one predictor. STK4900/ Lecture 7. Program
Logstc regresson wth one redctor STK49/99 - Lecture 7 Program. Logstc regresson wth one redctor 2. Maxmum lkelhood estmaton 3. Logstc regresson wth several redctors 4. Devance and lkelhood rato tests 5.
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationUncertainty as the Overlap of Alternate Conditional Distributions
Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant
More informationENG 8801/ Special Topics in Computer Engineering: Pattern Recognition. Memorial University of Newfoundland Pattern Recognition
EG 880/988 - Specal opcs n Computer Engneerng: Pattern Recognton Memoral Unversty of ewfoundland Pattern Recognton Lecture 7 May 3, 006 http://wwwengrmunca/~charlesr Offce Hours: uesdays hursdays 8:30-9:30
More informationMIMA Group. Chapter 2 Bayesian Decision Theory. School of Computer Science and Technology, Shandong University. Xin-Shun SDU
Group M D L M Chapter Bayesan Decson heory Xn-Shun Xu @ SDU School of Computer Scence and echnology, Shandong Unversty Bayesan Decson heory Bayesan decson theory s a statstcal approach to data mnng/pattern
More informationSpeech and Language Processing
Speech and Language rocessng Lecture 3 ayesan network and ayesan nference Informaton and ommuncatons Engneerng ourse Takahro Shnozak 08//5 Lecture lan (Shnozak s part) I gves the frst 6 lectures about
More informationAN ASYMMETRIC GENERALIZED FGM COPULA AND ITS PROPERTIES
Pa. J. Statst. 015 Vol. 31(1), 95-106 AN ASYMMETRIC GENERALIZED FGM COPULA AND ITS PROPERTIES Berzadeh, H., Parham, G.A. and Zadaram, M.R. Deartment of Statstcs, Shahd Chamran Unversty, Ahvaz, Iran. Corresondng
More informationChapter 9: Statistical Inference and the Relationship between Two Variables
Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,
More informationBasically, if you have a dummy dependent variable you will be estimating a probability.
ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy
More informationCase A. P k = Ni ( 2L i k 1 ) + (# big cells) 10d 2 P k.
THE CELLULAR METHOD In ths lecture, we ntroduce the cellular method as an approach to ncdence geometry theorems lke the Szemeréd-Trotter theorem. The method was ntroduced n the paper Combnatoral complexty
More informationHomework 10 Stat 547. Problem ) Z D!
Homework 0 Stat 547 Problem 74 Notaton: h s the hazard rate for the aneulod grou, h s the hazard rate for the dlod grou (a Log-rank test s erformed: H 0 : h (t = h (t Sgnfcance level α = 005 Test statstc
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationMaurice Cox, Peter Harris National Physical Laboratory, UK Steve Ellison LGC Ltd, UK
CCQM/11-18 Use of an excess-varance aroach for the estmaton of a key comarson reference value, assocated standard uncertanty and degrees of equvalence for CCQM key comarson data Maurce Cox, Peter Harrs
More informationSPANC -- SPlitpole ANalysis Code User Manual
Functonal Descrpton of Code SPANC -- SPltpole ANalyss Code User Manual Author: Dale Vsser Date: 14 January 00 Spanc s a code created by Dale Vsser for easer calbratons of poston spectra from magnetc spectrometer
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationTopology optimization of plate structures subject to initial excitations for minimum dynamic performance index
th World Congress on Structural and Multdsclnary Otmsaton 7 th -2 th, June 25, Sydney Australa oology otmzaton of late structures subject to ntal exctatons for mnmum dynamc erformance ndex Kun Yan, Gengdong
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationINF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018
INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationOnline Classification: Perceptron and Winnow
E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng
More information6. Hamilton s Equations
6. Hamlton s Equatons Mchael Fowler A Dynamcal System s Path n Confguraton Sace and n State Sace The story so far: For a mechancal system wth n degrees of freedom, the satal confguraton at some nstant
More informationTime-Varying Systems and Computations Lecture 6
Tme-Varyng Systems and Computatons Lecture 6 Klaus Depold 14. Januar 2014 The Kalman Flter The Kalman estmaton flter attempts to estmate the actual state of an unknown dscrete dynamcal system, gven nosy
More informationAlgorithms for factoring
CSA E0 235: Crytograhy Arl 9,2015 Instructor: Arta Patra Algorthms for factorng Submtted by: Jay Oza, Nranjan Sngh Introducton Factorsaton of large ntegers has been a wdely studed toc manly because of
More informationOn New Selection Procedures for Unequal Probability Sampling
Int. J. Oen Problems Comt. Math., Vol. 4, o. 1, March 011 ISS 1998-66; Coyrght ICSRS Publcaton, 011 www.-csrs.org On ew Selecton Procedures for Unequal Probablty Samlng Muhammad Qaser Shahbaz, Saman Shahbaz
More informationQUANTITATIVE RISK MANAGEMENT TECHNIQUES USING INTERVAL ANALYSIS, WITH APPLICATIONS TO FINANCE AND INSURANCE
QANTITATIVE RISK MANAGEMENT TECHNIQES SING INTERVA ANAYSIS WITH APPICATIONS TO FINANCE AND INSRANCE Slva DED Ph.D. Bucharest nversty of Economc Studes Deartment of Aled Mathematcs; Romanan Academy Insttute
More informationPsychology 282 Lecture #24 Outline Regression Diagnostics: Outliers
Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.
More informationDependent variable for case i with variance σ 2 g i. Number of distinct cases. Number of independent variables
REGRESSION Notaton Ths rocedure erforms multle lnear regresson wth fve methods for entry and removal of varables. It also rovdes extensve analyss of resdual and nfluental cases. Caseweght (CASEWEIGHT)
More informationOn the correction of the h-index for career length
1 On the correcton of the h-ndex for career length by L. Egghe Unverstet Hasselt (UHasselt), Campus Depenbeek, Agoralaan, B-3590 Depenbeek, Belgum 1 and Unverstet Antwerpen (UA), IBW, Stadscampus, Venusstraat
More informationQuantitative Discrimination of Effective Porosity Using Digital Image Analysis - Implications for Porosity-Permeability Transforms
2004, 66th EAGE Conference, Pars Quanttatve Dscrmnaton of Effectve Porosty Usng Dgtal Image Analyss - Implcatons for Porosty-Permeablty Transforms Gregor P. Eberl 1, Gregor T. Baechle 1, Ralf Weger 1,
More informationBayesian predictive Configural Frequency Analysis
Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse
More informationMultilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata
Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,
More informationResearch Journal of Pure Algebra -2(12), 2012, Page: Available online through ISSN
Research Journal of Pure Algebra (, 0, Page: 37038 Avalable onlne through www.rja.nfo ISSN 48 9037 A NEW GENERALISATION OF SAMSOLAI S MULTIVARIATE ADDITIVE EXPONENTIAL DISTRIBUTION* Dr. G. S. Davd Sam
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.
More informationMLE and Bayesian Estimation. Jie Tang Department of Computer Science & Technology Tsinghua University 2012
MLE and Bayesan Estmaton Je Tang Department of Computer Scence & Technology Tsnghua Unversty 01 1 Lnear Regresson? As the frst step, we need to decde how we re gong to represent the functon f. One example:
More informationOn the Connectedness of the Solution Set for the Weak Vector Variational Inequality 1
Journal of Mathematcal Analyss and Alcatons 260, 15 2001 do:10.1006jmaa.2000.7389, avalable onlne at htt:.dealbrary.com on On the Connectedness of the Soluton Set for the Weak Vector Varatonal Inequalty
More informationImprovement in Estimating the Population Mean Using Exponential Estimator in Simple Random Sampling
Bulletn of Statstcs & Economcs Autumn 009; Volume 3; Number A09; Bull. Stat. Econ. ISSN 0973-70; Copyrght 009 by BSE CESER Improvement n Estmatng the Populaton Mean Usng Eponental Estmator n Smple Random
More informationCHAPTER 3: BAYESIAN DECISION THEORY
HATER 3: BAYESIAN DEISION THEORY Decson mang under uncertanty 3 Data comes from a process that s completely not nown The lac of nowledge can be compensated by modelng t as a random process May be the underlyng
More informationThe Gaussian classifier. Nuno Vasconcelos ECE Department, UCSD
he Gaussan classfer Nuno Vasconcelos ECE Department, UCSD Bayesan decson theory recall that we have state of the world X observatons g decson functon L[g,y] loss of predctng y wth g Bayes decson rule s
More informationMaximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models
ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Mamum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models for
More informationSupplementary Material for Spectral Clustering based on the graph p-laplacian
Sulementary Materal for Sectral Clusterng based on the grah -Lalacan Thomas Bühler and Matthas Hen Saarland Unversty, Saarbrücken, Germany {tb,hen}@csun-sbde May 009 Corrected verson, June 00 Abstract
More informationANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)
Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of
More informationMultiple Regression Analysis
Multle Regresson Analss Roland Szlág Ph.D. Assocate rofessor Correlaton descres the strength of a relatonsh, the degree to whch one varale s lnearl related to another Regresson shows us how to determne
More informationChapter 13: Multiple Regression
Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to
More information1 The Mistake Bound Model
5-850: Advanced Algorthms CMU, Sprng 07 Lecture #: Onlne Learnng and Multplcatve Weghts February 7, 07 Lecturer: Anupam Gupta Scrbe: Bryan Lee,Albert Gu, Eugene Cho he Mstake Bound Model Suppose there
More information