Study of Classification Methods Based on Three Learning Criteria and Two Basis Functions

Size: px
Start display at page:

Download "Study of Classification Methods Based on Three Learning Criteria and Two Basis Functions"

Transcription

1 Study of Classfcaton Methods Based on hree Learnng Crtera and wo Bass Functons Jae Kyu Suhr Abstract - hs paper nvestgates several classfcaton ethods based on the three learnng crtera and two bass functons. he three learnng crtera are the least squares error, the total error rate, and the area under the recever operatng characterstc curve. he two bass functons are reduced ultal odel and sngle-hdden layer feedforward neural networks. In the experent, fve classfcaton ethods were evaluated by usng the UCI database and sutable data ralzaton procedures for each bass functon were dscussed.. Introducton Pattern classfcaton has been a key coponent for decson akng n any research felds. In pattern classfcaton, eprcal learnng consttutes a aor paradg []. Under ths paradg, a classfer s desgned to nze a certan cost functon (crteron) wth a tranng set. Least Squares Error (LSE) has been wdely used as a cost functon for eprcal classfer learnng. he reasons for the popularty of LSE are ts splcty, clear physcal eanng, and tractablty for analyss. he ebedent of nlneartes nto lnear odels has wdened ts applcaton. Recently, two effcent bass functons were proposed: Reduced ultal Model () [] and Extree Learnng Machne () [3]. uses a reduced verson of full polyal and uses a Sngle-hdden Layer Feedforward Neural networks (SLFNs). However, the an proble when usng the LSE cost functon s that t tres to nze the fttng error rather than the classfcaton error durng the learnng process. herefore, three an approaches have been adopted to overcoe ths drawback of the LSE cost functon. hese approaches are the dscrnant approach (Fsher Dscrnant Analyss and Generalzed Dscrnant Analyss), the structural approach (Support Vector Machne), and the classfcaton-error approach. In the thrd approach, two cost functons were recently proposed. One s otal Error Rate (ER) [4,5] and the other s Area under the recever operatng characterstcs curve (AUC) [6]. he an breakthrough of these three papers s a sooth approxate forulaton for calculatng ER and AUC. he step functon used for the countng process s approxated by a quadratc functon nstead of a sgod functon. hs akes t possble to have closed-for soluton. In ths paper, the above entoned fve ethods (,,,, and ) were nvestgated. Fve two-class probles n the UCI database were used for the experent snce can only be appled to two-class proble so far. he paper s organzed as follows: In Secton, the nvestgated classfcaton ethods are brefly descrbed. In Secton 3, experental setup, data ralzaton ssue, and evaluaton results are presented. Fnally, ths paper s concluded wth a suary n Secton 4. Method descrpton. LSE-based ethod Least Squares Error (LSE) s a wdely used cost functon for eprcal classfer learnng due to the several reasons as enton n the prevous secton. Consder an l-densonal nput x and the followng paraetrc odel adoptng a bass expanson ter: K g(, x) = p ( x) = p( x) () k = k k where pk ( x ) corresponds to the kth bass ter of the row vector p( x) = [ p( x), p( x),..., pk ( x )], and = [,,... ] K s a colun paraeter vector to be estated. When we have learnng data pars p( x ) vector can be K extended to P ( ) and a kwn label can be deted by y ( ). In ths case, the LSE cost functon becoes b J ( ) = y P () where b controls the weghtng of the regularzaton factor. he estated tranng output s gven by yˆ = P ˆ, where the soluton for ˆ, whch nzes J s ˆ = ( PPbI) Py (3) hs LSE cost functon has been wden ts applcaton by ebeddng nlneartes nto lnear odels. Recently, two effcent nlnear bass functons were ntroduced. One s the Reduced ultvarate Model () [] and the other s the Extree Learnng Machne () [3]. uses reduced verson of full polyal as a bass functon whch can be expressed as r l r ˆ k f (, x) = x ( x x x ) k rl l k= = = r x x x xl = l r ( )( ),,. uses a Sngle-hdden Layer Feedforward Neural networks (SLFNs) as a bass functon. It randoly chooses hdden des and analytcally deternes the output weghts of SLFNs. Gven arbtrary dstnct saples ( x, y ), =,, where x s a d-densonal nput vector and y s a (4)

2 q-densonal output target vector. he standard SLFNs wth p hdden des and actvaton functon φ can be odeled as p p β φ ( x ) = β φ ( w x b ) = g, =,..., = = where w s the weght vector connectng the th hdden de to the nput des, β s the weght vector connectng the th hdden de to the output des, b s the threshold of the th hdden de, and g s the q-densonal network output. he equatons above can be wrtten ore copactly as Hβ = y (6) where φ( w x b) φ( wp xp bp) H = φ( b) φ( p bp) w x w x p and y s a kwn label. Snce H can be treated as a P atrx n Eq. (), a least-squares soluton can be obtaned for the output weghng paraeters as ˆ β = Hy= ( HH) Hy (8). ER-based ethod Although, LSE -based ethods (for exaple and ) has been wdely used, ts ltaton becoes apparent when hgh accuracy s requred. It s anly because the LSE cost functon nzes the fttng error rather than the classfcaton error whch s ostly desred to be nzed n classfcaton tasks. o overcoe ths drawback, the otal Error Rate (ER) cost functon was proposed [4,5]. ER can be calculated by sung the false postve rate (FP) and the false negatve rate (FN). FP and FN can be expressed as Eq. (9) by detng the postve and negatve exaples of varables by the superscrpts and -, respectvely. FP = Lg ( ( ) τ ), FN Lg ( ( ) τ ) x = x (9) = = where L s a zero-one step loss functon. Let g( x) = g(, x ) wth adustable paraeters operatng on the feature vector x, then the ER can be wrtten as ER(, x, x ) () = Lg ( (, ) τ) L( τ g(, )) x x = = A natural choce to approxate a step functon s a sgod functon [7]. However, ths leads two aor probles. Frst, the forulaton becoes nlnear wth respect to the learnng paraeters. hs akes t dffcult to fnd the optal paraeters. Second, the obectve functon can becoe llcondtoned due to the uch local plateaus. hs causes the optzaton procedure to be te-consung. herefore, a quadratc functon was proposed to approxate a step functon. hs s very effectve snce t results a closedfor classfcaton-error-based soluton. By usng ths, (5) (7) approxaton and g(, x) = p( x), the ER n Eq. () can be rewrtten as b ER(, x, x ) = ( ) τ η px = () τ ( ) η px = where η s a postve offset of a quadratc functon. he LSE soluton of Eq. () calculated by lettng the dervatve of ER(, x, x ) wth respect to the paraeter be zero s ( τ η) ( τ η) = b I p p p p p p () where I s and dentty atrx of K K sze. In a ore copact atrx for, Eq. () can be wrtten as = bi PP P P (3) ( τ η) ( τ η) P P where P and P are the sae as P n Eq. () except that they are produced by usng the postve and negatve saples, respectvely, and = [,...,], = [,...,]. hs ER cost functon based on the quadratc approxaton was appled to the and bass functon n [4,5]. hey are called as and, respectvely..3 AUC-based ethod he recever operatng characterstcs (ROC) curve has been extensvely adopted for evaluatng the classfer perforance. However, the processes of classfer desgn optzaton and the fnal ROC perforance evaluaton are usually conducted separately. hs s anly because the ROC does t have a well-posed structure due to the error countng pont of vew. o overcoe ths drawback, a sooth approxate forulaton by usng a quadratc functon was proposed to calculate the Area under ROC curve (AUC) [6]. hs enables a drect optzaton of the AUC wth respect to the classfer paraeters. he AUC [8] for the gven tranng exaples can be expressed as AUC( x, x ) = (4) g( ) > g( ) x x = = where the ter corresponds to a whenever the g( x ) > g( x ) eleents g( x ) > g( x ), and otherwse. Let g( x) = g(, x ) wth adustable paraeters operatng on the feature vector x, then the goal to optze the classfer s dscrnaton perforance can be treated as to axze the AUC: arg ax AUC(, x, x ) = arg ax u g(, ) g(, ) x x = = (5) where u s a unt step functon. Maxzng the AUC s equal to nzng the Area Above ROC curve (AAC) n Eq. (6).

3 argn AAC(, x, x ) = arg n u g(, ) g(, ) x x = = (6) Snce the approxaton by usng a sgod functon can cause soe probles as entoned n the prevous secton, a quadratc functon was used to approxate a step functon. he approxated AAC after lettng g(, x) = p( x) s argn AAC(, x, x ) b (7) arg n ( ( ) ( )) η px px = = where η s a postve offset of a quadratc functon. he optal paraeter can be obtaned as Eq. (8) by lettng the dervatve of Eq. (7) wth respect to be zero. = bi ( ) ( ) p p p p = = η ( ) p p = = (8) where I s and dentty atrx of K K sze. he optal threshold n the sense of the total error rate crteron [9] can be calculated as τ = ( ) ( ) px px (9) = = 3. Experents 3. Experental setup In the experent, we appled fve dfferent classfcaton ethods to fve two-class probles n the UCI database []. Suares of classfcaton ethods and data sets are shown n able and able, respectvely. he experental setup s as follows. Mn-ax ralzaton was appled to P atrx n case of -based ethods and to the orgnal feature vector n case of based ethods. he reason for ths procedure wll be dscussed later. runs of -fold cross valdaton were perfored. ~ orders for -based ethods and ~ nuber of hdden neurons for -based ethods. A sgod functon was used for -based ethods as an actvaton functon due to ts popularty and effectveness. We fxed τ = η =.5 for and and η = for. he regularzaton factor b was set to -4 for all ethods. o evaluate and copare each ethod, two crtera were used. One s the test classfcaton error rate and the other s the (= -log AUC). For both crtera, a lower percentage eans better perforance. able. Suary of Classfcaton Methods Cost func. LSE ER AUC Bass func. [] [4] [6] SLFNs [3] [5] - able. Suary of Data Sets DB nae #case #feature #class #ss Pa-dabetes No SPEC-heart 67 No StatLog-heart 7 3 No c-tac-toe No No 3. Experental results he experental results consst of two parts. Frst, the feature vector ralzaton procedure s dscussed and then the results of fve dfferent classfcaton ethods are dscussed. For ralzng the feature vectors, we used the n-ax ralzaton ethod whch s kwn as a sple and effectve technque []. Snce several ralzaton ethods have been already copared n that paper, we tred to analyze how to use t when usng dfferent bass functons ( and ). We appled the n-ax ralzaton technque n three dfferent ways: ralzaton, ralzaton before akng P or H atrx, and ralzaton after akng P or H atrx before P atrx after P atrx order(~) sngular value rato (log scale) order(~) Fg.. Coparson of three ralzaton procedures for n ters of classfcaton error and sngular value rato. Fg. shows the classfcaton error rates of wth three dfferent ralzaton procedures by usng data set. In Fg., ralzaton after akng P atrx produces the best perforance and ralzaton produces the worst perforance. he reason that the results of ralzaton before and after akng P atrx are better than the result of ralzaton s the nstablty of the paraeter estaton caused by a sngularty of P P atrx whch should be nverted. hs sngularty coes when generatng P atrx. Snce P atrx s produced by ultplyng and addng any s, the coponent values of ths atrx are qute unbalanced f s are t ralzed. hs akes P P atrx close to sngular n case of ralzaton. One ore portant thng we can tce fro Fg. s that the result of ralzaton after akng P atrx produces a better perforance than the result of ralzaton before akng P atrx. hs s also due to the sngularty of 5 5 before P atrx after P atrx

4 P P atrx. Even though s are ralzed before akng P atrx, the ralzed s are ultpled and added each other whle akng P atrx. hs also produces unbalanced coponent values n P atrx whch causes a sngularty of P P atrx. Fg. shows how uch sngular P P atrx s n each ralzaton procedure and each order. he degree of sngularty was easured by usng a rato between the largest and sallest sngular values of P P atrx. Snce a sngular value ndcates an portance of correspondng sngular vector, t can be sad that a larger sngular value rato eans ore sngularty. Sngular value ratos are presented n log scale. In cases of ralzaton and ralzaton before akng P atrx, sngular value ratos ncrease wth order. hs s because a hgher order leads ore unbalanced coponents. However, ths rato s alost constant n case of ralzaton after P atrx. hs s the reason that ralzaton after akng P atrx produces a good and stable perforance n all orders before H atrx after H atrx hdden neuron (~) hdden neuron (~) Fg.. Coparson of three ralzaton procedures for n ters of classfcaton error and kurtoss value. Fg. shows classfcaton error rates of wth three dfferent ralzaton procedures by usng data set. In Fg., the results of ralzaton and ralzaton before akng H atrx are alost the sae and the result of ralzaton before akng H atrx s the best. he reason that ralzaton after akng H atrx and ralzaton produce slar results are due to the nature of a sgod functon used as an actvaton functon. Fg. 3 shows hstogras of s at each step of n case of ralzaton after akng H atrx. Fg. 3 shows a hstogra of the orgnal s. hese feature values are n the range of about ~35. Fg. 3 shows a hstogra of s after the rando weght and bas. Most of the values are n the range of about -~. Fg. 3 (c) shows a hstogra of s after applyng a sgod functon. Snce a sgod functon stretches out the values around zero and akes negatve sall values zeros and postve large values ones, ost of the s are concentrated on zero and one. herefore, t s alost eanngless to ralze s after akng H atrx. In Fg., the result of ralzaton before akng H atrx s better than the results of others. hs s because the pre-ralzaton akes the s bounded near zero before applyng a sgod functon. Fg. 4,, and (c) show a hstogra of the orgnal s, a hstogra kurtoss value 4 3 before H atrx after H atrx of ralzed s before akng H atrx and a hstogra of s after the rando weght and bas, respectvely. It can be seen that the s after the rando weght and bas are n the range of about -4~5. herefore, after applyng a sgod functon, the feature values becoe uch rcher (Fg. 4 (d)) than those of the ralzaton after akng H atrx (Fg. 3 (d)). hs s the reason that ralzaton before akng H atrx outperfors than the others. Fg. shows kurtoss of the s of H atrx n each ralzaton procedure and each order. hs value s near 3 for Gaussan dstrbuton and for sub-gaussan dstrbuton. Snce the dstrbuton n cases of ralzaton and ralzaton after akng H atrx are shaped lke sub-gaussan, the kurtoss values are close to. In case of ralzaton before akng H atrx, the dstrbuton s shaped lke Gaussan, so the kurtoss values are close to 3. nuber of occurrence nuber of occurrence (c) (d) Fg. 3. Hstogras of s at each step of n case of ralzaton after akng H atrx. nuber of occurrence nuber of occurrence (c) nuber of occurrence nuber of occurrence nuber of occurrence nuber of occurrence (d)

5 Fg. 4. Hstogras of s at each step of n case of ralzaton before akng H atrx order(~), hdden neuron(~) wdbc Pa-dabetes order(~), hdden neuron(~) SPEC-heart order(~), hdden neuron(~) StatLog-heart order(~), hdden neuron(~) tc-tac-toe order(~), hdden neuron(~) order(~), hdden neuron(~) wdbc order(~), hdden neuron(~) Fg. 5. ER and of,,,, and n fve data sets. and coluns show ER and n each data set, respectvely. he experental results of fve ethods by usng fve twoclass probles are shown n Fg. 5. and coluns n ths fgure show the test classfcaton error rates and values n each data set, respectvely. he perforances of all ethods are very slar. Especally, and shows alost the sae perforance n ters of test error rate and. hs s because fnds the optal order(~), hdden neuron(~) SPEC-heart order(~), hdden neuron(~) StatLog-heart order(~), hdden neuron(~) tc-tac-toe Pa-dabetes paraeters wth a fxed threshold to nze the total error rate and fnd the optal threshold wth the fxed paraeters to nze the total error rate. est error rate and value show very slar trend. It ght be sad that there s a strong correlaton between test error rate and. 4. Conclusons In ths paper, fve dfferent classfcaton ethods based on three learnng crtera and two bass functons were nvestgated. Approprate ralzaton procedures for and -based ethods were also dscussed. For two-class probles, fve ethods showed slar perforances, especally, and were qute slar. he results also showed that the classfcaton error rate and are hghly correlated. For data ralzaton, t was shown that the ralzaton should be appled after akng P atrx n case of and before akng H atrx n case of. hs can be generalzed as follows: the data ralzaton procedure should be chosen wth a consderaton of a bass functon propertes. References [] R.O. Duda, P.E. Hart, and D.G. Stork, Pattern Classfcaton, seconded. John Wley & Sons,. [] K.-A. oh, Q.-L. ran, and D. Srnvasan, Bencharkng a reduced ultvarate polyal pattern classfer, IEEE rans. Pattern Analyss and Machne Intellgence, vol. 6,. 6, pp , 4. [3] Huang, G.-B., Zhu, Q.-Y., & Sew, C.-K. (6). Extree learnng achne: heory and applcatons. Neurocoputng, 7, [4] K.-A. oh and H.-L. Eng, Between classfcaton-error approxaton and weghted least-squares learnng, IEEE rans. Pattern Analyss and Machne Intellgence, vol. 3,. 4, pp , 8. [5] K.-A. oh, Deternstc Neural Classfcaton, Neural Coputaton, 8. [6] K.-A. oh, J. K and S. Lee, Maxzng Area Under ROC Curve for Boetrc Scores Fuson, Pattern Recognton, 8. [7] K.-A. oh, Learnng fro arget Kwledge Approxaton, Proc. Frst IEEE Conf. Industral Electroncs and Applcatons, pp. 85-8, May 6. [8] J.A. Hanley, B.J. McNel, he eanng and use of the area under a recever operatng characterstc (ROC) curve, Radology 43 (98) [9] K.-A. oh, Between AUC Based and Error Rate Based Learnng, he 3rd IEEE Conference on Industral Electroncs and Applcatons (ICIEA), Sngapore, June 8. [] D.J. Newan, S. Hettch, C.L. Blake, and C.J. Merz, UCI Repostory of Machne Learnng Databases, Unv. of Calforna, Dept. of Inforaton and Coputer Scences,

6 [] A. K. Jan, K. Nandakuar and A. Ross, Score Noralzaton n Multodal Boetrc Systes, Pattern Recognton, Vol. 38, No., pp. 7-85, Deceber 5.

Study on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr

Study on Classification Methods Based on Three Different Learning Criteria. Jae Kyu Suhr Study on Classification Methods Based on Three Different Learning Criteria Jae Kyu Suhr Contents Introduction Three learning criteria LSE, TER, AUC Methods based on three learning criteria LSE:, ELM TER:

More information

Excess Error, Approximation Error, and Estimation Error

Excess Error, Approximation Error, and Estimation Error E0 370 Statstcal Learnng Theory Lecture 10 Sep 15, 011 Excess Error, Approxaton Error, and Estaton Error Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton So far, we have consdered the fnte saple

More information

System in Weibull Distribution

System in Weibull Distribution Internatonal Matheatcal Foru 4 9 no. 9 94-95 Relablty Equvalence Factors of a Seres-Parallel Syste n Webull Dstrbuton M. A. El-Dacese Matheatcs Departent Faculty of Scence Tanta Unversty Tanta Egypt eldacese@yahoo.co

More information

XII.3 The EM (Expectation-Maximization) Algorithm

XII.3 The EM (Expectation-Maximization) Algorithm XII.3 The EM (Expectaton-Maxzaton) Algorth Toshnor Munaata 3/7/06 The EM algorth s a technque to deal wth varous types of ncoplete data or hdden varables. It can be appled to a wde range of learnng probles

More information

Applied Mathematics Letters

Applied Mathematics Letters Appled Matheatcs Letters 2 (2) 46 5 Contents lsts avalable at ScenceDrect Appled Matheatcs Letters journal hoepage: wwwelseverco/locate/al Calculaton of coeffcents of a cardnal B-splne Gradr V Mlovanovć

More information

What is LP? LP is an optimization technique that allocates limited resources among competing activities in the best possible manner.

What is LP? LP is an optimization technique that allocates limited resources among competing activities in the best possible manner. (C) 998 Gerald B Sheblé, all rghts reserved Lnear Prograng Introducton Contents I. What s LP? II. LP Theor III. The Splex Method IV. Refneents to the Splex Method What s LP? LP s an optzaton technque that

More information

1 Review From Last Time

1 Review From Last Time COS 5: Foundatons of Machne Learnng Rob Schapre Lecture #8 Scrbe: Monrul I Sharf Aprl 0, 2003 Revew Fro Last Te Last te, we were talkng about how to odel dstrbutons, and we had ths setup: Gven - exaples

More information

PROBABILITY AND STATISTICS Vol. III - Analysis of Variance and Analysis of Covariance - V. Nollau ANALYSIS OF VARIANCE AND ANALYSIS OF COVARIANCE

PROBABILITY AND STATISTICS Vol. III - Analysis of Variance and Analysis of Covariance - V. Nollau ANALYSIS OF VARIANCE AND ANALYSIS OF COVARIANCE ANALYSIS OF VARIANCE AND ANALYSIS OF COVARIANCE V. Nollau Insttute of Matheatcal Stochastcs, Techncal Unversty of Dresden, Gerany Keywords: Analyss of varance, least squares ethod, odels wth fxed effects,

More information

1 Definition of Rademacher Complexity

1 Definition of Rademacher Complexity COS 511: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture #9 Scrbe: Josh Chen March 5, 2013 We ve spent the past few classes provng bounds on the generalzaton error of PAClearnng algorths for the

More information

COS 511: Theoretical Machine Learning

COS 511: Theoretical Machine Learning COS 5: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture #0 Scrbe: José Sões Ferrera March 06, 203 In the last lecture the concept of Radeacher coplexty was ntroduced, wth the goal of showng that

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fttng of Data Davd Eberly Geoetrc Tools, LLC http://www.geoetrctools.co/ Copyrght c 1998-2014. All Rghts Reserved. Created: July 15, 1999 Last Modfed: February 9, 2008 Contents 1 Lnear Fttng

More information

Computational and Statistical Learning theory Assignment 4

Computational and Statistical Learning theory Assignment 4 Coputatonal and Statstcal Learnng theory Assgnent 4 Due: March 2nd Eal solutons to : karthk at ttc dot edu Notatons/Defntons Recall the defnton of saple based Radeacher coplexty : [ ] R S F) := E ɛ {±}

More information

y new = M x old Feature Selection: Linear Transformations Constraint Optimization (insertion)

y new = M x old Feature Selection: Linear Transformations Constraint Optimization (insertion) Feature Selecton: Lnear ransforatons new = M x old Constrant Optzaton (nserton) 3 Proble: Gven an objectve functon f(x) to be optzed and let constrants be gven b h k (x)=c k, ovng constants to the left,

More information

LECTURE :FACTOR ANALYSIS

LECTURE :FACTOR ANALYSIS LCUR :FACOR ANALYSIS Rta Osadchy Based on Lecture Notes by A. Ng Motvaton Dstrbuton coes fro MoG Have suffcent aount of data: >>n denson Use M to ft Mture of Gaussans nu. of tranng ponts If

More information

Gradient Descent Learning and Backpropagation

Gradient Descent Learning and Backpropagation Artfcal Neural Networks (art 2) Chrstan Jacob Gradent Descent Learnng and Backpropagaton CSC 533 Wnter 200 Learnng by Gradent Descent Defnton of the Learnng roble Let us start wth the sple case of lnear

More information

Recap: the SVM problem

Recap: the SVM problem Machne Learnng 0-70/5-78 78 Fall 0 Advanced topcs n Ma-Margn Margn Learnng Erc Xng Lecture 0 Noveber 0 Erc Xng @ CMU 006-00 Recap: the SVM proble We solve the follong constraned opt proble: a s.t. J 0

More information

Least Squares Fitting of Data

Least Squares Fitting of Data Least Squares Fttng of Data Davd Eberly Geoetrc Tools, LLC http://www.geoetrctools.co/ Copyrght c 1998-2015. All Rghts Reserved. Created: July 15, 1999 Last Modfed: January 5, 2015 Contents 1 Lnear Fttng

More information

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup

BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS. Dariusz Biskup BAYESIAN CURVE FITTING USING PIECEWISE POLYNOMIALS Darusz Bskup 1. Introducton The paper presents a nonparaetrc procedure for estaton of an unknown functon f n the regresson odel y = f x + ε = N. (1) (

More information

Chapter 12 Lyes KADEM [Thermodynamics II] 2007

Chapter 12 Lyes KADEM [Thermodynamics II] 2007 Chapter 2 Lyes KDEM [Therodynacs II] 2007 Gas Mxtures In ths chapter we wll develop ethods for deternng therodynac propertes of a xture n order to apply the frst law to systes nvolvng xtures. Ths wll be

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

Designing Fuzzy Time Series Model Using Generalized Wang s Method and Its application to Forecasting Interest Rate of Bank Indonesia Certificate

Designing Fuzzy Time Series Model Using Generalized Wang s Method and Its application to Forecasting Interest Rate of Bank Indonesia Certificate The Frst Internatonal Senar on Scence and Technology, Islac Unversty of Indonesa, 4-5 January 009. Desgnng Fuzzy Te Seres odel Usng Generalzed Wang s ethod and Its applcaton to Forecastng Interest Rate

More information

Regularized Discriminant Analysis for Face Recognition

Regularized Discriminant Analysis for Face Recognition 1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths

More information

General Averaged Divergence Analysis

General Averaged Divergence Analysis General Averaged Dvergence Analyss Dacheng ao, Xuelong 2, Xndong u 3,, and Stephen J Maybank 2 Departent of Coputng, Hong Kong Polytechnc Unversty, Hong Kong 2 Sch Coputer Scence & Inforaton Systes, Brkbeck,

More information

COMP th April, 2007 Clement Pang

COMP th April, 2007 Clement Pang COMP 540 12 th Aprl, 2007 Cleent Pang Boostng Cobnng weak classers Fts an Addtve Model Is essentally Forward Stagewse Addtve Modelng wth Exponental Loss Loss Functons Classcaton: Msclasscaton, Exponental,

More information

Collaborative Filtering Recommendation Algorithm

Collaborative Filtering Recommendation Algorithm Vol.141 (GST 2016), pp.199-203 http://dx.do.org/10.14257/astl.2016.141.43 Collaboratve Flterng Recoendaton Algorth Dong Lang Qongta Teachers College, Haou 570100, Chna, 18689851015@163.co Abstract. Ths

More information

Multilayer Perceptron (MLP)

Multilayer Perceptron (MLP) Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method

More information

Three Algorithms for Flexible Flow-shop Scheduling

Three Algorithms for Flexible Flow-shop Scheduling Aercan Journal of Appled Scences 4 (): 887-895 2007 ISSN 546-9239 2007 Scence Publcatons Three Algorths for Flexble Flow-shop Schedulng Tzung-Pe Hong, 2 Pe-Yng Huang, 3 Gwoboa Horng and 3 Chan-Lon Wang

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

,..., k N. , k 2. ,..., k i. The derivative with respect to temperature T is calculated by using the chain rule: & ( (5) dj j dt = "J j. k i.

,..., k N. , k 2. ,..., k i. The derivative with respect to temperature T is calculated by using the chain rule: & ( (5) dj j dt = J j. k i. Suppleentary Materal Dervaton of Eq. 1a. Assue j s a functon of the rate constants for the N coponent reactons: j j (k 1,,..., k,..., k N ( The dervatve wth respect to teperature T s calculated by usng

More information

Our focus will be on linear systems. A system is linear if it obeys the principle of superposition and homogenity, i.e.

Our focus will be on linear systems. A system is linear if it obeys the principle of superposition and homogenity, i.e. SSTEM MODELLIN In order to solve a control syste proble, the descrptons of the syste and ts coponents ust be put nto a for sutable for analyss and evaluaton. The followng ethods can be used to odel physcal

More information

Denote the function derivatives f(x) in given points. x a b. Using relationships (1.2), polynomials (1.1) are written in the form

Denote the function derivatives f(x) in given points. x a b. Using relationships (1.2), polynomials (1.1) are written in the form SET OF METHODS FO SOUTION THE AUHY POBEM FO STIFF SYSTEMS OF ODINAY DIFFEENTIA EUATIONS AF atypov and YuV Nulchev Insttute of Theoretcal and Appled Mechancs SB AS 639 Novosbrs ussa Introducton A constructon

More information

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede. ) with a symmetric Pcovariance matrix of the y( x ) measurements V

Fall 2012 Analysis of Experimental Measurements B. Eisenstein/rev. S. Errede. ) with a symmetric Pcovariance matrix of the y( x ) measurements V Fall Analyss o Experental Measureents B Esensten/rev S Errede General Least Squares wth General Constrants: Suppose we have easureents y( x ( y( x, y( x,, y( x wth a syetrc covarance atrx o the y( x easureents

More information

Machine Learning. What is a good Decision Boundary? Support Vector Machines

Machine Learning. What is a good Decision Boundary? Support Vector Machines Machne Learnng 0-70/5 70/5-78 78 Sprng 200 Support Vector Machnes Erc Xng Lecture 7 March 5 200 Readng: Chap. 6&7 C.B book and lsted papers Erc Xng @ CMU 2006-200 What s a good Decson Boundar? Consder

More information

An Accurate Measure for Multilayer Perceptron Tolerance to Weight Deviations

An Accurate Measure for Multilayer Perceptron Tolerance to Weight Deviations Neural Processng Letters 10: 121 130, 1999. 1999 Kluwer Acadec Publshers. Prnted n the Netherlands. 121 An Accurate Measure for Multlayer Perceptron Tolerance to Weght Devatons JOSE L. BERNIER, J. ORTEGA,

More information

Xiangwen Li. March 8th and March 13th, 2001

Xiangwen Li. March 8th and March 13th, 2001 CS49I Approxaton Algorths The Vertex-Cover Proble Lecture Notes Xangwen L March 8th and March 3th, 00 Absolute Approxaton Gven an optzaton proble P, an algorth A s an approxaton algorth for P f, for an

More information

Slobodan Lakić. Communicated by R. Van Keer

Slobodan Lakić. Communicated by R. Van Keer Serdca Math. J. 21 (1995), 335-344 AN ITERATIVE METHOD FOR THE MATRIX PRINCIPAL n-th ROOT Slobodan Lakć Councated by R. Van Keer In ths paper we gve an teratve ethod to copute the prncpal n-th root and

More information

Pop-Click Noise Detection Using Inter-Frame Correlation for Improved Portable Auditory Sensing

Pop-Click Noise Detection Using Inter-Frame Correlation for Improved Portable Auditory Sensing Advanced Scence and Technology Letters, pp.164-168 http://dx.do.org/10.14257/astl.2013 Pop-Clc Nose Detecton Usng Inter-Frame Correlaton for Improved Portable Audtory Sensng Dong Yun Lee, Kwang Myung Jeon,

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

The Parity of the Number of Irreducible Factors for Some Pentanomials

The Parity of the Number of Irreducible Factors for Some Pentanomials The Party of the Nuber of Irreducble Factors for Soe Pentanoals Wolfra Koepf 1, Ryul K 1 Departent of Matheatcs Unversty of Kassel, Kassel, F. R. Gerany Faculty of Matheatcs and Mechancs K Il Sung Unversty,

More information

Elastic Collisions. Definition: two point masses on which no external forces act collide without losing any energy.

Elastic Collisions. Definition: two point masses on which no external forces act collide without losing any energy. Elastc Collsons Defnton: to pont asses on hch no external forces act collde thout losng any energy v Prerequstes: θ θ collsons n one denson conservaton of oentu and energy occurs frequently n everyday

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

AN ANALYSIS OF A FRACTAL KINETICS CURVE OF SAVAGEAU

AN ANALYSIS OF A FRACTAL KINETICS CURVE OF SAVAGEAU AN ANALYI OF A FRACTAL KINETIC CURE OF AAGEAU by John Maloney and Jack Hedel Departent of Matheatcs Unversty of Nebraska at Oaha Oaha, Nebraska 688 Eal addresses: aloney@unoaha.edu, jhedel@unoaha.edu Runnng

More information

Gadjah Mada University, Indonesia. Yogyakarta State University, Indonesia Karangmalang Yogyakarta 55281

Gadjah Mada University, Indonesia. Yogyakarta State University, Indonesia Karangmalang Yogyakarta 55281 Reducng Fuzzy Relatons of Fuzzy Te Seres odel Usng QR Factorzaton ethod and Its Applcaton to Forecastng Interest Rate of Bank Indonesa Certfcate Agus aan Abad Subanar Wdodo 3 Sasubar Saleh 4 Ph.D Student

More information

Integral Transforms and Dual Integral Equations to Solve Heat Equation with Mixed Conditions

Integral Transforms and Dual Integral Equations to Solve Heat Equation with Mixed Conditions Int J Open Probles Copt Math, Vol 7, No 4, Deceber 214 ISSN 1998-6262; Copyrght ICSS Publcaton, 214 www-csrsorg Integral Transfors and Dual Integral Equatons to Solve Heat Equaton wth Mxed Condtons Naser

More information

Several generation methods of multinomial distributed random number Tian Lei 1, a,linxihe 1,b,Zhigang Zhang 1,c

Several generation methods of multinomial distributed random number Tian Lei 1, a,linxihe 1,b,Zhigang Zhang 1,c Internatonal Conference on Appled Scence and Engneerng Innovaton (ASEI 205) Several generaton ethods of ultnoal dstrbuted rando nuber Tan Le, a,lnhe,b,zhgang Zhang,c School of Matheatcs and Physcs, USTB,

More information

ITERATIVE ESTIMATION PROCEDURE FOR GEOSTATISTICAL REGRESSION AND GEOSTATISTICAL KRIGING

ITERATIVE ESTIMATION PROCEDURE FOR GEOSTATISTICAL REGRESSION AND GEOSTATISTICAL KRIGING ESE 5 ITERATIVE ESTIMATION PROCEDURE FOR GEOSTATISTICAL REGRESSION AND GEOSTATISTICAL KRIGING Gven a geostatstcal regresson odel: k Y () s x () s () s x () s () s, s R wth () unknown () E[ ( s)], s R ()

More information

halftoning Journal of Electronic Imaging, vol. 11, no. 4, Oct Je-Ho Lee and Jan P. Allebach

halftoning Journal of Electronic Imaging, vol. 11, no. 4, Oct Je-Ho Lee and Jan P. Allebach olorant-based drect bnary search» halftonng Journal of Electronc Iagng, vol., no. 4, Oct. 22 Je-Ho Lee and Jan P. Allebach School of Electrcal Engneerng & oputer Scence Kyungpook Natonal Unversty Abstract

More information

Fermi-Dirac statistics

Fermi-Dirac statistics UCC/Physcs/MK/EM/October 8, 205 Fer-Drac statstcs Fer-Drac dstrbuton Matter partcles that are eleentary ostly have a type of angular oentu called spn. hese partcles are known to have a agnetc oent whch

More information

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD

CHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD CHALMERS, GÖTEBORGS UNIVERSITET SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 35, FIM 72 GU, PhD Tme: Place: Teachers: Allowed materal: Not allowed: January 2, 28, at 8 3 2 3 SB

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

A KERNEL FUZZY DISCRIMINANT ANALYSIS MINIMUM DISTANCE-BASED APPROACH FOR THE CLASSIFICATION OF FACE IMAGES

A KERNEL FUZZY DISCRIMINANT ANALYSIS MINIMUM DISTANCE-BASED APPROACH FOR THE CLASSIFICATION OF FACE IMAGES Journal of Matheatcal Scences: Advances and Applcatons Volue 29, 204, Pages 75-97 A KERNEL FUZZY DISCRIMINAN ANALYSIS MINIMUM DISANCE-BASED APPROACH FOR HE CLASSIFICAION OF FACE IMAGES College of Coputer

More information

Quantum Particle Motion in Physical Space

Quantum Particle Motion in Physical Space Adv. Studes Theor. Phys., Vol. 8, 014, no. 1, 7-34 HIKARI Ltd, www.-hkar.co http://dx.do.org/10.1988/astp.014.311136 Quantu Partcle Moton n Physcal Space A. Yu. Saarn Dept. of Physcs, Saara State Techncal

More information

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

More information

Lecture Slides for. ETHEM ALPAYDIN The MIT Press,

Lecture Slides for. ETHEM ALPAYDIN The MIT Press, ecture Sldes for ETHEM APAYDI The MIT Press, 00 alpaydn@boun.edu.tr http://www.cpe.boun.edu.tr/~ethe/le Introducton Questons: Assessent of the expected error of a learnng algorth: Is the error rate of

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

EXAMPLES of THEORETICAL PROBLEMS in the COURSE MMV031 HEAT TRANSFER, version 2017

EXAMPLES of THEORETICAL PROBLEMS in the COURSE MMV031 HEAT TRANSFER, version 2017 EXAMPLES of THEORETICAL PROBLEMS n the COURSE MMV03 HEAT TRANSFER, verson 207 a) What s eant by sotropc ateral? b) What s eant by hoogeneous ateral? 2 Defne the theral dffusvty and gve the unts for the

More information

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland

More information

Centroid Uncertainty Bounds for Interval Type-2 Fuzzy Sets: Forward and Inverse Problems

Centroid Uncertainty Bounds for Interval Type-2 Fuzzy Sets: Forward and Inverse Problems Centrod Uncertanty Bounds for Interval Type-2 Fuzzy Sets: Forward and Inverse Probles Jerry M. Mendel and Hongwe Wu Sgnal and Iage Processng Insttute Departent of Electrcal Engneerng Unversty of Southern

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

A Differential Evaluation Markov Chain Monte Carlo algorithm for Bayesian Model Updating M. Sherri a, I. Boulkaibet b, T. Marwala b, M. I.

A Differential Evaluation Markov Chain Monte Carlo algorithm for Bayesian Model Updating M. Sherri a, I. Boulkaibet b, T. Marwala b, M. I. A Dfferental Evaluaton Markov Chan Monte Carlo algorth for Bayesan Model Updatng M. Sherr a, I. Boulkabet b, T. Marwala b, M. I. Frswell c, a Departent of Mechancal Engneerng Scence, Unversty of Johannesburg,

More information

An Optimal Bound for Sum of Square Roots of Special Type of Integers

An Optimal Bound for Sum of Square Roots of Special Type of Integers The Sxth Internatonal Syposu on Operatons Research and Its Applcatons ISORA 06 Xnang, Chna, August 8 12, 2006 Copyrght 2006 ORSC & APORC pp. 206 211 An Optal Bound for Su of Square Roots of Specal Type

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng

More information

Solving Fuzzy Linear Programming Problem With Fuzzy Relational Equation Constraint

Solving Fuzzy Linear Programming Problem With Fuzzy Relational Equation Constraint Intern. J. Fuzz Maeatcal Archve Vol., 0, -0 ISSN: 0 (P, 0 0 (onlne Publshed on 0 Septeber 0 www.researchasc.org Internatonal Journal of Solvng Fuzz Lnear Prograng Proble W Fuzz Relatonal Equaton Constrant

More information

CHAPTER 6 CONSTRAINED OPTIMIZATION 1: K-T CONDITIONS

CHAPTER 6 CONSTRAINED OPTIMIZATION 1: K-T CONDITIONS Chapter 6: Constraned Optzaton CHAPER 6 CONSRAINED OPIMIZAION : K- CONDIIONS Introducton We now begn our dscusson of gradent-based constraned optzaton. Recall that n Chapter 3 we looked at gradent-based

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

SDMML HT MSc Problem Sheet 4

SDMML HT MSc Problem Sheet 4 SDMML HT 06 - MSc Problem Sheet 4. The recever operatng characterstc ROC curve plots the senstvty aganst the specfcty of a bnary classfer as the threshold for dscrmnaton s vared. Let the data space be

More information

FUZZY MODEL FOR FORECASTING INTEREST RATE OF BANK INDONESIA CERTIFICATE

FUZZY MODEL FOR FORECASTING INTEREST RATE OF BANK INDONESIA CERTIFICATE he 3 rd Internatonal Conference on Quanttatve ethods ISBN 979-989 Used n Econoc and Busness. June 6-8, 00 FUZZY ODEL FOR FORECASING INERES RAE OF BANK INDONESIA CERIFICAE Agus aan Abad, Subanar, Wdodo

More information

Evaluation of classifiers MLPs

Evaluation of classifiers MLPs Lecture Evaluaton of classfers MLPs Mlos Hausrecht mlos@cs.ptt.edu 539 Sennott Square Evaluaton For any data set e use to test the model e can buld a confuson matrx: Counts of examples th: class label

More information

A goodness-of-fit measure for a system-of-equations model

A goodness-of-fit measure for a system-of-equations model Songklanakarn J. Sc. Technol. 32 (5), 519-525, Sep. - Oct. 2010 Orgnal Artcle A goodness-of-ft easure for a syste-of-equatons odel Jrawan Jtthavech* School of Appled Statstcs Natonal Insttute of Developent

More information

On Pfaff s solution of the Pfaff problem

On Pfaff s solution of the Pfaff problem Zur Pfaff scen Lösung des Pfaff scen Probles Mat. Ann. 7 (880) 53-530. On Pfaff s soluton of te Pfaff proble By A. MAYER n Lepzg Translated by D. H. Delpenc Te way tat Pfaff adopted for te ntegraton of

More information

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata

Multilayer Perceptrons and Backpropagation. Perceptrons. Recap: Perceptrons. Informatics 1 CG: Lecture 6. Mirella Lapata Multlayer Perceptrons and Informatcs CG: Lecture 6 Mrella Lapata School of Informatcs Unversty of Ednburgh mlap@nf.ed.ac.uk Readng: Kevn Gurney s Introducton to Neural Networks, Chapters 5 6.5 January,

More information

Reliability estimation in Pareto-I distribution based on progressively type II censored sample with binomial removals

Reliability estimation in Pareto-I distribution based on progressively type II censored sample with binomial removals Journal of Scentfc esearch Developent (): 08-3 05 Avalable onlne at wwwjsradorg ISSN 5-7569 05 JSAD elablty estaton n Pareto-I dstrbuton based on progressvely type II censored saple wth bnoal reovals Ilhan

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Optimal Marketing Strategies for a Customer Data Intermediary. Technical Appendix

Optimal Marketing Strategies for a Customer Data Intermediary. Technical Appendix Optal Marketng Strateges for a Custoer Data Interedary Techncal Appendx oseph Pancras Unversty of Connectcut School of Busness Marketng Departent 00 Hllsde Road, Unt 04 Storrs, CT 0669-04 oseph.pancras@busness.uconn.edu

More information

Introducing Entropy Distributions

Introducing Entropy Distributions Graubner, Schdt & Proske: Proceedngs of the 6 th Internatonal Probablstc Workshop, Darstadt 8 Introducng Entropy Dstrbutons Noel van Erp & Peter van Gelder Structural Hydraulc Engneerng and Probablstc

More information

On the Eigenspectrum of the Gram Matrix and the Generalisation Error of Kernel PCA (Shawe-Taylor, et al. 2005) Ameet Talwalkar 02/13/07

On the Eigenspectrum of the Gram Matrix and the Generalisation Error of Kernel PCA (Shawe-Taylor, et al. 2005) Ameet Talwalkar 02/13/07 On the Egenspectru of the Gra Matr and the Generalsaton Error of Kernel PCA Shawe-aylor, et al. 005 Aeet alwalar 0/3/07 Outlne Bacground Motvaton PCA, MDS Isoap Kernel PCA Generalsaton Error of Kernel

More information

On Syndrome Decoding of Punctured Reed-Solomon and Gabidulin Codes 1

On Syndrome Decoding of Punctured Reed-Solomon and Gabidulin Codes 1 Ffteenth Internatonal Workshop on Algebrac and Cobnatoral Codng Theory June 18-24, 2016, Albena, Bulgara pp. 35 40 On Syndroe Decodng of Punctured Reed-Soloon and Gabduln Codes 1 Hannes Bartz hannes.bartz@tu.de

More information

PHYS 1443 Section 002 Lecture #20

PHYS 1443 Section 002 Lecture #20 PHYS 1443 Secton 002 Lecture #20 Dr. Jae Condtons for Equlbru & Mechancal Equlbru How to Solve Equlbru Probles? A ew Exaples of Mechancal Equlbru Elastc Propertes of Solds Densty and Specfc Gravty lud

More information

VERIFICATION OF FE MODELS FOR MODEL UPDATING

VERIFICATION OF FE MODELS FOR MODEL UPDATING VERIFICATION OF FE MODELS FOR MODEL UPDATING G. Chen and D. J. Ewns Dynacs Secton, Mechancal Engneerng Departent Iperal College of Scence, Technology and Medcne London SW7 AZ, Unted Kngdo Eal: g.chen@c.ac.uk

More information

Scattering by a perfectly conducting infinite cylinder

Scattering by a perfectly conducting infinite cylinder Scatterng by a perfectly conductng nfnte cylnder Reeber that ths s the full soluton everywhere. We are actually nterested n the scatterng n the far feld lt. We agan use the asyptotc relatonshp exp exp

More information

1. Statement of the problem

1. Statement of the problem Volue 14, 010 15 ON THE ITERATIVE SOUTION OF A SYSTEM OF DISCRETE TIMOSHENKO EQUATIONS Peradze J. and Tsklaur Z. I. Javakhshvl Tbls State Uversty,, Uversty St., Tbls 0186, Georga Georgan Techcal Uversty,

More information

A Knowledge-Based Feature Selection Method for Text Categorization

A Knowledge-Based Feature Selection Method for Text Categorization A Knowledge-Based Feature Selecton Method for Text Categorzaton Yan Xu,2, JnTao L, Bn Wang,ChunMng Sun,2 Insttute of Coputng Technology,Chnese Acadey of Scences No.6 Kexueyuan South Road, Zhongguancun,Hadan

More information

arxiv: v2 [math.co] 3 Sep 2017

arxiv: v2 [math.co] 3 Sep 2017 On the Approxate Asyptotc Statstcal Independence of the Peranents of 0- Matrces arxv:705.0868v2 ath.co 3 Sep 207 Paul Federbush Departent of Matheatcs Unversty of Mchgan Ann Arbor, MI, 4809-043 Septeber

More information

The Non-equidistant New Information Optimizing MGM(1,n) Based on a Step by Step Optimum Constructing Background Value

The Non-equidistant New Information Optimizing MGM(1,n) Based on a Step by Step Optimum Constructing Background Value Appl. Math. Inf. Sc. 6 No. 3 745-750 (0) 745 Appled Matheatcs & Inforaton Scences An Internatonal Journal The Non-equdstant New Inforaton Optzng MGM(n) Based on a Step by Step Optu Constructng Background

More information

Towards strong security in embedded and pervasive systems: energy and area optimized serial polynomial multipliers in GF(2 k )

Towards strong security in embedded and pervasive systems: energy and area optimized serial polynomial multipliers in GF(2 k ) Towards strong securty n ebedded and pervasve systes: energy and area optzed seral polynoal ultplers n GF( k ) Zoya Dyka, Peter Langendoerfer, Frank Vater and Steffen Peter IHP, I Technologepark 5, D-53

More information

Learning Theory: Lecture Notes

Learning Theory: Lecture Notes Learnng Theory: Lecture Notes Lecturer: Kamalka Chaudhur Scrbe: Qush Wang October 27, 2012 1 The Agnostc PAC Model Recall that one of the constrants of the PAC model s that the data dstrbuton has to be

More information

Solving Nonlinear Differential Equations by a Neural Network Method

Solving Nonlinear Differential Equations by a Neural Network Method Solvng Nonlnear Dfferental Equatons by a Neural Network Method Luce P. Aarts and Peter Van der Veer Delft Unversty of Technology, Faculty of Cvlengneerng and Geoscences, Secton of Cvlengneerng Informatcs,

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

PGM Learning Tasks and Metrics

PGM Learning Tasks and Metrics Probablstc Graphcal odels Learnng Overvew PG Learnng Tasks and etrcs Learnng doan epert True dstrbuton P* aybe correspondng to a PG * dataset of nstances D{d],...d]} sapled fro P* elctaton Network Learnng

More information

Pattern Classification

Pattern Classification Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher

More information

CS 548: Computer Vision Machine Learning - Part 1. Spring 2016 Dr. Michael J. Reale

CS 548: Computer Vision Machine Learning - Part 1. Spring 2016 Dr. Michael J. Reale CS 548: Coputer Vson Machne Learnng - Part Sprng 206 Dr. Mchael J. Reale Credt Where Credt Is Due ROC tutoral: http://g.unc.edu/dxtests/roc3.ht OpenCV tutorals on Machne Learnng: http://docs.opencv.org/trunk/doc/py_tutorals/py_l/py_table_of_contents

More information

Chapter 25: Machining Centers, Machine Tool Structures and Machining Economics

Chapter 25: Machining Centers, Machine Tool Structures and Machining Economics Manufacturng Engneerng echnology n SI Unts, 6 th Edton Chapter 25: Machnng Centers, Machne ool Structures and Machnng Econocs Copyrght 200 Pearson Educaton South Asa Pte Ltd Chapter Outlne 2 Introducton

More information

CHAPTER 7 CONSTRAINED OPTIMIZATION 1: THE KARUSH-KUHN-TUCKER CONDITIONS

CHAPTER 7 CONSTRAINED OPTIMIZATION 1: THE KARUSH-KUHN-TUCKER CONDITIONS CHAPER 7 CONSRAINED OPIMIZAION : HE KARUSH-KUHN-UCKER CONDIIONS 7. Introducton We now begn our dscusson of gradent-based constraned optzaton. Recall that n Chapter 3 we looked at gradent-based unconstraned

More information

Machine Learning. Support Vector Machines. Eric Xing. Lecture 4, August 12, Reading: Eric CMU,

Machine Learning. Support Vector Machines. Eric Xing. Lecture 4, August 12, Reading: Eric CMU, Machne Learnng Support Vector Machnes Erc Xng Lecture 4 August 2 200 Readng: Erc Xng @ CMU 2006-200 Erc Xng @ CMU 2006-200 2 What s a good Decson Boundar? Wh e a have such boundares? Irregular dstrbuton

More information

Solutions for Homework #9

Solutions for Homework #9 Solutons for Hoewor #9 PROBEM. (P. 3 on page 379 n the note) Consder a sprng ounted rgd bar of total ass and length, to whch an addtonal ass s luped at the rghtost end. he syste has no dapng. Fnd the natural

More information