Improving the performance of radial basis function classifiers in condition monitoring and fault diagnosis applications where unknown faults may occur
|
|
- Oswald Mason
- 6 years ago
- Views:
Transcription
1 Improvng the performance of radal bass functon classfers n condton montorng and fault dagnoss applcatons where unknown faults may occur Yuhua L, Mchael J. Pont and N. Barre Jones Control & Instrumentaton Research Secton, Department of Engneerng, Unversty of Lecester, Lecester, LE 7RH, Unted Kngdom Abstract: Ths paper presents a novel technque whch may be used to determne an approprate threshold for nterpretng the outputs of a traned Radal Bass Functon (RBF) classfer. Results from two experments demonstrate that ths method can be used to mprove the performance of RBF classfers n practcal applcatons. Keywords: Radal bass functon network, condton montorng, novelty detecton, threshold determnaton Pre-prnt of: L, Y., Pont, M.J. and Jones, N.B. (2002) Improvng the performance of radal bass functon classfers n condton montorng and fault dagnoss applcatons where unknown faults may occur, Pattern Recognton Letters, 23: To whom correspondence should be addressed.
2 . Introducton Radal Bass Functon (RBF) neural networks (Broomhead and Lowe, 988) have been wdely used n classfcaton problems such as speech recognton, medcal dagnoss, handwrtng recognton, mage processng, and fault dagnoss. In such applcatons, RBF networks are frequently used wth a wnner takes all (WTA) output rule (Scholkopf et al., 997). Ths rule s smple to use, and s often an approprate soluton (Cordella et al., 995). However, n ths paper, we are prmarly concerned wth condton montorng and fault dagnoss (CMFD) systems: for such systems, the WTA rule s not always approprate. Instead, the outputs of the classfer may be nterpreted by applyng a threshold to the output vector: f the value of an output neuron exceeds the gven threshold, then an example of the correspondng class s sad to have occurred (e.g. Josh et al., 997; Cheon et al., 993). There are two reasons why threshold-based output nterpretatons are approprate n CMFD systems. The frst reason s that a CMFD classfer s often requred not only to classfy known normal and fault nput vectors, but also to recognse that a partcular nput s nether normal, nor a member of one of the exstng fault categores: ths process s often called novelty detecton (Bshop, 994). By usng a threshold-based classfer, outputs may be readly nterpreted as an unknown fault when none of the normal or fault output neurons exceeds the threshold. The second reason for usng threshold-based output nterpretatons s that the WTA classfer s unsutable for use n applcatons where an nput vector may need to be assgned to more than one class: ths s often the case n CMFD applcatons, where multple faults may occur smultaneously (see, for example, Cheon et al., 993). Despte these potental advantages, use of a threshold classfer for CMFD applcatons can be problematc, because the overall performance of the system depends on the use of an approprate threshold value: n most publshed work n ths area, thresholds are smply emprcally set, usually at values of 0.5 (Watanabe et al., 994), a process whch does not necessarly result n optmal classfer performance (Josh et al., 997). To mprove the performance of threshold classfers n CMFD and other applcaton areas, methods for dentfyng the optmal threshold values are requred. 2
3 In ths paper, we present a novel technque for determnng an approprate threshold for RBF classfers: ths technque s based on an analyss of the relatonshp between the behavour of a well-traned RBF classfer and ts response to the data set. The paper s organsed as follows: n Secton 2, a method for determnng a sutable threshold s derved, based on theoretcal consderatons. Followng that, the results of two emprcal tests are presented n Secton 3. The results are dscussed n Secton Theoretcal Consderatons A technque for relable threshold selecton n RBF classfers wll be derved n ths secton. The problem of threshold selecton for neural network classfers s frst formulated n Secton 2.. In Secton 2.2, the decson behavour of RBF classfers s mathematcally and geometrcally analysed. The relable threshold selecton method for RBF classfers s then proposed n Secton The general problem A basc learnng problem can be represented by sx components (Smolensky et al., 996): X, Y, A, H, P, and L. The frst four components are the nstance(nput vector), outcome, decson, and decson rule spaces, respectvely. X s an arbtrary set, Y, A { 0, }, and H s a famly of functons from X nto A. The ffth component, P, s a famly of jont probablty dstrbutons of Z = X Y. These represent the possble states of nature that mght be governng the generaton of examples. The last component, the loss functon L, s a mappng from Y A nto the real number set R. Ths paper manly concerns the behavour of A and L wth respect to the threshold. Assume that the classfer has been well traned wth samples, z r ( z,, z m ), where z = ( x, y ) Z, drawn ndependently at random accordng to some probablty dstrbuton P P. After tranng, the classfer s appled to a set of samples whose class s known. Suppose the outcome of output neurons s $y, a threshold s appled to $y, then we have the hypothess h H that specfes the approprate acton a A as: h: X A, a( x, τ) = θ($ y τ ) () 3
4 0 f δ 0 where: θ( δ) = f δ > 0 ths gves () the followng values: 0 a( x, τ) = θ($ y τ) = f f y$ τ y$ > τ (2) a equals one for ndcatng the class occurred, zero otherwse. We consder the followng loss functon: 0 L( z, τ) = L( y, a( x, τ)) = f f a = y a y (3) and the rsk functon (Scholkopf et al., 997): I( τ) = L( z, τ) dp( z) (4) snce the probablty dstrbuton functon P(z) s unknown but a random, ndependent sample of pars z = ( x, y ) s gven, then we have nstead the emprcal rsk functon from (4) m I( τ ) = L( y, a ) m = (5) Thus the expected rsk of the decson rule (or hypothess) s just the probablty that t predcts ncorrectly. Our goal s then to make the rsk functon approach mnmum, by determnng an optmum value of the threshold, τ (see Secton 2.3). 2.2 Behavour of the RBF classfer A radal bass functon (RBF) network wth k hdden neurons has the form (Bshop, 995): T y( x) = w φ ( x) + b (6) where φ are bass functons: these can take several forms (see below). The weght coeffcents w combne the bass functons nto an output value. b s a bas term 2. The mostly commonly used RBF s the Gaussan bass functon: k y( x) = w φ ( x) + b = k x µ = w exp 2 σ = 2 b + (7) 2 The sgnfcance of ths bas s that the desred output values of the classfer have non-zero mean. 4
5 Thus n (7) φ s th Gaussan bass functon wth centre µ and varance σ 2. Snce the receptve felds of radal bass functons n the hdden layer have a localsed response behavour n the nput space, ts response approaches zero at large rad,.e., φ x µ 0. Consderng the most common form of codng scheme for classfcaton usng neural networks (Tarassenko and Roberts, 994), the output value y s f the nput pattern x belongs to the class and 0 otherwse, that s, y [ 0, ] on the tranng data 3. As an RBF classfer obtans ts parameters through tranng, ths gves the followng statement about the nterval of bas n the output layer: T For classfcaton usng RBFN, y = w φ( x) + b, f the output s set as y [ 0, ] on tranng data, then a well traned RBF classfer satsfy: b [ 0, ]. The valdty of the above statement can be proved as follows: Snce y [ 0,], and φ (0,], w are fnte numbers, f an nput vector s far from the centers of the radal bass functons, we have, =, L, k, x µ φ 0, also then from (7) we have lm φ = 0 y b T assume b [ 0, ], say b >, then y = w φ( x) + b b >, whch contradcts the condton of Thus b [0,] y [ 0, ]. Thus b. Smlarly we can prove that b 0. To understand the physcal meanng of outcomes and the bas n the output layer, consder a geometrc nterpretaton. Suppose there s a two-class problem wth nputs dstrbutng n two dmensons as n Fgure (a). An RBF 3 The condton y [0,] apples for the tranng data only, usng the codng scheme dscussed above. Even when an optmally traned network s used for classfcaton, the network output may stll fall outsde the nterval [0,]. In ths case, the classfer stll can perform well, when an approprate threshold s appled to the network output. 5
6 classfer wth two nput and two output neurons s used for ths classfcaton problem. After tranng, the values of output neurons wth respect to the nput varable x are as n Fgure (b). The correspondng output neuron has a hgh value f the nput s wthn the regon of the class, whle the other output neuron has a low value. Outcomes of all output neurons wll be asymptotc to ther bas f the nput s out of ther class regon. It s notable that the bas value of a traned RBF classfer may le out the nterval [0, ] n some mplementatons. However, such classfers wll perform very poorly, as demonstrated n Sectons 3 and 4. In these crcumstances the RBF classfer wll generally need to be re-traned by adjustng the tranng parameters (that s, the number and the spread constant of the RBFs). 2.3 Relable threshold selecton for RBF classfers As dscussed n Secton 2., for the problem of classfcaton, our goal s to determne a threshold that tends to mnmse the probablty of erroneous classfcaton n a gven class. To derve the optmal threshold for a radal bass functon classfer, the general model (Smolensky et al., 996) s used. The goal of desgnng a network s to fnd that network whch provdes the most lkely explanaton of the observed data set. To do ths t s necessary to try to maxmse the probablty: P ( N D) ( D N ) P( N ) P = (8) P( D) where N represents the network (wth all of the weghts and bases specfed), D represents the observed data set, and P( D N ) s the probablty that the network N would have produced the observed data D. Applyng the monotonc logarthm transformaton (Bshop, 995) to (8), we have: ( N D) ( D N) ( N) ( D) ln P = ln P + ln P ln P (9) Thus, maxmsng (9) s equvalent to maxmsng (8). Snce the probablty dstrbuton of the data s not dependent on the network, ln P( D) wll have no contrbuton to the maxmsng soluton of ln P( N D ) and s dropped from (9). The second term of (9) s a representaton of the probablty of the network tself; that s, t s the a pror probablty or a pror constrant on the network: snce our method assumes that the classfer has been well traned - and our purpose s to determne an optmal threshold for the traned RBF classfer - we wll also drop ths term. The frst term represents the probablty of the data gven the network: that s, t s a measure of how well the 6
7 classfer accounts for the data. Therefore the threshold selecton problem s equvalent to a requrement to maxmse ln P( D N ). Further, f the data are broken nto two parts: the output y and the nput x, then: ( D N ) = P( ( x y) N) = ln P( y x N ) + ln P( x) ln P ln, (0) Fnally, suppose that the nput x does not depend on the network, the last term of (0) has no effect n maxmsng ln P( D N ). Therefore, we need only maxmse the frst term ln P( y x N ). For the classfcaton problem, the output vectors, a, defned n (2), consst of a sequence of 0 s and s. In ths case, we magne that each element of classfer output, a, represents the probablty that the correspondng element of the desred output y takes on the value. Then the probablty of the data gven the network, for c classes problem, s represented by the bnomal dstrbuton (Flemng and Nells, 994): ( ) y P y x N = a ( a ) c = y () Thus we need to maxmse the log of () gven by: c F = y a + y a = ( ln ( ) ln( )) (2) By dfferentatng (2) wth respect to decson acton a, we then obtan: F a y = a a ( a ) (3) To obtan the statonary ponts of F, we set (3) to zero. We then have: y a = 0 (4) Snce to the left of the above statonary pont, F s postve and to the rght a F s negatve, the statonary a pont from (4) s the maxmum pont of F. Snce for an RBF network: T y$ ( x) = w φ ( x) + b (5) As dscussed n Secton 2.2, b [0,], and the classfer output as n (5) satsfes yˆ b for an nput pattern located far from the tranng patterns. To satsfy (4), we wsh a to concde wth the actual y when τ 7
8 (assumng, here, that each output node of the RBF network s allowed to have a dfferent threshold) s appled to ŷ, as n (2). In other word, a should be f the nput pattern s wthn the tranng class, and should be 0 f t s out of the tranng class. Therefore, we must select: τ = b (6) The threshold wth the value gven by (6) decdes whether samples are classfed nto that class. If there s nose or dsturbance n the data, we expect that the turnng pont wll be less senstve to the data set, and we therefore ncrease (6) by a small amount, ε: τ = b + ε (7) The ntroducton of ε s to make the classfer robust to nose and dsturbance whle havng lttle ncrease on the msclassfcaton rate. Fnally, f we use a sngle-threshold RBF classfer, then we obtan: τ = max( b ) + ε (8) Here, ε s a very small postve constant to make τ slghtly greater than max(b): of course, τ must not exceed. 3. Expermental Tests In ths secton, the threshold determnaton technque derved n Secton 2 s assessed n two expermental studes. The chosen data sets were obtaned, frstly, from a mathematcal model smulatng statc fault dagnoss and, secondly, from a non-lnear model of a desel engne coolng system. In both cases, the nput varables were normalsed to [0, ], and the classfers were traned usng the orthogonal least square algorthm (Chen et al., 99) n the Matlab Neural Network Toolbox. Please note that, n addton to determnng the threshold value, mplementng an effectve RBF classfer for a gven task nvolves determnng two further mportant parameters: a) the maxmum number (me) of radal bass functons to use n the hdden layer; b) the spread constant (sc) of the radal bass functon. For each of the followng two experments, the classfer was traned usng a range of possble values for me and sc. The traned classfer was then tested both on the (seen) tranng set and on (unseen) test sets. The classfcaton error rate on each data set s estmated usng (5). 8
9 The key purpose of each study was to explore the mpact of the threshold values. To ths end a tradtonal threshold value (τ 0 ) of 0.5 was used. Ths was compared wth a threshold value (τ ) determned accordng to (8). More explctly, ε was selected as: ε = max( b ), to provde a threshold slghtly greater than max(b). 3. Mathematcal Model Data Set The followng model represents a large class of statc dagnoss problems and s adapted from that ntroduced n (Leonard and Kramer, 990), and descrbed as follows: Y = Y0 + α p + v (9) Here Y represents the measurement vector of the plant, Y 0 represents the plant nomnal steady state. Measurement Y s a functon of plant physcal parameters p, and suffers from measurement nose v. α s a dstrbuton matrx of parameter effects on the measurement vector. Plant faults are caused by devaton of parameters. Here we assume that Y has two measurements y and y 2, p has two parameters p and p 2, and: α = Ths means that p fault causes y and y 2 to devate n the same drecton, and p 2 fault causes y and y 2 to devate n opposte drectons. The classes were defned as: Normal (C 0 ): p < 0. 05, p < Fault (C ): p > 0. 05, p < Fault 2 (C 2 ): p > 0. 05, p < v, v2 ~ N( 0, 0. 05) One set of tranng data was generated wth values of p and p 2 sampled from the normal dstrbuton N(0, 0.25). In total, 600 nput/output pars were generated from Equaton (9) and were used for tranng all networks. Two addtonal sets of test data each wth 300 nput/output pars were also generated, desgnated Test Set and Test Set 2. These data sets were ntended to explore how ths approach performs, both n terms of nterpolaton (Test Set ) and extrapolaton (Test Set 2). Test Set had the same dstrbuton as the tranng set. Test Set 2 had values dstrbuted over the whole parameter space. For ths set, samples wthn the regon of tranng data were assgned to one of the known classes, all other samples were assumed to belong to unknown faults. Usng the tranng data set, RBF classfers were traned by changng the number and the spread constant of 9
10 radal bass functons n the hdden layer. Table lsts the msclassfcaton rate of traned RBF classfers. In the table, tranng data, Test Set and Test Set 2 are represented by D 0, D and D 2, respectvely. Row 2 s the maxmum number of hdden neurons (me) and the spread constant (sc). Elements n row b are bases of correspondng output neurons. The followng rows are the msclassfcaton rate for the data sets (D 0, D and D 2 ) usng τ 0 =0.5 and τ =max(b)+ε, respectvely. As noted above, ε was set to ε = max( b ). From the table, t s clear that when usng ether τ 0 or τ, the classfers provde a very smlar msclassfcaton rate on the tranng set and test set. However, for samples out of the tranng regon the classfer usng τ produces a lower msclassfcaton rate. These results may be readly understood. They arse because the approprately traned classfer s ntended to produce a hgh output value (close to ) for samples n the class whle a low output value (close to 0) for samples n other classes. Thus there s a large nterval for threshold selecton. Theoretcally one could use any threshold between 0 and for a perfectly-traned classfer whch s requred only to classfy samples wthn the tranng range. To further explore the effects of the value of the threshold on the msclassfcaton rate, dfferent thresholds were used for the classfer 5 n Table. Fgure 2 shows the msclassfcaton rate versus the threshold. These emprcal results confrm the fndngs n (8). Specfcally, as s apparent n Fgure 2, the classfer wth a threshold slghtly larger than max(b), say τ, produces a mnmum msclassfcaton rate for Test 2 whle a near mnmum msclassfcaton rate for the tranng set and Test Set, and τ s the turnng pont of msclassfcaton rate on Test Set 2. It s frequently a requrement (n CMFD applcatons) that the classfer wll provde us not only wth a hgh performance for known condtons but also good performance n the presence of unknown faults. For ths purpose, these emprcal results support the use of a classfer wth threshold τ. 3.2 Desel engne coolng system dagnoss data set The second data set used n ths study was generated from a non-lnear model of the coolng system for a desel 0
11 engne, developed elsewhere n our research group (Twddle, 999). The model can smulate varous faults, ncludng those consdered n ths study: fan fault (that s, the radator fan s permanently off), thermostat fault (that s, the thermostat s stuck open) and pump fault (that s, the coolant pump has 50 percent of the capacty of the normal condton). To detect these faults, ths experment used sx measurements: the ambent temperature; the engne block temperature, the coolant temperature at engne block nlet; the coolant temperature at the engne block outlet; the coolant temperature at the radator outlet; and the engne speed. Usng the model, a tranng data set (D 0 ) wth three states (normal, fan fault, and thermostat fault) was generated. Two dfferent test data sets were also generated. One test set (Test Set, D ) has the same three classes as the tranng data. Another test data set (Test Set 2, D 2 ) has the same three classes plus pump fault, here pump fault was consdered as an unknown fault. Each data set has 300 samples. In each case, the data sets conssted of equal numbers of samples for each class. Table 2 lsts the msclassfcaton rate of classfers wth dfferent tranng parameters. To explore, agan, the effects of the value of the threshold on the msclassfcaton rate, dfferent thresholds were used for the classfers 3 and 5 n Table 2. Fgure 3 shows the msclassfcaton rate versus the threshold. Table 3 lsts the msclassfcaton rates versus thresholds around max(b). Here ε n Equaton (8) was set to be proportonal to max(b),.e., ε 0.}. = λ max( b ), where λ s a small coeffcent n {-0., -0.05, -0.0, 0.0, 0.0, 0.05, The results show that the classfers wth τ 0 produce a lower msclassfcaton rate for tranng data and Test Set than that wth τ, whle classfers wth τ can gve a better msclassfcaton rate for Test Set 2. As demonstrated n Experment, the classfer wth τ gves a near mnmum msclassfcaton rate for samples wthn the range of tranng data, so τ can stll be used n such applcatons. However, f we are concerned wth the classfer performance n applcatons wth possble unknown classes, τ s more sutable, at the cost of the rsk of some msclassfcaton of known classes.
12 4. Dscusson and Conclusons In ths paper, a technque for determnng a relable threshold for RBF classfers has been derved, and assessed n two emprcal studes. The approach s easy to use and s seen to be partcularly effectve n classfcaton problems where there may be possble new classes or unknown faults. Overall, these results suggest a two-phase approach to RBF classfer use n stuatons where unknown faults may occur. In the frst phase, we deal wth the possblty of unknown faults. Snce the threshold, τ (determned by the method descrbed n ths paper), results n a measurable mprovement of classfer performance for cases wth unknown faults, τ s used n these crcumstances. However, as the ablty to detect unknown faults may be at the expense of a decrease n the classfcaton performance for known classes, we also perform a second phase. In ths second phase, the classfer threshold wll modfed n order to translate the unknown faults nto known faults. In other words, when we have collected suffcent data about unknown faults, the classfer wll be re-traned usng all avalable data. A tradtonal threshold value of τ 0 may then be appled to the re-traned classfer, snce unknown faults are less lkely to occur. It should be noted that the results n ths paper have also demonstrated that the bas, b, at the output layer of a well traned RBF classfer should satsfy the condton b [ 0, ]. Ths result may be used to check whether a partcular classfer has traned successfully, and the classfer should be dscarded f max(b)> or mn(b)<0. Thus, on ths bass, t would be sensble to abandon classfers numbered 8 and 9 n Table. Fnally, t should also be noted that care must be exercsed when acceptng the tranng f max(b) s close to, because the classfer wll leave lttle margn for response to prevously unseen samples. An deal RBF classfer wll therefore have a low msclassfcaton rate, and a small value for max(b). 2
13 References Bshop, C.M., 994. Novelty detecton and neural-network valdaton, IEE Proceedngs-Vson Image and Sgnal Processng 4, Bshop, C.M., 995. Neural Networks for Pattern Recognton. Clarendon Press, Oxford. Broomhead, D.S., D. Lowe Multvarable functon nterpolaton and adaptve networks. Complex Systems 2, Chen, S., F.N. Cowan, P.M. Grant. 99. Orthogonal least squares learnng algorthm for radal bass functon networks. IEEE Transactons on Neural Networks 2, Cheon, S.W., S.H. Chang, H.Y. Chung, Z.N. Ben Applcaton of neural networks to multple alarm processng and dagnoss n nuclear-power-plants. IEEE Transactons on Nuclear Scence 40, -20 Cordella, L.P., C. Destefano, F. Tortorella, M. Vento A method for mprovng classfcaton relablty of multlayer perceptrons. IEEE Transactons on Neural Networks 6, Flemng, M.C., J.G. Nells Prncples of Appled Statstcs. Routledge. Josh, A., N. Ramakrshman, E.N. Housts, J.R. Rce On neurobologcal, neuro-fuzzy, machne learnng, and statstcal pattern recognton technques. IEEE Transactons on Neural Networks 8, 8-3. Leonard, J.A., M.A. Kramer Classfyng process behavour wth neural networks: strateges for mproved tranng and generalzaton. Proceedngs of the Amercan Control Conference, San Dego, CA, USA, Scholkopf, B., K.K. Sung, C.J.C. Burges, F. Gros, P. Nyog, T. Poggo, V. Vapnk Comparng support vector machnes wth Gaussan kernels to radal bass functon classfers. IEEE Transactons on Sgnal Processng 45, Smolensky, P., M. C. Mozer, D. E. Rumelhart Mathematcal Perspectves on Neural Networks. Lawrence Erlbaum Assocates, Publshers, Mahwal, New Jersey. Tarassenko, L., S. Roberts 994. Supervsed and unsupervsed learnng n radal bass functon classfers. IEE Proceedngs-Vson Image and Sgnal Processng 4, Twddle, J.A., 999. Fuzzy model based fault dagnoss of a desel engne coolng system. Department of Engneerng, Unversty of Lecester, Report 99-. Watanabe, K., S. Hrota, L. Hou, D.M. Hmmelblau Dagnoss of multple smultaneous fault va herarchcal artfcal neural networks. AICHE Journal 40,
14 x 2 Class 2 (a) Class x Output Output 2 b (b) b 2 x 4
15 0.8 Tranng Test Test 2 Error rate τ Threshold 5
16 Error rate 0.5 τ Tranng Test Test 2 (a) Error rate 0.5 Tranng Test Test 2 τ (b) Threshold 6
17 Fgure Outcomes of output neurons Fgure 2 Msclassfcaton rate vs. Threshold (classfer 5) Fgure 3 Msclassfcaton rate vs. threshold (a) for classfer 3 and (b) for classfer 5 7
18 Table Classfcaton error rate (%) for the mathematcal model data Classfer Number me/sc 50/ / /0. 00/ / / / /0. 00/0.2 b D 0 wth D 0 wth τ D wth D wth τ D 2 wth D 2 wth τ
19 Table 2 Classfcaton error rate (%) for coolng system dagnoss Classfer Number me/sc 00/0.5 00/0.2 00/ / / /0.3 b D 0 wth D 0 wth τ D wth D wth τ D 2 wth D 2 wth τ
20 Table 3 Classfcaton error rate (%) usng threshold around max(b) τ = max( b) + ε = max( b) + λ max( b) Classfer 3 Classfer 5 λ D 0 D D 2 D 0 D D
Support Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationMultigradient for Neural Networks for Equalizers 1
Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More informationANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)
Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More informationEEE 241: Linear Systems
EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they
More informationSolving Nonlinear Differential Equations by a Neural Network Method
Solvng Nonlnear Dfferental Equatons by a Neural Network Method Luce P. Aarts and Peter Van der Veer Delft Unversty of Technology, Faculty of Cvlengneerng and Geoscences, Secton of Cvlengneerng Informatcs,
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationReport on Image warping
Report on Image warpng Xuan Ne, Dec. 20, 2004 Ths document summarzed the algorthms of our mage warpng soluton for further study, and there s a detaled descrpton about the mplementaton of these algorthms.
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationPsychology 282 Lecture #24 Outline Regression Diagnostics: Outliers
Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.
More informationNon-linear Canonical Correlation Analysis Using a RBF Network
ESANN' proceedngs - European Smposum on Artfcal Neural Networks Bruges (Belgum), 4-6 Aprl, d-sde publ., ISBN -97--, pp. 57-5 Non-lnear Canoncal Correlaton Analss Usng a RBF Network Sukhbnder Kumar, Elane
More informationComparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method
Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method
More informationCSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography
CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve
More informationAppendix B: Resampling Algorithms
407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles
More informationRegularized Discriminant Analysis for Face Recognition
1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths
More informationBoostrapaggregating (Bagging)
Boostrapaggregatng (Baggng) An ensemble meta-algorthm desgned to mprove the stablty and accuracy of machne learnng algorthms Can be used n both regresson and classfcaton Reduces varance and helps to avod
More informationPop-Click Noise Detection Using Inter-Frame Correlation for Improved Portable Auditory Sensing
Advanced Scence and Technology Letters, pp.164-168 http://dx.do.org/10.14257/astl.2013 Pop-Clc Nose Detecton Usng Inter-Frame Correlaton for Improved Portable Audtory Sensng Dong Yun Lee, Kwang Myung Jeon,
More information3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X
Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number
More informationKernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan
Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationVapnik-Chervonenkis theory
Vapnk-Chervonenks theory Rs Kondor June 13, 2008 For the purposes of ths lecture, we restrct ourselves to the bnary supervsed batch learnng settng. We assume that we have an nput space X, and an unknown
More informationPredictive Analytics : QM901.1x Prof U Dinesh Kumar, IIMB. All Rights Reserved, Indian Institute of Management Bangalore
Sesson Outlne Introducton to classfcaton problems and dscrete choce models. Introducton to Logstcs Regresson. Logstc functon and Logt functon. Maxmum Lkelhood Estmator (MLE) for estmaton of LR parameters.
More informationSimulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests
Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth
More informationNumerical Heat and Mass Transfer
Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and
More informationOnline Classification: Perceptron and Winnow
E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationSupporting Information
Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to
More informationStatistics II Final Exam 26/6/18
Statstcs II Fnal Exam 26/6/18 Academc Year 2017/18 Solutons Exam duraton: 2 h 30 mn 1. (3 ponts) A town hall s conductng a study to determne the amount of leftover food produced by the restaurants n the
More informationP R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /
Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons
More informationOn the Multicriteria Integer Network Flow Problem
BULGARIAN ACADEMY OF SCIENCES CYBERNETICS AND INFORMATION TECHNOLOGIES Volume 5, No 2 Sofa 2005 On the Multcrtera Integer Network Flow Problem Vassl Vasslev, Marana Nkolova, Maryana Vassleva Insttute of
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More information4 Analysis of Variance (ANOVA) 5 ANOVA. 5.1 Introduction. 5.2 Fixed Effects ANOVA
4 Analyss of Varance (ANOVA) 5 ANOVA 51 Introducton ANOVA ANOVA s a way to estmate and test the means of multple populatons We wll start wth one-way ANOVA If the populatons ncluded n the study are selected
More informationNatural Language Processing and Information Retrieval
Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support
More informationA Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach
A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland
More informationChapter 13: Multiple Regression
Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to
More informationStatistical Foundations of Pattern Recognition
Statstcal Foundatons of Pattern Recognton Learnng Objectves Bayes Theorem Decson-mang Confdence factors Dscrmnants The connecton to neural nets Statstcal Foundatons of Pattern Recognton NDE measurement
More informationThe Order Relation and Trace Inequalities for. Hermitian Operators
Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence
More informationINF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018
INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton
More informationSee Book Chapter 11 2 nd Edition (Chapter 10 1 st Edition)
Count Data Models See Book Chapter 11 2 nd Edton (Chapter 10 1 st Edton) Count data consst of non-negatve nteger values Examples: number of drver route changes per week, the number of trp departure changes
More informationCOMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
More informationUNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours
UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x
More informationBayesian predictive Configural Frequency Analysis
Psychologcal Test and Assessment Modelng, Volume 54, 2012 (3), 285-292 Bayesan predctve Confgural Frequency Analyss Eduardo Gutérrez-Peña 1 Abstract Confgural Frequency Analyss s a method for cell-wse
More informationMultilayer Perceptron (MLP)
Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne
More informationThe Minimum Universal Cost Flow in an Infeasible Flow Network
Journal of Scences, Islamc Republc of Iran 17(2): 175-180 (2006) Unversty of Tehran, ISSN 1016-1104 http://jscencesutacr The Mnmum Unversal Cost Flow n an Infeasble Flow Network H Saleh Fathabad * M Bagheran
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationCHAPTER-5 INFORMATION MEASURE OF FUZZY MATRIX AND FUZZY BINARY RELATION
CAPTER- INFORMATION MEASURE OF FUZZY MATRI AN FUZZY BINARY RELATION Introducton The basc concept of the fuzz matr theor s ver smple and can be appled to socal and natural stuatons A branch of fuzz matr
More informationA Robust Method for Calculating the Correlation Coefficient
A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal
More informationMore metrics on cartesian products
More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of
More informationDepartment of Computer Science Artificial Intelligence Research Laboratory. Iowa State University MACHINE LEARNING
MACHINE LEANING Vasant Honavar Bonformatcs and Computatonal Bology rogram Center for Computatonal Intellgence, Learnng, & Dscovery Iowa State Unversty honavar@cs.astate.edu www.cs.astate.edu/~honavar/
More informationStatistics for Economics & Business
Statstcs for Economcs & Busness Smple Lnear Regresson Learnng Objectves In ths chapter, you learn: How to use regresson analyss to predct the value of a dependent varable based on an ndependent varable
More informationCHALMERS, GÖTEBORGS UNIVERSITET. SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS. COURSE CODES: FFR 135, FIM 720 GU, PhD
CHALMERS, GÖTEBORGS UNIVERSITET SOLUTIONS to RE-EXAM for ARTIFICIAL NEURAL NETWORKS COURSE CODES: FFR 35, FIM 72 GU, PhD Tme: Place: Teachers: Allowed materal: Not allowed: January 2, 28, at 8 3 2 3 SB
More informationComparison of Regression Lines
STATGRAPHICS Rev. 9/13/2013 Comparson of Regresson Lnes Summary... 1 Data Input... 3 Analyss Summary... 4 Plot of Ftted Model... 6 Condtonal Sums of Squares... 6 Analyss Optons... 7 Forecasts... 8 Confdence
More informationCopyright 2017 by Taylor Enterprises, Inc., All Rights Reserved. Adjusted Control Limits for P Charts. Dr. Wayne A. Taylor
Taylor Enterprses, Inc. Control Lmts for P Charts Copyrght 2017 by Taylor Enterprses, Inc., All Rghts Reserved. Control Lmts for P Charts Dr. Wayne A. Taylor Abstract: P charts are used for count data
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationx = , so that calculated
Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to
More informationLECTURE 9 CANONICAL CORRELATION ANALYSIS
LECURE 9 CANONICAL CORRELAION ANALYSIS Introducton he concept of canoncal correlaton arses when we want to quantfy the assocatons between two sets of varables. For example, suppose that the frst set of
More informationTemperature. Chapter Heat Engine
Chapter 3 Temperature In prevous chapters of these notes we ntroduced the Prncple of Maxmum ntropy as a technque for estmatng probablty dstrbutons consstent wth constrants. In Chapter 9 we dscussed the
More informationHomework Assignment 3 Due in class, Thursday October 15
Homework Assgnment 3 Due n class, Thursday October 15 SDS 383C Statstcal Modelng I 1 Rdge regresson and Lasso 1. Get the Prostrate cancer data from http://statweb.stanford.edu/~tbs/elemstatlearn/ datasets/prostate.data.
More informationBasically, if you have a dummy dependent variable you will be estimating a probability.
ECON 497: Lecture Notes 13 Page 1 of 1 Metropoltan State Unversty ECON 497: Research and Forecastng Lecture Notes 13 Dummy Dependent Varable Technques Studenmund Chapter 13 Bascally, f you have a dummy
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 6. Rdge regresson The OLSE s the best lnear unbased
More information= z 20 z n. (k 20) + 4 z k = 4
Problem Set #7 solutons 7.2.. (a Fnd the coeffcent of z k n (z + z 5 + z 6 + z 7 + 5, k 20. We use the known seres expanson ( n+l ( z l l z n below: (z + z 5 + z 6 + z 7 + 5 (z 5 ( + z + z 2 + z + 5 5
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationEnsemble Methods: Boosting
Ensemble Methods: Boostng Ncholas Ruozz Unversty of Texas at Dallas Based on the sldes of Vbhav Gogate and Rob Schapre Last Tme Varance reducton va baggng Generate new tranng data sets by samplng wth replacement
More informationLOGIT ANALYSIS. A.K. VASISHT Indian Agricultural Statistics Research Institute, Library Avenue, New Delhi
LOGIT ANALYSIS A.K. VASISHT Indan Agrcultural Statstcs Research Insttute, Lbrary Avenue, New Delh-0 02 amtvassht@asr.res.n. Introducton In dummy regresson varable models, t s assumed mplctly that the dependent
More informationLecture 6: Introduction to Linear Regression
Lecture 6: Introducton to Lnear Regresson An Manchakul amancha@jhsph.edu 24 Aprl 27 Lnear regresson: man dea Lnear regresson can be used to study an outcome as a lnear functon of a predctor Example: 6
More informationNegative Binomial Regression
STATGRAPHICS Rev. 9/16/2013 Negatve Bnomal Regresson Summary... 1 Data Input... 3 Statstcal Model... 3 Analyss Summary... 4 Analyss Optons... 7 Plot of Ftted Model... 8 Observed Versus Predcted... 10 Predctons...
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More information2.3 Nilpotent endomorphisms
s a block dagonal matrx, wth A Mat dm U (C) In fact, we can assume that B = B 1 B k, wth B an ordered bass of U, and that A = [f U ] B, where f U : U U s the restrcton of f to U 40 23 Nlpotent endomorphsms
More informationDepartment of Statistics University of Toronto STA305H1S / 1004 HS Design and Analysis of Experiments Term Test - Winter Solution
Department of Statstcs Unversty of Toronto STA35HS / HS Desgn and Analyss of Experments Term Test - Wnter - Soluton February, Last Name: Frst Name: Student Number: Instructons: Tme: hours. Ads: a non-programmable
More informationCONTRAST ENHANCEMENT FOR MIMIMUM MEAN BRIGHTNESS ERROR FROM HISTOGRAM PARTITIONING INTRODUCTION
CONTRAST ENHANCEMENT FOR MIMIMUM MEAN BRIGHTNESS ERROR FROM HISTOGRAM PARTITIONING N. Phanthuna 1,2, F. Cheevasuvt 2 and S. Chtwong 2 1 Department of Electrcal Engneerng, Faculty of Engneerng Rajamangala
More informationBounds on the Generalization Performance of Kernel Machines Ensembles
Bounds on the Generalzaton Performance of Kernel Machnes Ensembles Theodoros Evgenou theos@a.mt.edu Lus Perez-Breva lpbreva@a.mt.edu Massmlano Pontl pontl@a.mt.edu Tomaso Poggo tp@a.mt.edu Center for Bologcal
More informationMaximum Likelihood Estimation (MLE)
Maxmum Lkelhood Estmaton (MLE) Ken Kreutz-Delgado (Nuno Vasconcelos) ECE 175A Wnter 01 UCSD Statstcal Learnng Goal: Gven a relatonshp between a feature vector x and a vector y, and d data samples (x,y
More informationStructure and Drive Paul A. Jensen Copyright July 20, 2003
Structure and Drve Paul A. Jensen Copyrght July 20, 2003 A system s made up of several operatons wth flow passng between them. The structure of the system descrbes the flow paths from nputs to outputs.
More informationChapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems
Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons
More informationChapter 5 Multilevel Models
Chapter 5 Multlevel Models 5.1 Cross-sectonal multlevel models 5.1.1 Two-level models 5.1.2 Multple level models 5.1.3 Multple level modelng n other felds 5.2 Longtudnal multlevel models 5.2.1 Two-level
More informationUncertainty as the Overlap of Alternate Conditional Distributions
Uncertanty as the Overlap of Alternate Condtonal Dstrbutons Olena Babak and Clayton V. Deutsch Centre for Computatonal Geostatstcs Department of Cvl & Envronmental Engneerng Unversty of Alberta An mportant
More informationCHAPTER 14 GENERAL PERTURBATION THEORY
CHAPTER 4 GENERAL PERTURBATION THEORY 4 Introducton A partcle n orbt around a pont mass or a sphercally symmetrc mass dstrbuton s movng n a gravtatonal potental of the form GM / r In ths potental t moves
More informationUncertainty in measurements of power and energy on power networks
Uncertanty n measurements of power and energy on power networks E. Manov, N. Kolev Department of Measurement and Instrumentaton, Techncal Unversty Sofa, bul. Klment Ohrdsk No8, bl., 000 Sofa, Bulgara Tel./fax:
More informationCOS 521: Advanced Algorithms Game Theory and Linear Programming
COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton
More informationprinceton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg
prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there
More informationChapter 9: Statistical Inference and the Relationship between Two Variables
Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,
More informationRelevance Vector Machines Explained
October 19, 2010 Relevance Vector Machnes Explaned Trstan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introducton Ths document has been wrtten n an attempt to make Tppng s [1] Relevance Vector Machnes
More informationLecture 12: Discrete Laplacian
Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly
More informationFeb 14: Spatial analysis of data fields
Feb 4: Spatal analyss of data felds Mappng rregularly sampled data onto a regular grd Many analyss technques for geophyscal data requre the data be located at regular ntervals n space and/or tme. hs s
More informationSpeeding up Computation of Scalar Multiplication in Elliptic Curve Cryptosystem
H.K. Pathak et. al. / (IJCSE) Internatonal Journal on Computer Scence and Engneerng Speedng up Computaton of Scalar Multplcaton n Ellptc Curve Cryptosystem H. K. Pathak Manju Sangh S.o.S n Computer scence
More informationCOS 511: Theoretical Machine Learning. Lecturer: Rob Schapire Lecture #16 Scribe: Yannan Wang April 3, 2014
COS 511: Theoretcal Machne Learnng Lecturer: Rob Schapre Lecture #16 Scrbe: Yannan Wang Aprl 3, 014 1 Introducton The goal of our onlne learnng scenaro from last class s C comparng wth best expert and
More informationDUE: WEDS FEB 21ST 2018
HOMEWORK # 1: FINITE DIFFERENCES IN ONE DIMENSION DUE: WEDS FEB 21ST 2018 1. Theory Beam bendng s a classcal engneerng analyss. The tradtonal soluton technque makes smplfyng assumptons such as a constant
More informationChapter Newton s Method
Chapter 9. Newton s Method After readng ths chapter, you should be able to:. Understand how Newton s method s dfferent from the Golden Secton Search method. Understand how Newton s method works 3. Solve
More informationSTAT 511 FINAL EXAM NAME Spring 2001
STAT 5 FINAL EXAM NAME Sprng Instructons: Ths s a closed book exam. No notes or books are allowed. ou may use a calculator but you are not allowed to store notes or formulas n the calculator. Please wrte
More informationChapter 6 Support vector machine. Séparateurs à vaste marge
Chapter 6 Support vector machne Séparateurs à vaste marge Méthode de classfcaton bnare par apprentssage Introdute par Vladmr Vapnk en 1995 Repose sur l exstence d un classfcateur lnéare Apprentssage supervsé
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We
More informationThe Geometry of Logit and Probit
The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.
More informationLimited Dependent Variables
Lmted Dependent Varables. What f the left-hand sde varable s not a contnuous thng spread from mnus nfnty to plus nfnty? That s, gven a model = f (, β, ε, where a. s bounded below at zero, such as wages
More informationAtmospheric Environmental Quality Assessment RBF Model Based on the MATLAB
Journal of Envronmental Protecton, 01, 3, 689-693 http://dxdoorg/10436/jep0137081 Publshed Onlne July 01 (http://wwwscrporg/journal/jep) 689 Atmospherc Envronmental Qualty Assessment RBF Model Based on
More information