Efficient and Robust Feature Extraction by Maximum Margin Criterion

Size: px
Start display at page:

Download "Efficient and Robust Feature Extraction by Maximum Margin Criterion"

Transcription

1 Effcent and Robust Feature Extracton by Maxmum Margn Crteron Hafeng L Tao Jang Department of Computer Scence Unversty of Calforna Rversde, CA 95 {hl,jang}@cs.ucr.edu Keshu Zhang Department of Electrcal Engneerng Unversty of New Orleans New Orleans, LA 7048 kzhang@uno.edu Abstract A new feature extracton crteron, maxmum margn crteron (MMC), s proposed n ths paper. Ths new crteron s general n the sense that, when combned wth a sutable constrant, t can actually gve rse to the most popular feature extractor n the lterature, lnear dscrmnate analyss (LDA). We derve a new feature extractor based on MMC usng a dfferent constrant that does not depend on the nonsngularty of the wthn-class scatter matrx S w. Such a dependence s a major drawback of LDA especally when the sample sze s small. The kernelzed (nonlnear) counterpart of ths lnear feature extractor s also establshed n ths paper. Our prelmnary expermental results on face mages demonstrate that the new feature extractors are effcent and stable. Introducton In statstcal pattern recognton, the hgh-dmensonalty s a major cause of the practcal lmtatons of many pattern recognton technologes. In the past several decades, many dmensonalty reducton technques have been proposed. Lnear dscrmnant analyss (LDA, also called Fsher s Lnear Dscrmnant) [] s one of the most popular lnear dmensonalty reducton method. In many applcatons, LDA has been proven to be very powerful. LDA s gven by a lnear transformaton matrx W R D d maxmzng the so-called Fsher crteron (a knd of Raylegh coeffcent) J F (W) = WT S b W W T S w W where S b = c = p (m m)(m m) T and S w = c = p S are the betweenclass scatter matrx and the wthn-class scatter matrx, respectvely; c s the number of classes; m and p are the mean vector and a pror probablty of class, respectvely; m = c = p m s the overall mean vector; S s the wthn-class scatter matrx of class ; D and d are the dmensonaltes of the data before and after the transformaton, respectvely. To maxmze (), the transformaton matrx W must be consttuted by the largest egenvectors of S w S b. The purpose of LDA s to maxmze the between-class scatter whle smultaneously mnmzng the wthn-class scatter. The two-class LDA has a close connecton to optmal lnear Bayes classfers. In the two-class case, the transformaton matrx W s just a vector, whch s n the same drecton as the dscrmnant n the correspondng optmal Bayes classfer. However, t has been shown that LDA s suboptmal for mult-class problems []. A major drawback of LDA s that t cannot be appled when S w s sngular because of the small sample sze problem [3]. The small sample sze problem ()

2 arses whenever the number of samples s smaller than the dmensonalty of the samples. For example, a mage n a face recognton system has 4096 dmensons, whch requres more than 4096 tranng ponts to ensure that S w s nonsngular. So, LDA s not a stable method n practce when the tranng data s scarce. In recent years, many researchers have notced ths problem and tred to overcome the computatonal dffculty wth LDA. Tan et al. [4] used the pseudo-nverse matrx S + w nstead of the nverse matrx S w. For the same purpose, Hong and Yang [5] tred to add a sngular value perturbaton to S w to make t nonsngular. Nether of these methods are theoretcally sound because Fsher s crteron s not vald when S w s sngular. When S w s sngular, any postve S b makes Fsher s crteron nfntely large. Thus, these nave attempts to calculate the (pseudo or approxmate) nverse of S w may lead to arbtrary (meanngless) results. Besdes, t s also known that an egenvector could be very senstve to small perturbaton f ts correspondng egenvalue s close to another egenvalue of the same matrx [6]. In 99, Lu et al. [7] modfed Fsher s crteron by usng the total scatter matrx S t = S b + S w as the denomnator nstead of S w. It has been proven that the modfed crteron s exactly equvalent to Fsher s crteron. However, when S w s sngular, the modfed crteron reaches the maxmum value (.e., ) no matter what the transformaton W s. Such an arbtrary transformaton cannot guarantee the maxmum class separablty unless W T S b W s maxmzed. Besdes, ths method need stll calculate an nverse matrx, whch s tme consumng. In 000, Chen et al. [8] proposed the LDA+PCA method. When S w s of full rank, the LDA+PCA method just calculates the maxmum egenvectors of S t S b to form the transformaton matrx. Otherwse, a two-stage procedure s employed. Frst, the data are transformed nto the null space V 0 of S w. Second, t tres to maxmze the betweenclass scatter n V 0, whch s accomplshed by performng prncpal component analyss (PCA) on the between-class scatter matrx n V 0. Although ths method solves the small sample sze problem, t s obvously suboptmal because t maxmzes the between-class scatter n the null space of S w nstead of the orgnal nput space. Besdes, the performance of the LDA+PCA method drops sgnfcantly when n c s close to the dmensonalty D, where n s the number of samples and c s the number of classes. The reason s that the dmensonalty of the null space V 0 s too small n ths stuaton, and too much nformaton s lost when we try to extract the dscrmnant vectors n V 0. LDA+PCA also need calculate the rank of S w, whch s an ll-defned operaton due to floatng-pont mprecsons. At last, ths method s complcated and slow because too much calculaton s nvolved. Kernel Fsher s Dscrmnant (KFD) [9] s a well-known nonlnear extenson to LDA. The nstablty problem s more severe for KFD because S w n the (nonlnear) feature space F s always sngular (the rank of S w s n c). Smlar to [5], KFD smply adds a perturbaton µi to S w. Of course, t has the same stablty problem as that n [5] because egenvectors are senstve to small perturbaton. Although the authors also argued that ths perturbaton acts as some knd of regularzaton,.e., a capacty control n F, the real nfluence n ths settng of regularzaton s not yet fully understood. Besdes, t s hard to determne an optmal µ snce there are no theoretcal gudelnes. In ths paper, a smpler, more effcent, and stable method s proposed to calculate the most dscrmnant vectors based on a new feature extracton crteron, the maxmum margn crteron (MMC). Based on MMC, new lnear and nonlnear feature extractors are establshed. It can be shown that MMC represents class separablty better than PCA. As a connecton to Fsher s crteron, we may also derve LDA from MMC by ncorporatng some sutable constrant. On the other hand, the new feature extractors derved above (based on MMC) do not suffer from the small sample sze problem, whch s known to cause serous stablty problems for LDA (based on Fsher s crteron). Dfferent from LDA+PCA, the new feature extractors based on MMC maxmze the between-class scatter n the nput space nstead of the null space of S w. Hence, t has a better overall performance than LDA+PCA, as confrmed by our prelmnary expermental results.

3 Maxmum Margn Crteron Suppose that we are gven emprcal data (x, y ),..., (x n, y n ) X {C,..., C c } Here, the doman X R D s some nonempty set that the patterns x are taken from. The y s are called labels or targets. By studyng these samples, we want to predct the label y {C,..., C c } of some new pattern x X. In other words, we choose y such that (x, y) s n some sense smlar to the tranng examples. For ths purpose, some measure need be employed to assess smlarty or dssmlarty. We want to keep such smlarty/dssmlarty nformaton as much as possble after the dmensonalty reducton,.e., transformng x from R D to R d, where d D. If some dstance metrc s used to measure the dssmlarty, we would hope that a pattern s close to those n the same class but far from those n dfferent classes. So, a good feature extractor should maxmze the dstances between classes after the transformaton. Therefore, we may defne the feature extracton crteron as J = p p j d(c, C j ) () We call () the maxmum margn crteron (MMC). It s actually the summaton of c(c ) nterclass margns. Lke the weghted parwse Fsher s crtera n [], one may also defne a weghted maxmum margn crteron. Due to the page lmt, we omt the dscusson n ths paper. One may use the dstance between mean vectors as the dstance between classes,.e. d(c, C j ) = d(m, m j ) (3) where m and m j are the mean vectors of the class C and the class C j, respectvely. However, (3) s not sutable snce t neglects the scatter of classes. Even f the dstance between the mean vectors s large, t s not easy to separate two classes that have the large spread and overlap wth each other. By consderng the scatter of classes, we defne the nterclass dstance (or margn) as d(c, C j ) = d(m, m j ) s(c ) s(c j ) (4) where s(c ) s some measure of the scatter of the class C. In statstcs, we usually use the generalzed varance S or overall varance tr(s ) to measure the scatter of data. In ths paper, we use the overall varance tr(s ) because t s easy to analyze. The weakness of the overall varance s that t gnores covarance structure altogether. Note that, by employng the overall/generalzed varance, the expresson (4) measures the average margn between two classes whle the mnmum margn s used n support vector machnes (SVMs) [0]. Wth (4) and s(c ) beng tr(s ), we may decompose () nto two parts J = p p j (d(m, m j ) tr(s ) tr(s j )) = p p j d(m, m j ) p p j (tr(s ) + tr(s j )) The second part s easly smplfed to tr(s w ) ( ) p p j (tr(s ) + tr(s j )) = p tr(s ) = tr p S = tr(s w ) (5) = =

4 By employng the Eucldean dstance, we may also smplfy the frst part to tr(s b ) as follows p p j d(m, m j ) = p p j (m m j ) T (m m j ) = p p j (m m + m m j ) T (m m + m m j ) After expandng t, we can smplfy the above equaton to c = p (m m) T (m m) by usng the fact c j= p j(m m j ) = 0. So ( ) p p j d(m, m j ) = tr p (m m)(m m) T = tr(s b ) (6) = Now we obtan J = tr(s b S w ) (7) Snce tr(s b ) measures the overall varance of the class mean vectors, a large tr(s b ) mples that the class mean vectors scatter n a large space. On the other hand, a small tr(s w ) mples that every class has a small spread. Thus, a large J ndcates that patterns are close to each other f they are from the same class but are far from each other f they are from dfferent classes. Thus, ths crteron may represent class separablty better than PCA. Recall that PCA tres to maxmze the total scatter after a lnear transformaton. But the data set wth a large wthn-class scatter can also have a large total scatter even when t has a small between-class scatter because S t = S b + S w. Obvously, such data are not easy to classfy. Compared wth LDA+PCA, we maxmze the between-class scatter n nput space rather than the null space of S w when S w s sngular. So, our method can keep more dscrmnatve nformaton than LDA+PCA does. 3 Lnear Feature Extracton When performng dmensonalty reducton, we want to fnd a (lnear or nonlnear) mappng from the measurement space M to some feature space F such that J s maxmzed after the transformaton. In ths secton, we dscuss how to fnd an optmal lnear feature extractor. In the next secton, we wll generalze t to the nonlnear case. Consder a lnear mappng W R D d. We would lke to maxmze J(W) = tr(s W b S W w ) where S W b and S W w are the between-class scatter matrx and wthn-class scatter matrx n the feature space F. Snce W s a lnear mappng, t s easy to show S W b = W T S b W and = W T S w W. So, we have S W w J(W) = tr ( W T (S b S w )W ) (8) In ths formulaton, we have the freedom to multply W wth some nonzero constant. Thus, we addtonally requre that W s consttuted by the unt vectors,.e. W = [w w... w d ] and wk T w k =. Ths means that we need solve the followng constraned optmzaton max wk T (S b S w )w k subject to w T k w k = 0 k =,..., d

5 Note that, we may also use other constrants n the above. For example, we may requre tr ( W T S w W ) = and then maxmze tr ( W T S b W ). It s easy to show that maxmzng MMC wth such a constrant n fact results n LDA. The only dfference s that t nvolves a constraned optmzaton whereas the tradtonal LDA solves an unconstraned optmzaton. The motvaton for usng the constrant w T k w k = s that t allows us to avod calculatng the nverse of S w and thus the potental small sample sze problem. To solve the above optmzaton problem, we may ntroduce a Lagrangan L(w k, λ k ) = wk T (S b S w )w k λ k (wk T w k ) (9) wth multplers λ k. The Lagrangan L has to be maxmzed wth respect to λ k and w k. The condton that at the statonary pont, the dervatves of L wth respect to w k must vansh L(w k, λ k ) = ((S w b S w ) λ k I)w k = 0 k k =,..., d (0) leads to (S b S w )w k =λ k w k k =,..., d () whch means that the λ k s are the egenvalues of S b S w and the w k s are the correspondng egenvectors. Thus J(W) = wk T (S b S w )w k = λ k wk T w k = λ k () Therefore, J(W) s maxmzed when W s composed of the frst d largest egenvectors of S b S w. Here, we need not calculate the nverse of S w, whch allows us to avod the small sample sze problem easly. We may also requre W to be orthonormal, whch may help preserve the shape of the dstrbuton. 4 Nonlnear Feature Extracton wth Kernel In ths secton, we follow the approach of nonlnear SVMs [0] to kernelze the above lnear feature extractor. More precsely, we frst reformulate the maxmum margn crteron n terms of only dot-product Φ(x), Φ(y) of nput patterns. Then we replace the dot-product by some postve defnte kernel k(x, y), e.g. Gaussan kernel e γ x y. Consder the maxmum margn crteron n the feature space F J Φ (W) = wk T (S Φ b SΦ w)w k where S Φ b and S Φ w are the between-class scatter matrx and wthn-class scatter matrx n F,.e., S Φ b = c = p (m Φ m Φ )(m Φ m Φ ) T, S Φ w = c = p S Φ and S Φ = n n j= (Φ(x() j ) mφ )(Φ(x() j ) mφ )T wth m Φ = n n j= Φ(x() j ), mφ = c = p m Φ, and x() j s the pattern of class C that has n samples. For us, an mportant fact s that each w k les n the span of Φ(x ), Φ(x ),..., Φ(x n ). Therefore, we can fnd an expanson for w k n the form w k = n l= α(k) l Φ(x l ). Usng ths expanson and the defnton of m Φ, we have n wk T m Φ = α (k) n l Φ(x n l ), Φ(x () j ) l= j=

6 Replacng the dot-product by some kernel functon k(x, y) and defnng ( m ) l = n n j= k(x l, x () j ), we get wt k mφ = α T k m wth (α k ) l = α (k) l. Smlarly, we have w T k m Φ = w T k p m Φ = α T k = p m = α T k m wth m = c = p m. Ths means w T k (mφ mφ ) = α T k ( m m). and wk T S Φ b w k = = = = = p (wk T (m Φ m Φ ))(wk T (m Φ m Φ )) T p T α T k ( m m)( m m) T α k = where S b = c = p ( m m)( m m) T. α T S k b α k Smlarly, one can smplfy W T S Φ ww. Frst, we have wk T (Φ(x() j ) mφ ) = αt k (k() j m ) wth (k () j ) l = k(x l, x () j ). Consderng wt k SΦ w k = n n j= (wt k (Φ(x() j ) m Φ ))(wt k (Φ(x() j ) mφ ))T, we have w T k S Φ w k = n n j= α T k (k () j m )(k () j m ) T α k = n α T n S k (e j n )(e j n ) T ST n n α k j= = n α T n S k (e j e T j e j T n n n e T j + j= n n n T n ) S T α k = n α T k S (I n n n n T n ) S T α k where ( S ) lj = k(x l, x () j ), I n n s the n n dentty matrx, n s the n -dmensonal vector of s, and e j s the canoncal bass vector of n dmensons. Thus, we obtan wk T S Φ ww k = = α T k ( = = p n α T k S (I n n n T n ) S T α k p n S (I n n n T n ) S T ) α k = α T S k w α k where S w = c = p n S (I n n n T n ) S T. So the maxmum crteron n the feature space F s J(W) = α T k ( S b S w )α k (3) Smlar to the observatons n Secton 3, the above crteron s maxmzed by the largest egenvectors of S b S w.

7 error rate RAW LDA+PCA MMC KMMC class no. class no. (a) The comparson n term of error rate. (b) The comparson n term of tranng tme. Fgure : Expermental results obtaned usng a lnear SVM on the orgnal data (RAW), and the data extracted by LDA+PCA, the lnear feature extractor based on MMC (MMC) and the nonlnear feature extractor based on MMC (KMMC), whch employs the Gaussan kernel wth γ = tranng tme (second) LDA+PCA MMC KMMC Experments To evaluate the performance of our new methods (both lnear and nonlnear feature extractors), we ran both LDA+PCA and our methods on the ORL face dataset []. The ORL dataset conssts of 0 face mages from 40 subjects for a total of 400 mages, wth some varaton n pose, facal expresson and detals. The resoluton of the mages s 9, wth 56 gray-levels. Frst, we reszed the mages to 8 3 to save the expermental tme. Then, we reduced the dmensonalty of each mage set to c, where c s the number of classes. At last we traned and tested a lnear SVM on the dmensonalty-reduced data. As a control, we also traned and tested a lnear SVM on the orgnal data before ts dmensonalty was reduced. In order to demonstrate the effectveness and the effcency of our methods, we conducted a seres of experments and compared our results wth those obtaned usng LDA+PCA. The error rates are shown n Fg.(a). When traned wth 3 samples and tested wth 7 other samples for each class, our method s generally better than LDA+PCA. In fact, our method s usually better than LDA+PCA on other numbers of tranng samples. To save space, we do not show all the results here. Note that our methods can even acheve lower error rates than a lnear SVM on the orgnal data (wthout dmensonalty reducton). However, LDA+PCA does not demonstrate such a clear superorty over RAW. Fg. (a) also shows that the kernelzed (nonlnear) feature extractor based on MMC s sgnfcantly better than the lnear one, n partcular when the number of classes c s large. Besdes accuracy, our methods are also much more effcent than LDA+PCA n the sense of the tranng tme requred. Fg. (b) shows that our lnear feature extractor s about 4 tmes faster than LDA+PCA. The same speedup was observed on other numbers of tranng samples. Note that our nonlnear feature extractor s also faster than LDA+PCA n ths case although t s very tme-consumng to calculate the kernel matrx n general. An explanaton of the speedup s that the kernel matrx sze equals the number of samples, whch s pretty small n ths case. Furthermore, our method performs much better than LDA+PCA when n c s close to the dmensonalty D. Because the amount of tranng data was lmted, we reszed the mages to 68 dmensons to create such a stuaton. The expermental results are shown n Fg.. In ths stuaton, the performance of LDA+PCA drops sgnfcantly because the null space of S w has a small dmensonalty. When LDA+PCA tres to maxmze the between-class scatter n ths small null space, t loses a lot of nformaton. On the other hand, our method tres to maxmze the between-class scatter n the orgnal nput space. From Fg., we can

8 error rate LDA+PCA MMC KMMC class no. class no. (a) Each class contans three tranng samples. (b) Each class contans four tranng samples. Fgure : Comparson between our new methods and LDA+PCA when n c s close to D. error rate LDA+PCA MMC KMMC see that LDA+PCA s neffectve n ths stuaton because t s even worse than a random guess. But our method stll produced acceptable results. Thus, the expermental results show that our method s better than LDA+PCA n terms of both accuracy and effcency. 6 Concluson In ths paper, we proposed both lnear and nonlnear feature extractors based on the maxmum margn crteron. The new methods do not suffer from the small sample sze problem. The expermental results show that t s very effcent, accurate, and robust. Acknowledgments We thank D. Gunopulos, C. Domencon, and J. Peng for valuable dscussons and comments. Ths work was partally supported by NSF grants CCR and ACI References [] R. A. Fsher. The use of multple measurements n taxonomc problems. Annual of Eugencs, 7:79 88, 936. [] M. Loog, R. P. W. Dun, and R. Haeb-Umbach. Multclass lnear dmenson reducton by weghted parwse fsher crtera. IEEE Transactons on Pattern Analyss and Machne Intellgence, 3(7):76 766, 00. [3] K. Fukunaga. Introducton to Statstcal Pattern Recognton. Academc Press, New York, nd edton, 990. [4] Q. Tan, M. Barbero, Z. Gu, and S. Lee. Image classfcaton by the foley-sammon transform. Optcal Engneerng, 5(7): , 986. [5] Z. Hong and J. Yang. Optmal dscrmnant plane for a small number of samples and desgn method of classfer on the plane. Pattern Recognton, 4(4):37 34, 99. [6] G. W. Stewart. Introducton to Matrx Computatons. Academc Press, New York, 973. [7] K. Lu, Y. Cheng, and J. Yang. A generalzed optmal set of dscrmnant vectors. Pattern Recognton, 5(7):73 739, 99. [8] L. Chen, H. Lao, M.Ko, J. Ln, and G. Yu. A new LDA-based face recognton system whch can solve the small sample sze problem. Pattern Recognton, 33(0):73 76, 000. [9] S. Mka, G. Rätsch, J. Weston, B. Schölkopf, and K.-R. Müller. Fsher dscrmnant analyss wth kernels. In Y.-H. Hu, J. Larsen, E. Wlson, and S. Douglas, edtors, Neural Networks for Sgnal Processng IX, pages IEEE, 999. [0] V. N. Vapnk. Statstcal Learnng Theory. John Wley & Sons, New York, 998. [] F. Samara and A. Harter. Parametersaton of a stochastc model for human face dentfcaton. In Proceedngs of nd IEEE Workshop on Applcatons of Computer Vson, Sarasota, FL, 994.

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

Regularized Discriminant Analysis for Face Recognition

Regularized Discriminant Analysis for Face Recognition 1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths

More information

Lecture 10 Support Vector Machines II

Lecture 10 Support Vector Machines II Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed

More information

Fisher Linear Discriminant Analysis

Fisher Linear Discriminant Analysis Fsher Lnear Dscrmnant Analyss Max Wellng Department of Computer Scence Unversty of Toronto 10 Kng s College Road Toronto, M5S 3G5 Canada wellng@cs.toronto.edu Abstract Ths s a note to explan Fsher lnear

More information

CSE 252C: Computer Vision III

CSE 252C: Computer Vision III CSE 252C: Computer Vson III Lecturer: Serge Belonge Scrbe: Catherne Wah LECTURE 15 Kernel Machnes 15.1. Kernels We wll study two methods based on a specal knd of functon k(x, y) called a kernel: Kernel

More information

Lecture 12: Classification

Lecture 12: Classification Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna

More information

Feature Extraction by Maximizing the Average Neighborhood Margin

Feature Extraction by Maximizing the Average Neighborhood Margin Feature Extracton by Maxmzng the Average Neghborhood Margn Fe Wang, Changshu Zhang State Key Laboratory of Intellgent Technologes and Systems Department of Automaton, Tsnghua Unversty, Bejng, Chna. 184.

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

Unified Subspace Analysis for Face Recognition

Unified Subspace Analysis for Face Recognition Unfed Subspace Analyss for Face Recognton Xaogang Wang and Xaoou Tang Department of Informaton Engneerng The Chnese Unversty of Hong Kong Shatn, Hong Kong {xgwang, xtang}@e.cuhk.edu.hk Abstract PCA, LDA

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

A Novel Biometric Feature Extraction Algorithm using Two Dimensional Fisherface in 2DPCA subspace for Face Recognition

A Novel Biometric Feature Extraction Algorithm using Two Dimensional Fisherface in 2DPCA subspace for Face Recognition A Novel ometrc Feature Extracton Algorthm usng wo Dmensonal Fsherface n 2DPA subspace for Face Recognton R. M. MUELO, W.L. WOO, and S.S. DLAY School of Electrcal, Electronc and omputer Engneerng Unversty

More information

Statistical pattern recognition

Statistical pattern recognition Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

The Order Relation and Trace Inequalities for. Hermitian Operators

The Order Relation and Trace Inequalities for. Hermitian Operators Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Non-linear Canonical Correlation Analysis Using a RBF Network

Non-linear Canonical Correlation Analysis Using a RBF Network ESANN' proceedngs - European Smposum on Artfcal Neural Networks Bruges (Belgum), 4-6 Aprl, d-sde publ., ISBN -97--, pp. 57-5 Non-lnear Canoncal Correlaton Analss Usng a RBF Network Sukhbnder Kumar, Elane

More information

Which Separator? Spring 1

Which Separator? Spring 1 Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Kernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan

Kernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems

More information

Pattern Classification

Pattern Classification Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Supporting Information

Supporting Information Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to

More information

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix

Lectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could

More information

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal

Inner Product. Euclidean Space. Orthonormal Basis. Orthogonal Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,

More information

Support Vector Machines

Support Vector Machines Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x n class

More information

Errors for Linear Systems

Errors for Linear Systems Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch

More information

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)

ANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U) Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of

More information

Linear Classification, SVMs and Nearest Neighbors

Linear Classification, SVMs and Nearest Neighbors 1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush

More information

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography

CSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve

More information

MULTISPECTRAL IMAGE CLASSIFICATION USING BACK-PROPAGATION NEURAL NETWORK IN PCA DOMAIN

MULTISPECTRAL IMAGE CLASSIFICATION USING BACK-PROPAGATION NEURAL NETWORK IN PCA DOMAIN MULTISPECTRAL IMAGE CLASSIFICATION USING BACK-PROPAGATION NEURAL NETWORK IN PCA DOMAIN S. Chtwong, S. Wtthayapradt, S. Intajag, and F. Cheevasuvt Faculty of Engneerng, Kng Mongkut s Insttute of Technology

More information

Support Vector Machines

Support Vector Machines /14/018 Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x

More information

Numerical Heat and Mass Transfer

Numerical Heat and Mass Transfer Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and

More information

Natural Language Processing and Information Retrieval

Natural Language Processing and Information Retrieval Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support

More information

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k)

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k) ISSN 1749-3889 (prnt), 1749-3897 (onlne) Internatonal Journal of Nonlnear Scence Vol.17(2014) No.2,pp.188-192 Modfed Block Jacob-Davdson Method for Solvng Large Sparse Egenproblems Hongy Mao, College of

More information

Research Article Green s Theorem for Sign Data

Research Article Green s Theorem for Sign Data Internatonal Scholarly Research Network ISRN Appled Mathematcs Volume 2012, Artcle ID 539359, 10 pages do:10.5402/2012/539359 Research Artcle Green s Theorem for Sgn Data Lous M. Houston The Unversty of

More information

A kernel method for canonical correlation analysis

A kernel method for canonical correlation analysis A kernel method for canoncal correlaton analyss Shotaro Akaho AIST Neuroscence Research Insttute, Central 2, - Umezono, Tsukuba, Ibarak 3058568, Japan s.akaho@ast.go.jp http://staff.ast.go.jp/s.akaho/

More information

Physics 5153 Classical Mechanics. Principle of Virtual Work-1

Physics 5153 Classical Mechanics. Principle of Virtual Work-1 P. Guterrez 1 Introducton Physcs 5153 Classcal Mechancs Prncple of Vrtual Work The frst varatonal prncple we encounter n mechancs s the prncple of vrtual work. It establshes the equlbrum condton of a mechancal

More information

14 Lagrange Multipliers

14 Lagrange Multipliers Lagrange Multplers 14 Lagrange Multplers The Method of Lagrange Multplers s a powerful technque for constraned optmzaton. Whle t has applcatons far beyond machne learnng t was orgnally developed to solve

More information

The Minimum Universal Cost Flow in an Infeasible Flow Network

The Minimum Universal Cost Flow in an Infeasible Flow Network Journal of Scences, Islamc Republc of Iran 17(2): 175-180 (2006) Unversty of Tehran, ISSN 1016-1104 http://jscencesutacr The Mnmum Unversal Cost Flow n an Infeasble Flow Network H Saleh Fathabad * M Bagheran

More information

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach

A Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Singular Value Decomposition: Theory and Applications

Singular Value Decomposition: Theory and Applications Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real

More information

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018

INF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018 INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton

More information

The Prncpal Component Transform The Prncpal Component Transform s also called Karhunen-Loeve Transform (KLT, Hotellng Transform, oregenvector Transfor

The Prncpal Component Transform The Prncpal Component Transform s also called Karhunen-Loeve Transform (KLT, Hotellng Transform, oregenvector Transfor Prncpal Component Transform Multvarate Random Sgnals A real tme sgnal x(t can be consdered as a random process and ts samples x m (m =0; ;N, 1 a random vector: The mean vector of X s X =[x0; ;x N,1] T

More information

Relevance Vector Machines Explained

Relevance Vector Machines Explained October 19, 2010 Relevance Vector Machnes Explaned Trstan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introducton Ths document has been wrtten n an attempt to make Tppng s [1] Relevance Vector Machnes

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

CHAPTER III Neural Networks as Associative Memory

CHAPTER III Neural Networks as Associative Memory CHAPTER III Neural Networs as Assocatve Memory Introducton One of the prmary functons of the bran s assocatve memory. We assocate the faces wth names, letters wth sounds, or we can recognze the people

More information

APPENDIX A Some Linear Algebra

APPENDIX A Some Linear Algebra APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

x = , so that calculated

x = , so that calculated Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to

More information

Semi-supervised Classification with Active Query Selection

Semi-supervised Classification with Active Query Selection Sem-supervsed Classfcaton wth Actve Query Selecton Jao Wang and Swe Luo School of Computer and Informaton Technology, Beng Jaotong Unversty, Beng 00044, Chna Wangjao088@63.com Abstract. Labeled samples

More information

Tensor Subspace Analysis

Tensor Subspace Analysis Tensor Subspace Analyss Xaofe He 1 Deng Ca Partha Nyog 1 1 Department of Computer Scence, Unversty of Chcago {xaofe, nyog}@cs.uchcago.edu Department of Computer Scence, Unversty of Illnos at Urbana-Champagn

More information

Image classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?

Image classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them? Image classfcaton Gven te bag-of-features representatons of mages from dfferent classes ow do we learn a model for dstngusng tem? Classfers Learn a decson rule assgnng bag-offeatures representatons of

More information

Report on Image warping

Report on Image warping Report on Image warpng Xuan Ne, Dec. 20, 2004 Ths document summarzed the algorthms of our mage warpng soluton for further study, and there s a detaled descrpton about the mplementaton of these algorthms.

More information

LECTURE 9 CANONICAL CORRELATION ANALYSIS

LECTURE 9 CANONICAL CORRELATION ANALYSIS LECURE 9 CANONICAL CORRELAION ANALYSIS Introducton he concept of canoncal correlaton arses when we want to quantfy the assocatons between two sets of varables. For example, suppose that the frst set of

More information

1 Convex Optimization

1 Convex Optimization Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,

More information

Support Vector Machines CS434

Support Vector Machines CS434 Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

COS 521: Advanced Algorithms Game Theory and Linear Programming

COS 521: Advanced Algorithms Game Theory and Linear Programming COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton

More information

The exam is closed book, closed notes except your one-page cheat sheet.

The exam is closed book, closed notes except your one-page cheat sheet. CS 89 Fall 206 Introducton to Machne Learnng Fnal Do not open the exam before you are nstructed to do so The exam s closed book, closed notes except your one-page cheat sheet Usage of electronc devces

More information

Lecture 10: Dimensionality reduction

Lecture 10: Dimensionality reduction Lecture : Dmensonalt reducton g The curse of dmensonalt g Feature etracton s. feature selecton g Prncpal Components Analss g Lnear Dscrmnant Analss Intellgent Sensor Sstems Rcardo Guterrez-Osuna Wrght

More information

Support Vector Machines CS434

Support Vector Machines CS434 Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? + + + + + + + + + Intuton of Margn Consder ponts

More information

Advanced Introduction to Machine Learning

Advanced Introduction to Machine Learning Advanced Introducton to Machne Learnng 10715, Fall 2014 The Kernel Trck, Reproducng Kernel Hlbert Space, and the Representer Theorem Erc Xng Lecture 6, September 24, 2014 Readng: Erc Xng @ CMU, 2014 1

More information

CS 468 Lecture 16: Isometry Invariance and Spectral Techniques

CS 468 Lecture 16: Isometry Invariance and Spectral Techniques CS 468 Lecture 16: Isometry Invarance and Spectral Technques Justn Solomon Scrbe: Evan Gawlk Introducton. In geometry processng, t s often desrable to characterze the shape of an object n a manner that

More information

Subspace Learning Based on Tensor Analysis. by Deng Cai, Xiaofei He, and Jiawei Han

Subspace Learning Based on Tensor Analysis. by Deng Cai, Xiaofei He, and Jiawei Han Report No. UIUCDCS-R-2005-2572 UILU-ENG-2005-1767 Subspace Learnng Based on Tensor Analyss by Deng Ca, Xaofe He, and Jawe Han May 2005 Subspace Learnng Based on Tensor Analyss Deng Ca Xaofe He Jawe Han

More information

Chapter 6 Support vector machine. Séparateurs à vaste marge

Chapter 6 Support vector machine. Séparateurs à vaste marge Chapter 6 Support vector machne Séparateurs à vaste marge Méthode de classfcaton bnare par apprentssage Introdute par Vladmr Vapnk en 1995 Repose sur l exstence d un classfcateur lnéare Apprentssage supervsé

More information

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015

CS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015 CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research

More information

Support Vector Machines

Support Vector Machines CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at

More information

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system

Transfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z ) C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z

More information

Inexact Newton Methods for Inverse Eigenvalue Problems

Inexact Newton Methods for Inverse Eigenvalue Problems Inexact Newton Methods for Inverse Egenvalue Problems Zheng-jan Ba Abstract In ths paper, we survey some of the latest development n usng nexact Newton-lke methods for solvng nverse egenvalue problems.

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 6. Rdge regresson The OLSE s the best lnear unbased

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers

Psychology 282 Lecture #24 Outline Regression Diagnostics: Outliers Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.

More information

arxiv:cs.cv/ Jun 2000

arxiv:cs.cv/ Jun 2000 Correlaton over Decomposed Sgnals: A Non-Lnear Approach to Fast and Effectve Sequences Comparson Lucano da Fontoura Costa arxv:cs.cv/0006040 28 Jun 2000 Cybernetc Vson Research Group IFSC Unversty of São

More information

PARTICIPATION FACTOR IN MODAL ANALYSIS OF POWER SYSTEMS STABILITY

PARTICIPATION FACTOR IN MODAL ANALYSIS OF POWER SYSTEMS STABILITY POZNAN UNIVE RSITY OF TE CHNOLOGY ACADE MIC JOURNALS No 86 Electrcal Engneerng 6 Volodymyr KONOVAL* Roman PRYTULA** PARTICIPATION FACTOR IN MODAL ANALYSIS OF POWER SYSTEMS STABILITY Ths paper provdes a

More information

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013

ISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013 ISSN: 2277-375 Constructon of Trend Free Run Orders for Orthogonal rrays Usng Codes bstract: Sometmes when the expermental runs are carred out n a tme order sequence, the response can depend on the run

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Estimating the Fundamental Matrix by Transforming Image Points in Projective Space 1

Estimating the Fundamental Matrix by Transforming Image Points in Projective Space 1 Estmatng the Fundamental Matrx by Transformng Image Ponts n Projectve Space 1 Zhengyou Zhang and Charles Loop Mcrosoft Research, One Mcrosoft Way, Redmond, WA 98052, USA E-mal: fzhang,cloopg@mcrosoft.com

More information

Lecture 3: Dual problems and Kernels

Lecture 3: Dual problems and Kernels Lecture 3: Dual problems and Kernels C4B Machne Learnng Hlary 211 A. Zsserman Prmal and dual forms Lnear separablty revsted Feature mappng Kernels for SVMs Kernel trck requrements radal bass functons SVM

More information

Efficient, General Point Cloud Registration with Kernel Feature Maps

Efficient, General Point Cloud Registration with Kernel Feature Maps Effcent, General Pont Cloud Regstraton wth Kernel Feature Maps Hanchen Xong, Sandor Szedmak, Justus Pater Insttute of Computer Scence Unversty of Innsbruck 30 May 2013 Hanchen Xong (Un.Innsbruck) 3D Regstraton

More information

Chapter 12 Analysis of Covariance

Chapter 12 Analysis of Covariance Chapter Analyss of Covarance Any scentfc experment s performed to know somethng that s unknown about a group of treatments and to test certan hypothess about the correspondng treatment effect When varablty

More information

Composite Hypotheses testing

Composite Hypotheses testing Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter

More information

VQ widely used in coding speech, image, and video

VQ widely used in coding speech, image, and video at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng

More information

Pattern Recognition 42 (2009) Contents lists available at ScienceDirect. Pattern Recognition. journal homepage:

Pattern Recognition 42 (2009) Contents lists available at ScienceDirect. Pattern Recognition. journal homepage: Pattern Recognton 4 (9) 764 -- 779 Contents lsts avalable at ScenceDrect Pattern Recognton ournal homepage: www.elsever.com/locate/pr Perturbaton LDA: Learnng the dfference between the class emprcal mean

More information

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family

Using T.O.M to Estimate Parameter of distributions that have not Single Exponential Family IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran

More information

ECE559VV Project Report

ECE559VV Project Report ECE559VV Project Report (Supplementary Notes Loc Xuan Bu I. MAX SUM-RATE SCHEDULING: THE UPLINK CASE We have seen (n the presentaton that, for downlnk (broadcast channels, the strategy maxmzng the sum-rate

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Comparative Studies of Law of Conservation of Energy. and Law Clusters of Conservation of Generalized Energy

Comparative Studies of Law of Conservation of Energy. and Law Clusters of Conservation of Generalized Energy Comparatve Studes of Law of Conservaton of Energy and Law Clusters of Conservaton of Generalzed Energy No.3 of Comparatve Physcs Seres Papers Fu Yuhua (CNOOC Research Insttute, E-mal:fuyh1945@sna.com)

More information

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI

Logistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton

More information

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0

n α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0 MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector

More information

Some Comments on Accelerating Convergence of Iterative Sequences Using Direct Inversion of the Iterative Subspace (DIIS)

Some Comments on Accelerating Convergence of Iterative Sequences Using Direct Inversion of the Iterative Subspace (DIIS) Some Comments on Acceleratng Convergence of Iteratve Sequences Usng Drect Inverson of the Iteratve Subspace (DIIS) C. Davd Sherrll School of Chemstry and Bochemstry Georga Insttute of Technology May 1998

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Laboratory 3: Method of Least Squares

Laboratory 3: Method of Least Squares Laboratory 3: Method of Least Squares Introducton Consder the graph of expermental data n Fgure 1. In ths experment x s the ndependent varable and y the dependent varable. Clearly they are correlated wth

More information

Lecture 3. Ax x i a i. i i

Lecture 3. Ax x i a i. i i 18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest

More information

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k ANOVA Model and Matrx Computatons Notaton The followng notaton s used throughout ths chapter unless otherwse stated: N F CN Y Z j w W Number of cases Number of factors Number of covarates Number of levels

More information