Efficient and Robust Feature Extraction by Maximum Margin Criterion
|
|
- Alexia Booth
- 5 years ago
- Views:
Transcription
1 Effcent and Robust Feature Extracton by Maxmum Margn Crteron Hafeng L Tao Jang Department of Computer Scence Unversty of Calforna Rversde, CA 95 {hl,jang}@cs.ucr.edu Keshu Zhang Department of Electrcal Engneerng Unversty of New Orleans New Orleans, LA 7048 kzhang@uno.edu Abstract A new feature extracton crteron, maxmum margn crteron (MMC), s proposed n ths paper. Ths new crteron s general n the sense that, when combned wth a sutable constrant, t can actually gve rse to the most popular feature extractor n the lterature, lnear dscrmnate analyss (LDA). We derve a new feature extractor based on MMC usng a dfferent constrant that does not depend on the nonsngularty of the wthn-class scatter matrx S w. Such a dependence s a major drawback of LDA especally when the sample sze s small. The kernelzed (nonlnear) counterpart of ths lnear feature extractor s also establshed n ths paper. Our prelmnary expermental results on face mages demonstrate that the new feature extractors are effcent and stable. Introducton In statstcal pattern recognton, the hgh-dmensonalty s a major cause of the practcal lmtatons of many pattern recognton technologes. In the past several decades, many dmensonalty reducton technques have been proposed. Lnear dscrmnant analyss (LDA, also called Fsher s Lnear Dscrmnant) [] s one of the most popular lnear dmensonalty reducton method. In many applcatons, LDA has been proven to be very powerful. LDA s gven by a lnear transformaton matrx W R D d maxmzng the so-called Fsher crteron (a knd of Raylegh coeffcent) J F (W) = WT S b W W T S w W where S b = c = p (m m)(m m) T and S w = c = p S are the betweenclass scatter matrx and the wthn-class scatter matrx, respectvely; c s the number of classes; m and p are the mean vector and a pror probablty of class, respectvely; m = c = p m s the overall mean vector; S s the wthn-class scatter matrx of class ; D and d are the dmensonaltes of the data before and after the transformaton, respectvely. To maxmze (), the transformaton matrx W must be consttuted by the largest egenvectors of S w S b. The purpose of LDA s to maxmze the between-class scatter whle smultaneously mnmzng the wthn-class scatter. The two-class LDA has a close connecton to optmal lnear Bayes classfers. In the two-class case, the transformaton matrx W s just a vector, whch s n the same drecton as the dscrmnant n the correspondng optmal Bayes classfer. However, t has been shown that LDA s suboptmal for mult-class problems []. A major drawback of LDA s that t cannot be appled when S w s sngular because of the small sample sze problem [3]. The small sample sze problem ()
2 arses whenever the number of samples s smaller than the dmensonalty of the samples. For example, a mage n a face recognton system has 4096 dmensons, whch requres more than 4096 tranng ponts to ensure that S w s nonsngular. So, LDA s not a stable method n practce when the tranng data s scarce. In recent years, many researchers have notced ths problem and tred to overcome the computatonal dffculty wth LDA. Tan et al. [4] used the pseudo-nverse matrx S + w nstead of the nverse matrx S w. For the same purpose, Hong and Yang [5] tred to add a sngular value perturbaton to S w to make t nonsngular. Nether of these methods are theoretcally sound because Fsher s crteron s not vald when S w s sngular. When S w s sngular, any postve S b makes Fsher s crteron nfntely large. Thus, these nave attempts to calculate the (pseudo or approxmate) nverse of S w may lead to arbtrary (meanngless) results. Besdes, t s also known that an egenvector could be very senstve to small perturbaton f ts correspondng egenvalue s close to another egenvalue of the same matrx [6]. In 99, Lu et al. [7] modfed Fsher s crteron by usng the total scatter matrx S t = S b + S w as the denomnator nstead of S w. It has been proven that the modfed crteron s exactly equvalent to Fsher s crteron. However, when S w s sngular, the modfed crteron reaches the maxmum value (.e., ) no matter what the transformaton W s. Such an arbtrary transformaton cannot guarantee the maxmum class separablty unless W T S b W s maxmzed. Besdes, ths method need stll calculate an nverse matrx, whch s tme consumng. In 000, Chen et al. [8] proposed the LDA+PCA method. When S w s of full rank, the LDA+PCA method just calculates the maxmum egenvectors of S t S b to form the transformaton matrx. Otherwse, a two-stage procedure s employed. Frst, the data are transformed nto the null space V 0 of S w. Second, t tres to maxmze the betweenclass scatter n V 0, whch s accomplshed by performng prncpal component analyss (PCA) on the between-class scatter matrx n V 0. Although ths method solves the small sample sze problem, t s obvously suboptmal because t maxmzes the between-class scatter n the null space of S w nstead of the orgnal nput space. Besdes, the performance of the LDA+PCA method drops sgnfcantly when n c s close to the dmensonalty D, where n s the number of samples and c s the number of classes. The reason s that the dmensonalty of the null space V 0 s too small n ths stuaton, and too much nformaton s lost when we try to extract the dscrmnant vectors n V 0. LDA+PCA also need calculate the rank of S w, whch s an ll-defned operaton due to floatng-pont mprecsons. At last, ths method s complcated and slow because too much calculaton s nvolved. Kernel Fsher s Dscrmnant (KFD) [9] s a well-known nonlnear extenson to LDA. The nstablty problem s more severe for KFD because S w n the (nonlnear) feature space F s always sngular (the rank of S w s n c). Smlar to [5], KFD smply adds a perturbaton µi to S w. Of course, t has the same stablty problem as that n [5] because egenvectors are senstve to small perturbaton. Although the authors also argued that ths perturbaton acts as some knd of regularzaton,.e., a capacty control n F, the real nfluence n ths settng of regularzaton s not yet fully understood. Besdes, t s hard to determne an optmal µ snce there are no theoretcal gudelnes. In ths paper, a smpler, more effcent, and stable method s proposed to calculate the most dscrmnant vectors based on a new feature extracton crteron, the maxmum margn crteron (MMC). Based on MMC, new lnear and nonlnear feature extractors are establshed. It can be shown that MMC represents class separablty better than PCA. As a connecton to Fsher s crteron, we may also derve LDA from MMC by ncorporatng some sutable constrant. On the other hand, the new feature extractors derved above (based on MMC) do not suffer from the small sample sze problem, whch s known to cause serous stablty problems for LDA (based on Fsher s crteron). Dfferent from LDA+PCA, the new feature extractors based on MMC maxmze the between-class scatter n the nput space nstead of the null space of S w. Hence, t has a better overall performance than LDA+PCA, as confrmed by our prelmnary expermental results.
3 Maxmum Margn Crteron Suppose that we are gven emprcal data (x, y ),..., (x n, y n ) X {C,..., C c } Here, the doman X R D s some nonempty set that the patterns x are taken from. The y s are called labels or targets. By studyng these samples, we want to predct the label y {C,..., C c } of some new pattern x X. In other words, we choose y such that (x, y) s n some sense smlar to the tranng examples. For ths purpose, some measure need be employed to assess smlarty or dssmlarty. We want to keep such smlarty/dssmlarty nformaton as much as possble after the dmensonalty reducton,.e., transformng x from R D to R d, where d D. If some dstance metrc s used to measure the dssmlarty, we would hope that a pattern s close to those n the same class but far from those n dfferent classes. So, a good feature extractor should maxmze the dstances between classes after the transformaton. Therefore, we may defne the feature extracton crteron as J = p p j d(c, C j ) () We call () the maxmum margn crteron (MMC). It s actually the summaton of c(c ) nterclass margns. Lke the weghted parwse Fsher s crtera n [], one may also defne a weghted maxmum margn crteron. Due to the page lmt, we omt the dscusson n ths paper. One may use the dstance between mean vectors as the dstance between classes,.e. d(c, C j ) = d(m, m j ) (3) where m and m j are the mean vectors of the class C and the class C j, respectvely. However, (3) s not sutable snce t neglects the scatter of classes. Even f the dstance between the mean vectors s large, t s not easy to separate two classes that have the large spread and overlap wth each other. By consderng the scatter of classes, we defne the nterclass dstance (or margn) as d(c, C j ) = d(m, m j ) s(c ) s(c j ) (4) where s(c ) s some measure of the scatter of the class C. In statstcs, we usually use the generalzed varance S or overall varance tr(s ) to measure the scatter of data. In ths paper, we use the overall varance tr(s ) because t s easy to analyze. The weakness of the overall varance s that t gnores covarance structure altogether. Note that, by employng the overall/generalzed varance, the expresson (4) measures the average margn between two classes whle the mnmum margn s used n support vector machnes (SVMs) [0]. Wth (4) and s(c ) beng tr(s ), we may decompose () nto two parts J = p p j (d(m, m j ) tr(s ) tr(s j )) = p p j d(m, m j ) p p j (tr(s ) + tr(s j )) The second part s easly smplfed to tr(s w ) ( ) p p j (tr(s ) + tr(s j )) = p tr(s ) = tr p S = tr(s w ) (5) = =
4 By employng the Eucldean dstance, we may also smplfy the frst part to tr(s b ) as follows p p j d(m, m j ) = p p j (m m j ) T (m m j ) = p p j (m m + m m j ) T (m m + m m j ) After expandng t, we can smplfy the above equaton to c = p (m m) T (m m) by usng the fact c j= p j(m m j ) = 0. So ( ) p p j d(m, m j ) = tr p (m m)(m m) T = tr(s b ) (6) = Now we obtan J = tr(s b S w ) (7) Snce tr(s b ) measures the overall varance of the class mean vectors, a large tr(s b ) mples that the class mean vectors scatter n a large space. On the other hand, a small tr(s w ) mples that every class has a small spread. Thus, a large J ndcates that patterns are close to each other f they are from the same class but are far from each other f they are from dfferent classes. Thus, ths crteron may represent class separablty better than PCA. Recall that PCA tres to maxmze the total scatter after a lnear transformaton. But the data set wth a large wthn-class scatter can also have a large total scatter even when t has a small between-class scatter because S t = S b + S w. Obvously, such data are not easy to classfy. Compared wth LDA+PCA, we maxmze the between-class scatter n nput space rather than the null space of S w when S w s sngular. So, our method can keep more dscrmnatve nformaton than LDA+PCA does. 3 Lnear Feature Extracton When performng dmensonalty reducton, we want to fnd a (lnear or nonlnear) mappng from the measurement space M to some feature space F such that J s maxmzed after the transformaton. In ths secton, we dscuss how to fnd an optmal lnear feature extractor. In the next secton, we wll generalze t to the nonlnear case. Consder a lnear mappng W R D d. We would lke to maxmze J(W) = tr(s W b S W w ) where S W b and S W w are the between-class scatter matrx and wthn-class scatter matrx n the feature space F. Snce W s a lnear mappng, t s easy to show S W b = W T S b W and = W T S w W. So, we have S W w J(W) = tr ( W T (S b S w )W ) (8) In ths formulaton, we have the freedom to multply W wth some nonzero constant. Thus, we addtonally requre that W s consttuted by the unt vectors,.e. W = [w w... w d ] and wk T w k =. Ths means that we need solve the followng constraned optmzaton max wk T (S b S w )w k subject to w T k w k = 0 k =,..., d
5 Note that, we may also use other constrants n the above. For example, we may requre tr ( W T S w W ) = and then maxmze tr ( W T S b W ). It s easy to show that maxmzng MMC wth such a constrant n fact results n LDA. The only dfference s that t nvolves a constraned optmzaton whereas the tradtonal LDA solves an unconstraned optmzaton. The motvaton for usng the constrant w T k w k = s that t allows us to avod calculatng the nverse of S w and thus the potental small sample sze problem. To solve the above optmzaton problem, we may ntroduce a Lagrangan L(w k, λ k ) = wk T (S b S w )w k λ k (wk T w k ) (9) wth multplers λ k. The Lagrangan L has to be maxmzed wth respect to λ k and w k. The condton that at the statonary pont, the dervatves of L wth respect to w k must vansh L(w k, λ k ) = ((S w b S w ) λ k I)w k = 0 k k =,..., d (0) leads to (S b S w )w k =λ k w k k =,..., d () whch means that the λ k s are the egenvalues of S b S w and the w k s are the correspondng egenvectors. Thus J(W) = wk T (S b S w )w k = λ k wk T w k = λ k () Therefore, J(W) s maxmzed when W s composed of the frst d largest egenvectors of S b S w. Here, we need not calculate the nverse of S w, whch allows us to avod the small sample sze problem easly. We may also requre W to be orthonormal, whch may help preserve the shape of the dstrbuton. 4 Nonlnear Feature Extracton wth Kernel In ths secton, we follow the approach of nonlnear SVMs [0] to kernelze the above lnear feature extractor. More precsely, we frst reformulate the maxmum margn crteron n terms of only dot-product Φ(x), Φ(y) of nput patterns. Then we replace the dot-product by some postve defnte kernel k(x, y), e.g. Gaussan kernel e γ x y. Consder the maxmum margn crteron n the feature space F J Φ (W) = wk T (S Φ b SΦ w)w k where S Φ b and S Φ w are the between-class scatter matrx and wthn-class scatter matrx n F,.e., S Φ b = c = p (m Φ m Φ )(m Φ m Φ ) T, S Φ w = c = p S Φ and S Φ = n n j= (Φ(x() j ) mφ )(Φ(x() j ) mφ )T wth m Φ = n n j= Φ(x() j ), mφ = c = p m Φ, and x() j s the pattern of class C that has n samples. For us, an mportant fact s that each w k les n the span of Φ(x ), Φ(x ),..., Φ(x n ). Therefore, we can fnd an expanson for w k n the form w k = n l= α(k) l Φ(x l ). Usng ths expanson and the defnton of m Φ, we have n wk T m Φ = α (k) n l Φ(x n l ), Φ(x () j ) l= j=
6 Replacng the dot-product by some kernel functon k(x, y) and defnng ( m ) l = n n j= k(x l, x () j ), we get wt k mφ = α T k m wth (α k ) l = α (k) l. Smlarly, we have w T k m Φ = w T k p m Φ = α T k = p m = α T k m wth m = c = p m. Ths means w T k (mφ mφ ) = α T k ( m m). and wk T S Φ b w k = = = = = p (wk T (m Φ m Φ ))(wk T (m Φ m Φ )) T p T α T k ( m m)( m m) T α k = where S b = c = p ( m m)( m m) T. α T S k b α k Smlarly, one can smplfy W T S Φ ww. Frst, we have wk T (Φ(x() j ) mφ ) = αt k (k() j m ) wth (k () j ) l = k(x l, x () j ). Consderng wt k SΦ w k = n n j= (wt k (Φ(x() j ) m Φ ))(wt k (Φ(x() j ) mφ ))T, we have w T k S Φ w k = n n j= α T k (k () j m )(k () j m ) T α k = n α T n S k (e j n )(e j n ) T ST n n α k j= = n α T n S k (e j e T j e j T n n n e T j + j= n n n T n ) S T α k = n α T k S (I n n n n T n ) S T α k where ( S ) lj = k(x l, x () j ), I n n s the n n dentty matrx, n s the n -dmensonal vector of s, and e j s the canoncal bass vector of n dmensons. Thus, we obtan wk T S Φ ww k = = α T k ( = = p n α T k S (I n n n T n ) S T α k p n S (I n n n T n ) S T ) α k = α T S k w α k where S w = c = p n S (I n n n T n ) S T. So the maxmum crteron n the feature space F s J(W) = α T k ( S b S w )α k (3) Smlar to the observatons n Secton 3, the above crteron s maxmzed by the largest egenvectors of S b S w.
7 error rate RAW LDA+PCA MMC KMMC class no. class no. (a) The comparson n term of error rate. (b) The comparson n term of tranng tme. Fgure : Expermental results obtaned usng a lnear SVM on the orgnal data (RAW), and the data extracted by LDA+PCA, the lnear feature extractor based on MMC (MMC) and the nonlnear feature extractor based on MMC (KMMC), whch employs the Gaussan kernel wth γ = tranng tme (second) LDA+PCA MMC KMMC Experments To evaluate the performance of our new methods (both lnear and nonlnear feature extractors), we ran both LDA+PCA and our methods on the ORL face dataset []. The ORL dataset conssts of 0 face mages from 40 subjects for a total of 400 mages, wth some varaton n pose, facal expresson and detals. The resoluton of the mages s 9, wth 56 gray-levels. Frst, we reszed the mages to 8 3 to save the expermental tme. Then, we reduced the dmensonalty of each mage set to c, where c s the number of classes. At last we traned and tested a lnear SVM on the dmensonalty-reduced data. As a control, we also traned and tested a lnear SVM on the orgnal data before ts dmensonalty was reduced. In order to demonstrate the effectveness and the effcency of our methods, we conducted a seres of experments and compared our results wth those obtaned usng LDA+PCA. The error rates are shown n Fg.(a). When traned wth 3 samples and tested wth 7 other samples for each class, our method s generally better than LDA+PCA. In fact, our method s usually better than LDA+PCA on other numbers of tranng samples. To save space, we do not show all the results here. Note that our methods can even acheve lower error rates than a lnear SVM on the orgnal data (wthout dmensonalty reducton). However, LDA+PCA does not demonstrate such a clear superorty over RAW. Fg. (a) also shows that the kernelzed (nonlnear) feature extractor based on MMC s sgnfcantly better than the lnear one, n partcular when the number of classes c s large. Besdes accuracy, our methods are also much more effcent than LDA+PCA n the sense of the tranng tme requred. Fg. (b) shows that our lnear feature extractor s about 4 tmes faster than LDA+PCA. The same speedup was observed on other numbers of tranng samples. Note that our nonlnear feature extractor s also faster than LDA+PCA n ths case although t s very tme-consumng to calculate the kernel matrx n general. An explanaton of the speedup s that the kernel matrx sze equals the number of samples, whch s pretty small n ths case. Furthermore, our method performs much better than LDA+PCA when n c s close to the dmensonalty D. Because the amount of tranng data was lmted, we reszed the mages to 68 dmensons to create such a stuaton. The expermental results are shown n Fg.. In ths stuaton, the performance of LDA+PCA drops sgnfcantly because the null space of S w has a small dmensonalty. When LDA+PCA tres to maxmze the between-class scatter n ths small null space, t loses a lot of nformaton. On the other hand, our method tres to maxmze the between-class scatter n the orgnal nput space. From Fg., we can
8 error rate LDA+PCA MMC KMMC class no. class no. (a) Each class contans three tranng samples. (b) Each class contans four tranng samples. Fgure : Comparson between our new methods and LDA+PCA when n c s close to D. error rate LDA+PCA MMC KMMC see that LDA+PCA s neffectve n ths stuaton because t s even worse than a random guess. But our method stll produced acceptable results. Thus, the expermental results show that our method s better than LDA+PCA n terms of both accuracy and effcency. 6 Concluson In ths paper, we proposed both lnear and nonlnear feature extractors based on the maxmum margn crteron. The new methods do not suffer from the small sample sze problem. The expermental results show that t s very effcent, accurate, and robust. Acknowledgments We thank D. Gunopulos, C. Domencon, and J. Peng for valuable dscussons and comments. Ths work was partally supported by NSF grants CCR and ACI References [] R. A. Fsher. The use of multple measurements n taxonomc problems. Annual of Eugencs, 7:79 88, 936. [] M. Loog, R. P. W. Dun, and R. Haeb-Umbach. Multclass lnear dmenson reducton by weghted parwse fsher crtera. IEEE Transactons on Pattern Analyss and Machne Intellgence, 3(7):76 766, 00. [3] K. Fukunaga. Introducton to Statstcal Pattern Recognton. Academc Press, New York, nd edton, 990. [4] Q. Tan, M. Barbero, Z. Gu, and S. Lee. Image classfcaton by the foley-sammon transform. Optcal Engneerng, 5(7): , 986. [5] Z. Hong and J. Yang. Optmal dscrmnant plane for a small number of samples and desgn method of classfer on the plane. Pattern Recognton, 4(4):37 34, 99. [6] G. W. Stewart. Introducton to Matrx Computatons. Academc Press, New York, 973. [7] K. Lu, Y. Cheng, and J. Yang. A generalzed optmal set of dscrmnant vectors. Pattern Recognton, 5(7):73 739, 99. [8] L. Chen, H. Lao, M.Ko, J. Ln, and G. Yu. A new LDA-based face recognton system whch can solve the small sample sze problem. Pattern Recognton, 33(0):73 76, 000. [9] S. Mka, G. Rätsch, J. Weston, B. Schölkopf, and K.-R. Müller. Fsher dscrmnant analyss wth kernels. In Y.-H. Hu, J. Larsen, E. Wlson, and S. Douglas, edtors, Neural Networks for Sgnal Processng IX, pages IEEE, 999. [0] V. N. Vapnk. Statstcal Learnng Theory. John Wley & Sons, New York, 998. [] F. Samara and A. Harter. Parametersaton of a stochastc model for human face dentfcaton. In Proceedngs of nd IEEE Workshop on Applcatons of Computer Vson, Sarasota, FL, 994.
Support Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationKernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationRegularized Discriminant Analysis for Face Recognition
1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationFisher Linear Discriminant Analysis
Fsher Lnear Dscrmnant Analyss Max Wellng Department of Computer Scence Unversty of Toronto 10 Kng s College Road Toronto, M5S 3G5 Canada wellng@cs.toronto.edu Abstract Ths s a note to explan Fsher lnear
More informationCSE 252C: Computer Vision III
CSE 252C: Computer Vson III Lecturer: Serge Belonge Scrbe: Catherne Wah LECTURE 15 Kernel Machnes 15.1. Kernels We wll study two methods based on a specal knd of functon k(x, y) called a kernel: Kernel
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationFeature Extraction by Maximizing the Average Neighborhood Margin
Feature Extracton by Maxmzng the Average Neghborhood Margn Fe Wang, Changshu Zhang State Key Laboratory of Intellgent Technologes and Systems Department of Automaton, Tsnghua Unversty, Bejng, Chna. 184.
More informationP R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /
Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons
More informationUnified Subspace Analysis for Face Recognition
Unfed Subspace Analyss for Face Recognton Xaogang Wang and Xaoou Tang Department of Informaton Engneerng The Chnese Unversty of Hong Kong Shatn, Hong Kong {xgwang, xtang}@e.cuhk.edu.hk Abstract PCA, LDA
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationA Novel Biometric Feature Extraction Algorithm using Two Dimensional Fisherface in 2DPCA subspace for Face Recognition
A Novel ometrc Feature Extracton Algorthm usng wo Dmensonal Fsherface n 2DPA subspace for Face Recognton R. M. MUELO, W.L. WOO, and S.S. DLAY School of Electrcal, Electronc and omputer Engneerng Unversty
More informationStatistical pattern recognition
Statstcal pattern recognton Bayes theorem Problem: decdng f a patent has a partcular condton based on a partcular test However, the test s mperfect Someone wth the condton may go undetected (false negatve
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have
More informationThe Order Relation and Trace Inequalities for. Hermitian Operators
Internatonal Mathematcal Forum, Vol 3, 08, no, 507-57 HIKARI Ltd, wwwm-hkarcom https://doorg/0988/mf088055 The Order Relaton and Trace Inequaltes for Hermtan Operators Y Huang School of Informaton Scence
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationNon-linear Canonical Correlation Analysis Using a RBF Network
ESANN' proceedngs - European Smposum on Artfcal Neural Networks Bruges (Belgum), 4-6 Aprl, d-sde publ., ISBN -97--, pp. 57-5 Non-lnear Canoncal Correlaton Analss Usng a RBF Network Sukhbnder Kumar, Elane
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationKernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan
Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems
More informationPattern Classification
Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationSupporting Information
Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationInner Product. Euclidean Space. Orthonormal Basis. Orthogonal
Inner Product Defnton 1 () A Eucldean space s a fnte-dmensonal vector space over the reals R, wth an nner product,. Defnton 2 (Inner Product) An nner product, on a real vector space X s a symmetrc, blnear,
More informationSupport Vector Machines
Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x n class
More informationErrors for Linear Systems
Errors for Lnear Systems When we solve a lnear system Ax b we often do not know A and b exactly, but have only approxmatons  and ˆb avalable. Then the best thng we can do s to solve ˆx ˆb exactly whch
More informationANSWERS. Problem 1. and the moment generating function (mgf) by. defined for any real t. Use this to show that E( U) var( U)
Econ 413 Exam 13 H ANSWERS Settet er nndelt 9 deloppgaver, A,B,C, som alle anbefales å telle lkt for å gøre det ltt lettere å stå. Svar er gtt . Unfortunately, there s a prntng error n the hnt of
More informationLinear Classification, SVMs and Nearest Neighbors
1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush
More informationCSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography
CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve
More informationMULTISPECTRAL IMAGE CLASSIFICATION USING BACK-PROPAGATION NEURAL NETWORK IN PCA DOMAIN
MULTISPECTRAL IMAGE CLASSIFICATION USING BACK-PROPAGATION NEURAL NETWORK IN PCA DOMAIN S. Chtwong, S. Wtthayapradt, S. Intajag, and F. Cheevasuvt Faculty of Engneerng, Kng Mongkut s Insttute of Technology
More informationSupport Vector Machines
/14/018 Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x
More informationNumerical Heat and Mass Transfer
Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and
More informationNatural Language Processing and Information Retrieval
Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support
More informationHongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k)
ISSN 1749-3889 (prnt), 1749-3897 (onlne) Internatonal Journal of Nonlnear Scence Vol.17(2014) No.2,pp.188-192 Modfed Block Jacob-Davdson Method for Solvng Large Sparse Egenproblems Hongy Mao, College of
More informationResearch Article Green s Theorem for Sign Data
Internatonal Scholarly Research Network ISRN Appled Mathematcs Volume 2012, Artcle ID 539359, 10 pages do:10.5402/2012/539359 Research Artcle Green s Theorem for Sgn Data Lous M. Houston The Unversty of
More informationA kernel method for canonical correlation analysis
A kernel method for canoncal correlaton analyss Shotaro Akaho AIST Neuroscence Research Insttute, Central 2, - Umezono, Tsukuba, Ibarak 3058568, Japan s.akaho@ast.go.jp http://staff.ast.go.jp/s.akaho/
More informationPhysics 5153 Classical Mechanics. Principle of Virtual Work-1
P. Guterrez 1 Introducton Physcs 5153 Classcal Mechancs Prncple of Vrtual Work The frst varatonal prncple we encounter n mechancs s the prncple of vrtual work. It establshes the equlbrum condton of a mechancal
More information14 Lagrange Multipliers
Lagrange Multplers 14 Lagrange Multplers The Method of Lagrange Multplers s a powerful technque for constraned optmzaton. Whle t has applcatons far beyond machne learnng t was orgnally developed to solve
More informationThe Minimum Universal Cost Flow in an Infeasible Flow Network
Journal of Scences, Islamc Republc of Iran 17(2): 175-180 (2006) Unversty of Tehran, ISSN 1016-1104 http://jscencesutacr The Mnmum Unversal Cost Flow n an Infeasble Flow Network H Saleh Fathabad * M Bagheran
More informationA Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach
A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland
More informationEEE 241: Linear Systems
EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they
More informationSingular Value Decomposition: Theory and Applications
Sngular Value Decomposton: Theory and Applcatons Danel Khashab Sprng 2015 Last Update: March 2, 2015 1 Introducton A = UDV where columns of U and V are orthonormal and matrx D s dagonal wth postve real
More informationINF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018
INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton
More informationThe Prncpal Component Transform The Prncpal Component Transform s also called Karhunen-Loeve Transform (KLT, Hotellng Transform, oregenvector Transfor
Prncpal Component Transform Multvarate Random Sgnals A real tme sgnal x(t can be consdered as a random process and ts samples x m (m =0; ;N, 1 a random vector: The mean vector of X s X =[x0; ;x N,1] T
More informationRelevance Vector Machines Explained
October 19, 2010 Relevance Vector Machnes Explaned Trstan Fletcher www.cs.ucl.ac.uk/staff/t.fletcher/ Introducton Ths document has been wrtten n an attempt to make Tppng s [1] Relevance Vector Machnes
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationChapter 11: Simple Linear Regression and Correlation
Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests
More informationCHAPTER III Neural Networks as Associative Memory
CHAPTER III Neural Networs as Assocatve Memory Introducton One of the prmary functons of the bran s assocatve memory. We assocate the faces wth names, letters wth sounds, or we can recognze the people
More informationAPPENDIX A Some Linear Algebra
APPENDIX A Some Lnear Algebra The collecton of m, n matrces A.1 Matrces a 1,1,..., a 1,n A = a m,1,..., a m,n wth real elements a,j s denoted by R m,n. If n = 1 then A s called a column vector. Smlarly,
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationx = , so that calculated
Stat 4, secton Sngle Factor ANOVA notes by Tm Plachowsk n chapter 8 we conducted hypothess tests n whch we compared a sngle sample s mean or proporton to some hypotheszed value Chapter 9 expanded ths to
More informationSemi-supervised Classification with Active Query Selection
Sem-supervsed Classfcaton wth Actve Query Selecton Jao Wang and Swe Luo School of Computer and Informaton Technology, Beng Jaotong Unversty, Beng 00044, Chna Wangjao088@63.com Abstract. Labeled samples
More informationTensor Subspace Analysis
Tensor Subspace Analyss Xaofe He 1 Deng Ca Partha Nyog 1 1 Department of Computer Scence, Unversty of Chcago {xaofe, nyog}@cs.uchcago.edu Department of Computer Scence, Unversty of Illnos at Urbana-Champagn
More informationImage classification. Given the bag-of-features representations of images from different classes, how do we learn a model for distinguishing i them?
Image classfcaton Gven te bag-of-features representatons of mages from dfferent classes ow do we learn a model for dstngusng tem? Classfers Learn a decson rule assgnng bag-offeatures representatons of
More informationReport on Image warping
Report on Image warpng Xuan Ne, Dec. 20, 2004 Ths document summarzed the algorthms of our mage warpng soluton for further study, and there s a detaled descrpton about the mplementaton of these algorthms.
More informationLECTURE 9 CANONICAL CORRELATION ANALYSIS
LECURE 9 CANONICAL CORRELAION ANALYSIS Introducton he concept of canoncal correlaton arses when we want to quantfy the assocatons between two sets of varables. For example, suppose that the frst set of
More information1 Convex Optimization
Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We
More informationChapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems
Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons
More informationCOS 521: Advanced Algorithms Game Theory and Linear Programming
COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton
More informationThe exam is closed book, closed notes except your one-page cheat sheet.
CS 89 Fall 206 Introducton to Machne Learnng Fnal Do not open the exam before you are nstructed to do so The exam s closed book, closed notes except your one-page cheat sheet Usage of electronc devces
More informationLecture 10: Dimensionality reduction
Lecture : Dmensonalt reducton g The curse of dmensonalt g Feature etracton s. feature selecton g Prncpal Components Analss g Lnear Dscrmnant Analss Intellgent Sensor Sstems Rcardo Guterrez-Osuna Wrght
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? + + + + + + + + + Intuton of Margn Consder ponts
More informationAdvanced Introduction to Machine Learning
Advanced Introducton to Machne Learnng 10715, Fall 2014 The Kernel Trck, Reproducng Kernel Hlbert Space, and the Representer Theorem Erc Xng Lecture 6, September 24, 2014 Readng: Erc Xng @ CMU, 2014 1
More informationCS 468 Lecture 16: Isometry Invariance and Spectral Techniques
CS 468 Lecture 16: Isometry Invarance and Spectral Technques Justn Solomon Scrbe: Evan Gawlk Introducton. In geometry processng, t s often desrable to characterze the shape of an object n a manner that
More informationSubspace Learning Based on Tensor Analysis. by Deng Cai, Xiaofei He, and Jiawei Han
Report No. UIUCDCS-R-2005-2572 UILU-ENG-2005-1767 Subspace Learnng Based on Tensor Analyss by Deng Ca, Xaofe He, and Jawe Han May 2005 Subspace Learnng Based on Tensor Analyss Deng Ca Xaofe He Jawe Han
More informationChapter 6 Support vector machine. Séparateurs à vaste marge
Chapter 6 Support vector machne Séparateurs à vaste marge Méthode de classfcaton bnare par apprentssage Introdute par Vladmr Vapnk en 1995 Repose sur l exstence d un classfcateur lnéare Apprentssage supervsé
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationSupport Vector Machines
CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at
More informationTransfer Functions. Convenient representation of a linear, dynamic model. A transfer function (TF) relates one input and one output: ( ) system
Transfer Functons Convenent representaton of a lnear, dynamc model. A transfer functon (TF) relates one nput and one output: x t X s y t system Y s The followng termnology s used: x y nput output forcng
More informationGlobal Sensitivity. Tuesday 20 th February, 2018
Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values
More informationC4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )
C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z
More informationInexact Newton Methods for Inverse Eigenvalue Problems
Inexact Newton Methods for Inverse Egenvalue Problems Zheng-jan Ba Abstract In ths paper, we survey some of the latest development n usng nexact Newton-lke methods for solvng nverse egenvalue problems.
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 31 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 6. Rdge regresson The OLSE s the best lnear unbased
More informationNUMERICAL DIFFERENTIATION
NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the
More informationPsychology 282 Lecture #24 Outline Regression Diagnostics: Outliers
Psychology 282 Lecture #24 Outlne Regresson Dagnostcs: Outlers In an earler lecture we studed the statstcal assumptons underlyng the regresson model, ncludng the followng ponts: Formal statement of assumptons.
More informationarxiv:cs.cv/ Jun 2000
Correlaton over Decomposed Sgnals: A Non-Lnear Approach to Fast and Effectve Sequences Comparson Lucano da Fontoura Costa arxv:cs.cv/0006040 28 Jun 2000 Cybernetc Vson Research Group IFSC Unversty of São
More informationPARTICIPATION FACTOR IN MODAL ANALYSIS OF POWER SYSTEMS STABILITY
POZNAN UNIVE RSITY OF TE CHNOLOGY ACADE MIC JOURNALS No 86 Electrcal Engneerng 6 Volodymyr KONOVAL* Roman PRYTULA** PARTICIPATION FACTOR IN MODAL ANALYSIS OF POWER SYSTEMS STABILITY Ths paper provdes a
More informationISSN: ISO 9001:2008 Certified International Journal of Engineering and Innovative Technology (IJEIT) Volume 3, Issue 1, July 2013
ISSN: 2277-375 Constructon of Trend Free Run Orders for Orthogonal rrays Usng Codes bstract: Sometmes when the expermental runs are carred out n a tme order sequence, the response can depend on the run
More informationEcon107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)
I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes
More informationEstimating the Fundamental Matrix by Transforming Image Points in Projective Space 1
Estmatng the Fundamental Matrx by Transformng Image Ponts n Projectve Space 1 Zhengyou Zhang and Charles Loop Mcrosoft Research, One Mcrosoft Way, Redmond, WA 98052, USA E-mal: fzhang,cloopg@mcrosoft.com
More informationLecture 3: Dual problems and Kernels
Lecture 3: Dual problems and Kernels C4B Machne Learnng Hlary 211 A. Zsserman Prmal and dual forms Lnear separablty revsted Feature mappng Kernels for SVMs Kernel trck requrements radal bass functons SVM
More informationEfficient, General Point Cloud Registration with Kernel Feature Maps
Effcent, General Pont Cloud Regstraton wth Kernel Feature Maps Hanchen Xong, Sandor Szedmak, Justus Pater Insttute of Computer Scence Unversty of Innsbruck 30 May 2013 Hanchen Xong (Un.Innsbruck) 3D Regstraton
More informationChapter 12 Analysis of Covariance
Chapter Analyss of Covarance Any scentfc experment s performed to know somethng that s unknown about a group of treatments and to test certan hypothess about the correspondng treatment effect When varablty
More informationComposite Hypotheses testing
Composte ypotheses testng In many hypothess testng problems there are many possble dstrbutons that can occur under each of the hypotheses. The output of the source s a set of parameters (ponts n a parameter
More informationVQ widely used in coding speech, image, and video
at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng
More informationPattern Recognition 42 (2009) Contents lists available at ScienceDirect. Pattern Recognition. journal homepage:
Pattern Recognton 4 (9) 764 -- 779 Contents lsts avalable at ScenceDrect Pattern Recognton ournal homepage: www.elsever.com/locate/pr Perturbaton LDA: Learnng the dfference between the class emprcal mean
More informationUsing T.O.M to Estimate Parameter of distributions that have not Single Exponential Family
IOSR Journal of Mathematcs IOSR-JM) ISSN: 2278-5728. Volume 3, Issue 3 Sep-Oct. 202), PP 44-48 www.osrjournals.org Usng T.O.M to Estmate Parameter of dstrbutons that have not Sngle Exponental Famly Jubran
More informationECE559VV Project Report
ECE559VV Project Report (Supplementary Notes Loc Xuan Bu I. MAX SUM-RATE SCHEDULING: THE UPLINK CASE We have seen (n the presentaton that, for downlnk (broadcast channels, the strategy maxmzng the sum-rate
More informationDr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur
Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed
More informationComparative Studies of Law of Conservation of Energy. and Law Clusters of Conservation of Generalized Energy
Comparatve Studes of Law of Conservaton of Energy and Law Clusters of Conservaton of Generalzed Energy No.3 of Comparatve Physcs Seres Papers Fu Yuhua (CNOOC Research Insttute, E-mal:fuyh1945@sna.com)
More informationLogistic Regression. CAP 5610: Machine Learning Instructor: Guo-Jun QI
Logstc Regresson CAP 561: achne Learnng Instructor: Guo-Jun QI Bayes Classfer: A Generatve model odel the posteror dstrbuton P(Y X) Estmate class-condtonal dstrbuton P(X Y) for each Y Estmate pror dstrbuton
More informationn α j x j = 0 j=1 has a nontrivial solution. Here A is the n k matrix whose jth column is the vector for all t j=0
MODULE 2 Topcs: Lnear ndependence, bass and dmenson We have seen that f n a set of vectors one vector s a lnear combnaton of the remanng vectors n the set then the span of the set s unchanged f that vector
More informationSome Comments on Accelerating Convergence of Iterative Sequences Using Direct Inversion of the Iterative Subspace (DIIS)
Some Comments on Acceleratng Convergence of Iteratve Sequences Usng Drect Inverson of the Iteratve Subspace (DIIS) C. Davd Sherrll School of Chemstry and Bochemstry Georga Insttute of Technology May 1998
More informationLecture 12: Discrete Laplacian
Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly
More informationLaboratory 3: Method of Least Squares
Laboratory 3: Method of Least Squares Introducton Consder the graph of expermental data n Fgure 1. In ths experment x s the ndependent varable and y the dependent varable. Clearly they are correlated wth
More informationLecture 3. Ax x i a i. i i
18.409 The Behavor of Algorthms n Practce 2/14/2 Lecturer: Dan Spelman Lecture 3 Scrbe: Arvnd Sankar 1 Largest sngular value In order to bound the condton number, we need an upper bound on the largest
More informationNumber of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k
ANOVA Model and Matrx Computatons Notaton The followng notaton s used throughout ths chapter unless otherwse stated: N F CN Y Z j w W Number of cases Number of factors Number of covarates Number of levels
More information