Using Support Vector Machines to Enhance the Performance of X-Ray Diffraction Data Analysis in Crystalline Materials Cubic Structure Identification
|
|
- Theodora Murphy
- 5 years ago
- Views:
Transcription
1 194 IJCSNS Internatonal Journal of Computer Scence and Network Securty, VOL.7 No.7, July 007 Usng Support Vector Machnes to Enhance the Performance of X-Ray Dffracton Data Analyss n Crystallne Materals Cubc Structure Identfcaton Mohammad Syukur Faculty of Mathematcs and Natural Scences, Unversty of Sumatera Utara, 0155 Medan, Sumut, Indonesa Summary Crystallne materals cubc structure dentfcaton s very mportant n crystallography and materal scence research. For a long tme researchers n the feld have used manual approach n matchng the result data from X-Ray Dffracton (XRD) method wth the known fngerprnt. These manual matchng processes are complcated and sometmes are tedous because the dffracted data are complex and may have more than one fngerprnt nsde. Ths paper proposes the use of support vector machnes to enhance the performance of the matchng process between the dffracted data of crystallne materal and the fngerprnts. It s demonstrated, through experments, that support vector machnes gves more accurate and relable dentfcaton results compared to the use of neural network. Key words: Support Vector Machne, X-Ray Dffracton Data Analyss, Cubc Structure Identfcaton, Materal Scences, Artfcal Intellgence Applcaton. 1. Background Crystallography s one of the areas of research n physcs that deals wth the scentfc study of crystals. It has always been one of the most challengng research felds snce eghteenth century. The sgnfcant dscovery of X- ray by Röntgen n 1895 [8] had yelded to a new way of dong crystallography research. Snce then X-ray dffracton method has been proposed and appled to many dfferent sub-area of crystallography such as dentfcaton of crystallne phases, qualtatve and quanttatve analyss of mxtures and mnor consttuents, dstncton between crystallne and amorphous states, sde-chan packng of proten structures, dentfcaton of crystallne materal, etc. The last area s the focus of ths paper. X-ray dffracton data nterpretaton for most crystallne materals s a very complex and dffcult task. Ths s due to the condton that dfferent crystallne materal may contan more than one cubc structures component type and after beng dffracted usng X-ray dffracton method, the dffracted data are complex. Hence, the data can be very ambguous and s not easy to track and understand. Numerous artfcal ntellgence technques and applcaton have been appled and developed to solve the problems n varous domans. In our prevous work [5], an attempted has been made to use neural network to perform automatc cubc structure dentfcaton on the crystallne materals. Though t was a success, there s stll left room for mprovement. Ths paper proposes the use of support vector machne (SVM) to enhance the performance of crystallne materals cubc structure dentfcaton. The result of usng SVM s compared wth the result usng neural network. Support vector machnes have been proven to be a powerful method to solve dentfcaton and classfcaton problems. It ncludes gene dentfcaton [10], paraphrase dentfcaton [11], proten classfcaton usng X-ray crystallography [9], and many more. The man ntent of ths paper s to showcase the superor results on the use of support vector machne over neural network n crystallne materals cubc structure dentfcaton.. Cubc Structure Identfcaton usng X-Ray Dffracton Data In prncple, there are four cubc structures type for crystallne materals, the Smple Cubc (SC), Body Centered Cubc (BCC), Face Centered Cubc (FCC) and Damond [1]. In our prevous work [], a formula has been proposed to calculate the fngerprnts for these four cubc structures. The formula utlzes the Mller ndex (h,k,l) [6]. The proposed formula can be wrtten as follows: ( h + k + l ) sn θ = λ 4a (1) Snce the wavelength of the ncomng X-ray (λ) and lattce constant (a) are both constants, we can elmnate these Manuscrpt receved July 5, 007 Manuscrpt revsed July 5, 007
2 IJCSNS Internatonal Journal of Computer Scence and Network Securty, VOL.7 No.7, July quanttes from Eq. 1 and derve the rato of two sn θ as follows: sn θ h + k + l m m m m R = = m, n sn θ h + k + l n n n n () where θ m and θ n are the dffractng angles for two peak assocated wth the dffractng planes {h m,k m,l m } and {h n,k n,l n } respectvely for rato R m,n. The fngerprnt for each crystallne materal s cubc structure s calculated by takng 10 peaks from the X-ray dffracton data and calculates the two peaks combnatons from that 10 peaks. That s, each fngerprnt s actually contans 45 values of the quadratc snus rato from Eq. C 10 snce equal to 45. Table 1,, 3 and 4 depct the fngerprnt for Face Centered Cubc (FCC), Damond, Body Centered Cubc (BCC) and Smple Cubc (SC) respectvely. R 1, = R 1, = Table 1: Face Centered Cubc (FCC) Fngerprnt R,3 = R,4 = R,5 = R,6 = 0.50 R,7 = R 3,9 = R,8 = R 3,10 = R,9 = R 5,9 = R,10 = R 5,10 = Table : Damond Fngerprnt R,3 = R,4 = R,5 = R,6 = R,7 = R 3,9 = R,8 = R 3,10 = R,9 = R 5,9 = R,10 = R 5,10 = R 1, = R 1, = Table 3: Body Centered Cubc (BCC) Fngerprnt R,3 = R,4 = R,5 = 0.48 R,6 = R,7 = R 3,9 = R,8 = R 3,10 = R,9 = R 5,9 = R,10 = R 5,10 = Table 4: Smple Cubc (SC) Fngerprnt R,3 = R,4 = R,5 = R,6 = R,7 = R 3,9 = R,8 = R 3,10 = R,9 = R 5,9 = 0.00 R,10 = R 5,10 = Support Vector Machne Support Vector Machne (SVM) can be regarded as an excellent statstcal learnng performance and superor classfcaton performance. Smply put, the support vector machne that we used n ths paper can be summarzed as follows: t dvdes two specfed tranng samples whch belong to two dfferent categores through constructng an optmal separatng hyperplane ether n the orgnal space or n the mapped hgher dmensonal space [4]. The basc dea of constructng ths optmal separatng hyperplane s to guarantee maxmum dstance between each tranng sample and the separatng hyperplane. The SVM algorthm and ts learnng procedure can be comprehended as follows: If the data are lnearly separable n nput space, a bnary classfcaton task s taken nto account. Let {( x, y )}(1 N) be a lnearly separable set, where
3 196 IJCSNS Internatonal Journal of Computer Scence and Network Securty, VOL.7 No.7, July 007 d x R, y { 1,1} and y are labels of categores. The general expresson of the lnear dscrmnaton functon n d-dmenson space s defned as g ( x) = w x + b, and the correspondng equaton of separatng hyperplane s w x + b = 0 where w s normal to the hyperplane, b / w s the perpendcular dstance from the hyperplane to the orgn, and w s the Eucldean norm of w. Further normalze g(x) and let all the x meet g(x) 1 wll results to the samples whch are closest to optmal separatng hyperplane that meet g(x) = 1. Hence, the separatng nterval s equal to / w and solvng the optmal separatng hyperplane tself s equvalent to mnmzng w. The objectve functon used for ths s as follows: 1 mn Φ ( w ) = w (3) subject to the constrants: y ( w x + b) 1, = 1,..., N (4) The constrants n Eq. 4 can be replaced by the Lagrange multplers of the Lagrangan algorthm to further solved the constrants problem where w = α y x and x are the samples only appearng n the separatng nterval planes. These samples are called support vectors and ts classfcaton functon s defned as follows: f ( x) = sgn α y x x + b (5) If data are not lnearly separable n the nput space, smply use the followng objectve functon: N 1 mn Φ( w, ξ ) = w + C ξ (6) = 1 Where, ξ s slack varable and C s a penalty factor. Smultaneously, through a non-lnear transform Φ ( ), the nput space s mapped nto a hgher dmensonal space called feature space n whch optmal separatng hyperplane can be solved. In addton, the nner product calculaton s changed nto K( x, x j ) = Φ( x ) Φ( x j ) where K ( x, x j ) kernel functon s defned as nner product n Hlbert space. Thus the fnal classfcaton functon can be represented as follows: 4. Experments As mentoned earler, the man ntent of ths paper s to showcase the superor results on the use of support vector machne over neural network n crystallne materals cubc structure dentfcaton. In our prevous work [5], we had appled a back-propagaton neural network to dentfy the cubc structure of three samples, Alumnum (Al), Slcon (S), and a mxture of Al and S. In ths experment, more samples are used n addton to that three samples such as Iron (Fe), Tungsten (W), Sodumchlorde (NaCl), and CopperZnc (CuZn). From the new samples, a mxture sample s also created from three samples Al, S, and Fe to evaluate the performance of support vector machne n dentfyng a more complex pattern data. The experments are carred out usng the same equpments and envronment settngs as wth our prevous work whch s usng a Phlps X-Ray devce dffractometer control PW1710. The devce was also usng PW179 seres of X-Ray generator, anode Cu tube, and PCAPD (PW1877) software verson 3 ntegrated wth our proposed applcaton nstalled. In the overall experments, we use the same tube voltage and also the same tube current,.e. 30kV and 0mA respectvely. The SVMlght package [3] was used to construct the support vector machne (SVM) classfers. For a gven set of bnary-labeled tranng examples, SVM maps the nput space nto a hgher dmensonal feature space and seeks a hyperplane n the feature space to separate the postve data nstances from the negatve ones. Dfferent values for the γ and C parameters were tested to optmze the performance of SVM to dentfy the cubc structures from the presented crystallne materal s sample. Snce the fngerprnt tranng dataset was mbalanced, the cost factor was set to 5.8 for gvng more weght to tranng errors on postve examples than errors on negatve ones. In the end, after expermentng several dfferent value for C and γ, we use γ = and C = 0.55 as the SVM parameter n ths experment. Whle the rest of the parameters were set to ther default values as specfed n SVMlght package. As for the neural network, we kept the same settng from our prevous work [5] for comparson wse. That s, the neural network was constructed wth 8 nodes n the hdden layer, 5 n the frst hdden layer and 3 n the second hdden layer. It uses standard teratve gradent algorthm for back-propagaton tranng. f ( x) = sgn α y K( x, x ) x + b (7)
4 IJCSNS Internatonal Journal of Computer Scence and Network Securty, VOL.7 No.7, July Analyss & Results After feedng all the samples nto the X-Ray dffractometer, the resulted dffracton data for each sample and ts rato R m,n values are shown on Table 5, 6, 7, 8 and 9 for Iron (Fe), Tungsten (W), Sodumchlorde (NaCl), CopperZnc (CuZn), and the mxture of Al, S, and Fe respectvely. Whle the R m,n rato values for Al, S and mxture of Al and S are avalable at [5]. Table 5: The resulted R m,n rato values for the sample of Iron (Fe) R 1, = R,4 = R,5 = R,3 = R,6 = Table 6: The resulted R m,n rato values for the sample of Tungsten (W) R 1, = R,3 = R,4 = 0.50 R,5 = R,6 = 0.65 R,7 = R,8 = Table 7: The resulted R m,n rato values for the sample of Sodumchlorde (NaCl) R 1, = R,3 = R,4 = R 4,9 = 0.73 R,5 = 0.41 R 4,10 = R,6 = R,7 = 0.10 R 3,9 = R,8 = 0.00 R 3,10 = R,9 = R 5,9 = 0.15 R,10 = R 5,10 = Table 8: The resulted R m,n rato values for the sample of CopperZnc (CuZn) R 1, = R,3 = R,4 = R 4,9 = R,5 = R 4,10 = R 6, 10 = R,6 = R,7 = 0.51 R 3,9 = R,8 = 0.3 R 3,10 = R 7, 10 = 0.76 R,9 = 0.01 R 5,9 = R, 10 = 0.18 R 5, 10 = Table 9: The resulted R m,n rato values for the mxture sample of Alumnum (Al), Slcon (S), and Iron (Fe) R 1, = R,3 = R,4 = R,5 = R,6 = R 5,9 = R,7 = R 4,9 = R 5,10 = R,8 = 0.86 R 3,9 = R 4,10 = R 5,11 = R,9 = 0.75 R 3,10 = R 4,11 = R 5,1 = R,10 = 0.50 R 3,11 = 0.99 R 4,1 = 0.96 R 5,13 = R,11 = 0.6 R 3,1 = 0.66 R 4,13 = 0.77 R 5,14 = R 1, 11 = 0.15 R,1 = 0.01 R 3,13 = 0.50 R 4,14 = 0.50 R 5,15 = 0.31 R 1,1 = R,13 = R 3,14 = 0.4 R 4,15 = 0.34 R 5,16 = R 1,13 = R,14 = R 3,15 = 0.10 R 4,16 = 0.8 R 5,17 = R 1,14 = R,15 = R 3,16 = 0.05 R 4,17 = 0. R 5,18 = 0.75 R 1,15 = R,16 = R 3,17 = 0.00 R 4,18 = 0.00 R 5,19 = 0.50 R 1,16 = R,17 = R 3,18 = R 4,19 = R 10,11 = R 1,17 = R,18 = R 3,19 = R 10,1 = R 1,18 = R,19 = R 9,11 = 0.83 R 10,13 = R 1,19 = R 9,1 = 0.73 R 10,14 = R 8,11 = 0.79 R 9,13 = R 10,15 = R 8,1 = R 9,14 = R 10,16 = 0.615
5 198 IJCSNS Internatonal Journal of Computer Scence and Network Securty, VOL.7 No.7, July R 6,11 = R 6,1 = R 6,13 = R 6,14 = 0.45 R 6,15 = 0.41 R 6,16 = R 6,17 = R 6,18 = R 6,19 = R 11,1 = R 11,13 = R 11,14 = R 11,15 = 0.70 R 11,16 = R 11,17 = R 11,18 = R 11,19 = R 7,11 = R 7,1 = 0.59 R 7,13 = R 7,14 = R 7,15 = R 7,16 = R 7,17 = R 7,18 = R 7,19 = R 1,13 = R 1,14 = 0.84 R 1,15 = R 1,16 = R 1,17 = 0.75 R 1,18 = R 1,19 = 0.65 R 8,13 = R 8,14 = R 8,15 = R 8,16 = 0.54 R 8,17 = 0.59 R 8,18 = R 8,19 = R 13,14 = R 13,15 = 0.84 R 13,16 = 0.81 R 13,17 = R 13,18 = R 13,19 = R 14,15 = R 14,16 = R 14,17 = R 14,18 = R 14,19 = 0.74 R 9,15 = R 9,16 = R 9,17 = R 9,18 = R 9, 19 = R 15,16 = R 15,17 = R 15,18 = R 15,19 = 0.79 R 16,17 = R 16,18 = R 16,19 = 0.81 R 17,18 = R 17,19 = R 18,19 = 0.97 R 10,17 = R 10,18 = R 10,19 = Table 10: SVM and Neural Network Comparson Results SVM Neural Network Sample Name Type Accuracy Type Accuracy Al + S Al + S + Fe FCC 91% FCC 86% Damond 9% Damond 84% FCC 90% FCC 78% Damond 8% Damond 67% BCC 73% SC 51% Al FCC 93% FCC 9% S Damond 9% Damond 91% Fe BCC 89% BCC 78% W BCC 91% BCC 89% NaCl FCC 94% FCC 9% CuZn SC 90% SC 87% The last step s to pass the resulted data for each crystallne materal sample nto the traned SVM and neural network. Table 10 shows the comparson results of SVM and neural network. From result outlned on Table 10, we can clearly see that support vector machne (SVM) outperform neural network n terms of the accuracy of dentfyng the cubc structure type from each crystallne materal sample. Both SVM and neural network successfully detect the correct cubc structure type for sngle component crystallne materal, but neural network suffers and faled to detect the correct cubc structure type for mult component crystallne materal especally the one wth three cubc structure components nsde. In the experment wth the mxture sample of Alumnum (Al), Slcon (S), and Iron (Fe), neural network dentfy that the sample contan three cubc structure component of FCC, Damond and SC wth accuracy rate of 78%, 67%, and 51% respectvely. Snce the cubc structure type of Iron (Fe) s BCC, the result gven by neural network s wrong. Whle SVM gves more accurate result of FCC, Damond and BCC wth the accuracy rate of 90%, 8%, and 73% respectvely. Neural network falure s probably due to the fact that the fngerprnt structure of BCC and SC s almost the same and among the mxture sample component, Iron (Fe) s the crystallne materal that has the smallest number of R m,n rato values. And after beng mxed wth the two other components, t makes ts rato pattern dffcult to track. SVM results also outperform the neural network n the mxture of two component of Alumnum (Al) and Slcon (S). Though both successfully detect the correct cubc structure types (FCC and damond), but the accuracy of SVM s slghtly hgher, 91% and 9% respectvely for FCC and damond. As for sngle component crystallne materal samples, both SVM and neural network gves equvalent results. 6. Summary In ths paper we propose the use of support vector machne to enhance the performance of cubc structures dentfcaton on mult component crystallne materal. The complexty of mult component crystallne materal R m,n rato needs a sophstcated and powerful methods to accurately classfy ts cubc structure type. And support vector machne whch exhbts more excellent performances such as no local optmum problem, no overft or under-ft problem, better convergence property, less tranng samples, hgher correct dentfcaton rate, and hgher relablty suts ths job.
6 IJCSNS Internatonal Journal of Computer Scence and Network Securty, VOL.7 No.7, July Ths work proves that the use of artfcal ntellgence technques such as support vector machne and neural network towards crystallne materals cubc structure dentfcaton research to be very useful and can smplfy and fasten up the work on crystallography research area. Our future research drecton ncludes applyng these artfcal ntellgence technques to other problems n crystallography research area. Acknowledgments Author would lke to acknowledge Unversty of Sumatera Utara for provdng space and the equpments used n ths research. Also, author would lke to thank Muhammad Ferm Pasha for the frutful dscusson and nsghtful comments on the support vector machne part. Mohammad Syukur receved the B.Sc. degree n Physcs from Unversty of Sumatera Utara, Indonesa and M.Sc. degrees n Materals Physcs from Bandung Insttute of Technology, Indonesa n 1974 and 198, respectvely. He s currently a senor lecturer and the head of Crystallography & Radaton Physcs Research Laboratory at Faculty of Mathematcs and Natural Scences, Unversty of Sumatera Utara, Indonesa. Hs research nterests nclude Materal Physcs, X-Ray Dffracton methods, Artfcal Intellgence Applcaton and Crystallography. He has wrtten two book chapters and numerous refereed research papers. References [1] Suryanarayana, C. and Norton, M.G. X-Ray Dffracton - A Practcal Approach. Plenum Press - NewYork, [] Syukur, M. and Zarls, M. A Standard Calculaton for Lattce Analyss on Cubc Structures Sold Materals usng X-Ray Dffracton Method and ts Expermental Study. Journal of Sold State Scence and Technology Letter, Vol. 6, No., [3] Joachms, T. Makng large-scale SVM Learnng Practcal. Advances n Kernel Methods - Support Vector Learnng, B. Schölkopf and C. Burges and A. Smola (Eds.), MIT Press, [4] Vapnk, V.N. Statstcal Learnng Theory. New York: John Wley and Sons, [5] Syukur, M., Pasha, M.F. and Budarto, R. A Neural Network-Based Applcaton to Identfy Cubc Structures n Mult Component Crystallne Materals usng X-Ray Dffracton Data. Internatonal Journal of Computer Scence and Network Securty (IJCSNS), 7(): 49-54, 007. [6] Cullty, B.D. Elements of X-ray Dffracton. Addson- Wesley - London, [7] The Internatonal Center for Dffracton Data, [8] Ewald, P.P. Ffty Years of X-Ray Dffracton. Internatonal Unon of Crystallography, [9] Gubb, J., Shlton, A., Parker, M. and Palanswam, M. Proten Topology Classfcaton usng Two-Stage Support Vector Machnes. Genome Informatcs Vol.17, No., pp , 006. [10] Krause, L., McHardy, A.C., Nattkemper, T.W., Pühler, A., Stoye, J. and Meyer, F. GISMO - gene dentfcaton usng a support vector machne for ORF classfcaton. Nuclec Acds Research, 35(): , 006. [11] Brockett, C. and Dolan, W.B. Support vector machnes for paraphrase dentfcaton and corpus constructon. Proceedngs of The 3rd Internatonal Workshop on Paraphrasng (IWP005), 005.
Kernel Methods and SVMs Extension
Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general
More informationSupport Vector Machines. Vibhav Gogate The University of Texas at dallas
Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest
More informationMultilayer Perceptron (MLP)
Multlayer Perceptron (MLP) Seungjn Cho Department of Computer Scence and Engneerng Pohang Unversty of Scence and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjn@postech.ac.kr 1 / 20 Outlne
More informationWhich Separator? Spring 1
Whch Separator? 6.034 - Sprng 1 Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng Whch Separator? Mamze the margn to closest ponts 6.034 - Sprng 3 Margn of a pont " # y (w $ + b) proportonal
More information1 Convex Optimization
Convex Optmzaton We wll consder convex optmzaton problems. Namely, mnmzaton problems where the objectve s convex (we assume no constrants for now). Such problems often arse n machne learnng. For example,
More informationLecture 10 Support Vector Machines II
Lecture 10 Support Vector Machnes II 22 February 2016 Taylor B. Arnold Yale Statstcs STAT 365/665 1/28 Notes: Problem 3 s posted and due ths upcomng Frday There was an early bug n the fake-test data; fxed
More informationNatural Language Processing and Information Retrieval
Natural Language Processng and Informaton Retreval Support Vector Machnes Alessandro Moschtt Department of nformaton and communcaton technology Unversty of Trento Emal: moschtt@ds.untn.t Summary Support
More informationFor now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.
Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson
More informationMaximal Margin Classifier
CS81B/Stat41B: Advanced Topcs n Learnng & Decson Makng Mamal Margn Classfer Lecturer: Mchael Jordan Scrbes: Jana van Greunen Corrected verson - /1/004 1 References/Recommended Readng 1.1 Webstes www.kernel-machnes.org
More informationTHE SUMMATION NOTATION Ʃ
Sngle Subscrpt otaton THE SUMMATIO OTATIO Ʃ Most of the calculatons we perform n statstcs are repettve operatons on lsts of numbers. For example, we compute the sum of a set of numbers, or the sum of the
More informationCS 3710: Visual Recognition Classification and Detection. Adriana Kovashka Department of Computer Science January 13, 2015
CS 3710: Vsual Recognton Classfcaton and Detecton Adrana Kovashka Department of Computer Scence January 13, 2015 Plan for Today Vsual recognton bascs part 2: Classfcaton and detecton Adrana s research
More informationProblem Set 9 Solutions
Desgn and Analyss of Algorthms May 4, 2015 Massachusetts Insttute of Technology 6.046J/18.410J Profs. Erk Demane, Srn Devadas, and Nancy Lynch Problem Set 9 Solutons Problem Set 9 Solutons Ths problem
More informationLecture 10 Support Vector Machines. Oct
Lecture 10 Support Vector Machnes Oct - 20-2008 Lnear Separators Whch of the lnear separators s optmal? Concept of Margn Recall that n Perceptron, we learned that the convergence rate of the Perceptron
More informationSection 8.3 Polar Form of Complex Numbers
80 Chapter 8 Secton 8 Polar Form of Complex Numbers From prevous classes, you may have encountered magnary numbers the square roots of negatve numbers and, more generally, complex numbers whch are the
More informationKristin P. Bennett. Rensselaer Polytechnic Institute
Support Vector Machnes and Other Kernel Methods Krstn P. Bennett Mathematcal Scences Department Rensselaer Polytechnc Insttute Support Vector Machnes (SVM) A methodology for nference based on Statstcal
More informationThe Expectation-Maximization Algorithm
The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.
More informationKernels in Support Vector Machines. Based on lectures of Martin Law, University of Michigan
Kernels n Support Vector Machnes Based on lectures of Martn Law, Unversty of Mchgan Non Lnear separable problems AND OR NOT() The XOR problem cannot be solved wth a perceptron. XOR Per Lug Martell - Systems
More informationMMA and GCMMA two methods for nonlinear optimization
MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons
More informationPattern Classification
Pattern Classfcaton All materals n these sldes ere taken from Pattern Classfcaton (nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wley & Sons, 000 th the permsson of the authors and the publsher
More informationChapter 6 Support vector machine. Séparateurs à vaste marge
Chapter 6 Support vector machne Séparateurs à vaste marge Méthode de classfcaton bnare par apprentssage Introdute par Vladmr Vapnk en 1995 Repose sur l exstence d un classfcateur lnéare Apprentssage supervsé
More informationSupport Vector Machines
CS 2750: Machne Learnng Support Vector Machnes Prof. Adrana Kovashka Unversty of Pttsburgh February 17, 2016 Announcement Homework 2 deadlne s now 2/29 We ll have covered everythng you need today or at
More informationLecture 12: Classification
Lecture : Classfcaton g Dscrmnant functons g The optmal Bayes classfer g Quadratc classfers g Eucldean and Mahalanobs metrcs g K Nearest Neghbor Classfers Intellgent Sensor Systems Rcardo Guterrez-Osuna
More informationOnline Classification: Perceptron and Winnow
E0 370 Statstcal Learnng Theory Lecture 18 Nov 8, 011 Onlne Classfcaton: Perceptron and Wnnow Lecturer: Shvan Agarwal Scrbe: Shvan Agarwal 1 Introducton In ths lecture we wll start to study the onlne learnng
More informationUncertainty in measurements of power and energy on power networks
Uncertanty n measurements of power and energy on power networks E. Manov, N. Kolev Department of Measurement and Instrumentaton, Techncal Unversty Sofa, bul. Klment Ohrdsk No8, bl., 000 Sofa, Bulgara Tel./fax:
More informationA Hybrid Variational Iteration Method for Blasius Equation
Avalable at http://pvamu.edu/aam Appl. Appl. Math. ISSN: 1932-9466 Vol. 10, Issue 1 (June 2015), pp. 223-229 Applcatons and Appled Mathematcs: An Internatonal Journal (AAM) A Hybrd Varatonal Iteraton Method
More informationReport on Image warping
Report on Image warpng Xuan Ne, Dec. 20, 2004 Ths document summarzed the algorthms of our mage warpng soluton for further study, and there s a detaled descrpton about the mplementaton of these algorthms.
More informationEEE 241: Linear Systems
EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they
More informationSupport Vector Machines
Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x n class
More informationTemperature. Chapter Heat Engine
Chapter 3 Temperature In prevous chapters of these notes we ntroduced the Prncple of Maxmum ntropy as a technque for estmatng probablty dstrbutons consstent wth constrants. In Chapter 9 we dscussed the
More informationLecture Notes on Linear Regression
Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume
More informationU.C. Berkeley CS294: Beyond Worst-Case Analysis Luca Trevisan September 5, 2017
U.C. Berkeley CS94: Beyond Worst-Case Analyss Handout 4s Luca Trevsan September 5, 07 Summary of Lecture 4 In whch we ntroduce semdefnte programmng and apply t to Max Cut. Semdefnte Programmng Recall that
More informationA new Approach for Solving Linear Ordinary Differential Equations
, ISSN 974-57X (Onlne), ISSN 974-5718 (Prnt), Vol. ; Issue No. 1; Year 14, Copyrght 13-14 by CESER PUBLICATIONS A new Approach for Solvng Lnear Ordnary Dfferental Equatons Fawz Abdelwahd Department of
More informationGeneralized Linear Methods
Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set
More informationIntroduction to the Introduction to Artificial Neural Network
Introducton to the Introducton to Artfcal Neural Netork Vuong Le th Hao Tang s sldes Part of the content of the sldes are from the Internet (possbly th modfcatons). The lecturer does not clam any onershp
More informationWeek 5: Neural Networks
Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple
More informationCSC 411 / CSC D11 / CSC C11
18 Boostng s a general strategy for learnng classfers by combnng smpler ones. The dea of boostng s to take a weak classfer that s, any classfer that wll do at least slghtly better than chance and use t
More informationLinear Feature Engineering 11
Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19
More informationLINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity
LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have
More informationThe Geometry of Logit and Probit
The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.
More informationCSci 6974 and ECSE 6966 Math. Tech. for Vision, Graphics and Robotics Lecture 21, April 17, 2006 Estimating A Plane Homography
CSc 6974 and ECSE 6966 Math. Tech. for Vson, Graphcs and Robotcs Lecture 21, Aprl 17, 2006 Estmatng A Plane Homography Overvew We contnue wth a dscusson of the major ssues, usng estmaton of plane projectve
More informationComparative Studies of Law of Conservation of Energy. and Law Clusters of Conservation of Generalized Energy
Comparatve Studes of Law of Conservaton of Energy and Law Clusters of Conservaton of Generalzed Energy No.3 of Comparatve Physcs Seres Papers Fu Yuhua (CNOOC Research Insttute, E-mal:fuyh1945@sna.com)
More informationSupport Vector Machines
/14/018 Separatng boundary, defned by w Support Vector Machnes CISC 5800 Professor Danel Leeds Separatng hyperplane splts class 0 and class 1 Plane s defned by lne w perpendcular to plan Is data pont x
More informationarxiv:cs.cv/ Jun 2000
Correlaton over Decomposed Sgnals: A Non-Lnear Approach to Fast and Effectve Sequences Comparson Lucano da Fontoura Costa arxv:cs.cv/0006040 28 Jun 2000 Cybernetc Vson Research Group IFSC Unversty of São
More information1 Matrix representations of canonical matrices
1 Matrx representatons of canoncal matrces 2-d rotaton around the orgn: ( ) cos θ sn θ R 0 = sn θ cos θ 3-d rotaton around the x-axs: R x = 1 0 0 0 cos θ sn θ 0 sn θ cos θ 3-d rotaton around the y-axs:
More informationAnalytical Chemistry Calibration Curve Handout
I. Quck-and Drty Excel Tutoral Analytcal Chemstry Calbraton Curve Handout For those of you wth lttle experence wth Excel, I ve provded some key technques that should help you use the program both for problem
More informationLinear Classification, SVMs and Nearest Neighbors
1 CSE 473 Lecture 25 (Chapter 18) Lnear Classfcaton, SVMs and Nearest Neghbors CSE AI faculty + Chrs Bshop, Dan Klen, Stuart Russell, Andrew Moore Motvaton: Face Detecton How do we buld a classfer to dstngush
More information10-701/ Machine Learning, Fall 2005 Homework 3
10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40
More informationprinceton univ. F 17 cos 521: Advanced Algorithm Design Lecture 7: LP Duality Lecturer: Matt Weinberg
prnceton unv. F 17 cos 521: Advanced Algorthm Desgn Lecture 7: LP Dualty Lecturer: Matt Wenberg Scrbe: LP Dualty s an extremely useful tool for analyzng structural propertes of lnear programs. Whle there
More informationThe internal structure of natural numbers and one method for the definition of large prime numbers
The nternal structure of natural numbers and one method for the defnton of large prme numbers Emmanul Manousos APM Insttute for the Advancement of Physcs and Mathematcs 3 Poulou str. 53 Athens Greece Abstract
More informationSupporting Information
Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? + + + + + + + + + Intuton of Margn Consder ponts
More informationComparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method
Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method
More informationAdvanced Circuits Topics - Part 1 by Dr. Colton (Fall 2017)
Advanced rcuts Topcs - Part by Dr. olton (Fall 07) Part : Some thngs you should already know from Physcs 0 and 45 These are all thngs that you should have learned n Physcs 0 and/or 45. Ths secton s organzed
More informationSupport Vector Machines CS434
Support Vector Machnes CS434 Lnear Separators Many lnear separators exst that perfectly classfy all tranng examples Whch of the lnear separators s the best? Intuton of Margn Consder ponts A, B, and C We
More informationA Bayes Algorithm for the Multitask Pattern Recognition Problem Direct Approach
A Bayes Algorthm for the Multtask Pattern Recognton Problem Drect Approach Edward Puchala Wroclaw Unversty of Technology, Char of Systems and Computer etworks, Wybrzeze Wyspanskego 7, 50-370 Wroclaw, Poland
More informationWeek3, Chapter 4. Position and Displacement. Motion in Two Dimensions. Instantaneous Velocity. Average Velocity
Week3, Chapter 4 Moton n Two Dmensons Lecture Quz A partcle confned to moton along the x axs moves wth constant acceleraton from x =.0 m to x = 8.0 m durng a 1-s tme nterval. The velocty of the partcle
More informationFundamental loop-current method using virtual voltage sources technique for special cases
Fundamental loop-current method usng vrtual voltage sources technque for specal cases George E. Chatzaraks, 1 Marna D. Tortorel 1 and Anastasos D. Tzolas 1 Electrcal and Electroncs Engneerng Departments,
More informationSolving Nonlinear Differential Equations by a Neural Network Method
Solvng Nonlnear Dfferental Equatons by a Neural Network Method Luce P. Aarts and Peter Van der Veer Delft Unversty of Technology, Faculty of Cvlengneerng and Geoscences, Secton of Cvlengneerng Informatcs,
More informationMultigradient for Neural Networks for Equalizers 1
Multgradent for Neural Netorks for Equalzers 1 Chulhee ee, Jnook Go and Heeyoung Km Department of Electrcal and Electronc Engneerng Yonse Unversty 134 Shnchon-Dong, Seodaemun-Ku, Seoul 1-749, Korea ABSTRACT
More informationADVANCED MACHINE LEARNING ADVANCED MACHINE LEARNING
1 ADVANCED ACHINE LEARNING ADVANCED ACHINE LEARNING Non-lnear regresson technques 2 ADVANCED ACHINE LEARNING Regresson: Prncple N ap N-dm. nput x to a contnuous output y. Learn a functon of the type: N
More informationSupport Vector Machines. Jie Tang Knowledge Engineering Group Department of Computer Science and Technology Tsinghua University 2012
Support Vector Machnes Je Tang Knowledge Engneerng Group Department of Computer Scence and Technology Tsnghua Unversty 2012 1 Outlne What s a Support Vector Machne? Solvng SVMs Kernel Trcks 2 What s a
More informationCOMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS
More informationFUZZY GOAL PROGRAMMING VS ORDINARY FUZZY PROGRAMMING APPROACH FOR MULTI OBJECTIVE PROGRAMMING PROBLEM
Internatonal Conference on Ceramcs, Bkaner, Inda Internatonal Journal of Modern Physcs: Conference Seres Vol. 22 (2013) 757 761 World Scentfc Publshng Company DOI: 10.1142/S2010194513010982 FUZZY GOAL
More informationFeature Selection: Part 1
CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?
More informationCOS 521: Advanced Algorithms Game Theory and Linear Programming
COS 521: Advanced Algorthms Game Theory and Lnear Programmng Moses Charkar February 27, 2013 In these notes, we ntroduce some basc concepts n game theory and lnear programmng (LP). We show a connecton
More informationInstance-Based Learning (a.k.a. memory-based learning) Part I: Nearest Neighbor Classification
Instance-Based earnng (a.k.a. memory-based learnng) Part I: Nearest Neghbor Classfcaton Note to other teachers and users of these sldes. Andrew would be delghted f you found ths source materal useful n
More informationFoundations of Arithmetic
Foundatons of Arthmetc Notaton We shall denote the sum and product of numbers n the usual notaton as a 2 + a 2 + a 3 + + a = a, a 1 a 2 a 3 a = a The notaton a b means a dvdes b,.e. ac = b where c s an
More informationModule 9. Lecture 6. Duality in Assignment Problems
Module 9 1 Lecture 6 Dualty n Assgnment Problems In ths lecture we attempt to answer few other mportant questons posed n earler lecture for (AP) and see how some of them can be explaned through the concept
More informationLecture 12: Discrete Laplacian
Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly
More informationLagrange Multipliers Kernel Trick
Lagrange Multplers Kernel Trck Ncholas Ruozz Unversty of Texas at Dallas Based roughly on the sldes of Davd Sontag General Optmzaton A mathematcal detour, we ll come back to SVMs soon! subject to: f x
More informationPhysics 5153 Classical Mechanics. Principle of Virtual Work-1
P. Guterrez 1 Introducton Physcs 5153 Classcal Mechancs Prncple of Vrtual Work The frst varatonal prncple we encounter n mechancs s the prncple of vrtual work. It establshes the equlbrum condton of a mechancal
More informationArmy Ants Tunneling for Classical Simulations
Electronc Supplementary Materal (ESI) for Chemcal Scence. Ths journal s The Royal Socety of Chemstry 2014 electronc supplementary nformaton (ESI) for Chemcal Scence Army Ants Tunnelng for Classcal Smulatons
More informationModule 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur
Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:
More informationBézier curves. Michael S. Floater. September 10, These notes provide an introduction to Bézier curves. i=0
Bézer curves Mchael S. Floater September 1, 215 These notes provde an ntroducton to Bézer curves. 1 Bernsten polynomals Recall that a real polynomal of a real varable x R, wth degree n, s a functon of
More informationOn the correction of the h-index for career length
1 On the correcton of the h-ndex for career length by L. Egghe Unverstet Hasselt (UHasselt), Campus Depenbeek, Agoralaan, B-3590 Depenbeek, Belgum 1 and Unverstet Antwerpen (UA), IBW, Stadscampus, Venusstraat
More informationSpeeding up Computation of Scalar Multiplication in Elliptic Curve Cryptosystem
H.K. Pathak et. al. / (IJCSE) Internatonal Journal on Computer Scence and Engneerng Speedng up Computaton of Scalar Multplcaton n Ellptc Curve Cryptosystem H. K. Pathak Manju Sangh S.o.S n Computer scence
More informationLinear Approximation with Regularization and Moving Least Squares
Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...
More informationC/CS/Phy191 Problem Set 3 Solutions Out: Oct 1, 2008., where ( 00. ), so the overall state of the system is ) ( ( ( ( 00 ± 11 ), Φ ± = 1
C/CS/Phy9 Problem Set 3 Solutons Out: Oct, 8 Suppose you have two qubts n some arbtrary entangled state ψ You apply the teleportaton protocol to each of the qubts separately What s the resultng state obtaned
More informationOutline and Reading. Dynamic Programming. Dynamic Programming revealed. Computing Fibonacci. The General Dynamic Programming Technique
Outlne and Readng Dynamc Programmng The General Technque ( 5.3.2) -1 Knapsac Problem ( 5.3.3) Matrx Chan-Product ( 5.3.1) Dynamc Programmng verson 1.4 1 Dynamc Programmng verson 1.4 2 Dynamc Programmng
More informationBezier curves. Michael S. Floater. August 25, These notes provide an introduction to Bezier curves. i=0
Bezer curves Mchael S. Floater August 25, 211 These notes provde an ntroducton to Bezer curves. 1 Bernsten polynomals Recall that a real polynomal of a real varable x R, wth degree n, s a functon of the
More informationThe Study of Teaching-learning-based Optimization Algorithm
Advanced Scence and Technology Letters Vol. (AST 06), pp.05- http://dx.do.org/0.57/astl.06. The Study of Teachng-learnng-based Optmzaton Algorthm u Sun, Yan fu, Lele Kong, Haolang Q,, Helongang Insttute
More informationIntroduction to Vapor/Liquid Equilibrium, part 2. Raoult s Law:
CE304, Sprng 2004 Lecture 4 Introducton to Vapor/Lqud Equlbrum, part 2 Raoult s Law: The smplest model that allows us do VLE calculatons s obtaned when we assume that the vapor phase s an deal gas, and
More information2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification
E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton
More informationSome Comments on Accelerating Convergence of Iterative Sequences Using Direct Inversion of the Iterative Subspace (DIIS)
Some Comments on Acceleratng Convergence of Iteratve Sequences Usng Drect Inverson of the Iteratve Subspace (DIIS) C. Davd Sherrll School of Chemstry and Bochemstry Georga Insttute of Technology May 1998
More informationWe present the algorithm first, then derive it later. Assume access to a dataset {(x i, y i )} n i=1, where x i R d and y i { 1, 1}.
CS 189 Introducton to Machne Learnng Sprng 2018 Note 26 1 Boostng We have seen that n the case of random forests, combnng many mperfect models can produce a snglodel that works very well. Ths s the dea
More informationLectures - Week 4 Matrix norms, Conditioning, Vector Spaces, Linear Independence, Spanning sets and Basis, Null space and Range of a Matrix
Lectures - Week 4 Matrx norms, Condtonng, Vector Spaces, Lnear Independence, Spannng sets and Bass, Null space and Range of a Matrx Matrx Norms Now we turn to assocatng a number to each matrx. We could
More informationLOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin
Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence
More informationMore metrics on cartesian products
More metrcs on cartesan products If (X, d ) are metrc spaces for 1 n, then n Secton II4 of the lecture notes we defned three metrcs on X whose underlyng topologes are the product topology The purpose of
More informationStatistical machine learning and its application to neonatal seizure detection
19/Oct/2009 Statstcal machne learnng and ts applcaton to neonatal sezure detecton Presented by Andry Temko Department of Electrcal and Electronc Engneerng Page 2 of 42 A. Temko, Statstcal Machne Learnng
More informationIntegrals and Invariants of Euler-Lagrange Equations
Lecture 16 Integrals and Invarants of Euler-Lagrange Equatons ME 256 at the Indan Insttute of Scence, Bengaluru Varatonal Methods and Structural Optmzaton G. K. Ananthasuresh Professor, Mechancal Engneerng,
More informationFisher Linear Discriminant Analysis
Fsher Lnear Dscrmnant Analyss Max Wellng Department of Computer Scence Unversty of Toronto 10 Kng s College Road Toronto, M5S 3G5 Canada wellng@cs.toronto.edu Abstract Ths s a note to explan Fsher lnear
More informationHidden Markov Models
Hdden Markov Models Namrata Vaswan, Iowa State Unversty Aprl 24, 204 Hdden Markov Model Defntons and Examples Defntons:. A hdden Markov model (HMM) refers to a set of hdden states X 0, X,..., X t,...,
More informationVQ widely used in coding speech, image, and video
at Scalar quantzers are specal cases of vector quantzers (VQ): they are constraned to look at one sample at a tme (memoryless) VQ does not have such constrant better RD perfomance expected Source codng
More informationThe Synchronous 8th-Order Differential Attack on 12 Rounds of the Block Cipher HyRAL
The Synchronous 8th-Order Dfferental Attack on 12 Rounds of the Block Cpher HyRAL Yasutaka Igarash, Sej Fukushma, and Tomohro Hachno Kagoshma Unversty, Kagoshma, Japan Emal: {garash, fukushma, hachno}@eee.kagoshma-u.ac.jp
More informationLecture 3: Dual problems and Kernels
Lecture 3: Dual problems and Kernels C4B Machne Learnng Hlary 211 A. Zsserman Prmal and dual forms Lnear separablty revsted Feature mappng Kernels for SVMs Kernel trck requrements radal bass functons SVM
More informationHongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k)
ISSN 1749-3889 (prnt), 1749-3897 (onlne) Internatonal Journal of Nonlnear Scence Vol.17(2014) No.2,pp.188-192 Modfed Block Jacob-Davdson Method for Solvng Large Sparse Egenproblems Hongy Mao, College of
More informationCanonical transformations
Canoncal transformatons November 23, 2014 Recall that we have defned a symplectc transformaton to be any lnear transformaton M A B leavng the symplectc form nvarant, Ω AB M A CM B DΩ CD Coordnate transformatons,
More informationINF 5860 Machine learning for image classification. Lecture 3 : Image classification and regression part II Anne Solberg January 31, 2018
INF 5860 Machne learnng for mage classfcaton Lecture 3 : Image classfcaton and regresson part II Anne Solberg January 3, 08 Today s topcs Multclass logstc regresson and softma Regularzaton Image classfcaton
More informationStructure and Drive Paul A. Jensen Copyright July 20, 2003
Structure and Drve Paul A. Jensen Copyrght July 20, 2003 A system s made up of several operatons wth flow passng between them. The structure of the system descrbes the flow paths from nputs to outputs.
More information9 Derivation of Rate Equations from Single-Cell Conductance (Hodgkin-Huxley-like) Equations
Physcs 171/271 - Chapter 9R -Davd Klenfeld - Fall 2005 9 Dervaton of Rate Equatons from Sngle-Cell Conductance (Hodgkn-Huxley-lke) Equatons We consder a network of many neurons, each of whch obeys a set
More information