FITTING A RECTANGULAR FUNCTION BY GAUSSIANS AND APPLICATION TO THE MULTIVARIATE NORMAL INTEGRALS

Size: px
Start display at page:

Download "FITTING A RECTANGULAR FUNCTION BY GAUSSIANS AND APPLICATION TO THE MULTIVARIATE NORMAL INTEGRALS"

Transcription

1 Appl. Comput. Math., V.4, N.2, 25, pp FITTING A RECTANGULAR FUNCTION BY GAUSSIANS AND APPLICATION TO THE MULTIVARIATE NORMAL INTEGRALS HATEM A. FAYED, AMIR F. ATIYA 2, ASHRAF H. BADAWI 3 Abstract. Ths artcle ntroduces a new scheme to express a rectangular functon as a lnear combnaton of Gaussan functons. The man dea of ths scheme s based on fttng samples of the rectangular functon by adaptng the well-known clusterng algorthm, Gaussan mxture models (GMM). Ths method has several advantages compared to other exstng fttng algorthms. Frst, t ncorporates an effcent algorthm that can ft more Gaussan functons. Second, weghts of the lnear combnaton are already constraned n the algorthm to le n the nterval [,], whch avods large/small values that cause numercal nstablty. Thrd, almost the entre ftted Gaussan functons le wthn the nterval of the rectangular functon, whch can be utlzed effcently to approxmate dffcult defnte ntegrals such as the multvarate normal ntegral. Experments show that t s effcent when low accuracy s requred (error of order of 4 ) especally for small values of the correlaton coeffcents. Keywords: Functon Approxmaton, Gaussan Functons, Gaussan Mxture Models, Multvarate Normal Integrals. AMS Subject Classfcaton: 4Axx, 65D5, 65D3.. Introducton The problem of functon approxmaton s wdely used n many areas of scence and engneerng where numercal technques are ncorporated. Common approaches nvolve Taylor seres, orthogonal polynomals (Chebyshev, Hermte, Legendre, etc.), Gaussan functons, and Fourer seres. An mportant applcaton of functon approxmaton s the numercal approxmaton of ntegrals that do not have closed forms. The multvarate normal complementary ntegral s one of the sgnfcant ntegrals that appear n many engneerng and statstcs computatons. A lot of research was nvestgated to approxmate t, however no unque approxmaton s used for all dmensons and a predefned accuracy. The ntegral s defned by: L(h, Σ) = (2π) d Σ h h d exp } 2 xt Σ x dx, where x = (x, x 2, x d ), and Σ s an d d symmetrc postve defnte covarance matrx. The problem has receved consderable attenton n lterature [3, 4, 6]. For d = 2, there are some seres expressons [8, 8, 2] and effcent numercal technques [4 6, ]. For the multvarate case, d > 2, there exst several powerful numercal methods based on multvarate ntegraton technques that rely on ordnary Monte-Carlo methods, along wth some common varance reducton technques [2, 3]. Another group of algorthms s based on computng upper Department of Engneerng Mathematcs and Physcs, Caro Unversty, 263, Caro, Egypt & Unversty of Scence and Technology, Zewal Cty of Scence and Technology, 2588, Caro, Egypt e-mal: h fayed@eng.cu.edu.eg, hfayed@zewalcty.edu.eg 2 Department of Computer Engneerng, Caro Unversty, 263, Caro Unversty, 263, Caro, Egypt e-mal: amr@alumn.caltech.edu 3 Unversty of Scence and Technology, Zewal Cty of Scence and Technology, 2588, Caro, Egypt e-mal: abadaw@zewalcty.edu.eg Manuscrpt receved 2 November

2 H. FAYED et al.: FITTING A RECTANGULAR FUNCTION BY GAUSSIANS and lower bounds on the probablty (see [, 2] for a survey of these methods). Recently, Mwa [7] developed an algorthm that evaluates the multple ntegral by transformng t nto recursve evaluaton of a one-dmensonal ntegraton over a fne grd of ponts. Ths method s consdered among the most effcent methods for d. Fayed and Atya [9] derved a seres expanson based on Fourer seres that s more effcent up to d = 7. In most studes, the case when the components of x are equcorrelated, that s ρ j = ρ for all jand ρ =, s often used as a benchmark. Ths case can be evaluated as ( [23], p. 92): L(h, Σ) = 2π exp( t2 2 ) d = ( h + t ) ρ Φ dt. ρ 2. Gaussan mxture models as an approxmaton One of the approaches to approxmate a functon by a lnear combnaton of Gaussan s to sample the functon, and then use the radal bass functon networks (RBF networks) where the radal bass functons are chosen to be Gaussan functons [5]. However, to acheve a good accuracy n approxmatng a rectangular functon, a large network s often requred. Another approach s to use the method proposed n [], however ths method faled to obtan good accuracy for the rectangular functon due to the abrupt change at the begnnng and end of the functon. Nonlnear regresson s also another alternatve, but t becomes consderably slow f the number of components s moderately large, and t requres to bounhts to avod very small/large values that often deterorate the approxmaton n the multvarate case. So we proposed the combnaton weg a smple method that crcumvents the above problems leadng to a good approxmaton to a rectangular functon. Let us consder the one-dmensonal problem for approxmatng the followng rectangular functon: R(x, T ) = x T otherwse. The nterval [, T ] s sampled unformly to generate M data ponts. These data ponts are used to ft the rectangular functon by the followng Gaussan mxture model: R(x, T ) = T ω p (x λ ), = where K s the number of mxture components, λ = µ, σ 2 }, p (x λ ) N(µ, σ 2 ) s the normal dstrbuton wth mean µ and varance σ 2, and ω s the component weght n the mxture. So, for a predetermned number of mxtures (K), the mxture s parameters can be estmated teratvely usng the expectaton-maxmzaton (EM) algorthm as follows [7, 22]. Let X = x m R; m =,, M} be the sample sequence, and θ = ω, µ, σ 2 }, =,, K denotes, respectvely, the component weght, the mean and varance of the th normal component. To fnd the optmum values of the normal components usng the EM algorthm, maxmzaton of the followng lkelhood functon s performed: f (Θ) = ln p (x m, m Θ)} = m= ln p (x m m, Θ) P m }, where Θ = θ,, θ K }, m,, K} denotes that x m was generated from component. At each teraton j of EM algorthm, two steps are performed: the expectaton step (E-step), and the maxmzaton step (M-step) as descrbed below. m=

3 96 APPL. COMPUT. MATH., V.4, N.2, 25 E-step: Takng the expectaton of f (Θ) based on the current estmate Θ j, Q (Θ, Θ j ) } = E ln p (x m m, Θ j ) P m } = m= = M P ( m x m, Θ j ) ln p (x m m, Θ j ) P m }, m= m = where Q (Θ, Θ j ) s a functon of Θ, assumng that Θ j s fxed. The notaton can now be smplfed by droppng the ndex m from m. Ths s because, for each m, we sum up over all possble values of m, and these are the same for all m. Note also that P becomes smply the component weght ω. However, for GMM we have: } p (x m, Θ j ) = (2πσ ) exp (x m µ ) 2 2 d 2σ 2. So Q (Θ, Θ j ) = m= = P ( x m, Θ j ) d 2 ln ( 2πσ 2 ) 2σ 2 (x m µ ) 2 + ln ω }. M-step: By maxmzng the Q (Θ, Θ j ) wth respect to ω, µ, σ 2, we get: ω (j) = M P ( x m, Θ j ), m= (j) σ 2 = µ (j) = m= P ( x m, Θ j ) x m, P ( x m, Θ j ) m= m= ( P ( x m, Θ j ) x m µ (j) m= P ( x m, Θ j ) = P ( x m, Θ j ) k= ) 2 ω(j ) P (x m, Θ j ) ω (j ) k P (x m k, Θ j ) Fg. shows the results of fttng R(x, 4) usng a sample step of., and dfferent values of K. The shown results are the best of runs wth dfferent random ntalzaton that led to the mnmum mean absolute error:,. MAE = M R(x, T ). m= A smple form can also be obtaned by constranng the Gaussan functons to have the same varance σ 2 and equally spaced means;.e. µ = µ + ( ) δ. Thereby, the tradtonal EM

4 H. FAYED et al.: FITTING A RECTANGULAR FUNCTION BY GAUSSIANS algorthm s modfed to obtan µ and δ from the followng equatons: M ( ) P ( x m, Θ j ) ( ) = m= µ ( ) P ( x m, Θ j ) ( ) 2 = δ P ( x m, Θ j ) = m= = m= x m = m= ( ) P ( x m, Θ j ) x m = m= and the varance formula s reduced to: (j) σ 2 = Md = m= ( P ( x m, Θ j ) x m µ (j) Fg. 2 shows the results of fttng R(x, 4) usng a sample step of. and dfferent values of K usng these constrants. ) 2. or 3. The normal ntegrals Suppose that t s requred to approxmate the normal ntegral: I(h) = ) exp ( x2 dx. 2π 2 It can be approxmated usng the rectangular functon R (x h, T ) as: I(h) = 2π I(h) = T K 2π whch can smply be obtaned as: I(h) = = ω h R (x h, T ) exp T K 2π = N ( µ + h, σ 2 ω + σ 2 exp where T s chosen such that the Gaussan functon 2π exp ) ( x2 dx, 2 ) ) exp ( x2 dx, 2 } (µ + h) 2 2 ( ) + σ 2, ( T 2 2 to choose T = 4). For the multvarate case, suppose that t s requred to approxmate: L(h, Σ) = (2π) d Σ h h d exp } 2 xt Σ x dx. ) = (for h, t s reasonable One way to approxmate ths multple ntegral s to sample data n d-dmensonal space, and apply EM for the sampled data as before. However ths approach led to poor results as the effcency of EM degrades as both the number of samples and the number of Gaussan functons ncrease. Alternatvely, t was approxmated along each dmenson separately as:

5 98 APPL. COMPUT. MATH., V.4, N.2, (a) K = 5, MAE =.3 (b) K = 6, MAE = (c) K =, MAE =.49 (d) K = 2, MAE = (e) K = 3, MAE =.24 (f) K = 5, MAE =.9 Fgure : Results of fttng R(x, 4) (sold lne) usng the general form of GMM (dashed lne).

6 H. FAYED et al.: FITTING A RECTANGULAR FUNCTION BY GAUSSIANS (g) K = 5, MAE =.28 (h) K = 6, MAE = () K =, MAE =.85 (j) K = 2, MAE = (k) K = 3, MAE =.45 (l) K = 5, MAE =.33 Fgure 2: Results of fttng R(x, 4) (sold lne) usng the smple form of GMM (dashed lne).

7 2 APPL. COMPUT. MATH., V.4, N.2, 25 L(h, Σ) = whch can be evaluated as [2]: T d (2π) d Σ = exp 2 xt Σ x } dx d = ω ω d N(µ + h, Σ ) where L(h, Σ) = T d (2π) d = d = ω ω d exp (α ) Σ + Σ, α = 2 (µ + h) T Σ h h =. h d ( Σ + Σ, µ = µ. µ d ) Σ, Σ = (µ + h) 2 (µ + h) T Σ (µ + h), σ 2 σ σ 2 d Ths form, lke all other exstng algorthms used n approxmatng the normal ntegral, suffers from the curse of dmensonalty. However, to attan a smple expresson that can speed t up sgnfcantly, we used the smple form descrbed above; that s, the Gaussan functons are constraned to have the same varance σ 2 and µ l = µ + ( l ) δ, l d. Hence the ntegral can be approxmated by: L(h, Σ) = T d (2π) d Σ + σ 2 I = d = } ω ω d exp 2σ 2 C (µ + h) 2 2, where C s an upper trangular matrx obtaned from Cholesky decomposton of the matrx [ I Σ ( Σ + σ 2 I ) ]. In ths way, the computatonal complexty can be reduced to be of order O(d 2 K d ) flops. 4. Expermental results The orthant probabltes, L(, Σ) s evaluated usng the proposed method (GMM) for 7 d, and s compared wth Mwa s algorthm (avalable at: As a benchmark, we used the Gauss-Kronrod (7, 5) par quadrature ntegraton method for the equcorrelated case [23]. The probabltes are evaluated for ρ.,.2,.9}. For GMM, T = 4 s used to sample the rectangular functon R(x, T ), and K = 5 and K = 6 are nvestgated. For Mwa s algorthm, grd szes examned are G = 8 and G = 6 (whch have comparable runnng tmes wth the proposed approach). We used C language n our mplementaton on Wndows 7 runnng on Pentum 2.4GHz PC wth 3GB RAM. The absolute error and the elapsed tme are reported n Table to Table 4. It can be notced that GMM s comparable to Mwa s algorthm n accuracy, especally for ρ.7, whle consderably faster for d 8. Moreover, as the dmenson ncreases, processng of Mwa s algorthm becomes much slower than GMM. Thus, for 7 d, when low accuracy s requred, GMM becomes a reasonable choce, however, f hgh accuracy s needed, Mwa s algorthm s recommended.

8 H. FAYED et al.: FITTING A RECTANGULAR FUNCTION BY GAUSSIANS Conclusons In ths paper, an approxmaton of a rectangular functon s obtaned by adaptng the wellknown Gaussan mxture models to express t as a lnear combnaton of Gaussan functons. The proposed approxmaton s used to derve an approxmate expresson for the multvarate normal ntegral. The obtaned expresson s found to be fast f an error of order of 4 s plausble, especally for d 7 and ρ.7. Moreover, for d, t s consderably faster and thus more approprate and feasble than Mwa s algorthm. In future work, explorng other optmzaton strateges than the EM algorthm may be nvestgated, as t may yeld further mprovements [9]. Table. Results of the orthant probabltes for d = 7 Elapsed Tme Error ρ Mwa GMM Mwa GMM G = 8 G = 6 K = 5 K = 6 G = 8 G = 6 K = 5 K = E-4 3.6E-5.76E-4.3E E E-5.77E-4.34E E-4 2.5E E-5.5E E E-5.7E-4.52E E-4 2.E E-4 2.7E E-4.84E-5.47E-3 7.6E E-4.69E-5 2.9E-3.35E E-4.54E E-3.69E E-4.2E E E-4 Table 2. Results of the orthant probabltes for d = 8 Elapsed Tme Error ρ Mwa GMM Mwa GMM G = 8 G = 6 K = 5 K = 6 G = 8 G = 6 K = 5 K = E E-5.23E-4 8.4E E-4 6.E-5.25E-4.6E E E E-5 8.9E E E E E E E E-4 3.6E E E-5.67E-3 7.6E E-3 8.5E-5 3.3E-3.33E E E E-3.64E E-3 9.5E E E-4 Table 3. Results of the orthant probabltes for d = 9 Elapsed Tme Error ρ Mwa GMM Mwa GMM G = 8 G = 6 K = 5 K = 6 G = 8 G = 6 K = 5 K = E-3 8.6E E E E-3.8E-4 9.E E E-3.3E E-6 7.7E E-3.53E E E E-3.76E-4 8.8E-4 3.2E E-3 2.E-4.79E E E E-4 3.3E-3.25E E E-4 6.5E-3.4E E-3 2.9E-4.E-2 3.5E-5

9 22 APPL. COMPUT. MATH., V.4, N.2, 25 Table 4. Results of the orthant probabltes for d = Elapsed Tme Error ρ Mwa GMM Mwa GMM G = 8 G = 6 K = 5 K = 6 G = 8 G = 6 K = 5 K = E-3.5E E-5 4.4E E-3.67E E-5 7.5E E E-4.83E E E E E E E-3 3.5E E E E-3 4.2E-4.79E E E E-4 3.3E-3.7E E-3 5.9E E-3.E E-2 7.2E-4.7E E-5 References [] Calcaterra, C. Lnear combnaton of Gaussans wth a sngle varance are dense n L2, Proceedngs of the World Congress on Engneerng WCE, V.2, 28. [2] Deák, I. Three dgt accurate multple normal probabltes, Numer. Math., V.35, 98, pp [3] Deák, I. Random Number Generators and Smulaton, Akadéma Kadó, 99. [4] Dvg, D.R. Calculaton of unvarate and bvarate normal probablty functons, Ann. Stat., V.7, N.4, 979, pp [5] Drezner, Z. Computaton of the bvarate normal ntegral, Math. Comput., V.32, 978, pp [6] Drezner, Z., Wesolowsky, G.O. The computaton of the bvarate normal ntegral, J. Stat. Comput. Smul., V. 35, 99, pp.-7. [7] Duda, R.O., Hart, P.E., Stork, D.G. Pattern Classfcaton, 2nd ed., Wley, New York, 2. [8] Fayed, H.A., Atya, A.F. An evaluaton of the ntegral of the product of the error functon and the normal probablty densty, wth applcaton to the bvarate normal ntegral, Math. Comput., V.83, N.285, 24, pp [9] Fayed, H.A., Atya, A.F. A novel seres expanson for the multvarate normal probablty ntegrals based on fourer seres, Math. Comput., V.83, N.289, 24, pp [] Gassmann, H. Multvarate normal probabltes: Implementng an old dea of Plackett s, J. Comp. Graph. Stat., V.2, N.3, 23, pp [] Genz, A. Numercal computaton of rectangular bvarate and trvarate normal and t probabltes, Stat. Comput., V.4, N.3, 24, pp [2] Genz, A. Comparson of methods for the computaton of multvarate normal probabltes, Comp. Sc. Stat., V.25, 993, pp [3] Gupta, S.S. Probablty ntegrals of multvarate normal and multvarate t, Ann. Math. Statst., V.34, 963, pp [4] Harrs, B., Soms,A.P. The use of the tetrachorc seres for evaluatng multvarate normal probabltes, J. Multvarate Anal., V., 98, pp [5] Haykn, S. Neural Networks: a Comprehensve Foundaton, 2nd ed., Prentce-Hall, 999. [6] Kendall, M.G. Proof of relatons connected wth the tetrachorc seres and ts generalzatons, Bometrka, V.32, 94, pp [7] Mwa, T., Hayter, A.J., Kurk, S. The evaluaton of general non-centred orthant probabltes, J. R. Statst. Soc. B, V.65, 23, pp [8] Owen, D.B. Tables for computng bvarate normal probabltes, Ann. Math. Stat., V.27, N.4, 956, pp [9] Pardalos, M.P. and Chnchuluun, A. Some recent developments n determnstc global optmzaton (survey), Appl. Comput. Math., V.5, N., 26, pp [2] Pearson, K. Mathematcal contrbutons to the theory of evoluton. VII. on the correlaton of characters not quanttatvely, Phlos. Trans. R. Soc. S-A, V.96, 9, pp.-47. [2] Petersen, K.B. and Pedersen, M.S. The matrx cookbook, Nov 22, verson 225. [Onlne]. Avalable: detals.php?d=3274 [22] Theodords, S., Koutroumbas, K. Pattern Recognton, 2nd ed., Elsever, New York, 23. [23] Tong, Y.L. The Multvarate Normal Dstrbuton, Sprnger Seres n Statstcs, Sprnger-Verlag, New York, 99.

10 H. FAYED et al.: FITTING A RECTANGULAR FUNCTION BY GAUSSIANS Hatem A. Fayed - s an Assocate Professor at the Engneerng Mathematcs and Physcs Department, Caro Unversty, and currently a seconded Assocate Professor at Zewal Cty of Scence and Technology. He receved hs Ph.D. from the Engneerng Mathematcs and Physcs Department, Caro Unversty, 25. Hs research nterests are n the areas of machne learnng, tme seres forecastng, neural networks, optmzaton technques, and mage ptrocessng. Amr F. Atya - receved hs B.S. and M.S. degrees from Caro Unversty, and hs M.S. and Ph.D. degrees from Caltech, Pasadena, CA, all n electrcal engneerng. Dr. Atya s currently a Professor at the Department of Computer Engneerng, Caro Unversty. Hs research nterests are n the areas of machne learnng, theory of forecastng, computatonal fnance, dynamc prcng, and Monte Carlo Methods. Ashraf H. Badaw - s a Dean of Student Affars, Assstant Professor at the Center of Nanotechnology at Zewal Cty. He s also the Drector for the Learnng Center of Learnng Technologes. Pror to jonng SMART Ashraf was the lead WMAX Solutons Specalsts for Intel n the Mddle East and Afrca. He was an assstant professor, Caro Unversty, Engneerng Math and Physcs department from 22 tll 29. He graduated from the Systems and Bomedcal Engneerng Department n 99 n Caro, where he started pursung hs M.Sc. n Engneerng Physcs. He then traveled to Wnnpeg, Canada to pursue hs PhD n Electrcal Engneerng from the Unversty of Mantoba.

A new Approach for Solving Linear Ordinary Differential Equations

A new Approach for Solving Linear Ordinary Differential Equations , ISSN 974-57X (Onlne), ISSN 974-5718 (Prnt), Vol. ; Issue No. 1; Year 14, Copyrght 13-14 by CESER PUBLICATIONS A new Approach for Solvng Lnear Ordnary Dfferental Equatons Fawz Abdelwahd Department of

More information

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement

Markov Chain Monte Carlo (MCMC), Gibbs Sampling, Metropolis Algorithms, and Simulated Annealing Bioinformatics Course Supplement Markov Chan Monte Carlo MCMC, Gbbs Samplng, Metropols Algorthms, and Smulated Annealng 2001 Bonformatcs Course Supplement SNU Bontellgence Lab http://bsnuackr/ Outlne! Markov Chan Monte Carlo MCMC! Metropols-Hastngs

More information

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems

Chapter 5. Solution of System of Linear Equations. Module No. 6. Solution of Inconsistent and Ill Conditioned Systems Numercal Analyss by Dr. Anta Pal Assstant Professor Department of Mathematcs Natonal Insttute of Technology Durgapur Durgapur-713209 emal: anta.bue@gmal.com 1 . Chapter 5 Soluton of System of Lnear Equatons

More information

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications

Durban Watson for Testing the Lack-of-Fit of Polynomial Regression Models without Replications Durban Watson for Testng the Lack-of-Ft of Polynomal Regresson Models wthout Replcatons Ruba A. Alyaf, Maha A. Omar, Abdullah A. Al-Shha ralyaf@ksu.edu.sa, maomar@ksu.edu.sa, aalshha@ksu.edu.sa Department

More information

Global Sensitivity. Tuesday 20 th February, 2018

Global Sensitivity. Tuesday 20 th February, 2018 Global Senstvty Tuesday 2 th February, 28 ) Local Senstvty Most senstvty analyses [] are based on local estmates of senstvty, typcally by expandng the response n a Taylor seres about some specfc values

More information

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results.

For now, let us focus on a specific model of neurons. These are simplified from reality but can achieve remarkable results. Neural Networks : Dervaton compled by Alvn Wan from Professor Jtendra Malk s lecture Ths type of computaton s called deep learnng and s the most popular method for many problems, such as computer vson

More information

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010

Parametric fractional imputation for missing data analysis. Jae Kwang Kim Survey Working Group Seminar March 29, 2010 Parametrc fractonal mputaton for mssng data analyss Jae Kwang Km Survey Workng Group Semnar March 29, 2010 1 Outlne Introducton Proposed method Fractonal mputaton Approxmaton Varance estmaton Multple mputaton

More information

IV. Performance Optimization

IV. Performance Optimization IV. Performance Optmzaton A. Steepest descent algorthm defnton how to set up bounds on learnng rate mnmzaton n a lne (varyng learnng rate) momentum learnng examples B. Newton s method defnton Gauss-Newton

More information

Support Vector Machines. Vibhav Gogate The University of Texas at dallas

Support Vector Machines. Vibhav Gogate The University of Texas at dallas Support Vector Machnes Vbhav Gogate he Unversty of exas at dallas What We have Learned So Far? 1. Decson rees. Naïve Bayes 3. Lnear Regresson 4. Logstc Regresson 5. Perceptron 6. Neural networks 7. K-Nearest

More information

A Robust Method for Calculating the Correlation Coefficient

A Robust Method for Calculating the Correlation Coefficient A Robust Method for Calculatng the Correlaton Coeffcent E.B. Nven and C. V. Deutsch Relatonshps between prmary and secondary data are frequently quantfed usng the correlaton coeffcent; however, the tradtonal

More information

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin

LOW BIAS INTEGRATED PATH ESTIMATORS. James M. Calvin Proceedngs of the 007 Wnter Smulaton Conference S G Henderson, B Bller, M-H Hseh, J Shortle, J D Tew, and R R Barton, eds LOW BIAS INTEGRATED PATH ESTIMATORS James M Calvn Department of Computer Scence

More information

A Hybrid Variational Iteration Method for Blasius Equation

A Hybrid Variational Iteration Method for Blasius Equation Avalable at http://pvamu.edu/aam Appl. Appl. Math. ISSN: 1932-9466 Vol. 10, Issue 1 (June 2015), pp. 223-229 Applcatons and Appled Mathematcs: An Internatonal Journal (AAM) A Hybrd Varatonal Iteraton Method

More information

EEE 241: Linear Systems

EEE 241: Linear Systems EEE : Lnear Systems Summary #: Backpropagaton BACKPROPAGATION The perceptron rule as well as the Wdrow Hoff learnng were desgned to tran sngle layer networks. They suffer from the same dsadvantage: they

More information

Lecture Notes on Linear Regression

Lecture Notes on Linear Regression Lecture Notes on Lnear Regresson Feng L fl@sdueducn Shandong Unversty, Chna Lnear Regresson Problem In regresson problem, we am at predct a contnuous target value gven an nput feature vector We assume

More information

Numerical Heat and Mass Transfer

Numerical Heat and Mass Transfer Master degree n Mechancal Engneerng Numercal Heat and Mass Transfer 06-Fnte-Dfference Method (One-dmensonal, steady state heat conducton) Fausto Arpno f.arpno@uncas.t Introducton Why we use models and

More information

NUMERICAL DIFFERENTIATION

NUMERICAL DIFFERENTIATION NUMERICAL DIFFERENTIATION 1 Introducton Dfferentaton s a method to compute the rate at whch a dependent output y changes wth respect to the change n the ndependent nput x. Ths rate of change s called the

More information

Linear Approximation with Regularization and Moving Least Squares

Linear Approximation with Regularization and Moving Least Squares Lnear Approxmaton wth Regularzaton and Movng Least Squares Igor Grešovn May 007 Revson 4.6 (Revson : March 004). 5 4 3 0.5 3 3.5 4 Contents: Lnear Fttng...4. Weghted Least Squares n Functon Approxmaton...

More information

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS Avalable onlne at http://sck.org J. Math. Comput. Sc. 3 (3), No., 6-3 ISSN: 97-537 COMPARISON OF SOME RELIABILITY CHARACTERISTICS BETWEEN REDUNDANT SYSTEMS REQUIRING SUPPORTING UNITS FOR THEIR OPERATIONS

More information

Speeding up Computation of Scalar Multiplication in Elliptic Curve Cryptosystem

Speeding up Computation of Scalar Multiplication in Elliptic Curve Cryptosystem H.K. Pathak et. al. / (IJCSE) Internatonal Journal on Computer Scence and Engneerng Speedng up Computaton of Scalar Multplcaton n Ellptc Curve Cryptosystem H. K. Pathak Manju Sangh S.o.S n Computer scence

More information

Gaussian Mixture Models

Gaussian Mixture Models Lab Gaussan Mxture Models Lab Objectve: Understand the formulaton of Gaussan Mxture Models (GMMs) and how to estmate GMM parameters. You ve already seen GMMs as the observaton dstrbuton n certan contnuous

More information

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1

On an Extension of Stochastic Approximation EM Algorithm for Incomplete Data Problems. Vahid Tadayon 1 On an Extenson of Stochastc Approxmaton EM Algorthm for Incomplete Data Problems Vahd Tadayon Abstract: The Stochastc Approxmaton EM (SAEM algorthm, a varant stochastc approxmaton of EM, s a versatle tool

More information

Small Area Interval Estimation

Small Area Interval Estimation .. Small Area Interval Estmaton Partha Lahr Jont Program n Survey Methodology Unversty of Maryland, College Park (Based on jont work wth Masayo Yoshmor, Former JPSM Vstng PhD Student and Research Fellow

More information

Chapter 9: Statistical Inference and the Relationship between Two Variables

Chapter 9: Statistical Inference and the Relationship between Two Variables Chapter 9: Statstcal Inference and the Relatonshp between Two Varables Key Words The Regresson Model The Sample Regresson Equaton The Pearson Correlaton Coeffcent Learnng Outcomes After studyng ths chapter,

More information

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu

BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS. M. Krishna Reddy, B. Naveen Kumar and Y. Ramu BOOTSTRAP METHOD FOR TESTING OF EQUALITY OF SEVERAL MEANS M. Krshna Reddy, B. Naveen Kumar and Y. Ramu Department of Statstcs, Osmana Unversty, Hyderabad -500 007, Inda. nanbyrozu@gmal.com, ramu0@gmal.com

More information

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method

Comparison of the Population Variance Estimators. of 2-Parameter Exponential Distribution Based on. Multiple Criteria Decision Making Method Appled Mathematcal Scences, Vol. 7, 0, no. 47, 07-0 HIARI Ltd, www.m-hkar.com Comparson of the Populaton Varance Estmators of -Parameter Exponental Dstrbuton Based on Multple Crtera Decson Makng Method

More information

A linear imaging system with white additive Gaussian noise on the observed data is modeled as follows:

A linear imaging system with white additive Gaussian noise on the observed data is modeled as follows: Supplementary Note Mathematcal bacground A lnear magng system wth whte addtve Gaussan nose on the observed data s modeled as follows: X = R ϕ V + G, () where X R are the expermental, two-dmensonal proecton

More information

One-sided finite-difference approximations suitable for use with Richardson extrapolation

One-sided finite-difference approximations suitable for use with Richardson extrapolation Journal of Computatonal Physcs 219 (2006) 13 20 Short note One-sded fnte-dfference approxmatons sutable for use wth Rchardson extrapolaton Kumar Rahul, S.N. Bhattacharyya * Department of Mechancal Engneerng,

More information

A MODIFIED METHOD FOR SOLVING SYSTEM OF NONLINEAR EQUATIONS

A MODIFIED METHOD FOR SOLVING SYSTEM OF NONLINEAR EQUATIONS Journal of Mathematcs and Statstcs 9 (1): 4-8, 1 ISSN 1549-644 1 Scence Publcatons do:1.844/jmssp.1.4.8 Publshed Onlne 9 (1) 1 (http://www.thescpub.com/jmss.toc) A MODIFIED METHOD FOR SOLVING SYSTEM OF

More information

Week 5: Neural Networks

Week 5: Neural Networks Week 5: Neural Networks Instructor: Sergey Levne Neural Networks Summary In the prevous lecture, we saw how we can construct neural networks by extendng logstc regresson. Neural networks consst of multple

More information

Regularized Discriminant Analysis for Face Recognition

Regularized Discriminant Analysis for Face Recognition 1 Regularzed Dscrmnant Analyss for Face Recognton Itz Pma, Mayer Aladem Department of Electrcal and Computer Engneerng, Ben-Guron Unversty of the Negev P.O.Box 653, Beer-Sheva, 845, Israel. Abstract Ths

More information

Research Article Green s Theorem for Sign Data

Research Article Green s Theorem for Sign Data Internatonal Scholarly Research Network ISRN Appled Mathematcs Volume 2012, Artcle ID 539359, 10 pages do:10.5402/2012/539359 Research Artcle Green s Theorem for Sgn Data Lous M. Houston The Unversty of

More information

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k)

Hongyi Miao, College of Science, Nanjing Forestry University, Nanjing ,China. (Received 20 June 2013, accepted 11 March 2014) I)ϕ (k) ISSN 1749-3889 (prnt), 1749-3897 (onlne) Internatonal Journal of Nonlnear Scence Vol.17(2014) No.2,pp.188-192 Modfed Block Jacob-Davdson Method for Solvng Large Sparse Egenproblems Hongy Mao, College of

More information

10-701/ Machine Learning, Fall 2005 Homework 3

10-701/ Machine Learning, Fall 2005 Homework 3 10-701/15-781 Machne Learnng, Fall 2005 Homework 3 Out: 10/20/05 Due: begnnng of the class 11/01/05 Instructons Contact questons-10701@autonlaborg for queston Problem 1 Regresson and Cross-valdaton [40

More information

Primer on High-Order Moment Estimators

Primer on High-Order Moment Estimators Prmer on Hgh-Order Moment Estmators Ton M. Whted July 2007 The Errors-n-Varables Model We wll start wth the classcal EIV for one msmeasured regressor. The general case s n Erckson and Whted Econometrc

More information

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2)

MATH 829: Introduction to Data Mining and Analysis The EM algorithm (part 2) 1/16 MATH 829: Introducton to Data Mnng and Analyss The EM algorthm (part 2) Domnque Gullot Departments of Mathematcal Scences Unversty of Delaware Aprl 20, 2016 Recall 2/16 We are gven ndependent observatons

More information

Generalized Linear Methods

Generalized Linear Methods Generalzed Lnear Methods 1 Introducton In the Ensemble Methods the general dea s that usng a combnaton of several weak learner one could make a better learner. More formally, assume that we have a set

More information

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4)

Econ107 Applied Econometrics Topic 3: Classical Model (Studenmund, Chapter 4) I. Classcal Assumptons Econ7 Appled Econometrcs Topc 3: Classcal Model (Studenmund, Chapter 4) We have defned OLS and studed some algebrac propertes of OLS. In ths topc we wll study statstcal propertes

More information

Hidden Markov Models & The Multivariate Gaussian (10/26/04)

Hidden Markov Models & The Multivariate Gaussian (10/26/04) CS281A/Stat241A: Statstcal Learnng Theory Hdden Markov Models & The Multvarate Gaussan (10/26/04) Lecturer: Mchael I. Jordan Scrbes: Jonathan W. Hu 1 Hdden Markov Models As a bref revew, hdden Markov models

More information

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980

Chat eld, C. and A.J.Collins, Introduction to multivariate analysis. Chapman & Hall, 1980 MT07: Multvarate Statstcal Methods Mke Tso: emal mke.tso@manchester.ac.uk Webpage for notes: http://www.maths.manchester.ac.uk/~mkt/new_teachng.htm. Introducton to multvarate data. Books Chat eld, C. and

More information

Lecture 12: Discrete Laplacian

Lecture 12: Discrete Laplacian Lecture 12: Dscrete Laplacan Scrbe: Tanye Lu Our goal s to come up wth a dscrete verson of Laplacan operator for trangulated surfaces, so that we can use t n practce to solve related problems We are mostly

More information

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests

Simulated Power of the Discrete Cramér-von Mises Goodness-of-Fit Tests Smulated of the Cramér-von Mses Goodness-of-Ft Tests Steele, M., Chaselng, J. and 3 Hurst, C. School of Mathematcal and Physcal Scences, James Cook Unversty, Australan School of Envronmental Studes, Grffth

More information

Chapter 11: Simple Linear Regression and Correlation

Chapter 11: Simple Linear Regression and Correlation Chapter 11: Smple Lnear Regresson and Correlaton 11-1 Emprcal Models 11-2 Smple Lnear Regresson 11-3 Propertes of the Least Squares Estmators 11-4 Hypothess Test n Smple Lnear Regresson 11-4.1 Use of t-tests

More information

Non-linear Canonical Correlation Analysis Using a RBF Network

Non-linear Canonical Correlation Analysis Using a RBF Network ESANN' proceedngs - European Smposum on Artfcal Neural Networks Bruges (Belgum), 4-6 Aprl, d-sde publ., ISBN -97--, pp. 57-5 Non-lnear Canoncal Correlaton Analss Usng a RBF Network Sukhbnder Kumar, Elane

More information

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14

APPROXIMATE PRICES OF BASKET AND ASIAN OPTIONS DUPONT OLIVIER. Premia 14 APPROXIMAE PRICES OF BASKE AND ASIAN OPIONS DUPON OLIVIER Prema 14 Contents Introducton 1 1. Framewor 1 1.1. Baset optons 1.. Asan optons. Computng the prce 3. Lower bound 3.1. Closed formula for the prce

More information

Application of B-Spline to Numerical Solution of a System of Singularly Perturbed Problems

Application of B-Spline to Numerical Solution of a System of Singularly Perturbed Problems Mathematca Aeterna, Vol. 1, 011, no. 06, 405 415 Applcaton of B-Splne to Numercal Soluton of a System of Sngularly Perturbed Problems Yogesh Gupta Department of Mathematcs Unted College of Engneerng &

More information

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity

LINEAR REGRESSION ANALYSIS. MODULE IX Lecture Multicollinearity LINEAR REGRESSION ANALYSIS MODULE IX Lecture - 30 Multcollnearty Dr. Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur 2 Remedes for multcollnearty Varous technques have

More information

MMA and GCMMA two methods for nonlinear optimization

MMA and GCMMA two methods for nonlinear optimization MMA and GCMMA two methods for nonlnear optmzaton Krster Svanberg Optmzaton and Systems Theory, KTH, Stockholm, Sweden. krlle@math.kth.se Ths note descrbes the algorthms used n the author s 2007 mplementatons

More information

Inductance Calculation for Conductors of Arbitrary Shape

Inductance Calculation for Conductors of Arbitrary Shape CRYO/02/028 Aprl 5, 2002 Inductance Calculaton for Conductors of Arbtrary Shape L. Bottura Dstrbuton: Internal Summary In ths note we descrbe a method for the numercal calculaton of nductances among conductors

More information

A New Refinement of Jacobi Method for Solution of Linear System Equations AX=b

A New Refinement of Jacobi Method for Solution of Linear System Equations AX=b Int J Contemp Math Scences, Vol 3, 28, no 17, 819-827 A New Refnement of Jacob Method for Soluton of Lnear System Equatons AX=b F Naem Dafchah Department of Mathematcs, Faculty of Scences Unversty of Gulan,

More information

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur

Module 3 LOSSY IMAGE COMPRESSION SYSTEMS. Version 2 ECE IIT, Kharagpur Module 3 LOSSY IMAGE COMPRESSION SYSTEMS Verson ECE IIT, Kharagpur Lesson 6 Theory of Quantzaton Verson ECE IIT, Kharagpur Instructonal Objectves At the end of ths lesson, the students should be able to:

More information

Finding Dense Subgraphs in G(n, 1/2)

Finding Dense Subgraphs in G(n, 1/2) Fndng Dense Subgraphs n Gn, 1/ Atsh Das Sarma 1, Amt Deshpande, and Rav Kannan 1 Georga Insttute of Technology,atsh@cc.gatech.edu Mcrosoft Research-Bangalore,amtdesh,annan@mcrosoft.com Abstract. Fndng

More information

NON-CENTRAL 7-POINT FORMULA IN THE METHOD OF LINES FOR PARABOLIC AND BURGERS' EQUATIONS

NON-CENTRAL 7-POINT FORMULA IN THE METHOD OF LINES FOR PARABOLIC AND BURGERS' EQUATIONS IJRRAS 8 (3 September 011 www.arpapress.com/volumes/vol8issue3/ijrras_8_3_08.pdf NON-CENTRAL 7-POINT FORMULA IN THE METHOD OF LINES FOR PARABOLIC AND BURGERS' EQUATIONS H.O. Bakodah Dept. of Mathematc

More information

Parameter Estimation for Dynamic System using Unscented Kalman filter

Parameter Estimation for Dynamic System using Unscented Kalman filter Parameter Estmaton for Dynamc System usng Unscented Kalman flter Jhoon Seung 1,a, Amr Atya F. 2,b, Alexander G.Parlos 3,c, and Klto Chong 1,4,d* 1 Dvson of Electroncs Engneerng, Chonbuk Natonal Unversty,

More information

Feature Selection: Part 1

Feature Selection: Part 1 CSE 546: Machne Learnng Lecture 5 Feature Selecton: Part 1 Instructor: Sham Kakade 1 Regresson n the hgh dmensonal settng How do we learn when the number of features d s greater than the sample sze n?

More information

Supporting Information

Supporting Information Supportng Informaton The neural network f n Eq. 1 s gven by: f x l = ReLU W atom x l + b atom, 2 where ReLU s the element-wse rectfed lnear unt, 21.e., ReLUx = max0, x, W atom R d d s the weght matrx to

More information

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression

MACHINE APPLIED MACHINE LEARNING LEARNING. Gaussian Mixture Regression 11 MACHINE APPLIED MACHINE LEARNING LEARNING MACHINE LEARNING Gaussan Mture Regresson 22 MACHINE APPLIED MACHINE LEARNING LEARNING Bref summary of last week s lecture 33 MACHINE APPLIED MACHINE LEARNING

More information

DETERMINATION OF TEMPERATURE DISTRIBUTION FOR ANNULAR FINS WITH TEMPERATURE DEPENDENT THERMAL CONDUCTIVITY BY HPM

DETERMINATION OF TEMPERATURE DISTRIBUTION FOR ANNULAR FINS WITH TEMPERATURE DEPENDENT THERMAL CONDUCTIVITY BY HPM Ganj, Z. Z., et al.: Determnaton of Temperature Dstrbuton for S111 DETERMINATION OF TEMPERATURE DISTRIBUTION FOR ANNULAR FINS WITH TEMPERATURE DEPENDENT THERMAL CONDUCTIVITY BY HPM by Davood Domr GANJI

More information

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification

2E Pattern Recognition Solutions to Introduction to Pattern Recognition, Chapter 2: Bayesian pattern classification E395 - Pattern Recognton Solutons to Introducton to Pattern Recognton, Chapter : Bayesan pattern classfcaton Preface Ths document s a soluton manual for selected exercses from Introducton to Pattern Recognton

More information

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur

Dr. Shalabh Department of Mathematics and Statistics Indian Institute of Technology Kanpur Analyss of Varance and Desgn of Experment-I MODULE VII LECTURE - 3 ANALYSIS OF COVARIANCE Dr Shalabh Department of Mathematcs and Statstcs Indan Insttute of Technology Kanpur Any scentfc experment s performed

More information

Lecture 16 Statistical Analysis in Biomaterials Research (Part II)

Lecture 16 Statistical Analysis in Biomaterials Research (Part II) 3.051J/0.340J 1 Lecture 16 Statstcal Analyss n Bomaterals Research (Part II) C. F Dstrbuton Allows comparson of varablty of behavor between populatons usng test of hypothess: σ x = σ x amed for Brtsh statstcan

More information

Appendix B: Resampling Algorithms

Appendix B: Resampling Algorithms 407 Appendx B: Resamplng Algorthms A common problem of all partcle flters s the degeneracy of weghts, whch conssts of the unbounded ncrease of the varance of the mportance weghts ω [ ] of the partcles

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours

UNIVERSITY OF TORONTO Faculty of Arts and Science. December 2005 Examinations STA437H1F/STA1005HF. Duration - 3 hours UNIVERSITY OF TORONTO Faculty of Arts and Scence December 005 Examnatons STA47HF/STA005HF Duraton - hours AIDS ALLOWED: (to be suppled by the student) Non-programmable calculator One handwrtten 8.5'' x

More information

Chapter 13: Multiple Regression

Chapter 13: Multiple Regression Chapter 13: Multple Regresson 13.1 Developng the multple-regresson Model The general model can be descrbed as: It smplfes for two ndependent varables: The sample ft parameter b 0, b 1, and b are used to

More information

ANOMALIES OF THE MAGNITUDE OF THE BIAS OF THE MAXIMUM LIKELIHOOD ESTIMATOR OF THE REGRESSION SLOPE

ANOMALIES OF THE MAGNITUDE OF THE BIAS OF THE MAXIMUM LIKELIHOOD ESTIMATOR OF THE REGRESSION SLOPE P a g e ANOMALIES OF THE MAGNITUDE OF THE BIAS OF THE MAXIMUM LIKELIHOOD ESTIMATOR OF THE REGRESSION SLOPE Darmud O Drscoll ¹, Donald E. Ramrez ² ¹ Head of Department of Mathematcs and Computer Studes

More information

U-Pb Geochronology Practical: Background

U-Pb Geochronology Practical: Background U-Pb Geochronology Practcal: Background Basc Concepts: accuracy: measure of the dfference between an expermental measurement and the true value precson: measure of the reproducblty of the expermental result

More information

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k

Number of cases Number of factors Number of covariates Number of levels of factor i. Value of the dependent variable for case k ANOVA Model and Matrx Computatons Notaton The followng notaton s used throughout ths chapter unless otherwse stated: N F CN Y Z j w W Number of cases Number of factors Number of covarates Number of levels

More information

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M

CIS526: Machine Learning Lecture 3 (Sept 16, 2003) Linear Regression. Preparation help: Xiaoying Huang. x 1 θ 1 output... θ M x M CIS56: achne Learnng Lecture 3 (Sept 6, 003) Preparaton help: Xaoyng Huang Lnear Regresson Lnear regresson can be represented by a functonal form: f(; θ) = θ 0 0 +θ + + θ = θ = 0 ote: 0 s a dummy attrbute

More information

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X

3.1 Expectation of Functions of Several Random Variables. )' be a k-dimensional discrete or continuous random vector, with joint PMF p (, E X E X1 E X Statstcs 1: Probablty Theory II 37 3 EPECTATION OF SEVERAL RANDOM VARIABLES As n Probablty Theory I, the nterest n most stuatons les not on the actual dstrbuton of a random vector, but rather on a number

More information

Curve Fitting with the Least Square Method

Curve Fitting with the Least Square Method WIKI Document Number 5 Interpolaton wth Least Squares Curve Fttng wth the Least Square Method Mattheu Bultelle Department of Bo-Engneerng Imperal College, London Context We wsh to model the postve feedback

More information

A Bayesian Approach to Arrival Rate Forecasting for Inhomogeneous Poisson Processes for Mobile Calls

A Bayesian Approach to Arrival Rate Forecasting for Inhomogeneous Poisson Processes for Mobile Calls A Bayesan Approach to Arrval Rate Forecastng for Inhomogeneous Posson Processes for Moble Calls Mchael N. Nawar Department of Computer Engneerng Caro Unversty Caro, Egypt mchaelnawar@eee.org Amr F. Atya

More information

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering /

P R. Lecture 4. Theory and Applications of Pattern Recognition. Dept. of Electrical and Computer Engineering / Theory and Applcatons of Pattern Recognton 003, Rob Polkar, Rowan Unversty, Glassboro, NJ Lecture 4 Bayes Classfcaton Rule Dept. of Electrcal and Computer Engneerng 0909.40.0 / 0909.504.04 Theory & Applcatons

More information

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z )

C4B Machine Learning Answers II. = σ(z) (1 σ(z)) 1 1 e z. e z = σ(1 σ) (1 + e z ) C4B Machne Learnng Answers II.(a) Show that for the logstc sgmod functon dσ(z) dz = σ(z) ( σ(z)) A. Zsserman, Hlary Term 20 Start from the defnton of σ(z) Note that Then σ(z) = σ = dσ(z) dz = + e z e z

More information

Linear Feature Engineering 11

Linear Feature Engineering 11 Lnear Feature Engneerng 11 2 Least-Squares 2.1 Smple least-squares Consder the followng dataset. We have a bunch of nputs x and correspondng outputs y. The partcular values n ths dataset are x y 0.23 0.19

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm The Expectaton-Maxmaton Algorthm Charles Elan elan@cs.ucsd.edu November 16, 2007 Ths chapter explans the EM algorthm at multple levels of generalty. Secton 1 gves the standard hgh-level verson of the algorthm.

More information

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 )

Yong Joon Ryang. 1. Introduction Consider the multicommodity transportation problem with convex quadratic cost function. 1 2 (x x0 ) T Q(x x 0 ) Kangweon-Kyungk Math. Jour. 4 1996), No. 1, pp. 7 16 AN ITERATIVE ROW-ACTION METHOD FOR MULTICOMMODITY TRANSPORTATION PROBLEMS Yong Joon Ryang Abstract. The optmzaton problems wth quadratc constrants often

More information

Copyright 2014 Tech Science Press CMC, vol.43, no.2, pp.87-95, 2014

Copyright 2014 Tech Science Press CMC, vol.43, no.2, pp.87-95, 2014 Copyrght 2014 Tech Scence Press CMC, vol.43, no.2, pp.87-95, 2014 Analytcal Treatment of the Isotropc and Tetragonal Lattce Green Functons for the Face-centered Cubc, Body-centered Cubc and Smple Cubc

More information

Testing for seasonal unit roots in heterogeneous panels

Testing for seasonal unit roots in heterogeneous panels Testng for seasonal unt roots n heterogeneous panels Jesus Otero * Facultad de Economía Unversdad del Rosaro, Colomba Jeremy Smth Department of Economcs Unversty of arwck Monca Gulett Aston Busness School

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Maxmum Lkelhood Estmaton INFO-2301: Quanttatve Reasonng 2 Mchael Paul and Jordan Boyd-Graber MARCH 7, 2017 INFO-2301: Quanttatve Reasonng 2 Paul and Boyd-Graber Maxmum Lkelhood Estmaton 1 of 9 Why MLE?

More information

CIVL 8/7117 Chapter 10 - Isoparametric Formulation 42/56

CIVL 8/7117 Chapter 10 - Isoparametric Formulation 42/56 CIVL 8/77 Chapter 0 - Isoparametrc Formulaton 4/56 Newton-Cotes Example Usng the Newton-Cotes method wth = ntervals (n = 3 samplng ponts), evaluate the ntegrals: x x cos dx 3 x x dx 3 x x dx 4.3333333

More information

Suppose that there s a measured wndow of data fff k () ; :::; ff k g of a sze w, measured dscretely wth varable dscretzaton step. It s convenent to pl

Suppose that there s a measured wndow of data fff k () ; :::; ff k g of a sze w, measured dscretely wth varable dscretzaton step. It s convenent to pl RECURSIVE SPLINE INTERPOLATION METHOD FOR REAL TIME ENGINE CONTROL APPLICATIONS A. Stotsky Volvo Car Corporaton Engne Desgn and Development Dept. 97542, HA1N, SE- 405 31 Gothenburg Sweden. Emal: astotsky@volvocars.com

More information

Least squares cubic splines without B-splines S.K. Lucas

Least squares cubic splines without B-splines S.K. Lucas Least squares cubc splnes wthout B-splnes S.K. Lucas School of Mathematcs and Statstcs, Unversty of South Australa, Mawson Lakes SA 595 e-mal: stephen.lucas@unsa.edu.au Submtted to the Gazette of the Australan

More information

International Journal of Pure and Applied Sciences and Technology

International Journal of Pure and Applied Sciences and Technology Int. J. Pure Appl. Sc. Technol., 4() (03), pp. 5-30 Internatonal Journal of Pure and Appled Scences and Technology ISSN 9-607 Avalable onlne at www.jopaasat.n Research Paper Schrödnger State Space Matrx

More information

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin

Finite Mixture Models and Expectation Maximization. Most slides are from: Dr. Mario Figueiredo, Dr. Anil Jain and Dr. Rong Jin Fnte Mxture Models and Expectaton Maxmzaton Most sldes are from: Dr. Maro Fgueredo, Dr. Anl Jan and Dr. Rong Jn Recall: The Supervsed Learnng Problem Gven a set of n samples X {(x, y )},,,n Chapter 3 of

More information

Chapter 8 Indicator Variables

Chapter 8 Indicator Variables Chapter 8 Indcator Varables In general, e explanatory varables n any regresson analyss are assumed to be quanttatve n nature. For example, e varables lke temperature, dstance, age etc. are quanttatve n

More information

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore

8/25/17. Data Modeling. Data Modeling. Data Modeling. Patrice Koehl Department of Biological Sciences National University of Singapore 8/5/17 Data Modelng Patrce Koehl Department of Bologcal Scences atonal Unversty of Sngapore http://www.cs.ucdavs.edu/~koehl/teachng/bl59 koehl@cs.ucdavs.edu Data Modelng Ø Data Modelng: least squares Ø

More information

Simulation and Random Number Generation

Simulation and Random Number Generation Smulaton and Random Number Generaton Summary Dscrete Tme vs Dscrete Event Smulaton Random number generaton Generatng a random sequence Generatng random varates from a Unform dstrbuton Testng the qualty

More information

Solving Nonlinear Differential Equations by a Neural Network Method

Solving Nonlinear Differential Equations by a Neural Network Method Solvng Nonlnear Dfferental Equatons by a Neural Network Method Luce P. Aarts and Peter Van der Veer Delft Unversty of Technology, Faculty of Cvlengneerng and Geoscences, Secton of Cvlengneerng Informatcs,

More information

arxiv:cs.cv/ Jun 2000

arxiv:cs.cv/ Jun 2000 Correlaton over Decomposed Sgnals: A Non-Lnear Approach to Fast and Effectve Sequences Comparson Lucano da Fontoura Costa arxv:cs.cv/0006040 28 Jun 2000 Cybernetc Vson Research Group IFSC Unversty of São

More information

4DVAR, according to the name, is a four-dimensional variational method.

4DVAR, according to the name, is a four-dimensional variational method. 4D-Varatonal Data Assmlaton (4D-Var) 4DVAR, accordng to the name, s a four-dmensonal varatonal method. 4D-Var s actually a drect generalzaton of 3D-Var to handle observatons that are dstrbuted n tme. The

More information

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models

Maximum Likelihood Estimation of Binary Dependent Variables Models: Probit and Logit. 1. General Formulation of Binary Dependent Variables Models ECO 452 -- OE 4: Probt and Logt Models ECO 452 -- OE 4 Maxmum Lkelhood Estmaton of Bnary Dependent Varables Models: Probt and Logt hs note demonstrates how to formulate bnary dependent varables models

More information

Classification as a Regression Problem

Classification as a Regression Problem Target varable y C C, C,, ; Classfcaton as a Regresson Problem { }, 3 L C K To treat classfcaton as a regresson problem we should transform the target y nto numercal values; The choce of numercal class

More information

= z 20 z n. (k 20) + 4 z k = 4

= z 20 z n. (k 20) + 4 z k = 4 Problem Set #7 solutons 7.2.. (a Fnd the coeffcent of z k n (z + z 5 + z 6 + z 7 + 5, k 20. We use the known seres expanson ( n+l ( z l l z n below: (z + z 5 + z 6 + z 7 + 5 (z 5 ( + z + z 2 + z + 5 5

More information

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD

Matrix Approximation via Sampling, Subspace Embedding. 1 Solving Linear Systems Using SVD Matrx Approxmaton va Samplng, Subspace Embeddng Lecturer: Anup Rao Scrbe: Rashth Sharma, Peng Zhang 0/01/016 1 Solvng Lnear Systems Usng SVD Two applcatons of SVD have been covered so far. Today we loo

More information

Lecture 21: Numerical methods for pricing American type derivatives

Lecture 21: Numerical methods for pricing American type derivatives Lecture 21: Numercal methods for prcng Amercan type dervatves Xaoguang Wang STAT 598W Aprl 10th, 2014 (STAT 598W) Lecture 21 1 / 26 Outlne 1 Fnte Dfference Method Explct Method Penalty Method (STAT 598W)

More information

Convexity preserving interpolation by splines of arbitrary degree

Convexity preserving interpolation by splines of arbitrary degree Computer Scence Journal of Moldova, vol.18, no.1(52), 2010 Convexty preservng nterpolaton by splnes of arbtrary degree Igor Verlan Abstract In the present paper an algorthm of C 2 nterpolaton of dscrete

More information

Kernel Methods and SVMs Extension

Kernel Methods and SVMs Extension Kernel Methods and SVMs Extenson The purpose of ths document s to revew materal covered n Machne Learnng 1 Supervsed Learnng regardng support vector machnes (SVMs). Ths document also provdes a general

More information

COMPOSITE BEAM WITH WEAK SHEAR CONNECTION SUBJECTED TO THERMAL LOAD

COMPOSITE BEAM WITH WEAK SHEAR CONNECTION SUBJECTED TO THERMAL LOAD COMPOSITE BEAM WITH WEAK SHEAR CONNECTION SUBJECTED TO THERMAL LOAD Ákos Jósef Lengyel, István Ecsed Assstant Lecturer, Professor of Mechancs, Insttute of Appled Mechancs, Unversty of Mskolc, Mskolc-Egyetemváros,

More information

The Geometry of Logit and Probit

The Geometry of Logit and Probit The Geometry of Logt and Probt Ths short note s meant as a supplement to Chapters and 3 of Spatal Models of Parlamentary Votng and the notaton and reference to fgures n the text below s to those two chapters.

More information

An Improved multiple fractal algorithm

An Improved multiple fractal algorithm Advanced Scence and Technology Letters Vol.31 (MulGraB 213), pp.184-188 http://dx.do.org/1.1427/astl.213.31.41 An Improved multple fractal algorthm Yun Ln, Xaochu Xu, Jnfeng Pang College of Informaton

More information

OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION. Christophe De Luigi and Eric Moreau

OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION. Christophe De Luigi and Eric Moreau OPTIMAL COMBINATION OF FOURTH ORDER STATISTICS FOR NON-CIRCULAR SOURCE SEPARATION Chrstophe De Lug and Erc Moreau Unversty of Toulon LSEET UMR CNRS 607 av. G. Pompdou BP56 F-8362 La Valette du Var Cedex

More information