An Optimization Method for Numerically Solving Three-point BVPs of Linear Second-order ODEs with Variable Coefficients

Size: px
Start display at page:

Download "An Optimization Method for Numerically Solving Three-point BVPs of Linear Second-order ODEs with Variable Coefficients"

Transcription

1 An Optimization Method for Numericall Solving Three-point BVPs of Linear Second-order ODEs with Variable Coefficients Liaocheng Universit School of Mathematics Sciences , Liaocheng P.R. CHINA Abstract: It is known that most numerical methods for solving differential equations are based on iterative methods or Talor epansion methods. This paper tries to stud a numerical method from a new perspective optimization method. B means of the idea of kernel ε-svr, the paper constructs an optimization model for a class of threepoint boundar value problems (BVPs) of linear second-order ordinar differential equations (ODEs) with variable coefficients and proposes a novel numerical method for solving them. The proposed method has a certain versatilit and can be used to solve some other kinds of differential equations and integral equations. In order to verif the effectiveness of the proposed method, comparative eperiments with si specific linear second-order ODEs are performed. Eperimental results show that the proposed method has a good approimation propert. Ke Words: Optimization modeling, numerical method, ordinar differential equation, kernel ε-support vector regression, Lagrange function Introduction Differential equations can be found in the mathematical formulation of phsical phenomena in a wide variet of applications especiall in science and engineering. Depending on the form of the boundar conditions to be satisfied b the solution, problems involving ordinar differential equations (ODEs) can be divided into two main categories, initial value problems (IVPs) and boundar value problems (BVPs). Eact solutions for these problems are not generall available and hence numerical methods must be applied. Multi-point BVPs for ODEs can arise in solving linear partial differential equations b using the separation variable method []. Moreover, in the engineering problem to increase the stabilit of a rod, one also imposes a fied interior point ecept for the ends of the rod [2]. Along this line, the solutions of multipoint BVPs for ODEs have great significance in mathematical theor and practical applications [3-4]. It is seen that the three-point BVPs of nonlinear secondorder ODEs have attracted much attention [5-8] and it is significant to provide the solutions and Green s functions for these problems [9-0]. As shown in mentioned works, the eistence and uniqueness of the solution are alwas focused on b using a fied point theorem. However, from the viewpoint of practical Corresponding author. applications, one should provide eplicit solutions or s for multi-point BVPs. In this paper, inspired b work of S. Mehrkanoon et. al [], we tr from a new perspective to stud a class of threepoint BVPs of linear second-order ODEs with variable coefficients () + p() () + q()() = g(), [a, b], (a) = p 0, (b) + λ(µ) = q 0, µ (a, b), () where the known functions p() C [a, b], q(), g() C[a, b] and λ, µ are the constants, and research how to obtain its approimal solutions b using an optimization method based on kernel support vector regression (KSVR). Kernel support vector regressions (KSVRs) based on optimization modelings are a powerful methodolog for solving pattern recognition and function estimation problems, which based on Vapnik and Chervonenkis structural risk imization principle [2]. The show better generalization abilit compared to other machine learning methods on a wide variet of real-world problems, such as optimal control [3], image segmentation [4], image denoising [5], time series prediction [6], cber securit problems [7], network traffic predictive [8] and so on. E-ISSN: Volume, 205

2 KSVRs are used to deal with nonlinear regression problems b means of kernel skill. Data can be mapped into a high dimensional feature space b using a kernel function and then linear regression problems can be solved, which results in solving quadratic programg problems (QPPs). The main challenge in developing a useful regression modeling is to capture accuratel the underling functional relationship between the given inputs and their output values. Once the resulting model is obtained, it can be used as a tool for analsis, simulation and prediction. KSVR aims at detering a regressor for a given set of data points and a kernel function. The basic idea of kernel ε-svr with ε-insensitive loss function proposed b Vapnik [2] is to find a linear function () in the higher dimensional feature space such that, on the one hand, more mapped data samples locate in the ε- intensive tube between () ε and () + ε and on the other hand, the function () is as flat as possible, which leads to introduce the regularization term. Thus, the structural risk imization principle is implemented. The paper mainl researches how to use kernel ε- SVR method to obtain numerical solutions of Eq.(). It can be seen from the derivation process that some IVPs and BVPs of general linear ODEs ma be studied b using the similar wa. The paper does not involve nonlinear ODEs, which will be our net work. The remainder of this paper is organized as follows: Section 2 briefl reviews kernel ε-svr and some basic concepts. Section 3 presents a new approimation method for the solution of Eq.(). In order to verif the effectiveness of the presented method, a series of comparative eperiments with si specific linear second-order ODEs are performed in Section 4 and some concluding remarks are given in Section 5. 2 Kernel ε-svr This section briefl recalls kernel ε-svr, for details see [2]. Let T = {( i, i )} l i= be a set of sample data, where i R n and i R are input and output of the ith sample data respectivel. Let k : R n R n R be a kernel function with the reproducing kernel Hilbert space (RKHS) H and the nonlinear feature mapping φ : R n H. The space H is a higher dimensional space and can be epressed as an epended space of the mapped inputs {φ( i )} l i=, that is, H = span{φ( ),, φ( l )}. It is known that k(u, v) = φ(u), φ(v) for all u, v R n, where, denotes the inner product in RKBS H. Let K = [k( i, j )] R l l be the kernel matri and K i denote the ith column of K. ε-svr is based on ε-insensitive loss function defined b L ε () = { 0 if f () ε, f () ε otherwise, where f () and are the predicted value and true value of the input R n, respectivel. The regression function obtained b using kernel ε-svr has the form f () = w T φ() + b, where unknown normal vector w H and bias b R can be obtained b solving the following optimization problem: w,b 2 w 2 + C l (ξ i + ξ 2i ), i= s.t. w T φ( i ) + b i ε + ξ i, i =,, l, i (w T φ( i ) + b) ε + ξ 2i, i =,, l, ξ i, ξ 2i 0, i =,, l, (2) where C > 0 is a pre-specified value and ξ i, ξ 2i, i =,, l are slack variables. Due to H = span{φ( ),, φ( l )}, we can set w = l i= β i φ( i ). Put β = (β,, β l ) T, = (,, l ) T, e l = (,, ) T R l, ξ = (ξ,, ξ l ) T R l, ξ 2 = (ξ 2,, ξ 2l ) T R l, then w T φ( i ) = Ki T β and the problem (2) can be rewritten as the matri form: w,b 2 βt Kβ + Ce T l (ξ + ξ 2 ), s.t. Kβ + be l εe l + ξ, (Kβ + be l ) εe l + ξ 2, ξ, ξ 2 0. B solving the Wolfe dual form of the problem (3): α,α 2 2 (α α 2 ) T K(α α 2 ) T (α α 2 ) +εe T l (α + α 2 ) s.t. e T l (α α 2 ) = 0, 0 α, α 2 Ce l, we can obtain the optimal solution (α, α 2 ) and then or β = α α 2, b = i K T i β + ε for some i : 0 < α i < C, b = k K T k β ε for some k : 0 < α 2k < C. Therefore, the optimal regression function is f () = [k(, ),, k( l, )]β + b. (3) E-ISSN: Volume, 205

3 3 A numerical method based on Kernel ε-svr It is known that KSVRs based on optimization model are a powerful methodolog for solving function estimation problems and show better generalization abilit than other machine learning methods on a wide variet of real-world problems. But the main challenge in developing a useful regression model is to capture the underling functional relationship between the given inputs and their output values accuratel. This section tries to stud how to seek the numerical solutions of Eq.() b using kernel ε-svr. Specificall, it assumes that numerical solutions of Eq.() have the form ŷ() = w T φ() + b for a given kernel function k : R R R with RKHS H and feature mapping φ : R H and then learns unknown w H and b R b kernel ε-svr. For this end, the domain [a, b] of Eq.() needs to be discretized into a set of collocation points a = < 2 < < l = b with the same stepsize h = b a l such that µ = n (, l ), and then the points { i } l = are selected as input values. Since there are no available output values to learn (w, b) from Eq.(), we have to substitute the function ŷ() = w T φ() + b into Eq.(). So, it needs to define the derivative of the kernel function. Making use of the Mercer s Theorem [9], derivatives of the feature map can be written in terms of derivatives of the kernel function [20]. Define the following differential operator which will be used in subsequent sections, m n k(u, v) = n+m k(u, v) u n, u, v R, n, m =, 2,. vm B the definition of kernel function, it has and then m n k(u, v) = (φ (n) (u)) T φ (m) (v), ŷ() = w T φ() + b = l j= β j φ( j ) T φ() + b = l j= β j k( j, ) + b, ŷ () = w T φ () = l j= β j φ( j ) T φ () = l j= β j 0 k( j, ), ŷ () = w T φ () = l j= β j φ( j ) T φ () = l j= β j 2 0 k( j, ). In the sequel, put K 0 0 = [ 0 0 k( i, j )] = [k( i, j )] = K R l l, K m n = [ m n k( i, j )] = [ m n k(u, v) u=i,v= j ] R l l, and denote b (K m n ) i and K i the ith columns of the matrices K m n and K, respectivel. Let L(()) = () + p() () + q()(). If ŷ() is an eact solution of Eq.(), then ŷ(a) = p 0, ŷ(b) + λŷ(µ) = q 0 and L(ŷ( i )) = g( i ) for all i =,, l. But in general, ŷ(a), ŷ(b) + λŷ(µ) and L(ŷ( i )) are not necessaril equal to p 0, q 0 and g( i ), respectivel. So, we hope the smaller the better of the absolute values ŷ(a) p 0, ŷ(b) + λŷ(µ) q 0 and L(ŷ( i )) g( i ) for i =,, l. For this end, choose {g( i )} l i= as output values and construct the following optimization problem: w,b 2 w 2 + C l i= (ξ i + ξ 2i ), s.t. L(ŷ( i )) g( i ) ε + ξ i, ξ i 0, i =,, l, g( i ) L(ŷ( i )) ε + ξ 2i, ξ 2i 0, i =,, l, ŷ( ) = p 0, ŷ( l ) + λŷ( n ) = q 0, (4) where C > 0 is a pre-specified value and ξ i, ξ 2i, i =,, l are slack variables. In the problem (4), it assumes the boundar conditions being satisfied and hopes the most values of {L(ŷ( i ))} l i= satisfing L(ŷ( i)) g( i ) < ε for some given sufficientl small ε > 0, which indicates that the errors between eact values and estimated values at most of discrete points { i } l i= are smaller than ε. B discussion in the previous section, it knows that the numerical solution ŷ() = w T φ() + b is continuous if the kernel function k : R R R is continuous. According to the smbol propert of continuous functions, it can conclude that the error between eact solution and numerical solution in the whole domain [a, b] is smaller than ε as long as the stepsize h is taken sufficientl small. In general, ε is often taken as In this paper, take ε = 0 2. According to the discussion in Section 2, we can let w = l j= β j φ( j ) and then deduce that ŷ( i ) = l j= β j k( j, i ) + b = K T i β + b, ŷ( l ) + λŷ( n ) = (K l + λk n ) T β + ( + λ)b, ŷ ( i ) = (K 0 )T i β, ŷ ( i ) = (K 2 0 )T i β, L(ŷ( i )) = [(K 2 0 ) i + p( i )(K 0 ) i + q( i )K i ] T β + q( i )b, i =,, l. Put m i = (K 2 0 ) i + p( i )(K 0 ) i + q( i )K i R l, i =,, l, M = [m,, m l ] R l l, g = (g( ),, g( l )) T R l, q = (q( ),, q( l )) T R l, then the problem (4) can be rewritten as the matri E-ISSN: Volume, 205

4 form: β,b,ξ,ξ 2 βt Kβ + Ce T l (ξ + ξ 2 ) 2 s.t. M T β + qb g εe l + ξ, ξ 0, g M T β qb εe l + ξ 2, ξ 2 0, K T β + b = p 0, (K l + λk n ) T β + ( + λ)b = q 0. (5) Considering the Lagrange function of the problem (5) L(β, b, ξ, ξ 2, η, η 2, ς, ς 2, α, α 2 ) = 2 βt Kβ + Ce T l (ξ + ξ 2 ) +η T (MT β + qb g εe l ξ ) +η T 2 (g MT β qb εe l ξ 2 ) ς T ξ ς 2 T ξ 2 + α (K T β + b p 0) +α 2 ((K l + λk n ) T β + ( + λ)b q 0 ), where η, η 2, ς, ς 2 R l and α, α 2 R are the vectors of Lagrange multipliers, and letting L L ξ 2 = 0, it can deduce that β = L b = L ξ = Kβ + M(η η 2 ) + α K + α 2 (K l + λk n ) = 0, q T (η η 2 ) + α + ( + λ)α 2 = 0, Ce l η ς = 0 0 η Ce l, Ce l η 2 ς 2 = 0 0 η 2 Ce l. (6) Since the kernel matri K is smmetric and nonnegative definite, it can be assumed as a nonsingular matri without loss of generalit. Otherwise, to take care of problems due to possible ill-conditioning, a regularization term K + δi l can be introduced with a sufficientl small number δ > 0 and a l order unit matri I l. Put D = [M, M, K, K l + λk n ] R l (2l+2), γ = [η T, ηt 2, α, α 2 ] T R 2l+2, d = [q T, q T,, ( + λ)] T R 2l+2. It can get β = K Dγ from (6). Substituting (6) and β = K Dγ into the Lagrange function, it has where L(γ) = 2 γt D T K Dγ z T γ, z = ((εe l + g) T, (εe l g) T, p 0, q 0 ) T R 2l+2. Consequentl, the Wolfe dual form of the problem (5) can be written as γ 2 γt D T K Dγ + z T γ s.t. d T γ = 0, 0 γ j C, j =,, 2l. (7) After getting the optimal solution γ of the problem (7), it has β = K Dγ, b = p 0 K T β, and then the numerical solution of Eq.() is ŷ() = [k(, ),, k( l, )]β + b. The specific procedure is as follows. Algorithm. Step. Discrete the domain [a, b] b a = < 2 < < l = b with a sufficientl small same stepsize h = b a l such that µ = n (, l ). Step 2. Select a proper kernel function. Step 3. Choose proper kernel parameters and model parameters. Step 4. Solve the problem (7) and obtain the optimal solution γ. Step 5. Compute β = K Dγ and b = p 0 K T β. Step 6. construct the numerical solution of Eq.() b ŷ() = [k(, ),, k( l, )]β + b. 4 Numerical eamples In order to demonstrate the effectiveness of Algorithm, in this section, a series of comparative eperiments are performed between numerical solutions and eact solutions in the following si elaborated three-point BVPs of linear second-order ODEs with variable coefficients. The first five eamples come from a the second kind Fredholm integral equation [0] ϕ() 2 0 ( + )e ϕ()d = e e (+), 0, and have the same eact solution ϕ() = e. The sith eample is also taken from [0] and has the eact solution ϕ() = e. Eample. Consider a three-point BVP of linear { ϕ () + ϕ () + ϕ() = e, 0, ϕ(0) =, ϕ() + ϕ(/2) = e + e /2. Eample 2. Consider a three-point BVP of linear { ϕ () ( + sin )ϕ() = e sin, 0, ϕ(0) =, ϕ() + ϕ(/2) = e + e /2. Eample 3. Consider a three-point BVP of linear { ϕ () + 2 ϕ () ϕ() = ( 2 )e, 0, ϕ(0) =, ϕ() + ϕ(/2) = e + e /2. E-ISSN: Volume, 205

5 Eample 4. Consider a three-point BVP of linear { ϕ () + ( + )ϕ () = e, 0, ϕ(0) =, ϕ() + ϕ(/2) = e + e /2. Eample 5. Consider a three-point BVP of linear ϕ () + sin ϕ () + 2 ϕ() = ( sin + 2 )e, 0, ϕ(0) =, ϕ() + ϕ(/2) = e + e /2. Eample 6. Consider a three-point BVP for a { ϕ () ( + sin )ϕ() = e sin, [0, ], ϕ(0) =, ϕ() + ϕ(/2) = e + e /2. All computations are implemented in Matlab 200b on a PC with 2.5 GHz CPU and 4 G btes memor. The Gaussian RBF kernel function k(u, v) = ep{ (u v)2 2σ 2 }, u, v R is chosen in all eperiments and the model parameter C and kernel parameter σ are taken as C = and σ = 2. The eperiment results are listed in Table and the diagrams of comparisons are listed in Figure. Because most of known methods for solving differential equations numericall are focus on Talor epansion or discrete approaches, almost no use of optimization method, it can onl compare the errors between numerical solutions and eact solutions in this paper. From Table, it can see that the proposed numerical method has good accurac, which can be seen more obvious from Figure. According to the eperiment results, we can conclude that the proposed method is feasible and effective for solving numerical three-point BVPs of linear second-order ODEs with variable coefficients. 5 Conclusions This paper tries to seek the numerical solutions of three-point BVPs of linear second-order ODEs with variable coefficients b using optimization technolog. Although kernel ε-svr based on optimization model is a powerful methodolog for solving function estimation problems, the main challenge in developing a useful regression model is to capture the underling functional relationship between the given inputs and their output values accuratel. In order to construct optimization model corresponding to kernel ε-svr, this paper selects the discrete points of the domain of Eq.() as input values and selects values of the function L(ŷ()) at the discrete points as output values. B solving the Wolfe dual problem of the model, a numerical method is proposed. eperiment results indicates that the proposed method has a good approimation propert. From the derivation process, it can be seen that the proposed method has a certain versatilit and can be used to stud some other kinds of linear ODEs. The paper onl involves linear ODEs and nonlinear ODEs problems will be our net work. 6 Acknowledgements The research was supported b Science and Technolog Project of Higher Education of Shandong Province, P.R. China (grant No.L3LI0). References: [] F.V. Atkinson, Multiparameter Eigenvalue Problems, Academic Press, New York, 972. [2] A. Boucherif, Nonlinear three-point boundar value problems, J. Math. Anal. Appl. vol. 77, pp , 980. [3] C.Z. Bai, J.X. Fang, Eistence of multiple positive solutions for nonlinear m-point boundarvalue problems, Appl. Math. Comput. vol. 40, pp , [4] X.J. Liu, J.Q. Liu, Three positive solutions for second-order m-point boundar value problems, Appl. Math. Comput. vol. 56, no. 3, pp , [5] X. Xu, Multiplicit results for positive solutions of some semi-positone three-point boundar value problems, J. Math. Anal. Appl. vol. 29, pp , [6] X. Xu, Three solutions for three-point boundar value problems, Nonlinear Anal. vol. 62, pp , [7] H. Chen, B. Li, Three-point boundar value problems for second-order ordinar differential equations in Banach spaces, Comput. Math. Appl. vol. 56, pp , [8] P.D. Phung, L.X. Truong, On the eistence of a three point boundar value problem at resonance in Rn, J. Math. Anal. Appl. vol. 46, pp , 204. E-ISSN: Volume, 205

6 (a) Comparison for Eample. Eact solution Numerical solution Absolute error Table : Comparison for si eamples (b) Comparison for Eample 2. Eact solution Numerical solution Absolute error (c) Comparison for Eample 3. Eact solution Numerical solution Absolute error (e) Comparison for Eample 5. Eact solution Numerical solution Absolute error (d) Comparison for Eample 4. Eact solution Numerical solution Absolute error (f) Comparison for Eample 6. Eact solution Numerical solution Absolute error E-ISSN: Volume, 205

7 WSEAS TRANSACTIONS on SIGNAL PROCESSING eact solution eact solution. eact solution (a) Comparison for Eample (b) Comparison for Eample (c) Comparison for Eample 3. eact solution eact solution eact solution (d) Comparison for Eample (e) Comparison for Eample (f) Comparison for Eample 6. Figure : diagrams of comparisons for si eamples [9] Z. Zhao, Solutions and Greens functions for some linear second-order three-point boundar value problems, Comput. Math. Appl. vol. 56, pp. 04-3, [0] Xian-Ci Zhong, Qiong-Ao Huang, Approimate solution of three-point boundar value problems for second-order ordinar differential equations with variable coefficients, Applied Mathematics and Computation, vol. 247, pp. 8-29, 204. [] S. Mehrkanoon, T. Falck, J.A.K. Sukens, Approimate Solutions to Ordinar Differential Equations Using Least Squares Support Vector Machines, IEEE Transactions on Neural Networks and Learning Sstems, vol. 23, no. 9, pp , 202. [2] V. N.Vapnik, The Nature of Statistical Learning Theor, Springer, New York, [3] J. A. K. Sukens, J. Vandewalle, B. De Moor, Optimal control b least squares support vector machines, Neural Networks, vol. 4, no., pp , 200. [4] S. Chen, M. Wang, Seeking multi-thresholds directl from support vectors for image segmentation, Neurocomputing, vol. 67, pp , [5] Z. Liu, S. Xu, C. L. P. Chen, Y. Zhang, X. Chen, Y. Wang, A three-domain fuzz support vector regression for image denoising and eperimental studies. IEEE Transactions on Cbernetics, vol. 44, no. 4, pp , 204. [6] K. W. Lau, Q. H. Wu,. Local prediction of nonlinear time series using support vector regression, Pattern Recognition, vol. 4, no. 5, pp , [7] Y. Shang, Optimal control strategies for virus spreading in inhomogeneous epidemic dnamics. Canadian Mathematical Bulletin, vol. 56, no. 3, pp , 203. [8] A. K. Sharma, P. S. Parihar, An effective DoS prevention sstem to analsis and prediction of network traffic using support vector machine learning. International Journal of Application or Innovation in Engineering and Management, vol. 2, no. 7, pp , 203. [9] V. Vapnik, Statistical learning theor, New York: Wile, 998. [20] M. Lázaro, I. Santamaria, F. Pérez-Cruz, A. Artés-Rodriguez, Support vector regression for the simultaneous learning of a multivariate function and its derivative, Neurocomputing, vol. 69, pp. 42-6, E-ISSN: Volume, 205

Stability Analysis for Linear Systems under State Constraints

Stability Analysis for Linear Systems under State Constraints Stabilit Analsis for Linear Sstems under State Constraints Haijun Fang Abstract This paper revisits the problem of stabilit analsis for linear sstems under state constraints New and less conservative sufficient

More information

Duality, Geometry, and Support Vector Regression

Duality, Geometry, and Support Vector Regression ualit, Geometr, and Support Vector Regression Jinbo Bi and Kristin P. Bennett epartment of Mathematical Sciences Rensselaer Poltechnic Institute Tro, NY 80 bij@rpi.edu, bennek@rpi.edu Abstract We develop

More information

Introduction to Differential Equations. National Chiao Tung University Chun-Jen Tsai 9/14/2011

Introduction to Differential Equations. National Chiao Tung University Chun-Jen Tsai 9/14/2011 Introduction to Differential Equations National Chiao Tung Universit Chun-Jen Tsai 9/14/011 Differential Equations Definition: An equation containing the derivatives of one or more dependent variables,

More information

Krylov Integration Factor Method on Sparse Grids for High Spatial Dimension Convection Diffusion Equations

Krylov Integration Factor Method on Sparse Grids for High Spatial Dimension Convection Diffusion Equations DOI.7/s95-6-6-7 Krlov Integration Factor Method on Sparse Grids for High Spatial Dimension Convection Diffusion Equations Dong Lu Yong-Tao Zhang Received: September 5 / Revised: 9 March 6 / Accepted: 8

More information

SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE

SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE SUPPORT VECTOR MACHINE FOR THE SIMULTANEOUS APPROXIMATION OF A FUNCTION AND ITS DERIVATIVE M. Lázaro 1, I. Santamaría 2, F. Pérez-Cruz 1, A. Artés-Rodríguez 1 1 Departamento de Teoría de la Señal y Comunicaciones

More information

Applications of Gauss-Radau and Gauss-Lobatto Numerical Integrations Over a Four Node Quadrilateral Finite Element

Applications of Gauss-Radau and Gauss-Lobatto Numerical Integrations Over a Four Node Quadrilateral Finite Element Avaiable online at www.banglaol.info angladesh J. Sci. Ind. Res. (), 77-86, 008 ANGLADESH JOURNAL OF SCIENTIFIC AND INDUSTRIAL RESEARCH CSIR E-mail: bsir07gmail.com Abstract Applications of Gauss-Radau

More information

AN EFFICIENT MOVING MESH METHOD FOR A MODEL OF TURBULENT FLOW IN CIRCULAR TUBES *

AN EFFICIENT MOVING MESH METHOD FOR A MODEL OF TURBULENT FLOW IN CIRCULAR TUBES * Journal of Computational Mathematics, Vol.27, No.2-3, 29, 388 399. AN EFFICIENT MOVING MESH METHOD FOR A MODEL OF TURBULENT FLOW IN CIRCULAR TUBES * Yin Yang Hunan Ke Laborator for Computation and Simulation

More information

Iteratively Reweighted Least Square for Asymmetric L 2 -Loss Support Vector Regression

Iteratively Reweighted Least Square for Asymmetric L 2 -Loss Support Vector Regression Iteratively Reweighted Least Square for Asymmetric L 2 -Loss Support Vector Regression Songfeng Zheng Department of Mathematics Missouri State University Springfield, MO 65897 SongfengZheng@MissouriState.edu

More information

Introduction to Differential Equations

Introduction to Differential Equations Math0 Lecture # Introduction to Differential Equations Basic definitions Definition : (What is a DE?) A differential equation (DE) is an equation that involves some of the derivatives (or differentials)

More information

c 1999 Society for Industrial and Applied Mathematics

c 1999 Society for Industrial and Applied Mathematics SIAM J. SCI. COMPUT. Vol., No. 6, pp. 978 994 c 999 Societ for Industrial and Applied Mathematics A STUDY OF MONITOR FUNCTIONS FOR TWO-DIMENSIONAL ADAPTIVE MESH GENERATION WEIMING CAO, WEIZHANG HUANG,

More information

OPTIMAL CONTROL FOR A PARABOLIC SYSTEM MODELLING CHEMOTAXIS

OPTIMAL CONTROL FOR A PARABOLIC SYSTEM MODELLING CHEMOTAXIS Trends in Mathematics Information Center for Mathematical Sciences Volume 6, Number 1, June, 23, Pages 45 49 OPTIMAL CONTROL FOR A PARABOLIC SYSTEM MODELLING CHEMOTAXIS SANG UK RYU Abstract. We stud the

More information

520 Chapter 9. Nonlinear Differential Equations and Stability. dt =

520 Chapter 9. Nonlinear Differential Equations and Stability. dt = 5 Chapter 9. Nonlinear Differential Equations and Stabilit dt L dθ. g cos θ cos α Wh was the negative square root chosen in the last equation? (b) If T is the natural period of oscillation, derive the

More information

Research Article Chaotic Attractor Generation via a Simple Linear Time-Varying System

Research Article Chaotic Attractor Generation via a Simple Linear Time-Varying System Discrete Dnamics in Nature and Societ Volume, Article ID 836, 8 pages doi:.//836 Research Article Chaotic Attractor Generation via a Simple Linear Time-Varing Sstem Baiu Ou and Desheng Liu Department of

More information

Discontinuous Galerkin method for a class of elliptic multi-scale problems

Discontinuous Galerkin method for a class of elliptic multi-scale problems INTERNATIONAL JOURNAL FOR NUMERICAL METHODS IN FLUIDS Int. J. Numer. Meth. Fluids 000; 00: 6 [Version: 00/09/8 v.0] Discontinuous Galerkin method for a class of elliptic multi-scale problems Ling Yuan

More information

In applications, we encounter many constrained optimization problems. Examples Basis pursuit: exact sparse recovery problem

In applications, we encounter many constrained optimization problems. Examples Basis pursuit: exact sparse recovery problem 1 Conve Analsis Main references: Vandenberghe UCLA): EECS236C - Optimiation methods for large scale sstems, http://www.seas.ucla.edu/ vandenbe/ee236c.html Parikh and Bod, Proimal algorithms, slides and

More information

Solving Variable-Coefficient Fourth-Order ODEs with Polynomial Nonlinearity by Symmetric Homotopy Method

Solving Variable-Coefficient Fourth-Order ODEs with Polynomial Nonlinearity by Symmetric Homotopy Method Applied and Computational Mathematics 218; 7(2): 58-7 http://www.sciencepublishinggroup.com/j/acm doi: 1.11648/j.acm.21872.14 ISSN: 2328-565 (Print); ISSN: 2328-5613 (Online) Solving Variable-Coefficient

More information

THE HEATED LAMINAR VERTICAL JET IN A LIQUID WITH POWER-LAW TEMPERATURE DEPENDENCE OF DENSITY. V. A. Sharifulin.

THE HEATED LAMINAR VERTICAL JET IN A LIQUID WITH POWER-LAW TEMPERATURE DEPENDENCE OF DENSITY. V. A. Sharifulin. THE HEATED LAMINAR VERTICAL JET IN A LIQUID WITH POWER-LAW TEMPERATURE DEPENDENCE OF DENSITY 1. Introduction V. A. Sharifulin Perm State Technical Universit, Perm, Russia e-mail: sharifulin@perm.ru Water

More information

Initial Value Problems for. Ordinary Differential Equations

Initial Value Problems for. Ordinary Differential Equations Initial Value Problems for Ordinar Differential Equations INTRODUCTION Equations which are composed of an unnown function and its derivatives are called differential equations. It becomes an initial value

More information

16.5. Maclaurin and Taylor Series. Introduction. Prerequisites. Learning Outcomes

16.5. Maclaurin and Taylor Series. Introduction. Prerequisites. Learning Outcomes Maclaurin and Talor Series 6.5 Introduction In this Section we eamine how functions ma be epressed in terms of power series. This is an etremel useful wa of epressing a function since (as we shall see)

More information

8.1 Exponents and Roots

8.1 Exponents and Roots Section 8. Eponents and Roots 75 8. Eponents and Roots Before defining the net famil of functions, the eponential functions, we will need to discuss eponent notation in detail. As we shall see, eponents

More information

A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong

A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES. Wei Chu, S. Sathiya Keerthi, Chong Jin Ong A GENERAL FORMULATION FOR SUPPORT VECTOR MACHINES Wei Chu, S. Sathiya Keerthi, Chong Jin Ong Control Division, Department of Mechanical Engineering, National University of Singapore 0 Kent Ridge Crescent,

More information

* τσ σκ. Supporting Text. A. Stability Analysis of System 2

* τσ σκ. Supporting Text. A. Stability Analysis of System 2 Supporting Tet A. Stabilit Analsis of Sstem In this Appendi, we stud the stabilit of the equilibria of sstem. If we redefine the sstem as, T when -, then there are at most three equilibria: E,, E κ -,,

More information

Introduction to Differential Equations

Introduction to Differential Equations Introduction to Differential Equations. Definitions and Terminolog.2 Initial-Value Problems.3 Differential Equations as Mathematical Models Chapter in Review The words differential and equations certainl

More information

ABOUT THE WEAK EFFICIENCIES IN VECTOR OPTIMIZATION

ABOUT THE WEAK EFFICIENCIES IN VECTOR OPTIMIZATION ABOUT THE WEAK EFFICIENCIES IN VECTOR OPTIMIZATION CRISTINA STAMATE Institute of Mathematics 8, Carol I street, Iasi 6600, Romania cstamate@mail.com Abstract: We present the principal properties of the

More information

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature

Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature Support Vector Regression (SVR) Descriptions of SVR in this discussion follow that in Refs. (2, 6, 7, 8, 9). The literature suggests the design variables should be normalized to a range of [-1,1] or [0,1].

More information

Elliptic Equations. Chapter Definitions. Contents. 4.2 Properties of Laplace s and Poisson s Equations

Elliptic Equations. Chapter Definitions. Contents. 4.2 Properties of Laplace s and Poisson s Equations 5 4. Properties of Laplace s and Poisson s Equations Chapter 4 Elliptic Equations Contents. Neumann conditions the normal derivative, / = n u is prescribed on the boundar second BP. In this case we have

More information

CONTINUOUS SPATIAL DATA ANALYSIS

CONTINUOUS SPATIAL DATA ANALYSIS CONTINUOUS SPATIAL DATA ANALSIS 1. Overview of Spatial Stochastic Processes The ke difference between continuous spatial data and point patterns is that there is now assumed to be a meaningful value, s

More information

GROUP CLASSIFICATION OF LINEAR SECOND-ORDER DELAY ORDINARY DIFFERENTIAL EQUATION

GROUP CLASSIFICATION OF LINEAR SECOND-ORDER DELAY ORDINARY DIFFERENTIAL EQUATION Proceedings of the nd IMT-GT Regional onference on Mathematics Statistics and Applications Universiti Sains Malasia GROUP LASSIFIATION OF LINEAR SEOND-ORDER DELAY ORDINARY DIFFERENTIAL EQUATION Prapart

More information

COMPRESSIVE (CS) [1] is an emerging framework,

COMPRESSIVE (CS) [1] is an emerging framework, 1 An Arithmetic Coding Scheme for Blocked-based Compressive Sensing of Images Min Gao arxiv:1604.06983v1 [cs.it] Apr 2016 Abstract Differential pulse-code modulation (DPCM) is recentl coupled with uniform

More information

Numerical Simulation of Stochastic Differential Equations: Lecture 2, Part 1. Recap: SDE. Euler Maruyama. Lecture 2, Part 1: Euler Maruyama

Numerical Simulation of Stochastic Differential Equations: Lecture 2, Part 1. Recap: SDE. Euler Maruyama. Lecture 2, Part 1: Euler Maruyama Numerical Simulation of Stochastic Differential Equations: Lecture, Part 1 Des Higham Department of Mathematics Universit of Strathclde Lecture, Part 1: Euler Maruama Definition of Euler Maruama Method

More information

A new chaotic attractor from general Lorenz system family and its electronic experimental implementation

A new chaotic attractor from general Lorenz system family and its electronic experimental implementation Turk J Elec Eng & Comp Sci, Vol.18, No.2, 2010, c TÜBİTAK doi:10.3906/elk-0906-67 A new chaotic attractor from general Loren sstem famil and its electronic eperimental implementation İhsan PEHLİVAN, Yılma

More information

Math 214 Spring problem set (a) Consider these two first order equations. (I) dy dx = x + 1 dy

Math 214 Spring problem set (a) Consider these two first order equations. (I) dy dx = x + 1 dy Math 4 Spring 08 problem set. (a) Consider these two first order equations. (I) d d = + d (II) d = Below are four direction fields. Match the differential equations above to their direction fields. Provide

More information

On the Extension of Goal-Oriented Error Estimation and Hierarchical Modeling to Discrete Lattice Models

On the Extension of Goal-Oriented Error Estimation and Hierarchical Modeling to Discrete Lattice Models On the Etension of Goal-Oriented Error Estimation and Hierarchical Modeling to Discrete Lattice Models J. T. Oden, S. Prudhomme, and P. Bauman Institute for Computational Engineering and Sciences The Universit

More information

4 Inverse function theorem

4 Inverse function theorem Tel Aviv Universit, 2013/14 Analsis-III,IV 53 4 Inverse function theorem 4a What is the problem................ 53 4b Simple observations before the theorem..... 54 4c The theorem.....................

More information

CIS 520: Machine Learning Oct 09, Kernel Methods

CIS 520: Machine Learning Oct 09, Kernel Methods CIS 520: Machine Learning Oct 09, 207 Kernel Methods Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the lecture They may or may not cover all the material discussed

More information

STABILITY AND NEIMARK-SACKER BIFURCATION OF A SEMI-DISCRETE POPULATION MODEL

STABILITY AND NEIMARK-SACKER BIFURCATION OF A SEMI-DISCRETE POPULATION MODEL Journal of Applied Analsis and Computation Website:http://jaac-online.com/ Volume 4, Number 4, November 14 pp. 419 45 STABILITY AND NEIMARK-SACKER BIFURCATION OF A SEMI-DISCRETE POPULATION MODEL Cheng

More information

RELATIONS AND FUNCTIONS through

RELATIONS AND FUNCTIONS through RELATIONS AND FUNCTIONS 11.1.2 through 11.1. Relations and Functions establish a correspondence between the input values (usuall ) and the output values (usuall ) according to the particular relation or

More information

Fully Discrete Energy Stable High Order Finite Difference Methods for Hyperbolic Problems in Deforming Domains: An Initial Investigation

Fully Discrete Energy Stable High Order Finite Difference Methods for Hyperbolic Problems in Deforming Domains: An Initial Investigation Full Discrete Energ Stable High Order Finite Difference Methods for Hperbolic Problems in Deforming Domains: An Initial Investigation Samira Nikkar and Jan Nordström Abstract A time-dependent coordinate

More information

An improved ADI-DQM based on Bernstein polynomial for solving two-dimensional convection-diffusion equations

An improved ADI-DQM based on Bernstein polynomial for solving two-dimensional convection-diffusion equations Mathematical Theor and Modeling ISSN -50 (Paper) ISSN 5-05(Online) Vol 3, No.1, 013 An improved ADI-DQM based on Bernstein polnomial for solving two-dimensional convection-diffusion equations A.S.J. Al-

More information

Engineering Mathematics I

Engineering Mathematics I Engineering Mathematics I_ 017 Engineering Mathematics I 1. Introduction to Differential Equations Dr. Rami Zakaria Terminolog Differential Equation Ordinar Differential Equations Partial Differential

More information

SEPARABLE EQUATIONS 2.2

SEPARABLE EQUATIONS 2.2 46 CHAPTER FIRST-ORDER DIFFERENTIAL EQUATIONS 4. Chemical Reactions When certain kinds of chemicals are combined, the rate at which the new compound is formed is modeled b the autonomous differential equation

More information

INTRODUCTION TO DIFFERENTIAL EQUATIONS

INTRODUCTION TO DIFFERENTIAL EQUATIONS INTRODUCTION TO DIFFERENTIAL EQUATIONS. Definitions and Terminolog. Initial-Value Problems.3 Differential Equations as Mathematical Models CHAPTER IN REVIEW The words differential and equations certainl

More information

Dynamics of multiple pendula without gravity

Dynamics of multiple pendula without gravity Chaotic Modeling and Simulation (CMSIM) 1: 57 67, 014 Dnamics of multiple pendula without gravit Wojciech Szumiński Institute of Phsics, Universit of Zielona Góra, Poland (E-mail: uz88szuminski@gmail.com)

More information

Support Vector Machine Regression for Volatile Stock Market Prediction

Support Vector Machine Regression for Volatile Stock Market Prediction Support Vector Machine Regression for Volatile Stock Market Prediction Haiqin Yang, Laiwan Chan, and Irwin King Department of Computer Science and Engineering The Chinese University of Hong Kong Shatin,

More information

2.2 SEPARABLE VARIABLES

2.2 SEPARABLE VARIABLES 44 CHAPTER FIRST-ORDER DIFFERENTIAL EQUATIONS 6 Consider the autonomous DE 6 Use our ideas from Problem 5 to find intervals on the -ais for which solution curves are concave up and intervals for which

More information

Generalized Least-Squares Regressions III: Further Theory and Classication

Generalized Least-Squares Regressions III: Further Theory and Classication Generalized Least-Squares Regressions III: Further Theor Classication NATANIEL GREENE Department of Mathematics Computer Science Kingsborough Communit College, CUNY Oriental Boulevard, Brookln, NY UNITED

More information

On the Missing Modes When Using the Exact Frequency Relationship between Kirchhoff and Mindlin Plates

On the Missing Modes When Using the Exact Frequency Relationship between Kirchhoff and Mindlin Plates On the Missing Modes When Using the Eact Frequenc Relationship between Kirchhoff and Mindlin Plates C.W. Lim 1,*, Z.R. Li 1, Y. Xiang, G.W. Wei 3 and C.M. Wang 4 1 Department of Building and Construction,

More information

CONSERVATION LAWS AND CONSERVED QUANTITIES FOR LAMINAR RADIAL JETS WITH SWIRL

CONSERVATION LAWS AND CONSERVED QUANTITIES FOR LAMINAR RADIAL JETS WITH SWIRL Mathematical and Computational Applications,Vol. 15, No. 4, pp. 742-761, 21. c Association for Scientific Research CONSERVATION LAWS AND CONSERVED QUANTITIES FOR LAMINAR RADIAL JETS WITH SWIRL R. Naz 1,

More information

METHODS IN Mathematica FOR SOLVING ORDINARY DIFFERENTIAL EQUATIONS. Keçiören, Ankara 06010, Turkey.

METHODS IN Mathematica FOR SOLVING ORDINARY DIFFERENTIAL EQUATIONS. Keçiören, Ankara 06010, Turkey. Mathematical and Computational Applications, Vol. 16, No. 4, pp. 784-796, 2011. Association for Scientific Research METHODS IN Mathematica FOR SOLVING ORDINARY DIFFERENTIAL EQUATIONS Ünal Göktaş 1 and

More information

On Maximizing the Second Smallest Eigenvalue of a State-dependent Graph Laplacian

On Maximizing the Second Smallest Eigenvalue of a State-dependent Graph Laplacian 25 American Control Conference June 8-, 25. Portland, OR, USA WeA3.6 On Maimizing the Second Smallest Eigenvalue of a State-dependent Graph Laplacian Yoonsoo Kim and Mehran Mesbahi Abstract We consider

More information

MA2264 -NUMERICAL METHODS UNIT V : INITIAL VALUE PROBLEMS FOR ORDINARY DIFFERENTIAL. By Dr.T.Kulandaivel Department of Applied Mathematics SVCE

MA2264 -NUMERICAL METHODS UNIT V : INITIAL VALUE PROBLEMS FOR ORDINARY DIFFERENTIAL. By Dr.T.Kulandaivel Department of Applied Mathematics SVCE MA64 -NUMERICAL METHODS UNIT V : INITIAL VALUE PROBLEMS FOR ORDINARY DIFFERENTIAL EQUATIONS B Dr.T.Kulandaivel Department of Applied Matematics SVCE Numerical ordinar differential equations is te part

More information

MATRIX TRANSFORMATIONS

MATRIX TRANSFORMATIONS CHAPTER 5. MATRIX TRANSFORMATIONS INSTITIÚID TEICNEOLAÍOCHTA CHEATHARLACH INSTITUTE OF TECHNOLOGY CARLOW MATRIX TRANSFORMATIONS Matri Transformations Definition Let A and B be sets. A function f : A B

More information

BASE VECTORS FOR SOLVING OF PARTIAL DIFFERENTIAL EQUATIONS

BASE VECTORS FOR SOLVING OF PARTIAL DIFFERENTIAL EQUATIONS BASE VECTORS FOR SOLVING OF PARTIAL DIFFERENTIAL EQUATIONS J. Roubal, V. Havlena Department of Control Engineering, Facult of Electrical Engineering, Czech Technical Universit in Prague Abstract The distributed

More information

MMJ1153 COMPUTATIONAL METHOD IN SOLID MECHANICS PRELIMINARIES TO FEM

MMJ1153 COMPUTATIONAL METHOD IN SOLID MECHANICS PRELIMINARIES TO FEM B Course Content: A INTRODUCTION AND OVERVIEW Numerical method and Computer-Aided Engineering; Phsical problems; Mathematical models; Finite element method;. B Elements and nodes, natural coordinates,

More information

Theory of two-dimensional transformations

Theory of two-dimensional transformations Calhoun: The NPS Institutional Archive DSpace Repositor Facult and Researchers Facult and Researchers Collection 1998-1 Theor of two-dimensional transformations Kanaama, Yutaka J.; Krahn, Gar W. http://hdl.handle.net/1945/419

More information

Indirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina

Indirect Rule Learning: Support Vector Machines. Donglin Zeng, Department of Biostatistics, University of North Carolina Indirect Rule Learning: Support Vector Machines Indirect learning: loss optimization It doesn t estimate the prediction rule f (x) directly, since most loss functions do not have explicit optimizers. Indirection

More information

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential

CHAPTER 2: Partial Derivatives. 2.2 Increments and Differential CHAPTER : Partial Derivatives.1 Definition of a Partial Derivative. Increments and Differential.3 Chain Rules.4 Local Etrema.5 Absolute Etrema 1 Chapter : Partial Derivatives.1 Definition of a Partial

More information

New approach to study the van der Pol equation for large damping

New approach to study the van der Pol equation for large damping Electronic Journal of Qualitative Theor of Differential Equations 2018, No. 8, 1 10; https://doi.org/10.1422/ejqtde.2018.1.8 www.math.u-szeged.hu/ejqtde/ New approach to stud the van der Pol equation for

More information

Linear programming: Theory

Linear programming: Theory Division of the Humanities and Social Sciences Ec 181 KC Border Convex Analsis and Economic Theor Winter 2018 Topic 28: Linear programming: Theor 28.1 The saddlepoint theorem for linear programming The

More information

Iran University of Science and Technology, Tehran 16844, IRAN

Iran University of Science and Technology, Tehran 16844, IRAN DECENTRALIZED DECOUPLED SLIDING-MODE CONTROL FOR TWO- DIMENSIONAL INVERTED PENDULUM USING NEURO-FUZZY MODELING Mohammad Farrokhi and Ali Ghanbarie Iran Universit of Science and Technolog, Tehran 6844,

More information

Hopf Bifurcation and Control of Lorenz 84 System

Hopf Bifurcation and Control of Lorenz 84 System ISSN 79-3889 print), 79-3897 online) International Journal of Nonlinear Science Vol.63) No.,pp.5-3 Hopf Bifurcation and Control of Loren 8 Sstem Xuedi Wang, Kaihua Shi, Yang Zhou Nonlinear Scientific Research

More information

Ordinary Differential Equations n

Ordinary Differential Equations n Numerical Analsis MTH63 Ordinar Differential Equations Introduction Talor Series Euler Method Runge-Kutta Method Predictor Corrector Method Introduction Man problems in science and engineering when formulated

More information

On the Spectral Theory of Operator Pencils in a Hilbert Space

On the Spectral Theory of Operator Pencils in a Hilbert Space Journal of Nonlinear Mathematical Phsics ISSN: 1402-9251 Print 1776-0852 Online Journal homepage: http://www.tandfonline.com/loi/tnmp20 On the Spectral Theor of Operator Pencils in a Hilbert Space Roman

More information

Existence of periodic solutions in predator prey and competition dynamic systems

Existence of periodic solutions in predator prey and competition dynamic systems Nonlinear Analsis: Real World Applications 7 (2006) 1193 1204 www.elsevier.com/locate/na Eistence of periodic solutions in predator pre and competition dnamic sstems Martin Bohner a,, Meng Fan b, Jimin

More information

Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine

Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support Vector Machine Commun. Theor. Phys. (Beijing, China) 48 (2007) pp. 117 124 c International Academic Publishers Vol. 48, No. 1, July 15, 2007 Discussion of Some Problems About Nonlinear Time Series Prediction Using ν-support

More information

QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS

QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS Journal of the Operations Research Societ of Japan 008, Vol. 51, No., 191-01 QUADRATIC AND CONVEX MINIMAX CLASSIFICATION PROBLEMS Tomonari Kitahara Shinji Mizuno Kazuhide Nakata Toko Institute of Technolog

More information

Local Phase Portrait of Nonlinear Systems Near Equilibria

Local Phase Portrait of Nonlinear Systems Near Equilibria Local Phase Portrait of Nonlinear Sstems Near Equilibria [1] Consider 1 = 6 1 1 3 1, = 3 1. ( ) (a) Find all equilibrium solutions of the sstem ( ). (b) For each equilibrium point, give the linear approimating

More information

Stability Analysis of Mathematical Model for New Hydraulic Bilateral Rolling Shear

Stability Analysis of Mathematical Model for New Hydraulic Bilateral Rolling Shear SJ nternational, Vol. 56 (06), SJ nternational, No. Vol. 56 (06), No., pp. 88 9 Stabilit Analsis of Mathematical Model for New Hdraulic Bilateral Rolling Shear QingXue HUANG, ) Jia L, ) HongZhou L, ) HeYong

More information

Ordinary Differential Equations

Ordinary Differential Equations 58229_CH0_00_03.indd Page 6/6/6 2:48 PM F-007 /202/JB0027/work/indd & Bartlett Learning LLC, an Ascend Learning Compan.. PART Ordinar Differential Equations. Introduction to Differential Equations 2. First-Order

More information

Vibration Analysis of Isotropic and Orthotropic Plates with Mixed Boundary Conditions

Vibration Analysis of Isotropic and Orthotropic Plates with Mixed Boundary Conditions Tamkang Journal of Science and Engineering, Vol. 6, No. 4, pp. 7-6 (003) 7 Vibration Analsis of Isotropic and Orthotropic Plates with Mied Boundar Conditions Ming-Hung Hsu epartment of Electronic Engineering

More information

Discussion About Nonlinear Time Series Prediction Using Least Squares Support Vector Machine

Discussion About Nonlinear Time Series Prediction Using Least Squares Support Vector Machine Commun. Theor. Phys. (Beijing, China) 43 (2005) pp. 1056 1060 c International Academic Publishers Vol. 43, No. 6, June 15, 2005 Discussion About Nonlinear Time Series Prediction Using Least Squares Support

More information

Proper Orthogonal Decomposition Extensions For Parametric Applications in Transonic Aerodynamics

Proper Orthogonal Decomposition Extensions For Parametric Applications in Transonic Aerodynamics Proper Orthogonal Decomposition Etensions For Parametric Applications in Transonic Aerodnamics T. Bui-Thanh, M. Damodaran Singapore-Massachusetts Institute of Technolog Alliance (SMA) School of Mechanical

More information

Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights

Linear, threshold units. Linear Discriminant Functions and Support Vector Machines. Biometrics CSE 190 Lecture 11. X i : inputs W i : weights Linear Discriminant Functions and Support Vector Machines Linear, threshold units CSE19, Winter 11 Biometrics CSE 19 Lecture 11 1 X i : inputs W i : weights θ : threshold 3 4 5 1 6 7 Courtesy of University

More information

An Optimal Evader Strategy in a Two-Pursuer One-Evader Problem

An Optimal Evader Strategy in a Two-Pursuer One-Evader Problem An Optimal Evader Strateg in a Two-Pursuer One-Evader Problem Wei Sun and Panagiotis Tsiotras 2 Abstract We consider a rela pursuit-evasion problem with two pursuers and one evader. We reduce the problem

More information

Research Article Chaos and Control of Game Model Based on Heterogeneous Expectations in Electric Power Triopoly

Research Article Chaos and Control of Game Model Based on Heterogeneous Expectations in Electric Power Triopoly Discrete Dnamics in Nature and Societ Volume 29, Article ID 469564, 8 pages doi:.55/29/469564 Research Article Chaos and Control of Game Model Based on Heterogeneous Epectations in Electric Power Triopol

More information

Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron

Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron CS446: Machine Learning, Fall 2017 Lecture 14 : Online Learning, Stochastic Gradient Descent, Perceptron Lecturer: Sanmi Koyejo Scribe: Ke Wang, Oct. 24th, 2017 Agenda Recap: SVM and Hinge loss, Representer

More information

Functions of Several Variables

Functions of Several Variables Chapter 1 Functions of Several Variables 1.1 Introduction A real valued function of n variables is a function f : R, where the domain is a subset of R n. So: for each ( 1,,..., n ) in, the value of f is

More information

Kernel Methods. Machine Learning A W VO

Kernel Methods. Machine Learning A W VO Kernel Methods Machine Learning A 708.063 07W VO Outline 1. Dual representation 2. The kernel concept 3. Properties of kernels 4. Examples of kernel machines Kernel PCA Support vector regression (Relevance

More information

14.1 Systems of Linear Equations in Two Variables

14.1 Systems of Linear Equations in Two Variables 86 Chapter 1 Sstems of Equations and Matrices 1.1 Sstems of Linear Equations in Two Variables Use the method of substitution to solve sstems of equations in two variables. Use the method of elimination

More information

Control Schemes to Reduce Risk of Extinction in the Lotka-Volterra Predator-Prey Model

Control Schemes to Reduce Risk of Extinction in the Lotka-Volterra Predator-Prey Model Journal of Applied Mathematics and Phsics, 2014, 2, 644-652 Published Online June 2014 in SciRes. http://www.scirp.org/journal/jamp http://d.doi.org/10.4236/jamp.2014.27071 Control Schemes to Reduce Risk

More information

Glossary. Also available at BigIdeasMath.com: multi-language glossary vocabulary flash cards. An equation that contains an absolute value expression

Glossary. Also available at BigIdeasMath.com: multi-language glossary vocabulary flash cards. An equation that contains an absolute value expression Glossar This student friendl glossar is designed to be a reference for ke vocabular, properties, and mathematical terms. Several of the entries include a short eample to aid our understanding of important

More information

VCAA Bulletin. VCE, VCAL and VET. Supplement 2

VCAA Bulletin. VCE, VCAL and VET. Supplement 2 VCE, VCAL and VET VCAA Bulletin Regulations and information about curriculum and assessment for the VCE, VCAL and VET No. 87 April 20 Victorian Certificate of Education Victorian Certificate of Applied

More information

Outliers Treatment in Support Vector Regression for Financial Time Series Prediction

Outliers Treatment in Support Vector Regression for Financial Time Series Prediction Outliers Treatment in Support Vector Regression for Financial Time Series Prediction Haiqin Yang, Kaizhu Huang, Laiwan Chan, Irwin King, and Michael R. Lyu Department of Computer Science and Engineering

More information

Unit 12 Study Notes 1 Systems of Equations

Unit 12 Study Notes 1 Systems of Equations You should learn to: Unit Stud Notes Sstems of Equations. Solve sstems of equations b substitution.. Solve sstems of equations b graphing (calculator). 3. Solve sstems of equations b elimination. 4. Solve

More information

MULTI-SCROLL CHAOTIC AND HYPERCHAOTIC ATTRACTORS GENERATED FROM CHEN SYSTEM

MULTI-SCROLL CHAOTIC AND HYPERCHAOTIC ATTRACTORS GENERATED FROM CHEN SYSTEM International Journal of Bifurcation and Chaos, Vol. 22, No. 2 (212) 133 ( pages) c World Scientific Publishing Compan DOI: 1.1142/S21812741332 MULTI-SCROLL CHAOTIC AND HYPERCHAOTIC ATTRACTORS GENERATED

More information

MATRIX KERNELS FOR THE GAUSSIAN ORTHOGONAL AND SYMPLECTIC ENSEMBLES

MATRIX KERNELS FOR THE GAUSSIAN ORTHOGONAL AND SYMPLECTIC ENSEMBLES Ann. Inst. Fourier, Grenoble 55, 6 (5), 197 7 MATRIX KERNELS FOR THE GAUSSIAN ORTHOGONAL AND SYMPLECTIC ENSEMBLES b Craig A. TRACY & Harold WIDOM I. Introduction. For a large class of finite N determinantal

More information

Kernel PCA, clustering and canonical correlation analysis

Kernel PCA, clustering and canonical correlation analysis ernel PCA, clustering and canonical correlation analsis Le Song Machine Learning II: Advanced opics CSE 8803ML, Spring 2012 Support Vector Machines (SVM) 1 min w 2 w w + C j ξ j s. t. w j + b j 1 ξ j,

More information

SMO vs PDCO for SVM: Sequential Minimal Optimization vs Primal-Dual interior method for Convex Objectives for Support Vector Machines

SMO vs PDCO for SVM: Sequential Minimal Optimization vs Primal-Dual interior method for Convex Objectives for Support Vector Machines vs for SVM: Sequential Minimal Optimization vs Primal-Dual interior method for Convex Objectives for Support Vector Machines Ding Ma Michael Saunders Working paper, January 5 Introduction In machine learning,

More information

Stochastic Interpretation for the Arimoto Algorithm

Stochastic Interpretation for the Arimoto Algorithm Stochastic Interpretation for the Arimoto Algorithm Serge Tridenski EE - Sstems Department Tel Aviv Universit Tel Aviv, Israel Email: sergetr@post.tau.ac.il Ram Zamir EE - Sstems Department Tel Aviv Universit

More information

Support Vector Machines

Support Vector Machines Wien, June, 2010 Paul Hofmarcher, Stefan Theussl, WU Wien Hofmarcher/Theussl SVM 1/21 Linear Separable Separating Hyperplanes Non-Linear Separable Soft-Margin Hyperplanes Hofmarcher/Theussl SVM 2/21 (SVM)

More information

Implicit Second Derivative Hybrid Linear Multistep Method with Nested Predictors for Ordinary Differential Equations

Implicit Second Derivative Hybrid Linear Multistep Method with Nested Predictors for Ordinary Differential Equations American Scientiic Research Journal or Engineering, Technolog, and Sciences (ASRJETS) ISSN (Print) -44, ISSN (Online) -44 Global Societ o Scientiic Research and Researchers http://asretsournal.org/ Implicit

More information

A PROPOSAL OF ROBUST DESIGN METHOD APPLICABLE TO DIVERSE DESIGN PROBLEMS

A PROPOSAL OF ROBUST DESIGN METHOD APPLICABLE TO DIVERSE DESIGN PROBLEMS A PROPOSAL OF ROBUST DESIGN METHOD APPLICABLE TO DIVERSE DESIGN PROBLEMS Satoshi Nakatsuka Takeo Kato Yoshiki Ujiie Yoshiuki Matsuoka Graduate School of Science and Technolog Keio Universit Yokohama Japan

More information

Feedforward Neural Networks

Feedforward Neural Networks Chapter 4 Feedforward Neural Networks 4. Motivation Let s start with our logistic regression model from before: P(k d) = softma k =k ( λ(k ) + w d λ(k, w) ). (4.) Recall that this model gives us a lot

More information

Applications of Proper Orthogonal Decomposition for Inviscid Transonic Aerodynamics

Applications of Proper Orthogonal Decomposition for Inviscid Transonic Aerodynamics Applications of Proper Orthogonal Decomposition for Inviscid Transonic Aerodnamics Bui-Thanh Tan, Karen Willco and Murali Damodaran Abstract Two etensions to the proper orthogonal decomposition (POD) technique

More information

LATERAL BUCKLING ANALYSIS OF ANGLED FRAMES WITH THIN-WALLED I-BEAMS

LATERAL BUCKLING ANALYSIS OF ANGLED FRAMES WITH THIN-WALLED I-BEAMS Journal of arine Science and J.-D. Technolog, Yau: ateral Vol. Buckling 17, No. Analsis 1, pp. 9-33 of Angled (009) Frames with Thin-Walled I-Beams 9 ATERA BUCKING ANAYSIS OF ANGED FRAES WITH THIN-WAED

More information

Machine Learning Techniques

Machine Learning Techniques Machine Learning Techniques ( 機器學習技巧 ) Lecture 6: Kernel Models for Regression Hsuan-Tien Lin ( 林軒田 ) htlin@csie.ntu.edu.tw Department of Computer Science & Information Engineering National Taiwan University

More information

Towards Higher-Order Schemes for Compressible Flow

Towards Higher-Order Schemes for Compressible Flow WDS'6 Proceedings of Contributed Papers, Part I, 5 4, 6. ISBN 8-867-84- MATFYZPRESS Towards Higher-Order Schemes for Compressible Flow K. Findejs Charles Universit, Facult of Mathematics and Phsics, Prague,

More information

DIFFERENTIAL EQUATIONS First Order Differential Equations. Paul Dawkins

DIFFERENTIAL EQUATIONS First Order Differential Equations. Paul Dawkins DIFFERENTIAL EQUATIONS First Order Paul Dawkins Table of Contents Preface... First Order... 3 Introduction... 3 Linear... 4 Separable... 7 Eact... 8 Bernoulli... 39 Substitutions... 46 Intervals of Validit...

More information

Fractal dimension of the controlled Julia sets of the output duopoly competing evolution model

Fractal dimension of the controlled Julia sets of the output duopoly competing evolution model Available online at www.isr-publications.com/jmcs J. Math. Computer Sci. 1 (1) 1 71 Research Article Fractal dimension of the controlled Julia sets of the output duopol competing evolution model Zhaoqing

More information

Twin Support Vector Machine in Linear Programs

Twin Support Vector Machine in Linear Programs Procedia Computer Science Volume 29, 204, Pages 770 778 ICCS 204. 4th International Conference on Computational Science Twin Support Vector Machine in Linear Programs Dewei Li Yingjie Tian 2 Information

More information