Adaptive Neural Filters with Fixed Weights
|
|
- Iris Gregory
- 6 years ago
- Views:
Transcription
1 Adaptive Neural Filters wit Fixed Weigts James T. Lo and Justin Nave Department of Matematics and Statistics University of Maryland Baltimore County Baltimore, MD 150, U.S.A. Abstract By te fundamental neural filtering teorem, a properly trained recursive neural filter wit fixed weigts tat processes only te measurement process generates recursively te conditional expectation of te signal process wit respect to te joint probability distributions of te signal and measurement processes and any uncertain environmental process involved. Tis means tat said recursive neural filter wit fixed weigts as te ability to adapt to te uncertain environmental parameter. Tis ability is called accommodative ability. Tis paper sows tat if te uncertain environmental process is observable (not necessarily constant) from te measurement process, ten te estimate of te signal process generated by said recursive neural filter wit fixed weigts approaces te estimate of te signal process tat would be generated as if te precise value of te uncertain environmental process were given and processed togeter wit te measurement process by a minimal-variance filter. 1 Introduction If te signal and measurement processes involve an uncertain environmental process, an adaptive filter is needed to adapt to te environmental process in processing te measurements to estimate te signals. An adaptive filter usually requires online adjustment of its parameters, wic is difficult or impossible in many applications especially if te precised values of te signals are unavailable. In 199, a very general fundamental teorem on recursive neural filtering was proven [, 6, 7] in a most general formulation of filtering. Te teorem states tat a recursive neural network exists tat inputs te measurement process and outputs an estimate of te signal process, were te estimate can be made as close as desired to te conditional expectation of te signal process given te past istory of te measurement process tat as been processed. It was observed [3, 10] tat as an immediate corollary of tis fundamental teorem, a properly trained recurrent neural network wit fixed weigts can adapt to an environmental process tat is observable from te measurement process. In fact, suc adaptation can be viewed as a manifestation of an estimation of te observable environmental process performed internally inside te recurrent neural network. Tis adaptive capability of recurrent neural networks wit fixed weigts was used for active engine exaust noise control [4] and engine idle speed control and time series prediction [1]. To distinguis adaptive capability to adapt witout online processor adjustment from te ordinary adaptive capability to adapt wit online processor adjustment, te former is called accommodative capability. A neural network wit accommodative capability is called an accommodative neural network. In [5], te accuracy of accommodative neural networks for adaptive identification of dynamical systems was analyzed. It was found tat under mild conditions, if an uncertain environmental parameter is observable from te measurement process or if it is nonobservable but constant, ten a recursive neural network wit fixed weigts exists tat inputs te current dynamical state of a dynamical system and outputs a predicted value of te dynamical state for te next time point, were te predicted value approaces to te predicted value tat could be generated as if te precise value of te uncertain environmental parameter were given and processed togeter wit te current dynamical state. In tis paper, some similar results for accommodative filtering, namely adaptive filtering by recursive neural networks wit fixed weigts, are obtained. Altoug it was discussed in [3, 10, 8] tat an accommodative neural network does not generalize as well as an adaptive neural network (wit long- and sort-term memories), not requiring online adjustment for adaptation is a igly desirable advantage especially wen online training data is not available. In recent years, anoter approac to nonlinear filtering, called particle filtering, as been developed. However, it performs Monte Carlo online and tus involves excessive amount of online computation. Moreover, for adaptive filtering, te particle filter as to augment te signal process to include te environmental process and
2 estimate it online, wic increases te dimensionality of te signal process and tus requires even more particles and online computation. In contrast, an accommodative neural filter is syntesized prior to its deployment and functions recursively muc like te Kalman filter does for linear signal and measurement processes [9]. Te syntesis of te adaptive neural filter can be viewed as Monte Carlo, but no Monte Carlo is needed to run te filter online. How well an environmental process can be adapted to by an adaptive processor depends on ow muc information te input to te adaptive processor contains about te environmental process. Obviously, te best performance tat te adaptive processor can acieve is te performance acievable given te precise value of te environmental process. If an adaptive processor exists wose performance converges to tis best performance acievable given te value of te environmental process, te environmental process is said to be adaptation-fit forte processing involved. A precise statement of tis definition of adaptation-fitness for filtering is given in Section 3aftertefiltering problem is stated in Section. It is proven in Section 3 tat tere exists an accommodative neural filter wit a filtering accuracy allowed by te degree of adaptation-fitness of te environmental process. Te proof is based on te fundamental neural filtering teorem[,6,7]. Observability of a stocastic environmental process from a measuremet process is defined in Section 4. As expected, te degree of adaptation-fitness of te environmental process depends on te degree of its observability as well as te smootness of te conditional expectation of te signal process wit respect to te environmental process. Te main teorem in Section 4 gives an intuitively appealing quantification of te dependence. A numerical example is given in Section 5. In te example, te signal process is a Henon system wit an observable environmental parameter, and te measurement process is te cubic sensor. Te Monte Carlo test results are consistent wit te teoretical results discussed above. Problem of Adaptive Filtering A signal process to be estimated is described by te vector equation: For t =0, 1,..., x(t, θ t ) = f(x t 1 t p (θ),θ t,w t ) (1) x t 1 t p(θ) : = (x (t 1,θ t 1 ),...,x(t p, θ t p )) () wit te initial condition (or state), (x(0,θ),...,x(1 p, θ)) = (x 0,...,x 1 p ), (3) were te vector-valued function f and te integer p are given; x(t) is a known input vector at time t; w is a random vector sequence wit given joint probability distributions; θ (t) denotes te vector-valued uncertain environmental parameter at time t; and te initial state, (x 0,...,x 1 p ), is a random vector wit given probability distribution (reflecting te relative frequencies of te actual initial states of te system, (1), in operations). A measurement y (t) of te vector output x (t) is made available at time t, tat satisfies y(t, θ t )= x t t q+1(θ),θ t,v(t), (4) were v is a random vector sequence wit given joint probability distributions. Te problem of adaptive filtering is to design and implement an adaptive recursive processor tat adapts to te uncertain operating environment represented by te equations, (1), (3) and (4), and produce an estimate of x (t, θ t ) tat makes a best use of te information about θ contained in te measurements y. During te operation of suc an adaptive processor, no part of te signal process is directly measurable (i.e., signal plus noise unavailable) and at time t, only te measurement, y (t, θ t ), is available for processing by te adaptive processor. Te adaptive processor inputs y(t, θ t ) and outputs an estimate ˆx (t, θ t ) of x (t, θ t ) at eac time t =1,,,T,wereT is a positive integer or infinity. Te most widely used estimation error criterion is te conditional mean square error criterion, kx (t, θ t ) ˆx (t)k y t (θ) i, were y t (θ) := {y (s, θ s ),s=1,...,t}, ˆx (t) denotes an estimate of x (t, θ t ) given y t (θ), [ y t (θ)] is te conditional expectation given y t. An optimal estimate wit tis criterion is [x (t, θ t ) y t (θ)], wic is called te minimum-variance estimate. For notational simplicity, x (t, θ t ), y(t, θ t ), x t s(θ) and y t (θ) will be denoted by x(t), y(t), x t s and y t respectively in te sequel. 3 Accommodating an Adaptation- Fit nvironmental Process Successful adaptation to an uncertain environmental processrequiressufficient information about it. If sufficient information about te environmental process is contained in te measurement process suc tat an adaptive filter exists tat generates an estimate of te signal process tat approaces tat acievable wit a filter given te precise value of te unvironmental process, te environmental process is called adaptation-fit forfiltering. A rigorous definition of an adaptation-fit environmental process for filtering is stated as Definition 1 below. In te main teorem of tis section, it is proven tat if te environmental process is adaptation-fit forfiltering to a certain accuracy, adaptive filtering to te same accuracy can be realized
3 wit a recurrent neural network wit fixed weigts. Suc a recurrent neural network does not require online weigt adjustment and is called an accommodative neural filter. Definition 1. Letx, y and θ be signal, measurement and environmental processes respectively, and let ˆx (t) and ˆx (t θ) denote te conditional expectations [x (t) y t ] and x (t) y t,θ t,wereθ t := {θ τ,τ =1,...,t} and y t := {y τ,τ =1,...,t}. Assume tat θ is a stocastic process, wic is not necessarily time-varying. Te environmental process θ is said to be adaptation-fit towitinanerrorof ε for filtering, if tere is a positive integer N (ε) suc tat for all t>n(ε), kˆx (t θ) ˆx (t)k i <ε We need te fundamental neural filtering teorem [, 6, 7] to prove te main teorem of tis section. Teorem 1. Consider an n-dimensional stocastic process x(t) and an m-dimensional stocastic process y(t), t =1,,T defined on a probability space (Ω, A,P). Assume tat te range {y(t, ω) t =1,,T,ω Ω} R m is compact and ψ is an arbitrary k-dimensional Borel function of x(t) wit finite second moments [kψ(x(t))k ], t =1,,T.Let α(t) denote te k-dimensional output at time t of a recurrent neural network wic as taken te inputs, y(1),,y(t), integivenorder. 1. Given >0, tere exists a recurrent neural network wit one idden layer of fully interconnected neurons suc tat 1 T TX [kα(t) [ψ(x(t)) y t ]k ] <. t=1. If te recurrent neural network as one idden layer of N neurons, wic are fully interconnected, and te output α(t) is written as α(t; N) ere to indicate its dependency on N, ten r(n) :=min w 1 T TX [kα(t, N) [ψ(x(t)) y t ]k ] (5) t=1 is monotone decreasing and converges to 0 as N approaces infinity. Te following corollary follows immediately te above fundamental neural filtering teorem. Corollary 1. Consider an n-dimensional stocastic process x(t) and an m-dimensional stocastic process y(t), t =1,,T defined on a probability space (Ω, A,P). Assume tat te range {y(t, ω) t =1,,T,ω Ω} R m is compact and ψ is an arbitrary k-dimensional Borel function of x(t) wit finite second moments [kψ(x(t))k ], t =1,,T.Letα(t) denote te k-dimensional output at time t of a recurrent neural network wic as taken te inputs, y(1),,y(t), integivenorder. Given >0, tere exists a recurrent neural network wit one idden layer of fully interconnected neurons suc tat for all t =1,...,T, [kα(t) [ψ(x(t)) y t ]k ] <. We are now ready to state te first main teorem of tis paper: Teorem. Let x, y and θ be signal, measurement and environmental processes respectively, were θ is adaptation-fit towitinanerrorofε for filtering. For filtering over a time interval, 1 t T,afixed-weigt MLPWIN, tat as only one idden layer of neurons; inputs y(t) and outputs an estimate α (t) of te signal x(t, θ t ) at time t, existsasanadaptivefilter suc tat kˆx(t θ) α (t)k i <ε for all t less tan T and greater tan some positive integer N (ε), wereˆx(t θ) denotes te conditional expectation x (t) y t,θ t. Proof. By Definition 1, for te given ε, tere is a positive integer N (ε) suc tat for all t>n(ε), kˆx (t θ) ˆx (t)k i <ε (6) Let ε 1 be a positive number less tan ε kˆx (t θ) ˆx (t)k i > 0. Notice tat bot ˆx (t) and α (t) aremeasurablewitrespecttoy t. It follows tat (ˆx (t θ) ˆx (t)) T (ˆx (t) α (t)) y ti = (ˆx (t θ) ˆx (t)) T y ti (ˆx (t) α (t)) = ˆx T (t θ) y t ˆx T (t) ³ˆx ³ (t θ) ˆx t ˆθ = ˆx T (t) ˆx T (t) ³ˆx ³ (t θ) ˆx t ˆθ =0 (7) By te smooting property of te conditional expectation, we ave, for t>n(ε), kˆx (t θ) α (t)k i = k(ˆx (t θ) ˆx (t)) + (ˆx (t) α (t))k i = kˆx (t θ) ˆx (t)k i + kˆx (t) α (t)k i + (ˆx (t θ) ˆx (t)) T (ˆx (t) α (t)) y tii = kˆx (t θ) ˆx (t)k i + kˆx (t) α (t)k i <ε+ kˆx (t) α (t)k i
4 were te last equality and last inequality follow from (7) and (6) respectively. By Corollary 1, for any ε 1 > 0, tere is a recurrent neural network wose output α (t) satisfies kˆx (t) α (t)k i <ε 1 for all t =1,...,T. It follows tat kˆx (t θ) α (t)k i <ε, wic completes te proof. 4 Observability Implies Adaptation-Fitness Observability is a well-developed concept in te statespace system teory. However, it is defined only for deterministic systems. For our purpose, we adopt te following definition of an observable environmental process for a stocastic system: Definition. Let an environmental process be a stocastic process and let te conditional expectation [θ t y t ] of θ t given y t := {y τ,τ =1,...,t} be denoted by ˆθ t. Ten te environmental process θ is said to be observable from te measurement process y to witin an error of ε>0, if tere is a positive integer N (ε) suc tat for all t>n(ε), θt ˆθ t <ε Teorem 3. If te environmental process θ is observable from te measurement process y to witin an error of ε, and if te uclidean norm kd [x (t) y t,θ t ] /dθ t k <M uniformly for all y t and θ t,wered [x (t) y t,θ t ] /dθ t is te derivative of te conditional expectation [x (t) y t,θ t ] wit respect to θ t, ten te environmental process is adaptation-fit to witin an error of M ε for filtering. Proof. It follows from (1) tat ˆx (t θ) := x (t) y t,θ t = [x (t) y t,θ t ]. Hence ˆx (t θ) is a Borel measurable function of y t and θ t. Denoting tis function by g (y t,θ t ) and substituting ˆθ t for θ t in it yield g ³y t, t ˆθ,wicismeasurable wit respect to y t.notetat ³ˆx (ˆx (t θ) ˆx (t)) T (t) g ³y t, t ˆθ ³ˆx = (ˆx (t θ) ˆx (t)) T (t) g ³y t, t ˆθ y tii = (ˆx (t θ) ˆx (t)) T y ti³ˆx (t) g ³y t, t i ˆθ = ˆx T (t θ) y t ˆx T (t) ³ˆx (t) g ³y t, t i ˆθ = ˆx T (t) ˆx T (t) ³ˆx (t) g ³y t, t i ˆθ =0 Tis sows tat te cross term resulting from expanding te second expression below is zero and establises te second equality below. ˆx (t θ) g ³y t, t ˆθ ³ = (ˆx (t θ) ˆx (t)) + ˆx (t) g ³y t, t ˆθ ³y t, t ˆθ = kˆx (t θ) ˆx (t)k + ˆx (t) g Hence, kˆx (t θ) ˆx (t)k ˆx t θ t g ³y t, t ˆθ = g y t ³,θ t g y t, t ˆθ By te mean value teorem, g y t,θ t g ³y t, t ˆθ ³ dg y t, θ θt /dθ t ˆθ t M θ t ˆθ t. Terefore, kˆx (t θ) ˆx (t)k i θt M ˆθ t M ε,forsomen (ε), completing te proof. Remark. ε in Teorem 3 reflects te predictability of θ, and M reflects te smootness of te conditional expectation [x (t) y t,θ t ] wit respect to θ t. Teorem 3 only provides a quantification of our intuition tat adaptationfitness increases as te predictability of θ and te smootness of te conditional expectation [x (t) y t,θ t ] wit respect to θ t increase. 5 An Numerical xample In te numerical example, te signal process is te wellknown Henon system: x (t +1) = bx (t 1) + 1 θx (t)+0.1w(t) y (t) = x 3 (t)+0.1v(t) were w and v are wite sequences wit w (t) and v (t) aving normal distributions wit mean 0 and variance 1 and wit any samples greater tan or equal to 3 in absolute value discarded; b =0.3;andθ [0, 0.5] is te uncertain environmental parameter. Te training dataset consisted of 1000 realizations of x and y tat were 30 time points long including 30 for priming. Te teta values were cosen randomly from te set {0.1, 0., 0.3, 0.4, 0.5}. Starting values for x (0) and x ( 1) were selected from [0.5, 0.9]. Anoter dataset was generated similarly and used for cross-validation during training. Te weigts tat produced te minimum RMS for te cross-validation data set were used. Training was <
5 preformed on various arcitectures, but a recurrent neural network wit a single idden layer of 15 neurons was selected as te accommodative neural filter. To see weter and ow te output of te accommodative neural filter approaces te conditional expectation of x (t) given bot y t and θ, a separate MLPWIN was similarly trained for eac given teta value from te set { , 0.5, 0.35, 0.45}. A testing data set was generated wit 500 streams also 30 time points long. It contained 100 streams for eac of te teta values in te set { }. Te RMS of te accommodative neural filter on tis test data set is Te RMS of te MLPWINs trained for te exact teta values is Te standard deviation of te signal is Tese numbers sow tat te accommodative neural filter performed satisfactorily. Attaced are two plots. Figure 1 sows te RMS vs. time averaged over all te streams in te testing data set. Figure sows one realization of te signal process at θ = 0.5 and its estimate generated by te accommodative neural filter over 50 time points. 6 Conclusion Te adaptive capability of recurrent neural networks wit fixed weigts was observed as a consequence of te fundamental neural filtering teorem in 1994 [3, 10]. Tis capability is called accommodative capability. Tis paper defines adaptation-fitness and observability of an environmental process, and proves tat if te environmental process is adaptation-fit, a recurrent neural network wit fixed weigts exists wose filtering performance approaces te performance tat would be acievable as if te environmental process were given, and sows tat observability of an environmental process implies it being adaptation-fit. Altoug recurrent neural networks do not generalize as well as adaptive neural networks (wit long- and sortterm memories), te former ave te unique advantage of not requiring online adjustment. Tis advantage is important in many applications, especially wen no data is available online for adjustment of te parameters or weigts of a processor (e.g., filter or controller) to adapt to an uncertain environmental process. [] J. T. Lo. Neural network approac to optimal filtering. Invited paper presented at te 199 World Congress of Nonlinear Analysts, Tampa, Florida, Aug [3] J. T. Lo. Neural network approac to optimal filtering. Tecnical Report RL-TR , Rome Laboratory, Air Force Material Command, [4] J. T. Lo. Neural network approac to active engine noise cancellation. Tecnical report, Maryland Tecnology Corporation, [5] J. T. Lo. Matematical underpinning of adaptive capability of recurrent neural networks wit fixed weigts. In Proceedings of te 003 International Joint Conference on Neural Networks, pages , Portland, Oregon, July 003. [6] J. T. Lo. Syntetic approac to optimal filtering. In Proceedings of te 199 International Simulation Tecnology Conference and 199 Worksop on Neural Networks, pages , Clear Lake, Texas, November 199. [7] J. T. Lo. Syntetic approac to optimal filtering. I Transactions on Neural Networks, 5: , September [8] J. T. Lo and D. Bassu. Adaptive versus accommodative neural networks for adaptive identification of dynamic systems. In Proceedings of te 001 International Joint Conference on Neural Networks, Wasington, D.C., July, 001. [9] J. T. Lo and L. Yu. Recursive neural filters and dynamical range transformers. Proceedings of Te I, 9, No. 3: , 004. [10] J. T. Lo and L. Yu. Adaptive neural filtering by using te innovations process. Proceedings of te 1995 World Congress on Neural Networks, II:9 35, July References [1] L. A. Feldkamp and G. V. Puskorious. Training of robust neural controllers. In Proceedings of te 33rd Conference on Decision and Control, pages , Lake Buena Vista, Florida, December 1994.
6 Figure 1: Te RMSs of te estimates generated by te accommodative neural filter compared wit tose generated using given teta values. Figure : A single realization of te signal process compared wit its estimate generated by te accommodative neural filter.
A = h w (1) Error Analysis Physics 141
Introduction In all brances of pysical science and engineering one deals constantly wit numbers wic results more or less directly from experimental observations. Experimental observations always ave inaccuracies.
More informationPolynomial Interpolation
Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximatinga function fx, wose values at a set of distinct points x, x, x,, x n are known, by a polynomial P x suc
More informationVolume 29, Issue 3. Existence of competitive equilibrium in economies with multi-member households
Volume 29, Issue 3 Existence of competitive equilibrium in economies wit multi-member ouseolds Noriisa Sato Graduate Scool of Economics, Waseda University Abstract Tis paper focuses on te existence of
More information4. The slope of the line 2x 7y = 8 is (a) 2/7 (b) 7/2 (c) 2 (d) 2/7 (e) None of these.
Mat 11. Test Form N Fall 016 Name. Instructions. Te first eleven problems are wort points eac. Te last six problems are wort 5 points eac. For te last six problems, you must use relevant metods of algebra
More informationEFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS
Statistica Sinica 24 2014, 395-414 doi:ttp://dx.doi.org/10.5705/ss.2012.064 EFFICIENCY OF MODEL-ASSISTED REGRESSION ESTIMATORS IN SAMPLE SURVEYS Jun Sao 1,2 and Seng Wang 3 1 East Cina Normal University,
More information232 Calculus and Structures
3 Calculus and Structures CHAPTER 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS FOR EVALUATING BEAMS Calculus and Structures 33 Copyrigt Capter 17 JUSTIFICATION OF THE AREA AND SLOPE METHODS 17.1 THE
More informationNumerical Differentiation
Numerical Differentiation Finite Difference Formulas for te first derivative (Using Taylor Expansion tecnique) (section 8.3.) Suppose tat f() = g() is a function of te variable, and tat as 0 te function
More information1 The concept of limits (p.217 p.229, p.242 p.249, p.255 p.256) 1.1 Limits Consider the function determined by the formula 3. x since at this point
MA00 Capter 6 Calculus and Basic Linear Algebra I Limits, Continuity and Differentiability Te concept of its (p.7 p.9, p.4 p.49, p.55 p.56). Limits Consider te function determined by te formula f Note
More informationContinuity and Differentiability of the Trigonometric Functions
[Te basis for te following work will be te definition of te trigonometric functions as ratios of te sides of a triangle inscribed in a circle; in particular, te sine of an angle will be defined to be te
More informationUniversity Mathematics 2
University Matematics 2 1 Differentiability In tis section, we discuss te differentiability of functions. Definition 1.1 Differentiable function). Let f) be a function. We say tat f is differentiable at
More informationFunction Composition and Chain Rules
Function Composition and s James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 8, 2017 Outline 1 Function Composition and Continuity 2 Function
More informationCombining functions: algebraic methods
Combining functions: algebraic metods Functions can be added, subtracted, multiplied, divided, and raised to a power, just like numbers or algebra expressions. If f(x) = x 2 and g(x) = x + 2, clearly f(x)
More informationA SHORT INTRODUCTION TO BANACH LATTICES AND
CHAPTER A SHORT INTRODUCTION TO BANACH LATTICES AND POSITIVE OPERATORS In tis capter we give a brief introduction to Banac lattices and positive operators. Most results of tis capter can be found, e.g.,
More informationNatural Language Understanding. Recap: probability, language models, and feedforward networks. Lecture 12: Recurrent Neural Networks and LSTMs
Natural Language Understanding Lecture 12: Recurrent Neural Networks and LSTMs Recap: probability, language models, and feedforward networks Simple Recurrent Networks Adam Lopez Credits: Mirella Lapata
More informationPolynomial Interpolation
Capter 4 Polynomial Interpolation In tis capter, we consider te important problem of approximating a function f(x, wose values at a set of distinct points x, x, x 2,,x n are known, by a polynomial P (x
More informationNotes on Neural Networks
Artificial neurons otes on eural etwors Paulo Eduardo Rauber 205 Consider te data set D {(x i y i ) i { n} x i R m y i R d } Te tas of supervised learning consists on finding a function f : R m R d tat
More informationMath 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006
Mat 102 TEST CHAPTERS 3 & 4 Solutions & Comments Fall 2006 f(x+) f(x) 10 1. For f(x) = x 2 + 2x 5, find ))))))))) and simplify completely. NOTE: **f(x+) is NOT f(x)+! f(x+) f(x) (x+) 2 + 2(x+) 5 ( x 2
More informationCopyright c 2008 Kevin Long
Lecture 4 Numerical solution of initial value problems Te metods you ve learned so far ave obtained closed-form solutions to initial value problems. A closedform solution is an explicit algebriac formula
More information3.4 Worksheet: Proof of the Chain Rule NAME
Mat 1170 3.4 Workseet: Proof of te Cain Rule NAME Te Cain Rule So far we are able to differentiate all types of functions. For example: polynomials, rational, root, and trigonometric functions. We are
More informationThe derivative function
Roberto s Notes on Differential Calculus Capter : Definition of derivative Section Te derivative function Wat you need to know already: f is at a point on its grap and ow to compute it. Wat te derivative
More informationLecture XVII. Abstract We introduce the concept of directional derivative of a scalar function and discuss its relation with the gradient operator.
Lecture XVII Abstract We introduce te concept of directional derivative of a scalar function and discuss its relation wit te gradient operator. Directional derivative and gradient Te directional derivative
More informationLecture 15. Interpolation II. 2 Piecewise polynomial interpolation Hermite splines
Lecture 5 Interpolation II Introduction In te previous lecture we focused primarily on polynomial interpolation of a set of n points. A difficulty we observed is tat wen n is large, our polynomial as to
More informationJournal of Computational and Applied Mathematics
Journal of Computational and Applied Matematics 94 (6) 75 96 Contents lists available at ScienceDirect Journal of Computational and Applied Matematics journal omepage: www.elsevier.com/locate/cam Smootness-Increasing
More informationHomework 1 Due: Wednesday, September 28, 2016
0-704 Information Processing and Learning Fall 06 Homework Due: Wednesday, September 8, 06 Notes: For positive integers k, [k] := {,..., k} denotes te set of te first k positive integers. Wen p and Y q
More informationBob Brown Math 251 Calculus 1 Chapter 3, Section 1 Completed 1 CCBC Dundalk
Bob Brown Mat 251 Calculus 1 Capter 3, Section 1 Completed 1 Te Tangent Line Problem Te idea of a tangent line first arises in geometry in te context of a circle. But before we jump into a discussion of
More informationTHE hidden Markov model (HMM)-based parametric
JOURNAL OF L A TEX CLASS FILES, VOL. 6, NO. 1, JANUARY 2007 1 Modeling Spectral Envelopes Using Restricted Boltzmann Macines and Deep Belief Networks for Statistical Parametric Speec Syntesis Zen-Hua Ling,
More informationHow to Find the Derivative of a Function: Calculus 1
Introduction How to Find te Derivative of a Function: Calculus 1 Calculus is not an easy matematics course Te fact tat you ave enrolled in suc a difficult subject indicates tat you are interested in te
More informationAMS 147 Computational Methods and Applications Lecture 09 Copyright by Hongyun Wang, UCSC. Exact value. Effect of round-off error.
Lecture 09 Copyrigt by Hongyun Wang, UCSC Recap: Te total error in numerical differentiation fl( f ( x + fl( f ( x E T ( = f ( x Numerical result from a computer Exact value = e + f x+ Discretization error
More informationHOMEWORK HELP 2 FOR MATH 151
HOMEWORK HELP 2 FOR MATH 151 Here we go; te second round of omework elp. If tere are oters you would like to see, let me know! 2.4, 43 and 44 At wat points are te functions f(x) and g(x) = xf(x)continuous,
More informationch (for some fixed positive number c) reaching c
GSTF Journal of Matematics Statistics and Operations Researc (JMSOR) Vol. No. September 05 DOI 0.60/s4086-05-000-z Nonlinear Piecewise-defined Difference Equations wit Reciprocal and Cubic Terms Ramadan
More informationApplications of the van Trees inequality to non-parametric estimation.
Brno-06, Lecture 2, 16.05.06 D/Stat/Brno-06/2.tex www.mast.queensu.ca/ blevit/ Applications of te van Trees inequality to non-parametric estimation. Regular non-parametric problems. As an example of suc
More informationConsider a function f we ll specify which assumptions we need to make about it in a minute. Let us reformulate the integral. 1 f(x) dx.
Capter 2 Integrals as sums and derivatives as differences We now switc to te simplest metods for integrating or differentiating a function from its function samples. A careful study of Taylor expansions
More informationMathematics 5 Worksheet 11 Geometry, Tangency, and the Derivative
Matematics 5 Workseet 11 Geometry, Tangency, and te Derivative Problem 1. Find te equation of a line wit slope m tat intersects te point (3, 9). Solution. Te equation for a line passing troug a point (x
More informationFunction Composition and Chain Rules
Function Composition an Cain Rules James K. Peterson Department of Biological Sciences an Department of Matematical Sciences Clemson University November 2, 2018 Outline Function Composition an Continuity
More information7 Semiparametric Methods and Partially Linear Regression
7 Semiparametric Metods and Partially Linear Regression 7. Overview A model is called semiparametric if it is described by and were is nite-dimensional (e.g. parametric) and is in nite-dimensional (nonparametric).
More informationSymmetry Labeling of Molecular Energies
Capter 7. Symmetry Labeling of Molecular Energies Notes: Most of te material presented in tis capter is taken from Bunker and Jensen 1998, Cap. 6, and Bunker and Jensen 2005, Cap. 7. 7.1 Hamiltonian Symmetry
More information5 Ordinary Differential Equations: Finite Difference Methods for Boundary Problems
5 Ordinary Differential Equations: Finite Difference Metods for Boundary Problems Read sections 10.1, 10.2, 10.4 Review questions 10.1 10.4, 10.8 10.9, 10.13 5.1 Introduction In te previous capters we
More informationDifferential Calculus (The basics) Prepared by Mr. C. Hull
Differential Calculus Te basics) A : Limits In tis work on limits, we will deal only wit functions i.e. tose relationsips in wic an input variable ) defines a unique output variable y). Wen we work wit
More informationDifferentiation in higher dimensions
Capter 2 Differentiation in iger dimensions 2.1 Te Total Derivative Recall tat if f : R R is a 1-variable function, and a R, we say tat f is differentiable at x = a if and only if te ratio f(a+) f(a) tends
More informationNUMERICAL DIFFERENTIATION. James T. Smith San Francisco State University. In calculus classes, you compute derivatives algebraically: for example,
NUMERICAL DIFFERENTIATION James T Smit San Francisco State University In calculus classes, you compute derivatives algebraically: for example, f( x) = x + x f ( x) = x x Tis tecnique requires your knowing
More informationTHE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS. L. Trautmann, R. Rabenstein
Worksop on Transforms and Filter Banks (WTFB),Brandenburg, Germany, Marc 999 THE STURM-LIOUVILLE-TRANSFORMATION FOR THE SOLUTION OF VECTOR PARTIAL DIFFERENTIAL EQUATIONS L. Trautmann, R. Rabenstein Lerstul
More informationMAT 145. Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points
MAT 15 Test #2 Name Solution Guide Type of Calculator Used TI-89 Titanium 100 points Score 100 possible points Use te grap of a function sown ere as you respond to questions 1 to 8. 1. lim f (x) 0 2. lim
More informationREVIEW LAB ANSWER KEY
REVIEW LAB ANSWER KEY. Witout using SN, find te derivative of eac of te following (you do not need to simplify your answers): a. f x 3x 3 5x x 6 f x 3 3x 5 x 0 b. g x 4 x x x notice te trick ere! x x g
More informationHOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS
HOW TO DEAL WITH FFT SAMPLING INFLUENCES ON ADEV CALCULATIONS Po-Ceng Cang National Standard Time & Frequency Lab., TL, Taiwan 1, Lane 551, Min-Tsu Road, Sec. 5, Yang-Mei, Taoyuan, Taiwan 36 Tel: 886 3
More informationBootstrap confidence intervals in nonparametric regression without an additive model
Bootstrap confidence intervals in nonparametric regression witout an additive model Dimitris N. Politis Abstract Te problem of confidence interval construction in nonparametric regression via te bootstrap
More informationPoisson Equation in Sobolev Spaces
Poisson Equation in Sobolev Spaces OcMountain Dayligt Time. 6, 011 Today we discuss te Poisson equation in Sobolev spaces. It s existence, uniqueness, and regularity. Weak Solution. u = f in, u = g on
More informationNONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL. Georgeta Budura
NONLINEAR SYSTEMS IDENTIFICATION USING THE VOLTERRA MODEL Georgeta Budura Politenica University of Timisoara, Faculty of Electronics and Telecommunications, Comm. Dep., georgeta.budura@etc.utt.ro Abstract:
More informationContinuity. Example 1
Continuity MATH 1003 Calculus and Linear Algebra (Lecture 13.5) Maoseng Xiong Department of Matematics, HKUST A function f : (a, b) R is continuous at a point c (a, b) if 1. x c f (x) exists, 2. f (c)
More informationMath Spring 2013 Solutions to Assignment # 3 Completion Date: Wednesday May 15, (1/z) 2 (1/z 1) 2 = lim
Mat 311 - Spring 013 Solutions to Assignment # 3 Completion Date: Wednesday May 15, 013 Question 1. [p 56, #10 (a)] 4z Use te teorem of Sec. 17 to sow tat z (z 1) = 4. We ave z 4z (z 1) = z 0 4 (1/z) (1/z
More informationLyapunov characterization of input-to-state stability for semilinear control systems over Banach spaces
Lyapunov caracterization of input-to-state stability for semilinear control systems over Banac spaces Andrii Mironcenko a, Fabian Wirt a a Faculty of Computer Science and Matematics, University of Passau,
More information158 Calculus and Structures
58 Calculus and Structures CHAPTER PROPERTIES OF DERIVATIVES AND DIFFERENTIATION BY THE EASY WAY. Calculus and Structures 59 Copyrigt Capter PROPERTIES OF DERIVATIVES. INTRODUCTION In te last capter you
More informationTechnology-Independent Design of Neurocomputers: The Universal Field Computer 1
Tecnology-Independent Design of Neurocomputers: Te Universal Field Computer 1 Abstract Bruce J. MacLennan Computer Science Department Naval Postgraduate Scool Monterey, CA 9393 We argue tat AI is moving
More informationTe comparison of dierent models M i is based on teir relative probabilities, wic can be expressed, again using Bayes' teorem, in terms of prior probab
To appear in: Advances in Neural Information Processing Systems 9, eds. M. C. Mozer, M. I. Jordan and T. Petsce. MIT Press, 997 Bayesian Model Comparison by Monte Carlo Caining David Barber D.Barber@aston.ac.uk
More informationGELFAND S PROOF OF WIENER S THEOREM
GELFAND S PROOF OF WIENER S THEOREM S. H. KULKARNI 1. Introduction Te following teorem was proved by te famous matematician Norbert Wiener. Wiener s proof can be found in is book [5]. Teorem 1.1. (Wiener
More information(a) At what number x = a does f have a removable discontinuity? What value f(a) should be assigned to f at x = a in order to make f continuous at a?
Solutions to Test 1 Fall 016 1pt 1. Te grap of a function f(x) is sown at rigt below. Part I. State te value of eac limit. If a limit is infinite, state weter it is or. If a limit does not exist (but is
More informationFinancial Econometrics Prof. Massimo Guidolin
CLEFIN A.A. 2010/2011 Financial Econometrics Prof. Massimo Guidolin A Quick Review of Basic Estimation Metods 1. Were te OLS World Ends... Consider two time series 1: = { 1 2 } and 1: = { 1 2 }. At tis
More informationOSCILLATION OF SOLUTIONS TO NON-LINEAR DIFFERENCE EQUATIONS WITH SEVERAL ADVANCED ARGUMENTS. Sandra Pinelas and Julio G. Dix
Opuscula Mat. 37, no. 6 (2017), 887 898 ttp://dx.doi.org/10.7494/opmat.2017.37.6.887 Opuscula Matematica OSCILLATION OF SOLUTIONS TO NON-LINEAR DIFFERENCE EQUATIONS WITH SEVERAL ADVANCED ARGUMENTS Sandra
More informationOverdispersed Variational Autoencoders
Overdispersed Variational Autoencoders Harsil Sa, David Barber and Aleksandar Botev Department of Computer Science, University College London Alan Turing Institute arsil.sa.15@ucl.ac.uk, david.barber@ucl.ac.uk,
More informationTaylor Series and the Mean Value Theorem of Derivatives
1 - Taylor Series and te Mean Value Teorem o Derivatives Te numerical solution o engineering and scientiic problems described by matematical models oten requires solving dierential equations. Dierential
More informationClick here to see an animation of the derivative
Differentiation Massoud Malek Derivative Te concept of derivative is at te core of Calculus; It is a very powerful tool for understanding te beavior of matematical functions. It allows us to optimize functions,
More informationAPPLICATION OF A DIRAC DELTA DIS-INTEGRATION TECHNIQUE TO THE STATISTICS OF ORBITING OBJECTS
APPLICATION OF A DIRAC DELTA DIS-INTEGRATION TECHNIQUE TO THE STATISTICS OF ORBITING OBJECTS Dario Izzo Advanced Concepts Team, ESTEC, AG Noordwijk, Te Neterlands ABSTRACT In many problems related to te
More informationContinuous Stochastic Processes
Continuous Stocastic Processes Te term stocastic is often applied to penomena tat vary in time, wile te word random is reserved for penomena tat vary in space. Apart from tis distinction, te modelling
More information5.1 We will begin this section with the definition of a rational expression. We
Basic Properties and Reducing to Lowest Terms 5.1 We will begin tis section wit te definition of a rational epression. We will ten state te two basic properties associated wit rational epressions and go
More informationDerivatives. if such a limit exists. In this case when such a limit exists, we say that the function f is differentiable.
Derivatives 3. Derivatives Definition 3. Let f be a function an a < b be numbers. Te average rate of cange of f from a to b is f(b) f(a). b a Remark 3. Te average rate of cange of a function f from a to
More informationINTRODUCTION AND MATHEMATICAL CONCEPTS
Capter 1 INTRODUCTION ND MTHEMTICL CONCEPTS PREVIEW Tis capter introduces you to te basic matematical tools for doing pysics. You will study units and converting between units, te trigonometric relationsips
More informationModel Specification Testing in Nonparametric and Semiparametric Time Series Econometrics 1
Model Specification Testing in Nonparametric and Semiparametric Time Series Econometrics 1 By Jiti Gao 2 and Maxwell King 3 Abstract We propose a simultaneous model specification procedure for te conditional
More information2.8 The Derivative as a Function
.8 Te Derivative as a Function Typically, we can find te derivative of a function f at many points of its domain: Definition. Suppose tat f is a function wic is differentiable at every point of an open
More informationArtificial Neural Network Model Based Estimation of Finite Population Total
International Journal of Science and Researc (IJSR), India Online ISSN: 2319-7064 Artificial Neural Network Model Based Estimation of Finite Population Total Robert Kasisi 1, Romanus O. Odiambo 2, Antony
More informationLIMITS AND DERIVATIVES CONDITIONS FOR THE EXISTENCE OF A LIMIT
LIMITS AND DERIVATIVES Te limit of a function is defined as te value of y tat te curve approaces, as x approaces a particular value. Te limit of f (x) as x approaces a is written as f (x) approaces, as
More informationSin, Cos and All That
Sin, Cos and All Tat James K. Peterson Department of Biological Sciences and Department of Matematical Sciences Clemson University Marc 9, 2017 Outline Sin, Cos and all tat! A New Power Rule Derivatives
More informationFINITE DIFFERENCE APPROXIMATIONS FOR NONLINEAR FIRST ORDER PARTIAL DIFFERENTIAL EQUATIONS
UNIVERSITATIS IAGELLONICAE ACTA MATHEMATICA, FASCICULUS XL 2002 FINITE DIFFERENCE APPROXIMATIONS FOR NONLINEAR FIRST ORDER PARTIAL DIFFERENTIAL EQUATIONS by Anna Baranowska Zdzis law Kamont Abstract. Classical
More informationThe Priestley-Chao Estimator
Te Priestley-Cao Estimator In tis section we will consider te Pristley-Cao estimator of te unknown regression function. It is assumed tat we ave a sample of observations (Y i, x i ), i = 1,..., n wic are
More informationDiscontinuous Galerkin Methods for Relativistic Vlasov-Maxwell System
Discontinuous Galerkin Metods for Relativistic Vlasov-Maxwell System He Yang and Fengyan Li December 1, 16 Abstract e relativistic Vlasov-Maxwell (RVM) system is a kinetic model tat describes te dynamics
More informationNumerical analysis of a free piston problem
MATHEMATICAL COMMUNICATIONS 573 Mat. Commun., Vol. 15, No. 2, pp. 573-585 (2010) Numerical analysis of a free piston problem Boris Mua 1 and Zvonimir Tutek 1, 1 Department of Matematics, University of
More informationRecursive Neural Filters and Dynamical Range Transformers
Recursive Neural Filters and Dynamical Range Transformers JAMES T. LO AND LEI YU Invited Paper A recursive neural filter employs a recursive neural network to process a measurement process to estimate
More informationChapter 5 FINITE DIFFERENCE METHOD (FDM)
MEE7 Computer Modeling Tecniques in Engineering Capter 5 FINITE DIFFERENCE METHOD (FDM) 5. Introduction to FDM Te finite difference tecniques are based upon approximations wic permit replacing differential
More informationGlobal Existence of Classical Solutions for a Class Nonlinear Parabolic Equations
Global Journal of Science Frontier Researc Matematics and Decision Sciences Volume 12 Issue 8 Version 1.0 Type : Double Blind Peer Reviewed International Researc Journal Publiser: Global Journals Inc.
More informationConvexity and Smoothness
Capter 4 Convexity and Smootness 4.1 Strict Convexity, Smootness, and Gateaux Differentiablity Definition 4.1.1. Let X be a Banac space wit a norm denoted by. A map f : X \{0} X \{0}, f f x is called a
More information. If lim. x 2 x 1. f(x+h) f(x)
Review of Differential Calculus Wen te value of one variable y is uniquely determined by te value of anoter variable x, ten te relationsip between x and y is described by a function f tat assigns a value
More informationComplexity of Decoding Positive-Rate Reed-Solomon Codes
Complexity of Decoding Positive-Rate Reed-Solomon Codes Qi Ceng 1 and Daqing Wan 1 Scool of Computer Science Te University of Oklaoma Norman, OK73019 Email: qceng@cs.ou.edu Department of Matematics University
More informationPhysically Based Modeling: Principles and Practice Implicit Methods for Differential Equations
Pysically Based Modeling: Principles and Practice Implicit Metods for Differential Equations David Baraff Robotics Institute Carnegie Mellon University Please note: Tis document is 997 by David Baraff
More informationLogarithmic functions
Roberto s Notes on Differential Calculus Capter 5: Derivatives of transcendental functions Section Derivatives of Logaritmic functions Wat ou need to know alread: Definition of derivative and all basic
More information2.11 That s So Derivative
2.11 Tat s So Derivative Introduction to Differential Calculus Just as one defines instantaneous velocity in terms of average velocity, we now define te instantaneous rate of cange of a function at a point
More informationAnalytic Functions. Differentiable Functions of a Complex Variable
Analytic Functions Differentiable Functions of a Complex Variable In tis capter, we sall generalize te ideas for polynomials power series of a complex variable we developed in te previous capter to general
More informationAverage Rate of Change
Te Derivative Tis can be tougt of as an attempt to draw a parallel (pysically and metaporically) between a line and a curve, applying te concept of slope to someting tat isn't actually straigt. Te slope
More informationMath 242: Principles of Analysis Fall 2016 Homework 7 Part B Solutions
Mat 22: Principles of Analysis Fall 206 Homework 7 Part B Solutions. Sow tat f(x) = x 2 is not uniformly continuous on R. Solution. Te equation is equivalent to f(x) = 0 were f(x) = x 2 sin(x) 3. Since
More informationQuantum Mechanics Chapter 1.5: An illustration using measurements of particle spin.
I Introduction. Quantum Mecanics Capter.5: An illustration using measurements of particle spin. Quantum mecanics is a teory of pysics tat as been very successful in explaining and predicting many pysical
More informationMANY scientific and engineering problems can be
A Domain Decomposition Metod using Elliptical Arc Artificial Boundary for Exterior Problems Yajun Cen, and Qikui Du Abstract In tis paper, a Diriclet-Neumann alternating metod using elliptical arc artificial
More informationGlobal Output Feedback Stabilization of a Class of Upper-Triangular Nonlinear Systems
9 American Control Conference Hyatt Regency Riverfront St Louis MO USA June - 9 FrA4 Global Output Feedbac Stabilization of a Class of Upper-Triangular Nonlinear Systems Cunjiang Qian Abstract Tis paper
More informationChapter 1D - Rational Expressions
- Capter 1D Capter 1D - Rational Expressions Definition of a Rational Expression A rational expression is te quotient of two polynomials. (Recall: A function px is a polynomial in x of degree n, if tere
More informationof measurement uncertainty
Unscented propagation of measurement uncertainty Jan eter Hessling, Jörgen Stenarson, Tomas Svensson S Tecnical Researc Institute of Sweden, Sweden Of current interest Non-linear propagation of uncertainty
More information2.3 Algebraic approach to limits
CHAPTER 2. LIMITS 32 2.3 Algebraic approac to its Now we start to learn ow to find its algebraically. Tis starts wit te simplest possible its, and ten builds tese up to more complicated examples. Fact.
More informationNew families of estimators and test statistics in log-linear models
Journal of Multivariate Analysis 99 008 1590 1609 www.elsevier.com/locate/jmva ew families of estimators and test statistics in log-linear models irian Martín a,, Leandro Pardo b a Department of Statistics
More informationMath 312 Lecture Notes Modeling
Mat 3 Lecture Notes Modeling Warren Weckesser Department of Matematics Colgate University 5 7 January 006 Classifying Matematical Models An Example We consider te following scenario. During a storm, a
More informationMath 31A Discussion Notes Week 4 October 20 and October 22, 2015
Mat 3A Discussion Notes Week 4 October 20 and October 22, 205 To prepare for te first midterm, we ll spend tis week working eamples resembling te various problems you ve seen so far tis term. In tese notes
More informationMath 161 (33) - Final exam
Name: Id #: Mat 161 (33) - Final exam Fall Quarter 2015 Wednesday December 9, 2015-10:30am to 12:30am Instructions: Prob. Points Score possible 1 25 2 25 3 25 4 25 TOTAL 75 (BEST 3) Read eac problem carefully.
More informationWYSE Academic Challenge 2004 Sectional Mathematics Solution Set
WYSE Academic Callenge 00 Sectional Matematics Solution Set. Answer: B. Since te equation can be written in te form x + y, we ave a major 5 semi-axis of lengt 5 and minor semi-axis of lengt. Tis means
More informationThe Krewe of Caesar Problem. David Gurney. Southeastern Louisiana University. SLU 10541, 500 Western Avenue. Hammond, LA
Te Krewe of Caesar Problem David Gurney Souteastern Louisiana University SLU 10541, 500 Western Avenue Hammond, LA 7040 June 19, 00 Krewe of Caesar 1 ABSTRACT Tis paper provides an alternative to te usual
More information2.1 THE DEFINITION OF DERIVATIVE
2.1 Te Derivative Contemporary Calculus 2.1 THE DEFINITION OF DERIVATIVE 1 Te grapical idea of a slope of a tangent line is very useful, but for some uses we need a more algebraic definition of te derivative
More informationA MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES
A MONTE CARLO ANALYSIS OF THE EFFECTS OF COVARIANCE ON PROPAGATED UNCERTAINTIES Ronald Ainswort Hart Scientific, American Fork UT, USA ABSTRACT Reports of calibration typically provide total combined uncertainties
More information