Computing on Virtual Slow Manifolds of Fast Stochastic Systems.

Similar documents
Nested Stochastic Simulation Algorithm for Chemical Kinetic Systems with Disparate Rates. Abstract

Introduction to multiscale modeling and simulation. Explicit methods for ODEs : forward Euler. y n+1 = y n + tf(y n ) dy dt = f(y), y(0) = y 0

PARAMETERIZATION OF NON-LINEAR MANIFOLDS

Fast Probability Generating Function Method for Stochastic Chemical Reaction Networks

STOCHASTIC CHEMICAL KINETICS

Efficient Leaping Methods for Stochastic Chemical Systems

Gillespie s Algorithm and its Approximations. Des Higham Department of Mathematics and Statistics University of Strathclyde

Vector spaces. DS-GA 1013 / MATH-GA 2824 Optimization-based Data Analysis.

Introduction to asymptotic techniques for stochastic systems with multiple time-scales

which arises when we compute the orthogonal projection of a vector y in a subspace with an orthogonal basis. Hence assume that P y = A ij = x j, x i

1. Introduction Systems evolving on widely separated time-scales represent a challenge for numerical simulations. A generic example is

Getting Started with Communications Engineering

Linear Algebra Massoud Malek

Exploring the energy landscape

Comparison of Modern Stochastic Optimization Algorithms

PHYSICAL REVIEW LETTERS

Lecture Notes 1: Vector spaces

MODULE 8 Topics: Null space, range, column space, row space and rank of a matrix

, b = 0. (2) 1 2 The eigenvectors of A corresponding to the eigenvalues λ 1 = 1, λ 2 = 3 are

Focus was on solving matrix inversion problems Now we look at other properties of matrices Useful when A represents a transformations.

Region Description for Recognition

arxiv: v1 [hep-ph] 5 Sep 2017

Upon successful completion of MATH 220, the student will be able to:

Gene regulatory networks: A coarse-grained, equation-free approach to multiscale computation

Problem set 7 Math 207A, Fall 2011 Solutions

Nested Stochastic Simulation Algorithm for KMC with Multiple Time-Scales

DS-GA 1002 Lecture notes 0 Fall Linear Algebra. These notes provide a review of basic concepts in linear algebra.

(17) (18)

Lecture 1: Brief Review on Stochastic Processes

Matrices and Matrix Algebra.

Least Squares Approximation

THEODORE VORONOV DIFFERENTIABLE MANIFOLDS. Fall Last updated: November 26, (Under construction.)

Notes on Time Series Modeling

Extending the Tools of Chemical Reaction Engineering to the Molecular Scale

Linear Algebra. Preliminary Lecture Notes

Coarse bifurcation analysis of kinetic Monte Carlo simulations: A lattice-gas model with lateral interactions

ISOMETRIES OF R n KEITH CONRAD

exceptional value turns out to be, "RAC" CURVATURE calling attention especially to the fundamental Theorem III, and to emphasize

Using the computer to select the right variables

Dimension Reduction Techniques. Presented by Jie (Jerry) Yu

EE731 Lecture Notes: Matrix Computations for Signal Processing

Linear Systems. Carlo Tomasi

MATH 23a, FALL 2002 THEORETICAL LINEAR ALGEBRA AND MULTIVARIABLE CALCULUS Solutions to Final Exam (in-class portion) January 22, 2003

Introduction, basic but important concepts

Unconstrained optimization

Stiffness in stochastic chemically reacting systems: The implicit tau-leaping method

12. Cholesky factorization

Jim Lambers MAT 610 Summer Session Lecture 2 Notes

Manifold Regularization

Lecture: Adaptive Filtering

Math 302 Outcome Statements Winter 2013

Linear & nonlinear classifiers

Multi-Robotic Systems

Linear Algebra. Preliminary Lecture Notes

22.4. Numerical Determination of Eigenvalues and Eigenvectors. Introduction. Prerequisites. Learning Outcomes

Chapter 9 Implicit integration, incompressible flows

The natural numbers. Definition. Let X be any inductive set. We define the set of natural numbers as N = C(X).

5.3 The Power Method Approximation of the Eigenvalue of Largest Module

Variable-free exploration of stochastic models: A gene regulatory network example

Connection of Local Linear Embedding, ISOMAP, and Kernel Principal Component Analysis

Course Summary Math 211

New attacks on Keccak-224 and Keccak-256

MA 323 Geometric Modelling Course Notes: Day 07 Parabolic Arcs

1 Inner Product and Orthogonality

Introduction - Motivation. Many phenomena (physical, chemical, biological, etc.) are model by differential equations. f f(x + h) f(x) (x) = lim

Maximum variance formulation

Linear Algebra Review. Vectors

Beyond Scalar Affinities for Network Analysis or Vector Diffusion Maps and the Connection Laplacian

Linear Systems. Carlo Tomasi. June 12, r = rank(a) b range(a) n r solutions

Lie Groups for 2D and 3D Transformations

Zoology of Fatou sets

Statistical Pattern Recognition

Implicitely and Densely Discrete Black-Box Optimization Problems

+E x, y ρ. (fw (x) y) 2] (4)

Blind Source Separation with a Time-Varying Mixing Matrix

Session-Based Queueing Systems

M. A. Gusmão IF-UFRGS

An Introduction to Stochastic Simulation

COMP 558 lecture 18 Nov. 15, 2010

SPARSE signal representations have gained popularity in recent

The Adaptive Explicit-Implicit Tau-Leaping Method with Automatic Tau Selection

Discretization of SDEs: Euler Methods and Beyond

Classroom Tips and Techniques: Eigenvalue Problems for ODEs - Part 1

Sequential Monte Carlo methods for filtering of unobservable components of multidimensional diffusion Markov processes

Discriminative Direction for Kernel Classifiers

= w. These evolve with time yielding the

This can be accomplished by left matrix multiplication as follows: I

Numerical Simulation for Biochemical Kinetics

A lower bound for scheduling of unit jobs with immediate decision on parallel machines

Root systems and optimal block designs

Econ Slides from Lecture 7

Singular Value Decomposition

ECE133A Applied Numerical Computing Additional Lecture Notes

Applied Linear Algebra in Geoscience Using MATLAB

Decoherence and the Classical Limit

Manifold Coarse Graining for Online Semi-supervised Learning

Intersection of Objects with Linear and Angular Velocities using Oriented Bounding Boxes

Kernel Methods. Charles Elkan October 17, 2007

arxiv: v1 [math.ra] 13 Jan 2009

ELEMENTS OF PROBABILITY THEORY

Transcription:

European Society of Computational Methods in Sciences and Engineering (ESCMSE) Journal of Numerical Analysis, Industrial and Applied Mathematics (JNAIAM) vol. 1, no. 1, nnnn, pp. 1-3 ISSN 179 814 Computing on Virtual Slow Manifolds of Fast Stochastic Systems. C. W. Gear, 1 D. Givon and I. G. Kevrekidis Department of Chemical Engineering, Princeton University, Princeton, NJ, USA Received nn.nn.nn Abstract: The persistently fast evolutionary behavior of certain differential systems may have intrinsically slow features. We consider systems whose solution trajectories are slowly changing distributions and assume that we do not have access to the equations of the system, only to a simulator or legacy code that performs step-by-step time integration of the system. We characterize the set of all possible instantaneous solutions by points on a low-dimensional virtual slow manifold (VSM) and show how, when there is a sufficiently large gap between the time scales of the fast and slow behaviors, we can restrict the fast behavior observed numerically through simulation to the VSM and perform useful computations more efficiently there. c nnnn European Society of Computational Methods in Sciences and Engineering Keywords: Projective Integration, Steady State, Legacy Codes, Equation Free Mathematics Subject Classification: 65L5,65M99 1 Introduction As a trivial illustrative example of the type of systems of concern to this paper, we consider the chemical reaction k 3 ϕ Y (1) Y k 1 Z + Z (2) k 2 (Here, the species Y and Z are present in non-negative integer quantities Y and Z. The meaning of (1) is that new Y s are arriving at a Poisson rate k 3 and (2) indicates that Y s are changing into two Z s at a Poisson rate k 1 Y and the reverse process occurs at a Poisson rate k 2 Z(Z 1)/2.) 1 Corresponding author. E-mail: wgear@princeton.edu

2 C. W. Gear, D. Givon, I. G. Kevrekidis If the values Y and Z are small, this system has to be simulated as a discrete event system - the chemical master equation. As the number of each species present increases, the variables Y and Z can be treated as real values satisfying stochastic ODEs - the chemical Langevin equations. As the numbers get arbitrarily large, the law of large numbers implies that Y and Z approach the solutions of deterministic ODEs - the chemical kinetic equations. Consider the above example with k 1 =.25 and k 2 = 2. Ignoring, for the moment, the steady growth in Y due to (1) we have a stochastic partial equilibrium[1] between Y and Z when k 1 Y = k 2 Z(Z 1)/2 is satisfied in its mean. If, for example, Y = 1 6 then Z is about 5. The latter is a small number, so Z will be stochastic with that average, but 1 6 is large enough that Y is noise free to the first four or five digits. Figure 1 shows the solution of this problem with k 3 = 1 5 and the initial condition Y = 1 6, Z =. (This was computed using discrete event simulation with the Gillespie SSA algorithm[2].) Note that while Z starts at it almost immediately reaches its partial equilibrium average of around 5, and, as Y increases to about 2 1 6 over the simulation interval because of (1), the average of Z increases to about 5 2. Z remains stochastic but Y can be well approximated by the deterministic ODE Y = k 3. 2 x 16 12 1.9 1.8 1 1.7 8 # Y species 1.6 1.5 1.4 # Z species 6 1.3 4 1.2 2 1.1 1 1 2 3 4 5 6 7 8 9 1 Time 1 2 3 4 5 6 7 8 9 1 Time Figure 1: Simulation of reactions (1) and (2) with k 1 =.25, k 2 = 2 and k 3 = 1 5. The time scales in this problem are the time scales of the relative changes in Y and Z. The ratio of these is k 2 (Z 1)/(2k 3 /Y ) which is about 5. In the Y -Z plane, the solution over a very short time interval is a collection of points with almost the same value of Y but Z values drawn from a distribution. As the time scale ratio is increased the solution to a short simulation approaches a distribution in Z at a constant value of Y. The toy example above is extremely simple so the value of Y is almost independent of the randomness of Z, but in more complex cases, for example, if the reaction rate k 3 were non-linearly dependent on Z, the distribution of Z would affect the evolution of Y. For another chemically inspired example see [3]. In this paper we consider evolutionary problems whose instantaneous solutions are distributions on a manifold for which both the distribution and the manifold evolve slowly in time. We call the instantaneous solution (the distribution plus the manifold it lies on) a cloud, C(t). The key assumption is that, regardless of initial value, the fast random behavior reaches local stationarity on a time scale so much faster than the evolution of the description of C(t) that it can be considered instantaneous. Hence, we assume that the internal structure of the problem is equivalent to the system: y = g(y, µ y ) f(y, z)dµ y (3) z µ y

Virtual Slow Manifolds 3 where y is an n y -dimensional vector and µ y is a distribution (that depends only on y) of points, z, in an n z -dimensional space. (Alternatively the values of z - and their distribution µ y - might be specified by an extremely fast SDE so that its solution approaches samples from its local stationary distribution quickly.) The output from the legacy code or simulator that defines the system consists of its estimates of y and samples z, hence the dimension of the variables we observe in the problem is n = n y + n z. Methods exist for equations in the form (3) and essentially consist of averaging f over a suitable sample (see, for example [4, 5] and [6]). However, we further assume that neither can we access the internal structure of the problem nor can we directly observe the variables y and z. Instead, we assume that the observables are the unknown, non-linear, non-singular transformation of y and z w = w(y, z) (4) where w(, ) R n y R n z R n. (We use bold face symbols when we want to emphasize that the value is a vector in n-dimensional space since necessarily n > 1 in these problems.) Here we are not using observables in its control theory sense as (functions of) a subset of the internal state variables, but simply to indicate that these are the variables whose values we can obtain. In fact, because in a sense the state variables are y, these observables are functions of a larger set of variables. In the y-z space, C(t) for (3) is the distribution µ y on the linear manifold y = y(t) where y(t) is the solution (3). In w-space, C(t) s manifold is (generally) curved. The solution in w-space is a trajectory of clouds. In (3) the initial value of y = y is all that is needed to determine the trajectory. Hence, in the w-space, any initial value w = w(y, z) for arbitrary z i.e., any point in the initial cloud determines the same cloud trajectory. If we knew the transformation (4) (and its inverse) and the distribution µ y, the value of y(t) would enable C(t) to be calculated, so the set of all C(t) has n y dimensions. Unfortunately we do not know w(y, z) or µ y so our objective is to define a suitable n y -dimensional Virtual Slow Manifold (VSM) in w-space that characterizes C(t) in the sense that each point in the VSM is a (1-1) mapping of each cloud C(t) that is generated by (3) transformed by (4). Figure 2 shows 5 samples from each of ten clouds, C(t i ), at ten different time points, t i. These are drawn from the solution of an artificial example with n y = n z = 1 for which we know w and µ y (contrary to our general assumption). The knowledge of w and µ y allows us to obtain such samples computationally. Unfortunately under our original assumptions we can t directly compute such samples because the differential system is provided in the form of a simulator code or a legacy code, W, for the model (perhaps with a fixed step size) that approximates w at t n+1 = t n + h using w n+1 = W (w n, t n, h) (5) All this allows us to do is to produce an integration sequence I s (w ) = {w i }, i, = 1, N, which is the sequence of N consecutive values obtained from (5) starting from w. An example of such a sequence is shown as the set of magenta dots in Figure 3. Each point belongs to a different C(t i ). The sequence does not provide samples from a single cloud but contains statistical information about a sequence of clouds and hence about the VSM trajectory. We are interested in analyzing the slow behavior of the system - that is, the properties of the clouds. At the lowest level we want to determine their evolution in time by computing their derivatives. With that information we can use a zero finder to compute stationary states (stationary states of the unobservable y s, not of the output of the microscopic model which may continue to exhibit stochastic behavior even at its macroscopic stationary state). With these basic tools we can compute other information such as bifurcation diagrams or periodic solutions.

4 C. W. Gear, D. Givon, I. G. Kevrekidis.55.54.53.52.51 w 2.5.49.48.47.46.46.48.5.52.54.56.58 w 1 Figure 2: Samples from a sequence of ten clouds. The key assumption that we are making is that the stochastic behavior is sufficiently fast that the slow behavior can be well approximated as deterministic. We will also assume that we know the dimension, n z, of the fast stochastic behavior, although we will indicate how this can be estimated numerically. In Section 2 we will discuss how to define VSMs and then illustrate some ways to perform computation efficiently on a VSM in Section 3. In the last section we will present an example of accelerating the computation of cloud evolution using projective integration. 2 VSMs A VSM is not unique - we are free to define it in any way that provides a (1-1) mapping between points in the VSM and the clouds. The mapping from clouds to points in a VSM is a restriction in the sense used in [7], while the inverse mapping is the lifting step. We want to define a VSM such that the mapping is easy to compute that is, given a cloud it should be easy to compute a point in the VSM, and given a point in the VSM it should be easy to compute the corresponding cloud - or, more practically, a point in the cloud, since that serves to specify the cloud. In this paper we are going to limit ourselves to systems for which the cloud mean, given by m(c(t)) = m(t) = wdµ y(w), (6) w C(t) is a (1-1) map of a cloud. For the toy example given in the introduction the mean is clearly a (1-1) map and is itself a point in the cloud but the latter is not true in general. If the mean is a (1-1) map then m(t) lies in an n y -dimensional manifold in R n. (If we knew an n y -dimensional subset of the w coordinates that were guaranteed to parameterize this manifold we could compute in that lower dimensional space instead.) Note that the cloud mean may not, in general, be a (1-1) map of the clouds. Suppose, for example, that n y = n z = 1, µ y in (3) is the uniform distribution on [, 2π], and the transformed variable is w(y, z) = [y sin(z), y cos(z)] T In this case, the clouds in w-space are uniform distributions on concentric circles so their means

Virtual Slow Manifolds 5.54.53.52.51 w2.5.49.48.47.46.45.45.46.47.48.49.5.51.52 w1 Figure 3: Integration sequence, Is (magenta dots); Best linear fit to Is (blue line segment); Set back of Is to t (red dots) and set forward to tn (green dots) - see text Section 3.1. are constant. In such a case we would have to extract additional information such as the variance of a variable to uniquely identify a cloud. Although the VSM is an ny -dimensional manifold, the cloud mean is a point in the original Rn. If we had to include additional information, such as the variances or other moments, the VSM would be a sub manifold of an even larger space. Although we will generally work in Rn (or a larger space if we need additional moments or other information to characterize a cloud), the fact that the VSM is of lower dimension can be exploited. 3 Numerical Methods The two primary numerical tasks we want to perform are to compute the evolutionary behavior as inexpensively as possible, and to compute stationary states of the system. The most basic operation we have to implement is an estimation of the time derivative of the trajectory along the VSM, in this case dm/dt. We can make an approximation to the VSM trajectory locally by performing a linear least squares fit to each coordinate in the integration sequence Is as a function of time (higher order fits could be used) to get mf it (t) = m + st (7) (In http://www.princeton.edu/~wgear/bestslope.pdf it is shown that under some conditions an equally-weighted least squares fit is the approximation with the least variance.) To emphasize that m and s are functions of the initial condition w we will sometimes write them as m (w ) and s(w ). The result of a linear fit to the data in Figure 3 is plotted there as a blue line segment. The end points of that segment, mf it () and mf it (tn ), are approximations to m(t ) and m(tn ) while its slope, s, is an estimate of the local slope of the VSM trajectory. 3.1 Cloud sampling With an estimate of the slope of the VSM we can generate approximations to samples from a cloud at a fixed time by moving each point in the integration sequence forward (or back) in the direction of the VSM slope based on its time difference from the time of the desired cloud. (This assumes c nnnn European Society of Computational Methods in Sciences and Engineering (ESCMSE)

6 C. W. Gear, D. Givon, I. G. Kevrekidis that the cloud is not significantly changing shape or scale over the interval - part of our separation of times scales assumption.) Figure 3 also shows the points in the integration sequence of Figure 3 set back to the initial time (red points) and also set forward to the final time (green points). Since we get an estimate of the VSM derivative directly from the straight line approximation to the mean, we do not need cloud samples to do derivative estimation. We also get approximations to the cloud mean trajectory at any point directly from the least squares fit (7) so we do not need cloud samples for the mean calculation either. We will, however, need cloud samples to find regions of high density because, for reasons to be discussed, these will be the favored working points on a cloud. 3.2 Stationary state computation Finding a stationary cloud requires that we solve s(w ) = (8) The direction s approximately lies in the local tangent hyperplane (henceforth, TH) to the VSM and so it has n y degrees of freedom. Any change of w to another point in the manifold of the cloud that contains w does not change the value of s(w), so small perturbations to w in the local TH to the cloud manifold have no effect on s(w). Hence, although s and w are n-dimensional vectors, (8) is essentially an n y -dimensional problem. Consequently it makes sense to confine the search for solutions, w, to directions locally orthogonal to the cloud manifold at our working point. Let the n by n y matrix N p span these directions so that we can look for solutions of the form w = w init + N p α where w init is the initial approximation for an iterative solution and α is the n y -dimensional unknown. We would also like to project (8) into a suitable n y -dimensional linear manifold to reduce its dimension to n y. Ideally we should project (8) onto the local TH of the VSM but we do not know that TH. Instead we will project onto the hyperplane, N p, orthogonal to the cloud manifold at our working point. N p is spanned by the matrix N p discussed in the previous paragraph. This will allow us to solve (8) provided N p contains no directions orthogonal to the VSM locally. While we cannot guarantee that this is true, its falsity would mean that at least one direction in the VSM is a direction in the TH to the cloud. The VSM, given by the mean over a cloud, is most strongly influenced by the regions of high density in the cloud. Consequently, it is unlikely that mean will move in a direction that is in the TH of the cloud near a region of high density where we will chose our working point. Hence we solve N T p s(w init + N p α ) = (9) for α where w init is near a region of high cloud density and N p is orthogonal to the TH to the cloud near w init. If we started at w init in the neighborhood of the solution we could compute N p once and then leave the process to any non-linear equation package to solve (9). If we are not close to the solution we may need to recompute N p from time to time. Thus, we need a way to find a point in the cloud that is in a region of high density and also to compute an approximation to the TH to the cloud and its orthogonal complement at that point. 3.3 Evolutionary computation The coarse projective integration method [7] can be used to essentially operate in the VSM to capture the slow behavior of these systems. It has the following structure: 1. Start with an initial value w and j =.

Virtual Slow Manifolds 7 2. Generate an integration sequence starting from w j. 3. Compute m(w j ) and s(w j ). 4. Estimate the mean, m fit (T ), at time T = t j + (M + 1)Nh from (7) 5. Find a w j+1 such that m(w j+1 ) m fit (T ) (1) 6. w j+1 is a point on the initial cloud for the next outer step and m(w j+1 ) is its mapping in the VSM. 7. j j + 1 and repeat steps 2 to 7 until done. Solving (1), which is the lifting step in [7], is similar to solving (8) in the stationary problem (since m, like s, is an n-dimensional object lying on an n y -dimensional manifold) although it has an additional complication. The true m (w j+1 ) lies in the n y -dimensional VSM (although its computed approximation will probably not) while m fit (T ) is an approximation that almost certainly does not lie in the VSM. Hence, it is in general impossible to solve this equation exactly. Again, we project (1) onto the n y -dimensional hyperplane orthogonal to the TH to the cloud manifold at the most recent operating point and solve for w also restricted to an n y -dimensional linear manifold: N T p (m(w init + N p α) m fit (T )) =. (11) This must be solved for α with an iterative method for which the initial guess, w init, is needed. In addition to the TH we also need to compute a good initial estimate, w init, for (11). Because we have typically already computed some information about the solution, we can get a good approximation, unlike the stationary problem discussed above where we have no basis for choosing an initial guess unless we have some external information (e.g., a solution at a nearby parameter value in a continuation). 3.4 Other computational tasks Approximating a TH to a cloud at a working point w Let I s be the integration sequence starting at w and define wi = w i t i s(w ). (12) If C(t) moves linearly in time without change in shape or distribution, wi is a sample of points from C(); otherwise it is only an approximation. We call these setback points. They are shown for the I s in Figure 2b as red dots. A point at which a good approximation of the local TH to the cloud can be computed should be in a region of high sampling density. To locate such a point, after computing the wi points we define a point, w c, called their center, that has many near neighbors in the sample (with a technique to be discussed later). Next, we find the N s nearest neighbors, w j, j = 1 N s of w c by looking for those with the smallest w c w j 2, where N s is a small fraction of the number of points in the sample. Form the n by N s matrix W whose i-th column is w i w c. Since the columns of W are the directions to near neighbors of w c, they approximately lie in the TH at w c. Let QRP = W be the QR factorization of W where P is a permutation matrix such that R has non-increasing diagonal elements. Then the first n z columns of Q approximately span the TH at w c and the remaining n y columns are orthogonal. (If n z is initially unknown, this process can be used to estimate n z by looking for the first significant decrease in the relative sizes of the diagonal elements of R.) Figure 4 shows an example of computing the TH to a cloud. The problem has n y = 1 and n z = 2. The red dots are samples from the initial cloud (found by setback from an integration sequence) and the two black lines are orthogonal vectors in the TH at the center point of the cloud. The blue line is orthogonal to the TH.

8 C. W. Gear, D. Givon, I. G. Kevrekidis.7.6.5.4.3.2.1.1.2.3 1.5.5 1 1.2 1.8.6.4.2.2 Figure 4: Initial and final clouds (red and green), span of tangent hyperplane to red cloud (black lines) and vector orthogonal to those (blue line) in clouds generated from an example with n y = 1 and n z = 2. Finding the center of a cloud sample We define the center of a cloud to be the point with a maximum value of a property we call centerness designated as c i for point i and defined as j λp ijc j where P ij is a proximity measure which is 1 for co-located points and zero for distant points. We use P ij = exp( σd 2 ij ) where d ij is the Euclidean distance between points i and j and σ is a scaling factor although other functions could be used. If P = {P ij } and c = {c i } then P c = λc for the largest eigenvalue, λ, of P. This vector can be found easily with a few iterations of a power method. Figure 5 shows the result of this algorithm applied to a cloud sample. 6.3 5.25.2 4.15 3.1 2.5.2.4 1.1.5.5.1.15.2.2 Figure 5: Center computation for an example with n y = 1 and n y = 2 - color indicates centerness of sample points in a cloud. Computing w init A good estimate for w init can be obtained by projecting a point forward from the initial cloud of the previous step using the same projective step as for the mean. We choose to

Virtual Slow Manifolds 9 project the center point, w c w init = w c + st (13) In fact, this is often a sufficiently good starting point and can be used without further iteration. Although its error may be larger than in the solution of (11), this error can be reduced by using a shorter projective step, and this may be less costly, overall, than using a larger projective step plus the expensive solution of (11). 4 A Numerical Example We illustrate the use of (13) as the projective integration formula below - that is, instead of solving (11) for w(t ) we set w(t ) = w init from (13). The following artificial example was used for this and other tests. It is defined in slow-fast form and then subject to an invertible non-linear transformation. Knowledge of this transformation is used to generate the correct solution; yet the numerical method under test does not make use of that knowledge. y 1 = y 3 1 y ny + β z 2 /n z y i = y 3 i + y i 1 + β z 2 /n z, i = 2,, n y (14) z j = N[(1 + mean(y))σ, mean(y)ρ], j = 1,, n z where N[s, m] is a Gaussian distribution with mean m and standard deviation s, mean(x) is the mean of the components of x, and β and ρ are parameters. The (presumed unknown) non-linear transformation to the w variables is given by the steps: u i = y i + γ(2 + y 1 )i z 2, i = 1,, n y w j = mean(u)z j, j = 1,, n z (15) w nz+i = 2u i + mean(w j, j = 1 n z )i where γ is a parameter that controls the amount of curvature in the cloud (which is parabolic). The y s in (15) satisfy a deterministic equation that can be found by averaging over the distributions of z and this is used to generate the true solution via the use of a high order numerical integrator. For this test example, the microscopic simulator for w also makes use of its knowledge of the transformation (15). Given w it computes the y s from (15) and then executes one approximate integration step for the y s by choosing random values of the z s according to the distributions given and then using a forward Euler step for the y s where the derivatives are evaluated using the previous y values and the just generated random values of z. In Figure 6a we show three steps of a projective integration that projects the center of the initial cloud in the direction of the mean. In this example, n y = 3, n z = 1, β =.3, σ =.1, ρ =.1, and γ = 1. An integration sequence of 2, steps was taken to estimate the linear approximation to the trajectory in the VSM, while the first 2, points in the integration sequence after setback were used to estimate the center of the initial cloud. The microscopic integrator for the system used fixed step size h = 1 7. The projective factor, M, was 19 so the projective step length was H = 4 1 2. The problem is 4-dimensional; the last three components of w are plotted. The center projections are shown as straight lines. The very different directions of the three projective steps indicate significant sample noise. However, for clouds of low curvature, the predominant component of the noise at the next initial point lies within the new cloud. This can be seen by considering the case of a linear w(y, z) in which case we can just consider the y-z slow-fast system. All of the stochastic noise is in z and has no impact on the accuracy of the projective integration of y.

1 C. W. Gear, D. Givon, I. G. Kevrekidis 1.9 1.85 1.8 1.75 1.7 1.65 1.6 1.55.15.2.25.3.35.5.45 time.4.35.3 Figure 6: Result of 3 steps of projective integration of the cloud center point. The axes are the variables w i, i = 2, 3, 4 in (15) transformed by (15) with n y = 3 and n z = 1. The initial cloud is red - generated from the first 2, points of the 2, point I s. The line segments start at the center of the initial cloud, are in the direction of the best linear fit to C(t), and end at the initial point for the next I s. The final cloud is green - generated from the last 2 points. We ran the projective integration over the interval [,1] - which would have taken 1 8 inner steps if the projective process was not used, but instead took 5 1 6 inner steps. The equation for y was integrated to 6-digit accuracy using ode45 in Matlab for comparison. The results from the projective method were transformed back to y and z so we could show the errors in y with time in Figure 7. The projective integration is almost M + 1 = 2 times more efficient (there is the overhead for least squares and center calculation, so the actual speedup depends on the relative cost of integration steps). We also ran the same problem with M = 99 so that it was almost 1 times faster. The results are shown in Figure 8, which plots the three components of y in solid lines from the true solution and circles from the projective solution. While the errors are larger, most of the error is due to the first order nature of the particular projective method we used. To illustrate this, we integrated the slow equations directly with forward Euler using the same step size as the projective step. The errors in the projective solution for the y components are shown as solid lines in Figure 9 while the differences between the projective solution and the Forward Euler solution are shown as dashed lines. From this figure we can see that the error due to the first order integration formula is the dominant part of error most of the time. 5 Conclusion We have indicated how computation in a certain type of multi-scale problems can be sped up by definition of a suitable VSM. In this paper we have restricted ourselves to problems in which the slow components are not stochastic (which requires a very large difference in time scales or, in the case of chemically reacting systems, large differences in species concentrations). If the slow components are stochastic, a natural extension to this work would be to fit an SDE to the slow behavior of, for example, the cloud mean.

Virtual Slow Manifolds 11.2.2 1 2 3 4 5 6 7 8 9.2.2 1 2 3 4 5 6 7 8 9.2.2 1 2 3 4 5 6 7 8 9 Figure 7: Errors as functions of time in the each of the three y components recovered from the projective stochastic computation of (15) transformed by (15). Acknowledgment This work was partially supported by the US DOE. References [1] Y. Cao, D. T. Gillespie and L. Petzold, Multiscale stochastic simulation algorithm with stochastic partial euilibrium assumption for chemically reacting systems, J. Com. Phys., 26, 395-411 (25) [2] D. T. Gillespie, Exact stochastic simulation of coupled chemical reactions, J. Phys. Chem., 81, 234-2361 (1977). [3] A. Singer, R. Erban, I. G. Kevrekidis and R. R. Coifman, Detecting intrinsoc slow variables in stochastic dynamical systems by anisotropic diffusion maps, PNAS, 16, 169-1695 (29) [4] R. Z. Khasminski, On the principle of averaging the Ito s stochastic differential equations. Kybernetika (Prague), 4,26-279 (1968). [5] A. Yu. Veretennikov, On an averaging principle for systems of stochastic differential equations. Mat. Sb., 181(2), 256-268 (199). [6] W. E, D. Liu, and E. Vanden-Eijnden, Analysis of multiscale methods for stochastic differential equations, Communications on Pure and Applied Mathematics 58, 1544 1585 (25). [7] I. G. Kevrekidis, C. W. Gear, J. M. Hyman, P. G. Kevrekidis, O. Runborg, and K. Theodoropouloss, Equation-free coarse-grained multiscale computation: enabling microscopic simulators to perform system-level tasks, Comm. Math. Sciences 1, 715 762 (23). [8] C. W. Gear, and I. G. Kevrekidis, Projective Methods for Stiff Differential Equations, SIAM J. Sci. Comp 24, 191 116 (23).

12 C. W. Gear, D. Givon, I. G. Kevrekidis 1.8.6.4.2.2.4.6.8 1 1 2 3 4 5 6 7 8 9 1 Figure 8: True solution for each component of y (solid line) and projective solution (circles) with M = 99. The dashed line is the Forward Euler solution using step size H - the length of the projective step. (See also the caption for Figure 7.).1.5.5.1 1 2 3 4 5 6 7 8 9.1.5.5.1 1 2 3 4 5 6 7 8 9.1.5.5.1 1 2 3 4 5 6 7 8 9 Figure 9: Errors in projective solution (solid lines), and the difference between the projective solution and the Forward Euler at same step size (dashed line). (See also caption for Figure 7.)