Methods of Data Analysis Random numbers, Monte Carlo integration, and Stochastic Simulation Algorithm (SSA / Gillespie)

Size: px
Start display at page:

Download "Methods of Data Analysis Random numbers, Monte Carlo integration, and Stochastic Simulation Algorithm (SSA / Gillespie)"

Transcription

1 Methods of Data Analysis Random numbers, Monte Carlo integration, and Stochastic Simulation Algorithm (SSA / Gillespie) Week 1 1 Motivation Random numbers (RNs) are of course only pseudo-random when generated on our computers: they are determined uniquely by the seed value and the algorithm for producing the next pseudorandom number from the last in the generated sequence. Care must be taken to use a high-quality RNG, and to always know what is going on with the seed (especially if you run multiple jobs on the cluster). RNs are very useful for a series of numerical procedures, three of which we will look at today. Next week, we will also use RNs for Metropolis Monte Carlo simulations. 2 Goals 1. Generating random numbers from a non-uniform distribution by using uniform random numbers (URNs). 2. Monte Carlo integration, especially when the integrals involved are in high dimension. 3. Gillespie SSA simulation: illustrating the relation between a deterministic approximation to an underlying stochastic system, and true stochastic dynamics. 3 Generating random numbers from a given distribution As a rule, standard RNGs generate random numbers on an interval [0, 1) distributed uniformly, but one might require random numbers from some other distribution, e.g., Gaussian (for which a lot of software already provides the generator, for instance, Matlab s randn), or some arbitrary distribution function for which no such generator is provided. 3.1 Transformation method Suppose you can generate URN on unit interval: let x be uniformly distributed such that p(x) = 1. We would like random numbers y generated according to a distribution p(y) = f(y), with some given f, satisfying f(y) 0 and dyf(y) = 1. We would like to produce y as a functional transformation of URNs x. Consequently, the probabilities must transform as: p(y) = p(x) dx dy. (1) 1

2 Figure 1: Left: Generating points in a sphere where each of the spherical coordinates is drawn uniformly on its domain; points are not uniformly distributed in the volume. Right: the correct prescription by the transformation method of Eq. (6). As p(x) = 1, this implies a differential equation for f(y), f(y) = dx dy, which is solved by or x = y f(y )dy (2) y = F 1 (x), (3) where F (y) = f(y )dy is the cumulative distribution function (CDF) of the desired distribution that we want to generate, f. So to be able to generate random numbers according to f quickly, we need to be able to evaluate quickly the inverse of the CDF of f; how easy this is depends on the desired distribution. The graphical intuition behind this construction is simple: a particular value of URN x is drawn in the range of the CDF for f (which is always [0, 1]), and then y is identified as the value that contains the drawn total probability weight to the left of y. Note that this also allows us also to generate random numbers over infinite domains, for instance, for Gaussian random numbers. For details, see Numerical Recipes chapter 7.2 Example: Generating random numbers in a sphere. Suppose we want to generate points uniformly distributed in the volume of a sphere with radius R (and V = 4πR 3 /3), i.e., we would like P (V ) = 1/V, such that dv P (V ) = 1. Points in 3D spherical volume are compactly parametrized in spherical coordinates, (r, θ, φ), where r [0, R], θ [0, π], φ [0, 2π); these coordinates map into the cartesian system via: x = r sin(θ) cos(φ) y = r sin(θ) sin(φ) z = r cos(θ). (4) It would be incorrect, however, to conclude that generating points uniformly sampled in volume 2

3 amounts to generating URNs for each of the three coordinates, r, θ, φ, within their respective domains. This is clearly illustrated in Fig. 1 (left), and it leads to overly concentrated points towards the center of the sphere as well as its north/south poles. The functional transformation between the spherical volume element and spherical coordinates is given by the determinant of the Jacobian of the system in Eq. (4), i.e., dv = r 2 drd(cos θ)dφ. If, then, we want to draw spherical coordinates independently from distributions P θ (θ), P r (r), and P φ (φ) that we have yet to idenfiy, we must have: P (V )dv = P r (r)drp φ (φ)dφp θ (θ)dθ. (5) Given that P (V ) = 1/V is uniform and we know how dv transforms, we can insert both quantities into the above equation and recover: P r (r) = 3r2 R 3 P θ (θ) = 1 2 sin θ P φ (φ) = 1 2π. You can verify that the above distributions are properly normalized so that their integrals over the variable domains are 1 (note also that the units of the distribution for r are, correctly, those of inverse length). We next need to compute the CDFs of these distributions and invert them to get the prescription for generating points uniformly in a spherical volume; this gives (if z are URNs different for each coordinate): 3.2 Rejection sampling θ = arccos(1 2z) r = Rz 1 3 φ = 2πz. (6) Another easy approach to generate samples according to a desired distribution f(y) does so by selectively discarding samples. Here we use RNG to generate N random number pairs uniformly on a square [0, A] [0, B] which fully covers the distribution f(y) on y [0, A] domain (the uniform distribution in this case is referred to as the trial function ), to obtain a sequence (y i, z i ), for i = 1,..., N. Take each pair in turn and check whether z i < f(y i ); if yes, retain y i, otherwise discard it. The resulting set of y i is distributed according to f(y). Graphically, the procedure will uniformly cover the desired distribution f(y), so the number of samples in some small interval [y, y + dy] will be f(y)dy, which by definition means that samples are distributed as desired. You can also think about this as follows. Initially, the trial points are drawn according to P 0 (y, z) = (AB) 1, i.e., uniformly on the square that covers the desired distribution. Each such trial point is kept ( accepted ) with acceptance function P acc = Θ(f(y) z), where Θ is a Heaviside step function, Θ(x) = 1 for x 0 and is 0 otherwise. Then, the distribution of values of the y coordinate of accepted points only will be: dz P0 (y, z)p acc P (y) =. (7) dy dz P0 (y, z)p acc 3

4 You can convince yourself that the denominator (the normalizing factor) is just the area under the desired distribution (or the fraction of accepted points), and the numerator is the conditional distribution of accepted points. If you plug in for P 0 the uniform distribution and the acceptance function, you will find that P (y) = f(y), as desired. Generating a single desired random number hence takes 2 URNs and a single evaluation of f( ). Rejection sampling is more efficient if you can generate trial points uniformly under a function that covers the desired f(y) as tightly as possible; in other words, the trial function need not be uniform. This is in particular useful for generating random numbers on unbounded domains; in this case, transformation method can generate e.g., Lorenzian or Gaussian distributed numbers which can be made to cover the desired distribution (e.g., Gamma or Poisson) as trial functions, from which you subsequently discard the samples (cf. Numerical Recipes Chapter 7.3). 4 Monte Carlo integration Often, various integrals are impossible to evaluate analytically. On the other hand, analytically solvable limits may be of insufficient precision, while numerical quadratures quickly become unfeasible for integrals in more than a few dimensions due to the curse of dimensionality. One can then resort to Monte Carlo integration. The main idea is very similar to the idea of sample discarding described above. Suppose we need to evaluate Z = f(x)d D x, (8) D where D is a D-dimensional domain, and f is the integrand. Monte Carlo integration consists of drawing T random numbers x i either directly from domain D if possible, or some larger domain D such that D D, if the larger domain is easier to sample from (in this case, define f(x) = f(x) if x D and f(x) = 0 if not; in other words, f is zero outside of the domain of integration, D). The integral can then be approximated as follows: Z V T f(x i ), (9) T i=1 i.e., as an average of the integrand over the domain over the sample points. V is the volume of the domain D. If the domain from which random numbers are drawn is tight, the relative error of such an approximation to the true integral should decrease as T 1/2, regardless of the dimensionality of the domain. 5 Stochastic Simulation Algorithm (SSA) Proposed by Gillespie in a series of papers (e.g., Ref [1]), this is a way of exactly simulating stochastic chemical reaction processes taking place in well-mixed solution. 5.1 Introduction and example Let s start with an example. Suppose we have a chemical system consisting of four reactions (R 1,..., R 4 ) among three different chemical species (X, Y, Z): R 1 : X + Y Z 4

5 R 2 : R 3 : Z Y + 2Z X R 4 : Y (10) The usual way to approach the mathematical modeling of this problem would be to specify the initial concentrations of the three chemical species, c X, c Y, c Z at time t = 0, specify the reaction rates for each of the four reactions, k 1,..., k 4 and write down chemical reaction kinetics, a set of ODE (also known as mass-action equations): dc X dt dc Y dt dc Z dt = k 1 c X c Y + k 3 c Y c 2 Z = k 1 c X c Y k 3 c Y c 2 Z + k 4 = k 1 c X c Y k 2 c Z 2k 3 c Y c 2 Z. (11) We could then go on and, say, numerically integrate this system of equations starting with the initial condition. Instead of doing that, let us rather reexamine the assumptions behind chemical kinetics and pose the following questions: The actual numbers of molecules in a finite volume are integers (e.g., there is x molecules of type X in a reaction vessel of volume V ), while the Eqs. (11) model them effectively as concentrations, c X = x/v. How reasonable is this approximation? Even in a well-mixed system, a molecular encounter is stochastic, as are then the reaction times; this means that the evolution of a system is random. Eqs. (11), on the other hand, are deterministic. How do we properly describe the stochasticity? To address both questions, we can move to a probabilistic description of the time-evolution of the reaction system, Eqs. (10), by defining a probability distribution of observing x molecules of type X, y molecules of type Y and z molecules of type Z at time t, written as P (x, y, z t), given the initial condition, P (x, y, z t = 0) = δ(x, x 0 )δ(y, y 0 )δ(z, z 0 ), where δ(a, b) = 1 if a = b and 0 otherwise. The temporal evolution of P is given by the Master equation, which is a linear equation in P : d dt P ( t) = ˆLP ( t), (12) where ˆL is a linear operator that we can construct from the reaction scheme, Eqs. (10), as we will show next. But let s first pause and think about the Master equation and P. First, in theory, P is a vector of infinite dimensionality, because the molecule counts can take on any nonnegative integer value. Consequently, ˆL is of dimension of that squared, since it prescribes the transitions from some values of molecular counts to any other value. In practice (but not always, e.g., in exponentially growing populations) the numbers will be bounded and the probability will sharply decrease for very high counts, suggesting that P (and thus ˆL) can be truncated into a finite dimensional vector (or matrix). However, already a few interacting chemical species, even those truncated vectors can be of huge dimensionality, preventing us from even writing them down. On the other hand, the Master equation contains the full information about chemical reactions in a well-mixed system. For example, the information of all moments is there, obtainable 5

6 by simply marginalizing over the distribution: c X (t) = x(t) V = y,z xp (x, y, z t) (13) (although you may look at some caveats of equating the marginals of P with the deterministic mass-action kinetics of Eqs. (11) in Gillespie s original paper). Of course, the Master equation also contains all information about correlations etc. as well. How does one build the operator ˆL from the reaction system? Consider the reaction X +Y Z. This reaction consumes one X and one Y and produces one Z. The two terms that correspond to it (and every chemical reaction) in the Master equation are: dp (x, y, z t) = c 1 (x + 1)(y + 1)P (x + 1, y + 1, z 1 t) c 1 xyp (x, y, z t) +. (14) dt This is because there are two ways in which the probability of having (x, y, z) molecules can change (the term on the right): this probability will increase if the reaction takes place from the state (x + 1, y + 1, z 1), which will happen with the rate that is proportional to the number of reactants in that state and the molecular reaction rate c 1 ; or, this probability will decrease because the reaction takes place from the state (x, y, z). In this stochastic interpretation, c µ for each reaction µ are the so-called molecular or stochastic reaction rates: probabilities per unit time that a particular molecular configuration for reaction µ will react. So, for our four reactions, the probabilities per unit time a µ for each to take place in the state (x, y, z) are: R 1 : a 1 = c 1 xy (15) R 2 : a 2 = c 2 z (16) R 3 : a 3 = z(z 1) c 3 y 2 (17) R 4 : a 4 = c 4. (18) Note the binomial factor for R 3 that counts in how many ways can I choose its reactants: one molecule of Y and two (identical) molecules of Z. This becomes important at small number of molecules: for example, with only one Z molecule available, that reaction cannot take place, whereas the ODE system in Eqs. (11) would still assign a non-zero rate to such a reaction. These factors of order ±1, as well as combinatorial factors (e.g., of 2 in equation 3), are different between the stochastic reaction rates, c µ, and the mass-action constants, k µ in Eqs. (11). For further details see Gillespie s original paper and a classic reference on stochastic reaction kinetics [2]. 5.2 SSA Algorithm The logic behind the algorithm proposed by Gillespie, is general and applicable to the simulation of continuous-time discrete space Markov processes. For example, in the homework, we will use it to simulate a predator-prey system. Consider a vessel with various chemical species, S i, i = 1,..., N. There are X i (t = 0) molecules of every species at t = 0, where X i are non-negative whole numbers. The reaction system is specified by listing possible reactions R µ, µ = 1,..., M, as above, each prescribing (i) the reactants (i.e. a list of chemical species and integer quantities of how much of each is needed); (ii) the products of the reaction, again a subset of S and the corresponding integer quantities; (iii) the intrinsic reaction rate c µ. We want to simulate exactly various possible time-courses of X i (t). Note that the time-courses are stochastic, because the 6

7 exact times and ordering of reactions are random, due to randomness associated with diffusion that can bring two (or more) reactants in close enough proximity that the reaction can take place. Thus, starting from the same initial condition, and using the same reaction schemes, the exact integer trajectories will be different across many runs of the SSA. Going back to the stochastic simulation, a naive way to simulate this process would be as follows: divide the time in very small intervals t. In every interval (if it is small enough, which means that the product of reaction rates and t is always much less than 1), the probability that a reaction happens is, for each reaction, given by the product of the rate, the reactant concentrations, and t. Thus, at every time step, you can flip a die using RNG, and determine if reaction µ should occur: if so, you update the table of chemical species numbers, and repeat. This, however, can be very inefficient, since most of the time no reaction will be drawn. SSA does away with this, by making a clever argument that the time to any next reaction is exponentially distributed. So instead of chunking time into small bins, I draw one exponential random variable (time-to-next reaction), another random variable that tells me which reaction I should carry out, and then I can execute the reaction and update the table of chemical species numbers. More formally, SSA defines a reaction probability density function, P (τ, µ)dτ as the probability that, given state X (a vector of chemical species counts) at time t, the next reaction µ will happen in time interval [t + τ, t + τ + dτ]. Gillespie then shows that one can write P (τ, µ)dτ = P 0 (τ)a µ dτ, where P 0 (τ) is the probability that no reaction took place in time interval τ and a µ dτ is the probability that that next reaction is µ. One can show that the distribution P 0 for τ is exponential, P 0 (τ) exp( a µ τ), (19) µ by verifying that it satisfies the recursion for P 0 (τ) which is easy to write down (can you try? the probability for no reaction until t + dt is the probability for no reaction until t times the probability that no reaction takes place in dt, which is just 1 the probability that any reaction takes place in dt, which is µ a µdτ). Note that a µ depend on the state of the system, X! This leads to a very simple simulation scheme: At t = 0, set the system state to the initial condition, X(t = 0), and input stochastic reaction rates, c µ for all reactions µ = 1,..., M, and the termination time T. Compute a µ from c µ and the state X, compute also µ a µ. Draw two URNs and use them to generate exponentially distributed time-to-next-reaction, τ (with parameter µ a µ), and reaction index µ, from P (τ, µ). Update the state, X(t), into X(t+τ) by executing reaction µ, and advance time t t+τ. Repeat (2) and terminate if t T. Pros and cons of SSA: + Exact for well-mixed systems in thermal (not not necessarily chemical) equilibrium (if both equilibria hold, we can sample stationary statistics of chemical reactions). No need for the t 0 limit in the numerical simulation. + Gives 1 stochastic trace per run, to estimate low-order moments (e.g. mean concentration as a function of time) not that many traces usually need to be simulated. 7

8 + It can simulate rare events correctly. + Small memory requirement. Lots of computation time in stiff systems, where a subset of reactions is much faster than other reactions (e.g. dimerization/monomerization reactions in the cell). One is tempted to use SSA only for slow reactions and do mass-action kinetics for fast ones, but this requires a lot of care to do correctly. Produces a lot of simulation data (and takes lots of time) if reaction systems have many species / reactions; used for small systems usually. If space matters, e.g., if the system is not well-mixed as when the diffusion of reactants is slow, completely new effects can emerge and SSA should not be used. Alternative schemes that take space into account are clearly much slower. 6 Study literature Wikipedia on Lotka-Volterra: equation. The paper introducing Gillespie SSA algorithm in Ref [1]. Numerical Recipes in C, chapters 7.2 (random numbers by transformation method), 7.3 (by rejection method), 7.6 (Monte Carlo integration) A simple introduction to Euler method: method. 7 Homework 1. Drawing random numbers distributed according the exponential distribution. The uniform RNG generates random numbers on [0, 1) according to P (x) = 1. Using the inversion method and given that you have access to uniform RNG distributed numbers, how can you generate random numbers y that are distributed according to P (y) = α exp( αy) for y [0, ) and any positive parameter α? Implement this prescription, draw 10 6 exponentially distributed random numbers with α = 1, and show their properly normalized PDF on the y-logarithmic plot. Make sure you understand the Box-Muller method for generating Gaussian-distributed random numbers in Numerical Recipes Chapter Computing the center-of-mass of a ball with a cylindrical hole. Consider a homogenous ball with radius R = 1 centered on coordinates r 0 = 0. We drill a cylindrical hole into the ball, whose center line is aligned with the z axis, of radius 0.5, and passing through a point (0.5, 0, 0). Compute the center of mass of this object using Monte Carlo sampling. Hint: you can generate 3D random numbers uniformly in a box [ 1, 1] 3, discard all points that do not fall into the object, and evaluate the center of mass, r = dv r/ dv, simply as an average of coordinates r = (x, y, z) of all the points that fall into the object. To check that you are generating the points correctly, plot the projection of the points onto the (x, y) plane. Report the center of mass coordinates. Due to symmetry, the correct answer for two center-of-mass coordinates, y, z has to be 0; you can use this exact result to 8

9 estimate the error of your Monte Carlo integral. Evaluate the Monte Carlo integral using 10, 100, 1000,..., (however much is feasible on your machine, on Matlab you should easily go up to 10 6 in a few seconds) sampling points, and for each number of sampling points, repeat the estimation 10 or more times; plot the average error in y or z as a function of the number of sampling points, on a log-log plot. How does this error depend on the number of points? If you generate random points uniformly in a sphere, and only discard those that fall into the hole, does that integration go slower or faster than if you discard samples uniformly sampled in a box? Time the two implementations and compare. 3. Deterministic Lotka-Volterra system for the dynamics of predators and prey. We are given a system of two ordinary differential equations, describing the interactions between the prey (x), and predators (y), as follows: ẋ = λ x δ xy (20) ẏ = ɛ y + β xy (21) Here, the prey is intrinsically multiplying at a rate λ (a first-order reaction ), but gets predated on in a second-order reaction with coefficient δ by predators y. Similarly, the predators intrinsically decay with rate ɛ, but grow in numbers if they catch the prey (rate β). This well-known system has two fixed points (points where the time derivatives are 0). One is trivially (x = 0, y = 0); find the second fixed point (x 0, y 0 ). It appears as if the system has 4 free parameters (λ, δ, ɛ, β), but one can reduce this to a one-parameter system, by a proper choice of units. Show that if you measure the predators in units of y 0, prey in units of x 0, and time in units of t 0 = (λɛ) 1/2, the system reduces to: ωẋ = x xy (22) ω 1 ẏ = y + xy (23) Remember, in this system, x and y are measured in new units of x 0 and y 0, rather then in absolute numbers. What is the parameter ω in terms of the original parameters of the problem? Plot a phase portrait for ω = 1. A phase portrait is a plane (x, y), where at every point of the plane you draw a little vector pointing into the direction of (ẋ, ẏ). What do you think the dynamics will be like? Choose an initial point (t = 0, x = 0.5, y = 0.5) and draw a trajectory until T = 10 on the (x, y) plane; separately also show x(t) and y(t). You can compute the trajectories either by doing a very simple numerical Euler integration (where you change the differential Eqs (23) into difference equations for a very small t 1, and step forward into the future; see Study literature), or, if you know how to, more efficient schemes (like Runge-Kutta). You can also use solvers in Matlab / Mathematica, if you know how. 4. Stochastic simulation of the LV system. In reality, the numbers of predators and prey are not continuous variables but discrete quantities, and due to random nature of encounters between predators and prey it can happen that the prey or predators die out. In the first case the system will settle to the (0, 0) attractor, in the latter prey starts to multiply exponentially. Indeed, the LV system can be viewed as a reaction network with the following reactions, assuming δ = β: X 2X (24) 9

10 Y (25) X + Y 2Y (26) (Note that X 2X stands for one individual of species X becoming two, not a doubling of the whole population.) This is just one way of representing the same deterministic dynamics (think, for instance, about what have we assumed here regarding how the prey multiply). Implement Gillespie s SSA algorithm for this reaction network. Since the fluctuations matter now (as the system can go extinct), and fluctuations depend on all rates, the behavior of this system is no longer determined only by a single effective parameter ω, as before. Select a few sets of parameters for which ω = 1 in the deterministic regime, and run the simulation many times, with the same initial condition (that corresponds to the deterministic (0.5, 0.5), but different random seed, and plot the histogram of times until predators go extinct. Plot sample trajectories in the (x, y) plane, in dimensionless units introduced above, for different parameter sets. Which parameter sets make the average time to extinction longer (and the system thus more similar to the deterministic one that never goes extinct)? References [1] Gillespie D (1977) Exact stochastic simulation of coupled chemical reactions. J Phys Chem 81: [2] van Kampen NG (2011) Stochastic processes in physics and chemistry (North-Holland). 10

Advanced Monte Carlo Methods Problems

Advanced Monte Carlo Methods Problems Advanced Monte Carlo Methods Problems September-November, 2012 Contents 1 Integration with the Monte Carlo method 2 1.1 Non-uniform random numbers.......................... 2 1.2 Gaussian RNG..................................

More information

Markov Processes. Stochastic process. Markov process

Markov Processes. Stochastic process. Markov process Markov Processes Stochastic process movement through a series of well-defined states in a way that involves some element of randomness for our purposes, states are microstates in the governing ensemble

More information

Efficient Leaping Methods for Stochastic Chemical Systems

Efficient Leaping Methods for Stochastic Chemical Systems Efficient Leaping Methods for Stochastic Chemical Systems Ioana Cipcigan Muruhan Rathinam November 18, 28 Abstract. Well stirred chemical reaction systems which involve small numbers of molecules for some

More information

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6

MTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6 MTH739U/P: Topics in Scientific Computing Autumn 16 Week 6 4.5 Generic algorithms for non-uniform variates We have seen that sampling from a uniform distribution in [, 1] is a relatively straightforward

More information

Physics 200 Lecture 4. Integration. Lecture 4. Physics 200 Laboratory

Physics 200 Lecture 4. Integration. Lecture 4. Physics 200 Laboratory Physics 2 Lecture 4 Integration Lecture 4 Physics 2 Laboratory Monday, February 21st, 211 Integration is the flip-side of differentiation in fact, it is often possible to write a differential equation

More information

19 : Slice Sampling and HMC

19 : Slice Sampling and HMC 10-708: Probabilistic Graphical Models 10-708, Spring 2018 19 : Slice Sampling and HMC Lecturer: Kayhan Batmanghelich Scribes: Boxiang Lyu 1 MCMC (Auxiliary Variables Methods) In inference, we are often

More information

Kinetic Theory 1 / Probabilities

Kinetic Theory 1 / Probabilities Kinetic Theory 1 / Probabilities 1. Motivations: statistical mechanics and fluctuations 2. Probabilities 3. Central limit theorem 1 Reading check Main concept introduced in first half of this chapter A)Temperature

More information

1. Introductory Examples

1. Introductory Examples 1. Introductory Examples We introduce the concept of the deterministic and stochastic simulation methods. Two problems are provided to explain the methods: the percolation problem, providing an example

More information

Stochastic Chemical Kinetics

Stochastic Chemical Kinetics Stochastic Chemical Kinetics Joseph K Scott November 10, 2011 1 Introduction to Stochastic Chemical Kinetics Consider the reaction I + I D The conventional kinetic model for the concentration of I in a

More information

Stochastic Simulation.

Stochastic Simulation. Stochastic Simulation. (and Gillespie s algorithm) Alberto Policriti Dipartimento di Matematica e Informatica Istituto di Genomica Applicata A. Policriti Stochastic Simulation 1/20 Quote of the day D.T.

More information

STOCHASTIC CHEMICAL KINETICS

STOCHASTIC CHEMICAL KINETICS STOCHASTIC CHEICAL KINETICS Dan Gillespie GillespieDT@mailaps.org Current Support: Caltech (NIGS) Caltech (NIH) University of California at Santa Barbara (NIH) Past Support: Caltech (DARPA/AFOSR, Beckman/BNC))

More information

Simulation. Where real stuff starts

Simulation. Where real stuff starts 1 Simulation Where real stuff starts ToC 1. What is a simulation? 2. Accuracy of output 3. Random Number Generators 4. How to sample 5. Monte Carlo 6. Bootstrap 2 1. What is a simulation? 3 What is a simulation?

More information

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006

Data Analysis I. Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK. 10 lectures, beginning October 2006 Astronomical p( y x, I) p( x, I) p ( x y, I) = p( y, I) Data Analysis I Dr Martin Hendry, Dept of Physics and Astronomy University of Glasgow, UK 10 lectures, beginning October 2006 4. Monte Carlo Methods

More information

Lecture 20/Lab 21: Systems of Nonlinear ODEs

Lecture 20/Lab 21: Systems of Nonlinear ODEs Lecture 20/Lab 21: Systems of Nonlinear ODEs MAR514 Geoffrey Cowles Department of Fisheries Oceanography School for Marine Science and Technology University of Massachusetts-Dartmouth Coupled ODEs: Species

More information

Random Number Generation. Stephen Booth David Henty

Random Number Generation. Stephen Booth David Henty Random Number Generation Stephen Booth David Henty Introduction Random numbers are frequently used in many types of computer simulation Frequently as part of a sampling process: Generate a representative

More information

Random processes and probability distributions. Phys 420/580 Lecture 20

Random processes and probability distributions. Phys 420/580 Lecture 20 Random processes and probability distributions Phys 420/580 Lecture 20 Random processes Many physical processes are random in character: e.g., nuclear decay (Poisson distributed event count) P (k, τ) =

More information

Random variables. DS GA 1002 Probability and Statistics for Data Science.

Random variables. DS GA 1002 Probability and Statistics for Data Science. Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities

More information

Kinetic Theory 1 / Probabilities

Kinetic Theory 1 / Probabilities Kinetic Theory 1 / Probabilities 1. Motivations: statistical mechanics and fluctuations 2. Probabilities 3. Central limit theorem 1 The need for statistical mechanics 2 How to describe large systems In

More information

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm

Markov Chain Monte Carlo The Metropolis-Hastings Algorithm Markov Chain Monte Carlo The Metropolis-Hastings Algorithm Anthony Trubiano April 11th, 2018 1 Introduction Markov Chain Monte Carlo (MCMC) methods are a class of algorithms for sampling from a probability

More information

2 Random Variable Generation

2 Random Variable Generation 2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods

More information

Graded Project #1. Part 1. Explicit Runge Kutta methods. Goals Differential Equations FMN130 Gustaf Söderlind and Carmen Arévalo

Graded Project #1. Part 1. Explicit Runge Kutta methods. Goals Differential Equations FMN130 Gustaf Söderlind and Carmen Arévalo 2008-11-07 Graded Project #1 Differential Equations FMN130 Gustaf Söderlind and Carmen Arévalo This homework is due to be handed in on Wednesday 12 November 2008 before 13:00 in the post box of the numerical

More information

Partial Differential Equations (PDEs)

Partial Differential Equations (PDEs) 10.34 Numerical Methods Applied to Chemical Engineering Fall 2015 Final Exam Review Partial Differential Equations (PDEs) 1. Classification (a) Many PDEs encountered by chemical engineers are second order

More information

16 : Markov Chain Monte Carlo (MCMC)

16 : Markov Chain Monte Carlo (MCMC) 10-708: Probabilistic Graphical Models 10-708, Spring 2014 16 : Markov Chain Monte Carlo MCMC Lecturer: Matthew Gormley Scribes: Yining Wang, Renato Negrinho 1 Sampling from low-dimensional distributions

More information

Monte Carlo and cold gases. Lode Pollet.

Monte Carlo and cold gases. Lode Pollet. Monte Carlo and cold gases Lode Pollet lpollet@physics.harvard.edu 1 Outline Classical Monte Carlo The Monte Carlo trick Markov chains Metropolis algorithm Ising model critical slowing down Quantum Monte

More information

Solution. This is a routine application of the chain rule.

Solution. This is a routine application of the chain rule. EXAM 2 SOLUTIONS 1. If z = e r cos θ, r = st, θ = s 2 + t 2, find the partial derivatives dz ds chain rule. Write your answers entirely in terms of s and t. dz and dt using the Solution. This is a routine

More information

Chapter 4: Monte Carlo Methods. Paisan Nakmahachalasint

Chapter 4: Monte Carlo Methods. Paisan Nakmahachalasint Chapter 4: Monte Carlo Methods Paisan Nakmahachalasint Introduction Monte Carlo Methods are a class of computational algorithms that rely on repeated random sampling to compute their results. Monte Carlo

More information

Masters Comprehensive Examination Department of Statistics, University of Florida

Masters Comprehensive Examination Department of Statistics, University of Florida Masters Comprehensive Examination Department of Statistics, University of Florida May 6, 003, 8:00 am - :00 noon Instructions: You have four hours to answer questions in this examination You must show

More information

Multivariate distributions

Multivariate distributions CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density

More information

In this chapter we study elliptical PDEs. That is, PDEs of the form. 2 u = lots,

In this chapter we study elliptical PDEs. That is, PDEs of the form. 2 u = lots, Chapter 8 Elliptic PDEs In this chapter we study elliptical PDEs. That is, PDEs of the form 2 u = lots, where lots means lower-order terms (u x, u y,..., u, f). Here are some ways to think about the physical

More information

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.

Perhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows. Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage

More information

Physics 342 Lecture 4. Probability Density. Lecture 4. Physics 342 Quantum Mechanics I

Physics 342 Lecture 4. Probability Density. Lecture 4. Physics 342 Quantum Mechanics I Physics 342 Lecture 4 Probability Density Lecture 4 Physics 342 Quantum Mechanics I Monday, February 4th, 28 We review the basic notions of probability, in particular the role of probability density in

More information

Simulation. Where real stuff starts

Simulation. Where real stuff starts Simulation Where real stuff starts March 2019 1 ToC 1. What is a simulation? 2. Accuracy of output 3. Random Number Generators 4. How to sample 5. Monte Carlo 6. Bootstrap 2 1. What is a simulation? 3

More information

Probability theory basics

Probability theory basics Probability theory basics Michael Franke Basics of probability theory: axiomatic definition, interpretation, joint distributions, marginalization, conditional probability & Bayes rule. Random variables:

More information

Numerical Methods I Monte Carlo Methods

Numerical Methods I Monte Carlo Methods Numerical Methods I Monte Carlo Methods Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course G63.2010.001 / G22.2420-001, Fall 2010 Dec. 9th, 2010 A. Donev (Courant Institute) Lecture

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x

More information

1 Introduction. Systems 2: Simulating Errors. Mobile Robot Systems. System Under. Environment

1 Introduction. Systems 2: Simulating Errors. Mobile Robot Systems. System Under. Environment Systems 2: Simulating Errors Introduction Simulating errors is a great way to test you calibration algorithms, your real-time identification algorithms, and your estimation algorithms. Conceptually, the

More information

2008 Winton. Review of Statistical Terminology

2008 Winton. Review of Statistical Terminology 1 Review of Statistical Terminology 2 Formal Terminology An experiment is a process whose outcome is not known with certainty The experiment s sample space S is the set of all possible outcomes. A random

More information

Monte Carlo Simulations

Monte Carlo Simulations Monte Carlo Simulations What are Monte Carlo Simulations and why ones them? Pseudo Random Number generators Creating a realization of a general PDF The Bootstrap approach A real life example: LOFAR simulations

More information

(x 3)(x + 5) = (x 3)(x 1) = x + 5. sin 2 x e ax bx 1 = 1 2. lim

(x 3)(x + 5) = (x 3)(x 1) = x + 5. sin 2 x e ax bx 1 = 1 2. lim SMT Calculus Test Solutions February, x + x 5 Compute x x x + Answer: Solution: Note that x + x 5 x x + x )x + 5) = x )x ) = x + 5 x x + 5 Then x x = + 5 = Compute all real values of b such that, for fx)

More information

So we will instead use the Jacobian method for inferring the PDF of functionally related random variables; see Bertsekas & Tsitsiklis Sec. 4.1.

So we will instead use the Jacobian method for inferring the PDF of functionally related random variables; see Bertsekas & Tsitsiklis Sec. 4.1. 2011 Page 1 Simulating Gaussian Random Variables Monday, February 14, 2011 2:00 PM Readings: Kloeden and Platen Sec. 1.3 Why does the Box Muller method work? How was it derived? The basic idea involves

More information

Physics 403 Probability Distributions II: More Properties of PDFs and PMFs

Physics 403 Probability Distributions II: More Properties of PDFs and PMFs Physics 403 Probability Distributions II: More Properties of PDFs and PMFs Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Last Time: Common Probability Distributions

More information

Eco517 Fall 2004 C. Sims MIDTERM EXAM

Eco517 Fall 2004 C. Sims MIDTERM EXAM Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering

More information

Stat 516, Homework 1

Stat 516, Homework 1 Stat 516, Homework 1 Due date: October 7 1. Consider an urn with n distinct balls numbered 1,..., n. We sample balls from the urn with replacement. Let N be the number of draws until we encounter a ball

More information

Problem 1: Lagrangians and Conserved Quantities. Consider the following action for a particle of mass m moving in one dimension

Problem 1: Lagrangians and Conserved Quantities. Consider the following action for a particle of mass m moving in one dimension 105A Practice Final Solutions March 13, 01 William Kelly Problem 1: Lagrangians and Conserved Quantities Consider the following action for a particle of mass m moving in one dimension S = dtl = mc dt 1

More information

This is a Gaussian probability centered around m = 0 (the most probable and mean position is the origin) and the mean square displacement m 2 = n,or

This is a Gaussian probability centered around m = 0 (the most probable and mean position is the origin) and the mean square displacement m 2 = n,or Physics 7b: Statistical Mechanics Brownian Motion Brownian motion is the motion of a particle due to the buffeting by the molecules in a gas or liquid. The particle must be small enough that the effects

More information

Sampling Distributions Allen & Tildesley pgs and Numerical Recipes on random numbers

Sampling Distributions Allen & Tildesley pgs and Numerical Recipes on random numbers Sampling Distributions Allen & Tildesley pgs. 345-351 and Numerical Recipes on random numbers Today I explain how to generate a non-uniform probability distributions. These are used in importance sampling

More information

Reaction time distributions in chemical kinetics: Oscillations and other weird behaviors

Reaction time distributions in chemical kinetics: Oscillations and other weird behaviors Introduction The algorithm Results Summary Reaction time distributions in chemical kinetics: Oscillations and other weird behaviors Ramon Xulvi-Brunet Escuela Politécnica Nacional Outline Introduction

More information

MATH20411 PDEs and Vector Calculus B

MATH20411 PDEs and Vector Calculus B MATH2411 PDEs and Vector Calculus B Dr Stefan Güttel Acknowledgement The lecture notes and other course materials are based on notes provided by Dr Catherine Powell. SECTION 1: Introctory Material MATH2411

More information

Introduction to First Order Equations Sections

Introduction to First Order Equations Sections A B I L E N E C H R I S T I A N U N I V E R S I T Y Department of Mathematics Introduction to First Order Equations Sections 2.1-2.3 Dr. John Ehrke Department of Mathematics Fall 2012 Course Goals The

More information

Stat 451 Lecture Notes Simulating Random Variables

Stat 451 Lecture Notes Simulating Random Variables Stat 451 Lecture Notes 05 12 Simulating Random Variables Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 22 in Lange, and Chapter 2 in Robert & Casella 2 Updated:

More information

Today: Fundamentals of Monte Carlo

Today: Fundamentals of Monte Carlo Today: Fundamentals of Monte Carlo What is Monte Carlo? Named at Los Alamos in 940 s after the casino. Any method which uses (pseudo)random numbers as an essential part of the algorithm. Stochastic - not

More information

Lecture 4 The stochastic ingredient

Lecture 4 The stochastic ingredient Lecture 4 The stochastic ingredient Luca Bortolussi 1 Alberto Policriti 2 1 Dipartimento di Matematica ed Informatica Università degli studi di Trieste Via Valerio 12/a, 34100 Trieste. luca@dmi.units.it

More information

Solutions for the Practice Final - Math 23B, 2016

Solutions for the Practice Final - Math 23B, 2016 olutions for the Practice Final - Math B, 6 a. True. The area of a surface is given by the expression d, and since we have a parametrization φ x, y x, y, f x, y with φ, this expands as d T x T y da xy

More information

Lecture 10 - Moment of Inertia

Lecture 10 - Moment of Inertia Lecture 10 - oment of Inertia A Puzzle... Question For any object, there are typically many ways to calculate the moment of inertia I = r 2 dm, usually by doing the integration by considering different

More information

Introduction to Probability and Statistics (Continued)

Introduction to Probability and Statistics (Continued) Introduction to Probability and Statistics (Continued) Prof. icholas Zabaras Center for Informatics and Computational Science https://cics.nd.edu/ University of otre Dame otre Dame, Indiana, USA Email:

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg

Statistics for Data Analysis. Niklaus Berger. PSI Practical Course Physics Institute, University of Heidelberg Statistics for Data Analysis PSI Practical Course 2014 Niklaus Berger Physics Institute, University of Heidelberg Overview You are going to perform a data analysis: Compare measured distributions to theoretical

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

7 Pendulum. Part II: More complicated situations

7 Pendulum. Part II: More complicated situations MATH 35, by T. Lakoba, University of Vermont 60 7 Pendulum. Part II: More complicated situations In this Lecture, we will pursue two main goals. First, we will take a glimpse at a method of Classical Mechanics

More information

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Cheng Soon Ong & Christian Walder. Canberra February June 2018 Cheng Soon Ong & Christian Walder Research Group and College of Engineering and Computer Science Canberra February June 2018 Outlines Overview Introduction Linear Algebra Probability Linear Regression

More information

Monte Carlo Integration I

Monte Carlo Integration I Monte Carlo Integration I Digital Image Synthesis Yung-Yu Chuang with slides by Pat Hanrahan and Torsten Moller Introduction L o p,ωo L e p,ωo s f p,ωo,ω i L p,ω i cosθ i i d The integral equations generally

More information

1 The pendulum equation

1 The pendulum equation Math 270 Honors ODE I Fall, 2008 Class notes # 5 A longer than usual homework assignment is at the end. The pendulum equation We now come to a particularly important example, the equation for an oscillating

More information

17 : Markov Chain Monte Carlo

17 : Markov Chain Monte Carlo 10-708: Probabilistic Graphical Models, Spring 2015 17 : Markov Chain Monte Carlo Lecturer: Eric P. Xing Scribes: Heran Lin, Bin Deng, Yun Huang 1 Review of Monte Carlo Methods 1.1 Overview Monte Carlo

More information

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015 Probability Refresher Kai Arras, University of Freiburg Winter term 2014/2015 Probability Refresher Introduction to Probability Random variables Joint distribution Marginalization Conditional probability

More information

General Principles in Random Variates Generation

General Principles in Random Variates Generation General Principles in Random Variates Generation E. Moulines and G. Fort Telecom ParisTech June 2015 Bibliography : Luc Devroye, Non-Uniform Random Variate Generator, Springer-Verlag (1986) available on

More information

Nonlinear dynamics & chaos BECS

Nonlinear dynamics & chaos BECS Nonlinear dynamics & chaos BECS-114.7151 Phase portraits Focus: nonlinear systems in two dimensions General form of a vector field on the phase plane: Vector notation: Phase portraits Solution x(t) describes

More information

DYNAMICAL SYSTEMS

DYNAMICAL SYSTEMS 0.42 DYNAMICAL SYSTEMS Week Lecture Notes. What is a dynamical system? Probably the best way to begin this discussion is with arguably a most general and yet least helpful statement: Definition. A dynamical

More information

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008

Gaussian processes. Chuong B. Do (updated by Honglak Lee) November 22, 2008 Gaussian processes Chuong B Do (updated by Honglak Lee) November 22, 2008 Many of the classical machine learning algorithms that we talked about during the first half of this course fit the following pattern:

More information

Ordinary Differential Equations

Ordinary Differential Equations Chapter 13 Ordinary Differential Equations We motivated the problem of interpolation in Chapter 11 by transitioning from analzying to finding functions. That is, in problems like interpolation and regression,

More information

Gillespie s Algorithm and its Approximations. Des Higham Department of Mathematics and Statistics University of Strathclyde

Gillespie s Algorithm and its Approximations. Des Higham Department of Mathematics and Statistics University of Strathclyde Gillespie s Algorithm and its Approximations Des Higham Department of Mathematics and Statistics University of Strathclyde djh@maths.strath.ac.uk The Three Lectures 1 Gillespie s algorithm and its relation

More information

Predator-Prey Population Dynamics

Predator-Prey Population Dynamics Predator-Prey Population Dynamics Gonzalo Mateos Dept. of ECE and Goergen Institute for Data Science University of Rochester gmateosb@ece.rochester.edu http://www.ece.rochester.edu/~gmateosb/ October 2,

More information

Math Stochastic Processes & Simulation. Davar Khoshnevisan University of Utah

Math Stochastic Processes & Simulation. Davar Khoshnevisan University of Utah Math 5040 1 Stochastic Processes & Simulation Davar Khoshnevisan University of Utah Module 1 Generation of Discrete Random Variables Just about every programming language and environment has a randomnumber

More information

Question: My computer only knows how to generate a uniform random variable. How do I generate others? f X (x)dx. f X (s)ds.

Question: My computer only knows how to generate a uniform random variable. How do I generate others? f X (x)dx. f X (s)ds. Simulation Question: My computer only knows how to generate a uniform random variable. How do I generate others?. Continuous Random Variables Recall that a random variable X is continuous if it has a probability

More information

CHEM-UA 127: Advanced General Chemistry I

CHEM-UA 127: Advanced General Chemistry I 1 CHEM-UA 127: Advanced General Chemistry I Notes for Lecture 11 Nowthatwehaveintroducedthebasicconceptsofquantummechanics, wecanstarttoapplythese conceptsto build up matter, starting from its most elementary

More information

ter. on Can we get a still better result? Yes, by making the rectangles still smaller. As we make the rectangles smaller and smaller, the

ter. on Can we get a still better result? Yes, by making the rectangles still smaller. As we make the rectangles smaller and smaller, the Area and Tangent Problem Calculus is motivated by two main problems. The first is the area problem. It is a well known result that the area of a rectangle with length l and width w is given by A = wl.

More information

Statistics for scientists and engineers

Statistics for scientists and engineers Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3

More information

From Random Numbers to Monte Carlo. Random Numbers, Random Walks, Diffusion, Monte Carlo integration, and all that

From Random Numbers to Monte Carlo. Random Numbers, Random Walks, Diffusion, Monte Carlo integration, and all that From Random Numbers to Monte Carlo Random Numbers, Random Walks, Diffusion, Monte Carlo integration, and all that Random Walk Through Life Random Walk Through Life If you flip the coin 5 times you will

More information

Numerical integration and importance sampling

Numerical integration and importance sampling 2 Numerical integration and importance sampling 2.1 Quadrature Consider the numerical evaluation of the integral I(a, b) = b a dx f(x) Rectangle rule: on small interval, construct interpolating function

More information

Robot Control Basics CS 685

Robot Control Basics CS 685 Robot Control Basics CS 685 Control basics Use some concepts from control theory to understand and learn how to control robots Control Theory general field studies control and understanding of behavior

More information

Sampling Distributions Allen & Tildesley pgs and Numerical Recipes on random numbers

Sampling Distributions Allen & Tildesley pgs and Numerical Recipes on random numbers Sampling Distributions Allen & Tildesley pgs. 345-351 and Numerical Recipes on random numbers Today I explain how to generate a non-uniform probability distributions. These are used in importance sampling

More information

CS 365 Introduction to Scientific Modeling Fall Semester, 2011 Review

CS 365 Introduction to Scientific Modeling Fall Semester, 2011 Review CS 365 Introduction to Scientific Modeling Fall Semester, 2011 Review Topics" What is a model?" Styles of modeling" How do we evaluate models?" Aggregate models vs. individual models." Cellular automata"

More information

Fast Probability Generating Function Method for Stochastic Chemical Reaction Networks

Fast Probability Generating Function Method for Stochastic Chemical Reaction Networks MATCH Communications in Mathematical and in Computer Chemistry MATCH Commun. Math. Comput. Chem. 71 (2014) 57-69 ISSN 0340-6253 Fast Probability Generating Function Method for Stochastic Chemical Reaction

More information

Statistics 100A Homework 1 Solutions

Statistics 100A Homework 1 Solutions Problem Statistics 00A Homework Solutions Ryan Rosario Suppose we flip a fair coin 4 times independently. () What is the sample space? By definition, the sample space, denoted as Ω, is the set of all possible

More information

16. Working with the Langevin and Fokker-Planck equations

16. Working with the Langevin and Fokker-Planck equations 16. Working with the Langevin and Fokker-Planck equations In the preceding Lecture, we have shown that given a Langevin equation (LE), it is possible to write down an equivalent Fokker-Planck equation

More information

Monte Carlo. Lecture 15 4/9/18. Harvard SEAS AP 275 Atomistic Modeling of Materials Boris Kozinsky

Monte Carlo. Lecture 15 4/9/18. Harvard SEAS AP 275 Atomistic Modeling of Materials Boris Kozinsky Monte Carlo Lecture 15 4/9/18 1 Sampling with dynamics In Molecular Dynamics we simulate evolution of a system over time according to Newton s equations, conserving energy Averages (thermodynamic properties)

More information

Computational statistics

Computational statistics Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated

More information

1.2. Direction Fields: Graphical Representation of the ODE and its Solution Let us consider a first order differential equation of the form dy

1.2. Direction Fields: Graphical Representation of the ODE and its Solution Let us consider a first order differential equation of the form dy .. Direction Fields: Graphical Representation of the ODE and its Solution Let us consider a first order differential equation of the form dy = f(x, y). In this section we aim to understand the solution

More information

functions Poisson distribution Normal distribution Arbitrary functions

functions Poisson distribution Normal distribution Arbitrary functions Physics 433: Computational Physics Lecture 6 Random number distributions Generation of random numbers of various distribuition functions Normal distribution Poisson distribution Arbitrary functions Random

More information

Modelling in Systems Biology

Modelling in Systems Biology Modelling in Systems Biology Maria Grazia Vigliotti thanks to my students Anton Stefanek, Ahmed Guecioueur Imperial College Formal representation of chemical reactions precise qualitative and quantitative

More information

Metric spaces and metrizability

Metric spaces and metrizability 1 Motivation Metric spaces and metrizability By this point in the course, this section should not need much in the way of motivation. From the very beginning, we have talked about R n usual and how relatively

More information

Workshop on Heterogeneous Computing, 16-20, July No Monte Carlo is safe Monte Carlo - more so parallel Monte Carlo

Workshop on Heterogeneous Computing, 16-20, July No Monte Carlo is safe Monte Carlo - more so parallel Monte Carlo Workshop on Heterogeneous Computing, 16-20, July 2012 No Monte Carlo is safe Monte Carlo - more so parallel Monte Carlo K. P. N. Murthy School of Physics, University of Hyderabad July 19, 2012 K P N Murthy

More information

Simulation methods for stochastic models in chemistry

Simulation methods for stochastic models in chemistry Simulation methods for stochastic models in chemistry David F. Anderson anderson@math.wisc.edu Department of Mathematics University of Wisconsin - Madison SIAM: Barcelona June 4th, 21 Overview 1. Notation

More information

STABILITY. Phase portraits and local stability

STABILITY. Phase portraits and local stability MAS271 Methods for differential equations Dr. R. Jain STABILITY Phase portraits and local stability We are interested in system of ordinary differential equations of the form ẋ = f(x, y), ẏ = g(x, y),

More information

Langevin Methods. Burkhard Dünweg Max Planck Institute for Polymer Research Ackermannweg 10 D Mainz Germany

Langevin Methods. Burkhard Dünweg Max Planck Institute for Polymer Research Ackermannweg 10 D Mainz Germany Langevin Methods Burkhard Dünweg Max Planck Institute for Polymer Research Ackermannweg 1 D 55128 Mainz Germany Motivation Original idea: Fast and slow degrees of freedom Example: Brownian motion Replace

More information

SAMPLING SPECIAL DISTRIBUTIONS

SAMPLING SPECIAL DISTRIBUTIONS SAMPLING SPECIAL DISTRIBUTIONS M. Ragheb 4/9/4 INTRODUCTION Monte Carlo simulations require the sampling of various distributions at different steps in the simulation process. There exist a multitude of

More information

Data Analysis and Monte Carlo Methods

Data Analysis and Monte Carlo Methods Lecturer: Allen Caldwell, Max Planck Institute for Physics & TUM Recitation Instructor: Oleksander (Alex) Volynets, MPP & TUM General Information: - Lectures will be held in English, Mondays 16-18:00 -

More information

Review of Statistical Terminology

Review of Statistical Terminology Review of Statistical Terminology An experiment is a process whose outcome is not known with certainty. The experiment s sample space S is the set of all possible outcomes. A random variable is a function

More information

Solutions to Sample Questions for Final Exam

Solutions to Sample Questions for Final Exam olutions to ample Questions for Final Exam Find the points on the surface xy z 3 that are closest to the origin. We use the method of Lagrange Multipliers, with f(x, y, z) x + y + z for the square of the

More information

5. Advection and Diffusion of an Instantaneous, Point Source

5. Advection and Diffusion of an Instantaneous, Point Source 1 5. Advection and Diffusion of an Instantaneous, Point Source In this chapter consider the combined transport by advection and diffusion for an instantaneous point release. We neglect source and sink

More information

Electro Magnetic Field Dr. Harishankar Ramachandran Department of Electrical Engineering Indian Institute of Technology Madras

Electro Magnetic Field Dr. Harishankar Ramachandran Department of Electrical Engineering Indian Institute of Technology Madras Electro Magnetic Field Dr. Harishankar Ramachandran Department of Electrical Engineering Indian Institute of Technology Madras Lecture - 7 Gauss s Law Good morning. Today, I want to discuss two or three

More information