In manycomputationaleconomicapplications, one must compute thede nite n

Similar documents
Integration, differentiation, and root finding. Phys 420/580 Lecture 7

CS 450 Numerical Analysis. Chapter 8: Numerical Integration and Differentiation

Differentiation and Integration

Nonlinear equations can take one of two forms. In the nonlinear root nding n n

Integration. Topic: Trapezoidal Rule. Major: General Engineering. Author: Autar Kaw, Charlie Barker.

Monte Carlo Methods for Statistical Inference: Variance Reduction Techniques

Chapter 1: A Brief Review of Maximum Likelihood, GMM, and Numerical Tools. Joan Llull. Microeconometrics IDEA PhD Program

Section 6.6 Gaussian Quadrature

5 Numerical Integration & Dierentiation

Economics 241B Review of Limit Theorems for Sequences of Random Variables

IDL Advanced Math & Stats Module

GENG2140, S2, 2012 Week 7: Curve fitting

Mathematics for Engineers. Numerical mathematics

8.3 Numerical Quadrature, Continued

Chapter 5 continued. Chapter 5 sections

3. Numerical Quadrature. Where analytical abilities end...

Basic Sampling Methods

Lectures 9-10: Polynomial and piecewise polynomial interpolation

8 STOCHASTIC SIMULATION

Bindel, Spring 2012 Intro to Scientific Computing (CS 3220) Week 12: Monday, Apr 16. f(x) dx,

Numerical Methods I: Numerical Integration/Quadrature

Time Series Models and Inference. James L. Powell Department of Economics University of California, Berkeley

Ch. 03 Numerical Quadrature. Andrea Mignone Physics Department, University of Torino AA

Stat 451 Lecture Notes Numerical Integration

In numerical analysis quadrature refers to the computation of definite integrals.

NUMERICAL METHODS. x n+1 = 2x n x 2 n. In particular: which of them gives faster convergence, and why? [Work to four decimal places.

Hochdimensionale Integration

Today: Fundamentals of Monte Carlo

Parametric Inference on Strong Dependence

Nonuniform Random Variate Generation

Numerical integration and importance sampling

Lecture 5: Random numbers and Monte Carlo (Numerical Recipes, Chapter 7) Motivations for generating random numbers

Numerical Integration (Quadrature) Another application for our interpolation tools!

Chapter 4: Interpolation and Approximation. October 28, 2005

Multivariate Distributions

Introduction to Smoothing spline ANOVA models (metamodelling)

Hands-on Generating Random

Scientific Computing: Numerical Integration

you expect to encounter difficulties when trying to solve A x = b? 4. A composite quadrature rule has error associated with it in the following form

Numerical integration and differentiation. Unit IV. Numerical Integration and Differentiation. Plan of attack. Numerical integration.

Numerical Methods with MATLAB

Numerical Integration

Chapter 4. Continuous Random Variables 4.1 PDF

4. Numerical Quadrature. Where analytical abilities end... continued

Numerical integration - I. M. Peressi - UniTS - Laurea Magistrale in Physics Laboratory of Computational Physics - Unit V

Today: Fundamentals of Monte Carlo

Multiple Random Variables

Lecture 2: Repetition of probability theory and statistics

Homework and Computer Problems for Math*2130 (W17).

Preliminary statistics

PROJECTION METHODS FOR DYNAMIC MODELS

Bayesian Methods with Monte Carlo Markov Chains II

Information geometry for bivariate distribution control

Jim Lambers MAT 460/560 Fall Semester Practice Final Exam

Preliminary Examination in Numerical Analysis

CAM Ph.D. Qualifying Exam in Numerical Analysis CONTENTS

PowerPoints organized by Dr. Michael R. Gustafson II, Duke University

Lecture Notes Part 2: Matrix Algebra

Numerical Analysis for Statisticians

Irr. Statistical Methods in Experimental Physics. 2nd Edition. Frederick James. World Scientific. CERN, Switzerland

MA2501 Numerical Methods Spring 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

MA6452 STATISTICS AND NUMERICAL METHODS UNIT IV NUMERICAL DIFFERENTIATION AND INTEGRATION

1.12 Multivariate Random Variables

Romberg Integration and Gaussian Quadrature

Elements of Asymptotic Theory. James L. Powell Department of Economics University of California, Berkeley

April 15 Math 2335 sec 001 Spring 2014

Scientific Computing

LECTURE NOTES ELEMENTARY NUMERICAL METHODS. Eusebius Doedel

TAKEHOME FINAL EXAM e iω e 2iω e iω e 2iω

18.01 EXERCISES. Unit 3. Integration. 3A. Differentials, indefinite integration. 3A-1 Compute the differentials df(x) of the following functions.

Ordered Sample Generation

ECE 673-Random signal analysis I Final

2tdt 1 y = t2 + C y = which implies C = 1 and the solution is y = 1

Chapter 4 Integration

Mathematical Preliminaries

Outline Introduction: Problem Description Diculties Algebraic Structure: Algebraic Varieties Rank Decient Toeplitz Matrices Constructing Lower Rank St

PATTERN RECOGNITION AND MACHINE LEARNING CHAPTER 2: PROBABILITY DISTRIBUTIONS

The Standard Market Risk Model of the Swiss Solvency Test: an Analytical Solution

Lecture 17: The Exponential and Some Related Distributions

COURSE Numerical integration of functions (continuation) 3.3. The Romberg s iterative generation method

Lecture 4: Two-point Sampling, Coupon Collector s problem

F denotes cumulative density. denotes probability density function; (.)

Dynamic System Identification using HDMR-Bayesian Technique

Engg. Math. II (Unit-IV) Numerical Analysis

Numerical Methods. King Saud University

Iterative procedure for multidimesional Euler equations Abstracts A numerical iterative scheme is suggested to solve the Euler equations in two and th

Today: Fundamentals of Monte Carlo

Monte Carlo Integration. Computer Graphics CMU /15-662, Fall 2016

MAS223 Statistical Inference and Modelling Exercises

A Polynomial Chaos Approach to Robust Multiobjective Optimization

An idea how to solve some of the problems. diverges the same must hold for the original series. T 1 p T 1 p + 1 p 1 = 1. dt = lim

Algorithms for Uncertainty Quantification

Approximating Integrals for Stochastic Problems

CSCI-6971 Lecture Notes: Monte Carlo integration

Pascal s Triangle on a Budget. Accuracy, Precision and Efficiency in Sparse Grids

Outline. 1 Numerical Integration. 2 Numerical Differentiation. 3 Richardson Extrapolation

Extrapolation in Numerical Integration. Romberg Integration

We consider the problem of finding a polynomial that interpolates a given set of values:

8.7 MacLaurin Polynomials

Transcription:

Chapter 6 Numerical Integration In manycomputationaleconomicapplications, one must compute thede nite n integral of a real-valued function f de ned on some interval I of<. Some applications call for nding the area under the function, as in computing the change in consumer welfare caused by a price change. In these instances, the integral takes the form f(x) dx: I Other applications call for computing the expectation of a function of a random variable with a continuous probability densityp. In these instances the integral takes the form f(x)p(x) dx: I We will discuss only the three numericalintegration techniques most commonly encountered in practice. Newton-Cotes quadrature techniques employ a strategy that is a straightforward generalization of Riemann integration principles. Newton-Cotes methods are easy to implement. They are not, however, the most computationally e±cient way of computing the expectation of a smooth function. In such applications, the moment-matching strategy embodied by Gaussian quadrature is often substantially superior. Finally, Monte Carlo integration methods are discussed. Such methods are simple to implement and are particularly useful for complex high-dimensional integration, but should only be used as a last resort. 1

6.1 Newton-Cotes Quadrature Consider the problem of computing the de nite integral f(x) dx [a;b] of a real-valued functionf over a bounded interval [a;b] on the real line. Newton-Cotes quadrature rules approximatethe integrand f with a piecewise polynomial functionf^ and take the integral off^ as an approximation of the integral off. Two Newton-Cotes rules are widely used in practice: the trapezoidal rule and Simpson's rule. Both rules are very easy to implement and are typically adequate for computing the area under a curve. The trapezoidal rule approximates f using a piecewise linear function. To implement the trapezoidal rule, the integration interval [a;b] is partitioned into n subintervals of equal lengthh = (b a)=n. The function is evaluated at the partition points x = a + (i=n)h, yielding the values y = f(x ) for i i i i = 0; 1;:::;n. The function f is approximated over each subinterval by linearlyinterpolating these function values. The space under each line de nes a trapezoid with an areas of the form 0:5h(y + y ). The areas of all the i i+1 trapezoids are then summed to form an estimate of the integral of f: [a;b] h f(x)¼ (y 0 + 2y 1 + 2y 2 +::: + 2y n 1 +y n): 2 Figure 6.1: Trapezoidal rule. The trapezoidal rule is not only simple but also fairly robust. Other Newton-Cotes methods will be more accurate if the integrand f is smooth. However, the trapezoidal rule will often be more accurate if the integrand exhibits discontinuities in its rst derivative, which can occur in dynamic economic applications. The trapezoidal rule is said to be rst order exact because it exactly computes the integral of any rst order polynomial, that is, a line. In general, if theintegrand is smooth, thetrapezoidalrule yields an 2 approximation error that is O(1=n ), that is, the error shrinks quadratically with the number of partitions. 2

Simpson's rule is similar to the trapezoidal rule, except that it approximates the integrand f using a piecewise quadratic, rather than a piecewise linear function. As before, the integration interval is partitioned into n subintervals of equal length h = (b a)=n, where, in this case, n is required to be even. The function is evaluated at the partition points x = a + (i=n)h, i yielding the valuesy i =f(x i) fori = 0; 1;:::;n. A series of locally quadratic approximations to the integrand are formed by interpolating the function values at successive triplets of the partition points. The area under each h quadratic piece is of the form 3 pieces are then summed to form an estimate of the integral of f: h f(x)¼ (y + 4y + 2y + 4y ::: + 2y + 4y +y ): [a;b] 3 (y + 4y +y ). The areas of all these i i+1 i+2 0 1 2 3 n 2 n 1 n Figure 6.2: Simpson's rule. Simpson's rule is almost as simple as the trapezoidal rule, and thus not much harder to program. Simpson's rule will yield more accurate approximation than the trapezoidal rule if the integrand is smooth. Even though Simpson's rule is based on locally quadratic approximation of the integrand, it is third order exact. That is, it exactly computes the integral of any third order (e.g., cubic) polynomial. In general, if the integrand is smooth, Simp- 4 son's rule yields an approximation error that is O(1=n ), and thus falls at twice the geometric rate as the error associated with the trapezoidal rule. Simpson's rule is the Newton-Cotes rule most often used in practice because it retains algorithmic simplicity while o ering an adequate degree of approximation. Newton-Cotes rules of higher order may be de ned, but are more di±cult to work with and thus are rarely used. In most computational economic applications, it is not possible to determine a priori how many partition points are needed to compute an inde nite integral to a desired level of accuracy. One solution to this problem is to use an adaptive quadrature strategy whereby one increases the number of points at which the integrand is evaluated until the sequence of estimates of the definite integral converge. E±cient adaptive Newton-Cotes quadrature schemes are specially easy to implement. One simple, but powerful, scheme calls for the number of intervals to be doubled with each iteration. Because the new 3

partition points include the partition points used in the previous iteration, the computational e ort required to form the new integral estimate is cut in half. More sophisticated adaptive Newton-Cotes quadrature techniques relax the requirement that the intervals be equally spaced and concentrate new evaluation points in those areas where the integrand appears to be most irregular. Matlab o ers two Newton-Cotes quadrature routines, quad and quad8 both of which are based on an adaptive Simpson's rule. 6.2 Gaussian Quadrature Consider the problem of computing the expectation of a real-valued function f of a random variablex~ with probability density function p( ): Ef(X) ~ = f(x)p(x) dx: Gaussian quadrature is a numerical integration technique that calls forx~ to be replaced with a discrete random variable whose distribution matches that of X~ as closely as possible. Speci cally, in a Gaussian quadrature scheme, the mass points x ;x ;:::;x and probabilities p ;p ;:::;p of the 1 2 n 1 2 n discrete approximant are chosen in such a way that the approximant pos- ~ sesses the same moments of order 2n 1 and below asx: nx px k =E(X ~ k ) i=1 i i k = 0;:::;2n 1: Given the mass points and probabilities of the discrete approximant, the ~ expectation of f(x) is approximated as follows: ~ Ef(X) = n X f(x)p(x) dx¼ f(x )p: i i i=1 Computing the n-degree Gaussian mass points xi and probabilities pi for a random variable involves solving 2n nonlinear equations in as many unknowns. E±cient, specialized numerical routines for computing Gaussian mass points and probabilities are available for virtually every major distribution, including the normal, uniform, gamma, exponential, Chi-square, and beta distributions. In our applications, we will be exclusively concerned with computing expectations of functions of normal random variates and related 4

random variates. For this purpose, we supply a Matlab routine gaussnrm that may be used to compute, in vector format, the n-degree Gaussian mass points x and probabilities p for normal random variates. The routine is called as follows: [x,p] = gaussnrm(n,mu,sigma); Here, n is the degree of approximation and mu and sigma are the mean and standard deviation of the distribution, respectively. If mu and sigma are not speci ed, it is assumed that the mean is zero and the standard deviation is one. By design, an n-point Gaussian quadrature rule will compute the expectation off(x) ~ exactly iff is a polynomial of order 2n 1 or less. Thus, iff can be closely approximated by a polynomial, the Gaussian quadrature rule should provide an accurate estimate of the expectations. Gaussian quadrature rules are consistent for Riemann integrable functions. That is, if f is Riemann integrable, then the approximation a orded by Gaussian quadrature can bemade arbitrarily precise simply by increasing thenumber of mass points in the discretizing distribution. Gaussian quadrature is the numerical integration method of choice when the integrand is bounded and possesses continuous derivatives, but should be applied with great caution otherwise. If the integrand is unbounded, it is often possible to transform the integration problem into an equivalent one with bounded integrand. If the function possesses known kink points, it is often possible to break the integral into the sum of two integrals of smooth functions. If these or similar steps do not produce smooth, bounded integrands, then Newton-Cotes quadrature methods may be more accurate than Gaussian quadrature because they contain the error caused by the kinks and singularities to the interval in they occur. The Gaussian quadrature scheme for normal variates may be used to de- ~ velop a good scheme for lognormal random variates. By de nition, Y is lognormally distributed with parameters¹ and¾if, and only if, it is distributed ~ ~ as exp(x) werex is normally distributed with mean¹ and standard deviation ¾. It follows that iffx;pg are Gaussian mass points and probabilities i i for a normal(¹;¾) distribution, then fy;pg, where y = exp(x ), provides a i i i i reasonable a discrete approximant for a lognormal(¹;¾) distribution. Given this discrete approximant for the lognormal distribution, one can estimate 5

the expectation of a function of Y~ follows: ~ Ef(Y ) = n X f(y)p(y) dy¼ f(y )p: i i i=1 This integration rule for lognormal distributions will be exact if f is a polynomial of degree 2n 1 and less in log(y). Through the use of tensor product principles, the Gaussian quadrature scheme for the univariate normal distribution may be used to construct a Gaussian quadrature scheme for the multivariate normal distribution. Suppose, for example, that X ~ = (X ~ ;X ~ 1 2) is a bivariate normal random variable in row form with (row) mean vector ¹ and variance-covariance matrix. Then X~ is distributed as ¹ + R ~ where R is the Cholesky square root of 0 (e.g., = RR) and ~ is a row 2-vector of independent standard normal variates. If fz;pg i i are the degree n Gaussian mass points and weights for a 2 standard normal variate, then an n degree discrete approximant to X~ may be constructed by forming the mass points x = (¹ +R z +R z ;¹ +R z +R z ) ij 1 11 i 21 j 2 12 i 22 j and associating to them the probabilities p ij =pp i j: A similar procedure may be used for multivariate normal random variates of higher dimension. 6.3 Monte Carlo Integration Monte Carlo integration methods are motivated by the Strong Law of Large Numbers. One version of the Law states that if x 1;x 2;::: are independent realizations of a random variable X~ and f is a continuous function, then 1 nx lim f(x ) =Ef(X) ~ n!1n i=1 i with probability one. The Monte Carlo integration scheme is thus a simple one. To compute an approximation to the integral of f(x), ~ one draws a random sample x ;x ;:::;x from the distribution of X~ 1 2 n and sets 1 nx Ef(X)¼ ~ f(x ): n i=1 i 6

A fundamental problem that arises with Monte Carlo integration is that it is almost impossible to generate a truly random sample of variates for any distribution. Most compilers and vector processing packages provideintrinsic routines for computing so-called random numbers. These routines, however, employ iteration rules that generate a purely deterministic, not random, sequence of numbers. In particular, if the generator is repeatedly initiated at the same point, it will return the same sequence of `random' variates each time. About all that can be said of numerical random number generators is that good ones will generate sequences that appear to be random, in that they pass certain statistical tests for randomness. For this reason, numerical random number generators are often more accurately said to generate sequences of `pseudo-random' rather than random numbers. Matlab o ers two intrinsic random number generators. The routine rand generates a random sample from the uniform(0,1) distribution stored in either vector or matrix format. Similarly, the routine randn generates a random sample from the standard normal distribution stored in either vector or matrix format. In particular, a call of the form x=rand(n) or x=randn(n) generates a random sample of n realizations and stores it in a row vector. The uniform random number generator is useful for generating random samples from other distributions. Suppose X~ has a cumulative distribution function ~ F(x) = Pr(X x) ~ whose inverse has a well-de ned closed form. If U is uniformly distributed on (0; 1), then 1 ~ F (U) is distributed as X. ~ Thus, to generate a random samplex 1;x 2;:::;xn from the X~ distribution, one generates a random sample u 1;u 2;:::;un from the 1 uniform distribution and sets x i = F (u i). The standard normal random number generator is useful for generating random samples from related distributions. For example, to generate a random sample of n lognormal variates, one may use the script x = exp(mu+sigma*randn(n)); where mu and sigma are the parameters of the distribution. To generate a random sample of n d-dimensional normal variates one may use the script 7

x = mu+randn(d,n)*chol(s); where S is the d by d variance-covariance matrix and mu is the mean vector in row form. Monte Carlo integration is easy to implement and may be preferred over Gaussian quadratureifthea routine for computing thegaussian mass points and probabilities is not readily available or if the integration is over many dimensions. Monte Carlo integration, however, is subject to a sampling error that cannot be bounded with certainty. The approximation can be made more accurate, in a statistical sense, by increasing the size of the random sample, but this can be expensive if evaluating f or generating the pseudorandom variate is costly. Approximations generated by Monte Carlo integration will vary from one integration to the next, unless initiated at the same point, making the use of Monte Carlo integration in conjunction within other iterative schemes, such as dynamic programming or maximum likelihood estimation, problematic. So-called quasi Monte-Carlo methods can circumvent some of the problems associated with Monte-Carlo integration. 8