Nguyen, N., Rock, D., and Ying, V. Math 470 Spring 2011 Project 5 Problem 3.24
|
|
- Junior Hubbard
- 5 years ago
- Views:
Transcription
1 Problem 3.24 Here are some modifications of Example 3.8 with consequences that may not be apparent at first. Consider. (a) For and, compare the Riemann approximation with the Monte Carlo approximation. Modify the method in Example 3.8 appropriately, perhaps writing a program that incorporates d = 1/2 and h = x^d, to facilitate easy changes in the parts that follow. Find. The Riemann approximation yields a value of The Monte Carlo approximation yields a value of Also, we have. Code and Output: m = a = 0; b = 1; w = (b - a)/m d = 1/2 x = seq(a + w/2, b-w/2, length=m) h = x^d rect.areas = w*h sum(rect.areas) # Riemann # set.seed(1212) u = runif(m, a, b) h = u^d; y = (b - a)*h mean(y) 2*sd(y)/sqrt(m) var(y) # Monte Carlo # MC margin of error # MC variance > sum(rect.areas) [1] > mean(y) [1] > var(y) # MC variance [1] # Riemann # Monte Carlo (b) What assumptions of Section 3.4 fails for? What is the value of? Of? Try running the two approximations. How do you explain the unexpectedly good behavior of the Monte Carlo simulation? For, the assumption of the function we want to integrate ( ) being bounded is violated. The true value of J is The value of is Running the two approximations, we get an approximation of for the Riemann method and for the Monte Carlo method. The unexpectedly good behavior of the Monte Carlo simulation has to do with the asymptotic behavior of near Page 1 of 28
2 ; only values of very close to 0 yield values of that are so large that they are troublesome? The code for this part is trivial and is not included. (c) Repeat part (b), but with. Comment. For, assumption of the function we want to integrate we want to integrate ( ) being bounded is again violated. The true value of J in this case is ] The value of is The Riemann method yields an approximation of The Monte Carlo method yields an approximation of The bad behavior of the Monte Carlo simulation has to do with the asymptotic behavior of near ; here, values of that weren t a problem in part (b) are so high that even random points are not enough to capture the area under the curve for values of x near 0. The code for this part is trivial and is not included. Page 2 of 28
3 Problem 3.25 This problem shows how the rapid oscillation of a function can affect the accuracy of a Riemann approximation. (a) Let and be a positive integer. Then use calculus to show that. Use the code below to plot k = 4 x = seq(0,1, by = 0.01); plot(x, h, type="l") ] for on. h = abs(sin(k*pi*x)) Using calculus, we have [ ] ( ) ( ) (1) as required. Plot Figure plot of h on [0,1] for k=4 Page 3 of 28
4 (b) Modify the program of Example 3.8 to integrate. Use throughout, and make separate runs with and. Compare the accuracy of the resulting Riemann and Monte Carlo approximations, and explain the behavior of the Riemann approximation. The Riemann approximation is less accurate than the Monte Carlo approximation (as can be seen in the table below) maybe because the non-random points of evaluation used in Riemann approximation are too evenly spaced to accurately evaluate the evenly spaced oscillation of the. function m Riemann approximation Riemann error Monte Carlo approximation Monte Carlo error Table comparison of accuracy of Riemann and Monte Carlo approximations Code: k = 5000 m.vector = c(2500,5000,10000,15000,20000) true.answer = 2/pi for( m in m.vector) #Riemann a = 0; b = 1; w = (b - a)/m x = seq(a + w/2, b-w/2, length=m) h = abs(sin(k*pi*x)) rect.areas = w*h riemann.answer = sum(rect.areas) # Riemann #Monte Carlo #set.seed(1212) #uncomment to reproduce results exactly u = runif(m, a, b) h = abs(sin(k*pi*u)); y = (b - a)*h monte.carlo.answer = mean(y) # Monte Carlo monte.carlo.answer #accuracy of approximations riemann.error = abs( riemann.answer - true.answer) monte.carlo.error = abs( monte.carlo.answer - true.answer) print( paste( "m=", m, ". Riemann approximation: ", riemann.answer, ". Riemann error: ", riemann.error, ". Monte Carlo approximation: ", monte.carlo.answer, ". Monte Carlo error: ", monte.carlo.error, ".", sep="")) Page 4 of 28
5 (c) Use calculus to show that. How accurately is this value approximated by simulation? If, find the margin of error for the Monte Carlo approximation in part (b) based on and the Central Limit Theorem. Are your results consistent with this margin of error? To verify the equality, we first we need the following lemma: Lemma: [ ] [( ) ( )] Then, from calculus, we have ( ) ] ] [( ) ] ( ) (2) ( ) ( ) as required. This value is closely approximated by simulation, as confirmed in the table below. m Monte Carlo approximation of variance Table accuracy when approximating V(Y) For, the (95%) margin of error for the Monte Carlo approximation in part (b) based on is Note that the true answer is, so that the 95% CI based on is. Hence, Page 5 of 28
6 our Mote Carlo approximation (front part (b) with margin of error. ) of is consistent with this Code and Output: k = 5000 m.vector = c(2500,5000,10000,15000,20000) #accuracy in approximating variance using Monte Carlo method (for part c) for( m in m.vector) #Monte Carlo #set.seed(1212) #uncomment to reproduce results exactly u = runif(m, a, b) h = abs(sin(k*pi*u)); y = (b - a)*h monte.carlo.answer = mean(y) # Monte Carlo monte.carlo.answer #accuracy of approximation of variance riemann.error = abs( riemann.answer - true.answer) monte.carlo.error = abs( monte.carlo.answer - true.answer) print( paste( "m=", m, ". Monte Carlo approximatin of variance: ", var(y), ".", sep="")) #margin of error for Monte Carlo approximation in part (b) based on SD(Y) and CLT, when m=10000 m = #set.seed(1212) #uncomment to reproduce results exactly u = runif(m, a, b) h = abs(sin(k*pi*u)); y = (b - a)*h 2*sd(y)/sqrt(m) # MC margin of error > 2*sd(y)/sqrt(m)# MC margin of error [1] Page 6 of 28
7 Problem 3.26 The integral cannot be evaluated analytically, but advanced analytic methods yield. (a) Assuming this result, show that. Use R to plot both integrands on, obtaining results as in Figure We have ] ( ( )) [ ] as required. Plots: Figure plot of first integrand Page 7 of 28
8 Figure plot of second integrand Codes: par( mfrow= c(2,1)) curve( (sin(1/x))^2, from=0, to=1, n=10000, main="first Integrand") curve( (sin(x)/x)^2, from=0, to=1, n=10000, main="second Integrand") par( mfrow= c(1,1)) (b) Use both Riemann and Monte Carlo approximations to evaluate as originally defined. Then evaluate using the equation in part (a). Try both methods with, and iterations. What do you believe is the best answer? Comment on differences between methods and between equations. The result of each method and each equation are given in tables below. It is hard to determine which answer is the best. Regarding differences when using the equation for as originally defined, both methods seem to have performed similarly; when using the equation in part (a), however, the Riemann method is more consistent. m Riemann approximation Monte Carlo approximation Table Evaluating as originally defined m Riemann approximation Monte Carlo approximation Table Evaluating using equation in part (a) Page 8 of 28
9 Codes: m.vector = c(100,1000,1001,10000) #evaluating J as originally defined: for( m in m.vector) #Riemann a = 0; b = 1; w = (b - a)/m x = seq(a + w/2, b-w/2, length=m) h = (sin(1/x))^2 rect.areas = w*h riemann.answer = sum(rect.areas) # Riemann Monte Carlo #Monte Carlo set.seed(1212) #uncomment to reproduce results exactly u = runif(m, a, b) h = (sin(1/u))^2; y = (b - a)*h monte.carlo.answer = mean(y) # monte.carlo.answer #accuracy of approximations print( paste( "m=", m, ". Riemann approximation: ", riemann.answer, ". Monte Carlo approximation: ", monte.carlo.answer, ".", sep="")) #evaluating J as defined in part (a): for( m in m.vector) #Riemann a = 0; b = 1; w = (b - a)/m x = seq(a + w/2, b-w/2, length=m) h = pi/2 - (x^-2)*(sin(x))^2 rect.areas = w*h riemann.answer = sum(rect.areas) # Riemann Monte Carlo #Monte Carlo set.seed(1212) #uncomment to reproduce results exactly u = runif(m, a, b) h = pi/2 - (u^-2)*(sin(u))^2; y = (b - a)*h monte.carlo.answer = mean(y) # monte.carlo.answer #accuracy of approximations print( paste( "m=", m, ". Riemann approximation: ", riemann.answer, ". Monte Carlo approximation: ", monte.carlo.answer, ".", sep="")) Page 9 of 28
10 Problem 3.27 Modify the program of Example 3.9 to approximate the volume beneath the bivariate standard normal density surface and above two additional regions of integration as specified below. Use both the Riemann and Monte Carlo methods in parts (a) and (b), with. (a) Evaluate. Because and are independent standard normal random variables, we know that this probability is. For each method, say whether it would have been better to use points to find and then square the answer. For the Riemann approximation method, it would have been better to use points to find and then square the answer. However, for the Monte Carlo method it would have been better to find the answer directly. Surprisingly, the Riemann method performed better than the Monte Carlo method. algorithm Riemann approx. Riemann error MC approx. MC error Table Comparison of methods and algorithms Code and Outputs: true.answer = (pnorm(1) - 0.5)^2 #== ## METHOD OF DIRECTLY FINDING P0 < Z1 <= 1, 0 < Z2 <= 1: # Riemann approximation m = g = round(sqrt(m)) # no. of grid pts on each axis x1 = rep((1:g - 1/2)/g, times=g) # these two lines give x2 = rep((1:g - 1/2)/g, each=g) # coordinates of grid points hx = dnorm(x1)*dnorm(x2) riemann.approx = sum(hx[x1 <= 1 & x2 <= 1])/g^2 # Riemann approximation riemann.approx riemann.error = abs( true.answer - riemann.approx) riemann.error > riemann.approx [1] > riemann.error [1] e-07 # Monte Carlo approximation set.seed(pi) # uncomment to reproduce results exactly u1 = runif(m) # these two lines give a random u2 = runif(m) # point in the unit square hu = dnorm(u1)*dnorm(u2) hu.acc = hu[u1 <= 1 & u2 <= 1] # heights above accepted points mc.approx = mean(hu.acc) # Monte Carlo result mc.approx mc.error = abs( true.answer - mc.approx) Page 10 of 28
11 mc.error > mc.approx [1] > mc.error [1] e-05 ## # METHOD OF FINDING P0 < Z <= 1 AND THEN SQUARING: # Riemann approximation (works) m = a = 0; b = 1; w = (b - a)/m x = seq(a + w/2, b - w/2, length=m) h = dnorm(x) rect.areas = w*h riemann.approx = (sum(rect.areas))^2 riemann.error = abs( true.answer - riemann.approx) riemann.error > riemann.approx [1] > riemann.error [1] e-11 # Alternative Riemann approximation (equivalent to above) m=10000 x1 = (1:m - 1/2)/m hx = dnorm(x1) riemann.approx = (sum(hx[x1 < 1])/m)^2 riemann.approx riemann.error = abs( true.answer - riemann.approx) riemann.error > riemann.approx [1] > riemann.error [1] e-11 # MC approximation: #set.seed(pi) # uncomment to reproduce results exactly m=10000 u1 = runif(m,0,1) hu = dnorm(u1) hu.acc = hu[u1 < 1] mc.approx = mean(hu.acc)^2 # mc.approx = (sum( hu[u1 < 1] ) / sum(u1<1))^2 #same as above mc.approx mc.error = abs( true.answer - mc.approx) mc.error > mc.approx [1] Page 11 of 28
12 > mc.error [1] (b) Evaluate. Here the region of integration does not have area 1, so remember to multiply by an appropriate constant. Because CHISQ(2), the exact answer can be found with pchisq(1,2). Note that the true answer is pchisq(1,2)= The Riemann method yields an approximation of ; the error of this approximation is The Monte Carlo method yields an approximation of ; the error of this approximation is Codes: #true answer: true.answer = pchisq(1,2) > true.answer [1] # riemann method (correct?): m = g = round(sqrt(m)) x = rep((1:g)/g,times=g) y = rep((1:g)/g,each=g) hx = dnorm(x)*dnorm(y) riemann.approx = sum( hx[ (x^2) + (y^2) < 1 ]) / m * 4 riemann.approx #== riemann.error = abs( riemann.approx - true.answer) riemann.error #== > riemann.approx [1] > riemann.error [1] # MC approximation 1 (works): set.seed(1237) #uncomment to reproduce results exactly m = u1 = runif(m,-1,1) u2 = runif(m,-1,1) hu = dnorm(u1)*dnorm(u2) hu.acc = hu[ (u1^2) + (u2^2) < 1 ] mc.approx = mean(hu.acc) * pi mc.approx #== mc.error = abs( mc.approx - true.answer) mc.error #== > mc.approx #== [1] > mc.error #== [1] Page 12 of 28
13 # MC approximation 2 (this works even better than above): set.seed(1237) #uncomment to reproduce results exactly m = u1 = runif(m,0,1) u2 = runif(m,0,1) hu = dnorm(u1)*dnorm(u2) hu.acc = hu[ (u1^2) + (u2^2) < 1 ] mc.approx = mean(hu.acc) * pi / 4 * 4 mc.approx #== mc.error = abs( mc.approx - true.answer) mc.error #== > mc.approx #== [1] > mc.error #== [1] (c) The joint density function of has circular contour lines centered at the origin, so that probabilities of regions do not change if they are rotated about the origin. Use this fact to argue, which was approximated in Example 3.9, can be found that the exact value of with (pnorm(1/sqrt(2)) 0.5)^2. Consider the series of rotations about the origin illustrated in Figures through This response will continue after the break Figure Page 13 of
14 Figure = 1/ = 1/ = 1/ = 1/ Figure Since probabilities of regions do not change if they are rotated about the origin, then the probability of the triangular region in Figure , must be equal to the probability of the square region in Figure This probability is just ( ), where standard normal random variable. However, this is equivalent to ( is the density function of the ) (, where is the cdf of a standard normal random variable), which in turn is equivalent to ( )., which was approximated in Example 3.9, can be found with Thus the exact value of (pnorm(1/sqrt(2)) 0.5)^2, as required. Page 14 of 28
15 Problem 3.28 Here we extend the idea of Example 3.9 to three dimensions. Suppose three items are drawn at random from a population of items with weights (in grams) distributed as NORM(100,10). (a) Using the R function pnorm (cumulative distribution function of a standard normal), find the probability that the sum of the three weights is less than 310 g. Also find the probability that the minimum weight of these three items exceeds 100 g. For the probability that the sum is less than 310 g, note that since then and so we have (3.28.1) where is a standard normal random variable. pnorm(1/sqrt(3)) = , as required. Thus the answer for the sum is ( ) For the probability that the minimum exceeds 100 g, we have where the last equality comes from the R command pnorm(100, mean=100, sd=10, lower.tail=f)^3. Similarly, in terms of standard normal random variables, we could have used ( ) as required. Page 15 of 28 (3.28.2)
16 (b) Using both Riemann and Monte Carlo methods, approximate the probability that (simultaneously) the minimum of the weights exceeds 100 g and their sum is less than 310 g. The suggested procedure is to (i) express this problem in terms of three standard normal random variables, and (ii) modify appropriate parts of the program in Example 3.9 to approximate the required integral over a triangular cone (of area 1/6) in the unit cube using. Here is R code to make the three grid vectors needed for the Riemann approximation: [for the R code, see page 84 of the text] The probability that the minimum of the weights exceeds 100 g can be expressed as: (3.28.3) where the last equality follows since. The probability that the sum is less than 310 g can be expressed as: (3.28.4) where the s are standard normal random variables. Since the three grid vectors provided in the problem statement contain points with components all greater than 0, we don t need to worry about defining a region of integration that satisfies equation (3.28.3). We only need to worry about equation (3.28.4). Hence, in our R code below, we only need to worry about accepting points that satisfy the condition x1 + x2 + x3 < 1. Here are our results: True answer from book Riemann Approximation MC Approximation Approximately Table probability that (simultaneously) the minimum of the weights exceeds 100 g and their sum is less than 310 g. Code: ### Riemann approximation ## g = 25 m = g^3 x1 = rep((1:g-1/2)/g,each=g^2) x2 = rep(rep((1:g-1/2)/g,each=g),times=g) x3 = rep((1:g-1/2)/g,times=g^2) hx = dnorm(x1)*dnorm(x2)*dnorm(x3) riemann.approx = sum(hx[x1 + x2 + x3 < 1])/g^3 riemann.approx #== ### Monte Carlo Simulation ### set.seed(pi) #uncomment to reporduce results exactly m = u1 = runif(m) u2 = runif(m) u3 = runif(m) hu = dnorm(u1)*dnorm(u2)*dnorm(u3) hu.acc = hu[u1 + u2 + u3 < 1] mc.approx = (1/6) * mean(hu.acc) mc.approx #== Page 16 of 28
17 Problem 3.29 For and, use Monte Carlo approximation with m= to find the probability that a variate (independent) standard normal distribution places in the part of the -dimensional unit ball with all coordinates positive. To do this, imitate the program in Example 3.9, and use the fact that the entire 4-dimensional unit ball has hypervolume. Compare the results with appropriate values computed using the chi-squared distribution with degrees of freedom; use the R function pchisq. Results are presented in Table Code is after that exact result using pchisq Monte Carlo result Monte Carlo error e Table Code: fark = function(d) # Input: d - the dimention of the unit ball whose hypervolume we wish to calculuate # Returns: the hypervolume of the d-dimensional unit ball hypervolume = (pi^(d/2)) / gamma( (d+2)/2) return( hypervolume) m = ## # for d = 2 d=2 # true answer true.answer = pchisq(1,d) #== true.answer # MC answer (works): set.seed(pi) #uncomment to reproduce results exactly u1 = runif(m,-1,1) u2 = runif(m,-1,1) hu = dnorm(u1)*dnorm(u2) hu.acc = hu[ (u1^2) + (u2^2) < 1 ] mc.approx = mean(hu.acc) * fark(d) mc.approx > mc.approx [1] > mc.error [1] e-05 ## # for d = 3 Page 17 of 28
18 d=3 # true answer true.answer = pchisq(1,d) #== true.answer # MC answer (works): set.seed(pi) u1 = runif(m,-1,1) u2 = runif(m,-1,1) u3 = runif(m,-1,1) hu = dnorm(u1)*dnorm(u2)*dnorm(u3) hu.acc = hu[ (u1^2) + (u2^2) + (u3^2) < 1 ] mc.approx = mean(hu.acc) * fark(d) mc.approx > mc.approx [1] > mc.error [1] ## # for d = 4 d=4 # true answer true.answer = pchisq(1,d) #== true.answer # MC answer (works): set.seed(pi) u1 = runif(m,-1,1) u2 = runif(m,-1,1) u3 = runif(m,-1,1) u4 = runif(m,-1,1) hu = dnorm(u1)*dnorm(u2)*dnorm(u3)*dnorm(u4) hu.acc = hu[ (u1^2) + (u2^2) + (u3^2) + (u4^2) < 1 ] mc.answer = mean(hu.acc) * fark(d) mc.answer > mc.answer [1] > mc.error [1] Page 18 of 28
19 Problem 6.1 In each part below, consider the three Markov chains specified as follows: (i), ; (ii), ; and (iii),. (a) Find and, for. For we have and Hence, we have (i) and. (ii) and. (iii) and. (b) Use the given values of and and means of geometric distributions to find the average cycle length for each chain. The average cycle length of a chain is (i) The average cycle length is (ii) The average cycle length is (iii) The average cycle length is Hence for (c) For each chain, modify the program of Example 6.1 to find the long-run fraction of steps in state 1. The long-run fraction of steps in state 1 are: (i) , (ii) 0.273, and (iii) Codes (for parts (c) and (d)) Pause <- function () cat("hit <enter> to continue...") readline() invisible() # Preliminaries set.seed(1237) #uncomment to reproduce results exactly m = 2000; n = 1:m; x = numeric(m); x[1] = 0 #alpha = 1; beta = 0.44 alpha = c(0.3,0.15,0.03); beta = c(0.7,0.35,0.07) label = c( "(i) alpha = 0.3, beta = 0.7", "(ii) alpha = 0.15, beta = 0.35", "(iii) alpha = 0.03, beta = 0.07", sep="") # Simulation for (j in 1:3) for (i in 2:m) Page 19 of 28
20 if (x[i-1]==0) x[i] = rbinom(1, 1, alpha[j]) else x[i] = rbinom(1, 1, 1 - beta[j]) y = cumsum(x)/n # Running fractions of Long eruptions # Results lrf = y[m]; # long run fraction of steps in state 1 among m. Same as: mean(x) print( paste( "long run fraction of steps in state 1 for ", label[j], ": ", lrf, sep="")) Pause() a = sum(x[1:(m-1)]==0 & x[2:m]==1); a # No. of cycles acl = m/a; acl # Average cycle length plot(x[1:round(6*acl)], type="b", xlab="step", ylab="state", main=label[j]) # Similar to Fig. 6.3 Pause() # ---- (New plot, similar to Fig. 6.4) plot(y, type="l", ylim=c(0,1), xlab="step", ylab="proportion Long", main=label[j]) Pause() # ---- (New plot, similar to Fig. 6.5; additional output) acf(x, main=label[j]) # Autocorrelation plot acf(x, plot=f) # Printed output Pause() (d) For each chain, make and interpret plots similar to Figures 6.3 (where the number of steps is chosen to illustrate the behavior clearly), 6.4, and 6.5. For (i) we have: Page 20 of 28
21 For (ii), we have: For (iii), we have Codes See part (c). (e) In summary, what do these three chains have in common, and in what respects are they most remarkably different. Is any one of these chains an independent process, and how do you know? In summary, all three chains have in common the fact that they converge to a stationary distribution. They are most remarkably different in terms of their cycle lengths. The only chain that is an independent process is the one from part (i); we know this from the essentially zero autocorrelation. Page 21 of 28
22 Problem 6.2 There is more information in the joint distribution of two random variables than can be discerned by looking only at their marginal distributions. Consider two random variables and, each distributed as BINOM(1, ), where. (a) In general, show that. In particular, evaluate in three cases: where and are independent, where, and where. Case 1: When and are independent, we have (where the last equality follows since the. Then, since, we have. Hence we have Case 2: When, we have. Hence we have where the left-hand (in)equality follows since. Case 3: When, we have Hence we have where the right-hand (in)equality follows since. Thus, in each case, we have as required. (b) For each case in part (a), evaluate. Case 1: When and are independent, we have where the last equality follows since the. Case 2: When, we have. Case 3: When, we have. Page 22 of 28
23 (c) If and, then express,, and in terms of and. First, we have (3) as required. Second, we have (4) Last, we have ( ) (5) (d) In part (c), find the correlation, recalling that. First, we have where the last equality follows from (5). Second (since ), we have, so that Page 23 of 28 ( ) where the last equality comes from equation (3). Third (since ), we have, so that (6) (7)
24 Thus the correlation is ( ) ( ) (8) ( ) ( ) ( ) ( ) ( ) ( ) as required. (9) Page 24 of 28
25 Problem 6.3 Geometric distributions. Consider a coin with. A geometric random variable can be used to model the number of independent tosses required until the first Head is seen. (a) Show that the probability function For any, we have, for as required. (b) Show that the geometric series The sum of the geometric series sums to 1, so that one is sure to see a Head eventually. is (12) Multiplying equation (12) by, yields (13) Finally, subtracting (13) from (12) yields ] Thus the geometric series sums to 1, and one is sure to see a Head eventually. (c) Show that the moment generating function of hence that. ], and is (You may assume that the limits involved in differentiation and summing an infinite series can be interchanged as required.) The mgf of is (14) Multiplying (14) by yields (15) Finally, subtracting (15) from (14) and solving for and hence, we get, as required. Page 25 of 28
26 Problem 6.4 Suppose the weather for a day is either Dry (0) or Rainy (1) according to a homogeneous 2-state Markov chain with and. Today is Monday and the weather is Dry. (a) What is the probability that both tomorrow and Wednesday will be Dry? We have (b) What is the probability that it will be Dry on Wednesday? We have (c) Use equation (6.5) [(sic)] to find the probability that it will be Dry two weeks from Wednesday. The probability that it will be Dry two weeks from Wednesday ( ), given that it is dry today ( ) can be found using the 16-step transition matrix. Specifically, this probably,, is the upper-left entry of the following 16-step transition matrix [ ] * + Thus the probability that it will be Dry two weeks from Wednesday is (d) Modify the R code of Example 6.2 to find the probability that it will be Dry two weeks from Wednesday. The answer is, again, Code and Outputs: P = matrix(c( 0.9, 0.1, 0.5, 0.5), nrow=2, ncol=2, byrow=t) P Page 26 of 28
27 P2 = P %*% P; P2 P4 = P2 %*% P2; P4 P8 = P4 %*% P4; P8 P16 = P8 %*% P8; P16 > P = matrix(c( 0.9, 0.1, + 0.5, 0.5), nrow=2, ncol=2, byrow=t) > P [,1] [,2] [1,] [2,] > P2 = P %*% P; P2 [,1] [,2] [1,] [2,] > P4 = P2 %*% P2; P4 [,1] [,2] [1,] [2,] > P8 = P4 %*% P4; P8 [,1] [,2] [1,] [2,] > P16 = P8 %*% P8; P16 [,1] [,2] [1,] [2,] (e) Over the long run, what will be the proportion of Rainy days? Modify the R code of Example 6.1 to simulate the chain and find an approximate answer. The predicted proportion of Rainy days we get from simulation is The modification to the R code of Example 6.1 is rudimentary and so it is not included here. (f) What is the average length of runs of Rainy days? Note that we have. Thus the average length of a run of Rainy (1) days is. (g) How do the answers above change if and? If and, then (a') the probability that both tomorrow and Wednesday will be Dry becomes. (b') the probability that it will be Dry on Wednesday becomes. (c') the probability that it will be Dry two weeks from Wednesday becomes (d') modifying the R code of Example 6.2 to find the probability that it will be Dry two weeks from Wednesday yields an answer of The changes to the code in part (d) are rudimentary and so the code is not included. Page 27 of 28
28 (e') modifying the code of Example 6.1 to simulate the chain suggests that, over the long run, the proportion of Rainy days will be approximately The changes to the code in part (e) is rudimentary and so the code is not included. (f') we have., so that the average length of a run of Rainy (1) days is Page 28 of 28
Question Points Score Total: 76
Math 447 Test 2 March 17, Spring 216 No books, no notes, only SOA-approved calculators. true/false or fill-in-the-blank question. You must show work, unless the question is a Name: Question Points Score
More informationMarkov Chains. Chapter 16. Markov Chains - 1
Markov Chains Chapter 16 Markov Chains - 1 Why Study Markov Chains? Decision Analysis focuses on decision making in the face of uncertainty about one future event. However, many decisions need to consider
More informationProbability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology Kharagpur
Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology Kharagpur Lecture No. #13 Probability Distribution of Continuous RVs (Contd
More informationThis exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.
TEST #3 STA 5326 December 4, 214 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. (You will have access to
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationMATH Notebook 5 Fall 2018/2019
MATH442601 2 Notebook 5 Fall 2018/2019 prepared by Professor Jenny Baglivo c Copyright 2004-2019 by Jenny A. Baglivo. All Rights Reserved. 5 MATH442601 2 Notebook 5 3 5.1 Sequences of IID Random Variables.............................
More informationProbability Methods in Civil Engineering Prof. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur
Probability Methods in Civil Engineering Prof. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur Lecture No. # 12 Probability Distribution of Continuous RVs (Contd.)
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationStatistical Computing: Exam 1 Chapters 1, 3, 5
Statistical Computing: Exam Chapters, 3, 5 Instructions: Read each question carefully before determining the best answer. Show all work; supporting computer code and output must be attached to the question
More informationJoint Distribution of Two or More Random Variables
Joint Distribution of Two or More Random Variables Sometimes more than one measurement in the form of random variable is taken on each member of the sample space. In cases like this there will be a few
More information7 Random samples and sampling distributions
7 Random samples and sampling distributions 7.1 Introduction - random samples We will use the term experiment in a very general way to refer to some process, procedure or natural phenomena that produces
More information1 Introduction. P (n = 1 red ball drawn) =
Introduction Exercises and outline solutions. Y has a pack of 4 cards (Ace and Queen of clubs, Ace and Queen of Hearts) from which he deals a random of selection 2 to player X. What is the probability
More informationMCMC notes by Mark Holder
MCMC notes by Mark Holder Bayesian inference Ultimately, we want to make probability statements about true values of parameters, given our data. For example P(α 0 < α 1 X). According to Bayes theorem:
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationComputational statistics
Computational statistics Markov Chain Monte Carlo methods Thierry Denœux March 2017 Thierry Denœux Computational statistics March 2017 1 / 71 Contents of this chapter When a target density f can be evaluated
More informationEco517 Fall 2004 C. Sims MIDTERM EXAM
Eco517 Fall 2004 C. Sims MIDTERM EXAM Answer all four questions. Each is worth 23 points. Do not devote disproportionate time to any one question unless you have answered all the others. (1) We are considering
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More informationAlgebra 2: Semester 2 Practice Final Unofficial Worked Out Solutions by Earl Whitney
Algebra 2: Semester 2 Practice Final Unofficial Worked Out Solutions by Earl Whitney 1. The key to this problem is recognizing cubes as factors in the radicands. 24 16 5 2. The key to this problem is recognizing
More informationMath Stochastic Processes & Simulation. Davar Khoshnevisan University of Utah
Math 5040 1 Stochastic Processes & Simulation Davar Khoshnevisan University of Utah Module 1 Generation of Discrete Random Variables Just about every programming language and environment has a randomnumber
More informationMath Week 1 notes
Math 2270-004 Week notes We will not necessarily finish the material from a given day's notes on that day. Or on an amazing day we may get farther than I've predicted. We may also add or subtract some
More informationthe law of large numbers & the CLT
the law of large numbers & the CLT Probability/Density 0.000 0.005 0.010 0.015 0.020 n = 4 0.0 0.2 0.4 0.6 0.8 1.0 x-bar 1 sums of random variables If X,Y are independent, what is the distribution of Z
More informationEXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY (formerly the Examinations of the Institute of Statisticians) GRADUATE DIPLOMA, 2004
EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY (formerly the Examinations of the Institute of Statisticians) GRADUATE DIPLOMA, 004 Statistical Theory and Methods I Time Allowed: Three Hours Candidates should
More informationBayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007
Bayesian inference Fredrik Ronquist and Peter Beerli October 3, 2007 1 Introduction The last few decades has seen a growing interest in Bayesian inference, an alternative approach to statistical inference.
More informationCourse: ESO-209 Home Work: 1 Instructor: Debasis Kundu
Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear
More informationSYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions
SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu
More informationRandomized Algorithms
Randomized Algorithms Prof. Tapio Elomaa tapio.elomaa@tut.fi Course Basics A new 4 credit unit course Part of Theoretical Computer Science courses at the Department of Mathematics There will be 4 hours
More informationMore on Distribution Function
More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution
More informationSTOCHASTIC MODELING OF ENVIRONMENTAL TIME SERIES. Richard W. Katz LECTURE 5
STOCHASTIC MODELING OF ENVIRONMENTAL TIME SERIES Richard W Katz LECTURE 5 (1) Hidden Markov Models: Applications (2) Hidden Markov Models: Viterbi Algorithm (3) Non-Homogeneous Hidden Markov Model (1)
More information2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.
CS450 Final Review Problems Fall 08 Solutions or worked answers provided Problems -6 are based on the midterm review Identical problems are marked recap] Please consult previous recitations and textbook
More informationName: SOLUTIONS Exam 01 (Midterm Part 2 take home, open everything)
Name: SOLUTIONS Exam 01 (Midterm Part 2 take home, open everything) To help you budget your time, questions are marked with *s. One * indicates a straightforward question testing foundational knowledge.
More informationSimulation. Where real stuff starts
1 Simulation Where real stuff starts ToC 1. What is a simulation? 2. Accuracy of output 3. Random Number Generators 4. How to sample 5. Monte Carlo 6. Bootstrap 2 1. What is a simulation? 3 What is a simulation?
More informationGOV 2001/ 1002/ Stat E-200 Section 1 Probability Review
GOV 2001/ 1002/ Stat E-200 Section 1 Probability Review Solé Prillaman Harvard University January 28, 2015 1 / 54 LOGISTICS Course Website: j.mp/g2001 lecture notes, videos, announcements Canvas: problem
More informationChapter 6: Functions of Random Variables
Chapter 6: Functions of Random Variables We are often interested in a function of one or several random variables, U(Y 1,..., Y n ). We will study three methods for determining the distribution of a function
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More informationTMA 4265 Stochastic Processes Semester project, fall 2014 Student number and
TMA 4265 Stochastic Processes Semester project, fall 2014 Student number 730631 and 732038 Exercise 1 We shall study a discrete Markov chain (MC) {X n } n=0 with state space S = {0, 1, 2, 3, 4, 5, 6}.
More informationProbability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur
Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institute of Technology, Kharagpur Lecture No. # 33 Probability Models using Gamma and Extreme Value
More informationMath 416 Lecture 3. The average or mean or expected value of x 1, x 2, x 3,..., x n is
Math 416 Lecture 3 Expected values The average or mean or expected value of x 1, x 2, x 3,..., x n is x 1 x 2... x n n x 1 1 n x 2 1 n... x n 1 n 1 n x i p x i where p x i 1 n is the probability of x i
More informationBayesian Methods for Machine Learning
Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),
More informationStat 451 Lecture Notes Monte Carlo Integration
Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:
More informationCS 361: Probability & Statistics
October 17, 2017 CS 361: Probability & Statistics Inference Maximum likelihood: drawbacks A couple of things might trip up max likelihood estimation: 1) Finding the maximum of some functions can be quite
More information1, 0 r 1 f R (r) = 0, otherwise. 1 E(R 2 ) = r 2 f R (r)dr = r 2 3
STAT 5 4.43. We are given that a circle s radius R U(, ). the pdf of R is {, r f R (r), otherwise. The area of the circle is A πr. The mean of A is E(A) E(πR ) πe(r ). The second moment of R is ( ) r E(R
More informationTransformations of Standard Uniform Distributions
Transformations of Standard Uniform Distributions We have seen that the R function runif uses a random number generator to simulate a sample from the standard uniform distribution UNIF(0, 1). All of our
More informationR Functions for Probability Distributions
R Functions for Probability Distributions Young W. Lim 2018-03-22 Thr Young W. Lim R Functions for Probability Distributions 2018-03-22 Thr 1 / 15 Outline 1 R Functions for Probability Distributions Based
More informationProbability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur
Probability Methods in Civil Engineering Prof. Dr. Rajib Maity Department of Civil Engineering Indian Institution of Technology, Kharagpur Lecture No. # 36 Sampling Distribution and Parameter Estimation
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationStat751 / CSI771 Midterm October 15, 2015 Solutions, Comments. f(x) = 0 otherwise
Stat751 / CSI771 Midterm October 15, 2015 Solutions, Comments 1. 13pts Consider the beta distribution with PDF Γα+β ΓαΓβ xα 1 1 x β 1 0 x < 1 fx = 0 otherwise, for fixed constants 0 < α, β. Now, assume
More informationTopic 15: Simple Hypotheses
Topic 15: November 10, 2009 In the simplest set-up for a statistical hypothesis, we consider two values θ 0, θ 1 in the parameter space. We write the test as H 0 : θ = θ 0 versus H 1 : θ = θ 1. H 0 is
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationPart IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationMonte Carlo Integration. Computer Graphics CMU /15-662, Fall 2016
Monte Carlo Integration Computer Graphics CMU 15-462/15-662, Fall 2016 Talk Announcement Jovan Popovic, Senior Principal Scientist at Adobe Research will be giving a seminar on Character Animator -- Monday
More informationThe biggest Markov chain in the world
The biggest Markov chain in the world Randy s web-surfing behavior: From whatever page he s viewing, he selects one of the links uniformly at random and follows it. Defines a Markov chain in which the
More informationIntroduction and Overview STAT 421, SP Course Instructor
Introduction and Overview STAT 421, SP 212 Prof. Prem K. Goel Mon, Wed, Fri 3:3PM 4:48PM Postle Hall 118 Course Instructor Prof. Goel, Prem E mail: goel.1@osu.edu Office: CH 24C (Cockins Hall) Phone: 614
More informationName of the Student:
SUBJECT NAME : Probability & Queueing Theory SUBJECT CODE : MA 6453 MATERIAL NAME : Part A questions REGULATION : R2013 UPDATED ON : November 2017 (Upto N/D 2017 QP) (Scan the above QR code for the direct
More informationCS 6820 Fall 2014 Lectures, October 3-20, 2014
Analysis of Algorithms Linear Programming Notes CS 6820 Fall 2014 Lectures, October 3-20, 2014 1 Linear programming The linear programming (LP) problem is the following optimization problem. We are given
More informationSTAT 414: Introduction to Probability Theory
STAT 414: Introduction to Probability Theory Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical Exercises
More information1 Probability Distributions
1 Probability Distributions A probability distribution describes how the values of a random variable are distributed. For example, the collection of all possible outcomes of a sequence of coin tossing
More informationTABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1
TABLE OF CONTENTS CHAPTER 1 COMBINATORIAL PROBABILITY 1 1.1 The Probability Model...1 1.2 Finite Discrete Models with Equally Likely Outcomes...5 1.2.1 Tree Diagrams...6 1.2.2 The Multiplication Principle...8
More informationMidterm Examination. Mth 136 = Sta 114. Wednesday, 2000 March 8, 2:20 3:35 pm
Midterm Examination Mth 136 = Sta 114 Wednesday, 2000 March 8, 2:20 3:35 pm This is a closed-book examination so please do not refer to your notes, the text, or to any other books. You may use a two-sided
More informationStat 5102 Notes: Markov Chain Monte Carlo and Bayesian Inference
Stat 5102 Notes: Markov Chain Monte Carlo and Bayesian Inference Charles J. Geyer April 6, 2009 1 The Problem This is an example of an application of Bayes rule that requires some form of computer analysis.
More informationStat 5102 Notes: Markov Chain Monte Carlo and Bayesian Inference
Stat 5102 Notes: Markov Chain Monte Carlo and Bayesian Inference Charles J. Geyer March 30, 2012 1 The Problem This is an example of an application of Bayes rule that requires some form of computer analysis.
More informationChapter 4 Sampling Distributions and Limits
Chapter 4 Sampling Distributions and Limits CHAPTER OUTLINE Section 1 Sampling Distributions Section Convergence in Probability Section 3 Convergence with Probability 1 Section 4 Convergence in Distribution
More informationMAT137 - Term 2, Week 4
MAT137 - Term 2, Week 4 Reminders: Your Problem Set 6 is due tomorrow at 3pm. Test 3 is next Friday, February 3, at 4pm. See the course website for details. Today we will: Talk more about substitution.
More informationThis exam contains 6 questions. The questions are of equal weight. Print your name at the top of this page in the upper right hand corner.
GROUND RULES: This exam contains 6 questions. The questions are of equal weight. Print your name at the top of this page in the upper right hand corner. This exam is closed book and closed notes. Show
More informationThe data was collected from the website and then converted to a time-series indexed from 1 to 86.
Introduction For our group project, we analyzed the S&P 500s futures from 30 November, 2015 till 28 March, 2016. The S&P 500 futures give a reasonable estimate of the changes in the market in the short
More informationCSC 446 Notes: Lecture 13
CSC 446 Notes: Lecture 3 The Problem We have already studied how to calculate the probability of a variable or variables using the message passing method. However, there are some times when the structure
More informationBTRY 4090: Spring 2009 Theory of Statistics
BTRY 4090: Spring 2009 Theory of Statistics Guozhang Wang September 25, 2010 1 Review of Probability We begin with a real example of using probability to solve computationally intensive (or infeasible)
More informationConvergence Rate of Markov Chains
Convergence Rate of Markov Chains Will Perkins April 16, 2013 Convergence Last class we saw that if X n is an irreducible, aperiodic, positive recurrent Markov chain, then there exists a stationary distribution
More informationMath/Stats 425, Sec. 1, Fall 04: Introduction to Probability. Final Exam: Solutions
Math/Stats 45, Sec., Fall 4: Introduction to Probability Final Exam: Solutions. In a game, a contestant is shown two identical envelopes containing money. The contestant does not know how much money is
More information0, otherwise. U = Y 1 Y 2 Hint: Use either the method of distribution functions or a bivariate transformation. (b) Find E(U).
1. Suppose Y U(0, 2) so that the probability density function (pdf) of Y is 1 2, 0 < y < 2 (a) Find the pdf of U = Y 4 + 1. Make sure to note the support. (c) Suppose Y 1, Y 2,..., Y n is an iid sample
More information10.1 Generate n = 100 standard lognormal data points via the R commands
STAT 675 Statistical Computing Solutions to Homework Exercises Chapter 10 Note that some outputs may differ, depending on machine settings, generating seeds, random variate generation, etc. 10.1 Generate
More informationMathematical Methods for Computer Science
Mathematical Methods for Computer Science Computer Science Tripos, Part IB Michaelmas Term 2016/17 R.J. Gibbens Problem sheets for Probability methods William Gates Building 15 JJ Thomson Avenue Cambridge
More informationStatistical Methods in Particle Physics
Statistical Methods in Particle Physics Lecture 3 October 29, 2012 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2012 / 13 Outline Reminder: Probability density function Cumulative
More information57:022 Principles of Design II Final Exam Solutions - Spring 1997
57:022 Principles of Design II Final Exam Solutions - Spring 1997 Part: I II III IV V VI Total Possible Pts: 52 10 12 16 13 12 115 PART ONE Indicate "+" if True and "o" if False: + a. If a component's
More informationNumerical Analysis for Statisticians
Kenneth Lange Numerical Analysis for Statisticians Springer Contents Preface v 1 Recurrence Relations 1 1.1 Introduction 1 1.2 Binomial CoefRcients 1 1.3 Number of Partitions of a Set 2 1.4 Horner's Method
More informationSystematic uncertainties in statistical data analysis for particle physics. DESY Seminar Hamburg, 31 March, 2009
Systematic uncertainties in statistical data analysis for particle physics DESY Seminar Hamburg, 31 March, 2009 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan
More informationSC7/SM6 Bayes Methods HT18 Lecturer: Geoff Nicholls Lecture 2: Monte Carlo Methods Notes and Problem sheets are available at http://www.stats.ox.ac.uk/~nicholls/bayesmethods/ and via the MSc weblearn pages.
More informationToday. Next lecture. (Ch 14) Markov chains and hidden Markov models
Today (Ch 14) Markov chains and hidden Markov models Graphical representation Transition probability matrix Propagating state distributions The stationary distribution Next lecture (Ch 14) Markov chains
More informationSVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices)
Chapter 14 SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Today we continue the topic of low-dimensional approximation to datasets and matrices. Last time we saw the singular
More informationMarkov Processes Hamid R. Rabiee
Markov Processes Hamid R. Rabiee Overview Markov Property Markov Chains Definition Stationary Property Paths in Markov Chains Classification of States Steady States in MCs. 2 Markov Property A discrete
More informationP (A G) dp G P (A G)
First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume
More informationMath 308 Midterm Answers and Comments July 18, Part A. Short answer questions
Math 308 Midterm Answers and Comments July 18, 2011 Part A. Short answer questions (1) Compute the determinant of the matrix a 3 3 1 1 2. 1 a 3 The determinant is 2a 2 12. Comments: Everyone seemed to
More informationMath 180B Problem Set 3
Math 180B Problem Set 3 Problem 1. (Exercise 3.1.2) Solution. By the definition of conditional probabilities we have Pr{X 2 = 1, X 3 = 1 X 1 = 0} = Pr{X 3 = 1 X 2 = 1, X 1 = 0} Pr{X 2 = 1 X 1 = 0} = P
More informationSTA 294: Stochastic Processes & Bayesian Nonparametrics
MARKOV CHAINS AND CONVERGENCE CONCEPTS Markov chains are among the simplest stochastic processes, just one step beyond iid sequences of random variables. Traditionally they ve been used in modelling a
More informationEstimating a population mean
Introductory Statistics Lectures Estimating a population mean Confidence intervals for means Department of Mathematics Pima Community College Redistribution of this material is prohibited without written
More informationMA20226: STATISTICS 2A 2011/12 Assessed Coursework Sheet One. Set: Lecture 10 12:15-13:05, Thursday 3rd November, 2011 EB1.1
MA20226: STATISTICS 2A 2011/12 Assessed Coursework Sheet One Preamble Set: Lecture 10 12:15-13:05, Thursday 3rd November, 2011 EB1.1 Due: Please hand the work in to the coursework drop-box on level 1 of
More informationChapter 5: Integrals
Chapter 5: Integrals Section 5.3 The Fundamental Theorem of Calculus Sec. 5.3: The Fundamental Theorem of Calculus Fundamental Theorem of Calculus: Sec. 5.3: The Fundamental Theorem of Calculus Fundamental
More informationTest Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics
Test Code: STA/STB (Short Answer Type) 2013 Junior Research Fellowship for Research Course in Statistics The candidates for the research course in Statistics will have to take two shortanswer type tests
More informationCh. 7 Statistical Intervals Based on a Single Sample
Ch. 7 Statistical Intervals Based on a Single Sample Before discussing the topics in Ch. 7, we need to cover one important concept from Ch. 6. Standard error The standard error is the standard deviation
More informationDesigning Information Devices and Systems I Spring 2018 Homework 11
EECS 6A Designing Information Devices and Systems I Spring 28 Homework This homework is due April 8, 28, at 23:59. Self-grades are due April 2, 28, at 23:59. Submission Format Your homework submission
More informationBayesian Methods with Monte Carlo Markov Chains II
Bayesian Methods with Monte Carlo Markov Chains II Henry Horng-Shing Lu Institute of Statistics National Chiao Tung University hslu@stat.nctu.edu.tw http://tigpbp.iis.sinica.edu.tw/courses.htm 1 Part 3
More informationOrder Statistics and Distributions
Order Statistics and Distributions 1 Some Preliminary Comments and Ideas In this section we consider a random sample X 1, X 2,..., X n common continuous distribution function F and probability density
More informationSTA 2201/442 Assignment 2
STA 2201/442 Assignment 2 1. This is about how to simulate from a continuous univariate distribution. Let the random variable X have a continuous distribution with density f X (x) and cumulative distribution
More informationLecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora
princeton univ. F 13 cos 521: Advanced Algorithm Design Lecture 14: SVD, Power method, and Planted Graph problems (+ eigenvalues of random matrices) Lecturer: Sanjeev Arora Scribe: Today we continue the
More informationSTAT 418: Probability and Stochastic Processes
STAT 418: Probability and Stochastic Processes Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical
More information3 Random Samples from Normal Distributions
3 Random Samples from Normal Distributions Statistical theory for random samples drawn from normal distributions is very important, partly because a great deal is known about its various associated distributions
More informationProbability Distributions Columns (a) through (d)
Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)
More informationEpidemiology Principle of Biostatistics Chapter 11 - Inference about probability in a single population. John Koval
Epidemiology 9509 Principle of Biostatistics Chapter 11 - Inference about probability in a single population John Koval Department of Epidemiology and Biostatistics University of Western Ontario What is
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 08 SUBJECT NAME : Probability & Random Processes SUBJECT CODE : MA645 MATERIAL NAME : University Questions REGULATION : R03 UPDATED ON : November 07 (Upto N/D 07 Q.P) (Scan the
More informationSTAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed.
STAT 302 Introduction to Probability Learning Outcomes Textbook: A First Course in Probability by Sheldon Ross, 8 th ed. Chapter 1: Combinatorial Analysis Demonstrate the ability to solve combinatorial
More informationBayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017
Chalmers April 6, 2017 Bayesian philosophy Bayesian philosophy Bayesian statistics versus classical statistics: War or co-existence? Classical statistics: Models have variables and parameters; these are
More information