IEOR 4703: Homework 2 Solutions
|
|
- Cecily Short
- 6 years ago
- Views:
Transcription
1 IEOR 4703: Homework 2 Solutions Exercises for which no programming is required Let U be uniformly distributed on the interval (0, 1); P (U x) = x, x (0, 1). We assume that your computer can sequentially generate such U indepently upon demand. 1. Let X = (b a)u + a. Show that X is uniformly distributed on the interval (a, b); that is, show that P (X x) = (x a)/(b a), x (a, b). Thus, if your computer can generate a U, then it can in fact generate any rv X that is uniformly distributed on the interval (a, b) for any desired a and b. SOLUTION: If x (a, b), then for y = (x a)/(b a), it follows that y (0, 1) and thus P (U y) = y, that is, P (U (x a)/(b a)) = (x a)/(b a). We now use this fact: 2. Continuation: X has cdf given by P ((b a)u + a x) = P (U (x a)/(b a)) = (x a)/(b a). x a b a, if x (a, b); F (x) = 1, if x b; 0, if x a. Show that in fact setting X = (b a)u + a is just the inverse transform method. SOLUTION: It is easily sen that F 1 (y) = (b a)y + a, y (0, 1). 3. Let V = 1 U. Show that V is also uniformly distributed on the interval (0, 1). SOLUTION: We must show that P (V x) = x, x (0, 1). To this : P (V x) = P (1 U x) = P (U 1 x) = 1 (1 x) = x. 4. Recall that conditional probabilities are defined as: P (A B) = P (A B)/P (B), where A B denotes the intersection of the two events. Conditional on the event B = {U.5}, show that U is uniformly distributed on the smaller interval (0, 0.50). That is, P (U x U.50) = x/(.50), x (0,.50). More generally, conditional on the event B = {U (a, b)}, where 0 < a < b < 1, show that U is uniformly distributed on the smaller interval (a, b). P (U x U (a, b)) = (x a)/(b a), x (a, b). SOLUTION: With A = {U x} and B = {U.5}, and x (0,.50), we have A B = {U x}. Thus P (A B)/P (B) = P (U x)/(p (U.50)) = x/(.50). More generally, when B = {U (a, b)}, and A = {U x} where 0 < a < x < b < 1, we have A B = {U (a, x]} and thus P (A B)/P (B) = P (U (a, x])/(p (U (a, b))) = (x a)/(b a). 5. Give an algorithm that uses two uniforms U 1, U 2 for generating the following distribution: F (x) = 0.40(1 e 2x ) (1 e 2 x ), x 0. 1
2 SOLUTION: This is an example of a mixture, F (x) = pf 1 (x) + (1 p)f 2 (x) where p = 0.40, F 1 is exponential at rate 2 and F 2 is Weibull, with an easily found inverse F2 1 (y) = (1/4)[ln(1 y)] 2, y (0, 1). So we can use the composition method: (a) Generate U 1 and U 2. (b) If U 1 0.4, then set X = ( 1/2) ln(u 2 ); otherwise set X = (1/4)[ln(U 2 )] Let X 1,..., X n be iid rvs distributed as F (x) = P (X x), and assume that F 1 (y) is known in closed form. (a) Give an algorithm for generating a rv m = min{x 1, X 2,..., X n } that uses only ONE uniform U. SOLUTION: This is analogous to simulating the maximum, which we went thru in class lecture, same idea using the inverse transform method. Let G(x) = P (m x). Then P (m > x) = P (X 1 > x,..., X n > x) = (1 F (x)) n. Thus G(x) = 1 (1 F (x)) n. Thus solving y = G(x) for x in terms of y yields G 1 (y) = F 1 (1 (1 y) 1/n ). Since 1 U is itself uniform we obtain m = F 1 (1 (U) 1/n ) as our algorithm. (b) Give the algorithm in the special case when F (x) = 1 e λx, x 0; what kind of distribution does m have here? SOLUTION: In this case m exp(nλ), so m = (1/(nλ)) ln U. (c) Give the algorithm in the special case when F (x) = x, x (0, 1) is the uniform (0, 1) distribution; what kind of distribution does m have here? SOLUTION: F y = y, so we obtain X = (1 (U) 1/n ). G(x) = 1 (1 x) n, so taking the derivative yields the Beta density g(x) = n(1 x) n 1, x (0, 1). 7. Suppose that X 1 and X 2 are iid exponentials; P (X i x) = 1 e λx, x 0. Prove that U def = X 1 /(X 1 + X 2 ) is uniform on (0, 1) and is indepent of the sum X 1 + X 2. (Hint: consider the change of variables u = x 1 + x 2, v = x 1 /(x 1 + x 2 ); now find the probability density function g(u, v) using the fact that the joint density of X 1, X 2 is given by f(x 1, x 2 ) = λ 2 e λ(x 1+x 2 ). Jacobian, etc.) SOLUTION: Noting that the sum V = X 1 + X 2 is an Erlang distribution with density g(v) = λ 2 ve λv, v > 0, and recalling that the uniform density on (0, 1) is given by g(u) = 1, u (0, 1), it must be shown that the joint density of U = X 1 /(X 1 + X 2 ) and V = X 1 + X 2 is given by g(u, v) = 1 λ 2 ve λv, u (0, 1), v > 0. (The product of the two because of indepence.) We consider the function h(x 1, x 2 ) = (h 1 (x 1, x 2 ), h 2 (x 1, x 2 )), where h 1 (x 1, x 2 ) = u = x 1 /(x 1 + x 2 ) and h 2 (x 1, x 2 ) = v = x 1 + x 2 ). From general multidimensional calculus, changing variables in this manner involves computing the determinant of the 2 by 2 Jacobian matrix of partial derivatives ( ( h1 x 1 h 2 h 1 x 2 h 2 x 1 x ) = x 2 (x 1 +x 2 ) 2 x 1 (x 1 +x 2 ) 2 ). 2
3 The determinant is thus J(x 1, x 2 ) = (x 1 + x 2 )/(x 1 + x 2 ) 2 = 1/(x 1 + x 2 ) = v 1. The general formula for computing g is g(u, v) = J(x 1, x 2 ) 1 f(x 1, x 2 ) = vf(x 1, x 2 ) = yielding the result. vλ 2 e λ(x 1+x 2 ) = λ 2 e λv, Exercises for which programming is required 1. Program up the code (in MATLAB) for generating a Poisson rv X with mean α (below); then using α = 50, simulate via Monte Carlo simulation to estimate P (X = 50), the probability that X takes on its own mean as its value. (This is the Mode.) Generate, n = 1000 iid samples, X 1,..., X n and use the estimate 1 n n X i. i=1 Repeat for n = 5, 000 and then n = 10, 000.(The exact answer is of course e α α 50 /50!.) %This algorithm generates a Poisson rv with inputed mean alpha %User inputs alpha alpha = input( Value of alpha: ); X = 0; P = 1; u = rand; P = P*u; while (P >= exp(-alpha)) X u = rand; P = P*u; X= X+1; &This code estimates P(X=50) for a Poisson rv with mean alpha=50 alpha = input( Value of alpha: ); count = 0; %this variable keeps track of the number of times that X equals 50 over the ite 3
4 n = 10000; for i=1:n X = 0; P = 1; u = rand; P = P*u; while (P >= exp(-alpha)) u=rand; P=P*u; X=X+1; if ( X == 50) count = count+1; count / n 2. Simulate a compound Poisson process up to time T = 10, when λ = 5 and the iid jumps J are distributed as P (J = 1) = 0.30, P (J = 1) = 0.10, P (J = 4) = Use the following (but write up in MATLAB): (a) t = 0, N = 0, X = 0. (b) Generate a U (c) t = t + [ (1/λ) ln (U)]. If t > T, then stop. (d) Generate J distributed as G. (e) Set N = N + 1 and set X = X + J (f) Go back to 2. t=0; N=0; X=0; lambda=5; T=10; while (t<=t) t=t+(-1/lambda*log(rand)); Y=rand; 4
5 J=1*(Y<=0.3)-1*(Y>0.3)*(Y<=0.4)+4*(Y>0.4); N=N+1; X=X+J; X N 5
h(x) = f(x)/g(x) = (1 + x)e (1/2)x.
IEOR 4404 Homework 3 1. TEXT, Page 94, Exercise 20. (HINT: Consider g(x) = λe λx, for an appropriate value of λ.) We want to simulate from the density f(x) = 1 2 (1 + x)e x, x > 0. We can use an exponential
More information2 Random Variable Generation
2 Random Variable Generation Most Monte Carlo computations require, as a starting point, a sequence of i.i.d. random variables with given marginal distribution. We describe here some of the basic methods
More information1 Acceptance-Rejection Method
Copyright c 2016 by Karl Sigman 1 Acceptance-Rejection Method As we already know, finding an explicit formula for F 1 (y), y [0, 1], for the cdf of a rv X we wish to generate, F (x) = P (X x), x R, is
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More information1 Inverse Transform Method and some alternative algorithms
Copyright c 2016 by Karl Sigman 1 Inverse Transform Method and some alternative algorithms Assuming our computer can hand us, upon demand, iid copies of rvs that are uniformly distributed on (0, 1), it
More informationIEOR 4404: Homework 6
IEOR 4404: Homework 6. Consider the FIFO GI/GI/ queue as a model for a printer which prints one job at a time according to the distribution G(x) = P (S x) = e x, x 0 ( service time distribution ), and
More informationSTAT 430/510: Lecture 15
STAT 430/510: Lecture 15 James Piette June 23, 2010 Updates HW4 is up on my website. It is due next Mon. (June 28th). Starting today back at section 6.4... Conditional Distribution: Discrete Def: The conditional
More informationISyE 3044 Fall 2017 Test #1a Solutions
1 NAME ISyE 344 Fall 217 Test #1a Solutions This test is 75 minutes. You re allowed one cheat sheet. Good luck! 1. Suppose X has p.d.f. f(x) = 4x 3, < x < 1. Find E[ 2 X 2 3]. Solution: By LOTUS, we have
More informationSTAT 430/510 Probability
STAT 430/510 Probability Hui Nie Lecture 16 June 24th, 2009 Review Sum of Independent Normal Random Variables Sum of Independent Poisson Random Variables Sum of Independent Binomial Random Variables Conditional
More informationPROBABILITY DISTRIBUTIONS
Review of PROBABILITY DISTRIBUTIONS Hideaki Shimazaki, Ph.D. http://goo.gl/visng Poisson process 1 Probability distribution Probability that a (continuous) random variable X is in (x,x+dx). ( ) P x < X
More informationISyE 6644 Fall 2015 Test #1 Solutions (revised 10/5/16)
1 NAME ISyE 6644 Fall 2015 Test #1 Solutions (revised 10/5/16) You have 85 minutes. You get one cheat sheet. Put your succinct answers below. All questions are 3 points, unless indicated. You get 1 point
More informationS6880 #7. Generate Non-uniform Random Number #1
S6880 #7 Generate Non-uniform Random Number #1 Outline 1 Inversion Method Inversion Method Examples Application to Discrete Distributions Using Inversion Method 2 Composition Method Composition Method
More informationISyE 3044 Fall 2015 Test #1 Solutions
1 NAME ISyE 3044 Fall 2015 Test #1 Solutions You have 85 minutes. You get one cheat sheet. Put your succinct answers below. All questions are 3 points, unless indicated. You get 1 point for writing your
More informationOrdered Sample Generation
Ordered Sample Generation Xuebo Yu November 20, 2010 1 Introduction There are numerous distributional problems involving order statistics that can not be treated analytically and need to simulated through
More informationGenerating Random Variates 2 (Chapter 8, Law)
B. Maddah ENMG 6 Simulation /5/08 Generating Random Variates (Chapter 8, Law) Generating random variates from U(a, b) Recall that a random X which is uniformly distributed on interval [a, b], X ~ U(a,
More informationSlides 5: Random Number Extensions
Slides 5: Random Number Extensions We previously considered a few examples of simulating real processes. In order to mimic real randomness of events such as arrival times we considered the use of random
More informationOrder Statistics and Distributions
Order Statistics and Distributions 1 Some Preliminary Comments and Ideas In this section we consider a random sample X 1, X 2,..., X n common continuous distribution function F and probability density
More informationBMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs
Lecture #7 BMIR Lecture Series on Probability and Statistics Fall 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University 7.1 Function of Single Variable Theorem
More informationLecture 21: Convergence of transformations and generating a random variable
Lecture 21: Convergence of transformations and generating a random variable If Z n converges to Z in some sense, we often need to check whether h(z n ) converges to h(z ) in the same sense. Continuous
More information2. The CDF Technique. 1. Introduction. f X ( ).
Week 5: Distributions of Function of Random Variables. Introduction Suppose X,X 2,..., X n are n random variables. In this chapter, we develop techniques that may be used to find the distribution of functions
More informationMTH 202 : Probability and Statistics
MTH 202 : Probability and Statistics Homework 3 3th January, 207 ( Any point in the interval [0, can be represented by its decimal expansion 0.x x 2.... Suppose a point is chosen at random from the interval
More informationSTAT 801: Mathematical Statistics. Distribution Theory
STAT 81: Mathematical Statistics Distribution Theory Basic Problem: Start with assumptions about f or CDF of random vector X (X 1,..., X p ). Define Y g(x 1,..., X p ) to be some function of X (usually
More informationLecture 3 Continuous Random Variable
Lecture 3 Continuous Random Variable 1 Cumulative Distribution Function Definition Theorem 3.1 For any random variable X, 2 Continuous Random Variable Definition 3 Example Suppose we have a wheel of circumference
More information5.6 Logarithmic and Exponential Equations
SECTION 5.6 Logarithmic and Exponential Equations 305 5.6 Logarithmic and Exponential Equations PREPARING FOR THIS SECTION Before getting started, review the following: Solving Equations Using a Graphing
More information2.6 Logarithmic Functions. Inverse Functions. Question: What is the relationship between f(x) = x 2 and g(x) = x?
Inverse Functions Question: What is the relationship between f(x) = x 3 and g(x) = 3 x? Question: What is the relationship between f(x) = x 2 and g(x) = x? Definition (One-to-One Function) A function f
More informationClosed book and notes. 60 minutes. Cover page and four pages of exam. No calculators.
IE 230 Seat # Closed book and notes. 60 minutes. Cover page and four pages of exam. No calculators. Score Exam #3a, Spring 2002 Schmeiser Closed book and notes. 60 minutes. 1. True or false. (for each,
More information1 Exercises for lecture 1
1 Exercises for lecture 1 Exercise 1 a) Show that if F is symmetric with respect to µ, and E( X )
More informationSampling Random Variables
Sampling Random Variables Introduction Sampling a random variable X means generating a domain value x X in such a way that the probability of generating x is in accordance with p(x) (respectively, f(x)),
More informationUCSD ECE 153 Handout #20 Prof. Young-Han Kim Thursday, April 24, Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei)
UCSD ECE 53 Handout #0 Prof. Young-Han Kim Thursday, April 4, 04 Solutions to Homework Set #3 (Prepared by TA Fatemeh Arbabjolfaei). Time until the n-th arrival. Let the random variable N(t) be the number
More informationNonuniform Random Variate Generation
Nonuniform Random Variate Generation 1 Suppose we have a generator of i.i.d. U(0, 1) r.v. s. We want to generate r.v. s from other distributions, such as normal, Weibull, Poisson, etc. Or other random
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationRandom Number Generators
1/18 Random Number Generators Professor Karl Sigman Columbia University Department of IEOR New York City USA 2/18 Introduction Your computer generates" numbers U 1, U 2, U 3,... that are considered independent
More information1 Probability and Random Variables
1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationQualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama
Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationStat 451 Lecture Notes Simulating Random Variables
Stat 451 Lecture Notes 05 12 Simulating Random Variables Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 22 in Lange, and Chapter 2 in Robert & Casella 2 Updated:
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationPoisson Processes. Particles arriving over time at a particle detector. Several ways to describe most common model.
Poisson Processes Particles arriving over time at a particle detector. Several ways to describe most common model. Approach 1: a) numbers of particles arriving in an interval has Poisson distribution,
More informationLecture 3. David Aldous. 31 August David Aldous Lecture 3
Lecture 3 David Aldous 31 August 2015 This size-bias effect occurs in other contexts, such as class size. If a small Department offers two courses, with enrollments 90 and 10, then average class (faculty
More informationSolutions to Homework Set #3 (Prepared by Yu Xiang) Let the random variable Y be the time to get the n-th packet. Find the pdf of Y.
Solutions to Homework Set #3 (Prepared by Yu Xiang). Time until the n-th arrival. Let the random variable N(t) be the number of packets arriving during time (0,t]. Suppose N(t) is Poisson with pmf p N
More informationStat 426 : Homework 1.
Stat 426 : Homework 1. Moulinath Banerjee University of Michigan Announcement: The homework carries 120 points and contributes 10 points to the total grade. (1) A geometric random variable W takes values
More informationMAT 4872 Midterm Review Spring 2007 Prof. S. Singh
MAT 487 Midterm Review Spring 007 Prof S Singh Answer all questions carefully and give references for material used from journals Sketch all state spaces for all transformations You are required to work
More informationSTAT 380 Markov Chains
STAT 380 Markov Chains Richard Lockhart Simon Fraser University Spring 2016 Richard Lockhart (Simon Fraser University) STAT 380 Markov Chains Spring 2016 1 / 38 1/41 PoissonProcesses.pdf (#2) Poisson Processes
More informationIntro to Probability Instructor: Alexandre Bouchard
www.stat.ubc.ca/~bouchard/courses/stat302-sp2017-18/ Intro to Probability Instructor: Alexandre Bouchard Info on midterm CALCULATOR: only NON-programmable, NON-scientific, NON-graphing (and of course,
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationIEOR 3106: Introduction to Operations Research: Stochastic Models. Professor Whitt. SOLUTIONS to Homework Assignment 2
IEOR 316: Introduction to Operations Research: Stochastic Models Professor Whitt SOLUTIONS to Homework Assignment 2 More Probability Review: In the Ross textbook, Introduction to Probability Models, read
More informationWinter 2019 Math 106 Topics in Applied Mathematics. Lecture 8: Importance Sampling
Winter 2019 Math 106 Topics in Applied Mathematics Data-driven Uncertainty Quantification Yoonsang Lee (yoonsang.lee@dartmouth.edu) Lecture 8: Importance Sampling 8.1 Importance Sampling Importance sampling
More informationMath Stochastic Processes & Simulation. Davar Khoshnevisan University of Utah
Math 5040 1 Stochastic Processes & Simulation Davar Khoshnevisan University of Utah Module 1 Generation of Discrete Random Variables Just about every programming language and environment has a randomnumber
More information, find P(X = 2 or 3) et) 5. )px (1 p) n x x = 0, 1, 2,..., n. 0 elsewhere = 40
Assignment 4 Fall 07. Exercise 3.. on Page 46: If the mgf of a rom variable X is ( 3 + 3 et) 5, find P(X or 3). Since the M(t) of X is ( 3 + 3 et) 5, X has a binomial distribution with n 5, p 3. The probability
More informationLecture 7: Optimal Smoothing
Department of Biomedical Engineering and Computational Science Aalto University March 17, 2011 Contents 1 What is Optimal Smoothing? 2 Bayesian Optimal Smoothing Equations 3 Rauch-Tung-Striebel Smoother
More informationQuestion 1. Question Solution to final homework!! Friday 5 May Please report any typos immediately for a timely response.
1 Please report any typos immediately for a timely response. Call EW tn W tn 1 near u n 1,...,..., W t1 near u 1 ) LHS for short. Add the quantity W tn 1 W tn 1 into the expected value of the LHS. Apply
More informationApplied Statistics and Probability for Engineers. Sixth Edition. Chapter 4 Continuous Random Variables and Probability Distributions.
Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 4 Continuous Random Variables and Probability Distributions 4 Continuous CHAPTER OUTLINE Random
More informationChapter 4 Continuous Random Variables and Probability Distributions
Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 4 Continuous Random Variables and Probability Distributions 4 Continuous CHAPTER OUTLINE 4-1
More informationf X (x) = λe λx, , x 0, k 0, λ > 0 Γ (k) f X (u)f X (z u)du
11 COLLECTED PROBLEMS Do the following problems for coursework 1. Problems 11.4 and 11.5 constitute one exercise leading you through the basic ruin arguments. 2. Problems 11.1 through to 11.13 but excluding
More informationCh3. Generating Random Variates with Non-Uniform Distributions
ST4231, Semester I, 2003-2004 Ch3. Generating Random Variates with Non-Uniform Distributions This chapter mainly focuses on methods for generating non-uniform random numbers based on the built-in standard
More informationDecision making and problem solving Lecture 1. Review of basic probability Monte Carlo simulation
Decision making and problem solving Lecture 1 Review of basic probability Monte Carlo simulation Why probabilities? Most decisions involve uncertainties Probability theory provides a rigorous framework
More informationToday. Calculus. Linear Regression. Lagrange Multipliers
Today Calculus Lagrange Multipliers Linear Regression 1 Optimization with constraints What if I want to constrain the parameters of the model. The mean is less than 10 Find the best likelihood, subject
More informationThree hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.
Three hours To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER EXTREME VALUES AND FINANCIAL RISK Examiner: Answer QUESTION 1, QUESTION
More informationSTAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS
STAT/MA 4 Answers Homework November 5, 27 Solutions by Mark Daniel Ward PROBLEMS Chapter Problems 2a. The mass p, corresponds to neither of the first two balls being white, so p, 8 7 4/39. The mass p,
More informationConditional densities, mass functions, and expectations
Conditional densities, mass functions, and expectations Jason Swanson April 22, 27 1 Discrete random variables Suppose that X is a discrete random variable with range {x 1, x 2, x 3,...}, and that Y is
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional
More informationChapter Learning Objectives. Probability Distributions and Probability Density Functions. Continuous Random Variables
Chapter 4: Continuous Random Variables and Probability s 4-1 Continuous Random Variables 4-2 Probability s and Probability Density Functions 4-3 Cumulative Functions 4-4 Mean and Variance of a Continuous
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationISyE 6644 Fall 2016 Test #1 Solutions
1 NAME ISyE 6644 Fall 2016 Test #1 Solutions This test is 85 minutes. You re allowed one cheat sheet. Good luck! 1. Suppose X has p.d.f. f(x) = 3x 2, 0 < x < 1. Find E[3X + 2]. Solution: E[X] = 1 0 x 3x2
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder
More informationSTAT 450: Statistical Theory. Distribution Theory. Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6.
STAT 450: Statistical Theory Distribution Theory Reading in Casella and Berger: Ch 2 Sec 1, Ch 4 Sec 1, Ch 4 Sec 6. Example: Why does t-statistic have t distribution? Ingredients: Sample X 1,...,X n from
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationChapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS
Chapter 4: CONTINUOUS RANDOM VARIABLES AND PROBABILITY DISTRIBUTIONS Part 4: Gamma Distribution Weibull Distribution Lognormal Distribution Sections 4-9 through 4-11 Another exponential distribution example
More informationSUFFICIENT STATISTICS
SUFFICIENT STATISTICS. Introduction Let X (X,..., X n ) be a random sample from f θ, where θ Θ is unknown. We are interested using X to estimate θ. In the simple case where X i Bern(p), we found that the
More informationChange Of Variable Theorem: Multiple Dimensions
Change Of Variable Theorem: Multiple Dimensions Moulinath Banerjee University of Michigan August 30, 01 Let (X, Y ) be a two-dimensional continuous random vector. Thus P (X = x, Y = y) = 0 for all (x,
More informationIE 230 Probability & Statistics in Engineering I. Closed book and notes. 60 minutes.
Closed book and notes. 60 minutes. A summary table of some univariate continuous distributions is provided. Four Pages. In this version of the Key, I try to be more complete than necessary to receive full
More informationNumerical Methods I Monte Carlo Methods
Numerical Methods I Monte Carlo Methods Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 Course G63.2010.001 / G22.2420-001, Fall 2010 Dec. 9th, 2010 A. Donev (Courant Institute) Lecture
More informationGaussian random variables inr n
Gaussian vectors Lecture 5 Gaussian random variables inr n One-dimensional case One-dimensional Gaussian density with mean and standard deviation (called N, ): fx x exp. Proposition If X N,, then ax b
More informationGenerating pseudo- random numbers
Generating pseudo- random numbers What are pseudo-random numbers? Numbers exhibiting statistical randomness while being generated by a deterministic process. Easier to generate than true random numbers?
More informationSTATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero
STATISTICAL METHODS FOR SIGNAL PROCESSING c Alfred Hero 1999 32 Statistic used Meaning in plain english Reduction ratio T (X) [X 1,..., X n ] T, entire data sample RR 1 T (X) [X (1),..., X (n) ] T, rank
More informationRecursive Estimation
Recursive Estimation Raffaello D Andrea Spring 08 Problem Set 3: Extracting Estimates from Probability Distributions Last updated: April 9, 08 Notes: Notation: Unless otherwise noted, x, y, and z denote
More informationIE 303 Discrete-Event Simulation
IE 303 Discrete-Event Simulation 1 L E C T U R E 5 : P R O B A B I L I T Y R E V I E W Review of the Last Lecture Random Variables Probability Density (Mass) Functions Cumulative Density Function Discrete
More information1 Simulating normal (Gaussian) rvs with applications to simulating Brownian motion and geometric Brownian motion in one and two dimensions
Copyright c 2007 by Karl Sigman 1 Simulating normal Gaussian rvs with applications to simulating Brownian motion and geometric Brownian motion in one and two dimensions Fundamental to many applications
More informationChapter 4. Continuous Random Variables 4.1 PDF
Chapter 4 Continuous Random Variables In this chapter we study continuous random variables. The linkage between continuous and discrete random variables is the cumulative distribution (CDF) which we will
More informationPreliminaries Lectures. Dr. Abdulla Eid. Department of Mathematics MATHS 101: Calculus I
Preliminaries 2 1 2 Lectures Department of Mathematics http://www.abdullaeid.net/maths101 MATHS 101: Calculus I (University of Bahrain) Prelim 1 / 35 Pre Calculus MATHS 101: Calculus MATHS 101 is all about
More informationTwo hours. To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER.
Two hours MATH38181 To be supplied by the Examinations Office: Mathematical Formula Tables and Statistical Tables THE UNIVERSITY OF MANCHESTER EXTREME VALUES AND FINANCIAL RISK Examiner: Answer any FOUR
More informationComments. x > w = w > x. Clarification: this course is about getting you to be able to think as a machine learning expert
Logistic regression Comments Mini-review and feedback These are equivalent: x > w = w > x Clarification: this course is about getting you to be able to think as a machine learning expert There has to be
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More informationTheory of Statistical Tests
Ch 9. Theory of Statistical Tests 9.1 Certain Best Tests How to construct good testing. For simple hypothesis H 0 : θ = θ, H 1 : θ = θ, Page 1 of 100 where Θ = {θ, θ } 1. Define the best test for H 0 H
More informationContinuous Probability Distributions. Uniform Distribution
Continuous Probability Distributions Uniform Distribution Important Terms & Concepts Learned Probability Mass Function (PMF) Cumulative Distribution Function (CDF) Complementary Cumulative Distribution
More informationECE 510 Lecture 7 Goodness of Fit, Maximum Likelihood. Scott Johnson Glenn Shirley
ECE 510 Lecture 7 Goodness of Fit, Maximum Likelihood Scott Johnson Glenn Shirley Confidence Limits 30 Jan 2013 ECE 510 S.C.Johnson, C.G.Shirley 2 Binomial Confidence Limits (Solution 6.2) UCL: Prob of
More informationMath 3338, Homework 4. f(u)du. (Note that f is discontinuous, whereas F is continuous everywhere.)
Math 3338, Homework 4. Define a function f by f(x) = { if x < if x < if x. Calculate an explicit formula for F(x) x f(u)du. (Note that f is discontinuous, whereas F is continuous everywhere.) 2. Define
More informationContinuous Random Variables
Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability. Often, there is interest in random variables
More information1 Probability Model. 1.1 Types of models to be discussed in the course
Sufficiency January 11, 2016 Debdeep Pati 1 Probability Model Model: A family of distributions {P θ : θ Θ}. P θ (B) is the probability of the event B when the parameter takes the value θ. P θ is described
More informationRandom variables. DS GA 1002 Probability and Statistics for Data Science.
Random variables DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Motivation Random variables model numerical quantities
More informationNotes 9 : Infinitely divisible and stable laws
Notes 9 : Infinitely divisible and stable laws Math 733 - Fall 203 Lecturer: Sebastien Roch References: [Dur0, Section 3.7, 3.8], [Shi96, Section III.6]. Infinitely divisible distributions Recall: EX 9.
More informationStatistics 3657 : Moment Approximations
Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the
More informationHW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]
HW7 Solutions. 5 pts.) James Bond James Bond, my favorite hero, has again jumped off a plane. The plane is traveling from from base A to base B, distance km apart. Now suppose the plane takes off from
More informationMTH739U/P: Topics in Scientific Computing Autumn 2016 Week 6
MTH739U/P: Topics in Scientific Computing Autumn 16 Week 6 4.5 Generic algorithms for non-uniform variates We have seen that sampling from a uniform distribution in [, 1] is a relatively straightforward
More informationfor every x in the gomain of g
Section.7 Definition of Inverse Function Let f and g be two functions such that f(g(x)) = x for every x in the gomain of g and g(f(x)) = x for every x in the gomain of f Under these conditions, the function
More informationGeneration from simple discrete distributions
S-38.3148 Simulation of data networks / Generation of random variables 1(18) Generation from simple discrete distributions Note! This is just a more clear and readable version of the same slide that was
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 17: Continuous random variables: conditional PDF Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin
More information