1 Solution to Problem 2.1
|
|
- Ella Bridges
- 5 years ago
- Views:
Transcription
1 Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto itself, so the support of the distribution) of Y is, ). f Y y) f ) X y /3 d Checking the integral, dy y/3 2y 5/3 y /3) 3 y 2/3 y y /3), < y <. y y /3) dy y y /3 ) dy [ 2 y2 3 ] 7 y7/3 [ 2 3 7] [ 7 6 ] [ ]. b) X Y 3)/ is -. The interval, ), which is the support of X, is mapped to 3, ) under Y X + 3, so the support of Y is 3, ). Computing the pdf for Y : d f Y y) f X y 3)/) y 3)/ dy Checking the integral, 3 y 7 exp 7y 3)/) 7 exp 7y 3)/), y > 3. 7 exp 7y 3)/) dy
2 , 7 exp 7x/) dx substitute x y 3) since the latter expression is the integral over its entire range of the exponential, mean /7, pdf. c) The function y gx) x 2 is - on x, ), and maps, ) onto, ), so the support of Y is, ). As X Y /2, Checking the integral, f Y y) f X y /2 ) d dy y/2 3y y /2) 2 2 y /2 5y /2 y /2) 2, < y <. 5y /2 y /2) 2 dy 5 5 y /2 2y /2 + y ) dy y /2 2y + y 3/2) dy [ y3/2 y ] 5 y5/2 [ ] + 2 [ ] y 2 Solution to Problem 2.2 We apply Theorem 2..5 in each case. 2
3 a) X Y /2 so dx /2)y /2 dy. Note that as x ranges over, ), x 2 also ranges over, ). f Y y) f ) X y /2 /2)y /2, < y <, /2)y /2, < y <. b) If Y log X, then X e Y and dx e y dy. As x ranges over, ), y ranges over, ), so we have f Y y) f X e y ) e y, y >, n + m + )! e ny e y) m e y, n!m! y >, n + m + )! e n+)y e y) m n!m! y >. c) If Y e X, then X log Y and dx y dy. Also, as x ranges over, ), y ranges over, ). Hence, f Y y) f X log y)y, y >, [ σ log y exp ] log 2 y)2 y, y >, 2σ2 [ log y exp ] log σ 2 y)2, y >. y 2σ2 3 Solution to Problem 2.9 The cdf for < x < 3 is Fx) x x y )/2 dy y/2 dy [ y 2 / ] x y x ) 2 /. 3
4 Thus, the cdf over the whole real line is if x <, Fx) x ) 2 / if x < 3, if 3 x. Note that Fx) is continuous at x as x ) 2 / at x, and also at x 3 as x ) 2 / at x 3. Thus, U X ) 2 / has a Unif,) distribution. Solution to Problem 2. a) Find E[X 2 ] directly must mean integrating against the N, ) pdf: x 2 e x2 /2 dx x xe x2 /2 ) dx. ) Applying integration by parts: udv uv vdu u x dv xe x2 /2 dx du dx v e x2 /2 we obtain x 2 e x2 /2 dx xe x2 /2, e x2 /2 dx + x e x2 /2 dx so we obtain E[X 2 ] after we put the / factor back in the pdf.
5 Now, using the pdf of Example 2..7: if Y X 2 then f Y y) 2 y [f X y) + f X y)] 2 y 2 e y/2 y /2 e y/2, y >. Now, E[Y ] y y /2 e y/2 dy y /2 e y/2 dy 2 y /2 e y/2 y + y /2 e y/2 dy, 2) where we have used integration by parts with u y /2 du 2 y /2 dy dv e y/2 dy v 2e y/2 Observe that the first term on the r.h.s. of 2) is, since y /2 at y and lim y y /2 e y/2. Also, the second term on the r.h.s. of 2) is since it is the integral of f Y y) over its whole range. b) The pdf of Y X is similar but easier) to the derivation in Example To give the details: assume y, F Y y) P[Y y] P[ X y] P[ y X y] P[ y < X y] since X is a cont. rv F X y) F X y). 5
6 Differentiating, we get f Y y) f X y) + f X y) 2f X y) since by symmetry f X y) f X y) recall X N, )). Thus, f Y y) 2 π e y2 /2, y >. We have E[Y ] 2 π 2 π 2 π. ye y2 /2 dy e w dw substitute w y 2 /2, dw ydy We have E[Y 2 ] E[X 2 ], so Var[Y ] E[Y 2 ] E[Y ]) 2 2/π Solution to Problem 2. a) There are various approaches to this problem under certain assumptions one can use integration by parts. However, the easiest approach by far is to use the following trick: Fx) P[X > x] x fy) dy. Plugging this into the expression on the r.h.s. of the proposed formula: Fx) dx x y 6 fy) dy dx fy) dxdy.
7 In the last expression, we simply switched the order of integration. Evaluating the double integral can be done by iterated integration either way: first over x or first over y. Note that we are integrating in the plane IR 2 ) over the region {x, y) : x y < }. Since fy) is a constant when integrating with respect to x, the last integral is yfy) dy E[X], since f ) is the pdf for X. b) A similar argument works here: so Fk) jk+ fj), [ Fk)] k fj) k jk+ j fj) j k jfj) j E[X]. 6 Solution to Problem 2.22 I guess I had too much spare time and prepared a solution for this exercise as well. a) Clear fx), as β >, x 2, and e x2 /2 >. Checking that it integrates to, β 3 π x2 x2 /β2 e dx 2 π y2 e y2 /2 dy 3) 2 y 2 e y2 /2 dy 7
8 y 2 e y2 /2 dy ). 5) where at 3), we substituted y 2x/β, at ) we used the symmetry of the N, ) pdf, and at 5) we used the fact that E[Y 2 ] when Y N, ), something we already proved back in Exercise 2.. b) Applying the same substitution as at 3), we have E[X] β 3 π x3 e x2 /β2 β π y3 e y2 /2 dy β y 2 e y2 /2 π 2β π y where at 6) we did an integration by parts with dx + 2β π ye y2 /2 dy 6) e w dw, 7) 2β π. 8) u y 2 dv ye y2 /2 dy du 2y dy v e y2 /2, and at 7) we substituted w y 2 /2. Proceeding on to the next moment, E[X 2 ] β 3 π x x2 /β2 e dx β 2 y e y2 /2 dy β2 y 3 e y2 /2 3β2 2 3β2 2. y y 2 e y2 /2 dy 8 + 3β2 y 2 e y2 /2 dy
9 Our integration by parts here was u y 3 dv ye y2 /2 dy du 3y 2 dy v e y2 /2 Thus, we have ) 2 Var[X] 3β2 2β 2 π [ 3 2 π] β β 2. 7 Solution to Problem 2.32 We have by the chain rule for differentiation, d d log Mt) dt Mt) dt Mt). Since M), we obtain that S ) M )/M) M ) E[X]. For the second derivative, we apply the rule for differentiating a quotient: so S 2) ) M)M ) M )) 2 M 2 ) S 2) t) d M t) dt Mt) Mt)M t) M t)) 2, M 2 t) M ) M )) 2 E[X 2 ] E[X]) 2 Var[X]. 8 Solution to Problem 2.33 a) Mgf for the Poissonλ): Mt) k e tkλk k! e λ e λ λe t ) k k! k e λ exp [ λe t] 9) exp [ λ e t )]. 9
10 At 9), we used the formula for the Taylor series for the exponential function. Since this converges everywhere, there are no restrictions on t, i.e. the mgf is finite everywhere. Using the trick from the previous problem, we have E[X] S ) d dt λ e t ) λe t t t λ, and Var[X] S ) d dt λet t λ. b) Mgf for the Geometricp): where we used the formula for summing a geometric series: Mt) e kt p) k p k p e t p) ) k k p e t p), ) k r k, r <. r The region for convergence r < translates into e t p) < t < log p). Note that log p) > for < p <, so the mgf is finite in a neighborhood of. Computing the moments: S t) d dt Substituting in t gives Also, [ log p log [ e t p) ]] e t p)/ [ e t p) ]. E[X] p)/ [ p)] p)/p. S t) d dt et p)/ [ e t p) ] d dt p)/ [ e t p) ] p)e t / [ e t p) ] 2,
11 and so evaluating this at t gives Var[X] p)/p 2. c) As noted in lecture, it is much easier to start with the N, ) case and then extend to Nµ, σ 2 ) using transformation formulae as in Theorem So, let Z N, ). Then M Z t) e t2 /2. Now, X σz + µ is Nµ, σ 2 ), and e tz e z2 /2 dz [ exp ] 2 z t)2 + t 2 /2 dz M X t) E[e tx ] E[exptσZ + tµ)] expµt)e[exptσz)] expµt)m Z σt) expµt) expσ 2 t 2 /2) expµt + σ 2 t 2 /2), as desired. Now the mean and variance computations are especially easy with the trick from Exercise 2.32: E[X] d µt + σ 2 t 2 /2 ) dt t and Var[X] d µ + σ 2 t ) dt µ + σ 2 t ) t µ, t σ 2.
BMIR Lecture Series on Probability and Statistics Fall 2015 Discrete RVs
Lecture #7 BMIR Lecture Series on Probability and Statistics Fall 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University 7.1 Function of Single Variable Theorem
More informationStatistics STAT:5100 (22S:193), Fall Sample Final Exam B
Statistics STAT:5 (22S:93), Fall 25 Sample Final Exam B Please write your answers in the exam books provided.. Let X, Y, and Y 2 be independent random variables with X N(µ X, σ 2 X ) and Y i N(µ Y, σ 2
More informationStat410 Probability and Statistics II (F16)
Stat4 Probability and Statistics II (F6 Exponential, Poisson and Gamma Suppose on average every /λ hours, a Stochastic train arrives at the Random station. Further we assume the waiting time between two
More informationLecture 5: Moment generating functions
Lecture 5: Moment generating functions Definition 2.3.6. The moment generating function (mgf) of a random variable X is { x e tx f M X (t) = E(e tx X (x) if X has a pmf ) = etx f X (x)dx if X has a pdf
More informationSTAT 3610: Review of Probability Distributions
STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random
More informationReview 1: STAT Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics. August 25, 2015
Review : STAT 36 Mark Carpenter, Ph.D. Professor of Statistics Department of Mathematics and Statistics August 25, 25 Support of a Random Variable The support of a random variable, which is usually denoted
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationMoments. Raw moment: February 25, 2014 Normalized / Standardized moment:
Moments Lecture 10: Central Limit Theorem and CDFs Sta230 / Mth 230 Colin Rundel Raw moment: Central moment: µ n = EX n ) µ n = E[X µ) 2 ] February 25, 2014 Normalized / Standardized moment: µ n σ n Sta230
More informationChapter 4. Chapter 4 sections
Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation
More informationExpectation, variance and moments
Expectation, variance and moments John Appleby Contents Expectation and variance Examples 3 Moments and the moment generating function 4 4 Examples of moment generating functions 5 5 Concluding remarks
More informationp. 6-1 Continuous Random Variables p. 6-2
Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability (>). Often, there is interest in random variables
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections Chapter 3 - continued 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions
More informationChapter 3 sections. SKIP: 3.10 Markov Chains. SKIP: pages Chapter 3 - continued
Chapter 3 sections 3.1 Random Variables and Discrete Distributions 3.2 Continuous Distributions 3.3 The Cumulative Distribution Function 3.4 Bivariate Distributions 3.5 Marginal Distributions 3.6 Conditional
More informationE[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =
Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of
More informationSTA 4321/5325 Solution to Homework 5 March 3, 2017
STA 4/55 Solution to Homework 5 March, 7. Suppose X is a RV with E(X and V (X 4. Find E(X +. By the formula, V (X E(X E (X E(X V (X + E (X. Therefore, in the current setting, E(X V (X + E (X 4 + 4 8. Therefore,
More informationStat 366 A1 (Fall 2006) Midterm Solutions (October 23) page 1
Stat 366 A1 Fall 6) Midterm Solutions October 3) page 1 1. The opening prices per share Y 1 and Y measured in dollars) of two similar stocks are independent random variables, each with a density function
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationSampling Distributions
Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationExpectation, Variance and Standard Deviation for Continuous Random Variables Class 6, Jeremy Orloff and Jonathan Bloom
Expectation, Variance and Standard Deviation for Continuous Random Variables Class 6, 8.5 Jeremy Orloff and Jonathan Bloom Learning Goals. Be able to compute and interpret expectation, variance, and standard
More informationSDS 321: Introduction to Probability and Statistics
SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationBrief Review of Probability
Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationTest Problems for Probability Theory ,
1 Test Problems for Probability Theory 01-06-16, 010-1-14 1. Write down the following probability density functions and compute their moment generating functions. (a) Binomial distribution with mean 30
More informationContinuous Distributions
A normal distribution and other density functions involving exponential forms play the most important role in probability and statistics. They are related in a certain way, as summarized in a diagram later
More informationStatistics 3657 : Moment Generating Functions
Statistics 3657 : Moment Generating Functions A useful tool for studying sums of independent random variables is generating functions. course we consider moment generating functions. In this Definition
More informationStat 100a, Introduction to Probability.
Stat 100a, Introduction to Probability. Outline for the day: 1. Geometric random variables. 2. Negative binomial random variables. 3. Moment generating functions. 4. Poisson random variables. 5. Continuous
More informationSampling Distributions
In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example,
More informationChapter 5. Random Variables (Continuous Case) 5.1 Basic definitions
Chapter 5 andom Variables (Continuous Case) So far, we have purposely limited our consideration to random variables whose ranges are countable, or discrete. The reason for that is that distributions on
More informationTopic 4: Continuous random variables
Topic 4: Continuous random variables Course 3, 216 Page Continuous random variables Definition (Continuous random variable): An r.v. X has a continuous distribution if there exists a non-negative function
More informationFunctions of Random Variables
Functions of Random Variables Statistics 110 Summer 2006 Copyright c 2006 by Mark E. Irwin Functions of Random Variables As we ve seen before, if X N(µ, σ 2 ), then Y = ax + b is also normally distributed.
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationProblem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},
ECE32 Spring 25 HW Solutions April 6, 25 Solutions to HW Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More informationBell-shaped curves, variance
November 7, 2017 Pop-in lunch on Wednesday Pop-in lunch tomorrow, November 8, at high noon. Please join our group at the Faculty Club for lunch. Means If X is a random variable with PDF equal to f (x),
More information1.6 Families of Distributions
Your text 1.6. FAMILIES OF DISTRIBUTIONS 15 F(x) 0.20 1.0 0.15 0.8 0.6 Density 0.10 cdf 0.4 0.05 0.2 0.00 a b c 0.0 x Figure 1.1: N(4.5, 2) Distribution Function and Cumulative Distribution Function for
More informationMore on Distribution Function
More on Distribution Function The distribution of a random variable X can be determined directly from its cumulative distribution function F X. Theorem: Let X be any random variable, with cumulative distribution
More informationContinuous Random Variables
Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability. Often, there is interest in random variables
More informationMathematical Statistics 1 Math A 6330
Mathematical Statistics 1 Math A 6330 Chapter 3 Common Families of Distributions Mohamed I. Riffi Department of Mathematics Islamic University of Gaza September 28, 2015 Outline 1 Subjects of Lecture 04
More informationSTA 256: Statistics and Probability I
Al Nosedal. University of Toronto. Fall 2017 My momma always said: Life was like a box of chocolates. You never know what you re gonna get. Forrest Gump. Exercise 4.1 Let X be a random variable with p(x)
More informationContinuous Distributions
Chapter 3 Continuous Distributions 3.1 Continuous-Type Data In Chapter 2, we discuss random variables whose space S contains a countable number of outcomes (i.e. of discrete type). In Chapter 3, we study
More information1.1 Review of Probability Theory
1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,
More informationUses of Asymptotic Distributions: In order to get distribution theory, we need to norm the random variable; we usually look at n 1=2 ( X n ).
1 Economics 620, Lecture 8a: Asymptotics II Uses of Asymptotic Distributions: Suppose X n! 0 in probability. (What can be said about the distribution of X n?) In order to get distribution theory, we need
More informationContinuous r.v practice problems
Continuous r.v practice problems SDS 321 Intro to Probability and Statistics 1. (2+2+1+1 6 pts) The annual rainfall (in inches) in a certain region is normally distributed with mean 4 and standard deviation
More information18.440: Lecture 19 Normal random variables
18.440 Lecture 19 18.440: Lecture 19 Normal random variables Scott Sheffield MIT Outline Tossing coins Normal random variables Special case of central limit theorem Outline Tossing coins Normal random
More information4 Moment generating functions
4 Moment generating functions Moment generating functions (mgf) are a very powerful computational tool. They make certain computations much shorter. However, they are only a computational tool. The mgf
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationi=1 k i=1 g i (Y )] = k
Math 483 EXAM 2 covers 2.4, 2.5, 2.7, 2.8, 3.1, 3.2, 3.3, 3.4, 3.8, 3.9, 4.1, 4.2, 4.3, 4.4, 4.5, 4.6, 4.9, 5.1, 5.2, and 5.3. The exam is on Thursday, Oct. 13. You are allowed THREE SHEETS OF NOTES and
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationMATH2715: Statistical Methods
MATH2715: Statistical Methods Exercises IV (based on lectures 7-8, work week 5, hand in lecture Mon 30 Oct) ALL questions count towards the continuous assessment for this module. Q1. If a random variable
More informationBASICS OF PROBABILITY
October 10, 2018 BASICS OF PROBABILITY Randomness, sample space and probability Probability is concerned with random experiments. That is, an experiment, the outcome of which cannot be predicted with certainty,
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More information1 Review of di erential calculus
Review of di erential calculus This chapter presents the main elements of di erential calculus needed in probability theory. Often, students taking a course on probability theory have problems with concepts
More informationThe story of the film so far... Mathematics for Informatics 4a. Jointly distributed continuous random variables. José Figueroa-O Farrill
The story of the film so far... Mathematics for Informatics 4a José Figueroa-O Farrill Lecture 2 2 March 22 X a c.r.v. with p.d.f. f and g : R R: then Y g(x is a random variable and E(Y g(f(d variance:
More informationThe Multivariate Normal Distribution 1
The Multivariate Normal Distribution 1 STA 302 Fall 2017 1 See last slide for copyright information. 1 / 40 Overview 1 Moment-generating Functions 2 Definition 3 Properties 4 χ 2 and t distributions 2
More informationMATH Solutions to Probability Exercises
MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe
More informationExponential Distribution and Poisson Process
Exponential Distribution and Poisson Process Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 215 Outline Introduction Exponential
More informationChapter 4. Continuous Random Variables 4.1 PDF
Chapter 4 Continuous Random Variables In this chapter we study continuous random variables. The linkage between continuous and discrete random variables is the cumulative distribution (CDF) which we will
More informationChapter 4 Multiple Random Variables
Review for the previous lecture Theorems and Examples: How to obtain the pmf (pdf) of U = g ( X Y 1 ) and V = g ( X Y) Chapter 4 Multiple Random Variables Chapter 43 Bivariate Transformations Continuous
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More informationSOLUTION FOR HOMEWORK 12, STAT 4351
SOLUTION FOR HOMEWORK 2, STAT 435 Welcome to your 2th homework. It looks like this is the last one! As usual, try to find mistakes and get extra points! Now let us look at your problems.. Problem 7.22.
More informationHW4 : Bivariate Distributions (1) Solutions
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 7 Néhémy Lim HW4 : Bivariate Distributions () Solutions Problem. The joint probability mass function of X and Y is given by the following table : X Y
More information1 Review of Probability
1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x
More informationContinuous random variables
Continuous random variables Continuous r.v. s take an uncountably infinite number of possible values. Examples: Heights of people Weights of apples Diameters of bolts Life lengths of light-bulbs We cannot
More informationWe introduce methods that are useful in:
Instructor: Shengyu Zhang Content Derived Distributions Covariance and Correlation Conditional Expectation and Variance Revisited Transforms Sum of a Random Number of Independent Random Variables more
More informationOrder Statistics. The order statistics of a set of random variables X 1, X 2,, X n are the same random variables arranged in increasing order.
Order Statistics The order statistics of a set of random variables 1, 2,, n are the same random variables arranged in increasing order. Denote by (1) = smallest of 1, 2,, n (2) = 2 nd smallest of 1, 2,,
More informationStatistics 100A Homework 5 Solutions
Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More information15 Discrete Distributions
Lecture Note 6 Special Distributions (Discrete and Continuous) MIT 4.30 Spring 006 Herman Bennett 5 Discrete Distributions We have already seen the binomial distribution and the uniform distribution. 5.
More informationReview: mostly probability and some statistics
Review: mostly probability and some statistics C2 1 Content robability (should know already) Axioms and properties Conditional probability and independence Law of Total probability and Bayes theorem Random
More informationMath 222 Spring 2013 Exam 3 Review Problem Answers
. (a) By the Chain ule, Math Spring 3 Exam 3 eview Problem Answers w s w x x s + w y y s (y xy)() + (xy x )( ) (( s + 4t) (s 3t)( s + 4t)) ((s 3t)( s + 4t) (s 3t) ) 8s 94st + 3t (b) By the Chain ule, w
More informationAnnouncements. Topics: Homework:
Announcements Topics: - sections 7.3 (the definite integral +area), 7.4 (FTC), 7.5 (additional techniques of integration) * Read these sections and study solved examples in your textbook! Homework: - review
More informationACM 116: Lectures 3 4
1 ACM 116: Lectures 3 4 Joint distributions The multivariate normal distribution Conditional distributions Independent random variables Conditional distributions and Monte Carlo: Rejection sampling Variance
More informationMultivariate distributions
CHAPTER Multivariate distributions.. Introduction We want to discuss collections of random variables (X, X,..., X n ), which are known as random vectors. In the discrete case, we can define the density
More informationECO227: Term Test 2 (Solutions and Marking Procedure)
ECO7: Term Test (Solutions and Marking Procedure) January 6, 9 Question 1 Random variables X and have the joint pdf f X, (x, y) e x y, x > and y > Determine whether or not X and are independent. [1 marks]
More informationMathematics 426 Robert Gross Homework 9 Answers
Mathematics 4 Robert Gross Homework 9 Answers. Suppose that X is a normal random variable with mean µ and standard deviation σ. Suppose that PX > 9 PX
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More information1, 0 r 1 f R (r) = 0, otherwise. 1 E(R 2 ) = r 2 f R (r)dr = r 2 3
STAT 5 4.43. We are given that a circle s radius R U(, ). the pdf of R is {, r f R (r), otherwise. The area of the circle is A πr. The mean of A is E(A) E(πR ) πe(r ). The second moment of R is ( ) r E(R
More informationMATH 56A SPRING 2008 STOCHASTIC PROCESSES 197
MATH 56A SPRING 8 STOCHASTIC PROCESSES 197 9.3. Itô s formula. First I stated the theorem. Then I did a simple example to make sure we understand what it says. Then I proved it. The key point is Lévy s
More informationStatistics 3657 : Moment Approximations
Statistics 3657 : Moment Approximations Preliminaries Suppose that we have a r.v. and that we wish to calculate the expectation of g) for some function g. Of course we could calculate it as Eg)) by the
More informationDiscrete Distributions
A simplest example of random experiment is a coin-tossing, formally called Bernoulli trial. It happens to be the case that many useful distributions are built upon this simplest form of experiment, whose
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More informationAPPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2
APPM/MATH 4/5520 Solutions to Exam I Review Problems. (a) f X (x ) f X,X 2 (x,x 2 )dx 2 x 2e x x 2 dx 2 2e 2x x was below x 2, but when marginalizing out x 2, we ran it over all values from 0 to and so
More informationIntegration by Parts
Calculus 2 Lia Vas Integration by Parts Using integration by parts one transforms an integral of a product of two functions into a simpler integral. Divide the initial function into two parts called u
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationLecture 14: Multivariate mgf s and chf s
Lecture 14: Multivariate mgf s and chf s Multivariate mgf and chf For an n-dimensional random vector X, its mgf is defined as M X (t) = E(e t X ), t R n and its chf is defined as φ X (t) = E(e ıt X ),
More information8 Laws of large numbers
8 Laws of large numbers 8.1 Introduction We first start with the idea of standardizing a random variable. Let X be a random variable with mean µ and variance σ 2. Then Z = (X µ)/σ will be a random variable
More informationMAT 122 Homework 7 Solutions
MAT 1 Homework 7 Solutions Section 3.3, Problem 4 For the function w = (t + 1) 100, we take the inside function to be z = t + 1 and the outside function to be z 100. The derivative of the inside function
More information1 Expectation of a continuously distributed random variable
OCTOBER 3, 204 LECTURE 9 EXPECTATION OF A CONTINUOUSLY DISTRIBUTED RANDOM VARIABLE, DISTRIBUTION FUNCTION AND CHANGE-OF-VARIABLE TECHNIQUES Expectation of a continuously distributed random variable Recall
More informationStochastic Processes and Monte-Carlo Methods. University of Massachusetts: Spring Luc Rey-Bellet
Stochastic Processes and Monte-Carlo Methods University of Massachusetts: Spring 2010 Luc Rey-Bellet Contents 1 Random variables and Monte-Carlo method 3 1.1 Review of probability............................
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationHW7 Solutions. f(x) = 0 otherwise. 0 otherwise. The density function looks like this: = 20 if x [10, 90) if x [90, 100]
HW7 Solutions. 5 pts.) James Bond James Bond, my favorite hero, has again jumped off a plane. The plane is traveling from from base A to base B, distance km apart. Now suppose the plane takes off from
More informationDIFFERENTIATION RULES
3 DIFFERENTIATION RULES DIFFERENTIATION RULES 3. The Product and Quotient Rules In this section, we will learn about: Formulas that enable us to differentiate new functions formed from old functions by
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationChapter 3, 4 Random Variables ENCS Probability and Stochastic Processes. Concordia University
Chapter 3, 4 Random Variables ENCS6161 - Probability and Stochastic Processes Concordia University ENCS6161 p.1/47 The Notion of a Random Variable A random variable X is a function that assigns a real
More informationStat 35, Introduction to Probability.
Stat 35, Introduction to Probability. Outline for the day: 1. Harman/Negreanu and running it twice. 2. Uniform random variables. 3. Exponential random variables. 4. Normal random variables. 5. Functions
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More information1 Acceptance-Rejection Method
Copyright c 2016 by Karl Sigman 1 Acceptance-Rejection Method As we already know, finding an explicit formula for F 1 (y), y [0, 1], for the cdf of a rv X we wish to generate, F (x) = P (X x), x R, is
More information