THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA
|
|
- Grace Byrd
- 5 years ago
- Views:
Transcription
1 THE ROYAL STATISTICAL SOCIETY 4 EXAINATIONS SOLUTIONS GRADUATE DIPLOA PAPER I STATISTICAL THEORY & ETHODS The Societ provides these solutions to assist candidates preparing for the examinations in future ears and for the information of an other persons using the examinations The solutions should NOT be seen as "model answers" Rather, the have been written out in considerable detail and are intended as learning aids Users of the solutions should alwas be aware that in man cases there are valid alternative methods Also, in the man cases where discussion is called for, there ma be other valid points that could be made While ever care has been taken with the preparation of these solutions, the Societ will not be responsible for an errors or omissions The Societ will not enter into an correspondence in respect of these solutions RSS 4
2 Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question X and Y are independent, each with probabilit function given b P(W = w) = ( θ) w θ, for w =,, and where < θ < (i) Let U = X + Y U ma take the values, 3, B the independence of X and Y, u P ( U = u) = P( X = x and Y = u x) for u =, 3, x= Hence u { θ θ} {( θ) θ} θ ( θ) θ ( θ) u x u x u u, PU= u = = = u x= x= for u =, 3, [this is a negative binomial distribution] (ii) If U (= X +Y) = u, X must take values in the range,,, u Then ( = and = ) PU ( = u) P X x Y u x P X = xu = u = x {( ) } ( ) ( u ) θ ( θ) u x { } θ θ θ θ = = u u (for x =,,, u ) This is a discrete uniform distribution on the integers,,, u (iii) Suppose And takes X attempts and Bob takes Y attempts Then X and Y are independent geometric random variables with θ = 4 Thus from part (i), P(X + Y = 6) = (6 )(4) (6) 4 = 37 From part (ii), P(X < Y X + Y = 6) = P(X = or X + Y = 6) = = 5
3 Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question (i) (a) We have P(U = u) = /m, for u =,,, m So EU [ ] upu ( m) Also, m m+ m + = = = = m m ( + )( + ) ( + )( + ) m m m m m E U = u P( u) = ( m ) = = m m 6 6 and therefore Var(U) = E[U ] {E[U]} m + m + m+ m+ m m m ( 4m 3m 3) + = = + = = 6 m + (b) V = U + k, so E [ V] = + k and Var ( V ) m = (ii) Given that Y =, the first rolls must have been 6s and the final roll not a 6 Thus X is a discrete uniform random variable with, using the notation of part (i)(b), 5+ 5 k = 6( ) and m = 5 Hence E[ X] = + 6 ( ) = 6 3 and Var ( X ) = = These are the values conditional on Y = 6 5/6 5 Now, Y is a geometric random variable with EY [ ] = = and Var ( Y ) 6 = = E[ X] = E E( X Y) = E[ 6Y 3] = 6E[ Y] 3 = 3 = 5 5 We also need [ ] E Var X Y = E = and ( E X Y ) ( Y ) ( Y) Var = Var 6 3 = 36 Var = 6 / 5, from which (using the result quoted in the question) 6 66 Var ( X ) = + = 5 5
4 Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question 3 (i) θ u e θ E U = u du = u e du Γ Γ α α θu α m m α+ m θu [put t = θu] ( α) ( α) α m ( α ) θ Γ( a) ( α m) α α θ α + m t θ Γ + = t e dt = + α+ m Γ θ Hence EU [ ] = α Also θ ( + ) = and so Var(U) = E[U ] {E[U]} = E U α α θ α θ (ii) e x f X x = 8xe d = 8x = 4xe = x (for x > ) This is gamma with α = and θ = = = = fy xe dx e x e (for > ) x This is gamma with α = 3 and θ = Therefore from part (i), E[X] = and Var(X) = /, and E[Y] = 3/ and Var(Y) = 3/4 [ ] 8 ( 8 = x= ) 4 E XY x xe dxd e x dx d = = = e d [put t = ] 4 8 t t dt 4! e 3 6 = = Γ 5 = = ( XY) E( X) EY Cov, = = =, and the correlation between X and Y is ( XY) ( X) Var ( Y) 3 4 Cov, = = Var 3
5 Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question 4 X and Y are uniform(, ) U = X Y, V = X + Y So X = ½(U + V) and Y = ½(V U) (i) The limits of U are (, ) and of V are (, ) The diagram shows the square region where f (u, v) v u The Jacobian of the transformation is J x u u x v v = = = Therefore, for u, v in the region shown, and using the independence of X and Y, f u, v = J f x, = f x f = = (ii) For u, For u, For v, For v, + u u f u = dv= + u u u f u = dv= u f v = du = v v v v f v = du = v v (f (u) and f (v) are zero elsewhere) These distributions are identical except for a shift of one unit in the x-direction (iii) We now have W = a + tx and Z = b + ty (t > ) So W + Z = (a + b) + t(x + Y) = (a + b) + tv where V is as in parts (i) and (ii) So, using the result for V in part (ii), t t / / P W + Z < a+ b+ = P tv < = P V < = vdv= v = 8
6 Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question 5 (i) The method is to complete the square in the exponent in the integral and then recognise that this re-creates the pdf of a Normal distribution so that the value of the integral is simpl tx tx x µ X () t = E e = e exp dx Thus the exponent is σ π σ { } x µ tx = ( x µ x + µ σ tx) = x ( µ + σ t) µσ t σ t σ σ σ 4 x µ σ t µ t+ σ t + Hence X () t = e exp dx = exp µ t σ t + σ π σ tax ( b) bt ( atx ) bt (ii) ax b () t E e + + = e E e = = e X at Taking a = /σ and b = µ/σ we have Z, and so µ σ µ t σ t Z () t = e + = e σ σ t/ t / exp Putting µ = and σ = in the result of part (i), we see that this is the moment generating function of the standard Normal distribution Therefore (assuming without proof the uniqueness propert between distributions and their moment generating functions) Z follows the standard Normal distribution (iii) We expand Z (t) as a power series in t and use the result that the rth moment of Z about the origin, E[Z r ], is the coefficient of t r /r! in this expansion [Alternativel, use the result that it is given b the rth derivative evaluated at t = ] () ( ) ( ) 3 t t 4 6 t t t Z t = + t = ! 3! 8 48 We need E[Z ] and E[Z 4 ], which we see are and 3 respectivel So Var(Z ) = E[Z 4 ] {E[Z ]} = 3 =
7 Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question 6 (i) F( x) = P( X x) = P( X x bo) P( bo) + P( X x girl) P( girl) = { } F x + F x d = = + dx Hence the pdf of X is f ( x) F( x) { f( x) f( x) } ( ) ( ) E X = xf x dx= xf x dx+ xf x dx = µ + µ (ii) [ ] ( ) = = + E X x f x dx x f x dx x f x dx [now use σ = E[X ] {E[X]} ] {( σ ) } µ ( σ µ σ µ µ ) = = + + X = σ + µ + µ µ + µ = σ + µ + µ µ µ 4 4 and thus ( ) Var ( ) = σ + ( µ ) µ 4 (iii) As the sample is large, we ma use the central limit theorem to sa that the distribution of X is approximatel Normal, whatever the distributions f and f r The mean of X is equal to the mean of X, ie r µ + µ Var n n σ 4 µ µ ( X ) r = Var X = + (iv) Let the (independent) sample means for bos and girls be X and X Their means are µ and Xst = X+ X has mean µ, and their variances are each σ /n So σ σ σ ( µ + µ ) and variance + = Thus both X r and 4 n n n the latter has smaller variance; it is more precise X st are unbiased but
8 Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question 7 (i) (a) The distribution of X is x 3 P(x) 64/5 48/5 /5 /5 cdf F(x) < 5 and so corresponds to x = 6789 is between 5 and 896 and so corresponds to x = 355 < 5 and so corresponds to x = 898 < 5 and so corresponds to x = So the sample is,,, (b) x 4 4 x 4 3 t / / / t x 4 / F x = t e dt = e = e Thus the inverse cdf method gives u = e x, where u is the random number, so x= log( u) /4 u u log( u) [ log( u)] / These are the required random numbers (ii) Inter-arrival times, X, are exponential with parameter ½ Given u (as above), the inverse cdf method leads to u = e x/, ie x = log( u) [as is tabulated in column 3 above] First arrival time is at 8963 Service starts immediatel and lasts (uniform distribution on (5, 5)) = 789, and thus ends at = 375 Second arrival time is at = 7739 Service cannot start until 375 and then lasts = 7898, and thus ends at = 4865 Expressing the simulated results to the nearest minute after 9 am, the first customer arrives at 9 and leaves at 3, the second arrives at 8 and leaves at 49
9 (i) Graduate Diploma, Statistical Theor & ethods, Paper I, 4 Question 8 Consider the number of balls in urn A, an integer from to Define the p = P j balls in urn A at step n i balls in urn A at step n transition probabilit ij Then we have p =, p = for j ; j i i for i =,,,, pii =, pii+ =, pij = for j i or i+ ; p =, p = for j j (ii) Let the stationar probabilities be π, π,, π We have π = p π Hence j ij i i= j+ j+ π = π; for j =,,,, π j = π j + π j+ ; π = π It is required to show that π j = j ( j =,,, ) satisf these equations Immediatel, = = and π π π / π = = / Thus π = π and π = π, as required To confirm that the general equation is satisfied, consider j+ j+ j+! j+! + = + j j+ j! j+! j+! j! ( j j )!!!! = + = + = = j! j! j! j! j! j! j! j! j ultipling b (/), we see that the general equation ( j =,,, ) is satisfied (iii) Let X be the number of balls in urn A The given probabilities form a binomial distribution with parameters (6, ½) So E(X) = 6 ½ = 3 and Var(X) = 6 ½ ½ = 5 P(X = 34) can be approximated b P(335 < N(3, 5) < 345) Standardising to N(, ) gives P(937 < N(, ) < 69) = = 65
EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY (formerly the Examinations of the Institute of Statisticians) GRADUATE DIPLOMA, 2004
EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY (formerly the Examinations of the Institute of Statisticians) GRADUATE DIPLOMA, 004 Statistical Theory and Methods I Time Allowed: Three Hours Candidates should
More informationEXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY
EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 2016 MODULE 1 : Probability distributions Time allowed: Three hours Candidates should answer FIVE questions. All questions carry equal marks.
More informationTHE ROYAL STATISTICAL SOCIETY 2016 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE 5
THE ROAL STATISTICAL SOCIET 6 EAMINATIONS SOLUTIONS HIGHER CERTIFICATE MODULE The Society is providing these solutions to assist candidates preparing for the examinations in 7. The solutions are intended
More informationTHE ROYAL STATISTICAL SOCIETY 2009 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA MODULAR FORMAT MODULE 3 STOCHASTIC PROCESSES AND TIME SERIES
THE ROYAL STATISTICAL SOCIETY 9 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA MODULAR FORMAT MODULE 3 STOCHASTIC PROCESSES AND TIME SERIES The Society provides these solutions to assist candidates preparing
More informationTHE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE EXAMINATION MODULE 2
THE ROYAL STATISTICAL SOCIETY HIGHER CERTIFICATE EXAMINATION NEW MODULAR SCHEME introduced from the eaminations in 7 MODULE SPECIMEN PAPER B AND SOLUTIONS The time for the eamination is ½ hours The paper
More informationEXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY
EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 2015 MODULE 1 : Probability distributions Time allowed: Three hours Candidates should answer FIVE questions. All questions carry equal marks.
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationTHE ROYAL STATISTICAL SOCIETY 2007 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE (MODULAR FORMAT) MODULE 2 PROBABILITY MODELS
THE ROYAL STATISTICAL SOCIETY 007 EXAMINATIONS SOLUTIONS HIGHER CERTIFICATE (MODULAR FORMAT) MODULE ROBABILITY MODELS The Society provides these solutions to assist candidates preparing for the examinations
More informationSTAT/MA 416 Answers Homework 6 November 15, 2007 Solutions by Mark Daniel Ward PROBLEMS
STAT/MA 4 Answers Homework November 5, 27 Solutions by Mark Daniel Ward PROBLEMS Chapter Problems 2a. The mass p, corresponds to neither of the first two balls being white, so p, 8 7 4/39. The mass p,
More informationRelationship between probability set function and random variable - 2 -
2.0 Random Variables A rat is selected at random from a cage and its sex is determined. The set of possible outcomes is female and male. Thus outcome space is S = {female, male} = {F, M}. If we let X be
More informationTHE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA STATISTICAL THEORY AND METHODS PAPER I
THE ROYAL STATISTICAL SOCIETY 5 EXAMINATIONS SOLUTIONS GRADUATE DIPLOMA STATISTICAL THEORY AND METHODS PAPER I The Society provides these solutios to assist cadidates preparig for the examiatios i future
More informationExponential Distribution and Poisson Process
Exponential Distribution and Poisson Process Stochastic Processes - Lecture Notes Fatih Cavdur to accompany Introduction to Probability Models by Sheldon M. Ross Fall 215 Outline Introduction Exponential
More informationCHAPTER 5. Jointly Probability Mass Function for Two Discrete Distributed Random Variables:
CHAPTER 5 Jointl Distributed Random Variable There are some situations that experiment contains more than one variable and researcher interested in to stud joint behavior of several variables at the same
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationECE 302 Division 2 Exam 2 Solutions, 11/4/2009.
NAME: ECE 32 Division 2 Exam 2 Solutions, /4/29. You will be required to show your student ID during the exam. This is a closed-book exam. A formula sheet is provided. No calculators are allowed. Total
More informationMathematics 426 Robert Gross Homework 9 Answers
Mathematics 4 Robert Gross Homework 9 Answers. Suppose that X is a normal random variable with mean µ and standard deviation σ. Suppose that PX > 9 PX
More information6.041/6.431 Fall 2010 Quiz 2 Solutions
6.04/6.43: Probabilistic Systems Analysis (Fall 200) 6.04/6.43 Fall 200 Quiz 2 Solutions Problem. (80 points) In this problem: (i) X is a (continuous) uniform random variable on [0, 4]. (ii) Y is an exponential
More informationLecture 5: Moment generating functions
Lecture 5: Moment generating functions Definition 2.3.6. The moment generating function (mgf) of a random variable X is { x e tx f M X (t) = E(e tx X (x) if X has a pmf ) = etx f X (x)dx if X has a pdf
More informationRandom Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay
1 / 13 Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay August 8, 2013 2 / 13 Random Variable Definition A real-valued
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More information18.440: Lecture 28 Lectures Review
18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is
More informationMath 407: Probability Theory 5/10/ Final exam (11am - 1pm)
Math 407: Probability Theory 5/10/2013 - Final exam (11am - 1pm) Name: USC ID: Signature: 1. Write your name and ID number in the spaces above. 2. Show all your work and circle your final answer. Simplify
More informationIIT JAM : MATHEMATICAL STATISTICS (MS) 2013
IIT JAM : MATHEMATICAL STATISTICS (MS 2013 Question Paper with Answer Keys Ctanujit Classes Of Mathematics, Statistics & Economics Visit our website for more: www.ctanujit.in IMPORTANT NOTE FOR CANDIDATES
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 08 SUBJECT NAME : Probability & Random Processes SUBJECT CODE : MA645 MATERIAL NAME : University Questions REGULATION : R03 UPDATED ON : November 07 (Upto N/D 07 Q.P) (Scan the
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More information1 Solution to Problem 2.1
Solution to Problem 2. I incorrectly worked this exercise instead of 2.2, so I decided to include the solution anyway. a) We have X Y /3, which is a - function. It maps the interval, ) where X lives) onto
More informationChapter 4. Chapter 4 sections
Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation
More informationChapter 3 Common Families of Distributions
Lecture 9 on BST 631: Statistical Theory I Kui Zhang, 9/3/8 and 9/5/8 Review for the previous lecture Definition: Several commonly used discrete distributions, including discrete uniform, hypergeometric,
More informationChapter 4 Multiple Random Variables
Review for the previous lecture Theorems and Examples: How to obtain the pmf (pdf) of U = g ( X Y 1 ) and V = g ( X Y) Chapter 4 Multiple Random Variables Chapter 43 Bivariate Transformations Continuous
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationTest Problems for Probability Theory ,
1 Test Problems for Probability Theory 01-06-16, 010-1-14 1. Write down the following probability density functions and compute their moment generating functions. (a) Binomial distribution with mean 30
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More informationMath 362, Problem set 1
Math 6, roblem set Due //. (4..8) Determine the mean variance of the mean X of a rom sample of size 9 from a distribution having pdf f(x) = 4x, < x
More informationMATH 151, FINAL EXAM Winter Quarter, 21 March, 2014
Time: 3 hours, 8:3-11:3 Instructions: MATH 151, FINAL EXAM Winter Quarter, 21 March, 214 (1) Write your name in blue-book provided and sign that you agree to abide by the honor code. (2) The exam consists
More informationMATH Solutions to Probability Exercises
MATH 5 9 MATH 5 9 Problem. Suppose we flip a fair coin once and observe either T for tails or H for heads. Let X denote the random variable that equals when we observe tails and equals when we observe
More informationSlides 8: Statistical Models in Simulation
Slides 8: Statistical Models in Simulation Purpose and Overview The world the model-builder sees is probabilistic rather than deterministic: Some statistical model might well describe the variations. An
More informationMath Spring Practice for the final Exam.
Math 4 - Spring 8 - Practice for the final Exam.. Let X, Y, Z be three independnet random variables uniformly distributed on [, ]. Let W := X + Y. Compute P(W t) for t. Honors: Compute the CDF function
More informationSTA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/
STA263/25//24 Tutorial letter 25// /24 Distribution Theor ry II STA263 Semester Department of Statistics CONTENTS: Examination preparation tutorial letterr Solutions to Assignment 6 2 Dear Student, This
More information1 Review of Probability and Distributions
Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote
More informationSOLUTIONS TO MATH68181 EXTREME VALUES AND FINANCIAL RISK EXAM
SOLUTIONS TO MATH68181 EXTREME VALUES AND FINANCIAL RISK EXAM Solutions to Question A1 a) The joint cdf of X Y is F X,Y x, ) x 0 0 4uv + u + v + 1 dudv 4 u v + u + uv + u ] x dv 0 1 4 0 1 x v + x + xv
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationMoments. Raw moment: February 25, 2014 Normalized / Standardized moment:
Moments Lecture 10: Central Limit Theorem and CDFs Sta230 / Mth 230 Colin Rundel Raw moment: Central moment: µ n = EX n ) µ n = E[X µ) 2 ] February 25, 2014 Normalized / Standardized moment: µ n σ n Sta230
More informationSampling Distributions
Sampling Distributions In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of
More information2. This problem is very easy if you remember the exponential cdf; i.e., 0, y 0 F Y (y) = = ln(0.5) = ln = ln 2 = φ 0.5 = β ln 2.
FALL 8 FINAL EXAM SOLUTIONS. Use the basic counting rule. There are n = number of was to choose 5 white balls from 7 = n = number of was to choose ellow ball from 5 = ( ) 7 5 ( ) 5. there are n n = ( )(
More informationStat 100a, Introduction to Probability.
Stat 100a, Introduction to Probability. Outline for the day: 1. Geometric random variables. 2. Negative binomial random variables. 3. Moment generating functions. 4. Poisson random variables. 5. Continuous
More informationSolutions to Assignment #8 Math 501 1, Spring 2006 University of Utah
Solutions to Assignment #8 Math 5, Spring 26 University of Utah Problems:. A man and a woman agree to meet at a certain location at about 2:3 p.m. If the man arrives at a time uniformly distributed between
More informationContinuous Distributions
Continuous Distributions 1.8-1.9: Continuous Random Variables 1.10.1: Uniform Distribution (Continuous) 1.10.4-5 Exponential and Gamma Distributions: Distance between crossovers Prof. Tesler Math 283 Fall
More informationSTA 584 Supplementary Examples (not to be graded) Fall, 2003
Page 1 of 8 Central Michigan University Department of Mathematics STA 584 Supplementary Examples (not to be graded) Fall, 003 1. (a) If A and B are independent events, P(A) =.40 and P(B) =.70, find (i)
More informationContinuous Distributions
A normal distribution and other density functions involving exponential forms play the most important role in probability and statistics. They are related in a certain way, as summarized in a diagram later
More informationBrief Review of Probability
Maura Department of Economics and Finance Università Tor Vergata Outline 1 Distribution Functions Quantiles and Modes of a Distribution 2 Example 3 Example 4 Distributions Outline Distribution Functions
More informationMAS113 Introduction to Probability and Statistics. Proofs of theorems
MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a
More informationTHE QUEEN S UNIVERSITY OF BELFAST
THE QUEEN S UNIVERSITY OF BELFAST 0SOR20 Level 2 Examination Statistics and Operational Research 20 Probability and Distribution Theory Wednesday 4 August 2002 2.30 pm 5.30 pm Examiners { Professor R M
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More information18 Bivariate normal distribution I
8 Bivariate normal distribution I 8 Example Imagine firing arrows at a target Hopefully they will fall close to the target centre As we fire more arrows we find a high density near the centre and fewer
More informationChapter 5 Joint Probability Distributions
Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 5 Joint Probability Distributions 5 Joint Probability Distributions CHAPTER OUTLINE 5-1 Two
More informationFINAL EXAM: Monday 8-10am
ECE 30: Probabilistic Methods in Electrical and Computer Engineering Fall 016 Instructor: Prof. A. R. Reibman FINAL EXAM: Monday 8-10am Fall 016, TTh 3-4:15pm (December 1, 016) This is a closed book exam.
More informationName of the Student: Problems on Discrete & Continuous R.Vs
Engineering Mathematics 05 SUBJECT NAME : Probability & Random Process SUBJECT CODE : MA6 MATERIAL NAME : University Questions MATERIAL CODE : JM08AM004 REGULATION : R008 UPDATED ON : Nov-Dec 04 (Scan
More informationSOLUTION FOR HOMEWORK 11, ACTS 4306
SOLUTION FOR HOMEWORK, ACTS 36 Welcome to your th homework. This is a collection of transformation, Central Limit Theorem (CLT), and other topics.. Solution: By definition of Z, Var(Z) = Var(3X Y.5). We
More informationWrite your Registration Number, Test Centre, Test Code and the Number of this booklet in the appropriate places on the answersheet.
2016 Booklet No. Test Code : PSA Forenoon Questions : 30 Time : 2 hours Write your Registration Number, Test Centre, Test Code and the Number of this booklet in the appropriate places on the answersheet.
More informationSampling Distributions
In statistics, a random sample is a collection of independent and identically distributed (iid) random variables, and a sampling distribution is the distribution of a function of random sample. For example,
More informationPosted June 26, 2013 In the solution of Problem 4 in Practice Examination 8, the fourth item in the list of six events was mistyped as { X 1
ASM Stud Manual for Exam P, Sixth Edition B Dr. Krzsztof M. Ostaszewski, FSA, CFA, MAAA Web site: http://www.krzsio.net E-mail: krzsio@krzsio.net Errata Effective Jul 5, 3, onl the latest edition of this
More informationLIST OF FORMULAS FOR STK1100 AND STK1110
LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function
More informationBMIR Lecture Series on Probability and Statistics Fall, 2015 Uniform Distribution
Lecture #5 BMIR Lecture Series on Probability and Statistics Fall, 2015 Department of Biomedical Engineering and Environmental Sciences National Tsing Hua University s 5.1 Definition ( ) A continuous random
More informationNotes for Math 324, Part 19
48 Notes for Math 324, Part 9 Chapter 9 Multivariate distributions, covariance Often, we need to consider several random variables at the same time. We have a sample space S and r.v. s X, Y,..., which
More informationSystem Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models
System Simulation Part II: Mathematical and Statistical Models Chapter 5: Statistical Models Fatih Cavdur fatihcavdur@uludag.edu.tr March 20, 2012 Introduction Introduction The world of the model-builder
More informationGEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs
STATISTICS 4 Summary Notes. Geometric and Exponential Distributions GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs P(X = x) = ( p) x p x =,, 3,...
More informationContinuous random variables
Continuous random variables Continuous r.v. s take an uncountably infinite number of possible values. Examples: Heights of people Weights of apples Diameters of bolts Life lengths of light-bulbs We cannot
More informationSTAT 3610: Review of Probability Distributions
STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More information18.440: Lecture 28 Lectures Review
18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems
More informationSTAT FINAL EXAM
STAT101 2013 FINAL EXAM This exam is 2 hours long. It is closed book but you can use an A-4 size cheat sheet. There are 10 questions. Questions are not of equal weight. You may need a calculator for some
More informationChapter 6: Functions of Random Variables
Chapter 6: Functions of Random Variables We are often interested in a function of one or several random variables, U(Y 1,..., Y n ). We will study three methods for determining the distribution of a function
More information1 Basic continuous random variable problems
Name M362K Final Here are problems concerning material from Chapters 5 and 6. To review the other chapters, look over previous practice sheets for the two exams, previous quizzes, previous homeworks and
More informationApart from this page, you are not permitted to read the contents of this question paper until instructed to do so by an invigilator.
B. Sc. Examination by course unit 216 MTH6934: Topics in Probability & Stochastic Processes[SOLUTIONS] Duration: 2 hours Date and time: To Be Determined Apart from this page, you are not permitted to read
More informationMATH c UNIVERSITY OF LEEDS Examination for the Module MATH2715 (January 2015) STATISTICAL METHODS. Time allowed: 2 hours
MATH2750 This question paper consists of 8 printed pages, each of which is identified by the reference MATH275. All calculators must carry an approval sticker issued by the School of Mathematics. c UNIVERSITY
More informationChing-Han Hsu, BMES, National Tsing Hua University c 2015 by Ching-Han Hsu, Ph.D., BMIR Lab. = a + b 2. b a. x a b a = 12
Lecture 5 Continuous Random Variables BMIR Lecture Series in Probability and Statistics Ching-Han Hsu, BMES, National Tsing Hua University c 215 by Ching-Han Hsu, Ph.D., BMIR Lab 5.1 1 Uniform Distribution
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationCIVL Continuous Distributions
CIVL 3103 Continuous Distributions Learning Objectives - Continuous Distributions Define continuous distributions, and identify common distributions applicable to engineering problems. Identify the appropriate
More informationApplied Statistics and Probability for Engineers. Sixth Edition. Chapter 4 Continuous Random Variables and Probability Distributions.
Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 4 Continuous Random Variables and Probability Distributions 4 Continuous CHAPTER OUTLINE Random
More informationChapter 4 Continuous Random Variables and Probability Distributions
Applied Statistics and Probability for Engineers Sixth Edition Douglas C. Montgomery George C. Runger Chapter 4 Continuous Random Variables and Probability Distributions 4 Continuous CHAPTER OUTLINE 4-1
More informationExercises and Answers to Chapter 1
Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationSTA 256: Statistics and Probability I
Al Nosedal. University of Toronto. Fall 2017 My momma always said: Life was like a box of chocolates. You never know what you re gonna get. Forrest Gump. There are situations where one might be interested
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationChapter 2. Continuous random variables
Chapter 2 Continuous random variables Outline Review of probability: events and probability Random variable Probability and Cumulative distribution function Review of discrete random variable Introduction
More informationMath 180A. Lecture 16 Friday May 7 th. Expectation. Recall the three main probability density functions so far (1) Uniform (2) Exponential.
Math 8A Lecture 6 Friday May 7 th Epectation Recall the three main probability density functions so far () Uniform () Eponential (3) Power Law e, ( ), Math 8A Lecture 6 Friday May 7 th Epectation Eample
More informationSTAT/MATH 395 A - PROBABILITY II UW Winter Quarter Moment functions. x r p X (x) (1) E[X r ] = x r f X (x) dx (2) (x E[X]) r p X (x) (3)
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 07 Néhémy Lim Moment functions Moments of a random variable Definition.. Let X be a rrv on probability space (Ω, A, P). For a given r N, E[X r ], if it
More informationChapter 5. Statistical Models in Simulations 5.1. Prof. Dr. Mesut Güneş Ch. 5 Statistical Models in Simulations
Chapter 5 Statistical Models in Simulations 5.1 Contents Basic Probability Theory Concepts Discrete Distributions Continuous Distributions Poisson Process Empirical Distributions Useful Statistical Models
More informationE[X n ]= dn dt n M X(t). ). What is the mgf? Solution. Found this the other day in the Kernel matching exercise: 1 M X (t) =
Chapter 7 Generating functions Definition 7.. Let X be a random variable. The moment generating function is given by M X (t) =E[e tx ], provided that the expectation exists for t in some neighborhood of
More informationLimiting Distributions
Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the
More informationDecember 2010 Mathematics 302 Name Page 2 of 11 pages
December 2010 Mathematics 302 Name Page 2 of 11 pages [9] 1. An urn contains red balls, 10 green balls and 1 yellow balls. You randomly select balls, without replacement. (a What ( is( the probability
More informationProblem Y is an exponential random variable with parameter λ = 0.2. Given the event A = {Y < 2},
ECE32 Spring 25 HW Solutions April 6, 25 Solutions to HW Note: Most of these solutions were generated by R. D. Yates and D. J. Goodman, the authors of our textbook. I have added comments in italics where
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More information5.2 Continuous random variables
5.2 Continuous random variables It is often convenient to think of a random variable as having a whole (continuous) interval for its set of possible values. The devices used to describe continuous probability
More informationExpectation, Variance and Standard Deviation for Continuous Random Variables Class 6, Jeremy Orloff and Jonathan Bloom
Expectation, Variance and Standard Deviation for Continuous Random Variables Class 6, 8.5 Jeremy Orloff and Jonathan Bloom Learning Goals. Be able to compute and interpret expectation, variance, and standard
More information(4) Suppose X is a random variable such that E(X) = 5 and Var(X) = 16. Find
Chapter 5 Exercises A random variable X has the distribution: X 6 PX 6 5 6 Find the expected value, variance, and standard deviation of X. A number Y is chosen at random from the set S = {,, }. Find the
More informationHW4 : Bivariate Distributions (1) Solutions
STAT/MATH 395 A - PROBABILITY II UW Winter Quarter 7 Néhémy Lim HW4 : Bivariate Distributions () Solutions Problem. The joint probability mass function of X and Y is given by the following table : X Y
More informationSTAT 515 MIDTERM 2 EXAM November 14, 2018
STAT 55 MIDTERM 2 EXAM November 4, 28 NAME: Section Number: Instructor: In problems that require reasoning, algebraic calculation, or the use of your graphing calculator, it is not sufficient just to write
More informationECE Lecture #9 Part 2 Overview
ECE 450 - Lecture #9 Part Overview Bivariate Moments Mean or Expected Value of Z = g(x, Y) Correlation and Covariance of RV s Functions of RV s: Z = g(x, Y); finding f Z (z) Method : First find F(z), by
More information