Statistics Assignment 2 HET551 Design and Development Project 1
|
|
- Chester Lambert
- 5 years ago
- Views:
Transcription
1 Statistics Assignment HET Design and Development Project Michael Allwright Haddon O Neill 7396 Monday, 3 June
2 Simple Stochastic Processes Mean, Variance and Covariance Derivation The following sections show the derivation of various parameters of the data set Y n Equation This data set is defined in y n = { n i= X i n =,, n = () Mean Derivation m(n) = E Y n n m(n) = E X i i= m(n) = n E X i m(n) = nµ Variation Derivation V (n) = Var (Y n ) ( n ) V (n) = Var X i Since the variance of sum is the sum of variances for a random variable, this expression can be rewritten as: i= V (n) = n var (X i ) i= V (n) = nσ 3 Covariance Derivation L cov(y L, Y K ) = Cov X i, i= K j= X j cov(y L, Y K ) = L i= K Cov (X i, X j ) j=
3 Representing this expression in a matrix it can be shown that across the diagonal we have the covariance of a single random variable with itself, and outside the diagonal we the covariance of two indepent random variables Cov (X, X ) Cov (X, X ) Cov (X, X K ) Cov (X, X ) Cov (X, X ) Cov (X, X K ) Cov (X L, X ) Cov (X L, X ) Cov (X L, X K ) j K Using the basic rules of Covariance, Cov (X, X) = Var (X), and Cov (X, Y ) = (for X and Y being indepent random variables), the matrix can be simplified as follows Var (X ) Var (X ) Var ( ) X L K Since Var (X) = σ the cov(y K, Y L ) = Kσ given that K < L i L Covariance Simulation of the Binomial Distribution Process Simulation for the case of X i being binomially distributed To simulate this process, the MATLAB script in Listing was to generate a few trajectories These trajectories are shown in Figure few = ; % Few being the amount of trajectories plotted p =, n =; % Graph variables defined colors = b g r c k ; figure (), clf (), axis ( 4), xlabel ( n Trials ), ylabel ( Trial success ); hold for on; i =:: few Xi = binornd (,p,,n ); Y =; Y= cumsum (Xi ); stairs (Y, sprintf ( %s,colors (i ))); hold off ; title ( sprintf ( Number of successes for Probability of %3 f over -% i trials,p, n )); saveas (gcf, sprintf ( p =%3 f n=%i pdf,p,n )); Listing : MATLAB Source Code to plot of several trajectories of the counting process Estimation of the Covariance of two variables in the sequence Y n through Simulation To estimate the covariance through simulation, a MATLAB script which utilized, trials was executed This script is shown in Listings, this script relies on the sample covariance function defined in Listings 3 The script returned the value 937 for the values indexed at and 8 in Y n (arbitrarily chosen) 3
4 4 3 3 Trial success n Trials Figure : Number of successes for Probability of over - trials From Section 3 the analytical expression for Covariance is: cov(y K, Y L ) = Kσ given that K < L For the Bernoulli case is σ = p( p), substituting this into the expression yields cov(y K, Y L ) = Kσ = Kp( p) which for K = 49 and L = 8, results in the analytical calculation of 987, this approximately matches the result from simulation 3 Calculation of the Joint Probability of the Expression P (X 3 =, X 4 = 7) The probability of S n = y occurring given that S n = y is given by ( ) ( n n P (S n = y, S n = y ) = n y y y ) p y ( p) n y For our chosen probability, that of p = P (Y 3 =, Y 4 = 7) = ( ) ( 3 ) 7 ( ) Using the MATLAB script in Listings 4, this analytical result was confirmed by simulation which gave the result P (Y 3 =, Y 4 = 7) 96 3 Stochastic Process of a Uniformly Distributed Variable From Equation we know that y n is a cumulative summation of the randomly distributed variable X i To determine the function in Equation, we first analytically determine the mean and variance of y(n) ỹ t (n) = y(n) m(n) v(n) () 4
5 % Defining variables p =, n =, trials = ; Z = zeros ( trials,(n +)); acc = ; for i =:: trials Xi = binornd (,p,,n ); % Inserts N ()= Y =; Y= cumsum (Xi ); % Returns if (Y (3)== && Y (4)==7) acc = acc +; Z(i,:)= Y; % Using cov as Sample Variance % Cov (Yk, Yl) where k < l k =49; l =8; SampleCov (Z(:,k+),Z(:,l +)) % Comparing to Sample Variance cov (Z(:,k+),Z(:,l +)) Listing : MATLAB Source Code to determine the covariance of two variables in the counting process function covariance = SampleCov (x,y) acc =; for i = :: size (x,) acc = acc +(( x(i)- mean (x ))*( y(i)- mean (y ))); covariance = acc /( size (x,) -); Listing 3: MATLAB Source Code to Calculate the Sample Covariance 3 Mean Calculation for the Random Variable X i Uniform(, ) The Probability Density Function (pdf) of a uniform random variable in the range, is easily defined It is simply a rectangular area which covers the range, with a height of, This ensures that the area under the pdf is unity f(x) =, x otherwise The expectation of X is then defined as: Substituting in the derived density function: ˆ EX = x f(x) dx EX = ˆ x dx = 4 x = (3)
6 % Finding P( X_n = y, X_n = y) %P( X_n = y, X_n = y) via calculation % where n =3, y =, n =4, y =7; P = nchoosek ((n -n ),(y -y ))* nchoosek (n,y )*p^y *( -p)^( n -y) % Defining variables p =, n =, trials =, Z= zeros ( trials,(n+)), acc =; for i =:: trials Xi = binornd (,p,,n ); Y =; Y= cumsum (Xi ); if (Y (3)== && Y (4)==7) acc = acc +; Z(i,:)= Y; % Results from Simulation, ProbSim = acc / trials Listing 4: MATLAB code to prove probability through simulation 3 Variance Calculation for the Random Variable X i Uniform(, ) Using the result in Equation 3, the Variance can now be defined such that: V ar(x i ) = ˆ V ar(x i ) = ˆ ( x x + ) dx = (x EX) f(x) dx V ar(x i ) = 3 = x3 x + x = ( ) Distribution of the Stochastic Process The transformation of the random variable shown in Equation can now be written numerically given X i Uniform(, ) ỹ t (n) = (y(n) ) 3 (4) As shown in Figure, as n increases there is convergence towards a limiting distribution From Figure, It appears that this distribution is Gaussian in shape, which is expected due to central limit theorem The central limit theorem states that the summation of indepent random variables leads to a Gaussian distribution 6
7 8 Frequency 6 4 n Yt(n) bins 3 4 Figure : 3D Histogram showing change in distribution as n increases The Multi-Dimensional Normal Distribution Probability Density Function for the Bi-variate Normal Distribution The Probability Density Function (pdf) for the Bi-variate Normal Distribution is as follows: f(x, y) = π e ρ σ σ ( ρ ) x σ xyρ + y σ σ σ () Calculating the Correlation Coefficient The expression for the covariance is shown in Equation 6 To determine the covariance of the pdf in Equation, The mean values µ x and µ y are set to zero and the pdf is substituted for the f(x, y) term Cov(x, y) = Cov(x, y) = ˆ ˆ ˆ ˆ (x µ x ) (y µ y ) f(x, y) dx dy (6) xy π e ρ σ x σ y ( ρ ) x σx xyρ σxσy + y σy dx dy (7) To avoid any mistakes in evaluating this integral, this expression was evaluated and simplified using Wolfram s Mathematica Software Package To get a clean solution to this integral, some logical assumptions were made about the variables ρ, σ x and σ y these are as follows: > ρ > 7
8 σ x > σ y > Given these assumptions, the covariance of the pdf in Equation is defined by the expression in Equation 8 Cov(x, y) = ρσ x σ y (8) Comparing the result in Equation 8 with the relationship for the covariance and correlation coefficient in Equation 9, it shown that ρ is the correlation coefficient Corr(X, Y ) = Cov(X, Y ) σ x σ y (9) 3 Proof that the Bivariate Normal Distribution is a special case of the Multivariate Normal Distribution The definition of the pdf for a multivariate joint Gaussian distribution is defined by: Where f X (x) f X, X,X n = exp{ (x m)t K (x m)} (π) n/ K / x = x x x n, m = m m m n = EX EX EX n, K = Var (X ) Cov (X, X ) Cov (X, X n ) Cov (X, X ) Var (X ) Cov (X, X n ) Cov (X n, X ) Cov (X n, X ) Var (X n ) Substituting X and Y into the covariance matrix for n =, K = Var (X) Cov (X, Y ) Cov (Y, X) Var (Y ) K = σ ρσ σ ρσ σ σ The inverse of the covariance matrix is defined as follows K = σ σ ( ρ ) σ ρσ σ ρσ σ σ Substituting this back into the joint Gaussian multivariate pdf, where the term ϕ is the exponent, in the term e ϕ, such that: 8
9 z frequency 3 y bins y x bins (a) 3D Histogram of the Rayleigh Distribution x (b) pdf of the Bivariate Normal Distribution Figure 3: Comparison of the Gaussian and Rayleigh Distributions σ ρσ σ ρσ σ σ σ (x m ) ρσ σ (y m ) ρσ σ (x m ) + σ (y m ) ϕ= (x m, y m ) σ σ ( ρ ) (x m, y m ) ϕ= σ σ ( ρ ) ϕ= x m σ ) (y m ) ρ (x m + σ σ ( ρ ) y m σ x m y m () Them and m are the centering points for the X, Y and n = normal distribution, setting the centering pointsm and m to zero standardizes the expression in Equation giving the standard bivariate normal distribution s exponent from the pdf ϕ= 4 x σ ρxy σ σ + y σ ( ρ ) Rayleigh Distribution The Rayleigh method generates indepent pairs of normally distributed variables The MATLAB code in Listing, generates these values and plots a 3D histogram showing the distribution for the random process For comparison, the code also generates a surface plot of the Probability Density Function for the Bivariate Normal Distribution The plots for the Rayleigh Histogram and Bivariate Normal pdf surface are shown in Figures 3a and 3b respectively As can be from the plots in Figures 3a and 3b, Both of these plots appear to share the Gaussian shape in both dimensions, centered around the same mean 9
10 function myrayleigh ( sigma, numvals ) nqz = e - ; % not quite zero constant angle = random ( unif,, * pi,, numvals ); radius = sigma * sqrt ( - * log ( random ( unif,nqz,,, numvals ))); x y = polcart ( angle, radius ); figure (), clf (), hist3 (x y, 6 6); xlabel ( x- bins ), ylabel ( y- bins ), zlabel ( frequency ); set (gcf, rerer, opengl ); set ( get (gca, child ), FaceColor, interp, CDataMode, auto ); pause ; mx = ; Cx = ; ; x = -::; for i =: length (x), for j =: length (x), f(i,j )=(/(* pi*det (Cx )^/))* exp (( -/)* (x(i) x(j) -mx )* inv (Cx )*( x(i);x(j) - mx )); figure (), clf (), surf (x,x,f); xlabel ( x ), ylabel ( y ), zlabel ( z ); Listing : MATLAB Source Code to generate a histogram of the Rayleigh Distribution and Standard Bivariate Normal Plane for Comparison
Bivariate distributions
Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient
More information18 Bivariate normal distribution I
8 Bivariate normal distribution I 8 Example Imagine firing arrows at a target Hopefully they will fall close to the target centre As we fire more arrows we find a high density near the centre and fewer
More informationMultiple Random Variables
Multiple Random Variables This Version: July 30, 2015 Multiple Random Variables 2 Now we consider models with more than one r.v. These are called multivariate models For instance: height and weight An
More informationChapter 5 continued. Chapter 5 sections
Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationCovariance. Lecture 20: Covariance / Correlation & General Bivariate Normal. Covariance, cont. Properties of Covariance
Covariance Lecture 0: Covariance / Correlation & General Bivariate Normal Sta30 / Mth 30 We have previously discussed Covariance in relation to the variance of the sum of two random variables Review Lecture
More information4. Distributions of Functions of Random Variables
4. Distributions of Functions of Random Variables Setup: Consider as given the joint distribution of X 1,..., X n (i.e. consider as given f X1,...,X n and F X1,...,X n ) Consider k functions g 1 : R n
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More informationMultivariate Random Variable
Multivariate Random Variable Author: Author: Andrés Hincapié and Linyi Cao This Version: August 7, 2016 Multivariate Random Variable 3 Now we consider models with more than one r.v. These are called multivariate
More informationProperties of Summation Operator
Econ 325 Section 003/004 Notes on Variance, Covariance, and Summation Operator By Hiro Kasahara Properties of Summation Operator For a sequence of the values {x 1, x 2,..., x n, we write the sum of x 1,
More informationStatistics, Data Analysis, and Simulation SS 2015
Statistics, Data Analysis, and Simulation SS 2015 08.128.730 Statistik, Datenanalyse und Simulation Dr. Michael O. Distler Mainz, 27. April 2015 Dr. Michael O. Distler
More informationP (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n
JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are
More informationRandom vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.
Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just
More informationIntroduction to Probability and Stocastic Processes - Part I
Introduction to Probability and Stocastic Processes - Part I Lecture 2 Henrik Vie Christensen vie@control.auc.dk Department of Control Engineering Institute of Electronic Systems Aalborg University Denmark
More informationACM 116: Lectures 3 4
1 ACM 116: Lectures 3 4 Joint distributions The multivariate normal distribution Conditional distributions Independent random variables Conditional distributions and Monte Carlo: Rejection sampling Variance
More informationLecture 2: Review of Probability
Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................
More information5 Operations on Multiple Random Variables
EE360 Random Signal analysis Chapter 5: Operations on Multiple Random Variables 5 Operations on Multiple Random Variables Expected value of a function of r.v. s Two r.v. s: ḡ = E[g(X, Y )] = g(x, y)f X,Y
More informationPreliminary statistics
1 Preliminary statistics The solution of a geophysical inverse problem can be obtained by a combination of information from observed data, the theoretical relation between data and earth parameters (models),
More informationNotes for Math 324, Part 19
48 Notes for Math 324, Part 9 Chapter 9 Multivariate distributions, covariance Often, we need to consider several random variables at the same time. We have a sample space S and r.v. s X, Y,..., which
More informationEE4601 Communication Systems
EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationLet X and Y denote two random variables. The joint distribution of these random
EE385 Class Notes 9/7/0 John Stensby Chapter 3: Multiple Random Variables Let X and Y denote two random variables. The joint distribution of these random variables is defined as F XY(x,y) = [X x,y y] P.
More informationLecture 25: Review. Statistics 104. April 23, Colin Rundel
Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April
More informationEEL 5544 Noise in Linear Systems Lecture 30. X (s) = E [ e sx] f X (x)e sx dx. Moments can be found from the Laplace transform as
L30-1 EEL 5544 Noise in Linear Systems Lecture 30 OTHER TRANSFORMS For a continuous, nonnegative RV X, the Laplace transform of X is X (s) = E [ e sx] = 0 f X (x)e sx dx. For a nonnegative RV, the Laplace
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationContinuous Random Variables
1 / 24 Continuous Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay February 27, 2013 2 / 24 Continuous Random Variables
More informationClass 8 Review Problems solutions, 18.05, Spring 2014
Class 8 Review Problems solutions, 8.5, Spring 4 Counting and Probability. (a) Create an arrangement in stages and count the number of possibilities at each stage: ( ) Stage : Choose three of the slots
More informationFormulas for probability theory and linear models SF2941
Formulas for probability theory and linear models SF2941 These pages + Appendix 2 of Gut) are permitted as assistance at the exam. 11 maj 2008 Selected formulae of probability Bivariate probability Transforms
More informationEcon 371 Problem Set #1 Answer Sheet
Econ 371 Problem Set #1 Answer Sheet 2.1 In this question, you are asked to consider the random variable Y, which denotes the number of heads that occur when two coins are tossed. a. The first part of
More informationProbability and Statistics
Probability and Statistics 1 Contents some stochastic processes Stationary Stochastic Processes 2 4. Some Stochastic Processes 4.1 Bernoulli process 4.2 Binomial process 4.3 Sine wave process 4.4 Random-telegraph
More informationLecture 1: August 28
36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random
More informationBasics on Probability. Jingrui He 09/11/2007
Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability
More informationChapter 5. Chapter 5 sections
1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions
More informationSTAT Chapter 5 Continuous Distributions
STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example
More informationSection 8.1. Vector Notation
Section 8.1 Vector Notation Definition 8.1 Random Vector A random vector is a column vector X = [ X 1 ]. X n Each Xi is a random variable. Definition 8.2 Vector Sample Value A sample value of a random
More informationFINAL EXAM: Monday 8-10am
ECE 30: Probabilistic Methods in Electrical and Computer Engineering Fall 016 Instructor: Prof. A. R. Reibman FINAL EXAM: Monday 8-10am Fall 016, TTh 3-4:15pm (December 1, 016) This is a closed book exam.
More informationStatistics 351 Probability I Fall 2006 (200630) Final Exam Solutions. θ α β Γ(α)Γ(β) (uv)α 1 (v uv) β 1 exp v }
Statistics 35 Probability I Fall 6 (63 Final Exam Solutions Instructor: Michael Kozdron (a Solving for X and Y gives X UV and Y V UV, so that the Jacobian of this transformation is x x u v J y y v u v
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order
More informationAsymptotic Statistics-III. Changliang Zou
Asymptotic Statistics-III Changliang Zou The multivariate central limit theorem Theorem (Multivariate CLT for iid case) Let X i be iid random p-vectors with mean µ and and covariance matrix Σ. Then n (
More informationProbability Models. 4. What is the definition of the expectation of a discrete random variable?
1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions
More informationEstimation of uncertainties using the Guide to the expression of uncertainty (GUM)
Estimation of uncertainties using the Guide to the expression of uncertainty (GUM) Alexandr Malusek Division of Radiological Sciences Department of Medical and Health Sciences Linköping University 2014-04-15
More informationECE 541 Stochastic Signals and Systems Problem Set 9 Solutions
ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions Problem Solutions : Yates and Goodman, 9.5.3 9.1.4 9.2.2 9.2.6 9.3.2 9.4.2 9.4.6 9.4.7 and Problem 9.1.4 Solution The joint PDF of X and Y
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More informationMultivariate probability distributions and linear regression
Multivariate probability distributions and linear regression Patrik Hoyer 1 Contents: Random variable, probability distribution Joint distribution Marginal distribution Conditional distribution Independence,
More informationME 597: AUTONOMOUS MOBILE ROBOTICS SECTION 2 PROBABILITY. Prof. Steven Waslander
ME 597: AUTONOMOUS MOBILE ROBOTICS SECTION 2 Prof. Steven Waslander p(a): Probability that A is true 0 pa ( ) 1 p( True) 1, p( False) 0 p( A B) p( A) p( B) p( A B) A A B B 2 Discrete Random Variable X
More informationOn the Expected Absolute Value of a Bivariate Normal Distribution
Journal of Statistical Theory and Applications Volume, Number 4, 0, pp. 37-377 ISSN 538-7887 On the Expected Absolute Value of a Bivariate Normal Distribution S. Reza H. Shojaie, Mina Aminghafari and Adel
More informationEE/CpE 345. Modeling and Simulation. Fall Class 10 November 18, 2002
EE/CpE 345 Modeling and Simulation Class 0 November 8, 2002 Input Modeling Inputs(t) Actual System Outputs(t) Parameters? Simulated System Outputs(t) The input data is the driving force for the simulation
More informationECON Fundamentals of Probability
ECON 351 - Fundamentals of Probability Maggie Jones 1 / 32 Random Variables A random variable is one that takes on numerical values, i.e. numerical summary of a random outcome e.g., prices, total GDP,
More informationMore than one variable
Chapter More than one variable.1 Bivariate discrete distributions Suppose that the r.v. s X and Y are discrete and take on the values x j and y j, j 1, respectively. Then the joint p.d.f. of X and Y, to
More informationContents 1. Contents
Contents 1 Contents 6 Distributions of Functions of Random Variables 2 6.1 Transformation of Discrete r.v.s............. 3 6.2 Method of Distribution Functions............. 6 6.3 Method of Transformations................
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More informationEXAMINATIONS OF THE HONG KONG STATISTICAL SOCIETY GRADUATE DIPLOMA, Statistical Theory and Methods I. Time Allowed: Three Hours
EXAMINATIONS OF THE HONG KONG STATISTICAL SOCIETY GRADUATE DIPLOMA, 008 Statistical Theory and Methods I Time Allowed: Three Hours Candidates should answer FIVE questions. All questions carry equal marks.
More informationEXAM # 3 PLEASE SHOW ALL WORK!
Stat 311, Summer 2018 Name EXAM # 3 PLEASE SHOW ALL WORK! Problem Points Grade 1 30 2 20 3 20 4 30 Total 100 1. A socioeconomic study analyzes two discrete random variables in a certain population of households
More informationEE/CpE 345. Modeling and Simulation. Fall Class 9
EE/CpE 345 Modeling and Simulation Class 9 208 Input Modeling Inputs(t) Actual System Outputs(t) Parameters? Simulated System Outputs(t) The input data is the driving force for the simulation - the behavior
More informationSTAT2201 Assignment 3 Semester 1, 2017 Due 13/4/2017
Class Example 1. Single Sample Descriptive Statistics (a) Summary Statistics and Box-Plots You are working in factory producing hand held bicycle pumps and obtain a sample of 174 bicycle pump weights in
More informationStatistics STAT:5100 (22S:193), Fall Sample Final Exam B
Statistics STAT:5 (22S:93), Fall 25 Sample Final Exam B Please write your answers in the exam books provided.. Let X, Y, and Y 2 be independent random variables with X N(µ X, σ 2 X ) and Y i N(µ Y, σ 2
More informationReview of probability and statistics 1 / 31
Review of probability and statistics 1 / 31 2 / 31 Why? This chapter follows Stock and Watson (all graphs are from Stock and Watson). You may as well refer to the appendix in Wooldridge or any other introduction
More informationThis exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.
TEST #3 STA 536 December, 00 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. You will have access to a copy
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationChapter 2: Random Variables
ECE54: Stochastic Signals and Systems Fall 28 Lecture 2 - September 3, 28 Dr. Salim El Rouayheb Scribe: Peiwen Tian, Lu Liu, Ghadir Ayache Chapter 2: Random Variables Example. Tossing a fair coin twice:
More informationECON 3150/4150, Spring term Lecture 6
ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture
More informationContinuous Probability Distributions
1 Chapter 5 Continuous Probability Distributions 5.1 Probability density function Example 5.1.1. Revisit Example 3.1.1. 11 12 13 14 15 16 21 22 23 24 25 26 S = 31 32 33 34 35 36 41 42 43 44 45 46 (5.1.1)
More informationStatistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University
Statistics for Economists Lectures 6 & 7 Asrat Temesgen Stockholm University 1 Chapter 4- Bivariate Distributions 41 Distributions of two random variables Definition 41-1: Let X and Y be two random variables
More informationUQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables
UQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables To be provided to students with STAT2201 or CIVIL-2530 (Probability and Statistics) Exam Main exam date: Tuesday, 20 June 1
More informationChapter 2. Discrete Distributions
Chapter. Discrete Distributions Objectives ˆ Basic Concepts & Epectations ˆ Binomial, Poisson, Geometric, Negative Binomial, and Hypergeometric Distributions ˆ Introduction to the Maimum Likelihood Estimation
More informationDistributions of Functions of Random Variables. 5.1 Functions of One Random Variable
Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique
More informationRandom Vectors 1. STA442/2101 Fall See last slide for copyright information. 1 / 30
Random Vectors 1 STA442/2101 Fall 2017 1 See last slide for copyright information. 1 / 30 Background Reading: Renscher and Schaalje s Linear models in statistics Chapter 3 on Random Vectors and Matrices
More informationSection 9.1. Expected Values of Sums
Section 9.1 Expected Values of Sums Theorem 9.1 For any set of random variables X 1,..., X n, the sum W n = X 1 + + X n has expected value E [W n ] = E [X 1 ] + E [X 2 ] + + E [X n ]. Proof: Theorem 9.1
More informationn! (k 1)!(n k)! = F (X) U(0, 1). (x, y) = n(n 1) ( F (y) F (x) ) n 2
Order statistics Ex. 4. (*. Let independent variables X,..., X n have U(0, distribution. Show that for every x (0,, we have P ( X ( < x and P ( X (n > x as n. Ex. 4.2 (**. By using induction or otherwise,
More informationUC Berkeley Department of Electrical Engineering and Computer Science. EE 126: Probablity and Random Processes. Problem Set 8 Fall 2007
UC Berkeley Department of Electrical Engineering and Computer Science EE 6: Probablity and Random Processes Problem Set 8 Fall 007 Issued: Thursday, October 5, 007 Due: Friday, November, 007 Reading: Bertsekas
More informationBivariate Distributions. Discrete Bivariate Distribution Example
Spring 7 Geog C: Phaedon C. Kyriakidis Bivariate Distributions Definition: class of multivariate probability distributions describing joint variation of outcomes of two random variables (discrete or continuous),
More informationSTOR Lecture 16. Properties of Expectation - I
STOR 435.001 Lecture 16 Properties of Expectation - I Jan Hannig UNC Chapel Hill 1 / 22 Motivation Recall we found joint distributions to be pretty complicated objects. Need various tools from combinatorics
More informationIntroduction to Computational Finance and Financial Econometrics Probability Review - Part 2
You can t see this text! Introduction to Computational Finance and Financial Econometrics Probability Review - Part 2 Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Probability Review - Part 2 1 /
More informationPart IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015
Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.
More informationSTAT 430/510 Probability
STAT 430/510 Probability Hui Nie Lecture 16 June 24th, 2009 Review Sum of Independent Normal Random Variables Sum of Independent Poisson Random Variables Sum of Independent Binomial Random Variables Conditional
More informationENGG2430A-Homework 2
ENGG3A-Homework Due on Feb 9th,. Independence vs correlation a For each of the following cases, compute the marginal pmfs from the joint pmfs. Explain whether the random variables X and Y are independent,
More informationMATH Notebook 4 Fall 2018/2019
MATH442601 2 Notebook 4 Fall 2018/2019 prepared by Professor Jenny Baglivo c Copyright 2004-2019 by Jenny A. Baglivo. All Rights Reserved. 4 MATH442601 2 Notebook 4 3 4.1 Expected Value of a Random Variable............................
More informationProperties of Random Variables
Properties of Random Variables 1 Definitions A discrete random variable is defined by a probability distribution that lists each possible outcome and the probability of obtaining that outcome If the random
More informationf X, Y (x, y)dx (x), where f(x,y) is the joint pdf of X and Y. (x) dx
INDEPENDENCE, COVARIANCE AND CORRELATION Independence: Intuitive idea of "Y is independent of X": The distribution of Y doesn't depend on the value of X. In terms of the conditional pdf's: "f(y x doesn't
More informationMath 510 midterm 3 answers
Math 51 midterm 3 answers Problem 1 (1 pts) Suppose X and Y are independent exponential random variables both with parameter λ 1. Find the probability that Y < 7X. P (Y < 7X) 7x 7x f(x, y) dy dx e x e
More informationLIST OF FORMULAS FOR STK1100 AND STK1110
LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function
More informationECON 5350 Class Notes Review of Probability and Distribution Theory
ECON 535 Class Notes Review of Probability and Distribution Theory 1 Random Variables Definition. Let c represent an element of the sample space C of a random eperiment, c C. A random variable is a one-to-one
More informationProbability Distributions Columns (a) through (d)
Discrete Probability Distributions Columns (a) through (d) Probability Mass Distribution Description Notes Notation or Density Function --------------------(PMF or PDF)-------------------- (a) (b) (c)
More informationGaussian vectors and central limit theorem
Gaussian vectors and central limit theorem Samy Tindel Purdue University Probability Theory 2 - MA 539 Samy T. Gaussian vectors & CLT Probability Theory 1 / 86 Outline 1 Real Gaussian random variables
More informationRandom Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.
Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example
More informationIntroduction to Computational Finance and Financial Econometrics Matrix Algebra Review
You can t see this text! Introduction to Computational Finance and Financial Econometrics Matrix Algebra Review Eric Zivot Spring 2015 Eric Zivot (Copyright 2015) Matrix Algebra Review 1 / 54 Outline 1
More informationMULTIVARIATE DISTRIBUTIONS
Chapter 9 MULTIVARIATE DISTRIBUTIONS John Wishart (1898-1956) British statistician. Wishart was an assistant to Pearson at University College and to Fisher at Rothamsted. In 1928 he derived the distribution
More informationECE 636: Systems identification
ECE 636: Systems identification Lectures 3 4 Random variables/signals (continued) Random/stochastic vectors Random signals and linear systems Random signals in the frequency domain υ ε x S z + y Experimental
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationBasic concepts of probability theory
Basic concepts of probability theory Random variable discrete/continuous random variable Transform Z transform, Laplace transform Distribution Geometric, mixed-geometric, Binomial, Poisson, exponential,
More informationIntroduction to Normal Distribution
Introduction to Normal Distribution Nathaniel E. Helwig Assistant Professor of Psychology and Statistics University of Minnesota (Twin Cities) Updated 17-Jan-2017 Nathaniel E. Helwig (U of Minnesota) Introduction
More informationStationary independent increments. 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process.
Stationary independent increments 1. Random changes of the form X t+h X t fixed h > 0 are called increments of the process. 2. If each set of increments, corresponding to non-overlapping collection of
More informationProblem Solving. Correlation and Covariance. Yi Lu. Problem Solving. Yi Lu ECE 313 2/51
Yi Lu Correlation and Covariance Yi Lu ECE 313 2/51 Definition Let X and Y be random variables with finite second moments. the correlation: E[XY ] Yi Lu ECE 313 3/51 Definition Let X and Y be random variables
More informationDennis Bricker Dept of Mechanical & Industrial Engineering The University of Iowa
Dennis Bricker Dept of Mechanical & Industrial Engineering The University of Iowa dennis-bricker@uiowa.edu Probability Theory Results page 1 D.Bricker, U. of Iowa, 2002 Probability of simultaneous occurrence
More informationJoint p.d.f. and Independent Random Variables
1 Joint p.d.f. and Independent Random Variables Let X and Y be two discrete r.v. s and let R be the corresponding space of X and Y. The joint p.d.f. of X = x and Y = y, denoted by f(x, y) = P(X = x, Y
More informationFor a stochastic process {Y t : t = 0, ±1, ±2, ±3, }, the mean function is defined by (2.2.1) ± 2..., γ t,
CHAPTER 2 FUNDAMENTAL CONCEPTS This chapter describes the fundamental concepts in the theory of time series models. In particular, we introduce the concepts of stochastic processes, mean and covariance
More informationSTAT 516 Midterm Exam 3 Friday, April 18, 2008
STAT 56 Midterm Exam 3 Friday, April 8, 2008 Name Purdue student ID (0 digits). The testing booklet contains 8 questions. 2. Permitted Texas Instruments calculators: BA-35 BA II Plus BA II Plus Professional
More informationContinuous Probability Distributions
1 Chapter 5 Continuous Probability Distributions 5.1 Probability density function Example 5.1.1. Revisit Example 3.1.1. 11 12 13 14 15 16 21 22 23 24 25 26 S = 31 32 33 34 35 36 41 42 43 44 45 46 (5.1.1)
More information