September Statistics for MSc Weeks 1-2 Probability and Distribution Theories

Size: px
Start display at page:

Download "September Statistics for MSc Weeks 1-2 Probability and Distribution Theories"

Transcription

1 September Statistics for MSc Weeks 1-2 Probability and Distribution Theories Ali C. Tasiran Department of Economics, Mathematics and Statistics Malet Street, London WC1E 7HX September 2014

2 Contents 1 Introduction Textbooks Some preliminaries Problems Probability Probability definitions and concepts Classical definition of probability Frequency definition of probability Subjective definition of probability Axiomatic definition of probability Problems Random variables and probability distributions Random variables, densities, and cumulative distribution functions Discrete Distributions Continuous Distributions Example Problems Problems Expectations and moments Mathematical Expectation and Moments Mathematical Expectation Moments Problems

3 Chapter 1 Introduction 1.1 Textbooks Lecture notes are provided. However, these are not a substitute for a textbook. I do not recommend any particular text, but in the past students have found the following useful. Greene, W.H., (2004) Econometric Analysis, 5rd edition, Prentice-Hall. A good summary of much of the material can be found in Appendix. Hogg, R. V. and Craig A. T., (1995) Introduction to Mathematical Statistics, 5th edition, Prentice Hall. A popular textbook, even though it is slightly dated. Mittelhammer. R. C., (1999) Mathematical Statistics for Economics and Business, Springer Verlag. A good text. A good mathematical statistics textbook for economists, it is useful especially for further econometric studies. Mood. A.M., Graybill, and Boes D.C., (1974) Introduction to the Theory of Statistics, 3rd edition, McGraw-Hall. Spanos. A., (1999) Probability Theory and Statistical Inference, Econometric Modeling with Observational Data, Cambridge University Press. Wackerly. D., Mendenhall W., and Scheaffer. R., (1996) Mathematical Statistics with Applications, 5th edition, Duxbury Press. Those who plan to take forthcoming courses in Econometrics may buy the book by Green (2004). Welcome to this course. Ali Tasiran atasiran@ems.bbk.ac.uk 1.2 Some preliminaries Statistics is the science of observing data and making inferences about the characteristics of a random mechanism that has generated data. It is also called as science of uncertainty. 2

4 September Statistics 3 In Economics, theoretical models are used to analyze economic behavior. Economic theoretical models are deterministic functions but in real world, the relationships are not exact and deterministic rather than uncertain and stochastic. We thus employ distribution functions to make approximations to the actual processes that generate the observed data. The process that generates data is known as the data generating process (DGP or Super Population). In Econometrics, to study the economic relationships, we estimate statistical models, which are build under guidance of the theoretical economic models and by taking into account the properties in the data generating process. Using parameters of estimated statistical models, we make generalisations about the characteristics of a random mechanism that has generated data. In Econometrics, we use observed data in the samples to draw conclusions about the populations. Populations are either real which the data came or conceptual as processes by which the data were generated. The inference in the first case is called design-based (for experimental data) and used mainly to study samples from populations with known frames. The inference in the second case is called model-based (for observational data) and used mainly to study stochastic relationships. The statistical theory that used for such analyses is called as the Classical inference one will be followed in this course. It is based on two premises: 1. The sample data constitute the only relevant information 2. The construction and assessment on the different procedures for inference are based on long-run behavior under similar circumstances. The starting point of an investigation is an experiment. An experiment is a random experiment if it satisfies the following conditions: - all possible distinct outcomes are known ahead of time - the outcome of a particular trial is not known a priori - the experiment can be duplicated. The totality of all possible outcomes of the experiment is referred to as the sample space (denoted by S) and its distinct individual elements are called the sample points or elementary events. An event, is a subset of a sample space and is a set of sample points that represents several possible outcomes of an experiment. A sample space with a finite or countably infinite sample points (with a one to one correspondence to positive integers) is called a discrete space. A continuous space is one with an uncountable infinite number of sample points (that is, it has as many elements as there are real numbers). Events are generally represented by sets, and some important concepts can be explained by using the algebra of sets (known as Boolean Algebra). Definition 1 The sample space is denoted by S. A = S implies that the events in A must always occur. The empty set is a set with no elements and is denoted by. A = implies that the events in A do not occur. The set of all elements not in A is called the complement of A and is denoted by Ā. Thus, Ā occurs if and only if A does not occur. The set of all points in either a set A or a set B or both is called the union of the two sets and is denoted by. A B means that either the event A or the event B or both occur. Note: A Ā = S.

5 September Statistics 4 The set of all elements in both A and B is called the intersection of the two sets and is represented by. A B means that both the events A and B occur simultaneously. A B = means that A and B cannot occur together. A and B are said to be disjoint or mutually exclusive. Note: A Ā =. A B means that A is contained in B or that A is a subset of B, that is, every element of A is an element of B. In other words, if an event A has occurred, then B must have occurred also. Sometimes it is useful to divide elements of a set A into several subsets that are disjoint. Such a division is known as a partition. If A 1 and A 2 are such partitions, then A 1 A 2 = and A 1 A 2 = A. This can be generalized to n partitions; A = n 1 A i with A i A j = for i j. Some postulates according to the Boolean Algebra: Identity: There exist unique sets and S such that, for every set A, A S = A and A = A. Complementation: For each A we can define a unique set Ā such that A Ā = and A Ā = S. Closure: For every pair of sets A and B, we can define unique sets A B and A B. Commutative: A B = B A; A B = B A. Associative: (A B) C = A (B C). Also (A B) C = A (B C). Distributive: A (B C) = (A B) (A C). Also, A (B C) = (A B) (A C). Morgan s Laws: A B) = Ā B. (A B) = Ā B. Problems 1. Let the set S contains the ordered combination of sexes of two children S = {F F, F M, MF, MM}. Let A denote the subset of possibilities containing no males, B the subset of two males, and C the subset containing at least one male. List the elements of A, B, C, A B, A B, A C, A C, B C, B C, and C B. 2. Verify Morgan s Laws by drawing Venn Diagrams. A B = Ā B. (A B) = Ā B.

6 Chapter 2 Probability 2.1 Probability definitions and concepts Classical definition of probability If an experiment has n(n < ) mutually exclusive and equally likely outcomes, and if n A of these outcomes have an attribute A (that is, the event A occurs in n A possible ways), then the probability of A is n A /n, denoted as P (A) = n A /n Frequency definition of probability Let n A be the number of times the event A occurs in n trials of an experiment. If there exists a real number p such that p = lim n (n A /n), then p is called the probability of A and is denoted as P (A). (Examples are histograms for frequency distribution of variables) Subjective definition of probability Our personal judgments to assess the relative likelihood of various outcomes. They are based on our educated guesses or intuitions. The weather will be rainy with a probability 0.6, tomorrow Axiomatic definition of probability The probability of an event A Ϝ is a real number such that 1) P (A) 0 for every A Ϝ, 2) the probability of the entire sample space S is 1, that is P (S) = 1, and 3) if A 1, A 2,..., A n are mutually exclusive events (that is, A i A j = for all i j), then P (A 1 A 2...A n ) = i P (A i), and this holds for n = also. Where Ϝ is a set of any collection of sub-sets in the sample space, S. The triple (S, Ϝ, P ( )) is referred to as the probability space, and P ( ) is a probability measure. We can derive the following theorems by using the axiomatic Definition of probability. Theorem 2 P (Ā) = 1 P (A). Theorem 3 P (A) 1. 5

7 September Statistics 6 Theorem 4 P ( ) = 0. Theorem 5 If A B, then P (A) P (B). Theorem 6 P (A B) = P (A) + P (B) P (A B). Definition 7 Let A and B be two events in a probability space (S, Ϝ, P (.)) such that P (B) > 0. The conditional probability of A given that B has occurred, denoted by P (A B), is given by P (A B)/P (B). (It should be noted that the original probability space (S, Ϝ, P ( )) remains unchanged even though we focus our attention on the subspace, this is (S, Ϝ, P ( B)) Theorem 8 Bonferroni s Theorem: Let A and B be two events in a sample space S. Then P (A B) 1 P (Ā) P ( B). Theorem 9 Bayes Theorem: If Aand B are two events with positive probabilities, then P (A B) = P (A) P (B A) P (B) Law of total probability Assume that S = A 1 A 2... A n where A i A j = for i j. Then for any event B S n P (B) = P (A i )P (B A i ). i=1 Theorem 10 Extended Bayes Theorem: If A 1, A 2,..., A n constitute a partition of the sample space, so that A i A j = for i j and i A i = S, and P (A i ) 0 for any i, then for a given event B with P (B) > 0, P (A i B) = P (A i) P (B A i ) i P (A i) P (B A i ) Definition 11 Two events A and B with positive probabilities are said to be statistically independent if and only if P (A B) = P (A). Equivalently, P (B A) = P (B) and P (A B) = P (A)P (B). The other type of statistical inference is called Bayesian inference where sample information is combined with prior information. This is expressed of a probability distribution known as the prior distribution. When it is combined with the sample information then a posterior distribution of parameters is obtained. It can be derived by using Bayes Theorem. If we substitute Model (the model that generated the observed data) for A and Data (Observed Data) for B, then we have P (Model Data) = P (Data Model)P (Model) P (Data) (2.1) where P (Data Model) is the probability of observing data given that the Model is true. This is usually called the likelihood, (sample information). P (Model) is the probability

8 September Statistics 7 that the Model is true before observing the data (usually called the prior probability). P (Model Data) is the probability that the Model is true after observing the data (usually called posterior probability). P (Data) is the unconditional probability of observing data (whether the Model is true or not). Hence, the relation can be written P (Model Data) P (Data Model)P (Model) (2.2) That is, that Posterior probability is proportional to likelihood (sample information) times prior probability. The inverse of an estimator s variance is called as the precision. In Classical Inference, we use only parameter s variances but in Bayesian Inference, we have both sample precision and prior precision. Also, the precision (or inverse of the variance) of the posterior distribution of a parameter is the sum of sample precision and prior precision. For example, the posterior mean will lie between the sample mean and the prior mean. The posterior variance will be less than the both the sample and prior variances. These are the reasons behind the increasing popularity of Bayesian Inference in the practical econometric applications. When we speak in econometrics of models to be estimated or tested, we refer to sets of DGPs in Classical Inference context. In design-based inference, we restrict our attention to a particular sample size and characterize a DGP by the law of probability that governs the random variables in a sample of that size. In model based inference, we refer to a limiting process in which the sample size goes to infinity, it is clear that such a restricted characterization will no longer suffice. When we indulge in asymptotic theory, the DGPs in question must be stochastic processes. A stochastic process is a collection of random variables indexed by some suitable index set. This index set may be finite, in which case we have no more than a vector of random variables, or it may be infinite, with either a discrete or a continuous infinity of elements. In order to define a DGP, we must be able to specify the joint distribution of the set of random variables corresponding to the observations contained in a sample of arbitrarily large size. This is a very strong requirement. In econometrics, or any other empirical discipline for that matter, we deal with finite samples. How then can we, even theoretically, treat infinite samples? We must in some way create a rule that allows one to generalize from finite samples to an infinite stochastic process. Unfortunately, for any observational framework, there is an infinite number of ways in which such a rule can be constructed, and different rules can lead to widely asymptotic conclusions. In the process of estimating an econometric model, what we are doing is to try to obtain some estimated characterization of the DGP that actually did generate the data. Let us denote an econometric model that is to be estimated, tested, or both, as M and a typical DGP belonging to M as µ. The simplest model in econometrics is the linear regression model, one possibility is to write y = Xβ + u, u N(0, σ 2 I n ) (2.3) where y and u are n-vectors and X is a nonrandom nxk matrix and y follows the N(Xβ, σ 2 I n ) distribution. This distribution is unique if the parameters β and σ 2 are specified. We may therefore say that the DGP is completely characterized by the model parameters. In other words, knowledge of the model parameters β and σ 2 uniquely identify an element of µ in M. On the other hand, the linear regression model can also be written as y = Xβ + u, u IID(0, σ 2 I n ) (2.4)

9 September Statistics 8 with no assumption of normality. Many aspects of the theory of linear regressions are just applicable, the OLS estimator is unbiased, and its covariance matrix is σ 2 (X X) 1. But the distribution of the vector u, and hence also that of y, is now only partially characterized even when β and σ 2 are known. For example, errors u could be skewed to the left or to the right, could have fourth moments larger or smaller than 3σ 4.Let us call the sets of DGPs associated these regressions M 1 and M 2., respectively. M 1 being in fact a proper subset of M 2. For a given β and σ 2 there is an infinite number of DGPs in M 2 (only one of which is M 1 ) that all correspond to the same β and σ 2. Thus we must consider these models as different models even though the parameters used in them are the same. In either case, it must be possible to associate a parameter vector in a unique way to any DGP µ in the model M, even if the same parameter vector is associated with many DGPs. We call the model M with its associated parameter-defining mapping θ as a parametrized model The main task in our practical work is to build the association between the DGPs of a model and the model parameters. For example, in the Generalized Method of Moments (GMM) context, there are many possible ways of choosing the econometric model, i.e., the underlying set of DGPs. One of the advantages of GMM as an estimation method is that it permits models which consist of a very large number of DGPs. In striking contrast to Maximum Likelihood estimation, where the model must be completely specified, any DGP is admissible if it satisfies a relatively small number of restrictions or regularity conditions. Sometimes, the existence of the moments used to define the parameters is the only requirement needed for a model to be well defined. Problems 1. A sample space consists of five simple events E 1, E 2, E 3, E 4, and E 5. (a) If P (E 1 ) = P (E 2 ) = 0.15, P (E 3 ) = 0.4 and P (E 4 ) = 2P (E 5 ), find the probabilities of E 4 and E 5. (b) If P (E 1 ) = 3P (E 2 ) = 0.3, find the remaining simple events if you know that the remaining events are equally probable. 2. A business office orders paper supplies from one of three vendors, V 1, V 2, and V 3. Orders are to be placed on two successive days, one order per day. Thus (V 2, V 3 ) might denote that vendor V 2 gets the order on the first day and vendor V 3 gets the order on the second day. (a) List the sample points in this experiment of ordering paper on two successive days. (b) Assume the vendors are selected at random each day and assign a probability to each sample point. (c) Let A denote the event that the same vendor gets both orders and B the event that V 2 gets at least one order. Find P (A), P (B), P (A B), and P (A B) by summing probabilities of the sample points in these events.

10 Chapter 3 Random variables and probability distributions 3.1 Random variables, densities, and cumulative distribution functions A random variable X, is a function whose domain is the sample space and whose range is a set of real numbers. Definition 12 In simple terms, a random variable (also referred as a stochastic variable) is a real-valued set function whose value is a real number determined by the outcome of an experiment. The range of a random variable is the set of all the values it can assume. The particular values observed are called realisations x. If these are countable, x 1, x 2,..., it is said to be discrete with associated probabilities P (X = x i ) = p(x i ) 0, p(x i ) = 1; (3.1) and cumulative distribution P (X x j ) = j i=1 p(x i). For a continuous random variable, defined over the real line, the cumulative distribution function is F (x) = P (X x) = where denotes the probability density function f(x) = x df (x) dx i f(u)d(u), (3.2) (3.3) and f(x)d(x) = 1. Also note that the cumulative distribution function satisfies lim x F (x) = 1 and lim x F (x) = 0. Definition 13 The real-valued function F (x) such that F (x) = P x {(, x]} for each x R is called the distribution function, also known as the cumulative distribution (or cumulative density) function, or CDF. 9

11 September Statistics 10 Theorem 14 P (a X b) = F (b) F (a) Theorem 15 For each x R, F (x) is continuous to the right of x. Theorem 16 If F (x) is continuous at x R, then P (X = x) = 0. Although f(x) is defined at a point, P (X = x) = 0 for a continuous random variable. The support of a distribution is the range over which f(x) 0. Let f be a function from R k to R. Let x 0 be a vector in R k and let y = f(x 0 ) be its image. The function f is continuous at x 0 if whenever {x n } n=1 is a sequence in Rk which converges to x 0, then the sequence {f(x n )} n=1 converge to f(x 0). The function f is said to be continuous if it is continuous at each point in its domain. All polynomial functions are continuous. As an example of a function that is not continuous consider { 1, if x > 0, f(x) = 0, if x 0. If both g and f are continuous functions, then g(f(x)) is continuous Discrete Distributions Definition 17 For a discrete random variable X, let f(x) = P x (X = x). The function f(x) is called the probability function (or as probability mass function). The Bernoulli Distribution f(x; θ) = f(x; p) = p x (1 p) 1 x for x = 0, 1(failure, success) and 0 p 1. The Binomial Distribution f(x; θ) = B(x; n, p) = ( ) n p x (1 p) n x = x x = 0, 1,..., n (X is the number of success in n trials) 0 p Continuous Distributions n! x! (n x)! px (1 p) n x (3.4) Definition 18 For a random variable X if there exists a nonnegative function f(x), defined on the real line, such that for any interval B, P (X B) = B f(x) dx (3.5) then X is said to have a continuous distribution and the function f(x) is called the probability density function or simply density function (or pdf). The following can be written for the continuous random variables: F (x) = x f(u) d(u) (3.6)

12 September Statistics 11 f(x) = F (x) = F (x) x (3.7) + F (b) F (a) = Uniform Distribution on an Interval A random variable X with the density function f(u) d(u) = 1 (3.8) b a f(u) d(u) (3.9) f(x; a, b) = 1 (b a) in the interval a X b is called the uniform distribution on an interval. (3.10) The Normal Distribution A random variable X with the density function f(x; µ, σ) = 1 1 σ (2π) e 2 is called a Normal (Gaussian) distributed variable. (x µ)2 σ 2 (3.11) Example 1. Toss of a single fair coin. X =number of heads 0, if x < 0 1 F (x) = 2, if 0 x < 1 1, if x 1 the cumulative distribution function (cdf) of discrete random variables are always step functions because the cdf increases only at a countable of number of points. { 1 f(x) = 2, if x = 0 1 2, if x = 1 F (x) = x j x f(x j )

13 September Statistics Problems 1. Write P (a x b) in terms of integrals and draw a picture for it. 2. Assume the probability density function for x is: { cx, if 0 x 2 f(x) = 0, elsewhere (a) Find the value of c for which f(x) is a pdf. (b) Compute F (x). (c) Compute P (1 x 2). 3. The large lot of electrical is supposed to contain only 5 percent defectives assuming a binomial model. If n = 20 fuses are randomly sampled from this lot, find the probability that at least three defectives will be observed. 4. Let the distribution function of a random variable X be given by F (x) = (a) Find the density function (i.e., pdf) of x. (b) Find P (1 x 3) (c) Find P (x 3) (d) Find P (x 1 x 3). 0, x < 0 x 8, 0 x < 2 x 2 16, 2 x < 4 1, x 4

14 Chapter 4 Expectations and moments 4.1 Mathematical Expectation and Moments The probability density and the cumulative distributions functions determine the probabilities of random variables at various points or in different intervals. Very often we are interested in summary measures of where the distribution is located, how it is dispersed around some average measure, whether it is symmetric around some point, and so on Mathematical Expectation Definition 19 Let X be a random variable with f(x) as the PMF, or PDF, and g(x) be a single-valued-function. The integral is the expected value (or mathematical expectation) of g(x) and is denoted by E[g(X)]. In the case of a discrete random variable this takes the form E[g(X)] = + g(x)f(x i), and in the continuous case, E[g(X)] = + g(x)f(x)dx Mean of a Distribution For the special case of g(x) = X, the mean of a distribution is µ = E(X). Theorem 20 If c is a constant, E(c) = c. Theorem 21 If c is constant, E[cg(X)] = ce[g(x)]. Theorem 22 E[u(X) + v(x)] = E[u(X)] + E[v(X)]. Theorem 23 E(X µ) = 0, where µ = E(X). Examples: Ex1: Let X have the probability density function x f(x)

15 September Statistics 14 E(x) = x xf(x) = 1 ( ) ( ) ( ) ( ) ( 10 = 23 10). Ex2: Let X have the pdf f(x) = { 4x 3, 0 < x < 1 0, elsewhere. E(x) = + xf(x)dx = 1 0 x(4x3 )dx = 4 [ ] 1 0 x4 dx = 4 x = 4 [ ] = 4 5. Moments of a Distribution The mean of a distribution is the expected value of the random variable X. If the following integral exists µ m = E(X m ) = + x m df (4.1) it is called the mth moment around the origin, and it is denoted by µ m. Moments can also be obtained around the mean or the central moments (denoted by µ m ) µ m = E[(X µ) m ] = Variance and Standard Deviation + (x µ) m df (4.2) The central moment of a distribution that corresponds to m = 2 is called the variance of this distribution, and is denoted by σ 2 or V ar(x). The positive square root of the variance is called standard deviation and is denoted by σ or Std(x). The variance is an average of the squared deviation from the mean. There are many deviations from the mean but only one standard deviation. The variance shows the dispersion of a distribution and by squaring deviations one treats positive and negative deviations symmetrically. Mean and Variance of a Normal Distribution A random variable X is normal distributed as N(µ, σ 2 ) the mean is µ, and variance is σ 2. The operation of substracting the mean and dividing by the standard deviation is called standardizing. Then the standardized variable Z = (X µ)/σ is SN(0, 1). Mean and Variance of a Binomial Distribution The random variable X is binomial distributed B(n, p) with the mean np and a variance with np(1 p). (Show this!) Theorem 24 If E(X)=µ and Var(X)=σ 2, and a and b are constants, then V ar(a + bx) = b 2 σ 2. (Show this!)

16 September Statistics 15 Example: Ex3: Let X have the probability density function f(x) = { 4x 3, 0 < x < 1 0, elsewhere. E(x) = 4 5. V ar(x) = E(x 2 ) E 2 (x) = 1 0 x2 (4x 3 )dx [ ] [ ] = 4 x [ ] = = 2 75 = Expectations and Probabilities Any probability can be interpreted as an expectation. Define the variable Z which is equal to 1 if event A occurs, and equal to zero if event A does not occur. Then it is easy to see that P r(a) = E(Z). How much information about the probability distribution of a random variable X is provided by the expectation and variance of X? There are three useful theorems here. Theorem 25 Markov s Inequality If X is nonnegative random variable, that is, if P r(x < 0) = 0, and any k is any constant, then P r(x k) E(X)/k. Theorem 26 Chebyshev s Inequality Let b a positive constant and h(x) be a nonnegative measurable function of the random variable X. Then Pr(h(X) b) 1 b E[h(X)] For any constant c > 0 and σ 2 = V ar(x), Corollary 27 P r( X µ c) σ2 c 2 ( ) σ 2 Corollary 28 P r( X µ c) 1 c 2 Corollary 29 P r( X µ kσ) 1 k 2 For linear functions the expectation of the function is the function of the expectation. But if Y = h(x) is nonlinear, then in general E(Y ) h[e(x)]. The direction of the inequality may depend on the distribution of X. For certain functions, we can be more definite. Theorem 30 Jensen s Inequality If Y = h(x) is concave and E(X) = µ, then E(Y ) h(µ). For example, the logarithmic function is concave, so E[log(X)] log[e(x)] regardless of the distribution of X. Similarly, if Y = h(x) is convex, so that it lies everywhere above its tangent line, then E(Y ) h(µ). For example, the square function is convex, so E(X 2 ) [E(X)] 2 regardless of the distribution of X.

17 September Statistics 16 Approximate Mean and Variance of g(x) Suppose X is a random variable defined on (S, Ϝ, P ( )) with E(X) = µ and V ar(x) = σ 2, and let g(x) be a differentiable and measurable function of X. We first take a linear approximation of g(x) in the neighborhood of µ. This is given by g(x) g(µ) + g (µ)(x µ) (4.3) provided g(µ) and g (µ) exist. Since the second term zero expectation E[g(X)] g(µ), and variance is V ar[g(x)] σ 2 [g (µ)] 2. Mode of a Distribution The point(s) for which f(x) is maximum are called mode. It is the most frequently observed value of X. Median, Upper and Lower Quartiles, and Percentiles A value of x such that P (X < x) (1/2), and P (X (x)) (1/2) is called a median of the distribution. If the point is unique, then it is the median. Thus the median is the point on either side of which lies 50 percent of the distribution. We often prefer median as an average measure because the arithmetic average can be misleading if extreme values are present. The point(s) with an area 1/4 to the left is (are) called the lower quartile(s), and the point(s) corresponding to 3/4 is (are) called upper quartile(s). For any probability p, the values of X, for which the area to the right is p are called the upper pth percentiles (also referred to as quantiles). Coefficient of Variation The coefficient of variation is defined as the ratio (σ/µ)100, where the numerator is the standard deviation and the denominator is the mean. It is a measure of the dispersion of a distribution relative to its mean and useful in the estimation of relationships. We usually say that the variable X does not vary much if the coefficient of variation is less than 5 percent. This is also helpful to make comparison between two variables that are measured with different scales. Skewness and Kurtosis If a continuous density f(x) has the property that f(µ + a) = f(µ a) for all a (µ being the mean of the distribution), then f(x) is said to be symmetric around the mean. If a distribution is not symmetric about the mean, then it is called skewed. A commonly used measure of skewness is α 3 = E[(X µ) 3 /σ 3 ]. For a symmetric distribution such as the normal, this is zero(µ = α 3 = 0). [Positive skewed (µ > α 3 > 0), to the right with a long tail, negative skewed (µ < α 3 < 0), to the left with a long tail]. The peaknedness of a distribution is called kurtosis. One measure of kurtosis is α 4 = E[(X µ) 4 /σ 4 ]. For a normal distribution, kurtosis is called mesokurtic (α 4 = 3). A narrow distribution is called leptokurtic (α 4 > 3) and a flat distribution is called

18 September Statistics 17 platykurtic (α 4 < 3). The value E[(X µ) 4 /σ 4 ] 3 is often referred to as excess kurtosis Moments Mathematical Expectation The concept of mathematical expectation is easily extended to bivariate random variables. We have E[g(X, Y )] = g(x, y)df (x, y) (4.4) where the integral is over the (X, Y ) space. Moments The rth moment of X is E(X r ) = x r df (x) (4.5) Joint Moments E(X r Y s ) = x r y s df (x, y) Let X and Y be independent random variables and let u(x) be a function of X only and v(y ) be a function of Y only. Then, Covariance Covariance between X and Y is defined as E[u(X)v(Y )] = E[u(X)]E[v(Y )] (4.6) σ XY = Cov(X, Y ) = E[(X µ x )(Y µ y )] = E(XY ) µ x µ y (4.7) In the continuous case this takes the form: σ XY = (x µ x )(y µ y )f(x, y)dxdy (4.8) and in the discrete case it is σ XY = x (x µ x )(y µ y )f(x, y) (4.9) y Although the covariance measure is useful in identifying the nature of the association between X and Y, it has a serious problem, namely, the numerical value is very sensitive to the units of measurement. To avoid this problem, a normalized covariance measure is used. This measure is called the correlation coefficient.

19 September Statistics 18 Correlation The quantity ρ XY = σ XY σ X σ Y = Cov(X, Y ) V ar(x) V ar(y ) (4.10) is called correlation coefficient between X and Y. If Cov(X, Y ) = 0, then Cor(X, Y ) = 0, in which case X and Y are said to be uncorrelated. Two random variables are independent then σ XY = 0 and ρ XY = 0. The converse need not to be true. Theorem 31 ρ XY 1 that is, 1 ρ XY 1. The inequality [Cov(X, Y )] 2 V ar(x)v ar(y )is called Cauchy-Schwarz Inequality or ρ 2 XY 1 that is, 1 ρ XY 1. It should be emphasized that ρ XY measures only a linear relationship between X and Y. It is possible to have an exact relation but a correlation less than 1, even 0. Example: To illustrate, consider random variable X which is distributed as Uniform [ θ, θ] and the transformation Y = X 2. Cov(X, Y ) = E(X 3 ) E(X)E(X 2 ) = 0 because the distribution is symmetric around the origin and hence all the odd moments about the origin are zero. It follows that X and Y are uncorrelated even though there is an exact relation between them. In fact, this result holds for any distribution that is symmetric around the origin. Definition 32 Conditional Expectation: Let X and Y be continuous random variables and g(y ) be a continuous function. Then the conditional expectation (or conditional mean) of g(y ) given X = x, denoted by E Y X [g(y ) X], is given by g(y) f(y x) dy wheref(y x) is the conditional density of Y given X. Note that E[g(Y ) X = x] is a function of x and is not a random variable because x is fixed. The special case of E(Y X) is called the regression of Y on X. Theorem 33 Law of Iterated Expectation: E XY [g(y )] = E X [E Y X {g(y ) X}]. That is, the unconditional expectation is the expectation of the conditional expectation. Definition 34 Conditional Variance: Let µ Y X = E(Y X) = µ (X) be the conditional mean of Y given X. Then the conditional variance of Y given X is defined as V ar(y X) = E Y X [(Y µ ) 2 X}]. This is a function of X. Theorem 35 V ar Y X (Y ) = E X [V ar(y X)] + V ar X [E(Y X)], that is, the variance of Y is the mean of its conditional variance plus the variance of its conditional mean. Theorem 36 V ar(ax + by ) = a 2 V ar(x) + 2abCov(X, Y ) + b 2 V ar(y ).

20 September Statistics 19 Approximate Mean and Variance for g(x, Y ) After obtaining a linear approximation of the function g(x, Y ) [ ] [ ] g g g(x, Y ) g(µ x, µ y ) + (X µ X ) + (Y µ Y ) (4.11) X Y its mean can be written E[g(X, Y )] g(µ X, µ Y ). Its variance is [ ] g 2 [ ] g 2 [ ] [ ] g g V ar[g(x, Y )] σx 2 + σy 2 + 2ρ σ X σ Y X Y X Y (4.12) Note that approximations may be grossly in error. You should be especially careful with the variance and covariance approximations. Problems 1. For certain ore samples the proportion Y of impurities per sample is a random variable with density function given by { ( 3 ) f(y) = 2 y 2 + y, 0 y 1 0, elsewhere. The dollar value of each sample is W = 5 0.5Y. Find the mean and variance of W. 2. The random variable Y has the following probability density function { ( 3 ) f(y) = 8 (7 y) 2, 5 y 7 0, elsewhere. (a) Find E(Y ) and V ar(y ). (b) Find an interval shorter than (5, 7) in which least 3/4 of the Y values must lie. (c) Would you expect to see a measurement below 5.5 very often? Why?

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

Week 2. Review of Probability, Random Variables and Univariate Distributions

Week 2. Review of Probability, Random Variables and Univariate Distributions Week 2 Review of Probability, Random Variables and Univariate Distributions Probability Probability Probability Motivation What use is Probability Theory? Probability models Basis for statistical inference

More information

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix

More information

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES

IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the

More information

Review of Probability Theory

Review of Probability Theory Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

MULTIVARIATE PROBABILITY DISTRIBUTIONS

MULTIVARIATE PROBABILITY DISTRIBUTIONS MULTIVARIATE PROBABILITY DISTRIBUTIONS. PRELIMINARIES.. Example. Consider an experiment that consists of tossing a die and a coin at the same time. We can consider a number of random variables defined

More information

Set Theory Digression

Set Theory Digression 1 Introduction to Probability 1.1 Basic Rules of Probability Set Theory Digression A set is defined as any collection of objects, which are called points or elements. The biggest possible collection of

More information

Math-Stat-491-Fall2014-Notes-I

Math-Stat-491-Fall2014-Notes-I Math-Stat-491-Fall2014-Notes-I Hariharan Narayanan October 2, 2014 1 Introduction This writeup is intended to supplement material in the prescribed texts: Introduction to Probability Models, 10th Edition,

More information

Lecture 2: Review of Probability

Lecture 2: Review of Probability Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Preliminary Statistics Lecture 3: Probability Models and Distributions (Outline) prelimsoas.webs.com

Preliminary Statistics Lecture 3: Probability Models and Distributions (Outline) prelimsoas.webs.com 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 3: Probability Models and Distributions (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics,

More information

Random Variables and Expectations

Random Variables and Expectations Inside ECOOMICS Random Variables Introduction to Econometrics Random Variables and Expectations A random variable has an outcome that is determined by an experiment and takes on a numerical value. A procedure

More information

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu

Course: ESO-209 Home Work: 1 Instructor: Debasis Kundu Home Work: 1 1. Describe the sample space when a coin is tossed (a) once, (b) three times, (c) n times, (d) an infinite number of times. 2. A coin is tossed until for the first time the same result appear

More information

BASICS OF PROBABILITY

BASICS OF PROBABILITY October 10, 2018 BASICS OF PROBABILITY Randomness, sample space and probability Probability is concerned with random experiments. That is, an experiment, the outcome of which cannot be predicted with certainty,

More information

Actuarial Science Exam 1/P

Actuarial Science Exam 1/P Actuarial Science Exam /P Ville A. Satopää December 5, 2009 Contents Review of Algebra and Calculus 2 2 Basic Probability Concepts 3 3 Conditional Probability and Independence 4 4 Combinatorial Principles,

More information

Northwestern University Department of Electrical Engineering and Computer Science

Northwestern University Department of Electrical Engineering and Computer Science Northwestern University Department of Electrical Engineering and Computer Science EECS 454: Modeling and Analysis of Communication Networks Spring 2008 Probability Review As discussed in Lecture 1, probability

More information

Lecture 11. Probability Theory: an Overveiw

Lecture 11. Probability Theory: an Overveiw Math 408 - Mathematical Statistics Lecture 11. Probability Theory: an Overveiw February 11, 2013 Konstantin Zuev (USC) Math 408, Lecture 11 February 11, 2013 1 / 24 The starting point in developing the

More information

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

1.1 Review of Probability Theory

1.1 Review of Probability Theory 1.1 Review of Probability Theory Angela Peace Biomathemtics II MATH 5355 Spring 2017 Lecture notes follow: Allen, Linda JS. An introduction to stochastic processes with applications to biology. CRC Press,

More information

Quick Tour of Basic Probability Theory and Linear Algebra

Quick Tour of Basic Probability Theory and Linear Algebra Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra CS224w: Social and Information Network Analysis Fall 2011 Quick Tour of and Linear Algebra Quick Tour of and Linear Algebra Outline Definitions

More information

CS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro

CS37300 Class Notes. Jennifer Neville, Sebastian Moreno, Bruno Ribeiro CS37300 Class Notes Jennifer Neville, Sebastian Moreno, Bruno Ribeiro 2 Background on Probability and Statistics These are basic definitions, concepts, and equations that should have been covered in your

More information

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities

PCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets

More information

Recap of Basic Probability Theory

Recap of Basic Probability Theory 02407 Stochastic Processes Recap of Basic Probability Theory Uffe Høgsbro Thygesen Informatics and Mathematical Modelling Technical University of Denmark 2800 Kgs. Lyngby Denmark Email: uht@imm.dtu.dk

More information

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable

Distributions of Functions of Random Variables. 5.1 Functions of One Random Variable Distributions of Functions of Random Variables 5.1 Functions of One Random Variable 5.2 Transformations of Two Random Variables 5.3 Several Random Variables 5.4 The Moment-Generating Function Technique

More information

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample

More information

Recap of Basic Probability Theory

Recap of Basic Probability Theory 02407 Stochastic Processes? Recap of Basic Probability Theory Uffe Høgsbro Thygesen Informatics and Mathematical Modelling Technical University of Denmark 2800 Kgs. Lyngby Denmark Email: uht@imm.dtu.dk

More information

Expectation. DS GA 1002 Statistical and Mathematical Models. Carlos Fernandez-Granda

Expectation. DS GA 1002 Statistical and Mathematical Models.   Carlos Fernandez-Granda Expectation DS GA 1002 Statistical and Mathematical Models http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall16 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean, variance,

More information

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory

Chapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory Chapter 1 Statistical Reasoning Why statistics? Uncertainty of nature (weather, earth movement, etc. ) Uncertainty in observation/sampling/measurement Variability of human operation/error imperfection

More information

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.

2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y. CS450 Final Review Problems Fall 08 Solutions or worked answers provided Problems -6 are based on the midterm review Identical problems are marked recap] Please consult previous recitations and textbook

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Week 12-13: Discrete Probability

Week 12-13: Discrete Probability Week 12-13: Discrete Probability November 21, 2018 1 Probability Space There are many problems about chances or possibilities, called probability in mathematics. When we roll two dice there are possible

More information

ECON 5350 Class Notes Review of Probability and Distribution Theory

ECON 5350 Class Notes Review of Probability and Distribution Theory ECON 535 Class Notes Review of Probability and Distribution Theory 1 Random Variables Definition. Let c represent an element of the sample space C of a random eperiment, c C. A random variable is a one-to-one

More information

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables

Why study probability? Set theory. ECE 6010 Lecture 1 Introduction; Review of Random Variables ECE 6010 Lecture 1 Introduction; Review of Random Variables Readings from G&S: Chapter 1. Section 2.1, Section 2.3, Section 2.4, Section 3.1, Section 3.2, Section 3.5, Section 4.1, Section 4.2, Section

More information

1: PROBABILITY REVIEW

1: PROBABILITY REVIEW 1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following

More information

Exam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition)

Exam P Review Sheet. for a > 0. ln(a) i=0 ari = a. (1 r) 2. (Note that the A i s form a partition) Exam P Review Sheet log b (b x ) = x log b (y k ) = k log b (y) log b (y) = ln(y) ln(b) log b (yz) = log b (y) + log b (z) log b (y/z) = log b (y) log b (z) ln(e x ) = x e ln(y) = y for y > 0. d dx ax

More information

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces.

Probability Theory. Introduction to Probability Theory. Principles of Counting Examples. Principles of Counting. Probability spaces. Probability Theory To start out the course, we need to know something about statistics and probability Introduction to Probability Theory L645 Advanced NLP Autumn 2009 This is only an introduction; for

More information

Random Variables. P(x) = P[X(e)] = P(e). (1)

Random Variables. P(x) = P[X(e)] = P(e). (1) Random Variables Random variable (discrete or continuous) is used to derive the output statistical properties of a system whose input is a random variable or random in nature. Definition Consider an experiment

More information

Recitation 2: Probability

Recitation 2: Probability Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions

More information

1 Review of Probability

1 Review of Probability 1 Review of Probability Random variables are denoted by X, Y, Z, etc. The cumulative distribution function (c.d.f.) of a random variable X is denoted by F (x) = P (X x), < x

More information

MAT 271E Probability and Statistics

MAT 271E Probability and Statistics MAT 71E Probability and Statistics Spring 013 Instructor : Class Meets : Office Hours : Textbook : Supp. Text : İlker Bayram EEB 1103 ibayram@itu.edu.tr 13.30 1.30, Wednesday EEB 5303 10.00 1.00, Wednesday

More information

Bivariate distributions

Bivariate distributions Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Expectation. DS GA 1002 Probability and Statistics for Data Science. Carlos Fernandez-Granda

Expectation. DS GA 1002 Probability and Statistics for Data Science.   Carlos Fernandez-Granda Expectation DS GA 1002 Probability and Statistics for Data Science http://www.cims.nyu.edu/~cfgranda/pages/dsga1002_fall17 Carlos Fernandez-Granda Aim Describe random variables with a few numbers: mean,

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

1 Probability theory. 2 Random variables and probability theory.

1 Probability theory. 2 Random variables and probability theory. Probability theory Here we summarize some of the probability theory we need. If this is totally unfamiliar to you, you should look at one of the sources given in the readings. In essence, for the major

More information

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n

P (x). all other X j =x j. If X is a continuous random vector (see p.172), then the marginal distributions of X i are: f(x)dx 1 dx n JOINT DENSITIES - RANDOM VECTORS - REVIEW Joint densities describe probability distributions of a random vector X: an n-dimensional vector of random variables, ie, X = (X 1,, X n ), where all X is are

More information

ECON Fundamentals of Probability

ECON Fundamentals of Probability ECON 351 - Fundamentals of Probability Maggie Jones 1 / 32 Random Variables A random variable is one that takes on numerical values, i.e. numerical summary of a random outcome e.g., prices, total GDP,

More information

Probability. Table of contents

Probability. Table of contents Probability Table of contents 1. Important definitions 2. Distributions 3. Discrete distributions 4. Continuous distributions 5. The Normal distribution 6. Multivariate random variables 7. Other continuous

More information

3. Probability and Statistics

3. Probability and Statistics FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important

More information

Lecture 4: Probability and Discrete Random Variables

Lecture 4: Probability and Discrete Random Variables Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 4: Probability and Discrete Random Variables Wednesday, January 21, 2009 Lecturer: Atri Rudra Scribe: Anonymous 1

More information

Stat 5101 Notes: Algorithms

Stat 5101 Notes: Algorithms Stat 5101 Notes: Algorithms Charles J. Geyer January 22, 2016 Contents 1 Calculating an Expectation or a Probability 3 1.1 From a PMF........................... 3 1.2 From a PDF...........................

More information

Exercises and Answers to Chapter 1

Exercises and Answers to Chapter 1 Exercises and Answers to Chapter The continuous type of random variable X has the following density function: a x, if < x < a, f (x), otherwise. Answer the following questions. () Find a. () Obtain mean

More information

Lectures on Elementary Probability. William G. Faris

Lectures on Elementary Probability. William G. Faris Lectures on Elementary Probability William G. Faris February 22, 2002 2 Contents 1 Combinatorics 5 1.1 Factorials and binomial coefficients................. 5 1.2 Sampling with replacement.....................

More information

LIST OF FORMULAS FOR STK1100 AND STK1110

LIST OF FORMULAS FOR STK1100 AND STK1110 LIST OF FORMULAS FOR STK1100 AND STK1110 (Version of 11. November 2015) 1. Probability Let A, B, A 1, A 2,..., B 1, B 2,... be events, that is, subsets of a sample space Ω. a) Axioms: A probability function

More information

Probability and statistics; Rehearsal for pattern recognition

Probability and statistics; Rehearsal for pattern recognition Probability and statistics; Rehearsal for pattern recognition Václav Hlaváč Czech Technical University in Prague Czech Institute of Informatics, Robotics and Cybernetics 166 36 Prague 6, Jugoslávských

More information

UQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables

UQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables UQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables To be provided to students with STAT2201 or CIVIL-2530 (Probability and Statistics) Exam Main exam date: Tuesday, 20 June 1

More information

Review of Statistics I

Review of Statistics I Review of Statistics I Hüseyin Taştan 1 1 Department of Economics Yildiz Technical University April 17, 2010 1 Review of Distribution Theory Random variables, discrete vs continuous Probability distribution

More information

Introduction to Probability Theory for Graduate Economics Fall 2008

Introduction to Probability Theory for Graduate Economics Fall 2008 Introduction to Probability Theory for Graduate Economics Fall 008 Yiğit Sağlam October 10, 008 CHAPTER - RANDOM VARIABLES AND EXPECTATION 1 1 Random Variables A random variable (RV) is a real-valued function

More information

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results

More information

2 (Statistics) Random variables

2 (Statistics) Random variables 2 (Statistics) Random variables References: DeGroot and Schervish, chapters 3, 4 and 5; Stirzaker, chapters 4, 5 and 6 We will now study the main tools use for modeling experiments with unknown outcomes

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

Chapter 2. Probability

Chapter 2. Probability 2-1 Chapter 2 Probability 2-2 Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance with certainty. Examples: rolling a die tossing

More information

Deep Learning for Computer Vision

Deep Learning for Computer Vision Deep Learning for Computer Vision Lecture 3: Probability, Bayes Theorem, and Bayes Classification Peter Belhumeur Computer Science Columbia University Probability Should you play this game? Game: A fair

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Learning Objectives for Stat 225

Learning Objectives for Stat 225 Learning Objectives for Stat 225 08/20/12 Introduction to Probability: Get some general ideas about probability, and learn how to use sample space to compute the probability of a specific event. Set Theory:

More information

7 Random samples and sampling distributions

7 Random samples and sampling distributions 7 Random samples and sampling distributions 7.1 Introduction - random samples We will use the term experiment in a very general way to refer to some process, procedure or natural phenomena that produces

More information

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R

Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R Random Variables Definition: A random variable X is a real valued function that maps a sample space S into the space of real numbers R. X : S R As such, a random variable summarizes the outcome of an experiment

More information

EE514A Information Theory I Fall 2013

EE514A Information Theory I Fall 2013 EE514A Information Theory I Fall 2013 K. Mohan, Prof. J. Bilmes University of Washington, Seattle Department of Electrical Engineering Fall Quarter, 2013 http://j.ee.washington.edu/~bilmes/classes/ee514a_fall_2013/

More information

Math Bootcamp 2012 Miscellaneous

Math Bootcamp 2012 Miscellaneous Math Bootcamp 202 Miscellaneous Factorial, combination and permutation The factorial of a positive integer n denoted by n!, is the product of all positive integers less than or equal to n. Define 0! =.

More information

Introduction to probability theory

Introduction to probability theory Introduction to probability theory Fátima Sánchez Cabo Institute for Genomics and Bioinformatics, TUGraz f.sanchezcabo@tugraz.at 07/03/2007 - p. 1/35 Outline Random and conditional probability (7 March)

More information

Statistics for Economists. Lectures 3 & 4

Statistics for Economists. Lectures 3 & 4 Statistics for Economists Lectures 3 & 4 Asrat Temesgen Stockholm University 1 CHAPTER 2- Discrete Distributions 2.1. Random variables of the Discrete Type Definition 2.1.1: Given a random experiment with

More information

Refresher on Discrete Probability

Refresher on Discrete Probability Refresher on Discrete Probability STAT 27725/CMSC 25400: Machine Learning Shubhendu Trivedi University of Chicago October 2015 Background Things you should have seen before Events, Event Spaces Probability

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

ECON 3150/4150, Spring term Lecture 6

ECON 3150/4150, Spring term Lecture 6 ECON 3150/4150, Spring term 2013. Lecture 6 Review of theoretical statistics for econometric modelling (II) Ragnar Nymoen University of Oslo 31 January 2013 1 / 25 References to Lecture 3 and 6 Lecture

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS

EXAM. Exam #1. Math 3342 Summer II, July 21, 2000 ANSWERS EXAM Exam # Math 3342 Summer II, 2 July 2, 2 ANSWERS i pts. Problem. Consider the following data: 7, 8, 9, 2,, 7, 2, 3. Find the first quartile, the median, and the third quartile. Make a box and whisker

More information

Module 3. Function of a Random Variable and its distribution

Module 3. Function of a Random Variable and its distribution Module 3 Function of a Random Variable and its distribution 1. Function of a Random Variable Let Ω, F, be a probability space and let be random variable defined on Ω, F,. Further let h: R R be a given

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

Chapter 2 Random Variables

Chapter 2 Random Variables Stochastic Processes Chapter 2 Random Variables Prof. Jernan Juang Dept. of Engineering Science National Cheng Kung University Prof. Chun-Hung Liu Dept. of Electrical and Computer Eng. National Chiao Tung

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality

Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality Lecture 13 (Part 2): Deviation from mean: Markov s inequality, variance and its properties, Chebyshev s inequality Discrete Structures II (Summer 2018) Rutgers University Instructor: Abhishek Bhrushundi

More information

STOR Lecture 16. Properties of Expectation - I

STOR Lecture 16. Properties of Expectation - I STOR 435.001 Lecture 16 Properties of Expectation - I Jan Hannig UNC Chapel Hill 1 / 22 Motivation Recall we found joint distributions to be pretty complicated objects. Need various tools from combinatorics

More information

Topic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr.

Topic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr. Topic 2: Probability & Distributions ECO220Y5Y: Quantitative Methods in Economics Dr. Nick Zammit University of Toronto Department of Economics Room KN3272 n.zammit utoronto.ca November 21, 2017 Dr. Nick

More information

Some Concepts of Probability (Review) Volker Tresp Summer 2018

Some Concepts of Probability (Review) Volker Tresp Summer 2018 Some Concepts of Probability (Review) Volker Tresp Summer 2018 1 Definition There are different way to define what a probability stands for Mathematically, the most rigorous definition is based on Kolmogorov

More information

1 Probability and Random Variables

1 Probability and Random Variables 1 Probability and Random Variables The models that you have seen thus far are deterministic models. For any time t, there is a unique solution X(t). On the other hand, stochastic models will result in

More information

More than one variable

More than one variable Chapter More than one variable.1 Bivariate discrete distributions Suppose that the r.v. s X and Y are discrete and take on the values x j and y j, j 1, respectively. Then the joint p.d.f. of X and Y, to

More information

Discrete Distributions

Discrete Distributions Chapter 2 Discrete Distributions 2.1 Random Variables of the Discrete Type An outcome space S is difficult to study if the elements of S are not numbers. However, we can associate each element/outcome

More information

Appendix A : Introduction to Probability and stochastic processes

Appendix A : Introduction to Probability and stochastic processes A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of

More information

STAT 418: Probability and Stochastic Processes

STAT 418: Probability and Stochastic Processes STAT 418: Probability and Stochastic Processes Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical

More information

Algorithms for Uncertainty Quantification

Algorithms for Uncertainty Quantification Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example

More information

Statistical Methods in Particle Physics

Statistical Methods in Particle Physics Statistical Methods in Particle Physics Lecture 3 October 29, 2012 Silvia Masciocchi, GSI Darmstadt s.masciocchi@gsi.de Winter Semester 2012 / 13 Outline Reminder: Probability density function Cumulative

More information

Joint Distribution of Two or More Random Variables

Joint Distribution of Two or More Random Variables Joint Distribution of Two or More Random Variables Sometimes more than one measurement in the form of random variable is taken on each member of the sample space. In cases like this there will be a few

More information

Discrete Probability Refresher

Discrete Probability Refresher ECE 1502 Information Theory Discrete Probability Refresher F. R. Kschischang Dept. of Electrical and Computer Engineering University of Toronto January 13, 1999 revised January 11, 2006 Probability theory

More information

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University

Statistics for Economists Lectures 6 & 7. Asrat Temesgen Stockholm University Statistics for Economists Lectures 6 & 7 Asrat Temesgen Stockholm University 1 Chapter 4- Bivariate Distributions 41 Distributions of two random variables Definition 41-1: Let X and Y be two random variables

More information

Econ 325: Introduction to Empirical Economics

Econ 325: Introduction to Empirical Economics Econ 325: Introduction to Empirical Economics Lecture 2 Probability Copyright 2010 Pearson Education, Inc. Publishing as Prentice Hall Ch. 3-1 3.1 Definition Random Experiment a process leading to an uncertain

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

STAT 414: Introduction to Probability Theory

STAT 414: Introduction to Probability Theory STAT 414: Introduction to Probability Theory Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical Exercises

More information