Your pure maths needs to be far stronger for S4 than in any other Statistics module. You must be strong on general binomial expansion from C4.

Size: px
Start display at page:

Download "Your pure maths needs to be far stronger for S4 than in any other Statistics module. You must be strong on general binomial expansion from C4."

Transcription

1 OCR Statistics Module Revision Sheet The S exam is hour 0 minutes long. You are allowed a graphics calculator. Before you go into the exam make sureyou are fully aware of the contents of theformula booklet you receive. Also be sure not to panic; it is not uncommon to get stuck on a question (I ve been there!). Just continue with what you can do and return at the end to the question(s) you have found hard. If you have time check all your work, especially the first question you attempted...always an area prone to error. Preliminaries J.M.S. Your pure maths needs to be far stronger for S than in any other Statistics module. You must be strong on general binomial expansion from C. (+x) n = +nx+ n(n ) x + n(n )(n )! x + n(n )(n )(n ) x + This is valid only for x <. This is important for probability/moment generating functions. In particular you must be good at plucking out specific coefficients (which may represent probabilities). For example find the x coefficient in x(+x ) +x. x(+x ) +x = (x+x )(+x) = (x+x ) ((+ x ) ) = ( (x+x ) = (x+x ) ) + x ( x x , x So the x coefficient will be 9 6 ( 6, + 9 ) = 0 thinking ahead about what coefficients you are going to need. 5, ). It helps hugely to be Recall from S that E(g(X)) = g(x i )p i for discrete random variables and E(g(X)) = g(x)f(x)dx for continuous random variables. Recall also that Var(X) E(X ) (E(X)). Probability There are three very useful ways of representing information in probability questions. Venn diagrams, tree diagrams and two-way tables. You must think hard about which approach is going to be most helpful in the question you are to answer. Read the whole question before you start! Set theory is very important in probability. Know the following J.M.Stone

2 A B is the intersection of the sets A and B. The overlap between the two sets. AND A B is the union of the sets A and B. Anything that lies in either A or B (or both). OR A means not A. Everything outside A. {} (or ) denotes the empty set. For example A A = {} Events A and B are mutually exclusive if both A and B cannot both happen. Represented by a Venn diagram of non-overlapping circles. Here P(A B) = P(A)+P(B). However in the general case where A and B are not mutually exclusive we have P(A B) = P(A)+P(B) P(A B). This is because we are overcounting the overlap. It is called the addition law. For three events the addition law becomes A, B and C we have (in general) P(A B C) = P(A)+P(B)+P(C) P(A B) P(A C) P(B C)+P(A B C). Again this drops out easily from a Venn diagram. Events A,A,... are said to be exhaustive if P(A A...) =. In other words the events A,A,... contain all the possibilities. If A and B are independent events then P(A B) = P(A) P(B). We read P(A B) as the probability of A given that B has occured. It is defined P(A B) = P(A B). P(B) However this formula is not always easy to apply, so Mr Stone s patented collapsing universes approach from a Venn or tree diagram is often more intuitive. Using P(A B) = P(A B) P(B) and P(B A) = P(A B) P(A) we discover P(A B) = P(A)P(B A) = P(B)P(A B). This is called the multiplication law of probability and is incredibly useful in converting P(A B) into P(B A) and vice versa. The multiplication law drops out readily from a tree diagram. Bayes Theorem states P(A j B) = P(A j)p(b A j ) P(B) = P(A j )P(B A j ) all i P(A i)p(b A i ). This looks scary, but drops out from a tree diagram. The formal statement is not required for S, but is very important. Reverend Thomas Bayes from my home town of Tunbridge Wells. Wrote a document defending Newton s calculus hence a rather good bloke. J.M.Stone

3 Non-Parametric Tests AllofthehypothesistestsstudiedinStats&requiredknowledge(orattheveryleast an assumption) of some kind of underlying distribution for you to carry out the test. However sometimes you have no knowledge about the underlying population. Statisticians therefore developed a series of non-parametric tests for situations where you have no knowledge of the underlying population. The sign test is a test about the median (i.e. the point at which you have an equal number of data points either side). If H 0 : median = 0, say, then under H 0, whether a random piece of data lies above or below 0 has probability. For n pieces of data we therefore have a binomial B(n, ). Rather than work out critical values, the best approach is probably to calculate (under H 0 ) the probability of what you have observed and anything more extreme. For example test at the 5% level whether the median of the data,,,, 6, 7,, 9, 9, 9, 0, 0,, is 5. Note that there are four pieces of data less than 5. H 0 : The median of the data is 5. H : The median of the data is not 5. α = 5%. Two tailed test. Under H 0, X B(, ). P(X ) = 0.09 > 0.05, so at the 5% level there is insufficient evidence to reject H 0 and we conclude that the median of the data is probably 5. [You could have also gone through the rigmarole of demonstrating that the critical value is (or ) but my way is quicker and life s short.] Although there is no example in your textbook I see no reason why they couldn t ask a question where you had a large enough sample to require the normal approximation to B(n, )...don t forget your continuity correction. The sign test is a very crude test because it takes absolutely no account of how far away the data lies on either side of the median. If you want to take account of the magnitude of the deviations you need to use......the Wilcoxon signed-rank test. Here it is assumed that the data is symmetric; therefore it is a test about both the median or the mean because for symmetric data the median and mean are the same. You calculate the deviations from the median/mean, rank the size of the deviations and then sum the positive ranks to get P and sum the negative ranks to get Q. The test statistic is T, where T is the smaller of P or Q. For example test at the 5% level whether the mean of.,., 7.,.9,.,.6, 5.6, 5.7 is. The data sort of looks symmetric, so OK to proceed with Wilcoxon. H 0 : The mean of the data is. H : The mean of the data is greater than. α = 5%. One tailed test. J.M.Stone

4 Data Deviation Rank Signed Rank So P = 7, Q = 9, so T obs = 9. The lower T is, the worse it is for H 0 and the tables give the largest value at which you would reject H 0. T crit = 5. 9 > 5, so at the 5% level we have insufficient evidence to reject H 0 and conclude that the mean is probably. For large samples (i.e. when the tables don t give the values you want; running out of values) a normal approximation can be used where Z = T +0.5 n(n+). n(n+)(n+) Note that because T is the smaller of P and Q that Z will always be negative (both Z crit and Z obs ). For example if you had 00 pieces of data and you were testing at the % level whether the mean was some value (against H of the mean not being some value) and P = 000 and Q = 050 then T = 000. So Z obs = T obs +0.5 n(n+) n(n+)(n+) = =.0 Because it is a two-tailed % test we reverse look-up to obtain Z crit =.576. Finally.0 >.576, so at the % level there is insufficient evidence to reject H 0 and conclude that the mean is probably whatever we thought it was under H 0. The Wilcoxon rank-sum test is the non-parametric equivalent of the two-sample t-test from S. It tests whether two different sets of data are drawn from identical populations. The central idea for the theory is that if X and Y are drawn from identical distributions, then P(X < Y) =. The tables are then constructed from tedious consideration of all the possible arrangements of the ranks (called the sampling distribution ). Given two sets of data, let m be the number of pieces of data from the smaller data set and n be the number of pieces of data from the larger data set (if they are both the same size it s up to you which is m and which n). Then rank all the data and sum the ranks of the m population; call this total R m. Also calculate m(n+m+) R m and let the test statistic W be the smaller of R m and m(n+m+) R m. The smaller W is, the more likely we are to reject H 0 and the tables give the largest W at which we reject H 0. For example test at the 5% level whether the following are drawn from identical populations. A 0 0 B H 0 : Data drawn from identical distributions. H : Data not drawn from identical distributions. α = 5%. Two tailed test. J.M.Stone

5 Data Rank So m =, n = 6, R m =, m(n+m+) R m = 6, W obs =. Looking at the tables we see W crit =, and >, so at the 5% level there is insufficient evidence to reject H 0 and we conclude that the data is probably drawn from identical distributions. For large samples (i.e. when the tables don t give the values you want; running out of values) a normal approximation can be used where Probability Generating Functions Z = W +0.5 m(m+n+). mn(m+n+) In Stats & you met discrete random variables (DRVs) such that each outcome had a probability attached. Sometimes there were rules which related the probability to the outcome (binomial, geometric, Poisson). However, in general we had: x x x x x... P(X = x) p p p p... Recall that p i = becausethesumof all theprobabilities must total and that E(X) = pi x i. Also E(f(X)) = p i f(x i ) from Stats and Var(X) = E(X ) (E(X)) = pi x i ( p i x i ) from Stats. At some point some bright spark decided to consider the properties of G X (t) = E(t X ) = p i t x i = p t x +p t x +p t x +p t x + where t is a dummy variable unrelated to x. You can see that this will create either a finite or infinite series. This is called the probability generating function of X. It is a single function that contains within it all of the (potentially infinite) probabilities of X. For example given x 0 P(X = x) 6 the generating function is G X (t) = p t x + p t x + p t x + p t x + = 6 t + t + + t + t. We can therefore see that if (say) we saw a term 5 t6, then we can see that P(X = 6) = 5. Note that if you see a constant term then that tells you P(X = 0) because t 0 =. An important property is that G X () = because G X () is just the sum of all the probabilities of X, i.e. p i. Another useful thing to do is consider the derivative G X (t) with respect to t; G X(t) = p i x i t x i = p x t x +p x t x +p x t x + Again, if we consider G () we obtain G () = p i x i = p x +p x +p x +p x + = E(X). 5 J.M.Stone

6 Variances can also be calculated by Var(X) = G X()+G X() (G X()). Some standard pgfs are given in the formula book: Distribution B(n, p) P o(λ) Geo(p) pgf ( p+pt) n eλ(t ) pt Any good candidate should be able to derive these... ( p)t For two independent random variables X and Y (with pgfs G X (t) and G Y (t) respectively) the pgf of X+Y is G X+Y (t) = G X (t) G Y (t). This extends to three or more independent random variables. Moment Generating Functions You will recall from FP that the Maclaurin expansion for e x is e x = +x+ x! + x! + x! + x5 5! +. This is valid for all values of x (and you should know why from your Pure teachings). An alternative notation used is e x exp(x). The nth moment of a distribution is E(X n ). So the first moment is just E(X). The second moment is E(X ), which is useful in calculating variances. The zeroth moment is E(X 0 ) = E() =. The moment generating function(mgf) is defined for by x x x x x... P(X = x) p p p p... M X (t) = E(e tx ) = p i e x it = p e x t +p e x t +p e x t +p e x t + = p +p x t+p x t! +p +p x t+p x t! +p +p x t+p x t! +p x t! +p x t! +p x t! x +p +p x t+p t x +p t +!! = (p +p +p +p + ) +(p x +p x +p x +p x + ) t + ( p x +p x +p x +p x + ) t! + ( p x +p x +p x +p x + ) t + = E()+E(X)t+E ( X ) t!! +E( X ) t! +E( X ) t! J.M.Stone

7 So you can see that the constant term of M X (t) should always be E() = because it represents thesum of theprobabilities. Thecoefficient of twill bee(x) andthe coefficient of t! (not just the coefficient of t ) will be E(X ). In general the coefficient of tn n! will be E(X n ), that is, the nth moment. As with pgfs, differentiating mgfs (with respect to t) is a good thing. However, instead of letting t = we let t = 0 (because a 0 = ). So differentiating M X (t) we find: So M X (0) = E(X). Differentiating again we find: M X (t) = p e x t +p e x t +p e x t +p e x t + M X (t) = p x e x t +p x e x t +p x e x t +p x e x t + M X(0) = p x +p x +p x +p x + = x i p i = E(X). M X (t) = p x e x t +p x e x t +p x e x t +p x e x t + M X(t) = p x e x t +p x e x t +p x e x t +p x e x t + M X (0) = p x +p x +p x +p x + = x ip i = E(X ). So using Var(X) = E(X ) (E(X)) we find Var(X) = M X (0) (M X (0)). Notice that with mgfs there are two ways to obtain the expectation and variance of your random variable. All things being equal I would choose the differentiation method, but you must ensure that your mgf is defined for t = 0. Also read the question carefully to see what they are wanting. Moment generating functions can also be defined for continuous random variables: M X (t) = f(x)e tx dx. As before M X (0) =, M X (0) = E(X), M X (0) = E(X ). Convergence issues can arise EXAMPLE!!!!!! Some standard mgfs are given in the formula book: Distribution Uniform on [a,b] Exponential N(µ,σ ) mgf e bt e at (b a)t Any good candidate should be able to derive these too... λ λ t e µt+ σ t As with pgfs, for two independent random variables X and Y (with mgfs G X (t) and G Y (t) respectively) the mgf of X +Y is M X+Y (t) = M X (t) M Y (t). This extends to three or more independent random variables. Estimators It is vital to recall here that E(X) = µ and Var(X) = σ (by definition). 7 J.M.Stone

8 Given a population there may be many parameters that we may wish to know. For example we might like to know the mean µ, the variance σ, the median M, the maximum or minimum, the IQR, etc. In general we shall call this parameter θ. Usually we will never know θ because we won t have the whole population. But we will be able to take a random sample from the population. From this sample we can calculate a quantity U which we shall use to estimate θ. We call U an estimator of θ. U is said to be an unbiased estimator of θ if E(U) = θ. i.e. if we take an average of all possible U (remember that U is a random variable) we will get the desired θ. If E(U) θ then the estimator is said to be biased (not giving the desired result on average). For example to show that K = X +X +5X is an unbiased estimator of µ we merely consider E(K) and keep whittling down as far as we can go (using S expectation and variance algebra) ( ) X +X +5X E(K) = E ( X = E + X + 5X ) = E(X )+ E(X )+ 5 E(X ) = µ+ µ+ 5 µ = µ. For continuous random variables just remember that E(X) = xf(x)dx. For example find the value of k which makes L = k(x +X ) an unbiased estimator of θ for { ( ) f(x) = θ x θ 0 x θ 0 otherwise First calculate E(X) à la S: E(X) = xf(x)dx = θ 0 x θ So for L to be unbiased we need E(L) = θ, so ( x ) [ ] x θ dx = θ θ x θ = θ 0. E(L) = θ E(k(X +X ) = θ k(e(x )+E(X )) = θ k( E(X)) = θ ( ) θ k = θ k =. Given two unbiased estimators the most efficient estimator (of the two) is the one where Var(U) is smaller. A smaller variance is a good thing. J.M.Stone

9 Sometimes you may need calculus to work out the most efficient estimator from an infinite family. For example X, X and X are three independent measurements of X. S = ax +X +X a+6 (with a 6) is suggested as an estimator for µ. Prove that S is unbiased whatever the value of a and find the value of a which makes S most efficient. So ( ) ax +X +X E(S) = E a+6 = a+6 E(aX +X +X ) = a+6 [ae(x )+E(X )+E(X )] = a+6 [aµ+µ+µ] = µ (a+6) = µ. a+6 So S is unbiased for all values of a. Now consider ( ) ax +X +X Var(S) = Var a+6 = (a+6) Var(aX +X +X ) = (a+6) [a Var(X )+Var(X )+6Var(X )] = a +0 (a+6) σ. To minimise Var(S) we need d davar(s) = 0. So 0 = d ( a ) +0 da (a+6) σ = a(a+6) (a+6)(a +0) (a+6) σ So a = 6 or a = 0 So 0 = a(a+6) (a+6)(a +0) 0 = (a+6)[a(a+6) (a +0)] 0 = (a+6)(6a 0)., but a 6 so a = 0 is the value of a that makes S most efficient. Here is a tough type of problem that caught me out the first two (or three (or four (...))) times I saw it. Slot away the method just in case. For example consider f(x) = { x θ 0 x θ 0 otherwise An estimate of θ is required and a suggestion is made to calculate 5L where L is the maximum of two independent observations of X (X and X ). Show that this estimator is unbiased. The thing to remember is that for L to be the maximum of X and X, then X and X must both be less than or equal to L; i.e. we are going to calculate a cdf. So P(L l) = P(X l) P(X l). I suppose we should consider the second derivative to show that this value of a minimises rather than maximises the variance, but life s too short J.M.Stone

10 (This can be extended to three or more independent samplings of X.) By sketching f(x) we can see that the probability that one observation is less than or equal to l is given by a triangle in this case of area l θ (or by the integral l 0 f(x)dx for a more general f(x)). So P(L l) = P(X l) P(X l) = l l = l. Differentiating θ θ θ wrt to l we find the pdf of l to be f(l) = { l θ 0 l θ 0 otherwise Therefore we calculate E ( ) 5L as follows: ( ) 5L E = 5 E(L) θ = 5 l l 0 θ dl = 5 [ ] l 5 θ 5θ = θ. Therefore 5L is an unbiased estimator of θ. I will leave it as an exercise for the reader to demonstrate that Var ( ) 5L = θ. Discrete Bivariate Distributions The discrete random variables you have met thus far have been in one variable only. For example x 5 7 P(X = x) However we can have discrete bivariate distributions. For example 9 0 X 5 0 Y From this we can see, say, P(X =,Y = 5) = 0. The marginal distribution is what one obtains if one of the variables is ignored. In the above example the marginal distribution of X can be written 0 0 x 5 P(X = x) 5 This can be added to the bivariate distribution thus: 0 X Y J.M.Stone

11 E(X) and Var(X) can be calculated in the usual way obtaining E(X) = 0 and Var(X) = (do it!). Similarly you can work out the marginal distribution of Y if you are so inclined. The conditional distribution of a bivariate distribution can be calculated given that one of the variables (X or Y) has taken a specific value. For the above example the distribution of X conditional on Y = is calculated by rewriting the row with all the values divided by P(Y = ) = 5. x 5 5 P(X = x Y = ) This is all from our friend P(A B) = P(A B) P(B). A way to check whether X and Y are independent of each other in a bivariate distribution is to check whether every entry in the distribution is the product of the two relevant marginal probabilities. For example Y X Here we see P(X =,Y = ) = 9 is the same as P(X = ) P(Y = ) = = 9. The same is true for every entry in the table, so X and Y are independent. It only takes one entry not to satisfy this to ensure X and Y are not independent. The covariance of a discrete bivariate distribution is defined Cov(X,Y) E((X µ X )(Y µ Y )). However this tends to be cumbersome to calculate so we use the equivalent formula Cov(X,Y) = E(XY) µ X µ Y. The covariance can be thought of as the correlation coefficient (r from Stats ) for two probability distributions (sort of). The covariance can be both positive or negative (like the correlation coefficient). To calculate the covariance, first create the marginal distributions: Y 5 6 X 0 Y 5 X Then use the marginal distributions to calculate µ X and µ Y. µ X = E(X) = xp = + + = 7. µ Y = E(Y) = yp = =. Now we use this to calculate the covariance thus: Cov(X,Y) = E(XY) µ X µ Y = ( )+( 5 6 )+( )+( 5 )+( 0)+( 5 ) 7 = 5. J.M.Stone

12 If X and Y are independent then Cov(X,Y) = 0. However, if Cov(X,Y) = 0 this does not necessarily mean that X and Y are independent. But if Cov(X,Y) 0 then X and Y cannot be independent. With an understanding of covariance we can write the relationship for Var(aX ±by) when X and Y are not independent: Var(aX ±by) = a Var(X)+b Var(Y)±abCov(X,Y). Notice the extra term at the end of the formula we are used to from S for independent X and Y. J.M.Stone

Core 1 Module Revision Sheet J MS. 1. Basic Algebra

Core 1 Module Revision Sheet J MS. 1. Basic Algebra Core 1 Module Revision Sheet The C1 exam is 1 hour 0 minutes long and is in two sections Section A (6 marks) 8 10 short questions worth no more than 5 marks each Section B (6 marks) questions worth 12

More information

Nonparametric hypothesis tests and permutation tests

Nonparametric hypothesis tests and permutation tests Nonparametric hypothesis tests and permutation tests 1.7 & 2.3. Probability Generating Functions 3.8.3. Wilcoxon Signed Rank Test 3.8.2. Mann-Whitney Test Prof. Tesler Math 283 Fall 2018 Prof. Tesler Wilcoxon

More information

GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs

GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs STATISTICS 4 Summary Notes. Geometric and Exponential Distributions GEOMETRIC -discrete A discrete random variable R counts number of times needed before an event occurs P(X = x) = ( p) x p x =,, 3,...

More information

STATISTICS 1 REVISION NOTES

STATISTICS 1 REVISION NOTES STATISTICS 1 REVISION NOTES Statistical Model Representing and summarising Sample Data Key words: Quantitative Data This is data in NUMERICAL FORM such as shoe size, height etc. Qualitative Data This is

More information

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416)

SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) SUMMARY OF PROBABILITY CONCEPTS SO FAR (SUPPLEMENT FOR MA416) D. ARAPURA This is a summary of the essential material covered so far. The final will be cumulative. I ve also included some review problems

More information

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R

Random Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample

More information

Chapter 4. Chapter 4 sections

Chapter 4. Chapter 4 sections Chapter 4 sections 4.1 Expectation 4.2 Properties of Expectations 4.3 Variance 4.4 Moments 4.5 The Mean and the Median 4.6 Covariance and Correlation 4.7 Conditional Expectation SKIP: 4.8 Utility Expectation

More information

Week 2: Review of probability and statistics

Week 2: Review of probability and statistics Week 2: Review of probability and statistics Marcelo Coca Perraillon University of Colorado Anschutz Medical Campus Health Services Research Methods I HSMP 7607 2017 c 2017 PERRAILLON ALL RIGHTS RESERVED

More information

1 Review of Probability and Distributions

1 Review of Probability and Distributions Random variables. A numerically valued function X of an outcome ω from a sample space Ω X : Ω R : ω X(ω) is called a random variable (r.v.), and usually determined by an experiment. We conventionally denote

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Definitions. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Definitions Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline.

Random Variables. Cumulative Distribution Function (CDF) Amappingthattransformstheeventstotherealline. Random Variables Amappingthattransformstheeventstotherealline. Example 1. Toss a fair coin. Define a random variable X where X is 1 if head appears and X is if tail appears. P (X =)=1/2 P (X =1)=1/2 Example

More information

MAS113 Introduction to Probability and Statistics. Proofs of theorems

MAS113 Introduction to Probability and Statistics. Proofs of theorems MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a

More information

Preliminary Statistics. Lecture 3: Probability Models and Distributions

Preliminary Statistics. Lecture 3: Probability Models and Distributions Preliminary Statistics Lecture 3: Probability Models and Distributions Rory Macqueen (rm43@soas.ac.uk), September 2015 Outline Revision of Lecture 2 Probability Density Functions Cumulative Distribution

More information

Math Review Sheet, Fall 2008

Math Review Sheet, Fall 2008 1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the

More information

Counting principles, including permutations and combinations.

Counting principles, including permutations and combinations. 1 Counting principles, including permutations and combinations. The binomial theorem: expansion of a + b n, n ε N. THE PRODUCT RULE If there are m different ways of performing an operation and for each

More information

Fourier and Stats / Astro Stats and Measurement : Stats Notes

Fourier and Stats / Astro Stats and Measurement : Stats Notes Fourier and Stats / Astro Stats and Measurement : Stats Notes Andy Lawrence, University of Edinburgh Autumn 2013 1 Probabilities, distributions, and errors Laplace once said Probability theory is nothing

More information

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019

Lecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019 Lecture 10: Probability distributions DANIEL WELLER TUESDAY, FEBRUARY 19, 2019 Agenda What is probability? (again) Describing probabilities (distributions) Understanding probabilities (expectation) Partial

More information

Probability and Distributions

Probability and Distributions Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated

More information

APPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2

APPM/MATH 4/5520 Solutions to Exam I Review Problems. f X 1,X 2. 2e x 1 x 2. = x 2 APPM/MATH 4/5520 Solutions to Exam I Review Problems. (a) f X (x ) f X,X 2 (x,x 2 )dx 2 x 2e x x 2 dx 2 2e 2x x was below x 2, but when marginalizing out x 2, we ran it over all values from 0 to and so

More information

MAS113 Introduction to Probability and Statistics. Proofs of theorems

MAS113 Introduction to Probability and Statistics. Proofs of theorems MAS113 Introduction to Probability and Statistics Proofs of theorems Theorem 1 De Morgan s Laws) See MAS110 Theorem 2 M1 By definition, B and A \ B are disjoint, and their union is A So, because m is a

More information

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3

Probability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3 Probability Paul Schrimpf January 23, 2018 Contents 1 Definitions 2 2 Properties 3 3 Random variables 4 3.1 Discrete........................................... 4 3.2 Continuous.........................................

More information

2 (Statistics) Random variables

2 (Statistics) Random variables 2 (Statistics) Random variables References: DeGroot and Schervish, chapters 3, 4 and 5; Stirzaker, chapters 4, 5 and 6 We will now study the main tools use for modeling experiments with unknown outcomes

More information

EXPECTED VALUE of a RV. corresponds to the average value one would get for the RV when repeating the experiment, =0.

EXPECTED VALUE of a RV. corresponds to the average value one would get for the RV when repeating the experiment, =0. EXPECTED VALUE of a RV corresponds to the average value one would get for the RV when repeating the experiment, independently, infinitely many times. Sample (RIS) of n values of X (e.g. More accurately,

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY GRADUATE DIPLOMA, 2015 MODULE 1 : Probability distributions Time allowed: Three hours Candidates should answer FIVE questions. All questions carry equal marks.

More information

Review for Exam Spring 2018

Review for Exam Spring 2018 Review for Exam 1 18.05 Spring 2018 Extra office hours Tuesday: David 3 5 in 2-355 Watch web site for more Friday, Saturday, Sunday March 9 11: no office hours March 2, 2018 2 / 23 Exam 1 Designed to be

More information

CME 106: Review Probability theory

CME 106: Review Probability theory : Probability theory Sven Schmit April 3, 2015 1 Overview In the first half of the course, we covered topics from probability theory. The difference between statistics and probability theory is the following:

More information

STAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed.

STAT 302 Introduction to Probability Learning Outcomes. Textbook: A First Course in Probability by Sheldon Ross, 8 th ed. STAT 302 Introduction to Probability Learning Outcomes Textbook: A First Course in Probability by Sheldon Ross, 8 th ed. Chapter 1: Combinatorial Analysis Demonstrate the ability to solve combinatorial

More information

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors

5. Random Vectors. probabilities. characteristic function. cross correlation, cross covariance. Gaussian random vectors. functions of random vectors EE401 (Semester 1) 5. Random Vectors Jitkomut Songsiri probabilities characteristic function cross correlation, cross covariance Gaussian random vectors functions of random vectors 5-1 Random vectors we

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Limiting Distributions

Limiting Distributions Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the

More information

Final Exam # 3. Sta 230: Probability. December 16, 2012

Final Exam # 3. Sta 230: Probability. December 16, 2012 Final Exam # 3 Sta 230: Probability December 16, 2012 This is a closed-book exam so do not refer to your notes, the text, or any other books (please put them on the floor). You may use the extra sheets

More information

Review of Statistics

Review of Statistics Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Introduction to Probability and Statistics Lecture 17: Continuous random variables: conditional PDF Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin

More information

Probability Models. 4. What is the definition of the expectation of a discrete random variable?

Probability Models. 4. What is the definition of the expectation of a discrete random variable? 1 Probability Models The list of questions below is provided in order to help you to prepare for the test and exam. It reflects only the theoretical part of the course. You should expect the questions

More information

Basic Probability Reference Sheet

Basic Probability Reference Sheet February 27, 2001 Basic Probability Reference Sheet 17.846, 2001 This is intended to be used in addition to, not as a substitute for, a textbook. X is a random variable. This means that X is a variable

More information

ALL TEXTS BELONG TO OWNERS. Candidate code: glt090 TAKEN FROM

ALL TEXTS BELONG TO OWNERS. Candidate code: glt090 TAKEN FROM How are Generating Functions used in finding the closed form of sequences involving recurrence relations and in the analysis of probability distributions? Mathematics Extended Essay Word count: 3865 Abstract

More information

IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008

IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 IEOR 6711: Stochastic Models I SOLUTIONS to the First Midterm Exam, October 7, 2008 Justify your answers; show your work. 1. A sequence of Events. (10 points) Let {B n : n 1} be a sequence of events in

More information

Jointly Distributed Random Variables

Jointly Distributed Random Variables Jointly Distributed Random Variables CE 311S What if there is more than one random variable we are interested in? How should you invest the extra money from your summer internship? To simplify matters,

More information

3. Probability and Statistics

3. Probability and Statistics FE661 - Statistical Methods for Financial Engineering 3. Probability and Statistics Jitkomut Songsiri definitions, probability measures conditional expectations correlation and covariance some important

More information

Bivariate distributions

Bivariate distributions Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient

More information

1 Random Variable: Topics

1 Random Variable: Topics Note: Handouts DO NOT replace the book. In most cases, they only provide a guideline on topics and an intuitive feel. 1 Random Variable: Topics Chap 2, 2.1-2.4 and Chap 3, 3.1-3.3 What is a random variable?

More information

Chapter 2. Probability

Chapter 2. Probability 2-1 Chapter 2 Probability 2-2 Section 2.1: Basic Ideas Definition: An experiment is a process that results in an outcome that cannot be predicted in advance with certainty. Examples: rolling a die tossing

More information

Chapter 2. Continuous random variables

Chapter 2. Continuous random variables Chapter 2 Continuous random variables Outline Review of probability: events and probability Random variable Probability and Cumulative distribution function Review of discrete random variable Introduction

More information

IEOR 4701: Stochastic Models in Financial Engineering. Summer 2007, Professor Whitt. SOLUTIONS to Homework Assignment 9: Brownian motion

IEOR 4701: Stochastic Models in Financial Engineering. Summer 2007, Professor Whitt. SOLUTIONS to Homework Assignment 9: Brownian motion IEOR 471: Stochastic Models in Financial Engineering Summer 27, Professor Whitt SOLUTIONS to Homework Assignment 9: Brownian motion In Ross, read Sections 1.1-1.3 and 1.6. (The total required reading there

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 18-27 Review Scott Sheffield MIT Outline Outline It s the coins, stupid Much of what we have done in this course can be motivated by the i.i.d. sequence X i where each X i is

More information

Fundamental Tools - Probability Theory II

Fundamental Tools - Probability Theory II Fundamental Tools - Probability Theory II MSc Financial Mathematics The University of Warwick September 29, 2015 MSc Financial Mathematics Fundamental Tools - Probability Theory II 1 / 22 Measurable random

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

Week Topics of study Home/Independent Learning Assessment (If in addition to homework) 7 th September 2015

Week Topics of study Home/Independent Learning Assessment (If in addition to homework) 7 th September 2015 Week Topics of study Home/Independent Learning Assessment (If in addition to homework) 7 th September Functions: define the terms range and domain (PLC 1A) and identify the range and domain of given functions

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

STAT2201. Analysis of Engineering & Scientific Data. Unit 3

STAT2201. Analysis of Engineering & Scientific Data. Unit 3 STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random

More information

Topic 3: The Expectation of a Random Variable

Topic 3: The Expectation of a Random Variable Topic 3: The Expectation of a Random Variable Course 003, 2017 Page 0 Expectation of a discrete random variable Definition (Expectation of a discrete r.v.): The expected value (also called the expectation

More information

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results

More information

STA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/

STA2603/205/1/2014 /2014. ry II. Tutorial letter 205/1/ STA263/25//24 Tutorial letter 25// /24 Distribution Theor ry II STA263 Semester Department of Statistics CONTENTS: Examination preparation tutorial letterr Solutions to Assignment 6 2 Dear Student, This

More information

Class 26: review for final exam 18.05, Spring 2014

Class 26: review for final exam 18.05, Spring 2014 Probability Class 26: review for final eam 8.05, Spring 204 Counting Sets Inclusion-eclusion principle Rule of product (multiplication rule) Permutation and combinations Basics Outcome, sample space, event

More information

7 Random samples and sampling distributions

7 Random samples and sampling distributions 7 Random samples and sampling distributions 7.1 Introduction - random samples We will use the term experiment in a very general way to refer to some process, procedure or natural phenomena that produces

More information

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise.

(y 1, y 2 ) = 12 y3 1e y 1 y 2 /2, y 1 > 0, y 2 > 0 0, otherwise. 54 We are given the marginal pdfs of Y and Y You should note that Y gamma(4, Y exponential( E(Y = 4, V (Y = 4, E(Y =, and V (Y = 4 (a With U = Y Y, we have E(U = E(Y Y = E(Y E(Y = 4 = (b Because Y and

More information

Joint Probability Distributions, Correlations

Joint Probability Distributions, Correlations Joint Probability Distributions, Correlations What we learned so far Events: Working with events as sets: union, intersection, etc. Some events are simple: Head vs Tails, Cancer vs Healthy Some are more

More information

Cambridge Pre-U Syllabus. Cambridge International Level 3 Pre-U Certificate in FURTHER MATHEMATICS. For examination in 2010, 2011 and 2012

Cambridge Pre-U Syllabus. Cambridge International Level 3 Pre-U Certificate in FURTHER MATHEMATICS. For examination in 2010, 2011 and 2012 www.xtremepapers.com Cambridge International Level 3 Pre-U Certificate in FURTHER MATHEMATICS For examination in 00, 0 and 0 Further Mathematics (9795) Cambridge International Level 3 Pre-U Certificate

More information

18.440: Lecture 28 Lectures Review

18.440: Lecture 28 Lectures Review 18.440: Lecture 28 Lectures 17-27 Review Scott Sheffield MIT 1 Outline Continuous random variables Problems motivated by coin tossing Random variable properties 2 Outline Continuous random variables Problems

More information

University of Regina. Lecture Notes. Michael Kozdron

University of Regina. Lecture Notes. Michael Kozdron University of Regina Statistics 252 Mathematical Statistics Lecture Notes Winter 2005 Michael Kozdron kozdron@math.uregina.ca www.math.uregina.ca/ kozdron Contents 1 The Basic Idea of Statistics: Estimating

More information

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015

Part IA Probability. Theorems. Based on lectures by R. Weber Notes taken by Dexter Chua. Lent 2015 Part IA Probability Theorems Based on lectures by R. Weber Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly) after lectures.

More information

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.

This exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text. TEST #3 STA 5326 December 4, 214 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. (You will have access to

More information

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3.

t x 1 e t dt, and simplify the answer when possible (for example, when r is a positive even number). In particular, confirm that EX 4 = 3. Mathematical Statistics: Homewor problems General guideline. While woring outside the classroom, use any help you want, including people, computer algebra systems, Internet, and solution manuals, but mae

More information

1 Presessional Probability

1 Presessional Probability 1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional

More information

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Math 224 Fall 2017 Homework 1 Drew Armstrong Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Section 1.1, Exercises 4,5,6,7,9,12. Solutions to Book Problems.

More information

p. 6-1 Continuous Random Variables p. 6-2

p. 6-1 Continuous Random Variables p. 6-2 Continuous Random Variables Recall: For discrete random variables, only a finite or countably infinite number of possible values with positive probability (>). Often, there is interest in random variables

More information

Calculus first semester exam information and practice problems

Calculus first semester exam information and practice problems Calculus first semester exam information and practice problems As I ve been promising for the past year, the first semester exam in this course encompasses all three semesters of Math SL thus far. It is

More information

Disjointness and Additivity

Disjointness and Additivity Midterm 2: Format Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten

More information

STAT 3610: Review of Probability Distributions

STAT 3610: Review of Probability Distributions STAT 3610: Review of Probability Distributions Mark Carpenter Professor of Statistics Department of Mathematics and Statistics August 25, 2015 Support of a Random Variable Definition The support of a random

More information

Lecture 16. Lectures 1-15 Review

Lecture 16. Lectures 1-15 Review 18.440: Lecture 16 Lectures 1-15 Review Scott Sheffield MIT 1 Outline Counting tricks and basic principles of probability Discrete random variables 2 Outline Counting tricks and basic principles of probability

More information

Limiting Distributions

Limiting Distributions We introduce the mode of convergence for a sequence of random variables, and discuss the convergence in probability and in distribution. The concept of convergence leads us to the two fundamental results

More information

SDS 321: Introduction to Probability and Statistics

SDS 321: Introduction to Probability and Statistics SDS 321: Introduction to Probability and Statistics Lecture 14: Continuous random variables Purnamrita Sarkar Department of Statistics and Data Science The University of Texas at Austin www.cs.cmu.edu/

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley

Midterm 2 Review. CS70 Summer Lecture 6D. David Dinh 28 July UC Berkeley Midterm 2 Review CS70 Summer 2016 - Lecture 6D David Dinh 28 July 2016 UC Berkeley Midterm 2: Format 8 questions, 190 points, 110 minutes (same as MT1). Two pages (one double-sided sheet) of handwritten

More information

Part (A): Review of Probability [Statistics I revision]

Part (A): Review of Probability [Statistics I revision] Part (A): Review of Probability [Statistics I revision] 1 Definition of Probability 1.1 Experiment An experiment is any procedure whose outcome is uncertain ffl toss a coin ffl throw a die ffl buy a lottery

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

CS 237: Probability in Computing

CS 237: Probability in Computing CS 237: Probability in Computing Wayne Snyder Computer Science Department Boston University Lecture 13: Normal Distribution Exponential Distribution Recall that the Normal Distribution is given by an explicit

More information

Review of Basic Probability Theory

Review of Basic Probability Theory Review of Basic Probability Theory James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) 1 / 35 Review of Basic Probability Theory

More information

3 Multiple Discrete Random Variables

3 Multiple Discrete Random Variables 3 Multiple Discrete Random Variables 3.1 Joint densities Suppose we have a probability space (Ω, F,P) and now we have two discrete random variables X and Y on it. They have probability mass functions f

More information

Joint Probability Distributions, Correlations

Joint Probability Distributions, Correlations Joint Probability Distributions, Correlations What we learned so far Events: Working with events as sets: union, intersection, etc. Some events are simple: Head vs Tails, Cancer vs Healthy Some are more

More information

EE4601 Communication Systems

EE4601 Communication Systems EE4601 Communication Systems Week 2 Review of Probability, Important Distributions 0 c 2011, Georgia Institute of Technology (lect2 1) Conditional Probability Consider a sample space that consists of two

More information

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). For example P(X.04) =.8508. For z < 0 subtract the value from,

More information

MATHEMATICS. Higher 2 (Syllabus 9740)

MATHEMATICS. Higher 2 (Syllabus 9740) MATHEMATICS Higher (Syllabus 9740) CONTENTS Page AIMS ASSESSMENT OBJECTIVES (AO) USE OF GRAPHING CALCULATOR (GC) 3 LIST OF FORMULAE 3 INTEGRATION AND APPLICATION 3 SCHEME OF EXAMINATION PAPERS 3 CONTENT

More information

Discrete Probability Refresher

Discrete Probability Refresher ECE 1502 Information Theory Discrete Probability Refresher F. R. Kschischang Dept. of Electrical and Computer Engineering University of Toronto January 13, 1999 revised January 11, 2006 Probability theory

More information

May 2015 Timezone 2 IB Maths Standard Exam Worked Solutions

May 2015 Timezone 2 IB Maths Standard Exam Worked Solutions May 015 Timezone IB Maths Standard Exam Worked Solutions 015, Steve Muench steve.muench@gmail.com @stevemuench Please feel free to share the link to these solutions http://bit.ly/ib-sl-maths-may-015-tz

More information

Algorithms for Uncertainty Quantification

Algorithms for Uncertainty Quantification Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example

More information

S n = x + X 1 + X X n.

S n = x + X 1 + X X n. 0 Lecture 0 0. Gambler Ruin Problem Let X be a payoff if a coin toss game such that P(X = ) = P(X = ) = /2. Suppose you start with x dollars and play the game n times. Let X,X 2,...,X n be payoffs in each

More information

Examples of random experiment (a) Random experiment BASIC EXPERIMENT

Examples of random experiment (a) Random experiment BASIC EXPERIMENT Random experiment A random experiment is a process leading to an uncertain outcome, before the experiment is run We usually assume that the experiment can be repeated indefinitely under essentially the

More information

REVIEW OF MAIN CONCEPTS AND FORMULAS A B = Ā B. Pr(A B C) = Pr(A) Pr(A B C) =Pr(A) Pr(B A) Pr(C A B)

REVIEW OF MAIN CONCEPTS AND FORMULAS A B = Ā B. Pr(A B C) = Pr(A) Pr(A B C) =Pr(A) Pr(B A) Pr(C A B) REVIEW OF MAIN CONCEPTS AND FORMULAS Boolean algebra of events (subsets of a sample space) DeMorgan s formula: A B = Ā B A B = Ā B The notion of conditional probability, and of mutual independence of two

More information

MATH 3510: PROBABILITY AND STATS June 15, 2011 MIDTERM EXAM

MATH 3510: PROBABILITY AND STATS June 15, 2011 MIDTERM EXAM MATH 3510: PROBABILITY AND STATS June 15, 2011 MIDTERM EXAM YOUR NAME: KEY: Answers in Blue Show all your work. Answers out of the blue and without any supporting work may receive no credit even if they

More information

Random variables (discrete)

Random variables (discrete) Random variables (discrete) Saad Mneimneh 1 Introducing random variables A random variable is a mapping from the sample space to the real line. We usually denote the random variable by X, and a value that

More information

Math/Stats 425, Sec. 1, Fall 04: Introduction to Probability. Final Exam: Solutions

Math/Stats 425, Sec. 1, Fall 04: Introduction to Probability. Final Exam: Solutions Math/Stats 45, Sec., Fall 4: Introduction to Probability Final Exam: Solutions. In a game, a contestant is shown two identical envelopes containing money. The contestant does not know how much money is

More information

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable

Lecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed

More information

Chapter 1 Review of Equations and Inequalities

Chapter 1 Review of Equations and Inequalities Chapter 1 Review of Equations and Inequalities Part I Review of Basic Equations Recall that an equation is an expression with an equal sign in the middle. Also recall that, if a question asks you to solve

More information

STATISTICS 4, S4 (4769) A2

STATISTICS 4, S4 (4769) A2 (4769) A2 Objectives To provide students with the opportunity to explore ideas in more advanced statistics to a greater depth. Assessment Examination (72 marks) 1 hour 30 minutes There are four options

More information

Moment Generating Functions

Moment Generating Functions MATH 382 Moment Generating Functions Dr. Neal, WKU Definition. Let X be a random variable. The moment generating function (mgf) of X is the function M X : R R given by M X (t ) = E[e X t ], defined for

More information

Sample Spaces, Random Variables

Sample Spaces, Random Variables Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted

More information