Probability. Table of contents
|
|
- Olivia Heath
- 5 years ago
- Views:
Transcription
1 Probability Table of contents 1. Important definitions 2. Distributions 3. Discrete distributions 4. Continuous distributions 5. The Normal distribution 6. Multivariate random variables 7. Other continuous distributions L. Trapani MSc Induction - Probability 1
2 Overview After having examined descriptive statistics in TN 2, we are now ready to tackle the foundations of probability and random variables, covering the tools which are necessary for regression L. Trapani MSc Induction - Probability 2
3 1 Probability: some important definitions The following definitions are important since they form the vocabulary that is commonly used in statistics L. Trapani MSc Induction - Probability 3
4 Definition: outcomes The first important notion that needs to be introduced is that of outcome, which is the building block for probability theory Any experiment can produce a certain number of results, which will be termed the outcomes of the experiment Examples can be very simple or very complicated: Tossing a coin produces the two outcomes {H,T}, head or tail Rolling a dice can result in 6 possible outcomes, namely {1,2,3,4,5,6} The random extraction of a real number (using e.g. Excel) in the interval [0,1] can result in a quite large number of outcomes L. Trapani MSc Induction - Probability 4
5 Definition: sample space The set of all possible outcomes is called the sample space In the previous examples, the sample spaces are, respectively: The set {H,T} which contains only two elements; The set {1,2,3,4,5,6} which contains 6 elements The interval [0,1] which contains an infinite number of elements L. Trapani MSc Induction - Probability 5
6 Definition: events Events are combinations of elementary outcomes, or, to phrase this alternatively, combinations of elements in the sample space of an experiment It makes sense to consider events: consider the case of rolling a dice one could be interested in betting on a single outcome, i.e. one may wish to bet on 4 to come out but it is also possible that one may wish to bet on the more complicated event that an even number comes out this would mean betting on the subset {2,4,6} of the sample space L. Trapani MSc Induction - Probability 6
7 Thus, as said events are combinations of elementary outcomes, which enable the statistician to work out the probability associated to all possible results of an experiment, not merely the elementary outcomes Other than all the possible combinations of all the possible elementary outcomes, there are two special events that are considered in probability theory: the so-called null event, represented as an empty set: this means that as a result of the experiment nothing happens (roughly speaking, you roll the dice and the dice disappears) the sample space itself is an event: it means that as a result of the experiment something happens (rather obviously, this event happens for sure) L. Trapani MSc Induction - Probability 7
8 The collection of all possible events (including the sample space itself and the null event) is also referred to as the σ-algebra of the experiment Probabilities will be calculated with respect to events Note that identifying all possible events could be tedious Particularly, it could be proven that if the number of elementary outcomes is n, then the total number of possible events is 2 n there are 4 events when one flips a coin: {nothing happens}, {head}, {tail} and {either head or tail} there are as many as 64 events when one rolls a dice there is an infinite number of events when sampling randomly from the interval [0,1] L. Trapani MSc Induction - Probability 8
9 Definition: random variable Before defining probabilities it is necessary to introduce a useful, albeit rather abstract, notion: the concept of random variable Formally, let Σ denote the σ-algebra of an experiment and let w denote one event in Σ Then we can define the random variable X as a function that is fed with an event in Σ and returns a real number; formally X X : Σ R ( w) R for all w Σ L. Trapani MSc Induction - Probability 9
10 Much less rigorously, but more intuitively, a random variable is a device that sticks a bar-code onto each and every event: all events in Σ must be processed by the random variable each event will have its own unique identification code the notion of random variable is formal, but convenient: instead of describing events as e.g. I flip a coin and I get either head or tail, it provides a concise representation, by means of a number, of each event L. Trapani MSc Induction - Probability 10
11 Definition: discrete versus continuous random variables The examples discussed above (coin, dice, random number generator) differ in one important point: the dimension of the sample space (and, therefore, the dimension of the σ-algebra) When the number of elements in the sample space (and thus in the σ-algebra) is either finite or at least countable, the corresponding random variable is said to be a discrete random variable Otherwise, if the elements in the sample space are not countable (not merely infinity, but so many that cannot even be counted), the random variable is referred to as a continuous random variable L. Trapani MSc Induction - Probability 11
12 Definition: probability Consider an experiment and its associated random variable X it s not important whether X is continuous or discrete, but for simplicity assume X is discrete Then the function P(X=x) is the probability that X takes the value x, or the likelihood that the event represented by x happens if P(X=x) is a number in the interval [0,1] the sum of the probabilities across all outcomes is 1 this corresponds to the sample space, i.e. the event something happens, and something should happen for sure and some other higher level conditions L. Trapani MSc Induction - Probability 12
13 Roughly speaking, to calculate the probability of an event (which is ultimately what one wants to do) it is necessary to start from the outcomes, define events and random variables, then calculate P The expression P(X=x)=0.35, for example, reads the chance that X takes the value x, and thus the chance that the event represented by x happens, is 35% L. Trapani MSc Induction - Probability 13
14 2 Distributions Distributions are another important building block in order to compute probabilities L. Trapani MSc Induction - Probability 14
15 Cumulative distribution functions L. Trapani MSc Induction - Probability 15
16 Exhibit 1 Cumulative probability distribution and the additivity of probability (the probability of sales growing between 20 and 30% 34%) Prob % 0% 10% 20% 30% 40% 50% L. Trapani MSc Induction - Probability 16
17 Cumulative distribution functions We can define the cumulative distribution function (CDF; also known as the distribution function) of a random variable X, as the probability that, in a given random experiment, the random variable X will take a value (also known as a realization) which is equal or less than (x) Formally, the CDF is a function F such that F(x)=P[X<x] Thus, one should expect that F(x) takes values in the interval [0,1] Note that here (x) indicates values on the x-axis (see Exhibit 1) L. Trapani MSc Induction - Probability 17
18 The additivity of probability The CDF can never decrease because this would imply negative probabilities, and probabilities are measured on a [0,1] scale Consider the following example, illustrated in Exhibit 1: you estimate that sales of a given product have a 50% probability of growing by 20% or less over the next year also, you estimate that the probability of sales growing by 30% or less cannot be lower than 50% L. Trapani MSc Induction - Probability 18
19 The additivity of probability This is an intuitive consequence of the additivity of probability The probability of a sales growth of 30% or less equals the probability of sales growing at 20% or less plus the probability of sales growing between 20 and 30% L. Trapani MSc Induction - Probability 19
20 Formally, this means that let b be a real number, and a be another real number such that a<b then the event X<b can be decomposed as {X<a OR a<x<b} intuition leads to [ X < b] = P[ X < a] + P[ a < X b] P < and therefore F ( b) = F( a) + P[ a < X < b] [ < X < b] = F( b) F( a) P a L. Trapani MSc Induction - Probability 20
21 Exhibit 1 Cumulative probability distribution and the additivity of probability (the probability of sales growing between 20 and 30% 34%) Prob % 0% 10% 20% 30% 40% 50% L. Trapani MSc Induction - Probability 21
22 3 Discrete distributions. Moments The general notion of CDF will be spelt out more rigorously, by making a distinction between discrete and continuous random variables. The binomial distribution is used to illustrate the features of the CDF of a discrete random variable and also to introduce the concepts of expected value, variance and standard deviation of discrete probability distribution The binomial distribution has a large number of applications in business, especially finance (binomial models for the pricing of options and for investment valuation) L. Trapani MSc Induction - Probability 22
23 Discrete distributions We have been using discrete random variables with the cointossing and dice-rolling examples We know that a discrete random variable can take only finite values Each of the possible values can be attributed a probability measure, known as probability mass function (PMF) The probabilities attributed to all the possible outcomes must clearly add up to one because probability is measured on a [0,1] scale L. Trapani MSc Induction - Probability 23
24 Binomial distribution This distribution calculates the probabilities of the possible outcomes in problems: with a fixed number of tests or trials (e.g. we flip a coin a few times) when the possible outcome of a single trial can be only success (X = 1) or failure (X = 0) when trials are independent (non correlated) and when the probability of success (p) is constant throughout the experiment p = Prob [X = 1] L. Trapani MSc Induction - Probability 24
25 The binomial distribution is a parametric distribution, with two parameters: n = number on trials p = probability of success. In the simple case of tossing a fair coin: P[X = 1 (head)] = 0.5 As a preliminary example, consider a binomial random variable with parameters (n = 10, p = 0.5), as visualized in Exhibits 2, 3 and 4 Probabilities were computed using the BINOMDIST Excel function L. Trapani MSc Induction - Probability 25
26 Exhibit 2 Probability mass function and cumulative probability of tossing 10 times a fair coin (n = 10, p = 50%) No of heads Prob mass Cumulative prob % % % % % % % % % % % % % % % % % % % % % % L. Trapani MSc Induction - Probability 26
27 Exhibit 3 Binomial probability mass function of tossing 10 times a fair coin (n = 10, p = 50%). 30% Prob 20% 10% 0% Successes L. Trapani MSc Induction - Probability 27
28 Exhibit 4 Cumulative probability of tossing 10 times a fair coin (n = 10, p = 50%). 100% Prob 80% 60% 40% 20% 0% Successes L. Trapani MSc Induction - Probability 28
29 Observe the relationship between the CDF (exhibit 4) and the PMF (exhibit 3): they are actually completely equivalent, in that they are defined up to a one-to-one transformation Consider e.g. the probability that X=6 this can be calculated immediately from the PMF, and the table in exhibit 2 shows that P[X=6]=20.5%; however, exactly the same result could be obtained by employing the CDF: P [ X = 6] = P[ X 6] P[ X 5] = F ( 6) F( 5) = 82.8% 62.3% = 20.5% L. Trapani MSc Induction - Probability 29
30 It is interesting to note what the CDF looks like: its graph is not continuous, it has discontinuities at each event the size of each jump is the probability that the corresponding event will take place This is a property of discrete random variable: it can be said that a discrete random variable is a random variable whose CDF is piecewise continuous (actually, piecewise flat) with jumps corresponding to each event Also, note that as previously observed the CDF = 0 on the left of the lowest possible outcome the CDF = 1 on the right of the highest possible outcome L. Trapani MSc Induction - Probability 30
31 Binomial distribution We can now study some features of the binomial distributions We can see that, with p = 0.5, the probability mass function is perfectly symmetrical around the mean see exhibit 3 However, the binomial distribution can be used to model random variables where p 0.5 the case p=0.5 is just a possible example and it is a good representation for cases such flipping a fair coin, however, e.g. in management applications it is often the case that p needs not be 0.5. Or in finance, if you buy 10 shares and all you are interested in is whether you make or lose money L. Trapani MSc Induction - Probability 31
32 Binomial distribution In exhibits 5 & 6 the probability of success are respectively p=0.4 and p=0.7, This generates skewed mass functions L. Trapani MSc Induction - Probability 32
33 Exhibit 5 Binomial probability mass function for (n = 10, p = 40%). 30% Prob 20% 10% 0% Successes L. Trapani MSc Induction - Probability 33
34 Exhibit 6 Binomial probability mass function for (n = 10, p = 70%). 30% Prob 20% 10% 0% Successes L. Trapani MSc Induction - Probability 34
35 Moments of a discrete random variable In what follows, we shall give some very general definitions that apply to all possible discrete random variables, and subsequently apply them to the binomial distribution Formally, consider a random variable which can take values X 1,,X n ; we define uncentered moments of order k, denoted as E[X k ], as E [ ] k X = p( x ) n j= 1 centered (around the mean μ) moments of order k, denoted as E[(X-μ) k ], as E i x k i [( ) ] n k X μ = p( x )( x μ) j= 1 i i k L. Trapani MSc Induction - Probability 35
36 Thus, moments are weighted averages of X k, where weights are given by the probability that the random variable X takes the value X j Therefore, moments are similar to descriptive statistics: think of the centered moment of order k=2; this uses squared deviations (X j -μ) 2, and thus it is possible to make a parallel with the sample variance however, whilst the sample variance does not weigh the terms (X j -μ) 2, moments attach a small weight (i.e. a small probability) onto unlikely events and a bigger weight onto more likely events L. Trapani MSc Induction - Probability 36
37 Expected value of a discrete random variable The expected value is defined as the uncentered moment of order k=1, and the equation is E n [ X ] = p( ) j= 1 x j x j L. Trapani MSc Induction - Probability 37
38 The expected value is the analogue to the mean; only, in this case, each event is weighted by its probability Note that the expected value needs not be one of the values taken by the random variable, and thus it should not be interpreted as the most likely value to come out of the experiment This can be seen if one calculates the expected outcome from rolling a dice, which is 3.5 (a number that can obviously never come out) We have done together the exercise based on EVENT X=1 X=2 X=3 probability L. Trapani MSc Induction - Probability 38
39 Possible interpretations are: If each event is viewed as a mass point whose weight is given by the probability, then the expected value is the equilibrium point of the probability distribution The expected value is a rational prediction of the outcome of an experiment L. Trapani MSc Induction - Probability 39
40 Variance of a discrete random variable The variance of a discrete random variable is the centered moment of order 2 It is defined as the weighted average of the squared deviations of the possible n outcomes from their expected value Var [ ] n 2 = p( x )( ) j x j μ ( X ) = E ( X μ) j= 1 2 L. Trapani MSc Induction - Probability 40
41 Some algebra could lead to the following (quite useful and widely used) alternative formulation in terms of the expected value and the second order uncentered moment Var = E = E = E [ ] 2 ( X ) = E ( X μ) = [ ] 2 2 X 2μX + μ [ ] [ ] [ ] 2 2 X 2μE X + E μ [ ] 2 2 X μ L. Trapani MSc Induction - Probability 41
42 Last, like in the case of the variance as a descriptive statistic, it is common to employ the standard deviation as well, defined as σ = Var( X ) L. Trapani MSc Induction - Probability 42
43 The variance is the analogue to the sample variance; only, in this case, each event is weighted by its probability, like in the expected value case Note that the variance is always non-negative particularly, the variance is equal to zero if there is no deviation from the mean, i.e. if there is only one event tat occurs for sure (with probability one) The variance is interpreted as the degree of dispersion around the mean: the smaller the variance, the less likely extreme events are to happen thus, the variance is customarily employed as a measure of risk/uncertainty L. Trapani MSc Induction - Probability 43
44 Linear transformations Main question: what are expected value and variance of ax + b L. Trapani MSc Induction - Probability 44
45 L. Trapani MSc Induction - Probability 45 From the formula that defines the expected value, it is possible to observe that the expected value of a linear transformation of a random variable is the linear transformation of the expected value, i.e. [ ] ( )[ ] ( ) ( ) ( ) ( ) [ ] b X ae x p b x x p a b x p ax x p b ax x p b ax E n j j n j j j n j j n j j j n j j j + = + = + = + = + = = = = =
46 Also, the following useful formula for variance the of a linear transformation of a random variable could be proved ( ) 2 ax + b = a Var( X ) for all a b Var, This formula states that the variance increases quadratically with changes of scale of magnitude a and that it is invariant to shifts of any size b Think of the following intuitive argument: (1) when one buys 10 shares, one s risk is 100 times as much (don t put all your eggs in just one basket) (2) adding an asset with no risk at all can not make your portfolio more or less risky L. Trapani MSc Induction - Probability 46
47 Higher order moments Two frequently employed centered moments are the skewness, which is the centered moment of order 3 E [( ) ] n 3 X μ = p( x )( ) j x j μ the kurtosis, which is the centered moment of order 4 E j= 1 [( ) ] n 4 X μ = p( x )( ) j x j μ j= L. Trapani MSc Induction - Probability 47
48 The skewness is commonly interpreted as a measure of asymmetry of the distribution the distribution is symmetric if the skewness is equal to zero (and vice versa, skewness is equal to zero if the distribution is symmetric) a positive skewness implies that the right tail of the distribution is longer than the left tail, or, in more intuitive terms, that extreme positive events are more likely than extreme negative events likewise, a negative skewness entails a longer left tails L. Trapani MSc Induction - Probability 48
49 The kurtosis is interpreted as a measure of the heaviness of tails or, in more intuitive terms, as a measure of the impact of extreme events Note that the kurtosis is never negative: when the kurtosis is close to zero, this means that all events are clustered together around the mean: there are no extreme events the larger the kurtosis, the larger the likelihood of significantly extreme events (no matter whether these are positive or negative) thus, kurtosis is interpreted as another possible measure of risk L. Trapani MSc Induction - Probability 49
50 Quite often, these moments are rescaled by the variance, being defined as Skewness: 3 n X μ 1 E = [ Var( X )] 3/ 2 σ j= 1 p ( )( μ) x j x j 3 Kurtosis: 4 n X μ 1 E = [ Var( X )] 2 σ j= 1 p ( )( μ) x j x j 4 L. Trapani MSc Induction - Probability 50
51 4 Continuous random variables. Probability density functions In this section we shall examine how probability is defined for continuous random variables. This is an important issue, with applications in finance. Whilst the mathematics can be slightly more complicated than in the discrete random variable case, continuous random variables are used with extreme frequency and arise very often indeed in statistics and econometrics. L. Trapani MSc Induction - Probability 51
52 Continuous probability distributions Exhibit 7 shows a the CDF of a continuous random variable (it s a normal standard distribution) defined over the real line (the x- axis) L. Trapani MSc Induction - Probability 52
53 Exhibit 7 Continuous probability distribution. The numbers for this figure are calculated using the Excel function NORMSDIST L. Trapani MSc Induction - Probability 53
54 It is interesting to note that in this case the CDF is indeed continuous This is a property of continuous random variables: it can be said that a continuous random variable is a random variable whose CDF is continuous with no jumps etc Note that the probability of X being smaller or equal to a value (x) increases continuously with (x) moving from (- ) to (+ ) Note also that lim lim x x F F ( x) = 1 ( x) = 0 L. Trapani MSc Induction - Probability 54
55 The CDF has the same use and interpretation as in the discrete case It could be interesting to evaluate the probability that the random variable X is equal to a value x, i.e. P[X=x] This is equivalent to finding out the probability mass function, and it can be done by using the CDF Particularly, try to calculate P[X=1]; this would be given by P [ X = 1 ] = F( ) F( ) = 0 L. Trapani MSc Induction - Probability 55
56 Thus, the probability that the continuous random variable X is equal to a single value is always zero, i.e. P[X=x]=0 for all x We must recall that the real line (the x-axis) has an infinite number of points in any interval, no matter how small (for example from to ) Therefore, as an intuitive argument, it is impossible to attribute a probability value to any single point, say to X = 1.51 (the sum of such probabilities would be ) Therefore, continuous distributions do not have a mass probability function We can of course attribute a probability measure to intervals, such as from 1.50 to 1.60 L. Trapani MSc Induction - Probability 56
57 Probability density function A widely employed notion is that of probability density function (PDF) The PDF has a one-to-one correspondence with the CDF, and therefore it serves the same purpose and contains the same information Formally, the PDF of a random variable X is a function f(x) defined as d dx ( x) F( x) f = L. Trapani MSc Induction - Probability 57
58 Thus (once again formally), the PDF contains the same information as the CDF In order to see this, recall that the reverse operation of differentiation is the integral Thus, the probability that the random variable X lies between the real numbers a and b with a<b is equivalently calculated as P P [ a < X < b ] = F ( b ) F ( a ) [ a < X < b ] = f ( x ) a b dx L. Trapani MSc Induction - Probability 58
59 Exhibit 8 Probability density function of the standard normal distribution (μ = 0, σ = 1) Area = L. Trapani MSc Induction - Probability 59
60 The integral of the PDF also equals the area under the curve It s therefore common practice to visualize probabilities by shading the relevant area underneath the PDF Recall that probability is measured from zero to one [0, 1]. Formally, this means that in order for f(x) to be a PDF it must hold + f ( x) dx = 1 This implies that the area under the whole probability density function must equal one, as visualized in exhibit 8 L. Trapani MSc Induction - Probability 60
61 It should be emphasized that the function f(x) does NOT return the probability that the random variable X is equal to x: as we know, it holds that P[X=x]=0 for all possible x The PDF is merely a tool to compute the probability that X belongs to an interval; such calculation is performed by means of integrals L. Trapani MSc Induction - Probability 61
62 Moments of a continuous random variable Just like in the discrete case, moments play a pivotal role here as well Interpretation of moments is the same as in the discrete random variable case, albeit formulas differ uncentered moments of order k are defined as E [ ] k k X x f ( x) centered moments of order k are defined as E + = dx + [( ) ] k k X μ = ( x μ) f ( x) dx L. Trapani MSc Induction - Probability 62
63 The following moments are commonly employed Expected value + [ X ] = xf ( x ) μ = E dx Variance 2 σ + [( ) ] 2 2 X μ = ( x μ) f ( x) = E dx L. Trapani MSc Induction - Probability 63
64 L. Trapani MSc Induction - Probability 64 Skewness Kurtosis ( ) + = dx x f x X E 3 3 σ μ σ μ ( ) + = dx x f x X E 4 4 σ μ σ μ
65 5 Examples of continuous distribution. The normal distribution In this section we shall see the most frequently employed continuous distributions These are frequently employed in statistics; the most important one, that deserves particular attention, is the normal distribution L. Trapani MSc Induction - Probability 65
66 The normal distribution A random variable X which can take values over the whole real axis is said to have a normal distribution with mean μ and variance σ 2 if the equation of its PDF is f 1 1 x μ ( ) 2 σ x = e for x (, + ) 2πσ 2 2 L. Trapani MSc Induction - Probability 66
67 Exhibit 9 Normal density functions: f(z) is the standard normal, the other two curves are characterized by f(μ, σ) f(z) f(0, 0.5) f(1, 0.5) L. Trapani MSc Induction - Probability 67
68 Thus, the normal distribution is characterized by only two parameters: the mean μ and the standard deviation σ We denote a random variable following the normal distribution as X~N[μ,σ 2 ] Note that: the normal density function is perfectly symmetrical around the mean the higher the standard deviation the more the curve is spread around the mean L. Trapani MSc Induction - Probability 68
69 The standard normal distribution In the special case when the mean is zero (μ = 0) and the standard deviation is one (σ = 1), the distribution is known as standard normal distribution The equation in this cases reduces to f x ( ) 2 x = e for x (, + ) 2π L. Trapani MSc Induction - Probability 69
70 There is unfortunately no explicit equation for the CDF of the normal However, probabilities as P[X<x] can be computed by employing numerical methods this can be done with in Excel with the function NORMDIST (and NORMSDIST for the standard normal ) The cumulative normal distribution is usually denoted with F(x), and with Ф(x) for the standard normal L. Trapani MSc Induction - Probability 70
71 Properties of the normal distribution Moments it is not needed to calculate mean or variance, as these are given in the definition of the normal distribution: for a random variable X~N[μ,σ 2 ] we have E E [ X ] = μ ( X μ ) [ 2 ] 2 = σ L. Trapani MSc Induction - Probability 71
72 the normal distribution is symmetric around the mean μ, and therefore there is no need to calculate the skewness, which is zero it would be possible (but rather cumbersome) to calculate the kurtosis of X however, it can be shown that for each μ and σ 2 E 4 X μ σ = 3 L. Trapani MSc Induction - Probability 72
73 It is customary to employ the number 3 as a benchmark for kurtosis: random variables with kurtosis smaller than 3 have a steeper distribution than the normal, and the PDF is more clustered around the mean (these distributions are referred to as leptokurtic) thin tails random variables with kurtosis bigger than 3 have a flatter distribution than the normal, and the PDF is more dispersed around the mean (these distributions are referred to as platikurtic) fat tails L. Trapani MSc Induction - Probability 73
74 Probability under the tails The normal PDF is perfectly symmetrical, and therefore the probabilities under the left and the right tails are equal For example, the areas identified in exhibit 10 contain the following probabilities: Area Probability Left of -2 or right of Between -2 and Left of -1 or right of Between -1 and L. Trapani MSc Induction - Probability 74
75 Exhibit 10 Normal standard density function. Standard deviations are on the x-axis. Calculated with Excel function NORMSDIST L. Trapani MSc Induction - Probability 75
76 Alternatively, one can identify some probability values (usually: 1%, 2.5%, 5%) and then calculate the number of standard deviations from the mean (left or right) that correspond to these values From the table below, e.g., one can realize that 5% of the probability mass is in the interval [x< μ-2σ ] OR [x> μ+2σ] thus, the interval [μ-2σ, μ+2σ] contains 95% of the mass probability another name for these intervals is confidence intervals Probability N of std from mean 0.100% % % % L. Trapani MSc Induction - Probability 76
77 Standardization Most often, in order to compute probabilities from a random variable X~N[μ,σ 2 ], it is common to standardise X, i.e. to transform X into a N[0,1] variable Standardization is performed as follows subtract the mean μ from X, thereby getting the new, zero mean random variable X-μ rescale X-μ by the standard deviation, getting the new, unit variance random variable (X-μ)/σ then the new random variable (X-μ)/σ has a standard normal distribution L. Trapani MSc Induction - Probability 77
78 L. Trapani MSc Induction - Probability 78 This could be quickly proved recalling the formulas for expected value and variance of linear transformation of random variables This procedure can be used e.g. to calculate the probability P[a<X<b] for a given random variable X~N[μ,σ 2 ] [ ] = < < = < < σ μ σ μ π σ μ σ μ σ μ b a u du e b X a P b X P a
79 6 A note on multivariate random variables In this section we shall briefly see how to represent and calculate the probabilities for more than just one random variable L. Trapani MSc Induction - Probability 79
80 Multivariate random variables So far, the attention has been focused onto single, univariate random variables However, it may be interesting to consider the joint behaviour of two or more random variables e.g. consider a portfolio of 10 shares: returns to each of these shares could be viewed as 10 different random variables, and the portfolio is described by the joint behaviour of the 10 shares In what follows, we shall focus on the bivariate case, wherein only two random variables are analysed jointly, say X and Y; most of our analysis will consist of definitions L. Trapani MSc Induction - Probability 80
81 Some useful definitions the joint CDF of X and Y is a function F(x,y) which returns the probability that, at the same time, X<x AND Y<y formally: F ( x, y) = P[ X < x ANDY < y] L. Trapani MSc Induction - Probability 81
82 the joint PDF of X and Y is a function f(x,y) defined as f ( x, y) = 2 ( x, y) F x y the joint PDF can be used to calculate probabilities such as [ a < X < b, c < Y < d ] = f ( x y) P, b a d c dxdy L. Trapani MSc Induction - Probability 82
83 Independence A very important definition is that of independence between two random variables X and Y Intuitively, X and Y are said to be independent if the probability that X lies in any interval [a,b] is not affected by the value(s) taken by Y thus, knowing the behaviour of Y is not useful to estimate P[a<X<b] Formally, the two random variables X and Y are independent if their joint PDF is given by the product of the PDF of each variable, i.e. f ( x, y) = f ( x) f ( y) X Y L. Trapani MSc Induction - Probability 83
84 Conditioning Conditioning is an important notion that becomes useful in regression analysis Intuitively, when X and Y are NOT independent, knowing what values Y takes can indeed help to know the probability that X lies within an interval [a,b] The notation P[a<X<b Y=y] reads the probability that X lies in [a,b] given/knowing that Y has taken the value y The tool to compute such probabilities is the so-called conditional PDF of X given Y, denoted as f X Y (x y) and used as follows [ a < X < b Y = y] f ( x y) P = X Y b a dx L. Trapani MSc Induction - Probability 84
85 Moments of multivariate distributions Some useful rules on the moments of multivariate distributions: Expectations: it is worth pointing out that the expected value of the sum of random variables is the sum of the expected values, i.e. [ X + Y ] = E[ X ] E[ Y ] E + L. Trapani MSc Induction - Probability 85
86 Covariance: like in the descriptive statistics case, covariance is a measure of linear association between two random variables: E = the definition is Cov(X,Y)=E[(X-μ X )(Y-μ Y )], where μ represents the expected value the formula is [( X μ )( Y μ )] = ( x μ )( y μ ) f ( x, y) X Y [ ] μ ( ) XμY = xyf x, y dxdy μx Y E XY X Y μ dxdy L. Trapani MSc Induction - Probability 86
87 An important fact to bear in mind about the covariance is that if two random variables X and Y are independent, then their covariance is equal to zero However, the opposite does not generally hold i.e. covariance equal to zero is not enough to have independence Just like in the descriptive statistics case, one could define the correlation between X and Y, a number between -1 and 1 which represents the degree of linear association between X and Y ρ XY ( X, Y ) ( X ) Var ( Y ) Cov =, ρ XY Var [ 1,1] L. Trapani MSc Induction - Probability 87
88 Variance: it is important to remember that the variance of the sum of two value of the sum of independent random variables is given by ( X ± Y ) = Var( X ) + Var( Y ) 2Cov( X, Y ) Var ± here, independence would entail ( X ± Y ) = Var( X ) Var( Y ) Var + L. Trapani MSc Induction - Probability 88
89 Note that even though all notation for the multivariate case has been presented with respect to continuous random variables, it can be extended to the discrete case For example: consider flipping a coin, and introduce the random variable X which is equal to 1 if HEAD comes out and 0 if TAIL comes out of course P[X=1]=P[X=0]=0.5 thus, the expected value of X is E[X]=0.5 in light if this, consider a binomial random variable Y which counts the number of heads in 10 independent trials then E[Y]=E[X]+.+E[X]=10*E[X]=5 L. Trapani MSc Induction - Probability 89
90 7 Other examples of continuous distribution After discussing the normal distribution, in this section we shall see three other frequently employed continuous distributions All these random variables are derived from the standard normal, which shows the importance of this distribution L. Trapani MSc Induction - Probability 90
91 Other frequently employed distributions In statistics, there are some other distributions that arise quite frequently These are defined as transformations of the standard normal distribution, and they normally have rather complicated PDFs (which however do not need to be remembered) The most frequently used distributions are: the chi-squared distribution Student s t distribution Fisher s F distribution L. Trapani MSc Induction - Probability 91
92 The chi-squared distribution Consider the random variable X~N[0,1] The new random variable Y=X 2 is said to have a chi-squared distribution with one degree of freedom, denoted as χ 2 (1) More generally, consider n independent standard normal random variables X 1,,X n ; then the new variable Y = n j= 1 2 X j is said to have a chi-squared distribution with n degree of freedom, denoted as χ 2 (n) L. Trapani MSc Induction - Probability 92
93 Exhibit 11. The chi-square distribution for various degrees of freedom k L. Trapani MSc Induction - Probability 93
94 Some notes: independence is a crucial requirement; the chi-square takes only positive values (it s a square and thus no negative values are allowed); the chi-square has mean different from zero (and positive) and is not symmetric around the mean, as it can be seen in Exhibit 11 it could be proved that if X has a χ 2 (n) distribution, then E[X]=n it could be proved that if X has a χ 2 (n) distribution and n is very large, then the PDF of X is very similar to the PDF of a normally distributed random variable L. Trapani MSc Induction - Probability 94
95 Other distributions: Student s t distribution Consider two independent random variables X and Y, such that X~N[0,1] and Y has a χ 2 (n) distribution Then the new random variable Z defined as Z = X Y n is said to have a Student s t distribution with n degrees of freedom, denoted as t n L. Trapani MSc Induction - Probability 95
96 Exhibit 12. Student s t distribution for different values of the degree of freedom k L. Trapani MSc Induction - Probability 96
97 Some notes: the Student distribution can take positive and negative values; the Student distribution has mean strictly equal to zero is perfectly symmetric around the mean, just like the normal L. Trapani MSc Induction - Probability 97
98 it could be proved that if Z has a t n distribution, then its kurtosis is always greater than 3 thus, Student distribution could be useful to represent those situations where there are frequently extreme events it could be proved that if Z has a t n distribution and n is very large, then the PDF of Z is very similar to the PDF of a normally distributed random variable L. Trapani MSc Induction - Probability 98
99 Exhibit 13. Comparison between normal and Student s t distribution normal distribution t-distribution L. Trapani MSc Induction - Probability 99
100 Other distributions: Fisher s F distribution Consider two independent random variables X and Y, such that X has a χ 2 (n 1 ) distribution Y has a χ 2 (n 2 ) distribution Then the new random variable Z defined as X Z = Y n n 1 2 is said to have a Fisher s F distribution with n 1 and n 2 degrees of freedom, denoted as F n1,n2 L. Trapani MSc Induction - Probability 100
101 Exhibit 14. Fisher s F distribution for various values of the degree of freedom d 1 (numerator) and d 2 (denominator) L. Trapani MSc Induction - Probability 101
102 Some notes: the F distribution can take only positive values: it s the ratio of two random variables that can take only positive values has non zero mean and is not symmetric the quantities n 1 and n 2 are sometimes referred to as, respectively the degree of freedom of the numerator the degree of freedom of the denominator the F distribution is used very frequently in econometrics whenever hypothesis testing is carried out L. Trapani MSc Induction - Probability 102
Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables
Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results
More informationSummary statistics. G.S. Questa, L. Trapani. MSc Induction - Summary statistics 1
Summary statistics 1. Visualize data 2. Mean, median, mode and percentiles, variance, standard deviation 3. Frequency distribution. Skewness 4. Covariance and correlation 5. Autocorrelation MSc Induction
More informationIAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES
IAM 530 ELEMENTS OF PROBABILITY AND STATISTICS LECTURE 3-RANDOM VARIABLES VARIABLE Studying the behavior of random variables, and more importantly functions of random variables is essential for both the
More information1: PROBABILITY REVIEW
1: PROBABILITY REVIEW Marek Rutkowski School of Mathematics and Statistics University of Sydney Semester 2, 2016 M. Rutkowski (USydney) Slides 1: Probability Review 1 / 56 Outline We will review the following
More informationWeek 1 Quantitative Analysis of Financial Markets Distributions A
Week 1 Quantitative Analysis of Financial Markets Distributions A Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 October
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions
More informationPreliminary Statistics. Lecture 3: Probability Models and Distributions
Preliminary Statistics Lecture 3: Probability Models and Distributions Rory Macqueen (rm43@soas.ac.uk), September 2015 Outline Revision of Lecture 2 Probability Density Functions Cumulative Distribution
More informationStatistical inference
Statistical inference Contents 1. Main definitions 2. Estimation 3. Testing L. Trapani MSc Induction - Statistical inference 1 1 Introduction: definition and preliminary theory In this chapter, we shall
More informationL2: Review of probability and statistics
Probability L2: Review of probability and statistics Definition of probability Axioms and properties Conditional probability Bayes theorem Random variables Definition of a random variable Cumulative distribution
More informationBivariate distributions
Bivariate distributions 3 th October 017 lecture based on Hogg Tanis Zimmerman: Probability and Statistical Inference (9th ed.) Bivariate Distributions of the Discrete Type The Correlation Coefficient
More informationLecture 2: Review of Probability
Lecture 2: Review of Probability Zheng Tian Contents 1 Random Variables and Probability Distributions 2 1.1 Defining probabilities and random variables..................... 2 1.2 Probability distributions................................
More informationStatistical Concepts. Distributions of Data
Module : Review of Basic Statistical Concepts. Understanding Probability Distributions, Parameters and Statistics A variable that can take on any value in a range is called a continuous variable. Example:
More informationJoint Probability Distributions and Random Samples (Devore Chapter Five)
Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete
More informationLecture 10: Probability distributions TUESDAY, FEBRUARY 19, 2019
Lecture 10: Probability distributions DANIEL WELLER TUESDAY, FEBRUARY 19, 2019 Agenda What is probability? (again) Describing probabilities (distributions) Understanding probabilities (expectation) Partial
More informationSummary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016
8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying
More informationFourier and Stats / Astro Stats and Measurement : Stats Notes
Fourier and Stats / Astro Stats and Measurement : Stats Notes Andy Lawrence, University of Edinburgh Autumn 2013 1 Probabilities, distributions, and errors Laplace once said Probability theory is nothing
More informationBasics on Probability. Jingrui He 09/11/2007
Basics on Probability Jingrui He 09/11/2007 Coin Flips You flip a coin Head with probability 0.5 You flip 100 coins How many heads would you expect Coin Flips cont. You flip a coin Head with probability
More informationACE 562 Fall Lecture 2: Probability, Random Variables and Distributions. by Professor Scott H. Irwin
ACE 562 Fall 2005 Lecture 2: Probability, Random Variables and Distributions Required Readings: by Professor Scott H. Irwin Griffiths, Hill and Judge. Some Basic Ideas: Statistical Concepts for Economists,
More informationSample Spaces, Random Variables
Sample Spaces, Random Variables Moulinath Banerjee University of Michigan August 3, 22 Probabilities In talking about probabilities, the fundamental object is Ω, the sample space. (elements) in Ω are denoted
More informationLecture 2: Repetition of probability theory and statistics
Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationRandom Variables. Random variables. A numerically valued map X of an outcome ω from a sample space Ω to the real line R
In probabilistic models, a random variable is a variable whose possible values are numerical outcomes of a random phenomenon. As a function or a map, it maps from an element (or an outcome) of a sample
More informationAlgorithms for Uncertainty Quantification
Algorithms for Uncertainty Quantification Tobias Neckel, Ionuț-Gabriel Farcaș Lehrstuhl Informatik V Summer Semester 2017 Lecture 2: Repetition of probability theory and statistics Example: coin flip Example
More informationMultivariate Distributions
IEOR E4602: Quantitative Risk Management Spring 2016 c 2016 by Martin Haugh Multivariate Distributions We will study multivariate distributions in these notes, focusing 1 in particular on multivariate
More informationContinuum Probability and Sets of Measure Zero
Chapter 3 Continuum Probability and Sets of Measure Zero In this chapter, we provide a motivation for using measure theory as a foundation for probability. It uses the example of random coin tossing to
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More information7 Random samples and sampling distributions
7 Random samples and sampling distributions 7.1 Introduction - random samples We will use the term experiment in a very general way to refer to some process, procedure or natural phenomena that produces
More informationProbability Theory and Simulation Methods
Feb 28th, 2018 Lecture 10: Random variables Countdown to midterm (March 21st): 28 days Week 1 Chapter 1: Axioms of probability Week 2 Chapter 3: Conditional probability and independence Week 4 Chapters
More informationBasic Probability Reference Sheet
February 27, 2001 Basic Probability Reference Sheet 17.846, 2001 This is intended to be used in addition to, not as a substitute for, a textbook. X is a random variable. This means that X is a variable
More informationThe Delta Method and Applications
Chapter 5 The Delta Method and Applications 5.1 Local linear approximations Suppose that a particular random sequence converges in distribution to a particular constant. The idea of using a first-order
More informationDiscrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations
EECS 70 Discrete Mathematics and Probability Theory Fall 204 Anant Sahai Note 5 Random Variables: Distributions, Independence, and Expectations In the last note, we saw how useful it is to have a way of
More informationMath 105 Course Outline
Math 105 Course Outline Week 9 Overview This week we give a very brief introduction to random variables and probability theory. Most observable phenomena have at least some element of randomness associated
More informationSingle Maths B: Introduction to Probability
Single Maths B: Introduction to Probability Overview Lecturer Email Office Homework Webpage Dr Jonathan Cumming j.a.cumming@durham.ac.uk CM233 None! http://maths.dur.ac.uk/stats/people/jac/singleb/ 1 Introduction
More informationJoint Probability Distributions, Correlations
Joint Probability Distributions, Correlations What we learned so far Events: Working with events as sets: union, intersection, etc. Some events are simple: Head vs Tails, Cancer vs Healthy Some are more
More informationStatistics and Econometrics I
Statistics and Econometrics I Random Variables Shiu-Sheng Chen Department of Economics National Taiwan University October 5, 2016 Shiu-Sheng Chen (NTU Econ) Statistics and Econometrics I October 5, 2016
More informationProbability. Lecture Notes. Adolfo J. Rumbos
Probability Lecture Notes Adolfo J. Rumbos October 20, 204 2 Contents Introduction 5. An example from statistical inference................ 5 2 Probability Spaces 9 2. Sample Spaces and σ fields.....................
More informationPCMI Introduction to Random Matrix Theory Handout # REVIEW OF PROBABILITY THEORY. Chapter 1 - Events and Their Probabilities
PCMI 207 - Introduction to Random Matrix Theory Handout #2 06.27.207 REVIEW OF PROBABILITY THEORY Chapter - Events and Their Probabilities.. Events as Sets Definition (σ-field). A collection F of subsets
More informationSTAT2201. Analysis of Engineering & Scientific Data. Unit 3
STAT2201 Analysis of Engineering & Scientific Data Unit 3 Slava Vaisman The University of Queensland School of Mathematics and Physics What we learned in Unit 2 (1) We defined a sample space of a random
More informationRandom Variables and Expectations
Inside ECOOMICS Random Variables Introduction to Econometrics Random Variables and Expectations A random variable has an outcome that is determined by an experiment and takes on a numerical value. A procedure
More informationBASICS OF PROBABILITY
October 10, 2018 BASICS OF PROBABILITY Randomness, sample space and probability Probability is concerned with random experiments. That is, an experiment, the outcome of which cannot be predicted with certainty,
More informationProbabilities & Statistics Revision
Probabilities & Statistics Revision Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 January 6, 2017 Christopher Ting QF
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationMath 416 Lecture 3. The average or mean or expected value of x 1, x 2, x 3,..., x n is
Math 416 Lecture 3 Expected values The average or mean or expected value of x 1, x 2, x 3,..., x n is x 1 x 2... x n n x 1 1 n x 2 1 n... x n 1 n 1 n x i p x i where p x i 1 n is the probability of x i
More informationLectures on Statistics. William G. Faris
Lectures on Statistics William G. Faris December 1, 2003 ii Contents 1 Expectation 1 1.1 Random variables and expectation................. 1 1.2 The sample mean........................... 3 1.3 The sample
More informationECON Fundamentals of Probability
ECON 351 - Fundamentals of Probability Maggie Jones 1 / 32 Random Variables A random variable is one that takes on numerical values, i.e. numerical summary of a random outcome e.g., prices, total GDP,
More information1 Probability theory. 2 Random variables and probability theory.
Probability theory Here we summarize some of the probability theory we need. If this is totally unfamiliar to you, you should look at one of the sources given in the readings. In essence, for the major
More informationThis exam is closed book and closed notes. (You will have access to a copy of the Table of Common Distributions given in the back of the text.
TEST #3 STA 5326 December 4, 214 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. (You will have access to
More informationChapter 4: An Introduction to Probability and Statistics
Chapter 4: An Introduction to Probability and Statistics 4. Probability The simplest kinds of probabilities to understand are reflected in everyday ideas like these: (i) if you toss a coin, the probability
More informationDiscrete Random Variables
CPSC 53 Systems Modeling and Simulation Discrete Random Variables Dr. Anirban Mahanti Department of Computer Science University of Calgary mahanti@cpsc.ucalgary.ca Random Variables A random variable is
More informationLecture 1: Probability Fundamentals
Lecture 1: Probability Fundamentals IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge January 22nd, 2008 Rasmussen (CUED) Lecture 1: Probability
More information1 INFO Sep 05
Events A 1,...A n are said to be mutually independent if for all subsets S {1,..., n}, p( i S A i ) = p(a i ). (For example, flip a coin N times, then the events {A i = i th flip is heads} are mutually
More informationRecitation 2: Probability
Recitation 2: Probability Colin White, Kenny Marino January 23, 2018 Outline Facts about sets Definitions and facts about probability Random Variables and Joint Distributions Characteristics of distributions
More informationSystem Identification
System Identification Arun K. Tangirala Department of Chemical Engineering IIT Madras July 27, 2013 Module 3 Lecture 1 Arun K. Tangirala System Identification July 27, 2013 1 Objectives of this Module
More informationReview of Probability Theory
Review of Probability Theory Arian Maleki and Tom Do Stanford University Probability theory is the study of uncertainty Through this class, we will be relying on concepts from probability theory for deriving
More informationWe introduce methods that are useful in:
Instructor: Shengyu Zhang Content Derived Distributions Covariance and Correlation Conditional Expectation and Variance Revisited Transforms Sum of a Random Number of Independent Random Variables more
More information1 Random variables and distributions
Random variables and distributions In this chapter we consider real valued functions, called random variables, defined on the sample space. X : S R X The set of possible values of X is denoted by the set
More informationBasic Probability. Introduction
Basic Probability Introduction The world is an uncertain place. Making predictions about something as seemingly mundane as tomorrow s weather, for example, is actually quite a difficult task. Even with
More informationSYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions
SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu
More information1 Presessional Probability
1 Presessional Probability Probability theory is essential for the development of mathematical models in finance, because of the randomness nature of price fluctuations in the markets. This presessional
More information18.05 Practice Final Exam
No calculators. 18.05 Practice Final Exam Number of problems 16 concept questions, 16 problems. Simplifying expressions Unless asked to explicitly, you don t need to simplify complicated expressions. For
More informationStatistics 100A Homework 5 Solutions
Chapter 5 Statistics 1A Homework 5 Solutions Ryan Rosario 1. Let X be a random variable with probability density function a What is the value of c? fx { c1 x 1 < x < 1 otherwise We know that for fx to
More informationReview of Statistics
Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and
More informationProbability and Statistics
Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT
More informationINF FALL NATURAL LANGUAGE PROCESSING. Jan Tore Lønning
1 INF4080 2018 FALL NATURAL LANGUAGE PROCESSING Jan Tore Lønning 2 Probability distributions Lecture 5, 5 September Today 3 Recap: Bayes theorem Discrete random variable Probability distribution Discrete
More informationUniversity of Regina. Lecture Notes. Michael Kozdron
University of Regina Statistics 252 Mathematical Statistics Lecture Notes Winter 2005 Michael Kozdron kozdron@math.uregina.ca www.math.uregina.ca/ kozdron Contents 1 The Basic Idea of Statistics: Estimating
More informationIntroduction to Econometrics
Introduction to Econometrics Michael Bar October 3, 08 San Francisco State University, department of economics. ii Contents Preliminaries. Probability Spaces................................. Random Variables.................................
More informationMath Review Sheet, Fall 2008
1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the
More informationChapter 1 Statistical Reasoning Why statistics? Section 1.1 Basics of Probability Theory
Chapter 1 Statistical Reasoning Why statistics? Uncertainty of nature (weather, earth movement, etc. ) Uncertainty in observation/sampling/measurement Variability of human operation/error imperfection
More informationChapter 2 Random Variables
Stochastic Processes Chapter 2 Random Variables Prof. Jernan Juang Dept. of Engineering Science National Cheng Kung University Prof. Chun-Hung Liu Dept. of Electrical and Computer Eng. National Chiao Tung
More informationDiscrete Probability Refresher
ECE 1502 Information Theory Discrete Probability Refresher F. R. Kschischang Dept. of Electrical and Computer Engineering University of Toronto January 13, 1999 revised January 11, 2006 Probability theory
More informationBusiness Statistics PROBABILITY DISTRIBUTIONS
Business Statistics PROBABILITY DISTRIBUTIONS CONTENTS Probability distribution functions (discrete) Characteristics of a discrete distribution Example: uniform (discrete) distribution Example: Bernoulli
More informationNotes on Mathematics Groups
EPGY Singapore Quantum Mechanics: 2007 Notes on Mathematics Groups A group, G, is defined is a set of elements G and a binary operation on G; one of the elements of G has particularly special properties
More informationBrief Review of Probability
Brief Review of Probability Nuno Vasconcelos (Ken Kreutz-Delgado) ECE Department, UCSD Probability Probability theory is a mathematical language to deal with processes or experiments that are non-deterministic
More informationCME 106: Review Probability theory
: Probability theory Sven Schmit April 3, 2015 1 Overview In the first half of the course, we covered topics from probability theory. The difference between statistics and probability theory is the following:
More informationJoint Distribution of Two or More Random Variables
Joint Distribution of Two or More Random Variables Sometimes more than one measurement in the form of random variable is taken on each member of the sample space. In cases like this there will be a few
More informationProbability and Distributions
Probability and Distributions What is a statistical model? A statistical model is a set of assumptions by which the hypothetical population distribution of data is inferred. It is typically postulated
More informationProbability. Paul Schrimpf. January 23, Definitions 2. 2 Properties 3
Probability Paul Schrimpf January 23, 2018 Contents 1 Definitions 2 2 Properties 3 3 Random variables 4 3.1 Discrete........................................... 4 3.2 Continuous.........................................
More informationSet Theory Digression
1 Introduction to Probability 1.1 Basic Rules of Probability Set Theory Digression A set is defined as any collection of objects, which are called points or elements. The biggest possible collection of
More informationProbability & Statistics - FALL 2008 FINAL EXAM
550.3 Probability & Statistics - FALL 008 FINAL EXAM NAME. An urn contains white marbles and 8 red marbles. A marble is drawn at random from the urn 00 times with replacement. Which of the following is
More informationChapter 8: An Introduction to Probability and Statistics
Course S3, 200 07 Chapter 8: An Introduction to Probability and Statistics This material is covered in the book: Erwin Kreyszig, Advanced Engineering Mathematics (9th edition) Chapter 24 (not including
More informationClass 26: review for final exam 18.05, Spring 2014
Probability Class 26: review for final eam 8.05, Spring 204 Counting Sets Inclusion-eclusion principle Rule of product (multiplication rule) Permutation and combinations Basics Outcome, sample space, event
More information2 (Statistics) Random variables
2 (Statistics) Random variables References: DeGroot and Schervish, chapters 3, 4 and 5; Stirzaker, chapters 4, 5 and 6 We will now study the main tools use for modeling experiments with unknown outcomes
More informationMultiple Random Variables
Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x
More informationLecture 4: Probability and Discrete Random Variables
Error Correcting Codes: Combinatorics, Algorithms and Applications (Fall 2007) Lecture 4: Probability and Discrete Random Variables Wednesday, January 21, 2009 Lecturer: Atri Rudra Scribe: Anonymous 1
More informationProbability theory for Networks (Part 1) CS 249B: Science of Networks Week 02: Monday, 02/04/08 Daniel Bilar Wellesley College Spring 2008
Probability theory for Networks (Part 1) CS 249B: Science of Networks Week 02: Monday, 02/04/08 Daniel Bilar Wellesley College Spring 2008 1 Review We saw some basic metrics that helped us characterize
More informationPerhaps the simplest way of modeling two (discrete) random variables is by means of a joint PMF, defined as follows.
Chapter 5 Two Random Variables In a practical engineering problem, there is almost always causal relationship between different events. Some relationships are determined by physical laws, e.g., voltage
More information5. Conditional Distributions
1 of 12 7/16/2009 5:36 AM Virtual Laboratories > 3. Distributions > 1 2 3 4 5 6 7 8 5. Conditional Distributions Basic Theory As usual, we start with a random experiment with probability measure P on an
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More information2. Variance and Covariance: We will now derive some classic properties of variance and covariance. Assume real-valued random variables X and Y.
CS450 Final Review Problems Fall 08 Solutions or worked answers provided Problems -6 are based on the midterm review Identical problems are marked recap] Please consult previous recitations and textbook
More informationLecture Notes 1 Probability and Random Variables. Conditional Probability and Independence. Functions of a Random Variable
Lecture Notes 1 Probability and Random Variables Probability Spaces Conditional Probability and Independence Random Variables Functions of a Random Variable Generation of a Random Variable Jointly Distributed
More informationIntroduction to Probability Theory for Graduate Economics Fall 2008
Introduction to Probability Theory for Graduate Economics Fall 008 Yiğit Sağlam October 10, 008 CHAPTER - RANDOM VARIABLES AND EXPECTATION 1 1 Random Variables A random variable (RV) is a real-valued function
More informationp(z)
Chapter Statistics. Introduction This lecture is a quick review of basic statistical concepts; probabilities, mean, variance, covariance, correlation, linear regression, probability density functions and
More informationChapter 1. Probability, Random Variables and Expectations. 1.1 Axiomatic Probability
Chapter 1 Probability, Random Variables and Expectations Note: The primary reference for these notes is Mittelhammer (1999. Other treatments of probability theory include Gallant (1997, Casella & Berger
More informationAdvanced Herd Management Probabilities and distributions
Advanced Herd Management Probabilities and distributions Anders Ringgaard Kristensen Slide 1 Outline Probabilities Conditional probabilities Bayes theorem Distributions Discrete Continuous Distribution
More informationAppendix A : Introduction to Probability and stochastic processes
A-1 Mathematical methods in communication July 5th, 2009 Appendix A : Introduction to Probability and stochastic processes Lecturer: Haim Permuter Scribe: Shai Shapira and Uri Livnat The probability of
More informationRandom Variables. Saravanan Vijayakumaran Department of Electrical Engineering Indian Institute of Technology Bombay
1 / 13 Random Variables Saravanan Vijayakumaran sarva@ee.iitb.ac.in Department of Electrical Engineering Indian Institute of Technology Bombay August 8, 2013 2 / 13 Random Variable Definition A real-valued
More informationSlope Fields: Graphing Solutions Without the Solutions
8 Slope Fields: Graphing Solutions Without the Solutions Up to now, our efforts have been directed mainly towards finding formulas or equations describing solutions to given differential equations. Then,
More informationChapter 5. Means and Variances
1 Chapter 5 Means and Variances Our discussion of probability has taken us from a simple classical view of counting successes relative to total outcomes and has brought us to the idea of a probability
More information2 Functions of random variables
2 Functions of random variables A basic statistical model for sample data is a collection of random variables X 1,..., X n. The data are summarised in terms of certain sample statistics, calculated as
More informationMeasure and integration
Chapter 5 Measure and integration In calculus you have learned how to calculate the size of different kinds of sets: the length of a curve, the area of a region or a surface, the volume or mass of a solid.
More information