Boxplots A boxplot, or box-and-whisker plot, is a convenient way of summarising data, particularly when the data is made up of several groups.

Size: px
Start display at page:

Download "Boxplots A boxplot, or box-and-whisker plot, is a convenient way of summarising data, particularly when the data is made up of several groups."

Transcription

1 Boxplots A boxplot, or box-and-whisker plot, is a convenient way of summarising data, particularly when the data is made up of several groups Boxplot largest value (there are no extreme values) upper quartile median lower quartile smallest non extreme value extreme value The box extends from one quartile to the other, and the central line in the box is the median. The whiskers are drawn from the box to the most extreme observations that are no more than.5 IQR from the box. (Alternatively r IQR can be used for other values of r.) Observations which are more extreme than this are shown separately. 2 Gross national income per capita (World Bank data, 29) for 5 sovereign states in Europe. sovereign_states_in_europe_by_gni_(nominal)_per_capita Gross national income per capita (US dollars) Monaco Liechtenstein Now for 82 countries worldwide (including Europe) Monaco Liechtenstein Norway Luxembourg Switzerland Denmark Sweden GNI per capita log(gni per capita) 3 4

2 Parallel boxplots are often useful to show the differences between subgroups of the data. Below: InsectSprays data from R. A B C D E F Type of spray Insect count 5 Comparative boxplots of transformed GCSE scores by A-level chemistry exam score ( = worst, 2, 4, 6, 8, = best) and gender. Transformed GCSE score 5 M F M F 2 M F 4 M F 6 M F 8 M F 6 Comparing N(, ) and t distributions A t-distribution with r degrees of freedom has pdf f (x) ( + x 2 /r) (r+)/2, < x <. [More on t-distributions later.] Consider r = x f(x) N(,) pdf t 5 pdf 7 Suppose we simulate data (x,..., x 25 ) from a t 5 distribution. Using Q-Q plots we can consider the questions: is it reasonable to assume (x,..., x 25 ) is from a N(, )? is it reasonable to assume (x,..., x 25 ) is from a t 5? 8

3 Q Q Plot of data against a N(,) Theoretical Quantiles for a N(,) Sample Quantiles A N(, ) assumption is not good as expected Q Q Plot of data against a t5 Theoretical Quantiles for a t 5 Sample Quantiles A t 5 assumption is ok as expected. Normal Q-Q plots Michelson Morley (879) Speed of Light Data Experiment No. Speed of light (km/s minus 299,) true speed 2 observations from each experiment. Is a N(µ, σ 2 ) distribution plausible for these observations? Normal Q Q Plot for Michelson Morley data Theoretical Quantiles Sample Quantiles From the plot a normal distribution seems reasonable. 2

4 Below: precip data from R average precipitation for 7 US cities. Normal Q Q Plot Below: Newcomb s (882) speed of light data measurements are the time (in deviations from 248 nanoseconds) to travel about 74m. The currently accepted time (on this scale) is 33. Rainfall (inches per year) for 7 US cities Frequency Histogram of Newcomb's data Theoretical Quantiles A normal assumption doesn t look good problems in the lower tail. Time 3 4 This time the problems are different two (very small) outlying observations. If these are removed, a normal assumption looks ok. Example: Danish fire data (Davison, 23) Data on the times, and amounts, of major insurance claims due to fire in Denmark Q Q Plot of Newcomb's data Q Q Plot after deleting two points Sample Quantiles Sample Quantiles Claim amount Claim index Theoretical Quantiles Theoretical Quantiles Following Davison, let s consider the 254 largest claim amounts, and the interarrival times between these claims. 5 6

5 Is it reasonable to assume exponential interarrival times? See below inter-arrivals look fairly close to exponential. Is it reasonable to assume exponential claim amounts? See below an exponential assumption is not reasonable. Exponential Q Q Plot of interarrival times Exponential Q Q Plot of claim amounts Ordered interarrival times x[k] Ordered claim amounts log[ k/(n+)] log[ k/(n+)] 7 8 Is it reasonable to assume Pareto claim amounts? See below the Pareto fits fairly well. Pareto Q Q Plot of claim amounts Multivariate normal distribution [Multivariate/bivariate normal: recap from Part A Probability.] () Let Z,..., Z p iid N(, ). Then joint pdf of Z,..., Z p is log(ordered claim amounts) f (z) = = = p j= ( exp ) (2π) /2 2 z2 j ) exp (2π) p/2 exp (2π) p/2 ( 2 ( 2 zt z p z 2 j j= ), z R p We write Z N(, I) where is a p-vector of zeroes and I is the p p identity matrix. log[ k/(n+)] 9 2

6 z z z (2) Now let µ be a p -vector and let Σ be a p p symmetric, positive definite matrix. Then X = (X,..., X p ) is said to have a multivariate normal distribution with mean vector µ and covariance matrix Σ if its pdf is [ f (x) = (2π) p/2 exp ] det Σ /2 2 (x µ)t Σ (x µ). When X N(µ, Σ) µ is the mean vector: E(X j ) = µ j Σ is the covariance matrix: var(x j ) = Σ jj and cov(x j, X k ) = Σ jk the marginal distribution of X j is X j N(µ j, Σ jj ). We write X N(µ, Σ). This pdf reduces to the N(, I) pdf when µ = and Σ = I Example (bivariate normal distribution) Below: pdf of (X, X 2 ) for ρ =,.5,.9, and contour-plot for ρ =.5. rho = rho =.5 Suppose p = 2. Let < ρ < and µ = Then the pdf of (X, X 2 ) is ( ), Σ = ( ) ρ. ρ ( ) f (x, x 2 ) = 2π ρ exp 2 2( ρ 2 ) (x2 2ρx x 2 + x 2 2 ). The marginal distributions are X N(, ) and X 2 N(, ). The quantity ρ is the correlation between X and X 2. X and X 2 are independent ρ = x2 x2 2 2 rho = x x 2 2 x rho =.5.2 x x x 23 24

7 Opinion polls (revision) Smoking should be banned completely In an ICM poll for the BBC (Feb 26), 485 people agreed with this statement, and 5 disagreed. Let X i = { if ith person agreed otherwise and assume X,..., X n iid Bernoulli(p), with n = 986. So p = x. By CLT, ( P z α/2 < X p p( p)/n < z α/2 ) α. For the above data, and using α =.5, an approximate 95% CI for p is (.46,.523). Smoking should be banned in all public places but not completely 645 agreed, 34 disagreed This time the approximate 95% CI for p is (.624,.684). Estimating the variance p( p)/n by p( p)/n, we get an approximate α CI for p of (x ± z α/2 p( p)/n) Chi-squared pdfs t distribution pdfs f(y) r = 2 r = 4 r = 6 f(t) r = r = 2 r = 5 N(,) y t 27 28

8 Student s Sleep data Normal Q Q Plot of Sleep data Student = W.S. Gosset Below is half of Student s sleep data (98):.7,.6,.2,.2,., 3.4, 3.7,.8,., 2.. The data give the number of hours of sleep gained, by patients, following a low dose of a drug. [The other half of the data give the sleep gained following a normal dose of the drug.] Sample Quantiles 2 3 A point estimate of the sleep gained is x =.75 hours Theoretical Quantiles 29 3 Treating the sample as iid N(µ, σ 2 ), with µ and σ 2 unknown, a 95% CI for µ is ( x ± t n ( α 2 ) s ) = (.53, 2.3) n using x =.75, s 2 = 3.2, n =, α =.5, t 9 (.25) = The value of t 9 (.25) comes from statistical tables, or from R. Here, it would be incorrect to use a N(, ) distribution instead of a t 9. E.g. Suppose we assume σ 2 = s 2 = 3.2 (the sample variance) and calculate the interval ( ) 3.2 x ±.96 = (.36,.86). The interval (.53, 2.3) obtained using the t 9 distribution is wider than the interval (.36,.86). The interval from the t 9 distribution is the correct one here. Since σ 2 is unknown, we need to estimate it (our estimate is s 2 ). Since we are estimating σ 2, there is more uncertainty than if σ 2 were known, and the t 9 distribution correctly takes this uncertainty into account. 3 32

9 Sleep data (low dose) Number of hours of sleep gained, by patients:.7,.6,.2,.2,., 3.4, 3.7,.8,., 2.. Do the data support the conclusion that a low dose of the drug makes people sleep more, or not? We will start from the default position that the drug has no effect, and we will only reject this default position if the data contain sufficient evidence for us to reject it. So we would like to consider (i) the null hypothesis that the drug has no effect, and (ii) the alternative hypothesis that the drug makes people sleep more. We will denote the null hypothesis by H, and the alternative hypothesis by H Sleep data (normal dose) t-test (one sample) [Example from Dalgaard (28).] Data on the daily energy intake (in kj) of women: The other half of the sleep data is the number of hours of sleep gained, by the same patients, following a normal dose of the drug:.9,.8,.,.,., 4.4, 5.5,.6, 4.6, 3.4. Is there evidence that a normal dose of the drug makes people sleep more than not taking a drug at all, or not? 526, 547, 564, 68, 639, 655, 685, 755, 755, 823, 877. Do these values deviate from a recommended value of 7725 kj? We consider testing H : µ = µ against H : µ = µ, where µ = 7725, and we make the standard assumptions for a t-test. We have t obs = x µ s/ n = The p-value is p = 2P(t t obs ) =.8. So we conclude that there is good evidence to reject the null hypothesis that the mean intake is 7725 kj

10 t-test (two sample) Darwin s Zea mays data heights of young maize plants. Testing H : µ = 7725 against H : µ < 7725, the p-value is p = P(t t obs ) =.9. Conclusion: there is good evidence to reject H in favour of H. Testing H : µ = 7725 against H + : µ > 7725, the p-value is p + = P(t t obs ) =.99. Conclusion: there is no evidence to reject H in favour of H +. Height (eights of an inch) Crossed Self-fertilized Are the heights of the two types of plant the same? [In fact, the plants were in pairs one cross- and one self-fertilized in each pair we ignore this pairing for now. We ll see how to deal with pairing later.] Height Assume we have two independent samples X,..., X m iid N(µX, σ 2 ), and Y,..., Y n iid N(µY, σ 2 ), where σ 2 is unknown. Suppose we would like to test H : µ X = µ Y against H : µ X = µ Y. Let T = X Y S m + n where S 2 = m+n 2 [ (X i X) 2 + (Y i Y) 2 ]. Assuming H is true, we have T t m+n 2. Cross Self Type 39 4

11 t-test (paired) For the maize data, the observed value of T is t = x y s m + n = The alternative hypothesis (µ X = µ Y ) is two-sided, so the p-value of this test is p = 2P(t ) =.2. Conclusion: there is good evidence to reject the null hypothesis µ X = µ Y. Suppose we have pairs of RVs (X i, Y i ), i =..., n. Let D i = X i Y i. Suppose D,..., D n iid N(µ, σ 2 ), with σ 2 unknown, and that we want to test a hypothesis about µ. We can use the test statistic D µ S D / n which has a t n distribution under H : µ = µ. (Here, S 2 D is the sample variance of the D i.) The kind of situation where a paired test is used is when there are two measurements on the same experimental unit, e.g. in the sleep data, low and normal doses were given to the same patients Two sample t and paired t Is the amount of sleep gained with a low dose the same as the amount gained with a high dose? low (X) normal (Y) difference (D) Two sample t-test of H : µ X = µ Y against H : µ X = µ Y : the p-value is.79. Paired t-test (of µ = ), based on the differences D i : the p-value is.28. The paired test uses the information that the observations are paired: i.e. we have one low and one high dose observation per patient. The two sample test ignores this information. Prefer the paired test here. Could consider one-sided alternatives here. Hypothesis testing and confidence intervals For the maize data: the 95% (equal tail) confidence interval for µ X µ Y is (3.34, 38.53) (see Sheet 2, Question 5) when testing µ x = µ Y against µ x = µ Y, the p-value is.2. So, observe that (i) the p-value less than.5 (ii) the 95% confidence interval does not contain (= the value of µ X µ Y under H ). (i) and (ii) both being true is not a coincidence there is a connection between hypothesis tests and confidence intervals

12 Insect traps If the test has size α, then α = P( X i c H ). 33 insect traps were set out across sand dunes and the numbers of insects caught in a fixed time were counted (Gilchrist, 984). The number of traps containing various numbers of the taxa Staphylinoidea were as follows. Count Frequency Suppose X,..., X 33 iid Poisson(λ). Consider testing H : λ = against H : λ = λ, where λ >. The NP lemma leads to a test of the form reject H x i c. Under H, we have X i Poisson(33) exactly. However, instead of using this we can use a normal approximation: ( Xi 33 α = P c 33 ) H and, by the CLT, if H is true then X i ( c 33 α Φ ). 33 Hence c z α, so c 33 + z α 33. D N(, ), so So we have a critical region C = {x : x i 33 + z α 33}. Note that C does not depend on which value of λ > we are considering, so we actually have a UMP test of λ = against λ >. If α =. then c 47; if α =. then c 5. The observed value of x i is 54. So in both cases the observed value of 54 is c, so in both cases we d reject H. An alternative way of thinking about this is to calculate the p-value: p = P(we observe a value at least as extreme as 54 H ) = P( X i 54 H ).5 which is very strong evidence for rejecting H. Note that a test of size α rejects H if and only if α p. That is, the p-value is the smallest value of α for which H would be rejected. (This is true generally, not just in this particular example.) In practice, no-one tells us a value of α, we have to judge the situation for ourselves. Our conclusion here is that there is very strong evidence for rejecting H

13 Hardy Weinberg equilibrium In a sample from the Chinese population of Hong Kong, blood types occurred with the following frequencies (Rice, 995): Blood type M MN N Total Frequency If gene frequencies are in Hardy Weinberg equilibrium, then the probability of an individual having blood type M, MN, or N should be The observed frequencies are (n, n 2, n 3 ) = (342, 5, 87), with total n = n + n 2 + n 3 = 29. The likelihood is so the log-likelihood is L(θ) [( θ) 2 ] n [θ( θ)] n 2 [θ 2 ] n 3 l(θ) = (2n + n 2 ) log( θ) + (n 2 + 2n 3 ) log θ + constant from which we obtain P(M) = ( θ) 2 P(MN) = 2θ( θ) P(N) = θ 2. θ = n 2 + 2n 3 2n = So π ( θ) = ( θ) 2, π 2 ( θ) = 2 θ( θ), π 3 ( θ) = θ 2 and ( ) ni Λ = 2 n i log =.32. i nπ i ( θ) We compare Λ to a χ 2 p where p = dim Θ dim Θ = (3 ) =. The value Λ =.32 is much less than E(χ 2 ) =. The p-value is P(χ 2.32) =.86, so there is no reason to doubt the Hardy Weinberg model. Pearson s chi-squared statistic leads to the same conclusion P = [n i nπ i ( θ)] 2 nπ i ( θ) =.39. Insect counts (Bliss and Fisher, 953) [Example from Rice (995).] From each of 6 apple trees in an orchard that had been sprayed, 25 leaves were selected. On each of the leaves, the number of adult female red mites was counted. Number per leaf Observed frequency Does a Poisson(θ) model fit these data? As usual for a Poisson, θ = x =.47, and π i ( θ) = θ i e θ /i!, i =,,..., 7 π 8 ( θ) = 7 i= π i ( θ). The expected frequency in cell i is nπ i ( θ). 5 52

14 Some expected frequencies are very small: # per leaf Observed Expected The χ 2 approximation for the distribution of Λ applies when there are large counts. The usual rule-of-thumb is that the χ 2 approximation is good when the expected frequency in each cell is at least 5. To ensure this, we should pool some cells before calculating Λ or P. After pooling cells 3: # per leaf 2 3 Observed Expected Then Λ = 2 O i log ( O i E i ) = 26.6, and P = (Oi E i ) 2 /E i = These are to be compared with a χ 2 with (4 ) = 2 degrees of freedom. The p-value is p = P(χ ) 6, so there is clear evidence that a Poisson model is not suitable Hair and Eye Colour Relation between hair and eye colour The hair and eye colour of 592 statistics students at the University of Delaware were recorded (Snee, 974) dataset HairEyeColor in R. Black Brown Blue Hazel Green Eye colour Hair colour Brown Blue Hazel Green Black Brown Red Blond Are hair colour and eye colour independent? Hair Brown lond Red Eye 55 56

15 Bayesian Inference Λ = 2 r c i= j= dim(θ) = 6 = 5 ( ) nij n n ij log = 46.4 n i. n.j dim(θ ) = (4 ) + (4 ) = 6 Hence we compare Λ to a χ 2 p where p = 5 6 = 9. The p-value is P(χ ). So there is overwhelming evidence of an association between hair colour and eye colour (i.e. overwhelming evidence that they are not independent). [Pearson s chi-squared statistic is P = 38.3.] So far we have followed the classical/frequentist approach to statistics: we have treated unknown parameters as a fixed constants, and we have imagined repeated sampling from our model in order to evaluate properties of estimators, interpret confidence intervals, calculate p-values, etc. We now take a different approach: in Bayesian inference, unknown parameters are treated as random variables Example In subjective Bayesian inference, probability is a measure of the strength of belief. Before any data are available, there is uncertainty about the parameter θ. Suppose uncertainty about θ is expressed as a prior pdf (of pmf) for θ. Then, once data are available, we can use Bayes theorem to combine our prior beliefs with the data to obtain an updated posterior assessment of our beliefs about θ. Suppose we have a coin which we think might be a bit biased. Let θ be the probability of getting a head when we flip it. 59 6

16 Prior: Beta(5, 5). Data: 7 heads from flips. Prior: Beta(5, 5). Data: 7 heads from flips. Prior density Prior density Posterior density Posterior density Example (MRSA) [Example from Let θ denote the number of MRSA infections per, bed-days in a hospital. Suppose we observe y = 2 infections in 4, bed-days, i.e. in,n bed-days where N = 4. A simple estimate of θ is y/n = 5 infections per, bed-days. The MLE of θ is also θ = 5 if we assume that y is an observation from a Poisson distribution with mean θn, so f (y θ) = (θn) y e θn /y!. However, other evidence about θ may exist. Suppose this other information, on its own, suggests plausible values of θ of about per,, with 95% of the support for θ lying between 5 and 7. We can use a prior distribution to describe this. A Gamma pdf is convenient here: π(θ) = βα Γ(α) θα e βθ for θ >. Taking α =, β = gives approximately the properties above. The posterior combines the evidence from the data (i.e. the likelihood) and the other (i.e. prior) evidence. We can think of the posterior as a compromise between the likelihood and the prior. Calculated on board in lectures: the posterior is another Gamma

17 Prior density posterior likelihood prior Likelihood Posterior density Prior information Example How do we choose a prior π(θ)? (i) E.g. MRSA example, we might ask a scientific expert who might anticipate θ around, say with θ (5, 7) with probability.95. (ii) We might have little prior knowledge, so we might want a prior that expresses prior ignorance. E.g. if a probability is unknown, we might consider the prior θ U(, ). (iii) In the Bernoulli/Beta and Poisson/Gamma examples (Section 4.), the posterior was of same form as prior, i.e. Beta-Beta and Gamma-Gamma. This occurred because the likelihood and prior had the same functional form the prior and likelihood are said to be conjugate (convenient for doing calculations by hand). [Example from Carlin and Louis (28).] Product P old, standard. Product P newer, more expensive. Assumptions: the probability θ that a customer prefers P has prior π(θ) which is Beta(a, b) the number of customers X (out of n) that prefer P is X Binomial(n, θ). Let s say θ.6 means that P is a substantial improvement over P. So take H : θ.6 and H : θ <

18 We consider 3 possibile priors: Jeffreys prior: θ Beta(.5,.5). Uniform prior: θ Beta(, ). Sceptical prior: θ Beta(2, 2), i.e. favours values of θ near 2. prior density Jeffreys prior Uniform prior Sceptical prior Prior odds = P(H )/P(H ) where P(H ) = P(H ) =.6.6 B(a, b) θa ( θ) b dθ B(a, b) θa ( θ) b dθ. Suppose we have x = 3 successes from n = 6 customers. Then (Section 4.) the posterior π(θ x) is Beta(x + a, n x + b) with x = 3 and n = 6. Posterior odds = P(H x)/p(h x) where P(H x) = P(H x) =.6.6 B(x + a, n x + b) θx+a ( θ) n x+b dθ B(x + a, n x + b) θx+a ( θ) n x+b dθ. 7 72

19 posterior density Beta(3.5, 3.5) Beta(4, 4) Beta(5, 5) Prior Prior odds Posterior odds Bayes factor Beta(.5,.5) Beta(, ) Beta(2, 2) Conclusion: strong evidence for H Normal approx to posterior () Prior θ U(, ). Bernoulli likelihood: x = 3 successes out of n = 6 trials. Normal approx to posterior (2) Prior θ U(, ). Bernoulli likelihood: x = 3 successes out of n = 6 trials. posterior density exact (beta) approx (normal) posterior density exact (beta) approx (normal) θ θ 75 76

Bayesian Inference. Chapter 2: Conjugate models

Bayesian Inference. Chapter 2: Conjugate models Bayesian Inference Chapter 2: Conjugate models Conchi Ausín and Mike Wiper Department of Statistics Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in

More information

Foundations of Statistical Inference

Foundations of Statistical Inference Foundations of Statistical Inference Julien Berestycki Department of Statistics University of Oxford MT 2016 Julien Berestycki (University of Oxford) SB2a MT 2016 1 / 20 Lecture 6 : Bayesian Inference

More information

Exam 2 Practice Questions, 18.05, Spring 2014

Exam 2 Practice Questions, 18.05, Spring 2014 Exam 2 Practice Questions, 18.05, Spring 2014 Note: This is a set of practice problems for exam 2. The actual exam will be much shorter. Within each section we ve arranged the problems roughly in order

More information

Chapter 5. Bayesian Statistics

Chapter 5. Bayesian Statistics Chapter 5. Bayesian Statistics Principles of Bayesian Statistics Anything unknown is given a probability distribution, representing degrees of belief [subjective probability]. Degrees of belief [subjective

More information

Chapter 5. Chapter 5 sections

Chapter 5. Chapter 5 sections 1 / 43 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages Name No calculators. 18.05 Final Exam Number of problems 16 concept questions, 16 problems, 21 pages Extra paper If you need more space we will provide some blank paper. Indicate clearly that your solution

More information

Section 4.6 Simple Linear Regression

Section 4.6 Simple Linear Regression Section 4.6 Simple Linear Regression Objectives ˆ Basic philosophy of SLR and the regression assumptions ˆ Point & interval estimation of the model parameters, and how to make predictions ˆ Point and interval

More information

18.05 Practice Final Exam

18.05 Practice Final Exam No calculators. 18.05 Practice Final Exam Number of problems 16 concept questions, 16 problems. Simplifying expressions Unless asked to explicitly, you don t need to simplify complicated expressions. For

More information

simple if it completely specifies the density of x

simple if it completely specifies the density of x 3. Hypothesis Testing Pure significance tests Data x = (x 1,..., x n ) from f(x, θ) Hypothesis H 0 : restricts f(x, θ) Are the data consistent with H 0? H 0 is called the null hypothesis simple if it completely

More information

Statistics. Statistics

Statistics. Statistics The main aims of statistics 1 1 Choosing a model 2 Estimating its parameter(s) 1 point estimates 2 interval estimates 3 Testing hypotheses Distributions used in statistics: χ 2 n-distribution 2 Let X 1,

More information

Class 26: review for final exam 18.05, Spring 2014

Class 26: review for final exam 18.05, Spring 2014 Probability Class 26: review for final eam 8.05, Spring 204 Counting Sets Inclusion-eclusion principle Rule of product (multiplication rule) Permutation and combinations Basics Outcome, sample space, event

More information

F79SM STATISTICAL METHODS

F79SM STATISTICAL METHODS F79SM STATISTICAL METHODS SUMMARY NOTES 9 Hypothesis testing 9.1 Introduction As before we have a random sample x of size n of a population r.v. X with pdf/pf f(x;θ). The distribution we assign to X is

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1 Chapter 4 HOMEWORK ASSIGNMENTS These homeworks may be modified as the semester progresses. It is your responsibility to keep up to date with the correctly assigned homeworks. There may be some errors in

More information

Foundations of Statistical Inference

Foundations of Statistical Inference Foundations of Statistical Inference Jonathan Marchini Department of Statistics University of Oxford MT 2013 Jonathan Marchini (University of Oxford) BS2a MT 2013 1 / 27 Course arrangements Lectures M.2

More information

A Very Brief Summary of Statistical Inference, and Examples

A Very Brief Summary of Statistical Inference, and Examples A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015

STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots. March 8, 2015 STAT 135 Lab 6 Duality of Hypothesis Testing and Confidence Intervals, GLRT, Pearson χ 2 Tests and Q-Q plots March 8, 2015 The duality between CI and hypothesis testing The duality between CI and hypothesis

More information

Institute of Actuaries of India

Institute of Actuaries of India Institute of Actuaries of India Subject CT3 Probability & Mathematical Statistics May 2011 Examinations INDICATIVE SOLUTION Introduction The indicative solution has been written by the Examiners with the

More information

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio Estimation of reliability parameters from Experimental data (Parte 2) This lecture Life test (t 1,t 2,...,t n ) Estimate θ of f T t θ For example: λ of f T (t)= λe - λt Classical approach (frequentist

More information

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017

Bayesian inference. Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark. April 10, 2017 Bayesian inference Rasmus Waagepetersen Department of Mathematics Aalborg University Denmark April 10, 2017 1 / 22 Outline for today A genetic example Bayes theorem Examples Priors Posterior summaries

More information

Chapter 5 continued. Chapter 5 sections

Chapter 5 continued. Chapter 5 sections Chapter 5 sections Discrete univariate distributions: 5.2 Bernoulli and Binomial distributions Just skim 5.3 Hypergeometric distributions 5.4 Poisson distributions Just skim 5.5 Negative Binomial distributions

More information

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Review of Basic Probability The fundamentals, random variables, probability distributions Probability mass/density functions

More information

Lecture 12: Small Sample Intervals Based on a Normal Population Distribution

Lecture 12: Small Sample Intervals Based on a Normal Population Distribution Lecture 12: Small Sample Intervals Based on a Normal Population MSU-STT-351-Sum-17B (P. Vellaisamy: MSU-STT-351-Sum-17B) Probability & Statistics for Engineers 1 / 24 In this lecture, we will discuss (i)

More information

One-Way Tables and Goodness of Fit

One-Way Tables and Goodness of Fit Stat 504, Lecture 5 1 One-Way Tables and Goodness of Fit Key concepts: One-way Frequency Table Pearson goodness-of-fit statistic Deviance statistic Pearson residuals Objectives: Learn how to compute the

More information

1 Introduction. P (n = 1 red ball drawn) =

1 Introduction. P (n = 1 red ball drawn) = Introduction Exercises and outline solutions. Y has a pack of 4 cards (Ace and Queen of clubs, Ace and Queen of Hearts) from which he deals a random of selection 2 to player X. What is the probability

More information

Lecture 2: Repetition of probability theory and statistics

Lecture 2: Repetition of probability theory and statistics Algorithms for Uncertainty Quantification SS8, IN2345 Tobias Neckel Scientific Computing in Computer Science TUM Lecture 2: Repetition of probability theory and statistics Concept of Building Block: Prerequisites:

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior:

Unobservable Parameter. Observed Random Sample. Calculate Posterior. Choosing Prior. Conjugate prior. population proportion, p prior: Pi Priors Unobservable Parameter population proportion, p prior: π ( p) Conjugate prior π ( p) ~ Beta( a, b) same PDF family exponential family only Posterior π ( p y) ~ Beta( a + y, b + n y) Observed

More information

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3

Hypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3 Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest

More information

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003

Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Statistics Ph.D. Qualifying Exam: Part I October 18, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification

Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification Bayesian Statistics Part III: Building Bayes Theorem Part IV: Prior Specification Michael Anderson, PhD Hélène Carabin, DVM, PhD Department of Biostatistics and Epidemiology The University of Oklahoma

More information

15 Discrete Distributions

15 Discrete Distributions Lecture Note 6 Special Distributions (Discrete and Continuous) MIT 4.30 Spring 006 Herman Bennett 5 Discrete Distributions We have already seen the binomial distribution and the uniform distribution. 5.

More information

Learning Objectives for Stat 225

Learning Objectives for Stat 225 Learning Objectives for Stat 225 08/20/12 Introduction to Probability: Get some general ideas about probability, and learn how to use sample space to compute the probability of a specific event. Set Theory:

More information

STAT 4385 Topic 01: Introduction & Review

STAT 4385 Topic 01: Introduction & Review STAT 4385 Topic 01: Introduction & Review Xiaogang Su, Ph.D. Department of Mathematical Science University of Texas at El Paso xsu@utep.edu Spring, 2016 Outline Welcome What is Regression Analysis? Basics

More information

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary Patrick Breheny October 13 Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/25 Introduction Introduction What s wrong with z-tests? So far we ve (thoroughly!) discussed how to carry out hypothesis

More information

MATH4427 Notebook 2 Fall Semester 2017/2018

MATH4427 Notebook 2 Fall Semester 2017/2018 MATH4427 Notebook 2 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 2 MATH4427 Notebook 2 3 2.1 Definitions and Examples...................................

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata

Testing Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Chapter 10. Hypothesis Testing (I)

Chapter 10. Hypothesis Testing (I) Chapter 10. Hypothesis Testing (I) Hypothesis Testing, together with statistical estimation, are the two most frequently used statistical inference methods. It addresses a different type of practical problems

More information

STT 843 Key to Homework 1 Spring 2018

STT 843 Key to Homework 1 Spring 2018 STT 843 Key to Homework Spring 208 Due date: Feb 4, 208 42 (a Because σ = 2, σ 22 = and ρ 2 = 05, we have σ 2 = ρ 2 σ σ22 = 2/2 Then, the mean and covariance of the bivariate normal is µ = ( 0 2 and Σ

More information

Math 475. Jimin Ding. August 29, Department of Mathematics Washington University in St. Louis jmding/math475/index.

Math 475. Jimin Ding. August 29, Department of Mathematics Washington University in St. Louis   jmding/math475/index. istical A istic istics : istical Department of Mathematics Washington University in St. Louis www.math.wustl.edu/ jmding/math475/index.html August 29, 2013 istical August 29, 2013 1 / 18 istical A istic

More information

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices.

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices. Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices. 1.(10) What is usually true about a parameter of a model? A. It is a known number B. It is determined by the data C. It is an

More information

STAT Chapter 5 Continuous Distributions

STAT Chapter 5 Continuous Distributions STAT 270 - Chapter 5 Continuous Distributions June 27, 2012 Shirin Golchi () STAT270 June 27, 2012 1 / 59 Continuous rv s Definition: X is a continuous rv if it takes values in an interval, i.e., range

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October

More information

STAT 425: Introduction to Bayesian Analysis

STAT 425: Introduction to Bayesian Analysis STAT 425: Introduction to Bayesian Analysis Marina Vannucci Rice University, USA Fall 2017 Marina Vannucci (Rice University, USA) Bayesian Analysis (Part 1) Fall 2017 1 / 10 Lecture 7: Prior Types Subjective

More information

Stat 704 Data Analysis I Probability Review

Stat 704 Data Analysis I Probability Review 1 / 39 Stat 704 Data Analysis I Probability Review Dr. Yen-Yi Ho Department of Statistics, University of South Carolina A.3 Random Variables 2 / 39 def n: A random variable is defined as a function that

More information

Solutions to the Spring 2015 CAS Exam ST

Solutions to the Spring 2015 CAS Exam ST Solutions to the Spring 2015 CAS Exam ST (updated to include the CAS Final Answer Key of July 15) There were 25 questions in total, of equal value, on this 2.5 hour exam. There was a 10 minute reading

More information

Statistical Theory MT 2006 Problems 4: Solution sketches

Statistical Theory MT 2006 Problems 4: Solution sketches Statistical Theory MT 006 Problems 4: Solution sketches 1. Suppose that X has a Poisson distribution with unknown mean θ. Determine the conjugate prior, and associate posterior distribution, for θ. Determine

More information

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper

McGill University. Faculty of Science. Department of Mathematics and Statistics. Part A Examination. Statistics: Theory Paper McGill University Faculty of Science Department of Mathematics and Statistics Part A Examination Statistics: Theory Paper Date: 10th May 2015 Instructions Time: 1pm-5pm Answer only two questions from Section

More information

Mathematical Statistics

Mathematical Statistics Mathematical Statistics MAS 713 Chapter 8 Previous lecture: 1 Bayesian Inference 2 Decision theory 3 Bayesian Vs. Frequentist 4 Loss functions 5 Conjugate priors Any questions? Mathematical Statistics

More information

Lecture 1: Probability Fundamentals

Lecture 1: Probability Fundamentals Lecture 1: Probability Fundamentals IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge January 22nd, 2008 Rasmussen (CUED) Lecture 1: Probability

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Discrete Multivariate Statistics

Discrete Multivariate Statistics Discrete Multivariate Statistics Univariate Discrete Random variables Let X be a discrete random variable which, in this module, will be assumed to take a finite number of t different values which are

More information

A Very Brief Summary of Bayesian Inference, and Examples

A Very Brief Summary of Bayesian Inference, and Examples A Very Brief Summary of Bayesian Inference, and Examples Trinity Term 009 Prof Gesine Reinert Our starting point are data x = x 1, x,, x n, which we view as realisations of random variables X 1, X,, X

More information

Statistics Masters Comprehensive Exam March 21, 2003

Statistics Masters Comprehensive Exam March 21, 2003 Statistics Masters Comprehensive Exam March 21, 2003 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your answer

More information

CSC321 Lecture 18: Learning Probabilistic Models

CSC321 Lecture 18: Learning Probabilistic Models CSC321 Lecture 18: Learning Probabilistic Models Roger Grosse Roger Grosse CSC321 Lecture 18: Learning Probabilistic Models 1 / 25 Overview So far in this course: mainly supervised learning Language modeling

More information

(1) Introduction to Bayesian statistics

(1) Introduction to Bayesian statistics Spring, 2018 A motivating example Student 1 will write down a number and then flip a coin If the flip is heads, they will honestly tell student 2 if the number is even or odd If the flip is tails, they

More information

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n

Recall that in order to prove Theorem 8.8, we argued that under certain regularity conditions, the following facts are true under H 0 : 1 n Chapter 9 Hypothesis Testing 9.1 Wald, Rao, and Likelihood Ratio Tests Suppose we wish to test H 0 : θ = θ 0 against H 1 : θ θ 0. The likelihood-based results of Chapter 8 give rise to several possible

More information

MATH20802: STATISTICAL METHODS EXAMPLES

MATH20802: STATISTICAL METHODS EXAMPLES MATH20802: STATISTICAL METHODS EXAMPLES 1 1. If X N(µ, σ 2 ) show that its mgf is M X (t) = exp ( µt + σ2 t 2 2 2. If X 1 N(µ 1, σ 2 1 ) and X 2 N(µ 2, σ 2 2 ) are independent then show that ax 1 + bx

More information

Lecture 7: Confidence interval and Normal approximation

Lecture 7: Confidence interval and Normal approximation Lecture 7: Confidence interval and Normal approximation 26th of November 2015 Confidence interval 26th of November 2015 1 / 23 Random sample and uncertainty Example: we aim at estimating the average height

More information

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics DS-GA 100 Lecture notes 11 Fall 016 Bayesian statistics In the frequentist paradigm we model the data as realizations from a distribution that depends on deterministic parameters. In contrast, in Bayesian

More information

Statistical Theory MT 2007 Problems 4: Solution sketches

Statistical Theory MT 2007 Problems 4: Solution sketches Statistical Theory MT 007 Problems 4: Solution sketches 1. Consider a 1-parameter exponential family model with density f(x θ) = f(x)g(θ)exp{cφ(θ)h(x)}, x X. Suppose that the prior distribution has the

More information

Introduction to Applied Bayesian Modeling. ICPSR Day 4

Introduction to Applied Bayesian Modeling. ICPSR Day 4 Introduction to Applied Bayesian Modeling ICPSR Day 4 Simple Priors Remember Bayes Law: Where P(A) is the prior probability of A Simple prior Recall the test for disease example where we specified the

More information

STAT 418: Probability and Stochastic Processes

STAT 418: Probability and Stochastic Processes STAT 418: Probability and Stochastic Processes Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical

More information

Probability and Statistics Notes

Probability and Statistics Notes Probability and Statistics Notes Chapter Five Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Five Notes Spring 2011 1 / 37 Outline 1

More information

Lecture 10: Generalized likelihood ratio test

Lecture 10: Generalized likelihood ratio test Stat 200: Introduction to Statistical Inference Autumn 2018/19 Lecture 10: Generalized likelihood ratio test Lecturer: Art B. Owen October 25 Disclaimer: These notes have not been subjected to the usual

More information

Confidence Intervals for Normal Data Spring 2014

Confidence Intervals for Normal Data Spring 2014 Confidence Intervals for Normal Data 18.05 Spring 2014 Agenda Today Review of critical values and quantiles. Computing z, t, χ 2 confidence intervals for normal data. Conceptual view of confidence intervals.

More information

STATISTICS SYLLABUS UNIT I

STATISTICS SYLLABUS UNIT I STATISTICS SYLLABUS UNIT I (Probability Theory) Definition Classical and axiomatic approaches.laws of total and compound probability, conditional probability, Bayes Theorem. Random variable and its distribution

More information

Lecture 3. Univariate Bayesian inference: conjugate analysis

Lecture 3. Univariate Bayesian inference: conjugate analysis Summary Lecture 3. Univariate Bayesian inference: conjugate analysis 1. Posterior predictive distributions 2. Conjugate analysis for proportions 3. Posterior predictions for proportions 4. Conjugate analysis

More information

ECON Fundamentals of Probability

ECON Fundamentals of Probability ECON 351 - Fundamentals of Probability Maggie Jones 1 / 32 Random Variables A random variable is one that takes on numerical values, i.e. numerical summary of a random outcome e.g., prices, total GDP,

More information

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes

Hypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis

More information

Introduction to Bayesian Methods

Introduction to Bayesian Methods Introduction to Bayesian Methods Jessi Cisewski Department of Statistics Yale University Sagan Summer Workshop 2016 Our goal: introduction to Bayesian methods Likelihoods Priors: conjugate priors, non-informative

More information

Statistical Concepts

Statistical Concepts Statistical Concepts Ad Feelders May 19, 2015 1 Introduction Statistics is the science of collecting, organizing and drawing conclusions from data. How to properly produce and collect data is studied in

More information

Continuous random variables

Continuous random variables Continuous random variables Continuous r.v. s take an uncountably infinite number of possible values. Examples: Heights of people Weights of apples Diameters of bolts Life lengths of light-bulbs We cannot

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

18 Bivariate normal distribution I

18 Bivariate normal distribution I 8 Bivariate normal distribution I 8 Example Imagine firing arrows at a target Hopefully they will fall close to the target centre As we fire more arrows we find a high density near the centre and fewer

More information

STAT 414: Introduction to Probability Theory

STAT 414: Introduction to Probability Theory STAT 414: Introduction to Probability Theory Spring 2016; Homework Assignments Latest updated on April 29, 2016 HW1 (Due on Jan. 21) Chapter 1 Problems 1, 8, 9, 10, 11, 18, 19, 26, 28, 30 Theoretical Exercises

More information

9 Bayesian inference. 9.1 Subjective probability

9 Bayesian inference. 9.1 Subjective probability 9 Bayesian inference 1702-1761 9.1 Subjective probability This is probability regarded as degree of belief. A subjective probability of an event A is assessed as p if you are prepared to stake pm to win

More information

Statistics 135 Fall 2008 Final Exam

Statistics 135 Fall 2008 Final Exam Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations

More information

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama

Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Qualifying Exam CS 661: System Simulation Summer 2013 Prof. Marvin K. Nakayama Instructions This exam has 7 pages in total, numbered 1 to 7. Make sure your exam has all the pages. This exam will be 2 hours

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Part 2: One-parameter models

Part 2: One-parameter models Part 2: One-parameter models 1 Bernoulli/binomial models Return to iid Y 1,...,Y n Bin(1, ). The sampling model/likelihood is p(y 1,...,y n ) = P y i (1 ) n P y i When combined with a prior p( ), Bayes

More information

MATH4427 Notebook 4 Fall Semester 2017/2018

MATH4427 Notebook 4 Fall Semester 2017/2018 MATH4427 Notebook 4 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 4 MATH4427 Notebook 4 3 4.1 K th Order Statistics and Their

More information

First Year Examination Department of Statistics, University of Florida

First Year Examination Department of Statistics, University of Florida First Year Examination Department of Statistics, University of Florida August 19, 010, 8:00 am - 1:00 noon Instructions: 1. You have four hours to answer questions in this examination.. You must show your

More information

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam.

This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided with the exam. Probability and Statistics FS 2017 Session Exam 22.08.2017 Time Limit: 180 Minutes Name: Student ID: This exam contains 13 pages (including this cover page) and 10 questions. A Formulae sheet is provided

More information

UNIVERSITY OF TORONTO Faculty of Arts and Science

UNIVERSITY OF TORONTO Faculty of Arts and Science UNIVERSITY OF TORONTO Faculty of Arts and Science December 2013 Final Examination STA442H1F/2101HF Methods of Applied Statistics Jerry Brunner Duration - 3 hours Aids: Calculator Model(s): Any calculator

More information

Lecture 1: August 28

Lecture 1: August 28 36-705: Intermediate Statistics Fall 2017 Lecturer: Siva Balakrishnan Lecture 1: August 28 Our broad goal for the first few lectures is to try to understand the behaviour of sums of independent random

More information

HPD Intervals / Regions

HPD Intervals / Regions HPD Intervals / Regions The HPD region will be an interval when the posterior is unimodal. If the posterior is multimodal, the HPD region might be a discontiguous set. Picture: The set {θ : θ (1.5, 3.9)

More information

Compute f(x θ)f(θ) dθ

Compute f(x θ)f(θ) dθ Bayesian Updating: Continuous Priors 18.05 Spring 2014 b a Compute f(x θ)f(θ) dθ January 1, 2017 1 /26 Beta distribution Beta(a, b) has density (a + b 1)! f (θ) = θ a 1 (1 θ) b 1 (a 1)!(b 1)! http://mathlets.org/mathlets/beta-distribution/

More information

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013

UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013 UNIVERSITY OF TORONTO SCARBOROUGH Department of Computer and Mathematical Sciences FINAL EXAMINATION, APRIL 2013 STAB57H3 Introduction to Statistics Duration: 3 hours Last Name: First Name: Student number:

More information

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables.

Random vectors X 1 X 2. Recall that a random vector X = is made up of, say, k. X k. random variables. Random vectors Recall that a random vector X = X X 2 is made up of, say, k random variables X k A random vector has a joint distribution, eg a density f(x), that gives probabilities P(X A) = f(x)dx Just

More information

Normal (Gaussian) distribution The normal distribution is often relevant because of the Central Limit Theorem (CLT):

Normal (Gaussian) distribution The normal distribution is often relevant because of the Central Limit Theorem (CLT): Lecture Three Normal theory null distributions Normal (Gaussian) distribution The normal distribution is often relevant because of the Central Limit Theorem (CLT): A random variable which is a sum of many

More information

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution

Outline. Binomial, Multinomial, Normal, Beta, Dirichlet. Posterior mean, MAP, credible interval, posterior distribution Outline A short review on Bayesian analysis. Binomial, Multinomial, Normal, Beta, Dirichlet Posterior mean, MAP, credible interval, posterior distribution Gibbs sampling Revisit the Gaussian mixture model

More information

Hypothesis Testing Chap 10p460

Hypothesis Testing Chap 10p460 Hypothesis Testing Chap 1p46 Elements of a statistical test p462 - Null hypothesis - Alternative hypothesis - Test Statistic - Rejection region Rejection Region p462 The rejection region (RR) specifies

More information

Summary of Chapters 7-9

Summary of Chapters 7-9 Summary of Chapters 7-9 Chapter 7. Interval Estimation 7.2. Confidence Intervals for Difference of Two Means Let X 1,, X n and Y 1, Y 2,, Y m be two independent random samples of sizes n and m from two

More information

Lecture 11. Multivariate Normal theory

Lecture 11. Multivariate Normal theory 10. Lecture 11. Multivariate Normal theory Lecture 11. Multivariate Normal theory 1 (1 1) 11. Multivariate Normal theory 11.1. Properties of means and covariances of vectors Properties of means and covariances

More information