DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS
|
|
- Maximillian Wilson
- 5 years ago
- Views:
Transcription
1 DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS THUY ANH NGO 1. Introduction Statistics are easily come across in our daily life. Statements such as the average college student s debt after college is $2,700, or the average person spends 24 years of their life asleep, are results of statistical analysis. Statistical inference is one method of statistical analysis. It is the process of applying statistical methods in order to make probabilistic statements about some characteristics of the observed data. In cognitive science, statistical inference allows scientists to draw some conclusion about a psychological phenomenon by statistically analyzing a set of observations. For example, a social psychologist is interested in investigating whether the majority of the population shows change blindness (i.e., a phenomenon in which people often miss large changes in their visual field). To study this phenomenon, the scientist can collect data from the entire population of Americans. Such a task, however, is not economically feasible. Conveniently, statistical inference allows her to examine the change blindness phenomenon based on the data from a randomly selected pool of participants. There are different approaches to statistical inference, including hypothesis testing and Bayesian inference. Even though hypothesis testing is traditionally used by most cognitive scientists, Bayesian inference has garnered increased interest in recent years. This paper applies probability theory to explain these approaches, discusses the strengths and weaknesses of both approaches to explain the increasing use of Bayesian inference, especially in the field of cognitive science. 2. Two approaches to statistical inference 2.1. Important Statistical Concepts. We first start by defining key concepts. Definition 2.1 (Population vs. Sample). A population is a set of all observations that can possibly be made. A sample is a set of observations that are randomly selected from a population. Examples of a population include Trinity University students, people with anxiety disorder, trees in the state of Texas, etc. 1
2 2 NGO Examples of a sample include 70 students at Trinity chosen randomly, 20 clients with anxiety disorder at a hospital, 27 trees randomly selected in Texas, etc. Definition 2.2 (Parameter). A parameter is a measurable characteristic of a population. For example, when we examine whether a machine is functioning properly, we define the parameter to be the proportion of defective items in a large manufactured lot. Definition 2. (Random Variable). Let S be the sample space. A map X : S R is a random variable that assigns a real number X(s) to each possible outcome s S (DeGroot & Schervish, 2002). Definition 2.4 (Probability Function). Let X be a discrete random variable (X can take only a finite number k of different values x 1,..., x k ). The probability function (p.f.) of X is defined as the function f such that for every real number x, f(x) = P r(x = x). (DeGroot & Schervish, 2002) Definition 2.5 (Probability Density Function). Let X be a continuous random variable (X can take every value in an interval). The probability density function (p.d.f.) of X is the nonnegative function f, defined on R, such that for every subset A R, P r(x A) = f(x) dx. (DeGroot & Schervish, 2002) A Definition 2.6 (Statistical Inference). Statistical inference is the process of applying statistical methods in order to make probabilistic statements about some characteristics of the observed data Hypothesis Testing The Null and Alternative Hypotheses. In hypothesis testing approach, a parameter θ is an unknown value that lies in a parameter space Ω. A statistician s task is to decide whether the parameter value falls in one subset of Ω or its complement. We can then define our hypotheses as follows: Definition 2.7. Ω 0 and Ω 1 are disjoint subsets of the parameter space Ω with Ω 0 Ω 1 = Ω. The null and the alternative hypotheses are: H 0 : µ = µ 0, H A : µ µ 0.
3 STATISTICAL INFERENCE The hypothesis H 0 is called the null hypothesis and the hypothesis H A is called the alternative hypothesis. A statistician has to decide to accept either H 0 or H A. With only two possible decisions, accepting H 0 is equivalent to rejecting H A, and vice versa. In most statistical problems, the two hypotheses are treated quite differently. For example, in clinical trials, the null hypothesis almost always states that a particular treatment does not have an effect while the alternative hypothesis states that the treatment has an effect. Many cognitive scientists frame their research design and analysis in terms of rejecting the null hypothesis, a routine that has recently been debated. We will return to this debate later The Critical Region and Test Statistics. Suppose that X = (X 1,..., X n ) form a random sample from the population whose distribution involves the unknown parameter θ. Let S denote the set of all possible outcomes of the random sample X. A statistician then follows a test procedure that partitions S into two subsets, one of which contains values of X for which H 0 will be accepted while the other contains values of X for which H 0 will be rejected. Definition 2.8. The subset for which H 0 will be rejected is called the critical region, C. This region is defined in terms of a test statistic, T where T = f(x), a function of the observed data X. Example 1.1. Suppose that X = (X 1,..., X n ) is a random sample from a normal distribution with unknown mean µ and known variance σ 2. The hypotheses to be tested are: H 0 : µ = µ 0, H A : µ µ 0. H 0 is rejected if X n is far from µ 0. So if we define the test statistic to be T = X n µ 0, we then choose a test procedure that rejects H 0 if T c for some c > Level of Significance. Definition 2.9. The level of significance, α is determined by P r(t C) = α where C is the critical region. If T C, we reject H 0 in favor of H A on level α. Some standard values of α are 0.05, 0.01, Example 1.2. A psychologist wants to explore the change blindness phenomenon introduced above. She sets up an experiment in which an experimenter randomly approaches a pedestrian on a college campus to ask for directions (Simons & Levin, 1998). Shortly afterward, two men carrying a door pass in between them and interrupt the conversation. Meanwhile, the first experimenter is replaced with the second experimenter who has different height, different voice, and where different cloths from
4 4 NGO the first one. After the exchange, the second experimenter continues to talk to the pedestrian. Eventually, the pedestrian is asked if he/she noticed any changes throughout the interaction. Out of 40 pedestrians who were randomly selected to participate in the study, only 8 reported noticing the change. The psychologist would like to draw the conclusion whether at least two thirds of the people exhibit change blindness. Suppose that X = (X 1,..., X 40 ) is a random sample drawn from the population distribution with parameter p, where p is the probability of exhibiting change blindness. X i s are defined as follows: { 1, if the ith participant shows change blindness X i = 0, if the i th participant does not show change blindness for i {1, 2,..., 40}. X i = 1 with probability p and X 1 = 0 with probability q = p 1. Therefore, X 1,..., X 40 Bernoulli(p) and the p.f. of each X i can be written as follows: { p x i q f(x i p) = 1 x i, for x i = 0 or 1 0, otherwise The hypotheses are stated as follows: H 0 : p 2 (the majority doesn t exhibit change blindness), H A : p > 2 (the majority exhibits change blindness). The decision to reject or accept the null hypothesis first requires us to find the test statistic T and specify the critical region C. 40 Let the test statistic be Y = X i. So Y binomial(n = 40, p) and the p.f. is: i=1 f(y n, p) = ( ) n p y q n y, y for y = 0, 1,..., n 0, otherwise The larger p is, the larger we expect Y to be. Therefore, we choose to reject H 0 if Y c for some constant c. At the level of significance α 0 = 0.05, we reject H 0 if P r(y 2 p = 2 ) < α 0. We should now compute P r(y 2 p = 2 ): ( P r Y 2 p = 2 ) ( = 1 P r Y 1 p = 2 )
5 STATISTICAL INFERENCE 5 40 ( ) 40 = 1 p k (1 p) 40 k, p = 2 k k=0 40 = 1 = k=0 ( 40 k ) ( 2 ) k ( 1 2 ) 40 k Since P r(y 2 p = 2 ) = 2 ) < 0.05, we reject H 0 and conclude that the majority of people does exhibit change blindness. Remark. Rejecting the null hypothesis does not mean we prove that it is false. It is only that the data support a rejection. Also, in our example, we reject the null hypothesis at the 5% level of significance (α 0 = 0.05) but we cannot reject the null hypothesis at the 1% level of significance. In brief, our decision depends on our choice of the significance level, which can be quite arbitrary in nature. This is a typical criticism of hypothesis testing, and we will return to it later. 2.. Bayesian Inference. Bayesian inference relies on the same logic that is used by the famous fictional detective, Sherlock Holmes: when you have eliminated the impossible, whatever remains, however improbable, must be the truth. Likewise, Bayesian inference is the method in which we begin with a set of all possibilities, then based on collected data, we eliminate some possibilities and reallocate credibilities to the remaining ones. Consequently, a parameter is a random variable which has a probability distribution over the set of all candidate values. We say that the parameter has a prior distribution, ξ(θ). The prior distribution can be determined by utilizing previous information or knowledge about the parameter. After observing the data, reallocation of credibilities is reflected in the posterior distribution, which is conditioned on observed data. The posterior distribution is denoted as ξ(θ x) where x = (x 1,..., x n ) is the vector that represents the observed data. The relationship between the prior and the posterior distribution of a parameter is given by the Bayes theorem. Theorem 2.1 (Bayes Theorem). Let X 1,..., X n be i.i.d. random variables with p.d.f. f(x i θ) and joint p.d.f. f n (x θ). Let g n (x) = f n (x 1,..., x n θ)ξ(θ)dθ. Then, Ω ξ(θ x) = f n(x θ)ξ(θ) g n (x)
6 6 NGO Example 1.. (uniform prior) Consider the experiment described in Example 1.2. Let p be the probability of exhibiting change blindness and let X i = 1 if the ith participant shows change blindness and X i = 0 if not. Given any p, the distribution of each X i is: f(x i p) = { p x i (1 p) 1 x i, for x i = 0 or 1 0, otherwise Suppose that we have no prior knowledge about what values of p are more probable than others, we assume uniform distribution for p, so ξ(p) = 1 and its plot is given in Figure 2.1.a. We are interested in finding the distribution of p after observing some participants in the study. There were 40 pedestrians and 2 of them showed change blindness. The joint distribution of X 1,...X 40 is given by: f n (x p) = p x i (1 p) 40 x i The marginal joint p.d.f. g n (x) depends on the observed values x = x 1,..., x 40 but it does not depend on p and thus may be treated as a constant. Therefore, Equation 2.1 can be replaced with the following relation: ξ(p x) f n (x θξ(θ) p x i (1 p) 40 x i so that ξ(p x) is a beta distribution with 40 α = x i + 1 = i=1 40 β = n x i + 1 = 9 Figure 1.1.b shows the plot of beta(α =, β = 9). So ( P r p < 2 ) α =, β = 9 = 0.09 and thus, ( P r p 2 ) α =, β = 9 i=1 ( = 1 P r p < 2 = ) α =, β = 9 Remark. Bayesian approach tells us the probability that the parameter obtains values in a specified subset of the parameter space rather than merely supporting one decision over the other.
7 STATISTICAL INFERENCE 7 As discussed earlier, Bayesian inference takes into account prior knowledge about the distribution of the parameter. Example 1.4.(non-uniform prior) Suppose that there are n experiments of similar nature to the experiment in Example 1.2. These experiments had different sample sizes (N 1,..., N n ) and obtained different proportions of participants who exhibited change blindness (p 1,..., p n, respectively). We want to incorporate these information into the prior distribution. Suppose that the prior distribution of p is a beta distribution with parameters α and β. We may estimate α and β using the mean ˆµ and the variance s 2 of the observed values p i s, with i {1, 2,...n}. Since a larger sample size results in better estimation of the parameter, we may instead choose to estimate µ using the average of p i s weighted by the corresponding sample size: ˆµ = N 1p N n p n N N n Likewise, we may estimate σ 2 using the sample variance: n (p k µ) 2 s 2 = Since p beta(α, β), we also have: and σ 2 = k=1 n 1 µ = α α + β with n = 7 αβ (α + β) 2 (α + β + 1) Suppose that we solve the following system of equation and get αβ (α + β) 2 (α + β + 1) α α + β = N 1p N 7 p 7 N N 7 7 = α = 10 β = 10 k=1 (p k µ) Then, the prior distribution of p is beta(α = 10, β = 10). Figure 1.1.c gives the plot of beta(α = 10, β = 10). Apply Bayes Theorem, the posterior distribution of p becomes
8 8 NGO ξ(p x ) p x i (1 p) n xi p 9 (1 p) 9 p x i +9 (1 p) n+9 x i Therefore, the posterior distribution is a beta distribution with α = 6, β = 18 (see Figure 1.1.d) which gives P r(p 2 α = 6, β = 18) = a. b c. d. Figure 2.1 Remark. Comparing P r(p 2 α =, β = 9) and P r(p 2 α = 6, β = 18), we notice that incorporating previous knowledge into the prior distribution can have an impact on the posterior distribution. In this case, after taking into account previous information, we can say with lesser certainty that more than two thirds of the population exhibits change blindness. Moreover, with a uniform distribution, the conclusion is very similar to that of the hypothesis test but with another prior, it is very different Strengths and Weaknesses of Both Approaches. In the field of cognitive science, hypothesis testing has become the institutionalized approach to statistical analysis. Psychologists have been routinely trained to frame their research design and analysis in terms of rejecting null hypotheses. There are, however, several logical shortcomings pertaining to such an approach. The first problem, and the most important one, of the hypothesis testing is that nearly all null hypotheses are obviously false on a priori grounds. One example is testing if the proportion of patients showing no relapse after using the new drug is equal to the proportion of patients showing no relapse after using the old drug). Deming (1975)
9 STATISTICAL INFERENCE 9 comments that We do not perform an experiment to find out if two drugs are equal. We know in advance, without spending a dollar on an experiment, that they are not equal. The rejection therefore hardly gives any meaningful insights. Another argument is that the level of significance α is arbitrary and without theoretical basis. It merely classifies results into meaningless categories: significant and nonsignificant. More often than not, nonsignificant results are disregarded when they actually can potentially provide more insights than significant ones. Last but not least, the critical region, according to Kruschke (2010), is not well-defined because it is dependent upon the intentions of the experimenter. Different intentions result in different critical values. Figure 2.2 shows the sampling distribution of the test statistic for two groups when the null hypothesis is true, when the intention is to fix N = 6 for both groups regardless of how long it takes, or when the intention is to fix the duration of data collection at 2 weeks, when the mean rate is N = 6 per week. Suppose that we have data from two groups, with six values per group, and compute t to find that t = 2.5. We can only make the decision about rejection if we are aware of the experimenter s intentions. Such information, however, are not always accessible. While Bayesian inference approach can get away with the above mentioned problems of hypothesis testing approach, it is also subject to some level of subjectivity. First of all, the choice of prior distribution relies on subjective belief. For instance, in Example 1., the decision to estimate µ by putting weight the average of all p i s using the sample sizes relies on subjective judgment. Finally, the problems get complex, Bayesian approach may require much heavier computation than hypothesis testing approach. Despite these setbacks, Bayesian approach is superior in terms of its ability to incorporate previous knowledge into the analysis while hypothesis testing has no way of doing so. Moreover, the logic is different in the two approaches: hypothesis testing focuses on the probability of obtaining data as extreme as the observed data, P r(data hypothesis), whereas Bayesian inference is interested in how likely a hypothesis is in light of the data, P r(hypothesis data). The logic of Bayesian inference, therefore, could be argued to be more natural.
10 10 NGO a. b. Figure 2.2
11 STATISTICAL INFERENCE 11 References DeGroot, M. H., & Schervish, M. J. (2002). Probability and Statistics. Addison- Wesley. Kruschke, J. K. (2010). Bayesian data analysis. Wiley Interdisciplinary Reviews: Cognitive Science, 1(5), Simons, D. J., & Levin, D. T. (1998). Failure to detect changes to people during a real-world interaction. Psychonomic Bulletin & Review, 5(4),
Statistical Inference
Statistical Inference Classical and Bayesian Methods Class 6 AMS-UCSC Thu 26, 2012 Winter 2012. Session 1 (Class 6) AMS-132/206 Thu 26, 2012 1 / 15 Topics Topics We will talk about... 1 Hypothesis testing
More informationClassical and Bayesian inference
Classical and Bayesian inference AMS 132 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 8 1 / 11 The Prior Distribution Definition Suppose that one has a statistical model with parameter
More informationBAYESIAN DECISION THEORY
Last updated: September 17, 2012 BAYESIAN DECISION THEORY Problems 2 The following problems from the textbook are relevant: 2.1 2.9, 2.11, 2.17 For this week, please at least solve Problem 2.3. We will
More informationLecturer: Dr. Adote Anum, Dept. of Psychology Contact Information:
Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information: aanum@ug.edu.gh College of Education School of Continuing and Distance Education 2014/2015 2016/2017 Session Overview In this Session
More informationStatistical Inference: Estimation and Confidence Intervals Hypothesis Testing
Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire
More informationStatistical Inference: Uses, Abuses, and Misconceptions
Statistical Inference: Uses, Abuses, and Misconceptions Michael W. Trosset Indiana Statistical Consulting Center Department of Statistics ISCC is part of IU s Department of Statistics, chaired by Stanley
More informationHYPOTHESIS TESTING. Hypothesis Testing
MBA 605 Business Analytics Don Conant, PhD. HYPOTHESIS TESTING Hypothesis testing involves making inferences about the nature of the population on the basis of observations of a sample drawn from the population.
More informationSampling Distributions: Central Limit Theorem
Review for Exam 2 Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ)
More informationSampling Distributions
Sampling Distributions Sampling Distribution of the Mean & Hypothesis Testing Remember sampling? Sampling Part 1 of definition Selecting a subset of the population to create a sample Generally random sampling
More informationStructure learning in human causal induction
Structure learning in human causal induction Joshua B. Tenenbaum & Thomas L. Griffiths Department of Psychology Stanford University, Stanford, CA 94305 jbt,gruffydd @psych.stanford.edu Abstract We use
More informationReview: Statistical Model
Review: Statistical Model { f θ :θ Ω} A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced the data. The statistical model
More informationBios 6649: Clinical Trials - Statistical Design and Monitoring
Bios 6649: Clinical Trials - Statistical Design and Monitoring Spring Semester 2015 John M. Kittelson Department of Biostatistics & nformatics Colorado School of Public Health University of Colorado Denver
More informationNull Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017
Null Hypothesis Significance Testing p-values, significance level, power, t-tests 18.05 Spring 2017 Understand this figure f(x H 0 ) x reject H 0 don t reject H 0 reject H 0 x = test statistic f (x H 0
More informationClassical and Bayesian inference
Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli
More informationProbability and Probability Distributions. Dr. Mohammed Alahmed
Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about
More informationMachine Learning
Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 13, 2011 Today: The Big Picture Overfitting Review: probability Readings: Decision trees, overfiting
More informationChapter Three. Hypothesis Testing
3.1 Introduction The final phase of analyzing data is to make a decision concerning a set of choices or options. Should I invest in stocks or bonds? Should a new product be marketed? Are my products being
More information(1) Introduction to Bayesian statistics
Spring, 2018 A motivating example Student 1 will write down a number and then flip a coin If the flip is heads, they will honestly tell student 2 if the number is even or odd If the flip is tails, they
More information2 Belief, probability and exchangeability
2 Belief, probability and exchangeability We first discuss what properties a reasonable belief function should have, and show that probabilities have these properties. Then, we review the basic machinery
More informationLecture 5: Sampling Methods
Lecture 5: Sampling Methods What is sampling? Is the process of selecting part of a larger group of participants with the intent of generalizing the results from the smaller group, called the sample, to
More informationHypothesis testing: Steps
Review for Exam 2 Hypothesis testing: Steps Exam 2 Review 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region 3. Compute
More informationUniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence
Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Valen E. Johnson Texas A&M University February 27, 2014 Valen E. Johnson Texas A&M University Uniformly most powerful Bayes
More informationtheir contents. If the sample mean is 15.2 oz. and the sample standard deviation is 0.50 oz., find the 95% confidence interval of the true mean.
Math 1342 Exam 3-Review Chapters 7-9 HCCS **************************************************************************************** Name Date **********************************************************************************************
More informationDo not copy, post, or distribute
14 CORRELATION ANALYSIS AND LINEAR REGRESSION Assessing the Covariability of Two Quantitative Properties 14.0 LEARNING OBJECTIVES In this chapter, we discuss two related techniques for assessing a possible
More informationT.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS
ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only
More informationEnM Probability and Random Processes
Historical Note: EnM 503 - Probability and Random Processes Probability has its roots in games of chance, which have been played since prehistoric time. Games and equipment have been found in Egyptian
More informationPreliminary Statistics Lecture 5: Hypothesis Testing (Outline)
1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 5: Hypothesis Testing (Outline) Gujarati D. Basic Econometrics, Appendix A.8 Barrow M. Statistics
More informationLecture 1: Probability Fundamentals
Lecture 1: Probability Fundamentals IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge January 22nd, 2008 Rasmussen (CUED) Lecture 1: Probability
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationBayesian Statistics as an Alternative for Analyzing Data and Testing Hypotheses Benjamin Scheibehenne
Bayesian Statistics as an Alternative for Analyzing Data and Testing Hypotheses Benjamin Scheibehenne http://scheibehenne.de Can Social Norms Increase Towel Reuse? Standard Environmental Message Descriptive
More informationComputational Cognitive Science
Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning
More informationMAXIMUM LIKELIHOOD, SET ESTIMATION, MODEL CRITICISM
Eco517 Fall 2004 C. Sims MAXIMUM LIKELIHOOD, SET ESTIMATION, MODEL CRITICISM 1. SOMETHING WE SHOULD ALREADY HAVE MENTIONED A t n (µ, Σ) distribution converges, as n, to a N(µ, Σ). Consider the univariate
More information18.05 Practice Final Exam
No calculators. 18.05 Practice Final Exam Number of problems 16 concept questions, 16 problems. Simplifying expressions Unless asked to explicitly, you don t need to simplify complicated expressions. For
More informationNull Hypothesis Significance Testing p-values, significance level, power, t-tests
Null Hypothesis Significance Testing p-values, significance level, power, t-tests 18.05 Spring 2014 January 1, 2017 1 /22 Understand this figure f(x H 0 ) x reject H 0 don t reject H 0 reject H 0 x = test
More informationInference for a Population Proportion
Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist
More informationBayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida
Bayesian Statistical Methods Jeff Gill Department of Political Science, University of Florida 234 Anderson Hall, PO Box 117325, Gainesville, FL 32611-7325 Voice: 352-392-0262x272, Fax: 352-392-8127, Email:
More informationPreliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com
1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix
More informationStatistical Inference. Hypothesis Testing
Statistical Inference Hypothesis Testing Previously, we introduced the point and interval estimation of an unknown parameter(s), say µ and σ 2. However, in practice, the problem confronting the scientist
More informationPolitical Science 236 Hypothesis Testing: Review and Bootstrapping
Political Science 236 Hypothesis Testing: Review and Bootstrapping Rocío Titiunik Fall 2007 1 Hypothesis Testing Definition 1.1 Hypothesis. A hypothesis is a statement about a population parameter The
More information1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.
probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I
More informationTwo-stage Adaptive Randomization for Delayed Response in Clinical Trials
Two-stage Adaptive Randomization for Delayed Response in Clinical Trials Guosheng Yin Department of Statistics and Actuarial Science The University of Hong Kong Joint work with J. Xu PSI and RSS Journal
More informationThe t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary
Patrick Breheny October 13 Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/25 Introduction Introduction What s wrong with z-tests? So far we ve (thoroughly!) discussed how to carry out hypothesis
More informationMachine Learning
Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,
More informationIntroduction to Bayesian Statistics 1
Introduction to Bayesian Statistics 1 STA 442/2101 Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 42 Thomas Bayes (1701-1761) Image from the Wikipedia
More informationSingle Maths B: Introduction to Probability
Single Maths B: Introduction to Probability Overview Lecturer Email Office Homework Webpage Dr Jonathan Cumming j.a.cumming@durham.ac.uk CM233 None! http://maths.dur.ac.uk/stats/people/jac/singleb/ 1 Introduction
More informationHypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals
Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals Lecture 9 Justin Kern April 9, 2018 Measuring Effect Size: Cohen s d Simply finding whether a
More informationProblem Set #5. Due: 1pm on Friday, Nov 16th
1 Chris Piech CS 109 Problem Set #5 Due: 1pm on Friday, Nov 16th Problem Set #5 Nov 7 th, 2018 With problems by Mehran Sahami and Chris Piech For each problem, briefly explain/justify how you obtained
More informationFundamentals to Biostatistics. Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur
Fundamentals to Biostatistics Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur Statistics collection, analysis, interpretation of data development of new
More informationIntroductory Econometrics. Review of statistics (Part II: Inference)
Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will
More informationBayesian Inference for Normal Mean
Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density
More informationPubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH
PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH The First Step: SAMPLE SIZE DETERMINATION THE ULTIMATE GOAL The most important, ultimate step of any of clinical research is to do draw inferences;
More informationSlide 1 Math 1520, Lecture 21
Slide 1 Math 1520, Lecture 21 This lecture is concerned with a posteriori probability, which is the probability that a previous event had occurred given the outcome of a later event. Slide 2 Conditional
More informationHypothesis testing: Steps
Review for Exam 2 Hypothesis testing: Steps Repeated-Measures ANOVA 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region
More informationMATH 556: PROBABILITY PRIMER
MATH 6: PROBABILITY PRIMER 1 DEFINITIONS, TERMINOLOGY, NOTATION 1.1 EVENTS AND THE SAMPLE SPACE Definition 1.1 An experiment is a one-off or repeatable process or procedure for which (a there is a well-defined
More informationChapter 9: Hypothesis Testing Sections
1 / 22 : Hypothesis Testing Sections Skip: 9.2 Testing Simple Hypotheses Skip: 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.5 The t Test 9.6 Comparing the Means of Two Normal Distributions
More informationPurposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions
Part 1: Probability Distributions Purposes of Data Analysis True Distributions or Relationships in the Earths System Probability Distribution Normal Distribution Student-t Distribution Chi Square Distribution
More informationBasic Probabilistic Reasoning SEG
Basic Probabilistic Reasoning SEG 7450 1 Introduction Reasoning under uncertainty using probability theory Dealing with uncertainty is one of the main advantages of an expert system over a simple decision
More informationChapter 9 Inferences from Two Samples
Chapter 9 Inferences from Two Samples 9-1 Review and Preview 9-2 Two Proportions 9-3 Two Means: Independent Samples 9-4 Two Dependent Samples (Matched Pairs) 9-5 Two Variances or Standard Deviations Review
More informationCS 543 Page 1 John E. Boon, Jr.
CS 543 Machine Learning Spring 2010 Lecture 05 Evaluating Hypotheses I. Overview A. Given observed accuracy of a hypothesis over a limited sample of data, how well does this estimate its accuracy over
More informationStatistical Methods for Particle Physics (I)
Statistical Methods for Particle Physics (I) https://agenda.infn.it/conferencedisplay.py?confid=14407 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan
More informationStat 705: Completely randomized and complete block designs
Stat 705: Completely randomized and complete block designs Timothy Hanson Department of Statistics, University of South Carolina Stat 705: Data Analysis II 1 / 16 Experimental design Our department offers
More informationProbability and Statistics
Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT
More information2) There should be uncertainty as to which outcome will occur before the procedure takes place.
robability Numbers For many statisticians the concept of the probability that an event occurs is ultimately rooted in the interpretation of an event as an outcome of an experiment, others would interpret
More informationAgreement Coefficients and Statistical Inference
CHAPTER Agreement Coefficients and Statistical Inference OBJECTIVE This chapter describes several approaches for evaluating the precision associated with the inter-rater reliability coefficients of the
More informationBayesian Inference for Binomial Proportion
8 Bayesian Inference for Binomial Proportion Frequently there is a large population where π, a proportion of the population, has some attribute. For instance, the population could be registered voters
More informationReview of probability and statistics 1 / 31
Review of probability and statistics 1 / 31 2 / 31 Why? This chapter follows Stock and Watson (all graphs are from Stock and Watson). You may as well refer to the appendix in Wooldridge or any other introduction
More information18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages
Name No calculators. 18.05 Final Exam Number of problems 16 concept questions, 16 problems, 21 pages Extra paper If you need more space we will provide some blank paper. Indicate clearly that your solution
More informationClass 19. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700
Class 19 Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science Copyright 2017 by D.B. Rowe 1 Agenda: Recap Chapter 8.3-8.4 Lecture Chapter 8.5 Go over Exam. Problem Solving
More informationProbabilities & Statistics Revision
Probabilities & Statistics Revision Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 January 6, 2017 Christopher Ting QF
More informationHuman-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015
Probability Refresher Kai Arras, University of Freiburg Winter term 2014/2015 Probability Refresher Introduction to Probability Random variables Joint distribution Marginalization Conditional probability
More informationName: Exam: In-term Two Page: 1 of 8 Date: 12/07/2018. University of Texas at Austin, Department of Mathematics M358K - Applied Statistics TRUE/FALSE
Exam: In-term Two Page: 1 of 8 Date: 12/07/2018 Name: TRUE/FALSE 1.1 TRUE FALSE University of Texas at Austin, Department of Mathematics M358K - Applied Statistics MULTIPLE CHOICE 1.2 TRUE FALSE 1.3 TRUE
More informationShort Note: Naive Bayes Classifiers and Permanence of Ratios
Short Note: Naive Bayes Classifiers and Permanence of Ratios Julián M. Ortiz (jmo1@ualberta.ca) Department of Civil & Environmental Engineering University of Alberta Abstract The assumption of permanence
More informationHow Much Evidence Should One Collect?
How Much Evidence Should One Collect? Remco Heesen October 10, 2013 Abstract This paper focuses on the question how much evidence one should collect before deciding on the truth-value of a proposition.
More informationPhylogenetics: Bayesian Phylogenetic Analysis. COMP Spring 2015 Luay Nakhleh, Rice University
Phylogenetics: Bayesian Phylogenetic Analysis COMP 571 - Spring 2015 Luay Nakhleh, Rice University Bayes Rule P(X = x Y = y) = P(X = x, Y = y) P(Y = y) = P(X = x)p(y = y X = x) P x P(X = x 0 )P(Y = y X
More informationClassical and Bayesian inference
Classical and Bayesian inference AMS 132 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 8 1 / 8 Probability and Statistical Models Motivating ideas AMS 131: Suppose that the random variable
More informationAP Statistics Ch 12 Inference for Proportions
Ch 12.1 Inference for a Population Proportion Conditions for Inference The statistic that estimates the parameter p (population proportion) is the sample proportion p ˆ. p ˆ = Count of successes in the
More informationTopic 10: Hypothesis Testing
Topic 10: Hypothesis Testing Course 003, 2016 Page 0 The Problem of Hypothesis Testing A statistical hypothesis is an assertion or conjecture about the probability distribution of one or more random variables.
More informationNon-parametric Inference and Resampling
Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing
More informationthe amount of the data corresponding to the subinterval the width of the subinterval e x2 to the left by 5 units results in another PDF g(x) = 1 π
Math 10A with Professor Stankova Worksheet, Discussion #42; Friday, 12/8/2017 GSI name: Roy Zhao Problems 1. For each of the following distributions, derive/find all of the following: PMF/PDF, CDF, median,
More informationCS 540: Machine Learning Lecture 2: Review of Probability & Statistics
CS 540: Machine Learning Lecture 2: Review of Probability & Statistics AD January 2008 AD () January 2008 1 / 35 Outline Probability theory (PRML, Section 1.2) Statistics (PRML, Sections 2.1-2.4) AD ()
More informationDe Finetti s ultimate failure. Krzysztof Burdzy University of Washington
De Finetti s ultimate failure Krzysztof Burdzy University of Washington Does philosophy matter? Global temperatures will rise by 1 degree in 20 years with probability 80%. Reading suggestions Probability
More informationRelating Graph to Matlab
There are two related course documents on the web Probability and Statistics Review -should be read by people without statistics background and it is helpful as a review for those with prior statistics
More informationBeta statistics. Keywords. Bayes theorem. Bayes rule
Keywords Beta statistics Tommy Norberg tommy@chalmers.se Mathematical Sciences Chalmers University of Technology Gothenburg, SWEDEN Bayes s formula Prior density Likelihood Posterior density Conjugate
More informationParameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn
Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation
More informationHarvard University. Rigorous Research in Engineering Education
Statistical Inference Kari Lock Harvard University Department of Statistics Rigorous Research in Engineering Education 12/3/09 Statistical Inference You have a sample and want to use the data collected
More informationStatistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests
Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests http://benasque.org/2018tae/cgi-bin/talks/allprint.pl TAE 2018 Benasque, Spain 3-15 Sept 2018 Glen Cowan Physics
More informationUsing Probability to do Statistics.
Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving
More informationSTAT 830 Hypothesis Testing
STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These
More informationLecture Testing Hypotheses: The Neyman-Pearson Paradigm
Math 408 - Mathematical Statistics Lecture 29-30. Testing Hypotheses: The Neyman-Pearson Paradigm April 12-15, 2013 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 1 / 12 Agenda Example:
More informationModule 2 : Conditional Probability
Module 2 : Conditional Probability Ruben Zamar Department of Statistics UBC January 16, 2017 Ruben Zamar Department of Statistics UBC Module () 2 January 16, 2017 1 / 61 MOTIVATION The outcome could be
More informationBase Rates and Bayes Theorem
Slides to accompany Grove s handout March 8, 2016 Table of contents 1 Diagnostic and Prognostic Inference 2 Example inferences Does this patient have Alzheimer s disease, schizophrenia, depression, etc.?
More informationTheory of Maximum Likelihood Estimation. Konstantin Kashin
Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical
More informationProbability and Discrete Distributions
AMS 7L LAB #3 Fall, 2007 Objectives: Probability and Discrete Distributions 1. To explore relative frequency and the Law of Large Numbers 2. To practice the basic rules of probability 3. To work with the
More informationHow do we compare the relative performance among competing models?
How do we compare the relative performance among competing models? 1 Comparing Data Mining Methods Frequent problem: we want to know which of the two learning techniques is better How to reliably say Model
More informationSTAT 830 Hypothesis Testing
STAT 830 Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two
More informationNaïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability
Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish
More informationTwo-Sample Inferential Statistics
The t Test for Two Independent Samples 1 Two-Sample Inferential Statistics In an experiment there are two or more conditions One condition is often called the control condition in which the treatment is
More informationTopic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1
Topic -2 Probability Larson & Farber, Elementary Statistics: Picturing the World, 3e 1 Probability Experiments Experiment : An experiment is an act that can be repeated under given condition. Rolling a
More informationBayesian Statistics. Debdeep Pati Florida State University. February 11, 2016
Bayesian Statistics Debdeep Pati Florida State University February 11, 2016 Historical Background Historical Background Historical Background Brief History of Bayesian Statistics 1764-1838: called probability
More information