DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS

Size: px
Start display at page:

Download "DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS"

Transcription

1 DIFFERENT APPROACHES TO STATISTICAL INFERENCE: HYPOTHESIS TESTING VERSUS BAYESIAN ANALYSIS THUY ANH NGO 1. Introduction Statistics are easily come across in our daily life. Statements such as the average college student s debt after college is $2,700, or the average person spends 24 years of their life asleep, are results of statistical analysis. Statistical inference is one method of statistical analysis. It is the process of applying statistical methods in order to make probabilistic statements about some characteristics of the observed data. In cognitive science, statistical inference allows scientists to draw some conclusion about a psychological phenomenon by statistically analyzing a set of observations. For example, a social psychologist is interested in investigating whether the majority of the population shows change blindness (i.e., a phenomenon in which people often miss large changes in their visual field). To study this phenomenon, the scientist can collect data from the entire population of Americans. Such a task, however, is not economically feasible. Conveniently, statistical inference allows her to examine the change blindness phenomenon based on the data from a randomly selected pool of participants. There are different approaches to statistical inference, including hypothesis testing and Bayesian inference. Even though hypothesis testing is traditionally used by most cognitive scientists, Bayesian inference has garnered increased interest in recent years. This paper applies probability theory to explain these approaches, discusses the strengths and weaknesses of both approaches to explain the increasing use of Bayesian inference, especially in the field of cognitive science. 2. Two approaches to statistical inference 2.1. Important Statistical Concepts. We first start by defining key concepts. Definition 2.1 (Population vs. Sample). A population is a set of all observations that can possibly be made. A sample is a set of observations that are randomly selected from a population. Examples of a population include Trinity University students, people with anxiety disorder, trees in the state of Texas, etc. 1

2 2 NGO Examples of a sample include 70 students at Trinity chosen randomly, 20 clients with anxiety disorder at a hospital, 27 trees randomly selected in Texas, etc. Definition 2.2 (Parameter). A parameter is a measurable characteristic of a population. For example, when we examine whether a machine is functioning properly, we define the parameter to be the proportion of defective items in a large manufactured lot. Definition 2. (Random Variable). Let S be the sample space. A map X : S R is a random variable that assigns a real number X(s) to each possible outcome s S (DeGroot & Schervish, 2002). Definition 2.4 (Probability Function). Let X be a discrete random variable (X can take only a finite number k of different values x 1,..., x k ). The probability function (p.f.) of X is defined as the function f such that for every real number x, f(x) = P r(x = x). (DeGroot & Schervish, 2002) Definition 2.5 (Probability Density Function). Let X be a continuous random variable (X can take every value in an interval). The probability density function (p.d.f.) of X is the nonnegative function f, defined on R, such that for every subset A R, P r(x A) = f(x) dx. (DeGroot & Schervish, 2002) A Definition 2.6 (Statistical Inference). Statistical inference is the process of applying statistical methods in order to make probabilistic statements about some characteristics of the observed data Hypothesis Testing The Null and Alternative Hypotheses. In hypothesis testing approach, a parameter θ is an unknown value that lies in a parameter space Ω. A statistician s task is to decide whether the parameter value falls in one subset of Ω or its complement. We can then define our hypotheses as follows: Definition 2.7. Ω 0 and Ω 1 are disjoint subsets of the parameter space Ω with Ω 0 Ω 1 = Ω. The null and the alternative hypotheses are: H 0 : µ = µ 0, H A : µ µ 0.

3 STATISTICAL INFERENCE The hypothesis H 0 is called the null hypothesis and the hypothesis H A is called the alternative hypothesis. A statistician has to decide to accept either H 0 or H A. With only two possible decisions, accepting H 0 is equivalent to rejecting H A, and vice versa. In most statistical problems, the two hypotheses are treated quite differently. For example, in clinical trials, the null hypothesis almost always states that a particular treatment does not have an effect while the alternative hypothesis states that the treatment has an effect. Many cognitive scientists frame their research design and analysis in terms of rejecting the null hypothesis, a routine that has recently been debated. We will return to this debate later The Critical Region and Test Statistics. Suppose that X = (X 1,..., X n ) form a random sample from the population whose distribution involves the unknown parameter θ. Let S denote the set of all possible outcomes of the random sample X. A statistician then follows a test procedure that partitions S into two subsets, one of which contains values of X for which H 0 will be accepted while the other contains values of X for which H 0 will be rejected. Definition 2.8. The subset for which H 0 will be rejected is called the critical region, C. This region is defined in terms of a test statistic, T where T = f(x), a function of the observed data X. Example 1.1. Suppose that X = (X 1,..., X n ) is a random sample from a normal distribution with unknown mean µ and known variance σ 2. The hypotheses to be tested are: H 0 : µ = µ 0, H A : µ µ 0. H 0 is rejected if X n is far from µ 0. So if we define the test statistic to be T = X n µ 0, we then choose a test procedure that rejects H 0 if T c for some c > Level of Significance. Definition 2.9. The level of significance, α is determined by P r(t C) = α where C is the critical region. If T C, we reject H 0 in favor of H A on level α. Some standard values of α are 0.05, 0.01, Example 1.2. A psychologist wants to explore the change blindness phenomenon introduced above. She sets up an experiment in which an experimenter randomly approaches a pedestrian on a college campus to ask for directions (Simons & Levin, 1998). Shortly afterward, two men carrying a door pass in between them and interrupt the conversation. Meanwhile, the first experimenter is replaced with the second experimenter who has different height, different voice, and where different cloths from

4 4 NGO the first one. After the exchange, the second experimenter continues to talk to the pedestrian. Eventually, the pedestrian is asked if he/she noticed any changes throughout the interaction. Out of 40 pedestrians who were randomly selected to participate in the study, only 8 reported noticing the change. The psychologist would like to draw the conclusion whether at least two thirds of the people exhibit change blindness. Suppose that X = (X 1,..., X 40 ) is a random sample drawn from the population distribution with parameter p, where p is the probability of exhibiting change blindness. X i s are defined as follows: { 1, if the ith participant shows change blindness X i = 0, if the i th participant does not show change blindness for i {1, 2,..., 40}. X i = 1 with probability p and X 1 = 0 with probability q = p 1. Therefore, X 1,..., X 40 Bernoulli(p) and the p.f. of each X i can be written as follows: { p x i q f(x i p) = 1 x i, for x i = 0 or 1 0, otherwise The hypotheses are stated as follows: H 0 : p 2 (the majority doesn t exhibit change blindness), H A : p > 2 (the majority exhibits change blindness). The decision to reject or accept the null hypothesis first requires us to find the test statistic T and specify the critical region C. 40 Let the test statistic be Y = X i. So Y binomial(n = 40, p) and the p.f. is: i=1 f(y n, p) = ( ) n p y q n y, y for y = 0, 1,..., n 0, otherwise The larger p is, the larger we expect Y to be. Therefore, we choose to reject H 0 if Y c for some constant c. At the level of significance α 0 = 0.05, we reject H 0 if P r(y 2 p = 2 ) < α 0. We should now compute P r(y 2 p = 2 ): ( P r Y 2 p = 2 ) ( = 1 P r Y 1 p = 2 )

5 STATISTICAL INFERENCE 5 40 ( ) 40 = 1 p k (1 p) 40 k, p = 2 k k=0 40 = 1 = k=0 ( 40 k ) ( 2 ) k ( 1 2 ) 40 k Since P r(y 2 p = 2 ) = 2 ) < 0.05, we reject H 0 and conclude that the majority of people does exhibit change blindness. Remark. Rejecting the null hypothesis does not mean we prove that it is false. It is only that the data support a rejection. Also, in our example, we reject the null hypothesis at the 5% level of significance (α 0 = 0.05) but we cannot reject the null hypothesis at the 1% level of significance. In brief, our decision depends on our choice of the significance level, which can be quite arbitrary in nature. This is a typical criticism of hypothesis testing, and we will return to it later. 2.. Bayesian Inference. Bayesian inference relies on the same logic that is used by the famous fictional detective, Sherlock Holmes: when you have eliminated the impossible, whatever remains, however improbable, must be the truth. Likewise, Bayesian inference is the method in which we begin with a set of all possibilities, then based on collected data, we eliminate some possibilities and reallocate credibilities to the remaining ones. Consequently, a parameter is a random variable which has a probability distribution over the set of all candidate values. We say that the parameter has a prior distribution, ξ(θ). The prior distribution can be determined by utilizing previous information or knowledge about the parameter. After observing the data, reallocation of credibilities is reflected in the posterior distribution, which is conditioned on observed data. The posterior distribution is denoted as ξ(θ x) where x = (x 1,..., x n ) is the vector that represents the observed data. The relationship between the prior and the posterior distribution of a parameter is given by the Bayes theorem. Theorem 2.1 (Bayes Theorem). Let X 1,..., X n be i.i.d. random variables with p.d.f. f(x i θ) and joint p.d.f. f n (x θ). Let g n (x) = f n (x 1,..., x n θ)ξ(θ)dθ. Then, Ω ξ(θ x) = f n(x θ)ξ(θ) g n (x)

6 6 NGO Example 1.. (uniform prior) Consider the experiment described in Example 1.2. Let p be the probability of exhibiting change blindness and let X i = 1 if the ith participant shows change blindness and X i = 0 if not. Given any p, the distribution of each X i is: f(x i p) = { p x i (1 p) 1 x i, for x i = 0 or 1 0, otherwise Suppose that we have no prior knowledge about what values of p are more probable than others, we assume uniform distribution for p, so ξ(p) = 1 and its plot is given in Figure 2.1.a. We are interested in finding the distribution of p after observing some participants in the study. There were 40 pedestrians and 2 of them showed change blindness. The joint distribution of X 1,...X 40 is given by: f n (x p) = p x i (1 p) 40 x i The marginal joint p.d.f. g n (x) depends on the observed values x = x 1,..., x 40 but it does not depend on p and thus may be treated as a constant. Therefore, Equation 2.1 can be replaced with the following relation: ξ(p x) f n (x θξ(θ) p x i (1 p) 40 x i so that ξ(p x) is a beta distribution with 40 α = x i + 1 = i=1 40 β = n x i + 1 = 9 Figure 1.1.b shows the plot of beta(α =, β = 9). So ( P r p < 2 ) α =, β = 9 = 0.09 and thus, ( P r p 2 ) α =, β = 9 i=1 ( = 1 P r p < 2 = ) α =, β = 9 Remark. Bayesian approach tells us the probability that the parameter obtains values in a specified subset of the parameter space rather than merely supporting one decision over the other.

7 STATISTICAL INFERENCE 7 As discussed earlier, Bayesian inference takes into account prior knowledge about the distribution of the parameter. Example 1.4.(non-uniform prior) Suppose that there are n experiments of similar nature to the experiment in Example 1.2. These experiments had different sample sizes (N 1,..., N n ) and obtained different proportions of participants who exhibited change blindness (p 1,..., p n, respectively). We want to incorporate these information into the prior distribution. Suppose that the prior distribution of p is a beta distribution with parameters α and β. We may estimate α and β using the mean ˆµ and the variance s 2 of the observed values p i s, with i {1, 2,...n}. Since a larger sample size results in better estimation of the parameter, we may instead choose to estimate µ using the average of p i s weighted by the corresponding sample size: ˆµ = N 1p N n p n N N n Likewise, we may estimate σ 2 using the sample variance: n (p k µ) 2 s 2 = Since p beta(α, β), we also have: and σ 2 = k=1 n 1 µ = α α + β with n = 7 αβ (α + β) 2 (α + β + 1) Suppose that we solve the following system of equation and get αβ (α + β) 2 (α + β + 1) α α + β = N 1p N 7 p 7 N N 7 7 = α = 10 β = 10 k=1 (p k µ) Then, the prior distribution of p is beta(α = 10, β = 10). Figure 1.1.c gives the plot of beta(α = 10, β = 10). Apply Bayes Theorem, the posterior distribution of p becomes

8 8 NGO ξ(p x ) p x i (1 p) n xi p 9 (1 p) 9 p x i +9 (1 p) n+9 x i Therefore, the posterior distribution is a beta distribution with α = 6, β = 18 (see Figure 1.1.d) which gives P r(p 2 α = 6, β = 18) = a. b c. d. Figure 2.1 Remark. Comparing P r(p 2 α =, β = 9) and P r(p 2 α = 6, β = 18), we notice that incorporating previous knowledge into the prior distribution can have an impact on the posterior distribution. In this case, after taking into account previous information, we can say with lesser certainty that more than two thirds of the population exhibits change blindness. Moreover, with a uniform distribution, the conclusion is very similar to that of the hypothesis test but with another prior, it is very different Strengths and Weaknesses of Both Approaches. In the field of cognitive science, hypothesis testing has become the institutionalized approach to statistical analysis. Psychologists have been routinely trained to frame their research design and analysis in terms of rejecting null hypotheses. There are, however, several logical shortcomings pertaining to such an approach. The first problem, and the most important one, of the hypothesis testing is that nearly all null hypotheses are obviously false on a priori grounds. One example is testing if the proportion of patients showing no relapse after using the new drug is equal to the proportion of patients showing no relapse after using the old drug). Deming (1975)

9 STATISTICAL INFERENCE 9 comments that We do not perform an experiment to find out if two drugs are equal. We know in advance, without spending a dollar on an experiment, that they are not equal. The rejection therefore hardly gives any meaningful insights. Another argument is that the level of significance α is arbitrary and without theoretical basis. It merely classifies results into meaningless categories: significant and nonsignificant. More often than not, nonsignificant results are disregarded when they actually can potentially provide more insights than significant ones. Last but not least, the critical region, according to Kruschke (2010), is not well-defined because it is dependent upon the intentions of the experimenter. Different intentions result in different critical values. Figure 2.2 shows the sampling distribution of the test statistic for two groups when the null hypothesis is true, when the intention is to fix N = 6 for both groups regardless of how long it takes, or when the intention is to fix the duration of data collection at 2 weeks, when the mean rate is N = 6 per week. Suppose that we have data from two groups, with six values per group, and compute t to find that t = 2.5. We can only make the decision about rejection if we are aware of the experimenter s intentions. Such information, however, are not always accessible. While Bayesian inference approach can get away with the above mentioned problems of hypothesis testing approach, it is also subject to some level of subjectivity. First of all, the choice of prior distribution relies on subjective belief. For instance, in Example 1., the decision to estimate µ by putting weight the average of all p i s using the sample sizes relies on subjective judgment. Finally, the problems get complex, Bayesian approach may require much heavier computation than hypothesis testing approach. Despite these setbacks, Bayesian approach is superior in terms of its ability to incorporate previous knowledge into the analysis while hypothesis testing has no way of doing so. Moreover, the logic is different in the two approaches: hypothesis testing focuses on the probability of obtaining data as extreme as the observed data, P r(data hypothesis), whereas Bayesian inference is interested in how likely a hypothesis is in light of the data, P r(hypothesis data). The logic of Bayesian inference, therefore, could be argued to be more natural.

10 10 NGO a. b. Figure 2.2

11 STATISTICAL INFERENCE 11 References DeGroot, M. H., & Schervish, M. J. (2002). Probability and Statistics. Addison- Wesley. Kruschke, J. K. (2010). Bayesian data analysis. Wiley Interdisciplinary Reviews: Cognitive Science, 1(5), Simons, D. J., & Levin, D. T. (1998). Failure to detect changes to people during a real-world interaction. Psychonomic Bulletin & Review, 5(4),

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Class 6 AMS-UCSC Thu 26, 2012 Winter 2012. Session 1 (Class 6) AMS-132/206 Thu 26, 2012 1 / 15 Topics Topics We will talk about... 1 Hypothesis testing

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 8 1 / 11 The Prior Distribution Definition Suppose that one has a statistical model with parameter

More information

BAYESIAN DECISION THEORY

BAYESIAN DECISION THEORY Last updated: September 17, 2012 BAYESIAN DECISION THEORY Problems 2 The following problems from the textbook are relevant: 2.1 2.9, 2.11, 2.17 For this week, please at least solve Problem 2.3. We will

More information

Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information:

Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information: Lecturer: Dr. Adote Anum, Dept. of Psychology Contact Information: aanum@ug.edu.gh College of Education School of Continuing and Distance Education 2014/2015 2016/2017 Session Overview In this Session

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

Statistical Inference: Uses, Abuses, and Misconceptions

Statistical Inference: Uses, Abuses, and Misconceptions Statistical Inference: Uses, Abuses, and Misconceptions Michael W. Trosset Indiana Statistical Consulting Center Department of Statistics ISCC is part of IU s Department of Statistics, chaired by Stanley

More information

HYPOTHESIS TESTING. Hypothesis Testing

HYPOTHESIS TESTING. Hypothesis Testing MBA 605 Business Analytics Don Conant, PhD. HYPOTHESIS TESTING Hypothesis testing involves making inferences about the nature of the population on the basis of observations of a sample drawn from the population.

More information

Sampling Distributions: Central Limit Theorem

Sampling Distributions: Central Limit Theorem Review for Exam 2 Sampling Distributions: Central Limit Theorem Conceptually, we can break up the theorem into three parts: 1. The mean (µ M ) of a population of sample means (M) is equal to the mean (µ)

More information

Sampling Distributions

Sampling Distributions Sampling Distributions Sampling Distribution of the Mean & Hypothesis Testing Remember sampling? Sampling Part 1 of definition Selecting a subset of the population to create a sample Generally random sampling

More information

Structure learning in human causal induction

Structure learning in human causal induction Structure learning in human causal induction Joshua B. Tenenbaum & Thomas L. Griffiths Department of Psychology Stanford University, Stanford, CA 94305 jbt,gruffydd @psych.stanford.edu Abstract We use

More information

Review: Statistical Model

Review: Statistical Model Review: Statistical Model { f θ :θ Ω} A statistical model for some data is a set of distributions, one of which corresponds to the true unknown distribution that produced the data. The statistical model

More information

Bios 6649: Clinical Trials - Statistical Design and Monitoring

Bios 6649: Clinical Trials - Statistical Design and Monitoring Bios 6649: Clinical Trials - Statistical Design and Monitoring Spring Semester 2015 John M. Kittelson Department of Biostatistics & nformatics Colorado School of Public Health University of Colorado Denver

More information

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Spring 2017 Null Hypothesis Significance Testing p-values, significance level, power, t-tests 18.05 Spring 2017 Understand this figure f(x H 0 ) x reject H 0 don t reject H 0 reject H 0 x = test statistic f (x H 0

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 January 18, 2018 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 18, 2018 1 / 9 Sampling from a Bernoulli Distribution Theorem (Beta-Bernoulli

More information

Probability and Probability Distributions. Dr. Mohammed Alahmed

Probability and Probability Distributions. Dr. Mohammed Alahmed Probability and Probability Distributions 1 Probability and Probability Distributions Usually we want to do more with data than just describing them! We might want to test certain specific inferences about

More information

Machine Learning

Machine Learning Machine Learning 10-701 Tom M. Mitchell Machine Learning Department Carnegie Mellon University January 13, 2011 Today: The Big Picture Overfitting Review: probability Readings: Decision trees, overfiting

More information

Chapter Three. Hypothesis Testing

Chapter Three. Hypothesis Testing 3.1 Introduction The final phase of analyzing data is to make a decision concerning a set of choices or options. Should I invest in stocks or bonds? Should a new product be marketed? Are my products being

More information

(1) Introduction to Bayesian statistics

(1) Introduction to Bayesian statistics Spring, 2018 A motivating example Student 1 will write down a number and then flip a coin If the flip is heads, they will honestly tell student 2 if the number is even or odd If the flip is tails, they

More information

2 Belief, probability and exchangeability

2 Belief, probability and exchangeability 2 Belief, probability and exchangeability We first discuss what properties a reasonable belief function should have, and show that probabilities have these properties. Then, we review the basic machinery

More information

Lecture 5: Sampling Methods

Lecture 5: Sampling Methods Lecture 5: Sampling Methods What is sampling? Is the process of selecting part of a larger group of participants with the intent of generalizing the results from the smaller group, called the sample, to

More information

Hypothesis testing: Steps

Hypothesis testing: Steps Review for Exam 2 Hypothesis testing: Steps Exam 2 Review 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region 3. Compute

More information

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence Valen E. Johnson Texas A&M University February 27, 2014 Valen E. Johnson Texas A&M University Uniformly most powerful Bayes

More information

their contents. If the sample mean is 15.2 oz. and the sample standard deviation is 0.50 oz., find the 95% confidence interval of the true mean.

their contents. If the sample mean is 15.2 oz. and the sample standard deviation is 0.50 oz., find the 95% confidence interval of the true mean. Math 1342 Exam 3-Review Chapters 7-9 HCCS **************************************************************************************** Name Date **********************************************************************************************

More information

Do not copy, post, or distribute

Do not copy, post, or distribute 14 CORRELATION ANALYSIS AND LINEAR REGRESSION Assessing the Covariability of Two Quantitative Properties 14.0 LEARNING OBJECTIVES In this chapter, we discuss two related techniques for assessing a possible

More information

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS

T.I.H.E. IT 233 Statistics and Probability: Sem. 1: 2013 ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS ESTIMATION AND HYPOTHESIS TESTING OF TWO POPULATIONS In our work on hypothesis testing, we used the value of a sample statistic to challenge an accepted value of a population parameter. We focused only

More information

EnM Probability and Random Processes

EnM Probability and Random Processes Historical Note: EnM 503 - Probability and Random Processes Probability has its roots in games of chance, which have been played since prehistoric time. Games and equipment have been found in Egyptian

More information

Preliminary Statistics Lecture 5: Hypothesis Testing (Outline)

Preliminary Statistics Lecture 5: Hypothesis Testing (Outline) 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 5: Hypothesis Testing (Outline) Gujarati D. Basic Econometrics, Appendix A.8 Barrow M. Statistics

More information

Lecture 1: Probability Fundamentals

Lecture 1: Probability Fundamentals Lecture 1: Probability Fundamentals IB Paper 7: Probability and Statistics Carl Edward Rasmussen Department of Engineering, University of Cambridge January 22nd, 2008 Rasmussen (CUED) Lecture 1: Probability

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Bayesian Statistics as an Alternative for Analyzing Data and Testing Hypotheses Benjamin Scheibehenne

Bayesian Statistics as an Alternative for Analyzing Data and Testing Hypotheses Benjamin Scheibehenne Bayesian Statistics as an Alternative for Analyzing Data and Testing Hypotheses Benjamin Scheibehenne http://scheibehenne.de Can Social Norms Increase Towel Reuse? Standard Environmental Message Descriptive

More information

Computational Cognitive Science

Computational Cognitive Science Computational Cognitive Science Lecture 9: A Bayesian model of concept learning Chris Lucas School of Informatics University of Edinburgh October 16, 218 Reading Rules and Similarity in Concept Learning

More information

MAXIMUM LIKELIHOOD, SET ESTIMATION, MODEL CRITICISM

MAXIMUM LIKELIHOOD, SET ESTIMATION, MODEL CRITICISM Eco517 Fall 2004 C. Sims MAXIMUM LIKELIHOOD, SET ESTIMATION, MODEL CRITICISM 1. SOMETHING WE SHOULD ALREADY HAVE MENTIONED A t n (µ, Σ) distribution converges, as n, to a N(µ, Σ). Consider the univariate

More information

18.05 Practice Final Exam

18.05 Practice Final Exam No calculators. 18.05 Practice Final Exam Number of problems 16 concept questions, 16 problems. Simplifying expressions Unless asked to explicitly, you don t need to simplify complicated expressions. For

More information

Null Hypothesis Significance Testing p-values, significance level, power, t-tests

Null Hypothesis Significance Testing p-values, significance level, power, t-tests Null Hypothesis Significance Testing p-values, significance level, power, t-tests 18.05 Spring 2014 January 1, 2017 1 /22 Understand this figure f(x H 0 ) x reject H 0 don t reject H 0 reject H 0 x = test

More information

Inference for a Population Proportion

Inference for a Population Proportion Al Nosedal. University of Toronto. November 11, 2015 Statistical inference is drawing conclusions about an entire population based on data in a sample drawn from that population. From both frequentist

More information

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida

Bayesian Statistical Methods. Jeff Gill. Department of Political Science, University of Florida Bayesian Statistical Methods Jeff Gill Department of Political Science, University of Florida 234 Anderson Hall, PO Box 117325, Gainesville, FL 32611-7325 Voice: 352-392-0262x272, Fax: 352-392-8127, Email:

More information

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com

Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com 1 School of Oriental and African Studies September 2015 Department of Economics Preliminary Statistics Lecture 2: Probability Theory (Outline) prelimsoas.webs.com Gujarati D. Basic Econometrics, Appendix

More information

Statistical Inference. Hypothesis Testing

Statistical Inference. Hypothesis Testing Statistical Inference Hypothesis Testing Previously, we introduced the point and interval estimation of an unknown parameter(s), say µ and σ 2. However, in practice, the problem confronting the scientist

More information

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Political Science 236 Hypothesis Testing: Review and Bootstrapping Political Science 236 Hypothesis Testing: Review and Bootstrapping Rocío Titiunik Fall 2007 1 Hypothesis Testing Definition 1.1 Hypothesis. A hypothesis is a statement about a population parameter The

More information

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D.

1 A simple example. A short introduction to Bayesian statistics, part I Math 217 Probability and Statistics Prof. D. probabilities, we ll use Bayes formula. We can easily compute the reverse probabilities A short introduction to Bayesian statistics, part I Math 17 Probability and Statistics Prof. D. Joyce, Fall 014 I

More information

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials

Two-stage Adaptive Randomization for Delayed Response in Clinical Trials Two-stage Adaptive Randomization for Delayed Response in Clinical Trials Guosheng Yin Department of Statistics and Actuarial Science The University of Hong Kong Joint work with J. Xu PSI and RSS Journal

More information

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary

The t-distribution. Patrick Breheny. October 13. z tests The χ 2 -distribution The t-distribution Summary Patrick Breheny October 13 Patrick Breheny Biostatistical Methods I (BIOS 5710) 1/25 Introduction Introduction What s wrong with z-tests? So far we ve (thoroughly!) discussed how to carry out hypothesis

More information

Machine Learning

Machine Learning Machine Learning 10-601 Tom M. Mitchell Machine Learning Department Carnegie Mellon University August 30, 2017 Today: Decision trees Overfitting The Big Picture Coming soon Probabilistic learning MLE,

More information

Introduction to Bayesian Statistics 1

Introduction to Bayesian Statistics 1 Introduction to Bayesian Statistics 1 STA 442/2101 Fall 2018 1 This slide show is an open-source document. See last slide for copyright information. 1 / 42 Thomas Bayes (1701-1761) Image from the Wikipedia

More information

Single Maths B: Introduction to Probability

Single Maths B: Introduction to Probability Single Maths B: Introduction to Probability Overview Lecturer Email Office Homework Webpage Dr Jonathan Cumming j.a.cumming@durham.ac.uk CM233 None! http://maths.dur.ac.uk/stats/people/jac/singleb/ 1 Introduction

More information

Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals

Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals Hypothesis Testing and Confidence Intervals (Part 2): Cohen s d, Logic of Testing, and Confidence Intervals Lecture 9 Justin Kern April 9, 2018 Measuring Effect Size: Cohen s d Simply finding whether a

More information

Problem Set #5. Due: 1pm on Friday, Nov 16th

Problem Set #5. Due: 1pm on Friday, Nov 16th 1 Chris Piech CS 109 Problem Set #5 Due: 1pm on Friday, Nov 16th Problem Set #5 Nov 7 th, 2018 With problems by Mehran Sahami and Chris Piech For each problem, briefly explain/justify how you obtained

More information

Fundamentals to Biostatistics. Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur

Fundamentals to Biostatistics. Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur Fundamentals to Biostatistics Prof. Chandan Chakraborty Associate Professor School of Medical Science & Technology IIT Kharagpur Statistics collection, analysis, interpretation of data development of new

More information

Introductory Econometrics. Review of statistics (Part II: Inference)

Introductory Econometrics. Review of statistics (Part II: Inference) Introductory Econometrics Review of statistics (Part II: Inference) Jun Ma School of Economics Renmin University of China October 1, 2018 1/16 Null and alternative hypotheses Usually, we have two competing

More information

Statistical Inference

Statistical Inference Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will

More information

Bayesian Inference for Normal Mean

Bayesian Inference for Normal Mean Al Nosedal. University of Toronto. November 18, 2015 Likelihood of Single Observation The conditional observation distribution of y µ is Normal with mean µ and variance σ 2, which is known. Its density

More information

PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH

PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH PubH 7470: STATISTICS FOR TRANSLATIONAL & CLINICAL RESEARCH The First Step: SAMPLE SIZE DETERMINATION THE ULTIMATE GOAL The most important, ultimate step of any of clinical research is to do draw inferences;

More information

Slide 1 Math 1520, Lecture 21

Slide 1 Math 1520, Lecture 21 Slide 1 Math 1520, Lecture 21 This lecture is concerned with a posteriori probability, which is the probability that a previous event had occurred given the outcome of a later event. Slide 2 Conditional

More information

Hypothesis testing: Steps

Hypothesis testing: Steps Review for Exam 2 Hypothesis testing: Steps Repeated-Measures ANOVA 1. Determine appropriate test and hypotheses 2. Use distribution table to find critical statistic value(s) representing rejection region

More information

MATH 556: PROBABILITY PRIMER

MATH 556: PROBABILITY PRIMER MATH 6: PROBABILITY PRIMER 1 DEFINITIONS, TERMINOLOGY, NOTATION 1.1 EVENTS AND THE SAMPLE SPACE Definition 1.1 An experiment is a one-off or repeatable process or procedure for which (a there is a well-defined

More information

Chapter 9: Hypothesis Testing Sections

Chapter 9: Hypothesis Testing Sections 1 / 22 : Hypothesis Testing Sections Skip: 9.2 Testing Simple Hypotheses Skip: 9.3 Uniformly Most Powerful Tests Skip: 9.4 Two-Sided Alternatives 9.5 The t Test 9.6 Comparing the Means of Two Normal Distributions

More information

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions Part 1: Probability Distributions Purposes of Data Analysis True Distributions or Relationships in the Earths System Probability Distribution Normal Distribution Student-t Distribution Chi Square Distribution

More information

Basic Probabilistic Reasoning SEG

Basic Probabilistic Reasoning SEG Basic Probabilistic Reasoning SEG 7450 1 Introduction Reasoning under uncertainty using probability theory Dealing with uncertainty is one of the main advantages of an expert system over a simple decision

More information

Chapter 9 Inferences from Two Samples

Chapter 9 Inferences from Two Samples Chapter 9 Inferences from Two Samples 9-1 Review and Preview 9-2 Two Proportions 9-3 Two Means: Independent Samples 9-4 Two Dependent Samples (Matched Pairs) 9-5 Two Variances or Standard Deviations Review

More information

CS 543 Page 1 John E. Boon, Jr.

CS 543 Page 1 John E. Boon, Jr. CS 543 Machine Learning Spring 2010 Lecture 05 Evaluating Hypotheses I. Overview A. Given observed accuracy of a hypothesis over a limited sample of data, how well does this estimate its accuracy over

More information

Statistical Methods for Particle Physics (I)

Statistical Methods for Particle Physics (I) Statistical Methods for Particle Physics (I) https://agenda.infn.it/conferencedisplay.py?confid=14407 Glen Cowan Physics Department Royal Holloway, University of London g.cowan@rhul.ac.uk www.pp.rhul.ac.uk/~cowan

More information

Stat 705: Completely randomized and complete block designs

Stat 705: Completely randomized and complete block designs Stat 705: Completely randomized and complete block designs Timothy Hanson Department of Statistics, University of South Carolina Stat 705: Data Analysis II 1 / 16 Experimental design Our department offers

More information

Probability and Statistics

Probability and Statistics Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT

More information

2) There should be uncertainty as to which outcome will occur before the procedure takes place.

2) There should be uncertainty as to which outcome will occur before the procedure takes place. robability Numbers For many statisticians the concept of the probability that an event occurs is ultimately rooted in the interpretation of an event as an outcome of an experiment, others would interpret

More information

Agreement Coefficients and Statistical Inference

Agreement Coefficients and Statistical Inference CHAPTER Agreement Coefficients and Statistical Inference OBJECTIVE This chapter describes several approaches for evaluating the precision associated with the inter-rater reliability coefficients of the

More information

Bayesian Inference for Binomial Proportion

Bayesian Inference for Binomial Proportion 8 Bayesian Inference for Binomial Proportion Frequently there is a large population where π, a proportion of the population, has some attribute. For instance, the population could be registered voters

More information

Review of probability and statistics 1 / 31

Review of probability and statistics 1 / 31 Review of probability and statistics 1 / 31 2 / 31 Why? This chapter follows Stock and Watson (all graphs are from Stock and Watson). You may as well refer to the appendix in Wooldridge or any other introduction

More information

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages Name No calculators. 18.05 Final Exam Number of problems 16 concept questions, 16 problems, 21 pages Extra paper If you need more space we will provide some blank paper. Indicate clearly that your solution

More information

Class 19. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700

Class 19. Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science. Marquette University MATH 1700 Class 19 Daniel B. Rowe, Ph.D. Department of Mathematics, Statistics, and Computer Science Copyright 2017 by D.B. Rowe 1 Agenda: Recap Chapter 8.3-8.4 Lecture Chapter 8.5 Go over Exam. Problem Solving

More information

Probabilities & Statistics Revision

Probabilities & Statistics Revision Probabilities & Statistics Revision Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 January 6, 2017 Christopher Ting QF

More information

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015 Probability Refresher Kai Arras, University of Freiburg Winter term 2014/2015 Probability Refresher Introduction to Probability Random variables Joint distribution Marginalization Conditional probability

More information

Name: Exam: In-term Two Page: 1 of 8 Date: 12/07/2018. University of Texas at Austin, Department of Mathematics M358K - Applied Statistics TRUE/FALSE

Name: Exam: In-term Two Page: 1 of 8 Date: 12/07/2018. University of Texas at Austin, Department of Mathematics M358K - Applied Statistics TRUE/FALSE Exam: In-term Two Page: 1 of 8 Date: 12/07/2018 Name: TRUE/FALSE 1.1 TRUE FALSE University of Texas at Austin, Department of Mathematics M358K - Applied Statistics MULTIPLE CHOICE 1.2 TRUE FALSE 1.3 TRUE

More information

Short Note: Naive Bayes Classifiers and Permanence of Ratios

Short Note: Naive Bayes Classifiers and Permanence of Ratios Short Note: Naive Bayes Classifiers and Permanence of Ratios Julián M. Ortiz (jmo1@ualberta.ca) Department of Civil & Environmental Engineering University of Alberta Abstract The assumption of permanence

More information

How Much Evidence Should One Collect?

How Much Evidence Should One Collect? How Much Evidence Should One Collect? Remco Heesen October 10, 2013 Abstract This paper focuses on the question how much evidence one should collect before deciding on the truth-value of a proposition.

More information

Phylogenetics: Bayesian Phylogenetic Analysis. COMP Spring 2015 Luay Nakhleh, Rice University

Phylogenetics: Bayesian Phylogenetic Analysis. COMP Spring 2015 Luay Nakhleh, Rice University Phylogenetics: Bayesian Phylogenetic Analysis COMP 571 - Spring 2015 Luay Nakhleh, Rice University Bayes Rule P(X = x Y = y) = P(X = x, Y = y) P(Y = y) = P(X = x)p(y = y X = x) P x P(X = x 0 )P(Y = y X

More information

Classical and Bayesian inference

Classical and Bayesian inference Classical and Bayesian inference AMS 132 Claudia Wehrhahn (UCSC) Classical and Bayesian inference January 8 1 / 8 Probability and Statistical Models Motivating ideas AMS 131: Suppose that the random variable

More information

AP Statistics Ch 12 Inference for Proportions

AP Statistics Ch 12 Inference for Proportions Ch 12.1 Inference for a Population Proportion Conditions for Inference The statistic that estimates the parameter p (population proportion) is the sample proportion p ˆ. p ˆ = Count of successes in the

More information

Topic 10: Hypothesis Testing

Topic 10: Hypothesis Testing Topic 10: Hypothesis Testing Course 003, 2016 Page 0 The Problem of Hypothesis Testing A statistical hypothesis is an assertion or conjecture about the probability distribution of one or more random variables.

More information

Non-parametric Inference and Resampling

Non-parametric Inference and Resampling Non-parametric Inference and Resampling Exercises by David Wozabal (Last update. Juni 010) 1 Basic Facts about Rank and Order Statistics 1.1 10 students were asked about the amount of time they spend surfing

More information

the amount of the data corresponding to the subinterval the width of the subinterval e x2 to the left by 5 units results in another PDF g(x) = 1 π

the amount of the data corresponding to the subinterval the width of the subinterval e x2 to the left by 5 units results in another PDF g(x) = 1 π Math 10A with Professor Stankova Worksheet, Discussion #42; Friday, 12/8/2017 GSI name: Roy Zhao Problems 1. For each of the following distributions, derive/find all of the following: PMF/PDF, CDF, median,

More information

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics

CS 540: Machine Learning Lecture 2: Review of Probability & Statistics CS 540: Machine Learning Lecture 2: Review of Probability & Statistics AD January 2008 AD () January 2008 1 / 35 Outline Probability theory (PRML, Section 1.2) Statistics (PRML, Sections 2.1-2.4) AD ()

More information

De Finetti s ultimate failure. Krzysztof Burdzy University of Washington

De Finetti s ultimate failure. Krzysztof Burdzy University of Washington De Finetti s ultimate failure Krzysztof Burdzy University of Washington Does philosophy matter? Global temperatures will rise by 1 degree in 20 years with probability 80%. Reading suggestions Probability

More information

Relating Graph to Matlab

Relating Graph to Matlab There are two related course documents on the web Probability and Statistics Review -should be read by people without statistics background and it is helpful as a review for those with prior statistics

More information

Beta statistics. Keywords. Bayes theorem. Bayes rule

Beta statistics. Keywords. Bayes theorem. Bayes rule Keywords Beta statistics Tommy Norberg tommy@chalmers.se Mathematical Sciences Chalmers University of Technology Gothenburg, SWEDEN Bayes s formula Prior density Likelihood Posterior density Conjugate

More information

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn Parameter estimation and forecasting Cristiano Porciani AIfA, Uni-Bonn Questions? C. Porciani Estimation & forecasting 2 Temperature fluctuations Variance at multipole l (angle ~180o/l) C. Porciani Estimation

More information

Harvard University. Rigorous Research in Engineering Education

Harvard University. Rigorous Research in Engineering Education Statistical Inference Kari Lock Harvard University Department of Statistics Rigorous Research in Engineering Education 12/3/09 Statistical Inference You have a sample and want to use the data collected

More information

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests

Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests Statistical Methods for Particle Physics Lecture 1: parameter estimation, statistical tests http://benasque.org/2018tae/cgi-bin/talks/allprint.pl TAE 2018 Benasque, Spain 3-15 Sept 2018 Glen Cowan Physics

More information

Using Probability to do Statistics.

Using Probability to do Statistics. Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Richard Lockhart Simon Fraser University STAT 830 Fall 2018 Richard Lockhart (Simon Fraser University) STAT 830 Hypothesis Testing STAT 830 Fall 2018 1 / 30 Purposes of These

More information

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm

Lecture Testing Hypotheses: The Neyman-Pearson Paradigm Math 408 - Mathematical Statistics Lecture 29-30. Testing Hypotheses: The Neyman-Pearson Paradigm April 12-15, 2013 Konstantin Zuev (USC) Math 408, Lecture 29-30 April 12-15, 2013 1 / 12 Agenda Example:

More information

Module 2 : Conditional Probability

Module 2 : Conditional Probability Module 2 : Conditional Probability Ruben Zamar Department of Statistics UBC January 16, 2017 Ruben Zamar Department of Statistics UBC Module () 2 January 16, 2017 1 / 61 MOTIVATION The outcome could be

More information

Base Rates and Bayes Theorem

Base Rates and Bayes Theorem Slides to accompany Grove s handout March 8, 2016 Table of contents 1 Diagnostic and Prognostic Inference 2 Example inferences Does this patient have Alzheimer s disease, schizophrenia, depression, etc.?

More information

Theory of Maximum Likelihood Estimation. Konstantin Kashin

Theory of Maximum Likelihood Estimation. Konstantin Kashin Gov 2001 Section 5: Theory of Maximum Likelihood Estimation Konstantin Kashin February 28, 2013 Outline Introduction Likelihood Examples of MLE Variance of MLE Asymptotic Properties What is Statistical

More information

Probability and Discrete Distributions

Probability and Discrete Distributions AMS 7L LAB #3 Fall, 2007 Objectives: Probability and Discrete Distributions 1. To explore relative frequency and the Law of Large Numbers 2. To practice the basic rules of probability 3. To work with the

More information

How do we compare the relative performance among competing models?

How do we compare the relative performance among competing models? How do we compare the relative performance among competing models? 1 Comparing Data Mining Methods Frequent problem: we want to know which of the two learning techniques is better How to reliably say Model

More information

STAT 830 Hypothesis Testing

STAT 830 Hypothesis Testing STAT 830 Hypothesis Testing Hypothesis testing is a statistical problem where you must choose, on the basis of data X, between two alternatives. We formalize this as the problem of choosing between two

More information

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish

More information

Two-Sample Inferential Statistics

Two-Sample Inferential Statistics The t Test for Two Independent Samples 1 Two-Sample Inferential Statistics In an experiment there are two or more conditions One condition is often called the control condition in which the treatment is

More information

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1

Topic -2. Probability. Larson & Farber, Elementary Statistics: Picturing the World, 3e 1 Topic -2 Probability Larson & Farber, Elementary Statistics: Picturing the World, 3e 1 Probability Experiments Experiment : An experiment is an act that can be repeated under given condition. Rolling a

More information

Bayesian Statistics. Debdeep Pati Florida State University. February 11, 2016

Bayesian Statistics. Debdeep Pati Florida State University. February 11, 2016 Bayesian Statistics Debdeep Pati Florida State University February 11, 2016 Historical Background Historical Background Historical Background Brief History of Bayesian Statistics 1764-1838: called probability

More information