A Primer on Statistical Inference using Maximum Likelihood

Size: px
Start display at page:

Download "A Primer on Statistical Inference using Maximum Likelihood"

Transcription

1 A Primer on Statistical Inference using Maximum Likelihood November 3, Inference via Maximum Likelihood Statistical inference is the process of using observed data to estimate features of the population. In terms of distributions, statistical inference is using observed data to estimate parameters of the corresponding distribution. For example, if we assume that observed data Y 1,..., Y n follow a N pµ, σ 2 q distribution, then we would need to estimate the mean µ and variance σ 2 using the data Y 1,..., Y n. There are various ways of performing inference including method of moments estimation, generalized estimating equations and Bayesian inference. However, here were are going to focus on maximum likelihood estimation which is a very common form of estimation (inference) based on maximizing probabilities. 1.1 Intuition for Maximum Likelihood Before we get to the specifics of maximum likelihood estimation, let s start off with a simple example to get the intuition behind the inference technique. Suppose we have two six sided dice in a black box that are identical in all ways except their probabilities. Specifically, the probabilities associated with each dice are as follows: Outcome Die #1 1/6 1/6 1/6 1/6 1/6 1/6 Die #2 1/3 1/6 1/6 1/3 0 0 Now, pretend that we pull out one of the die and our goal is to figure out (just by rolling it or observing random outcomes ) which dice we have. Let Y i be the observation associated with the i th roll of the die. Notice that we can think of Y i as being a random variables with a distribution given by one of the rows in the table. The trouble is we don t know which row corresponds to the distribution of Y i so we need to roll the die n times and observe Y 1,..., Y n to try and figure it out. In other words, after rolling the die n times, we are going to use Y 1,..., Y n to infer which die we have. So, lets start rolling the die and try to make inference. 1

2 On our first roll we get Y 1 3: which die do we think it is? Well, lets calculate the probability of Y 1 3 under both die scenarios: PrpY 1 3 Die #1q 1{6 PrpY 1 3 Die #2q 1{6 so, under either die, the probability is the same. So after n 1 roll we really can t infer which die we have so lets roll again. On our second roll we get Y 2 1: which die do we think it is? Again, lets calculate the probabilities under each die: PrpY 1 3, Y 2 1 Die #1q 1{6 ˆ 1{6 1{36 PrpY 1 3, Y 2 1 Die #2q 1{6 ˆ 1{3 1{18 where the multiplication comes from assuming independence of rolls (which is quite reasonable in this case). Notice, that the probability of observing Y 1 3 and Y 2 1 under Die #2 is more likely so at this point we are going to infer that we have Die #2. In other words, we are going to choose the die that maximizes the probability of our data under each die. Just for be confident in our decision of die #2 from above, we decide to roll it again and get Y 3 6. Now which die do we think we have? The answer is obvious when calculating probabilities: PrpY 1 3, Y 2 1 Die #1q 1{6 ˆ 1{6 ˆ 1{6 1{216 PrpY 1 3, Y 2 1 Die #2q 1{6 ˆ 1{3 ˆ 0 0. At this point we know for sure which die we have: we have die #1 because we could never get a 6 under die #2. In this little example, we used our observed data Y 1, Y 2, Y 3 to make inference about which die we have. Specifically, we made the inference that maximized the probability of our observed data. This is what is referred to as maximum likelihood estimation. 1.2 Univariate Maximum Likelihood Estimation Suppose we observe n data points Y 1,..., Y n. Statistical inference for the population associated with Y 1,..., Y n proceeds as follows: (i) explore the data to determine what distribution shape is appropriate for Y 1,..., Y n, (ii) after determining a shape (e.g. normal) determine which parameters of the distribution are unknown, (iii) calculate the joint probability of Y 1,..., Y n under this distribution then (iv) infer the unknown parameters via maximizing the joint probability of the observed data. To get into the details a bit, lets walk through a real example together. Suppose I am trying to find out the success rate of a flu vaccine in preventing flu. In other words, I am trying to figure out the probability of contracting the flu given the person received the vaccine. Notationally, let p represent the probability of contracting the flu if a person receives the vaccine. Since our goal is to infer p, I give the flu vaccine to 100 people and find that 7 of them still got the flu. Let Y i be the outcome for person i where Y i 1 if person i got the flu and Y i 0 if person i didn t get the 2

3 flu. From the 100 people, I observed Y 1 Y 7 1 and Y 8 Y Lets perform statistical inference for p. Step #1 is to figure out an appropriate distribution for Y i. In this case, the Bernoulli distribution is a distribution that correspond to binary (0/1) outcomes so lets us that. Under the Bernoulli distribution, PrpY i pq p Y i p1 pq 1 Y i so that PrpY i 1q p and PrpY i 0q 1 p leaving us with p as the probability of contracting the flu when a person receives the vaccine. Step #2 is to figure out the unknown parameters associated with my chosen distribution. In this case, the only unknown parameter is p itself. Step #3 is to calculate the joint distribution of Y 1,..., Y n. To do this, lets assume independence of events so that, nź PrpY 1,..., Y n pq p Y i p1 pq 1 Y i i 1 p Y 1 p Y2 p Yn ˆ p1 pq 1 Y 1 p1 pq 1 Y2 p1 pq 1 Yn p ř n p1 pq ř n i 1 p1 Y iq p ř n p1 pq n řn. Taking a look at this last form, notice that ř n is the just the number of people who got the flu (take a minute to convince yourself of this if you don t see it). That means that n řn is the number of people who didn t get the flu. Notationally let n flu be the number of people who got the flu and n n flu are the number of people who didn t. We can rewrite the last equation above as PrpY 1,..., Y n pq p n flu p1 pq n n flu (1) Step #4 is to choose p that maximizes the joint probability that we just calculated (this is equivalent to choosing the die that maximizes the joint probability of the die outcomes above). The second we go to do this, notice that we are no longer thinking of Equation (1) as a function of Y i (which we did when we calculated it). Rather, we are considering Equation (1) as a function of the parameter p. For this reason we call (1) the likelihood of p (which is the same thing as saying the probability of Y 1,..., Y n ) because we thinking of it as a function of p (not Y i ). Hence, this is where we get the name maximum likelihood estimation. To maximize the likelihood, we need to take derivatives of (1) with respect to p. This derivative would be quite ugly so let s do something simpler using a calculus trick we remember from high school (or if you don t remember then let me remind you). The trick is to maximize the logarithm of (1) rather than the original function. Standard math results show us that the maximum of the natural logarithm is the same as the maximum. The natural log of (1) is Lppq n flu lnppq ` pn n flu q lnp1 pq (2) where I am using L to denote the log-likelihood of p. This log-likelihood is substantially easier to take derivatives of so lets do it, dlppq dp n flu p ` p 1qpn n fluq 1 p 3

4 where the 1 comes from the chain rule for derivatives. To maximize, we set the derivative equation to zero and solve for p. Lets do that, dlppq dp n flu p ` p 1qpn n fluq 1 p pn n fluq 1 p ñ n flu p ñ p1 pqn flu ppn n flu q ñ n flu pn flu ppn n flu q ñ n flu pn pn flu ` pn flu ñ n flu pn ñ p n flu n so that our inferred estimate of p is p n flu {n where I use p to denote that it is our estimate of p NOT p itself (p is the parameter while p is the statistic). Coincidentally, this is the exact thing you were taught in 121 to do when calculating p. Namely, number of successes/(total number). So, if you ever wondered why we teach you that then here you go: its the maximum likelihood estimate of the probability (or proportion). Now, its your turn to try and use maximum likelihood estimation on a real dataset. The file WomensHeights.txt contains measurements from 77 womens heights in inches. Your goal is to infer about the population of womens heights using this data. Do the following: 1. Complete steps #1 and #2 by drawing a histogram to confirm that the normal shape is appropriate for this data. Under the normal distribution, the unknown parameters would be the population mean µ and variance σ Write out the joint probability of Y 1,..., Y 7 7. Call this the likelihood of µ and σ 2. In case you have forgotten, if Y i is normal then " PrpY i µ, σ 2 1 q? exp 1 * 2πσ 2 2σ py 2 i µq 2. Technically the above equation is not a probability for Y i but a density of Y i. There is no harm, however, in thinking about it as a probability. 3. In preparation for Step #4 (maximizing), calculate the log-likelihood by taking the natural logarithm of your answer in #2. 4. Find the maximum likelihood estimate for µ but maximizing your log-likelihood in #3 by taking a derivative with respect to µ, setting it equal to 0, and then solving. What is the maximum likelihood estimate for µ? Do you recognize this from 121 (hint: you should)? 5. Now, use your data from the 77 women to get what the maximum likelihood estimate of µ is for this dataset. 0 4

5 1.3 Multivariate Maximum Likelihood Estimation Turn now to the situation where we have a multivariate observation Y i py i1,..., Y ip q 1 rather than a univariate one. The technique of maximizing the likelihood with a multivariate response is no different than in the univariate case. That is, we take exactly the same steps as we did before. Calculating the joint probability, though, can catch people who aren t careful (which isn t you of course). That is, our data are Y 1,..., Y n where each Y i is a vector of P units of information. So, assuming independent, calculating the joint probability of Y 1,..., Y n is PrpY 1,..., Y n parametersq nź PrpY i parametersq where PrpY i parametersq is itself a probability for the multivariate vector Y i. As an example of multivariate maximum likelihood estimation, consider the following example. According to the theory of left-brain or right-brain dominance, each side of the brain controls different types of thinking. Additionally, people are said to prefer one type of thinking over the other. For example, a person who is left-brained is often said to lean toward mathematical and quantitative thinking while a person who is right-brained is said to be creative and excel in verbal skills. Do people tend to be only left- or right-brained? To test this out, the ACT.txt dataset contains n 117 measurements of student ACT scores on the math and verbal section. Let Y i1 denote student i s score on the math section of the ACT and Y i2 denote the same student s score on the verbal section where i 1,..., n. We can test the sided brain theory by looking at the relationship between Y i1 and Y i2 as described by their joint distribution. So, if we know the joint distribution then we can see if math people tend to NOT be verbal and vice versa. Perform inference for the joint distribution of Y i py i1, Y i2 q 1 from the ACT.txt dataset by doing the following: 1. Complete Steps #1 and #2 by checking if the shape of the multivariate normal distribution for Y i is reasonable by (a) drawing histograms (or smoothed histograms called kernel density estimates ) of Y i1 and Y i2 individually. i 1 (b) drawing a 2-D kernel density estimate of the Y i. If the joint distribution of Y i is MVN then each distribution individually should be normal and the joint distribution should look similar to the pictures you drew in the primer on random variables and their distributions. The parameters of the multivariate normal distribution are the mean vector µ and the covariance matrix Σ. 2. Write out the joint probability of Y 1,..., Y n. To complete this problem, you should know that if Y i is multivariate normal then ˆ P {2 " * 1 PrpY i µ, Σq Σ 1{2 exp 1 2π 2 py i µq 1 Σ 1 py i µq 3. In preparation for Step #4 (maximizing) calculate the log-likelihood by taking the natural logarithm of your answer in #2. 5

6 4. Find the maximum likelihood estimate for µ by maximizing your log-likelihood in #3 by taking a derivative with respect to µ, setting it equal to 0, and then solving. What is the maximum likelihood estimate for µ? 5. Use your data to get the actual maximum likelihood estimate of µ for this problem. 2 Properties of Maximum Likelihood Estimators Maximum likelihood is particularly popular for inference because of a few really cool properties associated with the maximum likelihood estimates (which I ll abbreviate to MLE for short). First, notice that the MLEs are really just functions of data and, if you got new data, you would get a different answer. Because your data is a random variable, then so is the MLE (the randomness associated with the data gets passed onto the MLE). So, we can ask, if the MLE is a random variable, what then is its distribution? The answer, as it turns out is Normal as long as we have a large sample size. This is referred to as the central limit theorem of MLEs. This basically means that we can use the normal distribution to calculate probabilities associated with the MLE. This is particularly helpful when constructing confidence intervals for the population parameters. Second, the MLE is consistent. Basically, this means that as your sample size increase then the MLE will get closer and closer to the true parameter. You would think that should be a must have property of any estimate but there are some estimates out there for which this isn t true so we ll just be grateful that the MLE is consistent. Finally, the third property of the MLE is called invariance. This just means that the MLE of any function of a parameter is that same function of the MLE. For example, suppose we are interested in logppq from the flu example above. Under invariance of the MLE, the MLE of logppq would be logppq. Again, this seems like it should be obvious but there are techniques for which this isn t true. I only mention these properties here because we are likely (but not certain) to need these as we look at complicated data sets. We ll return to them and go into more detail as needed. 6

Ozone Project. Motivating Application. Model/Distribution Specification

Ozone Project. Motivating Application. Model/Distribution Specification Ozone Project Motivating Application Ozone po 3 q is a gas that occurs naturally in the atmosphere but comes in two varieties: good and bad. Good ozone is present in the upper atmosphere and protects us

More information

Solving with Absolute Value

Solving with Absolute Value Solving with Absolute Value Who knew two little lines could cause so much trouble? Ask someone to solve the equation 3x 2 = 7 and they ll say No problem! Add just two little lines, and ask them to solve

More information

Section 5.4. Ken Ueda

Section 5.4. Ken Ueda Section 5.4 Ken Ueda Students seem to think that being graded on a curve is a positive thing. I took lasers 101 at Cornell and got a 92 on the exam. The average was a 93. I ended up with a C on the test.

More information

Conditional probabilities and graphical models

Conditional probabilities and graphical models Conditional probabilities and graphical models Thomas Mailund Bioinformatics Research Centre (BiRC), Aarhus University Probability theory allows us to describe uncertainty in the processes we model within

More information

Confidence Intervals

Confidence Intervals Quantitative Foundations Project 3 Instructor: Linwei Wang Confidence Intervals Contents 1 Introduction 3 1.1 Warning....................................... 3 1.2 Goals of Statistics..................................

More information

Confidence Intervals and Hypothesis Tests

Confidence Intervals and Hypothesis Tests Confidence Intervals and Hypothesis Tests STA 281 Fall 2011 1 Background The central limit theorem provides a very powerful tool for determining the distribution of sample means for large sample sizes.

More information

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation PRE 905: Multivariate Analysis Spring 2014 Lecture 4 Today s Class The building blocks: The basics of mathematical

More information

Fitting a Straight Line to Data

Fitting a Straight Line to Data Fitting a Straight Line to Data Thanks for your patience. Finally we ll take a shot at real data! The data set in question is baryonic Tully-Fisher data from http://astroweb.cwru.edu/sparc/btfr Lelli2016a.mrt,

More information

Calculus II. Calculus II tends to be a very difficult course for many students. There are many reasons for this.

Calculus II. Calculus II tends to be a very difficult course for many students. There are many reasons for this. Preface Here are my online notes for my Calculus II course that I teach here at Lamar University. Despite the fact that these are my class notes they should be accessible to anyone wanting to learn Calculus

More information

Problem Solving. Kurt Bryan. Here s an amusing little problem I came across one day last summer.

Problem Solving. Kurt Bryan. Here s an amusing little problem I came across one day last summer. Introduction Problem Solving Kurt Bryan Here s an amusing little problem I came across one day last summer. Problem: Find three distinct positive integers whose reciprocals add up to one. Prove that the

More information

Sometimes the domains X and Z will be the same, so this might be written:

Sometimes the domains X and Z will be the same, so this might be written: II. MULTIVARIATE CALCULUS The first lecture covered functions where a single input goes in, and a single output comes out. Most economic applications aren t so simple. In most cases, a number of variables

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman.

Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Math 224 Fall 2017 Homework 1 Drew Armstrong Problems from Probability and Statistical Inference (9th ed.) by Hogg, Tanis and Zimmerman. Section 1.1, Exercises 4,5,6,7,9,12. Solutions to Book Problems.

More information

STA Why Sampling? Module 6 The Sampling Distributions. Module Objectives

STA Why Sampling? Module 6 The Sampling Distributions. Module Objectives STA 2023 Module 6 The Sampling Distributions Module Objectives In this module, we will learn the following: 1. Define sampling error and explain the need for sampling distributions. 2. Recognize that sampling

More information

Generative Learning. INFO-4604, Applied Machine Learning University of Colorado Boulder. November 29, 2018 Prof. Michael Paul

Generative Learning. INFO-4604, Applied Machine Learning University of Colorado Boulder. November 29, 2018 Prof. Michael Paul Generative Learning INFO-4604, Applied Machine Learning University of Colorado Boulder November 29, 2018 Prof. Michael Paul Generative vs Discriminative The classification algorithms we have seen so far

More information

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017

CPSC 340: Machine Learning and Data Mining. MLE and MAP Fall 2017 CPSC 340: Machine Learning and Data Mining MLE and MAP Fall 2017 Assignment 3: Admin 1 late day to hand in tonight, 2 late days for Wednesday. Assignment 4: Due Friday of next week. Last Time: Multi-Class

More information

Math 381 Midterm Practice Problem Solutions

Math 381 Midterm Practice Problem Solutions Math 381 Midterm Practice Problem Solutions Notes: -Many of the exercises below are adapted from Operations Research: Applications and Algorithms by Winston. -I have included a list of topics covered on

More information

Chapter 26: Comparing Counts (Chi Square)

Chapter 26: Comparing Counts (Chi Square) Chapter 6: Comparing Counts (Chi Square) We ve seen that you can turn a qualitative variable into a quantitative one (by counting the number of successes and failures), but that s a compromise it forces

More information

SAMPLE CHAPTER. Avi Pfeffer. FOREWORD BY Stuart Russell MANNING

SAMPLE CHAPTER. Avi Pfeffer. FOREWORD BY Stuart Russell MANNING SAMPLE CHAPTER Avi Pfeffer FOREWORD BY Stuart Russell MANNING Practical Probabilistic Programming by Avi Pfeffer Chapter 9 Copyright 2016 Manning Publications brief contents PART 1 INTRODUCING PROBABILISTIC

More information

Sampling Distribution Models. Chapter 17

Sampling Distribution Models. Chapter 17 Sampling Distribution Models Chapter 17 Objectives: 1. Sampling Distribution Model 2. Sampling Variability (sampling error) 3. Sampling Distribution Model for a Proportion 4. Central Limit Theorem 5. Sampling

More information

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI

Module 03 Lecture 14 Inferential Statistics ANOVA and TOI Introduction of Data Analytics Prof. Nandan Sudarsanam and Prof. B Ravindran Department of Management Studies and Department of Computer Science and Engineering Indian Institute of Technology, Madras Module

More information

MIT BLOSSOMS INITIATIVE

MIT BLOSSOMS INITIATIVE MIT BLOSSOMS INITIATIVE The Broken Stick Problem Taught by Professor Richard C. Larson Mitsui Professor of Engineering Systems and of Civil and Environmental Engineering Segment 1 Hi! My name is Dick Larson

More information

review session gov 2000 gov 2000 () review session 1 / 38

review session gov 2000 gov 2000 () review session 1 / 38 review session gov 2000 gov 2000 () review session 1 / 38 Overview Random Variables and Probability Univariate Statistics Bivariate Statistics Multivariate Statistics Causal Inference gov 2000 () review

More information

Descriptive Statistics (And a little bit on rounding and significant digits)

Descriptive Statistics (And a little bit on rounding and significant digits) Descriptive Statistics (And a little bit on rounding and significant digits) Now that we know what our data look like, we d like to be able to describe it numerically. In other words, how can we represent

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

In this unit we will study exponents, mathematical operations on polynomials, and factoring.

In this unit we will study exponents, mathematical operations on polynomials, and factoring. GRADE 0 MATH CLASS NOTES UNIT E ALGEBRA In this unit we will study eponents, mathematical operations on polynomials, and factoring. Much of this will be an etension of your studies from Math 0F. This unit

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Nondeterministic finite automata

Nondeterministic finite automata Lecture 3 Nondeterministic finite automata This lecture is focused on the nondeterministic finite automata (NFA) model and its relationship to the DFA model. Nondeterminism is an important concept in the

More information

Chapter 18. Sampling Distribution Models /51

Chapter 18. Sampling Distribution Models /51 Chapter 18 Sampling Distribution Models 1 /51 Homework p432 2, 4, 6, 8, 10, 16, 17, 20, 30, 36, 41 2 /51 3 /51 Objective Students calculate values of central 4 /51 The Central Limit Theorem for Sample

More information

Algebra & Trig Review

Algebra & Trig Review Algebra & Trig Review 1 Algebra & Trig Review This review was originally written for my Calculus I class, but it should be accessible to anyone needing a review in some basic algebra and trig topics. The

More information

Introduction to Matrix Algebra and the Multivariate Normal Distribution

Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Structural Equation Modeling Lecture #2 January 18, 2012 ERSH 8750: Lecture 2 Motivation for Learning the Multivariate

More information

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math.

Regression, part II. I. What does it all mean? A) Notice that so far all we ve done is math. Regression, part II I. What does it all mean? A) Notice that so far all we ve done is math. 1) One can calculate the Least Squares Regression Line for anything, regardless of any assumptions. 2) But, if

More information

Page 1. These are all fairly simple functions in that wherever the variable appears it is by itself. What about functions like the following, ( ) ( )

Page 1. These are all fairly simple functions in that wherever the variable appears it is by itself. What about functions like the following, ( ) ( ) Chain Rule Page We ve taken a lot of derivatives over the course of the last few sections. However, if you look back they have all been functions similar to the following kinds of functions. 0 w ( ( tan

More information

CS 124 Math Review Section January 29, 2018

CS 124 Math Review Section January 29, 2018 CS 124 Math Review Section CS 124 is more math intensive than most of the introductory courses in the department. You re going to need to be able to do two things: 1. Perform some clever calculations to

More information

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests:

One sided tests. An example of a two sided alternative is what we ve been using for our two sample tests: One sided tests So far all of our tests have been two sided. While this may be a bit easier to understand, this is often not the best way to do a hypothesis test. One simple thing that we can do to get

More information

Please bring the task to your first physics lesson and hand it to the teacher.

Please bring the task to your first physics lesson and hand it to the teacher. Pre-enrolment task for 2014 entry Physics Why do I need to complete a pre-enrolment task? This bridging pack serves a number of purposes. It gives you practice in some of the important skills you will

More information

Probability Distributions: Continuous

Probability Distributions: Continuous Probability Distributions: Continuous INFO-2301: Quantitative Reasoning 2 Michael Paul and Jordan Boyd-Graber FEBRUARY 28, 2017 INFO-2301: Quantitative Reasoning 2 Paul and Boyd-Graber Probability Distributions:

More information

We're in interested in Pr{three sixes when throwing a single dice 8 times}. => Y has a binomial distribution, or in official notation, Y ~ BIN(n,p).

We're in interested in Pr{three sixes when throwing a single dice 8 times}. => Y has a binomial distribution, or in official notation, Y ~ BIN(n,p). Sampling distributions and estimation. 1) A brief review of distributions: We're in interested in Pr{three sixes when throwing a single dice 8 times}. => Y has a binomial distribution, or in official notation,

More information

Chapter 1 Review of Equations and Inequalities

Chapter 1 Review of Equations and Inequalities Chapter 1 Review of Equations and Inequalities Part I Review of Basic Equations Recall that an equation is an expression with an equal sign in the middle. Also recall that, if a question asks you to solve

More information

Fifth Grade Science End-Of-Grade Test Preparation. Test-Taking Strategies per NCDPI Released Form E ( )

Fifth Grade Science End-Of-Grade Test Preparation. Test-Taking Strategies per NCDPI Released Form E ( ) Fifth Grade Science End-Of-Grade Test Preparation Test-Taking Strategies per NCDPI Released Form E (2008-2009) Note to Teacher: Use the following test-taking strategies to prepare for the fifth grade End-Of-Grade

More information

Take the Anxiety Out of Word Problems

Take the Anxiety Out of Word Problems Take the Anxiety Out of Word Problems I find that students fear any problem that has words in it. This does not have to be the case. In this chapter, we will practice a strategy for approaching word problems

More information

Algebra Exam. Solutions and Grading Guide

Algebra Exam. Solutions and Grading Guide Algebra Exam Solutions and Grading Guide You should use this grading guide to carefully grade your own exam, trying to be as objective as possible about what score the TAs would give your responses. Full

More information

What s the Average? Stu Schwartz

What s the Average? Stu Schwartz What s the Average? by Over the years, I taught both AP calculus and AP statistics. While many of the same students took both courses, rarely was there a problem with students confusing ideas from both

More information

Introduction. So, why did I even bother to write this?

Introduction. So, why did I even bother to write this? Introduction This review was originally written for my Calculus I class, but it should be accessible to anyone needing a review in some basic algebra and trig topics. The review contains the occasional

More information

Physics 6A Lab Experiment 6

Physics 6A Lab Experiment 6 Biceps Muscle Model Physics 6A Lab Experiment 6 Introduction This lab will begin with some warm-up exercises to familiarize yourself with the theory, as well as the experimental setup. Then you ll move

More information

Math Review Sheet, Fall 2008

Math Review Sheet, Fall 2008 1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the

More information

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2

Final Review Sheet. B = (1, 1 + 3x, 1 + x 2 ) then 2 + 3x + 6x 2 Final Review Sheet The final will cover Sections Chapters 1,2,3 and 4, as well as sections 5.1-5.4, 6.1-6.2 and 7.1-7.3 from chapters 5,6 and 7. This is essentially all material covered this term. Watch

More information

STAT/SOC/CSSS 221 Statistical Concepts and Methods for the Social Sciences. Random Variables

STAT/SOC/CSSS 221 Statistical Concepts and Methods for the Social Sciences. Random Variables STAT/SOC/CSSS 221 Statistical Concepts and Methods for the Social Sciences Random Variables Christopher Adolph Department of Political Science and Center for Statistics and the Social Sciences University

More information

Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2:

Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2: Math101, Sections 2 and 3, Spring 2008 Review Sheet for Exam #2: 03 17 08 3 All about lines 3.1 The Rectangular Coordinate System Know how to plot points in the rectangular coordinate system. Know the

More information

Generalized Linear Models for Non-Normal Data

Generalized Linear Models for Non-Normal Data Generalized Linear Models for Non-Normal Data Today s Class: 3 parts of a generalized model Models for binary outcomes Complications for generalized multivariate or multilevel models SPLH 861: Lecture

More information

1 Probabilities. 1.1 Basics 1 PROBABILITIES

1 Probabilities. 1.1 Basics 1 PROBABILITIES 1 PROBABILITIES 1 Probabilities Probability is a tricky word usually meaning the likelyhood of something occuring or how frequent something is. Obviously, if something happens frequently, then its probability

More information

CSC2515 Assignment #2

CSC2515 Assignment #2 CSC2515 Assignment #2 Due: Nov.4, 2pm at the START of class Worth: 18% Late assignments not accepted. 1 Pseudo-Bayesian Linear Regression (3%) In this question you will dabble in Bayesian statistics and

More information

CPSC 340: Machine Learning and Data Mining

CPSC 340: Machine Learning and Data Mining CPSC 340: Machine Learning and Data Mining MLE and MAP Original version of these slides by Mark Schmidt, with modifications by Mike Gelbart. 1 Admin Assignment 4: Due tonight. Assignment 5: Will be released

More information

Chapter 11. Regression with a Binary Dependent Variable

Chapter 11. Regression with a Binary Dependent Variable Chapter 11 Regression with a Binary Dependent Variable 2 Regression with a Binary Dependent Variable (SW Chapter 11) So far the dependent variable (Y) has been continuous: district-wide average test score

More information

Unit 4 Patterns and Algebra

Unit 4 Patterns and Algebra Unit 4 Patterns and Algebra In this unit, students will solve equations with integer coefficients using a variety of methods, and apply their reasoning skills to find mistakes in solutions of these equations.

More information

Probability and Inference. POLI 205 Doing Research in Politics. Populations and Samples. Probability. Fall 2015

Probability and Inference. POLI 205 Doing Research in Politics. Populations and Samples. Probability. Fall 2015 Fall 2015 Population versus Sample Population: data for every possible relevant case Sample: a subset of cases that is drawn from an underlying population Inference Parameters and Statistics A parameter

More information

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices.

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices. Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices. 1. What is the difference between a deterministic model and a probabilistic model? (Two or three sentences only). 2. What is the

More information

Introduction to Algebra: The First Week

Introduction to Algebra: The First Week Introduction to Algebra: The First Week Background: According to the thermostat on the wall, the temperature in the classroom right now is 72 degrees Fahrenheit. I want to write to my friend in Europe,

More information

The Conditions are Right

The Conditions are Right The Conditions are Right Standards Addressed in this Task MCC9-12.S.CP.2 Understand that two events A and B are independent if the probability of A and B occurring together is the product of their probabilities,

More information

Notes 6: Multivariate regression ECO 231W - Undergraduate Econometrics

Notes 6: Multivariate regression ECO 231W - Undergraduate Econometrics Notes 6: Multivariate regression ECO 231W - Undergraduate Econometrics Prof. Carolina Caetano 1 Notation and language Recall the notation that we discussed in the previous classes. We call the outcome

More information

Chapter 18. Sampling Distribution Models. Copyright 2010, 2007, 2004 Pearson Education, Inc.

Chapter 18. Sampling Distribution Models. Copyright 2010, 2007, 2004 Pearson Education, Inc. Chapter 18 Sampling Distribution Models Copyright 2010, 2007, 2004 Pearson Education, Inc. Normal Model When we talk about one data value and the Normal model we used the notation: N(μ, σ) Copyright 2010,

More information

DIFFERENTIAL EQUATIONS

DIFFERENTIAL EQUATIONS DIFFERENTIAL EQUATIONS Basic Concepts Paul Dawkins Table of Contents Preface... Basic Concepts... 1 Introduction... 1 Definitions... Direction Fields... 8 Final Thoughts...19 007 Paul Dawkins i http://tutorial.math.lamar.edu/terms.aspx

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Why should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4.

Why should you care?? Intellectual curiosity. Gambling. Mathematically the same as the ESP decision problem we discussed in Week 4. I. Probability basics (Sections 4.1 and 4.2) Flip a fair (probability of HEADS is 1/2) coin ten times. What is the probability of getting exactly 5 HEADS? What is the probability of getting exactly 10

More information

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas

CS839: Probabilistic Graphical Models. Lecture 7: Learning Fully Observed BNs. Theo Rekatsinas CS839: Probabilistic Graphical Models Lecture 7: Learning Fully Observed BNs Theo Rekatsinas 1 Exponential family: a basic building block For a numeric random variable X p(x ) =h(x)exp T T (x) A( ) = 1

More information

Chapter 27 Summary Inferences for Regression

Chapter 27 Summary Inferences for Regression Chapter 7 Summary Inferences for Regression What have we learned? We have now applied inference to regression models. Like in all inference situations, there are conditions that we must check. We can test

More information

PHY 101L - Experiments in Mechanics

PHY 101L - Experiments in Mechanics PHY 101L - Experiments in Mechanics introduction to error analysis What is Error? In everyday usage, the word error usually refers to a mistake of some kind. However, within the laboratory, error takes

More information

P (A) = P (B) = P (C) = P (D) =

P (A) = P (B) = P (C) = P (D) = STAT 145 CHAPTER 12 - PROBABILITY - STUDENT VERSION The probability of a random event, is the proportion of times the event will occur in a large number of repititions. For example, when flipping a coin,

More information

Steve Smith Tuition: Maths Notes

Steve Smith Tuition: Maths Notes Maths Notes : Discrete Random Variables Version. Steve Smith Tuition: Maths Notes e iπ + = 0 a + b = c z n+ = z n + c V E + F = Discrete Random Variables Contents Intro The Distribution of Probabilities

More information

Lesson 11-1: Parabolas

Lesson 11-1: Parabolas Lesson -: Parabolas The first conic curve we will work with is the parabola. You may think that you ve never really used or encountered a parabola before. Think about it how many times have you been going

More information

A Flag of Many Faces by Kelly Hashway

A Flag of Many Faces by Kelly Hashway Name: A Flag of Many Faces by Kelly Hashway Phoebe drummed her pencil on the desk as she stared at the blank paper in front of her. The assignment was easy enough. All she had to do was draw the flag of

More information

Inferring information about models from samples

Inferring information about models from samples Contents Inferring information about models from samples. Drawing Samples from a Probability Distribution............. Simple Samples from Matlab.................. 3.. Rejection Sampling........................

More information

We're in interested in Pr{three sixes when throwing a single dice 8 times}. => Y has a binomial distribution, or in official notation, Y ~ BIN(n,p).

We're in interested in Pr{three sixes when throwing a single dice 8 times}. => Y has a binomial distribution, or in official notation, Y ~ BIN(n,p). Sampling distributions and estimation. 1) A brief review of distributions: We're in interested in Pr{three sixes when throwing a single dice 8 times}. => Y has a binomial distribution, or in official notation,

More information

of 8 28/11/ :25

of 8 28/11/ :25 Paul's Online Math Notes Home Content Chapter/Section Downloads Misc Links Site Help Contact Me Differential Equations (Notes) / First Order DE`s / Modeling with First Order DE's [Notes] Differential Equations

More information

#29: Logarithm review May 16, 2009

#29: Logarithm review May 16, 2009 #29: Logarithm review May 16, 2009 This week we re going to spend some time reviewing. I say re- view since you ve probably seen them before in theory, but if my experience is any guide, it s quite likely

More information

Grades 7 & 8, Math Circles 10/11/12 October, Series & Polygonal Numbers

Grades 7 & 8, Math Circles 10/11/12 October, Series & Polygonal Numbers Faculty of Mathematics Waterloo, Ontario N2L G Centre for Education in Mathematics and Computing Introduction Grades 7 & 8, Math Circles 0//2 October, 207 Series & Polygonal Numbers Mathematicians are

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

Intersecting Two Lines, Part Two

Intersecting Two Lines, Part Two Module 1.5 Page 149 of 1390. Module 1.5: Intersecting Two Lines, Part Two In this module you will learn about two very common algebraic methods for intersecting two lines: the Substitution Method and the

More information

Confidence intervals

Confidence intervals Confidence intervals We now want to take what we ve learned about sampling distributions and standard errors and construct confidence intervals. What are confidence intervals? Simply an interval for which

More information

Proof Techniques (Review of Math 271)

Proof Techniques (Review of Math 271) Chapter 2 Proof Techniques (Review of Math 271) 2.1 Overview This chapter reviews proof techniques that were probably introduced in Math 271 and that may also have been used in a different way in Phil

More information

Bayesian Linear Regression [DRAFT - In Progress]

Bayesian Linear Regression [DRAFT - In Progress] Bayesian Linear Regression [DRAFT - In Progress] David S. Rosenberg Abstract Here we develop some basics of Bayesian linear regression. Most of the calculations for this document come from the basic theory

More information

Introduction to Maximum Likelihood Estimation

Introduction to Maximum Likelihood Estimation Introduction to Maximum Likelihood Estimation Eric Zivot July 26, 2012 The Likelihood Function Let 1 be an iid sample with pdf ( ; ) where is a ( 1) vector of parameters that characterize ( ; ) Example:

More information

Module 6: Methods of Point Estimation Statistics (OA3102)

Module 6: Methods of Point Estimation Statistics (OA3102) Module 6: Methods of Point Estimation Statistics (OA3102) Professor Ron Fricker Naval Postgraduate School Monterey, California Reading assignment: WM&S chapter 9.6-9.7 Revision: 1-12 1 Goals for this Module

More information

PMR Learning as Inference

PMR Learning as Inference Outline PMR Learning as Inference Probabilistic Modelling and Reasoning Amos Storkey Modelling 2 The Exponential Family 3 Bayesian Sets School of Informatics, University of Edinburgh Amos Storkey PMR Learning

More information

The General Linear Model. How we re approaching the GLM. What you ll get out of this 8/11/16

The General Linear Model. How we re approaching the GLM. What you ll get out of this 8/11/16 8// The General Linear Model Monday, Lecture Jeanette Mumford University of Wisconsin - Madison How we re approaching the GLM Regression for behavioral data Without using matrices Understand least squares

More information

Markov Chain Monte Carlo

Markov Chain Monte Carlo Department of Statistics The University of Auckland https://www.stat.auckland.ac.nz/~brewer/ Emphasis I will try to emphasise the underlying ideas of the methods. I will not be teaching specific software

More information

Open book, but no loose leaf notes and no electronic devices. Points (out of 200) are in parentheses. Put all answers on the paper provided to you.

Open book, but no loose leaf notes and no electronic devices. Points (out of 200) are in parentheses. Put all answers on the paper provided to you. ISQS 5347 Final Exam Spring 2017 Open book, but no loose leaf notes and no electronic devices. Points (out of 200) are in parentheses. Put all answers on the paper provided to you. 1. Recall the commute

More information

2. l = 7 ft w = 4 ft h = 2.8 ft V = Find the Area of a trapezoid when the bases and height are given. Formula is A = B = 21 b = 11 h = 3 A=

2. l = 7 ft w = 4 ft h = 2.8 ft V = Find the Area of a trapezoid when the bases and height are given. Formula is A = B = 21 b = 11 h = 3 A= 95 Section.1 Exercises Part A Find the Volume of a rectangular solid when the width, height and length are given. Formula is V=lwh 1. l = 4 in w = 2.5 in h = in V = 2. l = 7 ft w = 4 ft h = 2.8 ft V =.

More information

4.5 Linearization Calculus 4.5 LINEARIZATION. Notecards from Section 4.5: Linearization; Differentials. Linearization

4.5 Linearization Calculus 4.5 LINEARIZATION. Notecards from Section 4.5: Linearization; Differentials. Linearization 4.5 Linearization Calculus 4.5 LINEARIZATION Notecards from Section 4.5: Linearization; Differentials Linearization The goal of linearization is to approximate a curve with a line. Why? Because it s easier

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

Physics 6A Lab Experiment 6

Physics 6A Lab Experiment 6 Rewritten Biceps Lab Introduction This lab will be different from the others you ve done so far. First, we ll have some warmup exercises to familiarize yourself with some of the theory, as well as the

More information

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations

Discrete Mathematics and Probability Theory Fall 2014 Anant Sahai Note 15. Random Variables: Distributions, Independence, and Expectations EECS 70 Discrete Mathematics and Probability Theory Fall 204 Anant Sahai Note 5 Random Variables: Distributions, Independence, and Expectations In the last note, we saw how useful it is to have a way of

More information

base 2 4 The EXPONENT tells you how many times to write the base as a factor. Evaluate the following expressions in standard notation.

base 2 4 The EXPONENT tells you how many times to write the base as a factor. Evaluate the following expressions in standard notation. EXPONENTIALS Exponential is a number written with an exponent. The rules for exponents make computing with very large or very small numbers easier. Students will come across exponentials in geometric sequences

More information

Sums of Squares (FNS 195-S) Fall 2014

Sums of Squares (FNS 195-S) Fall 2014 Sums of Squares (FNS 195-S) Fall 014 Record of What We Did Drew Armstrong Vectors When we tried to apply Cartesian coordinates in 3 dimensions we ran into some difficulty tryiing to describe lines and

More information

Using Probability to do Statistics.

Using Probability to do Statistics. Al Nosedal. University of Toronto. November 5, 2015 Milk and honey and hemoglobin Animal experiments suggested that honey in a diet might raise hemoglobin level. A researcher designed a study involving

More information

Understanding Exponents Eric Rasmusen September 18, 2018

Understanding Exponents Eric Rasmusen September 18, 2018 Understanding Exponents Eric Rasmusen September 18, 2018 These notes are rather long, but mathematics often has the perverse feature that if someone writes a long explanation, the reader can read it much

More information

The Bayes classifier

The Bayes classifier The Bayes classifier Consider where is a random vector in is a random variable (depending on ) Let be a classifier with probability of error/risk given by The Bayes classifier (denoted ) is the optimal

More information

Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22

Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22 CS 70 Discrete Mathematics for CS Spring 2006 Vazirani Lecture 22 Random Variables and Expectation Question: The homeworks of 20 students are collected in, randomly shuffled and returned to the students.

More information

Student Activity: Finding Factors and Prime Factors

Student Activity: Finding Factors and Prime Factors When you have completed this activity, go to Status Check. Pre-Algebra A Unit 2 Student Activity: Finding Factors and Prime Factors Name Date Objective In this activity, you will find the factors and the

More information