Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Size: px
Start display at page:

Download "Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation"

Transcription

1 Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture 3: February 10, 2016

2 Today s Class The building blocks: The basics of mathematical statistics: Ø Random variables: definitions and types Ø Univariate distributions w General terminology w Univariate normal (aka, Gaussian) w Other (continuous) univariate distributions Ø Expected values: means, variances, and the algebra of expectations Ø Linear combinations of random variables The finished product: How the GLM fits within statistics Ø The GLM with the normal distribution Ø The statistical assumptions of the GLM Estimation of GLMs with Least Squares

3 RANDOM VARIABLES AND STATISTICAL DISTRIBUTIONS

4 Random Variables Random: situations in which the certainty of the outcome is unknown and is at least in part due to chance Variable: a value that may change given the scope of a given problem or set of operations = Random Variable: a variable whose outcome depends on chance + (possible values might represent the possible outcomes of a yet-to-beperformed experiment) Today we will denote a random variable with a lower-cased: x

5 Key Features of Random Variables Random variables each are described by a probability density/mass function (PDF) f(x) that indicates relative frequency of occurrence Ø A PDF is a mathematical function that gives a rough picture of the distribution from which a random variable is drawn The type of random variable dictates the name and nature of these functions: Ø Continuous random variables: w f(x) is called a probability density function w Area under curve must equal 1 (found by calculus integration) w Height of curve (the function value f x ): Can be any positive number Reflects relative likelihood of an observation occurring Ø Discrete random variables: w f(x) is called a probability mass function w Sum across all values must equal 1 w The function value f(x) is a probability (so must range from 0 to 1)

6 Uses of Distributions in Data Analysis Statistical models make distributional assumptions on various parameters and/or parts of data These assumptions govern: Ø How models are estimated Ø How inferences are made Ø How missing data may be imputed If data do not follow an assumed distribution, inferences may be inaccurate Ø Sometimes a very big problem, other times not so much Therefore, it can be helpful to check distributional assumptions prior to (or while) running statistical analyses

7 CONTINUOUS UNIVARIATE DISTRIBUTIONS

8 Continuous Univariate Distributions To demonstrate how continuous distributions work and look, we will discuss two: Ø Uniform distribution Ø Normal distribution Each are described a set of parameters, which we will later see are what give us our inferences when we analyze data What we then do is put constraints on those parameters based on hypothesized effects in data

9 Uniform Distribution The uniform distribution is shown to help set up how continuous distributions work For a continuous random variable x that ranges from a, b, the uniform probability density function is: f x = 1 b a The uniform distribution has two parameters: Ø a the lower limit Ø b the upper limit x U a, b In R: the dunif() function gives you the height: dunif(x, min = 0, max = 1, log = FALSE)

10 More on the Uniform Distribution To demonstrate how PDFs work, we will try a few values: x a b f(x) = = = =.1 The uniform PDF has the feature that all values of x are equally likely across the sample space of the distribution Ø Therefore, you do not see x in the PDF f x The mean of the uniform distribution is. (a + b) / The variance of the uniform distribution is. / b a./

11 Univariate Normal Distribution For a continuous random variable x (ranging from to ) the univariate normal distribution function is: 1 f x = exp x μ / < 2πσ 2σ The shape of the distribution is governed by two parameters: Ø The mean μ < Ø The variance σ Ø These parameters are called sufficient statistics (they contain all the information about the distribution) The skewness (lean) and kurtosis (peakedness) are fixed Standard notation for normal distributions is x N(μ <, σ ) Ø Read as: x follows a normal distribution with a mean μ < and a variance σ Linear combinations of random variables following normal distributions result in a random variable that is normally distributed Ø You ll see this later with respect to the GLM

12 Univariate Normal Distribution f(x) f x gives the height of the curve (relative frequency) for any value of x, μ <, and σ

13 More of the Univariate Normal Distribution To demonstrate how the normal distribution works, we will try a few values: x μ x σ2 x f(x) The values from f x were obtained by using Excel Ø Ø The =normdist() function Most statistics packages have a normal distribution function The mean of the normal distribution is μ < The variance of the normal distribution is σ In R the dnorm() function provides the height for a normal distribution: dnorm(x, mean = 0, sd = 1, log = FALSE)

14 EXPECTED VALUES AND THE ALGEBRA OF EXPECTATIONS

15 Expected Values Expected values are statistics taken the sample space of a random variable: they are essentially weighted averages The weights used in computing this average correspond to the densities for a continuous random variable (the heights of the distribution) Notation: the expected value is represented by: E(x) Ø The actual statistic that is being weighted by the PDF is put into the parentheses where x is now Expected values allow us to understand what a statistical model implies about data, for instance: Ø How a GLM specifies the (conditional) mean and variance of a DV

16 Expected Value Calculation For discrete random variables, the expected value is found by: E x = F xp(x = x) < For example, the expected value of a roll of a die is: E x = = 3.5 For continuous random variables, the expected value is found by: E x = L xf x dx < We won t be calculating theoretical expected values with calculus we use them only to see how models imply things about our data

17 Variance and Covariance As Expected Values A distribution s theoretical variance can also be written as an expected value: V x = E x E x / = E x μ < / Ø This formula will help us understand predictions made GLMs and how that corresponds to statistical parameters we interpret For a roll of a die, the theoretical variance is: V x = E x 3.5 / = Ȯ / + Ȯ / + Ȯ ( 3 3.5) / + Ȯ / + Ȯ / + Ȯ / = 2.92 Ø Likewise, the SD is then 2.92 = 1.71 Likewise, for a pair of random variables x and z, the covariance can be found from their joint distributions: Cov x, z = E xz E x E z = E xz μ < μ V

18 LINEAR COMBINATIONS OF RANDOM VARIABLES

19 Linear Combinations of Random Variables A linear combination is an expression constructed from a set of terms by multiplying each term by a constant and then adding the results x = a. v. + a / v / + + a X v X Ø The linear regression equation is a linear combination More generally, linear combinations of random variables have specific implications for the mean, variance, and possibly covariance of the new random variable As such, there are predicable ways in which the means, variances, and covariances change Ø These terms are called the algebra of expectations To guide us through this process, we will use the descriptive statistics from the height/weight/gender example from our 1 st class

20 Descriptive Statistics for Height/Weight Data Variable Mean SD Variance Height Weight , Female Diagonal: Variance Above Diagonal: Covariance Correlation /Covariance Height Weight Female Height Weight.798 3, Female Below Diagonal: Correlation

21 Algebra of Expectations Here are some properties of expected values: x and z are random variables, c and d constants Sums of Constants: E x + c = E x + c V x + c = V x Cov x + c, z = Cov(x, z) Products of Constants: E cx = ce x V cx = c / V x Cov cx, dz = cdcov(x, z) Sums of Random Variables: E cx + dz = ce x + de z V cx + dz = c / V x + d / V z + 2cd Cov x, z

22 Examples for Algebra of Expectations Image you wanted to convert weight from pounds to kilograms (where 1 pound = kg) Weight`a =.453Weight bc The mean (expected value) of weight in kg: E Weight`a = E.453Weight bc =.453E Weight bc =.453Weight bc = = 83.08kg The variance of weight in kg: V Weight`a = V.453Weight bc =.453 / V Weight bc =.453 / 3, = kg / The covariance of weight in kg with height in inches: Cov Weight`a, Height = Cov.453Weight bc, Height =.453Cov Weight bc, Height = = kg inches

23 Don t Take My Word For It R syntax for transforming weight in a DATA step: R syntax and output for marginal descriptive statistics and covariances:

24 Where We Use This The Dreaded ESTIMATE Statement The glht statement in R computes the expected value and standard error (square root of variance) for a new random variable Ø The new random variable is a linear combination of the original model parameters (the fixed effects) Ø The original model parameters are considered random here as their sampling distribution is used (assuming normal errors and a large N) Where: Estimate = 1 β r<srturxvrw + 1 β x/ r<srturxvrw Ø β r<srturxvrw has mean β r<srturxvrw y Ø β x/ r<srturxvrw has mean β x/ r<srturxvrw y Ø There exists a covariance between β r<srturxvrw y w We ll call this Cov β r<srturxvrw y, β x/ r<srturxvrw y and variance se β r<srturxvrw y / and variance se β x/ r<srturxvrw y / and β x/ r<srturxvrw y

25 THE GENERAL LINEAR MODEL WITH WHAT WE HAVE LEARNED TODAY

26 The General Linear Model, Revisited The general linear model for predicting Y from X and Z: Y s = β { + β. X s + β / Z s + β } X s Z s + e s In terms of random variables, under the GLM: e s is considered random: e s N 0,σ r / Y s is dependent on the linear combination of X s, Z s, and e s The GLM provides a model for the conditional distribution of the dependent variable, where the conditioning variables are the independent variables: f(y s X s, Z s ) Ø There are no assumptions made about X s and Z s - they are constants Ø The regression slopes β {, β., β /, β } are constants that are said to be fixed at their values (hence, called fixed effects)

27 Combining the GLM with Expectations Using the algebra of expectations predicting Y from X and Z: The expected value (mean) of f Y p X p, Z p : Y s = E Y s = E β { + β. X s + β / Z s + β } X s Z s + e s Constants Random Variable with E e s = 0 = β { + β. X s + β / Z s + β } X s Z s + E e s = β { + β. X s + β / Z s + β } X s Z s The variance of f Y p X p, Z p : V Y s = V β { + β. X s + β / Z s + β } X s Z s + e s = V e s = σ r /

28 Distribution of f Y s X s, Z s We just found the mean (expected value) and variance implied by the GLM for the conditional distribution of Y s given X s and Z s The next question: what is the distribution of f Y s X s, Z s? Linear combinations of random variables that are normally distributed result in variables that are normally distributed Because e s N 0, σ r / is the only random term in the GLM, the resulting conditional distribution of Y s is normally distributed: Y s X s, Z s N β { + β. X s + β / Z s + β } X s Z s, σ r / Model for the means: from fixed effects; literally gives mean of f Y s X s, Z s Model for the variances: from random effects; gives variance of f Y s X s, Z s

29 Examining What This Means in the Context of Data If you recall from our first lecture, the final model we decided to interpret: Model 5 W s = β { + β. H s H + β / F s + β } H s H F s + e s where e s N 0, σ r / From R:

30 Picturing the GLM with Distributions The distributional assumptions of the GLM are the reason why we do not need to worry if our dependent variable is normally distributed Our dependent variable should be conditionally normal We can check this assumption by checking our assumption about the residuals, e s N 0,σ r / More on this soon

31 More Pictures of the GLM Treating our estimated values of the slopes β {, β., β /, β } and the residual variance σ / r as the true values* we can now see what the theoretical* distribution of f Weight s Height s, Female s looks like for a given set of predictors *Note: these distributions change when sample estimates are used (think standard error of the prediction)

32 Behind the Pictures To emphasize the point that PDFs provide the height of the line, here is the normal PDF (with numbers) that produced those plots: f W s H s, F s = 1 2πσ r / exp = = 1 2πσ r / exp 1 2π 4.73 exp Model for the Variance W s Wˆs 2σ r / W s β { + β. H s H + β / F s + β } H s H F s / 2σ r / W s H s H 82.27F s 1.09 H s H F s The plots were created using the following value for the predictors: H = 67.9 Left plot: H s = 62; F s = 1 Right plot: H s = 76;F s = 0 Model for the Means / /

33 ESTIMATION OF GLMS USING LEAST SQUARES

34 Why Estimation is Important In applied statistics courses, estimation is not discussed very frequently Ø Can be very technical very intimidating Estimation is of critical importance Ø Quality and validity of estimates (and of inferences made from them) depends on how they were obtained Ø New estimation methods appear from time to time and get widespread use without anyone asking whether or not they are any good Consider an absurd example: Ø I say the mean for IQ should be 20 just from what I feel Ø Do you believe me? Do you feel like reporting this result? w Estimators need a basis in reality (in statistical theory)

35 How Estimation Works (More or Less) Most estimation routines do one of three things: 1. Minimize Something: Typically found with names that have least in the title. Forms of least squares include Generalized, Ordinary, Weighted, Diagonally Weighted, WLSMV, and Iteratively Reweighted. Typically the estimator of last resort 2. Maximize Something: Typically found with names that have maximum in the title. Forms include Maximum likelihood, ML, Residual Maximum Likelihood (REML), Robust ML. Typically the gold standard of estimators (and next week we ll see why).

36 Estimation of General Linear Models Recall our GLM (shown here for the prediction of a dependent variable Y s by two independent variables X s and Z s ): Y s = β { + β. X s + β / Z s + β } X s Z s + e s Traditionally (dating to circa 1840), general linear models can be estimated via a process called least squares Least squares attempts to find the GLM parameters (the βs) that minimize the squared residual terms: min {β {, β., β /, β } } F e s / sž.

37 Where We Are Going (and Why We Are Going There) Because the basics of estimation are critical to understanding the validity of the numbers you will use to make inferences from, we will detail the process of estimation Ø Today with Least Squares and then ending with Maximum Likelihood The LS estimation we will discuss is to get you to visualize functions of statistical parameters (the βs here) and data in order to show which estimates we would choose Ø To be repeated: In practice LS estimation for GLMs does not do this (by the magic of calculus and algebra) In the end, we would like for you to understand that not all estimators are created equally and that some can be trusted more than others

38 How Least Squares Estimation Works How Least Squares works is through minimizing the squared error terms but its what goes into error that drives the process: e s = Y s Y s = Y s β { + β. X s + β / Z s + β } X s Z s If you were to do this (and you wouldn t), the process called optimization would go like this: 1. Pick values for regression slopes 2. Calculate Y s and then e s for each person p 3. Calculate OF = e/ sž. s function) (letters OF stand for objective 4. Repeat 1-3 until you find the values of regression slopes

39 Today s Example Data Imagine an employer is looking to hire employees for a job where IQ is important Ø We will only use 5 observations so as to show the math behind the estimation calculations The employer collects two variables: Ø IQ scores Variable Mean SD Ø Job performance IQ Descriptive Statistics: Performance Observation IQ Performance Covariance Matrix IQ Performance

40 Visualizing the Data

41 Let s Play Pick the Parameters This slide is left as a placeholder for the recording we will now do a demonstration in R

42 And The Winner Is

43 Examining the Objective Function Surface

44 LS Estimates of GLMs The process of least squares estimation of GLMs does not need an iterative search Using calculus, a minimum of the objective function can be found Ø This involves taking the first derivative of the objective function with respect to each parameter w Derivative = slope of the tangent line for a given point Ø The first derivative is then set equal to zero w Flat slope = minimum (or maximum or saddle point neither apply here) Ø The equation is then solved for the parameter w Producing the equations you know and love For simple linear regression (one predictor): Ø β = œ š š œ = v žÿtuÿxvr ŸX š v žÿtuÿxvr ŸX = vt st v Ÿtr When we get to matrix algebra, you will know this as β = X X. X y

45 Ramifications of Closed-Form LS Equations Because least squares estimates have a closed form (equations that will provide statistics directly), they will work nearly every time Ø Only fail when collinearity is present (soon you ll know this to mean X X is singular and cannot be inverted) Virtually all other estimators you will encounter in statistics will not have a closed form Ø Even least squares estimators for other types of data (not continuous) Without a closed form, least squares estimates are found by search the objective function for its minimum Ø Like finding the drain of a pool.

46 Why LS is Still in Use Least squares estimates still make up the bulk of GLM analyses because: Ø They are easy to compute Ø They pretty much always give you an answer Ø They have been shown to have good statistical properties The good statistical properties actually come because LS estimates of GLMs match the Maximum Likelihood Estimates Ø We will learn more about maximum likelihood estimation next Ø For now, know that MLEs are the gold standard when it comes to estimates

47 Where LS Fails For all their flexibility, least squares estimates are somewhat limited Ø Only have good properties for basic univariate GLM for continuous data w Normally distributed error terms with homogeneous variance When data are not continuous/do not have normally distributed error terms, least squares estimates are not preferred For multivariate models with continuous data (repeated measures, longitudinal data, scales of any sort), least squares estimates quickly do not work Ø Cannot handle missing outcomes (deletes entire case) Ø Limited in the types of ways of modeling covariance between observations

48 AN INTRODUCTION TO GENERALIZED MODELS

49 A World View of Models Statistical models can be broadly organized as: Ø General (normal outcome) vs. Generalized (not normal outcome) Ø One dimension of sampling (one variance term per outcome) vs. multiple dimensions of sampling (multiple variance terms) w Fixed effects only vs. mixed (fixed and random effects = multilevel) All models have fixed effects, and then: Ø General Linear Models: conditionally normal distribution for data, fixed effects, no random effects Ø General Linear Mixed Models: conditionally normal distribution for data, fixed and random effects Ø Generalized Linear Models: any conditional distribution for data, fixed effects through link functions, no random effects Ø Generalized Linear Mixed Models: any conditional distribution for data, fixed and random effects through link functions Linear means the fixed effects predict the link-transformed DV in a linear combination of (effect*predictor) + (effect*predictor)

50 Observed outcomes: what you collect and evaluate based Unpacking the Big Picture Observed Outcomes (any format) Hypothesized Causal Process Model: Substantive Theory Substantive theory: what guides your study Hypothetical causal process: what the statistical model is testing (attempting to falsify) when estimated

51 The Goal of Generalized Models Generalized models map the substantive theory onto the sample space of the observed outcomes Ø Sample space = type/range/outcomes that are possible The general idea is that the statistical model will not approximate the outcome well if the assumed distribution is not a good fit to the sample space of the outcome Ø If model does not fit the outcome, the findings cannot be believed The key to making all of this work is the use of differing statistical distributions for the outcome Generalized models allow for different distributions for outcomes Ø The mean of the distribution is still modeled by the model for the means

52 What kind of outcome? Generalized vs. General Generalized Linear Models à General Linear Models whose residuals follow some not-normal distribution and in which a linktransformed Y is predicted instead of Y Many kinds of non-normally distributed outcomes have some kind of generalized linear model to go with them: Ø Binary (dichotomous) Ø Unordered categorical (nominal) Ø Ordered categorical (ordinal) Ø Counts (discrete, positive values) These two are often called multinomial inconsistently Ø Censored (piled up and cut off at one end left or right) Ø Zero-inflated (pile of 0 s, then some distribution after) Ø Continuous but skewed data (pile on one end, long tail)

53 Some Links/Distributions (from Wikipedia)

54 3 Parts of a Generalized Linear Model Link Function (main difference from GLM): Ø How a non-normal outcome gets transformed into something we can predict that is more continuous (unbounded) Ø For outcomes that are already normal, general linear models are just a special case with an identity link function (Y * 1) Model for the Means ( Structural Model ): Ø How predictors linearly relate to the link-transformed outcome Ø New link-transformed Y p = β 0 + β 1 X p + β 2 Z p + β 3 X p Z p Model for the Variance ( Sampling/Stochastic Model ): Ø If the errors aren t normally distributed, then what are they? Ø Family of alternative distributions at our disposal that map onto what the distribution of errors could possibly look like

55 A General Linear Model With Binary Outcomes? How can we have a linear relationship between X & Y? Probability of a 1 is bounded between 0 and 1, but predicted probabilities from a linear model aren t bounded Ø Impossible values Linear relationship needs to shut off somehow à made nonlinear Prob (Y=1) ???? Prob (Y=1) X Predictor X Predictor

56 WRAPPING UP

57 Wrapping Up Today discussed estimation, and in the process showed how differing estimators can give you different statistics The key today was to shake your statistical view point: Ø There are many more ways to arrive at statistical results than you may know The take home point is that not all estimators are created equal Ø If ever presented with estimates: ask how the numbers were attained Ø If ever getting estimates: get the best you can with your data Next week your world will further be expanded when we introduce maximum likelihood estimators

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation

Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation Univariate Normal Distribution; GLM with the Univariate Normal; Least Squares Estimation PRE 905: Multivariate Analysis Spring 2014 Lecture 4 Today s Class The building blocks: The basics of mathematical

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7 Introduction to Generalized Univariate Models: Models for Binary Outcomes EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7 EPSY 905: Intro to Generalized In This Lecture A short review

More information

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016 EPSY 905: Intro to Bayesian and MCMC Today s Class An

More information

Generalized Linear Models for Non-Normal Data

Generalized Linear Models for Non-Normal Data Generalized Linear Models for Non-Normal Data Today s Class: 3 parts of a generalized model Models for binary outcomes Complications for generalized multivariate or multilevel models SPLH 861: Lecture

More information

Introduction to Matrix Algebra and the Multivariate Normal Distribution

Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Matrix Algebra and the Multivariate Normal Distribution Introduction to Structural Equation Modeling Lecture #2 January 18, 2012 ERSH 8750: Lecture 2 Motivation for Learning the Multivariate

More information

Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood

Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood Maximum Likelihood Estimation; Robust Maximum Likelihood; Missing Data with Maximum Likelihood PRE 906: Structural Equation Modeling Lecture #3 February 4, 2015 PRE 906, SEM: Estimation Today s Class An

More information

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model EPSY 905: Multivariate Analysis Lecture 1 20 January 2016 EPSY 905: Lecture 1 -

More information

Hierarchical Generalized Linear Models. ERSH 8990 REMS Seminar on HLM Last Lecture!

Hierarchical Generalized Linear Models. ERSH 8990 REMS Seminar on HLM Last Lecture! Hierarchical Generalized Linear Models ERSH 8990 REMS Seminar on HLM Last Lecture! Hierarchical Generalized Linear Models Introduction to generalized models Models for binary outcomes Interpreting parameter

More information

A Introduction to Matrix Algebra and the Multivariate Normal Distribution

A Introduction to Matrix Algebra and the Multivariate Normal Distribution A Introduction to Matrix Algebra and the Multivariate Normal Distribution PRE 905: Multivariate Analysis Spring 2014 Lecture 6 PRE 905: Lecture 7 Matrix Algebra and the MVN Distribution Today s Class An

More information

Generalized Models: Part 1

Generalized Models: Part 1 Generalized Models: Part 1 Topics: Introduction to generalized models Introduction to maximum likelihood estimation Models for binary outcomes Models for proportion outcomes Models for categorical outcomes

More information

A Re-Introduction to General Linear Models

A Re-Introduction to General Linear Models A Re-Introduction to General Linear Models Today s Class: Big picture overview Why we are using restricted maximum likelihood within MIXED instead of least squares within GLM Linear model interpretation

More information

SEM with Non Normal Data: Robust Estimation and Generalized Models

SEM with Non Normal Data: Robust Estimation and Generalized Models SEM with Non Normal Data: Robust Estimation and Generalized Models Introduction to Structural Equation Modeling Lecture #10 April 4, 2012 ERSH 8750: Lecture 10 Today s Class Data assumptions of SEM Continuous

More information

Comparing IRT with Other Models

Comparing IRT with Other Models Comparing IRT with Other Models Lecture #14 ICPSR Item Response Theory Workshop Lecture #14: 1of 45 Lecture Overview The final set of slides will describe a parallel between IRT and another commonly used

More information

Introduction to Generalized Models

Introduction to Generalized Models Introduction to Generalized Models Today s topics: The big picture of generalized models Review of maximum likelihood estimation Models for binary outcomes Models for proportion outcomes Models for categorical

More information

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model

Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model Course Introduction and Overview Descriptive Statistics Conceptualizations of Variance Review of the General Linear Model PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 1: August 22, 2012

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

A Re-Introduction to General Linear Models (GLM)

A Re-Introduction to General Linear Models (GLM) A Re-Introduction to General Linear Models (GLM) Today s Class: You do know the GLM Estimation (where the numbers in the output come from): From least squares to restricted maximum likelihood (REML) Reviewing

More information

Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models

Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models Repeated Measures ANOVA Multivariate ANOVA and Their Relationship to Linear Mixed Models EPSY 905: Multivariate Analysis Spring 2016 Lecture #12 April 20, 2016 EPSY 905: RM ANOVA, MANOVA, and Mixed Models

More information

Review of the General Linear Model

Review of the General Linear Model Review of the General Linear Model EPSY 905: Multivariate Analysis Online Lecture #2 Learning Objectives Types of distributions: Ø Conditional distributions The General Linear Model Ø Regression Ø Analysis

More information

Joint Probability Distributions and Random Samples (Devore Chapter Five)

Joint Probability Distributions and Random Samples (Devore Chapter Five) Joint Probability Distributions and Random Samples (Devore Chapter Five) 1016-345-01: Probability and Statistics for Engineers Spring 2013 Contents 1 Joint Probability Distributions 2 1.1 Two Discrete

More information

A Primer on Statistical Inference using Maximum Likelihood

A Primer on Statistical Inference using Maximum Likelihood A Primer on Statistical Inference using Maximum Likelihood November 3, 2017 1 Inference via Maximum Likelihood Statistical inference is the process of using observed data to estimate features of the population.

More information

An Introduction to Matrix Algebra

An Introduction to Matrix Algebra An Introduction to Matrix Algebra EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #8 EPSY 905: Matrix Algebra In This Lecture An introduction to matrix algebra Ø Scalars, vectors, and matrices

More information

Generalized Linear Models for Count, Skewed, and If and How Much Outcomes

Generalized Linear Models for Count, Skewed, and If and How Much Outcomes Generalized Linear Models for Count, Skewed, and If and How Much Outcomes Today s Class: Review of 3 parts of a generalized model Models for discrete count or continuous skewed outcomes Models for two-part

More information

Statistics Boot Camp. Dr. Stephanie Lane Institute for Defense Analyses DATAWorks 2018

Statistics Boot Camp. Dr. Stephanie Lane Institute for Defense Analyses DATAWorks 2018 Statistics Boot Camp Dr. Stephanie Lane Institute for Defense Analyses DATAWorks 2018 March 21, 2018 Outline of boot camp Summarizing and simplifying data Point and interval estimation Foundations of statistical

More information

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis Path Analysis PRE 906: Structural Equation Modeling Lecture #5 February 18, 2015 PRE 906, SEM: Lecture 5 - Path Analysis Key Questions for Today s Lecture What distinguishes path models from multivariate

More information

An Introduction to Path Analysis

An Introduction to Path Analysis An Introduction to Path Analysis PRE 905: Multivariate Analysis Lecture 10: April 15, 2014 PRE 905: Lecture 10 Path Analysis Today s Lecture Path analysis starting with multivariate regression then arriving

More information

Linear algebra and differential equations (Math 54): Lecture 10

Linear algebra and differential equations (Math 54): Lecture 10 Linear algebra and differential equations (Math 54): Lecture 10 Vivek Shende February 24, 2016 Hello and welcome to class! As you may have observed, your usual professor isn t here today. He ll be back

More information

An Introduction to Mplus and Path Analysis

An Introduction to Mplus and Path Analysis An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression

More information

Notes on Random Variables, Expectations, Probability Densities, and Martingales

Notes on Random Variables, Expectations, Probability Densities, and Martingales Eco 315.2 Spring 2006 C.Sims Notes on Random Variables, Expectations, Probability Densities, and Martingales Includes Exercise Due Tuesday, April 4. For many or most of you, parts of these notes will be

More information

Vectors and Matrices Statistics with Vectors and Matrices

Vectors and Matrices Statistics with Vectors and Matrices Vectors and Matrices Statistics with Vectors and Matrices Lecture 3 September 7, 005 Analysis Lecture #3-9/7/005 Slide 1 of 55 Today s Lecture Vectors and Matrices (Supplement A - augmented with SAS proc

More information

An Introduction to Multilevel Models. PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 25: December 7, 2012

An Introduction to Multilevel Models. PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 25: December 7, 2012 An Introduction to Multilevel Models PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 25: December 7, 2012 Today s Class Concepts in Longitudinal Modeling Between-Person vs. +Within-Person

More information

review session gov 2000 gov 2000 () review session 1 / 38

review session gov 2000 gov 2000 () review session 1 / 38 review session gov 2000 gov 2000 () review session 1 / 38 Overview Random Variables and Probability Univariate Statistics Bivariate Statistics Multivariate Statistics Causal Inference gov 2000 () review

More information

Regression with Nonlinear Transformations

Regression with Nonlinear Transformations Regression with Nonlinear Transformations Joel S Steele Portland State University Abstract Gaussian Likelihood When data are drawn from a Normal distribution, N (µ, σ 2 ), we can use the Gaussian distribution

More information

Probability Theory for Machine Learning. Chris Cremer September 2015

Probability Theory for Machine Learning. Chris Cremer September 2015 Probability Theory for Machine Learning Chris Cremer September 2015 Outline Motivation Probability Definitions and Rules Probability Distributions MLE for Gaussian Parameter Estimation MLE and Least Squares

More information

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables

Chapter 2. Some Basic Probability Concepts. 2.1 Experiments, Outcomes and Random Variables Chapter 2 Some Basic Probability Concepts 2.1 Experiments, Outcomes and Random Variables A random variable is a variable whose value is unknown until it is observed. The value of a random variable results

More information

PROBABILITY THEORY REVIEW

PROBABILITY THEORY REVIEW PROBABILITY THEORY REVIEW CMPUT 466/551 Martha White Fall, 2017 REMINDERS Assignment 1 is due on September 28 Thought questions 1 are due on September 21 Chapters 1-4, about 40 pages If you are printing,

More information

Generalized Multilevel Models for Non-Normal Outcomes

Generalized Multilevel Models for Non-Normal Outcomes Generalized Multilevel Models for Non-Normal Outcomes Topics: 3 parts of a generalized (multilevel) model Models for binary, proportion, and categorical outcomes Complications for generalized multilevel

More information

Introduction to Within-Person Analysis and RM ANOVA

Introduction to Within-Person Analysis and RM ANOVA Introduction to Within-Person Analysis and RM ANOVA Today s Class: From between-person to within-person ANOVAs for longitudinal data Variance model comparisons using 2 LL CLP 944: Lecture 3 1 The Two Sides

More information

Model Estimation Example

Model Estimation Example Ronald H. Heck 1 EDEP 606: Multivariate Methods (S2013) April 7, 2013 Model Estimation Example As we have moved through the course this semester, we have encountered the concept of model estimation. Discussions

More information

Ron Heck, Fall Week 8: Introducing Generalized Linear Models: Logistic Regression 1 (Replaces prior revision dated October 20, 2011)

Ron Heck, Fall Week 8: Introducing Generalized Linear Models: Logistic Regression 1 (Replaces prior revision dated October 20, 2011) Ron Heck, Fall 2011 1 EDEP 768E: Seminar in Multilevel Modeling rev. January 3, 2012 (see footnote) Week 8: Introducing Generalized Linear Models: Logistic Regression 1 (Replaces prior revision dated October

More information

Describing Change over Time: Adding Linear Trends

Describing Change over Time: Adding Linear Trends Describing Change over Time: Adding Linear Trends Longitudinal Data Analysis Workshop Section 7 University of Georgia: Institute for Interdisciplinary Research in Education and Human Development Section

More information

MIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010

MIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010 MIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010 Part 1 of this document can be found at http://www.uvm.edu/~dhowell/methods/supplements/mixed Models for Repeated Measures1.pdf

More information

Introduction to Random Effects of Time and Model Estimation

Introduction to Random Effects of Time and Model Estimation Introduction to Random Effects of Time and Model Estimation Today s Class: The Big Picture Multilevel model notation Fixed vs. random effects of time Random intercept vs. random slope models How MLM =

More information

Introduction to Bayesian Learning. Machine Learning Fall 2018

Introduction to Bayesian Learning. Machine Learning Fall 2018 Introduction to Bayesian Learning Machine Learning Fall 2018 1 What we have seen so far What does it mean to learn? Mistake-driven learning Learning by counting (and bounding) number of mistakes PAC learnability

More information

Lecture #11: Classification & Logistic Regression

Lecture #11: Classification & Logistic Regression Lecture #11: Classification & Logistic Regression CS 109A, STAT 121A, AC 209A: Data Science Weiwei Pan, Pavlos Protopapas, Kevin Rader Fall 2016 Harvard University 1 Announcements Midterm: will be graded

More information

Model Assumptions; Predicting Heterogeneity of Variance

Model Assumptions; Predicting Heterogeneity of Variance Model Assumptions; Predicting Heterogeneity of Variance Today s topics: Model assumptions Normality Constant variance Predicting heterogeneity of variance CLP 945: Lecture 6 1 Checking for Violations of

More information

Linear Regression (9/11/13)

Linear Regression (9/11/13) STA561: Probabilistic machine learning Linear Regression (9/11/13) Lecturer: Barbara Engelhardt Scribes: Zachary Abzug, Mike Gloudemans, Zhuosheng Gu, Zhao Song 1 Why use linear regression? Figure 1: Scatter

More information

Generalized Linear Models

Generalized Linear Models Generalized Linear Models Advanced Methods for Data Analysis (36-402/36-608 Spring 2014 1 Generalized linear models 1.1 Introduction: two regressions So far we ve seen two canonical settings for regression.

More information

Introducing Generalized Linear Models: Logistic Regression

Introducing Generalized Linear Models: Logistic Regression Ron Heck, Summer 2012 Seminars 1 Multilevel Regression Models and Their Applications Seminar Introducing Generalized Linear Models: Logistic Regression The generalized linear model (GLM) represents and

More information

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari

MS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind

More information

Logistic Regression: Regression with a Binary Dependent Variable

Logistic Regression: Regression with a Binary Dependent Variable Logistic Regression: Regression with a Binary Dependent Variable LEARNING OBJECTIVES Upon completing this chapter, you should be able to do the following: State the circumstances under which logistic regression

More information

Generalized Linear. Mixed Models. Methods and Applications. Modern Concepts, Walter W. Stroup. Texts in Statistical Science.

Generalized Linear. Mixed Models. Methods and Applications. Modern Concepts, Walter W. Stroup. Texts in Statistical Science. Texts in Statistical Science Generalized Linear Mixed Models Modern Concepts, Methods and Applications Walter W. Stroup CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is an imprint

More information

22 Approximations - the method of least squares (1)

22 Approximations - the method of least squares (1) 22 Approximations - the method of least squares () Suppose that for some y, the equation Ax = y has no solutions It may happpen that this is an important problem and we can t just forget about it If we

More information

Lecture 01: Introduction

Lecture 01: Introduction Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction

More information

Applied Regression Analysis

Applied Regression Analysis Applied Regression Analysis Lecture 2 January 27, 2005 Lecture #2-1/27/2005 Slide 1 of 46 Today s Lecture Simple linear regression. Partitioning the sum of squares. Tests of significance.. Regression diagnostics

More information

Lecture 2: Linear regression

Lecture 2: Linear regression Lecture 2: Linear regression Roger Grosse 1 Introduction Let s ump right in and look at our first machine learning algorithm, linear regression. In regression, we are interested in predicting a scalar-valued

More information

DIFFERENTIAL EQUATIONS

DIFFERENTIAL EQUATIONS DIFFERENTIAL EQUATIONS Basic Concepts Paul Dawkins Table of Contents Preface... Basic Concepts... 1 Introduction... 1 Definitions... Direction Fields... 8 Final Thoughts...19 007 Paul Dawkins i http://tutorial.math.lamar.edu/terms.aspx

More information

HUDM4122 Probability and Statistical Inference. February 2, 2015

HUDM4122 Probability and Statistical Inference. February 2, 2015 HUDM4122 Probability and Statistical Inference February 2, 2015 Special Session on SPSS Thursday, April 23 4pm-6pm As of when I closed the poll, every student except one could make it to this I am happy

More information

Chapter 3: Examining Relationships

Chapter 3: Examining Relationships Chapter 3: Examining Relationships Most statistical studies involve more than one variable. Often in the AP Statistics exam, you will be asked to compare two data sets by using side by side boxplots or

More information

CHAPTER 4 VECTORS. Before we go any further, we must talk about vectors. They are such a useful tool for

CHAPTER 4 VECTORS. Before we go any further, we must talk about vectors. They are such a useful tool for CHAPTER 4 VECTORS Before we go any further, we must talk about vectors. They are such a useful tool for the things to come. The concept of a vector is deeply rooted in the understanding of physical mechanics

More information

Simple, Marginal, and Interaction Effects in General Linear Models

Simple, Marginal, and Interaction Effects in General Linear Models Simple, Marginal, and Interaction Effects in General Linear Models PRE 905: Multivariate Analysis Lecture 3 Today s Class Centering and Coding Predictors Interpreting Parameters in the Model for the Means

More information

Review of Multiple Regression

Review of Multiple Regression Ronald H. Heck 1 Let s begin with a little review of multiple regression this week. Linear models [e.g., correlation, t-tests, analysis of variance (ANOVA), multiple regression, path analysis, multivariate

More information

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed

More information

Interactions among Continuous Predictors

Interactions among Continuous Predictors Interactions among Continuous Predictors Today s Class: Simple main effects within two-way interactions Conquering TEST/ESTIMATE/LINCOM statements Regions of significance Three-way interactions (and beyond

More information

ACE 562 Fall Lecture 2: Probability, Random Variables and Distributions. by Professor Scott H. Irwin

ACE 562 Fall Lecture 2: Probability, Random Variables and Distributions. by Professor Scott H. Irwin ACE 562 Fall 2005 Lecture 2: Probability, Random Variables and Distributions Required Readings: by Professor Scott H. Irwin Griffiths, Hill and Judge. Some Basic Ideas: Statistical Concepts for Economists,

More information

Probability and Inference. POLI 205 Doing Research in Politics. Populations and Samples. Probability. Fall 2015

Probability and Inference. POLI 205 Doing Research in Politics. Populations and Samples. Probability. Fall 2015 Fall 2015 Population versus Sample Population: data for every possible relevant case Sample: a subset of cases that is drawn from an underlying population Inference Parameters and Statistics A parameter

More information

MATH 1150 Chapter 2 Notation and Terminology

MATH 1150 Chapter 2 Notation and Terminology MATH 1150 Chapter 2 Notation and Terminology Categorical Data The following is a dataset for 30 randomly selected adults in the U.S., showing the values of two categorical variables: whether or not the

More information

Overfitting, Bias / Variance Analysis

Overfitting, Bias / Variance Analysis Overfitting, Bias / Variance Analysis Professor Ameet Talwalkar Professor Ameet Talwalkar CS260 Machine Learning Algorithms February 8, 207 / 40 Outline Administration 2 Review of last lecture 3 Basic

More information

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014

Probability. Machine Learning and Pattern Recognition. Chris Williams. School of Informatics, University of Edinburgh. August 2014 Probability Machine Learning and Pattern Recognition Chris Williams School of Informatics, University of Edinburgh August 2014 (All of the slides in this course have been adapted from previous versions

More information

Naïve Bayes Introduction to Machine Learning. Matt Gormley Lecture 18 Oct. 31, 2018

Naïve Bayes Introduction to Machine Learning. Matt Gormley Lecture 18 Oct. 31, 2018 10-601 Introduction to Machine Learning Machine Learning Department School of Computer Science Carnegie Mellon University Naïve Bayes Matt Gormley Lecture 18 Oct. 31, 2018 1 Reminders Homework 6: PAC Learning

More information

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages: Glossary The ISI glossary of statistical terms provides definitions in a number of different languages: http://isi.cbs.nl/glossary/index.htm Adjusted r 2 Adjusted R squared measures the proportion of the

More information

Mean Vector Inferences

Mean Vector Inferences Mean Vector Inferences Lecture 5 September 21, 2005 Multivariate Analysis Lecture #5-9/21/2005 Slide 1 of 34 Today s Lecture Inferences about a Mean Vector (Chapter 5). Univariate versions of mean vector

More information

Part 6: Multivariate Normal and Linear Models

Part 6: Multivariate Normal and Linear Models Part 6: Multivariate Normal and Linear Models 1 Multiple measurements Up until now all of our statistical models have been univariate models models for a single measurement on each member of a sample of

More information

Logit Regression and Quantities of Interest

Logit Regression and Quantities of Interest Logit Regression and Quantities of Interest Stephen Pettigrew March 4, 2015 Stephen Pettigrew Logit Regression and Quantities of Interest March 4, 2015 1 / 57 Outline 1 Logistics 2 Generalized Linear Models

More information

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015

Human-Oriented Robotics. Probability Refresher. Kai Arras Social Robotics Lab, University of Freiburg Winter term 2014/2015 Probability Refresher Kai Arras, University of Freiburg Winter term 2014/2015 Probability Refresher Introduction to Probability Random variables Joint distribution Marginalization Conditional probability

More information

Midterm. Introduction to Machine Learning. CS 189 Spring Please do not open the exam before you are instructed to do so.

Midterm. Introduction to Machine Learning. CS 189 Spring Please do not open the exam before you are instructed to do so. CS 89 Spring 07 Introduction to Machine Learning Midterm Please do not open the exam before you are instructed to do so. The exam is closed book, closed notes except your one-page cheat sheet. Electronic

More information

This does not cover everything on the final. Look at the posted practice problems for other topics.

This does not cover everything on the final. Look at the posted practice problems for other topics. Class 7: Review Problems for Final Exam 8.5 Spring 7 This does not cover everything on the final. Look at the posted practice problems for other topics. To save time in class: set up, but do not carry

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

An introduction to biostatistics: part 1

An introduction to biostatistics: part 1 An introduction to biostatistics: part 1 Cavan Reilly September 6, 2017 Table of contents Introduction to data analysis Uncertainty Probability Conditional probability Random variables Discrete random

More information

, (1) e i = ˆσ 1 h ii. c 2016, Jeffrey S. Simonoff 1

, (1) e i = ˆσ 1 h ii. c 2016, Jeffrey S. Simonoff 1 Regression diagnostics As is true of all statistical methodologies, linear regression analysis can be a very effective way to model data, as along as the assumptions being made are true. For the regression

More information

Approximations - the method of least squares (1)

Approximations - the method of least squares (1) Approximations - the method of least squares () In many applications, we have to consider the following problem: Suppose that for some y, the equation Ax = y has no solutions It could be that this is an

More information

Advanced Regression Topics: Violation of Assumptions

Advanced Regression Topics: Violation of Assumptions Advanced Regression Topics: Violation of Assumptions Lecture 7 February 15, 2005 Applied Regression Analysis Lecture #7-2/15/2005 Slide 1 of 36 Today s Lecture Today s Lecture rapping Up Revisiting residuals.

More information

Estimation: Problems & Solutions

Estimation: Problems & Solutions Estimation: Problems & Solutions Edps/Psych/Stat 587 Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2017 Outline 1. Introduction: Estimation of

More information

Review of Statistics

Review of Statistics Review of Statistics Topics Descriptive Statistics Mean, Variance Probability Union event, joint event Random Variables Discrete and Continuous Distributions, Moments Two Random Variables Covariance and

More information

Review of Statistics 101

Review of Statistics 101 Review of Statistics 101 We review some important themes from the course 1. Introduction Statistics- Set of methods for collecting/analyzing data (the art and science of learning from data). Provides methods

More information

Naïve Bayes. Jia-Bin Huang. Virginia Tech Spring 2019 ECE-5424G / CS-5824

Naïve Bayes. Jia-Bin Huang. Virginia Tech Spring 2019 ECE-5424G / CS-5824 Naïve Bayes Jia-Bin Huang ECE-5424G / CS-5824 Virginia Tech Spring 2019 Administrative HW 1 out today. Please start early! Office hours Chen: Wed 4pm-5pm Shih-Yang: Fri 3pm-4pm Location: Whittemore 266

More information

Review of CLDP 944: Multilevel Models for Longitudinal Data

Review of CLDP 944: Multilevel Models for Longitudinal Data Review of CLDP 944: Multilevel Models for Longitudinal Data Topics: Review of general MLM concepts and terminology Model comparisons and significance testing Fixed and random effects of time Significance

More information

BAYESIAN DECISION THEORY

BAYESIAN DECISION THEORY Last updated: September 17, 2012 BAYESIAN DECISION THEORY Problems 2 The following problems from the textbook are relevant: 2.1 2.9, 2.11, 2.17 For this week, please at least solve Problem 2.3. We will

More information

STA 216: GENERALIZED LINEAR MODELS. Lecture 1. Review and Introduction. Much of statistics is based on the assumption that random

STA 216: GENERALIZED LINEAR MODELS. Lecture 1. Review and Introduction. Much of statistics is based on the assumption that random STA 216: GENERALIZED LINEAR MODELS Lecture 1. Review and Introduction Much of statistics is based on the assumption that random variables are continuous & normally distributed. Normal linear regression

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Rejection regions for the bivariate case

Rejection regions for the bivariate case Rejection regions for the bivariate case The rejection region for the T 2 test (and similarly for Z 2 when Σ is known) is the region outside of an ellipse, for which there is a (1-α)% chance that the test

More information

18 Bivariate normal distribution I

18 Bivariate normal distribution I 8 Bivariate normal distribution I 8 Example Imagine firing arrows at a target Hopefully they will fall close to the target centre As we fire more arrows we find a high density near the centre and fewer

More information

Latent Trait Reliability

Latent Trait Reliability Latent Trait Reliability Lecture #7 ICPSR Item Response Theory Workshop Lecture #7: 1of 66 Lecture Overview Classical Notions of Reliability Reliability with IRT Item and Test Information Functions Concepts

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Probability Distributions: Continuous

Probability Distributions: Continuous Probability Distributions: Continuous INFO-2301: Quantitative Reasoning 2 Michael Paul and Jordan Boyd-Graber FEBRUARY 28, 2017 INFO-2301: Quantitative Reasoning 2 Paul and Boyd-Graber Probability Distributions:

More information

Bayesian Inference for Regression Parameters

Bayesian Inference for Regression Parameters Bayesian Inference for Regression Parameters 1 Bayesian inference for simple linear regression parameters follows the usual pattern for all Bayesian analyses: 1. Form a prior distribution over all unknown

More information

TDA231. Logistic regression

TDA231. Logistic regression TDA231 Devdatt Dubhashi dubhashi@chalmers.se Dept. of Computer Science and Engg. Chalmers University February 19, 2016 Some data 5 x2 0 5 5 0 5 x 1 In the Bayes classifier, we built a model of each class

More information

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014

Regression. ECO 312 Fall 2013 Chris Sims. January 12, 2014 ECO 312 Fall 2013 Chris Sims Regression January 12, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported License What

More information