Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood

Size: px
Start display at page:

Download "Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood"

Transcription

1 Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal items in order to measure an underlying trait using a collection of observed responses to test questions or items. For binary items, a common IRT model is the two parameter logistic (2PL) form. Assume that we have observed p item responses for n individuals where a correct responses is designated with a 1 and an incorrect response a 0. Let the matrix Y = {Y ij : i = 1,..., n; j = 1,..., p} denote the set of binary item responses. The 2PL IRT form for the probability of a correct response on item j for participant i denoted by Y ij = 1 is the following, P (Y ij = 1 θ i, a j, b j ) = exp ( a j (θ i b j )) Here a j and b j are item specific parameters while θ i is an individual specific parameter. The parameter b in this framework is a measure of the item s difficulty. θ represents the individual s level of a latent trait. The parameters θ and b are on the same scale here and when θ > b the individual has a greater probability of answering the item correctly than answering it incorrectly. The item parameter a measures the degree to which an item differentiates among individuals of different levels of the latent trait. As a result, it is referred to as the discrimination parameter. The probability of a correct response is assumed to increase with θ in IRT models and a is restricted to be positive. Finally, the responses of an individual are assumed to be independent given their ability θ. The above interpretations make the linear form a j (θ i b j ) a useful one. When working with ordinal responses of more than two categories, IRT practitioners commonly rely upon the graded response model (GRM) or the generalized partial credit model (GPCM). We discuss the GPCM briefly. The GPCM (1) 1

2 is essentially an adjacent categories logit model. If we assume that item j has a total of K j categories, the probability that individual i is credited with item response k j is ( kj ) exp h=1 a j(θ i b jh ) P (Y ij = k j θ, a j, b j ) = ( Kj l=1 exp l ) (2) h=1 a j(θ i b jh ) where b j is now a vector of location or threshold parameters between categories. It is convention to set b j1 = 0 so that there are a total of K j 1 parameters. For instances where the observed response is continuous or is discrete but with a large number of categories, the manifest observations may be binned in some fashion in order to apply existing IRT models to such data. We are interested in estimating cognitive functioning in individuals using a set of their observed responses to cognitive testing items. As a result, we are primarily focused on estimation of the latent trait. These sets of outcomes for each individual may consist of binary, categorical (ordinal), counts, rightcensored counts and continuous outcomes. Thus we require a more general approach than the above. One common approach to latent variable models for more diverse types of outcomes, or mixed outcomes, includes latent variables in generalized linear models. Sammel, Ryan and Legler [4] developed a generalized linear model formulation that included a latent variable among the covariates with the latent variable also a function of covariates. Application was to testing results similar to ours. Although the model was formulated to handle outcomes from exponential families in general, the methods were demonstrated for binary and continuous data only. Moustaki and Knott [3] focus on the measurement model, generalizing the results of Sammel and Ryan to include multiple latent variables and to handle polytomous, Poisson and gamma distributed outcomes as well as binary and normal outcomes. While the above models relied on maximum likelihood for inference and employed the EM algorithm for estimation, Dunson [1] cast the problem in a Bayesian framework and extended the models to allow for repeated measurements, serial correlations in the latent variables and individual-specific response behavior. More recently, we have applied these generalized latent traits models in an IRT setting and, motivated by items on tests of cognitive functioning, we have extended these methods to allow for right-censored count outcomes as well as time-to-completion outcomes. We will refer to this model as the mixed outcomes IRT model. Of course, we could continue to further extend our model to accommodate many more types of outcomes that we may encounter. However our interest is ultimately in obtaining a good estimate of the latent ability, θ. Specification 2

3 of a diverse set of distributions F 1,..., F p to model the p mixed outcomes is merely a means to obtaining a good estimate of θ. Ideally however we could estimate θ without the need to specify the distributions F 1,..., F p. Hoff [2] produced the extended rank likelihood method to estimate parameters of a Gaussian copula for arbitrary marginal distributions and without the need to specify any assumptions regarding the marginal distributions. We seek to use the extended rank likelihood method in an IRT framework. Section 2 discusses the extended rank likelihood method and how we may use it to estimate latent traits. Section 3 discusses the Bayesian methods by which we are able to obtain estimates from this model. In Section 4, we demonstrate these methods as well as the mixed outcomes IRT model on simulated data. We conclude with discussion. 2 Methods We start by summarizing the methods of Hoff and then proceed to discuss how we may adapt them to estimate latent abilities provided a number of observed responses to items with unknown distributions. Hoff developed the extended rank likelihood methods to estimate dependence among multiple outcomes using a Gaussian copula without relying on assumptions regarding the marginal distributions. As above, let i = 1,..., n denote the ith participant, and let j = 1,..., p denote the jth item. Then, y ij represents the observed response of participant i to item j with marginal distribution F j. Thus we may also represent y ij = F 1 j (u ij ) where u ij is a uniform (0,1) random variable. Similarly we might formulate y ij = F 1 j [Φ(z ij )] where Φ denotes the normal CDF and z ij is distributed normal. If we assume that the correlation of z ij with z ij for all 1 j, j p is specified by the p p correlation matrix C, then we have the following Gaussian copula sampling model as presented by Hoff, z 1,..., z n C i.i.d. N(0, C) (3) where z i is the p-length vector of z ij for participant i. y i,j = F 1 j [Φ(z ij )] (4) The ultimate goal here is to estimate C rather than F 1,..., F p. Knowledge of the z ij s would allow us to estimate C; however these are unobserved. Nonetheless, we do have some information about the z ij s through the y ij s; we know that y ij < y ij implies z ij < z ij. If we let Z = (z 1,..., z n ) T and Y = (y 1,..., y n ) T, then Z D(Y) where D(Y) = {Z R n p : max{z kj : y kj < y ij } < z ij < min{z kj : y ij < y kj }}. (5) 3

4 Using this construction, one can construct a likelihood for C relying solely on Z based on the following probability, Pr(Z D(Y) C, F 1,..., F p ) = p(z C) dz D(Y) = Pr(Z D(Y) C) (6) that does not depend on F 1,..., F p. This enables the following decomposition of the density of Y, p(y C, F 1,..., F p ) = p(y, Z D(Y) C, F 1,..., F p ) (7) = Pr(Z D(Y) C, F 1,..., F p ) p(y Z D(Y), C, F 1,..., F p ). (8) This decomposition uses the fact that Z is conditionally independent of the marginal distributions as shown above and that the event Z D(Y) occurs whenever Y is observed. Thus one is able to estimate the dependence structure of Y through C without any knowledge or assumptions about the marginal distributions. In the context of item response theory modeling, we are not interested in explicitly estimating C. However we are interested in the characterizing the interdependencies in multivariate observed responses through a latent variable model. This difference is represented graphically in Figures 1 and 2. In Figure 1, we see the latent z ij s as directly correlated and this is parameterized through C in the Gaussian copula model. In Figure 2, we see the relation among the latent z ij s as being connected to a single latent trait θ i. z i1 z i2 z i3 z i4 z i5 y i1 y i2 y i3 y i4 y i5 Figure 1: Extended Rank Likelihood Applied To Gaussian Copula Using an IRT-type model, the data generating model now takes the form, θ i N(0, 1) (9) z i a, θ i N(aθ i, I) (10) z 1,..., z n a i.i.d. N(0, I + aa T ) (11) y ij = g j (z ij ). (12) 4

5 θ i z i1 z i2 z i3 z i4 z i5 y i1 y i2 y i3 y i4 y i5 Figure 2: Extended Rank Likelihood Applied To IRT Model Here the dependence among the z i s is restricted to I + aa T as opposed to the more general C in the Gaussian copula model. This restriction is based on our assumption that an IRT type model is appropriate for the data being analyzed. Our likelihood for the IRT model is hence based on Pr(Z D(Y) a, θ) with a and θ our parameters of interest. Finally notice that we have not included the difficulty parameters b in the above formulation. These location parameters are not thought to be identifiable because we could shift all z ij s by an arbitrary amount (as long as the amount is the same for each j) and still have a set of values entirely consistent with the rankings of the y ij s. As we are primarily interested in the estimation of the latent trait, this loss is not critical. 3 Estimation To estimate a and θ, we rely on Bayesian methods. Specifically, we rely upon Metropolis-Hastings (MH) and Gibbs sampling to obtain draws from the posterior distribution of a and θ. Because each a j is restricted to be positive, we specify a Lognormal(0, σ 2 a) prior for each. Thus, our complete model is, a j Lognormal(0, σ 2 a) θ i N(0, 1) z i a, θ i N(aθ i, I) y ij = g j (z ij ). To sample from p(a, θ Z D(Y)), we proceed by iterating through the following steps as in Hoff and as in Scott [5]. 5

6 1. Draw unobserved Z. For each i and j, sample z ij from p(z ij a, θ, Z ( i)( j), Z D(Y)). More specifically, for each j and within that for each y = unique{y 1j,..., y nj }, z ij TN (zl,z u)(a j θ i, 1) (13) where TN denotes truncated normal and z l, z u define the lower and upper truncation points, z l = max{z kj : y kj < y} (14) z u = min{z kj : y kj > y}. (15) 2. Draw latent abilities θ. For each i, we can sample directly from the conditional distribution for θ i as follows, ( a T ) z i θ i N a T a + 1, 1 a T. (16) a Draw item parameters a. To sample from p(a θ, Z D(Y), σ 2 a), we rely on MH sampling. Proposals for a are generated using the lognormal distribution. We have chosen to sample each a j individually rather than jointly. The implementation in R is presented in the Appendix. 4 Simulations We now demonstrate the above methods on simulated data. We generate 20 item responses for 600 individuals using the mixed outcome IRT model discussed in the introduction. The 20 item responses were of type and generating distribution listed in Table 1. Table 1: Types and associated distributions for the 20 simulated items. Item Type Distribution # of Items Binary Bernoulli 7 Count Poisson 3 Right Censored Count Right Censored Poisson 3 Positive Skewed Lognormal 3 Ordinal Categorical Multinomial (Adjacent Categories/GPCM) 4 6

7 In Figure 3, we presents histograms of the simulated data for two of the items. Figure 3a displays simulated responses for item 12, data that was generated using a right-censored Poisson model. One can see how the right-censored count outcomes exhibit a small ceiling effect. Figure 3b displays simulated responses for item 15. This data was produced using a lognormal distribution. We see the majority of responses clustered at the low end with a few responses dispersed among higher values. We truncated the vertical axis of the plot to provide a more granular view of the low frequency intervals; the first interval includes over 500 responses. Frequency Frequency (a) Item 12. (b) Item 15. Note that the height of the first bar has been truncated to provide a more granular view of the other intervals. It extends to over 500. Figure 3: Histograms Of Item Responses Having simulated this data, we now see how well the latent ability parameter can be recovered by different models. We consider a few different metrics including RMSE, average width of 95% credible interval, coverage properties of 95% credible interval (recognizing that coverage is a frequentist concept). However, because the location and scale of the latent variable is arbitrary and influenced by the choice of prior, it may be more meaningful to consider whether ranks of the estimated latent traits are consistent with those of the true, data-generating values. Hence we also calculate Spearman s ρ between the estimated and true values as well as the mean absolute difference in ranks. We consider three different methods of estimating the parameters. First we use the data-generating IRT model for estimation and naturally expect this to perform the best. Next we use the extended rank likelihood methods discussed above to estimate the ability parameters. While we do not expect this general 7

8 approach to perform as well as the data-generating model, we will be curious how close the performance is. Finally, we will use the data-generating model again but a misspecified version where we choose to model the lognormal outcomes as normally distributed. Comparison with the misspecified model will give us some idea of the robustness of the mixed outcome IRT model and the flexibility of the extended rank likelihood approach. We used 25,000 iterations for each Bayesian sampling scheme, discarding the first half of those as burn-in. Tuning parameters were selected for all MH steps so that acceptance rates generally fell within an interval of 25% to 50%. Trace plots did not appear to indicate any failure to converge. Table 2 presents the estimation metrics for θ by method. By every metric, the data generating mixed outcomes IRT model is not surprisingly superior. However, the extended rank likelihood method is quite competitive in all metrics, particularly the rank related metrics. The misspecified mixed outcomes model does not fare well in terms of RMSE and coverage. It does perform better in terms of the rank related metrics. The reason for this is that the simulated data for item 14 contains some very large values generated by the lognormal distribution skewing the outcomes heavily. When this is modeled with a normal distribution as in the misspecified model, the results suffer. The large simulated values for item 14 are perhaps not overly realistic but nonetheless provide some notion of the potential cost of misspecification. Table 2: Latent trait estimation metrics by estimation method. Metric Data Generating Extended Rank Misspecified RMSE % CI Coverage Mean 95% CI Width Spearman s ρ Mean Abs. Rank Diff Discussion We have applied the extended rank likelihood method of Hoff to estimate latent abilities as in a item response theory model but without the need to specify distributions for the observed item responses. In one simulation at least, the results produced by this method were quite favorable. The next step is to apply the extended rank likelihood IRT model to data from the subcortical ischemic 8

9 vascular dementia (SIVD) study. A primary goal of the SIVD study was to investigate the contribution of subcortical cerebrovascular disease to declines in cognitive functioning. Ultimately, we would like to relate the estimated latent abilities of participants in the SIVD study to MRI measured volumes of different brain matter. This extension should be straightforward as we can build upon the Bayesian model above in hierarchical fashion by specifying, where x i is a vector of covariates for participant i. θ i N(x T i β, I) (17) β N(0, σ 2 βi) (18) A number of areas require some further examination. I need to better understand identifiability of this model. Also the above model assumes a unidimensional latent trait. We would like to be able to test and/or relax this assumption. Finally the SIVD study is longitudinal and we would like to develop the methods here for longitudinal data. References [1] D.B. Dunson. Dynamic Latent Trait Models for Multidimensional Longitudinal Data. Journal of the American Statistical Association, 98(463): , [2] P.D. Hoff. Extending the rank likelihood for semiparametric copula estimation. Ann. Appl. Statist, 1(1): , [3] I. Moustaki and M. Knott. Generalized latent trait models. Psychometrika, 65(3): , [4] M.D. Sammel, L.M. Ryan, and J.M. Legler. Latent variable models for mixed discrete and continuous outcomes. Journal of the Royal Statistical Society. Series B (Methodological), pages , [5] J.G. Scott. Nonparametric Benchmarking with the Extended Rank Likelihood

Bayes methods for categorical data. April 25, 2017

Bayes methods for categorical data. April 25, 2017 Bayes methods for categorical data April 25, 2017 Motivation for joint probability models Increasing interest in high-dimensional data in broad applications Focus may be on prediction, variable selection,

More information

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent Latent Variable Models for Binary Data Suppose that for a given vector of explanatory variables x, the latent variable, U, has a continuous cumulative distribution function F (u; x) and that the binary

More information

Marginal Specifications and a Gaussian Copula Estimation

Marginal Specifications and a Gaussian Copula Estimation Marginal Specifications and a Gaussian Copula Estimation Kazim Azam Abstract Multivariate analysis involving random variables of different type like count, continuous or mixture of both is frequently required

More information

STA 216, GLM, Lecture 16. October 29, 2007

STA 216, GLM, Lecture 16. October 29, 2007 STA 216, GLM, Lecture 16 October 29, 2007 Efficient Posterior Computation in Factor Models Underlying Normal Models Generalized Latent Trait Models Formulation Genetic Epidemiology Illustration Structural

More information

Bayesian Multivariate Logistic Regression

Bayesian Multivariate Logistic Regression Bayesian Multivariate Logistic Regression Sean M. O Brien and David B. Dunson Biostatistics Branch National Institute of Environmental Health Sciences Research Triangle Park, NC 1 Goals Brief review of

More information

Comparison between conditional and marginal maximum likelihood for a class of item response models

Comparison between conditional and marginal maximum likelihood for a class of item response models (1/24) Comparison between conditional and marginal maximum likelihood for a class of item response models Francesco Bartolucci, University of Perugia (IT) Silvia Bacci, University of Perugia (IT) Claudia

More information

A Fully Nonparametric Modeling Approach to. BNP Binary Regression

A Fully Nonparametric Modeling Approach to. BNP Binary Regression A Fully Nonparametric Modeling Approach to Binary Regression Maria Department of Applied Mathematics and Statistics University of California, Santa Cruz SBIES, April 27-28, 2012 Outline 1 2 3 Simulation

More information

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence Bayesian Inference in GLMs Frequentists typically base inferences on MLEs, asymptotic confidence limits, and log-likelihood ratio tests Bayesians base inferences on the posterior distribution of the unknowns

More information

Anders Skrondal. Norwegian Institute of Public Health London School of Hygiene and Tropical Medicine. Based on joint work with Sophia Rabe-Hesketh

Anders Skrondal. Norwegian Institute of Public Health London School of Hygiene and Tropical Medicine. Based on joint work with Sophia Rabe-Hesketh Constructing Latent Variable Models using Composite Links Anders Skrondal Norwegian Institute of Public Health London School of Hygiene and Tropical Medicine Based on joint work with Sophia Rabe-Hesketh

More information

Gibbs Sampling in Latent Variable Models #1

Gibbs Sampling in Latent Variable Models #1 Gibbs Sampling in Latent Variable Models #1 Econ 690 Purdue University Outline 1 Data augmentation 2 Probit Model Probit Application A Panel Probit Panel Probit 3 The Tobit Model Example: Female Labor

More information

Fitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation

Fitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation Fitting Multidimensional Latent Variable Models using an Efficient Laplace Approximation Dimitris Rizopoulos Department of Biostatistics, Erasmus University Medical Center, the Netherlands d.rizopoulos@erasmusmc.nl

More information

Comparing IRT with Other Models

Comparing IRT with Other Models Comparing IRT with Other Models Lecture #14 ICPSR Item Response Theory Workshop Lecture #14: 1of 45 Lecture Overview The final set of slides will describe a parallel between IRT and another commonly used

More information

Part 8: GLMs and Hierarchical LMs and GLMs

Part 8: GLMs and Hierarchical LMs and GLMs Part 8: GLMs and Hierarchical LMs and GLMs 1 Example: Song sparrow reproductive success Arcese et al., (1992) provide data on a sample from a population of 52 female song sparrows studied over the course

More information

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Multilevel Statistical Models: 3 rd edition, 2003 Contents Multilevel Statistical Models: 3 rd edition, 2003 Contents Preface Acknowledgements Notation Two and three level models. A general classification notation and diagram Glossary Chapter 1 An introduction

More information

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff

Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff Summary of Extending the Rank Likelihood for Semiparametric Copula Estimation, by Peter Hoff David Gerard Department of Statistics University of Washington gerard2@uw.edu May 2, 2013 David Gerard (UW)

More information

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University

Motivation Scale Mixutres of Normals Finite Gaussian Mixtures Skew-Normal Models. Mixture Models. Econ 690. Purdue University Econ 690 Purdue University In virtually all of the previous lectures, our models have made use of normality assumptions. From a computational point of view, the reason for this assumption is clear: combined

More information

Bayesian non-parametric model to longitudinally predict churn

Bayesian non-parametric model to longitudinally predict churn Bayesian non-parametric model to longitudinally predict churn Bruno Scarpa Università di Padova Conference of European Statistics Stakeholders Methodologists, Producers and Users of European Statistics

More information

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements Jeffrey N. Rouder Francis Tuerlinckx Paul L. Speckman Jun Lu & Pablo Gomez May 4 008 1 The Weibull regression model

More information

Generalized Linear Models for Non-Normal Data

Generalized Linear Models for Non-Normal Data Generalized Linear Models for Non-Normal Data Today s Class: 3 parts of a generalized model Models for binary outcomes Complications for generalized multivariate or multilevel models SPLH 861: Lecture

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Department of Forestry & Department of Geography, Michigan State University, Lansing Michigan, U.S.A. 2 Biostatistics, School of Public

More information

Contents. Part I: Fundamentals of Bayesian Inference 1

Contents. Part I: Fundamentals of Bayesian Inference 1 Contents Preface xiii Part I: Fundamentals of Bayesian Inference 1 1 Probability and inference 3 1.1 The three steps of Bayesian data analysis 3 1.2 General notation for statistical inference 4 1.3 Bayesian

More information

PIRLS 2016 Achievement Scaling Methodology 1

PIRLS 2016 Achievement Scaling Methodology 1 CHAPTER 11 PIRLS 2016 Achievement Scaling Methodology 1 The PIRLS approach to scaling the achievement data, based on item response theory (IRT) scaling with marginal estimation, was developed originally

More information

Lecture 16: Mixtures of Generalized Linear Models

Lecture 16: Mixtures of Generalized Linear Models Lecture 16: Mixtures of Generalized Linear Models October 26, 2006 Setting Outline Often, a single GLM may be insufficiently flexible to characterize the data Setting Often, a single GLM may be insufficiently

More information

Default Priors and Effcient Posterior Computation in Bayesian

Default Priors and Effcient Posterior Computation in Bayesian Default Priors and Effcient Posterior Computation in Bayesian Factor Analysis January 16, 2010 Presented by Eric Wang, Duke University Background and Motivation A Brief Review of Parameter Expansion Literature

More information

Bayesian Methods for Machine Learning

Bayesian Methods for Machine Learning Bayesian Methods for Machine Learning CS 584: Big Data Analytics Material adapted from Radford Neal s tutorial (http://ftp.cs.utoronto.ca/pub/radford/bayes-tut.pdf), Zoubin Ghahramni (http://hunch.net/~coms-4771/zoubin_ghahramani_bayesian_learning.pdf),

More information

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features Yangxin Huang Department of Epidemiology and Biostatistics, COPH, USF, Tampa, FL yhuang@health.usf.edu January

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Computer Emulation With Density Estimation

Computer Emulation With Density Estimation Computer Emulation With Density Estimation Jake Coleman, Robert Wolpert May 8, 2017 Jake Coleman, Robert Wolpert Emulation and Density Estimation May 8, 2017 1 / 17 Computer Emulation Motivation Expensive

More information

Bayesian Linear Models

Bayesian Linear Models Bayesian Linear Models Sudipto Banerjee 1 and Andrew O. Finley 2 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. 2 Department of Forestry & Department

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

Generalized Linear Latent and Mixed Models with Composite Links and Exploded

Generalized Linear Latent and Mixed Models with Composite Links and Exploded Generalized Linear Latent and Mixed Models with Composite Links and Exploded Likelihoods Anders Skrondal 1 and Sophia Rabe-Hesketh 2 1 Norwegian Institute of Public Health, Oslo (anders.skrondal@fhi.no)

More information

Factor Analytic Models of Clustered Multivariate Data with Informative Censoring (refer to Dunson and Perreault, 2001, Biometrics 57, )

Factor Analytic Models of Clustered Multivariate Data with Informative Censoring (refer to Dunson and Perreault, 2001, Biometrics 57, ) Factor Analytic Models of Clustered Multivariate Data with Informative Censoring (refer to Dunson and Perreault, 2001, Biometrics 57, 302-308) Consider data in which multiple outcomes are collected for

More information

Bayesian Analysis of Latent Variable Models using Mplus

Bayesian Analysis of Latent Variable Models using Mplus Bayesian Analysis of Latent Variable Models using Mplus Tihomir Asparouhov and Bengt Muthén Version 2 June 29, 2010 1 1 Introduction In this paper we describe some of the modeling possibilities that are

More information

Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study

Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study Modeling and Interpolation of Non-Gaussian Spatial Data: A Comparative Study Gunter Spöck, Hannes Kazianka, Jürgen Pilz Department of Statistics, University of Klagenfurt, Austria hannes.kazianka@uni-klu.ac.at

More information

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness

A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A Bayesian Nonparametric Approach to Monotone Missing Data in Longitudinal Studies with Informative Missingness A. Linero and M. Daniels UF, UT-Austin SRC 2014, Galveston, TX 1 Background 2 Working model

More information

A Marginal Maximum Likelihood Procedure for an IRT Model with Single-Peaked Response Functions

A Marginal Maximum Likelihood Procedure for an IRT Model with Single-Peaked Response Functions A Marginal Maximum Likelihood Procedure for an IRT Model with Single-Peaked Response Functions Cees A.W. Glas Oksana B. Korobko University of Twente, the Netherlands OMD Progress Report 07-01. Cees A.W.

More information

39th Annual ISMS Marketing Science Conference University of Southern California, June 8, 2017

39th Annual ISMS Marketing Science Conference University of Southern California, June 8, 2017 Permuted and IROM Department, McCombs School of Business The University of Texas at Austin 39th Annual ISMS Marketing Science Conference University of Southern California, June 8, 2017 1 / 36 Joint work

More information

Multivariate Survival Analysis

Multivariate Survival Analysis Multivariate Survival Analysis Previously we have assumed that either (X i, δ i ) or (X i, δ i, Z i ), i = 1,..., n, are i.i.d.. This may not always be the case. Multivariate survival data can arise in

More information

Basic Sampling Methods

Basic Sampling Methods Basic Sampling Methods Sargur Srihari srihari@cedar.buffalo.edu 1 1. Motivation Topics Intractability in ML How sampling can help 2. Ancestral Sampling Using BNs 3. Transforming a Uniform Distribution

More information

Semiparametric Generalized Linear Models

Semiparametric Generalized Linear Models Semiparametric Generalized Linear Models North American Stata Users Group Meeting Chicago, Illinois Paul Rathouz Department of Health Studies University of Chicago prathouz@uchicago.edu Liping Gao MS Student

More information

Modelling geoadditive survival data

Modelling geoadditive survival data Modelling geoadditive survival data Thomas Kneib & Ludwig Fahrmeir Department of Statistics, Ludwig-Maximilians-University Munich 1. Leukemia survival data 2. Structured hazard regression 3. Mixed model

More information

A general mixed model approach for spatio-temporal regression data

A general mixed model approach for spatio-temporal regression data A general mixed model approach for spatio-temporal regression data Thomas Kneib, Ludwig Fahrmeir & Stefan Lang Department of Statistics, Ludwig-Maximilians-University Munich 1. Spatio-temporal regression

More information

STAT 518 Intro Student Presentation

STAT 518 Intro Student Presentation STAT 518 Intro Student Presentation Wen Wei Loh April 11, 2013 Title of paper Radford M. Neal [1999] Bayesian Statistics, 6: 475-501, 1999 What the paper is about Regression and Classification Flexible

More information

Introduction to Statistical Analysis

Introduction to Statistical Analysis Introduction to Statistical Analysis Changyu Shen Richard A. and Susan F. Smith Center for Outcomes Research in Cardiology Beth Israel Deaconess Medical Center Harvard Medical School Objectives Descriptive

More information

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution

MH I. Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution MH I Metropolis-Hastings (MH) algorithm is the most popular method of getting dependent samples from a probability distribution a lot of Bayesian mehods rely on the use of MH algorithm and it s famous

More information

CTDL-Positive Stable Frailty Model

CTDL-Positive Stable Frailty Model CTDL-Positive Stable Frailty Model M. Blagojevic 1, G. MacKenzie 2 1 Department of Mathematics, Keele University, Staffordshire ST5 5BG,UK and 2 Centre of Biostatistics, University of Limerick, Ireland

More information

COS513 LECTURE 8 STATISTICAL CONCEPTS

COS513 LECTURE 8 STATISTICAL CONCEPTS COS513 LECTURE 8 STATISTICAL CONCEPTS NIKOLAI SLAVOV AND ANKUR PARIKH 1. MAKING MEANINGFUL STATEMENTS FROM JOINT PROBABILITY DISTRIBUTIONS. A graphical model (GM) represents a family of probability distributions

More information

Basic IRT Concepts, Models, and Assumptions

Basic IRT Concepts, Models, and Assumptions Basic IRT Concepts, Models, and Assumptions Lecture #2 ICPSR Item Response Theory Workshop Lecture #2: 1of 64 Lecture #2 Overview Background of IRT and how it differs from CFA Creating a scale An introduction

More information

Bayesian linear regression

Bayesian linear regression Bayesian linear regression Linear regression is the basis of most statistical modeling. The model is Y i = X T i β + ε i, where Y i is the continuous response X i = (X i1,..., X ip ) T is the corresponding

More information

Bayesian Inference in the Multivariate Probit Model

Bayesian Inference in the Multivariate Probit Model Bayesian Inference in the Multivariate Probit Model Estimation of the Correlation Matrix by Aline Tabet A THESIS SUBMITTED IN PARTIAL FULFILMENT OF THE REQUIREMENTS FOR THE DEGREE OF Master of Science

More information

Discussion of Missing Data Methods in Longitudinal Studies: A Review by Ibrahim and Molenberghs

Discussion of Missing Data Methods in Longitudinal Studies: A Review by Ibrahim and Molenberghs Discussion of Missing Data Methods in Longitudinal Studies: A Review by Ibrahim and Molenberghs Michael J. Daniels and Chenguang Wang Jan. 18, 2009 First, we would like to thank Joe and Geert for a carefully

More information

General structural model Part 2: Categorical variables and beyond. Psychology 588: Covariance structure and factor models

General structural model Part 2: Categorical variables and beyond. Psychology 588: Covariance structure and factor models General structural model Part 2: Categorical variables and beyond Psychology 588: Covariance structure and factor models Categorical variables 2 Conventional (linear) SEM assumes continuous observed variables

More information

Whats beyond Concerto: An introduction to the R package catr. Session 4: Overview of polytomous IRT models

Whats beyond Concerto: An introduction to the R package catr. Session 4: Overview of polytomous IRT models Whats beyond Concerto: An introduction to the R package catr Session 4: Overview of polytomous IRT models The Psychometrics Centre, Cambridge, June 10th, 2014 2 Outline: 1. Introduction 2. General notations

More information

GENERALIZED LATENT TRAIT MODELS. 1. Introduction

GENERALIZED LATENT TRAIT MODELS. 1. Introduction PSYCHOMETRIKA VOL. 65, NO. 3, 391 411 SEPTEMBER 2000 GENERALIZED LATENT TRAIT MODELS IRINI MOUSTAKI AND MARTIN KNOTT LONDON SCHOOL OF ECONOMICS AND POLITICAL SCIENCE In this paper we discuss a general

More information

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) =

σ(a) = a N (x; 0, 1 2 ) dx. σ(a) = Φ(a) = Until now we have always worked with likelihoods and prior distributions that were conjugate to each other, allowing the computation of the posterior distribution to be done in closed form. Unfortunately,

More information

Review. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 770: Categorical Data Analysis

Review. Timothy Hanson. Department of Statistics, University of South Carolina. Stat 770: Categorical Data Analysis Review Timothy Hanson Department of Statistics, University of South Carolina Stat 770: Categorical Data Analysis 1 / 22 Chapter 1: background Nominal, ordinal, interval data. Distributions: Poisson, binomial,

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Part 2: One-parameter models

Part 2: One-parameter models Part 2: One-parameter models 1 Bernoulli/binomial models Return to iid Y 1,...,Y n Bin(1, ). The sampling model/likelihood is p(y 1,...,y n ) = P y i (1 ) n P y i When combined with a prior p( ), Bayes

More information

Analysing geoadditive regression data: a mixed model approach

Analysing geoadditive regression data: a mixed model approach Analysing geoadditive regression data: a mixed model approach Institut für Statistik, Ludwig-Maximilians-Universität München Joint work with Ludwig Fahrmeir & Stefan Lang 25.11.2005 Spatio-temporal regression

More information

A Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models

A Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models Journal of Data Science 8(2010), 43-59 A Nonparametric Approach Using Dirichlet Process for Hierarchical Generalized Linear Mixed Models Jing Wang Louisiana State University Abstract: In this paper, we

More information

Stat 5101 Lecture Notes

Stat 5101 Lecture Notes Stat 5101 Lecture Notes Charles J. Geyer Copyright 1998, 1999, 2000, 2001 by Charles J. Geyer May 7, 2001 ii Stat 5101 (Geyer) Course Notes Contents 1 Random Variables and Change of Variables 1 1.1 Random

More information

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j

Standard Errors & Confidence Intervals. N(0, I( β) 1 ), I( β) = [ 2 l(β, φ; y) β i β β= β j Standard Errors & Confidence Intervals β β asy N(0, I( β) 1 ), where I( β) = [ 2 l(β, φ; y) ] β i β β= β j We can obtain asymptotic 100(1 α)% confidence intervals for β j using: β j ± Z 1 α/2 se( β j )

More information

Bayesian hierarchical space time model applied to highresolution hindcast data of significant wave height

Bayesian hierarchical space time model applied to highresolution hindcast data of significant wave height Bayesian hierarchical space time model applied to highresolution hindcast data of Presented at Wave workshop 2013, Banff, Canada Erik Vanem Introduction The Bayesian hierarchical space-time models were

More information

Bayesian Regression (1/31/13)

Bayesian Regression (1/31/13) STA613/CBB540: Statistical methods in computational biology Bayesian Regression (1/31/13) Lecturer: Barbara Engelhardt Scribe: Amanda Lea 1 Bayesian Paradigm Bayesian methods ask: given that I have observed

More information

Bayesian inference for sample surveys. Roderick Little Module 2: Bayesian models for simple random samples

Bayesian inference for sample surveys. Roderick Little Module 2: Bayesian models for simple random samples Bayesian inference for sample surveys Roderick Little Module : Bayesian models for simple random samples Superpopulation Modeling: Estimating parameters Various principles: least squares, method of moments,

More information

The linear model is the most fundamental of all serious statistical models encompassing:

The linear model is the most fundamental of all serious statistical models encompassing: Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x

More information

Fitting Narrow Emission Lines in X-ray Spectra

Fitting Narrow Emission Lines in X-ray Spectra Outline Fitting Narrow Emission Lines in X-ray Spectra Taeyoung Park Department of Statistics, University of Pittsburgh October 11, 2007 Outline of Presentation Outline This talk has three components:

More information

Factor Analysis and Latent Structure of Categorical Data

Factor Analysis and Latent Structure of Categorical Data Factor Analysis and Latent Structure of Categorical Data Irini Moustaki Athens University of Economics and Business Outline Objectives Factor analysis model Literature Approaches Item Response Theory Models

More information

Generalized logit models for nominal multinomial responses. Local odds ratios

Generalized logit models for nominal multinomial responses. Local odds ratios Generalized logit models for nominal multinomial responses Categorical Data Analysis, Summer 2015 1/17 Local odds ratios Y 1 2 3 4 1 π 11 π 12 π 13 π 14 π 1+ X 2 π 21 π 22 π 23 π 24 π 2+ 3 π 31 π 32 π

More information

Research Projects. Hanxiang Peng. March 4, Department of Mathematical Sciences Indiana University-Purdue University at Indianapolis

Research Projects. Hanxiang Peng. March 4, Department of Mathematical Sciences Indiana University-Purdue University at Indianapolis Hanxiang Department of Mathematical Sciences Indiana University-Purdue University at Indianapolis March 4, 2009 Outline Project I: Free Knot Spline Cox Model Project I: Free Knot Spline Cox Model Consider

More information

Making rating curves - the Bayesian approach

Making rating curves - the Bayesian approach Making rating curves - the Bayesian approach Rating curves what is wanted? A best estimate of the relationship between stage and discharge at a given place in a river. The relationship should be on the

More information

Nonparametric Bayes Inference on Manifolds with Applications

Nonparametric Bayes Inference on Manifolds with Applications Nonparametric Bayes Inference on Manifolds with Applications Abhishek Bhattacharya Indian Statistical Institute Based on the book Nonparametric Statistics On Manifolds With Applications To Shape Spaces

More information

Using Estimating Equations for Spatially Correlated A

Using Estimating Equations for Spatially Correlated A Using Estimating Equations for Spatially Correlated Areal Data December 8, 2009 Introduction GEEs Spatial Estimating Equations Implementation Simulation Conclusion Typical Problem Assess the relationship

More information

Bayesian Linear Regression

Bayesian Linear Regression Bayesian Linear Regression Sudipto Banerjee 1 Biostatistics, School of Public Health, University of Minnesota, Minneapolis, Minnesota, U.S.A. September 15, 2010 1 Linear regression models: a Bayesian perspective

More information

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random

Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Calibration Estimation of Semiparametric Copula Models with Data Missing at Random Shigeyuki Hamori 1 Kaiji Motegi 1 Zheng Zhang 2 1 Kobe University 2 Renmin University of China Institute of Statistics

More information

Bayesian Nonparametric Rasch Modeling: Methods and Software

Bayesian Nonparametric Rasch Modeling: Methods and Software Bayesian Nonparametric Rasch Modeling: Methods and Software George Karabatsos University of Illinois-Chicago Keynote talk Friday May 2, 2014 (9:15-10am) Ohio River Valley Objective Measurement Seminar

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

Bayesian Inference. Chapter 9. Linear models and regression

Bayesian Inference. Chapter 9. Linear models and regression Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering

More information

Reconstruction of individual patient data for meta analysis via Bayesian approach

Reconstruction of individual patient data for meta analysis via Bayesian approach Reconstruction of individual patient data for meta analysis via Bayesian approach Yusuke Yamaguchi, Wataru Sakamoto and Shingo Shirahata Graduate School of Engineering Science, Osaka University Masashi

More information

Wageningen Summer School in Econometrics. The Bayesian Approach in Theory and Practice

Wageningen Summer School in Econometrics. The Bayesian Approach in Theory and Practice Wageningen Summer School in Econometrics The Bayesian Approach in Theory and Practice September 2008 Slides for Lecture on Qualitative and Limited Dependent Variable Models Gary Koop, University of Strathclyde

More information

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering

9/12/17. Types of learning. Modeling data. Supervised learning: Classification. Supervised learning: Regression. Unsupervised learning: Clustering Types of learning Modeling data Supervised: we know input and targets Goal is to learn a model that, given input data, accurately predicts target data Unsupervised: we know the input only and want to make

More information

Joint Modeling of Longitudinal Item Response Data and Survival

Joint Modeling of Longitudinal Item Response Data and Survival Joint Modeling of Longitudinal Item Response Data and Survival Jean-Paul Fox University of Twente Department of Research Methodology, Measurement and Data Analysis Faculty of Behavioural Sciences Enschede,

More information

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation

Biost 518 Applied Biostatistics II. Purpose of Statistics. First Stage of Scientific Investigation. Further Stages of Scientific Investigation Biost 58 Applied Biostatistics II Scott S. Emerson, M.D., Ph.D. Professor of Biostatistics University of Washington Lecture 5: Review Purpose of Statistics Statistics is about science (Science in the broadest

More information

Deep Poisson Factorization Machines: a factor analysis model for mapping behaviors in journalist ecosystem

Deep Poisson Factorization Machines: a factor analysis model for mapping behaviors in journalist ecosystem 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm

IEOR E4570: Machine Learning for OR&FE Spring 2015 c 2015 by Martin Haugh. The EM Algorithm IEOR E4570: Machine Learning for OR&FE Spring 205 c 205 by Martin Haugh The EM Algorithm The EM algorithm is used for obtaining maximum likelihood estimates of parameters when some of the data is missing.

More information

Bayesian Hypothesis Testing in GLMs: One-Sided and Ordered Alternatives. 1(w i = h + 1)β h + ɛ i,

Bayesian Hypothesis Testing in GLMs: One-Sided and Ordered Alternatives. 1(w i = h + 1)β h + ɛ i, Bayesian Hypothesis Testing in GLMs: One-Sided and Ordered Alternatives Often interest may focus on comparing a null hypothesis of no difference between groups to an ordered restricted alternative. For

More information

CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA

CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA Examples: Multilevel Modeling With Complex Survey Data CHAPTER 9 EXAMPLES: MULTILEVEL MODELING WITH COMPLEX SURVEY DATA Complex survey data refers to data obtained by stratification, cluster sampling and/or

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

Bayesian data analysis in practice: Three simple examples

Bayesian data analysis in practice: Three simple examples Bayesian data analysis in practice: Three simple examples Martin P. Tingley Introduction These notes cover three examples I presented at Climatea on 5 October 0. Matlab code is available by request to

More information

Item Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions

Item Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions R U T C O R R E S E A R C H R E P O R T Item Parameter Calibration of LSAT Items Using MCMC Approximation of Bayes Posterior Distributions Douglas H. Jones a Mikhail Nediak b RRR 7-2, February, 2! " ##$%#&

More information

CSC 2541: Bayesian Methods for Machine Learning

CSC 2541: Bayesian Methods for Machine Learning CSC 2541: Bayesian Methods for Machine Learning Radford M. Neal, University of Toronto, 2011 Lecture 4 Problem: Density Estimation We have observed data, y 1,..., y n, drawn independently from some unknown

More information

Simulating Random Variables

Simulating Random Variables Simulating Random Variables Timothy Hanson Department of Statistics, University of South Carolina Stat 740: Statistical Computing 1 / 23 R has many built-in random number generators... Beta, gamma (also

More information

Default Priors and Efficient Posterior Computation in Bayesian Factor Analysis

Default Priors and Efficient Posterior Computation in Bayesian Factor Analysis Default Priors and Efficient Posterior Computation in Bayesian Factor Analysis Joyee Ghosh Institute of Statistics and Decision Sciences, Duke University Box 90251, Durham, NC 27708 joyee@stat.duke.edu

More information

Modelling Operational Risk Using Bayesian Inference

Modelling Operational Risk Using Bayesian Inference Pavel V. Shevchenko Modelling Operational Risk Using Bayesian Inference 4y Springer 1 Operational Risk and Basel II 1 1.1 Introduction to Operational Risk 1 1.2 Defining Operational Risk 4 1.3 Basel II

More information

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Tihomir Asparouhov 1, Bengt Muthen 2 Muthen & Muthen 1 UCLA 2 Abstract Multilevel analysis often leads to modeling

More information

Partial factor modeling: predictor-dependent shrinkage for linear regression

Partial factor modeling: predictor-dependent shrinkage for linear regression modeling: predictor-dependent shrinkage for linear Richard Hahn, Carlos Carvalho and Sayan Mukherjee JASA 2013 Review by Esther Salazar Duke University December, 2013 Factor framework The factor framework

More information

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions

Pattern Recognition and Machine Learning. Bishop Chapter 2: Probability Distributions Pattern Recognition and Machine Learning Chapter 2: Probability Distributions Cécile Amblard Alex Kläser Jakob Verbeek October 11, 27 Probability Distributions: General Density Estimation: given a finite

More information

Modeling conditional distributions with mixture models: Theory and Inference

Modeling conditional distributions with mixture models: Theory and Inference Modeling conditional distributions with mixture models: Theory and Inference John Geweke University of Iowa, USA Journal of Applied Econometrics Invited Lecture Università di Venezia Italia June 2, 2005

More information

Learning Bayesian network : Given structure and completely observed data

Learning Bayesian network : Given structure and completely observed data Learning Bayesian network : Given structure and completely observed data Probabilistic Graphical Models Sharif University of Technology Spring 2017 Soleymani Learning problem Target: true distribution

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information