A multivariate model for ordinal trait analysis

Similar documents
Introduction to QTL mapping in model organisms

Introduction to QTL mapping in model organisms

Multiple QTL mapping

Introduction to QTL mapping in model organisms

THE data in the QTL mapping study are usually composed. A New Simple Method for Improving QTL Mapping Under Selective Genotyping INVESTIGATION

Gene mapping in model organisms

Introduction to QTL mapping in model organisms

Introduction to QTL mapping in model organisms

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences

Mapping multiple QTL in experimental crosses

On a multivariate implementation of the Gibbs sampler

Sample size calculations for logistic and Poisson regression models

Prediction of the Confidence Interval of Quantitative Trait Loci Location

Statistical issues in QTL mapping in mice

Bayes methods for categorical data. April 25, 2017

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

GENERALIZED LINEAR MIXED MODELS: AN APPLICATION

Mapping multiple QTL in experimental crosses

Bayesian Multivariate Logistic Regression

Multiple interval mapping for ordinal traits

A new simple method for improving QTL mapping under selective genotyping

Lecture 8. QTL Mapping 1: Overview and Using Inbred Lines

QTL model selection: key players

Combining dependent tests for linkage or association across multiple phenotypic traits

Binary trait mapping in experimental crosses with selective genotyping

Estimation of Parameters in Random. Effect Models with Incidence Matrix. Uncertainty

Longitudinal random effects models for genetic analysis of binary data with application to mastitis in dairy cattle

MODEL-FREE LINKAGE AND ASSOCIATION MAPPING OF COMPLEX TRAITS USING QUANTITATIVE ENDOPHENOTYPES

QTL Model Search. Brian S. Yandell, UW-Madison January 2017

Multivariate Versus Multinomial Probit: When are Binary Decisions Made Separately also Jointly Optimal?

Generalized Linear Models for Non-Normal Data

Analysis of variance, multivariate (MANOVA)

SYLLABUS Short course in EVOLUTIONARY QUANTITATIVE GENETICS 30 Jan 10 Feb, 2012, University of Uppsala

Lecture 4: Types of errors. Bayesian regression models. Logistic regression

Normal distribution We have a random sample from N(m, υ). The sample mean is Ȳ and the corrected sum of squares is S yy. After some simplification,

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Stat 5101 Lecture Notes

Supplementary Materials for Molecular QTL Discovery Incorporating Genomic Annotations using Bayesian False Discovery Rate Control

R/qtl workshop. (part 2) Karl Broman. Biostatistics and Medical Informatics University of Wisconsin Madison. kbroman.org

COMPOSITIONAL IDEAS IN THE BAYESIAN ANALYSIS OF CATEGORICAL DATA WITH APPLICATION TO DOSE FINDING CLINICAL TRIALS

Online Appendix to: Marijuana on Main Street? Estimating Demand in Markets with Limited Access

Generalized, Linear, and Mixed Models

Lecture 9. QTL Mapping 2: Outbred Populations

QTL mapping of Poisson traits: a simulation study

Supplemental Information Likelihood-based inference in isolation-by-distance models using the spatial distribution of low-frequency alleles

Nonparametric Modeling of Longitudinal Covariance Structure in Functional Mapping of Quantitative Trait Loci

Research Division Federal Reserve Bank of St. Louis Working Paper Series

STA 216, GLM, Lecture 16. October 29, 2007

Default Priors and Effcient Posterior Computation in Bayesian

Overview. Background

. Also, in this case, p i = N1 ) T, (2) where. I γ C N(N 2 2 F + N1 2 Q)

Lecture 9. Short-Term Selection Response: Breeder s equation. Bruce Walsh lecture notes Synbreed course version 3 July 2013

PQL Estimation Biases in Generalized Linear Mixed Models

Anumber of statistical methods are available for map- 1995) in some standard designs. For backcross populations,

The Admixture Model in Linkage Analysis

A NOTE ON ROBUST ESTIMATION IN LOGISTIC REGRESSION MODEL

Multinomial Logistic Regression Models

Tutorial Session 2. MCMC for the analysis of genetic data on pedigrees:

Lecture 11: Multiple trait models for QTL analysis

STA6938-Logistic Regression Model

A Fully Nonparametric Modeling Approach to. BNP Binary Regression

QTL Mapping I: Overview and using Inbred Lines

Marginal Specifications and a Gaussian Copula Estimation

Detection of multiple QTL with epistatic effects under a mixed inheritance model in an outbred population

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

Lecture 3. G. Cowan. Lecture 3 page 1. Lectures on Statistical Data Analysis

Generalized Linear Models (GLZ)

Quantitative Genomics and Genetics BTRY 4830/6830; PBSB

Bayesian networks with a logistic regression model for the conditional probabilities

Determining the number of components in mixture models for hierarchical data

Part 6: Multivariate Normal and Linear Models

Lecture 1 Basic Statistical Machinery

Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3

Fast Bayesian Methods for Genetic Mapping Applicable for High-Throughput Datasets

25 : Graphical induced structured input/output models

Kneib, Fahrmeir: Supplement to "Structured additive regression for categorical space-time data: A mixed model approach"

STAT 536: Genetic Statistics

Repeated ordinal measurements: a generalised estimating equation approach

Chapter 1 Statistical Inference

Charles E. McCulloch Biometrics Unit and Statistics Center Cornell University

Quantitative Genomics and Genetics BTRY 4830/6830; PBSB

MOST complex traits are influenced by interacting

Statistical Distribution Assumptions of General Linear Models

Answers to Problem Set #4

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Latent Variable Models for Binary Data. Suppose that for a given vector of explanatory variables x, the latent

Methods for QTL analysis

Using Estimating Equations for Spatially Correlated A

Biostat 2065 Analysis of Incomplete Data

EPSY 905: Fundamentals of Multivariate Modeling Online Lecture #7

Use of hidden Markov models for QTL mapping

On the limiting distribution of the likelihood ratio test in nucleotide mapping of complex disease

STATISTICAL ANALYSIS WITH MISSING DATA

Model Based Statistics in Biology. Part V. The Generalized Linear Model. Chapter 16 Introduction

Bayesian Methods for Machine Learning

On the mapping of quantitative trait loci at marker and non-marker locations

Bustamante et al., Supplementary Nature Manuscript # 1 out of 9 Information #

EM Algorithm II. September 11, 2018

STAT 425: Introduction to Bayesian Analysis

Chapter 2. Review of basic Statistical methods 1 Distribution, conditional distribution and moments

Transcription:

(00) 9, 09 1 & 00 Nature Publishing Group All rights reserved 0018-0X/0 $000 wwwnaturecom/hdy 1 Department of Botany and Plant Sciences, University of California, Riverside, CA 91, USA Many economically important characteristics of agricultural crops are measured as ordinal traits Statistical analysis of the genetic basis of ordinal traits appears to be quite different from regular quantitative traits The generalized linear model methodology implemented via the Newton Raphson algorithm offers improved efficiency in the analysis of such data, but does not take full advantage of the extensive theory developed in the linear model arena Instead, we develop a multivariate model for ordinal trait analysis and implement an EM algorithm for parameter estimation We also propose a method for calculating the variance-covariance matrix of the estimated parameters The EM equations turn out to be extremely similar to formulae seen in standard linear model analysis Computer simulations are performed to validate the EM algorithm A real data set is analyzed to demonstrate the application of the method The advantages of the EM algorithm over other methods are addressed Application of the method to QTL mapping for ordinal traits is demonstrated using a simulated baclcross (BC) population (00) 9, 09 1 doi:10108/sjhdy80088; published online 1 August 00 Keywords: EM algorithm; generalized linear model; ordinal traits; QTL; truncated normal distribution Introduction Many disease resistance traits in plants are scored in several ordered categories based on the magnitude of the disease symptoms For example, this approach has been used for sheath blight resistance in rice (Zou et al, 000), clubroot resistance in Brassica napus (Manzanares- Dauleux et al, 000) and cucumber mosaic virus resistance in pepper (Caranta et al, 00) Similarly, many characters in animals and humans are expressed as binary or ordinal traits, including the score for calving difficulty, expression of congential malformations, numbers of reproductive events and so on In other cases traits are actually continuously distributed, but, for technical reasons, measured as ordinal traits Special statistical methods are required to analyze traits measured on an ordinal scale (McCullagh and Nelder, 1989) The probability model of McIntyre et al (001) used the trait penetrances directly as the genetic parameters of interest The method can only be applied to QTL mapping for binary traits A generalized linear model is currently considered to be the most appropriate for ordinal data analysis method because, using a simple link function, we can adopt theory and methodology developed extensively in linear model for continuously distributed data The key to using a generalized linear model is the use of a hypothetic continuous latent variable (known as liability) The observed categorical phenotype depends on whether the liability exceeds one Correspondence: S Xu, Department of Botany and Plant Sciences, University of California, Riverside, CA 91, USA E-mail: xu@geneticsucredu The work was done when Dr Chenwu Xu was a postdoctoral research associate at UC Riverside He has returned to China 1 Current address: Department of Agronomy, Yangzhou University, Yangzhou, Jiangshu Province, PR China Received 18 December 00; accepted 1 July 00; published online 1 August 00 or more of an ordered threshold This generalized linear model is, therefore, also called the threshold model (Lynch and Walsh, 1998) Many statistical methods of estimation and hypothesis testing have been developed under the threshold model They include a maximum likelihood method (Aitchison and Silvery, 19; Ashford, 199) and a Bayesian method (Albert and Chib, 199; Sorensen et al, 199) Under the maximum likelihood framework, parameters are often estimated iteratively via the Fisher scoring algorithm or the Newton Raphson ridge algorithm (Ashford, 199) When the Bayesian method is applied, the posterior means or modes of parameters are often inferred from a posterior sample generated from a Markov chain Monte Carlo process (Sorensen et al, 199) The Bayesian method is more versatile than the ML method because it can handle more complicated models However, the ML method is more cost effective because no MCMC sampling is required Both methods are, therefore, currently being used in ordinal data analysis A thorough description of the statistical methods for ordinal data analysis may be found in McCullagh and Nelder (1989) and Fahrmeir and Tutz (199) When parameters are estimated using the Bayesian method via MCMC sampling, realizations of the latent variable are sampled from its conditional posterior distribution (a truncated normal distribution if the probit link function is used) Once the latent variable has been sampled, the problem of parameter estimation becomes that of parameter estimation in the usual linear model In this way, the latent variable is treated as a missing value in the Bayesian analysis For binary data, Xu et al (00) developed an EM algorithm to search for the MLE of parameters, also by treating the latent variable as a missing value Xu et al (00a, b) recently extended the EM algorithm for binary data to handle ordinal traits and multiple binary traits Compared to the Fisher scoring or the Newton Raphson

10 algorithm, the EM algorithm has the following desirable properties: (1) it is numerically stable, in stark contrast to the Newton Raphson algorithm that crashes easily when the thresholds are not well separated; () the steps of EM iterations are transparent and intuitive, and thus easily understood by biologists; () the EM algorithm takes full advantage of the results developed in the usual linear model analysis Unfortunately, the EM algorithm also has two undesirable properties: (1) the convergence process may be slow and () it does not automatically provide an estimate of the variance covariance matrix of the parameters The first undesirable property is no longer a problem, thanks to the ever-growing computing power The second problem has been circumvented by using the Louis (198) information matrix for EM Xu et al (00) developed the information matrix of parameters for binary trait analysis, which were extended recently by Xu et al (00a) to handle ordinal traits However, the method of Xu et al (00a) estimates parameters in two steps: estimating the regression coefficients conditional on the thresholds and estimating the thresholds conditional on the regression coefficients They called the method ECM algorithm (expectation and conditional maximization, Meng and Rubin (199)) Although the ECM method is a convenient approach for finding the ML solution, the information matrix of the parameters are hard to derive Therefore, Xu et al (00a) did not provide an estimate of the variance covariance matrix In this study, we propose an EM algorithm for parameter estimation in a single step, that is, solving for the thresholds and regression coefficients simultaneously As a consequence, the variance covariance matrix of the parameters can be found on the basis of the Louis information matrix Statistical models Univariate model Let s j be the ordinal data observed for subject j, 8j ¼ 1, y, n, where n is the sample size Let C be the number of ordered categories in the data set Variable s j is defined as S j ¼ k if j belongs to category k, 8k ¼ 1, y, C A set of fixed thresholds, t 1, t, y, t C 1, on an underlying scale define the observed categories on the ordinal scale Further define y j as an underlying latent variable for individual j The relationship between the latent variable and the thresholds is S j ¼ k if t k 1 oy 1 rt k where t 0 ¼ N and t C ¼ N Here, we define C þ 1 thresholds but only C 1 of them are parameters of interest and these thresholds are denoted by a vector t ¼ [t 1 t y t C 1 ] T For notational convenience, let m ¼ C 1 so that t is an m 1 vector The natural choice for the distribution of y is the normal distribution, under which the model is called the probit model The latent variable is described by the following linear model y j ¼ X j b þ e j ð1þ where b is a p 1 vector for the model effects, X j is a 1 p known design matrix, and e j is a residual error assumed to be N(0, 1) distributed This model is commonly used in ordinal data analysis, especially in QTL mapping for ordinal traits (Hackett and Weller, 199; Rao and Xu, 1998; Xu et al, 00a) It is called the univariate model or the threshold model As mentioned in the introduction, the variance covariance matrix of the estimated regression coefficients is hard to derive under the EM algorithm, although estimation of the regression coefficients themselves is relatively straightforward (Xu et al, 00a) Multivariate model An alternative model for ordinal traits is the so-called multivariate model or cumulative threshold model, in which we formulate the ordinal trait analysis as a multivariate problem The ordinal trait with C categories can be described by C 1 binary traits Each binary trait is controlled by its own latent variable with its own threshold These binary observations are defined as w jk ¼ 0 ifs j k ðþ 1 otherwise For each binary trait, we define a latent variable y jk that is linked to w jk by w jk ¼ 0 for y jk rt k and w jk ¼ 1 for y jk t k So, each subject is described by a vector w j ¼ [w j1 w j y w jm ] T which is controlled by a vector y j ¼ [y j1 y j y y jm ] T The relationship between S j under the univariate model and w j under the multivariate model is shown in Table 1 for the case where C ¼ The multivariate liabilities are described by the following multivariate linear model, y j1 1 0 0 y j y jm 0 1 0 t ¼ 0 0 1 t m x j1 x j x jp b 1 x j1 x j x jp b þ þ x j1 x j x jp b p In matrix notation, we have y j ¼ I m t þð1 m X j Þb þ e j ðþ where I m is a m m identity matrix, 1 m is an m 1 vector with all elements equal to 1 and # represents the Kronecker matrix product Let h ¼ t//b be an (m þ p) 1 vector for the parameters and Z j ¼ ( I m ) (1 m #X j )bean m (m þ p) design matrix, where the symbols // and represent vertical and horizontal matrix concatenations, respectively, a notation adopted from SAS/IML Table 1 Relationship between s j under the univariate model and w j under the multivariate model for C ¼ Univariate (s j ) Multivariate (w j ) t 1 e j1 e j e jm ðþ w j1 w j w j w j w j 1 0 0 0 0 0 1 0 0 0 0 1 1 0 0 0 1 1 1 0 0 1 1 1 1 0 1 1 1 1 1

(SAS Institute, 1999a) The above model (Equation ()) can be rewritten as y j ¼ Z j h þ e j ðþ The residual errors are assumed to be distributed as an m-dimensional independent normal, that is, e j BN m (0, I m ) Parameter estimation Under the multivariate model, we have formulated the thresholds as a subset of the regression coefficients As a result, we are able to estimate the entire parameter vector (including the thresholds and the original regression coefficients) simultaneously in a single step If we treat the liability vector as data, we have the following complete-data log likelihood LðhÞ ¼ 1 X n ðy j Z j hþ T ðy j Z j hþ Note that the variance covariance matrix of the residual errors is an identity matrix (constant), and thus it does not play a role in the maximum likelihood analysis In addition, this log likelihood function is not the observed likelihood function, which is a function of w j with y j integrated out (the actual form of the likelihood function is dealt with in the discussion) The EM algorithm requires maximization of el(h), the expected L(h), with respect to the parameters because the latent variable y is missing (not observable) The expectation is taken with respect to y conditional on the parameters (h) and the observed data (w), and has the following form, elðhþ ¼E½LðhÞjw; h ðtþ Š ¼ 1 X n h i E ðy j Z j hþ T ðy j Z j hþjw j ; h ðtþ ðþ The parameter values, however, are the quantities to be estimated To calculate the conditional expectation, we need to choose an arbitrary value of h from its legal domain to start the maximization process Once h ¼ h (t) is chosen for, t ¼ 0, we can use h (t) to compute the conditional expectation of the complete-data log likelihood (still a function of h) We then maximize the conditional expectation of the log likelihood (Equation ()) with respect to the parameters and obtain the following EM iteration formula for h, 1 h ðtþ1þ ¼ Xn Z T j Z X n j Z T j Eðy j jw j; h ðtþ Þ ð8þ where Eðy j1 jw j1 ; h ðtþ Þ Eðy j jw j ; h ðtþ Þ¼ Eðy jm jw jm ; h ðtþ Þ ð9þ is an m 1 vector for the conditional expectation of y j given the parameter values at the tth iteration and the ðþ observed ordinal trait The kth element of the above vector is Eðy jk jw jk ; h ðtþ Þ 8 >< ¼ t ðtþ k X j b ðtþ Þ þ X j b ðtþ þ fðtðtþ k FðX j b ðtþ t ðtþ k Þ for w jk ¼ 1 ð10þ t ðtþ k þ X j b ðtþ fðtðtþ k X j b ðtþ Þ >: Fðt ðtþ k X j b ðtþ Þ for w jk ¼ 0 where f(x) andf(x) represent the standardized normal density function and the cumulative standardized normal distribution function, respectively Equation (10) is the expectation of a truncated normal variable (Cohen, 1991) In summary, the EM algorithm requires initialization of the parameters with h ¼ h (t) for t ¼ 0 and the following two steps: E-Step: Calculate E(y j w j, h (t) ) using Equation (10); M-Step: Update parameter h with h ¼ h (t þ 1) using Equation (8) The E- and M-Steps are repeated several times until some criterion of convergence is satisfied Let t max be the number of iterations taken for the EM algorithm to converge The MLE of h is ^h ¼ h ðt maxþ Information matrix The observed information matrix for h can be found using the method of Louis (198), which requires the first and the second partial derivatives of the complete-data log likelihood The first partial derivative is q LðhÞ ¼Sðh; yþ ¼Xn Z T j qh y Xn j Z T j Z jh ð11þ and the second partial derivative is q LðhÞ ¼Bðh; yþ ¼ Xn Z T j qh Z j ð1þ The observed information matrix of Louis (198) is IðhÞ ¼E ½ Bðh; yþš E ½Sðh; yþs T ðh; yþš y y ¼ Xn þ Xn Z T j Z j Xn þ Xn Z T j Z jhh T Xn Z T j Z j Z T j Eðy Xn jþht Z T j Z j Z T j Z jh Xn Eðy T j ÞZ j E Xn Z T j y j X n y T j Z j ð1þ where E(y j ) is a short notation for E(y j w j, h) After extensive algebraic manipulation, we can show that IðhÞ ¼E y ½ Bðh; yþš E y ½Sðh; yþs T ðh; yþš ¼ Xn Z T j ð1þ I Vðy j Þ Z j 11

1 where V(y j ) is a short notation for V(y j w j, h) the conditional variance covariance matrix of y j It is an m m diagonal matrix with the kth diagonal element defined as Vðy jk jw jk ; hþ ¼ 1 Q j0ðq j0 þ x jk Þ for w jk ¼ 0 ð1þ 1 Q j1 ðq j1 x jk Þ for w jk ¼ 1 where 8 < x jk ¼ t k X j b Q j0 ¼ fðx jk Þ=Fðx jk Þ ð1þ : Q j1 ¼ fðx jk Þ=Fð x jk Þ¼fðx jk Þ=½1 Fðx jk ÞŠ (see Cohen, 1991 for the variance of a truncated normal distribution) Hypothesis tests An analytical form of the likelihood function for the observed data is needed only if a likelihood ratio test is to be performed Instead, under the multivariate model, we are able to formulate the variance-covariance matrix for the estimated parameters by taking the inverse of the information matrix, VðhÞ I 1 ðhþ ð1þ which can be used to derive the Wald test statistic (Fahrmeir and Tutz, 199) Therefore, the likelihood ratio test statistic is no longer required The Wald test for the null hypothesis of H 0 : b ¼ 0 is W ¼ ^b T ½VðbÞŠ 1^b; ð18þ where V(b) is simply a subset of matrix V(h) Under the null hypothesis, W will asymptotically follow a w - distribution with p degrees of freedom Each individual regression coefficient can also be tested separately with ^b W k ¼ k Vðb k Þ : ð19þ Under the null hypothesis H 0 : b k ¼ 0, W k will asymptotically follow a w -distribution with one degree of freedom Interval mapping of QTL The particular reason for developing this EM algorithm was to solve the problem of QTL mapping for ordinal traits The algorithm is sufficiently general, that we have been able to present it in a more general way We now demonstrate application of the model to QTL mapping Suppose that we collect the phenotypes of n backcross (BC) progeny and genotyped all the markers for these individuals We can scan QTL along the genome using the idea of interval mapping (Lander and Botstein, 1989; Haley and Knott, 199) The model in the context of interval mapping is y j1 1 0 0 t 1 x j e j1 y j ¼ 0 1 0 t x j þ e j b þ ð0þ 0 0 1 y jm where x j is the genotype indicator variable for individual j at a putative position of the genome and it is defined as x j ¼ 1 for one genotype and x j ¼ 1 for the alternative genotype (only two genotypes are present in an BC t m x j e jm population at any particular locus) The regression coefficient b ¼ m AA m Aa is the difference between the average values of the two genotypes Variable x is not observable but its probability distribution can be inferred from genotypes of flanking markers (ie interval mapping) Two approaches can be taken to incorporate the conditional probability distribution of variable x Oneis the mixture model approach (Lander and Botstein, 1989) by treating x j as a missing value A detailed algorithm for the mixture model for binary traits has been developed by Xu et al (00), which can be directly adopted here for ordinal traits without much modification The other approach is to adopt the idea of Haley and Knott (199) who replaced x j by xˆj ¼ p j q j ¼ p j 1, the conditional expectation of x j given marker information, where p j ¼ Pr(x j ¼ 1 flanking marker) and q j ¼ Pr(x j ¼ 1 flanking marker) ¼ 1 p j To map QTL in F populations, the model should be written as y j1 1 0 0 t 1 x j1 x j y j 0 1 0 t x j1 x j " # b 1 ¼ þ b y jm 0 0 1 t m x j1 x j e j1 e j þ ð1þ e jm where x j1 ¼ { þ 1,0, 1} and x j ¼ { 1, þ 1, 1} for genotype array {AA, Aa, aa}, and b 1 ¼ a ¼ 1 (m AA m aa ) (additive effect) and b ¼ d ¼ m Aa 1 (m AA þ m aa ) (dominance effect) Let 8 < p j ðþ1þ ¼Prðx j ¼þ1jflanking markerþ p j ð0þ ¼Prðx j ¼ 0jflanking markerþ : p j ð 1Þ ¼Prðx j ¼ 1jflanking markerþ The conditional expectations of the two x variables are ^x j1 ¼ p j ðþ1þ p j ð 1Þ ^x j ¼ p j ð0þ ½p j ðþ1þþp j ð 1ÞŠ which will replace x j1 and x j in model (1) for QTL mapping in F populations Illustrations Example 1: This example shows the analysis of a simulated data set with multiple replications We simulated four explanatory variables X j ¼ [x j1 x j x j x j ] The values of the four variables for each subject were generated from a multivariate normal distribution, that is, X j : N (0, R), where 1 0: 0 0:1 0: 1 0: 0: R ¼ 0 0: 1 0 0:1 0: 0 1 The true values of the regression coefficients were b ¼½b 1 b b b Š T ¼½0: 0 0: 1:0 Š T The liability for subject j was generated by y j ¼ X j b þ e j where e j was simulated from N(0, 1) The observed

ordinal measurement was converted from y j using thresholds t ¼½t 1 t t Š T ¼½ 1: 0 1: Š T There were three thresholds (excluding t 0 ¼ N and t C ¼þN) and four ordinal categories (C ¼ ) The sample size (n) was simulated at the following four levels: 100, 0, 00 and 0 For each simulated data set, the proposed EM algorithm was used to estimate the parameters Meanwhile, the variance-covariance matrix of the EM estimates was calculated using the inverse of the Louis (198) information matrix The simulation was replicated 1000 times, from which we were able to evaluate the property of the EM algorithm Note that the data were simulated under the univariate model, but analyzed under the proposed multivariate model We demonstrated that the multivariate model is a good approximation to the univariate model For comparison, we also analyzed each data set using the SAS procedure, PROC LOGISTIC (SAS Institute, 1999b), with the logit link function replaced by the probit link function We were also able to use PROC PROBIT (SAS Institute, 1999b), but for some reason the estimated parameters had signs opposite to the ones obtained from our EM estimates The PROC LOGISTIC program uses the univariate model but finds the MLE with the Newton Raphson ridge algorithm The program also calculates the variance covariance matrix using the observed information matrix Results of the Newton Raphson algorithm are considered as exact because the data were simulated using exactly the same univariate model Results from the simulated data analysis are listed in Table for the estimated parameters and the variances of the estimated parameters Tables and give the covariances of the estimated parameters Table shows (A) the average ^h ¼½^t 1 ^t ^t ^b1 ^b ^b ^b Š T calculated across i¼1 the 1000 replicates, denoted by ^h ¼ 1 ^h, 1000 (B) the variance of ^h calculated across the 1000 replicates, denoted by Vð^hÞ ¼ 1 1000 1 i¼1 ð^h ^hþ, and (C) the average V(h) calculated across the 1000 replicates, denoted by VðhÞ ¼ 1 1000 i¼1 VðhÞ The bias of parameter estimation for each method can be evaluated by comparing ^h with the true value of h When the sample size was small, say 100 0, slight bias has been observed for each estimated parameter There is no clear trend on which method has a larger bias than the 1 Table Estimated parameters and the variances of the estimates for the EM algorithm and the Newton Raphson algorithm in the simulated data analysis (example 1) Sample size Method t 1 t t b 1 b b b 100 EM A 1 0000 181 01 0008 018 10 B 000 00 000 001 00 008 009 C 0080 00 0010 00189 009 00191 001 Newton A 19 0008 18 0 0000 0 10 B 00 009 009 00 001 00 001 C 00 001 001 00 00 00 008 0 EM A 19 000 19 0080 0001 011 10188 B 0018 0009 001 00100 001 0008 0009 C 00189 0009 0019 0009 0010 000 0008 Newton A 190 00000 100 01 000 019 109 B 00181 0009 00188 0010 001 00081 0009 C 0019 00091 0010 0010 001 0008 0009 00 EM A 1080 00019 1108 001 0000 008 1009 B 0008 000 0008 0009 0001 000 000 C 00088 000 00089 0008 0008 000 0008 Newton A 1181 00008 111 0091 000 0011 101 B 0008 000 00090 000 000 000 0008 C 0008 000 0008 0000 000 0008 000 0 EM A 10 00008 10 099 000 001 1000 B 000 0008 0008 000 0000 0000 000 C 0009 0000 0008 000 0009 000 000 Newton A 119 0000 108 001 0000 008 100 B 0009 0009 000 0009 0008 0009 000 C 000 0009 000 0001 0009 0009 000 A: The average ^h calculated across the 1000 replicates using ^h ¼ 1 1000 i¼1 1000 1 B: The sample variance of ^h calculated across the 1000 replicates using Vð^hÞ ¼ 1 i¼1 ð^h ^hþ C: The average Vð^hÞ calculated across the 1000 replicates using VðhÞ ¼ 1 1000 i¼1 VðhÞ

1 Table Covariances of estimated parameters calculated from 1000 replicates for the EM algorithm (upper triangles) and the Newton Raphson algorithm (lower triangles) in the simulated data analysis (example 1) Sample size t 1 t t b 1 b b b 100 t 1 0011 00118 000 000 0009 001 t 000 000 0001 000 000 00000 t 000 000 0009 000 000 001 b 1 0001 000 0009 0011 000 000 b 0001 0008 000 0011 001 000 b 0009 0001 0009 0008 0018 00001 b 001 0000 001 0009 0009 0001 0 t 1 0000 0000 000 0000 000 0009 t 000 000 0000 0000 00001 0000 t 000 000 0008 0000 000 0001 b 1 0001 0000 0008 000 00018 000 b 0000 0000 0000 000 0000 0001 b 000 0000 0000 000 000 0001 b 000 0000 0000 0000 00019 00011 00 t 1 00010 00018 00011 0000 0001 000 t 0001 00010 0000 0000 0000 00001 t 00011 0001 0001 0000 00009 000 b 1 00010 0000 0001 000 00009 00018 b 0000 0000 0000 000 000 0001 b 0001 0000 00008 0001 000 0000 b 000 00000 000 000 00018 00001 0 t 1 00000 0001 00008 00001 00008 0001 t 00008 00000 00000 00001 00001 00000 t 0000 00009 0000 00001 00008 0001 b 1 00008 00000 0000 0001 0000 0001 b 00001 0000 00001 0001 0001 00008 b 00008 00001 00008 0000 00019 0000 b 0001 00000 0001 0001 00011 00001 A covariance between two estimated parameters was calculated from 1000 replicates using covð^y i ; ^y j Þ¼ 1 1000 1 k¼1 ð^y i ^yi Þð^y j ^yj Þ other for the regression coefficients, but the estimated thresholds, for example, t 1, appear to have a larger bias for the Newton method than for the EM algorithm The variance of each estimated parameter calculated from the sample of 1000 replicates, Vð^hÞ, is a good indication of the precision of the estimate When the sample size was small, say 100, the Newton method tends to have a consistently smaller variance than the EM algorithm, although the difference is barely noticeable The multivariate model appears to be a good approximation of the univariate model Recall that V(h) is the variance of parameters calculated from the information matrix for each replicate and VðhÞ is the average of V(h) calculated across the 1000 replicates If VðhÞ is close to Vð^hÞ, it means that the method for calculating V(h) is reasonable When the sample size was small (n ¼ 100), the EM algorithm has a smaller VðhÞ than Vð^hÞ, but the bias goes away quickly as the sample size increases The Newton method, however, always behave well, regardless of the sample size Table gives the covariances between estimated parameters for both the EM algorithm (upper triangular elements) and the Newton method (lower triangular elements) using covð^y i ; ^y j Þ¼ 1 1000 1 k¼1 ð^y i ^yi Þð^y j ^yj Þ Again, when the sample size was small (n ¼ 100), the two methods are slightly different for the covariances, but the differences diminish quickly as the sample size increases Table lists the averages of the covariances between parameters calculated from the 1000 replicates for both the EM algorithm (upper triangular elements) and the Newton method (lower triangular elements) using 1000 covðy i ; y j Þ¼ 1 k¼1 covðy i; y j Þ The conclusion of Table also applies to Table Comparing Table with Table, we conclude that the methods for calculating cov(y i, y j ) in both the EM algorithm and the Newton method are reasonably good Example : This example shows the analysis of a real data set The data were obtained from Koch and Edwards (1988) for a double-blind clinical trial investigating a new treatment for rheumatoid arthritis In this data set, there were n ¼ 8 subjects with different ages who received an active or placebo treatment for their arthritis pain, and the subsequent extent of improvement was recorded as marked, some, or none The dependent variable was an ordinal categorical observation with three categories (1 ¼ none, ¼ some and ¼ marked) The three explanatory variables were treatment (active or placebo), sex (male or female), and age (recorded as a continuous variable), respectively The design matrix was X j ¼ [x j1 x j x j ] where the three variables in the vector correspond to treatment, sex and age, respectively We analyzed this real data set using both the proposed EM

Table Averages of covariances among parameters calculated from 1000 replicates for the EM algorithm (upper triangles) and the Newton Raphson algorithm (lower triangles) in the simulated data analysis (example 1) 1 Sample size t 1 t t b 1 b b b 100 t 1 00088 000 000 00009 001 0011 t 0008 0008 0008 000 0000 00009 t 000 000 0011 0009 000 001 b 1 000 0008 0008 001 0001 0009 b 00011 0008 0008 001 000 0010 b 0008 000 0000 0008 000 00008 b 001 0001 0019 000 001 000 0 t 1 000 000 00018 0000 0009 000 t 0008 000 00009 00011 0000 00008 t 000 0008 000 0000 0000 0000 b 1 0001 00001 000 000 000 000 b 0000 00001 0000 0009 000 000 b 0000 0000 0001 000 000 00010 b 0008 0000 000 000 000 0000 00 t 1 00011 0001 00011 00001 0001 000 t 0001 0001 0000 00001 00000 00001 t 0001 0001 0001 0000 0000 000 b 1 0001 0000 0001 0001 0001 0001 b 0000 0000 0000 000 000 0001 b 0001 0000 0000 0001 000 00001 b 000 0000 000 000 00018 00001 0 t 1 0000 00010 00009 00000 00009 0001 t 00008 00008 00000 00001 00001 00000 t 00011 00008 00009 0000 00008 0001 b 1 00009 00001 00008 0001 0000 0001 b 00000 00001 00001 0000 00018 0001 b 00009 00001 00008 0000 00019 0000 b 0001 00001 0001 0001 00010 0000 The average of a covariance was calculated using covðy i ; y j Þ¼ 1 1000 k¼1 covðy i; y j Þ Table Estimated parameters and their variances for the EM algorithm and the Newton Raphson algorithm in the real data analysis (example ) Method Statistic t 1 t b 1 b b EM ^h 189 88 1000 0199 00 Var(h) 0 000 008 0098 00001 Newton ^h 910 8 1089 09 00 Var(h) 0 0 0089 0099 00001 algorithm and the Newton method implemented in the PROC LOGISTIC procedure of SAS Table gives the estimated parameters and their variances and Table give the covariances of the estimated parameters for both methods The estimated parameters are very similar for the two methods The variances of parameter estimates are different for the two methods The EM algorithm produced smaller variances than the Newton method, implying that the sample size (n ¼ 8) was not sufficiently large for the EM to provide accurate estimates for the variances, although the estimates of parameters are remarkably close to those of the Newton method Example : This example shows the analysis of a simulated BC population for interval mapping of Wald test 110 100 90 80 0 0 0 0 0 0 10 0 0 10 0 0 0 0 0 0 80 90 100 Genome (cm) Figure 1 Wald test statistic profile for the simulated chromosome The true position of the QTL was cm quantitative trait loci This simulation serves as a working example to demonstrate the method for QTL mapping We assumed that the liability has a zero mean and a unit residual variance A single QTL was placed at position cm of a 100 cm long chromosome covered by

1 Table Covariances of estimated parameters for the EM algorithm (upper traingles) and the Newton Raphson algorithm (lower triangles) obtained from the real data analysis (example ) t 1 t b 1 b b t 1 09 000 00 0009 t 0 00 0008 0000 b 1 0090 00 0011 00000 b 000 009 0019 00000 b 0000 0001 00001 00001 Table Estimated parameters (column ) and their covariance matrix (columns, upper triangular elements) at position cm of the simulated chromosome (example ) Estimate t 1 t t t b t 1 1009 0010 0000 0000 0000 00010 t 08 0008 000010 0000 0000 t 0 000 000019 00000 t 11 00118 000000 b 0811 000 11 evenly placed markers The effect of the QTL was a ¼ 00, which explained h ¼ a /(a þ 1) ¼ 0% of the liability variance (see Xu et al (00b) for the definition of h ) We simulated five ordered categories (C ¼ ) with four threshold values The four thresholds were chosen by trial and error so that the expected frequencies of the five categories occurring in the BC population had a ratio of 1::::1 The threshold values generating this ratio were t ¼½ 1:9 0:9 0:9 1:9 Š T The population size was n ¼ 00 Only one sample was generated and analyzed as a working example to demonstrate the method The chromosome was scanned from one end to the other with a one cm increment Figure 1 shows the Wald test statistic profile across the genome The peak of the test statistic profile occurs at position cm (the true position was cm) The estimated QTL parameters at position cm are given in Table, which are quite close to the true values Table also gives the covariance matrix Discussion The log likelihood function given in Equation () is the complete-data log likelihood function, which is only used to derive the EM algorithm The actual observed log likelihood function that can be used for deriving the likelihood ratio test statistic must have the latent variable y integrated out Such an observed log likelihood function has the following form, L o ðhþ ¼ Xn X C k¼1 w jk log e Fðx j b t k Þ þð1 w jk Þ log e Fðt k x j bþ : ðþ As we have proposed using the Wald test statistic for testing hypotheses, this observed log likelihood function is not required For people who prefer the likelihood ratio test, Equation () must be used The focus of this paper has been to derive the EM algorithm and the information matrix under the EM method As a result, hypothesis tests were only briefly mentioned in the manuscript and the actual tests were carried out in neither the simulation experiment nor the real data analysis Some advantages of the EM algorithm over the Newton Raphson algorithm were discussed in the introduction One of them is that the EM formulae are more transparent and intuitive to biologists who have little knowledge in advanced statistics except some basic background in linear modeling The derivations of the EM algorithm and the variance covariance matrix of the parameters are demanding to some extent, but the final results are extremely simple The estimated parameters and the variance covariance matrix of the parameters have the following expressions, 8 " # 1 " # ^h ¼ Pn P Z T j Z n >< j Z T j Eðy j Þ ( ) VðhÞ ¼ Pn h i 1 ðþ >: Z T j I Vðy j Þ Z j People will immediately recognize the similarity between this set of equations and those commonly seen in the usual linear model analysis If y j were observed variables, as in the usual linear regression analysis, then E(y j ) ¼ y j and V(y j ) ¼ 0 would hold The above equations would be 8 " # 1 " # ^h ¼ Pn P Z T j Z n j Z T j y j >< " # ðþ 1 >: VðhÞ ¼ Pn Z T j Z j which are exactly the least square estimates of the parameters and the variance covariance matrix of the estimates in the usual regression analysis The EM algorithm only requires substitutions of y j by E(y j ) and I by I V(y j ), where E(y j ) and V(y j ) are the means and variances of truncated normal variables (Cohen, 1991) This is a very desirable property of the proposed EM algorithm The exact form of the variance covariance matrix, V(y j ), is still unknown We ignored the covariance elements (assumed to be zero) of the matrix and used a diagonal approximation This approximation has caused biased (downward) estimates for the variances of the regression coefficients when the sample size was small As a result, the Wald test statistic is biased upward Therefore, the test statistic under the null model may not follow the assumed w -distribution In QTL mapping, however, the exact form of the distribution for the test statistic is not important because the critical value of the test statistic used to declare statistical significance is often drawn from a permutation test (Churchill and Doerge, 199) Therefore, the slightly biased Wald test will not alter the conclusion of QTL mapping relative to an unbiased test Developing a Bayesian method for ordinal traits is straightforward If an uninformative prior is assigned to the parameters, the conditional posterior distribution

of h given y is multivariate normal with mean ^h and variance covariance matrix V(h) Given h sampled from the normal distribution, y jk is a truncated normal variable with mean t k þ x j b and variance one Standard algorithms for sampling a truncated normal variable are available (Devroye, 198) Given the fact that the Bayesian method implemented via the MCMC is so simple, how do we justify the need of an EM algorithm? The reason is twofold, the cost effectiveness and the prevention of MCMC error Acknowledgements The research was supported by the National Institute of Health Grant R01-GM1 and the National Science Foundation Grant DBI-00 to SX References Aitchison J, Silvery SD (19) The generalization of probit analysis to the case of multiple responses Biometrika : 11 10 Albert JH, Chib S (199) Bayesian analysis of binary and polychotomous response data J Am Stat Assoc 88: 9 9 Ashford JR (199) An approach to the analysis of data for semiquantal responses in biology response Biometrics 1: 81 Caranta C, Plieger S, Lefebvre V, Daubeze AM, Thabuis A, Palloix A (00) QTLs involved in the restriction of cucumber mosaic virus (CMV) long-distance movement in pepper Theor Appl Genet 10: 8 91 Cohen AC (1991) Truncated and Censored Samples Marcel Dekker Inc: New York Churchill GA, Doerge RW (199) Empirical threshold values for quantitative trait mapping Genetics 18: 9 91 Devroye L (198) Non-Uniform Random Variable Generation Springer-Verlag: New York Fahrmeir L, Tutz G (199) Multivariate Statistical Modelling Based on Generalized Linear Models Spring-Verlag Inc: New York Hackett CA, Weller JI (199) Genetic mapping of quantitative trait loci for traits with ordinal distributions Biometrics 1: 1 1 Haley CS, Knott SA (199) A simple regression method for mapping quantitative trait loci in line crosses using flanking markers 9: 1 Koch G, Edwards S (1988) Clinical efficiency trials with categorical data In: Peace KE (ed) Biopharmaceutical Statistics for Drug Development Marcel Dekker: New York pp 0 1 Lander ES, Botstein D (1989) Mapping Mendelian factors underlying quantitative traits using RFLP linkage maps Genetics 11: 18 199 Louis TA (198) Finding the observed information matrix when using the EM algorithm J Roy Statist Soc B : Lynch M, Walsh B (1998) Genetics and Analysis of Quantitative Traits Sinauer Associates: Sunderland, MA Manzanares-Dauleux MJ, Delourme R, Baron F, Thomas G (000) Mapping of one major gene and of QTLs involved in resistance to clubroot in Brassica napus Theor Appl Genet 101: 88 891 McCullagh P, Nelder JA (1989) Generalized Linear Models Chapman and Hall: New York McIntyre LM, Coffman CJ, Doerge RW (001) Detecting and localization of a single binary trait locus in experimental populations Genet Res 8: 9 9 Meng XL, Rubin DB (199) Maximum likelihood estimation via the ECM algorithm Biometrika 80: 8 Rao S, Xu S (1998) Mapping quantitative trait loci for ordered categorical traits in four-way crosses 81: 1 SAS Institute (1999a) SAS/IML User s Guide, Version 8 SAS Institute Inc: Cary, North Carolina SAS Institute (1999b) SAS/SATA User s Guide, Version 8 SAS Institute Inc: Cary, North Carolina Sorensen DA, Andersen S, Gianola D, Korsgaard I (199) Bayesian inference in threshold models using Gibbs sampling Genet Sel Evol : 9 9 Xu C, Li Z, Xu S (00a) Joint mapping of quantitative trait loci for multiple binary characters Genetics 19: 10 109 Xu C, Zhang YM, Xu S (00b) An EM algorithm for mapping quantitative resistance loci 9: 119 18 Xu S, Yi N, Burke D, Galecki A, Miller RA (00) An EM algorithm for mapping binary disease loci: application to fibrosarcoma in a four-way cross mouse family Genet Res 8: 1 18 Zou J, Pan XB, Chen ZX, Xu JY, Lu JF, Zhai WX et al (000) Mapping quantitative trait loci controlling sheath blight resistance in two rice cultivars Theor Appl Genet 101: 9 1