Hierarchical Linear Models and Structural Equation Modelling for the Children of Siblings model

Size: px
Start display at page:

Download "Hierarchical Linear Models and Structural Equation Modelling for the Children of Siblings model"

Transcription

1 U.U.D.M. Project Report 2008:20 Hierarchical Linear Models and Structural Equation Modelling for the Children of Siblings model Ralf Kuja-Halkola Examensarbete i matematisk statistik, 30 hp Handledare: Paul Lichtenstein, Medicinsk epidemiologi och biostatistik, Karolinska Institutet Examinator: Ingemar Kaj December 2008 Department of Mathematics Uppsala University

2

3 Hierarchical Linear Models and Structural Equation Modelling for the Children of Siblings model Ralf Kuja-Halkola Abstract The main aim of this paper is to examine and apply the Children of Siblings model to normally distributed outcome data. The model refers to siblings who are mothers/aunts and/or fathers/uncles and on whose children an outcome variable is measured. This is done through two statistical models. First a Hierarchical Linear Model (also known as multi-level model) which takes into account clustering of data in three levels. The levels correspond to increasing sizes of a family: individual, nuclear family and extended family. This model makes it possible to examine whether a specific exposure is responsible for a change in the outcome variable after confounders and clustering have been considered. Furthermore the genetical relatedness is taken into account making it possible to draw conclusions regarding the familial effects on the association between exposure and outcome. The second model is a 2-level Structional Equation Model which aims at splitting up variance/covariance in three different predetermined components representing genetic influence, environmental similarity and environmental dissimilarity using two levels of clustering in the data. The clustering levels are within- and between- nuclear families, which also are considered hierarchical. This model investigates the importance of genetic heritability contra environmental heritability. An example study is performed using data for smoking habits of mothers during pregnancy and the psychological functioning capacity, a prognosis of the ability to cope with war-time stress, of their children as measured by Försvarsverket during the medical examination for military service. We found that smoking during pregnancy does not affect the psychological functioning capacity. The exposures effect on the outcome has familial confounding, and this is mainly explained through genetical relatedness 1

4 Contents 1 Introduction and aims Background Quasi-experimental study The Children of Siblings model Methods - Theory Hierarchical Linear Models CoS - a Hierarchical Linear Model The submodels included in the CoS Interaction terms Estimation and degrees of freedom Structural Equation Models What is Structural Equation Models? Matrix notation approach Model fit level SEM, ACE model level SEM for ACE Estimation: Maximum Likelihood using Expectation Maximization 21 3 Data The data The variables Results The problem of belonging to multiple families HLM results SEM results Discussion Limitations of the model Future improvements References 35 2

5 1 Introduction and aims The aim of this paper is to make the Children of Siblings (CoS) model accessible for researchers, both in theory and in practice. By using this model it is possible to draw conclusions from data that can be influenced by familial confounding. The confounding can be due to genetics or environment, and the model also tries to distinguish the importance of these two sources. An example, it is not unlikely that a child s reading ability is influenced by the environment in which he/she is brought up (access to books, parental encouragement). Probably his/her sibling will have a similar environment. Thus they will be more similar than when comparing to a random child taken from the same population (e.g. swedish children born a certain year). Furthermore they are genetically similar, but the similarity is different for full- and half- siblings. They will also share genes with their cousins (here the genetical similarity also differs depending on whether they are full- or half-cousins), but the environment in which cousins are brought up may differ. By considering all these (dis-)similarities we are hopefully able to isolate a specific exposure s influence on an outcome, and if there is familial confounding isolate the source of this. The outcome variable will be limited to be normally distributed. A study will be performed whether a (male) child s exposure to mother s Smoking During Pregnancy (SDP) influence his Psychological Functional capacity score (PF), as measured by Försvarsverket during the medical examination for military service (henceforth called conscript). The data concerns boys born between 1973 and 1988 (99% born between 1979 and 1988). We will try to explain the model in a conceptually simple way, and when examining the theory behind the model rather than focusing on all the small problems we try to keep it at an overviewlevel. This since a non-mathematician researcher should be able to grasp, and apply, the result. In the theory section some more advanced theoretical problems are addressed. The Background section introduces the idea of the model and indicates its applicability. The Methods - Theory section explains the models of CoS and deals with the mathematical theory behind the models and the obstacles to overcome. In the Data- Analysis- and Results-section we perform the analysis of how SDP influence the PF-score. The data is taken from a large data base called the MgrCrime data base which include information about relatedness, maternal and paternal social and economic status and birth-data. Merging of data from different sub-data bases and forming of variables are made with SAS, analyses are made with SAS and Mplus 1. Finally the Discussion section consider possible conclusions to be drawn from the anlyses. We also discuss and suggest improvements, both for the current study and for the model to be applicable to other problems. 1.1 Background Researchers in the epidemiological field often try to address causation. Does a mother s SDP influence the child s intellect? Does a father s bipolarity cause his infant child to 1 Data- and analysis- program code can be obtained from the author, raku9405@student.uu.se. 3

6 have a higher mortality rate? Does a certain lifestyle increase the risk of cardiovascular diseases? Such questions are hard to answer since the causality is confounded by the interplay between outcome and exposure; an exposed child s mother also provides genes and other environment. For example, the exposure to SDP might be a negligible part of the whole picture making the child perform poorly in e.g. a test of PF. The problem of parents providing (possibly) hazardous genes as well as the child having a likelier risk to be exposed to a hazardous environment is called Gene-Environment correlation (rge). The child is at double risk to be exposed to hazardous environment, both due to the upbringing and to her/his own genes making her/him more prone to put herself/himself in such environment 1. The causality in these studies are difficult to assess, but with designs such as sibling comparison or the Children of Twins (CoT) model there is a possibility to give some answers. The CoS model is another part of this tool-box. The model CoS extends the CoT model, which has been used for some time by researchers in e.g. epidemiology and psychology (see for example (D Onofrio, 2005; Harden et al., 2007)). By using CoS rather than CoT a greater population is made available for analysis, and using CoS rather than ordinary sibling comparison can help differentiate the possible underlying mechanisms (D Onofrio et al., In press). The CoT model, as well as the CoS model, aims at distinguishing if a risk factor effects the outcome, and whether this is environmentally- and/or genetically impelled. This is not a simple task because of the rge. In the CoS an important part consist of considering the genetic relatedness between siblings (full or half) and/or cousins (full or half) and some conclusions regarding genetical importance can, hopefully, be drawn. The approach that is used is similar to that used by Brian M. D Onofrio in (D Onofrio et al., In press; D Onofrio et al., 2008). The idea to use this design is also mentioned in (Rutter, 2007). 1.2 Quasi-experimental study The basis of statistical theory in general relies on the data being a valid statistical sample, e.g. acquired through an experimental design randomizing subjects to different exposures. For many studies, as the study that we have performed, this is not true. By treating the data as experimental a number of assumptions are made, which more easily can be violated in database studies or natural experiments (Rutter, 2007) as often used in epidemiology. By labeling the study quasi-experimental this is emphasized. Examples when then assumptions possibly will not hold are Voluntary or involuntary self-censorship. When data is collected through self-reporting the supplier of the information may be less likely to report certain variables. E.g. mothers smoking during pregnancy might feel ashamed and choose not to report this, making the data biased. 1 That the parent passing their genes also can create a hazardous environment is called passive geneenvironment correlation, that the child creates her/his own hazardous environment due to genes inherited from parents is called active gene-environment correlation (Rutter, 2007). 4

7 Sampling bias reflecting the possibility that data collected in a voluntary manner is susceptible to e.g. social bias; people of low socioeconomic status may have less incentive to supply wanted information. Often a randomized case-control design is out of the question, for example it would be unethical to randomize mothers to smoke or not to smoke during pregnancy. And randomizing parent to have a certain mental disorder and thus effecting their child is impossible. The second best option in this case is to use available data, and analyze this data as if it came from a designed experiment. Even though the researcher has to be aware of the possible pitfalls, these kind of studies fills an important role. And even if the inferences can be questionable they can at least be said to be hypotheses generating. 1.3 The Children of Siblings model Our goal is to explore the influence of an exposure on an outcome after controlling for confounders and clustering effects. The confounders are the covariates that influence the outcome and might affect the causation. In the CoS model the clusters are families and subfamilies. The extended families are clusters which consists of grandparents, their children and spouses and their grandchildren. Clusters within these clusters are the nuclear families consisting of mothers, fathers and their children. These clusters can be said to be hierarchical. Another level of the hierarchy is the individual level, consisting of the children. By approaching the model in this way it is reasonable to consider a model with three levels of hierarchicity. The CoS model is designed to detect if the (possible) influence of the exposure is due to genetics or environment. By assuming that siblings are raised equally (i.e. have the same environment in upbringing), and that cousins are not, a possibility to differentiate the variation arises. When comparing cousins any similarity in the correlation between exposure and outcome will be due to both genetic- and environmental (dis-)similarities (full cousins share of their genes and half cousins share of their genes). And when comparing siblings (0.5 shared genes) with maternal half siblings (0.25 shared genes) any detection of difference between the two groups will be (thought of to be) mainly due to genetics. The confounders are just nuisance and the in analyses and will not be investigated in depth. 2 Methods - Theory The CoS model is hierarchical by construct. To cope with this a Hierarchial Linear Model (HLM) will be applied. This model is able to incorporate the dependency structure within each cluster/family and divide the variance into level-specific parts. This way of taking care of the dependence using clusters can be thought of to take into acount unmeasured confounders apart from the measured ones included in the model (D Onofrio et al., 2008). The hierarchicity will also be incorporated into a Structural Equation Model (SEM). The SEM aims to explain the variability in the outcome variable in terms of genetical and 5

8 environmental similarity/dissimilarity. The model has two levels; one regarding children of two female siblings, within each of the siblings separately. The other between the two siblings. This model uses an approach which divides the variation into three components: The A-component which measures the genetical influence on likeness, the C-component measuring the environmental influence that makes the siblings similar and the E-component measuring the environmental influence that makes sibling different. When we are referring to the ACE-model this is the model to have in mind. 2.1 Hierarchical Linear Models The Hierarchical Linear Model (HLM) (also known as multi level model) is a way to cope with dependent and nested structures in data using clustering in many levels. The basis is an ordinary linear model. By letting e.g. the intercept vary randomly in each level the model produced represent a hierarchical structure (Raudenbush & Bryk, 2002). The variation of the intercept depends on cluster-belonging of the outcome variable. This structure allows for hierarchical nesting, e.g. in CoS the nuclear family is nested within extended families (or households) and offspring is nested within nuclear families. To conceptualize this model we imagine how the data is created. Having in mind a three level model, we try to understand the formation of the unit at the individual level. First, at the extended family level, a cluster is drawn from the extended family clusters according to some rules. The nuclear family level cluster is then drawn, and the information about the extended family level cluster is incorporated into this cluster. Finally the unit at the individual level is drawn, and this is affected by which clusters at the nuclear family- and extended family- level the unit belongs to. When comparing this unit to other units, the extended family level cluster of the unit is compared with the extended family level cluster of the unit to which it is compared, the same holds for the nuclear family level cluster CoS - a Hierarchical Linear Model The hierarchicity of the model enables a partitioning of the variance into parts corresponding to each level of the hierarchy. Let Y int denote the outcome of the i:th offspring in the n:th nuclear-family within the t:th extended family. The model without any covariates is Y int = λ 0 + e int + r nt + u t. (1) This model relies upon the assumptions that the random errors are normally distributed, e int N(0, σ 2 ), the errors at the nuclear family level are normally distributed as well, r nt N(0, τ 2 1 ), as are the errors at the extended family level u t N(0, τ 2 2 ). We interpret the equation (1) as: The outcome Y int equals a grand mean λ 0 and includes three random effects: e int - a random effect at the individual level for individual i, r nt - a random effect at the nuclear level for nuclear family n where the individual i is included and u t - a random effect at the extended family level for extended family t including the nuclear family n and individual i. Another way to describe the random parts of (1) is that the variation of the 6

9 offspring from the nuclear-family mean is measured by e int, the variation of the nuclearfamily from the extended family mean is accounted for in r nt and the variation for the extended family from the overall mean is found in u t. At each level there is a possibility to include covariates which are thought to have an impact on the outcome and therefore confounding the causality. At the individual level (level 1) consider a total of p covariates, and the simple linear model Y int = π 0nt + p π jnt (α jint ) + e int, (2) j=1 where the meaning of the notation π jnt (α jint ) is explained below. To include the other levels random errors as in (1), let the intercept π 0nt vary randomly while including covariates specific to the nuclear family level (level 2). With q covariates at the nuclear family level we obtain q π 0nt = β 0t + β jt (χ jnt ) + r nt. (3) j=1 Similarly, the extended family level (level 3) terms are introduced by letting the intercept vary randomly once again and including covariates that are similar for the whole extended family. Using the linear model with s covariates at the extended family level, β 0t = λ 0 + s λ j (ξ jt ) + u t. (4) j=1 By combining (2) - (4) and using the fact that only the intercepts, not the slopes, vary between clusters (i.e. π jnt = π j, n, t, similar for β), the model is Y int = λ 0 + p π j (α jint ) + j=1 q β j (χ jnt ) + j=1 s λ j (ξ jt ) + e int + r nt + u t. (5) Assumptions made are that Cov(e int, r nt ) = 0, Cov(e int, u t ) = 0 and Cov(r nt, u t ) = 0, i.e. all random effects are uncorrelated with each other. The variables α jint, χ knt and ξ lt represent indicator variables for covariate j, k and l at level 1, 2 and 3, respectively, for fixed effect covariates. In case of a continuous effect (considered to be measured without error) they represent a centered version of the varibles, creating contrast codes. E.g. the model containing (only) one regression coefficient π 1 associated with α 1int = continuous variable X 1int would look like Y int = λ 0 + π 1 (X 1int X 1.nt ) + e int + r nt + u t. But if the continuous variable were on the level 2, e.g. χ 1nt = X 1nt, the centering would look like Y int = λ 0 +β 1 (X 1nt X 1.t )+e int +r nt +u t. The reason for the cluster-level centering is that possible bias from correlation between the covariate and random effects will be (approximately) removed (Neuhaus & McCulloch, 2006), furthermore this contrast-coding approach yields parameters equivalent to fixed effects (Greene, 2003). There are possibilities of letting the slopes vary randomly as well, but in this application they are not, see (Raudenbush & 7 j=1

10 Bryk, 2002) for a more complete treatment of the HLM models. As can be seen the model in (5) is a linear mixed model, with a certain design of the random- and fixed-effects. The HLM design is reflected in the covariance matrix. For example let the outcome vector Y = (Y 111, Y 211, Y 121, Y 221, Y 112, Y 122 ) T be all the data from the three level HLM used in CoS. In this small example the first extended family consists of two nuclear families with two children in each and the second extended family exists of two nuclear families with one child in each. The covariance matrix have the blocked structure of (6). Cov(Y ) = σ 2 + τ1 2 + τ2 2 τ1 2 + τ2 2 τ2 2 τ τ1 2 + τ2 2 σ 2 + τ1 2 + τ2 2 τ2 2 τ τ2 2 τ2 2 σ 2 + τ1 2 + τ2 2 τ1 2 + τ τ2 2 τ2 2 τ1 2 + τ2 2 σ 2 + τ1 2 + τ σ 2 + τ1 2 + τ2 2 τ τ2 2 σ 2 + τ1 2 + τ2 2 (6) In the covariance matrix (6) the different families (nuclear and extended) can be found through looking at the covariance parameters of r 11, r 21, r 12, r 22 (τ 2 1 ) and u 1, u 2 (τ 2 2 ). Let us turn the attention to the different level of the covariates. We will describe each level, and exemplify with variables used in the study in this paper. The individual level, level 1, includes the outcome variable and confounders specific to the individual, e.g. a mother s SDP, a mother s age at childbirth and a child s age when measuring the outcome variable. The level 2 is the nuclear family level, variables common to a nuclear family belongs here. Examples are socioeconomic status for mother/father, whether the siblings are full or half and the cohabitation of parents. The level 3 is the extended family level, the variables common for the whole extended family is counted to this level, an example is whether the aunts are full- or half-siblings. In figure 1 a schematical illustration of the three levels of hirachicity is found. The outcome is measured at level 1, and the ρ 1 represents that the aunts have some correlation because of shared genes and environment. The ρ 2 points out that correlation between siblings is present, this time also due to genes and environment, while the ρ 3 represents a correlation between cousins that is (thought of to be) mainly due to genetics. The ρ 4 indicates correlation between half siblings, which is unlike that of ρ 2 in terms of genetic similarity. There are more correlations which have not been indicated in the figure. All of these correlations of exposure, confounding variables and outcome has to be taken into account as rigorous as possible and the CoS model deals with this through clustering and submodelling. 8

11 F M M F ρ 1 F M F Level 3 Level 2 M ρ 2 M F M ρ 4 M ρ 3 Level 1 Figure 1: An illustration of the different levels, M=male F= female The submodels included in the CoS The ingenuity of the CoS model lies in taking good care of the information about genetical similarity/dissimilarity in the individuals included in the data. This will be addressed in a number of submodels. The first model, M 1, is an ordinary HLM with no consideration of genetics or covariates, M 1 performs a simple test of any correlation between the exposure and the outcome. The next considered model, M 2, where all confounders are included, is an ordinary way of testing for significance for the exposure. But the idea of the CoS model is that there is a possibility that significance of exposure in model M 2 can be confounded by non-independence between e.g. cousins due to partly shared genetics. In model M 3 siblings are compared with siblings and cousins with cousins regarding the exposure. This means that the similarity of siblings and cousins are addressed, but there is no differentiation between environmental and genetical effects. A mean across all cousins included in an extended family supplies an unrelated comparison. In models M 4 (cousins) and M 5 (siblings) consideration will be taken to whether the cousins (for M 4 )/siblings (for M 5 ) are full or half. These models uses subsamples which include two sisters and one of their children in M 4 or two children per mother in M 5, where possible. The subsamples are indicated with superscripts (c) for cousins in M 4 and (s) for siblings in M 5. We favour the inclusion of mothers/sisters with diverse exposure where there is a possibilty of important comparisons. For example if a mother in M 5 has more than two children and of these just one has been exposed we will pick the one with exposure and randomly sample one of the others. In M 4 and M 5 the interaction term between the exposure-variable and the full-/half- sibling/cousin will be included in the models. This term measures the difference between the differentially exposed sibling/cousin in the full/half pairs yielding a genetically/environmentally informative result. A description of the submodels in equation (7) 9

12 - (11) below uses the exposure variable SDP, from the study performed, as an example, the confounders are unspecified though. Model 1, M 1 - An ordinary HLM, examines the magnitude of association between SDP (=0 or 1) and PF using the entire sample. Y int = λ 0 + π 1 (SDP int ) + e int + r nt + u t. (7) Model 2, M 2 - As above but controlling for covariates/confounders. p q s Y int = λ 0 +π 1 (SDP int )+ π j (α jint )+ β j (χ jnt )+ λ j (ξ jt )+e int +r nt +u t. (8) j=2 j=1 Model 3, M 3 - This model compares cousins (with cousins) and siblings (with siblings) who are differentially exposed to SDP, in these calculations we have included contrast coding. p Y int =λ 0 + π 1 (SDP int SDP.nt ) + π j (α jint ) + β 1 (SDP.nt SDP..t ) + j=2 q β j (χ jnt ) + j=2 The contrast codes are calculated as follows: j=1 s λ j (ξ jt ) + e int + r nt + u t. For siblings: the average SDP for a mother over all her pregnancies is subtracted from the SDP for each son, thus comparing siblings differently exposed to SDP within nuclear family. For cousins: the average SDP for sisters over all their pregnancies is subtracted from the mean SDP for each nuclear family, comparing nuclear families differently exposed to SDP, between nuclear families; within extended families. Model 4, M 4 - Compares cousins whose parents are full siblings with those who have parents that are half siblings to further examine the aspect of genetic relatedness. For this a subsample is created, including two adult siblings, and one of their children, within each extended family (where possible). The same approach to contrast coding as above is used, recalculated for the new subset, and including a full-/half-cousin variable (values 0 and 1). The interaction term between the contrast for smoking and the full-/half-sibling variable now estimates the difference between cousin pair types. j=1 Y (c) int = λ 0 + β 1 (SDP (c).nt SDP (c)..t ) + + q β j (χ jnt ) + j=2 p π j (α jint ) j=1 s λ j (ξ jt ) + e int + r nt + u t. j=1 (9) (10) 10

13 Model 5, M 5 - Compares full- and half- siblings. As above, but with sibling pairs and their types. Y (s) int = λ 0 + π 1 (SDP (s) int SDP(s).nt) + + q β j (χ jnt ) + j=1 p π j (α jint ) j=2 s λ j (ξ jt ) + e int + r nt + u t. j=1 (11) Of course each model is estimated separately, meaning that for example the grand mean λ 0 will (probably) have different values in each model. We have chosen to include all confounders in our model, and by stepwise elimination remove the non-significant ones, each step eliminating the least significant confounder. Another way to pick confounders to be in the models is to test for correlation prior to fitting the big models and thus come up with the risk factors/confounders (as in (D Onofrio et al., In press)). The contrast coding is of great importance. Since the mean over all cousins in an extended family is included we can find an estimate for unrelated individuals in each model. The differences when comparing within extended- or nuclear family are found in the contrast. E.g. siblings in a family in M 3 might have a high mean SDP, and also a lower mean result in PF, the contrast allows us to examine whether the non-sdp siblings are different from the SDP sibling within this family. By doing this we can eliminate the main family effect, and make statements regarding the comparisons of individuals with similar starting position Interaction terms When deciding which, if any, interaction terms to be included in the model there are a couple of things to be aware of. First, interaction between levels seems unfeasible (Montgomery, 2005) because all levels of factors at the lower level is possible not present for every higher level. In our analysis the interaction between the contrast coded SDP-variable and the full or half sibling/cousin-variable must be there because it is of a main interest for the conclusions. From (D Onofrio et al., In press) (as mentioned this is one of the papers using the CoS model which we are using as a template) it can be found that none of the interaction terms between confounders has been taken into consideration. This approach is also favorable when dealing with large datasets to lower the time for computation. Another way to approach this is to test for all possible interactions and stepwise eliminate the non-significant ones. Or perhaps the middle-way; pick interaction terms to be included using knowledge about them. No matter which approach to be taken an informed decision has to be made. 11

14 2.1.4 Estimation and degrees of freedom The mixed model of the type in (5) is to be estimated using SAS:s proc mixed. Proceeding from this to estimation of the parameters π, β and λ and the variance parameters is done using the General Linear Mixed Model. We rewrite the equation (5) in matrix form Y = Xβ + Zu + ε. (12) The X are the values of the fixed effects in β, the values in Z are the dummy variables which indicates which random effects in u that are of interest and ε are the random errors corresponding to the e int :s. To illustrate the transition from (5) to (12) we use the outcome vector Y = (Y 111, Y 211, Y 121, Y 221, Y 112, Y 122 ) T as we did when illustrating the covariance matrix (6). For each level of the hierarchy we include two effects, one fixed with two levels and one continuous. The β-vector specifying the covariates is β (10 1) = (λ 0, π 1 (1), π 1 (2), π 2, β 1 (1), β 1 (2), β 2, λ 1 (1), λ 1 (2), λ 2 ) T. (13) With made-up values for the fixed effect covariates π 1, β 1 and λ 1 and centered versions for the continuous covariates π 2, β 2 and λ 2 we have (X 2111 X 2.11 ) 1 0 (X 211 X 2.1 ) 1 0 (X 21 X 2. ) (X 2211 X 2.11 ) 1 0 (X 211 X 2.1 ) 1 0 (X 21 X 2. ) X (6 10) = (X 2121 X 2.21 ) 0 1 (X 221 X 2.1 ) 1 0 (X 21 X 2. ) (X 2221 X 2.21 ) 0 1 (X 221 X 2.1 ) 1 0 (X 21 X 2. ). (14) (X 2112 X 2.12 ) 1 0 (X 212 X 2.2 ) 0 1 (X 22 X 2. ) (X 2122 X 2.22 ) 1 0 (X 222 X 2.2 ) 0 1 (X 22 X 2. ) Within each of the two extended families we have two nuclear families, this yields a random effects vector u (6 1) = (r 11, r 21, r 12, r 22, u 1, u 2 ) T. (15) The specifications are dependent on which family each individual belongs to Z (6 6) = (16) The random errors vector is simply ε (6 1) = (e 111, e 211, e 121, e 221, e 112, e 122 ) T. (17) Following the notation in (Littell et al., 2006), let u N(0, G), ε N(0, R) and also Cov(u, ε) = 0 is an assumption. Proceeding from this to write a joint normal distribution 12

15 for u and ε 1 f(u, ε) = (2π) (n+g)/2 G 0 1/2 0 R ( exp 1 [ u 2 y Xβ Zu ] T [ G R 1 ] [ u y Xβ Zu ] ) (18) with n = the sample size and g = the number of elements in u. The joint distribution f(u, ε) can be thought of as a joint likelihood. But since the randomness in Zu is present the maximation of this (quasi-) likelihood would yield an incorrect solution, the estimation method starts out here and then refines the estimation. By maximizing the distribution (likelihood) with respect to β and u solutions β and ũ can be obtained for β and u, (see (Littell et al., 2006) for details) via partial derivatives the mixed models equations can be found: [ X T R 1 X X T R 1 Z Z T R 1 Z Z T R 1 Z + G 1 ] [ β ũ ] = [ X T R 1 X Z T R 1 Z ]. (19) This can be solved as [ ] [ ] [ ] β (X = T V 1 X) 1 X T V 1 y (X T ũ GZ T V ( 1 y X(X T V 1 X) 1 X T V 1 y ) V 1 X) 1 X T V 1 y = GZ T V (y 1 X β ) (20) where V = V ar(y) = ZGZ T +R. We have ignored the possibility of needing to deal with generalized inverses (to which there are solutions which are implemented in proc mixed.) These equations can of course be solved, but it can be time consuming to find the inverse of V and the estimation ũ of u is dependent on the estimation of β. Through a method SWEEP the time is considerably lowered, for explanation of the method see for example (Smith & Graser, 1986). The default method of estimation using SAS:s proc mixed for solutions to the HLM:s is Restricted Maximum Likelihood (REML). The reason for using REML rather than ordinary Maximum Likelihood (ML) is that ML tends to bias the estimation of the standard errors negatively (Littell et al., 2006). This results in smaller confidence intervals and in the end hypotheses tests that are more likely to be overly liberal. The Minimum Variance Quadratic Unbiased Estimators (MIVQUE) procedure is non-iterative, in contrast to ML and REML, which finds the quadratic variance/covariance estimators that comes as close as analytically possible to satisfying the Lehmann-Scheffé criterion for uniformly minimum variance (Slanger, 1996). This procedure provide unbiased estimations of the variance parameters. The reason for choosing MIVQUE over REML is that sometimes convergence can be hard to obtain using REML or ML and also to save computing time when handling large data sets 1. Sometimes there can be some trouble with convergence due to problems involving one or more confounders, a way to approach this problem is to use MIVQUE 1 In the current study the data sets for M 1, M 2 and M 3 can be considered large, and convergence sometimes fails using REML. 13

16 during the stepwise elimination of non-significant variables to identify problem-variables, but the final analysis should be made using REML. The REML procedure uses the values of MIVQUE as a starting point. REML actually is just a estimation technique for the covariance parameters, the fixed effects β are estimated at the REML estimates of the covariance parameters. Let us start with the ordinary ML method. The ML minimizes the -2 log likelihood, let θ be some covariance parameters, y be the observed sample, from (20) we can find a solution β(θ) for the fixed effects β(θ) = (X T V (θ) 1 X) 1 X T V (θ) 1 y. (21) Thus the -2 log likelihood will be 2l(θ; y) = (n + g) ln(2π) + ln V (θ) + (y X β(θ)) T V (θ) 1 (y X β(θ)). (22) By minimizing this function with regard to θ the ML solution can be found. The ML estimators for a sample y = (y 1,..., y n ) T independent identically distributed from N(µ, σ 2 ) with unknown µ and σ 2 are ˆµ = 1 n i y i and ˆσ 2 = 1 n i (y i y) 2. The ˆσ 2 is biased by n 1 σ 2 but if the mean was known the unbiased estimator σ 2 = 1 n i (y i µ) 2 would constitute the ML estimand for σ 2. Thus the unknown mean seems to be a source of bias for the covariance parameter, the REML incorporates this fact. The REML estimation procedure can be seen as a maximation of the likelihood of transposed data. Instead of, as in ordinary ML, minimizing the -2 log likelihood of Y REML minimized the -2 log likelihood of KY where K is chosen such that E(KY ) = 0, i.e. the maximization occurs on error contrasts. A known result shown by Harville 1974 (see (Littell et al., 2006) for details) is that the restricted likelihood for REML is Thus the -2 log likelihood is 1 L R (θ) = X T X 1/2 1 1 (2π) (n p)/2 V (θ) 1/2 X T V (θ) 1 X 1/2 ( exp 1 ) 2 (y X β(θ)) T V (θ) 1 (y X β(θ)). 2l R (θ; y) = ln V (θ) +ln X T V (θ) 1 X +(y X β(θ)) T V (θ) 1 (y X β(θ))+c R (24) where the extra term ln X T V (θ) 1 X is the only difference to the ML -2 log likelihood in terms of what is minimized. The c R term is a constant term which is not affected of the choice of matrix K, and is independent of θ. The minimization of (24) with regards to θ is performed iteratively since the covariance parameters V (θ) is included non-linearly. The solution of the fixed effects are found through the estimated covariance components θ as β( θ) = (X T V ( θ) 1 X) 1 X T V ( θ) 1 y. Proc mixed uses a ridge-stabilized Newton-Raphson algorithm for for minimization of the -2 log likelihood (Littell et al., 2006). Now we turn to the problem of picking the correct degrees of freedom for the F -tests and confidence interval construction for the fixed factors in the model. Since the regular 14 (23)

17 ANOVA assumptions (e.g. the independence of outcomes and homoscedasticity) do not hold when using the mixed model a correction of the degrees of freedom has to be made. If no corrections are made the resulting confidence intervals will be too narrow, and testing of significance will be overly liberal. Thus the importance of the correct degrees of freedom can not be stressed enough. There are a couple of ways to do this, one is the Fai-Cornelius (FC) method as implemented in the satterthwaithe option in specifying the model using proc mixed. Another option is available: Kenward-Rogers which often is more suiting when the covariance structure is more complex, but in this study the covariance structure is relatively simple and thus the FC option is better (Littell et al., 2006) (Schaalje et al., 2001). The FC uses a Sattertwaite approximation of the degrees of freedom found through spectral decomposition of the covariance matrix of the estimated β (for details see (Schaalje et al., 2001)). A third option is the between-within option, this is most suitable for highly unbalanced data. The best choice for the actual study is the sattertwaithe option, but sometimes there can be data that are too unbalanced and the between-within approach can be suitable. 2.2 Structural Equation Models We will use the SEM on data for siblings and their children to examine whether an exposure s correlation to an outcome is due to genetics, environment that make siblings similar or environment that makes siblings dissimilar. The model used to do this is called the ACE model, where the A represents genetical influence, C represent environment that makes siblings similar and E represent environment that makes siblings dissimilar What is Structural Equation Models? The Structural Equation Models are models which, through mean- and covariance matrixanalyses, depict relationships between observed variables (Schumacker & Lomax, 2004). Behind this fuzzy formulation lies a complex theory mainly developed by applied researchers in psychology, psychometrics and econometrics (Hays et al., 2005). The method is closely related with factor analysis. The difference is that while factor analysis focus on the covariance matrices and through this latent (unmeasured) factors, SEM includes more possible dependencies within variables, within latent factors and between variables and latent factors. SEM is able to cope with multiple latent variables in combination with measured indicator and outcome variables using dependence between and within the latent and indicator and outcome variables. Since the dependences can be complicated researchers often choose to use a type of diagram to indicate the models. An example is shown in figure 2 where the F 1 and F 2 represent latent (unobserved) independent variables, x 1 and x 2 represent observed dependent variables and y is the outcome variable. The V 1 and V 2 imply that the latent unobservable variables are estimated with an error. Further explanation of arrows are found below. 15

18 V 1 V 2 F 1 F 2 x 1 x 2 y Figure 2: An illustration of SEM. The illustrations often use circles to indicate latent unmeasured factors and squares for measured indicator and outcome variables. By using this presentation it is easy to create a picture of whats going on, which is quite suitable if you are more interested in drawing conclusions rather than understanding the theory behind. The figure 2, for example, says that the indicator variables x 1 and x 2 and the outcome variable y all three are manifests of two unmeasured factors. The indicator variables also has a correlation. In general the model consists of two parts, the measurement model and the structural model. In the measurement model the measured variables are included, in the structural model the structures of the latent variables are described Matrix notation approach Apart from the distinction between observed- and latent variables there is also a distinction between dependent- and independent- variables. The independent variables are variables that are not influenced by any other variable and the dependent variables are variables which are influenced by other variables. The SEM can be written in matrix notation, the notation follows that of (Schumacker & Lomax, 2004). First the structural model. Let η (m 1) be the latent dependent variables, a (m 1) vector, and ξ (n 1) be the latent independent variables, a (n 1) vector. Let the relationships between the latent variables be described by B (m m) for latent dependent variables amongst themselves and let Γ (m n) describe the relation of latent independent variables to latent dependent variables. Let the equation prediction errors be ζ (m 1). Then the model is η = Bη + Γξ + ζ. (25) The covariance matrix Cov(ξ) = Φ (n n) describes the variances and covariances among the independent latent variables, and Cov(ζ) = Ψ (m m) is the covariance matrix for the latent dependent prediction equation errors. Let us turn the attention to the measurement model. For the latent dependent variables η, let Y (p 1) be the observed measures (dependent or independent) and let Λ y (p m) be the relationship between observed variables and latent 16

19 dependent variables. Let the measurement errors for Y be ε (p 1), the measurement model will be Y = Λ y η + ε. (26) In the same manner the independent latent variables are found through the measured X (q 1), and the relationship between observed variables (dependent or independent) and latent dependent variables is defined to be found in Λ x (q n). Let the measurement errors be δ (q 1), the measurement model will be X = Λ x ξ + δ. (27) The covariance matrices of ε and δ can be labeled Cov(ε) = Θ ε (p p) and Cov(δ) = Θ δ (q q) and contain the covariances between errors of the observed dependent- and independent variables respectively. These rather complex notations to describe the model include eight matrices, of which four (Φ, Ψ, Θ ε, and Θ δ ) are random covariance matrices. All matrices can be estimated and used for inferences. This is done by letting parameters of the matrices be free, fixed or constrained. Which parameters that are of interest is of course dependent on how the model is formulated, all matrices might not be of interest for a model (Schumacker & Lomax, 2004) Model fit We will use two model fit measures, Root Mean Squared Error of Approximation (RMSEA) and Comparative Fit Index (CFI). The RMSEA is defined as (Muthén, ) RMSEA = max((2f ML (ˆπ)/d 1/n), 0) G (28) where F ML is the ML fitting function for G groups, d is the number of degrees of freedom for the model, n is the number of observations and ˆπ is the ML estimate under H 0. The fitting function is F ML (π) = ln L H 0 n + ln L H 1 (29) n where L H0 is the likelihood of the fitted model and L H1 is the likelihood of a unrestricted model. The unrestriction denotes that the means and covariances are not constrained. The RMSEA is a global measure, and it improves when more variables are added to the model. A value of 0.05 or less is deemed to be acceptable (Schumacker & Lomax, 2004). The CFI is defined as CF I = 1 max(χ 2 H 0 d H0, 0) max(χ 2 H 0 d H0, χ 2 B d B, 0) where H 0 and B represents the fitted- and baseline- model respectively. The baseline model has uncorrelated outcomes with unrestricted means. As can be seen the CFI utilizes the χ 2 - distribution, this relies on the normal assumption of the outcome variable. The CFI spans 17 (30)

20 from 0 to 1, where 0 indicates no fit and 1 perfect fit. A value over 0.90 indicates a good fit (Schumacker & Lomax, 2004). The χ 2 H for the models are based on that the estimated parameters in the covariance matrix have a normal error. The difference between the observed and the hypothesed parameters squared divided by the hypothesed parameter follows a χ 2 -distribution with degrees of freedom = number of estimated parameters - 1. The actual χ 2 H-statistic is more advanced though taking into account stratification, non-independence of observations due to cluster sampling, and/or unequal probability of selection. Subpopulation analysis is also available. (Muthén & Muthén, ). When modifying the model according to findings some parameters might be excluded. The estimation procedure includes the forming of confidence intervals constructed with standard errors of the parameters. Following (Schumacker & Lomax, 2004) the rules of exclusion of parameters is based on three criteria: The parameter should be in the expected direction, be statistically different from zero and make practical sense. But Schumacker & Lomax also stresses that If a parameter is not significant but is of substantive interest, then the parameter should probably remain in the model. The actual ACE-model is based on that the parameters associated with the respective parts A, C and E must be considered as a group. Thus exclusion of just one parameter is probably unwise, if one is excluded the rest must be excluded as well. This also applies in the opposite direction, if one of the parameters is slightly non significant but the rest is not then that parameter can, and should, stay in the model level SEM, ACE model The ACE model is described in figure 3 for two sisters, their SDP and their childrens PF. The left part of the figure represents one sibling, and the right part the other. The parameters connected to the A-part (V A, b A and A cov ) models the variation in outcome due to genetics. The parameters in connection with the C-factor represent the variation due to shared environment, the environmental influence that makes siblings similar. The parameters associated with the E-factor represents the variation due to unshared environment, making siblings dissimilar. The arrows with a 1 besides them says that a parameter is (perfectly) measured by the parameter to which the arrow points. The arrows with parameters b F (F = A, B, C) besides them are regression coefficients, e.g. PF 1 is regressed on A 1 and the resulting regression coefficient is b A. The arrows pointing in two directions indicates correlation/covariation, e.g. A 1 covaries with A 2 with value A cov. 18

21 A cov C cov V A V C V E V A V C V E A 1 C 1 E 1 A 2 C 2 E SDP 1 b A b C b E SDP 2 b A b C b E PF 1 PF 2 r Figure 3: An illustration of SEM. By using constraints on the parameters we can find the value of V A, V C and V E ; the variations of the A-, C- and E-parameters. We may also be interested in b A, b C and b E ; the regression parameters of PF on the A-, C- and E-parameters. The r is the covariation/correlation between the outcome for both cousins, this might be of interest as well. The ACE-parameters are independent latent variables (the covariances between the A:s and C:s are not considered to induce dependent variables) and the SDP:s and PF:s are dependent variables. The covariance matrices Φ and Θ δ are the ones of main interest in the model, since the latent independent variables are measured by all observed variables. The constraints to put upon the model are those that arises from the assumptions for the model: 1. The genetical likeness of full siblings are twice of that for half siblings. 2. The environment that makes siblings alike are equal for siblings. The constraints also include setting covariances to zero for variables that are not considered to covary in the model. This includes only independent latent variables, so all of the information will be handled by (27) and the vectors and matrices X, Λ x, ξ, δ, Φ and Θ δ. The structural equation of interest will be X = Λ x ξ + δ SDP 1 P F 1 SDP 2 P F 2 = b A b C b E b A b C b E A 1 C 1 E 1 A 2 C 2 E 2 + δ 1 δ 2 δ 3 δ 4 (31) 19

22 And the covariance matrix for the latent independent variables is V A 0 0 A cov V C 0 0 C cov 0 Φ = 0 0 V E A cov 0 0 V A C cov 0 0 V C V E (32) The last matrix is the observed dependent variables measurement error covariance matrix Θ δ = 0 ϕ P F1 0 r (33) 0 r 0 ϕ P F2 The model indicates that SDP is measured without error, but not PF. Further restrictions will also be laid upon A cov and C cov to fit our modelling assumptions. We assume equal environment for the siblings, this means that C cov = V c. For the genetic part the full- and half- sibling induce different restrictions, full siblings share 0.5 of their genes so A covfull = 1 2 V A. Half siblings share half of the amount shared by full siblings thus A covhalf = 1 2 A covfull = 1 4 V A. The last two of the restrictions are the most informative parts, by enforcing this restriction on the model the genetic influence can be measured since this differs between the group of full- and half- siblings level SEM for ACE The idea of multiple levels can be applied to SEM. The model is then splitted up in a within- and between-part, within and between refers to within- and between- nuclear families. This does not mean that each part is estimated separatly but as in the HLM they are estimated simultaneously. We apply this approach to the ACE-model to incorporate the nuclear families into the model. The model looks like figure 4 where the upper portion of the figure represent the within part, and the lower portion the between. 20

23 SDP 1 SDP 2 w w WITHIN PF 1 PF 2 BETWEEN A cov C cov V A V C V E V A V C V E A 1 C 1 E 1 A 2 C 2 E SDP 1 b A b C b E SDP 2 b A b C b E PF 1 PF 2 r Figure 4: An illustration of the 2-level SEM describing the ACE model. We want to explain how the variation can be splitted up between the ACE-parameters and from this draw conclusions regarding environmental and genetical relevance for the influence of SDP on PF. To do this we want to incorporate the confounders in the model, one simple way to do this is to perform a regression analysis for PF on the confounders and then do the SEM-analysis on the residuals. Since the SEM aims to use the familial information we perform a simple regression analysis without using clustering. The SDPvariable is not included in the regression since this is the indicator variable in the ACE model Estimation: Maximum Likelihood using Expectation Maximization The estimation procedure uses a ML-approach with Expectation Maximization. A full description of the procedure is out of the scope of this paper, but details are given in (Muthén, ). The model incorporates that two classes are possible, full- and halfsiblings (more classes can be included if wanted). The method maximizes the likelihood, which is a mixture distribution of the separate classes. Through conditioning on data and class belonging the likelihood is approximated and maximized. 21

24 3 Data 3.1 The data The main idea of this study is to examine whether SDP influences the childs Psychological Functioning capacity PF. This is a variable that is obtained from a standardized psychological interview for Swedish boys during conscript. PF is a prognosis of the ability to cope with stress in war-time (Nilsson et al., 2001). Any conclusions made by the study concerns how the exposure of prenatal nicotine affects the susceptibility to stress in late adolescence/early adulthood. The boys are born between 1973 and 1988 and were at conscript between 1997 and The ages varies between 18 and 30, but the majority are 18 or 19 ( of , 98.1%). To be able to make correct inferences regarding the association of interest other effect must be taken into account. To cope with the rge an analysis must be made with covariates which may affect the outcome. How to pick these is always a subjective task, and prior knowledge must be used to make correct decisions regarding possible confounders to include in the model. The data used for the CoS model is taken from a large dataset called the MgrCrime data base. This data base consist of many merged Swedish data bases, a couple of which is used in this study. The data bases used includes information collected from: BRÅ (Swedish National Council for Crime Prevention), EpC (Centre for Epidemiology at the National Board of Health and Welfare), Pliktverket (National Service Administration) and SCB (Statistics Sweden). Variables of importance to this study are extracted from the data base and re-merged as suitable for the models M 1 M 5. The data used is stored in different data sets as described in figure 5. Central is the Multi-Generation Register which enables connection between children and parents/grandparents through the usage of unique identification numbers for each individual (variables originating from this registry: lopnr, idfam, idextfam, idspouse, idgrandpa, lopnrmor, lopnrfar, lopnrmormor, lopnrmorfar, lopnrfarmor, lopnrfarfar). The National Crime Register is used to find out whether the parents are convicted of a crime (crimem, crimef). The Medical Birth Register provides information concerning the birth of the children such as weight (birthweight), gestational length (pregtime), mothers age at childbirth (agem) and mothers SDP (rok1, sdp). From the National Service Administration the outcome variable PF is collected, the age at conscript (conscriptage) is also a variable of interest. To find out about socioeconomic status (seim, seif), income (incomem, incomef) and cohabitation (cohab) of parents the Population and Housing Census of 1990 is used. Educational level (edum, eduf) for parents are obtained from the Register of Education. 22

25 National Crime Register crime. Medical Birth Register rok1, birthweight, agem, pregtime. National Service Administration pf, conscriptage. Multi-Generation Register lopnr, lopnrmor, lopnrfar, lopnrmormor, lopnrmorfar, lopnrfarmor, lopnrfarfar. Sample M 1, M 2, M 3 N = Population and Housing Census 1990 sei, income, cohab. Register of Education edu. Subsample cousin comparison M 4 N = Subsample sibling comparison M 5 N = Subsample ACE data N = Figure 5: The data flow The variables agem Mother s age at childbirth, categorical variable taking on 5 values; 1: less than 20 years, 2: years, 3: years, 4: years, 5: greater than 35 years. birthnr An integer which indicates the number in the birtorder the child has. birthnr2 As birthnr but recalculated for the M 5 -subsample. birthweight all The child s birthweight, centered for unrelated comparison created from bviktbs. 23

26 birthweight cou The child s birthweight, centered for cousin comparison. birthweight sib The child s birthweight, centered for sibling comparison. bviktbs The child s birthweight. crimef A variable indicating whether the father has been convicted for a crime between 1973 and crimem A variable indicating whether the mother has been convicted for a crime between 1973 and cohab The cohabitation of parents at the time of child s birth. 0: Lives toghether with child s father,1: Mother single, 2: Not married to/divorced from- child s father (no cohabitation data),. : no information conscriptage Child s age at conscript, a categorical variable taking on 3 values; 1: less than 17.5 years, 2: years, 3: greater than 18.5 years. eduf The education for father 2004, a categorical variable taking on 7 values; 1: Primary and lower secondary school, less than 9 years of education, 2: Primary and lower secondary school, 9 years of education, 3: Upper secondary school, 2-3 years of education, 4: Post secondary school, less than 2 years of education, 5: Post secondary school/college/university, 2-5 years of education, 6: Postgraduate education, 7: Unknown. edum The education for mother 2004, a categorical variable taking on values as above. fullhalf c A variable indicating if the cousins are full- or half- cousins. fullhalf s A variable indicating if the siblings are full- or half- siblings. idextfam The index of the extended family. idfam The index of the family. idgrandpa The id-number of the man with whom the grandmother with id-number idextfam has the children included in the extended family. idspouse The index of the man with whom to the mother with id-number idfam has the child. incomef The income of the father 1990, a categorical variable taking on 4 values; 1: less than 100 kkr, 2: kkr, 3: kkr, 4: greater than 300 kkr. incomem The income of the mother 1990, values as above. lopnr The index of the child of interest. 24

27 lopnrfar The index of the father of the child. lopnrfarfar The index of the paternal grandfather of the child. lopnrfarmor The index of the paternal grandmother of the child. lopnrmor The index of the mother of the child, the same as idfam. lopnrmorfar The index of the maternal grandfather of the child. lopnrmormor The index of the maternal grandmother of the child. pf The Psychological Functioning capacity as measured by Försvarsverket at conscript, this is a discrete variable taking on integer values 1 9 and has a stipulated mean of 5 and variance of 4. The distribution is called stanine. pregtime The gestational time, a categorical variable taking on 4 values; 1: weeks, 2: weeks, 3: weeks, 4: greater than 41 weeks. rok1, sdp A dichotomous variable (= 0, 1) indicating whether the mother smoked during pregnancy. sdp cou A contrast coded variable indicating the difference of a mother s mean SDP during her pregnancies and the extended families SDP mean during pregnancies. sdp mean cou The mean value of sdp for cousins in an extended family. sdp sib A contrast coded variable indicating the difference of a mother s SDP for a single child and the mean over all her pregnancies. seif The socioeconomic status of the father 1990, this is a categorical variable taking on values; 1: Blue collar worker (In production of goods), not specially trained, 2: Blue collar worker, specially trained, 3: White collar worker, lower level, 4: White collar worker, intermediate level, 5: White collar worker, upper level or Self-employed, academic work, 6: Self-employed, 7: Uncategorized employed or no information. seim The socioeconomic status of the mother 1990, a categorical variable taking on values as above. 4 Results Since we are considering normally distributed variables we will treat the PF variable as a (proposed) N(µ, σ 2 ) variable, where µ = 5 and σ 2 = 4 are the predefined values. No major problems has been encountered by doing this, but the F -tests, and consequently the inference and conclusions of the analyses, are questionable. We will only consider main effect of the confounders, not the interactions, to save computational time and simplify 25

28 Figure 6: Left: The M 1 quantile-quantile-plot. Right: The M 3 quantile-quantile-plot interpretation. In the left part of figure 6 the quantile-quantile plot shows that even though the data approximately follows the normal distribution the discrete nature of the variable is still present after controlling for SDP and clustering as in M 1. The deviations from the normal-line must be seen as major since the sample size is so large. A second q- q-plot is included, see right part of figure 6, regarding the model M 3, the model controlled for SDP, covariates and clustering effects. The data seems to connect better to the normalquantile line, but the deviations at the ends are still major. The histogram of the scaled (i.e. they are scaled as a proposed N(0, 1) distribution) residuals of M 4 are displayed in figure 7 together with the assumed normal distribution and a fitted kernel distribution. This also shows that there are deviances from the normal assumption. Ways to deal with Figure 7: The histogram of the residuals of M 4. 26

Confounding, mediation and colliding

Confounding, mediation and colliding Confounding, mediation and colliding What types of shared covariates does the sibling comparison design control for? Arvid Sjölander and Johan Zetterqvist Causal effects and confounding A common aim of

More information

MIXED MODELS THE GENERAL MIXED MODEL

MIXED MODELS THE GENERAL MIXED MODEL MIXED MODELS This chapter introduces best linear unbiased prediction (BLUP), a general method for predicting random effects, while Chapter 27 is concerned with the estimation of variances by restricted

More information

Describing Change over Time: Adding Linear Trends

Describing Change over Time: Adding Linear Trends Describing Change over Time: Adding Linear Trends Longitudinal Data Analysis Workshop Section 7 University of Georgia: Institute for Interdisciplinary Research in Education and Human Development Section

More information

Introduction to Structural Equation Modeling

Introduction to Structural Equation Modeling Introduction to Structural Equation Modeling Notes Prepared by: Lisa Lix, PhD Manitoba Centre for Health Policy Topics Section I: Introduction Section II: Review of Statistical Concepts and Regression

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

Multilevel Modeling: A Second Course

Multilevel Modeling: A Second Course Multilevel Modeling: A Second Course Kristopher Preacher, Ph.D. Upcoming Seminar: February 2-3, 2017, Ft. Myers, Florida What this workshop will accomplish I will review the basics of multilevel modeling

More information

Specifying Latent Curve and Other Growth Models Using Mplus. (Revised )

Specifying Latent Curve and Other Growth Models Using Mplus. (Revised ) Ronald H. Heck 1 University of Hawai i at Mānoa Handout #20 Specifying Latent Curve and Other Growth Models Using Mplus (Revised 12-1-2014) The SEM approach offers a contrasting framework for use in analyzing

More information

Mixed-Models. version 30 October 2011

Mixed-Models. version 30 October 2011 Mixed-Models version 30 October 2011 Mixed models Mixed models estimate a vector! of fixed effects and one (or more) vectors u of random effects Both fixed and random effects models always include a vector

More information

An Introduction to Path Analysis

An Introduction to Path Analysis An Introduction to Path Analysis PRE 905: Multivariate Analysis Lecture 10: April 15, 2014 PRE 905: Lecture 10 Path Analysis Today s Lecture Path analysis starting with multivariate regression then arriving

More information

An Introduction to Multilevel Models. PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 25: December 7, 2012

An Introduction to Multilevel Models. PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 25: December 7, 2012 An Introduction to Multilevel Models PSYC 943 (930): Fundamentals of Multivariate Modeling Lecture 25: December 7, 2012 Today s Class Concepts in Longitudinal Modeling Between-Person vs. +Within-Person

More information

Likelihood-Based Methods

Likelihood-Based Methods Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)

More information

An Introduction to Mplus and Path Analysis

An Introduction to Mplus and Path Analysis An Introduction to Mplus and Path Analysis PSYC 943: Fundamentals of Multivariate Modeling Lecture 10: October 30, 2013 PSYC 943: Lecture 10 Today s Lecture Path analysis starting with multivariate regression

More information

Introduction to Random Effects of Time and Model Estimation

Introduction to Random Effects of Time and Model Estimation Introduction to Random Effects of Time and Model Estimation Today s Class: The Big Picture Multilevel model notation Fixed vs. random effects of time Random intercept vs. random slope models How MLM =

More information

Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values. Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 2013

Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values. Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 2013 Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 013 1 Estimation of Var(A) and Breeding Values in General Pedigrees The classic

More information

The Common Factor Model. Measurement Methods Lecture 15 Chapter 9

The Common Factor Model. Measurement Methods Lecture 15 Chapter 9 The Common Factor Model Measurement Methods Lecture 15 Chapter 9 Today s Class Common Factor Model Multiple factors with a single test ML Estimation Methods New fit indices because of ML Estimation method

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

Statistical Distribution Assumptions of General Linear Models

Statistical Distribution Assumptions of General Linear Models Statistical Distribution Assumptions of General Linear Models Applied Multilevel Models for Cross Sectional Data Lecture 4 ICPSR Summer Workshop University of Colorado Boulder Lecture 4: Statistical Distributions

More information

Research Design: Topic 18 Hierarchical Linear Modeling (Measures within Persons) 2010 R.C. Gardner, Ph.d.

Research Design: Topic 18 Hierarchical Linear Modeling (Measures within Persons) 2010 R.C. Gardner, Ph.d. Research Design: Topic 8 Hierarchical Linear Modeling (Measures within Persons) R.C. Gardner, Ph.d. General Rationale, Purpose, and Applications Linear Growth Models HLM can also be used with repeated

More information

Chapter 5. Introduction to Path Analysis. Overview. Correlation and causation. Specification of path models. Types of path models

Chapter 5. Introduction to Path Analysis. Overview. Correlation and causation. Specification of path models. Types of path models Chapter 5 Introduction to Path Analysis Put simply, the basic dilemma in all sciences is that of how much to oversimplify reality. Overview H. M. Blalock Correlation and causation Specification of path

More information

1 Motivation for Instrumental Variable (IV) Regression

1 Motivation for Instrumental Variable (IV) Regression ECON 370: IV & 2SLS 1 Instrumental Variables Estimation and Two Stage Least Squares Econometric Methods, ECON 370 Let s get back to the thiking in terms of cross sectional (or pooled cross sectional) data

More information

Ignoring the matching variables in cohort studies - when is it valid, and why?

Ignoring the matching variables in cohort studies - when is it valid, and why? Ignoring the matching variables in cohort studies - when is it valid, and why? Arvid Sjölander Abstract In observational studies of the effect of an exposure on an outcome, the exposure-outcome association

More information

Econometrics Summary Algebraic and Statistical Preliminaries

Econometrics Summary Algebraic and Statistical Preliminaries Econometrics Summary Algebraic and Statistical Preliminaries Elasticity: The point elasticity of Y with respect to L is given by α = ( Y/ L)/(Y/L). The arc elasticity is given by ( Y/ L)/(Y/L), when L

More information

Mixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012

Mixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012 Mixed-Model Estimation of genetic variances Bruce Walsh lecture notes Uppsala EQG 01 course version 8 Jan 01 Estimation of Var(A) and Breeding Values in General Pedigrees The above designs (ANOVA, P-O

More information

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Institute of Statistics and Econometrics Georg-August-University Göttingen Department of Statistics

More information

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Department of Mathematics Carl von Ossietzky University Oldenburg Sonja Greven Department of

More information

Resemblance among relatives

Resemblance among relatives Resemblance among relatives Introduction Just as individuals may differ from one another in phenotype because they have different genotypes, because they developed in different environments, or both, relatives

More information

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed

More information

Centering Predictor and Mediator Variables in Multilevel and Time-Series Models

Centering Predictor and Mediator Variables in Multilevel and Time-Series Models Centering Predictor and Mediator Variables in Multilevel and Time-Series Models Tihomir Asparouhov and Bengt Muthén Part 2 May 7, 2018 Tihomir Asparouhov and Bengt Muthén Part 2 Muthén & Muthén 1/ 42 Overview

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

STOCKHOLM UNIVERSITY Department of Economics Course name: Empirical Methods Course code: EC40 Examiner: Lena Nekby Number of credits: 7,5 credits Date of exam: Friday, June 5, 009 Examination time: 3 hours

More information

Appendix A: Review of the General Linear Model

Appendix A: Review of the General Linear Model Appendix A: Review of the General Linear Model The generallinear modelis an important toolin many fmri data analyses. As the name general suggests, this model can be used for many different types of analyses,

More information

B. Weaver (24-Mar-2005) Multiple Regression Chapter 5: Multiple Regression Y ) (5.1) Deviation score = (Y i

B. Weaver (24-Mar-2005) Multiple Regression Chapter 5: Multiple Regression Y ) (5.1) Deviation score = (Y i B. Weaver (24-Mar-2005) Multiple Regression... 1 Chapter 5: Multiple Regression 5.1 Partial and semi-partial correlation Before starting on multiple regression per se, we need to consider the concepts

More information

STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013

STRUCTURAL EQUATION MODELING. Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013 STRUCTURAL EQUATION MODELING Khaled Bedair Statistics Department Virginia Tech LISA, Summer 2013 Introduction: Path analysis Path Analysis is used to estimate a system of equations in which all of the

More information

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis

Path Analysis. PRE 906: Structural Equation Modeling Lecture #5 February 18, PRE 906, SEM: Lecture 5 - Path Analysis Path Analysis PRE 906: Structural Equation Modeling Lecture #5 February 18, 2015 PRE 906, SEM: Lecture 5 - Path Analysis Key Questions for Today s Lecture What distinguishes path models from multivariate

More information

Introducing Generalized Linear Models: Logistic Regression

Introducing Generalized Linear Models: Logistic Regression Ron Heck, Summer 2012 Seminars 1 Multilevel Regression Models and Their Applications Seminar Introducing Generalized Linear Models: Logistic Regression The generalized linear model (GLM) represents and

More information

Random Intercept Models

Random Intercept Models Random Intercept Models Edps/Psych/Soc 589 Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Spring 2019 Outline A very simple case of a random intercept

More information

ECON 4160, Autumn term Lecture 1

ECON 4160, Autumn term Lecture 1 ECON 4160, Autumn term 2017. Lecture 1 a) Maximum Likelihood based inference. b) The bivariate normal model Ragnar Nymoen University of Oslo 24 August 2017 1 / 54 Principles of inference I Ordinary least

More information

Time-Invariant Predictors in Longitudinal Models

Time-Invariant Predictors in Longitudinal Models Time-Invariant Predictors in Longitudinal Models Today s Class (or 3): Summary of steps in building unconditional models for time What happens to missing predictors Effects of time-invariant predictors

More information

Ron Heck, Fall Week 8: Introducing Generalized Linear Models: Logistic Regression 1 (Replaces prior revision dated October 20, 2011)

Ron Heck, Fall Week 8: Introducing Generalized Linear Models: Logistic Regression 1 (Replaces prior revision dated October 20, 2011) Ron Heck, Fall 2011 1 EDEP 768E: Seminar in Multilevel Modeling rev. January 3, 2012 (see footnote) Week 8: Introducing Generalized Linear Models: Logistic Regression 1 (Replaces prior revision dated October

More information

Stat 579: Generalized Linear Models and Extensions

Stat 579: Generalized Linear Models and Extensions Stat 579: Generalized Linear Models and Extensions Mixed models Yan Lu March, 2018, week 8 1 / 32 Restricted Maximum Likelihood (REML) REML: uses a likelihood function calculated from the transformed set

More information

Analysis of variance using orthogonal projections

Analysis of variance using orthogonal projections Analysis of variance using orthogonal projections Rasmus Waagepetersen Abstract The purpose of this note is to show how statistical theory for inference in balanced ANOVA models can be conveniently developed

More information

MIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010

MIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010 MIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010 Part 1 of this document can be found at http://www.uvm.edu/~dhowell/methods/supplements/mixed Models for Repeated Measures1.pdf

More information

AGEC 661 Note Fourteen

AGEC 661 Note Fourteen AGEC 661 Note Fourteen Ximing Wu 1 Selection bias 1.1 Heckman s two-step model Consider the model in Heckman (1979) Y i = X iβ + ε i, D i = I {Z iγ + η i > 0}. For a random sample from the population,

More information

Model Assumptions; Predicting Heterogeneity of Variance

Model Assumptions; Predicting Heterogeneity of Variance Model Assumptions; Predicting Heterogeneity of Variance Today s topics: Model assumptions Normality Constant variance Predicting heterogeneity of variance CLP 945: Lecture 6 1 Checking for Violations of

More information

2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2

2/26/2017. PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 PSY 512: Advanced Statistics for Psychological and Behavioral Research 2 What is SEM? When should we use SEM? What can SEM tell us? SEM Terminology and Jargon Technical Issues Types of SEM Models Limitations

More information

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague

Econometrics. Week 8. Fall Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Econometrics Week 8 Institute of Economic Studies Faculty of Social Sciences Charles University in Prague Fall 2012 1 / 25 Recommended Reading For the today Instrumental Variables Estimation and Two Stage

More information

Technical Track Session I: Causal Inference

Technical Track Session I: Causal Inference Impact Evaluation Technical Track Session I: Causal Inference Human Development Human Network Development Network Middle East and North Africa Region World Bank Institute Spanish Impact Evaluation Fund

More information

6. Assessing studies based on multiple regression

6. Assessing studies based on multiple regression 6. Assessing studies based on multiple regression Questions of this section: What makes a study using multiple regression (un)reliable? When does multiple regression provide a useful estimate of the causal

More information

Final Exam - Solutions

Final Exam - Solutions Ecn 102 - Analysis of Economic Data University of California - Davis March 19, 2010 Instructor: John Parman Final Exam - Solutions You have until 5:30pm to complete this exam. Please remember to put your

More information

Introduction to Within-Person Analysis and RM ANOVA

Introduction to Within-Person Analysis and RM ANOVA Introduction to Within-Person Analysis and RM ANOVA Today s Class: From between-person to within-person ANOVAs for longitudinal data Variance model comparisons using 2 LL CLP 944: Lecture 3 1 The Two Sides

More information

REVIEW 8/2/2017 陈芳华东师大英语系

REVIEW 8/2/2017 陈芳华东师大英语系 REVIEW Hypothesis testing starts with a null hypothesis and a null distribution. We compare what we have to the null distribution, if the result is too extreme to belong to the null distribution (p

More information

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables

Structural Equation Modeling and Confirmatory Factor Analysis. Types of Variables /4/04 Structural Equation Modeling and Confirmatory Factor Analysis Advanced Statistics for Researchers Session 3 Dr. Chris Rakes Website: http://csrakes.yolasite.com Email: Rakes@umbc.edu Twitter: @RakesChris

More information

WISE International Masters

WISE International Masters WISE International Masters ECONOMETRICS Instructor: Brett Graham INSTRUCTIONS TO STUDENTS 1 The time allowed for this examination paper is 2 hours. 2 This examination paper contains 32 questions. You are

More information

Model Selection in GLMs. (should be able to implement frequentist GLM analyses!) Today: standard frequentist methods for model selection

Model Selection in GLMs. (should be able to implement frequentist GLM analyses!) Today: standard frequentist methods for model selection Model Selection in GLMs Last class: estimability/identifiability, analysis of deviance, standard errors & confidence intervals (should be able to implement frequentist GLM analyses!) Today: standard frequentist

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical

More information

Stat/F&W Ecol/Hort 572 Review Points Ané, Spring 2010

Stat/F&W Ecol/Hort 572 Review Points Ané, Spring 2010 1 Linear models Y = Xβ + ɛ with ɛ N (0, σ 2 e) or Y N (Xβ, σ 2 e) where the model matrix X contains the information on predictors and β includes all coefficients (intercept, slope(s) etc.). 1. Number of

More information

Survival Analysis for Case-Cohort Studies

Survival Analysis for Case-Cohort Studies Survival Analysis for ase-ohort Studies Petr Klášterecký Dept. of Probability and Mathematical Statistics, Faculty of Mathematics and Physics, harles University, Prague, zech Republic e-mail: petr.klasterecky@matfyz.cz

More information

Introduction to Confirmatory Factor Analysis

Introduction to Confirmatory Factor Analysis Introduction to Confirmatory Factor Analysis Multivariate Methods in Education ERSH 8350 Lecture #12 November 16, 2011 ERSH 8350: Lecture 12 Today s Class An Introduction to: Confirmatory Factor Analysis

More information

Lectures on Simple Linear Regression Stat 431, Summer 2012

Lectures on Simple Linear Regression Stat 431, Summer 2012 Lectures on Simple Linear Regression Stat 43, Summer 0 Hyunseung Kang July 6-8, 0 Last Updated: July 8, 0 :59PM Introduction Previously, we have been investigating various properties of the population

More information

Time-Invariant Predictors in Longitudinal Models

Time-Invariant Predictors in Longitudinal Models Time-Invariant Predictors in Longitudinal Models Topics: What happens to missing predictors Effects of time-invariant predictors Fixed vs. systematically varying vs. random effects Model building strategies

More information

Lecture 4: Linear panel models

Lecture 4: Linear panel models Lecture 4: Linear panel models Luc Behaghel PSE February 2009 Luc Behaghel (PSE) Lecture 4 February 2009 1 / 47 Introduction Panel = repeated observations of the same individuals (e.g., rms, workers, countries)

More information

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS

LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS LECTURE 4 PRINCIPAL COMPONENTS ANALYSIS / EXPLORATORY FACTOR ANALYSIS NOTES FROM PRE- LECTURE RECORDING ON PCA PCA and EFA have similar goals. They are substantially different in important ways. The goal

More information

Review of CLDP 944: Multilevel Models for Longitudinal Data

Review of CLDP 944: Multilevel Models for Longitudinal Data Review of CLDP 944: Multilevel Models for Longitudinal Data Topics: Review of general MLM concepts and terminology Model comparisons and significance testing Fixed and random effects of time Significance

More information

Correlation and regression

Correlation and regression 1 Correlation and regression Yongjua Laosiritaworn Introductory on Field Epidemiology 6 July 2015, Thailand Data 2 Illustrative data (Doll, 1955) 3 Scatter plot 4 Doll, 1955 5 6 Correlation coefficient,

More information

multilevel modeling: concepts, applications and interpretations

multilevel modeling: concepts, applications and interpretations multilevel modeling: concepts, applications and interpretations lynne c. messer 27 october 2010 warning social and reproductive / perinatal epidemiologist concepts why context matters multilevel models

More information

Machine Learning Linear Classification. Prof. Matteo Matteucci

Machine Learning Linear Classification. Prof. Matteo Matteucci Machine Learning Linear Classification Prof. Matteo Matteucci Recall from the first lecture 2 X R p Regression Y R Continuous Output X R p Y {Ω 0, Ω 1,, Ω K } Classification Discrete Output X R p Y (X)

More information

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that

This model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that Linear Regression For (X, Y ) a pair of random variables with values in R p R we assume that E(Y X) = β 0 + with β R p+1. p X j β j = (1, X T )β j=1 This model of the conditional expectation is linear

More information

LECTURE 15: SIMPLE LINEAR REGRESSION I

LECTURE 15: SIMPLE LINEAR REGRESSION I David Youngberg BSAD 20 Montgomery College LECTURE 5: SIMPLE LINEAR REGRESSION I I. From Correlation to Regression a. Recall last class when we discussed two basic types of correlation (positive and negative).

More information

Instrumental Variables

Instrumental Variables Instrumental Variables Department of Economics University of Wisconsin-Madison September 27, 2016 Treatment Effects Throughout the course we will focus on the Treatment Effect Model For now take that to

More information

Final Review. Yang Feng. Yang Feng (Columbia University) Final Review 1 / 58

Final Review. Yang Feng.   Yang Feng (Columbia University) Final Review 1 / 58 Final Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Final Review 1 / 58 Outline 1 Multiple Linear Regression (Estimation, Inference) 2 Special Topics for Multiple

More information

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models

Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Computationally Efficient Estimation of Multilevel High-Dimensional Latent Variable Models Tihomir Asparouhov 1, Bengt Muthen 2 Muthen & Muthen 1 UCLA 2 Abstract Multilevel analysis often leads to modeling

More information

Lecture 2: Linear and Mixed Models

Lecture 2: Linear and Mixed Models Lecture 2: Linear and Mixed Models Bruce Walsh lecture notes Introduction to Mixed Models SISG, Seattle 18 20 July 2018 1 Quick Review of the Major Points The general linear model can be written as y =

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

Models with multiple random effects: Repeated Measures and Maternal effects

Models with multiple random effects: Repeated Measures and Maternal effects Models with multiple random effects: Repeated Measures and Maternal effects 1 Often there are several vectors of random effects Repeatability models Multiple measures Common family effects Cleaning up

More information

Unless provided with information to the contrary, assume for each question below that the Classical Linear Model assumptions hold.

Unless provided with information to the contrary, assume for each question below that the Classical Linear Model assumptions hold. Economics 345: Applied Econometrics Section A01 University of Victoria Midterm Examination #2 Version 1 SOLUTIONS Spring 2015 Instructor: Martin Farnham Unless provided with information to the contrary,

More information

EDF 7405 Advanced Quantitative Methods in Educational Research. Data are available on IQ of the child and seven potential predictors.

EDF 7405 Advanced Quantitative Methods in Educational Research. Data are available on IQ of the child and seven potential predictors. EDF 7405 Advanced Quantitative Methods in Educational Research Data are available on IQ of the child and seven potential predictors. Four are medical variables available at the birth of the child: Birthweight

More information

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs)

36-309/749 Experimental Design for Behavioral and Social Sciences. Dec 1, 2015 Lecture 11: Mixed Models (HLMs) 36-309/749 Experimental Design for Behavioral and Social Sciences Dec 1, 2015 Lecture 11: Mixed Models (HLMs) Independent Errors Assumption An error is the deviation of an individual observed outcome (DV)

More information

Technical Track Session I:

Technical Track Session I: Impact Evaluation Technical Track Session I: Click to edit Master title style Causal Inference Damien de Walque Amman, Jordan March 8-12, 2009 Click to edit Master subtitle style Human Development Human

More information

Generalized Linear Models for Non-Normal Data

Generalized Linear Models for Non-Normal Data Generalized Linear Models for Non-Normal Data Today s Class: 3 parts of a generalized model Models for binary outcomes Complications for generalized multivariate or multilevel models SPLH 861: Lecture

More information

Hierarchical Generalized Linear Models. ERSH 8990 REMS Seminar on HLM Last Lecture!

Hierarchical Generalized Linear Models. ERSH 8990 REMS Seminar on HLM Last Lecture! Hierarchical Generalized Linear Models ERSH 8990 REMS Seminar on HLM Last Lecture! Hierarchical Generalized Linear Models Introduction to generalized models Models for binary outcomes Interpreting parameter

More information

Help! Statistics! Mediation Analysis

Help! Statistics! Mediation Analysis Help! Statistics! Lunch time lectures Help! Statistics! Mediation Analysis What? Frequently used statistical methods and questions in a manageable timeframe for all researchers at the UMCG. No knowledge

More information

Lecture 1 Introduction to Multi-level Models

Lecture 1 Introduction to Multi-level Models Lecture 1 Introduction to Multi-level Models Course Website: http://www.biostat.jhsph.edu/~ejohnson/multilevel.htm All lecture materials extracted and further developed from the Multilevel Model course

More information

STA 431s17 Assignment Eight 1

STA 431s17 Assignment Eight 1 STA 43s7 Assignment Eight The first three questions of this assignment are about how instrumental variables can help with measurement error and omitted variables at the same time; see Lecture slide set

More information

FIT CRITERIA PERFORMANCE AND PARAMETER ESTIMATE BIAS IN LATENT GROWTH MODELS WITH SMALL SAMPLES

FIT CRITERIA PERFORMANCE AND PARAMETER ESTIMATE BIAS IN LATENT GROWTH MODELS WITH SMALL SAMPLES FIT CRITERIA PERFORMANCE AND PARAMETER ESTIMATE BIAS IN LATENT GROWTH MODELS WITH SMALL SAMPLES Daniel M. McNeish Measurement, Statistics, and Evaluation University of Maryland, College Park Background

More information

Multilevel Structural Equation Modeling

Multilevel Structural Equation Modeling Multilevel Structural Equation Modeling Joop Hox Utrecht University j.hox@uu.nl http://www.joophox.net 14_15_mlevsem Multilevel Regression Three level data structure Groups at different levels may have

More information

A (Brief) Introduction to Crossed Random Effects Models for Repeated Measures Data

A (Brief) Introduction to Crossed Random Effects Models for Repeated Measures Data A (Brief) Introduction to Crossed Random Effects Models for Repeated Measures Data Today s Class: Review of concepts in multivariate data Introduction to random intercepts Crossed random effects models

More information

ISQS 5349 Spring 2013 Final Exam

ISQS 5349 Spring 2013 Final Exam ISQS 5349 Spring 2013 Final Exam Name: General Instructions: Closed books, notes, no electronic devices. Points (out of 200) are in parentheses. Put written answers on separate paper; multiple choices

More information

Measurement Invariance (MI) in CFA and Differential Item Functioning (DIF) in IRT/IFA

Measurement Invariance (MI) in CFA and Differential Item Functioning (DIF) in IRT/IFA Topics: Measurement Invariance (MI) in CFA and Differential Item Functioning (DIF) in IRT/IFA What are MI and DIF? Testing measurement invariance in CFA Testing differential item functioning in IRT/IFA

More information

Factor Analysis & Structural Equation Models. CS185 Human Computer Interaction

Factor Analysis & Structural Equation Models. CS185 Human Computer Interaction Factor Analysis & Structural Equation Models CS185 Human Computer Interaction MoodPlay Recommender (Andjelkovic et al, UMAP 2016) Online system available here: http://ugallery.pythonanywhere.com/ 2 3 Structural

More information

Causality II: How does causal inference fit into public health and what it is the role of statistics?

Causality II: How does causal inference fit into public health and what it is the role of statistics? Causality II: How does causal inference fit into public health and what it is the role of statistics? Statistics for Psychosocial Research II November 13, 2006 1 Outline Potential Outcomes / Counterfactual

More information

Contest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2.

Contest Quiz 3. Question Sheet. In this quiz we will review concepts of linear regression covered in lecture 2. Updated: November 17, 2011 Lecturer: Thilo Klein Contact: tk375@cam.ac.uk Contest Quiz 3 Question Sheet In this quiz we will review concepts of linear regression covered in lecture 2. NOTE: Please round

More information

Comparing Change Scores with Lagged Dependent Variables in Models of the Effects of Parents Actions to Modify Children's Problem Behavior

Comparing Change Scores with Lagged Dependent Variables in Models of the Effects of Parents Actions to Modify Children's Problem Behavior Comparing Change Scores with Lagged Dependent Variables in Models of the Effects of Parents Actions to Modify Children's Problem Behavior David R. Johnson Department of Sociology and Haskell Sie Department

More information

EMERGING MARKETS - Lecture 2: Methodology refresher

EMERGING MARKETS - Lecture 2: Methodology refresher EMERGING MARKETS - Lecture 2: Methodology refresher Maria Perrotta April 4, 2013 SITE http://www.hhs.se/site/pages/default.aspx My contact: maria.perrotta@hhs.se Aim of this class There are many different

More information

R E S E A R C H T R I A N G L E P A R K, N O R T H C A R O L I N A

R E S E A R C H T R I A N G L E P A R K, N O R T H C A R O L I N A R E S E A R C H T R I A N G L E P A R K, N O R T H C A R O L I N A Simultaneous Modeling of Multiple Outcomes over Time: A Hierarchical Modeling Approach Abhik Das1, Charlotte Gard1, Henrietta Bada2 and

More information

CS534 Machine Learning - Spring Final Exam

CS534 Machine Learning - Spring Final Exam CS534 Machine Learning - Spring 2013 Final Exam Name: You have 110 minutes. There are 6 questions (8 pages including cover page). If you get stuck on one question, move on to others and come back to the

More information

Stat 579: Generalized Linear Models and Extensions

Stat 579: Generalized Linear Models and Extensions Stat 579: Generalized Linear Models and Extensions Linear Mixed Models for Longitudinal Data Yan Lu April, 2018, week 15 1 / 38 Data structure t1 t2 tn i 1st subject y 11 y 12 y 1n1 Experimental 2nd subject

More information

Time-Invariant Predictors in Longitudinal Models

Time-Invariant Predictors in Longitudinal Models Time-Invariant Predictors in Longitudinal Models Today s Topics: What happens to missing predictors Effects of time-invariant predictors Fixed vs. systematically varying vs. random effects Model building

More information

Estimation: Problems & Solutions

Estimation: Problems & Solutions Estimation: Problems & Solutions Edps/Psych/Stat 587 Carolyn J. Anderson Department of Educational Psychology c Board of Trustees, University of Illinois Fall 2017 Outline 1. Introduction: Estimation of

More information

Describing Within-Person Change over Time

Describing Within-Person Change over Time Describing Within-Person Change over Time Topics: Multilevel modeling notation and terminology Fixed and random effects of linear time Predicted variances and covariances from random slopes Dependency

More information

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43

Panel Data. March 2, () Applied Economoetrics: Topic 6 March 2, / 43 Panel Data March 2, 212 () Applied Economoetrics: Topic March 2, 212 1 / 43 Overview Many economic applications involve panel data. Panel data has both cross-sectional and time series aspects. Regression

More information