Likelihood Methods. 1 Likelihood Functions. The multivariate normal distribution likelihood function is

Size: px
Start display at page:

Download "Likelihood Methods. 1 Likelihood Functions. The multivariate normal distribution likelihood function is"

Transcription

1 Likelihood Methods 1 Likelihood Functions The multivariate normal distribution likelihood function is The log of the likelihood, say L 1 is Ly = π.5n V.5 exp.5y Xb V 1 y Xb. L 1 = 0.5[N lnπ + ln V +y Xb V 1 y Xb]. The term N lnπ is a constant that does not involve any of the unknown variances or effects in the model, therefore, it is commonly omitted during maximization computations. Maximizing the log likelihood maximizes the original likelihood function. Previously, therefore, V = R Z R 1 Z + G 1 G, ln V = ln R + ln G + ln Z R 1 Z + G 1. If R = Iσ 0, then ln R = ln Iσ 0 = lnσ 0 N I = N ln σ 01. Similarly, if G = + Iσi, where i = 1 to s, then ln G = ln Iσi = q i ln σi. Except, that in animal models one of the G i is equal to Aσi. In that case, which is ln Aσ i = lnσ i q i A ln Aσ i = q i ln σ i A = q i ln σ i + ln A. Recall that so that X C = R 1 X X R 1 Z Z R 1 X Z R 1 Z + G 1, C = Z R 1 Z + G 1 X V 1 X ln C = ln Z R 1 Z + G 1 + ln X V 1 X. 1

2 Maximum Likelihood Method Hartley Rao 1967 described the maximum likelihood approach for the estimation of variance components. Let L be equivalent to L 1 except for the constant involving π. L = 0.5[ln V +y Xb V 1 y Xb]. The derivatives of L with respect to b to σ i for i = 0, 1,... s are L b = X V 1 Xb X V 1 y L σ i =.5 tr[v 1 V/ σ i ] +.5y Xb V 1 V/ σ i V 1 y Xb =.5 tr[v 1 V i ] +.5y Xb V 1 V i V 1 y Xb Equating the derivatives to zero gives Recall that ˆb = X V 1 X X V 1 y, tr[v 1 V i ] = y Xˆb V 1 V i V 1 y Xˆb. Py = V 1 y Xˆb, where P is the projection matrix, that V i = Z i Z i, then tr[v 1 Z i Z i] = y PV i Py. In usual mixed model theory, the solution vector for a rom factor may be written as so that Also, Let û i = G i Z ipy, y PV i Py = y PZ i G i G i G i Z ipy = û ig i û i = û iûi/σi 4. tr[v 1 V i ] = tr[r 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 Z i Z i]. T = Z R 1 Z + G 1 1, R = Iσ 0, G = + Iσ i,

3 then tr[v 1 V i ] = trz iz i σ 0 trz iztz Z i σ 4 0. If T can be partitioned into submatrices for each rom factor, then Tσ 0 Z Z + + Iαi = I, + TZ Zσ0 = I T Iσ i, TZ Z i σ 0 = I T ii σ i, which yields trz iztz Z i σ0 4 = trz iz i σ0 tri T ii σ i σi. Finally, Combining results gives for i = 1,,..., s, for i = 0 gives tr[v 1 V i ] = tri T ii σi σi = triσ i = q i σ i trt ii σ 4 i trt ii σ 4 i. ˆσ i = û iûi + trt iiˆσ 0/q i ˆσ 0 = y y ˆb X y û Z y/ N..1 The EM Algorithm EM sts for Expectation Maximization. The procedure alternates between calculating conditional expected values maximizing simplified likelihoods. The actual data y are called the incomplete data in the EM algorithm, the complete data are considered to be y the unobservable rom effects, u i. If the realized values of the unobservable rom effects were known, then their variance would be the average of their squared values, i.e., ˆσ i = u iu i /q i. However, in real life the realized values of the rom effects are unknown. The steps of the EM algorithm are as follows: Step 0. Decide on starting values for the variances set m = 0. 3

4 Step 1.E-step Calculate the conditional expectation of the sufficient statistics, conditional on the incomplete data. Eu iu i y = σ 4m i y P m Z i Z ip m y +trσ m i I σ 4m i Z iv m 1 Z i = ˆt m i Step.M-step Maximize the likelihood of the complete data, σ m+1 i = ˆt m i /q i, i = 0, 1,,..., s. Step 3. If convergence is reached, set ˆσ = σ m+1, otherwise increase m by one return to Step 1. This is equivalent to constructing solving the mixed model equations with a given set of variances, σ m, then σ m+1 0 = y y ˆb X y û Z y/n, σ m+1 i = û iûi + σ m+1 0 trt ii /q i. 4

5 3 Restricted Maximum Likelihood Method Restricted or Residual maximum likelihood REML, was first suggested by Thompson 196, described formally by Patterson Thompson The procedure requires that y has a multivariate normal distribution. The method is translation invariant. The maximum likelihood approach automatically keeps the estimator within the allowable parameter spacei.e. zero to plus infinity, therefore, REML is a biased procedure. REML was proposed as an improvement to ML in order to account for the degrees of freedom lost in estimating fixed effects. The likelihood function used in REML is that for a set of error contrasts i.e. residuals assumed to have a multivariate normal distribution. The multivariate normal distribution likelihood function for the residual contrasts, K y, where K X = 0, K has rank equal to N rx, is LK y = π.5n rx K VK.5 exp.5k y K VK 1 K y. The natural log of the likelihood function is L 3 =.5N rx lnπ.5 ln K VK.5y KK VK 1 K y. Notice that.5n rx lnπ is a constant that does not depend on the unknown variance components or factors in the model, therefore, can be ignored to give L 4. Searle 1979 showed that ln K VK = ln V + ln X V 1 X y KK VK 1 K y = y Py = y Xˆb V 1 y Xˆb for any K such that K X = 0. Hence, L 4 can be written as L 4 =.5 ln V.5 ln X V 1 X.5y Xˆb V 1 y Xˆb. REML can be calculated a number of different ways. 1. Derivative Free approach is a search technique to find the parameters that maximize the log likelihood function. Two techniques will be described here.. First Derivatives EM is where the first derivatives of the log likelihood are determined set to zero in order to maximize the likelihood function. Solutions need to be obtained by iteration because the resulting equations are non linear. 3. Second Derivatives are generally more computationally deming. Gradient methods are used to find the parameters that make the first derivatives equal to zero. Newton- Raphson involves the observed information matrix Fishers Method of Scoring involves the expected information matrix have been used. Lately, the average information algorithm averages the observed expected information matrices has been used to reduce the computational time. 5

6 3.1 Example Problem All of the approaches attempt to maximize the log likelihood function of the error contrasts. To illustrate the methods, consider a single trait model with three factors F, A, B, of which A B are rom factors. There were a total of 90 observations, the total sum of squares was 356,000. The least squares equations for this small example are shown below F 1 F A 1 A A 3 B 1 B B 3 B 4 = Derivative Free REML Imagine an s dimensional array containing the values of the likelihood function for every possible set of values of the ratios of the components to the residual variance. The technique is to search this array find the set of ratios for which the likelihood function is maximized. There is more than one way to conduct the search. Care must be taken to find the global maximum rather than one of possibly many local maxima. At the same time the number of likelihood evaluations to be computed must also be minimized. Various alternative forms of L 4 can be derived. Note that that that combining these results gives Now note that ln V = ln R + ln G + ln G 1 + Z R 1 Z ln X V 1 X = ln C ln Z R 1 Z + G 1 L 4 =.5 ln R.5 ln G.5 ln C.5y Py. ln R = ln Iσ 0 = N ln σ0, ln G = q i ln σi, ln C = ln X R 1 X + ln Z SZ + G 1 6

7 Then where ln X R 1 X = ln X Xσ0 = lnσ0 X X = ln X X rx ln σ0, Z SZ + G 1 = σ0 MZ + G 1 = σ 0 Z MZ + G 1 σ 0. ln C = ln X X rx ln σ 0 q ln σ 0 + ln Z MZ + G 1 σ 0, finally, the log-likelihood function becomes L 4 =.5N rx q ln σ0.5 q i ln σi.5 ln C.5y Py, where Note that C = X X X Z Z X Z Z + G 1 σ 0. q i ln σi = q i ln σ0/α i = q i ln σ0 ln α i so that L 4 =.5[N rx ln σ0 q i ln α i + ln C +y Py]. The quantity y Py is y y Xˆb Zû/σ0. The computations are achieved by constructing the following matrix, X X X Z X y Z X Z Z + G 1 σ0 Z y y X y Z y y = C W y y W y, y then by Gaussian elimination of one row at a time, the sum of the log of the non-zero pivots using the same ordering for each evaluation of the likelihood gives log C y y Xˆb Zû. Gaussian elimination, using sparse matrix techniques, requires less computing time than inverting the coefficient matrix of the mixed model equations. The ordering of factors within the equations could be critical to the computational process some experimentation may be necessary to determine the best ordering. The likelihood function can be evaluated without the calculation of solutions to the mixed model equations, without inverting the coefficient matrix of the mixed model equations, without computing any of the σi. The formulations for more general models multiple trait models are more complex, but follow the same ideas. Searching the array of likelihood values for various values of α i can be done in several different ways. One method is to fix the values of all but one of the s α i, then evaluate L for four 7

8 or more different values of the α i that were not fixed. Then one can use a quadratic regression analysis to determine the value of that one ratio which maximizes L given that the other ratios are fixed. This is repeated for each of the s ratios, the process is repeated until a maximum likelihood is obtained. The calculations are demonstrated in the example that follows. Begin by fixing the value of α B = 10 letting the value of α A take on the values of 5, 10, 0, 30, 40. Using L 4 to evaluate the likelihood, then the results were as follows: α A L For example, the likelihood value for α A = 40, would be L 4 = 1 [N rx ln σ 0 q A ln α A q B ln α B + ln C +y y Xˆb Zû/σ0] where ln C = , y Py = /σ0 = 88, q A ln α A = , q B ln α B = , σ0 = , ln σ0 = , N rx = 88, then L 4 = 0.5[ / ] = To find the value of α A that maximizes L 4 for α B = 10, let Q = Y = then ˆβ = Q Q 1 Q Y =

9 From this a prediction equation for L 4 can be written as L 4 = α A α A. This equation can be differentiated with respect to α A then equated to zero to find the value of the ratio that maximizes the prediction equation. This gives α A =.04489/ = Now keep α A = try a number of values of α B from to 10, which give the following results. α B L Applying the quadratic regression to these points gives α B = The next step would be to fix α B = 1.65 to try new values for α A, such as 5 to 40 by units of 1. The range of values becomes finer finer. To insure that one has found the global maximum, the entire process could be started with vastly different starting values for the ratios, such as α B = 50 let values for α A be 40, 50, 60, 70. The more components there are to estimate, the more evaluations of the likelihood that are going to be needed, the more probable that convergence might be to a local maximum rather than to the global maximum. Please refer to the literature for specification of the log likelihood function for particular models situations The Simplex Method The Simplex Method Nelder Mead, 1965 is a procedure for finding the minimum of a function i.e. the minimum of L 4 or the maximum of L 4 with respect to the unknown variances covariances. The best way to describe the method is using the example data from 9

10 the previous sections. Begin by constructing a set of points for which L 4 is to be evaluated. A point is a vector of values for the unknowns α A, α B, for example, θ 1 = , then form two more points by changing one unknown at a time. Let the three points be as shown in the following table. No. α A α B Now calculate L 4 for each point arrange from largest to lowest value. No. α A α B L The idea now is to find another point to replace the last onelowest L 4. This is done by a process called reflection. Compute the mean of all points excluding the one with the lowest L 4. θ m = , then the reflection step is θ 4 = θ m + r θ m θ last, where r is recommended by Nelder Mead 1965 to be 1, giving θ 4 = The corresponding L 4 for this point was Compared to those in the table it has the largest value, therefore, is a better point than the other three. No. α A α B L Given this success, the Simplex method calls for an expansion step, i.e. to make a bigger change. Thus, θ 5 = θ m + E θ 4 θ m, 10

11 where E is suggested to be equal to. Hence θ 5 = Then L 4 = , the exped point is better yet. Now drop θ 1 from the table put θ 5 at the top. No. α A α B L This completes one iteration. Begin the next iteration by computing the mean of all points excluding the point with the lowest L 4. θ m = Another reflection step gives θ 6 = θ m + r θ m θ last, = However, this gives L 4 = , which is between θ θ 4, can push out θ from the table. No. α A α B L Instead of an expansion step, a contraction step is needed because θ 6 did not give a greater L 4 than the first two. Thus, θ 7 = θ m + c θ 6 θ m, where c = 0.5 is recommended. Hence, θ 7 = Then L 4 = is better than that given by θ 4, but not by θ 5, thus the new table becomes as follows:. No. α A α B L

12 The following steps were taken in the next iteration. 1. The mean of the top two L 4 is θ m = A reflection step gives θ 8 = θ m + r θ m θ last, = , which gave L 4 = , which is better than θ Add θ 8 to the table drop θ 4. No. α A α B L Because L 4 for θ 8 was not larger than L 4 for θ 5 or smaller than L 4 for θ 7, then no expansion or contraction step is necessary. Begin the next iteration. The Simplex method continues in this manner until all point entries in the table are equal. The constants recommended by Nelder Mead 1965 for reflection, expansion, contraction could be adjusted for a particular data set. This method may converge to a local maximum, so different starting values are needed to see if it converges to the same point. The Simplex method does not work well with a large number of parameters to be estimated. 3. First Derivatives EM Algorithm To derive formulas for estimating the variance components take the derivatives of L 4 with respect to the unknown components. L 4 σ i 1 V =.5trV σi +.5y Xˆb V.5trX V 1 X X V 1 V σi V 1 y Xˆb 1 V σi V 1 X Combine the two terms involving the traces note that V 1 y Xˆb = Py, 1

13 then L 4 σ i for i = 1,..., s or =.5trV 1 V 1 XX V 1 X X V 1 V σ i =.5trPZ i Z i +.5y PZ i Z ipy =.5trP +.5y PPy for i = 0 for the residual component. Using P the fact that V 1 = R 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 +.5y P V σi Py then trpz i Z i = q i /σi trc ii σ0/σ i 4 trp = N rxσ0 û iûi/σi. The other terms, y PZ i Z i Py y PPy, were simplified by Henderson 1973 to show that they could be calculated from the Mixed Model Equations. Note that Henderson 1973 showed then which when G i = Iσ i gives Py = V 1 y Xˆb, ˆb = X V 1 X X V 1 y, û i = G i Z ipy, y PZ i Z ipy = y PZ i [G i G 1 i = y PZ i G i G i = û ig i û i û iûi/σ 4 i. Similarly for the residual component, Henderson showed that where α i = σ 0 /σ i. G 1 i G i ]Z ipy G i Z ipy y PPy = [y y ˆb X y û iz iy + û iûiα i ]/σ0, Equate the derivatives to zero incorporating the above simplifications obtain ˆσ i = û iûi + trc ii σ 0/q i, ˆσ 0 = y Py/N rx. As with ML, solutions using the EM algorithm must be computed iteratively. Convergence is usually very slow, if it occurs, the process may also diverge. 13

14 Notice the differences between REML ML. The denominator for ˆσ 0 is N rx rather than N, in ˆσ i is trc ii rather than trt ii. The quadratic forms, however, are identical in REML ML. Accounting for the degrees of freedom to estimate b has resulted in the REML algorithm. A major computing problem with the EM algorithm is the calculation of trc ii, which is the corresponding inverse elements of the mixed model equations for the i th rom factor. With most applications in animal breeding, the order of the mixed model equations are too large to be inverted, solutions to the equations are obtained by Gauss-Seidel iterations. However, there have been several attempts to approximate trc ii, but these have not been totally suitable. To demonstrate the EM algorithm let α A = 10 α B = 5 be the starting values of the ratios for factors A B, respectively. There were N = 90 total observations, rx =. The solution vector is Then F 1 F A 1 A A 3 B 1 B B 3 B 4 = from the inverse of the coefficient matrix, which give rise to the following estimates, New ratios are formed as y Xˆb + Zû = 347, trc AA =.16493, trc BB = ˆσ 0 = 356, , /88 = , ˆσ A = /3 = , ˆσ B = /4 = α A = / = , α B = / =

15 these are used to form the mixed model equations again, new solutions traces are calculated, so on, until the estimated ratios the prior values of the ratios are equal. The estimates converge to ˆσ 0 = , ˆσ A =.569, ˆσ B = or α A = , α B = Second Derivatives, Average Information Second derivatives of the log likelihood lead to the expectations of the quadratic forms. One technique, MIVQUE Minimum Variance Quadratic Unbiased Estimation equates the quadratic forms to their expectations. The estimates are unbiased if all variances remain positive, then convergence will be to the REML estimates. However, due to a shortage of data or an inappropriate model, the estimates derived in this manner can be negative. Computing the expectations of the quadratic forms requires the inverse of the mixed model equations coefficient matrix, then products crossproducts of various parts of the inverse. A gradient method using first second derivatives can be used Hofer, The gradient, d the vector of first derivatives of the log likelihood, is used to determine the direction towards the parameters that give the maximum of the log likelihood, such that θ t+1 = θ t + M t d t, where d t are the first derivatives evaluated at θ = θ t, M t in the Newton-RaphsonNR algorithm is the observed information matrix, in the Fisher Method of ScoringFS it is the expected information matrix. The first derivatives are as follows from earlier in these notes: L 4 σ i =.5trPZ i Z i +.5y PZ i Z ipy = 0 for i = 1,..., s or L 4 σ0 =.5trP +.5y PPy = 0 for the residual component. Then from earlier results, trpz i Z i = q i /σ i trc ii σ 0/σ 4 i, y PZ i Z ipy = û iûi/σ 4 i 15

16 which combined give 0.5û iûi/σi 4 q i /σi + trc ii σ0/σ i 4 = 0, for i = 1,..., s, trp = N rxσ0 û iûi/σi y PPy = [y y ˆb X y û iz iy + û iûiα i ]/σ0 which combined give 0.5[y y ˆb X y û iz iy + û iûiα i ]/σ0 N rxσ0 + û iûi/σi = 0, which simplifies to 0.5[y y ˆb X y û iz iy]/σ0 N rxσ0 = 0. The second derivatives give a matrix of quantities. The elements of the observed information matrix Gilmour et al are L 4 σ i σ 0 = 0.5y PZ i Z ipy/σ 4 0, L 4 σ i σ j L 4 σ 0 σ 0 = 0.5trPZ i Z j 0.5trPZ i Z ipz j Z j +y PZ i Z ipz j Z jpy/σ 0 0.5y PZ i Z jpy/σ 0, = y Py/σ N rx/σ 4 0. The elements of the expected information matrix Gilmour et al are E[ L 4 σi ] = 0.5trPZ i Z i/σ0, σ 0 E[ L 4 σi ] σ j = 0.5trPZ i Z ipz j Z j, E[ L 4 σ0 ] = 0.5N rx/σ0. 4 σ 0 As the name Average Information implies, average the observed expected information matrices to give the following matrix of elements. I[σ i, σ 0] = 0.5y PZ i Z ipy/σ 4 0, I[σ i, σ j ] = y PZ i Z ipz j Z jpy/σ 0, I[σ 0, σ 0] = 0.5y Py/σ

17 The first derivatives form the vector, d t, M t = I[σ, σ] 1. The rest of this method is computational detail to simplify the requirements for inverse elements solutions to MME. The calculations can not be illustrated very easily for the example data because the y-vector is not available. 3.4 Animal Models The model commonly applied to estimation of variance components in livestock genetics since 1989 has been an animal model. The animal model assumes a large, rom mating population, an infinite number of loci each with a small equal effect on the trait, only additive genetic effects, all relationships among animals are known tracible to an unselected base population somewhere in the past. Animals may have more than one record each. The equation of the model is y = Xb + Za + Zp + e, where a is the vector of animal additive genetic effects one per animal, p is a vector of permanent environmental p.e. effects associated with each animal. V ar Ey = Xb, a p = e Aσa Iσp Iσe The matrix A is called the numerator relationship matrix. Wright defined relationships among animals as correlations, but A is essentially relationships defined as covariances the numerators of the correlation coefficients. Also, these only represent the additive genetic relationships between animals. The MME for this model are X X X Z X Z Z X Z Z + A 1 k a Z Z Z X Z Z Z Z + Ik p ˆb â ˆp = Note that k a is the ratio of residual to additive genetic variances, k p is the ratio of residual to permanent environmental variances. Also, in MME the inverse of A is required. The EM-REML procedure gives. X y Z y Z y ˆσ e = y y ˆb X y â Z y ˆp Z y/n rx, ˆσ a = â A 1 â + tra 1 C aaˆσ e/n, ˆσ p = ˆp ˆp + trc ppˆσ e/n,. 17

18 where n is the total number of animals, N is the total number of records, C aa are the inverse elements of the MME for the animal additive genetic effects, C pp are the inverse elements of the MME for the animal permanent environmental effects. An example of this model will be given in later notes Quadratic Forms in an Animal Model A necessary quadratic form in an animal model is â A 1 â, this can be computed very easily. Note that the inverse of A may be written as A 1 = T 1 D T 1, where T 1 is an upper triangular matrix, diagonal matrix D has elements equal to 1,, or 4/3 in noninbred situations, values greater than in inbred situations. In Henderson 1975, this inverse was shown to be composed of just three numbers, i.e. 0, 1 s on the diagonals, -.5 corresponding to the parents of an animal. For example, T 1 = Then T 1â = ˆm = â i 0.5â s + â d, for the i th animal, â s â d are the sire dam estimated breeding values, respectively. Consequently, â A 1 â = â T 1 B T 1â = ˆm B ˆm q = ˆm i b ii, where b ii are the diagonal elements of B, q is the number of animals. 4 EXERCISES Below are pedigrees data on 0 animals. 18

19 Animal Sire Dam Group Record Construct A 1 set up the MME.. Apply EM-REML to the model, y ij = µ + g i + a j + e ij, where group, animal additive genetic, residual effects are rom. Let σ e/σ a = 1.5, σ e/σ g = 5.0, to start the iterations. Do five iterations of EM-REML. 3. Apply EM-REML again to the model with σ e/σ a = 3.0, σ e/σ g =.5, do five iterations of REML from these starting values. 4. Do the results of the previous two question tend to give similar answers? Comment on the results. 5. Use the Derivative Free method compute the log likelihoods for various sets of parameters. 19

Estimation of Variance Components in Animal Breeding

Estimation of Variance Components in Animal Breeding Estimation of Variance Components in Animal Breeding L R Schaeffer July 19-23, 2004 Iowa State University Short Course 2 Contents 1 Distributions 9 11 Random Variables 9 12 Discrete Random Variables 10

More information

Lecture Notes. Introduction

Lecture Notes. Introduction 5/3/016 Lecture Notes R. Rekaya June 1-10, 016 Introduction Variance components play major role in animal breeding and genetic (estimation of BVs) It has been an active area of research since early 1950

More information

Chapter 11 MIVQUE of Variances and Covariances

Chapter 11 MIVQUE of Variances and Covariances Chapter 11 MIVQUE of Variances and Covariances C R Henderson 1984 - Guelph The methods described in Chapter 10 for estimation of variances are quadratic, translation invariant, and unbiased For the balanced

More information

Repeated Records Animal Model

Repeated Records Animal Model Repeated Records Animal Model 1 Introduction Animals are observed more than once for some traits, such as Fleece weight of sheep in different years. Calf records of a beef cow over time. Test day records

More information

Reduced Animal Models

Reduced Animal Models Reduced Animal Models 1 Introduction In situations where many offspring can be generated from one mating as in fish poultry or swine or where only a few animals are retained for breeding the genetic evaluation

More information

MIXED MODELS THE GENERAL MIXED MODEL

MIXED MODELS THE GENERAL MIXED MODEL MIXED MODELS This chapter introduces best linear unbiased prediction (BLUP), a general method for predicting random effects, while Chapter 27 is concerned with the estimation of variances by restricted

More information

Estimation of Variances and Covariances

Estimation of Variances and Covariances Estimation of Variances and Covariances Variables and Distributions Random variables are samples from a population with a given set of population parameters Random variables can be discrete, having a limited

More information

Chapter 12 REML and ML Estimation

Chapter 12 REML and ML Estimation Chapter 12 REML and ML Estimation C. R. Henderson 1984 - Guelph 1 Iterative MIVQUE The restricted maximum likelihood estimator (REML) of Patterson and Thompson (1971) can be obtained by iterating on MIVQUE,

More information

Animal Models. Sheep are scanned at maturity by ultrasound(us) to determine the amount of fat surrounding the muscle. A model (equation) might be

Animal Models. Sheep are scanned at maturity by ultrasound(us) to determine the amount of fat surrounding the muscle. A model (equation) might be Animal Models 1 Introduction An animal model is one in which there are one or more observations per animal, and all factors affecting those observations are described including an animal additive genetic

More information

Mixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012

Mixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012 Mixed-Model Estimation of genetic variances Bruce Walsh lecture notes Uppsala EQG 01 course version 8 Jan 01 Estimation of Var(A) and Breeding Values in General Pedigrees The above designs (ANOVA, P-O

More information

3. Properties of the relationship matrix

3. Properties of the relationship matrix 3. Properties of the relationship matrix 3.1 Partitioning of the relationship matrix The additive relationship matrix, A, can be written as the product of a lower triangular matrix, T, a diagonal matrix,

More information

Maternal Genetic Models

Maternal Genetic Models Maternal Genetic Models In mammalian species of livestock such as beef cattle sheep or swine the female provides an environment for its offspring to survive and grow in terms of protection and nourishment

More information

Mixed-Models. version 30 October 2011

Mixed-Models. version 30 October 2011 Mixed-Models version 30 October 2011 Mixed models Mixed models estimate a vector! of fixed effects and one (or more) vectors u of random effects Both fixed and random effects models always include a vector

More information

REML Variance-Component Estimation

REML Variance-Component Estimation REML Variance-Component Estimation In the numerous forms of analysis of variance (ANOVA) discussed in previous chapters, variance components were estimated by equating observed mean squares to expressions

More information

Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values. Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 2013

Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values. Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 2013 Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 013 1 Estimation of Var(A) and Breeding Values in General Pedigrees The classic

More information

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8

Peter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8 Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall

More information

Animal Model. 2. The association of alleles from the two parents is assumed to be at random.

Animal Model. 2. The association of alleles from the two parents is assumed to be at random. Animal Model 1 Introduction In animal genetics, measurements are taken on individual animals, and thus, the model of analysis should include the animal additive genetic effect. The remaining items in the

More information

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences

Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Genet. Sel. Evol. 33 001) 443 45 443 INRA, EDP Sciences, 001 Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Louis Alberto GARCÍA-CORTÉS a, Daniel SORENSEN b, Note a

More information

VARIANCE COMPONENT ESTIMATION & BEST LINEAR UNBIASED PREDICTION (BLUP)

VARIANCE COMPONENT ESTIMATION & BEST LINEAR UNBIASED PREDICTION (BLUP) VARIANCE COMPONENT ESTIMATION & BEST LINEAR UNBIASED PREDICTION (BLUP) V.K. Bhatia I.A.S.R.I., Library Avenue, New Delhi- 11 0012 vkbhatia@iasri.res.in Introduction Variance components are commonly used

More information

Best unbiased linear Prediction: Sire and Animal models

Best unbiased linear Prediction: Sire and Animal models Best unbiased linear Prediction: Sire and Animal models Raphael Mrode Training in quantitative genetics and genomics 3 th May to th June 26 ILRI, Nairobi Partner Logo Partner Logo BLUP The MME of provided

More information

Chapter 5 Prediction of Random Variables

Chapter 5 Prediction of Random Variables Chapter 5 Prediction of Random Variables C R Henderson 1984 - Guelph We have discussed estimation of β, regarded as fixed Now we shall consider a rather different problem, prediction of random variables,

More information

Chapter 3 Best Linear Unbiased Estimation

Chapter 3 Best Linear Unbiased Estimation Chapter 3 Best Linear Unbiased Estimation C R Henderson 1984 - Guelph In Chapter 2 we discussed linear unbiased estimation of k β, having determined that it is estimable Let the estimate be a y, and if

More information

Lecture 9 Multi-Trait Models, Binary and Count Traits

Lecture 9 Multi-Trait Models, Binary and Count Traits Lecture 9 Multi-Trait Models, Binary and Count Traits Guilherme J. M. Rosa University of Wisconsin-Madison Mixed Models in Quantitative Genetics SISG, Seattle 18 0 September 018 OUTLINE Multiple-trait

More information

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model

Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives

More information

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012

Lecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed

More information

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1

Inverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1 Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is

More information

Outline of GLMs. Definitions

Outline of GLMs. Definitions Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density

More information

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.

STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review

More information

Chapter 3: Maximum Likelihood Theory

Chapter 3: Maximum Likelihood Theory Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood

More information

Regression. Oscar García

Regression. Oscar García Regression Oscar García Regression methods are fundamental in Forest Mensuration For a more concise and general presentation, we shall first review some matrix concepts 1 Matrices An order n m matrix is

More information

1 Mixed effect models and longitudinal data analysis

1 Mixed effect models and longitudinal data analysis 1 Mixed effect models and longitudinal data analysis Mixed effects models provide a flexible approach to any situation where data have a grouping structure which introduces some kind of correlation between

More information

Economics 620, Lecture 5: exp

Economics 620, Lecture 5: exp 1 Economics 620, Lecture 5: The K-Variable Linear Model II Third assumption (Normality): y; q(x; 2 I N ) 1 ) p(y) = (2 2 ) exp (N=2) 1 2 2(y X)0 (y X) where N is the sample size. The log likelihood function

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Lecture 32: Infinite-dimensional/Functionvalued. Functions and Random Regressions. Bruce Walsh lecture notes Synbreed course version 11 July 2013

Lecture 32: Infinite-dimensional/Functionvalued. Functions and Random Regressions. Bruce Walsh lecture notes Synbreed course version 11 July 2013 Lecture 32: Infinite-dimensional/Functionvalued Traits: Covariance Functions and Random Regressions Bruce Walsh lecture notes Synbreed course version 11 July 2013 1 Longitudinal traits Many classic quantitative

More information

Linear Methods for Prediction

Linear Methods for Prediction Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we

More information

Raphael Mrode. Training in quantitative genetics and genomics 30 May 10 June 2016 ILRI, Nairobi. Partner Logo. Partner Logo

Raphael Mrode. Training in quantitative genetics and genomics 30 May 10 June 2016 ILRI, Nairobi. Partner Logo. Partner Logo Basic matrix algebra Raphael Mrode Training in quantitative genetics and genomics 3 May June 26 ILRI, Nairobi Partner Logo Partner Logo Matrix definition A matrix is a rectangular array of numbers set

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Linear Models in Machine Learning

Linear Models in Machine Learning CS540 Intro to AI Linear Models in Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu We briefly go over two linear models frequently used in machine learning: linear regression for, well, regression,

More information

Summary INTRODUCTION. Running head : AI-REML FOR EQUAL DESIGN MATRICES. K. Meyer

Summary INTRODUCTION. Running head : AI-REML FOR EQUAL DESIGN MATRICES. K. Meyer Running head : AI-REML FOR EQUAL DESIGN MATRICES An average information Restricted Maximum Likelihood algorithm for estimating reduced rank genetic covariance matrices or covariance functions for animal

More information

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32 ANOVA Variance Component Estimation Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 32 We now consider the ANOVA approach to variance component estimation. The ANOVA approach

More information

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011

Lecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector

More information

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2

MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2 MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a

More information

Linear Algebra Review

Linear Algebra Review Linear Algebra Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Linear Algebra Review 1 / 45 Definition of Matrix Rectangular array of elements arranged in rows and

More information

Lecture 5 Basic Designs for Estimation of Genetic Parameters

Lecture 5 Basic Designs for Estimation of Genetic Parameters Lecture 5 Basic Designs for Estimation of Genetic Parameters Bruce Walsh. jbwalsh@u.arizona.edu. University of Arizona. Notes from a short course taught June 006 at University of Aarhus The notes for this

More information

Models with multiple random effects: Repeated Measures and Maternal effects

Models with multiple random effects: Repeated Measures and Maternal effects Models with multiple random effects: Repeated Measures and Maternal effects 1 Often there are several vectors of random effects Repeatability models Multiple measures Common family effects Cleaning up

More information

Likelihood-Based Methods

Likelihood-Based Methods Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)

More information

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32

ANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32 ANOVA Variance Component Estimation Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 32 We now consider the ANOVA approach to variance component estimation. The ANOVA approach

More information

Economics 620, Lecture 4: The K-Variable Linear Model I. y 1 = + x 1 + " 1 y 2 = + x 2 + " 2 :::::::: :::::::: y N = + x N + " N

Economics 620, Lecture 4: The K-Variable Linear Model I. y 1 = + x 1 +  1 y 2 = + x 2 +  2 :::::::: :::::::: y N = + x N +  N 1 Economics 620, Lecture 4: The K-Variable Linear Model I Consider the system y 1 = + x 1 + " 1 y 2 = + x 2 + " 2 :::::::: :::::::: y N = + x N + " N or in matrix form y = X + " where y is N 1, X is N

More information

1 Data Arrays and Decompositions

1 Data Arrays and Decompositions 1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is

More information

GBLUP and G matrices 1

GBLUP and G matrices 1 GBLUP and G matrices 1 GBLUP from SNP-BLUP We have defined breeding values as sum of SNP effects:! = #$ To refer breeding values to an average value of 0, we adopt the centered coding for genotypes described

More information

Chapter 4: Factor Analysis

Chapter 4: Factor Analysis Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.

More information

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error

Journal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error Journal of Multivariate Analysis 00 (009) 305 3 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Sphericity test in a GMANOVA MANOVA

More information

RESTRICTED M A X I M U M LIKELIHOOD TO E S T I M A T E GENETIC P A R A M E T E R S - IN PRACTICE

RESTRICTED M A X I M U M LIKELIHOOD TO E S T I M A T E GENETIC P A R A M E T E R S - IN PRACTICE RESTRICTED M A X I M U M LIKELIHOOD TO E S T I M A T E GENETIC P A R A M E T E R S - IN PRACTICE K. M e y e r Institute of Animal Genetics, Edinburgh University, W e s t M a i n s Road, Edinburgh EH9 3JN,

More information

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations

Chapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2

More information

6-1. Canonical Correlation Analysis

6-1. Canonical Correlation Analysis 6-1. Canonical Correlation Analysis Canonical Correlatin analysis focuses on the correlation between a linear combination of the variable in one set and a linear combination of the variables in another

More information

Properties of the least squares estimates

Properties of the least squares estimates Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares

More information

Lecture 15. Hypothesis testing in the linear model

Lecture 15. Hypothesis testing in the linear model 14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma

More information

Quantitative characters - exercises

Quantitative characters - exercises Quantitative characters - exercises 1. a) Calculate the genetic covariance between half sibs, expressed in the ij notation (Cockerham's notation), when up to loci are considered. b) Calculate the genetic

More information

The equivalence of the Maximum Likelihood and a modified Least Squares for a case of Generalized Linear Model

The equivalence of the Maximum Likelihood and a modified Least Squares for a case of Generalized Linear Model Applied and Computational Mathematics 2014; 3(5): 268-272 Published online November 10, 2014 (http://www.sciencepublishinggroup.com/j/acm) doi: 10.11648/j.acm.20140305.22 ISSN: 2328-5605 (Print); ISSN:

More information

ECON 331 Homework #2 - Solution. In a closed model the vector of external demand is zero, so the matrix equation writes:

ECON 331 Homework #2 - Solution. In a closed model the vector of external demand is zero, so the matrix equation writes: ECON 33 Homework #2 - Solution. (Leontief model) (a) (i) The matrix of input-output A and the vector of level of production X are, respectively:.2.3.2 x A =.5.2.3 and X = y.3.5.5 z In a closed model the

More information

Should genetic groups be fitted in BLUP evaluation? Practical answer for the French AI beef sire evaluation

Should genetic groups be fitted in BLUP evaluation? Practical answer for the French AI beef sire evaluation Genet. Sel. Evol. 36 (2004) 325 345 325 c INRA, EDP Sciences, 2004 DOI: 10.1051/gse:2004004 Original article Should genetic groups be fitted in BLUP evaluation? Practical answer for the French AI beef

More information

Multiple random effects. Often there are several vectors of random effects. Covariance structure

Multiple random effects. Often there are several vectors of random effects. Covariance structure Models with multiple random effects: Repeated Measures and Maternal effects Bruce Walsh lecture notes SISG -Mixed Model Course version 8 June 01 Multiple random effects y = X! + Za + Wu + e y is a n x

More information

MLES & Multivariate Normal Theory

MLES & Multivariate Normal Theory Merlise Clyde September 6, 2016 Outline Expectations of Quadratic Forms Distribution Linear Transformations Distribution of estimates under normality Properties of MLE s Recap Ŷ = ˆµ is an unbiased estimate

More information

PREDICTION OF BREEDING VALUES FOR UNMEASURED TRAITS FROM MEASURED TRAITS

PREDICTION OF BREEDING VALUES FOR UNMEASURED TRAITS FROM MEASURED TRAITS Libraries Annual Conference on Applied Statistics in Agriculture 1994-6th Annual Conference Proceedings PREDICTION OF BREEDING VALUES FOR UNMEASURED TRAITS FROM MEASURED TRAITS Kristin L. Barkhouse L.

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

1 Appendix A: Matrix Algebra

1 Appendix A: Matrix Algebra Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix

More information

5. Best Linear Unbiased Prediction

5. Best Linear Unbiased Prediction 5. Best Linear Unbiased Prediction Julius van der Werf Lecture 1: Best linear unbiased prediction Learning objectives On completion of Lecture 1 you should be able to: Understand the principle of mixed

More information

Lecture 7 Correlated Characters

Lecture 7 Correlated Characters Lecture 7 Correlated Characters Bruce Walsh. Sept 2007. Summer Institute on Statistical Genetics, Liège Genetic and Environmental Correlations Many characters are positively or negatively correlated at

More information

Multivariate Regression Analysis

Multivariate Regression Analysis Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x

More information

Decomposable and Directed Graphical Gaussian Models

Decomposable and Directed Graphical Gaussian Models Decomposable Decomposable and Directed Graphical Gaussian Models Graphical Models and Inference, Lecture 13, Michaelmas Term 2009 November 26, 2009 Decomposable Definition Basic properties Wishart density

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

4 Elementary matrices, continued

4 Elementary matrices, continued 4 Elementary matrices, continued We have identified 3 types of row operations and their corresponding elementary matrices. To repeat the recipe: These matrices are constructed by performing the given row

More information

Computational methods for mixed models

Computational methods for mixed models Computational methods for mixed models Douglas Bates Department of Statistics University of Wisconsin Madison March 27, 2018 Abstract The lme4 package provides R functions to fit and analyze several different

More information

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,

More information

ANALYSIS OF VARIANCE AND QUADRATIC FORMS

ANALYSIS OF VARIANCE AND QUADRATIC FORMS 4 ANALYSIS OF VARIANCE AND QUADRATIC FORMS The previous chapter developed the regression results involving linear functions of the dependent variable, β, Ŷ, and e. All were shown to be normally distributed

More information

Charles E. McCulloch Biometrics Unit and Statistics Center Cornell University

Charles E. McCulloch Biometrics Unit and Statistics Center Cornell University A SURVEY OF VARIANCE COMPONENTS ESTIMATION FROM BINARY DATA by Charles E. McCulloch Biometrics Unit and Statistics Center Cornell University BU-1211-M May 1993 ABSTRACT The basic problem of variance components

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

For more information about how to cite these materials visit

For more information about how to cite these materials visit Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/

More information

Economics 620, Lecture 4: The K-Varable Linear Model I

Economics 620, Lecture 4: The K-Varable Linear Model I Economics 620, Lecture 4: The K-Varable Linear Model I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 4: The K-Varable Linear Model I 1 / 20 Consider the system

More information

Review Packet 1 B 11 B 12 B 13 B = B 21 B 22 B 23 B 31 B 32 B 33 B 41 B 42 B 43

Review Packet 1 B 11 B 12 B 13 B = B 21 B 22 B 23 B 31 B 32 B 33 B 41 B 42 B 43 Review Packet. For each of the following, write the vector or matrix that is specified: a. e 3 R 4 b. D = diag{, 3, } c. e R 3 d. I. For each of the following matrices and vectors, give their dimension.

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

Multivariate Regression

Multivariate Regression Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the

More information

CPSC 540: Machine Learning

CPSC 540: Machine Learning CPSC 540: Machine Learning Multivariate Gaussians Mark Schmidt University of British Columbia Winter 2019 Last Time: Multivariate Gaussian http://personal.kenyon.edu/hartlaub/mellonproject/bivariate2.html

More information

Non-linear least squares

Non-linear least squares Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined

More information

Approximate Dynamic Programming

Approximate Dynamic Programming Approximate Dynamic Programming A. LAZARIC (SequeL Team @INRIA-Lille) Ecole Centrale - Option DAD SequeL INRIA Lille EC-RL Course Value Iteration: the Idea 1. Let V 0 be any vector in R N A. LAZARIC Reinforcement

More information

1 EM algorithm: updating the mixing proportions {π k } ik are the posterior probabilities at the qth iteration of EM.

1 EM algorithm: updating the mixing proportions {π k } ik are the posterior probabilities at the qth iteration of EM. Université du Sud Toulon - Var Master Informatique Probabilistic Learning and Data Analysis TD: Model-based clustering by Faicel CHAMROUKHI Solution The aim of this practical wor is to show how the Classification

More information

Outline for today. Maximum likelihood estimation. Computation with multivariate normal distributions. Multivariate normal distribution

Outline for today. Maximum likelihood estimation. Computation with multivariate normal distributions. Multivariate normal distribution Outline for today Maximum likelihood estimation Rasmus Waageetersen Deartment of Mathematics Aalborg University Denmark October 30, 2007 the multivariate normal distribution linear and linear mixed models

More information

Use of sparse matrix absorption in animal breeding

Use of sparse matrix absorption in animal breeding Original article Use of sparse matrix absorption in animal breeding B. Tier S.P. Smith University of New England, Anirreal Genetics and Breeding Unit, Ar!nidale, NSW 2351, Australia (received 1 March 1988;

More information

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction

ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction Acta Math. Univ. Comenianae Vol. LXV, 1(1996), pp. 129 139 129 ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES V. WITKOVSKÝ Abstract. Estimation of the autoregressive

More information

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)

1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?) Math 35 Exam Review SOLUTIONS Overview In this third of the course we focused on linear learning algorithms to model data. summarize: To. Background: The SVD and the best basis (questions selected from

More information

Lecture 2: Linear and Mixed Models

Lecture 2: Linear and Mixed Models Lecture 2: Linear and Mixed Models Bruce Walsh lecture notes Introduction to Mixed Models SISG, Seattle 18 20 July 2018 1 Quick Review of the Major Points The general linear model can be written as y =

More information

MIT Spring 2015

MIT Spring 2015 Regression Analysis MIT 18.472 Dr. Kempthorne Spring 2015 1 Outline Regression Analysis 1 Regression Analysis 2 Multiple Linear Regression: Setup Data Set n cases i = 1, 2,..., n 1 Response (dependent)

More information

Estimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1

Estimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1 Estimation and Model Selection in Mixed Effects Models Part I Adeline Samson 1 1 University Paris Descartes Summer school 2009 - Lipari, Italy These slides are based on Marc Lavielle s slides Outline 1

More information

Generalized Linear Models. Kurt Hornik

Generalized Linear Models. Kurt Hornik Generalized Linear Models Kurt Hornik Motivation Assuming normality, the linear model y = Xβ + e has y = β + ε, ε N(0, σ 2 ) such that y N(μ, σ 2 ), E(y ) = μ = β. Various generalizations, including general

More information

Matrix & Linear Algebra

Matrix & Linear Algebra Matrix & Linear Algebra Jamie Monogan University of Georgia For more information: http://monogan.myweb.uga.edu/teaching/mm/ Jamie Monogan (UGA) Matrix & Linear Algebra 1 / 84 Vectors Vectors Vector: A

More information

ECE 275A Homework 6 Solutions

ECE 275A Homework 6 Solutions ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =

More information

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38

Preliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38 Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters

More information

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research

Linear models. Linear models are computationally convenient and remain widely used in. applied econometric research Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y

More information

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2

Multilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do

More information

Advanced Econometrics I

Advanced Econometrics I Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics

More information