Likelihood Methods. 1 Likelihood Functions. The multivariate normal distribution likelihood function is
|
|
- Horatio Griffith
- 5 years ago
- Views:
Transcription
1 Likelihood Methods 1 Likelihood Functions The multivariate normal distribution likelihood function is The log of the likelihood, say L 1 is Ly = π.5n V.5 exp.5y Xb V 1 y Xb. L 1 = 0.5[N lnπ + ln V +y Xb V 1 y Xb]. The term N lnπ is a constant that does not involve any of the unknown variances or effects in the model, therefore, it is commonly omitted during maximization computations. Maximizing the log likelihood maximizes the original likelihood function. Previously, therefore, V = R Z R 1 Z + G 1 G, ln V = ln R + ln G + ln Z R 1 Z + G 1. If R = Iσ 0, then ln R = ln Iσ 0 = lnσ 0 N I = N ln σ 01. Similarly, if G = + Iσi, where i = 1 to s, then ln G = ln Iσi = q i ln σi. Except, that in animal models one of the G i is equal to Aσi. In that case, which is ln Aσ i = lnσ i q i A ln Aσ i = q i ln σ i A = q i ln σ i + ln A. Recall that so that X C = R 1 X X R 1 Z Z R 1 X Z R 1 Z + G 1, C = Z R 1 Z + G 1 X V 1 X ln C = ln Z R 1 Z + G 1 + ln X V 1 X. 1
2 Maximum Likelihood Method Hartley Rao 1967 described the maximum likelihood approach for the estimation of variance components. Let L be equivalent to L 1 except for the constant involving π. L = 0.5[ln V +y Xb V 1 y Xb]. The derivatives of L with respect to b to σ i for i = 0, 1,... s are L b = X V 1 Xb X V 1 y L σ i =.5 tr[v 1 V/ σ i ] +.5y Xb V 1 V/ σ i V 1 y Xb =.5 tr[v 1 V i ] +.5y Xb V 1 V i V 1 y Xb Equating the derivatives to zero gives Recall that ˆb = X V 1 X X V 1 y, tr[v 1 V i ] = y Xˆb V 1 V i V 1 y Xˆb. Py = V 1 y Xˆb, where P is the projection matrix, that V i = Z i Z i, then tr[v 1 Z i Z i] = y PV i Py. In usual mixed model theory, the solution vector for a rom factor may be written as so that Also, Let û i = G i Z ipy, y PV i Py = y PZ i G i G i G i Z ipy = û ig i û i = û iûi/σi 4. tr[v 1 V i ] = tr[r 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 Z i Z i]. T = Z R 1 Z + G 1 1, R = Iσ 0, G = + Iσ i,
3 then tr[v 1 V i ] = trz iz i σ 0 trz iztz Z i σ 4 0. If T can be partitioned into submatrices for each rom factor, then Tσ 0 Z Z + + Iαi = I, + TZ Zσ0 = I T Iσ i, TZ Z i σ 0 = I T ii σ i, which yields trz iztz Z i σ0 4 = trz iz i σ0 tri T ii σ i σi. Finally, Combining results gives for i = 1,,..., s, for i = 0 gives tr[v 1 V i ] = tri T ii σi σi = triσ i = q i σ i trt ii σ 4 i trt ii σ 4 i. ˆσ i = û iûi + trt iiˆσ 0/q i ˆσ 0 = y y ˆb X y û Z y/ N..1 The EM Algorithm EM sts for Expectation Maximization. The procedure alternates between calculating conditional expected values maximizing simplified likelihoods. The actual data y are called the incomplete data in the EM algorithm, the complete data are considered to be y the unobservable rom effects, u i. If the realized values of the unobservable rom effects were known, then their variance would be the average of their squared values, i.e., ˆσ i = u iu i /q i. However, in real life the realized values of the rom effects are unknown. The steps of the EM algorithm are as follows: Step 0. Decide on starting values for the variances set m = 0. 3
4 Step 1.E-step Calculate the conditional expectation of the sufficient statistics, conditional on the incomplete data. Eu iu i y = σ 4m i y P m Z i Z ip m y +trσ m i I σ 4m i Z iv m 1 Z i = ˆt m i Step.M-step Maximize the likelihood of the complete data, σ m+1 i = ˆt m i /q i, i = 0, 1,,..., s. Step 3. If convergence is reached, set ˆσ = σ m+1, otherwise increase m by one return to Step 1. This is equivalent to constructing solving the mixed model equations with a given set of variances, σ m, then σ m+1 0 = y y ˆb X y û Z y/n, σ m+1 i = û iûi + σ m+1 0 trt ii /q i. 4
5 3 Restricted Maximum Likelihood Method Restricted or Residual maximum likelihood REML, was first suggested by Thompson 196, described formally by Patterson Thompson The procedure requires that y has a multivariate normal distribution. The method is translation invariant. The maximum likelihood approach automatically keeps the estimator within the allowable parameter spacei.e. zero to plus infinity, therefore, REML is a biased procedure. REML was proposed as an improvement to ML in order to account for the degrees of freedom lost in estimating fixed effects. The likelihood function used in REML is that for a set of error contrasts i.e. residuals assumed to have a multivariate normal distribution. The multivariate normal distribution likelihood function for the residual contrasts, K y, where K X = 0, K has rank equal to N rx, is LK y = π.5n rx K VK.5 exp.5k y K VK 1 K y. The natural log of the likelihood function is L 3 =.5N rx lnπ.5 ln K VK.5y KK VK 1 K y. Notice that.5n rx lnπ is a constant that does not depend on the unknown variance components or factors in the model, therefore, can be ignored to give L 4. Searle 1979 showed that ln K VK = ln V + ln X V 1 X y KK VK 1 K y = y Py = y Xˆb V 1 y Xˆb for any K such that K X = 0. Hence, L 4 can be written as L 4 =.5 ln V.5 ln X V 1 X.5y Xˆb V 1 y Xˆb. REML can be calculated a number of different ways. 1. Derivative Free approach is a search technique to find the parameters that maximize the log likelihood function. Two techniques will be described here.. First Derivatives EM is where the first derivatives of the log likelihood are determined set to zero in order to maximize the likelihood function. Solutions need to be obtained by iteration because the resulting equations are non linear. 3. Second Derivatives are generally more computationally deming. Gradient methods are used to find the parameters that make the first derivatives equal to zero. Newton- Raphson involves the observed information matrix Fishers Method of Scoring involves the expected information matrix have been used. Lately, the average information algorithm averages the observed expected information matrices has been used to reduce the computational time. 5
6 3.1 Example Problem All of the approaches attempt to maximize the log likelihood function of the error contrasts. To illustrate the methods, consider a single trait model with three factors F, A, B, of which A B are rom factors. There were a total of 90 observations, the total sum of squares was 356,000. The least squares equations for this small example are shown below F 1 F A 1 A A 3 B 1 B B 3 B 4 = Derivative Free REML Imagine an s dimensional array containing the values of the likelihood function for every possible set of values of the ratios of the components to the residual variance. The technique is to search this array find the set of ratios for which the likelihood function is maximized. There is more than one way to conduct the search. Care must be taken to find the global maximum rather than one of possibly many local maxima. At the same time the number of likelihood evaluations to be computed must also be minimized. Various alternative forms of L 4 can be derived. Note that that that combining these results gives Now note that ln V = ln R + ln G + ln G 1 + Z R 1 Z ln X V 1 X = ln C ln Z R 1 Z + G 1 L 4 =.5 ln R.5 ln G.5 ln C.5y Py. ln R = ln Iσ 0 = N ln σ0, ln G = q i ln σi, ln C = ln X R 1 X + ln Z SZ + G 1 6
7 Then where ln X R 1 X = ln X Xσ0 = lnσ0 X X = ln X X rx ln σ0, Z SZ + G 1 = σ0 MZ + G 1 = σ 0 Z MZ + G 1 σ 0. ln C = ln X X rx ln σ 0 q ln σ 0 + ln Z MZ + G 1 σ 0, finally, the log-likelihood function becomes L 4 =.5N rx q ln σ0.5 q i ln σi.5 ln C.5y Py, where Note that C = X X X Z Z X Z Z + G 1 σ 0. q i ln σi = q i ln σ0/α i = q i ln σ0 ln α i so that L 4 =.5[N rx ln σ0 q i ln α i + ln C +y Py]. The quantity y Py is y y Xˆb Zû/σ0. The computations are achieved by constructing the following matrix, X X X Z X y Z X Z Z + G 1 σ0 Z y y X y Z y y = C W y y W y, y then by Gaussian elimination of one row at a time, the sum of the log of the non-zero pivots using the same ordering for each evaluation of the likelihood gives log C y y Xˆb Zû. Gaussian elimination, using sparse matrix techniques, requires less computing time than inverting the coefficient matrix of the mixed model equations. The ordering of factors within the equations could be critical to the computational process some experimentation may be necessary to determine the best ordering. The likelihood function can be evaluated without the calculation of solutions to the mixed model equations, without inverting the coefficient matrix of the mixed model equations, without computing any of the σi. The formulations for more general models multiple trait models are more complex, but follow the same ideas. Searching the array of likelihood values for various values of α i can be done in several different ways. One method is to fix the values of all but one of the s α i, then evaluate L for four 7
8 or more different values of the α i that were not fixed. Then one can use a quadratic regression analysis to determine the value of that one ratio which maximizes L given that the other ratios are fixed. This is repeated for each of the s ratios, the process is repeated until a maximum likelihood is obtained. The calculations are demonstrated in the example that follows. Begin by fixing the value of α B = 10 letting the value of α A take on the values of 5, 10, 0, 30, 40. Using L 4 to evaluate the likelihood, then the results were as follows: α A L For example, the likelihood value for α A = 40, would be L 4 = 1 [N rx ln σ 0 q A ln α A q B ln α B + ln C +y y Xˆb Zû/σ0] where ln C = , y Py = /σ0 = 88, q A ln α A = , q B ln α B = , σ0 = , ln σ0 = , N rx = 88, then L 4 = 0.5[ / ] = To find the value of α A that maximizes L 4 for α B = 10, let Q = Y = then ˆβ = Q Q 1 Q Y =
9 From this a prediction equation for L 4 can be written as L 4 = α A α A. This equation can be differentiated with respect to α A then equated to zero to find the value of the ratio that maximizes the prediction equation. This gives α A =.04489/ = Now keep α A = try a number of values of α B from to 10, which give the following results. α B L Applying the quadratic regression to these points gives α B = The next step would be to fix α B = 1.65 to try new values for α A, such as 5 to 40 by units of 1. The range of values becomes finer finer. To insure that one has found the global maximum, the entire process could be started with vastly different starting values for the ratios, such as α B = 50 let values for α A be 40, 50, 60, 70. The more components there are to estimate, the more evaluations of the likelihood that are going to be needed, the more probable that convergence might be to a local maximum rather than to the global maximum. Please refer to the literature for specification of the log likelihood function for particular models situations The Simplex Method The Simplex Method Nelder Mead, 1965 is a procedure for finding the minimum of a function i.e. the minimum of L 4 or the maximum of L 4 with respect to the unknown variances covariances. The best way to describe the method is using the example data from 9
10 the previous sections. Begin by constructing a set of points for which L 4 is to be evaluated. A point is a vector of values for the unknowns α A, α B, for example, θ 1 = , then form two more points by changing one unknown at a time. Let the three points be as shown in the following table. No. α A α B Now calculate L 4 for each point arrange from largest to lowest value. No. α A α B L The idea now is to find another point to replace the last onelowest L 4. This is done by a process called reflection. Compute the mean of all points excluding the one with the lowest L 4. θ m = , then the reflection step is θ 4 = θ m + r θ m θ last, where r is recommended by Nelder Mead 1965 to be 1, giving θ 4 = The corresponding L 4 for this point was Compared to those in the table it has the largest value, therefore, is a better point than the other three. No. α A α B L Given this success, the Simplex method calls for an expansion step, i.e. to make a bigger change. Thus, θ 5 = θ m + E θ 4 θ m, 10
11 where E is suggested to be equal to. Hence θ 5 = Then L 4 = , the exped point is better yet. Now drop θ 1 from the table put θ 5 at the top. No. α A α B L This completes one iteration. Begin the next iteration by computing the mean of all points excluding the point with the lowest L 4. θ m = Another reflection step gives θ 6 = θ m + r θ m θ last, = However, this gives L 4 = , which is between θ θ 4, can push out θ from the table. No. α A α B L Instead of an expansion step, a contraction step is needed because θ 6 did not give a greater L 4 than the first two. Thus, θ 7 = θ m + c θ 6 θ m, where c = 0.5 is recommended. Hence, θ 7 = Then L 4 = is better than that given by θ 4, but not by θ 5, thus the new table becomes as follows:. No. α A α B L
12 The following steps were taken in the next iteration. 1. The mean of the top two L 4 is θ m = A reflection step gives θ 8 = θ m + r θ m θ last, = , which gave L 4 = , which is better than θ Add θ 8 to the table drop θ 4. No. α A α B L Because L 4 for θ 8 was not larger than L 4 for θ 5 or smaller than L 4 for θ 7, then no expansion or contraction step is necessary. Begin the next iteration. The Simplex method continues in this manner until all point entries in the table are equal. The constants recommended by Nelder Mead 1965 for reflection, expansion, contraction could be adjusted for a particular data set. This method may converge to a local maximum, so different starting values are needed to see if it converges to the same point. The Simplex method does not work well with a large number of parameters to be estimated. 3. First Derivatives EM Algorithm To derive formulas for estimating the variance components take the derivatives of L 4 with respect to the unknown components. L 4 σ i 1 V =.5trV σi +.5y Xˆb V.5trX V 1 X X V 1 V σi V 1 y Xˆb 1 V σi V 1 X Combine the two terms involving the traces note that V 1 y Xˆb = Py, 1
13 then L 4 σ i for i = 1,..., s or =.5trV 1 V 1 XX V 1 X X V 1 V σ i =.5trPZ i Z i +.5y PZ i Z ipy =.5trP +.5y PPy for i = 0 for the residual component. Using P the fact that V 1 = R 1 R 1 ZZ R 1 Z + G 1 1 Z R 1 +.5y P V σi Py then trpz i Z i = q i /σi trc ii σ0/σ i 4 trp = N rxσ0 û iûi/σi. The other terms, y PZ i Z i Py y PPy, were simplified by Henderson 1973 to show that they could be calculated from the Mixed Model Equations. Note that Henderson 1973 showed then which when G i = Iσ i gives Py = V 1 y Xˆb, ˆb = X V 1 X X V 1 y, û i = G i Z ipy, y PZ i Z ipy = y PZ i [G i G 1 i = y PZ i G i G i = û ig i û i û iûi/σ 4 i. Similarly for the residual component, Henderson showed that where α i = σ 0 /σ i. G 1 i G i ]Z ipy G i Z ipy y PPy = [y y ˆb X y û iz iy + û iûiα i ]/σ0, Equate the derivatives to zero incorporating the above simplifications obtain ˆσ i = û iûi + trc ii σ 0/q i, ˆσ 0 = y Py/N rx. As with ML, solutions using the EM algorithm must be computed iteratively. Convergence is usually very slow, if it occurs, the process may also diverge. 13
14 Notice the differences between REML ML. The denominator for ˆσ 0 is N rx rather than N, in ˆσ i is trc ii rather than trt ii. The quadratic forms, however, are identical in REML ML. Accounting for the degrees of freedom to estimate b has resulted in the REML algorithm. A major computing problem with the EM algorithm is the calculation of trc ii, which is the corresponding inverse elements of the mixed model equations for the i th rom factor. With most applications in animal breeding, the order of the mixed model equations are too large to be inverted, solutions to the equations are obtained by Gauss-Seidel iterations. However, there have been several attempts to approximate trc ii, but these have not been totally suitable. To demonstrate the EM algorithm let α A = 10 α B = 5 be the starting values of the ratios for factors A B, respectively. There were N = 90 total observations, rx =. The solution vector is Then F 1 F A 1 A A 3 B 1 B B 3 B 4 = from the inverse of the coefficient matrix, which give rise to the following estimates, New ratios are formed as y Xˆb + Zû = 347, trc AA =.16493, trc BB = ˆσ 0 = 356, , /88 = , ˆσ A = /3 = , ˆσ B = /4 = α A = / = , α B = / =
15 these are used to form the mixed model equations again, new solutions traces are calculated, so on, until the estimated ratios the prior values of the ratios are equal. The estimates converge to ˆσ 0 = , ˆσ A =.569, ˆσ B = or α A = , α B = Second Derivatives, Average Information Second derivatives of the log likelihood lead to the expectations of the quadratic forms. One technique, MIVQUE Minimum Variance Quadratic Unbiased Estimation equates the quadratic forms to their expectations. The estimates are unbiased if all variances remain positive, then convergence will be to the REML estimates. However, due to a shortage of data or an inappropriate model, the estimates derived in this manner can be negative. Computing the expectations of the quadratic forms requires the inverse of the mixed model equations coefficient matrix, then products crossproducts of various parts of the inverse. A gradient method using first second derivatives can be used Hofer, The gradient, d the vector of first derivatives of the log likelihood, is used to determine the direction towards the parameters that give the maximum of the log likelihood, such that θ t+1 = θ t + M t d t, where d t are the first derivatives evaluated at θ = θ t, M t in the Newton-RaphsonNR algorithm is the observed information matrix, in the Fisher Method of ScoringFS it is the expected information matrix. The first derivatives are as follows from earlier in these notes: L 4 σ i =.5trPZ i Z i +.5y PZ i Z ipy = 0 for i = 1,..., s or L 4 σ0 =.5trP +.5y PPy = 0 for the residual component. Then from earlier results, trpz i Z i = q i /σ i trc ii σ 0/σ 4 i, y PZ i Z ipy = û iûi/σ 4 i 15
16 which combined give 0.5û iûi/σi 4 q i /σi + trc ii σ0/σ i 4 = 0, for i = 1,..., s, trp = N rxσ0 û iûi/σi y PPy = [y y ˆb X y û iz iy + û iûiα i ]/σ0 which combined give 0.5[y y ˆb X y û iz iy + û iûiα i ]/σ0 N rxσ0 + û iûi/σi = 0, which simplifies to 0.5[y y ˆb X y û iz iy]/σ0 N rxσ0 = 0. The second derivatives give a matrix of quantities. The elements of the observed information matrix Gilmour et al are L 4 σ i σ 0 = 0.5y PZ i Z ipy/σ 4 0, L 4 σ i σ j L 4 σ 0 σ 0 = 0.5trPZ i Z j 0.5trPZ i Z ipz j Z j +y PZ i Z ipz j Z jpy/σ 0 0.5y PZ i Z jpy/σ 0, = y Py/σ N rx/σ 4 0. The elements of the expected information matrix Gilmour et al are E[ L 4 σi ] = 0.5trPZ i Z i/σ0, σ 0 E[ L 4 σi ] σ j = 0.5trPZ i Z ipz j Z j, E[ L 4 σ0 ] = 0.5N rx/σ0. 4 σ 0 As the name Average Information implies, average the observed expected information matrices to give the following matrix of elements. I[σ i, σ 0] = 0.5y PZ i Z ipy/σ 4 0, I[σ i, σ j ] = y PZ i Z ipz j Z jpy/σ 0, I[σ 0, σ 0] = 0.5y Py/σ
17 The first derivatives form the vector, d t, M t = I[σ, σ] 1. The rest of this method is computational detail to simplify the requirements for inverse elements solutions to MME. The calculations can not be illustrated very easily for the example data because the y-vector is not available. 3.4 Animal Models The model commonly applied to estimation of variance components in livestock genetics since 1989 has been an animal model. The animal model assumes a large, rom mating population, an infinite number of loci each with a small equal effect on the trait, only additive genetic effects, all relationships among animals are known tracible to an unselected base population somewhere in the past. Animals may have more than one record each. The equation of the model is y = Xb + Za + Zp + e, where a is the vector of animal additive genetic effects one per animal, p is a vector of permanent environmental p.e. effects associated with each animal. V ar Ey = Xb, a p = e Aσa Iσp Iσe The matrix A is called the numerator relationship matrix. Wright defined relationships among animals as correlations, but A is essentially relationships defined as covariances the numerators of the correlation coefficients. Also, these only represent the additive genetic relationships between animals. The MME for this model are X X X Z X Z Z X Z Z + A 1 k a Z Z Z X Z Z Z Z + Ik p ˆb â ˆp = Note that k a is the ratio of residual to additive genetic variances, k p is the ratio of residual to permanent environmental variances. Also, in MME the inverse of A is required. The EM-REML procedure gives. X y Z y Z y ˆσ e = y y ˆb X y â Z y ˆp Z y/n rx, ˆσ a = â A 1 â + tra 1 C aaˆσ e/n, ˆσ p = ˆp ˆp + trc ppˆσ e/n,. 17
18 where n is the total number of animals, N is the total number of records, C aa are the inverse elements of the MME for the animal additive genetic effects, C pp are the inverse elements of the MME for the animal permanent environmental effects. An example of this model will be given in later notes Quadratic Forms in an Animal Model A necessary quadratic form in an animal model is â A 1 â, this can be computed very easily. Note that the inverse of A may be written as A 1 = T 1 D T 1, where T 1 is an upper triangular matrix, diagonal matrix D has elements equal to 1,, or 4/3 in noninbred situations, values greater than in inbred situations. In Henderson 1975, this inverse was shown to be composed of just three numbers, i.e. 0, 1 s on the diagonals, -.5 corresponding to the parents of an animal. For example, T 1 = Then T 1â = ˆm = â i 0.5â s + â d, for the i th animal, â s â d are the sire dam estimated breeding values, respectively. Consequently, â A 1 â = â T 1 B T 1â = ˆm B ˆm q = ˆm i b ii, where b ii are the diagonal elements of B, q is the number of animals. 4 EXERCISES Below are pedigrees data on 0 animals. 18
19 Animal Sire Dam Group Record Construct A 1 set up the MME.. Apply EM-REML to the model, y ij = µ + g i + a j + e ij, where group, animal additive genetic, residual effects are rom. Let σ e/σ a = 1.5, σ e/σ g = 5.0, to start the iterations. Do five iterations of EM-REML. 3. Apply EM-REML again to the model with σ e/σ a = 3.0, σ e/σ g =.5, do five iterations of REML from these starting values. 4. Do the results of the previous two question tend to give similar answers? Comment on the results. 5. Use the Derivative Free method compute the log likelihoods for various sets of parameters. 19
Estimation of Variance Components in Animal Breeding
Estimation of Variance Components in Animal Breeding L R Schaeffer July 19-23, 2004 Iowa State University Short Course 2 Contents 1 Distributions 9 11 Random Variables 9 12 Discrete Random Variables 10
More informationLecture Notes. Introduction
5/3/016 Lecture Notes R. Rekaya June 1-10, 016 Introduction Variance components play major role in animal breeding and genetic (estimation of BVs) It has been an active area of research since early 1950
More informationChapter 11 MIVQUE of Variances and Covariances
Chapter 11 MIVQUE of Variances and Covariances C R Henderson 1984 - Guelph The methods described in Chapter 10 for estimation of variances are quadratic, translation invariant, and unbiased For the balanced
More informationRepeated Records Animal Model
Repeated Records Animal Model 1 Introduction Animals are observed more than once for some traits, such as Fleece weight of sheep in different years. Calf records of a beef cow over time. Test day records
More informationReduced Animal Models
Reduced Animal Models 1 Introduction In situations where many offspring can be generated from one mating as in fish poultry or swine or where only a few animals are retained for breeding the genetic evaluation
More informationMIXED MODELS THE GENERAL MIXED MODEL
MIXED MODELS This chapter introduces best linear unbiased prediction (BLUP), a general method for predicting random effects, while Chapter 27 is concerned with the estimation of variances by restricted
More informationEstimation of Variances and Covariances
Estimation of Variances and Covariances Variables and Distributions Random variables are samples from a population with a given set of population parameters Random variables can be discrete, having a limited
More informationChapter 12 REML and ML Estimation
Chapter 12 REML and ML Estimation C. R. Henderson 1984 - Guelph 1 Iterative MIVQUE The restricted maximum likelihood estimator (REML) of Patterson and Thompson (1971) can be obtained by iterating on MIVQUE,
More informationAnimal Models. Sheep are scanned at maturity by ultrasound(us) to determine the amount of fat surrounding the muscle. A model (equation) might be
Animal Models 1 Introduction An animal model is one in which there are one or more observations per animal, and all factors affecting those observations are described including an animal additive genetic
More informationMixed-Model Estimation of genetic variances. Bruce Walsh lecture notes Uppsala EQG 2012 course version 28 Jan 2012
Mixed-Model Estimation of genetic variances Bruce Walsh lecture notes Uppsala EQG 01 course version 8 Jan 01 Estimation of Var(A) and Breeding Values in General Pedigrees The above designs (ANOVA, P-O
More information3. Properties of the relationship matrix
3. Properties of the relationship matrix 3.1 Partitioning of the relationship matrix The additive relationship matrix, A, can be written as the product of a lower triangular matrix, T, a diagonal matrix,
More informationMaternal Genetic Models
Maternal Genetic Models In mammalian species of livestock such as beef cattle sheep or swine the female provides an environment for its offspring to survive and grow in terms of protection and nourishment
More informationMixed-Models. version 30 October 2011
Mixed-Models version 30 October 2011 Mixed models Mixed models estimate a vector! of fixed effects and one (or more) vectors u of random effects Both fixed and random effects models always include a vector
More informationREML Variance-Component Estimation
REML Variance-Component Estimation In the numerous forms of analysis of variance (ANOVA) discussed in previous chapters, variance components were estimated by equating observed mean squares to expressions
More informationLecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values. Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 2013
Lecture 5: BLUP (Best Linear Unbiased Predictors) of genetic values Bruce Walsh lecture notes Tucson Winter Institute 9-11 Jan 013 1 Estimation of Var(A) and Breeding Values in General Pedigrees The classic
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More informationAnimal Model. 2. The association of alleles from the two parents is assumed to be at random.
Animal Model 1 Introduction In animal genetics, measurements are taken on individual animals, and thus, the model of analysis should include the animal additive genetic effect. The remaining items in the
More informationAlternative implementations of Monte Carlo EM algorithms for likelihood inferences
Genet. Sel. Evol. 33 001) 443 45 443 INRA, EDP Sciences, 001 Alternative implementations of Monte Carlo EM algorithms for likelihood inferences Louis Alberto GARCÍA-CORTÉS a, Daniel SORENSEN b, Note a
More informationVARIANCE COMPONENT ESTIMATION & BEST LINEAR UNBIASED PREDICTION (BLUP)
VARIANCE COMPONENT ESTIMATION & BEST LINEAR UNBIASED PREDICTION (BLUP) V.K. Bhatia I.A.S.R.I., Library Avenue, New Delhi- 11 0012 vkbhatia@iasri.res.in Introduction Variance components are commonly used
More informationBest unbiased linear Prediction: Sire and Animal models
Best unbiased linear Prediction: Sire and Animal models Raphael Mrode Training in quantitative genetics and genomics 3 th May to th June 26 ILRI, Nairobi Partner Logo Partner Logo BLUP The MME of provided
More informationChapter 5 Prediction of Random Variables
Chapter 5 Prediction of Random Variables C R Henderson 1984 - Guelph We have discussed estimation of β, regarded as fixed Now we shall consider a rather different problem, prediction of random variables,
More informationChapter 3 Best Linear Unbiased Estimation
Chapter 3 Best Linear Unbiased Estimation C R Henderson 1984 - Guelph In Chapter 2 we discussed linear unbiased estimation of k β, having determined that it is estimable Let the estimate be a y, and if
More informationLecture 9 Multi-Trait Models, Binary and Count Traits
Lecture 9 Multi-Trait Models, Binary and Count Traits Guilherme J. M. Rosa University of Wisconsin-Madison Mixed Models in Quantitative Genetics SISG, Seattle 18 0 September 018 OUTLINE Multiple-trait
More informationRestricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model
Restricted Maximum Likelihood in Linear Regression and Linear Mixed-Effects Model Xiuming Zhang zhangxiuming@u.nus.edu A*STAR-NUS Clinical Imaging Research Center October, 015 Summary This report derives
More informationLecture 3: Linear Models. Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012
Lecture 3: Linear Models Bruce Walsh lecture notes Uppsala EQG course version 28 Jan 2012 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector of observed
More informationInverse of a Square Matrix. For an N N square matrix A, the inverse of A, 1
Inverse of a Square Matrix For an N N square matrix A, the inverse of A, 1 A, exists if and only if A is of full rank, i.e., if and only if no column of A is a linear combination 1 of the others. A is
More informationOutline of GLMs. Definitions
Outline of GLMs Definitions This is a short outline of GLM details, adapted from the book Nonparametric Regression and Generalized Linear Models, by Green and Silverman. The responses Y i have density
More informationSTAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method.
STAT 135 Lab 13 (Review) Linear Regression, Multivariate Random Variables, Prediction, Logistic Regression and the δ-method. Rebecca Barter May 5, 2015 Linear Regression Review Linear Regression Review
More informationChapter 3: Maximum Likelihood Theory
Chapter 3: Maximum Likelihood Theory Florian Pelgrin HEC September-December, 2010 Florian Pelgrin (HEC) Maximum Likelihood Theory September-December, 2010 1 / 40 1 Introduction Example 2 Maximum likelihood
More informationRegression. Oscar García
Regression Oscar García Regression methods are fundamental in Forest Mensuration For a more concise and general presentation, we shall first review some matrix concepts 1 Matrices An order n m matrix is
More information1 Mixed effect models and longitudinal data analysis
1 Mixed effect models and longitudinal data analysis Mixed effects models provide a flexible approach to any situation where data have a grouping structure which introduces some kind of correlation between
More informationEconomics 620, Lecture 5: exp
1 Economics 620, Lecture 5: The K-Variable Linear Model II Third assumption (Normality): y; q(x; 2 I N ) 1 ) p(y) = (2 2 ) exp (N=2) 1 2 2(y X)0 (y X) where N is the sample size. The log likelihood function
More informationSTAT 100C: Linear models
STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix
More informationLecture 32: Infinite-dimensional/Functionvalued. Functions and Random Regressions. Bruce Walsh lecture notes Synbreed course version 11 July 2013
Lecture 32: Infinite-dimensional/Functionvalued Traits: Covariance Functions and Random Regressions Bruce Walsh lecture notes Synbreed course version 11 July 2013 1 Longitudinal traits Many classic quantitative
More informationLinear Methods for Prediction
Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we
More informationRaphael Mrode. Training in quantitative genetics and genomics 30 May 10 June 2016 ILRI, Nairobi. Partner Logo. Partner Logo
Basic matrix algebra Raphael Mrode Training in quantitative genetics and genomics 3 May June 26 ILRI, Nairobi Partner Logo Partner Logo Matrix definition A matrix is a rectangular array of numbers set
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationLinear Models in Machine Learning
CS540 Intro to AI Linear Models in Machine Learning Lecturer: Xiaojin Zhu jerryzhu@cs.wisc.edu We briefly go over two linear models frequently used in machine learning: linear regression for, well, regression,
More informationSummary INTRODUCTION. Running head : AI-REML FOR EQUAL DESIGN MATRICES. K. Meyer
Running head : AI-REML FOR EQUAL DESIGN MATRICES An average information Restricted Maximum Likelihood algorithm for estimating reduced rank genetic covariance matrices or covariance functions for animal
More informationANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32
ANOVA Variance Component Estimation Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 32 We now consider the ANOVA approach to variance component estimation. The ANOVA approach
More informationLecture 2: Linear Models. Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011
Lecture 2: Linear Models Bruce Walsh lecture notes Seattle SISG -Mixed Model Course version 23 June 2011 1 Quick Review of the Major Points The general linear model can be written as y = X! + e y = vector
More informationMATRIX ALGEBRA AND SYSTEMS OF EQUATIONS. + + x 1 x 2. x n 8 (4) 3 4 2
MATRIX ALGEBRA AND SYSTEMS OF EQUATIONS SYSTEMS OF EQUATIONS AND MATRICES Representation of a linear system The general system of m equations in n unknowns can be written a x + a 2 x 2 + + a n x n b a
More informationLinear Algebra Review
Linear Algebra Review Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Linear Algebra Review 1 / 45 Definition of Matrix Rectangular array of elements arranged in rows and
More informationLecture 5 Basic Designs for Estimation of Genetic Parameters
Lecture 5 Basic Designs for Estimation of Genetic Parameters Bruce Walsh. jbwalsh@u.arizona.edu. University of Arizona. Notes from a short course taught June 006 at University of Aarhus The notes for this
More informationModels with multiple random effects: Repeated Measures and Maternal effects
Models with multiple random effects: Repeated Measures and Maternal effects 1 Often there are several vectors of random effects Repeatability models Multiple measures Common family effects Cleaning up
More informationLikelihood-Based Methods
Likelihood-Based Methods Handbook of Spatial Statistics, Chapter 4 Susheela Singh September 22, 2016 OVERVIEW INTRODUCTION MAXIMUM LIKELIHOOD ESTIMATION (ML) RESTRICTED MAXIMUM LIKELIHOOD ESTIMATION (REML)
More informationANOVA Variance Component Estimation. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 32
ANOVA Variance Component Estimation Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 32 We now consider the ANOVA approach to variance component estimation. The ANOVA approach
More informationEconomics 620, Lecture 4: The K-Variable Linear Model I. y 1 = + x 1 + " 1 y 2 = + x 2 + " 2 :::::::: :::::::: y N = + x N + " N
1 Economics 620, Lecture 4: The K-Variable Linear Model I Consider the system y 1 = + x 1 + " 1 y 2 = + x 2 + " 2 :::::::: :::::::: y N = + x N + " N or in matrix form y = X + " where y is N 1, X is N
More information1 Data Arrays and Decompositions
1 Data Arrays and Decompositions 1.1 Variance Matrices and Eigenstructure Consider a p p positive definite and symmetric matrix V - a model parameter or a sample variance matrix. The eigenstructure is
More informationGBLUP and G matrices 1
GBLUP and G matrices 1 GBLUP from SNP-BLUP We have defined breeding values as sum of SNP effects:! = #$ To refer breeding values to an average value of 0, we adopt the centered coding for genotypes described
More informationChapter 4: Factor Analysis
Chapter 4: Factor Analysis In many studies, we may not be able to measure directly the variables of interest. We can merely collect data on other variables which may be related to the variables of interest.
More informationJournal of Multivariate Analysis. Sphericity test in a GMANOVA MANOVA model with normal error
Journal of Multivariate Analysis 00 (009) 305 3 Contents lists available at ScienceDirect Journal of Multivariate Analysis journal homepage: www.elsevier.com/locate/jmva Sphericity test in a GMANOVA MANOVA
More informationRESTRICTED M A X I M U M LIKELIHOOD TO E S T I M A T E GENETIC P A R A M E T E R S - IN PRACTICE
RESTRICTED M A X I M U M LIKELIHOOD TO E S T I M A T E GENETIC P A R A M E T E R S - IN PRACTICE K. M e y e r Institute of Animal Genetics, Edinburgh University, W e s t M a i n s Road, Edinburgh EH9 3JN,
More informationChapter 1: Systems of linear equations and matrices. Section 1.1: Introduction to systems of linear equations
Chapter 1: Systems of linear equations and matrices Section 1.1: Introduction to systems of linear equations Definition: A linear equation in n variables can be expressed in the form a 1 x 1 + a 2 x 2
More information6-1. Canonical Correlation Analysis
6-1. Canonical Correlation Analysis Canonical Correlatin analysis focuses on the correlation between a linear combination of the variable in one set and a linear combination of the variables in another
More informationProperties of the least squares estimates
Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares
More informationLecture 15. Hypothesis testing in the linear model
14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma
More informationQuantitative characters - exercises
Quantitative characters - exercises 1. a) Calculate the genetic covariance between half sibs, expressed in the ij notation (Cockerham's notation), when up to loci are considered. b) Calculate the genetic
More informationThe equivalence of the Maximum Likelihood and a modified Least Squares for a case of Generalized Linear Model
Applied and Computational Mathematics 2014; 3(5): 268-272 Published online November 10, 2014 (http://www.sciencepublishinggroup.com/j/acm) doi: 10.11648/j.acm.20140305.22 ISSN: 2328-5605 (Print); ISSN:
More informationECON 331 Homework #2 - Solution. In a closed model the vector of external demand is zero, so the matrix equation writes:
ECON 33 Homework #2 - Solution. (Leontief model) (a) (i) The matrix of input-output A and the vector of level of production X are, respectively:.2.3.2 x A =.5.2.3 and X = y.3.5.5 z In a closed model the
More informationShould genetic groups be fitted in BLUP evaluation? Practical answer for the French AI beef sire evaluation
Genet. Sel. Evol. 36 (2004) 325 345 325 c INRA, EDP Sciences, 2004 DOI: 10.1051/gse:2004004 Original article Should genetic groups be fitted in BLUP evaluation? Practical answer for the French AI beef
More informationMultiple random effects. Often there are several vectors of random effects. Covariance structure
Models with multiple random effects: Repeated Measures and Maternal effects Bruce Walsh lecture notes SISG -Mixed Model Course version 8 June 01 Multiple random effects y = X! + Za + Wu + e y is a n x
More informationMLES & Multivariate Normal Theory
Merlise Clyde September 6, 2016 Outline Expectations of Quadratic Forms Distribution Linear Transformations Distribution of estimates under normality Properties of MLE s Recap Ŷ = ˆµ is an unbiased estimate
More informationPREDICTION OF BREEDING VALUES FOR UNMEASURED TRAITS FROM MEASURED TRAITS
Libraries Annual Conference on Applied Statistics in Agriculture 1994-6th Annual Conference Proceedings PREDICTION OF BREEDING VALUES FOR UNMEASURED TRAITS FROM MEASURED TRAITS Kristin L. Barkhouse L.
More informationPOLI 8501 Introduction to Maximum Likelihood Estimation
POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,
More information1 Appendix A: Matrix Algebra
Appendix A: Matrix Algebra. Definitions Matrix A =[ ]=[A] Symmetric matrix: = for all and Diagonal matrix: 6=0if = but =0if 6= Scalar matrix: the diagonal matrix of = Identity matrix: the scalar matrix
More information5. Best Linear Unbiased Prediction
5. Best Linear Unbiased Prediction Julius van der Werf Lecture 1: Best linear unbiased prediction Learning objectives On completion of Lecture 1 you should be able to: Understand the principle of mixed
More informationLecture 7 Correlated Characters
Lecture 7 Correlated Characters Bruce Walsh. Sept 2007. Summer Institute on Statistical Genetics, Liège Genetic and Environmental Correlations Many characters are positively or negatively correlated at
More informationMultivariate Regression Analysis
Matrices and vectors The model from the sample is: Y = Xβ +u with n individuals, l response variable, k regressors Y is a n 1 vector or a n l matrix with the notation Y T = (y 1,y 2,...,y n ) 1 x 11 x
More informationDecomposable and Directed Graphical Gaussian Models
Decomposable Decomposable and Directed Graphical Gaussian Models Graphical Models and Inference, Lecture 13, Michaelmas Term 2009 November 26, 2009 Decomposable Definition Basic properties Wishart density
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More information4 Elementary matrices, continued
4 Elementary matrices, continued We have identified 3 types of row operations and their corresponding elementary matrices. To repeat the recipe: These matrices are constructed by performing the given row
More informationComputational methods for mixed models
Computational methods for mixed models Douglas Bates Department of Statistics University of Wisconsin Madison March 27, 2018 Abstract The lme4 package provides R functions to fit and analyze several different
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More informationANALYSIS OF VARIANCE AND QUADRATIC FORMS
4 ANALYSIS OF VARIANCE AND QUADRATIC FORMS The previous chapter developed the regression results involving linear functions of the dependent variable, β, Ŷ, and e. All were shown to be normally distributed
More informationCharles E. McCulloch Biometrics Unit and Statistics Center Cornell University
A SURVEY OF VARIANCE COMPONENTS ESTIMATION FROM BINARY DATA by Charles E. McCulloch Biometrics Unit and Statistics Center Cornell University BU-1211-M May 1993 ABSTRACT The basic problem of variance components
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationFor more information about how to cite these materials visit
Author(s): Kerby Shedden, Ph.D., 2010 License: Unless otherwise noted, this material is made available under the terms of the Creative Commons Attribution Share Alike 3.0 License: http://creativecommons.org/licenses/by-sa/3.0/
More informationEconomics 620, Lecture 4: The K-Varable Linear Model I
Economics 620, Lecture 4: The K-Varable Linear Model I Nicholas M. Kiefer Cornell University Professor N. M. Kiefer (Cornell University) Lecture 4: The K-Varable Linear Model I 1 / 20 Consider the system
More informationReview Packet 1 B 11 B 12 B 13 B = B 21 B 22 B 23 B 31 B 32 B 33 B 41 B 42 B 43
Review Packet. For each of the following, write the vector or matrix that is specified: a. e 3 R 4 b. D = diag{, 3, } c. e R 3 d. I. For each of the following matrices and vectors, give their dimension.
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationMultivariate Regression
Multivariate Regression The so-called supervised learning problem is the following: we want to approximate the random variable Y with an appropriate function of the random variables X 1,..., X p with the
More informationCPSC 540: Machine Learning
CPSC 540: Machine Learning Multivariate Gaussians Mark Schmidt University of British Columbia Winter 2019 Last Time: Multivariate Gaussian http://personal.kenyon.edu/hartlaub/mellonproject/bivariate2.html
More informationNon-linear least squares
Non-linear least squares Concept of non-linear least squares We have extensively studied linear least squares or linear regression. We see that there is a unique regression line that can be determined
More informationApproximate Dynamic Programming
Approximate Dynamic Programming A. LAZARIC (SequeL Team @INRIA-Lille) Ecole Centrale - Option DAD SequeL INRIA Lille EC-RL Course Value Iteration: the Idea 1. Let V 0 be any vector in R N A. LAZARIC Reinforcement
More information1 EM algorithm: updating the mixing proportions {π k } ik are the posterior probabilities at the qth iteration of EM.
Université du Sud Toulon - Var Master Informatique Probabilistic Learning and Data Analysis TD: Model-based clustering by Faicel CHAMROUKHI Solution The aim of this practical wor is to show how the Classification
More informationOutline for today. Maximum likelihood estimation. Computation with multivariate normal distributions. Multivariate normal distribution
Outline for today Maximum likelihood estimation Rasmus Waageetersen Deartment of Mathematics Aalborg University Denmark October 30, 2007 the multivariate normal distribution linear and linear mixed models
More informationUse of sparse matrix absorption in animal breeding
Original article Use of sparse matrix absorption in animal breeding B. Tier S.P. Smith University of New England, Anirreal Genetics and Breeding Unit, Ar!nidale, NSW 2351, Australia (received 1 March 1988;
More informationON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES. 1. Introduction
Acta Math. Univ. Comenianae Vol. LXV, 1(1996), pp. 129 139 129 ON VARIANCE COVARIANCE COMPONENTS ESTIMATION IN LINEAR MODELS WITH AR(1) DISTURBANCES V. WITKOVSKÝ Abstract. Estimation of the autoregressive
More information1. Background: The SVD and the best basis (questions selected from Ch. 6- Can you fill in the exercises?)
Math 35 Exam Review SOLUTIONS Overview In this third of the course we focused on linear learning algorithms to model data. summarize: To. Background: The SVD and the best basis (questions selected from
More informationLecture 2: Linear and Mixed Models
Lecture 2: Linear and Mixed Models Bruce Walsh lecture notes Introduction to Mixed Models SISG, Seattle 18 20 July 2018 1 Quick Review of the Major Points The general linear model can be written as y =
More informationMIT Spring 2015
Regression Analysis MIT 18.472 Dr. Kempthorne Spring 2015 1 Outline Regression Analysis 1 Regression Analysis 2 Multiple Linear Regression: Setup Data Set n cases i = 1, 2,..., n 1 Response (dependent)
More informationEstimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1
Estimation and Model Selection in Mixed Effects Models Part I Adeline Samson 1 1 University Paris Descartes Summer school 2009 - Lipari, Italy These slides are based on Marc Lavielle s slides Outline 1
More informationGeneralized Linear Models. Kurt Hornik
Generalized Linear Models Kurt Hornik Motivation Assuming normality, the linear model y = Xβ + e has y = β + ε, ε N(0, σ 2 ) such that y N(μ, σ 2 ), E(y ) = μ = β. Various generalizations, including general
More informationMatrix & Linear Algebra
Matrix & Linear Algebra Jamie Monogan University of Georgia For more information: http://monogan.myweb.uga.edu/teaching/mm/ Jamie Monogan (UGA) Matrix & Linear Algebra 1 / 84 Vectors Vectors Vector: A
More informationECE 275A Homework 6 Solutions
ECE 275A Homework 6 Solutions. The notation used in the solutions for the concentration (hyper) ellipsoid problems is defined in the lecture supplement on concentration ellipsoids. Note that θ T Σ θ =
More informationPreliminaries. Copyright c 2018 Dan Nettleton (Iowa State University) Statistics / 38
Preliminaries Copyright c 2018 Dan Nettleton (Iowa State University) Statistics 510 1 / 38 Notation for Scalars, Vectors, and Matrices Lowercase letters = scalars: x, c, σ. Boldface, lowercase letters
More informationLinear models. Linear models are computationally convenient and remain widely used in. applied econometric research
Linear models Linear models are computationally convenient and remain widely used in applied econometric research Our main focus in these lectures will be on single equation linear models of the form y
More informationMultilevel Models in Matrix Form. Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2
Multilevel Models in Matrix Form Lecture 7 July 27, 2011 Advanced Multivariate Statistical Methods ICPSR Summer Session #2 Today s Lecture Linear models from a matrix perspective An example of how to do
More informationAdvanced Econometrics I
Lecture Notes Autumn 2010 Dr. Getinet Haile, University of Mannheim 1. Introduction Introduction & CLRM, Autumn Term 2010 1 What is econometrics? Econometrics = economic statistics economic theory mathematics
More information