Inference in Constrained Linear Regression
|
|
- Alexina Walker
- 5 years ago
- Views:
Transcription
1 Inference in Constrained Linear Regression by Xinyu Chen A Thesis Submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUTE In partial fulfillment of the requirements for the Degree of Master of Science in Applied Statistics by May 07 APPROVED: Professor Thelge Buddika Peiris, Major Thesis Advisor
2 Abstract Regression analyses constitutes an important part of the statistical inference and has great applications in many areas. In some applications, we strongly believe that the regression function changes monotonically with some or all of the predictor variables in a region of interest. Deriving analyses under such constraints will be an enormous task. In this work, the restricted prediction interval for the mean of the regression function is constructed when two predictors are present. I use a modified likelihood ratio test (LRT) to construct prediction intervals. Keywords: Least favorable distribution, Restricted prediction interval, Chi-barsquare distribution, Likelihood ratio test
3 Acknowledgements I would like to express my gratitude to my advisor Thelge Buddika Peiris who helped me going through the constrained statistical inference in regression and instructed me in Latex and R code. My thanks are also due to my classmate, Jinxin Tao, who gave me a lot of support and shared the knowledge with me. Also, my special thanks are extended to all the professors who imparted to me with the valuable statistical knowledge and their assistance for the past four semesters. i
4 Contents Introduction. General Linear Regression Model and Assumptions Maximum Likelihood Estimator Confidence and Prediction Intervals Constrained Statistical Inference The Basics Likelihood Ratio Test First Order Model with Two Variables 8. Model and Assumptions Inferences for β 0 + β X 0 + β X Inference for β 0 + β X 0 + β X 0 3. Test when X 0 > 0 and X 0 > Test when X 0 < 0 and X 0 < Test when X 0 > 0 and X 0 < Test when X 0 < 0 and X 0 > Formulas for The Rejection Region 33 ii
5 4. when X 0 > 0 and X 0 > when X 0 < 0 and X 0 < when X 0 > 0 and X 0 < when X 0 < 0 and X 0 > When σ is Unknown 46 6 Prediction Intervals when X 0 > 0 and X 0 > when X 0 < 0 and X 0 < when X 0 > 0 and X 0 < when X 0 < 0 and X 0 > Example When σ is known When X 0 > 0 and X 0 > When X 0 < 0 and X 0 < When X 0 > 0 and X 0 < When X 0 < 0 and X 0 > When σ is unknown When X 0 > 0 and X 0 > When X 0 < 0 and X 0 < When X 0 > 0 and X 0 < When X 0 < 0 and X 0 > A R Codes in Example 64 iii
6 List of Figures 3. The region G 0 and boundary of the rejection region Dimensional illustration of S 3, S 4 and S Dimensional illustration of S 6, S 7 and S Dimensional view of disjoint regions S 9, S 0, S, S and S H 0 and Rejection region G 03 and Rejection region M 0 and Rejection region and dual cone M Dimensional view of M 0 and Rejection region H 03 and Rejection region G 04 and Rejection region iv
7 Chapter Introduction Techniques of statistical inference under order restrictions have been used in many applications. Regression analysis constitutes a large part of them. In many applications, experimenters believe that the regression function varies monotonically with the predictor variables in some region of interest. Usually the restricted regression analysis will consider the null hypothesis of the type Rβ = r versus Rβ r, Rβ r, for some matrix R, vector r when X N(θ, V ) and V is arbitrary. I do not have to consider the case for a general V separately in this thesis because the inference problem can be restated in terms of the identity covariance matrix. (Silvapulle and Sen (004)) And linear regression analysis is simple and efficient. Techniques of linear regression have been used for many areas and for a long time. However constrained regression analysis is more suitable and reasonable for the reality. If we use general linear regression inference techniques, we would not be able to get the benefit of our assumptions. In the fields such as economics and aerospace, constrained linear regression analysis might give more precise predictions which are important than regular one. Mukerjee and Tu (995) already discussed constrained simple linear
8 regression on a single variable. Commonly higher dimensional constrained inference is needed which is more practical. Peiris and Bhattacharya (06) has developed the techniques for point and interval estimators for model parameters and the mean response for two predictor variables model Y = β 0 + β X + β X with sign constraints on slope parameters β and β. Confidence intervals reflect the goodness of fitting and prediction intervals tell us that under certain probability the future observations will fall into the estimated intervals. In this thesis, I develop the formulas for the prediction intervals for the two predictor variables model with such constraints in slope parameters.. General Linear Regression Regression analysis is a statistical process to estimate the relation between two or more variables. Regression analysis techniques are often used to help understanding how the response variables change under the variation of the predictor variables. Regression analysis has three main purposes:.describle the relation between two or more variables,.use predictor variables to control response variables, and 3.use statistical relation to make predictions. Further, they are widely used to make forecasting in many areas, like biology, business, and data science. A few examples of applications are:. The length of patient stay in a hospital in days can be predicted by utilizing the relationship between patient s age in years and the time in the hospital.. The patient s blood pressure can be predicted by utilizing the relationship between the blood pressure and body weight.
9 .. Model and Assumptions I consider a general regression model where there are several predictor variables and the regression function is linear. The model can be stated as Y i = β 0 + β X i + β X i β p X pi + ɛ i, i =,,..., n, (.) where Y i is the repsonse for the i th trial, β 0, β, β...β p are parameters, X i, X i,...x pi are the values of predictor variables in i th trial, ɛ i is a random error term and ɛ i N n (0, σ ) for all i =,,..., n.. Maximum Likelihood Estimator In model (.), let p=, then ɛ = Y i (β 0 + β X i + β X i ) N (0, σ ), [ f Yi = exp πσ ( ) ] Yi (β 0 + β X i + β X i ). σ Then likelihood function is L(β 0, β, β, σ ) = n i= = (πσ ) (πσ ) [ exp / [ σ n/ exp σ (Y i β 0 β X i β X i ) ]. n (Y i β 0 β X i β X i ) i= ] To obtain the values of ˆβ0, ˆβ, ˆβ, and ˆσ that maximize the likelihood functon L(β 0, β, β, σ ), we let corresponding first derivatives equal zero and solve those simultaneous equations for MLEs. Then the MLEs of β 0, β, and β also can be obtained in matrix form as ˆβ = (X X) X Y, 3
10 Y Y n Y where Y n =. x x x x,and X n 3 =.... x n x n..3 Confidence and Prediction Intervals Confidence Interval One of linear regression analysis objects is to estimate the mean response E(Y ). Consider a study of the relationship between patient s blood pressure (Y) and body weight (X). The mean blood pressure at high and medium levels of body weight may be one of the purposes of analyzing the effect of overweight. Let X h as the level of X for which we wish to estimate the mean repsonse E(Y h ). Then the point estimator of E(Y h ) is Ŷ h = X h ˆβ, and Ŷh follows normal distribution with mean E(Ŷh) = E(Y h ) and variance σ (Ŷh) = σ ( n + (X h X) (Xi X) ). When σ is known, Ŷ h E(Y h ) σ(ŷh) N (0, ), where N(0,) is the standard normal distribution. Therefore 00( α)% condifence interval can be known as, Ŷ h ± Z( α/)σ(ŷh), where Z( α/) is the ( α/) 00 percentile of standard normal distribution. Usually σ is unknown, we replace σ (Ŷh) with the estimated variance s (Ŷh) = 4
11 MSE( n + (X h X) (Xi X) ), and we have Ŷ h E(Y h ) s(ŷh) t v, where t ν denotes the t-distribution with ν degree s of freedom. Therefore 00( α)% condifence interval is Ŷ h ± t ( α/;v) s(ŷh), where t ( α/;v) is the 00( α)percentile of t distribution with v degrees of freedom. Prediction Interval Now we consider the prediction of a new observation Y corresponding to a given level X of the predictor variable. When σ is known, Ŷ h(new) Ŷh σ{pred} N (0, ), where σ {pred} = σ + σ {Ŷh} = σ ( + n + (X h X) (Xi X) ). So prediction interval can be obtained as Ŷ h ± Z( α/)σ{pred}, When σ is unknown, Ŷ h(new) Ŷh s{pred} t v, where s {pred} = MSE +s {Ŷh} = MSE(+ n + (X h X) (Xi X) ). So prediction interval can be obtained as Ŷ h ± t( α/; v)s{pred} 5
12 . Constrained Statistical Inference Statistical inference has been used in many fields. The needs of developing for modeling and analysis of observational or experimental data in constrained enviroments are growing. In many applications, it is reasonable to assume that there are some constraints in our statistical models which means we have more information about our model parameter space. So the models will become more efficient than those wherein constraints are ignored if we properly incorporate those information... The Basics First we consider the observations X iid N (θ, V ). In order-restricted regression analysis, it is more common to consider inference under null hyphothesis of type Rβ = r versus Rβ r, Rβ r, for some matrix R, vector r when X iid N(θ, V ). We should consider, i V is a known positive definite matrix, ii V = σ U where U is a known positive definite matrix and σ is unknown, and iii V is unknown. (Silvapulle and Sen (004)) The following are some common terms in restricted inference, Convex: A set A R P is said to be convex if and only if {λx + ( λ)y} A where x, y A and 0 < λ <. Cone: A set A R P is said to be a cone with vector x 0 if and only if x 0 +k(x x 0 ) A for every x A and k 0. Further if the vertex x 0 is the origin O, then A is a cone simply. Fenchel Dual (or negative dual) Cone: C 0 = {α : α T θ 0 for every θ C} is called the dual cone of C with respect to the inner product. It can be shown that the boundaries of C 0 are the perpendiculars to the boundaries of C. Maximum likelihood estimation: 6
13 If X = (X, X ) N (θ, I), where I is the identity matrix and θ = (θ, θ ). Then for a single observation X, the kernel l(θ) of the loglikelihood is given by l(θ) = {(X θ ) + (X θ ) } = X θ, So in my work, I only use the kernel of the likelihood function to discuss our model... Likelihood Ratio Test Here is an simple example for likelihood ratio test. Let X = (X, X ) N (θ, I), where I is the identity matrix and θ = (θ, θ ). Consider the likelihood ratio test of H 0 : θ = θ = 0 vs H : θ 0, θ 0, LRT = X X θ, where θ {(θ, θ ) θ 0, θ 0} Then P r(lrt c) = = 4 P r(lrt c and X Q i ) i= 4 P r(lrt c X Q i )P r(x Q i ) i= where Q = {θ, θ : θ > 0, θ > 0}, Q = {θ, θ : θ < 0, θ > 0}, Q 3 = {θ, θ : θ < 0, θ < 0}, Q 4 = {θ, θ : θ > 0, θ < 0}. The null distribution of the LRT is the weighted sum of chi-square distributions, known as the chi-bar-square distribution. I use the similar method in the following chapter for hypothesis tests. 7
14 Chapter First Order Model with Two Variables. Model and Assumptions Consider the normal linear regression model with two predictor variables Y i = β 0 + β X i + β X i + ɛ i i =,,..., n, (.) or Y = Xβ + ɛ, where Y x x Y x x Y n =, X n 3 =, β 3 =.... Y n x n x n β 0 β β ɛ, ɛ ɛ n =,. and {ɛ i } are iid N (0, σ ). Let ˆβ 0, ˆβ and ˆβ be the unrestricted maximum likelihood estimaros (MLEs) of β 0, β and β respectively. Let S X = Xi, S X = Xi 8 ɛ n
15 and S = (Y i ˆβ 0 ˆβ X i ˆβ X i ) /v where v = n 3. Then we assume Xi = 0, X i = 0 and X i X i = 0 to simplify our model. Then, n cov( ˆβ) = σ n i= X i n i= X i n i= X i n i= X i n i= X ix i n i= X i n 0 0 n i= X ix i = 0 0, 0 0 S X n i= X i and so ˆβ 0, ˆβ and ˆβ are independent. Further cov( ˆβ, Y X ˆβ) =cov ( (X X) X Y, Y X(X X) X Y ) =cov ( (X X) X Y, (I n X(X X) X )Y ) =(X X) X (σ I)(I n X(X X) X ) =σ ( (X X) X (X X) X X(X X) X ) = 0. So ˆβ and Y X ˆβ are independent. Then ˆβ and S = (Y X ˆβ) (Y X ˆβ)/v are independent. Thus ˆβ 0, ˆβ, ˆβ and S are mutually independent. Following the properties of multivariate normal distribution, ˆβ0, ˆβ, ˆβ and S have normal distributions. They are unbiased estimators for β 0, β, β and σ. Hence ˆβ 0 N (β 0, σ /n), ˆβ N (β, σ / ), ˆβ N (β, σ /S X ), vs /σ χ v, where v = n 3. We consider the sign constraints for β and β. First I consider, β 0 and β 0 (.) 9
16 We can always make transformations of predictor variables for other constraints of β. The restricted MLEs of β 0, β, and β under the constraint (.) are given by β 0 = ˆβ 0, β = ˆβ + = max{ ˆβ, 0}, β = ˆβ + = max{ ˆβ, 0}, which is obvious and reasonable.. Inferences for β 0 + β X 0 + β X 0 We consider inferences for the mean response E(Y ) = β 0 + β X 0 + β X 0, for given point (X 0, X 0 ). For example, we already known the length of patient stay in a hospital in days (Y ) depends on the patient s age in years (X 0 ) and the infection risk (X 0 ). Given a patient s age and the infection risk, we want to know how long the patient will stay in hospital. Peiris and Bhattacharya (06) have already proposed formulas for the confidence interval for different signes of X 0 and X 0. 0
17 Chapter 3 Inference for β 0 + β X 0 + β X 0 Mentioned by Cox and Hinkley (97), inverting one-sdied tests for two-sample problems can derive the correct ( α)-coefficient prediction intervals for a new observation Y. Hence I derive the prediction intervals fro Y by inverting one-sided tests that test whether µ = E[Y (X 0, X 0 )] excceds or is exceeded by β 0 + β X 0 + β X 0, where the estimate of µ is to be obtained from the future observation Y and the estimate of β 0 + β X 0 + β X 0 is obtained from the observations in the past. Here there are four possible cases based on the signs of X 0 and X Test when X 0 > 0 and X 0 > 0 First we consider the hypothesis, G 0L : β 0 + β X 0 + β X 0 µ β 0, β 0 G 0U : β 0 + β X 0 + β X 0 µ β 0, β 0 (3.) and G : β 0, β 0
18 Define a ( α)-coefficient prediction interval (L P, U P ) for Y, where L P = min{y G 0L is accepted at level α/ against G a = G G 0L }, U P = max{y G 0U is accepted at level α/ against G a = G G 0U }. Then we use the transformation from β to r. Let r 0 = β 0 / + /n, r = β, ˆβ 0 r = S X β, then ˆr = ( ˆr 0, ˆr, ˆr ) = (, ˆβ, S X ˆβ ) N 3 (r, σ I) where ˆr is +/n the unrestricted MLE of r. Let b = µs X X 0, c = S X X 0 + /n > 0, d = X 0S X X 0 > 0. Then r 0 + /n + r X 0 + r X 0 S X µ, c r 0 + d r + r b, r b c r 0 d r. Hence our hypothesis can be restated in terms of r, G 0 : 0 r b c r 0 d r, r 0, G : r 0, r 0, (3.) and test G 0 against G a = G G 0. Suppose we use the same notation G 0 to denote the null hypothesis region. Here note that G 0 is a polyhedral cone with vertex L = (b /c, 0, 0). If we shift G 0 along the r 0 axis to the origin, we obtain a shifted cone K. K is a closed convex cone bounded by three hyperplanes {c r 0 + d r + r = 0, r 0, r 0}, {r 0 0, 0 r c r 0 d, r = 0}, {r 0 0, r = 0, 0 r c r 0 }. Then G 0 = K + L, bounded by {c r 0 + d r + r = b, r 0, r 0}, {r 0 b c, c r 0 + d r b, r = 0}, {r 0 b c, r = 0, c r 0 + r b }. Let G 0 = K + L, where K is the dual cone of K. It can be shown that the boundaries of K are the perpendiculars to the boundaries of K. So the Fenchel
19 dual cone K is bounded by three hyperplance {r 0 0, r d c r 0, r = c r 0 }, {r 0 0, r = d c r 0, r c r 0 }, {r 0 = 0, r d c r 0, r c r 0 }. Now let ˆr N 3 (r, σ I), where ˆr is the unrestricted MLE of r. Hence note that the restricted MLE of r in G is r = (r 0, r, r ) = ( ˆr 0, ˆr +, ˆr + ), and r is the equal weight projection of ˆr onto parameter space G. Let r be the MLE of r under G 0 and r is the equal weight prejection of ˆr onto G 0. When σ is known, the likelihood ratio test(lrt) rejects G 0 for large values of the test statistic χ 0 = logλ = ( ˆr r ˆr r )/σ, (3.3) where Λ is the kernel of LRT statistic. Figure 3.4, we consider several different cases when ˆr located in several different regions. With the boundaries of G 0 and G 0, we consider the whole region as an union of 3 disjoint regions. Depending on the signs of ˆr and ˆr, we discuss the test statistic χ 0 in each area separately. First when ˆr < 0 and ˆr < 0, we partition the region {(r 0, r, r ) : r < 0, r < 0} into S = {(r 0, r, r ) : r 0 < b c, r < 0, r < 0} and S = {(r 0, r, r ) : r 0 b c, r < 0, r < 0}. When ˆr S, r = r = ( ˆr 0, 0, 0). So χ 0 = r r /σ = 0, so S is inside the accept region. 3
20 Figure 3.: The region G 0 and boundary of the rejection region When ˆr S, r = ( ˆr 0, 0, 0) and r = ( b c, 0, 0). So χ 0 = ( ˆr 0 b c ) /σ which has chi-square distribution with degree of freedom. Furthere the boundary of rejection region is r 0 = b c + C α σ. When ˆr < 0 and ˆr 0, we can define regions S 3 = {(r 0, r, r ) : r 0 < b c c r, r < 0, r 0}, S 4 = {(r 0, r, r ) : r < 0, r max{b c r 0, c r 0 b c }}, and S 5 = {(r 0, r, r ) : r 0 > c r b c, r < 0, r 0} such that the disjoint union of S 3, S 4, and S 5 is the region {(r 0, r, r ) : r < 0, r 0}. 4
21 Figure 3.: -Dimensional illustration of S 3, S 4 and S 5 When ˆr S 3, r = ( ˆr 0, 0, ˆr ) and r = ( ˆr 0, 0, ˆr ).So χ 0 = r r /σ = 0. So S 3 is inside the accept region. When ˆr S 4, r = ( ˆr 0, 0, ˆr ) and r = ( ˆr 0, 0, ( ˆr u)u) where u is a unit vector along the line {c r 0 +r = b, r = 0}. χ 0 = r r /σ > C α which belongs to chi-square distribution with degree of freedom.then the boundary of the rejection region is c r 0 + r = b + + c C α σ. When ˆr S 5, r = ( ˆr 0, 0, ˆr ) and r = ( b c, 0, 0). χ 0 = (( ˆr 0 b c ) + ˆr )/σ > Cα which has chi-square distribution with degree of freedom. Then a part of the boundary of the rejection region is (r 0 b c ) + r = C ασ. And when ˆr 0 and ˆr < 0, we can define regions S 6 = {(r 0, r, r ) : 0 r < b d c d r 0, r < 0}, S 7 = {(r 0, r, r ) : r max{ b d c d r 0, d c r 0 b d }, r < 0}, and S 8 = {(r 0, r, r ) : 0 r < d c r 0 b d, r < 0} such that the disjoint union of S 6, S 7 and S 8 is the region {(r 0, r, r ) : r 0, r < 0}. c c 5
22 Figure 3.3: -Dimensional illustration of S 6, S 7 and S 8 When ˆr S 6, r = ( ˆr 0, ˆr, 0) and r = ( ˆr 0, ˆr, 0).So χ 0 = r r /σ = 0. So S 6 is inside the accept region. When ˆr S 7, r = ( ˆr 0, ˆr, 0) and r = (( ˆr 0, ˆr, 0) v)v where v is a unit vector along the line {c r 0 + d r = b, r = 0}. χ 0 = r r /σ > C α, which has chi-square distribution with degree of freedom.then the boundary of the rejection region is c r 0 + d r = b + c + d C α σ. When ˆr S 8, r = ( ˆr 0, ˆr, 0) and r = ( b c, 0, 0). χ 0 = (( ˆr 0 b c ) + ˆr )/σ > Cα which has chi-square distribution with degree of freedom. Then a part of the boundary of the rejection region is (r 0 b c ) + r = C ασ. And when ˆr 0 and ˆr 0, we can define regions S 9 = {(r 0, r, r ) : c r 0 + d r + r b, 0 r, 0 r }, S 0 = {(r 0, r, r ) : 0 r c d r +c 0 + d, r +c c r 0 b }, S c = {(r 0, r, r ) : r d c r 0 b d, 0 r c c r c 0 + d r +d c b }, +d c +d S 3 = {(r 0, r, r ) : 0 r < d c r 0 b d, 0 r c c r 0 b d }, S c = {(r 0, r, r ) : r 0, r 0} S 9 S 0 S S 3, such that the disjoint union of S 9, S 0, S, S, and S 3 is the region {(r 0, r, r ) : r 0, r 0}. 6
23 Figure 3.4: 3-Dimensional view of disjoint regions S 9, S 0, S, S and S 3 When ˆr S 9, r = ( ˆr 0, ˆr, ˆr ) and r = ( ˆr 0, ˆr, ˆr ). χ 0 = r r /σ = 0. So S 9 is inside the accept region. When ˆr S 0, r = ( ˆr 0, ˆr, ˆr ) and r = (( ˆr 0, ˆr, ˆr ) u)u where u is a unit vector along the line {c r 0 +r = b, r = 0}. χ 0 = r r /σ > C α, which has to chi-square distribution with degree of freedom. The boundary of the rejection region is a part of a rotated cylinder, r +[ r + c (r 0 b +c +c c )] = Cασ. When ˆr S, r = ( ˆr 0, ˆr, ˆr ) and r = (( ˆr 0, ˆr, ˆr ) v)v where v is a unit vector along the line {c r 0 + d r = b, r = 0}. χ 0 = r r /σ > Cα which belongs to chi-square distribution with degree of freedom. The boundary of the rejection region is a part of a rotated cylinder: r + [ d r + c (r 0 b c +d c +d c )] = Cασ. When ˆr S, r = ( ˆr 0, ˆr, ˆr ) and r = (( ˆr 0, ˆr, ˆr ) w)w where w is a unit vector along the line LB. χ 0 = r r /σ > Cα which belongs to chi-square distribution with degree of freedom. The boundary of the rejection region is hyperplane above G 0, which is c r 0 +d r +r = b + + c + d C α σ. When ˆr S 3, r = ( ˆr 0, ˆr, ˆr ) and r = ( b c, 0, 0). χ 0 = r r /σ = [( ˆr 0 b c ) + ( ˆr ) + ( ˆr ) ]/σ > Cα which belongs to chi-square distribution with 3 degree of freedom. The boundary of the 7
24 rejection region is a part of sphere surface( ˆr 0 b c ) + ˆr + ˆr = C ασ. The least favorable null value of χ 0 is attained at r = L = ( b c, 0, 0) and sup r G 0 P r r {ˆr : r r C ασ } = P r L { r r C α σ}. (3.4) THe proof of above result is given by Peiris and Bhattacharya (06). When ˆr is attained the least favorable null value, the distribution of LRT χ 0 is given by folloing formula. (See Peiris and Bhattacharya (06) for the proof and more details.) The least favorable null distribution of LRT is P r(lrt t ˆr = L) = 3 w i P (χ i t), i=0 where w 0 = (4π) (cos + cos d + cos + c c + d + c + d + cos d ), + c + d w = (4π) ( 3 π + cos d ( + c )(c + d ) ), + c w = (4π) (π + cos c + cos + d + c + d + c + d cos cos d ), + c c + d w 3 = (4π) ( 3 + c π c cos cos + d cos + c + d + c + d + c + d And the prediction upper bound is cos d cos d + c + d ( + c )(c + d ) ). (3.5) U P = max{y G 0U is accepted at level α/ against G a = G G 0U }. 8
25 Use the transformation from β to r. Let r 0 = β 0 / + /n, r = β, r = S X β then ˆr = ( ˆr 0, ˆr, ˆr ) = (β 0 / + /n, ˆβ, S X ˆβ ) N 3 (r, σ I) where ˆr is the unrestricted MLE of r. Let b = µs X X 0, c = S X X 0 + /n > 0, d = X 0S X X 0 > 0. r 0 + /n + r X 0 + r X 0 S X µ, c r 0 + d r + r b, r b c r 0 d r, Hence our hypothesis is H 0 : r b c r 0 d r, r 0, r 0, H : r 0 r 0, (3.6) and test H 0 against H a = H H 0. Similarly, to illustrate the construction of rejection region, we need the boundaries of H 0 in the r form. Let K be the shifted cone of H 0, and K be the dual cone of K. And H 0 = K +L and H 0 = K +L, where L = (b /c, 0, 0). Then we can get the 6 regions divided by the boundaries of K and K. Now let ˆr N 3 (r, σ I), where ˆr si the unrestricted MLE of r. And the restricted MLE of β is β = (β0, β, β) in section.. Hence we can define the restricted MLE of r is r = (r0, r, r) = ( ˆr 0, ˆr +, ˆr + ), and r is the equal weight projection of ˆr onto parameter space H. Let r be the MLE of r under H 0 and r is the equal weight prejection of ˆr onto H 0. When σ is known, the likelihood ratio test(lrt) rejects H 0 for large values of the test statistic is χ 0 = logλ = ( ˆr r ˆr r )/σ, (3.7) 9
26 where Λ is the kernel of LRT statistic. According to the discussion in Peiris and Bhattacharya (06), the least favorable null value of LRT(.8) is attained at lim t,s (b /c s c t, c t, c s) and sup P r r {ˆr : ˆr r ˆr r Dασ } r H 0 = lim P r (b t,s /c s c t,c t,c s){ χ 0 > Dασ } (3.8) Also, the null critical value is D α = χ,α. Figure 3.5: H 0 and Rejection region Shown in figure 3.5, we consider the whole region as an union of several disjoint areas. Depending on the signs of ˆr and ˆr, I discuss the test statistic χ 0 in each area separately. However I can not obtain the exact rejection region to get confidence and prediction intervals. Hence we need to modify our likelihood ratio test under same power. 0
27 Consider the hypothesis without the restrictions r 0 and r 0 H 0 : c r 0 + d r + r b, H : c r 0 + d r + r < b. Then the null hypothesis is excatly same as the unrestricted case. (3.9) LRT rejects H 0 for small values χ 03 = c ˆr 0 +d ˆr + ˆr b +c +d σ. So the rejection region of LRT is {ˆr : χ 03 < X α σ} Here rejection region for the unrestricted LRT contained that for the restricted LRT. So the unrestricted LRT is more powerful than the restricted LRT. But this creates a philosophical dilemma in some cases. In some cases, we will reject H 0 under unrestricted LRT but will not reject it under restricted LRT. So we need to modify LRT. Then consider following four regions. We use a similar idea as that for two dimensional model EY = β 0 + β X discussed in Mukerjee and Tu (995). We consider four regions in R 3 : S, S, S 3, and S 4, where S = {r : r d + c + d Z α σ, r 0}, S 3 = {r : r 0, r + c + d Z α σ}, S 4 = {r : r < 0, r < min{0, d r + c + d Z α σ}}, and S = R 3 S S 3 S 4. The boundary of the H 0 which is c r 0 + d r + r = b meets the hyperplane {r = 0} on the line {r = 0, c r 0 + d r = b } and the hyperplane {r = 0} on the line {r = 0, c r 0 + r = b }. Hyperplane c r 0 + d r + r = b + c + d Z α σ and hyperplane c r 0 + d r = b intersect on the line r = + c + d Z α σ. Hyperplane c r 0 + d r + r = b + c + d Z α σ and hyperplane c r 0 + r = b intersect on the hyperplane r = d + c + d Z α σ. To keep the same rejection level α, we modify LRT as follows, when ˆr S, we use the same boundary of the rejection region of the unrestricted case, which is c r 0 + d r + r = b + c + d Z α σ. When ˆr S, we already know the intersection of hyperplane c r 0 + d r + r = b + c + d Z α σ and S s boundary r = d + c + d Z α σ is the plane {c r 0 + r = b, r = d + c + d Z α σ}.
28 So let c r 0 + r = b as a part of the boundaries of rejection region in S. Similarly when ˆr S 4, we let c r 0 + d r = b as a part of boundaries of rejection region in S 4. when ˆr S 3, we let c r 0 = b as a part of boundaries of rejection region in S Test when X 0 < 0 and X 0 < 0 We consider the hypothesis, G 0L : β 0 + β X 0 + β X 0 µ β 0, β 0, G 0U : β 0 + β X 0 + β X 0 µ β 0, β 0, (3.0) and G : β 0, β 0. Define a ( α)-coefficient prediction interval (L P, U P ) for Y, where L P = min{y G 0L is accepted at level α/ against G a = G G 0L }, U P = max{y G 0U is accepted at level α/ against G a = G G 0U }. Then we can make a tranformation, where X 0 = X 0 > 0, X 0 = X 0 > 0, β 0 = β 0, and µ = µ. Then the new hypothesis will be, G 0L : β 0 + β X 0 + β X 0 µ β 0, β 0, G 0U : β 0 + β X 0 + β X 0 µ β 0, β 0, (3.) and G : β 0, β 0.
29 Define a ( α)-coefficient prediction interval (L P, U P ) for Y, where L P = min{y G 0L is accepted at level α/ against G a = G G 0L} = U P, U P = max{y G 0U is accepted at level α/ against G a = G G 0U} = L P, where L P and U P are in section 3.. Hence the formulas for rejection region and prediciton intervals can be obtained using the symetric property and are shown in next chapter. 3.3 Test when X 0 > 0 and X 0 < 0 Then we consider the hypothesis, G 0L : β 0 + β X 0 + β X 0 µ β 0, β 0, G 0U : β 0 + β X 0 + β X 0 µ β 0, β 0, (3.) and G : β 0, β 0. Define a ( α)-coefficient prediction interval (L P, U P ) for Y, where L P = min{y G 0L is accepted at level α/ against G a = G G 0L }, U P = max{y G 0U is accepted at level α/ against G a = G G 0U }. We use the transformation from β to r. Let r 0 = β 0 / + /n, r = β, r = S X β, then ˆr = ( ˆr 0, ˆr, ˆr ) = ( ˆβ 0 / + /n, ˆβ, S X ˆβ ) N 3 (r, σ I), where ˆr is the unrestricted MLE of r. 3
30 And b = µs X X 0, c = S X X 0 + /n < 0, d = X 0S X X 0 < 0, r 0 + /n + r X 0 + r X 0 S X µ c r 0 + d r + r b r b c r 0 d r Hence our hypothesis in terms of r is G 03 : r b c r 0 d r, r 0, r 0, G 3 : r 0 r 0. (3.3) Figure 3.6: G 03 and Rejection region Suppose we use the same notation G 03 to denote the null hypothesis region. Here note that G 03 is a polyhedral cone with vertex L = (b /c, 0, 0). If we shift G 03 along the r 0 axis to the origin, we obtain a shifted cone K. K is the closed 4
31 convex cone bounded by three hyperplanes {c r 0 + d r + r = 0, r 0, r 0}, {r = 0, c r 0 + d r + r 0, r 0}, {r = 0, c r 0 + d r + r 0, r 0}. Then G 03 = K + L, bounded by {c r 0 + d r + r = b, r 0, r b }, {r = 0, c r 0 + d r + r b, r 0}, {r = 0, c r 0 + d r + r b, r 0}. Recall the definition of dual cone, let G 03 = K + L, where K is the dual cone of K. It can be shown that the boundaries of K are the perpendiculars to the boundaries of K. So the Fenchel dual cone G 03 is bounded by three hyperplance {r 0 c d r b c, r 0 b c, r 0 c r = b c }, {r 0 c d r = b c, r 0 b c, r 0 c r b c }, {r 0 c d r b c, r 0 = b c, r 0 c r b c }. Now let ˆr N 3 (r, σ I), where ˆr si the unrestricted MLE of r. And the restricted MLE of β is β = (β 0, β, β ) in section.. Hence we can define the restricted MLE of r is r = (r 0, r, r ) = ( ˆr 0, ˆr +, ˆr + ), and r is the equal weight projection of ˆr onto parameter space G 3. Let r be the MLE of r under G 03 and r is the equal weight prejection of ˆr onto G 03. When σ is known, the likelihood ratio test(lrt) rejects G 03 for large values of the test statisti is χ 03 = logλ = ( ˆr r ˆr r )/σ, (3.4) where Λ is the kernel of LRT statistic. So the rejection region with a level α is {( ˆr r ˆr r ) > E ασ }, where E α is the critical value. Similar as previous section, when r 0, the rejection region is { ˆr r > E ασ }. We can obtain boundaries of the rejection region {c r 0 + r b + c F α σ, r < 0}, {r + ( r + c (r 0 b +c +c c )) = Fασ, 0 r c d r +c 0 + d r +c b d }, {c +c r 0 + d r + r = b + c + d F α σ, r max{0, c d r +c 0 + d r +c b d }}. +c But when r < 0, the rejection region has a complicated formulas and it is hard 5
32 to illstruate them with figures. We propose a new modified rejection region which is similar as previous section. The least favorable null value and the least favorable distribution is given in Peiris and Bhattacharya (06). So the least favorable null value of χ 03 is attained at infinity with lim r 0 (r 0, 0, b c r 0 ) and sup P r r {ˆr : ( ˆr r ˆr r )/σ Eα} = lim P r (r 0,0,b c r 0 ){ χ 03 > E r G 03 r 0 ασ }. The least favorable null distribution of LRT is, sup P r(lrt c) = ( r G θ π )P (χ 0 c) + P (χ c) + ( 4 θ π )P (χ c), where θ is the anlge between hyperplane C r 0 + d r + r = b and hyperplane r = 0 To obtain the modified LRT, first we consider hypothesis (.3) without the restriction r 0, M 0 : r b c r 0 d r r 0 against M : r 0. So the new LRT rejects M 0 for large values is, χ 03 = ( ˆr r ˆr r )/σ, where r is the MLE under M 0 and r is the MLE under M. 6
33 Figure 3.7: M 0 and Rejection region and dual cone M 0 We can have the projection of those regions boundaries to the hyperplane which the intersection line of regions is orthogonal to. Then the discussion for rejection region is similar to two predictor variables model case (Mukerjee and Tu (995)). Hence when r < 0, divide the region into two parts S and S, where S = {r : r < 0, c r 0 + r b }, S = {r : r < 0, c r 0 + r < b }, and obtain the center axis which is the intersection line of five regions, where {c r 0 + r = b, r = 0} When ˆr S, χ 03 = r r = 0 where r = r = ( ˆr 0, 0, ˆr ). So S is in the acceptance region. When ˆr S, χ 03 = r r = ( ˆr 0, 0, ˆr ) (( ˆr 0, 0, ˆr ) u)u Fασ where r = (( ˆr 0, 0, ˆr ) u)u and r = ( ˆr 0, 0, ˆr ) It has chi-square distribution with degree of freedom. The boundary of rejection region is a hyperplane parallely above the M 0 and has F α σ distance to the hyperplane c r 0 + r = b. 7
34 Figure 3.8: -Dimensional view of M 0 and Rejection region when r 0, the hyperplane c r 0 + d r + r = b and c d r 0 ( + c )r + d r = b d divide the region into three subregions S 3, S 4, and S 5. When ˆr S 3, χ 03 = r r where r = (( ˆr 0, ˆr, ˆr ) u)u and r = ( ˆr 0, ˆr, ˆr ), where u is a unit vector along the center axis. It has chi-square distribution with degree of freedom. The boundary of rejection region is a part of cylinder (center axis is the axis and radius is F α σ). When ˆr S 4, χ 03 = r r where r = (( ˆr 0, ˆr, ˆr ) w)w and r = ( ˆr 0, ˆr, ˆr ), where w is a unit vector along the projection of ˆr onto hyperplane c r 0 + d r + r = b. It has chi-square distribution with degree of freedom. The boundary of rejection region is above the hyperplane c r 0 + d r + r = b with distance F α σ. When ˆr S 5, χ 03 = r r = 0 where r = r = ( ˆr 0, r ˆ, ˆr ). So S 5 is in the acceptance region. Using the same argument as in the previous section, I propose a modified LRT 8
35 for (3.0), keep the same boundary of LRT for (3.0) when r + c E α σ. Note that the hyperplane c r 0 + r = b + c E α σ and r 0 = b c intersected at r = + c E α σ. When r < + c E α σ, modify the rejection region with cut-off. I propose a hyperplane r 0 = b c as the part boundary of the rejection region. And I propose a curved plane which is parallel to r axis and a hyperplane which is {c r 0 + d r = b ( + c + d + c )E α σ} We consider another hypothesis G 0U agiainst G a = G G 0U because U P = max{y G 0U is accepted at level α/ against G a = G G 0U }. We use the transformation from β to r. Let r 0 = β 0 / + /n, r = β, r = S X β, then ˆr = ( ˆr 0, ˆr, ˆr ) = ( ˆβ 0 / + /n, ˆβ, S X ˆβ ) N 3 (r, σ I) where ˆr is the unrestricted MLE of r. And b = µs X X 0, c = S X X 0 + /n < 0, d = X 0S X X 0 < 0, r 0 + /n + r X 0 + r X 0 S X µ, c r 0 + d r + r b, r b c r 0 d r. Hence our hypothesis in terms of r is H 03 : 0 r b c r 0 d r, r 0, H 3 : r 0 r 0. (3.5) 9
36 Figure 3.9: H 03 and Rejection region Here I note that the null region H 03 is a mirror image of the null region G 03 in the previous section. Considering hypothesis without the restriction r 0, keep the boudary of rejection region which is same as in (3.) when r d c + d K α σ. Then I propose a modified LRT when r < d c + d K α σ. and The least favorable null value of χ 03 is attained at infinity with lim (r 0, b c r 0 r 0 d, 0) sup P r r {ˆr : ( ˆr r ˆr r )/σ Kα} = r H 03 lim P r r 0 (r 0, b c r 0 d { χ 03 > K,0) ασ } The least favorable distribution of LRT is, P r(lrt c) = ( 4 + θ π )P (χ 0 c) + P (χ c) + ( 4 θ π )P (χ c), 30
37 where θ is the anlge between hyperplane C r 0 + d r + r = b and hyperplane r = 0 (See more details in Peiris and Bhattacharya (06)). 3.4 Test when X 0 < 0 and X 0 > 0 Then we consider the hypothesis, G 0L : β 0 + β X 0 + β X 0 µ β 0, β 0, G 0U : β 0 + β X 0 + β X 0 µ β 0, β 0, (3.6) and G : β 0, β 0. Define a ( α)-coefficient prediction interval (L P, U P ) for Y, where L P = min{y G 0L is accepted at level α/ against G a = G G 0L }, U P = max{y G 0U is accepted at level α/ against G a = G G 0U }. Then we can make a tranformation, where X 0 = X 0 > 0, X 0 = X 0 < 0, β 0 = β 0, and µ = µ. Then the new hypothesis will be, G 0L : β 0 + β X 0 + β X 0 µ β 0, β 0, G 0U : β 0 + β X 0 + β X 0 µ β 0, β 0, (3.7) and G : β 0, β 0. Define a ( α)-coefficient prediction interval (L P, U P ) for Y, where L P = min{y G 0L is accepted at level α/ against G a = G G 0L} = U P, 3
38 U P = max{y G 0U is accepted at level α/ against G a = G G 0U} = L P. where L P and U P are in section 3.3. Hence the formulas for rejection region and prediciton intervals can be obtained using symetric properties of these cases as shown in next chapter. Figure 3.0: G 04 and Rejection region 3
39 Chapter 4 Formulas for The Rejection Region 4. when X 0 > 0 and X 0 > 0 For hypothesis (3.) G 0 : 0 r b c r 0 d r, r 0, G : r 0, r 0. 33
40 The rejection region of r form,.{ ˆr 0 b c + C α/ σ, ˆr < 0, ˆr < 0},.{( ˆr 0 b ) + ˆr C c α/σ, ˆr < 0, 0 ˆr < ˆr 0 b }, c 3.{( ˆr 0 b ) + ˆr C c α/σ, 0 ˆr < d ˆr 0 b d, ˆr < 0}, c 4.{( ˆr 0 b ) + ˆr + ˆr C c α/σ, 0 ˆr < d ˆr 0 b d, 0 ˆr c c < ˆr 0 b }, c c 5.{c ˆr 0 + ˆr b + + c C α/ σ, ˆr < 0, ˆr ˆr 0 b }, c 6.{c ˆr 0 + d ˆr b c + d C α/ σ, ˆr d ˆr 0 b d, ˆr < 0}, c 7.{ ˆr c + ( ˆr + ( ˆr 0 b )) C + c + c c α/σ, 0 ˆr < c d ˆr + c 0 + d + c ˆr b d + c c c c c, ˆr ˆr 0 b }, c c 8.{ ˆr d c + ( ˆr + ( ˆr 0 b )) C c + d c + d c α/σ, 0 ˆr < c d ˆr c + d 0 + d ˆr c + d b, ˆr c + d d ˆr 0 b d }, c c 9.{c ˆr 0 + d ˆr + ˆr b C α/ σ + c + d, c d ˆr max{0, + c c d ˆr max{0, c + d ˆr 0 + d ˆr + c b d }, + c ˆr 0 + d ˆr c + d b }}. c + d Then we have the transformation from r to β. Let r 0 = β 0 / + /n, r = β, r = S X β, then ˆr = ( ˆr 0, ˆr, ˆr ) = (, ˆβ, S X ˆβ ) N 3 (r, σ I) +/n where ˆr is the unrestricted MLE of r. And b = µs X X 0, c = S X X 0 + /n > 0, d = X 0 S X X 0 > ˆβ 0
41 Then we transform to the original variables, rejection region of ˆβ form,.{ ˆβ 0 µ + C α/ σ + n, ˆβ < 0, ˆβ < 0},.{ + /n ( ˆβ 0 µ) + S X ˆβ C α/ σ, ˆβ < 0, 0 ˆβ < 3.{ + /n ( ˆβ 0 µ) + ˆβ C α/ σ, 0 ˆβ < 4.{ + /n ( ˆβ 0 µ) + ˆβ + SX ˆβ C α/ σ, 0 ˆβ < 0 ˆβ < +/n X 0 ( ˆβ 0 µ)}, S X X 0 5.{ ˆβ 0 + ˆβ X 0 > µ + S + + X n C α/σ, ˆβ < 0, ˆβ 6.{ ˆβ 0 + ˆβ X0 X 0 > µ + S + + X n C α/σ, ˆβ 7.{ ˆβ + ( ˆβ0 + ˆβ X 0 µ) 0 ˆβ < X 0 + n + X 0 S X C α/σ, ˆβ ( X 0 S X + + )( 0 + ˆβ X 0 µ), n 8.{S X ˆβ + ( ˆβ0 + ˆβ X 0 l) ˆβ + n + X 0 +/n X 0 ( ˆβ 0 µ), 0 ˆβ < X 0 S X 9.{ ˆβ 0 + ˆβ X 0 + ˆβ X 0 > µ + C α/ σ ˆβ max{0, X 0 ˆβ max{0, X 0 S X C α/σ, +/n X 0 S X ( ˆβ 0 µ)}, +/n X 0 ( ˆβ 0 µ), ˆβ < 0}, +/n X 0 ( ˆβ 0 µ), +/n X 0 S X ( ˆβ 0 µ)}, +/n X 0 ( ˆβ 0 µ), ˆβ < 0}, ˆβ +/n X 0 S X ( ˆβ 0 µ)}, ˆβ ( X 0 S X + )( 0 + ˆβ X 0 µ)}, n + n + X 0 + X 0 S X, ˆβ ( X 0 S X + + )( 0 + ˆβ X 0 µ)}, n ˆβ ( X 0 S X + + )( 0 + ˆβ X 0 µ)}}. n 35
42 For hypothesis (3.6) H 0 : r b c r 0 d r, r 0, r 0 H : r 0 r 0 rejection region in terms of ˆr,. {c ˆr 0 + ˆr b, ˆr + c + d d Z α/ σ, ˆr 0},. {c ˆr 0 + d ˆr b, ˆr 0, ˆr 3. { ˆr 0 < b, ˆr 0, ˆr min{0, d ˆr c 4. { ˆr 0 + d ˆr + ˆr b Then we transform to the original variables, rejection region in terms of ˆβ, + c + d Z α/ σ}, + c + d Z α/ σ}, + c + d Z α/ σ, otherwise}.. { ˆβ 0 + ˆβ X 0 µ, ˆβ < + X 0 n + X 0 S + X 0 X S Z α/σ, ˆβ 0}, X. { ˆβ 0 + ˆβ X 0 µ, ˆβ 0, ˆβ < 3. { ˆβ 0 µ, ˆβ < 0, ˆβ < min{0, 4. { ˆβ 0 + ˆβ X 0 + ˆβ X 0 µ X 0 X 0 + n + X 0 + X 0 S X Z α/σ}, + n + X 0 + X 0 S X Z α/σ}}, + n + X 0 + X 0 S X Z α/σ, otherwise}. 36
43 4. when X 0 < 0 and X 0 < 0 rejection region lower bound in section 4... { ˆβ 0 + ˆβ X0 µ, ˆβ < + X0 n + X 0 S + X 0 X S Z α/σ, ˆβ 0} X. { ˆβ 0 + ˆβ X0 µ, ˆβ 0, ˆβ < + X0 n + X 0 S + X 0 X S Z α/σ} X 3. { ˆβ 0 µ, ˆβ < 0, ˆβ < min{0, + X0 n + X 0 S + X 0 X S Z α/σ}} X 4. { ˆβ 0 + ˆβ X0 + ˆβ X0 µ + n + X 0 S + X 0 X S Z α/σ, otherwise} X Tranform to X 0 < 0 and X 0 < 0 case,. { ˆβ 0 + ˆβ X 0 µ, ˆβ < + X 0 n + X 0 S + X 0 X S Z α/σ, ˆβ 0}, X. { ˆβ 0 + ˆβ X 0 µ, ˆβ 0, ˆβ < + X 0 n + X 0 S + X 0 X S Z α/σ}, X 3. { ˆβ 0 µ, ˆβ < 0, ˆβ < min{0, + X 0 n + X 0 S + X 0 X S Z α/σ}}, X 4. { ˆβ 0 + ˆβ X 0 + ˆβ X 0 µ + + n + X 0 S + X 0 X S Z α/σ, otherwise}. X 37
44 rejection region upper bound in section 4.,.{ ˆβ 0 µ + C α/ σ + n, ˆβ < 0, ˆβ < 0},.{ + /n ( ˆβ 0 µ ) + S X ˆβ C α/ σ, ˆβ < 0, 0 ˆβ < 3.{ + /n ( ˆβ 0 µ ) + ˆβ C α/ σ, 0 ˆβ < 4.{ + /n ( ˆβ 0 µ ) + ˆβ + SX ˆβ C α/ σ, 0 ˆβ < +/n X 0 5.{ ˆβ 0 + ˆβ X 0 µ + 6.{ ˆβ 0 + ˆβ X 0 µ + ( ˆβ 0 µ ), 0 ˆβ < X0 +/n X 0 ( ˆβ 0 µ )}, S X S X + + n C α/σ, ˆβ < 0, ˆβ X n C α/σ, 7.{ ˆβ + ( ˆβ0 + ˆβ X 0 µ ) 0 ˆβ < X 0 ( X 0 S X + + n )( ˆβ 0 8.{S X ˆβ + ( ˆβ0 + ˆβ X 0 l) ˆβ + n + X 0 S X ˆβ C α/σ, + ˆβ X 0 µ ), + n + X 0 +/n X 0 ( ˆβ 0 µ ), 0 ˆβ < X 0 S X C α/σ, +/n X 0 ( ˆβ 0 µ )}, S X +/n X 0 ( ˆβ 0 µ ), ˆβ < 0}, +/n X 0 ( ˆβ 0 µ )}, S X +/n X 0 ( ˆβ 0 µ ), ˆβ < 0}, ˆβ ( X 0 + n )( ˆβ 0 9.{ ˆβ 0 + ˆβ X0 + ˆβ X0 µ + C α/ σ + n + X 0 S + X 0 X S, X ˆβ max{0, X 0 ˆβ max{0, X 0 S X ( X 0 S X + + n )( ˆβ 0 ( X n )( ˆβ 0 + ˆβ X 0 µ )}, + ˆβ X 0 µ )}}. +/n X 0 ( ˆβ 0 µ )}, S X + ˆβ X 0 µ )}, 38
45 Transform to X 0 < 0 and X 0 < 0 case,.{ ˆβ 0 µ C α/ σ + n, ˆβ < 0, ˆβ < 0},.{ + /n ( ˆβ 0 µ) + S X ˆβ C α/ σ, ˆβ < 0, 0 ˆβ < 3.{ + /n ( ˆβ 0 µ) + ˆβ C α/ σ, 0 ˆβ < 4.{ + /n ( ˆβ 0 µ) + ˆβ + SX ˆβ C α/ σ, 0 ˆβ +/n < X 0 ( ˆβ 0 µ), 0 ˆβ < X +/n 0 ( ˆβ 0 µ)}, S X 5.{ ˆβ 0 + ˆβ X 0 X 0 µ S + + X n C α/σ, ˆβ < 0, ˆβ 6.{ ˆβ 0 + ˆβ X 0 µ X n C α/σ, 7.{ ˆβ + ( ˆβ0 + ˆβ X 0 µ) + n + X 0 0 ˆβ < X 0 S X ˆβ C α/σ, ˆβ ( X 0 S X + + )( 0 + ˆβ X 0 µ), n 8.{S X ˆβ + ( ˆβ0 + ˆβ X 0 µ) + n + X 0 ˆβ +/n X 0 ( ˆβ 0 µ), 0 ˆβ < X 0 S X C α/σ, +/n X 0 S X ( ˆβ 0 µ)}, +/n X 0 ( ˆβ 0 µ), ˆβ < 0}, +/n X 0 S X ( ˆβ 0 µ)}, +/n X 0 ( ˆβ 0 µ), ˆβ < 0}, ˆβ +/n X 0 S X ( ˆβ 0 µ)}, ˆβ ( X 0 S X + )( 0 + ˆβ X 0 µ)}, n 9.{ ˆβ 0 + ˆβ X 0 + ˆβ X 0 µ C α/ σ + n + X 0 S + X 0 X S, X ˆβ max{0, X 0 ˆβ max{0, X 0 S X ˆβ ( X 0 S X + + )( 0 + ˆβ X 0 µ)}, n ˆβ ( X 0 S X + + )( 0 + ˆβ X 0 µ)}}. n 39
46 4.3 when X 0 > 0 and X 0 < 0 For hypothesis (3.3) G 03 : r b c r 0 d r, r 0, r 0 G 3 : r 0 r 0. rejection region lower bound of r form,. { ˆr 0 > b, ˆr < 0, ˆr < +c c E α/ σ}. {(c ˆr 0 (b + + c E α/ σ)) + ( + c ) ˆr ( + c )Eα/σ, 0 ˆr < c d ˆr + c 0 + d ˆr + c b d, ˆr + c < + c E α/ σ} 3. {c ˆr 0 + d ˆr b ( + c + d + c )E α/ σ, c d ˆr max{0, ˆr + c 0 + d ˆr + c b d }, ˆr + c < + c E α/ σ} 4. {c ˆr 0 + ˆr b + c, ˆr < 0, ˆr > + c E α/ σ} 5. { ˆr c + ( ˆr + ( ˆr 0 b )) E + c + c c α/σ, 0 < ˆr c d ˆr + c 0 + d + c 6. {c ˆr 0 + d ˆr + ˆr b c d ˆr max{0, + c ˆr b d, ˆr + c + c + d E α/ σ, ˆr 0 + d ˆr + c b d }, ˆr + c + c E α/ σ} + c E α/ σ} 40
47 We use the transformation from r to β. Let r 0 = β 0 / + /n, r = β, r = S X β, b = µs X X 0, c = S X X 0 + /n < 0, d = X 0S X X 0 < 0,. { ˆβ 0 > µ, ˆβ < 0, ˆβ <. {( ˆβ 0 µ ( X 0 S X + + n )E α/σ, X 0 X0 S + + X n E α/σ}, X0 S + + X n E α/σ) + ( X 0 S + + X n )S X ˆβ 0 ˆβ < ( ˆβ 0 µ + X 0 ˆβ ) X 0 ˆβ < X0 X 0 3. { ˆβ 0 + X 0 ˆβ µ + ( S + + X n E α/σ}, X 0 ˆβ ( ˆβ 0 µ + X 0 ˆβ ) X 0 X 0 S X + + n S + X 0 X S + + X n X 0 S X + + n, X 0, ˆβ < X 0 4. { ˆβ 0 + ˆβ X0 X 0 µ + S + + X n E α/σ, ˆβ < 0, ˆβ X0 X 0 S + + X n E α/σ}, 5. {( ˆβ 0 µ + X 0 ˆβ ) X 0 S X + + n 0 ˆβ < ( ˆβ 0 µ + X 0 ˆβ ) X 0 6. { ˆβ 0 + ˆβ X 0 + ˆβ X 0 µ + ˆβ ( ˆβ 0 µ + X 0 ˆβ ) X 0 X 0 + ˆβ E α/ σ, X 0 S X + + n + X 0 S X + + n E α/σ, X 0 S X + + n S + + X n )E α/σ, X 0 S X + + n E α/σ},, ˆβ X0 X 0 S + + X n E α/σ},, ˆβ X 0 X 0 S X + + n E α/σ}. 4
48 For hypothesis (3.5) H 03 : 0 r b c r 0 d r, r 0, H 3 : r 0 r 0. rejection region upper bound of r form,. { ˆr 0 b c,. {(c ˆr 0 (b ˆr < c + d d K α/ σ, ˆr < 0}, c + d K α/ σ)) + (c + d ) ˆr (c + d )Kα/σ, ˆr < d c + d K α/ σ 0 ˆr < c 3. {c ˆr 0 + ˆr b + ( + c + d c + d c + d )K α/ σ, ˆr 0 + d ˆr c + d b }, c + d ˆr < c + d d K α/ σ, ˆr c ˆr c + d 0 + d ˆr c + d b }, c + d 4. {c ˆr 0 + d ˆr b + c + d K α/ σ, ˆr c + d K α/ σ, ˆr < 0}, d d 5. { ˆr + ( ˆr + c + d c c + d ( ˆr 0 b c )) K α/σ, ˆr c + d d K α/ σ, 0 ˆr < c c + d 6. {c ˆr 0 + d ˆr + ˆr b + + c + d K α/ σ, ˆr 0 + d ˆr c + d b }, c + d ˆr c + d d K α/ σ, ˆr c ˆr c + d 0 + d ˆr c + d b }. c + d 4
49 rejection region upper bound ˆβ. { ˆβ 0 µ, ˆβ <. {( ˆβ 0 µ X 0 + n + X 0 K α/σ, ˆβ < 0}, + n + X 0 S K α/σ) + ( + X n + X 0 S ) S X ˆβ X ( + n + X 0 S )K α/σ, X ˆβ < + X 0 n + X 0 S K α/σ, X 0 ˆβ < ( ˆβ 0 + X 0 ˆβ µ) X 0 S X + n + X 0 3. { ˆβ 0 + X 0 ˆβ µ ( + n + X 0 S + X 0 X S X ˆβ < X 0 + n + X 0 S K α/σ, X ˆβ ( ˆβ 0 + X 0 ˆβ µ) X 0 S X 4. { ˆβ 0 + X 0 ˆβ µ + n + X 0 S K α/σ, X ˆβ + X 0 n + X 0 S K α/σ, ˆβ < 0}, X 5. {S X ˆβ (X 0 ˆβ + ˆβ 0 µ) Kα/σ, X 0 n S X ˆβ X 0 6. { ˆβ 0 + X 0 ˆβ + X 0 ˆβ µ }, + n + X 0 )K α/σ, + n + X 0 S K α/σ, 0 ˆβ < ( ˆβ 0 + X 0 ˆβ µ) X 0 X S X + n + X 0 + X 0 S X K α/σ, ˆβ X 0 + n + X 0 S K α/σ, X ˆβ ( ˆβ 0 + X 0 ˆβ µ) X 0 S X + n + X 0 }, + n + X 0 + n + X 0 }. }, 43
50 4.4 when X 0 < 0 and X 0 > 0 rejection region lower bound. { ˆβ 0 µ, ˆβ <. {( ˆβ 0 µ + X 0 + n + X 0 K α/σ, ˆβ < 0}, + n + X 0 S K α/σ) + ( + X n + X 0 S ) S X ˆβ X ( + n + X 0 S )K α/σ, X ˆβ < + X 0 n + X 0 S K α/σ, X 0 ˆβ < ( ˆβ 0 + X 0 ˆβ µ) X 0 S X + n + X 0 3. { ˆβ 0 + X 0 ˆβ µ + ( + n + X 0 S + X 0 X S X ˆβ < + X 0 n + X 0 }, + n + X 0 )K α/σ, S K α/σ, ˆβ ( ˆβ 0 + X 0 ˆβ µ) X 0 X S X 4. { ˆβ 0 + X 0 ˆβ µ + + n + X 0 S K α/σ, X ˆβ + X 0 n + X 0 S K α/σ, ˆβ < 0}, X 5. {S X ˆβ (X 0 ˆβ + ˆβ 0 µ) Kα/σ, X 0 n S X ˆβ + X 0 n + X 0 S K α/σ, X 0 ˆβ < ( ˆβ 0 + X 0 ˆβ µ) X 0 S X 6. { ˆβ 0 + X 0 ˆβ + X 0 ˆβ µ + + n + X 0 }, + n + X 0 + X 0 S X K α/σ, ˆβ + X 0 n + X 0 S K α/σ, ˆβ ( ˆβ 0 + X 0 ˆβ µ) X 0 X S X 44 + n + X 0 + n + X 0 }. },
51 rejection region upper bound. { ˆβ 0 > µ, ˆβ < 0, ˆβ <. {( ˆβ 0 µ ( X 0 S X + + n )E α/σ, X 0 X0 S + + X n E α/σ}, X0 S + + X n E α/σ) + ( X 0 S + + X n )S X ˆβ 0 ˆβ < ( ˆβ 0 µ + X 0 ˆβ ) X 0 3. { ˆβ 0 + X 0 ˆβ µ + ( X 0 ˆβ ( ˆβ 0 µ + X 0 ˆβ ) X 0 X 0 S X + + n S + X 0 X S + + X n X 0 X 0 S X + + n 4. { ˆβ 0 + ˆβ X 0 µ + S + + X n E α/σ, ˆβ < 0, ˆβ X0 X 0 S + + X n E α/σ}, 5. {( ˆβ 0 µ + X 0 ˆβ ) X 0 S X + + n 0 ˆβ < ( ˆβ 0 µ + X 0 ˆβ ) X 0 6. { ˆβ 0 + ˆβ X 0 + ˆβ X 0 µ + ˆβ ( ˆβ 0 µ + X 0 ˆβ ) X 0 X 0, ˆβ < X 0, ˆβ < X 0 + ˆβ E α/ σ, X 0 S X + + n X 0 + X 0 S X + + n E α/σ, X 0 S X + + n X0 S + + X n E α/σ}, S + + X n )E α/σ, X 0 S X + + n E α/σ},, ˆβ X0 X 0 S + + X n E α/σ},, ˆβ X 0 X 0 S X + + n E α/σ}. 45
52 Chapter 5 When σ is Unknown When σ is unknown, recall the hypothesis test (3.) G 0L : β 0 + β X 0 + β X 0 µ β 0, β 0 G 0U : β 0 + β X 0 + β X 0 µ β 0, β 0 (5.) and G : β 0, β 0 Test G 0L against G G 0L. In terms of r, the test becomes to G 0 : 0 r b c r 0 d r, 0 r against G : r 0, r 0. Then the LRT is Λ = ( σ σ )n /, where σ is the MLE of σ under G 0 and σ is the MLE of σ under G. Hence the LRT reject G 0 for large value of test statistic, λ = Λ /n = vs + ˆr r vs + ˆr r. Even we can change its form, Mukerjee and Tu (995) have shown some diffi- 46
53 culties of using this test statistic. Peiris and Bhattacharya (06) proposed a test statistic T which rejects G 0 with large values, where T = r r S and I use that for my forgoing discussion. The test reduced to the χ 0 test when σ is known by replacing S with σ. As Peiris and Bhattacharya (06) shown, the least favorable null distribution of LRT is P r(lrt C α ˆr = L) = 3 w i P (F i,n 3 Cα/i) i=0 where F i,n 3 is the F-distribution with i and n-3 degrees of freedom. If i=0, Let P (F i,n 3 C α/ /i) =. And the critical values C α can be computed using the equation α = w P (F,n 3 C α) + w P (F,n 3 C α/) + w 3 P (F 3,n 3 C α/3) Then the table of critical values are given by Peiris and Bhattacharya (06) in appendix. For other σ unknown cases, the rejection regions are very similar to the corresponding σ known cases with replacing σ with S and obtaining C α from above equation. I replace the Z α with t v,α in the boundaries of rejection region and prediction intervals when {X 0 > 0, X 0 > 0} and {X 0 < 0, X 0 < 0}. 47
54 Chapter 6 Prediction Intervals In this chapter, I summarize all the formulas for the prediction intervals for all the possible sign constraints of X 0 and X 0. When σ is known, the formulas for prediction intervals have similar formats as the formulas when σ is unknown. Hence, I only provide formulas of the prediction intervals when σ is unknown. 6. when X 0 > 0 and X 0 > 0 Lower Boundaries, L P =. ˆβ0 C α/ S + n if ˆβ < 0 ˆβ < 0,. ˆβ0 (Cα/ S S X ˆβ )( + n ) if ˆβ < 0 0 ˆβ X0 < C α/ S ( + )S n X 4 + X0S, X 3. ˆβ0 (Cα/ S ˆβ )( + n ) if 0 ˆβ X0 < C α/ S ( + )S n X 4 + X0S ˆβ < 0, X 48
55 4. ˆβ0 (Cα/ S ˆβ SX ˆβ )( + n ) if 0 ˆβ < C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S X 0 ˆβ < C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S, X 5. ˆβ0 + ˆβ X0 X 0 C α/ S S + + X n X0 if ˆβ < 0 ˆβ C α/ S ( + )S n X 4 + X0S, X 6. ˆβ0 + ˆβ X0 X 0 C α/ S S + + X n X0 if ˆβ C α/ S ( + )S n X 4 + X0S ˆβ < 0, X 7. ˆβ0 + ˆβ X 0 (C α/ S ˆβ )( + n + X 0 S ) X 8. ˆβ0 + ˆβ X 0 if 0 ˆβ < + S X 4 X 0 C α/ S ( X 0 S X + + n ) ˆβ > C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S, X (C α/ S S X ˆβ )( + n + X 0 S ) X if ˆβ > C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S X 0 ˆβ < C α/ σ, S X + S 4 X ( X X0 0 S X + + ) n 49
56 9. ˆβ0 + ˆβ X 0 + ˆβ X 0 C α/ S Upper Boundaries, + n + X 0 + X 0 S X if ˆβ > ˆβ > C α/ S S X + S 4 X ( X X0 0 S X + + ) n C α/ σ. S X + S 4 X ( X X0 0 S X + + ) n U P =. ˆβ0 + ˆβ X 0 if ˆβ < + X 0 n + X 0 S + X 0 X S T α/s ˆβ 0, X. ˆβ0 + ˆβ X 0 if ˆβ 0 ˆβ < 3. ˆβ0 if ˆβ < 0 ˆβ < min{0, 4. ˆβ0 + ˆβ X 0 + ˆβ X 0 + X 0 X 0 6. when X 0 < 0 and X 0 < 0 + n + X 0 + X 0 S X T α/s + n + X 0 + X 0 S X T α/s, + n + X 0 + X 0 S X T α/s}, otherwise. Lower Boundaries, L P =. ˆβ0 + ˆβ X 0 if ˆβ < + X 0 n + X 0 S + X 0 X S T α/s ˆβ 0, X. ˆβ0 + ˆβ X 0 if ˆβ 0 ˆβ < + X 0 n + X 0 S + X 0 X S T α/s, X 3. ˆβ0 if ˆβ < 0 ˆβ < min{0, + X 0 n + X 0 S + X 0 X S T α/s}, X 4. ˆβ0 + ˆβ X 0 + ˆβ X 0 + n + X 0 S + X 0 X S X 68 T α/s otherwise. 50
57 Upper Boundaries, U P =. ˆβ0 + C α/ S + if ˆβ < 0 ˆβ < 0 n. ˆβ0 + (Cα/ S S X ˆβ )( + n ) if ˆβ < 0 0 ˆβ X0 < C α/ S ( + )S n X 4 + X0S X 3. ˆβ0 + (Cα/ S ˆβ )( + n ) if 0 ˆβ X0 < C α/ S ( + )S n X 4 + X0S ˆβ < 0 X 4. ˆβ0 + (Cα/ S ˆβ SX ˆβ )( + n ) if 0 ˆβ < C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S X 0 ˆβ < C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S X 5. ˆβ0 + ˆβ X0 X 0 + C α/ S S + + X n X0 if ˆβ < 0 ˆβ C α/ S ( + )S n X 4 + X0S X 6. ˆβ0 + ˆβ X0 X 0 + C α/ S S + + X n X0 if ˆβ C α/ S ( + )S n X 4 + X0S ˆβ < 0 X 5
58 7. ˆβ0 + ˆβ X ˆβ0 + ˆβ X ˆβ0 + ˆβ X 0 + ˆβ X 0 + C α/ S if ˆβ > (C α/ S ˆβ )( + n + X 0 S X ) if 0 ˆβ C α/ S < S X + S 4 X ( X X0 0 S X + + ) n ˆβ > C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S X (C α/ S S X ˆβ )( + n + X 0 S ) X if ˆβ > C α/ S X 0( S X ˆβ ) Cα/ S ( + )S n X 4 + X0S X C α/ S 0 ˆβ < S X + S 4 X ( X X0 0 S X + + ) n + n + X 0 + X 0 S X ˆβ > C α/ σ S X + S 4 X ( X X0 0 S X + + ) n C α/ σ S X + S 4 X ( X X0 0 S X + + ) n 5
59 6.3 when X 0 > 0 and X 0 < 0 Lower Boundaries, L p =. ˆβ0 if ˆβ < 0 ˆβ < X 0 X0 S + X n + E α/s. ˆβ0 ( X 0 S + X n + )(E α/ S X ˆβ ) 0 S + X n + E α/s if 0 ˆβ < X ˆβ 0 + ( + + X 0 n S X + X 0 S X + + X 0 n S X )E α/ S X 0 ( + + X 0 n ˆβ < X0 X 0 S + X n + E α/s 3. ˆβ0 + ˆβ X0 X 0 ( S + X 0 X S + X n + X0 S + X n + )E α/s if ˆβ X ˆβ 0 + ( + + X 0 n S X + X 0 S X + + X 0 n S X )E α/ S S X X 0 ( + + X 0 n 4. ˆβ0 + ˆβ X0 X 0 S + X n + E α/s 5. ˆβ0 + ˆβ X 0 S X + X 0 S X ) S X + X 0 S X ) ˆβ < X 0 if ˆβ < 0 ˆβ X 0 ( X 0 S + X n + )(E α/ S ˆβ ) if 0 ˆβ < ˆβ X 0 S X + S 4 X X0 X 0 S X + n + E α/s X 0 S X + n + E α/s E α/ S ( X 0 S X + n + ) X 0 S X + n + E α/s 53
Ch 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationFormal Statement of Simple Linear Regression Model
Formal Statement of Simple Linear Regression Model Y i = β 0 + β 1 X i + ɛ i Y i value of the response variable in the i th trial β 0 and β 1 are parameters X i is a known constant, the value of the predictor
More informationSimple and Multiple Linear Regression
Sta. 113 Chapter 12 and 13 of Devore March 12, 2010 Table of contents 1 Simple Linear Regression 2 Model Simple Linear Regression A simple linear regression model is given by Y = β 0 + β 1 x + ɛ where
More informationBias Variance Trade-off
Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]
More informationSTAT 540: Data Analysis and Regression
STAT 540: Data Analysis and Regression Wen Zhou http://www.stat.colostate.edu/~riczw/ Email: riczw@stat.colostate.edu Department of Statistics Colorado State University Fall 205 W. Zhou (Colorado State
More informationInference in Normal Regression Model. Dr. Frank Wood
Inference in Normal Regression Model Dr. Frank Wood Remember We know that the point estimator of b 1 is b 1 = (Xi X )(Y i Ȳ ) (Xi X ) 2 Last class we derived the sampling distribution of b 1, it being
More information17: INFERENCE FOR MULTIPLE REGRESSION. Inference for Individual Regression Coefficients
17: INFERENCE FOR MULTIPLE REGRESSION Inference for Individual Regression Coefficients The results of this section require the assumption that the errors u are normally distributed. Let c i ij denote the
More informationDirection: This test is worth 250 points and each problem worth points. DO ANY SIX
Term Test 3 December 5, 2003 Name Math 52 Student Number Direction: This test is worth 250 points and each problem worth 4 points DO ANY SIX PROBLEMS You are required to complete this test within 50 minutes
More informationLikelihood Inference for Order Restricted Models
Likelihood Inference for Order Restricted Models by Afnan Almazmomi A Thesis Submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUTE In partial fulfillment of the requirements for the Degree of
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More informationLinear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,
Linear Regression In this problem sheet, we consider the problem of linear regression with p predictors and one intercept, y = Xβ + ɛ, where y t = (y 1,..., y n ) is the column vector of target values,
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationSummer School in Statistics for Astronomers V June 1 - June 6, Regression. Mosuk Chow Statistics Department Penn State University.
Summer School in Statistics for Astronomers V June 1 - June 6, 2009 Regression Mosuk Chow Statistics Department Penn State University. Adapted from notes prepared by RL Karandikar Mean and variance Recall
More informationLecture 14 Simple Linear Regression
Lecture 4 Simple Linear Regression Ordinary Least Squares (OLS) Consider the following simple linear regression model where, for each unit i, Y i is the dependent variable (response). X i is the independent
More informationDiagnostics can identify two possible areas of failure of assumptions when fitting linear models.
1 Transformations 1.1 Introduction Diagnostics can identify two possible areas of failure of assumptions when fitting linear models. (i) lack of Normality (ii) heterogeneity of variances It is important
More informationA Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when the Covariance Matrices are Unknown but Common
Journal of Statistical Theory and Applications Volume 11, Number 1, 2012, pp. 23-45 ISSN 1538-7887 A Test for Order Restriction of Several Multivariate Normal Mean Vectors against all Alternatives when
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7
MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is
More informationTesting Hypothesis. Maura Mezzetti. Department of Economics and Finance Università Tor Vergata
Maura Department of Economics and Finance Università Tor Vergata Hypothesis Testing Outline It is a mistake to confound strangeness with mystery Sherlock Holmes A Study in Scarlet Outline 1 The Power Function
More informationINTERVAL ESTIMATION AND HYPOTHESES TESTING
INTERVAL ESTIMATION AND HYPOTHESES TESTING 1. IDEA An interval rather than a point estimate is often of interest. Confidence intervals are thus important in empirical work. To construct interval estimates,
More informationHypothesis Test. The opposite of the null hypothesis, called an alternative hypothesis, becomes
Neyman-Pearson paradigm. Suppose that a researcher is interested in whether the new drug works. The process of determining whether the outcome of the experiment points to yes or no is called hypothesis
More informationSTAT763: Applied Regression Analysis. Multiple linear regression. 4.4 Hypothesis testing
STAT763: Applied Regression Analysis Multiple linear regression 4.4 Hypothesis testing Chunsheng Ma E-mail: cma@math.wichita.edu 4.4.1 Significance of regression Null hypothesis (Test whether all β j =
More informationInference in Regression Analysis
Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value
More informationProbability and Statistics Notes
Probability and Statistics Notes Chapter Seven Jesse Crawford Department of Mathematics Tarleton State University Spring 2011 (Tarleton State University) Chapter Seven Notes Spring 2011 1 / 42 Outline
More informationChapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests
Chapter 8: Hypothesis Testing Lecture 9: Likelihood ratio tests Throughout this chapter we consider a sample X taken from a population indexed by θ Θ R k. Instead of estimating the unknown parameter, we
More informationProblems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B
Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2
More informationLinear models and their mathematical foundations: Simple linear regression
Linear models and their mathematical foundations: Simple linear regression Steffen Unkel Department of Medical Statistics University Medical Center Göttingen, Germany Winter term 2018/19 1/21 Introduction
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part : Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationEstimating σ 2. We can do simple prediction of Y and estimation of the mean of Y at any value of X.
Estimating σ 2 We can do simple prediction of Y and estimation of the mean of Y at any value of X. To perform inferences about our regression line, we must estimate σ 2, the variance of the error term.
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationMaximum Likelihood Estimation
Maximum Likelihood Estimation Merlise Clyde STA721 Linear Models Duke University August 31, 2017 Outline Topics Likelihood Function Projections Maximum Likelihood Estimates Readings: Christensen Chapter
More informationMultiple Linear Regression
Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there
More informationTMA 4275 Lifetime Analysis June 2004 Solution
TMA 4275 Lifetime Analysis June 2004 Solution Problem 1 a) Observation of the outcome is censored, if the time of the outcome is not known exactly and only the last time when it was observed being intact,
More informationRegression diagnostics
Regression diagnostics Kerby Shedden Department of Statistics, University of Michigan November 5, 018 1 / 6 Motivation When working with a linear model with design matrix X, the conventional linear model
More informationLinear Models and Estimation by Least Squares
Linear Models and Estimation by Least Squares Jin-Lung Lin 1 Introduction Causal relation investigation lies in the heart of economics. Effect (Dependent variable) cause (Independent variable) Example:
More informationRegression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood
Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing
More informationSimple Linear Regression
Simple Linear Regression In simple linear regression we are concerned about the relationship between two variables, X and Y. There are two components to such a relationship. 1. The strength of the relationship.
More informationLECTURE 5 HYPOTHESIS TESTING
October 25, 2016 LECTURE 5 HYPOTHESIS TESTING Basic concepts In this lecture we continue to discuss the normal classical linear regression defined by Assumptions A1-A5. Let θ Θ R d be a parameter of interest.
More informationLecture 28: Asymptotic confidence sets
Lecture 28: Asymptotic confidence sets 1 α asymptotic confidence sets Similar to testing hypotheses, in many situations it is difficult to find a confidence set with a given confidence coefficient or level
More informationStatistics 135 Fall 2008 Final Exam
Name: SID: Statistics 135 Fall 2008 Final Exam Show your work. The number of points each question is worth is shown at the beginning of the question. There are 10 problems. 1. [2] The normal equations
More informationThis model of the conditional expectation is linear in the parameters. A more practical and relaxed attitude towards linear regression is to say that
Linear Regression For (X, Y ) a pair of random variables with values in R p R we assume that E(Y X) = β 0 + with β R p+1. p X j β j = (1, X T )β j=1 This model of the conditional expectation is linear
More informationAssociation studies and regression
Association studies and regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Association studies and regression 1 / 104 Administration
More information(a) (3 points) Construct a 95% confidence interval for β 2 in Equation 1.
Problem 1 (21 points) An economist runs the regression y i = β 0 + x 1i β 1 + x 2i β 2 + x 3i β 3 + ε i (1) The results are summarized in the following table: Equation 1. Variable Coefficient Std. Error
More informationMore Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction
Sankhyā : The Indian Journal of Statistics 2007, Volume 69, Part 4, pp. 700-716 c 2007, Indian Statistical Institute More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationNotes on the Multivariate Normal and Related Topics
Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions
More informationMatrix Approach to Simple Linear Regression: An Overview
Matrix Approach to Simple Linear Regression: An Overview Aspects of matrices that you should know: Definition of a matrix Addition/subtraction/multiplication of matrices Symmetric/diagonal/identity matrix
More informationReview: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.
1 Review: Let X 1, X,..., X n denote n independent random variables sampled from some distribution might not be normal!) with mean µ) and standard deviation σ). Then X µ σ n In other words, X is approximately
More informationMath 423/533: The Main Theoretical Topics
Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)
More informationCh 3: Multiple Linear Regression
Ch 3: Multiple Linear Regression 1. Multiple Linear Regression Model Multiple regression model has more than one regressor. For example, we have one response variable and two regressor variables: 1. delivery
More informationAMS 315/576 Lecture Notes. Chapter 11. Simple Linear Regression
AMS 315/576 Lecture Notes Chapter 11. Simple Linear Regression 11.1 Motivation A restaurant opening on a reservations-only basis would like to use the number of advance reservations x to predict the number
More informationWe like to capture and represent the relationship between a set of possible causes and their response, by using a statistical predictive model.
Statistical Methods in Business Lecture 5. Linear Regression We like to capture and represent the relationship between a set of possible causes and their response, by using a statistical predictive model.
More informationPh.D. Qualifying Exam Friday Saturday, January 3 4, 2014
Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently
More informationRegression and Statistical Inference
Regression and Statistical Inference Walid Mnif wmnif@uwo.ca Department of Applied Mathematics The University of Western Ontario, London, Canada 1 Elements of Probability 2 Elements of Probability CDF&PDF
More information14.30 Introduction to Statistical Methods in Economics Spring 2009
MIT OpenCourseWare http://ocw.mit.edu 4.0 Introduction to Statistical Methods in Economics Spring 009 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationLinear Methods for Prediction
Chapter 5 Linear Methods for Prediction 5.1 Introduction We now revisit the classification problem and focus on linear methods. Since our prediction Ĝ(x) will always take values in the discrete set G we
More informationwhere x and ȳ are the sample means of x 1,, x n
y y Animal Studies of Side Effects Simple Linear Regression Basic Ideas In simple linear regression there is an approximately linear relation between two variables say y = pressure in the pancreas x =
More informationThe linear model is the most fundamental of all serious statistical models encompassing:
Linear Regression Models: A Bayesian perspective Ingredients of a linear model include an n 1 response vector y = (y 1,..., y n ) T and an n p design matrix (e.g. including regressors) X = [x 1,..., x
More informationIntroduction to Simple Linear Regression
Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department
More informationHypothesis Testing. 1 Definitions of test statistics. CB: chapter 8; section 10.3
Hypothesis Testing CB: chapter 8; section 0.3 Hypothesis: statement about an unknown population parameter Examples: The average age of males in Sweden is 7. (statement about population mean) The lowest
More informationNonparametric Regression and Bonferroni joint confidence intervals. Yang Feng
Nonparametric Regression and Bonferroni joint confidence intervals Yang Feng Simultaneous Inferences In chapter 2, we know how to construct confidence interval for β 0 and β 1. If we want a confidence
More informationFirst Year Examination Department of Statistics, University of Florida
First Year Examination Department of Statistics, University of Florida August 20, 2009, 8:00 am - 2:00 noon Instructions:. You have four hours to answer questions in this examination. 2. You must show
More informationy ˆ i = ˆ " T u i ( i th fitted value or i th fit)
1 2 INFERENCE FOR MULTIPLE LINEAR REGRESSION Recall Terminology: p predictors x 1, x 2,, x p Some might be indicator variables for categorical variables) k-1 non-constant terms u 1, u 2,, u k-1 Each u
More informationChapter 1. Linear Regression with One Predictor Variable
Chapter 1. Linear Regression with One Predictor Variable 1.1 Statistical Relation Between Two Variables To motivate statistical relationships, let us consider a mathematical relation between two mathematical
More informationLecture 3: Multiple Regression
Lecture 3: Multiple Regression R.G. Pierse 1 The General Linear Model Suppose that we have k explanatory variables Y i = β 1 + β X i + β 3 X 3i + + β k X ki + u i, i = 1,, n (1.1) or Y i = β j X ji + u
More informationLet us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided
Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or
More informationStatistical Inference with Monotone Incomplete Multivariate Normal Data
Statistical Inference with Monotone Incomplete Multivariate Normal Data p. 1/4 Statistical Inference with Monotone Incomplete Multivariate Normal Data This talk is based on joint work with my wonderful
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More information3 Multiple Linear Regression
3 Multiple Linear Regression 3.1 The Model Essentially, all models are wrong, but some are useful. Quote by George E.P. Box. Models are supposed to be exact descriptions of the population, but that is
More informationPart IB Statistics. Theorems with proof. Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua. Lent 2015
Part IB Statistics Theorems with proof Based on lectures by D. Spiegelhalter Notes taken by Dexter Chua Lent 2015 These notes are not endorsed by the lecturers, and I have modified them (often significantly)
More informationMATH5745 Multivariate Methods Lecture 07
MATH5745 Multivariate Methods Lecture 07 Tests of hypothesis on covariance matrix March 16, 2018 MATH5745 Multivariate Methods Lecture 07 March 16, 2018 1 / 39 Test on covariance matrices: Introduction
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationPeter Hoff Linear and multilinear models April 3, GLS for multivariate regression 5. 3 Covariance estimation for the GLM 8
Contents 1 Linear model 1 2 GLS for multivariate regression 5 3 Covariance estimation for the GLM 8 4 Testing the GLH 11 A reference for some of this material can be found somewhere. 1 Linear model Recall
More informationChapter 7. Hypothesis Testing
Chapter 7. Hypothesis Testing Joonpyo Kim June 24, 2017 Joonpyo Kim Ch7 June 24, 2017 1 / 63 Basic Concepts of Testing Suppose that our interest centers on a random variable X which has density function
More informationSTAT 514 Solutions to Assignment #6
STAT 514 Solutions to Assignment #6 Question 1: Suppose that X 1,..., X n are a simple random sample from a Weibull distribution with density function f θ x) = θcx c 1 exp{ θx c }I{x > 0} for some fixed
More informationSummary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)
Summary of Chapter 7 (Sections 7.2-7.5) and Chapter 8 (Section 8.1) Chapter 7. Tests of Statistical Hypotheses 7.2. Tests about One Mean (1) Test about One Mean Case 1: σ is known. Assume that X N(µ, σ
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationSection 4.6 Simple Linear Regression
Section 4.6 Simple Linear Regression Objectives ˆ Basic philosophy of SLR and the regression assumptions ˆ Point & interval estimation of the model parameters, and how to make predictions ˆ Point and interval
More informationMISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30
MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationLecture 12 November 3
STATS 300A: Theory of Statistics Fall 2015 Lecture 12 November 3 Lecturer: Lester Mackey Scribe: Jae Hyuck Park, Christian Fong Warning: These notes may contain factual and/or typographic errors. 12.1
More informationMath 3330: Solution to midterm Exam
Math 3330: Solution to midterm Exam Question 1: (14 marks) Suppose the regression model is y i = β 0 + β 1 x i + ε i, i = 1,, n, where ε i are iid Normal distribution N(0, σ 2 ). a. (2 marks) Compute the
More informationLecture 6 Multiple Linear Regression, cont.
Lecture 6 Multiple Linear Regression, cont. BIOST 515 January 22, 2004 BIOST 515, Lecture 6 Testing general linear hypotheses Suppose we are interested in testing linear combinations of the regression
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More informationLinear Models Review
Linear Models Review Vectors in IR n will be written as ordered n-tuples which are understood to be column vectors, or n 1 matrices. A vector variable will be indicted with bold face, and the prime sign
More informationLinear Methods for Prediction
This work is licensed under a Creative Commons Attribution-NonCommercial-ShareAlike License. Your use of this material constitutes acceptance of that license and the conditions of use of materials on this
More informationTHE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay
THE UNIVERSITY OF CHICAGO Booth School of Business Business 41912, Spring Quarter 2016, Mr. Ruey S. Tsay Lecture 5: Multivariate Multiple Linear Regression The model is Y n m = Z n (r+1) β (r+1) m + ɛ
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationLecture 01: Introduction
Lecture 01: Introduction Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology Medical University of South Carolina Lecture 01: Introduction
More informationSimple linear regression
Simple linear regression Biometry 755 Spring 2008 Simple linear regression p. 1/40 Overview of regression analysis Evaluate relationship between one or more independent variables (X 1,...,X k ) and a single
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the
More informationLikelihood Ratio Test of a General Linear Hypothesis. Copyright c 2012 Dan Nettleton (Iowa State University) Statistics / 42
Likelihood Ratio Test of a General Linear Hypothesis Copyright c 2012 Dan Nettleton (Iowa State University) Statistics 611 1 / 42 Consider the Likelihood Ratio Test of H 0 : Cβ = d vs H A : Cβ d. Suppose
More informationQualifying Exam in Probability and Statistics. https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf
Part 1: Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More informationThe outline for Unit 3
The outline for Unit 3 Unit 1. Introduction: The regression model. Unit 2. Estimation principles. Unit 3: Hypothesis testing principles. 3.1 Wald test. 3.2 Lagrange Multiplier. 3.3 Likelihood Ratio Test.
More informationSTAT 135 Lab 5 Bootstrapping and Hypothesis Testing
STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,
More informationEconomics 520. Lecture Note 19: Hypothesis Testing via the Neyman-Pearson Lemma CB 8.1,
Economics 520 Lecture Note 9: Hypothesis Testing via the Neyman-Pearson Lemma CB 8., 8.3.-8.3.3 Uniformly Most Powerful Tests and the Neyman-Pearson Lemma Let s return to the hypothesis testing problem
More informationA Likelihood Ratio Test
A Likelihood Ratio Test David Allen University of Kentucky February 23, 2012 1 Introduction Earlier presentations gave a procedure for finding an estimate and its standard error of a single linear combination
More informationStatistical Inference
Statistical Inference Classical and Bayesian Methods Revision Class for Midterm Exam AMS-UCSC Th Feb 9, 2012 Winter 2012. Session 1 (Revision Class) AMS-132/206 Th Feb 9, 2012 1 / 23 Topics Topics We will
More informationMultivariate Statistics
Multivariate Statistics Chapter 2: Multivariate distributions and inference Pedro Galeano Departamento de Estadística Universidad Carlos III de Madrid pedro.galeano@uc3m.es Course 2016/2017 Master in Mathematical
More informationQualifying Exam in Probability and Statistics.
Part 1: Sample Problems for the Elementary Section of Qualifying Exam in Probability and Statistics https://www.soa.org/files/edu/edu-exam-p-sample-quest.pdf Part 2: Sample Problems for the Advanced Section
More information