Response surface designs using the generalized variance inflation factors

Similar documents
Mitigating collinearity in linear regression models using ridge, surrogate and raised estimators

A note on the unique solution of linear complementarity problem

Revisting Some Design Criteria

The plastic number and its generalized polynomial

Matrix l-algebras over l-fields

On a mixed interpolation with integral conditions at arbitrary nodes

Finding the strong defining hyperplanes of production possibility set with constant returns to scale using the linear independent vectors

Some aspects on hesitant fuzzy soft set

Graded fuzzy topological spaces

Derivation, f-derivation and generalized derivation of KUS-algebras

Petrović s inequality on coordinates and related results

Draft of an article prepared for the Encyclopedia of Social Science Research Methods, Sage Publications. Copyright by John Fox 2002

Non-linear unit root testing with arctangent trend: Simulation and applications in finance

Multicollinearity and A Ridge Parameter Estimation Approach

A detection for patent infringement suit via nanotopology induced by graph

Existence and uniqueness of a stationary and ergodic solution to stochastic recurrence equations via Matkowski s FPT

Response Surface Methodology

Wiley. Methods and Applications of Linear Models. Regression and the Analysis. of Variance. Third Edition. Ishpeming, Michigan RONALD R.

Useful Numerical Statistics of Some Response Surface Methodology Designs

ELECTRICAL & ELECTRONIC ENGINEERING RESEARCH ARTICLE

Efficient Choice of Biasing Constant. for Ridge Regression

Experimental designs for multiple responses with different models

Ridge Regression and Ill-Conditioning

Double domination in signed graphs

Classes of Second-Order Split-Plot Designs

Response Surface Methodology:

Relationship between ridge regression estimator and sample size when multicollinearity present among regressors

Minimizing Oblique Errors for Robust Estimating

Generalizations on Humbert polynomials and functions

Response Surface Methodology: Process and Product Optimization Using Designed Experiments, 3rd Edition

Geometric View of Measurement Errors

On some properties of p-ideals based on intuitionistic fuzzy sets

EIGENVALUES AND SINGULAR VALUE DECOMPOSITION

On the complex k-fibonacci numbers

Comparison of Some Improved Estimators for Linear Regression Model under Different Conditions

Bounds on Hankel determinant for starlike and convex functions with respect to symmetric points

Multivariate Statistical Analysis

Review of Linear Algebra

Definitive Screening Designs with Added Two-Level Categorical Factors *

Optimality of Central Composite Designs Augmented from One-half Fractional Factorial Designs

Regression Diagnostics for Survey Data

Intermediate Algebra. Gregg Waterman Oregon Institute of Technology

Linear Models 1. Isfahan University of Technology Fall Semester, 2014

Singular Value Decomposition Compared to cross Product Matrix in an ill Conditioned Regression Model

Mathematics. EC / EE / IN / ME / CE. for

The combined reproducing kernel method and Taylor series to solve nonlinear Abel s integral equations with weakly singular kernel

Single and multiple linear regression analysis

Minimax design criterion for fractional factorial designs

Dynamics of the equation complex plane

Moment Aberration Projection for Nonregular Fractional Factorial Designs

DIMENSION REDUCTION OF THE EXPLANATORY VARIABLES IN MULTIPLE LINEAR REGRESSION. P. Filzmoser and C. Croux

Some inequalities for unitarily invariant norms of matrices

8 RESPONSE SURFACE DESIGNS

Improved Ridge Estimator in Linear Regression with Multicollinearity, Heteroscedastic Errors and Outliers

The successive raising estimator and its relation with the ridge estimator

Available online at (Elixir International Journal) Statistics. Elixir Statistics 49 (2012)

((n r) 1) (r 1) ε 1 ε 2. X Z β+

I L L I N O I S UNIVERSITY OF ILLINOIS AT URBANA-CHAMPAIGN

Multilevel modeling and panel data analysis in educational research (Case study: National examination data senior high school in West Java)

University, Wuhan, China c College of Physical Science and Technology, Central China Normal. University, Wuhan, China Published online: 25 Apr 2014.

An Introduction to Multivariate Statistical Analysis

1 Data Arrays and Decompositions

Exercises * on Principal Component Analysis

Day 4: Shrinkage Estimators

component risk analysis

Multicollinearity Exercise

Linear Regression. In this problem sheet, we consider the problem of linear regression with p predictors and one intercept,

Lecture 1: OLS derivations and inference

APPLICATION OF RIDGE REGRESSION TO MULTICOLLINEAR DATA

Nonlinear least squares algorithm for identification of hazards

Appendix A: Matrices

Next is material on matrix rank. Please see the handout

EXAMINATIONS OF THE ROYAL STATISTICAL SOCIETY (formerly the Examinations of the Institute of Statisticians) GRADUATE DIPLOMA, 2007

Machine learning for pervasive systems Classification in high-dimensional spaces

Using Ridge Least Median Squares to Estimate the Parameter by Solving Multicollinearity and Outliers Problems

AN IMPROVEMENT TO THE ALIGNED RANK STATISTIC

Stochastic Design Criteria in Linear Models

Variable Selection in Restricted Linear Regression Models. Y. Tuaç 1 and O. Arslan 1

Theorems. Least squares regression

Forecast comparison of principal component regression and principal covariate regression

Researchers often record several characters in their research experiments where each character has a special significance to the experimenter.

Math Camp Lecture 4: Linear Algebra. Xiao Yu Wang. Aug 2010 MIT. Xiao Yu Wang (MIT) Math Camp /10 1 / 88

φ-contractive multivalued mappings in complex valued metric spaces and remarks on some recent papers

The Precise Effect of Multicollinearity on Classification Prediction

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

Psychology Seminar Psych 406 Dr. Jeffrey Leitzel

Vector control of asymmetrical six-phase synchronous motor

Linear Algebra: Matrix Eigenvalue Problems

USING REGULAR FRACTIONS OF TWO-LEVEL DESIGNS TO FIND BASELINE DESIGNS

ISyE 691 Data mining and analytics

JORDAN NORMAL FORM NOTES

SAMPLE OF THE STUDY MATERIAL PART OF CHAPTER 1 Introduction to Linear Algebra

Full terms and conditions of use:

Detection of Influential Observation in Linear Regression. R. Dennis Cook. Technometrics, Vol. 19, No. 1. (Feb., 1977), pp

MSE Performance and Minimax Regret Significance Points for a HPT Estimator when each Individual Regression Coefficient is Estimated

Effects of Outliers and Multicollinearity on Some Estimators of Linear Regression Model

Checking model assumptions with regression diagnostics

MS-C1620 Statistical inference

Linear Regression Models

Least Squares Estimation-Finite-Sample Properties

Transcription:

STATISTICS RESEARCH ARTICLE Response surface designs using the generalized variance inflation factors Diarmuid O Driscoll and Donald E Ramirez 2 * Received: 22 December 204 Accepted: 5 May 205 Published: 4 June 205 *Corresponding author: Donald E Ramirez, Department of Mathematics, University of Virginia, Charlottesville, VA 22904, USA E-mail: der@virginiaedu Reviewing editor: Guohua Zou, Chinese Academy of Sciences, China Additional information is available at the end of the article Abstract: We study response surface designs using the generalized variance inflation factors for subsets as an extension of the variance inflation factors Subjects: Mathematics & Statistics; Science; Statistical Computing; Statistical Theory & Methods; Statistics; Statistics & Probability Keywords: variance inflation factors; generalized variance inflation factors; response surface designs Introduction We consider a linear regression Y β + ε with a full rank n p matrix and (ε) N(0, σ 2 I n ) The variance inflation factor VIF, Belsley (986), measures the penalty for adding one non-orthogonal additional explanatory variable to a linear regression model, and they can be computed as a ratio of determinants The extension of VIF to a measure of the penalty for adding a subset of variables to a model is the generalized variance inflation factor GVIF of Fox and Monette (992), which will be used to study response surface designs, in particular, as the penalty for adding the quadratic terms to the model 2 Variance inflation factors For our linear model Y β + ε, let D be the diagonal matrix with entries on the diagonal D [i, i ( ) 2 When the design has been standardized D i,i, the VIFs are the diagonal entries of the inverse of S D ( )D That is, the VIFs are the ratios of the actual variances for ABOUT THE AUTHORS Diarmuid O Driscoll is the head of the Mathematics and Computer Studies Department at Mary Immaculate College, Limerick He was awarded a Travelling Studentship for his MSc at University College Cork in 977 He has taught at University College Cork, Cork Institute of Technology, University of Virginia, and Frostburg State University His research interests are in mathematical education, errors in variables regression, and design criteria In 204, he was awarded a Teaching Heroes Award by the National Forum for the Enhancement of Teaching and Learning (Ireland) Donald E Ramirez is a full professor in the Department of Mathematics at the University of Virginia in Charlottesville, Virginia He received his PhD in Mathematics from Tulane University in New Orleans, Louisiana His research is in harmonic analysis and mathematical statistics His current research interests are in statistical outliers and ridge regression PUBLIC INTEREST STATEMENT Response surface designs are a mainstay in applied statistics The variance inflation factors VIF are a measure of collinearity for a single variable in a linear regression model The generalization to subsets of variables is the generalized variance inflation factor GVIF This research introduces GVIF as a penalty measure for extending a linear response model to a response surface with the included quadratic terms The methodology is demonstrated with case studies, and, in particular, it is shown that using GVIF, the H30 design can be improved for the standard global optimality criteria of A, D, and E 205 The Author(s) This open access article is distributed under a Creative Commons Attribution (CC-BY) 40 license Page of

the explanatory variables to the ideal variances had the columns of been orthogonal Note that we follow Stewart (987) and do not necessarily center the explanatory variables For our linear model Y β + ε, view [ [p, x p with x p the p th column of and [p the matrix formed by the remaining columns The variance inflation factor VIF p measures the effect of adding column x p to [p For notational convenience, we demonstrate VIF p with the last column p An ideal column would be orthogonal to the previous columns with the entries in the off-diagonal elements of the p th row and p th column of all zeros Denote by M p the idealized moment matrix [ [p M p [p 0 p 0 p x x p p The VIFs are the diagonal entries of S D ( ) D It remains to note that the inverse, S, can be computed using cofactors C i,j In particular, VIF p [S p,p [D ( ) D p,p (x p x p ) 2 det(c p,p ) det( ) (x p x p ) 2 det(m p ) det( ) the ratio of the determinant of the idealized moment matrix M p to the determinant of the moment matrix This definition extends naturally to subsets and is discussed in the next section For an alternate view of the how collinearities in the explanatory variables inflate the model variances of the regression coefficients when compared to a fictitious orthogonal reference design, consider the formula for the model variance Var M ( β j ) where R 2 j is the square of the multiple correlation from the regression of the jth column of [x ij on the remaining columns as in Liao and Valliant (202) The first term σ 2 (x ij x j ) 2 is the model variance for β j had the j th explanatory variable been orthogonal to the remaining variables The second term ( R 2 j ) is a standard definition of the jth VIF as in Thiel (97) 3 Generalized variance inflation factors In this section, we introduce the GVIFs as an extension of the classical variance inflation factors VIF from Equation For the linear model Y β + ε, view [, 2 partitioned with of dimension n r usually consisting of the lower order terms and 2 of dimension n s usually consisting of the higher order terms The idealized moment matrix for the (r, s) partitioning of is M (r,s) σ 2 n i (x ij x j )2 R 2 j [ 0 r s 0 s r 2 2 () Following Equation, to measure the effect of adding 2 to the design, that is for 2, we define the generalized variance inflation factor as GVIF( 2 ) det(m (r,s) ) det( ) det( ) det( 2 2 ) det( ) as in Equation 0 of Fox and Monette (992), who compared the sizes of the joint confidence regions for β for partitioned designs and noted when [ [p, x p that GVIF[x p [p VIF p Equation 2 is in the spirit of the efficiency comparisons in linear inferences introduced in Theorems 4 and 5 of Jensen and Ramirez (993) A similar measure of collinearity is mentioned in Note 2 in Wichers (975), (2) Page 2 of

Theorem of Berk (977), and Garcia, Garcia, and Soto (20) For the simple linear regression model with p 2, Equation 2 gives VIF ρ 2 with ρ the correlation coefficient as required Fox and Monette (992) suggested that contains the variables which are of simultaneous interest, while 2 contains additional variables selected by the investigator We will set for the constant and main effects and set 2 the (optional) quadratic terms with values from Willan and Watts (978) measured the effect of collinearity using the ratio of the volume of the actual joint confidence region for β to the volume of the joint confidence region in the fictitious orthogonal reference design Their ratio is in the spirit of GVIF as det( ) is inversely proportional to the square of the volume of the joint confidence region for β They also introduced a measure of relative predictability and they note: The existence of near linear relations in the independent variables of the actual data reduces the overall predictive efficiency by this factor For a simple case study, consider the simple linear regression model with n 4, x [ 2,,, 2, and y [4,,, 4 The 95% prediction interval for x 0 is 25 ± 020 If the model also includes [ 200, 00, 00, 200, then the 95% prediction interval for (x, )(0, 0) is 25 ± 4602 demonstrating the loss of predictive efficiency due to the collinearity introduced by For the (r, s) partition of [, 2 with of dimension n r and 2 of dimension n s, set [ ( D (r,s) 2 ) 0 ( 0 ) 2, 2 2 and denote the canonical moment matrix as R D (r,s) ( )D (r,s) [ ( I r r ) 2 ( ) ( ) 2 ( 2 2 2 ) 2 2 2 ( ) ( 2 ) 2 ; I s s with determinant det(r) equivalently, det( ) det( ) det( 2 2 ) GVIF( 2 ) ; det(r) det(i r r B r s B s r )det(i s s B s r B r s ) where B r s ( ) 2 ( ) ( 2 2 2 2) (3) In the case {r p, s }, 2 x p is a n vector and the partitioned design [, x p has det(r) [x ( p ) ) x p (x x p p From standard facts for the inverse of a partitioned matrix, for example, Myers (990, p 459), VIF p [R p,p [D (p,) ( ) D (p,) p,p can be computed directly as ( ) 2 ( ) 2 x x p p ( ) p,p x x x x p p p p x x p p x ( p ) x p [x ( p ) ) x p (x x p p det(r) GVIF( ) 2 Page 3 of

We study the eigenvalue structure ( of M (r,s) in Appendix Let {λ λ 2 λ min(r,s) 0} be the non-negative singular values of ) 2 ( ) ( 2 2 2 2) It is shown in Appendix that an alternative formulation for GVIF is GVIF( 2 ) ( λ 2 i ) min(r,s) i (4) 4 Quadratic model with p 3 For the partitioning [ r s, the canonical moment matrix, Equation 3, has the identity matrices I r, I ( s down the diagonal and off-diagonal array ) 2 ( 2 2 2 2) For the quadratic model y β 0 + β x + β 2 and partitioning [, x, we have R 0 ρ 0 ρ 2 ρ ρ 2 From Equation 4, GVIF(, x) ( λ 2 ) where λ value of [ρ, ρ 2 Denote γ 2 ρ2 + ρ2 2 ρ 2 + ρ2 2 is the unique positive singular as the canonical index with GVIF(, x) Surprisingly, many higher order designs γ 2 det(r) also have the off-diagonal entry of the canonical moment matrix with a unique positive singular value with GVIF( 2 ) with the collinearity between the lower order terms and the upper γ 2 order terms as a function of the canonical index γ 2 5 Central composite and factorial designs for quadratic models (p 6) In this section, we compare the central composite design (CCD) of Box and Wilson (95) and the factorial design Z The design points are shown in Table B of Appendi Both designs are 9 6 and use the quadratic response model y β 0 + β x + β 2 + β + β 22 x2 + β x x + ε 2 2 2 The CCD traditionally uses the value a 2 in four entries, while the factorial design uses the value a To study the difference in the designs with these different values, we computed the GVIF to compare the orthogonality between the lower order terms of dimension 9 3 and the higher order quadratic terms 2 of dimension 9 3 The off-diagonal B 3 3 entry of R from Equation 3 in Section 3 has the form B 3 3 ρ ρ 2 0 0 0 0 0 0 0 Table CCD with parameter a, canonical index γ 2, and GVIF a GVIF( 2 ) γ 2 0000 04444 800 000 08000 5000 500 08858 8758 2 08889 9000 750 0854 6729 04444 800 Page 4 of

2+a 2 with ρ ρ 2 2 3 8+2a 4, canonical index γ2 ρ2 + ρ2 2 and GVIF( ) 2 as in the quadratic γ 2 model case with p 3 shown in the Section 4 For Table, if a, then ρ ρ 2 2 0, γ 2 8 0, and GVIF( 2 )5 Surprisingly, the classical choice of a 2 gives the largest value for GVIF( 2 ), that is the worst value, indicating the greatest collinearity between the lower and higher order terms, as noted in O Driscoll and Ramirez (in press) 6 Larger designs (p 0) We consider the quadratic response surface designs for y β 0 + β x + β 2 + β 3 + β + β 22 x2 2 + β 33 x2 3 + β 2 x + β 3 x + β 23 + ε with n responses and with partitioned into [ 2 with the four lower order terms (r 4) and 2 the six quadratic terms (s 6) Four popular designs are given in Appendi They are the hybrid designs (H30 and H3B) of Roquemore (976) Tables B2 and B3, the Box and Behnken (960) (BBD) design Table B4, and the CCD of Box and Wilson (95) Table B5 For each design, we compute the 0 0 canonical moment matrix It is striking that, for all these designs, the off-diagonal 4 6 array in R has only one non-zero singular value with its square the canonical index γ 2 It follows that GVIF( ) 2 γ 2 Table 2 reports that the design H30 is the most conditioned with respect to the GVIF with the least amount of collinearity between the lower and higher order terms 7 More complicated designs with ordered singular values Let be the minimal design of Box and Draper (974) BDD with n from Table B6, and let Z be the small composite design of Hartley (959) SCD with n from Table B7 for the quadratic response surface model (r 4 and s 6) as in Equation (5) Let α {α α r 0} and β {β β r 0} be the non-negative singular values of the off-diagonal array for R and R Z, respectively As α i β i ( i r) (Table 3), it follows that GVIF( 2 ) GVIF(Z 2 Z ) showing less collinearity between the lower and higher order terms for the BDD design Table 2 Hybrid designs H30, H3B, Box and Behnken BBD, and CCD Design n r γ 2 GVIF( 2 ) H30 4 0899 5553 H3B 4 0909 00 BBD 3 4 0923 300 CCD 5 4 09333 500 (5) Table 3 Singular values for off-diagonal array of R for BDD and SCD with GVIF BDD SCD 0907 09535 α 03424 β 06325 03424 06325 0877 06325 GVIF 74 5093 Page 5 of

8 An improved H30 design When the diagonal matrix Λ r s in Equation 6 in Appendix has only one non-zero entry, we have denoted the square of this value the canonical index We extend this definition to the case when ( ) 2 ( ) ( 2 2 2 2) has multiple positive singular values The Frobenious norm for a rectangular matrix A r s is defined by A 2 r s F i j a2 ij trace(a A) For a design matrix, we extend the definition of the canonical index with γ 2 Λ r s 2 F Alternatively, γ 2 trace(( ) 2 2 ( ) ( 2 ) ( 2 )) as in Equation 7 We examine, in detail, the H30 design matrix 0, Table B2 in Appendi, with our attention to the value of 0360 in row 2 for In succession, we will replace the values {736, 06386, 09273, 0000, 2906, 0360} by a free parameter and use γ 2 to determine an optimal value For example, replacing the four entries which are 736 with c, we calculate the minimum value for γ 2 0899 with c 768 denoted c min in Table 4 These values are within the four digit accuracy of the data We performed a similar calculation with c 2 using the four entries which are 06386; with c 3 with the four entries which are 09273; with c 4 with the eight entries which are ; and with c 5 with the single entry 2906 The original design has γ 2 0899 The entries in the H30 design are given to four significant digits With this precision, the original design is nearly optimal with respect to the canonical index γ 2 for the first five entries in Table 4 The sixth entry of c 6 0360 was not optimal with γ 2 088 with c min 00264, a magnitude value smaller Denote the improved H30 design as the H30 design with the value of c 6 00264 The improved H30 also has a unique positive singular value for the off-diagonal of R with its square the canonical index γ 2 All of the standard design criteria favor the improved H30 design over the H30 design, which was originally constructed based on the rotatability criterion to maintain equal variances for predicted responses for points that have the same distance from the design center As usual A() tr(( ) ), D() det(( ) ), and E() max{eigenvalues of ( ) } The small relative changes Δ in the design criteria are shown in Table 5 in Column 4 The abnormality of the second row in H30 has been noted in Jensen (998) who showed that the design is least sensitive to the second row of, the row containing the value c 6 0360 Table 4 Optimal values c min for γ 2 Design entries c min γ 2 c 06386-09273 2906 0360 768 0899 736 c 2-09273 2906 0360 06356 0899 736 06386 c 3 2906 0360 09303 0899 736 06386 09273 c 4 2906 0360 09975 0899 736 06386 09273 c 5 0360 2880 0899 736 06386 09273 2906 c 6 0026 088 Table 5 Design criteria for the improved H30 with Δ the relative change c 6 0360 c 6 00264 Δ A 2697 2688 033% D 0552 0 6 0542 0 6 064% E 086 0852 5% γ 2 0899 088 022% Page 6 of

9 Conclusions The VIF measure the penalty for adding a non-orthogonal variable to a linear regression The VIF can be computed as a ratio of determinant as in Equation A similar ratio criterion was studied by Fox and Monette (992) to measure the effect of adding a subset of new variables to a design and they dubbed it the generalized variance inflation factor GVIF, Equation 2 We have noted the relationship between GVIF and the singular values of the off-diagonal array in the canonical moment matrix and have used GFIV to study standard quadratic response designs The H30 design of Roquemorer (976) was shown not to be optimal with respect to GFIV and an improved H30 design was introduced which was favored over H30 using the standard design criteria A, D, and E Funding The authors received no direct funding for this research Author details Diarmuid O Driscoll E-mail: diarmuidodriscoll@miculie Donald E Ramirez 2 E-mail: der@virginiaedu Department of Mathematics and Computer Studies, Mary Immaculate College, Limerick, Ireland 2 Department of Mathematics, University of Virginia, Charlottesville, VA, 22904, USA Citation information Cite this article as: Response surface designs using the generalized variance inflation factors, Diarmuid O Driscoll & Donald E Ramirez, Cogent Mathematics (205), 2: 053728 References Belsley, D A (986) Centering, the constant, firstdifferencing, and assessing conditioning In E Kuh & D A Belsley (Eds), Model reliability (pp 7 53) Cambridge: MIT Press Berk, K (977) Tolerance and condition in regression computations Journal of the American Statistical Association, 72, 863 866 Box, G E P, & Behnken, D W (960) Some new threelevel designs for the study of quantitative variables Technometrics, 2, 455 475 Box, M J, & Draper, N R (974) On minimum-point second order design Technometrics, 6, 63 66 Appendix Box, G E P, & Wilson, K B (95) On the experimental attainment of optimum conditions Journal of the Royal Statistical Society, Series B, 3, 45 Eaton, M L (983) Multivariate statistics New York, NY: Wiley Fox, J, & Monette, G (992) Generalized collinearity diagnostics Journal of the American Statistical Association, 87, 78 83 Garcia, C B, Garcia, J, & Soto, J (20) The raise method: An alternative procedure to estimate the parameters in presence of collinearity Quality and Quantity, 45, 403 423 Hartley, H O (959) Smallest composite design for quadratic response surfaces Biometrics, 5, 6 624 Jensen, D R (998) Principal predictors and efficiency in small second-order designs Biometrical Journal, 40, 83 203 Jensen, D R, & Ramirez, D E (993) Efficiency comparisons in linear inference Journal of Statistical Planning and Inference, 37, 5 68 Liao, D, & Valliant, R (202) Variance inflation in the analysis of complex survey data Survey Methodology, 38, 53 62 Myers, R (990) Classical and modern regression with applications (2nd ed) Boston, MA: PWS-Kent O Driscoll, D, & Ramirez, D E (in press) Revisiting some design criteria (under review) Roquemorer, K G (976) Hybrid designs for quadratic response surfaces Technometrics, 8, 49 423 Stewart, G W (987) Collinearity and least squares regression Statistical Science, 2, 68 84 Thiel, H (97) Principles of econometrics New York, NY: Wiley Wichers, R (975) The detection of multicollinearity: A comment Review of Economics and Statistics, 57, 366 368 Willan, A R, & Watts, D G (978) Meaningful multicollinearity measures Technometrics, 20, 407 42 We study the eigenvalue ( structure of M (r,s) Let {λ λ 2 λ min(r,s) 0} be the non-negative singular values of ) 2 ( ) ( 2 2 2 2) As with the canonical correlation coefficients Eaton (983), write the off-diagonal rectangular array B r s of R as PΛQ with P and Q orthogonal matrices and Λ r s the rectangular diagonal matrix with the non-negative singular values down the diagonal Set [ Pr r 0 L r s 0 s r Q s s For notational convenience, we assume r s The matrix L is orthogonal and transforms R L RL into diagonal matrices: [ I r Λ s r Λ r s I s [ I r [SV r r 0 r (s r) [SV r r 0 r (s r) I s (A) Page 7 of

with Λ r s [SV r r 0 r (s r) where SV r r is the diagonal matrix of the non-negative singular values Since L is orthogonal, this transformation has not changed the eigenvalues To compute the determinant of R, convert the matrix in Equation 6 into an upper diagonal matrix by Gauss Elimination on Λ s r This changes r of the s on the diagonal in rows r + to r + r into λ 2 i, and thus det(r) min(r,s) ( λ 2 ) i i with GVIF( 2 ) min(r,s) i λ 2 i The singular values of R 2 ( ) 2 ( ) ( 2 2 2 2) are the non-negative square roots of the eigenvalues of Λ Λ denoted by eigvals(λ Λ)eigvals((Q R 2 P)(P R 2 Q))) eigvals( ( ) )( ) ( 2 ) 2 2 ( )) 2 If the trace of the inverse of the matrix in Equation 6 is required, then we note that [ I r Λ s r Λ r s I s with trace given by tr((l RL) ) r s + 2 min(r,s) i Appendi [ (I r Λ r s Λ s r ) Λ r s (I s Λ Λ s r r s ) Λ (I Λ s r r r s Λ s r ) (I s Λ Λ s r r s ) Table B The lower order matrix for the CCD with center run with a 2, n 9 and the lower order matrix for the factorial design with center run n 9 Central composite design Factorial design x x - - - - - - - - a 0 0 a 0-0 0 a 0 0 a 0-0 0 0 0 λ 2 i (A2) Page 8 of

Table B2 The lower order matrix for the hybrid (H30) design of Roquemore (976) with center run, n x 0 0 2906 0 0-0360 06386-06386 - 06386 - - 06386 736 0-09273 -736 0-09273 0 736-09273 0-736 -09273 0 0 0 Table B3 The lower order matrix for the hybrid (H3B) design of Roquemore (976) with center run, n x 0 0 24495 0 0-24495 2063 07507 2063-07507 - -2063-07507 -2063 07507-07507 2063-07507 -2063-07507 -2063 - -07507 2063 0 0 0 Table B4 The lower order matrix for the Box and Behnken (960) design (BBD) with center run, n 3 x 0-0 0 0 - - 0 - - 0-0 - 0-0 0-0 - 0 - - 0 0 0 Page 9 of

Table B5 The lower order matrix for the Box and Wilson (95) CCD for α 732 with center run, n 5 x - - - - - - 732 0 0-732 0 0 0 732 0 0-732 0 0 0 732 0 0-732 0 0 0 Table B6 The lower order matrix for the Box and Draper (974) minimal design (BDD) with center run, n x - - - - - - - - - 0925 0925-0925 - 0925-0925 0925-0292 -0292-0292 0 0 0 Table B7 The Lower order matrix for the small composite design of Hartley (959) (SCD) for α 732 with center run, n x - - - - - - - - - - - - 732 0 0-732 0 0 0 732 0 0-732 0 0 0 732 0 0-732 0 0 0 Page 0 of

205 The Author(s) This open access article is distributed under a Creative Commons Attribution (CC-BY) 40 license You are free to: Share copy and redistribute the material in any medium or format Adapt remix, transform, and build upon the material for any purpose, even commercially The licensor cannot revoke these freedoms as long as you follow the license terms Under the following terms: Attribution You must give appropriate credit, provide a link to the license, and indicate if changes were made You may do so in any reasonable manner, but not in any way that suggests the licensor endorses you or your use No additional restrictions You may not apply legal terms or technological measures that legally restrict others from doing anything the license permits Cogent Mathematics (ISSN: 233-835) is published by Cogent OA, part of Taylor & Francis Group Publishing with Cogent OA ensures: Immediate, universal access to your article on publication High visibility and discoverability via the Cogent OA website as well as Taylor & Francis Online Download and citation statistics for your article Rapid online publication Input from, and dialog with, expert editors and editorial boards Retention of full copyright of your article Guaranteed legacy preservation of your article Discounts and waivers for authors in developing regions Submit your manuscript to a Cogent OA journal at wwwcogentoacom Page of