7.3 Ridge Analysis of the Response Surface
|
|
- Berenice Boone
- 6 years ago
- Views:
Transcription
1 7.3 Ridge Analysis of the Response Surface When analyzing a fitted response surface, the researcher may find that the stationary point is outside of the experimental design region, but the researcher wants to locate the optimal response value (and its associated variable levels) within the boundaries of the region. Example: For a rising or falling ridge system, the researcher may want to calculate the absolute maximum or minimum value of the estimated response ŷ over the experimental region. The method of ridge analysis is a response surface method that enables the researcher to search for an optimal ŷ value in the experimental design region given that the canonical analysis does not provide a stationary point that is contained in the region or when the stationary point is a saddle point. Specifically, the method of ridge analysis is used to find the absolute maximum or minimum of ŷ on concentric spheres of varying radii, say R l (l = 1, 2,...) which are centered at the coded variable origin (x 1, x 2,..., x k ) = (0, 0,..., 0). For example, suppose there are two design variables, x 1 and x 2. In the following figure, three circles (l = 1, 2, 3) are drawn with different radii. The objective is to find the maximum value of ŷ on the circumference of each circle. Recall the fitted second-order response surface model in k variables is ŷ = b 0 + x b + x Bx Thus, our objective is to search for the coordinates of the variables x 1, x 2,..., x k on a sphere of radius R which maximize (or minimize) ŷ subject to the constraint k x 2 i = R 2 or equivalently x x R 2 = 0 i=1 We use the method of Lagrange multipliers to solve this problem. Let L = b 0 + x b + x Bx where µ is the Lagrange multiplier. Then we differentiate L with respect to the vector x and equate it to 0: This is equivalent to L x = Thus, for a fixed µ, a solution x to this equation is a constrained stationary point on the sphere of radius R = (x x) 1/2. 135
2 The problem is to choose µ which guarantees finding an absolute maximum or minimum ŷ in the experimental design region (and not a local optimum). Here are the following rules for selecting the value of µ: 1. To find an absolute maximum ŷ on a sphere of radius R, choose any value µ such that µ is 2. To find an absolute minimum ŷ on a sphere of radius R, choose any value µ such that µ is Warning: intermediate values of µ maximum or minimum. will direct us to a local and not an absolute Computer ridge analysis algorithms (such as RSREG in SAS) require the user to specify values of R (usually within the design region). The algorithm then converts these R values to µ values and then solves for x. The output of the ridge analysis is the set of coordinates of the maxima or minima along with the predicted response ŷ for each R. The results are displayed graphically by plotting the maximum and minimum ŷ against R and by plotting the optimum conditions against R. 136
3 137
4 If you wanted to minimize response y, then choose some µ < (which is the smallest eigenvalue), and repeat the process to find (x 1, x 2 ) ŷ, and R associated with that µ. Then ŷ is the absolute minimum predicted response value for all points that are a distance R from (0, 0) Two Ridge Analysis Examples Ridge Analysis for Example 1 Data: We will perform a ridge analysis of the data from Example 1. Based on the canonical analysis, the stationary point is a maximum and occurs at (x 1, x 2, x 3 ) = (.601,.155,.405) This point is R = (.601) 2 + (.155) 2 + (.405) 2 =.741. The minimum predicted response occurs on the boundary R = 3 at (x 1, x 2, x 3 ) = (.307,.338, 1.619) Both of these points can be observed on the ridge analysis plots. SAS Code For Ridge Analysis for Example 1 DM LOG; CLEAR; OUT; CLEAR; ; % ODS PRINTER PDF file= C:\COURSES\ST578\SAS\ridge1.PDF ; OPTIONS NODATE NONUMBER LS=78 PS=54; ODS LISTING; DATA DSGN; INPUT X1 X2 X3 LINES; ; PROC RSREG DATA=DSGN PLOTS=RIDGE; *PLOTS=(DIAGNOSTICS RIDGE SURFACE); MODEL STRENGTH = X1 X2 X3 / LACKFIT; RIDGE MAX MIN RADIUS=0 TO 1 BY.05; TITLE Ridge Analysis for Example 1 ; RUN; 138
5 RIDGE ANALYSIS USING RSREG Response Surface for Variable STRENGTH Response Mean Root MSE R-Square Coefficient of Variation Sum of Residual DF Squares Mean Square F Value Pr > F Lack of Fit Pure Error Total Error Parameter Estimate Standard from Coded Parameter DF Estimate Error t Value Pr > t Data Intercept < X X X X1*X X2*X X2*X X3*X X3*X X3*X Canonical Analysis of Response Surface Based on Coded Data Critical Value Factor Coded Uncoded X X X Predicted value at stationary point: Eigenvectors Eigenvalues X1 X2 X Stationary point is a maximum. 139
6 Estimated Ridge of Minimum Response for Variable STRENGTH Estimated Standard Uncoded Factor Values Coded Radius Response Error X1 X2 X > ^^^^^^^^ Absolute minimum Estimated Ridge of Maximum Response for Variable STRENGTH Estimated Standard Uncoded Factor Values Coded Radius Response Error X1 X2 X > ^^^^^^^^ Absolute maximum 140
7 Ridge Analysis for Example 1 The RSREG Procedure Ridge of Minimum STRENGTH Estimated STRENGTH Coded Factor Levels Radius Ridge X1 X2 X3 141
8 Ridge Analysis for Example 1 The RSREG Procedure Ridge of Maximum STRENGTH Estimated STRENGTH Coded Factor Levels Radius Ridge X1 X2 X3 142
9 Ridge Analysis for Example 3 Data: We will perform a ridge analysis of the data from Example 3. Based on the canonical analysis, the stationary point is a saddle point in the design space. Therefore, the minimum and maximum predicted values will occur on the boundary. The maximum predicted response occurs on the boundary R = 3 at (x 1, x 2, x 3 ) = (.544, 1.589,.089) The minimum predicted response occurs on the boundary R = 3 at (x 1, x 2, x 3 ) = (1.356,.248,.964) Both of these points can be observed on the ridge analysis plots. SAS Code for Ridge Analysis of Example 3 Data DM LOG; CLEAR; OUT; CLEAR; ; ODS LISTING; % ODS PRINTER PDF file= C:\COURSES\ST578\sas\ridge2.pdf ; OPTIONS NODATE NONUMBER PS=54 LS=78; TITLE Ridge Analysis for Example 3 (Saddle Point Example) ; DATA DSGN; INPUT X1 X2 X3 LINES; ; PROC RSREG DATA=DSGN PLOTS = RIDGE; MODEL YIELD = X1 X2 X3 / LACKFIT ; RIDGE MAX MIN RADIUS=0 TO 1 BY.05; RUN; SAS Output Ridge Analysis for Example 3 (Saddle Point Example) Coding Coefficients for the Independent Variables Factor Subtracted off Divided by X X X
10 Degrees of Type I Sum Regression Freedom of Squares R-Square F-Ratio Prob > F Linear Quadratic Crossproduct Total Regress Degrees of Sum of Residual Freedom Squares Mean Square F-Ratio Prob > F Lack of Fit Pure Error Total Error Degrees of Parameter Standard T for H0: Parameter Freedom Estimate Error Parameter=0 Prob > T INTERCEPT X X X X1*X X2*X X2*X X3*X X3*X X3*X Canonical Analysis of Response Surface (based on coded data) Critical Value Factor Coded Uncoded X X X Predicted value at stationary point Eigenvectors Eigenvalues X1 X2 X Stationary point is a saddle point. Estimated Ridge of Minimum Response for Variable YIELD 144
11 Estimated Standard Uncoded Factor Values Coded Radius Response Error X1 X2 X Estimated Ridge of Maximum Response for Variable YIELD Estimated Standard Uncoded Factor Values Coded Radius Response Error X1 X2 X
12 Ridge Analysis for Example 3 (Saddle Point Example) The RSREG Procedure 13.0 Ridge of Maximum YIELD Estimated YIELD Coded Factor Levels Radius Ridge X1 X2 X3 146
13 Ridge Analysis for Example 3 (Saddle Point Example) The RSREG Procedure Ridge of Minimum YIELD Estimated YIELD Coded Factor Levels Radius Ridge X1 X2 X3 147
7 The Analysis of Response Surfaces
7 The Analysis of Response Surfaces Goal: the researcher is seeking the experimental conditions which are most desirable, that is, determine optimum design variable levels. Once the researcher has determined
More informationResponse Surface Methodology
Response Surface Methodology Bruce A Craig Department of Statistics Purdue University STAT 514 Topic 27 1 Response Surface Methodology Interested in response y in relation to numeric factors x Relationship
More informationPerforming response surface analysis using the SAS RSREG procedure
Paper DV02-2012 Performing response surface analysis using the SAS RSREG procedure Zhiwu Li, National Database Nursing Quality Indicator and the Department of Biostatistics, University of Kansas Medical
More informationEXST Regression Techniques Page 1. We can also test the hypothesis H :" œ 0 versus H :"
EXST704 - Regression Techniques Page 1 Using F tests instead of t-tests We can also test the hypothesis H :" œ 0 versus H :" Á 0 with an F test.! " " " F œ MSRegression MSError This test is mathematically
More information4.8 Alternate Analysis as a Oneway ANOVA
4.8 Alternate Analysis as a Oneway ANOVA Suppose we have data from a two-factor factorial design. The following method can be used to perform a multiple comparison test to compare treatment means as well
More information1) Answer the following questions as true (T) or false (F) by circling the appropriate letter.
1) Answer the following questions as true (T) or false (F) by circling the appropriate letter. T F T F T F a) Variance estimates should always be positive, but covariance estimates can be either positive
More information5.3 Three-Stage Nested Design Example
5.3 Three-Stage Nested Design Example A researcher designs an experiment to study the of a metal alloy. A three-stage nested design was conducted that included Two alloy chemistry compositions. Three ovens
More informationResponse Surface Methodology III
LECTURE 7 Response Surface Methodology III 1. Canonical Form of Response Surface Models To examine the estimated regression model we have several choices. First, we could plot response contours. Remember
More informationLecture notes on Regression & SAS example demonstration
Regression & Correlation (p. 215) When two variables are measured on a single experimental unit, the resulting data are called bivariate data. You can describe each variable individually, and you can also
More information7.4 The c-chart (fixed sample size)
7.4 The c-chart (fixed sample size) Often there is greater concern for the total number of nonconformities (or defects) (C) than the fraction of nonconforming (defective) units (p). An inspection unit
More informationLecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is
Lecture 15 Multiple regression I Chapter 6 Set 2 Least Square Estimation The quadratic form to be minimized is Q = (Y i β 0 β 1 X i1 β 2 X i2 β p 1 X i.p 1 ) 2, which in matrix notation is Q = (Y Xβ) (Y
More information39.1 Absolute maxima/minima
Module 13 : Maxima, Minima Saddle Points, Constrained maxima minima Lecture 39 : Absolute maxima / minima [Section 39.1] Objectives In this section you will learn the following : The notion of absolute
More informationChapter 8 (More on Assumptions for the Simple Linear Regression)
EXST3201 Chapter 8b Geaghan Fall 2005: Page 1 Chapter 8 (More on Assumptions for the Simple Linear Regression) Your textbook considers the following assumptions: Linearity This is not something I usually
More informationCHAPTER II RESPONSE SURFACE METHODOLOGY
CHAPTER II RESPOSE SURFACE METHODOLOGY.1. Introduction Response surface methodology is useful for factorial eperiments with v quantitative factors that are under taken so as to determine the combination
More information7. Response Surface Methodology (Ch.10. Regression Modeling Ch. 11. Response Surface Methodology)
7. Response Surface Methodology (Ch.10. Regression Modeling Ch. 11. Response Surface Methodology) Hae-Jin Choi School of Mechanical Engineering, Chung-Ang University 1 Introduction Response surface methodology,
More informationSTOR 455 STATISTICAL METHODS I
STOR 455 STATISTICAL METHODS I Jan Hannig Mul9variate Regression Y=X β + ε X is a regression matrix, β is a vector of parameters and ε are independent N(0,σ) Es9mated parameters b=(x X) - 1 X Y Predicted
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationEXST7015: Estimating tree weights from other morphometric variables Raw data print
Simple Linear Regression SAS example Page 1 1 ********************************************; 2 *** Data from Freund & Wilson (1993) ***; 3 *** TABLE 8.24 : ESTIMATING TREE WEIGHTS ***; 4 ********************************************;
More informationMulticollinearity Exercise
Multicollinearity Exercise Use the attached SAS output to answer the questions. [OPTIONAL: Copy the SAS program below into the SAS editor window and run it.] You do not need to submit any output, so there
More informationVersion 1: Equality of Distributions. 3. F (x) and G(x) represent the distribution functions corresponding to the Xs and Y s, respectively.
4 Two-Sample Methods 4.1 The (Mann-Whitney) Wilcoxon Rank Sum Test Version 1: Equality of Distributions Assumptions: Given two independent random samples X 1, X 2,..., X n and Y 1, Y 2,..., Y m : 1. The
More informationCh 5.7: Series Solutions Near a Regular Singular Point, Part II
Ch 5.7: Series Solutions Near a Regular Singular Point, Part II! Recall from Section 5.6 (Part I): The point x 0 = 0 is a regular singular point of with and corresponding Euler Equation! We assume solutions
More information6 Designs with Split Plots
6 Designs with Split Plots Many factorial experimental designs are incorrectly analyzed because the assumption of complete randomization is not true. Many factorial experiments have one or more restrictions
More informationUnit 12: Response Surface Methodology and Optimality Criteria
Unit 12: Response Surface Methodology and Optimality Criteria STA 643: Advanced Experimental Design Derek S. Young 1 Learning Objectives Revisit your knowledge of polynomial regression Know how to use
More informationFailure Time of System due to the Hot Electron Effect
of System due to the Hot Electron Effect 1 * exresist; 2 option ls=120 ps=75 nocenter nodate; 3 title of System due to the Hot Electron Effect ; 4 * TIME = failure time (hours) of a system due to drift
More informationIES 612/STA 4-573/STA Winter 2008 Week 1--IES 612-STA STA doc
IES 612/STA 4-573/STA 4-576 Winter 2008 Week 1--IES 612-STA 4-573-STA 4-576.doc Review Notes: [OL] = Ott & Longnecker Statistical Methods and Data Analysis, 5 th edition. [Handouts based on notes prepared
More informationPerformance Surfaces and Optimum Points
CSC 302 1.5 Neural Networks Performance Surfaces and Optimum Points 1 Entrance Performance learning is another important class of learning law. Network parameters are adjusted to optimize the performance
More informationStatistical Techniques II EXST7015 Simple Linear Regression
Statistical Techniques II EXST7015 Simple Linear Regression 03a_SLR 1 Y - the dependent variable 35 30 25 The objective Given points plotted on two coordinates, Y and X, find the best line to fit the data.
More informationDesign and Analysis of Experiments
Design and Analysis of Experiments Part IX: Response Surface Methodology Prof. Dr. Anselmo E de Oliveira anselmo.quimica.ufg.br anselmo.disciplinas@gmail.com Methods Math Statistics Models/Analyses Response
More informationAnalysis of Variance. Source DF Squares Square F Value Pr > F. Model <.0001 Error Corrected Total
Math 221: Linear Regression and Prediction Intervals S. K. Hyde Chapter 23 (Moore, 5th Ed.) (Neter, Kutner, Nachsheim, and Wasserman) The Toluca Company manufactures refrigeration equipment as well as
More informationThe use of response surface analysis in obtaining maximum profit in oil palm industry
ORIGINAL ARTICLE The use of response surface analysis in obtaining maximum profit in oil palm industry Azme Khamis 1, Zuhaimy Ismail 2, Khalid Haron 3 and Ahmad Tarmizi Mohammed 4 Abstract Khamis, A.,
More informationJ. Response Surface Methodology
J. Response Surface Methodology Response Surface Methodology. Chemical Example (Box, Hunter & Hunter) Objective: Find settings of R, Reaction Time and T, Temperature that produced maximum yield subject
More informationGradient Descent. Dr. Xiaowei Huang
Gradient Descent Dr. Xiaowei Huang https://cgi.csc.liv.ac.uk/~xiaowei/ Up to now, Three machine learning algorithms: decision tree learning k-nn linear regression only optimization objectives are discussed,
More information15.S24 Sample Exam Solutions
5.S4 Sample Exam Solutions. In each case, determine whether V is a vector space. If it is not a vector space, explain why not. If it is, find basis vectors for V. (a) V is the subset of R 3 defined by
More informationExst7037 Multivariate Analysis Cancorr interpretation Page 1
Exst7037 Multivariate Analysis Cancorr interpretation Page 1 1 *** C03S3D1 ***; 2 ****************************************************************************; 3 *** The data set insulin shows data from
More informationLecture 3: Inference in SLR
Lecture 3: Inference in SLR STAT 51 Spring 011 Background Reading KNNL:.1.6 3-1 Topic Overview This topic will cover: Review of hypothesis testing Inference about 1 Inference about 0 Confidence Intervals
More informationOptimality Conditions
Chapter 2 Optimality Conditions 2.1 Global and Local Minima for Unconstrained Problems When a minimization problem does not have any constraints, the problem is to find the minimum of the objective function.
More information17. Example SAS Commands for Analysis of a Classic Split-Plot Experiment 17. 1
17 Example SAS Commands for Analysis of a Classic SplitPlot Experiment 17 1 DELIMITED options nocenter nonumber nodate ls80; Format SCREEN OUTPUT proc import datafile"c:\data\simulatedsplitplotdatatxt"
More informationEcon Slides from Lecture 14
Econ 205 Sobel Econ 205 - Slides from Lecture 14 Joel Sobel September 10, 2010 Theorem ( Lagrange Multipliers ) Theorem If x solves max f (x) subject to G(x) = 0 then there exists λ such that Df (x ) =
More informationOdor attraction CRD Page 1
Odor attraction CRD Page 1 dm'log;clear;output;clear'; options ps=512 ls=99 nocenter nodate nonumber nolabel FORMCHAR=" ---- + ---+= -/\*"; ODS LISTING; *** Table 23.2 ********************************************;
More informationThis is a Randomized Block Design (RBD) with a single factor treatment arrangement (2 levels) which are fixed.
EXST3201 Chapter 13c Geaghan Fall 2005: Page 1 Linear Models Y ij = µ + βi + τ j + βτij + εijk This is a Randomized Block Design (RBD) with a single factor treatment arrangement (2 levels) which are fixed.
More informationSOLUTION OF QUADRATIC EQUATIONS LESSON PLAN. A3 Topic Overview ALGEBRA
ALGEBRA A Topic Overview A SOLUTION OF QUADRATIC EQUATIONS This topic describes three methods of solving Quadratic equations. assumes you understand and have practised using the algebraic methods described
More informationTranspose & Dot Product
Transpose & Dot Product Def: The transpose of an m n matrix A is the n m matrix A T whose columns are the rows of A. So: The columns of A T are the rows of A. The rows of A T are the columns of A. Example:
More informationSTATISTICS 479 Exam II (100 points)
Name STATISTICS 79 Exam II (1 points) 1. A SAS data set was created using the following input statement: Answer parts(a) to (e) below. input State $ City $ Pop199 Income Housing Electric; (a) () Give the
More informationLecture 5: Comparing Treatment Means Montgomery: Section 3-5
Lecture 5: Comparing Treatment Means Montgomery: Section 3-5 Page 1 Linear Combination of Means ANOVA: y ij = µ + τ i + ɛ ij = µ i + ɛ ij Linear combination: L = c 1 µ 1 + c 1 µ 2 +...+ c a µ a = a i=1
More informationLecture 4 September 15
IFT 6269: Probabilistic Graphical Models Fall 2017 Lecture 4 September 15 Lecturer: Simon Lacoste-Julien Scribe: Philippe Brouillard & Tristan Deleu 4.1 Maximum Likelihood principle Given a parametric
More informationAdaptive Filtering. Squares. Alexander D. Poularikas. Fundamentals of. Least Mean. with MATLABR. University of Alabama, Huntsville, AL.
Adaptive Filtering Fundamentals of Least Mean Squares with MATLABR Alexander D. Poularikas University of Alabama, Huntsville, AL CRC Press Taylor & Francis Croup Boca Raton London New York CRC Press is
More informationDESIGN OF EXPERIMENT ERT 427 Response Surface Methodology (RSM) Miss Hanna Ilyani Zulhaimi
+ DESIGN OF EXPERIMENT ERT 427 Response Surface Methodology (RSM) Miss Hanna Ilyani Zulhaimi + Outline n Definition of Response Surface Methodology n Method of Steepest Ascent n Second-Order Response Surface
More informationBayes Estimators & Ridge Regression
Readings Chapter 14 Christensen Merlise Clyde September 29, 2015 How Good are Estimators? Quadratic loss for estimating β using estimator a L(β, a) = (β a) T (β a) How Good are Estimators? Quadratic loss
More informationTranspose & Dot Product
Transpose & Dot Product Def: The transpose of an m n matrix A is the n m matrix A T whose columns are the rows of A. So: The columns of A T are the rows of A. The rows of A T are the columns of A. Example:
More informationLecture 1 Linear Regression with One Predictor Variable.p2
Lecture Linear Regression with One Predictor Variablep - Basics - Meaning of regression parameters p - β - the slope of the regression line -it indicates the change in mean of the probability distn of
More informationSTAT Checking Model Assumptions
STAT 704 --- Checking Model Assumptions Recall we assumed the following in our model: (1) The regression relationship between the response and the predictor(s) specified in the model is appropriate (2)
More informationLecture 6 Positive Definite Matrices
Linear Algebra Lecture 6 Positive Definite Matrices Prof. Chun-Hung Liu Dept. of Electrical and Computer Engineering National Chiao Tung University Spring 2017 2017/6/8 Lecture 6: Positive Definite Matrices
More informationBSM510 Numerical Analysis
BSM510 Numerical Analysis Polynomial Interpolation Prof. Manar Mohaisen Department of EEC Engineering Review of Precedent Lecture Polynomial Regression Multiple Linear Regression Nonlinear Regression Lecture
More informationUC Berkeley Department of Electrical Engineering and Computer Science. EECS 227A Nonlinear and Convex Optimization. Solutions 5 Fall 2009
UC Berkeley Department of Electrical Engineering and Computer Science EECS 227A Nonlinear and Convex Optimization Solutions 5 Fall 2009 Reading: Boyd and Vandenberghe, Chapter 5 Solution 5.1 Note that
More informationWritten Examination
Division of Scientific Computing Department of Information Technology Uppsala University Optimization Written Examination 202-2-20 Time: 4:00-9:00 Allowed Tools: Pocket Calculator, one A4 paper with notes
More informationTaylor Series and stationary points
Chapter 5 Taylor Series and stationary points 5.1 Taylor Series The surface z = f(x, y) and its derivatives can give a series approximation for f(x, y) about some point (x 0, y 0 ) as illustrated in Figure
More informationSimple Linear Regression
Simple Linear Regression September 24, 2008 Reading HH 8, GIll 4 Simple Linear Regression p.1/20 Problem Data: Observe pairs (Y i,x i ),i = 1,...n Response or dependent variable Y Predictor or independent
More informationKing s Year 12 Medium Term Plan for LC1- A-Level Mathematics
King s Year 12 Medium Term Plan for LC1- A-Level Mathematics Modules Algebra, Geometry and Calculus. Materials Text book: Mathematics for A-Level Hodder Education. needed Calculator. Progress objectives
More informationEffect of Centering and Standardization in Moderation Analysis
Effect of Centering and Standardization in Moderation Analysis Raw Data The CORR Procedure 3 Variables: govact negemot Simple Statistics Variable N Mean Std Dev Sum Minimum Maximum Label govact 4.58699
More informationdm'log;clear;output;clear'; options ps=512 ls=99 nocenter nodate nonumber nolabel FORMCHAR=" = -/\<>*"; ODS LISTING;
dm'log;clear;output;clear'; options ps=512 ls=99 nocenter nodate nonumber nolabel FORMCHAR=" ---- + ---+= -/\*"; ODS LISTING; *** Table 23.2 ********************************************; *** Moore, David
More informationSeptember Math Course: First Order Derivative
September Math Course: First Order Derivative Arina Nikandrova Functions Function y = f (x), where x is either be a scalar or a vector of several variables (x,..., x n ), can be thought of as a rule which
More information5 Handling Constraints
5 Handling Constraints Engineering design optimization problems are very rarely unconstrained. Moreover, the constraints that appear in these problems are typically nonlinear. This motivates our interest
More informationChapter 8 Quantitative and Qualitative Predictors
STAT 525 FALL 2017 Chapter 8 Quantitative and Qualitative Predictors Professor Dabao Zhang Polynomial Regression Multiple regression using X 2 i, X3 i, etc as additional predictors Generates quadratic,
More information14.7: Maxima and Minima
14.7: Maxima and Minima Marius Ionescu October 29, 2012 Marius Ionescu () 14.7: Maxima and Minima October 29, 2012 1 / 13 Local Maximum and Local Minimum Denition Marius Ionescu () 14.7: Maxima and Minima
More informationSAS Example 3: Deliberately create numerical problems
SAS Example 3: Deliberately create numerical problems Four experiments 1. Try to fit this model, failing the parameter count rule. 2. Set φ 12 =0 to pass the parameter count rule, but still not identifiable.
More informationDeep Learning. Authors: I. Goodfellow, Y. Bengio, A. Courville. Chapter 4: Numerical Computation. Lecture slides edited by C. Yim. C.
Chapter 4: Numerical Computation Deep Learning Authors: I. Goodfellow, Y. Bengio, A. Courville Lecture slides edited by 1 Chapter 4: Numerical Computation 4.1 Overflow and Underflow 4.2 Poor Conditioning
More informationMath 3330: Solution to midterm Exam
Math 3330: Solution to midterm Exam Question 1: (14 marks) Suppose the regression model is y i = β 0 + β 1 x i + ε i, i = 1,, n, where ε i are iid Normal distribution N(0, σ 2 ). a. (2 marks) Compute the
More informationSimple linear regression
Simple linear regression Biometry 755 Spring 2008 Simple linear regression p. 1/40 Overview of regression analysis Evaluate relationship between one or more independent variables (X 1,...,X k ) and a single
More informationLogistic Regression Analysis
Logistic Regression Analysis Predicting whether an event will or will not occur, as well as identifying the variables useful in making the prediction, is important in most academic disciplines as well
More informationDensity Temp vs Ratio. temp
Temp Ratio Density 0.00 0.02 0.04 0.06 0.08 0.10 0.12 Density 0.0 0.2 0.4 0.6 0.8 1.0 1. (a) 170 175 180 185 temp 1.0 1.5 2.0 2.5 3.0 ratio The histogram shows that the temperature measures have two peaks,
More informationST 512-Practice Exam I - Osborne Directions: Answer questions as directed. For true/false questions, circle either true or false.
ST 512-Practice Exam I - Osborne Directions: Answer questions as directed. For true/false questions, circle either true or false. 1. A study was carried out to examine the relationship between the number
More informationTABLES AND FORMULAS FOR MOORE Basic Practice of Statistics
TABLES AND FORMULAS FOR MOORE Basic Practice of Statistics Exploring Data: Distributions Look for overall pattern (shape, center, spread) and deviations (outliers). Mean (use a calculator): x = x 1 + x
More informationLecture 18: Optimization Programming
Fall, 2016 Outline Unconstrained Optimization 1 Unconstrained Optimization 2 Equality-constrained Optimization Inequality-constrained Optimization Mixture-constrained Optimization 3 Quadratic Programming
More informationLecture 2 Simple Linear Regression STAT 512 Spring 2011 Background Reading KNNL: Chapter 1
Lecture Simple Linear Regression STAT 51 Spring 011 Background Reading KNNL: Chapter 1-1 Topic Overview This topic we will cover: Regression Terminology Simple Linear Regression with a single predictor
More informationa = 4 levels of treatment A = Poison b = 3 levels of treatment B = Pretreatment n = 4 replicates for each treatment combination
In Box, Hunter, and Hunter Statistics for Experimenters is a two factor example of dying times for animals, let's say cockroaches, using 4 poisons and pretreatments with n=4 values for each combination
More informationExtreme Values and Positive/ Negative Definite Matrix Conditions
Extreme Values and Positive/ Negative Definite Matrix Conditions James K. Peterson Department of Biological Sciences and Department of Mathematical Sciences Clemson University November 8, 016 Outline 1
More informationLinear regression. We have that the estimated mean in linear regression is. ˆµ Y X=x = ˆβ 0 + ˆβ 1 x. The standard error of ˆµ Y X=x is.
Linear regression We have that the estimated mean in linear regression is The standard error of ˆµ Y X=x is where x = 1 n s.e.(ˆµ Y X=x ) = σ ˆµ Y X=x = ˆβ 0 + ˆβ 1 x. 1 n + (x x)2 i (x i x) 2 i x i. The
More informationAddition and subtraction: element by element, and dimensions must match.
Matrix Essentials review: ) Matrix: Rectangular array of numbers. ) ranspose: Rows become columns and vice-versa ) single row or column is called a row or column) Vector ) R ddition and subtraction: element
More informationCHAPTER 2: QUADRATIC PROGRAMMING
CHAPTER 2: QUADRATIC PROGRAMMING Overview Quadratic programming (QP) problems are characterized by objective functions that are quadratic in the design variables, and linear constraints. In this sense,
More informationPolynomial functions right- and left-hand behavior (end behavior):
Lesson 2.2 Polynomial Functions For each function: a.) Graph the function on your calculator Find an appropriate window. Draw a sketch of the graph on your paper and indicate your window. b.) Identify
More informationThis module focuses on the logic of ANOVA with special attention given to variance components and the relationship between ANOVA and regression.
WISE ANOVA and Regression Lab Introduction to the WISE Correlation/Regression and ANOVA Applet This module focuses on the logic of ANOVA with special attention given to variance components and the relationship
More informationAn equation is a statement that states that two expressions are equal. For example:
Section 0.1: Linear Equations Solving linear equation in one variable: An equation is a statement that states that two expressions are equal. For example: (1) 513 (2) 16 (3) 4252 (4) 64153 To solve the
More informationMath 5311 Constrained Optimization Notes
ath 5311 Constrained Optimization otes February 5, 2009 1 Equality-constrained optimization Real-world optimization problems frequently have constraints on their variables. Constraints may be equality
More informationCircle the single best answer for each multiple choice question. Your choice should be made clearly.
TEST #1 STA 4853 March 6, 2017 Name: Please read the following directions. DO NOT TURN THE PAGE UNTIL INSTRUCTED TO DO SO Directions This exam is closed book and closed notes. There are 32 multiple choice
More informationIntegrated Math II Performance Level Descriptors
Limited Integrated Math II Performance Level Descriptors A student performing at the Limited Level demonstrates a minimal command of Ohio s Learning Standards for Integrated Math II. A student at this
More informationTable 1: Fish Biomass data set on 26 streams
Math 221: Multiple Regression S. K. Hyde Chapter 27 (Moore, 5th Ed.) The following data set contains observations on the fish biomass of 26 streams. The potential regressors from which we wish to explain
More informationWillmar Public Schools Curriculum Map
Subject Area Mathematics Senior High Course Name Advanced Algebra 2A (Prentice Hall Mathematics) Date April 2010 The Advanced Algebra 2A course parallels each other in content and time. The Advanced Algebra
More information6 Single Sample Methods for a Location Parameter
6 Single Sample Methods for a Location Parameter If there are serious departures from parametric test assumptions (e.g., normality or symmetry), nonparametric tests on a measure of central tendency (usually
More informationMath Maximum and Minimum Values, I
Math 213 - Maximum and Minimum Values, I Peter A. Perry University of Kentucky October 8, 218 Homework Re-read section 14.7, pp. 959 965; read carefully pp. 965 967 Begin homework on section 14.7, problems
More informationSOL Warm-Up Graphing Calculator Active
A.2c (a) Factoring polynomials SOL Warm-Up 1. Which of the following represents 12x 2 + 6x + 3 in simplified form after factoring out the greatest common factor? A 12(x 2 + 2x + 4) B x(12x 2 + 6x + 3)
More informationCorrelation and the Analysis of Variance Approach to Simple Linear Regression
Correlation and the Analysis of Variance Approach to Simple Linear Regression Biometry 755 Spring 2009 Correlation and the Analysis of Variance Approach to Simple Linear Regression p. 1/35 Correlation
More informationAMS 7 Correlation and Regression Lecture 8
AMS 7 Correlation and Regression Lecture 8 Department of Applied Mathematics and Statistics, University of California, Santa Cruz Suumer 2014 1 / 18 Correlation pairs of continuous observations. Correlation
More informationSTAT5044: Regression and Anova
STAT5044: Regression and Anova Inyoung Kim 1 / 25 Outline 1 Multiple Linear Regression 2 / 25 Basic Idea An extra sum of squares: the marginal reduction in the error sum of squares when one or several
More informationMIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010
MIXED MODELS FOR REPEATED (LONGITUDINAL) DATA PART 2 DAVID C. HOWELL 4/1/2010 Part 1 of this document can be found at http://www.uvm.edu/~dhowell/methods/supplements/mixed Models for Repeated Measures1.pdf
More informationAlgebra II Chapter 5
Algebra II Chapter 5 5.1 Quadratic Functions The graph of a quadratic function is a parabola, as shown at rig. Standard Form: f ( x) = ax2 + bx + c vertex: (x, y) = b 2a, f b 2a a < 0 graph opens down
More informationStatistical Modelling in Stata 5: Linear Models
Statistical Modelling in Stata 5: Linear Models Mark Lunt Arthritis Research UK Epidemiology Unit University of Manchester 07/11/2017 Structure This Week What is a linear model? How good is my model? Does
More informationGeneral Linear Model (Chapter 4)
General Linear Model (Chapter 4) Outcome variable is considered continuous Simple linear regression Scatterplots OLS is BLUE under basic assumptions MSE estimates residual variance testing regression coefficients
More informationInterpreting coefficients for transformed variables
Interpreting coefficients for transformed variables! Recall that when both independent and dependent variables are untransformed, an estimated coefficient represents the change in the dependent variable
More informationEXST 7015 Fall 2014 Lab 08: Polynomial Regression
EXST 7015 Fall 2014 Lab 08: Polynomial Regression OBJECTIVES Polynomial regression is a statistical modeling technique to fit the curvilinear data that either shows a maximum or a minimum in the curve,
More informationStatistics 203: Introduction to Regression and Analysis of Variance Penalized models
Statistics 203: Introduction to Regression and Analysis of Variance Penalized models Jonathan Taylor - p. 1/15 Today s class Bias-Variance tradeoff. Penalized regression. Cross-validation. - p. 2/15 Bias-variance
More information