A new algorithm for deriving optimal designs

Size: px
Start display at page:

Download "A new algorithm for deriving optimal designs"

Transcription

1 A new algorithm for deriving optimal designs Stefanie Biedermann, University of Southampton, UK Joint work with Min Yang, University of Illinois at Chicago 18 October 2012, DAE, University of Georgia, Athens, GA 1 / 45

2 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 2 / 45

3 Statistics and Optimal design of experiments The main goal of statistics is to gain information from collected data Design of experiments concerns the way of data collection An optimal/efficient design can reduce the sample size and cost needed for achieving a specified precision of estimation, or can improve the precision for a given sample size 3 / 45

4 Exact and approximate designs Exact design ξ n for sample size n: ξ n = { x1 x 2... x m n 1 /n n 2 /n... n m /n } ; n i integers, m n i = n i=1 Approximate design ξ: ξ = { } x1 x 2... x m ; ω ω 1 ω 2... ω i > 0, m m ω i = 1 Note: To run an approximate design, some rounding procedure is required i=1 4 / 45

5 Optimality criteria Suppose interest is in F(θ), a vector of functions of the model parameters The (asymptotic) covariance matrix of the MLE of F(θ) under design ξ, Σ ξ (F), can be written as Σ ξ (F) = F (θ) θ T (I ξ) F(θ) θ where I ξ is the information matrix of the design ξ Aim: Obtain precise estimates Minimise Φ(Σ ξ (F )), an optimality criterion 5 / 45

6 Optimality criteria Example 1: A D-optimal design minimises the determinant of the covariance matrix minimises the volume of a confidence ellipsoid for F(θ) Example 2: An A-optimal design minimises the trace of the covariance matrix minimises the sum of the variances of the elements of F(ˆθ) 6 / 45

7 Locally optimal designs A large and popular class of optimality criteria are the Φ p -criteria, of which A- and D-optimality are special cases: Φ p (Σ ξ (F)) = where 0 p < and Σ ξ (F ) IR v v [ ] 1/p 1 v trace(σ ξ(f )) p Problem: In many situations, Φ p -optimal designs depend on the unknown parameter vector θ locally optimal designs 7 / 45

8 Optimal multi-stage designs Use a sequential strategy: Suppose n 0 observations are already available, from design ξ n0, and another n observations can be made Estimate θ ˆθ 0 Design problem: Find ξ such that the combined design ξ = n 0 /(n 0 + n)ξ n0 + n/(n 0 + n)ξ minimises Φ(Σ ξ (F )) for ˆθ 0 Note: ξ and thus ξ are approximate designs 8 / 45

9 Aim of this project Provide a fast, general, convenient, and easy to use approach to find locally and multi-stage optimal designs 9 / 45

10 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 10 / 45

11 Current algorithms Three well-known algorithms are: Vertex direction method (VDM; Fedorov, 1972) Multiplicative algorithm (MA; Silvey, Titterington and Torsney, 1978) Vertex exchange method (VEM; Böhning, 1986) Yu (2011) combines the best features of these algorithms with a nearest neighbour exchange step to improve speed Cocktail algorithm 11 / 45

12 Computing time comparison I Model: Y = θ 1 e θ 2x + θ 3 e θ 4x + ε Design space [0, 3] discretised to X = {3i/N, i = 1,..., N} Computing time (in seconds) (Table 1 of Yu, 2011) N = 20 N = 50 N = 100 N = 200 N = 500 MA VEM Cocktail (VDM is omitted since it is known to be slower than VEM) 12 / 45

13 Computing time comparison II Y = θ 1 + θ 2 r + θ 3 r 2 + θ 4 s + θ 5 rs + ε Design space [ 1, 1] [0, 1] discretised to X = {(r i, s j ), r i = 2i/k 1, i = 1,..., k, s j = j/k, j = 1,..., k}, N = k 2 Computing time (in seconds) (Table 4 of Yu, 2011) N = 20 2 N = 50 2 N = N = MA aborted aborted VEM aborted Cocktail / 45

14 Summary of current algorithms All algorithms are relatively easy to implement The Cocktail algorithm leads to dramatically improved speed, sometimes by several orders of magnitude, relative to MA, VDM, and VEM. The Cocktail algorithm is restricted to D-optimal design for the full parameter vector Restricted to (one-stage) locally optimal designs 14 / 45

15 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 15 / 45

16 Goal of the new algorithm Any parameters (or functions thereof) can be of interest Any optimality criterion: Φ p -optimality (includes D-, A-, and E-optimality) One-stage and multi-stage Convenient to use Even faster 16 / 45

17 Rationale 1. The existing algorithms typically seek monotonic convergence May not be the fastest way to converge to the optimal weights 2. New idea: instead of seeking monotonic convergence, we derive the optimal weights directly, which can be done by solving a system of nonlinear equations 17 / 45

18 Notations for the new algorithm X = {x 1,..., x N }, the set of all possible design points S (t), the support after tth iteration ξ, design with support S (t) and optimal weights S (t) d(x, ξ ), sensitivity function of design ξ S (t) S (t) 18 / 45

19 The new algorithm Start with an initial design support, S (0). In the tth step: Derive ξ using the Newton-Raphson method S (t) Compute the sensitivity function d(x i, ξ ) for i = 1,..., N S (t) Select 1 i max N such that d(x imax, ξ S (t) ) = max i n N d(x i, ξ S (t) ) Stop when d(x imax, ξ S (t) ) ɛ, ɛ > 0 very small, otherwise S (t+1) = S (t) {x imax } 19 / 45

20 Convergence (I) Theorem 1: For a multi-stage design, let I ξ0 be non-singular. Then for any initial support S (0), the sequence of designs {ξ S (t) ; t 0} converges to an optimal design minimising Φ p (Σ ξ (F)) as t, provided the optimal weights are found in each updating step For a one-stage design, the above result holds if I ξs (0) and if the Jacobian of F is square and of full rank is non-singular, 20 / 45

21 Convergence (II) Theorem 2: For a multi-stage design, let p be an integer and I ξ0 be non-singular. For given support, Φ p (Σ ξ (F)) is minimised (w.r.t the weights) at any critical point in Ω = {w = (w 1,..., w m 1 ) T : w i 0, w w m 1 1} or its boundary. In addition, the Hessian of Φ p (Σ ξ (F )) is non-negative definite. Newton-Raphson method will converge provided the starting point is close enough to the critical point. Since Ω is compact, we can always find a critical point (given it exists) by using sufficiently many starting values 21 / 45

22 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 22 / 45

23 Example I Set up Consider model Y = θ 1 e θ 2x + θ 3 e θ 4x + ε. Let θ = (θ 1, θ 2, θ 3, θ 4 ) and X = {3i/N, i = 1,..., N}. Assume that θ = (1, 1, 1, 2). The same set up as that of Table 1 of Yu (2011) Algorithms are coded using SAS IML and computed at a Dell laptop (2.2GHz and 8Gb Ram) 23 / 45

24 Example I Computing time (in seconds) comparison I Computing time (in seconds) for the D-optimal design for the full parameter vector θ N = 500 N = 1000 N = 5000 N = Cocktail New algorithm / 45

25 Example I Locally A- and D-optimal designs Computing time (in seconds) under the new algorithm D A θ (θ 1, θ 3 ) (θ 2, θ 4 ) θ (θ 1, θ 3 ) (θ 2, θ 4 ) N = N = N = N = / 45

26 Example I Multi-stage A- and D-optimal designs Suppose we have an initial design ξ 0 = {(0, 0.25), (1, 0.25), (2, 0.25), (3, 0.25)} with n 0 = 40 Aim: Allocate the next 80 observations Computing time (in seconds) under the new algorithm D A θ (θ 1, θ 3 ) (θ 2, θ 4 ) θ (θ 1, θ 3 ) (θ 2, θ 4 ) N = N = N = N = / 45

27 Example II Set up Consider model Y = θ 1 + θ 2 r + θ 3 r 2 + θ 4 s + θ 5 rs + ε. Let θ = (θ 1, θ 2, θ 3, θ 4, θ 5 ) and X = {(r i, s j ), r i = 2i/k 1, i = 1,..., k, s j = j/k, j = 1,..., k}. The same set up as that of table 4 of Yu (2011) Algorithms are coded using SAS IML and computed at a Dell laptop (2.2GHz and 8Gb Ram) 27 / 45

28 Example II Computing time (in seconds) comparison II D-optimal design for the full parameter vector θ N = 20 2 N = 50 2 N = N = N = Cocktail new algorithm / 45

29 Example III Set up Consider multinomial logit model with 3 categories, and 3 variables: Response Y = (Y 1, Y 2, Y 3 ) T at experimental condition x with Y 1 + Y 2 + Y 3 = 1 and π i (x) = P(Y i = 1 x) = where g(x) T = (1, x 1, x 2, x 3 ) e g(x)t θ i 1 + e g(x)t θ 1 + e g(x) T θ 2, i = 1, / 45

30 Example III Modification to algorithm (example III only) Design space X = {(6i/k, 6j/k, 6l/k), i, j, l = 0, 1,..., k} The number of design points in X rapidly increases as k increases (N k 3 ) Start with a coarse grid Identify the best design based on the grid at that stage For the next stage, a finer grid is restricted to neighbourhoods of the best support points found at the current stage Continue until a specified accuracy for the design points is reached Verify optimality through the equivalence theorem 30 / 45

31 Example III Computing time (in seconds) example III Locally optimal design Multi-stage optimal design θ θ θ θ N D A D A D A D A θ T = (θ T 1, θt 2 ) where θ 1 = (1, 1, 1, 2) and θ 2 = ( 1, 2, 1, 1). θ = (θ 11, θ 12, θ 13, θ 21, θ 22, θ 23 ) 31 / 45

32 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 32 / 45

33 Summary The new algorithm is even faster than the Cocktail algorithm More importantly, the new algorithm is more widely applicable: Φ p-optimality Single- and Multi-stage designs interest in functions of the parameters Future work: Extend algorithm to find Bayesian and (hopefully) Minimax designs Compare its performance with further algorithms, e.g PSO 33 / 45

34 Thank You! 34 / 45

35 Selected references Böhning, D. (1986), A vertex-exchange-method in D-optimal design theory, Metrika, 33, Fedorov, V.V. (1972). Theory of optimal experiments (transl. and ed. by Studden WJ, Klimko EM), New York: Academic Press. Silvey, S.D., Titterington, D.M., and Torsney, B. (1978), An algorithm for optimal designs on a finite design space, Communications in Statistics - Theory and Methods, 14, Yang, M. and Biedermann, S. (2012). On optimal designs for nonlinear models: a general and efficient algorithm, Under revision. Yu, Y. (2011), D-optimal designs via a cocktail algorithm, Statistics and Computing, 21, / 45

36 36 / 45

37 VDM Starting with an initial design ξ (0) (uniform design over a set of support points). Let ξ (t) be the design with the corresponding weight vector ω (t). ξ (t+1) is updated through the following steps. Select 1 i max N such that d(i max, ξ (t) ) = max i n N d(i, ξ (t) ) Stop when d(i max, ξ (t) ) m + ɛ, otherwise { ω (t+1) (1 δ)ω (t) i = i, i i max (1 δ)ω (t) where δ = d(imax,ξ i + δ, i = i max (t) )/m 1 d(i max,ξ (t) ) / 45

38 MA Starting with an initial design ξ (0) (uniform design over a set of support points). Let ξ (t) be the design with the corresponding weight vector ω (t). ξ (t+1) is updated through the following steps. Stop when d(i max, ξ (t) ) m + ɛ, otherwise ω (t+1) i = ω (t) i d(i, ξ (t) )/m, where m is the dimension of I ξ. 38 / 45

39 VEM VEM (vertex exchange method) is close to VDM but performs better. Select i min and i max such that d(i min, ξ (t) ) = min{d(i, ξ (t) ) : w (t) i > 0}, and d(i max, ξ (t) ) = max i n N d(i, ξ (t) ) Stop when d(i max, ξ (t) ) m + ɛ, otherwise Set ω (t+1) = VE(i max, i min, ω (t) ). 39 / 45

40 Vertex exchanges (VE) VE(j, k, ξ (t) ) is to exchange the weight between x j and x k, where ω (t) ω (t+1) i, i j, k i = ω (t) i δ, i = j Here δ = min{ω j, max{ ω k, δ (j, k)}} ω (t) i + δ, i = k and δ (j, k) = d(k, ξ (t) ) d(j, ξ (t) ) ( 2 d(j, ξ (t) ) d(k, ξ (t) ) ( (f (x j ) T I 1 ξ (t) f (x k ) ) 2 ) 40 / 45

41 Nearest neighbour exchanges (NNE) Let i 1,..., i q+1 be the elements of {i : ω (t) i > 0} where q + 1 is the number of support points of ω (t). For each i j, j = 1,..., q, let ij be any index i {i j+1,..., i q+1 } such that x i x ij is minimized. Perform VE between x ij and x i j for j = 1,..., q in turn, i.e., ω (t+j/q)=ve(i j,i j,ξ (t+(j 1)/q) ) 41 / 45

42 42 / 45

43 Computing optimal weights Initial weight ω (t) 0 for S (t) can be updated through the following iteration: (a) ω (t) j = ω (t) j 1 α ( ) 1 2 Φ p(σ ξ0 +ξ(g)) Φp(Σ ξ0 (g)) +ξ ωω T (t) ω=ω ω (t) j 1 ω=ω j 1 (b) Check if there are non-positive components of ω (t) j (c) If answer of (b) is no, check whether Φp(Σ ξ 0 +ξ(g)) ω (t) ω=ω is less j than some pre-specified ɛ 1. If yes, ω (t) j holds the optimal weights. If no, start the next iteration. (d) If answer of (b) is yes, reduce α to α/2. Repeat (a) and (b) until α reaches a pre-specified value. Remove the support point with smallest weight, and repeat (a) and (b) 43 / 45

44 44 / 45

45 Sensitivity function: ( d(x i, ξ S ) = Tr[ I(x (t) i ) I ξ ( Σ ξ S (t) ) p 1 ( F (θ) For D-optimality, p = 0. θ T ) ( I ξ0 + ri ξ S ) (t) ( ) ] I + ri ξ0 ξ S (t) S (t) ) ( F(θ) θ T ) T 45 / 45

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

On construction of constrained optimum designs

On construction of constrained optimum designs On construction of constrained optimum designs Institute of Control and Computation Engineering University of Zielona Góra, Poland DEMA2008, Cambridge, 15 August 2008 Numerical algorithms to construct

More information

Optimal experimental design, an introduction, Jesús López Fidalgo

Optimal experimental design, an introduction, Jesús López Fidalgo Optimal experimental design, an introduction Jesus.LopezFidalgo@uclm.es University of Castilla-La Mancha Department of Mathematics Institute of Applied Mathematics to Science and Engineering Books (just

More information

Lecture 3 September 1

Lecture 3 September 1 STAT 383C: Statistical Modeling I Fall 2016 Lecture 3 September 1 Lecturer: Purnamrita Sarkar Scribe: Giorgio Paulon, Carlos Zanini Disclaimer: These scribe notes have been slightly proofread and may have

More information

Econometrics I, Estimation

Econometrics I, Estimation Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the

More information

Statistical Estimation

Statistical Estimation Statistical Estimation Use data and a model. The plug-in estimators are based on the simple principle of applying the defining functional to the ECDF. Other methods of estimation: minimize residuals from

More information

Efficient Bayesian Multivariate Surface Regression

Efficient Bayesian Multivariate Surface Regression Efficient Bayesian Multivariate Surface Regression Feng Li (joint with Mattias Villani) Department of Statistics, Stockholm University October, 211 Outline of the talk 1 Flexible regression models 2 The

More information

On Multiple-Objective Nonlinear Optimal Designs

On Multiple-Objective Nonlinear Optimal Designs On Multiple-Objective Nonlinear Optimal Designs Qianshun Cheng, Dibyen Majumdar, and Min Yang December 1, 2015 Abstract Experiments with multiple objectives form a staple diet of modern scientific research.

More information

Phasing via the Expectation Maximization (EM) Algorithm

Phasing via the Expectation Maximization (EM) Algorithm Computing Haplotype Frequencies and Haplotype Phasing via the Expectation Maximization (EM) Algorithm Department of Computer Science Brown University, Providence sorin@cs.brown.edu September 14, 2010 Outline

More information

Active Robot Calibration Algorithm

Active Robot Calibration Algorithm Active Robot Calibration Algorithm Yu Sun and John M. Hollerbach Abstract This paper presents a new updating algorithm to reduce the complexity of computing an observability index for kinematic calibration

More information

An Inexact Newton Method for Optimization

An Inexact Newton Method for Optimization New York University Brown Applied Mathematics Seminar, February 10, 2009 Brief biography New York State College of William and Mary (B.S.) Northwestern University (M.S. & Ph.D.) Courant Institute (Postdoc)

More information

Statistical Machine Learning Hilary Term 2018

Statistical Machine Learning Hilary Term 2018 Statistical Machine Learning Hilary Term 2018 Pier Francesco Palamara Department of Statistics University of Oxford Slide credits and other course material can be found at: http://www.stats.ox.ac.uk/~palamara/sml18.html

More information

Statistical Analysis of Binary Composite Endpoints with Missing Data in Components

Statistical Analysis of Binary Composite Endpoints with Missing Data in Components Statistical Analysis of Binary Composite Endpoints with Missing Data in Components Daowen Zhang zhang@stat.ncsu.edu http://www4.ncsu.edu/ dzhang Joint work with Hui Quan, Ji Zhang & Laure Devlamynck Sanofi-Aventis

More information

ML Estimation of Process Noise Variance in Dynamic Systems

ML Estimation of Process Noise Variance in Dynamic Systems ML Estimation of Process Noise Variance in Dynamic Systems Patrik Axelsson, Umut Orguner, Fredrik Gustafsson and Mikael Norrlöf {axelsson,umut,fredrik,mino} @isy.liu.se Division of Automatic Control Department

More information

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014

Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently

More information

MCMC algorithms for fitting Bayesian models

MCMC algorithms for fitting Bayesian models MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models

More information

Local&Bayesianoptimaldesigns in binary bioassay

Local&Bayesianoptimaldesigns in binary bioassay Local&Bayesianoptimaldesigns in binary bioassay D.M.Smith Office of Biostatistics & Bioinformatics Medicial College of Georgia. New Directions in Experimental Design (DAE 2003 Chicago) 1 Review of binary

More information

Optimal design of experiments

Optimal design of experiments Optimal design of experiments Session 4: Some theory Peter Goos / 40 Optimal design theory continuous or approximate optimal designs implicitly assume an infinitely large number of observations are available

More information

Numerical Methods I Solving Nonlinear Equations

Numerical Methods I Solving Nonlinear Equations Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)

More information

Efficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies

Efficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies Efficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies Tim Holland-Letz Ruhr-Universität Bochum Medizinische Fakultät 44780 Bochum, Germany email: tim.holland-letz@rub.de

More information

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Mai Zhou Yifan Yang Received: date / Accepted: date Abstract In this note

More information

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition

Last updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship

More information

D-optimal Designs for Multinomial Logistic Models

D-optimal Designs for Multinomial Logistic Models D-optimal Designs for Multinomial Logistic Models Jie Yang University of Illinois at Chicago Joint with Xianwei Bu and Dibyen Majumdar October 12, 2017 1 Multinomial Logistic Models Cumulative logit model:

More information

Appendix 8 Numerical Methods for Solving Nonlinear Equations 1

Appendix 8 Numerical Methods for Solving Nonlinear Equations 1 Appendix 8 Numerical Methods for Solving Nonlinear Equations 1 An equation is said to be nonlinear when it involves terms of degree higher than 1 in the unknown quantity. These terms may be polynomial

More information

Biostat 2065 Analysis of Incomplete Data

Biostat 2065 Analysis of Incomplete Data Biostat 2065 Analysis of Incomplete Data Gong Tang Dept of Biostatistics University of Pittsburgh October 20, 2005 1. Large-sample inference based on ML Let θ is the MLE, then the large-sample theory implies

More information

StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory

StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory S.V. N. (vishy) Vishwanathan Purdue University and Microsoft vishy@purdue.edu October 9, 2012 S.V. N. Vishwanathan (Purdue,

More information

OPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES

OPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES Statistica Sinica 21 (2011, 1415-1430 OPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES Min Yang, Bin Zhang and Shuguang Huang University of Missouri, University of Alabama-Birmingham

More information

Time Series Analysis

Time Series Analysis Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Regression based methods, 1st part: Introduction (Sec.

More information

Likelihood, MLE & EM for Gaussian Mixture Clustering. Nick Duffield Texas A&M University

Likelihood, MLE & EM for Gaussian Mixture Clustering. Nick Duffield Texas A&M University Likelihood, MLE & EM for Gaussian Mixture Clustering Nick Duffield Texas A&M University Probability vs. Likelihood Probability: predict unknown outcomes based on known parameters: P(x q) Likelihood: estimate

More information

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0

1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0 Numerical Analysis 1 1. Nonlinear Equations This lecture note excerpted parts from Michael Heath and Max Gunzburger. Given function f, we seek value x for which where f : D R n R n is nonlinear. f(x) =

More information

An Inexact Newton Method for Nonlinear Constrained Optimization

An Inexact Newton Method for Nonlinear Constrained Optimization An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results

More information

D-optimal Designs for Factorial Experiments under Generalized Linear Models

D-optimal Designs for Factorial Experiments under Generalized Linear Models D-optimal Designs for Factorial Experiments under Generalized Linear Models Jie Yang Department of Mathematics, Statistics, and Computer Science University of Illinois at Chicago Joint research with Abhyuday

More information

Research Article A Novel Differential Evolution Invasive Weed Optimization Algorithm for Solving Nonlinear Equations Systems

Research Article A Novel Differential Evolution Invasive Weed Optimization Algorithm for Solving Nonlinear Equations Systems Journal of Applied Mathematics Volume 2013, Article ID 757391, 18 pages http://dx.doi.org/10.1155/2013/757391 Research Article A Novel Differential Evolution Invasive Weed Optimization for Solving Nonlinear

More information

PROBABILITY AND STATISTICS Vol. III - Statistical Experiments and Optimal Design - Andrej Pázman STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN

PROBABILITY AND STATISTICS Vol. III - Statistical Experiments and Optimal Design - Andrej Pázman STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN Andrej Pázman Comenius University, Bratislava, Slovakia Keywords: Experiment design, linear statistical model, nonlinear regression, least squares, information

More information

SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD

SOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD BISECTION METHOD If a function f(x) is continuous between a and b, and f(a) and f(b) are of opposite signs, then there exists at least one root between a and b. It is shown graphically as, Let f a be negative

More information

Machine Learning 2017

Machine Learning 2017 Machine Learning 2017 Volker Roth Department of Mathematics & Computer Science University of Basel 21st March 2017 Volker Roth (University of Basel) Machine Learning 2017 21st March 2017 1 / 41 Section

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Gibbs Sampling in Linear Models #2

Gibbs Sampling in Linear Models #2 Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling

More information

Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics

Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html

More information

Multiplicative Algorithm for computing Optimum Designs

Multiplicative Algorithm for computing Optimum Designs for computing s DAE Conference 2012 Roberto Dorta-Guerra, Raúl Martín Martín, Mercedes Fernández-Guerrero and Licesio J. Rodríguez-Aragón Optimum Experimental Design group October 20, 2012 http://areaestadistica.uclm.es/oed

More information

Matrix Derivatives and Descent Optimization Methods

Matrix Derivatives and Descent Optimization Methods Matrix Derivatives and Descent Optimization Methods 1 Qiang Ning Department of Electrical and Computer Engineering Beckman Institute for Advanced Science and Techonology University of Illinois at Urbana-Champaign

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Logistic Regression. Seungjin Choi

Logistic Regression. Seungjin Choi Logistic Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Basic math for biology

Basic math for biology Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood

More information

Inexact Newton Methods and Nonlinear Constrained Optimization

Inexact Newton Methods and Nonlinear Constrained Optimization Inexact Newton Methods and Nonlinear Constrained Optimization Frank E. Curtis EPSRC Symposium Capstone Conference Warwick Mathematics Institute July 2, 2009 Outline PDE-Constrained Optimization Newton

More information

STAT 730 Chapter 4: Estimation

STAT 730 Chapter 4: Estimation STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

Spring 2012 Math 541B Exam 1

Spring 2012 Math 541B Exam 1 Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote

More information

Exponential families also behave nicely under conditioning. Specifically, suppose we write η = (η 1, η 2 ) R k R p k so that

Exponential families also behave nicely under conditioning. Specifically, suppose we write η = (η 1, η 2 ) R k R p k so that 1 More examples 1.1 Exponential families under conditioning Exponential families also behave nicely under conditioning. Specifically, suppose we write η = η 1, η 2 R k R p k so that dp η dm 0 = e ηt 1

More information

Efficient Solvers for Stochastic Finite Element Saddle Point Problems

Efficient Solvers for Stochastic Finite Element Saddle Point Problems Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite

More information

Master s Written Examination

Master s Written Examination Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth

More information

Optimal Designs for 2 k Experiments with Binary Response

Optimal Designs for 2 k Experiments with Binary Response 1 / 57 Optimal Designs for 2 k Experiments with Binary Response Dibyen Majumdar Mathematics, Statistics, and Computer Science College of Liberal Arts and Sciences University of Illinois at Chicago Joint

More information

1 One-way analysis of variance

1 One-way analysis of variance LIST OF FORMULAS (Version from 21. November 2014) STK2120 1 One-way analysis of variance Assume X ij = µ+α i +ɛ ij ; j = 1, 2,..., J i ; i = 1, 2,..., I ; where ɛ ij -s are independent and N(0, σ 2 ) distributed.

More information

Stochastic Design Criteria in Linear Models

Stochastic Design Criteria in Linear Models AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework

More information

Information in a Two-Stage Adaptive Optimal Design

Information in a Two-Stage Adaptive Optimal Design Information in a Two-Stage Adaptive Optimal Design Department of Statistics, University of Missouri Designed Experiments: Recent Advances in Methods and Applications DEMA 2011 Isaac Newton Institute for

More information

Optimal design of experiments

Optimal design of experiments Optimal design of experiments Session 5: Design construction Peter Goos 1 / 33 Constructing exact optimal designs direct approaches to maximize X T X Microsoft Excel mathematical programming run into problems

More information

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project

Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore

More information

COURSE Iterative methods for solving linear systems

COURSE Iterative methods for solving linear systems COURSE 0 4.3. Iterative methods for solving linear systems Because of round-off errors, direct methods become less efficient than iterative methods for large systems (>00 000 variables). An iterative scheme

More information

Bayesian estimation of the discrepancy with misspecified parametric models

Bayesian estimation of the discrepancy with misspecified parametric models Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012

More information

Computer Intensive Methods in Mathematical Statistics

Computer Intensive Methods in Mathematical Statistics Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of

More information

Estimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1

Estimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1 Estimation and Model Selection in Mixed Effects Models Part I Adeline Samson 1 1 University Paris Descartes Summer school 2009 - Lipari, Italy These slides are based on Marc Lavielle s slides Outline 1

More information

Efficient Bayesian Multivariate Surface Regression

Efficient Bayesian Multivariate Surface Regression Efficient Bayesian Multivariate Surface Regression Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Outline of the talk 1 Introduction to flexible

More information

Interior Point Algorithms for Constrained Convex Optimization

Interior Point Algorithms for Constrained Convex Optimization Interior Point Algorithms for Constrained Convex Optimization Chee Wei Tan CS 8292 : Advanced Topics in Convex Optimization and its Applications Fall 2010 Outline Inequality constrained minimization problems

More information

Sparse Approximation via Penalty Decomposition Methods

Sparse Approximation via Penalty Decomposition Methods Sparse Approximation via Penalty Decomposition Methods Zhaosong Lu Yong Zhang February 19, 2012 Abstract In this paper we consider sparse approximation problems, that is, general l 0 minimization problems

More information

Describing Contingency tables

Describing Contingency tables Today s topics: Describing Contingency tables 1. Probability structure for contingency tables (distributions, sensitivity/specificity, sampling schemes). 2. Comparing two proportions (relative risk, odds

More information

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception

LINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,

More information

Random Eigenvalue Problems Revisited

Random Eigenvalue Problems Revisited Random Eigenvalue Problems Revisited S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html

More information

Lecture 20: Linear model, the LSE, and UMVUE

Lecture 20: Linear model, the LSE, and UMVUE Lecture 20: Linear model, the LSE, and UMVUE Linear Models One of the most useful statistical models is X i = β τ Z i + ε i, i = 1,...,n, where X i is the ith observation and is often called the ith response;

More information

L11: Pattern recognition principles

L11: Pattern recognition principles L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction

More information

AGEC 661 Note Eleven Ximing Wu. Exponential regression model: m (x, θ) = exp (xθ) for y 0

AGEC 661 Note Eleven Ximing Wu. Exponential regression model: m (x, θ) = exp (xθ) for y 0 AGEC 661 ote Eleven Ximing Wu M-estimator So far we ve focused on linear models, where the estimators have a closed form solution. If the population model is nonlinear, the estimators often do not have

More information

Optimal discrimination designs

Optimal discrimination designs Optimal discrimination designs Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@ruhr-uni-bochum.de Stefanie Titoff Ruhr-Universität Bochum Fakultät

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

Least Squares Approximation

Least Squares Approximation Chapter 6 Least Squares Approximation As we saw in Chapter 5 we can interpret radial basis function interpolation as a constrained optimization problem. We now take this point of view again, but start

More information

Gaussian Process Approximations of Stochastic Differential Equations

Gaussian Process Approximations of Stochastic Differential Equations Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run

More information

Gaussian mean-shift algorithms

Gaussian mean-shift algorithms Gaussian mean-shift algorithms Miguel Á. Carreira-Perpiñán Dept. of Computer Science & Electrical Engineering, OGI/OHSU http://www.csee.ogi.edu/~miguel Gaussian mean-shift (GMS) as an EM algorithm Gaussian

More information

A geometric characterization of c-optimal designs for heteroscedastic regression

A geometric characterization of c-optimal designs for heteroscedastic regression A geometric characterization of c-optimal designs for heteroscedastic regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Tim Holland-Letz

More information

Image restoration: numerical optimisation

Image restoration: numerical optimisation Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context

More information

Minimax design criterion for fractional factorial designs

Minimax design criterion for fractional factorial designs Ann Inst Stat Math 205 67:673 685 DOI 0.007/s0463-04-0470-0 Minimax design criterion for fractional factorial designs Yue Yin Julie Zhou Received: 2 November 203 / Revised: 5 March 204 / Published online:

More information

A Study on Minimal-point Composite Designs

A Study on Minimal-point Composite Designs A Study on Minimal-point Composite Designs by Yin-Jie Huang Advisor Ray-Bing Chen Institute of Statistics, National University of Kaohsiung Kaohsiung, Taiwan 8 R.O.C. July 007 Contents Introduction Construction

More information

POLI 8501 Introduction to Maximum Likelihood Estimation

POLI 8501 Introduction to Maximum Likelihood Estimation POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,

More information

COMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure)

COMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure) STAT 52 Completely Randomized Designs COMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure) Completely randomized means no restrictions on the randomization

More information

Least squares problems

Least squares problems Least squares problems How to state and solve them, then evaluate their solutions Stéphane Mottelet Université de Technologie de Compiègne 30 septembre 2016 Stéphane Mottelet (UTC) Least squares 1 / 55

More information

Construction of designs for twophase experiments with Tabu search

Construction of designs for twophase experiments with Tabu search Construction of designs for twophase experiments with Tabu search Nha Vo-Thanh In collaboration with Professor Hans-Peter Piepho University of Hohenheim Common bean experiment A breeding program by CIAT

More information

EM Algorithm II. September 11, 2018

EM Algorithm II. September 11, 2018 EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data

More information

Nonlinear Error Correction Model and Multiple-Threshold Cointegration May 23, / 31

Nonlinear Error Correction Model and Multiple-Threshold Cointegration May 23, / 31 Nonlinear Error Correction Model and Multiple-Threshold Cointegration Man Wang Dong Hua University, China Joint work with N.H.Chan May 23, 2014 Nonlinear Error Correction Model and Multiple-Threshold Cointegration

More information

The consequences of misspecifying the random effects distribution when fitting generalized linear mixed models

The consequences of misspecifying the random effects distribution when fitting generalized linear mixed models The consequences of misspecifying the random effects distribution when fitting generalized linear mixed models John M. Neuhaus Charles E. McCulloch Division of Biostatistics University of California, San

More information

Easy and not so easy multifacility location problems... (In 20 minutes.)

Easy and not so easy multifacility location problems... (In 20 minutes.) Easy and not so easy multifacility location problems... (In 20 minutes.) MINLP 2014 Pittsburgh, June 2014 Justo Puerto Institute of Mathematics (IMUS) Universidad de Sevilla Outline 1 Introduction (In

More information

PHYS 410/555 Computational Physics Solution of Non Linear Equations (a.k.a. Root Finding) (Reference Numerical Recipes, 9.0, 9.1, 9.

PHYS 410/555 Computational Physics Solution of Non Linear Equations (a.k.a. Root Finding) (Reference Numerical Recipes, 9.0, 9.1, 9. PHYS 410/555 Computational Physics Solution of Non Linear Equations (a.k.a. Root Finding) (Reference Numerical Recipes, 9.0, 9.1, 9.4) We will consider two cases 1. f(x) = 0 1-dimensional 2. f(x) = 0 d-dimensional

More information

MATH 4211/6211 Optimization Basics of Optimization Problems

MATH 4211/6211 Optimization Basics of Optimization Problems MATH 4211/6211 Optimization Basics of Optimization Problems Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 A standard minimization

More information

Lecture 7 Unconstrained nonlinear programming

Lecture 7 Unconstrained nonlinear programming Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,

More information

Mahdi karbasian* & Zoubi Ibrahim

Mahdi karbasian* & Zoubi Ibrahim International Journal of Industrial Engineering & Production Research (010) pp. 105-110 September 010, Volume 1, Number International Journal of Industrial Engineering & Production Research ISSN: 008-4889

More information

Logistic Regression and Generalized Linear Models

Logistic Regression and Generalized Linear Models Logistic Regression and Generalized Linear Models Sridhar Mahadevan mahadeva@cs.umass.edu University of Massachusetts Sridhar Mahadevan: CMPSCI 689 p. 1/2 Topics Generative vs. Discriminative models In

More information

Statistical Inference of Covariate-Adjusted Randomized Experiments

Statistical Inference of Covariate-Adjusted Randomized Experiments 1 Statistical Inference of Covariate-Adjusted Randomized Experiments Feifang Hu Department of Statistics George Washington University Joint research with Wei Ma, Yichen Qin and Yang Li Email: feifang@gwu.edu

More information

Introduction to gradient descent

Introduction to gradient descent 6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our

More information

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.

11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly. C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a

More information

LASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape

LASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape LASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape Nikolaus Umlauf https://eeecon.uibk.ac.at/~umlauf/ Overview Joint work with Andreas Groll, Julien Hambuckers

More information

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices

Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.

More information

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria

Weak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then

More information

Stochastic Analogues to Deterministic Optimizers

Stochastic Analogues to Deterministic Optimizers Stochastic Analogues to Deterministic Optimizers ISMP 2018 Bordeaux, France Vivak Patel Presented by: Mihai Anitescu July 6, 2018 1 Apology I apologize for not being here to give this talk myself. I injured

More information

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)

A nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable) Nonlinear equations Definition A nonlinear equation is any equation of the form where f is a nonlinear function. Nonlinear equations x 2 + x + 1 = 0 (f : R R) f(x) = 0 (x cos y, 2y sin x) = (0, 0) (f :

More information