A new algorithm for deriving optimal designs
|
|
- Luke Pope
- 6 years ago
- Views:
Transcription
1 A new algorithm for deriving optimal designs Stefanie Biedermann, University of Southampton, UK Joint work with Min Yang, University of Illinois at Chicago 18 October 2012, DAE, University of Georgia, Athens, GA 1 / 45
2 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 2 / 45
3 Statistics and Optimal design of experiments The main goal of statistics is to gain information from collected data Design of experiments concerns the way of data collection An optimal/efficient design can reduce the sample size and cost needed for achieving a specified precision of estimation, or can improve the precision for a given sample size 3 / 45
4 Exact and approximate designs Exact design ξ n for sample size n: ξ n = { x1 x 2... x m n 1 /n n 2 /n... n m /n } ; n i integers, m n i = n i=1 Approximate design ξ: ξ = { } x1 x 2... x m ; ω ω 1 ω 2... ω i > 0, m m ω i = 1 Note: To run an approximate design, some rounding procedure is required i=1 4 / 45
5 Optimality criteria Suppose interest is in F(θ), a vector of functions of the model parameters The (asymptotic) covariance matrix of the MLE of F(θ) under design ξ, Σ ξ (F), can be written as Σ ξ (F) = F (θ) θ T (I ξ) F(θ) θ where I ξ is the information matrix of the design ξ Aim: Obtain precise estimates Minimise Φ(Σ ξ (F )), an optimality criterion 5 / 45
6 Optimality criteria Example 1: A D-optimal design minimises the determinant of the covariance matrix minimises the volume of a confidence ellipsoid for F(θ) Example 2: An A-optimal design minimises the trace of the covariance matrix minimises the sum of the variances of the elements of F(ˆθ) 6 / 45
7 Locally optimal designs A large and popular class of optimality criteria are the Φ p -criteria, of which A- and D-optimality are special cases: Φ p (Σ ξ (F)) = where 0 p < and Σ ξ (F ) IR v v [ ] 1/p 1 v trace(σ ξ(f )) p Problem: In many situations, Φ p -optimal designs depend on the unknown parameter vector θ locally optimal designs 7 / 45
8 Optimal multi-stage designs Use a sequential strategy: Suppose n 0 observations are already available, from design ξ n0, and another n observations can be made Estimate θ ˆθ 0 Design problem: Find ξ such that the combined design ξ = n 0 /(n 0 + n)ξ n0 + n/(n 0 + n)ξ minimises Φ(Σ ξ (F )) for ˆθ 0 Note: ξ and thus ξ are approximate designs 8 / 45
9 Aim of this project Provide a fast, general, convenient, and easy to use approach to find locally and multi-stage optimal designs 9 / 45
10 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 10 / 45
11 Current algorithms Three well-known algorithms are: Vertex direction method (VDM; Fedorov, 1972) Multiplicative algorithm (MA; Silvey, Titterington and Torsney, 1978) Vertex exchange method (VEM; Böhning, 1986) Yu (2011) combines the best features of these algorithms with a nearest neighbour exchange step to improve speed Cocktail algorithm 11 / 45
12 Computing time comparison I Model: Y = θ 1 e θ 2x + θ 3 e θ 4x + ε Design space [0, 3] discretised to X = {3i/N, i = 1,..., N} Computing time (in seconds) (Table 1 of Yu, 2011) N = 20 N = 50 N = 100 N = 200 N = 500 MA VEM Cocktail (VDM is omitted since it is known to be slower than VEM) 12 / 45
13 Computing time comparison II Y = θ 1 + θ 2 r + θ 3 r 2 + θ 4 s + θ 5 rs + ε Design space [ 1, 1] [0, 1] discretised to X = {(r i, s j ), r i = 2i/k 1, i = 1,..., k, s j = j/k, j = 1,..., k}, N = k 2 Computing time (in seconds) (Table 4 of Yu, 2011) N = 20 2 N = 50 2 N = N = MA aborted aborted VEM aborted Cocktail / 45
14 Summary of current algorithms All algorithms are relatively easy to implement The Cocktail algorithm leads to dramatically improved speed, sometimes by several orders of magnitude, relative to MA, VDM, and VEM. The Cocktail algorithm is restricted to D-optimal design for the full parameter vector Restricted to (one-stage) locally optimal designs 14 / 45
15 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 15 / 45
16 Goal of the new algorithm Any parameters (or functions thereof) can be of interest Any optimality criterion: Φ p -optimality (includes D-, A-, and E-optimality) One-stage and multi-stage Convenient to use Even faster 16 / 45
17 Rationale 1. The existing algorithms typically seek monotonic convergence May not be the fastest way to converge to the optimal weights 2. New idea: instead of seeking monotonic convergence, we derive the optimal weights directly, which can be done by solving a system of nonlinear equations 17 / 45
18 Notations for the new algorithm X = {x 1,..., x N }, the set of all possible design points S (t), the support after tth iteration ξ, design with support S (t) and optimal weights S (t) d(x, ξ ), sensitivity function of design ξ S (t) S (t) 18 / 45
19 The new algorithm Start with an initial design support, S (0). In the tth step: Derive ξ using the Newton-Raphson method S (t) Compute the sensitivity function d(x i, ξ ) for i = 1,..., N S (t) Select 1 i max N such that d(x imax, ξ S (t) ) = max i n N d(x i, ξ S (t) ) Stop when d(x imax, ξ S (t) ) ɛ, ɛ > 0 very small, otherwise S (t+1) = S (t) {x imax } 19 / 45
20 Convergence (I) Theorem 1: For a multi-stage design, let I ξ0 be non-singular. Then for any initial support S (0), the sequence of designs {ξ S (t) ; t 0} converges to an optimal design minimising Φ p (Σ ξ (F)) as t, provided the optimal weights are found in each updating step For a one-stage design, the above result holds if I ξs (0) and if the Jacobian of F is square and of full rank is non-singular, 20 / 45
21 Convergence (II) Theorem 2: For a multi-stage design, let p be an integer and I ξ0 be non-singular. For given support, Φ p (Σ ξ (F)) is minimised (w.r.t the weights) at any critical point in Ω = {w = (w 1,..., w m 1 ) T : w i 0, w w m 1 1} or its boundary. In addition, the Hessian of Φ p (Σ ξ (F )) is non-negative definite. Newton-Raphson method will converge provided the starting point is close enough to the critical point. Since Ω is compact, we can always find a critical point (given it exists) by using sufficiently many starting values 21 / 45
22 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 22 / 45
23 Example I Set up Consider model Y = θ 1 e θ 2x + θ 3 e θ 4x + ε. Let θ = (θ 1, θ 2, θ 3, θ 4 ) and X = {3i/N, i = 1,..., N}. Assume that θ = (1, 1, 1, 2). The same set up as that of Table 1 of Yu (2011) Algorithms are coded using SAS IML and computed at a Dell laptop (2.2GHz and 8Gb Ram) 23 / 45
24 Example I Computing time (in seconds) comparison I Computing time (in seconds) for the D-optimal design for the full parameter vector θ N = 500 N = 1000 N = 5000 N = Cocktail New algorithm / 45
25 Example I Locally A- and D-optimal designs Computing time (in seconds) under the new algorithm D A θ (θ 1, θ 3 ) (θ 2, θ 4 ) θ (θ 1, θ 3 ) (θ 2, θ 4 ) N = N = N = N = / 45
26 Example I Multi-stage A- and D-optimal designs Suppose we have an initial design ξ 0 = {(0, 0.25), (1, 0.25), (2, 0.25), (3, 0.25)} with n 0 = 40 Aim: Allocate the next 80 observations Computing time (in seconds) under the new algorithm D A θ (θ 1, θ 3 ) (θ 2, θ 4 ) θ (θ 1, θ 3 ) (θ 2, θ 4 ) N = N = N = N = / 45
27 Example II Set up Consider model Y = θ 1 + θ 2 r + θ 3 r 2 + θ 4 s + θ 5 rs + ε. Let θ = (θ 1, θ 2, θ 3, θ 4, θ 5 ) and X = {(r i, s j ), r i = 2i/k 1, i = 1,..., k, s j = j/k, j = 1,..., k}. The same set up as that of table 4 of Yu (2011) Algorithms are coded using SAS IML and computed at a Dell laptop (2.2GHz and 8Gb Ram) 27 / 45
28 Example II Computing time (in seconds) comparison II D-optimal design for the full parameter vector θ N = 20 2 N = 50 2 N = N = N = Cocktail new algorithm / 45
29 Example III Set up Consider multinomial logit model with 3 categories, and 3 variables: Response Y = (Y 1, Y 2, Y 3 ) T at experimental condition x with Y 1 + Y 2 + Y 3 = 1 and π i (x) = P(Y i = 1 x) = where g(x) T = (1, x 1, x 2, x 3 ) e g(x)t θ i 1 + e g(x)t θ 1 + e g(x) T θ 2, i = 1, / 45
30 Example III Modification to algorithm (example III only) Design space X = {(6i/k, 6j/k, 6l/k), i, j, l = 0, 1,..., k} The number of design points in X rapidly increases as k increases (N k 3 ) Start with a coarse grid Identify the best design based on the grid at that stage For the next stage, a finer grid is restricted to neighbourhoods of the best support points found at the current stage Continue until a specified accuracy for the design points is reached Verify optimality through the equivalence theorem 30 / 45
31 Example III Computing time (in seconds) example III Locally optimal design Multi-stage optimal design θ θ θ θ N D A D A D A D A θ T = (θ T 1, θt 2 ) where θ 1 = (1, 1, 1, 2) and θ 2 = ( 1, 2, 1, 1). θ = (θ 11, θ 12, θ 13, θ 21, θ 22, θ 23 ) 31 / 45
32 Outline 1 Introduction 2 Current algorithms 3 New algorithm 4 Examples Example I Example II Example III 5 Summary 32 / 45
33 Summary The new algorithm is even faster than the Cocktail algorithm More importantly, the new algorithm is more widely applicable: Φ p-optimality Single- and Multi-stage designs interest in functions of the parameters Future work: Extend algorithm to find Bayesian and (hopefully) Minimax designs Compare its performance with further algorithms, e.g PSO 33 / 45
34 Thank You! 34 / 45
35 Selected references Böhning, D. (1986), A vertex-exchange-method in D-optimal design theory, Metrika, 33, Fedorov, V.V. (1972). Theory of optimal experiments (transl. and ed. by Studden WJ, Klimko EM), New York: Academic Press. Silvey, S.D., Titterington, D.M., and Torsney, B. (1978), An algorithm for optimal designs on a finite design space, Communications in Statistics - Theory and Methods, 14, Yang, M. and Biedermann, S. (2012). On optimal designs for nonlinear models: a general and efficient algorithm, Under revision. Yu, Y. (2011), D-optimal designs via a cocktail algorithm, Statistics and Computing, 21, / 45
36 36 / 45
37 VDM Starting with an initial design ξ (0) (uniform design over a set of support points). Let ξ (t) be the design with the corresponding weight vector ω (t). ξ (t+1) is updated through the following steps. Select 1 i max N such that d(i max, ξ (t) ) = max i n N d(i, ξ (t) ) Stop when d(i max, ξ (t) ) m + ɛ, otherwise { ω (t+1) (1 δ)ω (t) i = i, i i max (1 δ)ω (t) where δ = d(imax,ξ i + δ, i = i max (t) )/m 1 d(i max,ξ (t) ) / 45
38 MA Starting with an initial design ξ (0) (uniform design over a set of support points). Let ξ (t) be the design with the corresponding weight vector ω (t). ξ (t+1) is updated through the following steps. Stop when d(i max, ξ (t) ) m + ɛ, otherwise ω (t+1) i = ω (t) i d(i, ξ (t) )/m, where m is the dimension of I ξ. 38 / 45
39 VEM VEM (vertex exchange method) is close to VDM but performs better. Select i min and i max such that d(i min, ξ (t) ) = min{d(i, ξ (t) ) : w (t) i > 0}, and d(i max, ξ (t) ) = max i n N d(i, ξ (t) ) Stop when d(i max, ξ (t) ) m + ɛ, otherwise Set ω (t+1) = VE(i max, i min, ω (t) ). 39 / 45
40 Vertex exchanges (VE) VE(j, k, ξ (t) ) is to exchange the weight between x j and x k, where ω (t) ω (t+1) i, i j, k i = ω (t) i δ, i = j Here δ = min{ω j, max{ ω k, δ (j, k)}} ω (t) i + δ, i = k and δ (j, k) = d(k, ξ (t) ) d(j, ξ (t) ) ( 2 d(j, ξ (t) ) d(k, ξ (t) ) ( (f (x j ) T I 1 ξ (t) f (x k ) ) 2 ) 40 / 45
41 Nearest neighbour exchanges (NNE) Let i 1,..., i q+1 be the elements of {i : ω (t) i > 0} where q + 1 is the number of support points of ω (t). For each i j, j = 1,..., q, let ij be any index i {i j+1,..., i q+1 } such that x i x ij is minimized. Perform VE between x ij and x i j for j = 1,..., q in turn, i.e., ω (t+j/q)=ve(i j,i j,ξ (t+(j 1)/q) ) 41 / 45
42 42 / 45
43 Computing optimal weights Initial weight ω (t) 0 for S (t) can be updated through the following iteration: (a) ω (t) j = ω (t) j 1 α ( ) 1 2 Φ p(σ ξ0 +ξ(g)) Φp(Σ ξ0 (g)) +ξ ωω T (t) ω=ω ω (t) j 1 ω=ω j 1 (b) Check if there are non-positive components of ω (t) j (c) If answer of (b) is no, check whether Φp(Σ ξ 0 +ξ(g)) ω (t) ω=ω is less j than some pre-specified ɛ 1. If yes, ω (t) j holds the optimal weights. If no, start the next iteration. (d) If answer of (b) is yes, reduce α to α/2. Repeat (a) and (b) until α reaches a pre-specified value. Remove the support point with smallest weight, and repeat (a) and (b) 43 / 45
44 44 / 45
45 Sensitivity function: ( d(x i, ξ S ) = Tr[ I(x (t) i ) I ξ ( Σ ξ S (t) ) p 1 ( F (θ) For D-optimality, p = 0. θ T ) ( I ξ0 + ri ξ S ) (t) ( ) ] I + ri ξ0 ξ S (t) S (t) ) ( F(θ) θ T ) T 45 / 45
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017
Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a
More informationOn construction of constrained optimum designs
On construction of constrained optimum designs Institute of Control and Computation Engineering University of Zielona Góra, Poland DEMA2008, Cambridge, 15 August 2008 Numerical algorithms to construct
More informationOptimal experimental design, an introduction, Jesús López Fidalgo
Optimal experimental design, an introduction Jesus.LopezFidalgo@uclm.es University of Castilla-La Mancha Department of Mathematics Institute of Applied Mathematics to Science and Engineering Books (just
More informationLecture 3 September 1
STAT 383C: Statistical Modeling I Fall 2016 Lecture 3 September 1 Lecturer: Purnamrita Sarkar Scribe: Giorgio Paulon, Carlos Zanini Disclaimer: These scribe notes have been slightly proofread and may have
More informationEconometrics I, Estimation
Econometrics I, Estimation Department of Economics Stanford University September, 2008 Part I Parameter, Estimator, Estimate A parametric is a feature of the population. An estimator is a function of the
More informationStatistical Estimation
Statistical Estimation Use data and a model. The plug-in estimators are based on the simple principle of applying the defining functional to the ECDF. Other methods of estimation: minimize residuals from
More informationEfficient Bayesian Multivariate Surface Regression
Efficient Bayesian Multivariate Surface Regression Feng Li (joint with Mattias Villani) Department of Statistics, Stockholm University October, 211 Outline of the talk 1 Flexible regression models 2 The
More informationOn Multiple-Objective Nonlinear Optimal Designs
On Multiple-Objective Nonlinear Optimal Designs Qianshun Cheng, Dibyen Majumdar, and Min Yang December 1, 2015 Abstract Experiments with multiple objectives form a staple diet of modern scientific research.
More informationPhasing via the Expectation Maximization (EM) Algorithm
Computing Haplotype Frequencies and Haplotype Phasing via the Expectation Maximization (EM) Algorithm Department of Computer Science Brown University, Providence sorin@cs.brown.edu September 14, 2010 Outline
More informationActive Robot Calibration Algorithm
Active Robot Calibration Algorithm Yu Sun and John M. Hollerbach Abstract This paper presents a new updating algorithm to reduce the complexity of computing an observability index for kinematic calibration
More informationAn Inexact Newton Method for Optimization
New York University Brown Applied Mathematics Seminar, February 10, 2009 Brief biography New York State College of William and Mary (B.S.) Northwestern University (M.S. & Ph.D.) Courant Institute (Postdoc)
More informationStatistical Machine Learning Hilary Term 2018
Statistical Machine Learning Hilary Term 2018 Pier Francesco Palamara Department of Statistics University of Oxford Slide credits and other course material can be found at: http://www.stats.ox.ac.uk/~palamara/sml18.html
More informationStatistical Analysis of Binary Composite Endpoints with Missing Data in Components
Statistical Analysis of Binary Composite Endpoints with Missing Data in Components Daowen Zhang zhang@stat.ncsu.edu http://www4.ncsu.edu/ dzhang Joint work with Hui Quan, Ji Zhang & Laure Devlamynck Sanofi-Aventis
More informationML Estimation of Process Noise Variance in Dynamic Systems
ML Estimation of Process Noise Variance in Dynamic Systems Patrik Axelsson, Umut Orguner, Fredrik Gustafsson and Mikael Norrlöf {axelsson,umut,fredrik,mino} @isy.liu.se Division of Automatic Control Department
More informationPh.D. Qualifying Exam Friday Saturday, January 3 4, 2014
Ph.D. Qualifying Exam Friday Saturday, January 3 4, 2014 Put your solution to each problem on a separate sheet of paper. Problem 1. (5166) Assume that two random samples {x i } and {y i } are independently
More informationMCMC algorithms for fitting Bayesian models
MCMC algorithms for fitting Bayesian models p. 1/1 MCMC algorithms for fitting Bayesian models Sudipto Banerjee sudiptob@biostat.umn.edu University of Minnesota MCMC algorithms for fitting Bayesian models
More informationLocal&Bayesianoptimaldesigns in binary bioassay
Local&Bayesianoptimaldesigns in binary bioassay D.M.Smith Office of Biostatistics & Bioinformatics Medicial College of Georgia. New Directions in Experimental Design (DAE 2003 Chicago) 1 Review of binary
More informationOptimal design of experiments
Optimal design of experiments Session 4: Some theory Peter Goos / 40 Optimal design theory continuous or approximate optimal designs implicitly assume an infinitely large number of observations are available
More informationNumerical Methods I Solving Nonlinear Equations
Numerical Methods I Solving Nonlinear Equations Aleksandar Donev Courant Institute, NYU 1 donev@courant.nyu.edu 1 MATH-GA 2011.003 / CSCI-GA 2945.003, Fall 2014 October 16th, 2014 A. Donev (Courant Institute)
More informationEfficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies
Efficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies Tim Holland-Letz Ruhr-Universität Bochum Medizinische Fakultät 44780 Bochum, Germany email: tim.holland-letz@rub.de
More informationA Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints
Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Mai Zhou Yifan Yang Received: date / Accepted: date Abstract In this note
More informationLast updated: Oct 22, 2012 LINEAR CLASSIFIERS. J. Elder CSE 4404/5327 Introduction to Machine Learning and Pattern Recognition
Last updated: Oct 22, 2012 LINEAR CLASSIFIERS Problems 2 Please do Problem 8.3 in the textbook. We will discuss this in class. Classification: Problem Statement 3 In regression, we are modeling the relationship
More informationD-optimal Designs for Multinomial Logistic Models
D-optimal Designs for Multinomial Logistic Models Jie Yang University of Illinois at Chicago Joint with Xianwei Bu and Dibyen Majumdar October 12, 2017 1 Multinomial Logistic Models Cumulative logit model:
More informationAppendix 8 Numerical Methods for Solving Nonlinear Equations 1
Appendix 8 Numerical Methods for Solving Nonlinear Equations 1 An equation is said to be nonlinear when it involves terms of degree higher than 1 in the unknown quantity. These terms may be polynomial
More informationBiostat 2065 Analysis of Incomplete Data
Biostat 2065 Analysis of Incomplete Data Gong Tang Dept of Biostatistics University of Pittsburgh October 20, 2005 1. Large-sample inference based on ML Let θ is the MLE, then the large-sample theory implies
More informationStreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory
StreamSVM Linear SVMs and Logistic Regression When Data Does Not Fit In Memory S.V. N. (vishy) Vishwanathan Purdue University and Microsoft vishy@purdue.edu October 9, 2012 S.V. N. Vishwanathan (Purdue,
More informationOPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES
Statistica Sinica 21 (2011, 1415-1430 OPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES Min Yang, Bin Zhang and Shuguang Huang University of Missouri, University of Alabama-Birmingham
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Regression based methods, 1st part: Introduction (Sec.
More informationLikelihood, MLE & EM for Gaussian Mixture Clustering. Nick Duffield Texas A&M University
Likelihood, MLE & EM for Gaussian Mixture Clustering Nick Duffield Texas A&M University Probability vs. Likelihood Probability: predict unknown outcomes based on known parameters: P(x q) Likelihood: estimate
More information1. Nonlinear Equations. This lecture note excerpted parts from Michael Heath and Max Gunzburger. f(x) = 0
Numerical Analysis 1 1. Nonlinear Equations This lecture note excerpted parts from Michael Heath and Max Gunzburger. Given function f, we seek value x for which where f : D R n R n is nonlinear. f(x) =
More informationAn Inexact Newton Method for Nonlinear Constrained Optimization
An Inexact Newton Method for Nonlinear Constrained Optimization Frank E. Curtis Numerical Analysis Seminar, January 23, 2009 Outline Motivation and background Algorithm development and theoretical results
More informationD-optimal Designs for Factorial Experiments under Generalized Linear Models
D-optimal Designs for Factorial Experiments under Generalized Linear Models Jie Yang Department of Mathematics, Statistics, and Computer Science University of Illinois at Chicago Joint research with Abhyuday
More informationResearch Article A Novel Differential Evolution Invasive Weed Optimization Algorithm for Solving Nonlinear Equations Systems
Journal of Applied Mathematics Volume 2013, Article ID 757391, 18 pages http://dx.doi.org/10.1155/2013/757391 Research Article A Novel Differential Evolution Invasive Weed Optimization for Solving Nonlinear
More informationPROBABILITY AND STATISTICS Vol. III - Statistical Experiments and Optimal Design - Andrej Pázman STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN
STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN Andrej Pázman Comenius University, Bratislava, Slovakia Keywords: Experiment design, linear statistical model, nonlinear regression, least squares, information
More informationSOLUTION OF ALGEBRAIC AND TRANSCENDENTAL EQUATIONS BISECTION METHOD
BISECTION METHOD If a function f(x) is continuous between a and b, and f(a) and f(b) are of opposite signs, then there exists at least one root between a and b. It is shown graphically as, Let f a be negative
More informationMachine Learning 2017
Machine Learning 2017 Volker Roth Department of Mathematics & Computer Science University of Basel 21st March 2017 Volker Roth (University of Basel) Machine Learning 2017 21st March 2017 1 / 41 Section
More informationLecture 3. Inference about multivariate normal distribution
Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates
More informationGibbs Sampling in Linear Models #2
Gibbs Sampling in Linear Models #2 Econ 690 Purdue University Outline 1 Linear Regression Model with a Changepoint Example with Temperature Data 2 The Seemingly Unrelated Regressions Model 3 Gibbs sampling
More informationRandom Matrix Eigenvalue Problems in Probabilistic Structural Mechanics
Random Matrix Eigenvalue Problems in Probabilistic Structural Mechanics S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html
More informationMultiplicative Algorithm for computing Optimum Designs
for computing s DAE Conference 2012 Roberto Dorta-Guerra, Raúl Martín Martín, Mercedes Fernández-Guerrero and Licesio J. Rodríguez-Aragón Optimum Experimental Design group October 20, 2012 http://areaestadistica.uclm.es/oed
More informationMatrix Derivatives and Descent Optimization Methods
Matrix Derivatives and Descent Optimization Methods 1 Qiang Ning Department of Electrical and Computer Engineering Beckman Institute for Advanced Science and Techonology University of Illinois at Urbana-Champaign
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationLogistic Regression. Seungjin Choi
Logistic Regression Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/
More informationBasic math for biology
Basic math for biology Lei Li Florida State University, Feb 6, 2002 The EM algorithm: setup Parametric models: {P θ }. Data: full data (Y, X); partial data Y. Missing data: X. Likelihood and maximum likelihood
More informationInexact Newton Methods and Nonlinear Constrained Optimization
Inexact Newton Methods and Nonlinear Constrained Optimization Frank E. Curtis EPSRC Symposium Capstone Conference Warwick Mathematics Institute July 2, 2009 Outline PDE-Constrained Optimization Newton
More informationSTAT 730 Chapter 4: Estimation
STAT 730 Chapter 4: Estimation Timothy Hanson Department of Statistics, University of South Carolina Stat 730: Multivariate Analysis 1 / 23 The likelihood We have iid data, at least initially. Each datum
More informationLinear Regression and Its Applications
Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start
More informationSpring 2012 Math 541B Exam 1
Spring 2012 Math 541B Exam 1 1. A sample of size n is drawn without replacement from an urn containing N balls, m of which are red and N m are black; the balls are otherwise indistinguishable. Let X denote
More informationExponential families also behave nicely under conditioning. Specifically, suppose we write η = (η 1, η 2 ) R k R p k so that
1 More examples 1.1 Exponential families under conditioning Exponential families also behave nicely under conditioning. Specifically, suppose we write η = η 1, η 2 R k R p k so that dp η dm 0 = e ηt 1
More informationEfficient Solvers for Stochastic Finite Element Saddle Point Problems
Efficient Solvers for Stochastic Finite Element Saddle Point Problems Catherine E. Powell c.powell@manchester.ac.uk School of Mathematics University of Manchester, UK Efficient Solvers for Stochastic Finite
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 05 Full points may be obtained for correct answers to eight questions Each numbered question (which may have several parts) is worth
More informationOptimal Designs for 2 k Experiments with Binary Response
1 / 57 Optimal Designs for 2 k Experiments with Binary Response Dibyen Majumdar Mathematics, Statistics, and Computer Science College of Liberal Arts and Sciences University of Illinois at Chicago Joint
More information1 One-way analysis of variance
LIST OF FORMULAS (Version from 21. November 2014) STK2120 1 One-way analysis of variance Assume X ij = µ+α i +ɛ ij ; j = 1, 2,..., J i ; i = 1, 2,..., I ; where ɛ ij -s are independent and N(0, σ 2 ) distributed.
More informationStochastic Design Criteria in Linear Models
AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework
More informationInformation in a Two-Stage Adaptive Optimal Design
Information in a Two-Stage Adaptive Optimal Design Department of Statistics, University of Missouri Designed Experiments: Recent Advances in Methods and Applications DEMA 2011 Isaac Newton Institute for
More informationOptimal design of experiments
Optimal design of experiments Session 5: Design construction Peter Goos 1 / 33 Constructing exact optimal designs direct approaches to maximize X T X Microsoft Excel mathematical programming run into problems
More informationPerformance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project
Performance Comparison of K-Means and Expectation Maximization with Gaussian Mixture Models for Clustering EE6540 Final Project Devin Cornell & Sushruth Sastry May 2015 1 Abstract In this article, we explore
More informationCOURSE Iterative methods for solving linear systems
COURSE 0 4.3. Iterative methods for solving linear systems Because of round-off errors, direct methods become less efficient than iterative methods for large systems (>00 000 variables). An iterative scheme
More informationBayesian estimation of the discrepancy with misspecified parametric models
Bayesian estimation of the discrepancy with misspecified parametric models Pierpaolo De Blasi University of Torino & Collegio Carlo Alberto Bayesian Nonparametrics workshop ICERM, 17-21 September 2012
More informationComputer Intensive Methods in Mathematical Statistics
Computer Intensive Methods in Mathematical Statistics Department of mathematics johawes@kth.se Lecture 16 Advanced topics in computational statistics 18 May 2017 Computer Intensive Methods (1) Plan of
More informationEstimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1
Estimation and Model Selection in Mixed Effects Models Part I Adeline Samson 1 1 University Paris Descartes Summer school 2009 - Lipari, Italy These slides are based on Marc Lavielle s slides Outline 1
More informationEfficient Bayesian Multivariate Surface Regression
Efficient Bayesian Multivariate Surface Regression Feng Li feng.li@cufe.edu.cn School of Statistics and Mathematics Central University of Finance and Economics Outline of the talk 1 Introduction to flexible
More informationInterior Point Algorithms for Constrained Convex Optimization
Interior Point Algorithms for Constrained Convex Optimization Chee Wei Tan CS 8292 : Advanced Topics in Convex Optimization and its Applications Fall 2010 Outline Inequality constrained minimization problems
More informationSparse Approximation via Penalty Decomposition Methods
Sparse Approximation via Penalty Decomposition Methods Zhaosong Lu Yong Zhang February 19, 2012 Abstract In this paper we consider sparse approximation problems, that is, general l 0 minimization problems
More informationDescribing Contingency tables
Today s topics: Describing Contingency tables 1. Probability structure for contingency tables (distributions, sensitivity/specificity, sampling schemes). 2. Comparing two proportions (relative risk, odds
More informationLINEAR MODELS FOR CLASSIFICATION. J. Elder CSE 6390/PSYC 6225 Computational Modeling of Visual Perception
LINEAR MODELS FOR CLASSIFICATION Classification: Problem Statement 2 In regression, we are modeling the relationship between a continuous input variable x and a continuous target variable t. In classification,
More informationRandom Eigenvalue Problems Revisited
Random Eigenvalue Problems Revisited S Adhikari Department of Aerospace Engineering, University of Bristol, Bristol, U.K. Email: S.Adhikari@bristol.ac.uk URL: http://www.aer.bris.ac.uk/contact/academic/adhikari/home.html
More informationLecture 20: Linear model, the LSE, and UMVUE
Lecture 20: Linear model, the LSE, and UMVUE Linear Models One of the most useful statistical models is X i = β τ Z i + ε i, i = 1,...,n, where X i is the ith observation and is often called the ith response;
More informationL11: Pattern recognition principles
L11: Pattern recognition principles Bayesian decision theory Statistical classifiers Dimensionality reduction Clustering This lecture is partly based on [Huang, Acero and Hon, 2001, ch. 4] Introduction
More informationAGEC 661 Note Eleven Ximing Wu. Exponential regression model: m (x, θ) = exp (xθ) for y 0
AGEC 661 ote Eleven Ximing Wu M-estimator So far we ve focused on linear models, where the estimators have a closed form solution. If the population model is nonlinear, the estimators often do not have
More informationOptimal discrimination designs
Optimal discrimination designs Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@ruhr-uni-bochum.de Stefanie Titoff Ruhr-Universität Bochum Fakultät
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationLeast Squares Approximation
Chapter 6 Least Squares Approximation As we saw in Chapter 5 we can interpret radial basis function interpolation as a constrained optimization problem. We now take this point of view again, but start
More informationGaussian Process Approximations of Stochastic Differential Equations
Gaussian Process Approximations of Stochastic Differential Equations Cédric Archambeau Dan Cawford Manfred Opper John Shawe-Taylor May, 2006 1 Introduction Some of the most complex models routinely run
More informationGaussian mean-shift algorithms
Gaussian mean-shift algorithms Miguel Á. Carreira-Perpiñán Dept. of Computer Science & Electrical Engineering, OGI/OHSU http://www.csee.ogi.edu/~miguel Gaussian mean-shift (GMS) as an EM algorithm Gaussian
More informationA geometric characterization of c-optimal designs for heteroscedastic regression
A geometric characterization of c-optimal designs for heteroscedastic regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Tim Holland-Letz
More informationImage restoration: numerical optimisation
Image restoration: numerical optimisation Short and partial presentation Jean-François Giovannelli Groupe Signal Image Laboratoire de l Intégration du Matériau au Système Univ. Bordeaux CNRS BINP / 6 Context
More informationMinimax design criterion for fractional factorial designs
Ann Inst Stat Math 205 67:673 685 DOI 0.007/s0463-04-0470-0 Minimax design criterion for fractional factorial designs Yue Yin Julie Zhou Received: 2 November 203 / Revised: 5 March 204 / Published online:
More informationA Study on Minimal-point Composite Designs
A Study on Minimal-point Composite Designs by Yin-Jie Huang Advisor Ray-Bing Chen Institute of Statistics, National University of Kaohsiung Kaohsiung, Taiwan 8 R.O.C. July 007 Contents Introduction Construction
More informationPOLI 8501 Introduction to Maximum Likelihood Estimation
POLI 8501 Introduction to Maximum Likelihood Estimation Maximum Likelihood Intuition Consider a model that looks like this: Y i N(µ, σ 2 ) So: E(Y ) = µ V ar(y ) = σ 2 Suppose you have some data on Y,
More informationCOMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure)
STAT 52 Completely Randomized Designs COMPLETELY RANDOMIZED DESIGNS (CRD) For now, t unstructured treatments (e.g. no factorial structure) Completely randomized means no restrictions on the randomization
More informationLeast squares problems
Least squares problems How to state and solve them, then evaluate their solutions Stéphane Mottelet Université de Technologie de Compiègne 30 septembre 2016 Stéphane Mottelet (UTC) Least squares 1 / 55
More informationConstruction of designs for twophase experiments with Tabu search
Construction of designs for twophase experiments with Tabu search Nha Vo-Thanh In collaboration with Professor Hans-Peter Piepho University of Hohenheim Common bean experiment A breeding program by CIAT
More informationEM Algorithm II. September 11, 2018
EM Algorithm II September 11, 2018 Review EM 1/27 (Y obs, Y mis ) f (y obs, y mis θ), we observe Y obs but not Y mis Complete-data log likelihood: l C (θ Y obs, Y mis ) = log { f (Y obs, Y mis θ) Observed-data
More informationNonlinear Error Correction Model and Multiple-Threshold Cointegration May 23, / 31
Nonlinear Error Correction Model and Multiple-Threshold Cointegration Man Wang Dong Hua University, China Joint work with N.H.Chan May 23, 2014 Nonlinear Error Correction Model and Multiple-Threshold Cointegration
More informationThe consequences of misspecifying the random effects distribution when fitting generalized linear mixed models
The consequences of misspecifying the random effects distribution when fitting generalized linear mixed models John M. Neuhaus Charles E. McCulloch Division of Biostatistics University of California, San
More informationEasy and not so easy multifacility location problems... (In 20 minutes.)
Easy and not so easy multifacility location problems... (In 20 minutes.) MINLP 2014 Pittsburgh, June 2014 Justo Puerto Institute of Mathematics (IMUS) Universidad de Sevilla Outline 1 Introduction (In
More informationPHYS 410/555 Computational Physics Solution of Non Linear Equations (a.k.a. Root Finding) (Reference Numerical Recipes, 9.0, 9.1, 9.
PHYS 410/555 Computational Physics Solution of Non Linear Equations (a.k.a. Root Finding) (Reference Numerical Recipes, 9.0, 9.1, 9.4) We will consider two cases 1. f(x) = 0 1-dimensional 2. f(x) = 0 d-dimensional
More informationMATH 4211/6211 Optimization Basics of Optimization Problems
MATH 4211/6211 Optimization Basics of Optimization Problems Xiaojing Ye Department of Mathematics & Statistics Georgia State University Xiaojing Ye, Math & Stat, Georgia State University 0 A standard minimization
More informationLecture 7 Unconstrained nonlinear programming
Lecture 7 Unconstrained nonlinear programming Weinan E 1,2 and Tiejun Li 2 1 Department of Mathematics, Princeton University, weinan@princeton.edu 2 School of Mathematical Sciences, Peking University,
More informationMahdi karbasian* & Zoubi Ibrahim
International Journal of Industrial Engineering & Production Research (010) pp. 105-110 September 010, Volume 1, Number International Journal of Industrial Engineering & Production Research ISSN: 008-4889
More informationLogistic Regression and Generalized Linear Models
Logistic Regression and Generalized Linear Models Sridhar Mahadevan mahadeva@cs.umass.edu University of Massachusetts Sridhar Mahadevan: CMPSCI 689 p. 1/2 Topics Generative vs. Discriminative models In
More informationStatistical Inference of Covariate-Adjusted Randomized Experiments
1 Statistical Inference of Covariate-Adjusted Randomized Experiments Feifang Hu Department of Statistics George Washington University Joint research with Wei Ma, Yichen Qin and Yang Li Email: feifang@gwu.edu
More informationIntroduction to gradient descent
6-1: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction to gradient descent Derivation and intuitions Hessian 6-2: Introduction to gradient descent Prof. J.C. Kao, UCLA Introduction Our
More information11 a 12 a 21 a 11 a 22 a 12 a 21. (C.11) A = The determinant of a product of two matrices is given by AB = A B 1 1 = (C.13) and similarly.
C PROPERTIES OF MATRICES 697 to whether the permutation i 1 i 2 i N is even or odd, respectively Note that I =1 Thus, for a 2 2 matrix, the determinant takes the form A = a 11 a 12 = a a 21 a 11 a 22 a
More informationLASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape
LASSO-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape Nikolaus Umlauf https://eeecon.uibk.ac.at/~umlauf/ Overview Joint work with Andreas Groll, Julien Hambuckers
More informationApplications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices
Applications of Randomized Methods for Decomposing and Simulating from Large Covariance Matrices Vahid Dehdari and Clayton V. Deutsch Geostatistical modeling involves many variables and many locations.
More informationWeak convergence. Amsterdam, 13 November Leiden University. Limit theorems. Shota Gugushvili. Generalities. Criteria
Weak Leiden University Amsterdam, 13 November 2013 Outline 1 2 3 4 5 6 7 Definition Definition Let µ, µ 1, µ 2,... be probability measures on (R, B). It is said that µ n converges weakly to µ, and we then
More informationStochastic Analogues to Deterministic Optimizers
Stochastic Analogues to Deterministic Optimizers ISMP 2018 Bordeaux, France Vivak Patel Presented by: Mihai Anitescu July 6, 2018 1 Apology I apologize for not being here to give this talk myself. I injured
More informationA nonlinear equation is any equation of the form. f(x) = 0. A nonlinear equation can have any number of solutions (finite, countable, uncountable)
Nonlinear equations Definition A nonlinear equation is any equation of the form where f is a nonlinear function. Nonlinear equations x 2 + x + 1 = 0 (f : R R) f(x) = 0 (x cos y, 2y sin x) = (0, 0) (f :
More information