Optimal experimental design, an introduction, Jesús López Fidalgo

Size: px
Start display at page:

Download "Optimal experimental design, an introduction, Jesús López Fidalgo"

Transcription

1 Optimal experimental design, an introduction University of Castilla-La Mancha Department of Mathematics Institute of Applied Mathematics to Science and Engineering

2 Books (just a few) Fedorov V.V. (1972). Theory of optimal experiments. Academic press. New York. Pazman A. (1986). Foundations of Optimum Experimental Design. D. Reidel Pub. New York. Pukelsheim F. (1993). Optimal Design of Experiments. John Wiley & Sons. New York. Fedorov V.V. and Hackl P. (1997). Model-oriented design of experiments. Springer. New York. Atkinson A.C., Doney A.N., Tobias R. (2007). Optimum Experimental Designs, With SAS. Oxford science publications. Oxford.

3 OPTIMAL EXPERIMENTAL DESIGN THEORY

4 Concepts and notation Process: Select a model, E(y x) = f T (x)θ, Var(y x) = σ 2 (linear model). f continuous with components linearly independent. Select an experimental condition x on a design space χ (typically a compact on a Euclidian space). Examples: y = α0 + α 1x + ε, f (x) = (1, x) T, x χ = [a, b]. y = α1x 1 + α 2x 2 + α 3x 3 + ε, f (x) = x = (x 1, x 2, x 3) T, x χ = [a 1, b 1] [a 2, b 2] [a 3, b 3]. y = α0 + α 1x + α 2x 2 + ε, f (x) = (1, x, x 2 ) T, x χ = [a, b]. ANOVA (1-way, 3 levels): y = β 0 + β 1x 1 + β 2x 2 + ε, f (x 1, x 2) = (1, x 1, x 2) T, x = (x 1, x 2) χ = {(0, 1), (1, 0), (1, 1)}

5 Process Perform the experiment and observe the response (r.v. y). Repeat the process for the experimental conditions x 1,..., x n obtaining responses y 1,..., y n. Estimate the parameters (MLE) θ 1,..., θ m, σ 2 and make inferences.

6 EXPERIMENTAL DESIGN Aim: Choosing x 1,..., x n (exact design of size n). Some may be repeated (probability measure): ξ n (x) = nx n, where n x = times x appears in the design. Covariance matrix: Σˆθ = σ2 (X T X ) 1 = σ 2 n 1 M 1 (ξ), M(ξ) = x χ f (x)f T (x)ξ n (x) (information matrix): Symmetric. Non-definite positive. Singular if the number of different points in the design k < m.

7 Extended concept of experimental design Approximate design (Kiefer): probability measure ζ on the Borel field generated by the open sets of χ. Information matrix: M(ζ) = χ f (x)f T (x)ζ(dx). Convex set of approximate designs: Ξ. If ζ is discrete and ξ the probability function associated: M(ξ) = f (x)f T (x)ξ(x) x χ If ζ is absolutely continuous there exists a density function (pdf), ξ, associated: M(ξ) = f (x)f T (x)ξ(x)dx Compact and convex set of information matrices: M χ

8 Caratheodory s theorem Caratheodory s theorem: For any information matrix there is always a design with at most 1 2m(m + 1) + 1 different points. Finite support, { } x1 x ξ = 2... x k p 1 p 2... p k Restricting Ξ to finite designs M is still the same (Caratheodory, definition of integral and continuity). In practice, perform n i nξ(x i ) experiments at x i, i n i = n. From now on we will use ξ or ζ as convenience.

9 Example y = α 0 + α 1 x + ε, x χ = [0, 1], f (x) = (1, x) T. ξ = { }. M(ξ) = = = 3 ( 1 xi i=1 x i x 2 i ) p i ( ) ( ) ( ) ( ) 0.4

10 Example ξ = { If just n = 12 experiments may be performed, possible rounding off: }. about = 2.4 (2, 3, 3) experiments at 0, about = 4.8 (5, 4, 5) experiments at 0.5, about = 4.8 (5, 5, 4) experiments at 1. Remark: It is not as simple as the usual rounding off.

11 CRITERION FUNCTION Φ : Ξ [0, + ) or Φ : M [0, + ) to be minimized (maximized). The first is more general. The second has to be considered as a function of M 1, has to be non-increasing (better estimates of the parameters): M N Φ(M) Φ(N), may be global or partial (interest on part of the parameters), may be positive homogeneous: Φ(δM) = 1 Φ(M), δ > 0, δ M is within the cone of the non-negative definite matrices in the Euclidean space of the symmetric matrices. Thus, the usual concept of differentiability applies here. A Φ optimal design ζ minimizes Φ. Convex: Φ[(1 ɛ)ζ + ɛζ ] (1 ɛ)φ(ζ) + ɛφ(ζ ).

12 Equivalence theorem for differentiable criteria Sensitive function: ψ(x, ξ) = Φ[M(ξ), f (x)f T (x)] = f T (x) Φ[M(ξ)]f (x) trm(ξ) Φ[M(ξ)]. Equivalence theorem: ξ is Φ optimal if and only if ψ(x, ξ ) 0, x χ. Equality for x S ξ.

13 Comments on the equivalence theorem Just for approximate designs. M is convex and it is the same for finite or general designs. From now on we assume finite designs and use the symbol ξ (pdf) instead of ζ (measure). Still valid for a restricted search in a convex subset.

14 Checking condition ψ(z, ξ ) = 0, z S ξ ( ψ(x, ξ ) ) x x=z E.g., for a two parameter model assume ξ = then there are two equations: = 0, z S ξ Int(χ) { x1 x 2 1 p p }, ψ(x 1, ξ) = 0, ψ(x 2, ξ) = 0 and 0, 1 or 2 of the equations: ( ) ( ) ψ(x, ξ) ψ(x, ξ) = 0, = 0 x x=x 1 x x=x 2 to be solved.

15 EFFICIENCY (goodness of a design) eff Φ (ξ) = Φ[M(ξ )] Φ[M(ξ)]. If Φ homogenous, e.g. for 60% efficiency and n observations: 0.6 = Φ[M(ξ )] Φ[M(ξ)] = Φ(σ 2 nσ 1 ) ˆθ Φ(σ 2 nσ 1 ˆθ 1 ) ˆθ ) = σ2n 1 Φ(Σ σ 2 n 1 Φ(Σ 1 ) = Φ(Σ Φ(Σ 1 ˆθ ). For n observations with ξ and n with ξ the efficiency is 1 if: 1 = Φ(Σ 1 ) ˆθ Φ(Σ 1 ˆθ ) = σ2n 1Φ[M(ξ )] σ 2 n 1 Φ[M(ξ)] ˆθ = σ2 n 1 σ 2 0.6, n 1 n = 0.6n (60%) observations would be enough with ξ. 1 ) ˆθ

16 General properties for global criteria A Φ optimal design has m points at least: Any design with less than m points gives a singular matrix. For strictly convex criteria there is always a Φ optimal design with no more than m(m+1) 2 points: The optimal must be in the boundary (dimension m(m+1) 1). 2 Caratheodory in the boundary gives a limit of m(m+1) points. 2 If σ 2 (x) is not constant, the whole theory is applicable for f (x) = f (x)/σ(x). If the observations are correlated (Σ y ), just exact designs. Information matrix: X T Σ 1 y X.

17 D OPTIMALITY Determinant of the inverse: { log det M Φ D [M(ξ)] = 1 (ξ) = log det M(ξ) if det M(ξ) 0, if det M(ξ) = 0. Homogeneous definition: { det M Φ D [M(ξ)] = 1/m (ξ) if det M(ξ) 0, if det M(ξ) = 0.

18 Properties Φ D continuous on M. Convex on M and strictly convex on M + (nonsingular information matrices). Differentiable on M + : ( log det M) = M 1 Minimum volume of the confidence elipsoide: ( ˆΘ Θ) T (X T X ) 1 ( ˆΘ Θ) mf m,n m,γ S 2 R c 2. Volume =c m V m [det M 1 (ξ)] 1/2, where V m is the volume of the m dimensional sphere.

19 G optimality Kirstine Smith (1918). Kiefer & Wolfovitz gave the name. Φ G [M(ξ)] = { maxx χ f T (x)m 1 (ξ)f (x) if det M(ξ) 0, if det M(ξ) = 0. Φ G continuous and convex on M. Φ G [M(ξ)] max x χ var ξ ŷ(x).

20 A optimality Trace of the inverse: { } trm Φ A [M(ξ)] = 1 (ξ) if det M(ξ) 0 if det M(ξ) = 0, The second equality is a consequence of if the parameters are estimable. Φ A continuous on M. var ξ ˆθi e t i M 1 (ξ)e i Convex on M and strictly convex on M +. Differentiable on M + : (trm 1 (ξ)) = M 2 (ξ) m var ξ ˆθi. i=1

21 L optimality Atwood (1976): Φ L [M(ξ)] = { trwm 1 (ξ) if det M(ξ) 0, if det M(ξ) = 0, where W is a positive definite matrix of dimension m. Φ L continuous on M. Convex on M and strictly convex on M +. Differentiable on M + : [trwm 1 (ξ)] = M 1 (ξ)wm 1 (ξ).

22 c optimality Variance of the estimator of c T θ, linear combination of the parameters. Φ c [M(ξ)] = c T M (ξ)c.

23 Elfving procedure in practice Plot the curve x(t) = f 1 (t), y(t) = f 2 (t), t χ and its symmetric through origin. Plot the convex hull of both curves. Plot the line through the origin defined by vector c. Optimal design defined by the boundary point.

24 Elfving plot (One point design) { t } 1

25 Elfving plot (Two point design) { } a t 1 p p

26 Elfving procedure for three parameters y = θ 0 + θ 1 t + θ 2 t 2 + ε, t [0, 1] One, two or three design points.

27 Equivalence theorem for D and G optimality ξ is D optimal iff it is G optimal: where d(x, ξ) = f T (x)m 1 (ξ)f (x). Efficiency bound: eff D [M(ξ)] 2 max d(x, x χ ξ ) = m, maxx d(x,ξ) m

28 Example: Checking condition for a quadratic model E[y] = θ T f (x) = θ 1 + θ 2 x + θ 3 x 2, σ 2 (x) = 1, x χ = [ 1, 2], f (x) = (1, x, x 2 ) T. M(ξ) = = x [ 1,2] x [ 1,2] f (x)f T (x)ξ(x) ξ(x) 1 x x 2 x x 2 x 3 x 2 x 3 x 4 Assume the D-optimal design is of the type: { } 1 z 2 ξ = 1/3 1/3 1/3,

29 Quadratic model Maximize the determinant: det M(ξ) = 1 3 ( z 2 + z + 2 ) 2. The maximum is reached at z = 0.5: { ξ = 1/3 1/3 1/3 Checking whether ξ is the D-optimal: d(x, ξ ) = f T (x)m 1 (ξ )f (x) }, = 0.89 ( x x ) ( x x )

30 Example: Elfving procedure E[y] = θ T f (x) = θ 1 x + θ 2 x 2, σ 2 (x) = 1, x χ = [0, 1], f (x) = (1, x, x 2 ) T. Plot the curve x(t) = t, y(t) = t 2, t [0, 1]:

31 Symmetric through the origin

32 Convex hull Tangential point, z: x(z) x 0 x (z) = y(z) y 0 y (z) z = z2 + 1, z = 2 1 2z

33 Line through the origin defined by vector c = (0.5, 0.3) (λ0.5) 2 = λ0.3, t = 3/5, ξ c = { 3/5 1 }

34 Line through the origin defined by vector c = (0.5, 1) Cutting point: { y = 2x, x = 1 4 ( 2 2 x 1 z+1 = y 1 z 2 +1, ) = 0.15, y = 0.30.

35 Weights Convex combination: (0.15, 0.30) = (1 p)( z, z 2 ) + p(1, 1), p = 1 4 ( 3 ) 2 = 0.30 { ξc = ( ) ( ) }.

36 THANK YOU

Optimal experimental designs applied to Biostatistics.

Optimal experimental designs applied to Biostatistics. Optimal experimental designs applied to Biostatistics. Thesis for the Degree of Doctor of Mathematics Santiago Campos Barreiro Academic advisor: Prof. Dr. Jesús López Fidalgo Institute of Mathematics Applied

More information

Approximate Optimal Designs for Multivariate Polynomial Regression

Approximate Optimal Designs for Multivariate Polynomial Regression Approximate Optimal Designs for Multivariate Polynomial Regression Fabrice Gamboa Collaboration with: Yohan de Castro, Didier Henrion, Roxana Hess, Jean-Bernard Lasserre Universität Potsdam 16th of February

More information

Optimal Design in Regression. and Spline Smoothing

Optimal Design in Regression. and Spline Smoothing Optimal Design in Regression and Spline Smoothing by Jaerin Cho A thesis submitted to the Department of Mathematics and Statistics in conformity with the requirements for the degree of Doctor of Philosophy

More information

ON THE REGULARIZATION OF SINGULAR C-OPTIMAL

ON THE REGULARIZATION OF SINGULAR C-OPTIMAL LABORATOIRE INFORMATIQUE, SIGNAUX ET SYSTÈMES DE SOPHIA ANTIPOLIS UMR 6070 ON THE REGULARIZATION OF SINGULAR C-OPTIMAL DESIGNS Luc Pronzato Equipe SYSTEMES Rapport de recherche ISRN I3S/RR 2008-15 FR Septembre

More information

PROBABILITY AND STATISTICS Vol. III - Statistical Experiments and Optimal Design - Andrej Pázman STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN

PROBABILITY AND STATISTICS Vol. III - Statistical Experiments and Optimal Design - Andrej Pázman STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN Andrej Pázman Comenius University, Bratislava, Slovakia Keywords: Experiment design, linear statistical model, nonlinear regression, least squares, information

More information

E-optimal approximate block designs for treatment-control comparisons

E-optimal approximate block designs for treatment-control comparisons E-optimal approximate block designs for treatment-control comparisons Samuel Rosa 1 1 Faculty of Mathematics, Physics and Informatics, Comenius University in Bratislava, Slovakia We study E-optimal block

More information

Optimal designs for estimating the slope of a regression

Optimal designs for estimating the slope of a regression Optimal designs for estimating the slope of a regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 4480 Bochum, Germany e-mail: holger.dette@rub.de Viatcheslav B. Melas St. Petersburg

More information

arxiv: v1 [stat.co] 7 Sep 2017

arxiv: v1 [stat.co] 7 Sep 2017 Computing optimal experimental designs with respect to a compound Bayes risk criterion arxiv:1709.02317v1 [stat.co] 7 Sep 2017 Radoslav Harman, and Maryna Prus September 8, 2017 Abstract We consider the

More information

A new algorithm for deriving optimal designs

A new algorithm for deriving optimal designs A new algorithm for deriving optimal designs Stefanie Biedermann, University of Southampton, UK Joint work with Min Yang, University of Illinois at Chicago 18 October 2012, DAE, University of Georgia,

More information

Optimal Design. By Anthony C. Atkinson. 1 Introduction

Optimal Design. By Anthony C. Atkinson. 1 Introduction By Anthony C. Atkinson Keywords: algorithms, blocking, D-optimal design, equivalence theorem, I-optimal design, mixture design, nonregular design region, prediction variance, response-surface design Abstract:

More information

Optimal design of experiments

Optimal design of experiments Optimal design of experiments Session 4: Some theory Peter Goos / 40 Optimal design theory continuous or approximate optimal designs implicitly assume an infinitely large number of observations are available

More information

A Study on Minimal-point Composite Designs

A Study on Minimal-point Composite Designs A Study on Minimal-point Composite Designs by Yin-Jie Huang Advisor Ray-Bing Chen Institute of Statistics, National University of Kaohsiung Kaohsiung, Taiwan 8 R.O.C. July 007 Contents Introduction Construction

More information

On construction of constrained optimum designs

On construction of constrained optimum designs On construction of constrained optimum designs Institute of Control and Computation Engineering University of Zielona Góra, Poland DEMA2008, Cambridge, 15 August 2008 Numerical algorithms to construct

More information

Stochastic Design Criteria in Linear Models

Stochastic Design Criteria in Linear Models AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework

More information

Bayesian Inference. Chapter 9. Linear models and regression

Bayesian Inference. Chapter 9. Linear models and regression Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering

More information

Model Selection and Geometry

Model Selection and Geometry Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model

More information

Ma 3/103: Lecture 24 Linear Regression I: Estimation

Ma 3/103: Lecture 24 Linear Regression I: Estimation Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the

More information

ECON 4117/5111 Mathematical Economics

ECON 4117/5111 Mathematical Economics Test 1 September 29, 2006 1. Use a truth table to show the following equivalence statement: (p q) (p q) 2. Consider the statement: A function f : X Y is continuous on X if for every open set V Y, the pre-image

More information

ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE

ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE Sankhyā : The Indian Journal of Statistics 999, Volume 6, Series B, Pt. 3, pp. 488 495 ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE By S. HUDA and A.A. AL-SHIHA King Saud University, Riyadh, Saudi Arabia

More information

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed

Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,

More information

Multiplicative Algorithm for computing Optimum Designs

Multiplicative Algorithm for computing Optimum Designs for computing s DAE Conference 2012 Roberto Dorta-Guerra, Raúl Martín Martín, Mercedes Fernández-Guerrero and Licesio J. Rodríguez-Aragón Optimum Experimental Design group October 20, 2012 http://areaestadistica.uclm.es/oed

More information

Frequentist-Bayesian Model Comparisons: A Simple Example

Frequentist-Bayesian Model Comparisons: A Simple Example Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal

More information

Optimal and efficient designs for Gompertz regression models

Optimal and efficient designs for Gompertz regression models Ann Inst Stat Math (2012) 64:945 957 DOI 10.1007/s10463-011-0340-y Optimal and efficient designs for Gompertz regression models Gang Li Received: 13 July 2010 / Revised: 11 August 2011 / Published online:

More information

Discussion Paper SFB 823. Nearly universally optimal designs for models with correlated observations

Discussion Paper SFB 823. Nearly universally optimal designs for models with correlated observations SFB 823 Nearly universally optimal designs for models with correlated observations Discussion Paper Holger Dette, Andrey Pepelyshev, Anatoly Zhigljavsky Nr. 3/212 Nearly universally optimal designs for

More information

Optimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models

Optimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models Optimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models Anthony C. Atkinson, Department of Statistics, London School of Economics, London WC2A 2AE, UK and Barbara Bogacka, School

More information

Randomisation, Replication, Response Surfaces. and. Rosemary

Randomisation, Replication, Response Surfaces. and. Rosemary Randomisation, Replication, Response Surfaces and Rosemary 1 A.C. Atkinson a.c.atkinson@lse.ac.uk Department of Statistics London School of Economics London WC2A 2AE, UK One joint publication RAB AND ME

More information

Optimal designs for estimating the slope in nonlinear regression

Optimal designs for estimating the slope in nonlinear regression Optimal designs for estimating the slope in nonlinear regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Viatcheslav B. Melas St.

More information

LOWELL JOURNAL. MUST APOLOGIZE. such communication with the shore as Is m i Boimhle, noewwary and proper for the comfort

LOWELL JOURNAL. MUST APOLOGIZE. such communication with the shore as Is m i Boimhle, noewwary and proper for the comfort - 7 7 Z 8 q ) V x - X > q - < Y Y X V - z - - - - V - V - q \ - q q < -- V - - - x - - V q > x - x q - x q - x - - - 7 -» - - - - 6 q x - > - - x - - - x- - - q q - V - x - - ( Y q Y7 - >»> - x Y - ] [

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there

More information

A geometric characterization of c-optimal designs for heteroscedastic regression

A geometric characterization of c-optimal designs for heteroscedastic regression A geometric characterization of c-optimal designs for heteroscedastic regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Tim Holland-Letz

More information

Lecture 32: Asymptotic confidence sets and likelihoods

Lecture 32: Asymptotic confidence sets and likelihoods Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence

More information

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator

Estimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were

More information

Orthogonal decompositions in growth curve models

Orthogonal decompositions in growth curve models ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 4, Orthogonal decompositions in growth curve models Daniel Klein and Ivan Žežula Dedicated to Professor L. Kubáček on the occasion

More information

Minimax design criterion for fractional factorial designs

Minimax design criterion for fractional factorial designs Ann Inst Stat Math 205 67:673 685 DOI 0.007/s0463-04-0470-0 Minimax design criterion for fractional factorial designs Yue Yin Julie Zhou Received: 2 November 203 / Revised: 5 March 204 / Published online:

More information

MAS223 Statistical Inference and Modelling Exercises

MAS223 Statistical Inference and Modelling Exercises MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,

More information

The Q Method for Symmetric Cone Programmin

The Q Method for Symmetric Cone Programmin The Q Method for Symmetric Cone Programming The Q Method for Symmetric Cone Programmin Farid Alizadeh and Yu Xia alizadeh@rutcor.rutgers.edu, xiay@optlab.mcma Large Scale Nonlinear and Semidefinite Progra

More information

Multivariable Calculus

Multivariable Calculus 2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)

More information

Designs for Generalized Linear Models

Designs for Generalized Linear Models Designs for Generalized Linear Models Anthony C. Atkinson David C. Woods London School of Economics and Political Science, UK University of Southampton, UK December 9, 2013 Email: a.c.atkinson@lse.ac.uk

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Review (Probability & Linear Algebra)

Review (Probability & Linear Algebra) Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint

More information

A new approach to optimal designs for models with correlated observations

A new approach to optimal designs for models with correlated observations A new approach to optimal designs for models with correlated observations Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Andrey Pepelyshev

More information

Lecture 25: Review. Statistics 104. April 23, Colin Rundel

Lecture 25: Review. Statistics 104. April 23, Colin Rundel Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April

More information

Multivariate Analysis and Likelihood Inference

Multivariate Analysis and Likelihood Inference Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density

More information

where r n = dn+1 x(t)

where r n = dn+1 x(t) Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution

More information

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30

MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30 MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)

More information

conditional cdf, conditional pdf, total probability theorem?

conditional cdf, conditional pdf, total probability theorem? 6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random

More information

D-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors

D-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors Journal of Data Science 920), 39-53 D-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors Chuan-Pin Lee and Mong-Na Lo Huang National Sun Yat-sen University Abstract: Central

More information

MAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS

MAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS Statistica Sinica 17(2007), 463-480 MAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS Holger Dette, Linda M. Haines and Lorens A. Imhof Ruhr-Universität, University of Cape Town and Bonn University

More information

Nonlinear Programming Models

Nonlinear Programming Models Nonlinear Programming Models Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Nonlinear Programming Models p. Introduction Nonlinear Programming Models p. NLP problems minf(x) x S R n Standard form:

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56

Cointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56 Cointegrated VAR s Eduardo Rossi University of Pavia November 2013 Rossi Cointegrated VAR s Financial Econometrics - 2013 1 / 56 VAR y t = (y 1t,..., y nt ) is (n 1) vector. y t VAR(p): Φ(L)y t = ɛ t The

More information

Mathematical Background

Mathematical Background A Mathematical Background Version: date: June 28, 2018 Copyright c 2018 by Nob Hill Publishing, LLC A.1 Introduction In this appendix we give a brief review of some concepts that we need. It is assumed

More information

[y i α βx i ] 2 (2) Q = i=1

[y i α βx i ] 2 (2) Q = i=1 Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

Lecture 7. Econ August 18

Lecture 7. Econ August 18 Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar

More information

Time Series Analysis

Time Series Analysis Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Regression based methods, 1st part: Introduction (Sec.

More information

MA102: Multivariable Calculus

MA102: Multivariable Calculus MA102: Multivariable Calculus Rupam Barman and Shreemayee Bora Department of Mathematics IIT Guwahati Differentiability of f : U R n R m Definition: Let U R n be open. Then f : U R n R m is differentiable

More information

Active Robot Calibration Algorithm

Active Robot Calibration Algorithm Active Robot Calibration Algorithm Yu Sun and John M. Hollerbach Abstract This paper presents a new updating algorithm to reduce the complexity of computing an observability index for kinematic calibration

More information

Asymptotics of minimax stochastic programs

Asymptotics of minimax stochastic programs Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming

More information

Penalized D-optimal design for dose finding

Penalized D-optimal design for dose finding Penalized for dose finding Luc Pronzato Laboratoire I3S, CNRS-Univ. Nice Sophia Antipolis, France 2/ 35 Outline 3/ 35 Bernoulli-type experiments: Y i {0,1} (success or failure) η(x,θ) = Prob(Y i = 1 x

More information

Introduction to Simple Linear Regression

Introduction to Simple Linear Regression Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department

More information

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology

Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna

More information

SCHOOL OF MATHEMATICS AND STATISTICS. Linear and Generalised Linear Models

SCHOOL OF MATHEMATICS AND STATISTICS. Linear and Generalised Linear Models SCHOOL OF MATHEMATICS AND STATISTICS Linear and Generalised Linear Models Autumn Semester 2017 18 2 hours Attempt all the questions. The allocation of marks is shown in brackets. RESTRICTED OPEN BOOK EXAMINATION

More information

Central Limit Theorem ( 5.3)

Central Limit Theorem ( 5.3) Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately

More information

Math 362, Problem set 1

Math 362, Problem set 1 Math 6, roblem set Due //. (4..8) Determine the mean variance of the mean X of a rom sample of size 9 from a distribution having pdf f(x) = 4x, < x

More information

SOLUTION FOR HOMEWORK 4, STAT 4352

SOLUTION FOR HOMEWORK 4, STAT 4352 SOLUTION FOR HOMEWORK 4, STAT 4352 Welcome to your fourth homework. Here we begin the study of confidence intervals, Errors, etc. Recall that X n := (X 1,...,X n ) denotes the vector of n observations.

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1

MA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical

More information

The EG-family of algorithms and procedures for solving linear differential and difference higher-order systems

The EG-family of algorithms and procedures for solving linear differential and difference higher-order systems The EG-family of algorithms and procedures for solving linear differential and difference higher-order systems S. Abramov, D. Khmelnov, A. Ryabenko Computing Centre of the Russian Academy of Sciences Moscow

More information

Optimal designs for enzyme inhibition kinetic models

Optimal designs for enzyme inhibition kinetic models Optimal designs for enzyme inhibition kinetic models Kirsten Schorning, Holger Dette, Katrin Kettelhake, Tilman Möller Fakultät für Mathematik Ruhr-Universität Bochum 44780 Bochum Germany Abstract In this

More information

Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India

Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India CHAPTER 9 BY Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India E-mail : mantusaha.bu@gmail.com Introduction and Objectives In the preceding chapters, we discussed normed

More information

Optimal discrimination designs

Optimal discrimination designs Optimal discrimination designs Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@ruhr-uni-bochum.de Stefanie Titoff Ruhr-Universität Bochum Fakultät

More information

Regression: Lecture 2

Regression: Lecture 2 Regression: Lecture 2 Niels Richard Hansen April 26, 2012 Contents 1 Linear regression and least squares estimation 1 1.1 Distributional results................................ 3 2 Non-linear effects and

More information

VARIATIONAL CALCULUS IN SPACE OF MEASURES AND OPTIMAL DESIGN

VARIATIONAL CALCULUS IN SPACE OF MEASURES AND OPTIMAL DESIGN Chapter 1 VARIATIONAL CALCULUS IN SPACE OF MEASURES AND OPTIMAL DESIGN Ilya Molchanov Department of Statistics University of Glasgow ilya@stats.gla.ac.uk www.stats.gla.ac.uk/ ilya Sergei Zuyev Department

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Statistical Inference of Covariate-Adjusted Randomized Experiments

Statistical Inference of Covariate-Adjusted Randomized Experiments 1 Statistical Inference of Covariate-Adjusted Randomized Experiments Feifang Hu Department of Statistics George Washington University Joint research with Wei Ma, Yichen Qin and Yang Li Email: feifang@gwu.edu

More information

Lecture 6: Linear models and Gauss-Markov theorem

Lecture 6: Linear models and Gauss-Markov theorem Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables

More information

Stat 710: Mathematical Statistics Lecture 12

Stat 710: Mathematical Statistics Lecture 12 Stat 710: Mathematical Statistics Lecture 12 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 12 Feb 18, 2009 1 / 11 Lecture 12:

More information

A Probability Review

A Probability Review A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in

More information

Chapter 5 Matrix Approach to Simple Linear Regression

Chapter 5 Matrix Approach to Simple Linear Regression STAT 525 SPRING 2018 Chapter 5 Matrix Approach to Simple Linear Regression Professor Min Zhang Matrix Collection of elements arranged in rows and columns Elements will be numbers or symbols For example:

More information

Theory of Statistics.

Theory of Statistics. Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased

More information

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings

Structural and Multidisciplinary Optimization. P. Duysinx and P. Tossings Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be

More information

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation

Clustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide

More information

9. Multivariate Linear Time Series (II). MA6622, Ernesto Mordecki, CityU, HK, 2006.

9. Multivariate Linear Time Series (II). MA6622, Ernesto Mordecki, CityU, HK, 2006. 9. Multivariate Linear Time Series (II). MA6622, Ernesto Mordecki, CityU, HK, 2006. References for this Lecture: Introduction to Time Series and Forecasting. P.J. Brockwell and R. A. Davis, Springer Texts

More information

6.3 Forecasting ARMA processes

6.3 Forecasting ARMA processes 6.3. FORECASTING ARMA PROCESSES 123 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss

More information

COHERENT DISPERSION CRITERIA FOR OPTIMAL EXPERIMENTAL DESIGN

COHERENT DISPERSION CRITERIA FOR OPTIMAL EXPERIMENTAL DESIGN The Annals of Statistics 1999, Vol. 27, No. 1, 65 81 COHERENT DISPERSION CRITERIA FOR OPTIMAL EXPERIMENTAL DESIGN By A. Philip Dawid and Paola Sebastiani 1 University College London and The Open University

More information

Mathematical statistics

Mathematical statistics October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter

More information

15.S24 Sample Exam Solutions

15.S24 Sample Exam Solutions 5.S4 Sample Exam Solutions. In each case, determine whether V is a vector space. If it is not a vector space, explain why not. If it is, find basis vectors for V. (a) V is the subset of R 3 defined by

More information

On the efficiency of two-stage adaptive designs

On the efficiency of two-stage adaptive designs On the efficiency of two-stage adaptive designs Björn Bornkamp (Novartis Pharma AG) Based on: Dette, H., Bornkamp, B. and Bretz F. (2010): On the efficiency of adaptive designs www.statistik.tu-dortmund.de/sfb823-dp2010.html

More information

Z-estimators (generalized method of moments)

Z-estimators (generalized method of moments) Z-estimators (generalized method of moments) Consider the estimation of an unknown parameter θ in a set, based on data x = (x,...,x n ) R n. Each function h(x, ) on defines a Z-estimator θ n = θ n (x,...,x

More information

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets

Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Confidence sets X: a sample from a population P P. θ = θ(p): a functional from P to Θ R k for a fixed integer k. C(X): a confidence

More information

ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA. Cross River State, Nigeria

ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA. Cross River State, Nigeria ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA Thomas Adidaume Uge and Stephen Seastian Akpan, Department Of Mathematics/Statistics And Computer

More information

Heteroscedastic T-Optimum Designs for Multiresponse Dynamic Models

Heteroscedastic T-Optimum Designs for Multiresponse Dynamic Models Heteroscedastic T-Optimum Designs for Multiresponse Dynamic Models Dariusz Uciński 1 and Barbara Bogacka 2 1 Institute of Control and Computation Engineering, University of Zielona Góra, ul. Podgórna 50,

More information

16.584: Random Vectors

16.584: Random Vectors 1 16.584: Random Vectors Define X : (X 1, X 2,..X n ) T : n-dimensional Random Vector X 1 : X(t 1 ): May correspond to samples/measurements Generalize definition of PDF: F X (x) = P[X 1 x 1, X 2 x 2,...X

More information

Error analysis for efficiency

Error analysis for efficiency Glen Cowan RHUL Physics 28 July, 2008 Error analysis for efficiency To estimate a selection efficiency using Monte Carlo one typically takes the number of events selected m divided by the number generated

More information

Variance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18

Variance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18 Variance reduction p. 1/18 Variance reduction Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Variance reduction p. 2/18 Example Use simulation to compute I = 1 0 e x dx We

More information

On V-orthogonal projectors associated with a semi-norm

On V-orthogonal projectors associated with a semi-norm On V-orthogonal projectors associated with a semi-norm Short Title: V-orthogonal projectors Yongge Tian a, Yoshio Takane b a School of Economics, Shanghai University of Finance and Economics, Shanghai

More information

Lecture 28: Asymptotic confidence sets

Lecture 28: Asymptotic confidence sets Lecture 28: Asymptotic confidence sets 1 α asymptotic confidence sets Similar to testing hypotheses, in many situations it is difficult to find a confidence set with a given confidence coefficient or level

More information

STA205 Probability: Week 8 R. Wolpert

STA205 Probability: Week 8 R. Wolpert INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and

More information

Characterization of c-, L- and φ k -optimal designs for a class of non-linear multiple-regression models

Characterization of c-, L- and φ k -optimal designs for a class of non-linear multiple-regression models J. R. Statist. Soc. B (2019) 81, Part 1, pp. 101 120 Characterization of c-, L- and φ k -optimal designs for a class of non-linear multiple-regression models Dennis Schmidt Otto-von-Guericke-Universität

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 477 Instructor: Tony Jebara Topic 5 Generalization Guarantees VC-Dimension Nearest Neighbor Classification (infinite VC dimension) Structural Risk Minimization Support Vector Machines

More information