Optimal experimental design, an introduction, Jesús López Fidalgo
|
|
- Jonah Dean
- 6 years ago
- Views:
Transcription
1 Optimal experimental design, an introduction University of Castilla-La Mancha Department of Mathematics Institute of Applied Mathematics to Science and Engineering
2 Books (just a few) Fedorov V.V. (1972). Theory of optimal experiments. Academic press. New York. Pazman A. (1986). Foundations of Optimum Experimental Design. D. Reidel Pub. New York. Pukelsheim F. (1993). Optimal Design of Experiments. John Wiley & Sons. New York. Fedorov V.V. and Hackl P. (1997). Model-oriented design of experiments. Springer. New York. Atkinson A.C., Doney A.N., Tobias R. (2007). Optimum Experimental Designs, With SAS. Oxford science publications. Oxford.
3 OPTIMAL EXPERIMENTAL DESIGN THEORY
4 Concepts and notation Process: Select a model, E(y x) = f T (x)θ, Var(y x) = σ 2 (linear model). f continuous with components linearly independent. Select an experimental condition x on a design space χ (typically a compact on a Euclidian space). Examples: y = α0 + α 1x + ε, f (x) = (1, x) T, x χ = [a, b]. y = α1x 1 + α 2x 2 + α 3x 3 + ε, f (x) = x = (x 1, x 2, x 3) T, x χ = [a 1, b 1] [a 2, b 2] [a 3, b 3]. y = α0 + α 1x + α 2x 2 + ε, f (x) = (1, x, x 2 ) T, x χ = [a, b]. ANOVA (1-way, 3 levels): y = β 0 + β 1x 1 + β 2x 2 + ε, f (x 1, x 2) = (1, x 1, x 2) T, x = (x 1, x 2) χ = {(0, 1), (1, 0), (1, 1)}
5 Process Perform the experiment and observe the response (r.v. y). Repeat the process for the experimental conditions x 1,..., x n obtaining responses y 1,..., y n. Estimate the parameters (MLE) θ 1,..., θ m, σ 2 and make inferences.
6 EXPERIMENTAL DESIGN Aim: Choosing x 1,..., x n (exact design of size n). Some may be repeated (probability measure): ξ n (x) = nx n, where n x = times x appears in the design. Covariance matrix: Σˆθ = σ2 (X T X ) 1 = σ 2 n 1 M 1 (ξ), M(ξ) = x χ f (x)f T (x)ξ n (x) (information matrix): Symmetric. Non-definite positive. Singular if the number of different points in the design k < m.
7 Extended concept of experimental design Approximate design (Kiefer): probability measure ζ on the Borel field generated by the open sets of χ. Information matrix: M(ζ) = χ f (x)f T (x)ζ(dx). Convex set of approximate designs: Ξ. If ζ is discrete and ξ the probability function associated: M(ξ) = f (x)f T (x)ξ(x) x χ If ζ is absolutely continuous there exists a density function (pdf), ξ, associated: M(ξ) = f (x)f T (x)ξ(x)dx Compact and convex set of information matrices: M χ
8 Caratheodory s theorem Caratheodory s theorem: For any information matrix there is always a design with at most 1 2m(m + 1) + 1 different points. Finite support, { } x1 x ξ = 2... x k p 1 p 2... p k Restricting Ξ to finite designs M is still the same (Caratheodory, definition of integral and continuity). In practice, perform n i nξ(x i ) experiments at x i, i n i = n. From now on we will use ξ or ζ as convenience.
9 Example y = α 0 + α 1 x + ε, x χ = [0, 1], f (x) = (1, x) T. ξ = { }. M(ξ) = = = 3 ( 1 xi i=1 x i x 2 i ) p i ( ) ( ) ( ) ( ) 0.4
10 Example ξ = { If just n = 12 experiments may be performed, possible rounding off: }. about = 2.4 (2, 3, 3) experiments at 0, about = 4.8 (5, 4, 5) experiments at 0.5, about = 4.8 (5, 5, 4) experiments at 1. Remark: It is not as simple as the usual rounding off.
11 CRITERION FUNCTION Φ : Ξ [0, + ) or Φ : M [0, + ) to be minimized (maximized). The first is more general. The second has to be considered as a function of M 1, has to be non-increasing (better estimates of the parameters): M N Φ(M) Φ(N), may be global or partial (interest on part of the parameters), may be positive homogeneous: Φ(δM) = 1 Φ(M), δ > 0, δ M is within the cone of the non-negative definite matrices in the Euclidean space of the symmetric matrices. Thus, the usual concept of differentiability applies here. A Φ optimal design ζ minimizes Φ. Convex: Φ[(1 ɛ)ζ + ɛζ ] (1 ɛ)φ(ζ) + ɛφ(ζ ).
12 Equivalence theorem for differentiable criteria Sensitive function: ψ(x, ξ) = Φ[M(ξ), f (x)f T (x)] = f T (x) Φ[M(ξ)]f (x) trm(ξ) Φ[M(ξ)]. Equivalence theorem: ξ is Φ optimal if and only if ψ(x, ξ ) 0, x χ. Equality for x S ξ.
13 Comments on the equivalence theorem Just for approximate designs. M is convex and it is the same for finite or general designs. From now on we assume finite designs and use the symbol ξ (pdf) instead of ζ (measure). Still valid for a restricted search in a convex subset.
14 Checking condition ψ(z, ξ ) = 0, z S ξ ( ψ(x, ξ ) ) x x=z E.g., for a two parameter model assume ξ = then there are two equations: = 0, z S ξ Int(χ) { x1 x 2 1 p p }, ψ(x 1, ξ) = 0, ψ(x 2, ξ) = 0 and 0, 1 or 2 of the equations: ( ) ( ) ψ(x, ξ) ψ(x, ξ) = 0, = 0 x x=x 1 x x=x 2 to be solved.
15 EFFICIENCY (goodness of a design) eff Φ (ξ) = Φ[M(ξ )] Φ[M(ξ)]. If Φ homogenous, e.g. for 60% efficiency and n observations: 0.6 = Φ[M(ξ )] Φ[M(ξ)] = Φ(σ 2 nσ 1 ) ˆθ Φ(σ 2 nσ 1 ˆθ 1 ) ˆθ ) = σ2n 1 Φ(Σ σ 2 n 1 Φ(Σ 1 ) = Φ(Σ Φ(Σ 1 ˆθ ). For n observations with ξ and n with ξ the efficiency is 1 if: 1 = Φ(Σ 1 ) ˆθ Φ(Σ 1 ˆθ ) = σ2n 1Φ[M(ξ )] σ 2 n 1 Φ[M(ξ)] ˆθ = σ2 n 1 σ 2 0.6, n 1 n = 0.6n (60%) observations would be enough with ξ. 1 ) ˆθ
16 General properties for global criteria A Φ optimal design has m points at least: Any design with less than m points gives a singular matrix. For strictly convex criteria there is always a Φ optimal design with no more than m(m+1) 2 points: The optimal must be in the boundary (dimension m(m+1) 1). 2 Caratheodory in the boundary gives a limit of m(m+1) points. 2 If σ 2 (x) is not constant, the whole theory is applicable for f (x) = f (x)/σ(x). If the observations are correlated (Σ y ), just exact designs. Information matrix: X T Σ 1 y X.
17 D OPTIMALITY Determinant of the inverse: { log det M Φ D [M(ξ)] = 1 (ξ) = log det M(ξ) if det M(ξ) 0, if det M(ξ) = 0. Homogeneous definition: { det M Φ D [M(ξ)] = 1/m (ξ) if det M(ξ) 0, if det M(ξ) = 0.
18 Properties Φ D continuous on M. Convex on M and strictly convex on M + (nonsingular information matrices). Differentiable on M + : ( log det M) = M 1 Minimum volume of the confidence elipsoide: ( ˆΘ Θ) T (X T X ) 1 ( ˆΘ Θ) mf m,n m,γ S 2 R c 2. Volume =c m V m [det M 1 (ξ)] 1/2, where V m is the volume of the m dimensional sphere.
19 G optimality Kirstine Smith (1918). Kiefer & Wolfovitz gave the name. Φ G [M(ξ)] = { maxx χ f T (x)m 1 (ξ)f (x) if det M(ξ) 0, if det M(ξ) = 0. Φ G continuous and convex on M. Φ G [M(ξ)] max x χ var ξ ŷ(x).
20 A optimality Trace of the inverse: { } trm Φ A [M(ξ)] = 1 (ξ) if det M(ξ) 0 if det M(ξ) = 0, The second equality is a consequence of if the parameters are estimable. Φ A continuous on M. var ξ ˆθi e t i M 1 (ξ)e i Convex on M and strictly convex on M +. Differentiable on M + : (trm 1 (ξ)) = M 2 (ξ) m var ξ ˆθi. i=1
21 L optimality Atwood (1976): Φ L [M(ξ)] = { trwm 1 (ξ) if det M(ξ) 0, if det M(ξ) = 0, where W is a positive definite matrix of dimension m. Φ L continuous on M. Convex on M and strictly convex on M +. Differentiable on M + : [trwm 1 (ξ)] = M 1 (ξ)wm 1 (ξ).
22 c optimality Variance of the estimator of c T θ, linear combination of the parameters. Φ c [M(ξ)] = c T M (ξ)c.
23 Elfving procedure in practice Plot the curve x(t) = f 1 (t), y(t) = f 2 (t), t χ and its symmetric through origin. Plot the convex hull of both curves. Plot the line through the origin defined by vector c. Optimal design defined by the boundary point.
24 Elfving plot (One point design) { t } 1
25 Elfving plot (Two point design) { } a t 1 p p
26 Elfving procedure for three parameters y = θ 0 + θ 1 t + θ 2 t 2 + ε, t [0, 1] One, two or three design points.
27 Equivalence theorem for D and G optimality ξ is D optimal iff it is G optimal: where d(x, ξ) = f T (x)m 1 (ξ)f (x). Efficiency bound: eff D [M(ξ)] 2 max d(x, x χ ξ ) = m, maxx d(x,ξ) m
28 Example: Checking condition for a quadratic model E[y] = θ T f (x) = θ 1 + θ 2 x + θ 3 x 2, σ 2 (x) = 1, x χ = [ 1, 2], f (x) = (1, x, x 2 ) T. M(ξ) = = x [ 1,2] x [ 1,2] f (x)f T (x)ξ(x) ξ(x) 1 x x 2 x x 2 x 3 x 2 x 3 x 4 Assume the D-optimal design is of the type: { } 1 z 2 ξ = 1/3 1/3 1/3,
29 Quadratic model Maximize the determinant: det M(ξ) = 1 3 ( z 2 + z + 2 ) 2. The maximum is reached at z = 0.5: { ξ = 1/3 1/3 1/3 Checking whether ξ is the D-optimal: d(x, ξ ) = f T (x)m 1 (ξ )f (x) }, = 0.89 ( x x ) ( x x )
30 Example: Elfving procedure E[y] = θ T f (x) = θ 1 x + θ 2 x 2, σ 2 (x) = 1, x χ = [0, 1], f (x) = (1, x, x 2 ) T. Plot the curve x(t) = t, y(t) = t 2, t [0, 1]:
31 Symmetric through the origin
32 Convex hull Tangential point, z: x(z) x 0 x (z) = y(z) y 0 y (z) z = z2 + 1, z = 2 1 2z
33 Line through the origin defined by vector c = (0.5, 0.3) (λ0.5) 2 = λ0.3, t = 3/5, ξ c = { 3/5 1 }
34 Line through the origin defined by vector c = (0.5, 1) Cutting point: { y = 2x, x = 1 4 ( 2 2 x 1 z+1 = y 1 z 2 +1, ) = 0.15, y = 0.30.
35 Weights Convex combination: (0.15, 0.30) = (1 p)( z, z 2 ) + p(1, 1), p = 1 4 ( 3 ) 2 = 0.30 { ξc = ( ) ( ) }.
36 THANK YOU
Optimal experimental designs applied to Biostatistics.
Optimal experimental designs applied to Biostatistics. Thesis for the Degree of Doctor of Mathematics Santiago Campos Barreiro Academic advisor: Prof. Dr. Jesús López Fidalgo Institute of Mathematics Applied
More informationApproximate Optimal Designs for Multivariate Polynomial Regression
Approximate Optimal Designs for Multivariate Polynomial Regression Fabrice Gamboa Collaboration with: Yohan de Castro, Didier Henrion, Roxana Hess, Jean-Bernard Lasserre Universität Potsdam 16th of February
More informationOptimal Design in Regression. and Spline Smoothing
Optimal Design in Regression and Spline Smoothing by Jaerin Cho A thesis submitted to the Department of Mathematics and Statistics in conformity with the requirements for the degree of Doctor of Philosophy
More informationON THE REGULARIZATION OF SINGULAR C-OPTIMAL
LABORATOIRE INFORMATIQUE, SIGNAUX ET SYSTÈMES DE SOPHIA ANTIPOLIS UMR 6070 ON THE REGULARIZATION OF SINGULAR C-OPTIMAL DESIGNS Luc Pronzato Equipe SYSTEMES Rapport de recherche ISRN I3S/RR 2008-15 FR Septembre
More informationPROBABILITY AND STATISTICS Vol. III - Statistical Experiments and Optimal Design - Andrej Pázman STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN
STATISTICAL EXPERIMENTS AND OPTIMAL DESIGN Andrej Pázman Comenius University, Bratislava, Slovakia Keywords: Experiment design, linear statistical model, nonlinear regression, least squares, information
More informationE-optimal approximate block designs for treatment-control comparisons
E-optimal approximate block designs for treatment-control comparisons Samuel Rosa 1 1 Faculty of Mathematics, Physics and Informatics, Comenius University in Bratislava, Slovakia We study E-optimal block
More informationOptimal designs for estimating the slope of a regression
Optimal designs for estimating the slope of a regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 4480 Bochum, Germany e-mail: holger.dette@rub.de Viatcheslav B. Melas St. Petersburg
More informationarxiv: v1 [stat.co] 7 Sep 2017
Computing optimal experimental designs with respect to a compound Bayes risk criterion arxiv:1709.02317v1 [stat.co] 7 Sep 2017 Radoslav Harman, and Maryna Prus September 8, 2017 Abstract We consider the
More informationA new algorithm for deriving optimal designs
A new algorithm for deriving optimal designs Stefanie Biedermann, University of Southampton, UK Joint work with Min Yang, University of Illinois at Chicago 18 October 2012, DAE, University of Georgia,
More informationOptimal Design. By Anthony C. Atkinson. 1 Introduction
By Anthony C. Atkinson Keywords: algorithms, blocking, D-optimal design, equivalence theorem, I-optimal design, mixture design, nonregular design region, prediction variance, response-surface design Abstract:
More informationOptimal design of experiments
Optimal design of experiments Session 4: Some theory Peter Goos / 40 Optimal design theory continuous or approximate optimal designs implicitly assume an infinitely large number of observations are available
More informationA Study on Minimal-point Composite Designs
A Study on Minimal-point Composite Designs by Yin-Jie Huang Advisor Ray-Bing Chen Institute of Statistics, National University of Kaohsiung Kaohsiung, Taiwan 8 R.O.C. July 007 Contents Introduction Construction
More informationOn construction of constrained optimum designs
On construction of constrained optimum designs Institute of Control and Computation Engineering University of Zielona Góra, Poland DEMA2008, Cambridge, 15 August 2008 Numerical algorithms to construct
More informationStochastic Design Criteria in Linear Models
AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework
More informationBayesian Inference. Chapter 9. Linear models and regression
Bayesian Inference Chapter 9. Linear models and regression M. Concepcion Ausin Universidad Carlos III de Madrid Master in Business Administration and Quantitative Methods Master in Mathematical Engineering
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More informationMa 3/103: Lecture 24 Linear Regression I: Estimation
Ma 3/103: Lecture 24 Linear Regression I: Estimation March 3, 2017 KC Border Linear Regression I March 3, 2017 1 / 32 Regression analysis Regression analysis Estimate and test E(Y X) = f (X). f is the
More informationECON 4117/5111 Mathematical Economics
Test 1 September 29, 2006 1. Use a truth table to show the following equivalence statement: (p q) (p q) 2. Consider the statement: A function f : X Y is continuous on X if for every open set V Y, the pre-image
More informationON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE
Sankhyā : The Indian Journal of Statistics 999, Volume 6, Series B, Pt. 3, pp. 488 495 ON D-OPTIMAL DESIGNS FOR ESTIMATING SLOPE By S. HUDA and A.A. AL-SHIHA King Saud University, Riyadh, Saudi Arabia
More informationNotes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed
18.466 Notes, March 4, 2013, R. Dudley Maximum likelihood estimation: actual or supposed 1. MLEs in exponential families Let f(x,θ) for x X and θ Θ be a likelihood function, that is, for present purposes,
More informationMultiplicative Algorithm for computing Optimum Designs
for computing s DAE Conference 2012 Roberto Dorta-Guerra, Raúl Martín Martín, Mercedes Fernández-Guerrero and Licesio J. Rodríguez-Aragón Optimum Experimental Design group October 20, 2012 http://areaestadistica.uclm.es/oed
More informationFrequentist-Bayesian Model Comparisons: A Simple Example
Frequentist-Bayesian Model Comparisons: A Simple Example Consider data that consist of a signal y with additive noise: Data vector (N elements): D = y + n The additive noise n has zero mean and diagonal
More informationOptimal and efficient designs for Gompertz regression models
Ann Inst Stat Math (2012) 64:945 957 DOI 10.1007/s10463-011-0340-y Optimal and efficient designs for Gompertz regression models Gang Li Received: 13 July 2010 / Revised: 11 August 2011 / Published online:
More informationDiscussion Paper SFB 823. Nearly universally optimal designs for models with correlated observations
SFB 823 Nearly universally optimal designs for models with correlated observations Discussion Paper Holger Dette, Andrey Pepelyshev, Anatoly Zhigljavsky Nr. 3/212 Nearly universally optimal designs for
More informationOptimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models
Optimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models Anthony C. Atkinson, Department of Statistics, London School of Economics, London WC2A 2AE, UK and Barbara Bogacka, School
More informationRandomisation, Replication, Response Surfaces. and. Rosemary
Randomisation, Replication, Response Surfaces and Rosemary 1 A.C. Atkinson a.c.atkinson@lse.ac.uk Department of Statistics London School of Economics London WC2A 2AE, UK One joint publication RAB AND ME
More informationOptimal designs for estimating the slope in nonlinear regression
Optimal designs for estimating the slope in nonlinear regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Viatcheslav B. Melas St.
More informationLOWELL JOURNAL. MUST APOLOGIZE. such communication with the shore as Is m i Boimhle, noewwary and proper for the comfort
- 7 7 Z 8 q ) V x - X > q - < Y Y X V - z - - - - V - V - q \ - q q < -- V - - - x - - V q > x - x q - x q - x - - - 7 -» - - - - 6 q x - > - - x - - - x- - - q q - V - x - - ( Y q Y7 - >»> - x Y - ] [
More informationMultiple Linear Regression
Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there
More informationA geometric characterization of c-optimal designs for heteroscedastic regression
A geometric characterization of c-optimal designs for heteroscedastic regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Tim Holland-Letz
More informationLecture 32: Asymptotic confidence sets and likelihoods
Lecture 32: Asymptotic confidence sets and likelihoods Asymptotic criterion In some problems, especially in nonparametric problems, it is difficult to find a reasonable confidence set with a given confidence
More informationEstimation Theory. as Θ = (Θ 1,Θ 2,...,Θ m ) T. An estimator
Estimation Theory Estimation theory deals with finding numerical values of interesting parameters from given set of data. We start with formulating a family of models that could describe how the data were
More informationOrthogonal decompositions in growth curve models
ACTA ET COMMENTATIONES UNIVERSITATIS TARTUENSIS DE MATHEMATICA Volume 4, Orthogonal decompositions in growth curve models Daniel Klein and Ivan Žežula Dedicated to Professor L. Kubáček on the occasion
More informationMinimax design criterion for fractional factorial designs
Ann Inst Stat Math 205 67:673 685 DOI 0.007/s0463-04-0470-0 Minimax design criterion for fractional factorial designs Yue Yin Julie Zhou Received: 2 November 203 / Revised: 5 March 204 / Published online:
More informationMAS223 Statistical Inference and Modelling Exercises
MAS223 Statistical Inference and Modelling Exercises The exercises are grouped into sections, corresponding to chapters of the lecture notes Within each section exercises are divided into warm-up questions,
More informationThe Q Method for Symmetric Cone Programmin
The Q Method for Symmetric Cone Programming The Q Method for Symmetric Cone Programmin Farid Alizadeh and Yu Xia alizadeh@rutcor.rutgers.edu, xiay@optlab.mcma Large Scale Nonlinear and Semidefinite Progra
More informationMultivariable Calculus
2 Multivariable Calculus 2.1 Limits and Continuity Problem 2.1.1 (Fa94) Let the function f : R n R n satisfy the following two conditions: (i) f (K ) is compact whenever K is a compact subset of R n. (ii)
More informationDesigns for Generalized Linear Models
Designs for Generalized Linear Models Anthony C. Atkinson David C. Woods London School of Economics and Political Science, UK University of Southampton, UK December 9, 2013 Email: a.c.atkinson@lse.ac.uk
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationReview (Probability & Linear Algebra)
Review (Probability & Linear Algebra) CE-725 : Statistical Pattern Recognition Sharif University of Technology Spring 2013 M. Soleymani Outline Axioms of probability theory Conditional probability, Joint
More informationA new approach to optimal designs for models with correlated observations
A new approach to optimal designs for models with correlated observations Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Andrey Pepelyshev
More informationLecture 25: Review. Statistics 104. April 23, Colin Rundel
Lecture 25: Review Statistics 104 Colin Rundel April 23, 2012 Joint CDF F (x, y) = P [X x, Y y] = P [(X, Y ) lies south-west of the point (x, y)] Y (x,y) X Statistics 104 (Colin Rundel) Lecture 25 April
More informationMultivariate Analysis and Likelihood Inference
Multivariate Analysis and Likelihood Inference Outline 1 Joint Distribution of Random Variables 2 Principal Component Analysis (PCA) 3 Multivariate Normal Distribution 4 Likelihood Inference Joint density
More informationwhere r n = dn+1 x(t)
Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution
More informationMISCELLANEOUS TOPICS RELATED TO LIKELIHOOD. Copyright c 2012 (Iowa State University) Statistics / 30
MISCELLANEOUS TOPICS RELATED TO LIKELIHOOD Copyright c 2012 (Iowa State University) Statistics 511 1 / 30 INFORMATION CRITERIA Akaike s Information criterion is given by AIC = 2l(ˆθ) + 2k, where l(ˆθ)
More informationconditional cdf, conditional pdf, total probability theorem?
6 Multiple Random Variables 6.0 INTRODUCTION scalar vs. random variable cdf, pdf transformation of a random variable conditional cdf, conditional pdf, total probability theorem expectation of a random
More informationD-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors
Journal of Data Science 920), 39-53 D-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors Chuan-Pin Lee and Mong-Na Lo Huang National Sun Yat-sen University Abstract: Central
More informationMAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS
Statistica Sinica 17(2007), 463-480 MAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS Holger Dette, Linda M. Haines and Lorens A. Imhof Ruhr-Universität, University of Cape Town and Bonn University
More informationNonlinear Programming Models
Nonlinear Programming Models Fabio Schoen 2008 http://gol.dsi.unifi.it/users/schoen Nonlinear Programming Models p. Introduction Nonlinear Programming Models p. NLP problems minf(x) x S R n Standard form:
More informationChapter 2: Fundamentals of Statistics Lecture 15: Models and statistics
Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:
More informationCointegrated VAR s. Eduardo Rossi University of Pavia. November Rossi Cointegrated VAR s Financial Econometrics / 56
Cointegrated VAR s Eduardo Rossi University of Pavia November 2013 Rossi Cointegrated VAR s Financial Econometrics - 2013 1 / 56 VAR y t = (y 1t,..., y nt ) is (n 1) vector. y t VAR(p): Φ(L)y t = ɛ t The
More informationMathematical Background
A Mathematical Background Version: date: June 28, 2018 Copyright c 2018 by Nob Hill Publishing, LLC A.1 Introduction In this appendix we give a brief review of some concepts that we need. It is assumed
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationLasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices
Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,
More informationLecture 7. Econ August 18
Lecture 7 Econ 2001 2015 August 18 Lecture 7 Outline First, the theorem of the maximum, an amazing result about continuity in optimization problems. Then, we start linear algebra, mostly looking at familiar
More informationTime Series Analysis
Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Regression based methods, 1st part: Introduction (Sec.
More informationMA102: Multivariable Calculus
MA102: Multivariable Calculus Rupam Barman and Shreemayee Bora Department of Mathematics IIT Guwahati Differentiability of f : U R n R m Definition: Let U R n be open. Then f : U R n R m is differentiable
More informationActive Robot Calibration Algorithm
Active Robot Calibration Algorithm Yu Sun and John M. Hollerbach Abstract This paper presents a new updating algorithm to reduce the complexity of computing an observability index for kinematic calibration
More informationAsymptotics of minimax stochastic programs
Asymptotics of minimax stochastic programs Alexander Shapiro Abstract. We discuss in this paper asymptotics of the sample average approximation (SAA) of the optimal value of a minimax stochastic programming
More informationPenalized D-optimal design for dose finding
Penalized for dose finding Luc Pronzato Laboratoire I3S, CNRS-Univ. Nice Sophia Antipolis, France 2/ 35 Outline 3/ 35 Bernoulli-type experiments: Y i {0,1} (success or failure) η(x,θ) = Prob(Y i = 1 x
More informationIntroduction to Simple Linear Regression
Introduction to Simple Linear Regression Yang Feng http://www.stat.columbia.edu/~yangfeng Yang Feng (Columbia University) Introduction to Simple Linear Regression 1 / 68 About me Faculty in the Department
More informationReview (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology
Review (probability, linear algebra) CE-717 : Machine Learning Sharif University of Technology M. Soleymani Fall 2012 Some slides have been adopted from Prof. H.R. Rabiee s and also Prof. R. Gutierrez-Osuna
More informationSCHOOL OF MATHEMATICS AND STATISTICS. Linear and Generalised Linear Models
SCHOOL OF MATHEMATICS AND STATISTICS Linear and Generalised Linear Models Autumn Semester 2017 18 2 hours Attempt all the questions. The allocation of marks is shown in brackets. RESTRICTED OPEN BOOK EXAMINATION
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationMath 362, Problem set 1
Math 6, roblem set Due //. (4..8) Determine the mean variance of the mean X of a rom sample of size 9 from a distribution having pdf f(x) = 4x, < x
More informationSOLUTION FOR HOMEWORK 4, STAT 4352
SOLUTION FOR HOMEWORK 4, STAT 4352 Welcome to your fourth homework. Here we begin the study of confidence intervals, Errors, etc. Recall that X n := (X 1,...,X n ) denotes the vector of n observations.
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1
MA 575 Linear Models: Cedric E Ginestet, Boston University Mixed Effects Estimation, Residuals Diagnostics Week 11, Lecture 1 1 Within-group Correlation Let us recall the simple two-level hierarchical
More informationThe EG-family of algorithms and procedures for solving linear differential and difference higher-order systems
The EG-family of algorithms and procedures for solving linear differential and difference higher-order systems S. Abramov, D. Khmelnov, A. Ryabenko Computing Centre of the Russian Academy of Sciences Moscow
More informationOptimal designs for enzyme inhibition kinetic models
Optimal designs for enzyme inhibition kinetic models Kirsten Schorning, Holger Dette, Katrin Kettelhake, Tilman Möller Fakultät für Mathematik Ruhr-Universität Bochum 44780 Bochum Germany Abstract In this
More informationProf. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India
CHAPTER 9 BY Prof. M. Saha Professor of Mathematics The University of Burdwan West Bengal, India E-mail : mantusaha.bu@gmail.com Introduction and Objectives In the preceding chapters, we discussed normed
More informationOptimal discrimination designs
Optimal discrimination designs Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@ruhr-uni-bochum.de Stefanie Titoff Ruhr-Universität Bochum Fakultät
More informationRegression: Lecture 2
Regression: Lecture 2 Niels Richard Hansen April 26, 2012 Contents 1 Linear regression and least squares estimation 1 1.1 Distributional results................................ 3 2 Non-linear effects and
More informationVARIATIONAL CALCULUS IN SPACE OF MEASURES AND OPTIMAL DESIGN
Chapter 1 VARIATIONAL CALCULUS IN SPACE OF MEASURES AND OPTIMAL DESIGN Ilya Molchanov Department of Statistics University of Glasgow ilya@stats.gla.ac.uk www.stats.gla.ac.uk/ ilya Sergei Zuyev Department
More informationStatistics 3858 : Maximum Likelihood Estimators
Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,
More informationStatistical Inference of Covariate-Adjusted Randomized Experiments
1 Statistical Inference of Covariate-Adjusted Randomized Experiments Feifang Hu Department of Statistics George Washington University Joint research with Wei Ma, Yichen Qin and Yang Li Email: feifang@gwu.edu
More informationLecture 6: Linear models and Gauss-Markov theorem
Lecture 6: Linear models and Gauss-Markov theorem Linear model setting Results in simple linear regression can be extended to the following general linear model with independently observed response variables
More informationStat 710: Mathematical Statistics Lecture 12
Stat 710: Mathematical Statistics Lecture 12 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 12 Feb 18, 2009 1 / 11 Lecture 12:
More informationA Probability Review
A Probability Review Outline: A probability review Shorthand notation: RV stands for random variable EE 527, Detection and Estimation Theory, # 0b 1 A Probability Review Reading: Go over handouts 2 5 in
More informationChapter 5 Matrix Approach to Simple Linear Regression
STAT 525 SPRING 2018 Chapter 5 Matrix Approach to Simple Linear Regression Professor Min Zhang Matrix Collection of elements arranged in rows and columns Elements will be numbers or symbols For example:
More informationTheory of Statistics.
Theory of Statistics. Homework V February 5, 00. MT 8.7.c When σ is known, ˆµ = X is an unbiased estimator for µ. If you can show that its variance attains the Cramer-Rao lower bound, then no other unbiased
More informationStructural and Multidisciplinary Optimization. P. Duysinx and P. Tossings
Structural and Multidisciplinary Optimization P. Duysinx and P. Tossings 2018-2019 CONTACTS Pierre Duysinx Institut de Mécanique et du Génie Civil (B52/3) Phone number: 04/366.91.94 Email: P.Duysinx@uliege.be
More informationClustering by Mixture Models. General background on clustering Example method: k-means Mixture model based clustering Model estimation
Clustering by Mixture Models General bacground on clustering Example method: -means Mixture model based clustering Model estimation 1 Clustering A basic tool in data mining/pattern recognition: Divide
More information9. Multivariate Linear Time Series (II). MA6622, Ernesto Mordecki, CityU, HK, 2006.
9. Multivariate Linear Time Series (II). MA6622, Ernesto Mordecki, CityU, HK, 2006. References for this Lecture: Introduction to Time Series and Forecasting. P.J. Brockwell and R. A. Davis, Springer Texts
More information6.3 Forecasting ARMA processes
6.3. FORECASTING ARMA PROCESSES 123 6.3 Forecasting ARMA processes The purpose of forecasting is to predict future values of a TS based on the data collected to the present. In this section we will discuss
More informationCOHERENT DISPERSION CRITERIA FOR OPTIMAL EXPERIMENTAL DESIGN
The Annals of Statistics 1999, Vol. 27, No. 1, 65 81 COHERENT DISPERSION CRITERIA FOR OPTIMAL EXPERIMENTAL DESIGN By A. Philip Dawid and Paola Sebastiani 1 University College London and The Open University
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More information15.S24 Sample Exam Solutions
5.S4 Sample Exam Solutions. In each case, determine whether V is a vector space. If it is not a vector space, explain why not. If it is, find basis vectors for V. (a) V is the subset of R 3 defined by
More informationOn the efficiency of two-stage adaptive designs
On the efficiency of two-stage adaptive designs Björn Bornkamp (Novartis Pharma AG) Based on: Dette, H., Bornkamp, B. and Bretz F. (2010): On the efficiency of adaptive designs www.statistik.tu-dortmund.de/sfb823-dp2010.html
More informationZ-estimators (generalized method of moments)
Z-estimators (generalized method of moments) Consider the estimation of an unknown parameter θ in a set, based on data x = (x,...,x n ) R n. Each function h(x, ) on defines a Z-estimator θ n = θ n (x,...,x
More informationChapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets
Chapter 7. Confidence Sets Lecture 30: Pivotal quantities and confidence sets Confidence sets X: a sample from a population P P. θ = θ(p): a functional from P to Θ R k for a fixed integer k. C(X): a confidence
More informationON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA. Cross River State, Nigeria
ON THE COMPARISON OF BOUNDARY AND INTERIOR SUPPORT POINTS OF A RESPONSE SURFACE UNDER OPTIMALITY CRITERIA Thomas Adidaume Uge and Stephen Seastian Akpan, Department Of Mathematics/Statistics And Computer
More informationHeteroscedastic T-Optimum Designs for Multiresponse Dynamic Models
Heteroscedastic T-Optimum Designs for Multiresponse Dynamic Models Dariusz Uciński 1 and Barbara Bogacka 2 1 Institute of Control and Computation Engineering, University of Zielona Góra, ul. Podgórna 50,
More information16.584: Random Vectors
1 16.584: Random Vectors Define X : (X 1, X 2,..X n ) T : n-dimensional Random Vector X 1 : X(t 1 ): May correspond to samples/measurements Generalize definition of PDF: F X (x) = P[X 1 x 1, X 2 x 2,...X
More informationError analysis for efficiency
Glen Cowan RHUL Physics 28 July, 2008 Error analysis for efficiency To estimate a selection efficiency using Monte Carlo one typically takes the number of events selected m divided by the number generated
More informationVariance reduction. Michel Bierlaire. Transport and Mobility Laboratory. Variance reduction p. 1/18
Variance reduction p. 1/18 Variance reduction Michel Bierlaire michel.bierlaire@epfl.ch Transport and Mobility Laboratory Variance reduction p. 2/18 Example Use simulation to compute I = 1 0 e x dx We
More informationOn V-orthogonal projectors associated with a semi-norm
On V-orthogonal projectors associated with a semi-norm Short Title: V-orthogonal projectors Yongge Tian a, Yoshio Takane b a School of Economics, Shanghai University of Finance and Economics, Shanghai
More informationLecture 28: Asymptotic confidence sets
Lecture 28: Asymptotic confidence sets 1 α asymptotic confidence sets Similar to testing hypotheses, in many situations it is difficult to find a confidence set with a given confidence coefficient or level
More informationSTA205 Probability: Week 8 R. Wolpert
INFINITE COIN-TOSS AND THE LAWS OF LARGE NUMBERS The traditional interpretation of the probability of an event E is its asymptotic frequency: the limit as n of the fraction of n repeated, similar, and
More informationCharacterization of c-, L- and φ k -optimal designs for a class of non-linear multiple-regression models
J. R. Statist. Soc. B (2019) 81, Part 1, pp. 101 120 Characterization of c-, L- and φ k -optimal designs for a class of non-linear multiple-regression models Dennis Schmidt Otto-von-Guericke-Universität
More informationMachine Learning 4771
Machine Learning 477 Instructor: Tony Jebara Topic 5 Generalization Guarantees VC-Dimension Nearest Neighbor Classification (infinite VC dimension) Structural Risk Minimization Support Vector Machines
More information