Efficient computation of Bayesian optimal discriminating designs

Size: px
Start display at page:

Download "Efficient computation of Bayesian optimal discriminating designs"

Transcription

1 Efficient computation of Bayesian optimal discriminating designs Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik Bochum, Germany Roman Guchenko, Viatcheslav B. Melas St. Petersburg State University Department of Mathematics St. Petersburg, Russia Abstract An efficient algorithm for the determination of Bayesian optimal discriminating designs for competing regression models is developed, where the main focus is on models with general distributional assumptions beyond the classical case of normally distributed homoscedastic errors. For this purpose we consider a Bayesian version of the Kullback- Leibler (KL) optimality criterion introduced by López-Fidalgo et al. (2007). Discretizing the prior distribution leads to local KL-optimal discriminating design problems for a large number of competing models. All currently available methods either require a large computation time or fail to calculate the optimal discriminating design, because they can only deal efficiently with a few model comparisons. In this paper we develop a new algorithm for the determination of Bayesian optimal discriminating designs with respect to the Kullback-Leibler criterion. It is demonstrated that the new algorithm is able to calculate the optimal discriminating designs with reasonable accuracy and computational time in situations where all currently available procedures are either slow or fail. Keyword and Phrases: Design of experiment; Bayesian optimal design; model discrimination; gradient methods; model uncertainty; Kullback-Leibler distance 1

2 1 Introduction Although optimal designs can provide a substantial improvement in the statistical accuracy without making any additional experiments, classical optimal design theory [see for example Pukelsheim (2006); Atkinson et al. (2007)] has been criticized, because it relies heavily on the specification of a particular model. In many cases a good design for a given model might be inefficient if it is used in a different setup. Most of the literature addressing the problem of model uncertainty in the design of experiments can be roughly divided into two parts, where all authors assume that a certain class of parametric models is available to describe the relation between the predictor and the response. One approach to obtain model robustness is to construct designs which allow the efficient estimation of parameters in all models under consideration. This is usually achieved by optimizing composite optimality criteria, which are defined as an average of the criteria for the different models [see Läuter (1974), Dette (1990); Biedermann et al. (2006); Dette et al. (2008)]. Alternatively, one can directly construct designs to discriminate between several competing models. An early reference is Stigler (1971) who determined designs for discriminating between two nested univariate polynomials by minimizing the volume of the confidence ellipsoid for the parameters corresponding to the extension of the smaller model. Since this seminal paper several authors have followed this line of research [see for example Dette and Haller (1998) or Song and Wong (1999) among others]. A completely different approach for the construction of optimal designs for model discrimination was suggested by Atkinson and Fedorov (1975a). The corresponding optimality criterion is called T -optimality criterion. To be precise, assume that the relation between the response Y and predictor x is described by a nonlinear regression model such that (1.1) E[Y x] = η(x, θ), Var(Y x) = v 2 (x, θ), and that the experimenter considers two rival models, say η 1, η 2, as candidates for the parametric form of the mean. Roughly speaking, Atkinson and Fedorov (1975a) assumed homoscedasticity, fixed one model, say η 1, and constructed the design such that the sum of squares for a lack of fit test against the alternative η 2 is large. The criterion was extended in several directions. For example, Atkinson and Fedorov (1975b) considered the problem of discriminating a selected model η 1 from a class of other regression models, say {η 2,..., η ν }, ν 2, and Tommasi (2009) combined the T -criterion with the approach introduced by Läuter (1974). Ucinski and Bogacka (2005) remarked that the criterion introduced by Atkinson and Fedorov (1975a) is only applicable in the case of homoscedastic errors in the regression model (1.1) and discussed an extension to the case of heteroscedasticity. More generally, López-Fidalgo et al. (2007) introduced a generalization of the T -optimality criterion which is applicable under general distributional 2

3 assumptions and called KL-optimality criterion. Meanwhile the determination of KL-optimal discriminating designs has been discussed by several authors [see Tommasi (2009); Tommasi and López-Fidalgo (2010) among others]. It is important to note here that the T -optimality criterion and its extensions are local optimality criteria in the sense of Chernoff (1953), because they require the explicit knowledge of the parameters in the model η 1. As a consequence, optimal designs with respect to the T -optimality criterion might be sensitive with respect to misspecification of the parameters [see Dette et al. (2012) for a striking example]. A standard approach to obtain robust designs [which was already mentioned by Atkinson and Fedorov (1975a)] is the use of a Bayesian T -optimality criterion. This criterion is defined as an expectation of various local T -optimality criteria with respect to a prior distribution. Dette et al. (2012) derived some explicit Bayesian T -optimal designs for polynomial regression models, but in general these designs have to be found numerically in nearly all cases of practical interest. Recently, Dette et al. (2015) pointed out that the numerical construction of Bayesian T -optimal designs is an extremely difficult optimization problem, because roughly speaking the Bayesian optimality criterion corresponds to an optimal design problem for model discrimination with an extremely large number of competing models. As a consequence, the commonly used algorithms for the calculation of optimal designs, such as exchange-type methods or multiplicative methods and their extensions, cannot be applied to determine the Bayesian T -optimal discriminating design in reasonable computational time. Dette et al. (2015) proposed a new algorithm for the calculation of Bayesian T -optimal discriminating designs and demonstrated its efficiency in several numerical examples. A drawback of this method consists still in the fact that it is only applicable to the classical Bayesian T -optimality criterion which refers to the nonlinear regression model (1.1) with homoscedastic and normally distributed responses, i.e. P Y x N (η(x, θ), v 2 (x, θ)). The purpose of the present paper is to extend the methodology introduced by Dette et al. (2015) to regression models with more general distributional assumptions. In Section 2 we will introduce a Bayesian KL-optimality criterion which extends the criterion introduced by López- Fidalgo et al. (2007) to address for uncertainty in the model parameters. The criterion has also been discussed in Tommasi and López-Fidalgo (2010), who considered only two competing regression models. The new algorithm is proposed in Section 3 and combines some features of the classical exchange type algorithms with gradient methods and quadratic programming. In Section 4 we illustrate the applicability of the new method in several examples. In particular, we determine optimal discriminating designs with respect to the Bayesian KL-optimality criterion in situations where all other methods fail to find the optimal design. Finally, the appendix contains a proof of an auxiliary result. 3

4 2 KL-optimal discriminating designs The regression model (1.1) is a special case of a more general model, where the distribution of the random variable Y has a density, say f(y, x), and x denotes an explanatory variable, which varies in a compact design space X. We assume that observations at different experimental conditions are independent. Following Kiefer (1974) we consider approximate designs that are defined as probability measures, say ξ, with finite support. The support points x 1,..., x k of a design ξ give the locations where observations are taken, while the weights ω 1,..., ω k describe the relative proportions of observations at these points. If an approximate design is given and n observations can be taken, a rounding procedure is applied to obtain integers n i (i = 1,..., k) from the not necessarily integer valued quantities ω i n such that k i=1 n i = n. Assume that the experimenter wants to choose a most appropriate model from a given class, say {f 1,..., f ν } of competing models, where f j (y, x, θ j ) denotes the density of the jth model with respect to a sigma-finite measure, say µ. The parameter θ j varies in a compact parameter space Θ j (j = 1,..., ν). The models may contain additional nuisance parameters, which will not be displayed in our notation. For two competing models, say f i and f j, we denote by (2.1) I i,j (x, θ i, θ j ) = f i (y, x, θ i ) log f i(y, x, θ i ) f j (y, x, θ j ) µ(dy) the Kullback-Leibler distance between f i and f j. If the model f i is assumed to be the true model with parameter θ i, then López-Fidalgo et al. (2007) defined a local KL-optimal discriminating design for the models f i and f j as a design maximizing the optimality criterion (2.2) KL i,j (ξ, θ i ) = inf I i,j (x, θ i, θ i,j )ξ(dx). θ i,j Θ j X This criterion can now easily be extended to construct optimal discriminating designs for more than two competing models. Following Tommasi and López-Fidalgo (2010) and Braess and Dette (2013) we denote by p i,j nonnegative weights reflecting the importance of the comparison between the the model f i and f j, where f i is assumed as the true model. The (symmetrized) KL-optimality criterion for more than ν 2 competing models f 1,..., f ν is then defined by (2.3) KL P (ξ) = p i,j KL i,j (ξ, θ i ) = p i,j inf I i,j (x, θ i, θ i,j )ξ(dx), θ i,j Θ j X and a design maximizing the criterion (2.3) is called local KL P -optimal discriminating design 4

5 for the models f 1,..., f ν. For a design ξ we also introduce the notation (2.4) Θ i,j(ξ) = arg inf I i,j (x, θ i, θ i,j )ξ(dx), θ i,j Θ j X Our first result characterizes local KL-optimal discriminating design and will be helpful to check the optimality of the numerically constructed designs. Its proof can be obtained by standard arguments and is therefore omitted. Theorem 2.1 Let Assumption 2.1 For each i = 1,..., ν the function f i (,, θ i ) is continuously differentiable with respect to the parameter θ i Θ i, be satisfied. A design ξ is a local KL P -optimal discriminating design, if and only if there exist distributions ρ ij on the sets Θ i,j(ξ ) defined in (2.4) such that the inequality (2.5) p i,j I i,j (x, θ i, θ i,j )ρ ij(dθ i,j ) KL P (ξ ) Θ i,j (ξ ) is satisfied for all x X. Moreover, there is equality in (2.5) for all support points of the local KL P -optimal discriminating design ξ. If, additionally, Assumption 2.2 For any design ξ such that KL P (ξ) > 0 and weight p i,j 0 the infima in (2.3) are attained at a unique points θ i,j = θ i,j (ξ) in the interior of the set Θ j, is satisfied, then all measures ρ ij in Theorem 2.1 are one-point measures and the left-hand side of inequality (2.5) simplifies to (2.6) Ψ(x, ξ) = p i,j I i,j (x, θ i, θ i,j ). Consequently, if ξ is not a local KL P -optimal discriminating design, it follows that there exists a point x X such that Ψ( x, ξ) > KL P (ξ). Note that the criterion (2.3) depends on the unknown parameters θ 1,..., θ ν, which have to be specified by the experimenter for the competing model f 1,..., f ν, respectively. Therefore 5

6 the criterion is a local one in the sense of Chernoff (1953). It was pointed out by Dette et al. (2012) that the optimal designs maximizing the criterion (2.3) are rather sensitive with respect to misspecification of these parameters. For this reason we will now propose a Bayesian version of the criterion in order to obtain robust discriminating designs for the competing models f 1,..., f ν. We denote by P i a prior distribution for the parameter θ i in model f i (i = 1,..., ν) and define a Bayesian KL-optimality criterion by (2.7) KL B P(ξ) = = p i,j KL i,j (ξ, θ i )P i (dθ i ), Θ i p i,j inf I i,j (x, θ i, θ i,j )ξ(dx)p i (dθ i ) Θ θ i,j Θ j i X Optimal designs maximizing this criterion will be called Bayesian KL-optimal discriminating designs throughout this paper. We also note that the criterion (2.7) has been considered before by Tommasi and López-Fidalgo (2010) in the case of two competing regression models. It was pointed out by Dette et al. (2015) that the determination of Bayesian optimal discriminating designs with respect to the criterion (2.7) is closely related to the problem of finding local optimal discriminating designs for a large class of competing regression models. To be precise, we note that in most applications the integral in (2.7) is evaluated by numerical integration approximating the prior distribution by a measure with finite support. Consequently, if the prior distribution P i in the criterion is given by a discrete measure with masses τ i1,... τ ili at the points λ i1,..., λ ili the criterion in (2.7) can be represented as (2.8) KL B P(ξ) = l i p i,j τ ik k=1 inf I i,j (x, λ ik, θ i,j )ξ(dx). θ i,j Θ j X which is a local KL-optimality criterion of the from (2.3), where the competing models are given by {f i (y, x, λ ik ) k = 1,..., l i ; i = 1,..., ν}. The only difference between the criterion obtained from the (discrete) Bayesian approach and the criterion (2.3) consists in the fact that - due to discretization of the prior distributions P 1,..., P ν - the criterion (2.8) involves substantially more comparisons of competing models f i (y, x, λ ik ). As a consequence the computation of Bayesian KL-optimal discriminating design is computationally very challenging, because for each support point of the prior distribution in the criterion (2.8) the infimum has to be calculated numerically. In the following section we will propose several new algorithms to address this problem. In Section 4 it will be demonstrated that these methods yield very satisfactory 6

7 results in cases where commonly used algorithms are either very slow or fail to determine the Bayesian KL-optimal discriminating design. 3 Efficient algorithms for Bayesian KL-optimal designs In this section we propose several algorithms for the calculation of Bayesian KL-optimal designs, which determine the optimal designs with reasonable accuracy and are computationally very efficient. As pointed out in Section 2 the Bayesian optimality criterion with a discrete prior distribution reduces to a local KL-optimality criterion of the form (2.3) with a large number of model comparisons. For this reason we will describe the numerical procedures in this section for the criterion (2.3). It is straightforward to extend the algorithms to the Bayesian criterion (2.8) and in the following Section 4 we will give some illustrations determining Bayesian KL-optimal discriminating designs by the new methods. Most of the algorithms proposed in the literature for the calculation of optimal designs are based on the fact which was mentioned in the paragraph following Theorem 2.1. More precisely, recall the definition of the function Ψ in (2.6) and assume that the design ξ is not a Bayesian KLoptimal discriminating design. It then follows under Assumption 2.2 that there exists a point x X, such that the inequality Ψ(x, ξ) > KL P (ξ) holds. López-Fidalgo et al. (2007) used this property to extend the algorithm of Atkinson and Fedorov (1975a) to the KL-optimality criterion. In the case of the local KL-optimality criterion (2.3) it reads as follows. Algorithm 3.1 Let ξ 0 denote a given (starting) design and let (α s ) s=0 be a sequence of positive numbers, such that lim s α s = 0, s=0 α s =, s=0 α2 s <. For s = 0, 1,... define where x s+1 = arg max x X Ψ(x, ξ s ). ξ s+1 = (1 α s )ξ s + α s ξ(x s+1 ), It can be shown that this algorithm yields a sequence of designs (ξ s ) s N converging in the sense that lim s KL P (ξ s ) = KL P (ξ ), where ξ denotes a local KL-optimal discriminating design. However, it turns out that the rate of convergence is very slow. In particular, if there are many models under consideration, the algorithm is very slow and fails in some models to determine the local KL-optimal discriminating design (see our numerical example in Section 4). One reason for these difficulties consists in the fact that Algorithm 3.1 usually yields a sequence of 7

8 designs with an increasing number of support points. As a consequence the resulting design (after applying some stopping criterion) is concentrated on a large set of points. In the case of normal distributed responses it is also demonstrated by Braess and Dette (2013) that Algorithm 3.1 requires a large number of iterations if it is used for the calculation of local KL-optimal discriminating designs for more than two competing models. Following Dette et al. (2015) we therefore propose an alternative procedure for the calculation of local KL-optimal discriminating designs, which separates the maximization with respect to the support points and weights in two steps. In the discussion below we will present two methods for the calculation of the weights in the second step [see Section 3.1 and 3.2 for details]. Algorithm 3.2 Let ξ 0 denote a starting design such that KL P (ξ 0 ) > 0 and define recursively a sequence of designs (ξ s ) s=0,1,... as follows: (1) Let S [s] denote the support of the design ξ s. Determine the set E [s] of all local maxima of the function Ψ(x, ξ s ) on the design space X and define S [s+1] = S [s] E [s]. (2) Define ξ = {S [s+1], ω} as the design supported at S [s+1] (with a normalized vector w of non-negative weights) and determine the local KL P -optimal design in the class of all designs supported at S [s+1]. In other words: we determine the vector ω [s+1] maximizing the function g(ω) = KL P ({S [s+1], ω}) = p i,j inf θ i,j Θ j x S [s+1] I i,j (x, θ i, θ i,j )w x (here w x denotes the weights at the point x S s+1 ). All points in S [s+1] with vanishing components in the vector of weights ω [s+1] will be be removed and the new set of support points will also be denoted by S [s+1]. Finally the design ξ s+1 is defined as the design with the set of support points S [s+1] and the corresponding nonzero weights. It follows by similar arguments as given in Dette et al. (2015) that the sequence (ξ s ) s=0,1,... of designs generated by Algorithm 3.2 converges to a local KL-optimal discriminating design. The crucial step in this algorithm is the second one, because it requires in particular if a large number of competing models are under consideration the calculation of numerous infima. In order to address this problem we propose a quadratic programming and a gradient method in the following two subsections. 8

9 3.1 Quadratic programming Let S [s+1] = {x 1,..., x n } denote the set obtained in the first step of Algorithm 3.2 and recall the definition of the Kullback-Leibler distance I i,j in (2.1). In Step 2 of Algorithm 3.2 a design ξ with masses ω 1,..., ω n at the points x 1,..., x n has to be determined such that the function g(ω) = p i,j n ω k k=1 { fi (y, x k, θ i ) } log f j (y, x k, θ f i (y, x k, θ i )dµ(y). i,j ) is maximal, where (3.1) θ i,j = θ i,j (ω) = arg inf θ i,j Θ j n ω k k=1 { } fi (y, x k, θ i ) log f i (y, x k, θ i ). f j (y, x k, θ i,j ) Define f i (y, x k, θ i ) J i,j (y) = J i,j ( θ i,j, y) = f j (y, x k, θ i,j ) R i,j = R i,j ( θ i,j ) = f j (y, x k, θ i,j ) θ i,j θi,j = θ i,j ( fj (y, x k, θ i,j ) θ i,j θi,j = θ i,j f i (y, x k, θ i ) n k=1 ) n f j (y, x k, θ i,j ) dµ(y) R n d j, k=1 R n d j, and consider a linearized version of the function g, that is (3.2) g(ω) = { n { } f i (y, x k, θ i ) p i,j min ω k log α i,j k=1 f j (y, x k, θ f i (y, x k, θ i )dµ(y) i,j ) + αi,j(r T T i,j) k 1 } α T 2 i,j(j T i,j(y)) k (J i,j (y)) k α i,j dµ(y). Note that the minimum with respect to the parameters α i,j R d j ( α i,j = ) 1 J T i,j(y)ωj i,j (y)dµ(dy) R T i,jω, is achieved for where the matrix Ω is defined by Ω = diag(ω 1,..., ω n ) and ω = (ω 1,..., ω n ) T. For the following discussion we define by = {ω R n ω i 0 (i = 1,..., n) n i=1 ω i = 1} the simplex in R n. Lemma 3.3 If Assumptions 2.1 and 2.2 are satisfied, then each maximizer of the function g( ) 9

10 in is a maximizer of g( ) in and vice versa. Moreover, max g(ω) = max g(ω). ω ω A proof of Lemma 3.3 can be found in the Section 5. With the notations b i,j = b i,j ( θ i,j ) = ( log { f i (y, x k, θ i ) f j (y, x k, θ i,j ) } f i (y, x k, θ i )µ(dy) ) n k=1 = ( I i,j (x k, θ i, θ ) n i,j ) k=1 R n we have g(ω) = b T ω ω T Q(ω)ω where the vector b R n and the n n matrix Q are defined by Q(ω) = R i,j ( ) 1 J T i,j(y)ω(ω)j i,j (y)µ(dy) R T i,j, and b = Σ ν p i,j b i,j, respectively. If we ignore the dependence of the matrix Q(ω) and consider this matrix as fixed for a given matrix Ω = diag(ω 1,..., ω n ), we obtain a quadratic programming problem, that is (3.3) φ(ω, ω) = ω T Q(ω) ω + b T ω max ω. This problem can now be solved iteratively substituting each time the solution obtained in the previous iteration instead of ω. Example 3.4 In this example we illustrate the calculation of the function Q(ω) under several distributional assumptions. (1) Ucinski and Bogacka (2005) considered the regression model with normal distributed heteroscedastic errors, that is f(y, x, θ) = 1 ( 2πv2 (x, θ) exp (y η(x, ) θ))2, 2v 2 (x, θ) where η(x, θ) and v 2 (x, θ) denotes the expectation and variance of the response at experimental condition x. In this case the Kullback-Leibler distance between the two densities 10

11 f i and f j is given by I i,j (x, θ i, θ i,j ) = [ ηi (x, θ i ) η j (x, θ i,j ) ] 2 v 2 j (x, θ i,j) + v2 i (x, θ i ) v 2 j (x, θ i,j) + log { v 2 } j (x, θ i,j ) vi 2(x, θ 1, i) and a straightforward calculation gives for the function ḡ in (3.2) the representation g(ω) = where Ω i = diag (ω 1,..., ω n ), { s i,j (x, θ i,j ) = v2 i (x, θ i ) v 2 } vj 2(x, θ i,j) + log j (x, θ i,j ) vi 2(x, θ i) ( ) hi,j (x k, θ i,j ) J i,j = R i,j = b i,j = θ i,j min α i,j [ α T i,j J T i,jω i J i,j α i,j + 2ω T R i,j α i,j + b T i,jω ], θi,j = θ i,j ( h i,j (x k, θ i,j ) h i,j(x k, θ i,j ) ( I i,j (x k, θ i, θ ) i,j ) k=1,...,n ; k=1,...,n θ i,j θi,j = θ i,j ; h i,j (x, θ i,j ) = η i(x, θ i ) η j (x, θ i,j ) ; v j (x, θ i,j ) s i,j (x k, θ i,j ) ) ; θ i,j θi,j = θ i,j k=1,...,n (3) López-Fidalgo et al. (2007) considered the regression model (1.1) with log-normal distribution with parameters µ(x, θ) and σ 2 (x, θ). This means that the mean and the variance are given by E[Y ] = { σ 2 (x, θ) } η(x, θ) = exp + µ(x, θ), 2 Var(Y ) = v 2 (x, θ) = η 2 (x, θ){exp{σ 2 (x, θ)} 1}, respectively, and the density of the response Y is given by f(y, x, θ) = 1 { x 2πσ(x, θ) exp {log(y) µ(x, } θ)}2 2σ 2 (x, θ) In the paper López-Fidalgo et al. (2007) it was shown that the Kullback-Leibler distance between two log-normal densities with parameters µ l (x, θ l ) and σ 2 l (x, θ l) (l = i, j) is given 11

12 by (3.4) where and I ij (x, θ i, θ i,j ) = 1 2 {s i,j (x, θ i,j ) + [µ i(x, θ i ) µ j (x, θ i,j )] 2 σ 2 i (x, θ i) [ σ 2 s i,j (x, θ i,j ) = log i (x, θ i ) ] σj 2(x, θ + σ2 j (x, θ i,j ) i,j) σi 2(x, θ i), σ 2 i (x, θ i ) = log [ 1 + v 2 i (x, θ i )/η 2 i (x, θ i ) ], µ i (x, θ i ) = log [η i (x, θ i )] σ 2 i (x, θ i )/2. } 1, Now a straightforward calculation gives for the function ḡ in (3.2) the representation g(ω) = 1 2 min α i,j [ α T i,j J T i,jω i J i,j α i,j 2ω T R i,j α i,j + b T i,jω ], where Ω i = diag (ω 1,..., ω n ), J i,j = µ j (x k,θ i,j ) θ i,j θi,j = θ i,j σ i (x 1, θ i ) k=1,...,n [ µi (x k, θ i ) µ j (x k, θ i,j ) ] µ i (x k,θ i,j ) θ R i,j = i,j σi 2(x k, θ i ) ( b i,j = I i,j (x k, θ i, θ ) i,j ) k=1,...,n ; θi,j = θ i,j 1 2 s i,j (x k, θ i,j ) θ i,j θi,j = θ i,j k=1,...,n ; 3.2 A gradient method In this section we describe a specialized gradient method for second step of Algorithm 3.2 function To be precise we introduce theunctions v k (ω) = p i,j I i,j (x k, θ i, θ i,j (ω)), k = 1,..., n, 12

13 where θ i,j = θ i,j (ω) is defined in (3.1). Next we iteratively calculate a sequence of vectors (ω (γ) ) γ=0,1,... starting with a vector ω (0) = ω (for example equal weights). For ω (γ) = (ω (γ),1,..., ω (γ),n ) we determine indices k and k corresponding to max 1 k n v k (ω (γ) ) and min 1 k n v k (ω (γ) ), respectively, and define (3.5) α = arg max g(ω (γ) (α)), 0 α ω (γ),k where the vector ω (γ) (α) = (ω (γ),1 (α),..., ω (γ),n (α)) is given by ω (γ),i + α ω (γ),i (α) = ω (γ),i α ω (γ),i if i = k if i = k else The vector ω (γ+1) of the next iteration is then defined by ω (γ+1) = ω (γ) (α ). It follows by similar arguments as in Dette et al. (2015) that the generated sequence of vectors converges to a maximizer of the function g. 4 Implementation and numerical examples In this section we illustrate the new algorithms calculating Bayesian KL-optimal discriminating designs for several models with non-normal errors. We begin giving a few more details regarding the implementation. (1) As pointed out in Section 2 a Bayesian KL-optimality criterion is reduced to a local criterion of the form (2.3) for a large number of model comparisons. For illustration purposes, consider the criterion (2.8), where ν = 2, p 1,2 = 1, p 2,1 = 0 and the prior for the parameter θ 1 puts masses τ 1,... τ l at the points λ 1,..., λ l. This criterion can be rewritten as a local criterion of the form (2.3), i.e. (4.1) KL P (ξ) = l+1 p i,j inf I i,j (x, θ i, θ i,j )ξ(dx), θ i,j Θ j X where p 1,l+1 = τ 1,..., p 1,l+1 = τ l and all other weights p i,j are 0. The extension of this approach to more than two models is easy and left to the reader. (2) In Step 1 of Algorithm 3.2 all local maxima of the function Ψ(x, ξ s ) are aded as possible support points of the design in the next iteration. In order to avoid the problem of 13

14 accumulating too many support points we remove in each iteration those points with a weight smaller than m 0.25, where m = is the working precision R. (3) In the implementation of the quadratic programming method for Step 2 of Algorithm 3.2 (see Section 3.1) we perform only a few iterations such that an improvement compared to the starting design is obtained. This speeds up the convergence of the procedure substantially without affecting the convergence in all examples under consideration. (4) In the implementation of the gradient method for Step 2 of Algorithm 3.2 (see Section 3.2) we use a linearization of the optimization problem in order to improve the speed of the procedure. We are now ready to demonstrate the advantages of the new method in several examples calculating Bayesian KL-optimal discriminating designs. For the sake of brevity we restrict ourselves to the case of non-linear regression models, where the response has a log-normal distribution with parameters µ(x, θ) and σ 2 (x, θ) as described in Example 3.4. Example 4.1 Our first example refers the problem of determining local KL-optimal designs for a situation investigated by López-Fidalgo et al. (2007). Motivated by pharmacokinetic practice [see Lindsey et al. (2001); Crawley (2002)] these authors determined local KL-optimal designs for two log-normal models with mean functions (4.2) η 1 (x, θ 1 ) = θ 1,1x θ 1,2 + x + θ 1,3x, η 2 (x, θ 2 ) = θ 2,1x θ 2,2 + x on the interval X = [0.1, 5]. They assumed equal and constant variances and considered model η 1 with parameter θ 1 = (1, 1, 1) as fixed. This corresponds to the choice ν = 2 and p 1,2 = 1, p 2,1 = 0 in the criterion (2.3). In Table 1 we resent the optimal design calculating by the new algorithms for various choices of the mean and variance function, that is (4.3) (1) v1(x, 2 θ 1 ) = v2(x, 2 θ 2 ) = 1 (2) σ1(x, 2 θ 1 ) = σ2(x, 2 θ 2 ) = 1 (3) vi 2 (x, θ i ) = exp(η i (x, θ i )) All designs have an efficiency that is at least 0.999, and we have used three methods for the calculation of the local KL-optimal design. The first procedure is a classical exchange type method as proposed by López-Fidalgo et al. (2007). The other methods are the two versions of the new Algorithm 3.2 with the modifications described in Section 3.1 (quadratic programming) and 3.2 (gradient method). For the case (2) of equal variances in (4.3) the corresponding 14

15 function in (3.4) simplifies to (µ 1 (x, θ 1 ) µ 2 (x, θ 1,2 )) 2. Consequently, one can use the procedure for the special case of a normal distributed response developed in Dette et al. (2015), where µ i (x, θ i ) = log η i (x, θ i ) (i = 1, 2), which works significantly faster. It should be noted that the calculated designs slightly differ from those in López-Fidalgo et al. (2007). In Table 1 we also show the computation time (CPU time in seconds on a standard PC with an intel core i7-4790k processor) for the different methods. We observe that the methods developed in this paper work substantially faster than the exchange type algorithm proposed in López-Fidalgo et al. (2007). For example, the new gradient methods are between 5 and 30 times faster, while the quadratic programming approach yield to a procedure which is between 25 and 120 times faster than the classical exchange type algorithm. In the case of two competing models the exchange type algorithm is still finding the local KL-optimal discriminating design in a reasonable time, but the difference become more important if a discriminating design has to be found for more than two competing models or if a Bayesian KL-optimal design has to be determined. Some of these situations are discussed in the following examples. case KL-opt. design AF grad quad (1) (2) (3) Table 1: Local KL-optimal discriminating designs for the models in (4.2). The responses are log-normal distributed with different specifications of the mean and variance - see (4.3). Column 3-5 show the computation time of the different algorithms (exchange type Algorithm 3.1 (AF) proposed in López-Fidalgo et al. (2007) and Algorithm 3.2 with a gradient (grad) and quadratic programming method (quad) in Step 2). Example 4.2 In our second example we calculate Bayesian KL-optimal discriminating design for two competing exponential models (4.4) η 1 (x, θ 1 ) = θ 1,1 θ 1,2 exp( θ 1,3 x θ 1,4 ); η 2 (x, θ 2 ) = θ 2,1 θ 2,2 exp( θ 2,3 x). on the interval [0, 10], where model η 1 is again fixed. Discriminating designs for these models have been determined by Dette et al. (2015) under the assumption of a normal distribution, 15

16 Psi Psi Psi x x x (1) (2) (3) Figure 1: The function on the left hand side of inequality (2.5) in the equivalence Theorem 2.1 for the numerically calculated Bayesian KL-optimal discriminating designs. The competing regression models are given in (4.4). and we will now investigate how the designs change for the log-normal distributed responses with mean and variance specified by (4.3). Following these authors we considered independent prior distributions supported at the points 0.3(i 3) (4.5) µ j +, i = 1,..., 5 ; j = 3, 4, 2 for the parameters θ 1,3 and θ 1,4 where µ 3 = 0.8, µ 4 = 1.5. The corresponding weights at these points are proportional (in both cases) to (4.6) 1 (i ) 3)2 exp ( ; i = 1,..., 5. 2π Note that the optimal discriminating designs do not depend on the linear parameters of η 1, for which we have chosen as θ 2,1 = 2 and θ 2,2 = 1. The Bayesian KL-optimal discriminating designs for log-normal distributed responses are displayed in Table 2 for the different specifications of the mean and variance in (4.3). In Figure 1 we show the function on the left hand side of inequality (2.5) in the equivalence Theorem 2.1. Comparing the computational times in Table 2 we observe again that using quadratic programming in Step 2 of Algorithm 3.2 is substantially faster than the gradient method. It might be of interest to compare the Bayesian optimal discriminating designs for the various log-normal distributed responses with the design for normal distributed responses determined in Dette et al. (2015). This design is supported at the five(!) points 0.000, 0.452, 1.747,

17 and with masses 0.207, 0.396, 0.292, and 0.102, respectively. The efficiencies Eff (j) KL P (ξ (i)) = KL(j) P (ξ (i) ) sup η KL (j) P (η). under misspecification of the distribution of the response are depicted in Table 3. For example, the efficiency of the design ξ(0) calculated under the assumption homoscedastic normal distributed responses in the model with log-normal distributed responses in (4.3)(2) is given by 95.3%. We observe that the Bayesian optimal discriminating designs calculated for normal distributed responses are rather robust and have good efficiencies for the log-normal distribution. (4.3) design AF grad quad (1) (2) (3) Table 2: Bayesian KL-optimal discriminating designs for the models in (4.4). The responses are log-normal distributed with different specifications of the mean and variance - see (4.3). The prior distribution is defined by (4.5) and (4.6). Column three and four show the computation time of the new Algorithm 3.2 proposed in this paper with a gradient (grad) and quadratic programming method (quad) in Step 2. (0) (1) (2) (3) (0) (1) (2) (3) Table 3: Efficiencies of Bayesian KL-optimal discriminating designs for the models in (4.4) under different distributional assumptions for the responses. (0): homoscedastic normal distribution; (1) - (3): log-normal distribution with different specifications of the mean and variance - see (4.3). 17

18 (4.9) design AF grad quad (1) (2) (3) Table 4: Bayesian KL-optimal discriminating designs for the competing dose response models in (4.7). The responses are log-normal distributed with different specifications of the mean and variance - see (4.9). The prior distribution is a uniform distribution on 81 points as specified in (4.8). Column three, four and five show the computation time of the exchange type algorithm (AF), the new Algorithm 3.2 proposed in this paper with a gradient (grad) and quadratic programming method (quad) in Step 2. Example 4.3 Our final example refers to the construction of Bayesian KL-optimal discriminating designs for several dose response curves, which have been recently proposed by Pinheiro et al. (2006) for modeling the dose response relationship of a Phase II clinical trial, that is (4.7) η 1 (x, θ 1 ) = θ 1,1 + θ 1,2 x; η 2 (x, θ 2 ) = θ 2,1 + θ 2,2 x(θ 2,3 x); η 3 (x, θ 3 ) = θ 3,1 + θ 3,2 x/(θ 3,3 + x); η 4 (x, θ 4 ) = θ 4,1 + θ 4,2 /(1 + exp(θ 4,3 x)/θ 4,4 ); where the designs space (dose range) is given by the interval X = [0, 500]. In this reference some prior information regarding the parameters for the models is also provided., that is θ 1 = (60, 0.56), θ 2 = (60, 7/2250, 600), θ 3 = (60, 294, 25), θ 4 = (49.62, , 150, 45.51). Dette et al. (2015) determined Bayesian KL-optimal discriminating designs for these models under the assumption of normal distributed responses, where they used p i,j = 1/6, (1 j < i 4) and they assumed that there exist only uncertainty for the parameter θ 4. We will now consider similar problems for log-normal distributed responses, where the prior distribution is a uniform distribution at 81 points in R 4, that is (4.8) ( c 1, c 2, c 3, c 4 ) with c 1, c 2, c 3, c 4 { 20, 0, 45}. Note that we cannot use the prior distribution considered in 18

19 Dette et al. (2015) because this would yield a negative mean η i (x, θ i ). The resulting Bayesian optimality criterion (2.8) consist of 246 model comparisons and Bayesain KL optimal discriminating designs are depicted in Table 4 for the cases (4.9) (1) v1(x, 2 θ 1 ) = 1, i = 1, 2, 3, 4; (2) σi 2 (x, θ i ) = 1, i = 1, 2, 3, 4; (3) vi 2 (x, θ i ) = exp(η i (x, θ i )/100), i = 1, 2, 3, 4. All calculated designs have at least efficiency 99.9% and the corresponding plots of the equivalence Theorem 2.1 are shown in Figure 2. In the models specified by (4.7) all new algorithms were able to find the Bayesian KL-optimal discriminating design, where the exchange type algorithm failed in the case (4.9)(2). Moreover, in the other cases the new methods are substantially faster than the exchange type method. For example, the gradient method yields only 25% 30% of the computational time, while the quadratic programming approach is about times faster. Psi Psi Psi x x x (1) (2) (3) Figure 2: The function on the left hand side of inequality (2.5) in the equivalence Theorem 2.1 for the numerically calculated Bayesian KL-optimal discriminating designs. The competing regression models are given in (4.7) and the scenarios for log-normal distribution specified in (4.9). 19

20 5 Appendix: Proof of Lemma 3.3 By the construction of the function g we have g(ω) g(ω) for all ω. Let ω arg max ω g(ω) and define the function Ψ i,j (θ i,j, ω) = n ω k k=1 { } fi (x k, y, θ i ) log f i (x k, y, θ i )dµ(dy) f j (x k, y, θ i,j ) If θ i,j = argmin θi,j Θ j Ψ i,j (θ i,j, ω ) is the miminizer of Ψ i,j it follows from Assumption 2.2 that Ψ i,j (θ i,j, ω ) = R i,j ( θ i,j )ω = 0, θ i,j θi,j = θ i,j and we obtain ( α i,j = ) 1 J T i,j(y)ωj i,j (y)dy R T i,j( θ i,j )ω = 0 Inserting this value in (3.2) gives g(ω ) = g(ω ), i.e. max ω g(ω) = max ω g(ω). Now let ω = arg max ω g(ω). From the above equality it follows that α i,j = 0 and therefore R i,j ( θ i,j )ω = 0, that is ω = arg max g(ω). Acknowledgements. Parts of this work were done during a visit of the second author at the Department of Mathematics, Ruhr-Universität Bochum, Germany. The authors would like to thank M. Stein who typed this manuscript with considerable technical expertise. The work of H. Dette and V. Melas was supported by the Deutsche Forschungsgemeinschaft (SFB 823: Statistik nichtlinearer dynamischer Prozesse, Teilprojekt C2). The research of H. Dette reported in this publication was also partially supported by the National Institute of General Medical Sciences of the National Institutes of Health under Award Number R01GM The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health. The work of V. Melas and R. Guchenko was also partially supported by by St. Petersburg State University (project Actual problems of design and analysis for regression models, ). 20

21 References Atkinson, A., Donev, A., and Tobias, R. (2007). Optimum Experimental Designs, with SAS (Oxford Statistical Science Series). Oxford University Press, USA, 2nd edition. Atkinson, A. C. and Fedorov, V. V. (1975a). The designs of experiments for discriminating between two rival models. Biometrika, 62: Atkinson, A. C. and Fedorov, V. V. (1975b). Optimal design: Experiments for discriminating between several models. Biometrika, 62: Biedermann, S., Dette, H., and Pepelyshev, A. (2006). Some robust design strategies for percentile estimation in binary response models. Canadian Journal of Statistics, 34: Braess, D. and Dette, H. (2013). Optimal discriminating designs for several competing regression models. Annals of Statistics, 41(2): Chernoff, H. (1953). Locally optimal designs for estimating parameters. Annals of Mathematical Statistics, 24: Crawley, M. J. (2002). Statistical Computing: an Introduction to Data Analysis using S-Plus. Wiley, New York. Dette, H. (1990). A generalization of D- and D1-optimal designs in polynomial regression. Annals of Statistics, 18: Dette, H., Bretz, F., Pepelyshev, A., and Pinheiro, J. (2008). Optimal designs for dose-finding studies. Journal of the American Statistical Association, 104(483): Dette, H. and Haller, G. (1998). Optimal designs for the identification of the order of a Fourier regression. Annals of Statistics, 26: Dette, H., Melas, V. B., and Guchenko, R. (2015). Bayesian T -optimal discriminating designs. Annals of Statistics, to appear,. Dette, H., Melas, V. B., and Shpilev, P. (2012). T-optimal designs for discrimination between two polynomial models. Annals of Statistics, 40(1): Kiefer, J. (1974). General equivalence theory for optimum designs (approximate theory). Annals of Statistics, 2(5): Läuter, E. (1974). Experimental design in a class of models. Math. Operationsforsch. Statist., 5(4 & 5): Lindsey, J. K., Jones, B., and Jarvis, P. (2001). Some statistical issues in modelling pharmacokinetic data. Statistics in Medicine, 20: López-Fidalgo, J., Tommasi, C., and Trandafir, P. C. (2007). An optimal experimental design criterion for discriminating between non-normal models. Journal of the Royal Statistical Society, Series B, 69: Pinheiro, J., Bretz, F., and Branson, M. (2006). Analysis of dose-response studies: Modeling approaches. In Ting, N., editor, Dose Finding in Drug Development, pages Springer-Verlag, New York. Pukelsheim, F. (2006). Optimal Design of Experiments. SIAM, Philadelphia. Song, D. and Wong, W. K. (1999). On the construction of g rm -optimal designs. Statistica Sinica, 9: Stigler, S. (1971). Optimal experimental design for polynomial regression. Journal of the American Statistical Association, 66: Tommasi, C. (2009). Optimal designs for both model discrimination and parameter estimation. Journal of 21

22 Statistical Planning and Inference, 139: Tommasi, C. and López-Fidalgo, J. (2010). Bayesian optimum designs for discriminating between models with any distribution. Computational Statistics & Data Analysis, 54(1): Ucinski, D. and Bogacka, B. (2005). T -optimum designs for discrimination between two multiresponse dynamic models. Journal of the Royal Statistical Society, Ser. B, 67:

Optimal discrimination designs

Optimal discrimination designs Optimal discrimination designs Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@ruhr-uni-bochum.de Stefanie Titoff Ruhr-Universität Bochum Fakultät

More information

Optimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models

Optimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models Optimum Designs for the Equality of Parameters in Enzyme Inhibition Kinetic Models Anthony C. Atkinson, Department of Statistics, London School of Economics, London WC2A 2AE, UK and Barbara Bogacka, School

More information

Optimal designs for rational regression models

Optimal designs for rational regression models Optimal designs for rational regression models Holger Dette, Christine Kiss Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany holger.dette@ruhr-uni-bochum.de tina.kiss12@googlemail.com

More information

Optimal designs for estimating the slope of a regression

Optimal designs for estimating the slope of a regression Optimal designs for estimating the slope of a regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 4480 Bochum, Germany e-mail: holger.dette@rub.de Viatcheslav B. Melas St. Petersburg

More information

A geometric characterization of c-optimal designs for heteroscedastic regression

A geometric characterization of c-optimal designs for heteroscedastic regression A geometric characterization of c-optimal designs for heteroscedastic regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Tim Holland-Letz

More information

Optimal designs for multi-response generalized linear models with applications in thermal spraying

Optimal designs for multi-response generalized linear models with applications in thermal spraying Optimal designs for multi-response generalized linear models with applications in thermal spraying Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum Germany email: holger.dette@ruhr-uni-bochum.de

More information

Optimal designs for estimating the slope in nonlinear regression

Optimal designs for estimating the slope in nonlinear regression Optimal designs for estimating the slope in nonlinear regression Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Viatcheslav B. Melas St.

More information

Efficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies

Efficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies Efficient algorithms for calculating optimal designs in pharmacokinetics and dose finding studies Tim Holland-Letz Ruhr-Universität Bochum Medizinische Fakultät 44780 Bochum, Germany email: tim.holland-letz@rub.de

More information

Compound Optimal Designs for Percentile Estimation in Dose-Response Models with Restricted Design Intervals

Compound Optimal Designs for Percentile Estimation in Dose-Response Models with Restricted Design Intervals Compound Optimal Designs for Percentile Estimation in Dose-Response Models with Restricted Design Intervals Stefanie Biedermann 1, Holger Dette 1, Wei Zhu 2 Abstract In dose-response studies, the dose

More information

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses

460 HOLGER DETTE AND WILLIAM J STUDDEN order to examine how a given design behaves in the model g` with respect to the D-optimality criterion one uses Statistica Sinica 5(1995), 459-473 OPTIMAL DESIGNS FOR POLYNOMIAL REGRESSION WHEN THE DEGREE IS NOT KNOWN Holger Dette and William J Studden Technische Universitat Dresden and Purdue University Abstract:

More information

Optimal designs for the Michaelis Menten model with correlated observations

Optimal designs for the Michaelis Menten model with correlated observations Optimal designs for the Michaelis Menten model with correlated observations Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Joachim Kunert

More information

MAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS

MAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS Statistica Sinica 17(2007), 463-480 MAXIMIN AND BAYESIAN OPTIMAL DESIGNS FOR REGRESSION MODELS Holger Dette, Linda M. Haines and Lorens A. Imhof Ruhr-Universität, University of Cape Town and Bonn University

More information

Exact optimal designs for weighted least squares analysis with correlated errors

Exact optimal designs for weighted least squares analysis with correlated errors Exact optimal designs for weighted least squares analysis with correlated errors Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Joachim Kunert

More information

Locally optimal designs for errors-in-variables models

Locally optimal designs for errors-in-variables models Locally optimal designs for errors-in-variables models Maria Konstantinou, Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany maria.konstantinou@ruhr-uni-bochum.de holger.dette@ruhr-uni-bochum.de

More information

EXACT OPTIMAL DESIGNS FOR WEIGHTED LEAST SQUARES ANALYSIS WITH CORRELATED ERRORS

EXACT OPTIMAL DESIGNS FOR WEIGHTED LEAST SQUARES ANALYSIS WITH CORRELATED ERRORS Statistica Sinica 18(2008), 135-154 EXACT OPTIMAL DESIGNS FOR WEIGHTED LEAST SQUARES ANALYSIS WITH CORRELATED ERRORS Holger Dette, Joachim Kunert and Andrey Pepelyshev Ruhr-Universität Bochum, Universität

More information

Discussion Paper SFB 823. Nearly universally optimal designs for models with correlated observations

Discussion Paper SFB 823. Nearly universally optimal designs for models with correlated observations SFB 823 Nearly universally optimal designs for models with correlated observations Discussion Paper Holger Dette, Andrey Pepelyshev, Anatoly Zhigljavsky Nr. 3/212 Nearly universally optimal designs for

More information

Discussion Paper SFB 823. Optimal designs for the EMAX, log-linear linear and exponential model

Discussion Paper SFB 823. Optimal designs for the EMAX, log-linear linear and exponential model SFB 823 Optimal designs for the EMAX, log-linear linear and exponential model Discussion Paper Holger Dette, Christine Kiss, Mirjana Bevanda, Frank Bretz Nr. 5/2009 Optimal designs for the EMAX, log-linear

More information

Optimal designs for the EMAX, log-linear and exponential model

Optimal designs for the EMAX, log-linear and exponential model Optimal designs for the EMAX, log-linear and exponential model Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@ruhr-uni-bochum.de Mirjana Bevanda

More information

Optimal designs for enzyme inhibition kinetic models

Optimal designs for enzyme inhibition kinetic models Optimal designs for enzyme inhibition kinetic models Kirsten Schorning, Holger Dette, Katrin Kettelhake, Tilman Möller Fakultät für Mathematik Ruhr-Universität Bochum 44780 Bochum Germany Abstract In this

More information

Optimum designs for model. discrimination and estimation. in Binary Response Models

Optimum designs for model. discrimination and estimation. in Binary Response Models Optimum designs for model discrimination and estimation in Binary Response Models by Wei-Shan Hsieh Advisor Mong-Na Lo Huang Department of Applied Mathematics National Sun Yat-sen University Kaohsiung,

More information

D-optimal Designs for Factorial Experiments under Generalized Linear Models

D-optimal Designs for Factorial Experiments under Generalized Linear Models D-optimal Designs for Factorial Experiments under Generalized Linear Models Jie Yang Department of Mathematics, Statistics, and Computer Science University of Illinois at Chicago Joint research with Abhyuday

More information

Optimal and efficient designs for Gompertz regression models

Optimal and efficient designs for Gompertz regression models Ann Inst Stat Math (2012) 64:945 957 DOI 10.1007/s10463-011-0340-y Optimal and efficient designs for Gompertz regression models Gang Li Received: 13 July 2010 / Revised: 11 August 2011 / Published online:

More information

AP-Optimum Designs for Minimizing the Average Variance and Probability-Based Optimality

AP-Optimum Designs for Minimizing the Average Variance and Probability-Based Optimality AP-Optimum Designs for Minimizing the Average Variance and Probability-Based Optimality Authors: N. M. Kilany Faculty of Science, Menoufia University Menoufia, Egypt. (neveenkilany@hotmail.com) W. A. Hassanein

More information

Model identification for dose response signal detection

Model identification for dose response signal detection SFB 823 Model identifi fication for dose response signal detection Discussion Paper Holger Dette, Stefanie Titoff, Stanislav Volgushev, Frank Bretz Nr. 35/2012 Model identification for dose response signal

More information

On Multiple-Objective Nonlinear Optimal Designs

On Multiple-Objective Nonlinear Optimal Designs On Multiple-Objective Nonlinear Optimal Designs Qianshun Cheng, Dibyen Majumdar, and Min Yang December 1, 2015 Abstract Experiments with multiple objectives form a staple diet of modern scientific research.

More information

MINIMAX OPTIMAL DESIGNS IN NONLINEAR REGRESSION MODELS

MINIMAX OPTIMAL DESIGNS IN NONLINEAR REGRESSION MODELS Statistica Sinica 8(998, 49-64 MINIMAX OPTIMAL DESIGNS IN NONLINEAR REGRESSION MODELS Holger Dette and Michael Sahm Ruhr-Universität Bochum Astract: We consider the maximum variance optimality criterion

More information

On construction of constrained optimum designs

On construction of constrained optimum designs On construction of constrained optimum designs Institute of Control and Computation Engineering University of Zielona Góra, Poland DEMA2008, Cambridge, 15 August 2008 Numerical algorithms to construct

More information

A new approach to optimal designs for models with correlated observations

A new approach to optimal designs for models with correlated observations A new approach to optimal designs for models with correlated observations Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@rub.de Andrey Pepelyshev

More information

Heteroscedastic T-Optimum Designs for Multiresponse Dynamic Models

Heteroscedastic T-Optimum Designs for Multiresponse Dynamic Models Heteroscedastic T-Optimum Designs for Multiresponse Dynamic Models Dariusz Uciński 1 and Barbara Bogacka 2 1 Institute of Control and Computation Engineering, University of Zielona Góra, ul. Podgórna 50,

More information

OPTIMAL DESIGN OF EXPERIMENTS WITH POSSIBLY FAILING TRIALS

OPTIMAL DESIGN OF EXPERIMENTS WITH POSSIBLY FAILING TRIALS Statistica Sinica 12(2002), 1145-1155 OPTIMAL DESIGN OF EXPERIMENTS WITH POSSIBLY FAILING TRIALS Lorens A. Imhof, Dale Song and Weng Kee Wong RWTH Aachen, Scirex Corporation and University of California

More information

Designing dose finding studies with an active control for exponential families

Designing dose finding studies with an active control for exponential families Designing dose finding studies with an active control for exponential families Holger Dette, Katrin Kettelhake Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum, Germany e-mail: holger.dette@ruhr-uni-bochum.de

More information

ON THE NUMBER OF SUPPORT POINTS OF MAXIMIN AND BAYESIAN OPTIMAL DESIGNS 1. BY DIETRICH BRAESS AND HOLGER DETTE Ruhr-Universität Bochum

ON THE NUMBER OF SUPPORT POINTS OF MAXIMIN AND BAYESIAN OPTIMAL DESIGNS 1. BY DIETRICH BRAESS AND HOLGER DETTE Ruhr-Universität Bochum The Annals of Statistics 2007, Vol. 35, No. 2, 772 792 DOI: 10.1214/009053606000001307 Institute of Mathematical Statistics, 2007 ON THE NUMBER OF SUPPORT POINTS OF MAXIMIN AND BAYESIAN OPTIMAL DESIGNS

More information

D-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors

D-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors Journal of Data Science 920), 39-53 D-Optimal Designs for Second-Order Response Surface Models with Qualitative Factors Chuan-Pin Lee and Mong-Na Lo Huang National Sun Yat-sen University Abstract: Central

More information

Stochastic Design Criteria in Linear Models

Stochastic Design Criteria in Linear Models AUSTRIAN JOURNAL OF STATISTICS Volume 34 (2005), Number 2, 211 223 Stochastic Design Criteria in Linear Models Alexander Zaigraev N. Copernicus University, Toruń, Poland Abstract: Within the framework

More information

A-optimal designs for generalized linear model with two parameters

A-optimal designs for generalized linear model with two parameters A-optimal designs for generalized linear model with two parameters Min Yang * University of Missouri - Columbia Abstract An algebraic method for constructing A-optimal designs for two parameter generalized

More information

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses

Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Ann Inst Stat Math (2009) 61:773 787 DOI 10.1007/s10463-008-0172-6 Generalized Neyman Pearson optimality of empirical likelihood for testing parameter hypotheses Taisuke Otsu Received: 1 June 2007 / Revised:

More information

OPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES

OPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES Statistica Sinica 21 (2011, 1415-1430 OPTIMAL DESIGNS FOR GENERALIZED LINEAR MODELS WITH MULTIPLE DESIGN VARIABLES Min Yang, Bin Zhang and Shuguang Huang University of Missouri, University of Alabama-Birmingham

More information

Computational Statistics and Data Analysis. Robustness of design for the testing of lack of fit and for estimation in binary response models

Computational Statistics and Data Analysis. Robustness of design for the testing of lack of fit and for estimation in binary response models Computational Statistics and Data Analysis 54 (200) 337 3378 Contents lists available at ScienceDirect Computational Statistics and Data Analysis journal homepage: www.elsevier.com/locate/csda Robustness

More information

Integrating Correlated Bayesian Networks Using Maximum Entropy

Integrating Correlated Bayesian Networks Using Maximum Entropy Applied Mathematical Sciences, Vol. 5, 2011, no. 48, 2361-2371 Integrating Correlated Bayesian Networks Using Maximum Entropy Kenneth D. Jarman Pacific Northwest National Laboratory PO Box 999, MSIN K7-90

More information

Response-adaptive dose-finding under model uncertainty

Response-adaptive dose-finding under model uncertainty Response-adaptive dose-finding under model uncertainty Björn Bornkamp Fakultät Statistik Technische Universität Dortmund 44221 Dortmund, Germany e-mail: bornkamp@statistik.tu-dortmund.de Frank Bretz Statistical

More information

Information in a Two-Stage Adaptive Optimal Design

Information in a Two-Stage Adaptive Optimal Design Information in a Two-Stage Adaptive Optimal Design Department of Statistics, University of Missouri Designed Experiments: Recent Advances in Methods and Applications DEMA 2011 Isaac Newton Institute for

More information

RATE-OPTIMAL GRAPHON ESTIMATION. By Chao Gao, Yu Lu and Harrison H. Zhou Yale University

RATE-OPTIMAL GRAPHON ESTIMATION. By Chao Gao, Yu Lu and Harrison H. Zhou Yale University Submitted to the Annals of Statistics arxiv: arxiv:0000.0000 RATE-OPTIMAL GRAPHON ESTIMATION By Chao Gao, Yu Lu and Harrison H. Zhou Yale University Network analysis is becoming one of the most active

More information

On the efficiency of two-stage adaptive designs

On the efficiency of two-stage adaptive designs On the efficiency of two-stage adaptive designs Björn Bornkamp (Novartis Pharma AG) Based on: Dette, H., Bornkamp, B. and Bretz F. (2010): On the efficiency of adaptive designs www.statistik.tu-dortmund.de/sfb823-dp2010.html

More information

Some Approximations of the Logistic Distribution with Application to the Covariance Matrix of Logistic Regression

Some Approximations of the Logistic Distribution with Application to the Covariance Matrix of Logistic Regression Working Paper 2013:9 Department of Statistics Some Approximations of the Logistic Distribution with Application to the Covariance Matrix of Logistic Regression Ronnie Pingel Working Paper 2013:9 June

More information

Optimal global rates of convergence for interpolation problems with random design

Optimal global rates of convergence for interpolation problems with random design Optimal global rates of convergence for interpolation problems with random design Michael Kohler 1 and Adam Krzyżak 2, 1 Fachbereich Mathematik, Technische Universität Darmstadt, Schlossgartenstr. 7, 64289

More information

DT-OPTIMAL DESIGNS FOR PROBIT MODELS IN CLINICAL TRIALS

DT-OPTIMAL DESIGNS FOR PROBIT MODELS IN CLINICAL TRIALS DT-OPTIMAL DESIGNS FOR PROBIT MODELS IN CLINICAL TRIALS A Thesis Submitted to the Graduate Faculty of the North Dakota State University of Agriculture and Applied Science By Andrew Michael Lexvold In Partial

More information

arxiv: v1 [stat.co] 7 Sep 2017

arxiv: v1 [stat.co] 7 Sep 2017 Computing optimal experimental designs with respect to a compound Bayes risk criterion arxiv:1709.02317v1 [stat.co] 7 Sep 2017 Radoslav Harman, and Maryna Prus September 8, 2017 Abstract We consider the

More information

Optimal Designs for 2 k Experiments with Binary Response

Optimal Designs for 2 k Experiments with Binary Response 1 / 57 Optimal Designs for 2 k Experiments with Binary Response Dibyen Majumdar Mathematics, Statistics, and Computer Science College of Liberal Arts and Sciences University of Illinois at Chicago Joint

More information

arxiv:math/ v1 [math.st] 3 Mar 2006

arxiv:math/ v1 [math.st] 3 Mar 2006 The Annals of Statistics 2005, Vol. 33, No. 6, 2758 2788 DOI: 10.1214/009053605000000552 c Institute of Mathematical Statistics, 2005 arxiv:math/0603075v1 [math.st] 3 Mar 2006 OPTIMAL DESIGNS FOR THREE-DIMENSIONAL

More information

Modified Simes Critical Values Under Positive Dependence

Modified Simes Critical Values Under Positive Dependence Modified Simes Critical Values Under Positive Dependence Gengqian Cai, Sanat K. Sarkar Clinical Pharmacology Statistics & Programming, BDS, GlaxoSmithKline Statistics Department, Temple University, Philadelphia

More information

Optimal Design for Nonlinear and Spatial Models: Introduction and Historical Overview

Optimal Design for Nonlinear and Spatial Models: Introduction and Historical Overview 12 Optimal Design for Nonlinear and Spatial Models: Introduction and Historical Overview Douglas P. Wiens CONTENTS 12.1 Introduction...457 12.2 Generalized Linear Models...457 12.3 Selected Nonlinear Models...460

More information

Multiplicative Algorithm for computing Optimum Designs

Multiplicative Algorithm for computing Optimum Designs for computing s DAE Conference 2012 Roberto Dorta-Guerra, Raúl Martín Martín, Mercedes Fernández-Guerrero and Licesio J. Rodríguez-Aragón Optimum Experimental Design group October 20, 2012 http://areaestadistica.uclm.es/oed

More information

Design of Screening Experiments with Partial Replication

Design of Screening Experiments with Partial Replication Design of Screening Experiments with Partial Replication David J. Edwards Department of Statistical Sciences & Operations Research Virginia Commonwealth University Robert D. Leonard Department of Information

More information

A strongly polynomial algorithm for linear systems having a binary solution

A strongly polynomial algorithm for linear systems having a binary solution A strongly polynomial algorithm for linear systems having a binary solution Sergei Chubanov Institute of Information Systems at the University of Siegen, Germany e-mail: sergei.chubanov@uni-siegen.de 7th

More information

Optimal experimental design, an introduction, Jesús López Fidalgo

Optimal experimental design, an introduction, Jesús López Fidalgo Optimal experimental design, an introduction Jesus.LopezFidalgo@uclm.es University of Castilla-La Mancha Department of Mathematics Institute of Applied Mathematics to Science and Engineering Books (just

More information

ROBUST DISCRIMINATION DESIGNS OVER HELLINGER NEIGHBOURHOODS 1. BY RUI HU AND DOUGLAS P. WIENS MacEwan University and University of Alberta

ROBUST DISCRIMINATION DESIGNS OVER HELLINGER NEIGHBOURHOODS 1. BY RUI HU AND DOUGLAS P. WIENS MacEwan University and University of Alberta The Annals of Statistics 2017, Vol. 45, No. 4, 1638 1663 DOI: 10.1214/16-AOS1503 Institute of Mathematical Statistics, 2017 ROBUST DISCRIMINATION DESIGNS OVER HELLINGER NEIGHBOURHOODS 1 BY RUI HU AND DOUGLAS

More information

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process

Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Applied Mathematical Sciences, Vol. 4, 2010, no. 62, 3083-3093 Sequential Procedure for Testing Hypothesis about Mean of Latent Gaussian Process Julia Bondarenko Helmut-Schmidt University Hamburg University

More information

E-optimal approximate block designs for treatment-control comparisons

E-optimal approximate block designs for treatment-control comparisons E-optimal approximate block designs for treatment-control comparisons Samuel Rosa 1 1 Faculty of Mathematics, Physics and Informatics, Comenius University in Bratislava, Slovakia We study E-optimal block

More information

arxiv: v1 [stat.me] 24 May 2010

arxiv: v1 [stat.me] 24 May 2010 The role of the nugget term in the Gaussian process method Andrey Pepelyshev arxiv:1005.4385v1 [stat.me] 24 May 2010 Abstract The maximum likelihood estimate of the correlation parameter of a Gaussian

More information

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints

A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Noname manuscript No. (will be inserted by the editor) A Recursive Formula for the Kaplan-Meier Estimator with Mean Constraints Mai Zhou Yifan Yang Received: date / Accepted: date Abstract In this note

More information

Model Selection versus Model Averaging in Dose Finding Studies

Model Selection versus Model Averaging in Dose Finding Studies Model Selection versus Model Averaging in Dose Finding Studies Schorning, Kirsten Ruhr-Universität Bochum Bornkamp, Björn Novartis Pharma AG arxiv:8.00281v1 [stat.ap] 2 Aug 2015 Fakultät für Mathematik

More information

Long-Run Covariability

Long-Run Covariability Long-Run Covariability Ulrich K. Müller and Mark W. Watson Princeton University October 2016 Motivation Study the long-run covariability/relationship between economic variables great ratios, long-run Phillips

More information

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion

Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Proceedings 59th ISI World Statistics Congress, 25-30 August 2013, Hong Kong (Session CPS020) p.3863 Model Selection for Semiparametric Bayesian Models with Application to Overdispersion Jinfang Wang and

More information

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix

Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Nonparametric Identification of a Binary Random Factor in Cross Section Data - Supplemental Appendix Yingying Dong and Arthur Lewbel California State University Fullerton and Boston College July 2010 Abstract

More information

1 The linear algebra of linear programs (March 15 and 22, 2015)

1 The linear algebra of linear programs (March 15 and 22, 2015) 1 The linear algebra of linear programs (March 15 and 22, 2015) Many optimization problems can be formulated as linear programs. The main features of a linear program are the following: Variables are real

More information

MULTIPLE-OBJECTIVE DESIGNS IN A DOSE-RESPONSE EXPERIMENT

MULTIPLE-OBJECTIVE DESIGNS IN A DOSE-RESPONSE EXPERIMENT New Developments and Applications in Experimental Design IMS Lecture Notes - Monograph Series (1998) Volume 34 MULTIPLE-OBJECTIVE DESIGNS IN A DOSE-RESPONSE EXPERIMENT BY WEI ZHU AND WENG KEE WONG 1 State

More information

Covariance function estimation in Gaussian process regression

Covariance function estimation in Gaussian process regression Covariance function estimation in Gaussian process regression François Bachoc Department of Statistics and Operations Research, University of Vienna WU Research Seminar - May 2015 François Bachoc Gaussian

More information

Dose-response modeling with bivariate binary data under model uncertainty

Dose-response modeling with bivariate binary data under model uncertainty Dose-response modeling with bivariate binary data under model uncertainty Bernhard Klingenberg 1 1 Department of Mathematics and Statistics, Williams College, Williamstown, MA, 01267 and Institute of Statistics,

More information

D-optimal designs for logistic regression in two variables

D-optimal designs for logistic regression in two variables D-optimal designs for logistic regression in two variables Linda M. Haines 1, M. Gaëtan Kabera 2, P. Ndlovu 2 and T. E. O Brien 3 1 Department of Statistical Sciences, University of Cape Town, Rondebosch

More information

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models

The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models The Behaviour of the Akaike Information Criterion when Applied to Non-nested Sequences of Models Centre for Molecular, Environmental, Genetic & Analytic (MEGA) Epidemiology School of Population Health

More information

ADAPTIVE EXPERIMENTAL DESIGNS. Maciej Patan and Barbara Bogacka. University of Zielona Góra, Poland and Queen Mary, University of London

ADAPTIVE EXPERIMENTAL DESIGNS. Maciej Patan and Barbara Bogacka. University of Zielona Góra, Poland and Queen Mary, University of London ADAPTIVE EXPERIMENTAL DESIGNS FOR SIMULTANEOUS PK AND DOSE-SELECTION STUDIES IN PHASE I CLINICAL TRIALS Maciej Patan and Barbara Bogacka University of Zielona Góra, Poland and Queen Mary, University of

More information

Single Index Quantile Regression for Heteroscedastic Data

Single Index Quantile Regression for Heteroscedastic Data Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR

More information

Basics of Point-Referenced Data Models

Basics of Point-Referenced Data Models Basics of Point-Referenced Data Models Basic tool is a spatial process, {Y (s), s D}, where D R r Chapter 2: Basics of Point-Referenced Data Models p. 1/45 Basics of Point-Referenced Data Models Basic

More information

A Posteriori Estimates for Cost Functionals of Optimal Control Problems

A Posteriori Estimates for Cost Functionals of Optimal Control Problems A Posteriori Estimates for Cost Functionals of Optimal Control Problems Alexandra Gaevskaya, Ronald H.W. Hoppe,2 and Sergey Repin 3 Institute of Mathematics, Universität Augsburg, D-8659 Augsburg, Germany

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

Learning Theory. Ingo Steinwart University of Stuttgart. September 4, 2013

Learning Theory. Ingo Steinwart University of Stuttgart. September 4, 2013 Learning Theory Ingo Steinwart University of Stuttgart September 4, 2013 Ingo Steinwart University of Stuttgart () Learning Theory September 4, 2013 1 / 62 Basics Informal Introduction Informal Description

More information

Designs for Generalized Linear Models

Designs for Generalized Linear Models Designs for Generalized Linear Models Anthony C. Atkinson David C. Woods London School of Economics and Political Science, UK University of Southampton, UK December 9, 2013 Email: a.c.atkinson@lse.ac.uk

More information

Methods of Estimation

Methods of Estimation Methods of Estimation MIT 18.655 Dr. Kempthorne Spring 2016 1 Outline Methods of Estimation I 1 Methods of Estimation I 2 X X, X P P = {P θ, θ Θ}. Problem: Finding a function θˆ(x ) which is close to θ.

More information

On Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence:

On Acceleration with Noise-Corrupted Gradients. + m k 1 (x). By the definition of Bregman divergence: A Omitted Proofs from Section 3 Proof of Lemma 3 Let m x) = a i On Acceleration with Noise-Corrupted Gradients fxi ), u x i D ψ u, x 0 ) denote the function under the minimum in the lower bound By Proposition

More information

Invariant HPD credible sets and MAP estimators

Invariant HPD credible sets and MAP estimators Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in

More information

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures

9. Model Selection. statistical models. overview of model selection. information criteria. goodness-of-fit measures FE661 - Statistical Methods for Financial Engineering 9. Model Selection Jitkomut Songsiri statistical models overview of model selection information criteria goodness-of-fit measures 9-1 Statistical models

More information

LOCAL TIMES OF RANKED CONTINUOUS SEMIMARTINGALES

LOCAL TIMES OF RANKED CONTINUOUS SEMIMARTINGALES LOCAL TIMES OF RANKED CONTINUOUS SEMIMARTINGALES ADRIAN D. BANNER INTECH One Palmer Square Princeton, NJ 8542, USA adrian@enhanced.com RAOUF GHOMRASNI Fakultät II, Institut für Mathematik Sekr. MA 7-5,

More information

arxiv: v1 [math.st] 14 Aug 2007

arxiv: v1 [math.st] 14 Aug 2007 The Annals of Statistics 2007, Vol. 35, No. 2, 772 792 DOI: 10.1214/009053606000001307 c Institute of Mathematical Statistics, 2007 arxiv:0708.1901v1 [math.st] 14 Aug 2007 ON THE NUMBER OF SUPPORT POINTS

More information

Linear, Generalized Linear, and Mixed-Effects Models in R. Linear and Generalized Linear Models in R Topics

Linear, Generalized Linear, and Mixed-Effects Models in R. Linear and Generalized Linear Models in R Topics Linear, Generalized Linear, and Mixed-Effects Models in R John Fox McMaster University ICPSR 2018 John Fox (McMaster University) Statistical Models in R ICPSR 2018 1 / 19 Linear and Generalized Linear

More information

An a posteriori error estimate and a Comparison Theorem for the nonconforming P 1 element

An a posteriori error estimate and a Comparison Theorem for the nonconforming P 1 element Calcolo manuscript No. (will be inserted by the editor) An a posteriori error estimate and a Comparison Theorem for the nonconforming P 1 element Dietrich Braess Faculty of Mathematics, Ruhr-University

More information

1 Directional Derivatives and Differentiability

1 Directional Derivatives and Differentiability Wednesday, January 18, 2012 1 Directional Derivatives and Differentiability Let E R N, let f : E R and let x 0 E. Given a direction v R N, let L be the line through x 0 in the direction v, that is, L :=

More information

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices

Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Article Lasso Maximum Likelihood Estimation of Parametric Models with Singular Information Matrices Fei Jin 1,2 and Lung-fei Lee 3, * 1 School of Economics, Shanghai University of Finance and Economics,

More information

Design of experiments for microbiological models

Design of experiments for microbiological models Design of experiments for microbiological models Holger Dette Ruhr-Universität Bochum Fakultät für Mathematik 44780 Bochum Germany email: holger.dette@ruhr-uni-bochum.de FAX: +49 2 34 70 94 559 Viatcheslav

More information

ORDER RESTRICTED STATISTICAL INFERENCE ON LORENZ CURVES OF PARETO DISTRIBUTIONS. Myongsik Oh. 1. Introduction

ORDER RESTRICTED STATISTICAL INFERENCE ON LORENZ CURVES OF PARETO DISTRIBUTIONS. Myongsik Oh. 1. Introduction J. Appl. Math & Computing Vol. 13(2003), No. 1-2, pp. 457-470 ORDER RESTRICTED STATISTICAL INFERENCE ON LORENZ CURVES OF PARETO DISTRIBUTIONS Myongsik Oh Abstract. The comparison of two or more Lorenz

More information

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach

Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Statistical Methods for Handling Incomplete Data Chapter 2: Likelihood-based approach Jae-Kwang Kim Department of Statistics, Iowa State University Outline 1 Introduction 2 Observed likelihood 3 Mean Score

More information

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown.

Now consider the case where E(Y) = µ = Xβ and V (Y) = σ 2 G, where G is diagonal, but unknown. Weighting We have seen that if E(Y) = Xβ and V (Y) = σ 2 G, where G is known, the model can be rewritten as a linear model. This is known as generalized least squares or, if G is diagonal, with trace(g)

More information

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games

Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Stability of Feedback Solutions for Infinite Horizon Noncooperative Differential Games Alberto Bressan ) and Khai T. Nguyen ) *) Department of Mathematics, Penn State University **) Department of Mathematics,

More information

Support Vector Machine via Nonlinear Rescaling Method

Support Vector Machine via Nonlinear Rescaling Method Manuscript Click here to download Manuscript: svm-nrm_3.tex Support Vector Machine via Nonlinear Rescaling Method Roman Polyak Department of SEOR and Department of Mathematical Sciences George Mason University

More information

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing

Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes with Killing Advances in Dynamical Systems and Applications ISSN 0973-5321, Volume 8, Number 2, pp. 401 412 (2013) http://campus.mst.edu/adsa Weighted Sums of Orthogonal Polynomials Related to Birth-Death Processes

More information

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract

COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS. Abstract Far East J. Theo. Stat. 0() (006), 179-196 COMPARISON OF GMM WITH SECOND-ORDER LEAST SQUARES ESTIMATION IN NONLINEAR MODELS Department of Statistics University of Manitoba Winnipeg, Manitoba, Canada R3T

More information

Bayesian inference for the location parameter of a Student-t density

Bayesian inference for the location parameter of a Student-t density Bayesian inference for the location parameter of a Student-t density Jean-François Angers CRM-2642 February 2000 Dép. de mathématiques et de statistique; Université de Montréal; C.P. 628, Succ. Centre-ville

More information

Convex Optimization & Parsimony of L p-balls representation

Convex Optimization & Parsimony of L p-balls representation Convex Optimization & Parsimony of L p -balls representation LAAS-CNRS and Institute of Mathematics, Toulouse, France IMA, January 2016 Motivation Unit balls associated with nonnegative homogeneous polynomials

More information

Response Surface Methods

Response Surface Methods Response Surface Methods 3.12.2014 Goals of Today s Lecture See how a sequence of experiments can be performed to optimize a response variable. Understand the difference between first-order and second-order

More information

Gaussian Estimation under Attack Uncertainty

Gaussian Estimation under Attack Uncertainty Gaussian Estimation under Attack Uncertainty Tara Javidi Yonatan Kaspi Himanshu Tyagi Abstract We consider the estimation of a standard Gaussian random variable under an observation attack where an adversary

More information

One-step ahead adaptive D-optimal design on a finite design. space is asymptotically optimal

One-step ahead adaptive D-optimal design on a finite design. space is asymptotically optimal Author manuscript, published in "Metrika (2009) 20" DOI : 10.1007/s00184-008-0227-y One-step ahead adaptive D-optimal design on a finite design space is asymptotically optimal Luc Pronzato Laboratoire

More information