Evaluation of the Fisher information matrix in nonlinear mixed eect models without linearization

Similar documents
Robust design in model-based analysis of longitudinal clinical data

Sebastian Ueckert, France Mentré. To cite this version: HAL Id: inserm

F. Combes (1,2,3) S. Retout (2), N. Frey (2) and F. Mentré (1) PODE 2012

Estimation and Model Selection in Mixed Effects Models Part I. Adeline Samson 1

DESIGN EVALUATION AND OPTIMISATION IN CROSSOVER PHARMACOKINETIC STUDIES ANALYSED BY NONLINEAR MIXED EFFECTS MODELS

Evaluation of the Fisher information matrix in nonlinear mixed effect models using adaptive Gaussian quadrature

Stochastic approximation EM algorithm in nonlinear mixed effects model for viral load decrease during anti-hiv treatment

Extension of the SAEM algorithm for nonlinear mixed models with 2 levels of random effects Panhard Xavière

STA 4273H: Statistical Machine Learning

Description of UseCase models in MDL

A comparison of estimation methods in nonlinear mixed effects models using a blind analysis

Extension of the SAEM algorithm for nonlinear mixed. models with two levels of random effects

Introduction to Bayesian Statistics and Markov Chain Monte Carlo Estimation. EPSY 905: Multivariate Analysis Spring 2016 Lecture #10: April 6, 2016

Pattern Recognition and Machine Learning. Bishop Chapter 11: Sampling Methods

Comparison of multiple imputation methods for systematically and sporadically missing multilevel data

An effective approach for obtaining optimal sampling windows for population pharmacokinetic experiments

Default Priors and Effcient Posterior Computation in Bayesian

Population Design in Nonlinear Mixed Effects Multiple Response Models: extension of PFIM and evaluation by simulation with NONMEM and MONOLIX

Accurate Maximum Likelihood Estimation for Parametric Population Analysis. Bob Leary UCSD/SDSC and LAPK, USC School of Medicine

Markov Chain Monte Carlo

MCMC Sampling for Bayesian Inference using L1-type Priors

Pharmacometrics : Nonlinear mixed effect models in Statistics. Department of Statistics Ewha Womans University Eun-Kyung Lee

Extension of the SAEM algorithm to left-censored data in nonlinear mixed-effects model: application to HIV dynamics model

Markov Chain Monte Carlo methods

Probabilistic Graphical Models Lecture 17: Markov chain Monte Carlo

Probabilistic Machine Learning

Nonlinear Mixed Effects Models

(5) Multi-parameter models - Gibbs sampling. ST440/540: Applied Bayesian Analysis

Metropolis-Hastings Algorithm

Generalizing the MCPMod methodology beyond normal, independent data

Riemann Manifold Methods in Bayesian Statistics

19 : Slice Sampling and HMC

Bayesian Inference on Joint Mixture Models for Survival-Longitudinal Data with Multiple Features. Yangxin Huang

Advanced Introduction to Machine Learning

Hamiltonian Monte Carlo

ML estimation: Random-intercepts logistic model. and z

Probabilistic Graphical Models

Multilevel Statistical Models: 3 rd edition, 2003 Contents

Bayesian model selection in graphs by using BDgraph package

Statistical Inference and Methods

Bayesian Inference and MCMC

Bayesian Estimation with Sparse Grids

Kernel adaptive Sequential Monte Carlo

MIT /30 Gelman, Carpenter, Hoffman, Guo, Goodrich, Lee,... Stan for Bayesian data analysis

Statistical Machine Learning Lecture 8: Markov Chain Monte Carlo Sampling

ST 740: Markov Chain Monte Carlo

Computer Intensive Methods in Mathematical Statistics

Learning the hyper-parameters. Luca Martino

Understanding Automatic Differentiation to Improve Performance

STAT 425: Introduction to Bayesian Analysis

Bayesian Methods for Machine Learning

Lecture 7 and 8: Markov Chain Monte Carlo

Introduction to Machine Learning CMU-10701

Variational Inference via Stochastic Backpropagation

Advanced Statistical Modelling

Monte Carlo Methods. Leon Gu CSD, CMU

Dynamic System Identification using HDMR-Bayesian Technique

Contents. Part I: Fundamentals of Bayesian Inference 1

Supplement to A Hierarchical Approach for Fitting Curves to Response Time Measurements

Generalizing the MCPMod methodology beyond normal, independent data

Computer Vision Group Prof. Daniel Cremers. 11. Sampling Methods: Markov Chain Monte Carlo

Generative Models and Stochastic Algorithms for Population Average Estimation and Image Analysis

Bayesian Estimation of Input Output Tables for Russia

Large Scale Bayesian Inference

Generalized Exponential Random Graph Models: Inference for Weighted Graphs

Graphical Models and Kernel Methods

Markov Chain Monte Carlo (MCMC) and Model Evaluation. August 15, 2017

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

Computer Vision Group Prof. Daniel Cremers. 10a. Markov Chain Monte Carlo

April 20th, Advanced Topics in Machine Learning California Institute of Technology. Markov Chain Monte Carlo for Machine Learning

Markov Chain Monte Carlo

Cheng Soon Ong & Christian Walder. Canberra February June 2018

Bayesian Inference in GLMs. Frequentists typically base inferences on MLEs, asymptotic confidence

Bayesian non-parametric model to longitudinally predict churn

Computer Vision Group Prof. Daniel Cremers. 14. Sampling Methods

Announcements. Inference. Mid-term. Inference by Enumeration. Reminder: Alarm Network. Introduction to Artificial Intelligence. V22.

Outline. Mixed models in R using the lme4 package Part 5: Generalized linear mixed models. Parts of LMMs carried over to GLMMs

CPSC 540: Machine Learning

Correction of the likelihood function as an alternative for imputing missing covariates. Wojciech Krzyzanski and An Vermeulen PAGE 2017 Budapest

6 Markov Chain Monte Carlo (MCMC)

STA414/2104 Statistical Methods for Machine Learning II

17 : Markov Chain Monte Carlo

Estimation and Inference on Dynamic Panel Data Models with Stochastic Volatility

On Markov chain Monte Carlo methods for tall data

Prerequisite: STATS 7 or STATS 8 or AP90 or (STATS 120A and STATS 120B and STATS 120C). AP90 with a minimum score of 3

Probing the covariance matrix

Probabilistic Graphical Models

Annealing Between Distributions by Averaging Moments

Introduction to Bayesian methods in inverse problems

Mixed models in R using the lme4 package Part 7: Generalized linear mixed models

Nonlinear mixed-effects models using Stata

Markov Chain Monte Carlo in Practice

Heterogeneous shedding of influenza by human subjects and. its implications for epidemiology and control

Monte Carlo methods for sampling-based Stochastic Optimization

Estimation, Model Selection and Optimal Design in Mixed Eects Models Applications to pharmacometrics. Marc Lavielle 1. Cemracs CIRM logo

Adaptive Monte Carlo methods

Bayesian Methods and Uncertainty Quantification for Nonlinear Inverse Problems

Online appendix to On the stability of the excess sensitivity of aggregate consumption growth in the US

STA 4273H: Statistical Machine Learning

Sampling from complex probability distributions

Transcription:

Evaluation of the Fisher information matrix in nonlinear mixed eect models without linearization Sebastian Ueckert, Marie-Karelle Riviere, and France Mentré INSERM, IAME, UMR 1137, F-75018 Paris, France; Univ Paris Diderot, Sorbonne Paris Cité, F-75018 Paris, France Riviere, Ueckert (UMRS 1137) FIM 1 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 1 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 1 / 35

Novel approximation for Fisher information matrix (FIM) for non-linear mixed eect models Riviere, Ueckert (UMRS 1137) FIM 2 / 35

Novel approximation for Fisher information matrix (FIM) for non-linear mixed eect models Again? Why? Riviere, Ueckert (UMRS 1137) FIM 2 / 35

Pharmacokinetics... characterization of the time course of drug absorption, distribution, metabolism and excretion, and with the relationship of these processes to the intensity and time course of therapeutic and adverse eects of drugs 1 Indomethacin concentration (mcg/ml) 0 2 4 6 8 0 2 4 6 8 0 2 4 6 8 1 4 2 5 2.5 2.0 1.5 1.0 0.5 0.0 0 2 4 6 8 0 2 4 6 8 1 Wagner 1981 Time since drug administration (hr) 6 3 0 2 4 6 8 Riviere, Ueckert (UMRS 1137) FIM 3 / 35

Pharmacokinetics Important growth period 1961-1972 (important concepts, journals,... ) Clinical use limited due to need for many measurments per subject New approach by Sheiner et al. 1972 Pharmacokinetic models (non-linear) Parameters with xed and random component Many advantages (sparse sampling, individualized dosing,... ) Non-linear mixed eect models (NLMEM) Riviere, Ueckert (UMRS 1137) FIM 4 / 35

Parameter estimation No analytic form for likelihood estimation challenging Sheiner et al. 1972 also proposed parameter estimation algorithm First order (FO) approximation 2 : Model linearized w.r.t. random eect parameters Random eect parameter evaluated at 0 Essentially treatment as linear mixed eect model Algorithm fast & from 1980 available in software (NONMEM) 2 MQL approximation in statistics literature Riviere, Ueckert (UMRS 1137) FIM 5 / 35

Design Design of population pharmacokinetic studies mostly heuristic & rarely took advantage of NLMEM Derivation of FIM approximation for NLMEM by Mentré et al. 1997 First-order approximation Treatment as linear mixed eect model Riviere, Ueckert (UMRS 1137) FIM 6 / 35

Evolution of estimation algorithms FO approximation for estimation Biased Asympthotic theory does not apply Problems already for simple models NLMEM estimation algorithms: Model linearization around mode (Lindström and Bates 1990) Laplacian approximation (Tierney and Kadane 1986) Importances sampling (Geweke 1989) Gaussian quadrature (Davidian and Gallant 1992) Adaptive gaussian quadrature (Pinheiro and Bates 1995) Stochastic approximation expectation maximization (Kuhn and Lavielle 2004) Riviere, Ueckert (UMRS 1137) FIM 7 / 35

Evolution of population FIM approximations FO approximation surprisingly accurate for a large class of models 3 Population FIM approximations First order conditional approximation (Retout and Mentré 2003) Non-block diagonal FO approximation (Foracchia et al. 2004) First order conditional mode approximation (Nyberg et al. 2012) FO approximation still most used (PFIM & PopED) 3 Nyberg et al. Br J Clin Pharmacol 79, 617 (2015). Riviere, Ueckert (UMRS 1137) FIM 8 / 35

Evolution of models Advances in estimation methods fostered development of more complex models Complex random eects structures Multiple outcomes Many parameters Discrete outcomes 4 FO approximation not or less appropriate 4 Mentré et al. CPT: pharmacomet. syst. pharmacol. 2, e46 (2013) Riviere, Ueckert (UMRS 1137) FIM 9 / 35

Objectives Apply integration algorithms, which have proven to be ecient for estimation, to evaluate the asymptotically exact FIM in NLMEM for both discrete and continuous outcomes Riviere, Ueckert (UMRS 1137) FIM 10 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 10 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 10 / 35

Nonlinear mixed eect models For continuous data: y i = f(g(µ, b i ), ξ i ) + ε i with For discrete data: n i P (y i b i ) = h(y ij, g(µ, b i ), ξ i ) j=1 y i = (y i1,..., y ini ) T response for individual i (i = 1,..., N) f, h structural model ξ i elementary design for subject i g individual parameters vector, function of µ and b i µ vector of xed eects b i vector of random eects for individual i, b i N (0, Ω) ε i vector of residual errors, ε i N (0, Σ) and Σ diagonal matrix Riviere, Ueckert (UMRS 1137) FIM 11 / 35

Fisher Information Matrix (FIM) Population FIM: N M(ψ, Ξ) = M(ψ, ξ i ) with i=1 ψ vector of all parameters (Ψ = (µ, Ω, Σ) T ) Ξ population design (Ξ = (ξ 1,..., ξ N ) T ) Individual FIM: M(ψ, ξ) = E ( log(l(y,ψ)) ψ with the likelihood: L(y, ψ) = p(y b, ψ)p(b)db where b log(l(y,ψ)) T ) ψ p(y b, ψ): conditional density of y given the random eects b p(b): density of b Riviere, Ueckert (UMRS 1137) FIM 12 / 35

M(ψ, ξ) = ( T log(l(y, ψ)) log(l(y, ψ)) E ψ ψ = ( T log(l(y, ψ)) log(l(y, ψ)) L(y, ψ)dy y ψ ψ }{{} A y M(ψ, ξ) = A y L(y, ψ)dy Monte Carlo - MC y M(ψ, ξ) = y r A yr a AGQ MCMC Riviere, Ueckert (UMRS 1137) FIM 13 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 13 / 35

Estimation of A y with AGQ 5 A y = log(l(y,ψ)) ψ b log(l(y,ψ)) ψ with L(y, ψ) = p(y b, ψ)p(b)db = L(y, ψ) (p(y b, ψ)p(b)) = db ψ b ψ L(y, ψ) p(y Ω 1 2 η, ψ) = φ(η) dη + ψ η ψ L(y, ψ) [ = 1 ψ 2 T r η η T = L(y,ψ) ψ p(y Ω 1 2 η, ψ)φ(η) dη ( ) 1 Ω Ω ψ k η T Ω 1 2 Adaptive Gaussian Quadrature - AGQ L(y,ψ) T ψ L(y, Ψ)) 2 (η def = Ω 1 2 b) ] Ω ψ k Ω 1 2 η p(y Ω 1 2 η, ψ)φ(η) dη 5 Nguyen & Mentré. Comp. Statistics & Data Analysis 80, 5769 (2014). Riviere, Ueckert (UMRS 1137) FIM 14 / 35

Gaussian Quadrature f(x)φ(z)dz Q w q f(z q ) q=1 Riviere, Ueckert (UMRS 1137) FIM 15 / 35

Adaptive Gaussian Quadrature Q f(x)φ(z)dz w q f(z q ) q=1 Center integration grid at mode of f(x)φ(z) Center integration grid at mode of f(x)φ(z) & scale by 2 z 2 f(x)φ(z) Riviere, Ueckert (UMRS 1137) FIM 16 / 35

MC-AGQ algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) Determine integration grid (III) Approximate integrals Riviere, Ueckert (UMRS 1137) FIM 17 / 35

MC-AGQ algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) Determine integration grid Determine ˆη = arg max η log p(y Ω 1 2 η, ψ)φ(η) through BFGS optimization starting at η. Calculate 2 log p(y Ω 1 η 2 2 η, ψ)φ(η) η=ˆη through numerical dierentiation. Calculate quadrature nodes [ ] 1 a q1,...,q d = ˆη + 2 η log p(y Ω 1 2 2 2 η, ψ)φ(η) a q1,...,q η=ˆη d and weights [ w q1,...,q d = ] 1 η log p(y Ω 1 2 2 2 η, ψ)φ(η) η=ˆη 2 (III) Approximate integrals d k=1 w q k φ(a qk ) φ(a q k ) Riviere, Ueckert (UMRS 1137) FIM 17 / 35

MC-AGQ algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) Determine integration grid Determine ˆη = arg max η log p(y Ω 1 2 η, ψ)φ(η) through BFGS optimization starting at η. Calculate 2 log p(y Ω 1 η 2 2 η, ψ)φ(η) η=ˆη through numerical dierentiation. Calculate quadrature nodes [ ] 1 a q1,...,q d = ˆη + 2 η log p(y Ω 1 2 2 2 η, ψ)φ(η) a q1,...,q η=ˆη d and weights [ w q1,...,q d = ] 1 η log p(y Ω 1 2 2 2 η, ψ)φ(η) η=ˆη 2 (III) Approximate integrals d k=1 w q k φ(a qk ) φ(a q k ) Riviere, Ueckert (UMRS 1137) FIM 17 / 35

MC-AGQ algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) Determine integration grid Determine ˆη = arg max η log p(y Ω 1 2 η, ψ)φ(η) through BFGS optimization starting at η. Calculate 2 log p(y Ω 1 η 2 2 η, ψ)φ(η) η=ˆη through numerical dierentiation. Calculate quadrature nodes [ ] 1 a q1,...,q d = ˆη + 2 η log p(y Ω 1 2 2 2 η, ψ)φ(η) a q1,...,q η=ˆη d and weights [ w q1,...,q d = ] 1 η log p(y Ω 1 2 2 2 η, ψ)φ(η) η=ˆη 2 (III) Approximate integrals d k=1 w q k φ(a qk ) φ(a q k ) Riviere, Ueckert (UMRS 1137) FIM 17 / 35

MC-AGQ algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) Determine integration grid (III) Approximate integrals Q q 1 =1 Q q d =1 w q 1,...,q d f(a q1,...,q d, ) Riviere, Ueckert (UMRS 1137) FIM 17 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 17 / 35

Reminder M(ψ, ξ) = ( T log(l(y, ψ)) log(l(y, ψ)) E ψ ψ = ( T log(l(y, ψ)) log(l(y, ψ)) L(y, ψ)dy y ψ ψ }{{} A y M(ψ, ξ) = A y L(y, ψ)dy Monte Carlo - MC M(ψ, ξ) = y r y A yr Riviere, Ueckert (UMRS 1137) FIM 18 / 35

Estimation of A y with MCMC After calculation... A y in the FIM: db 1. p(y b, ψ)p(b)db b 1 (log(p(y b 1,ψ)p(b 1 ))) ψ k p(y b 1, ψ)p(b 1 ) (log(p(y b 2,ψ)p(b 2 ))) p(y b 2, ψ)p(b 2 ) db ψ 2 b l 2 p(y b, ψ)p(b)db Riviere, Ueckert (UMRS 1137) FIM 19 / 35

Estimation of A y with MCMC After calculation... A y in the FIM: (log(p(y b 1,ψ)p(b 1 ))) p(y b 1, ψ)p(b 1 ) ψ b k 1 p(y b, ψ)p(b)db } {{ } conditional density of b given y db 1. (log(p(y b 2,ψ)p(b 2 ))) p(y b 2, ψ)p(b 2 ) ψ b l 2 p(y b, ψ)p(b)db } {{ } conditional density of b given y db 2 Markov Chains Monte Carlo - MCMC Riviere, Ueckert (UMRS 1137) FIM 19 / 35

MC-MCMC algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. Riviere, Ueckert (UMRS 1137) FIM 20 / 35

MC-MCMC algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) For each value of y sampled: Riviere, Ueckert (UMRS 1137) FIM 20 / 35

MC-MCMC algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) For each value of y sampled: (III) Using MCMC, draw two series of M-samples of b from its conditional density given y. Riviere, Ueckert (UMRS 1137) FIM 20 / 35

MC-MCMC algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) For each value of y sampled: (III) Using MCMC, draw two series of M-samples of b from its conditional density given y. (IV) Estimate b 1 and b 2 by the mean of the partial derivatives of the conditional log-likelihood taken in the samples of b drawn in step (III). Riviere, Ueckert (UMRS 1137) FIM 20 / 35

MC-MCMC algorithm for FIM evaluation (I) Draw an R-sample of y from its marginal distribution. (II) For each value of y sampled: (III) Using MCMC, draw two series of M-samples of b from its conditional density given y. (IV) Estimate b 1 and b 2 by the mean of the partial derivatives of the conditional log-likelihood taken in the samples of b drawn in step (III). (V) Using MC, estimate by the mean according to y of the product of the y previous partial derivatives. Riviere, Ueckert (UMRS 1137) FIM 20 / 35

Partial derivatives of the conditional log-likelihood (log(p(y b, ψ)p(b))) ψ k By hand. For continuous data: (log(p(y b, ψ)p(b))) = 1 [ ( ) T r V 1 V b b ψ k 2 ψ k (y E b ) T V 1 V b b V 1 b (y E b ) + T r ψ k with E b = f(g(µ, b), ξ) and V b = Σ 2(y E b ) T V 1 E b b ψ k ( 1 Ω Ω ψ k ) b T 1 Ω Ω Ω 1 b ψ k ] Numerically for all types of distributions Riviere, Ueckert (UMRS 1137) FIM 21 / 35

STAN for MCMC STAN Markov Chain Monte Carlo (MCMC) sampler (as JAGS, BUGS,...) To sample in posterior distributions Based on constructing a Markov chain that has the desired distribution as its stationary distribution STAN uses Hamiltonian Monte Carlo (HMC) vs. random walk Monte Carlo methods (Metropolis-Hastings, Gibbs sampling,...) More complex but more ecient, faster convergence Able to overcome some issues inherent in Gibbs sampling STAN calculates the gradient of the log probability function (necessary for HMC) Stan Development Team. Gelman, Carpenter,... Columbia University 2014. Stan: A C++ Library for Probability and Sampling, Version 2.5.0. http://mc-stan.org Riviere, Ueckert (UMRS 1137) FIM 22 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 22 / 35

FIM evaluation We compared 3 approaches: Linearization (FO) using PFIM 4.0 AGQ-based approach (AGQ) implemented in R (using statmod for nodes) MCMC-based approach (MCMC) implemented in R (using RSTAN) with clinical trial simulations (CTS): Simulate 1000 datasets Y with Ψ = Ψ T using R For each Y : estimate ˆΨ using Monolix 4.3 in terms of RSE / RRMSE: RRMSE = Calculation time 1 ( ˆΨ ΨT ) 2 /Ψ T 1000 Riviere, Ueckert (UMRS 1137) FIM 23 / 35

Example 1: PK Warfarin PKW: One compartment model with rst order absorption and elimination: f(φ = (k a, V, CL), t) = 70 V k a k a CL V ( e CL V t e kat) Riviere, Ueckert (UMRS 1137) FIM 24 / 35

Example 1: PK Warfarin PKW: One compartment model with rst order absorption and elimination: f(φ = (k a, V, CL), t) = 70 V k a k a CL V ( e CL V t e kat) Fixed eects: (µ ka, µ V, µ CL ) = (1.00, 8.00, 0.15) Exponential random eects with variances: (ωk 2 a, ωv 2, ω2 CL ) = (0.60, 0.02, 0.07) Proportional residual error: σ slope = 0.1 8 times: t = (0.5, 1, 2, 6, 24, 36, 72, 120) N = 32 patients Nyberg et al. Br J Clin Pharmacol 79, 617 (2015). Riviere, Ueckert (UMRS 1137) FIM 24 / 35

Example 1 - RSE/RRMSE Riviere, Ueckert (UMRS 1137) FIM 25 / 35

Example 2: Sigmoïd E max model SC1: Sigmoïd E max model: f(φ = (E 0, E max, ED 50, γ), d) = E 0 + E maxd γ ED γ 50 + dγ Riviere, Ueckert (UMRS 1137) FIM 26 / 35

Example 2: Sigmoïd E max model SC1: Sigmoïd E max model: f(φ = (E 0, E max, ED 50, γ), d) = E 0 + E maxd γ ED γ 50 + dγ Fixed eects: (µ E0, µ Emax, µ ED50, µ γ ) = (5, 30, 500, 3) Exponential random eects with variance-covariance: 0.09 0.06 0.06 0 Ω = 0.06 0.09 0.06 0 0.06 0.06 0.09 0 0 0 0 0.09 Combined residual error: (σ inter, σ slope ) = (0.2, 0.2) 4 doses: d = (0, 100, 300, 1000) N = 100 patients Riviere, Ueckert (UMRS 1137) FIM 26 / 35

Example 2 - RSE/RRMSE Riviere, Ueckert (UMRS 1137) FIM 27 / 35

Example 3: Repeated time-to-event RRTE: Exponential distribution for repeated time-to-event with constant hazard: Fixed eects: µ 1 = 1.0 P (y b) = λ 1 exp( λ 1 t) Exponential random eects: λ 1 = µ 1 exp(b) with variances: ω 2 1 = 0.1 Censoring time: 10 N = 50 patients Response (Cum. events) 10 8 6 4 2 Subject i=1 Subject i=2 Subject i=3 Subject i=4 0.0 2.5 5.0 7.5 10.00.0 2.5 5.0 7.5 10.00.0 2.5 5.0 7.5 10.00.0 2.5 5.0 7.5 10.0 Time Riviere, Ueckert (UMRS 1137) FIM 28 / 35

Example 3 - RSE/RRMSE Riviere, Ueckert (UMRS 1137) FIM 29 / 35

Example 4: Longitudinal binary LLB: Probability of response at time t: P (y = 1 b) = exp(β 1 + β 2 (1 µ 3 δ)t) 1 + exp(β 1 + β 2 (1 µ 3 δ)t) Fixed eects: (µ 1, µ 2, µ 3 ) = ( 1.0, 4.0, 0.4) Additive random eects with variances: (ω 2 1, ω 2 2) = (0.5, 4.0) 2 groups: δ = 0 and δ = 1 13 time points equally spaced between 0 and 1 time units for each patient N = 25 patients per group Subject i=1 Subject i=2 Subject i=26 Subject i=27 Response 1 0 δ 0 1 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 Time Riviere, Ueckert (UMRS 1137) FIM 30 / 35

Example 4 - RSE/RRMSE Riviere, Ueckert (UMRS 1137) FIM 31 / 35

Comparison: calculation time Time: PKW SC1 RTTE LLB CTS >5h >5h >5h >5h MCMC 6min 8min 2min 4min AGQ 2min 13min 10s 2min FO <5s <5s - - Riviere, Ueckert (UMRS 1137) FIM 32 / 35

Comparison: calculation time Time: PKW SC1 RTTE LLB CTS >5h >5h >5h >5h MCMC 6min 8min 2min 4min AGQ 2min 13min 10s 2min FO <5s <5s - - Riviere, Ueckert (UMRS 1137) FIM 32 / 35

Comparison: calculation time Time: PKW SC1 RTTE LLB CTS >5h >5h >5h >5h MCMC 6min 8min 2min 4min AGQ 2min 13min 10s 2min FO <5s <5s - - Riviere, Ueckert (UMRS 1137) FIM 32 / 35

Comparison: calculation time Time: PKW SC1 RTTE LLB CTS >5h >5h >5h >5h MCMC 6min 8min 2min 4min AGQ 2min 13min 10s 2min FO <5s <5s - - Riviere, Ueckert (UMRS 1137) FIM 32 / 35

Comparison: calculation time Time: PKW SC1 RTTE LLB CTS >5h >5h >5h >5h MCMC 6min 8min 2min 4min AGQ 2min 13min 10s 2min FO <5s <5s - - Riviere, Ueckert (UMRS 1137) FIM 32 / 35

Comparison: calculation time Time: PKW SC1 RTTE LLB CTS >5h >5h >5h >5h MCMC 6min 8min 2min 4min AGQ 2min 13min 10s 2min FO <5s <5s - - AGQ: time increases exponentially with the number of random parameters MCMC: time increases linearly Riviere, Ueckert (UMRS 1137) FIM 32 / 35

Outline 1 Introduction 2 Proposed approaches to compute the FIM General expression of the FIM AGQ-based approach MCMC-based approach 3 Evaluation by simulations 4 Summary Riviere, Ueckert (UMRS 1137) FIM 32 / 35

Summary Recent NLMEM (multivariate, complex, discrete,...) require improved method for FIM prediction Presented two complementing MC-based methods for calculating FIM Advantages: Adapted for discrete and continuous models No model linearization Very high agreement with clinical trial results Drawbacks: Much slower than FO approximation Monte-Carlo noise Riviere, Ueckert (UMRS 1137) FIM 33 / 35

Perspectives Publish R packages on CRAN Investigate design optimization: Handling of MC noise (stochastic approximation, simulated annealing,...) Adaptive approximation (rene approximation during optimization) Investigate alternative sampling methods (Latin hypercube sampling, quasi-random sampling,...) Riviere, Ueckert (UMRS 1137) FIM 34 / 35

Thank you for your attention! This work was supported by the Innovative Medicines Initiative Joint Undertaking under grant agreement no. 115156, resources of which are composed of nancial contributions from the European Union's Seventh Framework Program (FP7/2007 2013) and EFPIA companies' in kind contribution. The DDMoRe project is also supported by nancial contribution from Academic and SME partners. This work does not necessarily represent the view of all DDMoRe partners. This work has also received funding from the European Union's 7th Framework Programme for research, technological development and demonstration under Grant Agreement no 602552. Riviere, Ueckert (UMRS 1137) FIM 35 / 35