Pubh 8482: Sequential Analysis
|
|
- Prudence Jordan
- 5 years ago
- Views:
Transcription
1 Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 7
2 Course Summary To this point, we have discussed group sequential testing focusing on Maintaining the correct type-i error rate and power Decreasing the expected sample size These approaches only provide a yes or no answer as to whether or not we reject the null hypothesis Generally, more detail is provided when presenting results
3 Four-Number Summary In general, the following should always be reported when presenting results Point estimate Confidence Interval p-value
4 Impact of a Group Sequential Design Implementing a group sequential procedure will change the properties of standard point and interval estimators Group sequential procedures change the sampling distribution of standard estimators Confidence intervals derived from normal approximations will no longer have nominal coverage
5 Impact of a Group Sequential Design Implementing a group sequential procedure will change the properties of standard point and interval estimators Group sequential procedures change the sampling distribution of standard estimators Confidence intervals derived from normal approximations will no longer have nominal coverage We will start by considering distribution theory for group sequential design and then consider the implication for point and interval estimation
6 Set-up Let β be our ( parameter of interest and assueme that the sequence of estimates ˆβ1,..., ˆβ ) K follows a multivariate normal distribution with ( ) ˆβk N β, I 1 β,k for k = 1,..., K [ Cov ˆβk, ˆβ ] [ ] j = Var ˆβj = I 1 β,j for k j
7 Set-up ( ) Define ˆβ β0 = ˆθ β and (β β 0 ) = θ β. For Z k = ˆθ β,k Ik, the sequence of test statistics (Z 1,..., Z K ) follows a multivariate normal distribution with Z k N ( θ β Iβ,k, 1 ) for k = 1,..., K Cov [ Z k, Z j ] = Iβ,k /I β,j for k j
8 Notation Let T be the stage at which stopping occurs: T = min (k : Z k C k ) where C k is the continuation region at stage k and C K =
9 Notation Let Z (k) = (Z 1,..., Z k ) be the vector of the first k test statistics and for k = 1,..., K, define A k = {z (k) : z i C i, i = 1,..., k 1, and z k C k } i.e. A k is the set of sample paths that terminate at stage k.
10 Density of (Z 1,..., Z k ) The joint density of (Z 1,..., Z k ) follows a multivariate normal distribution as described above The joint density of (Z 1,..., Z k ) can also be written as a product of independent normal random variables by considering the following transformation
11 Transformations Consider the following transformation Let and y 1 = z 1 Iβ,1 1 = I β,1 y i = z i Iβ,i z i 1 Iβ,i 1 i = I β,i I β,i 1
12 Joint Density of (y 1,..., y k ) For i = 1 y 1 is normally distributed with ] E [y 1 ] = E [z 1 Iβ,1 = θ β,1 I β,1 = θ β 1 ] Var [y 1 ] = Var [z 1 Iβ,1 = I β,1 = 1 For i = 2,..., k y i is normally distributed with ] E [y i ] = E [z i Iβ,i z i 1 Iβ,i 1 = θ β (I β,i I β,i 1 ) = θ β i ] Var [y i ] = Var [z i Iβ,i z i 1 Iβ,i 1 = I β,1 = i
13 Joint Density of (y 1,..., y k ) More importantly, we know that the y i s are indendent. Cov [ y i, y j ] = 0 for i j Recall that the z i s have independent increments
14 Joint Density of (y 1,..., y k ) This means that we can write the joint density of (y 1,..., y k ) as the product of independent normally distributed random variables f T, yt (k, y k θ beta ) = k i=1 1 e ( y i i θ β) 2 2 i 2π i Therefore, the joint density of (z 1,..., z k ), f T, z ( k, z (k) θ β ), can be evaluated by evaluating f T, yt (k, y k θ beta ) for the correct y k.
15 Joint Density of (y 1,..., y k ): Example Assume you have the following sequence of test statistics: z (k) = (0.73,.25, 0.33, 0.10) and the following sequence of information I β,k = (3.53, 5.00, 6.12)
16 Joint Density of (y 1,..., y k ): Example The resulting sequence of y i s and i s are y k = (0.73, 3.83, 3.27, 1.31) and k = (3.54, 1.46, 1.12, 0.95)
17 Joint Density of (y 1,..., y k ): Example Therefore, the joint density of z (k) is f T,z (T ) ((0.73, 0.25, 0.33, 0.10), k θ β ) =f T, yt (k, (0.73, 3.83, 3.27, 1.31) θ β ) k 1 = e (y i i θ β) 2 2 i 2π i i=1 =
18 Equivalence of two joint distributions The preceding argument shows that the joint distribution of z (k) and y k are equivalent Therefore, we can simply study f T, yt to derive theoretical properties of z (k)
19 Joint Density of (y 1,..., y k ) We can re-write f T, yt (k, y k θ β ) as f T, yt (k, y k θ β ) = = k i=1 ( k 1 e ( y i i θ β) 2 2 i 2π i i=1 ) 1 e y k 2 2 i i θ β y i + 2 i θ2 β i=1 2 i 2π i ( k ) 1 = e y 2 i 2 i e θ βz k Iβ,k θ 2 β I β,k /2 2π i i=1 = h (k, y k, I 1,..., I k ) e θ βz k Iβ,k θ 2 β I β,k /2
20 Joint Density of (y 1,..., y k ) There are two primary implications from the previous results By factorization, we see that (Z T, T ) is sufficient for θ β Z T / I β,t is the MLE of θ β
21 Implications Implications of the sufficiency of (Z T, T ) for θ β The only information about θ β is contained in the stopping time and final Z That is, it only matters that you reached the kth stopping time. The exact path followed to the kth stopping time is irrelevant This should be somewhat intuitive given that the Z s have independent increments The final increment z k Iβ,k z k 1 Iβ,k 1 is independent of the first k 1 test statistics
22 Sub-densities of Z k To this point we have consider the joint density of (Z 1,..., Z k ) and (y 1,..., y k ) We might also consider the sub-densities of Z k, f (k, z k θ β ) The sub-densities can be found by integrating over all paths that result in terminating at the kth interim analysis
23 Sub-densities of Z k That is, the kth sub-density, f (k, z k θ β ) is defined as f (k, z k θ β ) = h (k, y k, I 1,..., I k ) e θ 2 βz k Iβ,k θβ I β,k /2 dy k 1... dy 1 B k( y) where B k ( y) is the set of all paths that result in terminating at the kth interim analysis
24 Sub-densities of Z k Note that if θ β = 0, f (k, z k θ β ) = h (k, y k, I 1,..., I k ) e θ 2 βz k Iβ,k θβ I β,k /2 dy k 1... dy 1 B k( y) = h (k, y k, I 1,..., I k ) dy k 1... dy 1 B k( y)
25 Sub-densities of Z k This implies that f (k, z k θ β ) = f (k, z k 0) e θ βz k Iβ,k θ 2 β I β,k /2 This is a helpful because it allows us to easily calculate sub-densities at multiple values of θ β.
26 Defining the sub-densities recursively The previous integral is potentially nasty Luckily, the sub-densities can be defined recursively, which aids in computation
27 Defining the sub-densities recursively The general form of the sub-densities is { g k (z θ β ) if z / C k f (k, z k θ β ) = 0 if z C k
28 Defining the sub-densities recursively Sub-density at the first interim analysis. At the first interim analysis, Z 1 is normally distributed with mean θ β I1 and variance 1 Therefore ) g 1 (z θ β ) = φ (z θ β I1
29 Defining the sub-densities recursively For k = 2,..., K, g k is defined recursively as ( Ik z I k u ) I k 1 k θ β g k (z θ β ) = g k 1 (u θ β ) φ du C k 1 k k
30 Defining the sub-densities recursively Essentially, each sub-density is the kernel of a normal density multiplied a factor accounting for the possibility of terminating early The inflation factor is determined by integrating over all sample paths that result in terminating at the kth interim analysis using the recursive procedure described before
31 Sub-Densities: Example Consider a group sequential design with O Brien-Fleming stopping boundaries and α = 0.10 k = 4 90% power to reject assuming that θ β = δ
32 Sub-Densities: Example theta = 0 sub densities z
33 Sub-Densities: Example theta =.5 * delta sub densities z
34 Sub-Densities: Example theta = delta sub densities z
35 Sub-Densities: Example theta = 1.5 * delta sub densities z
36 Sub-Densities and stopping times It should be noted that the sub-densities do not integrate to 1 Integrating each sub-density will give the probability of stopping at that interim analysis Pr (T = k θ β = θ) = f (k, z θ β = θ) dz z / C k In contrast, the sum of the k integrals will equal 1
37 Sub-Densities and stopping times: Example For example, if θ β = 0 and assuming the O Brien-Fleming design discussed before Pr (T = 1) = Pr (T = 2) = Pr (T = 3) = Pr (T = 2) =
38 Sub-Densities and stopping times: Example If θ β = δ Pr (T = 1) = Pr (T = 2) = Pr (T = 3) = Pr (T = 2) =
39 Estimating β To this point, we have considered distribution theory for a group sequential test of a general parameter β We are also interested in point and interval estimates of β In a fixed-sample test, point and interval estimates of β are based on a normal sampling distribution for ˆβ We have seen that implementing a group sequential procedure changes the sampling distribution of Z Group sequential procedures also change the sampling distribution of ˆβ and thus changes our approach to estimation after a group sequential test
40 Sampling distribution of ˆβ Previously, we defined the sub-densities of Z k, f (k, z k θ) It should be clear that the overall density is simply K f (z θ) = f (k, z k θ) How do we use this result to derive the sampling density of ˆβ? k=1
41 Sampling distribution of ˆβ Recall that Z k = ( ˆβ β0 ) Ik Therefore, the sampling density of ˆβ at ˆβ = y is f (y β) = K f k=1 ( k, (y β 0 ) I k θ) Ik Note that θ = (β β 0 ), in which case, conditioning on θ is synonimous to conditioning on β
42 Sampling distribution of ˆβ: example Consider the case where x 1, x 2,..., x 128 are i.i.d. N ( µ, σ 2 = 20 ) We want to complete a group sequential test of H 0 : µ = 0 In this case, ˆβ = X
43 Sampling distribution of ˆβ: example Consider a group sequential design with O Brien-Fleming stopping boundaries and α = 0.10 K = 4 In this case, I 1 I 4 = 1.6, 3.2, 4.8, 6.4
44 Density ˆβ: Example beta = 0 f(beta_hat) beta_hat
45 Density ˆβ: Example beta = 0.5 f(beta_hat) beta_hat
46 Density ˆβ: Example beta = 1 f(beta_hat) beta_hat
47 Density ˆβ: Example beta = 0 f(beta_hat) beta_hat
48 Density ˆβ: Example We see that the sampling distribution is substantially difference when a group sequential test is used The sampling distribution is no longer normal and, therefore, interval estimates based on the normal approximation are no longer valid The difference between the sampling density under the group sequential test and the usual sampling density becomes more dramatic as β moves away from the null hypothesis
49 Expected value of ˆβ The expected value of ˆβ after a group sequential test can be expressed as [ ] E β ˆβ = β 0 + K i=1 z / C k z Ik f (k, z β) dz For simplicity, we will now consider a two-stage design with cotinuation region C 1 = (a, b) in order to illustrate the bias due to a group sequential clinical trial
50 Expected value of ˆβ After a two-stage design The expected value of ˆβ after a two-stage design can be expressed as: [ a E ˆβ] = β z 1 I1 φ b a b z 1 φ (z 1 θ ) I 1 dz 1 I1 (z 1 θ I 1 ) dz 1 z 2 φ (z 1 θ ) ( ) z2 I2 z 1 I1 (I 2 I 2 ) θ I 1 φ dz 2 dz 1 I2 I 1 I2 I 1
51 Expected value of ˆβ After a two-stage design At stage 1, z 1 is a truncated normal random variable with mean θ I 1 and variance 1 and b a z 1 φ (z 1 θ ) I 1 dz 1 = θφ (a θ ) I 1 φ ( a θ ) I 1 I1 I1 z 1 I1 φ (z 1 θ I 1 ) dz 1 = θ ( 1 Φ (b θ I 1 )) + φ ( b θ I 1 ) I1
52 Expected value of ˆβ After a two-stage design Consider the double integral, we see that ( ) z 2 z2 I2 z 1 I1 (I 2 I 1 ) θ φ dz 2 I2 I 1 I2 I 1 Is simply the expected value of x I 2, where x is a normally distributed random variable with mean ( z 1 I1 I 2 I 1 θ ) and variance I 2 I 2. Therefore: z 2 I2 I 1 φ ( z2 I2 z 1 I1 (I 2 I 1 ) θ I2 I 1 ) dz 2 = z 1 I1 + (I 2 I 1 ) θ I 2
53 Expected value of ˆβ After a two-stage design Therefore, a b a = b = θi ( 1 Φ I 2 z 2 φ (z 1 θ ( ) ) z 2 I2 z 1 I1 (I 2 I 2 ) θ I 1 φ dz 2 dz 1 I2 I 1 I2 I 1 z 1 I1 + (I 2 I 1 ) θ I 2 (b θ ) I 1 Φ + (I 2 I 1 ) θ ( I 2 ( =θ Φ (b θ I 1 ) Φ Φ (b θ ) I 1 φ (z 1 θ ) I 1 dz 1 (a θ )) I 1 + ((φ(a θ ) ( I 1 φ b θ )) I 1 I1 /I 2 Φ (a θ )) I 1 (a θ )) I 1 + ((φ(a θ ) ( I 1 φ b θ )) I 1 I1 /I 2
54 Expected value of ˆβ After a two-stage design Summing everything up, we get a β b a + b z 1 φ (z 1 θ ) I 1 dz 1 I1 z 1 I1 φ (z 1 θ I 1 ) dz 1 z 2 φ (z 1 θ ( ) ) z 2 I2 z 1 I1 (I 2 I 2 ) θ I 1 φ dz 2 dz 1 I2 I 1 I2 I 1 =β 0 + θφ (a θ ) φ I 1 (a θ I 1 ) ( + θ 1 Φ (b θ )) φ (b θ ) I 1 I 1 + I1 I1 ( + θ Φ (b θ ) I 1 Φ (a θ )) I 1 + ((φ(a θ ) ( I 1 φ b θ )) I 1 I1 /I 2 =β + ((φ(b θ I 1 ) φ ( a θ I 1 )) I1 I 2 I1 I 2
55 Bias of ˆβ From the previous slide, we see that the bias in ˆβ is [ E ˆβ] = β + ((φ(b θ ) ( I 1 φ a θ )) I1 I 2 I 1 I1 I 2 = β + b (β) where the bias, b (β) depends on β a and b I 1 and I 2
56 Bias of ˆβ: Example Consider a two-stage design with O Brien-Fleming boundaries with α = 0.05 a 1 = 2.80 b 1 = 2.80 I 1 = 1.6 I 2 = 3.2
57 Bias of ˆβ: Example bias beta
58 Bias of ˆβ: Example What if we double the information? I 1 = 3.2 I 2 = 6.4
59 Bias of ˆβ: Example bias beta
60 Bias of ˆβ: Example Our first example considered symmetric bounds In this case, the bias was naturally symmetric about 0 What if we use asymmetric bounds? a 1 = 0.38 b 1 = 2.00 Information same as before
61 Bias of ˆβ: Example bias beta
62 Bias of ˆβ: Example Again, doubling the information I 1 = 3.2 I 2 = 6.4
63 Bias of ˆβ: Example bias beta
64 Bias of ˆβ: Summary Implementing a group sequential procedure results in substantial bias for ˆβ Bias is smallest at the extremes and in the middle of the continuation region Where the study either stops early or continues to full enrollment with high probability Bias is symmetric for symmetric bounds and asymmetric for asymmetric bounds
65 Correcting the Bias We will consider two estimators for correcting the biased caused by a group sequential design Whitehead s mean adjusted estimator UMVUE suggested by Emerson and Fleming
66 Whitehead s Mean adjusted Estimator Whitehead s mean adjusted estimator is defined as ˆβ w, such that ˆβ = ˆβ ( ) w + b ˆβw That is, whitehead s mean adjusted estimator is the true value of beta that results in an expectation equal to the observed ˆβ
67 Properties of Whitehead s Mean adjusted estimator ˆβw must be found by numerical search ˆβw is only biased adjusted and not unbiased
68 UMVUE Emerson and Fleming proposed the UMVUE defined as [ ˆβ umvue = E ˆβ 1 (T, Z T )] Where ˆβ 1 is the estimate of ˆβ 1 after stage 1 Note that ˆβ 1 is an unbiased estimator of β and we find the UMVUE by the Rao-Blackwell technique
69 Properties of the UMVUE ˆβumvue has the minimum variance among the class of unbiased estimators Unbiasedness is a restrictive property and the set of unbiased estimators is narrow This estimator has substantial bias and, in fact, has larger MSE than ˆβ w
70 Estimating β Implementing a group sequential design dramatically impacts the sampling distribution of ˆβ This results on substantial bias in ˆβ depending on the true value of β Unbiased or bias-reduced estimators have been proposed but we need to be mindful of the mean-variance trade-off when evaluating these estimators
Pubh 8482: Sequential Analysis
Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 8 P-values When reporting results, we usually report p-values in place of reporting whether or
More informationPubh 8482: Sequential Analysis
Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 5 Course Summary So far, we have discussed Group sequential procedures for two-sided tests Group
More informationPubh 8482: Sequential Analysis
Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 12 Review So far... We have discussed the role of phase III clinical trials in drug development
More informationInterim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming. and Optimal Stopping
Interim Monitoring of Clinical Trials: Decision Theory, Dynamic Programming and Optimal Stopping Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj
More informationPubh 8482: Sequential Analysis
Pubh 8482: Sequential Analysis Joseph S. Koopmeiners Division of Biostatistics University of Minnesota Week 10 Class Summary Last time... We began our discussion of adaptive clinical trials Specifically,
More informationGroup Sequential Designs: Theory, Computation and Optimisation
Group Sequential Designs: Theory, Computation and Optimisation Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj 8th International Conference
More informationOptimising Group Sequential Designs. Decision Theory, Dynamic Programming. and Optimal Stopping
: Decision Theory, Dynamic Programming and Optimal Stopping Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj InSPiRe Conference on Methodology
More informationFall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.
1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n
More informationEstimation in Flexible Adaptive Designs
Estimation in Flexible Adaptive Designs Werner Brannath Section of Medical Statistics Core Unit for Medical Statistics and Informatics Medical University of Vienna BBS and EFSPI Scientific Seminar on Adaptive
More informationRegression #3: Properties of OLS Estimator
Regression #3: Properties of OLS Estimator Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #3 1 / 20 Introduction In this lecture, we establish some desirable properties associated with
More information4. Issues in Trial Monitoring
4. Issues in Trial Monitoring 4.1 Elements of Trial Monitoring Monitoring trial process and quality Infrastructure requirements and DSMBs 4.2 Interim analyses: group sequential trial design 4.3 Group sequential
More informationGroup Sequential Tests for Delayed Responses. Christopher Jennison. Lisa Hampson. Workshop on Special Topics on Sequential Methodology
Group Sequential Tests for Delayed Responses Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Lisa Hampson Department of Mathematics and Statistics,
More informationMath 494: Mathematical Statistics
Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationMethods of evaluating estimators and best unbiased estimators Hamid R. Rabiee
Stochastic Processes Methods of evaluating estimators and best unbiased estimators Hamid R. Rabiee 1 Outline Methods of Mean Squared Error Bias and Unbiasedness Best Unbiased Estimators CR-Bound for variance
More informationOverrunning in Clinical Trials: a Methodological Review
Overrunning in Clinical Trials: a Methodological Review Dario Gregori Unit of Biostatistics, Epidemiology and Public Health Department of Cardiac, Thoracic and Vascular Sciences dario.gregori@unipd.it
More informationMonitoring clinical trial outcomes with delayed response: incorporating pipeline data in group sequential designs. Christopher Jennison
Monitoring clinical trial outcomes with delayed response: incorporating pipeline data in group sequential designs Christopher Jennison Department of Mathematical Sciences, University of Bath http://people.bath.ac.uk/mascj
More informationData Mining Stat 588
Data Mining Stat 588 Lecture 02: Linear Methods for Regression Department of Statistics & Biostatistics Rutgers University September 13 2011 Regression Problem Quantitative generic output variable Y. Generic
More information557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE
557: MATHEMATICAL STATISTICS II BIAS AND VARIANCE An estimator, T (X), of θ can be evaluated via its statistical properties. Typically, two aspects are considered: Expectation Variance either in terms
More informationEvaluating the Performance of Estimators (Section 7.3)
Evaluating the Performance of Estimators (Section 7.3) Example: Suppose we observe X 1,..., X n iid N(θ, σ 2 0 ), with σ2 0 known, and wish to estimate θ. Two possible estimators are: ˆθ = X sample mean
More informationInference in Regression Analysis
Inference in Regression Analysis Dr. Frank Wood Frank Wood, fwood@stat.columbia.edu Linear Regression Models Lecture 4, Slide 1 Today: Normal Error Regression Model Y i = β 0 + β 1 X i + ǫ i Y i value
More informationMS&E 226: Small Data. Lecture 11: Maximum likelihood (v2) Ramesh Johari
MS&E 226: Small Data Lecture 11: Maximum likelihood (v2) Ramesh Johari ramesh.johari@stanford.edu 1 / 18 The likelihood function 2 / 18 Estimating the parameter This lecture develops the methodology behind
More information[y i α βx i ] 2 (2) Q = i=1
Least squares fits This section has no probability in it. There are no random variables. We are given n points (x i, y i ) and want to find the equation of the line that best fits them. We take the equation
More informationBias Variance Trade-off
Bias Variance Trade-off The mean squared error of an estimator MSE(ˆθ) = E([ˆθ θ] 2 ) Can be re-expressed MSE(ˆθ) = Var(ˆθ) + (B(ˆθ) 2 ) MSE = VAR + BIAS 2 Proof MSE(ˆθ) = E((ˆθ θ) 2 ) = E(([ˆθ E(ˆθ)]
More informationUnbiased Estimation. Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others.
Unbiased Estimation Binomial problem shows general phenomenon. An estimator can be good for some values of θ and bad for others. To compare ˆθ and θ, two estimators of θ: Say ˆθ is better than θ if it
More informationBIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation
BIO5312 Biostatistics Lecture 13: Maximum Likelihood Estimation Yujin Chung November 29th, 2016 Fall 2016 Yujin Chung Lec13: MLE Fall 2016 1/24 Previous Parametric tests Mean comparisons (normality assumption)
More informationLecture 15 (Part 2): Logistic Regression & Common Odds Ratio, (With Simulations)
Lecture 15 (Part 2): Logistic Regression & Common Odds Ratio, (With Simulations) Dipankar Bandyopadhyay, Ph.D. BMTRY 711: Analysis of Categorical Data Spring 2011 Division of Biostatistics and Epidemiology
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 15: Examples of hypothesis tests (v5) Ramesh Johari ramesh.johari@stanford.edu 1 / 32 The recipe 2 / 32 The hypothesis testing recipe In this lecture we repeatedly apply the
More informationThe Design of a Survival Study
The Design of a Survival Study The design of survival studies are usually based on the logrank test, and sometimes assumes the exponential distribution. As in standard designs, the power depends on The
More informationGroup sequential designs with negative binomial data
Group sequential designs with negative binomial data Ekkehard Glimm 1 Tobias Mütze 2,3 1 Statistical Methodology, Novartis, Basel, Switzerland 2 Department of Medical Statistics, University Medical Center
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationMultiple Testing in Group Sequential Clinical Trials
Multiple Testing in Group Sequential Clinical Trials Tian Zhao Supervisor: Michael Baron Department of Mathematical Sciences University of Texas at Dallas txz122@utdallas.edu 7/2/213 1 Sequential statistics
More informationOriginal citation: Kimani, Peter K., Todd, Susan and Stallard, Nigel. (03) Conditionally unbiased estimation in phase II/III clinical trials with early stopping for futility. Statistics in Medicine. ISSN
More informationGraduate Econometrics I: Unbiased Estimation
Graduate Econometrics I: Unbiased Estimation Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Unbiased Estimation
More informationThe Design of Group Sequential Clinical Trials that Test Multiple Endpoints
The Design of Group Sequential Clinical Trials that Test Multiple Endpoints Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj Bruce Turnbull
More informationINTERVAL ESTIMATION AND HYPOTHESES TESTING
INTERVAL ESTIMATION AND HYPOTHESES TESTING 1. IDEA An interval rather than a point estimate is often of interest. Confidence intervals are thus important in empirical work. To construct interval estimates,
More informationCombining Biased and Unbiased Estimators in High Dimensions. (joint work with Ed Green, Rutgers University)
Combining Biased and Unbiased Estimators in High Dimensions Bill Strawderman Rutgers University (joint work with Ed Green, Rutgers University) OUTLINE: I. Introduction II. Some remarks on Shrinkage Estimators
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2009 Prof. Gesine Reinert Our standard situation is that we have data x = x 1, x 2,..., x n, which we view as realisations of random
More informationReview of probability and statistics 1 / 31
Review of probability and statistics 1 / 31 2 / 31 Why? This chapter follows Stock and Watson (all graphs are from Stock and Watson). You may as well refer to the appendix in Wooldridge or any other introduction
More informationCentral Limit Theorem ( 5.3)
Central Limit Theorem ( 5.3) Let X 1, X 2,... be a sequence of independent random variables, each having n mean µ and variance σ 2. Then the distribution of the partial sum S n = X i i=1 becomes approximately
More informationBios 6649: Clinical Trials - Statistical Design and Monitoring
Bios 6649: Clinical Trials - Statistical Design and Monitoring Spring Semester 2015 John M. Kittelson Department of Biostatistics & Informatics Colorado School of Public Health University of Colorado Denver
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some
More informationECE531 Lecture 10b: Maximum Likelihood Estimation
ECE531 Lecture 10b: Maximum Likelihood Estimation D. Richard Brown III Worcester Polytechnic Institute 05-Apr-2011 Worcester Polytechnic Institute D. Richard Brown III 05-Apr-2011 1 / 23 Introduction So
More informationIEOR E4703: Monte-Carlo Simulation
IEOR E4703: Monte-Carlo Simulation Output Analysis for Monte-Carlo Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Output Analysis
More informationApplied Statistics and Econometrics
Applied Statistics and Econometrics Lecture 6 Saul Lach September 2017 Saul Lach () Applied Statistics and Econometrics September 2017 1 / 53 Outline of Lecture 6 1 Omitted variable bias (SW 6.1) 2 Multiple
More informationRerandomization to Balance Covariates
Rerandomization to Balance Covariates Kari Lock Morgan Department of Statistics Penn State University Joint work with Don Rubin University of Minnesota Biostatistics 4/27/16 The Gold Standard Randomized
More informationPrevious lecture. Single variant association. Use genome-wide SNPs to account for confounding (population substructure)
Previous lecture Single variant association Use genome-wide SNPs to account for confounding (population substructure) Estimation of effect size and winner s curse Meta-Analysis Today s outline P-value
More informationRegression Estimation - Least Squares and Maximum Likelihood. Dr. Frank Wood
Regression Estimation - Least Squares and Maximum Likelihood Dr. Frank Wood Least Squares Max(min)imization Function to minimize w.r.t. β 0, β 1 Q = n (Y i (β 0 + β 1 X i )) 2 i=1 Minimize this by maximizing
More informationMultiple Linear Regression
Multiple Linear Regression ST 430/514 Recall: a regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates).
More informationSTATS 200: Introduction to Statistical Inference. Lecture 29: Course review
STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout
More informationMultiple regression. CM226: Machine Learning for Bioinformatics. Fall Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar
Multiple regression CM226: Machine Learning for Bioinformatics. Fall 2016 Sriram Sankararaman Acknowledgments: Fei Sha, Ameet Talwalkar Multiple regression 1 / 36 Previous two lectures Linear and logistic
More informationEconometrics Review questions for exam
Econometrics Review questions for exam Nathaniel Higgins nhiggins@jhu.edu, 1. Suppose you have a model: y = β 0 x 1 + u You propose the model above and then estimate the model using OLS to obtain: ŷ =
More informationMachine Learning for OR & FE
Machine Learning for OR & FE Supervised Learning: Regression I Martin Haugh Department of Industrial Engineering and Operations Research Columbia University Email: martin.b.haugh@gmail.com Some of the
More informationSAMPLE SIZE RE-ESTIMATION FOR ADAPTIVE SEQUENTIAL DESIGN IN CLINICAL TRIALS
Journal of Biopharmaceutical Statistics, 18: 1184 1196, 2008 Copyright Taylor & Francis Group, LLC ISSN: 1054-3406 print/1520-5711 online DOI: 10.1080/10543400802369053 SAMPLE SIZE RE-ESTIMATION FOR ADAPTIVE
More informationStatistical inference
Statistical inference Contents 1. Main definitions 2. Estimation 3. Testing L. Trapani MSc Induction - Statistical inference 1 1 Introduction: definition and preliminary theory In this chapter, we shall
More informationEstimation of Parameters
CHAPTER Probability, Statistics, and Reliability for Engineers and Scientists FUNDAMENTALS OF STATISTICAL ANALYSIS Second Edition A. J. Clark School of Engineering Department of Civil and Environmental
More informationStatistics Ph.D. Qualifying Exam: Part II November 9, 2002
Statistics Ph.D. Qualifying Exam: Part II November 9, 2002 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. 1 2 3 4 5 6 7 8 9 10 11 12 2. Write your
More informationSYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions
SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu
More informationORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing
ORF 245 Fundamentals of Statistics Chapter 9 Hypothesis Testing Robert Vanderbei Fall 2014 Slides last edited on November 24, 2014 http://www.princeton.edu/ rvdb Coin Tossing Example Consider two coins.
More informationEstimators as Random Variables
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until
More informationMultivariate Time Series: VAR(p) Processes and Models
Multivariate Time Series: VAR(p) Processes and Models A VAR(p) model, for p > 0 is X t = φ 0 + Φ 1 X t 1 + + Φ p X t p + A t, where X t, φ 0, and X t i are k-vectors, Φ 1,..., Φ p are k k matrices, with
More informationIntroduction to Maximum Likelihood Estimation
Introduction to Maximum Likelihood Estimation Eric Zivot July 26, 2012 The Likelihood Function Let 1 be an iid sample with pdf ( ; ) where is a ( 1) vector of parameters that characterize ( ; ) Example:
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2
MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and
More informationReview of Econometrics
Review of Econometrics Zheng Tian June 5th, 2017 1 The Essence of the OLS Estimation Multiple regression model involves the models as follows Y i = β 0 + β 1 X 1i + β 2 X 2i + + β k X ki + u i, i = 1,...,
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationUniversity of North Texas Health Science Center at Fort Worth
A Class of Parameter Functions for Which the Unbiased Estimator Does Not Έxist BY SHANDE CHEN University of North Texas Health Science Center at Fort Worth Unbiased estimation is a popular criterion in
More informationThe SEQDESIGN Procedure
SAS/STAT 9.2 User s Guide, Second Edition The SEQDESIGN Procedure (Book Excerpt) This document is an individual chapter from the SAS/STAT 9.2 User s Guide, Second Edition. The correct bibliographic citation
More informationThe regression model with one stochastic regressor.
The regression model with one stochastic regressor. 3150/4150 Lecture 6 Ragnar Nymoen 30 January 2012 We are now on Lecture topic 4 The main goal in this lecture is to extend the results of the regression
More informationf (1 0.5)/n Z =
Math 466/566 - Homework 4. We want to test a hypothesis involving a population proportion. The unknown population proportion is p. The null hypothesis is p = / and the alternative hypothesis is p > /.
More informationA General Overview of Parametric Estimation and Inference Techniques.
A General Overview of Parametric Estimation and Inference Techniques. Moulinath Banerjee University of Michigan September 11, 2012 The object of statistical inference is to glean information about an underlying
More informationA Type of Sample Size Planning for Mean Comparison in Clinical Trials
Journal of Data Science 13(2015), 115-126 A Type of Sample Size Planning for Mean Comparison in Clinical Trials Junfeng Liu 1 and Dipak K. Dey 2 1 GCE Solutions, Inc. 2 Department of Statistics, University
More informationA Very Brief Summary of Statistical Inference, and Examples
A Very Brief Summary of Statistical Inference, and Examples Trinity Term 2008 Prof. Gesine Reinert 1 Data x = x 1, x 2,..., x n, realisations of random variables X 1, X 2,..., X n with distribution (model)
More informationProblem Selected Scores
Statistics Ph.D. Qualifying Exam: Part II November 20, 2010 Student Name: 1. Answer 8 out of 12 problems. Mark the problems you selected in the following table. Problem 1 2 3 4 5 6 7 8 9 10 11 12 Selected
More informationInverse Sampling for McNemar s Test
International Journal of Statistics and Probability; Vol. 6, No. 1; January 27 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education Inverse Sampling for McNemar s Test
More informationROI ANALYSIS OF PHARMAFMRI DATA:
ROI ANALYSIS OF PHARMAFMRI DATA: AN ADAPTIVE APPROACH FOR GLOBAL TESTING Giorgos Minas, John A.D. Aston, Thomas E. Nichols and Nigel Stallard Department of Statistics and Warwick Centre of Analytical Sciences,
More informationLet us first identify some classes of hypotheses. simple versus simple. H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided
Let us first identify some classes of hypotheses. simple versus simple H 0 : θ = θ 0 versus H 1 : θ = θ 1. (1) one-sided H 0 : θ θ 0 versus H 1 : θ > θ 0. (2) two-sided; null on extremes H 0 : θ θ 1 or
More informationIndependent Increments in Group Sequential Tests: A Review
Independent Increments in Group Sequential Tests: A Review KyungMann Kim kmkim@biostat.wisc.edu University of Wisconsin-Madison, Madison, WI, USA July 13, 2013 Outline Early Sequential Analysis Independent
More informationV. Properties of estimators {Parts C, D & E in this file}
A. Definitions & Desiderata. model. estimator V. Properties of estimators {Parts C, D & E in this file}. sampling errors and sampling distribution 4. unbiasedness 5. low sampling variance 6. low mean squared
More informationAnswers to Problem Set #4
Answers to Problem Set #4 Problems. Suppose that, from a sample of 63 observations, the least squares estimates and the corresponding estimated variance covariance matrix are given by: bβ bβ 2 bβ 3 = 2
More informationSAS/STAT 15.1 User s Guide The SEQDESIGN Procedure
SAS/STAT 15.1 User s Guide The SEQDESIGN Procedure This document is an individual chapter from SAS/STAT 15.1 User s Guide. The correct bibliographic citation for this manual is as follows: SAS Institute
More informationEconometric Methods. Prediction / Violation of A-Assumptions. Burcu Erdogan. Universität Trier WS 2011/2012
Econometric Methods Prediction / Violation of A-Assumptions Burcu Erdogan Universität Trier WS 2011/2012 (Universität Trier) Econometric Methods 30.11.2011 1 / 42 Moving on to... 1 Prediction 2 Violation
More informationBios 6649: Clinical Trials - Statistical Design and Monitoring
Bios 6649: Clinical Trials - Statistical Design and Monitoring Spring Semester 2015 John M. Kittelson Department of Biostatistics & Informatics Colorado School of Public Health University of Colorado Denver
More informationAn Adaptive Futility Monitoring Method with Time-Varying Conditional Power Boundary
An Adaptive Futility Monitoring Method with Time-Varying Conditional Power Boundary Ying Zhang and William R. Clarke Department of Biostatistics, University of Iowa 200 Hawkins Dr. C-22 GH, Iowa City,
More informationProblem Set 6 Solution
Problem Set 6 Solution May st, 009 by Yang. Causal Expression of AR Let φz : αz βz. Zeros of φ are α and β, both of which are greater than in absolute value by the assumption in the question. By the theorem
More informationMaster s Written Examination
Master s Written Examination Option: Statistics and Probability Spring 016 Full points may be obtained for correct answers to eight questions. Each numbered question which may have several parts is worth
More informationStatement: With my signature I confirm that the solutions are the product of my own work. Name: Signature:.
MATHEMATICAL STATISTICS Homework assignment Instructions Please turn in the homework with this cover page. You do not need to edit the solutions. Just make sure the handwriting is legible. You may discuss
More informationTwo hours. To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER. 21 June :45 11:45
Two hours MATH20802 To be supplied by the Examinations Office: Mathematical Formula Tables THE UNIVERSITY OF MANCHESTER STATISTICAL METHODS 21 June 2010 9:45 11:45 Answer any FOUR of the questions. University-approved
More informationStat 535 C - Statistical Computing & Monte Carlo Methods. Lecture 15-7th March Arnaud Doucet
Stat 535 C - Statistical Computing & Monte Carlo Methods Lecture 15-7th March 2006 Arnaud Doucet Email: arnaud@cs.ubc.ca 1 1.1 Outline Mixture and composition of kernels. Hybrid algorithms. Examples Overview
More informationMaster s Written Examination - Solution
Master s Written Examination - Solution Spring 204 Problem Stat 40 Suppose X and X 2 have the joint pdf f X,X 2 (x, x 2 ) = 2e (x +x 2 ), 0 < x < x 2
More informationMA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7
MA 575 Linear Models: Cedric E. Ginestet, Boston University Midterm Review Week 7 1 Random Vectors Let a 0 and y be n 1 vectors, and let A be an n n matrix. Here, a 0 and A are non-random, whereas y is
More informationSpring 2012 Math 541A Exam 1. X i, S 2 = 1 n. n 1. X i I(X i < c), T n =
Spring 2012 Math 541A Exam 1 1. (a) Let Z i be independent N(0, 1), i = 1, 2,, n. Are Z = 1 n n Z i and S 2 Z = 1 n 1 n (Z i Z) 2 independent? Prove your claim. (b) Let X 1, X 2,, X n be independent identically
More informationMathematical statistics
October 4 th, 2018 Lecture 12: Information Where are we? Week 1 Week 2 Week 4 Week 7 Week 10 Week 14 Probability reviews Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter
More information1. (Rao example 11.15) A study measures oxygen demand (y) (on a log scale) and five explanatory variables (see below). Data are available as
ST 51, Summer, Dr. Jason A. Osborne Homework assignment # - Solutions 1. (Rao example 11.15) A study measures oxygen demand (y) (on a log scale) and five explanatory variables (see below). Data are available
More information5. Erroneous Selection of Exogenous Variables (Violation of Assumption #A1)
5. Erroneous Selection of Exogenous Variables (Violation of Assumption #A1) Assumption #A1: Our regression model does not lack of any further relevant exogenous variables beyond x 1i, x 2i,..., x Ki and
More informationVisual interpretation with normal approximation
Visual interpretation with normal approximation H 0 is true: H 1 is true: p =0.06 25 33 Reject H 0 α =0.05 (Type I error rate) Fail to reject H 0 β =0.6468 (Type II error rate) 30 Accept H 1 Visual interpretation
More informationTesting a secondary endpoint after a group sequential test. Chris Jennison. 9th Annual Adaptive Designs in Clinical Trials
Testing a secondary endpoint after a group sequential test Christopher Jennison Department of Mathematical Sciences, University of Bath, UK http://people.bath.ac.uk/mascj 9th Annual Adaptive Designs in
More informationReview Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the
Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then
More informationTerminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationReview. December 4 th, Review
December 4 th, 2017 Att. Final exam: Course evaluation Friday, 12/14/2018, 10:30am 12:30pm Gore Hall 115 Overview Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 6: Statistics and Sampling Distributions Chapter
More information