Inferences for the Ratio: Fieller s Interval, Log Ratio, and Large Sample Based Confidence Intervals

Similar documents
Linear models and their mathematical foundations: Simple linear regression

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

Finite Population Correction Methods

Confidence Intervals for the Process Capability Index C p Based on Confidence Intervals for Variance under Non-Normality

STA 2201/442 Assignment 2

Estimation and sample size calculations for correlated binary error rates of biometric identification devices

Bootstrap inference for the finite population total under complex sampling designs

Better Bootstrap Confidence Intervals

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

A Simulation Study on Confidence Interval Procedures of Some Mean Cumulative Function Estimators

The Nonparametric Bootstrap

Ch 2: Simple Linear Regression

Statistical Hypothesis Testing

IEOR E4703: Monte-Carlo Simulation

The assumptions are needed to give us... valid standard errors valid confidence intervals valid hypothesis tests and p-values

Reliable Inference in Conditions of Extreme Events. Adriana Cornea

Econ 371 Problem Set #1 Answer Sheet

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

Supporting Information for Estimating restricted mean. treatment effects with stacked survival models

Unit 14: Nonparametric Statistical Methods

Probability Theory and Statistics. Peter Jochumzen

Block Bootstrap Prediction Intervals for Vector Autoregression

Linear Models and Estimation by Least Squares

Confidence Intervals of the Simple Difference between the Proportions of a Primary Infection and a Secondary Infection, Given the Primary Infection

Heteroskedasticity-Robust Inference in Finite Samples

COMPOSITE RELIABILITY MODELS FOR SYSTEMS WITH TWO DISTINCT KINDS OF STOCHASTIC DEPENDENCES BETWEEN THEIR COMPONENTS LIFE TIMES

Estimation of AUC from 0 to Infinity in Serial Sacrifice Designs

Empirical Likelihood Inference for Two-Sample Problems

MFin Econometrics I Session 4: t-distribution, Simple Linear Regression, OLS assumptions and properties of OLS estimators

Kevin Ewans Shell International Exploration and Production

High-dimensional regression

Estimation of uncertainties using the Guide to the expression of uncertainty (GUM)

Algebra II Unit Breakdown (Curriculum Map Outline)

Summary and discussion of: Exact Post-selection Inference for Forward Stepwise and Least Angle Regression Statistics Journal Club

SUPPLEMENT TO PARAMETRIC OR NONPARAMETRIC? A PARAMETRICNESS INDEX FOR MODEL SELECTION. University of Minnesota

A Note on Coverage Probability of Confidence Interval for the Difference between Two Normal Variances

Asymptotic Statistics-VI. Changliang Zou

11. Bootstrap Methods

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Conservative variance estimation for sampling designs with zero pairwise inclusion probabilities

Bayesian nonparametric estimation of finite population quantities in absence of design information on nonsampled units

Chapter 8 - Statistical intervals for a single sample

Do not copy, post, or distribute

UQ, Semester 1, 2017, Companion to STAT2201/CIVL2530 Exam Formulae and Tables

Bootstrap (Part 3) Christof Seiler. Stanford University, Spring 2016, Stats 205

A GEOMETRIC APPROACH TO CONFIDENCE SETS FOR RATIOS: FIELLER S THEOREM, GENERALIZATIONS AND BOOTSTRAP

Review of Statistics

STAT 512 sp 2018 Summary Sheet

Asymptotic distribution of the sample average value-at-risk

1/24/2008. Review of Statistical Inference. C.1 A Sample of Data. C.2 An Econometric Model. C.4 Estimating the Population Variance and Other Moments

A Significance Test for the Lasso

Topic 2: Probability & Distributions. Road Map Probability & Distributions. ECO220Y5Y: Quantitative Methods in Economics. Dr.

2008 Winton. Statistical Testing of RNGs

Week 2: Review of probability and statistics

Simple Linear Regression (Part 3)

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Confidence Intervals for the Ratio of Two Exponential Means with Applications to Quality Control

Statistical Inference

A Bootstrap Test for Conditional Symmetry

n =10,220 observations. Smaller samples analyzed here to illustrate sample size effect.

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 26. Estimation: Regression and Least Squares

MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems

A noninformative Bayesian approach to domain estimation

Empirical Likelihood Methods for Sample Survey Data: An Overview

Review of Statistics 101

Applied Microeconometrics (L5): Panel Data-Basics

Simulating Uniform- and Triangular- Based Double Power Method Distributions

Course information: Instructor: Tim Hanson, Leconte 219C, phone Office hours: Tuesday/Thursday 11-12, Wednesday 10-12, and by appointment.

Monte Carlo Study on the Successive Difference Replication Method for Non-Linear Statistics

Hypothesis Testing hypothesis testing approach

Lecture 2: Basic Concepts and Simple Comparative Experiments Montgomery: Chapter 2

Motivation for multiple regression

An Introduction to Parameter Estimation

On robust and efficient estimation of the center of. Symmetry.

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

Bootstrap. Director of Center for Astrostatistics. G. Jogesh Babu. Penn State University babu.

UNIVERSITÄT POTSDAM Institut für Mathematik

Learning Objectives for Stat 225

Chapter 3. Comparing two populations

The Distribution of F

The Delta Method and Applications

Measuring the fit of the model - SSR

Simple linear regression

Summary of Chapter 7 (Sections ) and Chapter 8 (Section 8.1)

The Prediction of Monthly Inflation Rate in Romania 1

Bayesian Linear Regression

Inference for Regression

Mathematical statistics

Institute of Actuaries of India

Math 494: Mathematical Statistics

Math 423/533: The Main Theoretical Topics

Overall Plan of Simulation and Modeling I. Chapters

STT 843 Key to Homework 1 Spring 2018

ON THE NUMBER OF BOOTSTRAP REPETITIONS FOR BC a CONFIDENCE INTERVALS. DONALD W. K. ANDREWS and MOSHE BUCHINSKY COWLES FOUNDATION PAPER NO.

Exercises and Answers to Chapter 1

Confidence Intervals. Confidence interval for sample mean. Confidence interval for sample mean. Confidence interval for sample mean

Double Bootstrap Confidence Interval Estimates with Censored and Truncated Data

Lecture 1: August 28

Business Statistics. Lecture 10: Course Review

Transcription:

Inferences for the Ratio: Fieller s Interval, Log Ratio, and Large Sample Based Confidence Intervals Michael Sherman Department of Statistics, 3143 TAMU, Texas A&M University, College Station, Texas 77843, USA sherman@stat.tamu.edu Arnab Maity Department of Statistics, North Carolina State University, 2311 Stinson Drive, Raleigh, North Carolina 27695, U.S.A. amaity@ncsu.edu Suojin Wang Department of Statistics, 3143 TAMU, Texas A&M University, College Station, Texas 77843, USA sjwang@stat.tamu.edu Abstract In sample surveys and many other areas of application, the ratio of variables is often of great importance. This often occurs when one variable is available at the population level while another variable of interest is available for sample data only. In this case, using the sample ratio, we can often gather valuable information on the variable of interest for the unsampled observations. In many other studies, the ratio itself is of interest, for example when estimating proportions from a random number of observations. In this note we compare three confidence intervals for the population ratio: A large sample interval, a log based version of the large sample interval, and Fieller s interval. This is done through data analysis and through a small simulation experiment. The Fieller method has often been proposed as a superior interval for small sample sizes. We show through a data example and simulation experiments that Fieller s method often gives nonsensical and uninformative intervals when the observations are noisy relative to the mean of the data. The large sample interval does not similarly suffer and thus can be a more reliable method for small and large samples. Some key words: Fieller s interval, ratio estimation, variance estimation, sample surveys, small sample inference.

1 Introduction In sample surveys and many other areas of application, the ratio of variables is often of great importance. This often occurs when one variable is available at the population level while another variable of interest is available for sample data only. In this case, if the ratio of the two variables is estimable we gather valuable information on the variable of interest for the unsampled observations. In other studies, the ratio itself is of interest. One such example is the estimation of proportions where the number of observations is random, as is often the case in cluster sampling, for example. Other examples include applications to cost effectiveness in economics (Jiang, Wu, and Williams, 2000), studying the ratio of regression coefficients (Hirschberg and Lye, 2007) and comparing health outcomes across spatial domains (Beyene and Moineddin, 2005). For these reasons estimation of the true ratio between variables is of great interest. 2 Methodological Development There are two basic frameworks in which we draw inferences: the infinite population and the finite population settings. For the former, we consider our observations as coming from a bivariate distribution function, F (x, y), with correlation coefficient ρ. In the latter, our data observations come from a finite population (X, Y ), X = (x 1,..., x N ), Y = (y 1,..., y N ) of size N. In the former case the population ratio is defined to be R = E(Y )/E(X) = µ y /µ x, where the expectations, µ y and µ x, are the population means of the variables. In the latter case the population ratio is defined to be R = N i=1 y i / N i=1 x i = Ȳ / X. Our results are not specific to the finite or infinite population case, and in either case we denote a random sample by the sample elements chosen as S = (i 1,..., i n ). Then the sample data is (x, y), where x = x(s), y = y(s). For either the infinite population or finite population setting, we consider the sample ratio to be the ratio of sample means: r = i S y i/ i S x i = ȳ/ x where x, ȳ are the sample means of the observations x = x(s) and y = y(s). 1

We desire inferences on the population ratio, R, based on the sample ratio r. Define the sample standard deviations of x = x(s) and y = y(s) as s x and s y. Further, define c x x = n 1 s 2 x/ x 2, cȳȳ = n 1 s 2 y/ȳ 2 and cȳ x = n 1 s xy /( xȳ). Large sample approximations, e.g., Cochran (1977) or Lohr (2009), show that σ C 2 = i S (y i rx i ) 2 /{n(n 1) x 2 } = r 2 [cȳȳ + c x x 2cȳ x ] is a consistent variance estimator of r so that a large sample confidence interval is given by I C = (r ± t α/2,n 1 r cȳȳ + c x x 2cȳ x ), where t α,n 1 denotes the (1 α)100% quantile of a t n 1 distribution. Note that we have assumed that in the finite population situation that the sampling fraction n/n is small enough so that we can consider the finite population correction, fpc = 1 n/n, to be equal to one. We assume this throughout the present study. If this is not the case then all variance estimates can be adjusted accordingly. As an alternative to the previous large sample interval, one can use the log ratio as a pivotal quantity, construct a confidence interval around the estimated log ratio and exponentiate both end points to obtain an interval for the actual ratio. Specifically, using a Taylor s expansion we derive var{log(r)} = var{log(ȳ) log( x)} var{log(ȳ ) log( X) + (ȳ Ȳ )/Ȳ ( x X)/ X} = var(ȳ Ȳ )/Ȳ 2 + var( x X)/ X 2 2cov{(ȳ Ȳ ), ( x X)}/(Ȳ X). Hence we have the variance estimator var{log(r)} cȳȳ + c x x 2cȳ x. Thus the log ratio based interval is I LR = (r exp{ t α/2,n 1 cȳȳ + c x x 2cȳ x }, r exp{t α/2,n 1 cȳȳ + c x x 2cȳ x }). One motivation for this interval is that positive variables are often skewed to the right and the distribution of the sample ratio, r, is also often skewed to the right. The interval I C is by definition symmetric around r while the interval I LR allows for asymmetric behavior. 2

For small samples, another better known suggested improvement on I C is the so called Fieller interval, suggested initially by Fieller (1932). This interval has often been used in applications, for example in Heitjan (2000), Hirschberg and Lye (2007), Jiang, Wu, and Williams (2000), and Beyene and Moineddin (2005). To explore the behavior of this interval, it is natural to assume that the framework for sampling is simple random sampling from an infinite population. This is mainly for simplicity and clarity in our analytical discussions. This Fieller approach, however, is often used in the finite population setting where data are typically not assumed to come from a specific distribution. The Fieller interval assumes that the joint distribution of the infinite population, F (x, y), is bivariate normal. In this case, the pivotal statistic T = n 1/2 (ȳ R x)/(s 2 y 2Rs xy + R 2 s 2 x) 1/2 has an exact t-distribution. This is easily seen by defining the variables z i = y i Rx i, i S. Then the mean of the z variables is 0, and T = n 1/2 ( z)/s z has a t-distribution with n 1 degrees of freedom, where s z is the sample standard deviation of z i. A confidence interval for the ratio is given by the set of all R such that t α/2,n 1 < T < t 1 α/2,n 1. (1) Using this we can form the equation: t 2 α/2,n 1 > T 2. This expression is easily written as a quadratic inequality in R: ar 2 + br + c < 0 (2) with a = 1 t 2 α/2,n 1 c x x, b = 2r(1 t 2 α/2,n 1 c xȳ), c = r 2 (1 t 2 α/2,n 1 c ȳȳ). Suppose there are two real valued solutions, d 1 d 2, to the equation in (2) obtained by changing the inequality into an equality. If the coefficient a is such that a > 0 then the solution to the inequality (1) for R is (d 1, d 2 ). On the other hand, if a < 0, the solution in R is (, d 1 ) (d 2, ). Using the quadratic formula, and solving for R we find the two roots, d 1 and d 2 as functions of our sample data. This gives the endpoints of the Fieller interval. The endpoints of the 3

interval are given by I F = r [ (1 t 2 α/2,n 1 c xȳ) ± t α/2,n 1 {(c x x + cȳȳ 2cȳ x ) t 2 α/2,n 1 (c x xcȳȳ c 2 ȳ x)} 1/2 ]. 1 t 2 α/2,n 1 c x x In a small sample (n = 8) of observations, Efron and Tibshirani (1993) seek to estimate the population ratio in a study on bioequivalence. The observations are roughly compatible with normality and the Fieller interval is called exact and the gold standard in this case. This seems appropriate and several nonparametric intervals based on resampling are shown to be close to the Fieller interval with the better intervals closer to the Fieller interval. In the following section we show that qualitatively different behavior of the Fieller interval, I F, can occur. 3 Data Example In contrast to the Efron and Tibshirani example from Section 2, we show there can be quite different behavior of the Fieller interval. Consider the n = 8 observations from Lehtonen and Pahkinen (2004), p.103 of two variables (ue91, hou85), the number of unemployed people and the number of households in several provinces of central Finland. The data are given in Table 1. The goal is to estimate the true ratio of the ue91 to the hou85 variables. Although the sampling fraction is n/n = 8/32 in this case for simplicity we assume in our illustration that fpc = 1. Using the formula for the Fieller interval we find the two solutions to the quadratic equation to be 0.1379 and 0.1510. The implied interval seems to be reasonably precise. Note, however, that the sample ratio, r = 0.1603, and that this value is not included in the interval. Closer inspection shows that the right endpoint from the formula is given by 0.1379 and the left endpoint from the formula is given by 0.1510. This is a curious situation. If we seek to find the interval from (1), we find that the interval is given by (, 0.1379) (0.1510, ), the complement of the interval (0.1379, 0.1510). This suggests that we cannot simply switch the endpoints to obtain an appropriate interval. Further, we see that the correct Fieller interval 4

ue91 hou85 4123 26881 760 4896 721 3730 142 556 187 1463 331 1946 127 834 219 932 Table 1: A simple random sample without replacement of size n = 8 from the Province 91 population data as presented in Lehtonen and Pahkinen (2004), p.103. Presented are the two study variables: the number of unemployed persons (ue91) and the number of households (hou85). gives a very uninformative and completely nonsensical interval (which is actually the union of two intervals). Inspection of the formula shows that we have completely nonsensical intervals when either (c x x + cȳȳ 2cȳ x ) t 2 α/2,n 1 (c x xcȳȳ c 2 ȳ x) < 0 and/or 1 t 2 α/2,n 1 c x x < 0. The second situation occurs when the sample variance of the x s is large relative to the square of the sample mean. Note, also, that this is the case where the usual confidence interval for µ x contains zero. In this data set we find that c x x = 0.37 and this in turn leads to a value of 1 t 2 α/2,n 1 c x x = 1.088. Note that in stark contrast to the strange behavior of the Fieller interval, the usual large sample interval is given by I C = (0.1478, 0.1728) and the log ratio interval is given by I LR = (0.1483, 0.1733). Both of these intervals give informative inferences on the population ratio. 5

4 Simulation We now investigate how the three intervals, Fieller, log ratio and large sample perform in a simulation experiment. We are particularly interested in the small sample behavior of Fieller s interval, where it has been assumed to be particularly suited. The model is: X F X, Y = h(x)β + ɛ, where ɛ F ɛ. We consider two choices for h( ), namely, h(x) = x and h(x) = x 2. We set F X and F ɛ to be one of the three distributions (1) Normal(2,1), (2) Gamma(shape=3) and (3) Gamma(shape=1). Also we set β = 0, 0.5 and 1. For each of these cases, we consider sample sizes n = 5, n = 10 and n = 30. We calculate all three intervals and evaluate the coverage and mean interval length based on 10,000 data sets for each setting. To give some idea of what might be expected, we give the details of one particular simulation experiment. In this case, we set n = 5, and set F X and F ɛ to be the Normal(2,1) distribution with β = 0, so that in this case the t-distribution for T is exact. In one simulation we find that the right endpoint is less than the left in 921 cases. Further in 17 cases we find that (c x x + cȳȳ 2cȳ x ) t 2 α/2,n 1 (c x xcȳȳ c 2 ȳ x) < 0 leading to a negative square root. In these cases the interval is undefined. Using the interval (d 1, d 2 ) in all cases we find the coverage to be.857. However, including the 921 intervals in which this interval does not include the true ratio but (, d 1 ) (d 2, ) does contain R we find a coverage of.949 which is within simulation error of the nominal.950 (as it must be as the t-distribution is exact). It is not clear what to do in the 17 cases with a negative square root. In these cases, the left hand side of (2) is either > 0 or < 0 for all real values of R, which means that either there is no solution for R in (2) or the solution for R is the whole real line. By convention, we say there is no solution and in this case the interval fails to capture the true ratio. However, the practical impact of this situation is small in this case. The difficulty is in a significant minority the correct intervals necessary to get the nominal coverage are the absurd disjoint 6

intervals of the form (, d 1 ) (d 2, ). These intervals are nonsensical and would likely be interpreted inadvertently to actually be (d 1, d 2 ) which is not the correct interval. Note, in particular, that in the data example in Section 3 the interval obtained by switching the endpoints is not comparable to either the large sample interval or the log interval. A natural question is how likely it is to obtain the nonsensical interval. Recall that this occurs when the coefficient a = 1 t 2 α/2,n c x x is negative. In the case of bivariate normality where both X and Y follow Normal(2,1) distributions, we can explicitly give the probability of a < 0. Note that P [a < 0] = P [t 2 α/2,n 1c x x > 1] = P [n x 2 /s 2 x t 2 α/2,n 1]. Further, n x 2 /σx 2 has a noncentral χ 2 distribution with 1 degree of freedom and noncentrality parameter nµ 2 x/σx. 2 Now, (n 1)s 2 x/σx 2 has a (central) χ 2 distribution with n 1 degrees of freedom. Thus, n x 2 /s 2 x has a noncentral F -distribution with numerator degrees of freedom 1, denominator degrees of freedom n 1, and noncentrality parameter nµ 2 x/σx. 2 To evaluate this probability in our simulation experiment, with n = 5, µ x = 2 and σx 2 = 1. Thus we have nµ 2 x/σx 2 = 20 and we find for α = 0.05 P [F (1, 4, 20) < 7.71] = 0.0911, where F (a, b, c) denotes a random variable with the F -distribution with a and b numerator and denominator degrees of freedom and noncentrality parameter c. This value is close to our empirical result in the simulation experiment, where we observed 921/10000 =.0921 negative denominators in the Fieller interval. For other sample sizes, Figure 1 displays the chance of a negative denominator in the Fieller interval when sampling from a bivariate normal population where both X and Y follow Normal(2,1) distributions. We see from the left plot in Figure 1 that once the sample size is larger than n = 10 the chance is quite small. The sample size needed to be safe from the sign switch is in fact dependent on the inverse of the coefficient of variation (µ x /s x ), as nµ 2 x/σx 2 is the noncentrality of the F distribution. In the left plot in Figure 1 cv = 0.5 while in the right plot both X and Y follow Normal(1,1) 7

% of times a < 0 0 2 4 6 8 10 % of times a < 0 0 20 40 60 80 100 5 6 7 8 9 10 11 12 Sample size (n) 5 7 9 11 13 15 17 19 21 23 25 27 29 Sample size (n) Figure 1: Results from simulation study. Displayed are the chance of a negative denominator in the Fieller interval when sampling from a bivariate normal population with coefficient of variation 0.5 (left) and 1.0 (right), respectively. distributions so cv = 1.0. The results are much more severe in the right plot. For larger values of cv the Fieller interval performs badly even for larger sample sizes. We note that a potential problem with bigger cv is that the log ratio method can also become non-applicable for large cv s as x itself may become negative. Table 2 summarizes our simulation results. In all cases when two distinct roots d 1 and d 2 were found in the construction of the Fieller interval the interval was taken to be (d 1, d 2 ). We see that the Fieller interval, often motivated for good small sample behavior, behaves very poorly for n = 5 and moderately badly for n = 10. The reason for the undercoverage of Fieller s interval under bivariate normality is as discussed above. Fieller intervals tend to be the widest but in many cases still do not have coverage closest to nominal of the three intervals. For the larger sample size of n = 30 all three intervals perform well across all situations considered. We see that under gamma distributions the comparisons of the three intervals are qualitatively similar. 8

5 Conclusion We have studied the Fieller interval for a population ratio pointing out that care must be used when using the Fieller interval. We can obtain nonsensical answers from the formula. We see that this is not a rare occurrence. For bivariate normal observations Fieller s formula gives nonsensical results in approximately 10 percent of data sets when n = 5 and cv = 0.5. This sample size is small, but such sample sizes are common in biological applications. The larger the coefficient of variation the larger the proportion of nonsensical and uninformative intervals we obtain using Fieller s interval. The large sample approximations leading to the Cochran and log ratio intervals perform more stably in general for small samples. Although the log ratio method requires positive means to be usable, it appears to perform generally better when applicable than Cochran s large sample method, especially when the sample size is small. The bootstrap is a common alternative to large sample methods for small to moderate sample sizes. The analysis of a ratio, however, is a particularly difficult problem for the bootstrap and great care is necessary to draw reasonable inferences. See, e.g., Chapter 25 of Efron and Tibshirani (1993) where bias correction and calibration are necessary to make the intervals perform adequately. References Beyene, J. and Moineddin, R. (2005), Methods for Confidence Interval Estimation of a Ratio Parameter with Application to Location Quotients, BMC Medical Research Methodology. 5, 32. Cochran, W. G. (1977), Sampling Techniques, New York: Wiley. Efron, B. and Tibshirani, R.J. (1993), An Introduction to the Bootstrap, New York: Chapman Hall. Fieller, E.C. (1932), The Distribution of the Index in a Bivariate Normal Distribution, 9

Biometrika, 24, 428-440. Heitjan, D.F. (2000), Fieller s Method and Net Health Benefits, Health Economics, 9, 327-335. Hirschberg, J.G. and Lye, J.N. (2007), Providing Intuition to the Fieller Method with Two Geometric Representations Using STATA and EVIEWS, preprint. Jiang, G., Wu, J., and Williams, G.R. (2000), Fieller s Interval and the Bootstrap-Fieller Interval for the Incremental Cost-Effectiveness Ratio, Health Services and Outcomes Research Methodology, 1, 291-303. Lehtonen, R. and Pahkinen, E. (2004), Practical Methods for Design and Analysis of Complex Surveys, 2nd Edition, New York: Wiley. Lohr, S.L. (2009), Sampling: Design and Analysis, 2nd Edition, Pacific Grove: Brooks/Cole. 10

n = 5 n = 10 n = 30 β = 0 β = 0.5 β = 1 β = 0 β = 0.5 β = 1 β = 0 β = 0.5 β = 1 case: X Normal(2, 1), Y = Xβ + Normal(2, 1) Cochran 0.950 1.603 0.951 1.605 0.951 1.615 0.951 0.966 0.951 0.965 0.951 0.968 0.951 0.504 0.951 0.507 0.951 0.506 Fieller 0.868 2.193 0.868 2.201 0.868 2.213 0.950 1.067 0.950 1.065 0.950 1.069 0.950 0.517 0.950 0.519 0.950 0.519 log ratio 0.963 1.822 0.959 1.692 0.956 1.661 0.960 1.008 0.956 0.982 0.955 0.978 0.954 0.510 0.953 0.509 0.952 0.508 case: X Normal(2, 1), Y = X 2 β + Normal(2, 1) Cochran 0.951 1.606 0.957 1.360 0.939 1.779 0.951 0.966 0.956 0.825 0.936 1.111 0.951 0.508 0.953 0.440 0.941 0.617 Fieller 0.868 2.209 0.875 1.757 0.869 2.311 0.950 1.066 0.949 0.888 0.931 1.203 0.950 0.520 0.949 0.448 0.938 0.629 log ratio 0.964 1.828 0.963 1.382 0.947 1.800 0.960 1.009 0.961 0.830 0.942 1.116 0.954 0.513 0.955 0.441 0.943 0.618 case: X Normal(2, 1), Y = Xβ + Gamma(shape = 3) Cochran 0.929 2.499 0.929 2.494 0.929 2.490 0.935 1.531 0.935 1.529 0.935 1.528 0.945 0.814 0.945 0.813 0.945 0.816 Fieller 0.864 3.469 0.865 3.462 0.865 3.450 0.947 1.692 0.947 1.689 0.947 1.686 0.949 0.834 0.949 0.832 0.949 0.836 log ratio 0.950 2.850 0.943 2.675 0.939 2.602 0.949 1.605 0.944 1.567 0.942 1.553 0.949 0.824 0.948 0.818 0.946 0.819 case: X Normal(2, 1), Y = X 2 β + Gamma(shape = 3) Cochran 0.929 2.491 0.928 2.131 0.933 2.322 0.935 1.524 0.933 1.319 0.937 1.438 0.945 0.813 0.944 0.711 0.944 0.781 Fieller 0.864 3.461 0.851 2.815 0.855 3.022 0.948 1.681 0.935 1.429 0.931 1.554 0.949 0.833 0.944 0.725 0.942 0.795 log ratio 0.949 2.836 0.940 2.188 0.943 2.357 0.949 1.597 0.942 1.331 0.943 1.446 0.949 0.823 0.946 0.713 0.947 0.782 case: X Normal(2, 1), Y = Xβ + Gamma(shape = 1) Cochran 0.888 1.093 0.888 1.095 0.889 1.094 0.907 0.701 0.907 0.698 0.906 0.699 0.931 0.388 0.931 0.388 0.931 0.388 Fieller 0.851 1.491 0.853 1.490 0.852 1.488 0.929 0.768 0.929 0.768 0.929 0.765 0.939 0.397 0.939 0.397 0.939 0.397 log ratio 0.937 1.399 0.904 1.157 0.899 1.119 0.937 0.769 0.918 0.713 0.914 0.705 0.942 0.399 0.935 0.391 0.934 0.389 case: X Normal(2, 1), Y = X 2 β + Gamma(shape = 1) Cochran 0.888 1.092 0.920 1.258 0.915 1.939 0.906 0.702 0.926 0.792 0.920 1.233 0.931 0.388 0.939 0.434 0.935 0.684 Fieller 0.853 1.493 0.846 1.669 0.837 2.592 0.929 0.771 0.917 0.859 0.905 1.349 0.939 0.397 0.935 0.442 0.929 0.699 log ratio 0.937 1.401 0.935 1.288 0.929 1.977 0.936 0.770 0.935 0.800 0.928 1.242 0.942 0.399 0.942 0.435 0.937 0.686 Table 2: Results from the simulation study. For each setting (β = 0, 0.5 and 1) the coverage probability (first column) and interval length (second column) is reported. The nominal coverage is 95%. 11

n = 5 n = 10 n = 30 β = 0 β = 0.5 β = 1 β = 0 β = 0.5 β = 1 β = 0 β = 0.5 β = 1 case: X Gamma(shape = 3), Y = Xβ + Normal(2, 1) Cochran 0.946 1.147 0.945 1.147 0.946 1.147 0.947 0.692 0.946 0.693 0.946 0.691 0.948 0.363 0.948 0.364 0.949 0.363 Fieller 0.878 1.807 0.877 1.810 0.877 1.813 0.948 0.792 0.947 0.795 0.948 0.791 0.949 0.375 0.949 0.376 0.949 0.376 log ratio 0.950 1.321 0.946 1.197 0.946 1.170 0.949 0.727 0.947 0.704 0.946 0.696 0.949 0.368 0.949 0.365 0.948 0.364 case: X Gamma(shape = 3), Y = X 2 β + Normal(2, 1) Cochran 0.945 1.148 0.870 1.452 0.780 2.847 0.946 0.691 0.850 0.985 0.809 2.118 0.948 0.365 0.881 0.643 0.871 1.417 Fieller 0.878 1.808 0.800 2.116 0.657 4.445 0.948 0.791 0.836 1.100 0.767 2.421 0.949 0.378 0.868 0.663 0.850 1.468 log ratio 0.950 1.324 0.893 1.472 0.809 2.902 0.949 0.725 0.866 0.992 0.824 2.140 0.949 0.369 0.888 0.645 0.879 1.424 case: X Gamma(shape = 3), Y = Xβ + Gamma(shape = 3) Cochran 0.929 1.757 0.929 1.755 0.929 1.748 0.935 1.085 0.935 1.082 0.935 1.082 0.944 0.577 0.944 0.578 0.944 0.578 Fieller 0.885 2.799 0.885 2.785 0.886 2.781 0.953 1.240 0.953 1.238 0.953 1.238 0.950 0.597 0.950 0.598 0.951 0.598 log ratio 0.942 2.025 0.936 1.862 0.933 1.807 0.944 1.142 0.940 1.106 0.938 1.096 0.947 0.585 0.945 0.582 0.945 0.581 case: X Gamma(shape = 3), Y = X 2 β + Gamma(shape = 3) Cochran 0.929 1.754 0.900 1.791 0.832 2.920 0.935 1.083 0.897 1.148 0.833 2.054 0.944 0.578 0.911 0.671 0.877 1.353 Fieller 0.884 2.798 0.848 2.574 0.742 4.361 0.952 1.240 0.898 1.268 0.803 2.316 0.950 0.597 0.908 0.688 0.860 1.398 log ratio 0.943 2.028 0.914 1.820 0.858 2.971 0.944 1.140 0.907 1.156 0.847 2.070 0.947 0.586 0.915 0.672 0.883 1.357 case: X Gamma(shape = 3), Y = Xβ + Gamma(shape = 1) Cochran 0.889 0.746 0.887 0.745 0.889 0.750 0.907 0.481 0.908 0.482 0.907 0.483 0.932 0.267 0.932 0.267 0.932 0.267 Fieller 0.876 1.161 0.875 1.161 0.876 1.165 0.939 0.544 0.939 0.546 0.939 0.544 0.944 0.275 0.943 0.275 0.944 0.275 log ratio 0.932 0.966 0.898 0.771 0.895 0.760 0.935 0.531 0.914 0.489 0.912 0.485 0.942 0.274 0.934 0.269 0.934 0.267 case: X Gamma(shape = 3), Y = X 2 β + Gamma(shape = 1) Cochran 0.889 0.742 0.838 1.663 0.802 3.213 0.908 0.480 0.844 1.153 0.824 2.348 0.932 0.267 0.886 0.741 0.878 1.521 Fieller 0.877 1.160 0.727 2.536 0.659 5.132 0.939 0.542 0.810 1.306 0.775 2.698 0.944 0.274 0.868 0.766 0.855 1.578 log ratio 0.932 0.961 0.868 1.707 0.835 3.308 0.935 0.528 0.863 1.167 0.843 2.381 0.943 0.274 0.894 0.745 0.887 1.528 Table 2: continued. 12

n = 5 n = 10 n = 30 β = 0 β = 0.5 β = 1 β = 0 β = 0.5 β = 1 β = 0 β = 0.5 β = 1 case: X Gamma(shape = 1), Y = Xβ + Normal(2, 1) Cochran 0.940 4.956 0.939 4.965 0.939 4.999 0.940 2.984 0.940 2.981 0.940 2.974 0.944 1.578 0.944 1.581 0.945 1.576 Fieller 0.417 9.784 0.417 9.770 0.416 9.774 0.897 4.995 0.894 4.981 0.896 4.958 0.939 1.777 0.939 1.779 0.939 1.773 log ratio 0.936 6.456 0.934 5.847 0.934 5.603 0.937 3.277 0.935 3.162 0.935 3.098 0.942 1.620 0.942 1.608 0.942 1.594 case: X Gamma(shape = 1), Y = X 2 β + Normal(2, 1) Cochran 0.940 4.963 0.938 4.228 0.927 4.009 0.940 2.983 0.940 2.432 0.923 2.388 0.945 1.573 0.948 1.242 0.926 1.289 Fieller 0.417 9.747 0.458 8.715 0.497 8.449 0.897 4.980 0.923 3.899 0.928 3.553 0.940 1.774 0.950 1.373 0.945 1.396 log ratio 0.937 6.467 0.943 4.629 0.938 4.199 0.937 3.273 0.947 2.500 0.933 2.426 0.943 1.615 0.950 1.251 0.931 1.295 case: X Gamma(shape = 1), Y = Xβ + Gamma(shape = 3) Cochran 0.929 7.470 0.929 7.532 0.929 7.491 0.934 4.579 0.934 4.553 0.934 4.581 0.943 2.435 0.943 2.436 0.942 2.451 Fieller 0.424 14.385 0.425 14.346 0.425 14.350 0.905 7.618 0.906 7.567 0.906 7.626 0.944 2.739 0.943 2.745 0.944 2.755 log ratio 0.932 9.725 0.929 9.123 0.927 8.667 0.934 5.060 0.932 4.891 0.931 4.838 0.942 2.505 0.941 2.488 0.941 2.490 case: X Gamma(shape = 1), Y = X 2 β + Gamma(shape = 3) Cochran 0.928 7.481 0.921 6.705 0.920 6.131 0.934 4.564 0.928 3.963 0.927 3.658 0.943 2.432 0.943 2.044 0.939 1.926 Fieller 0.424 14.317 0.450 13.212 0.480 12.775 0.906 7.618 0.921 6.436 0.927 5.711 0.943 2.729 0.951 2.270 0.949 2.104 log ratio 0.932 9.754 0.926 7.572 0.930 6.580 0.935 5.041 0.934 4.132 0.936 3.744 0.942 2.501 0.946 2.067 0.943 1.938 case: X Gamma(shape = 1), Y = Xβ + Gamma(shape = 1) Cochran 0.891 2.822 0.892 2.827 0.892 2.825 0.911 1.800 0.911 1.805 0.910 1.790 0.933 0.991 0.934 0.983 0.934 0.985 Fieller 0.471 5.084 0.471 5.088 0.469 5.097 0.925 2.896 0.924 2.879 0.925 2.873 0.951 1.101 0.951 1.089 0.952 1.093 log ratio 0.924 4.104 0.904 3.299 0.898 3.081 0.931 2.082 0.919 1.918 0.915 1.855 0.941 1.033 0.936 1.000 0.935 0.995 case: X Gamma(shape = 1), Y = X 2 β + Gamma(shape = 1) Cochran 0.892 2.842 0.882 2.502 0.848 2.966 0.910 1.801 0.897 1.604 0.855 1.965 0.933 0.989 0.923 0.899 0.879 1.182 Fieller 0.471 5.080 0.531 5.162 0.515 5.714 0.925 2.890 0.907 2.345 0.863 2.766 0.951 1.097 0.935 0.969 0.889 1.271 log ratio 0.923 4.145 0.906 2.704 0.878 3.125 0.931 2.083 0.914 1.654 0.874 2.006 0.941 1.032 0.930 0.907 0.890 1.190 Table 2: continued. 13