Resampling Statistics

Size: px
Start display at page:

Download "Resampling Statistics"

Transcription

1 Resampling Statistics Rationale Much of modern statistics is anchored in the use of statistics and hypothesis tests that only have desirable and well-known properties when computed from populations that are normally distributed. While it is claimed that many such statistics and hypothesis tests are generally robust with respect to non-normality, other approaches that require an empirical investigation of the underlying population distribution or of the distribution of the statistic are possible and in some instances preferable. In instances when the distribution of a statistic, conceivably a very complicated statistic, is unknown, no recourse to a normal theory approach is available and alternative approaches are required. General Overview Resampling statistics refers to the use of the observed data or of a data generating mechanism (such as a die) to produce new hypothetical samples (resamples) that mimic the underlying population, the results of which can then be analyzed. With numerous cross-disciplinary applications especially in the sub-disciplines of the life science, resampling methods are widely used since they are options when parametric approaches are difficult to employ or otherwise do not apply. Resampled data is derived using a manual mechanism to simulate many pseudotrials. These approaches were difficult to utilize prior to 980s since these methods require many repetitions. With the incorporation of computers, the trials can be simulated in a few minutes and is why these methods have become widely used. The methods that will be discussed are used to make many statistical inferences about the underlying population. The most practical use of resampling methods is to derive confidence intervals and test hypotheses. This is accomplished by drawing simulated samples from the data themselves (resamples) or from a reference distribution based on the data; afterwards, you are able to observe how the statistic of interest in these resamples behaves. Resampling approaches can be used to substitute for traditional statistical (formulaic) approaches or when a traditional approach is difficult to apply. These methods are widely used because their ease of use. They generally require minimal mathematical formulas, needing a small amount of mathematical (algebraic) knowledge. These methods are easy to understand and stray away from choosing an incorrect formula in your diagnostics. Ecology and Evolution Publicized Applications This is a list of applications of resampling methods concatenated by Phillip Crowley, 99. Analysis of Null models, competition, and community structure Detecting Density Dependence Characterizing Spatial Patterns and Processes Estimating Population Size and Vital Rates Environmental Modeling Evolutionary Processes and Rates

2 Phylogeny Analysis In order of use, Crowley found this relationship when searching for publications: Monte Carlo > Bootstrap > Permutation & Jackknife Overall, resampling methods were increasing in significant use over the prior decade. Types of Resampling Methods I. Monte Carlo Simulation This is a method that derives data from a mechanism (such as a proportion) that models the process you wish to understand (the population). This produces new samples of simulated data, which can be examined as possible results. After doing many repetitions, Monte Carlo tests produce exact p-values that can be interpreted as an error rate; letting the number of repeats sharpens the critical region. II. III. IV. Randomization (Permutation) Test this is a type of statistical significance test, in which a reference distribution is obtained by calculating all possible values of the test statistic under rearrangements of the labels on the observed data points. Like other the Bootstrap and the Monte Carlo approach, permutation methods for significance testing also produce exact p-values. These tests are the oldest, simplest, and most common form of resampling tests and are suitable whenever the null hypothesis makes all permutations of the observed data equally likely. In this method, data is reassigned randomly without replacement. They are usually based off the Student t and Fisher s F test. Most non-parametric tests are based on permutations of rank orderings of the data. This method has become practical because of computers; without them, it may be impossible to derive all the possible permutations. This method should be employed when you are dealing with an unknown distribution. Bootstrapping This approach is based on the fact that all we know about the underlying population is what we derived in our samples. Becoming the most widely used resampling method, it estimates the sampling distribution of an estimator by sampling with replacement from the original estimate, most often with the purpose of deriving robust estimates of standard errors and confidence intervals of a population parameter. Like all Monte Carlo based methods, this approach can be used to define confidence Intervals and in hypothesis testing. This method is beneficial to side step problems with nonnormality or if the distribution parameters are unknown. This method can be used to calculate an appropriate sample size for experimental design. Jackknife This method is used in statistical inference to estimate the bias and standard error in a statistic, when a random sample of observations is used to calculate it. This method provides a systematic method of resampling

3 with a mild amount of calculations. It offers improved estimate of the sample parameter to create less sampling bias. The basic idea behind the jackknife estimator lies in systematically re-computing the statistic estimate leaving out one observation at a time from the sample set. From this new improved" sample statistic can be used to estimate the bias can be variance of the statistic. The Flow of Information for resampling methods In these methods, it is necessary to specify the universe to sample from (random numbers, an observed data set, true or false, etc.), specify the sampling procedure (number of samples, sizes of samples, sampling with or without replacement), and specify the statistic you wish to keep track of. The flow of information is as follows:. Input data. Resample from the inputted data 3. Calculate the statistic desired 4. Record statistic 5. Return to sample for (X) number of resamples; once reached to completed (X) times, continue to step 6 6. Calculate p-value by counting number of resamples that occur in desired extreme domains divided by the total number of resamples 7. Present/Print results Resampling Computer Programs To effectively use these methods, you should have a good program and a fast computer to handle the repetitions. Phillip L. Good has suggested the following programs, with the first four being recommended: I. R a programming language that is easy to manipulate. This program is free and scripts are precompiled throughout the Internet. However, be aware, you are on your own. II. III. IV. C ++ like R, this is a programming language that has great potential for those entering statistics with a great programming background. Also like R, this is do-it-yourself so you are once again on your own. Resampling Stats easy to use, this programming language seems very similar to BASIC programming language. It has all the resampling method functions already incorporated and is also available as a Microsoft Excel addin. It is cheap and easy to follow but can eventually become limited for intense practice of these methods. S Plus R based, this program has many built-in functions and pull-down menus, which make it easy to use. The program s designers offer much support; this package comes at an expensive price. 3

4 V. SAS commonly used in statistical analysis, this package is C based. Pricey and time consuming to debug. Monte Carlo Simulation Similar to what was outlined above, the general procedure of Monte Carlo Simulations is as follows: A. Make a simulated sample population utilizing a non-biased randomizing mechanism (cards, dice, or a computer program) which is based on the population whose behavior you wish to investigate. B. Create a pseudo-sample to simulate a real-life sample of interest. C. Repeat step B, (X) number of times. D. Calculate the probability of interest from the tabulation of outcomes of the resampling trials. Monte Carlo Example (derived from a Simon 997 example) On any given day, it is likely to be sunny 70% of the time. On a sunny day, the Redskins win 65% of their games. What is the likelihood of winning a game on a sunny day? This is a very simplistic example easily described by calculating the joint probability of the two, but serves as a good example of the BASIC programming involved in resampling. The outline:.put seven blue balls and three yellow balls into an urn labeled A (the Nice Day Urn). Put 65 green balls and 35 red balls into an urn labeled B (the Win/Lose Conditional to Nice day Urn)..Draw one ball from urn A. If it is blue, continue; otherwise record no and stop. 3. If you have drawn a blue ball from run A, now draw a ball from urn B, and if it is green, record yes ; otherwise write no. 4.Repeat steps (or more) times. 5.Count the number of trials. 6.Compute the proportion of yeses in the 000 samples. The Resampling Statistics programming looks like this: URN 7# 3#0 weather Create 0 days samples, 7 nice day URN 35#0 65# winlose Create 00 record sample, 55 wins REPEAT 000 Repeat following 000 SAMPLE weather a Sample of the days IF a = If a good day, continue; otherwise to skip if statement SAMPLE winlose b Sample winlose records IF b = If win, continue; otherwise skip SCORE b z Tally Good Day Wins END End if END End if 4

5 END Go back to repeat; after 000, End repeat COUNT z = k Count Good Day wins DIVIDE k 000 kk Then divide it by 000 PRINT kk Show on screen the result Using Resampling Stats plugin for Microsoft Excel, this is what the print out might look like:.454 = wins on a sunny day. The actual joint probability approach gives a probability of.455. Resampling Approaches in Estimation and Hypothesis Testing I. Hypothesis Testing A. Normal Theory Approach For illustration consider Student's t - test for differences in means when variances are unknown, but are considered to be equal. The hypothesis of interest is that H 0 : μ = μ. While several possible alternative hypotheses could be specified, for our purposes H A : μ < μ. Given two samples drawn from populations and, assuming that these are normally distributed populations with equal variances, and that the samples were drawn independently and at random from each population, then a statistic whose distribution is known can be elaborated to test H 0 : x s n x s + n, () where x, x, s, s, n, n are the respective sample means, variances and sample sizes. When the conditions stated above are strictly met and H 0 is true, () is distributed as Student's t with (n + n - ) degrees of freedom. As ( n + n ), t N(0,). 5

6 The reasons for making the assumptions specified above is to allow the investigator to make some statement about the likelihood of the computed t - value, and to make a decision as to whether to profess belief in either H 0 or H a. The percentiles of the t distribution with the computed degrees of freedom can be interpreted as the conditional probability of observing the computed t value or one larger (or smaller) given that H 0 is true: P( t T H 0 ) = α.. Therefore, we would probably wish to profess belief in H 0 for suitably large values of α and disbelief for suitably small values. Embedded in this probability we must also include the distributional assumptions mentioned above P( t T H, N( μ, σ ), ( μ, σ ), σ = σ ) = α. 0 N If a specific alternative hypothesis had been stated, for example H A : μ = μ - 5, then under the assumption of normality and equal variances, the t - statistic could be recomputed given the new estimate of μ under the alternative hypothesis μ = + 5 x. The conditional probability of obtaining the observed difference in t values as computed under the null and alternative hypotheses (t 0 - t a ), given specified α, the observed variances, and the particular alternative hypothesis could be computed: P( t 0 t A δ T α, σ, σ, H A ) = β. This probability is also conditioned on the assumption that both populations are normally distributed ( N( μ, σ ), N( μ, σ )). Recall that of these two conditional probabilities α is the Type I error rate, the probability of rejecting the tested null hypothesis when true, and that β is the Type II error rate, the probability of failing to reject the tested null hypothesis when false. I present this review to emphasize that the estimation of each of these probabilities, which are interpreted as error rates in the process of making a decision about nature, in the course of interpreting a specific statistical test, is totally contingent on assuming specific forms for the distribution of the underlying populations. To know these error rates exactly requires that all the conditions of these tests be met. B. A distribution-free approach One way to avoid these distributional assumptions has been the approach now called non - parametric, rank - order, rank - like, and distribution - free statistics. A series of 6

7 tests many of which apply in situations analogous to normal theory statistics have been elaborated (see ref,, 3 for expanded treatments of these procedures). The key to the function of these statistics is that they are based on the ranks of the actual observations in a joint ranking and not on the observations themselves. For example the Wilcoxon distribution - free rank sum test can be applied in place of the sample or separate groups Student's t test. The observations from both samples combined are ranked from least to greatest and the sum of the ranks assigned to the observations from either sample is computed. If both samples are comprised of observations that are of similar magnitude, then the ranks assigned to sample should be similar to the ranks assigned to sample. For fixed sample sizes a fixed number of ranks are possible, for n = 5 and n = 7, ranks will be assigned. Under the null hypothesis that the location of each population on the number line is identical, the sum of the ranks assigned to either sample should equal the sum obtained from randomly assigning ranks to the observations in each sample. In this example there are ways 5 of assigning ranks to sample and C7 or way to assign ranks to sample after assigning the ranks to sample. The total number of possible arrangements of the C 5 ranks is then and it is possible for each arrangement to compute the sum of the ranks. From this we can enumerate the distribution of the sum of the ranks, W. W in this example can range from 5 to 50. If we divide the number of arrangements with w = W C 5 then we have the probability of observing a particular value of w equal to P ( w = W H 0 ). To obtain the conditional probability that w > W given H 0, we simply tabulate the cumulative probabilities. The only additional assumption embedded in this approach is that the observations are independent, but this is also an assumption of the normal theory approach. Notice we make no assumption about the forms of the underlying populations about which we wish to make inferences, and that the exact distribution of the test statistic, W, is known because it is enumerated. These distribution - free statistics are usually criticized for being less "efficient" than the analogous test based on assuming the populations to be normally distributed. It is true that when the underlying populations are normally distributed then the asymptotic relative efficiencies (ratio of sample sizes of one test to another necessary to have equal power relative to a broad class of alternative hypotheses for fixed α) of distribution - free tests are generally lower than their normal theory analogs, but usually not markedly so. In instances where the underlying populations are non-normal then the distribution - free tests can be infinitely more efficient that their normal theory counterparts. In general, this means that distribution- C 5 7

8 free tests will have higher Type II error rates (β) than normal theory tests when the normal theory assumptions are met. Type I error rates will not be affected. However, if the underlying populations are not normally distributed then normal theory tests can lead to under estimation of both Type I and Type II error rate. C. Randomization So far we have used two approaches to estimating error rates in hypothesis testing that either require the assumption of a particular form of the distribution of the underlying population, or that require the investigator to be able to enumerate the distribution of the test statistic when the null hypothesis is true and under specific alternative hypotheses. What can be done when we neither wish to assume normality nor can we enumerate the distribution of the test statistic? Recall the analogy I used when describing how to generate the expected sum of ranks assigned to a particular sample under the null hypothesis of identical population locations on a number line. The analogy was to a process of randomly assigning ranks to observations independent of one's knowledge of which sample an observation is a member. A randomization test makes use of such a procedure, but does so by operating on the observations rather than the joint ranking of the observations. For this reason, the distribution of an analogous statistic (the sum of the observations in one sample) cannot be easily tabulated, although it is theoretically possible to enumerate such a distribution. From one instance to the next the observations may be of substantially different magnitude so a single tabulation of the probabilities of observing a specific sum of observations could not be made, a different tabulation would be required for each application of the test. A further problem arises if the sample sizes are large. In the example mentioned previously there are only C 5 = 79 possible arrangements of values so the exact distribution of the sum of observations in one sample could conceivably have been enumerated. Had our sample sizes been 0 and 5 then over 3. million arrangements would have been possible. If you have had any experience in combinatorial enumeration then you would know that this approach has rapidly become computationally impractical. With high-speed computers it is certainly possible to tally 3. million sums, but developing an efficient algorithm to be sure that each and every arrangement has been included, and included only once is prohibitive. What then? Sample. When the universe of possible arrangements is too large to enumerate why not sample arrangements from this universe independently and at random? The distribution of the test statistic over this series of samples can then be tabulated, its' mean and variance computed, and the error rate associated with an hypothesis test estimated. Table contains samples of n = 0, and n = 5, obtained from sampling from populations with μ = 00, σ = 40 and μ = 90, σ = 40, respectively. A normal theory t - test applied to these data yields a t = 3.36, df = 3, < p < The same data examined by the distribution - free Wilcoxon's rank sum test yields W* =.7735, 8

9 0.006 < p < Applying this randomization approach with 000 iterations of sampling without replacement first 0 and then 5 observations and computing the t statistic for each of these samples yields the distribution depicted in Figure. The normal theory t distribution is depicted as a smooth curve. According to the randomization procedure the probability of observing a t value greater than or equal to that actually observed (3.36) is < p < Remember that this sampling procedure, unlike an enumeration, allows each possible arrangement of values to be sampled more than once. The probability that on any iteration a particular arrangement 5 will be chosen is in this instance C 0 or approximately 3. x 0-7. After 000 such randomizations it is quite possible that some arrangements have been sampled more than once, but there is no reason to believe that particular arrangements yielding either low or high t - values should be systematically included or excluded from the 000 randomizations. This approach is obviously an empirical approach to learning something about the distribution of a test statistic under specified conditions. This Monte Carlo sample procedure would have to be performed anew for each new set of observations. One aspect that may be an advantage of this approach over normal theory approaches is that any ad hoc test statistic can be elaborated since a direct empirical investigation of its distributional properties accompanies each test. For example we could just use the difference in the sample means x x as one test statistic. Figure shows the distribution of x x over the same 000 randomizations. The actual difference between sample means is 8.95 and under the randomization approach the probability of observing a difference this large or larger is < p < The computed probability of observing the t or x x actually observed compares favorably with the normal theory estimates. Figure 3 and 4 illustrate the same procedure applied to two populations whose underlying distributions are exponential. Table presents the sample data generated from two populations with λ = 0 and λ = 00, respectively. A normal theory test on these data yields t = , df = 3, 0.0l <p < 0.05, while the randomization approach yielded a probability of obtaining the observed value of t or one greater of 0.07 < p < 0.08, and a probability of obtaining the observed or a greater difference in means of p > Here we see the normal theory test breaking down and convergence in the results obtained from the distribution - free and randomization tests. Is all of this kosher? We can see the parallel development of the distribution-free and the randomization tests, yet is the randomization test actually yielding a meaningful result? The answer is a resounding well-maybe-er-i-don't-know. The randomization procedure essentially asks the question, given observed samples n and n, if we assume that these samples came from the same underlying population whose distribution F is given by the n + n sampled values, with probability mass /(n +n ) for each observation, what is the chance of partitioning the observations into groups of the 9

10 size observed that have means that differ by an amount as large as that observed? Is this the best empirical estimate of the distribution of a test statistic? D. The Bootstrap The Bootstrap is another empirical approach to understanding the distributional properties of a test statistic, but is also useful as a means of estimating statistics and their standard errors. The bootstrap is very similar to the randomization procedure outlined above. The observed distribution of sample values is used as an estimate of the underlying probability distribution of the population F. Then, the distribution of a statistic for fixed sample sizes is obtained by repeatedly sampling from the distribution F, with each value receiving probability mass /(n + n ), but sampling values with replacement, so that instead of individual partitions of the data having the potential to occur more than once, the individual values themselves may appear repeatedly in a single sample. Under this resampling algorithm the number of possible sample arrangements is much greater than for the randomization approach. For example with a total sample size m =, with component samples of size 7 and 5, 7 x 5 = x 0 arrangements are possible. For m = 5, and n = 0, n = 5, 5 0 x 5 5 = x 0 34 arrangements are possible, factors of 0 0 and 0 8 more arrangements, respectively, compared to the randomization approach. Any test statistic averaged across a series of say 000 samples under this algorithm will have a larger standard error since sub-samples of F can deviate from F more than under the randomization algorithm. Figures 5 and 6 illustrate the distribution of t values and x x for 000 bootstrap samples of the empirical probability distribution presented in Table. For the normal populations the bootstrap estimates the probability of the observed t or one greater to be 0.00 < p < which, surprisingly is somewhat less than the randomization approach. This comparison is reversed when examining the differences between means. The bootstrap estimates the probability of the observed mean difference or one greater as 0.0 < p < 0.03, which is an order of magnitude greater than that estimated by the randomization approach, or for that matter for the t - statistic from the same group of bootstrap samples. Figures 7 and 8 provide similar data for the samples derived from exponentially distributed populations presented in Table. The bootstrap is more conservative than either the normal theory approach or the randomization approach when examining the t value obtained for the exponential populations. This is the result I would generally expect in a comparison of the bootstrap and randomization. Which approach is best? While the randomization approach can be seen to be analogous to the enumeration of distributions that characterizes distribution - free statistics, it is unrealistic in that the distribution of a test statistic across a series of randomized samples is restricted to sub-samples that contain exactly the same observations as the true samples, once each. In some instances this may be the appropriate procedure, but in general randomization may give unrealistically small standard errors for test statistics, so that the true Type I error rates will be greater than nominally stated and Type II error rates also will be greater than nominally stated. 0

11 However, in all the examples presented above the empirical randomization and bootstrap approaches compare favorably with the normal theory approach. More examples of Randomization and Bootstrap methods (Simon, 997): Simon produced a book Resampling: the New Statistics, an example based book on Monte Carlo, Permutation (Randomization) tests, and Bootstrap available for free on the Resampling Stats website. I found the following examples demonstrate the effectiveness of these methods. Bootstrap example in creating a confidence interval: Of 35 men of age with high cholesterol, 0 developed myocardial infarction. How much confidence should we have that if we were to take a much larger sample than was actually obtained, the sample mean (0/35 =.07) would be in some close vicinity of the actual mean? The general set up may be like this:. Construct a sample containing 35 representatives balls: 0 red representing infarction and 5 green representing no infarction. Mix, choose a ball, record its color, replace it, and repeat 35 times (to simulate a sample of 35 men). 3. Record the number of red balls among the 35 balls drawn. 4. Repeat steps -4 perhaps 000 times, and observe how much the total number of reds varies from sample to sample. How this looks in the basic programming: URN 0# 5#0 men which have myocardial infarction REPEAT 000 SAMPLE 35 men a replacement COUNT a = b infarctions occurred DIVIDE b 35 c SCORE c z sample mean END HISTOGRAM z PERCENTILE z ( ) k percentiles PRINT k Create a 35 based sample, 0 of Resample 35 times with Count how many myocardial Divide by 35 to get a sample mean Keep track of every resampled Plot the sample means in a histogram Calculate the.5 th and 97.5 th Print histogram and results Using Resampling Stats plug-in for Microsoft Excel, this is what the print out might look like:

12 Randomization in hypothesis testing: The following is the price of whisky in 6 monopoly states (where the state owns the liquor store) and 6 private-owned states are as follows: 6 monopoly states: $4.65, $4.55, $4., $4.5, $4.0, $4.55, $3.80, $4.00, $4.9, $4.75, $4.74, $4.50, $4.0, $4.00, $5.05, $4.0. Mean = $ private-ownership states: $4.8, $5.9, $4.89, $4.95, $4.55, $4.90, $5.5, $5.30, $4.9, $4.85, $4.54, $4.75, $4.85, $4.85, $4.50, $4.75, $4.79, $4.85, $4.79, $4.95, $4.95, $4.75, $5.0, $5.0, $4.80, $4.9. Mean = $4.84 There is a difference in prices of whisky between monopoly and private ownership states with a mean difference of.49. How likely is it that one would choose 6 and 6 states randomly that would have a difference of $.49? To answer this question, the basic programming would look like this: NUMBERS ( Create number array for private which is observed private-ownership sample ) priv NUMBERS ( Create number array for monopoly which is observed state owned sample ) mono CONCAT priv govt all Concatenate the two samples REPEAT 000 SHUFFLE 6 all priv$ Allocate 6 of concatenated samples at random w/o replacement to the private resample

13 SHUFFLE 6 all mono$ Allocate 6 of concatenated samples at random w/o replacement to the monopoly resample MEAN priv$ p Calculate the mean of the private resample MEAN mono$ g Calculate the mean of the monopoly resample SUBTRACT p g diff Subtract the private from monopoly SCORE diff z Score the difference per resample END Repeat 000x then stop COUNT z >=.49 k Count how many times a resample of equal or greater than.49 occurred; store in k COUNT z <= -.49 j Count how many times a resample of equal or less than -.49 occurred; store in j ADD k+j m Add k and j together in n DIVIDE m/000 n Calculate a p-value by dividing m by 000 PRINT m Print result Using Resampling Stats plug-in for Microsoft Excel, this is what the print out might look like: Probability of have a difference greater or equal to.49 is zero. A difference of.49 never occurred in 000 resamples of the data. Bootstrap approach in Hypothesis Testing, a difference in treatment If you believe there is a difference between treatments, act as though they come from different populations; the difference is not due to chance. This is completed by resampling with replacement from individual treatments. The example is as follows: mice that were treated with a certain antibiotic lived 94, 38, 3, 97, 99, 6 and 4 days after their surgery with a mean survival period of 86.8 days. The untreated control lived 5, 0, 40, 04, 5, 7, 46, 30, and 46 days with a mean survival 56. days. The difference between the two means is 30.6 days H 0 : There is a difference in mean survival between treated and untreated. The basic programming to test this hypothesis might look like this: NUMBERS ( ) treatmt NUMBERS ( ) control REPEAT 000 SAMPLE 7 treatmt treatmt$ SAMPLE 9 control control$ Treatment Sample Control Sample Sample 7 treatments w/r. Sample 9 controls w/r. 3

14 MEAN treatmt$ tmean Mean of Treatments MEAN control$ cmean Mean of Control SUBTRACT tmean cmean diff Subtract SCORE diff scrboard Score the Difference END Repeat 000x HISTOGRAM scrboard Draw a histogram PERCENTILE scrboard ( ) interval Get the 95% Confidence Interval PRINT interval Give the 95% Confidence Interval Using Resampling Stats plug-in for Microsoft Excel, this is what the print out might look like: The difference of 30.6 days does not suggest that there is a significant difference of post-operational mice that were treated versus untreated. II. Estimation A. The Jackknife In the course of applying each of the empirical techniques in the construction of hypothesis tests we could also have estimated test statistics and a suite of characteristics of the test statistics and the empirical distributions. In the test or means we obviously could estimate the means, the variances (or standard errors), and the medians (their standard errors), etc. We could also estimate the bias associated with each of these estimators. Define an estimator θˆ of the parameter θ, E ( ˆ θ ) = θ + c then the bias of the estimator is c. The sample mean, x, is an unbiased estimator of μ because E (x) = μ, even though different samples may give different estimates x i of μ they are all unbiased estimates. In general, however, most estimators are biased, and 4

15 the bias can be depicted as a Taylor series expansion of the estimator. So the bias of θˆ is ( ˆ a a a3 E θ θ ) = + + n n n If we define ˆ J θ n ˆ θ ( n ) ˆ θ, i =,,..., n = i J to be a new estimator of θ, then the bias of θˆ is ˆ a a + a E( θ J θ ) = n n 3..., which is less than the bias of θˆ J estimator θˆ is computed as since it eliminates the term of order /n. In practice the ˆ J θ = n i= ˆ θ i = n ˆ θ ( n ) n i= ˆ θ i where i =,... n. This is the first order jackknife estimator. It is useful in that it is a less biased estimator although being somewhat more variable than the un-jackknifed estimator, but this increased variability is at maximum SE( ˆ θ J ) = SE( ˆ)( θ + a n ). Since the standard error of an estimator decreases as n by a factor of n, the J estimator θˆ has dispersion greater by a factor of /n than θˆ, but usually only n -3/ J greater than θˆ. Therefore the reduction in bias achieved by using θˆ is not offset by a similar increase in the magnitude of the estimator's variance. If we depict the bootstrap B estimator as θˆ then the jackknife estimator of the standard error of θˆ is σ ϑ n ( ˆ B = VAR θ L )( F ) n where F B is the empirical bootstrap probability distribution of the random variables and θˆl is a linear approximation of the estimator θˆ on the empirical bootstrap probability 5

16 B distribution. This implies that the bootstrap estimator θˆ has a standard error that is [n /n - ] / J times less than the jackknife estimator θˆ. The jackknife estimator has bias β ϑ n = n [ ˆ B E( θ ( F ) θ )] q Where θˆq is a quadratic approximation of the estimator θˆ on the distribution F B and E indicates the expectation with respect to bootstrap sampling. This implies that the bootstrap estimate of bias is n/(n - ) times less than the jackknife estimate of bias (see ref. 4). In general then the bootstrap will provide estimators with less bias and variance than the jackknife. Table 3 shows a data set generated by sampling from two normally distributed populations with μ = 00, σ = 60, and μ = 00 and σ = 30. To test the hypothesis that the variances of these populations are equal, that is H 0 :σ = σ versus the alternative that H A :σ σ, we could use the normal theory approach, which is again conditioned, on the assumptions mentioned earlier and elaborate the test statistic based on the sample estimates of σ s s n n, which is F distributed with n - numerator degrees of freedom and n - denominator degrees of freedom. Alternatively we could use a jackknife or a bootstrap estimate of the same or a similar test statistic. The F statistic computed under normal theory assumptions is F 9,4 = 8.86, p <0.00, while the bootstrap estimate of the probability of obtaining the observed F or one greater is 0.07 < p < The jackknife test is performed on the natural logs of the jackknifed variances rather than the variances themselves. A full description of the computations is given in reference (3). The test statistic for the jackknife test on variances is Q = B A V + V, 6

17 where B and A are the averages of the natural logs of the variances across the n and n jackknifed estimates and V and V are the variances of the jackknifed estimates of the variances. For large samples (n +n > 0), Q is N(0,), but for small equal size samples it follows student's t distribution with n +n - degrees of a freedom. For this example Q = , < p < Figure 9 presents the distribution of the bootstrap estimates of F, and Table 4 presents the jackknifed pseudo-values their standard errors and bias. Table 5 and Figure 0 provide a similar test for two exponential populations. Under the assumption of normality F 9,4 =.783, 0.05 < p < The jackknife test, however, yields Q = 3.095, < p < 0.00, and the bootstrap yields p > In these instances the jackknife is the most powerful test. III. Recommendations for Use In 99, Philip Crowley gave recommendations for these methods:. Use a Large Number of Repetitions use large amount, which will smoothen out the distribution of the data. Use 000 to start to get a rough idea and use 0,000 as final.. In the absence of random sampling, two or more samples with equivalent distributions should be tested by randomization. 3. Don t use the jackknife approach in confidence intervals and hypothesis testing randomization and bootstrap approaches are superior. 4. With small sample sizes, be a skeptic of parametric analysis. IV. Prospectus Thus far we have presented in a non-rigorous fashion a number of computationally expensive, empirical approaches to estimation and hypothesis testing. The theory underlying some of these approaches is well developed and I refer you to the reference list for that material. However, much of what I have presented has no rigorous theoretical underpinnings, but can be shown to be quite useful particularly in situations where the assumption of normality is suspect. The prognosis among statisticians is that theory will catch up to our computational prowess, so that many of these procedures will be justified and should be adopted. In the interim, however, should you choose to employ one of the more radical of these procedures be prepared for considerable disagreement over its validity and usefulness. The prospects for further development of these kinds of procedures, and work to establish their limitations, advantages, and care and maintenance is considerable. At present, however, the burden of investigating the properties of one of these procedures, in its application to a particular situation and test statistic rests with the investigator. References Bradley, J.U Distribution-free statistical tests. Prentice-Hall, Inc: Englewood Cliff, N.J. Conover, W.J Practical Nonparametric statistics. John Wiley and Sons: New York. 7

18 Crowley, P. 99. Resampling Methods for Computation-Intensive Data Analysis in Ecology and Evolution. Annu. Rev. Ecol. Syst. 3: Efron, B. and G. Gong A leisurely look at the bootstrap, the jackknife, and crossvalidation. The American Statistician 37: Good, P.L Resampling Methods: 3rd Edition. Birkhauser. Hollander, M. and D.A. Wolfe Nonparametric Statistical Methods. John Wiley and Sons: New York. Simon, J.L Resampling: the New Statistics. Resampling Stats. FREE ONLINE! Other Readable Literature Miller, R.G The jackknife - a review. Biometrika 6: -5. Peters, S.C. and D.A. Freedman Some notes on the Bootstrap in regression problems. Journal of Business and Economic Statistics : Efron, B Bootstrap Methods: another look at the jackknife. Annals of Statistics 7: -6. Other Not So Readable Literature Arvesen, J.N Jackknifing U-Statistics. Annals of Mathematical Statistics 40: Miller, R.G A trustworthy jackknife. Annals of Mathematical Statistics 35: Miller, R.G Jackknifing variances. Annals of Mathematical Statistics 39: Quenouille, M.H Notes on bias in estimation. Biometrika 43: Some Applications Zahl, S Jackknifing an index of diversity. Ecology 58: Heltshe, J.F. and N.E. Forrester Statistical evaluation of the jackknife estimate of diversity when using quadrat samples. Ecology 66: 07-. Routledge, R.D Bias in estimating the diversity of large uncensused communities. Ecology 6: Originally by Dr. Edward Connor, modified by Eugenel Espiritu, June,

6 Single Sample Methods for a Location Parameter

6 Single Sample Methods for a Location Parameter 6 Single Sample Methods for a Location Parameter If there are serious departures from parametric test assumptions (e.g., normality or symmetry), nonparametric tests on a measure of central tendency (usually

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

ON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES

ON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES ON SMALL SAMPLE PROPERTIES OF PERMUTATION TESTS: INDEPENDENCE BETWEEN TWO SAMPLES Hisashi Tanizaki Graduate School of Economics, Kobe University, Kobe 657-8501, Japan e-mail: tanizaki@kobe-u.ac.jp Abstract:

More information

The Nonparametric Bootstrap

The Nonparametric Bootstrap The Nonparametric Bootstrap The nonparametric bootstrap may involve inferences about a parameter, but we use a nonparametric procedure in approximating the parametric distribution using the ECDF. We use

More information

Permutation Tests. Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods

Permutation Tests. Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods Permutation Tests Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods The Two-Sample Problem We observe two independent random samples: F z = z 1, z 2,, z n independently of

More information

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions

Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions Journal of Modern Applied Statistical Methods Volume 8 Issue 1 Article 13 5-1-2009 Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error

More information

STAT Section 5.8: Block Designs

STAT Section 5.8: Block Designs STAT 518 --- Section 5.8: Block Designs Recall that in paired-data studies, we match up pairs of subjects so that the two subjects in a pair are alike in some sense. Then we randomly assign, say, treatment

More information

11. Bootstrap Methods

11. Bootstrap Methods 11. Bootstrap Methods c A. Colin Cameron & Pravin K. Trivedi 2006 These transparencies were prepared in 20043. They can be used as an adjunct to Chapter 11 of our subsequent book Microeconometrics: Methods

More information

HYPOTHESIS TESTING. Hypothesis Testing

HYPOTHESIS TESTING. Hypothesis Testing MBA 605 Business Analytics Don Conant, PhD. HYPOTHESIS TESTING Hypothesis testing involves making inferences about the nature of the population on the basis of observations of a sample drawn from the population.

More information

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions

Purposes of Data Analysis. Variables and Samples. Parameters and Statistics. Part 1: Probability Distributions Part 1: Probability Distributions Purposes of Data Analysis True Distributions or Relationships in the Earths System Probability Distribution Normal Distribution Student-t Distribution Chi Square Distribution

More information

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review STATS 200: Introduction to Statistical Inference Lecture 29: Course review Course review We started in Lecture 1 with a fundamental assumption: Data is a realization of a random process. The goal throughout

More information

1 Overview. Coefficients of. Correlation, Alienation and Determination. Hervé Abdi Lynne J. Williams

1 Overview. Coefficients of. Correlation, Alienation and Determination. Hervé Abdi Lynne J. Williams In Neil Salkind (Ed.), Encyclopedia of Research Design. Thousand Oaks, CA: Sage. 2010 Coefficients of Correlation, Alienation and Determination Hervé Abdi Lynne J. Williams 1 Overview The coefficient of

More information

Psychology 282 Lecture #4 Outline Inferences in SLR

Psychology 282 Lecture #4 Outline Inferences in SLR Psychology 282 Lecture #4 Outline Inferences in SLR Assumptions To this point we have not had to make any distributional assumptions. Principle of least squares requires no assumptions. Can use correlations

More information

Hypothesis testing. Data to decisions

Hypothesis testing. Data to decisions Hypothesis testing Data to decisions The idea Null hypothesis: H 0 : the DGP/population has property P Under the null, a sample statistic has a known distribution If, under that that distribution, the

More information

STAT 536: Genetic Statistics

STAT 536: Genetic Statistics STAT 536: Genetic Statistics Tests for Hardy Weinberg Equilibrium Karin S. Dorman Department of Statistics Iowa State University September 7, 2006 Statistical Hypothesis Testing Identify a hypothesis,

More information

The bootstrap. Patrick Breheny. December 6. The empirical distribution function The bootstrap

The bootstrap. Patrick Breheny. December 6. The empirical distribution function The bootstrap Patrick Breheny December 6 Patrick Breheny BST 764: Applied Statistical Modeling 1/21 The empirical distribution function Suppose X F, where F (x) = Pr(X x) is a distribution function, and we wish to estimate

More information

The Statistics of Hypothesis-Testing With Measured Data

The Statistics of Hypothesis-Testing With Measured Data 270 Resampling: The New Statistics CHAPTER 18 The Statistics of Hypothesis-Testing With Measured Data Chapters 15 and 17 discussed testing a hypothesis with data that either arrive in dichotomized (yes-no)

More information

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2

MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 MA 575 Linear Models: Cedric E. Ginestet, Boston University Non-parametric Inference, Polynomial Regression Week 9, Lecture 2 1 Bootstrapped Bias and CIs Given a multiple regression model with mean and

More information

Unit 14: Nonparametric Statistical Methods

Unit 14: Nonparametric Statistical Methods Unit 14: Nonparametric Statistical Methods Statistics 571: Statistical Methods Ramón V. León 8/8/2003 Unit 14 - Stat 571 - Ramón V. León 1 Introductory Remarks Most methods studied so far have been based

More information

Answers and expectations

Answers and expectations Answers and expectations For a function f(x) and distribution P(x), the expectation of f with respect to P is The expectation is the average of f, when x is drawn from the probability distribution P E

More information

IENG581 Design and Analysis of Experiments INTRODUCTION

IENG581 Design and Analysis of Experiments INTRODUCTION Experimental Design IENG581 Design and Analysis of Experiments INTRODUCTION Experiments are performed by investigators in virtually all fields of inquiry, usually to discover something about a particular

More information

The exact bootstrap method shown on the example of the mean and variance estimation

The exact bootstrap method shown on the example of the mean and variance estimation Comput Stat (2013) 28:1061 1077 DOI 10.1007/s00180-012-0350-0 ORIGINAL PAPER The exact bootstrap method shown on the example of the mean and variance estimation Joanna Kisielinska Received: 21 May 2011

More information

Fourier and Stats / Astro Stats and Measurement : Stats Notes

Fourier and Stats / Astro Stats and Measurement : Stats Notes Fourier and Stats / Astro Stats and Measurement : Stats Notes Andy Lawrence, University of Edinburgh Autumn 2013 1 Probabilities, distributions, and errors Laplace once said Probability theory is nothing

More information

Probability and Statistics

Probability and Statistics Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT

More information

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares

Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares Keppel, G. & Wickens, T.D. Design and Analysis Chapter 2: Sources of Variability and Sums of Squares K&W introduce the notion of a simple experiment with two conditions. Note that the raw data (p. 16)

More information

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007)

CHAPTER 17 CHI-SQUARE AND OTHER NONPARAMETRIC TESTS FROM: PAGANO, R. R. (2007) FROM: PAGANO, R. R. (007) I. INTRODUCTION: DISTINCTION BETWEEN PARAMETRIC AND NON-PARAMETRIC TESTS Statistical inference tests are often classified as to whether they are parametric or nonparametric Parameter

More information

Multiple Regression Analysis

Multiple Regression Analysis Multiple Regression Analysis y = β 0 + β 1 x 1 + β 2 x 2 +... β k x k + u 2. Inference 0 Assumptions of the Classical Linear Model (CLM)! So far, we know: 1. The mean and variance of the OLS estimators

More information

Non-parametric Statistics

Non-parametric Statistics 45 Contents Non-parametric Statistics 45.1 Non-parametric Tests for a Single Sample 45. Non-parametric Tests for Two Samples 4 Learning outcomes You will learn about some significance tests which may be

More information

Physics 509: Bootstrap and Robust Parameter Estimation

Physics 509: Bootstrap and Robust Parameter Estimation Physics 509: Bootstrap and Robust Parameter Estimation Scott Oser Lecture #20 Physics 509 1 Nonparametric parameter estimation Question: what error estimate should you assign to the slope and intercept

More information

AP Statistics Cumulative AP Exam Study Guide

AP Statistics Cumulative AP Exam Study Guide AP Statistics Cumulative AP Eam Study Guide Chapters & 3 - Graphs Statistics the science of collecting, analyzing, and drawing conclusions from data. Descriptive methods of organizing and summarizing statistics

More information

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals

Last week: Sample, population and sampling distributions finished with estimation & confidence intervals Past weeks: Measures of central tendency (mean, mode, median) Measures of dispersion (standard deviation, variance, range, etc). Working with the normal curve Last week: Sample, population and sampling

More information

18.05 Practice Final Exam

18.05 Practice Final Exam No calculators. 18.05 Practice Final Exam Number of problems 16 concept questions, 16 problems. Simplifying expressions Unless asked to explicitly, you don t need to simplify complicated expressions. For

More information

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances Advances in Decision Sciences Volume 211, Article ID 74858, 8 pages doi:1.1155/211/74858 Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances David Allingham 1 andj.c.w.rayner

More information

Resampling and the Bootstrap

Resampling and the Bootstrap Resampling and the Bootstrap Axel Benner Biostatistics, German Cancer Research Center INF 280, D-69120 Heidelberg benner@dkfz.de Resampling and the Bootstrap 2 Topics Estimation and Statistical Testing

More information

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages: Glossary The ISI glossary of statistical terms provides definitions in a number of different languages: http://isi.cbs.nl/glossary/index.htm Adjusted r 2 Adjusted R squared measures the proportion of the

More information

A Monte-Carlo study of asymptotically robust tests for correlation coefficients

A Monte-Carlo study of asymptotically robust tests for correlation coefficients Biometrika (1973), 6, 3, p. 661 551 Printed in Great Britain A Monte-Carlo study of asymptotically robust tests for correlation coefficients BY G. T. DUNCAN AND M. W. J. LAYAKD University of California,

More information

DISTRIBUTIONS USED IN STATISTICAL WORK

DISTRIBUTIONS USED IN STATISTICAL WORK DISTRIBUTIONS USED IN STATISTICAL WORK In one of the classic introductory statistics books used in Education and Psychology (Glass and Stanley, 1970, Prentice-Hall) there was an excellent chapter on different

More information

Tutorial 4: Power and Sample Size for the Two-sample t-test with Unequal Variances

Tutorial 4: Power and Sample Size for the Two-sample t-test with Unequal Variances Tutorial 4: Power and Sample Size for the Two-sample t-test with Unequal Variances Preface Power is the probability that a study will reject the null hypothesis. The estimated probability is a function

More information

Contents. Acknowledgments. xix

Contents. Acknowledgments. xix Table of Preface Acknowledgments page xv xix 1 Introduction 1 The Role of the Computer in Data Analysis 1 Statistics: Descriptive and Inferential 2 Variables and Constants 3 The Measurement of Variables

More information

A nonparametric two-sample wald test of equality of variances

A nonparametric two-sample wald test of equality of variances University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 211 A nonparametric two-sample wald test of equality of variances David

More information

Glossary for the Triola Statistics Series

Glossary for the Triola Statistics Series Glossary for the Triola Statistics Series Absolute deviation The measure of variation equal to the sum of the deviations of each value from the mean, divided by the number of values Acceptance sampling

More information

Inferences for Regression

Inferences for Regression Inferences for Regression An Example: Body Fat and Waist Size Looking at the relationship between % body fat and waist size (in inches). Here is a scatterplot of our data set: Remembering Regression In

More information

Chapter 5 Confidence Intervals

Chapter 5 Confidence Intervals Chapter 5 Confidence Intervals Confidence Intervals about a Population Mean, σ, Known Abbas Motamedi Tennessee Tech University A point estimate: a single number, calculated from a set of data, that is

More information

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods

Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods Chapter 4 Confidence Intervals in Ridge Regression using Jackknife and Bootstrap Methods 4.1 Introduction It is now explicable that ridge regression estimator (here we take ordinary ridge estimator (ORE)

More information

Bootstrapping Australian inbound tourism

Bootstrapping Australian inbound tourism 19th International Congress on Modelling and Simulation, Perth, Australia, 12 16 December 2011 http://mssanz.org.au/modsim2011 Bootstrapping Australian inbound tourism Y.H. Cheunga and G. Yapa a School

More information

Tutorial 3: Power and Sample Size for the Two-sample t-test with Equal Variances. Acknowledgements:

Tutorial 3: Power and Sample Size for the Two-sample t-test with Equal Variances. Acknowledgements: Tutorial 3: Power and Sample Size for the Two-sample t-test with Equal Variances Anna E. Barón, Keith E. Muller, Sarah M. Kreidler, and Deborah H. Glueck Acknowledgements: The project was supported in

More information

Class 26: review for final exam 18.05, Spring 2014

Class 26: review for final exam 18.05, Spring 2014 Probability Class 26: review for final eam 8.05, Spring 204 Counting Sets Inclusion-eclusion principle Rule of product (multiplication rule) Permutation and combinations Basics Outcome, sample space, event

More information

Lecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series

Lecture Slides. Elementary Statistics. by Mario F. Triola. and the Triola Statistics Series Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 13 Nonparametric Statistics 13-1 Overview 13-2 Sign Test 13-3 Wilcoxon Signed-Ranks

More information

Topic 1. Definitions

Topic 1. Definitions S Topic. Definitions. Scalar A scalar is a number. 2. Vector A vector is a column of numbers. 3. Linear combination A scalar times a vector plus a scalar times a vector, plus a scalar times a vector...

More information

Lecture Slides. Section 13-1 Overview. Elementary Statistics Tenth Edition. Chapter 13 Nonparametric Statistics. by Mario F.

Lecture Slides. Section 13-1 Overview. Elementary Statistics Tenth Edition. Chapter 13 Nonparametric Statistics. by Mario F. Lecture Slides Elementary Statistics Tenth Edition and the Triola Statistics Series by Mario F. Triola Slide 1 Chapter 13 Nonparametric Statistics 13-1 Overview 13-2 Sign Test 13-3 Wilcoxon Signed-Ranks

More information

Two-sample Categorical data: Testing

Two-sample Categorical data: Testing Two-sample Categorical data: Testing Patrick Breheny April 1 Patrick Breheny Introduction to Biostatistics (171:161) 1/28 Separate vs. paired samples Despite the fact that paired samples usually offer

More information

Contents. Counting Methods and Induction

Contents. Counting Methods and Induction Contents Counting Methods and Induction Lesson 1 Counting Strategies Investigations 1 Careful Counting... 555 Order and Repetition I... 56 3 Order and Repetition II... 569 On Your Own... 573 Lesson Counting

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

Math 494: Mathematical Statistics

Math 494: Mathematical Statistics Math 494: Mathematical Statistics Instructor: Jimin Ding jmding@wustl.edu Department of Mathematics Washington University in St. Louis Class materials are available on course website (www.math.wustl.edu/

More information

Estimation and Hypothesis Testing in LAV Regression with Autocorrelated Errors: Is Correction for Autocorrelation Helpful?

Estimation and Hypothesis Testing in LAV Regression with Autocorrelated Errors: Is Correction for Autocorrelation Helpful? Journal of Modern Applied Statistical Methods Volume 10 Issue Article 13 11-1-011 Estimation and Hypothesis Testing in LAV Regression with Autocorrelated Errors: Is Correction for Autocorrelation Helpful?

More information

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES

BIOL Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES BIOL 458 - Biometry LAB 6 - SINGLE FACTOR ANOVA and MULTIPLE COMPARISON PROCEDURES PART 1: INTRODUCTION TO ANOVA Purpose of ANOVA Analysis of Variance (ANOVA) is an extremely useful statistical method

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8

CIVL /8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 CIVL - 7904/8904 T R A F F I C F L O W T H E O R Y L E C T U R E - 8 Chi-square Test How to determine the interval from a continuous distribution I = Range 1 + 3.322(logN) I-> Range of the class interval

More information

A Signed-Rank Test Based on the Score Function

A Signed-Rank Test Based on the Score Function Applied Mathematical Sciences, Vol. 10, 2016, no. 51, 2517-2527 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/ams.2016.66189 A Signed-Rank Test Based on the Score Function Hyo-Il Park Department

More information

Analysis of 2x2 Cross-Over Designs using T-Tests

Analysis of 2x2 Cross-Over Designs using T-Tests Chapter 234 Analysis of 2x2 Cross-Over Designs using T-Tests Introduction This procedure analyzes data from a two-treatment, two-period (2x2) cross-over design. The response is assumed to be a continuous

More information

2008 Winton. Statistical Testing of RNGs

2008 Winton. Statistical Testing of RNGs 1 Statistical Testing of RNGs Criteria for Randomness For a sequence of numbers to be considered a sequence of randomly acquired numbers, it must have two basic statistical properties: Uniformly distributed

More information

Power Calculations for Preclinical Studies Using a K-Sample Rank Test and the Lehmann Alternative Hypothesis

Power Calculations for Preclinical Studies Using a K-Sample Rank Test and the Lehmann Alternative Hypothesis Power Calculations for Preclinical Studies Using a K-Sample Rank Test and the Lehmann Alternative Hypothesis Glenn Heller Department of Epidemiology and Biostatistics, Memorial Sloan-Kettering Cancer Center,

More information

Practice Questions: Statistics W1111, Fall Solutions

Practice Questions: Statistics W1111, Fall Solutions Practice Questions: Statistics W, Fall 9 Solutions Question.. The standard deviation of Z is 89... P(=6) =..3. is definitely inside of a 95% confidence interval for..4. (a) YES (b) YES (c) NO (d) NO Questions

More information

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages

18.05 Final Exam. Good luck! Name. No calculators. Number of problems 16 concept questions, 16 problems, 21 pages Name No calculators. 18.05 Final Exam Number of problems 16 concept questions, 16 problems, 21 pages Extra paper If you need more space we will provide some blank paper. Indicate clearly that your solution

More information

Do not copy, post, or distribute

Do not copy, post, or distribute 14 CORRELATION ANALYSIS AND LINEAR REGRESSION Assessing the Covariability of Two Quantitative Properties 14.0 LEARNING OBJECTIVES In this chapter, we discuss two related techniques for assessing a possible

More information

Bootstrapping, Permutations, and Monte Carlo Testing

Bootstrapping, Permutations, and Monte Carlo Testing Bootstrapping, Permutations, and Monte Carlo Testing Problem: Population of interest is extremely rare spatially and you are interested in using a 95% CI to estimate total abundance. The sampling design

More information

A comparison study of the nonparametric tests based on the empirical distributions

A comparison study of the nonparametric tests based on the empirical distributions 통계연구 (2015), 제 20 권제 3 호, 1-12 A comparison study of the nonparametric tests based on the empirical distributions Hyo-Il Park 1) Abstract In this study, we propose a nonparametric test based on the empirical

More information

Samples and Populations Confidence Intervals Hypotheses One-sided vs. two-sided Statistical Significance Error Types. Statistiek I.

Samples and Populations Confidence Intervals Hypotheses One-sided vs. two-sided Statistical Significance Error Types. Statistiek I. Statistiek I Sampling John Nerbonne CLCG, Rijksuniversiteit Groningen http://www.let.rug.nl/nerbonne/teach/statistiek-i/ John Nerbonne 1/41 Overview 1 Samples and Populations 2 Confidence Intervals 3 Hypotheses

More information

Exam 2 (KEY) July 20, 2009

Exam 2 (KEY) July 20, 2009 STAT 2300 Business Statistics/Summer 2009, Section 002 Exam 2 (KEY) July 20, 2009 Name: USU A#: Score: /225 Directions: This exam consists of six (6) questions, assessing material learned within Modules

More information

Chapter 24. Comparing Means

Chapter 24. Comparing Means Chapter 4 Comparing Means!1 /34 Homework p579, 5, 7, 8, 10, 11, 17, 31, 3! /34 !3 /34 Objective Students test null and alternate hypothesis about two!4 /34 Plot the Data The intuitive display for comparing

More information

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14

Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 CS 70 Discrete Mathematics and Probability Theory Spring 2016 Rao and Walrand Note 14 Introduction One of the key properties of coin flips is independence: if you flip a fair coin ten times and get ten

More information

Inferential Statistics

Inferential Statistics Inferential Statistics Eva Riccomagno, Maria Piera Rogantin DIMA Università di Genova riccomagno@dima.unige.it rogantin@dima.unige.it Part G Distribution free hypothesis tests 1. Classical and distribution-free

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Chapter 15: Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics

Chapter 15: Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics Section 15.1: An Overview of Nonparametric Statistics Understand Difference between Parametric and Nonparametric Statistical Procedures Parametric statistical procedures inferential procedures that rely

More information

Fin285a:Computer Simulations and Risk Assessment Section 2.3.2:Hypothesis testing, and Confidence Intervals

Fin285a:Computer Simulations and Risk Assessment Section 2.3.2:Hypothesis testing, and Confidence Intervals Fin285a:Computer Simulations and Risk Assessment Section 2.3.2:Hypothesis testing, and Confidence Intervals Overview Hypothesis testing terms Testing a die Testing issues Estimating means Confidence intervals

More information

Probability Distributions.

Probability Distributions. Probability Distributions http://www.pelagicos.net/classes_biometry_fa18.htm Probability Measuring Discrete Outcomes Plotting probabilities for discrete outcomes: 0.6 0.5 0.4 0.3 0.2 0.1 NOTE: Area within

More information

INTERPENETRATING SAMPLES

INTERPENETRATING SAMPLES INTERPENETRATING SAMPLES Dr. A. C. Kulshreshtha UN Statistical Institute for Asia and Pacific Second RAP Regional Workshop on Building Training Resources for Improving Agricultural & Rural Statistics Sampling

More information

7 Estimation. 7.1 Population and Sample (P.91-92)

7 Estimation. 7.1 Population and Sample (P.91-92) 7 Estimation MATH1015 Biostatistics Week 7 7.1 Population and Sample (P.91-92) Suppose that we wish to study a particular health problem in Australia, for example, the average serum cholesterol level for

More information

Inferences About the Difference Between Two Means

Inferences About the Difference Between Two Means 7 Inferences About the Difference Between Two Means Chapter Outline 7.1 New Concepts 7.1.1 Independent Versus Dependent Samples 7.1. Hypotheses 7. Inferences About Two Independent Means 7..1 Independent

More information

Chapter 16. Simple Linear Regression and Correlation

Chapter 16. Simple Linear Regression and Correlation Chapter 16 Simple Linear Regression and Correlation 16.1 Regression Analysis Our problem objective is to analyze the relationship between interval variables; regression analysis is the first tool we will

More information

MIDTERM EXAMINATION (Spring 2011) STA301- Statistics and Probability

MIDTERM EXAMINATION (Spring 2011) STA301- Statistics and Probability STA301- Statistics and Probability Solved MCQS From Midterm Papers March 19,2012 MC100401285 Moaaz.pk@gmail.com Mc100401285@gmail.com PSMD01 MIDTERM EXAMINATION (Spring 2011) STA301- Statistics and Probability

More information

Appendix D INTRODUCTION TO BOOTSTRAP ESTIMATION D.1 INTRODUCTION

Appendix D INTRODUCTION TO BOOTSTRAP ESTIMATION D.1 INTRODUCTION Appendix D INTRODUCTION TO BOOTSTRAP ESTIMATION D.1 INTRODUCTION Bootstrapping is a general, distribution-free method that is used to estimate parameters ofinterest from data collected from studies or

More information

Math Review Sheet, Fall 2008

Math Review Sheet, Fall 2008 1 Descriptive Statistics Math 3070-5 Review Sheet, Fall 2008 First we need to know about the relationship among Population Samples Objects The distribution of the population can be given in one of the

More information

hypotheses. P-value Test for a 2 Sample z-test (Large Independent Samples) n > 30 P-value Test for a 2 Sample t-test (Small Samples) n < 30 Identify α

hypotheses. P-value Test for a 2 Sample z-test (Large Independent Samples) n > 30 P-value Test for a 2 Sample t-test (Small Samples) n < 30 Identify α Chapter 8 Notes Section 8-1 Independent and Dependent Samples Independent samples have no relation to each other. An example would be comparing the costs of vacationing in Florida to the cost of vacationing

More information

Resampling and the Bootstrap

Resampling and the Bootstrap Resampling and the Bootstrap Axel Benner Biostatistics, German Cancer Research Center INF 280, D-69120 Heidelberg benner@dkfz.de Resampling and the Bootstrap 2 Topics Estimation and Statistical Testing

More information

Nonparametric tests. Mark Muldoon School of Mathematics, University of Manchester. Mark Muldoon, November 8, 2005 Nonparametric tests - p.

Nonparametric tests. Mark Muldoon School of Mathematics, University of Manchester. Mark Muldoon, November 8, 2005 Nonparametric tests - p. Nonparametric s Mark Muldoon School of Mathematics, University of Manchester Mark Muldoon, November 8, 2005 Nonparametric s - p. 1/31 Overview The sign, motivation The Mann-Whitney Larger Larger, in pictures

More information

Single Sample Means. SOCY601 Alan Neustadtl

Single Sample Means. SOCY601 Alan Neustadtl Single Sample Means SOCY601 Alan Neustadtl The Central Limit Theorem If we have a population measured by a variable with a mean µ and a standard deviation σ, and if all possible random samples of size

More information

Two Correlated Proportions Non- Inferiority, Superiority, and Equivalence Tests

Two Correlated Proportions Non- Inferiority, Superiority, and Equivalence Tests Chapter 59 Two Correlated Proportions on- Inferiority, Superiority, and Equivalence Tests Introduction This chapter documents three closely related procedures: non-inferiority tests, superiority (by a

More information

STAT440/840: Statistical Computing

STAT440/840: Statistical Computing First Prev Next Last STAT440/840: Statistical Computing Paul Marriott pmarriott@math.uwaterloo.ca MC 6096 February 2, 2005 Page 1 of 41 First Prev Next Last Page 2 of 41 Chapter 3: Data resampling: the

More information

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z).

Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). Table of z values and probabilities for the standard normal distribution. z is the first column plus the top row. Each cell shows P(X z). For example P(X.04) =.8508. For z < 0 subtract the value from,

More information

Empirical Power of Four Statistical Tests in One Way Layout

Empirical Power of Four Statistical Tests in One Way Layout International Mathematical Forum, Vol. 9, 2014, no. 28, 1347-1356 HIKARI Ltd, www.m-hikari.com http://dx.doi.org/10.12988/imf.2014.47128 Empirical Power of Four Statistical Tests in One Way Layout Lorenzo

More information

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career. Introduction to Data and Analysis Wildlife Management is a very quantitative field of study Results from studies will be used throughout this course and throughout your career. Sampling design influences

More information

What is a Hypothesis?

What is a Hypothesis? What is a Hypothesis? A hypothesis is a claim (assumption) about a population parameter: population mean Example: The mean monthly cell phone bill in this city is μ = $42 population proportion Example:

More information

AIM HIGH SCHOOL. Curriculum Map W. 12 Mile Road Farmington Hills, MI (248)

AIM HIGH SCHOOL. Curriculum Map W. 12 Mile Road Farmington Hills, MI (248) AIM HIGH SCHOOL Curriculum Map 2923 W. 12 Mile Road Farmington Hills, MI 48334 (248) 702-6922 www.aimhighschool.com COURSE TITLE: Statistics DESCRIPTION OF COURSE: PREREQUISITES: Algebra 2 Students will

More information

Logistic Regression: Regression with a Binary Dependent Variable

Logistic Regression: Regression with a Binary Dependent Variable Logistic Regression: Regression with a Binary Dependent Variable LEARNING OBJECTIVES Upon completing this chapter, you should be able to do the following: State the circumstances under which logistic regression

More information

Political Science 236 Hypothesis Testing: Review and Bootstrapping

Political Science 236 Hypothesis Testing: Review and Bootstrapping Political Science 236 Hypothesis Testing: Review and Bootstrapping Rocío Titiunik Fall 2007 1 Hypothesis Testing Definition 1.1 Hypothesis. A hypothesis is a statement about a population parameter The

More information

Hotelling s One- Sample T2

Hotelling s One- Sample T2 Chapter 405 Hotelling s One- Sample T2 Introduction The one-sample Hotelling s T2 is the multivariate extension of the common one-sample or paired Student s t-test. In a one-sample t-test, the mean response

More information

Tutorial 1: Power and Sample Size for the One-sample t-test. Acknowledgements:

Tutorial 1: Power and Sample Size for the One-sample t-test. Acknowledgements: Tutorial 1: Power and Sample Size for the One-sample t-test Anna E. Barón, Keith E. Muller, Sarah M. Kreidler, and Deborah H. Glueck Acknowledgements: The project was supported in large part by the National

More information

Chapter 9 Inferences from Two Samples

Chapter 9 Inferences from Two Samples Chapter 9 Inferences from Two Samples 9-1 Review and Preview 9-2 Two Proportions 9-3 Two Means: Independent Samples 9-4 Two Dependent Samples (Matched Pairs) 9-5 Two Variances or Standard Deviations Review

More information

Math 20 Spring Discrete Probability. Midterm Exam

Math 20 Spring Discrete Probability. Midterm Exam Math 20 Spring 203 Discrete Probability Midterm Exam Thursday April 25, 5:00 7:00 PM Your name (please print): Instructions: This is a closed book, closed notes exam. Use of calculators is not permitted.

More information