RT Analysis 2 Contents Abstract 3 Analysis of Response Time Distributions 4 Characterization of Random Variables 5 Density Function

Size: px
Start display at page:

Download "RT Analysis 2 Contents Abstract 3 Analysis of Response Time Distributions 4 Characterization of Random Variables 5 Density Function"

Transcription

1 RT Analysis 1 Analysis of Response Time Distributions Trisha Van Zandt Ohio State University Running Head: RT Distributions Word Count: Approximately 25,000 Address correspondence to: Trisha Van Zandt Psychology Department Ohio State University 1885 Neil Avenue Mall Columbus, OH

2 RT Analysis 2 Contents Abstract 3 Analysis of Response Time Distributions 4 Characterization of Random Variables 5 Density Function Distribution Function Survivor and Hazard Functions Estimation 10 Properties of Estimators Parameter Estimation Nonparametric Function Estimation The Density Function The Hazard Function Model Fitting Summary Model and Hypothesis Testing 36 Orderings of Distributions RT Decomposition Mixture Distributions Concluding Remarks 60 References 64 Appendix 73

3 RT Analysis 3 Abstract This chapter addresses the problem of response time analysis, including estimates of the mean and variance, outlier techniques, estimation of distribution parameters, and function estimation. The use of distributional analysis in testing processing models is also discussed.

4 RT Analysis 4 Analysis of Response Time Distributions Response time is ubiquitous in experimental psychology. It is perhaps the most important measure used to investigate hypotheses about mental processing. Fifteen of the 24 studies published in the February, 2000 issue of the Journal of Experimental Psychology: Human Perception and Performance used some form of response time analysis. Not restricted to cognitive psychology, response times (RTs) are collected routinely in empirical investigations in biological, social, developmental, and clinical psychology as well. RTs are collected from both human and animal subjects. Over 27,000 abstracts in the PsychInfo database spanning from 1887 to the end of April, 2000 make reference to reaction or response time or latency. One reason that RTs are so important in experimental psychology is because they, like other physical measurements such as length or weight or force, are dened on a ratio scale (Townsend, 1992). This means that we can bring all possible mathematical machinery to bear on the analysis of RT. We can devise precise mathematical models of cognitive processing, make predictions about process durations as a function of physical (numerical) variables in the experimental context, and how those processes change with changes in these variables. Since Kinnebrook's brief tenure in Maskelyne's observatory in 1796 (Mollon & Perkins, 1996), variations in RT have seemed to be a clean and easy way to get at how mental processing unfolds over time. Although the use of RTs in cognitive research has been established for over a century, basic issues of analysis still arise. Techniques such as Neyman-Pearson null hypothesis testing are routinely applied to RTs, even though the basic assumptions required for such analyses, such as normality and independence, are known to be violated. 1 RT distributions are not normally distributed; rather, they are positively skewed. Individual RTs are not typically independent of one another; some trial-by-trial sequential eects persist even in the most carefully controlled experiments. For these and other reasons, RT analysis is not always as straightforward as it appears. 1 These violations are particularly obvious in the analysis of single-subject data, where raw RTs are subjected to Neyman-Pearson methods. However, even in the more common group analysis, where the mean RTs from individual subjects are used to compute the test statistics, the critical assumption of identically distributed observations is violated, even though the individual means can safely be assumed to be normal.

5 RT Analysis 5 For the most part, RT analyses occur at the level of the means. That is, hypotheses are formulated with regard to predicted average increases or decreases in RT, as opposed to eects on less obvious parameters of the RT distribution such as skew. Of the fteen studies in the issue of the journal cited above, twelve reported only mean RTs. The remaining three studies investigated the RT distributions in greater depth. This disparity actually reects a recent trend: it is becoming more and more important to consider the overall distribution, as many important and interesting ndings are typically hidden when collapsing observations into averages. Some of my colleagues have seen changes in patterns of RTs in their data across conditions but faced the frustration that ANOVAs on the mean RTs failed to show signicant main eects. They then performed the same ANOVAs on the RT medians, or on some trimmed or otherwise modied averages, in an attempt to demonstrate for some measure of central tendency what was obvious at the level of the distribution. The purpose of this chapter is to provide an outline of RT analyses, particularly at the distributional level. Space constrains me from writing a complete primer or handbook of such analyses, but appropriate references will be provided along the way. There are many statistical and mathematical packages that can help those who are less skilled at programming to perform these analyses, and pointers to these packages will also be provided. In particular, I have provided in an appendix to this chapter some MATLAB routines that can perform some of the analyses we will examine. I will begin by discussing RTs as random variables and the ways that random variables can be characterized. I will present problems of estimation, not only of central tendency but also of the parameters of a given theoretical distribution. I will also discuss estimation of the functional forms of the distribution. In the second half of the paper, I will give some thought to how distributional estimates can be used to test theories about the structure of processing. Characterization of Random Variables RTs are random variables. A sample of RTs, such as might be collected in an experiment (under xed conditions), is drawn at random from the entire population or distribution of possible RTs. The observations in the sample are assumed to be independent and

6 RT Analysis 6 identically distributed, or iid. That is, each observation was taken from exactly the same distribution as each other observation (there are no changes in the distributional parameters or in the form of the distribution across trials), and the observations are statistically independent from each other (so the probability of the i th observation is unaected by the value of the j th observation). The iid assumption is important for a number of reasons, not the least of which is the desire that the many trials within identical experimental conditions all have exactly the same eect on the human observer. (Note, however, that the independence assumption is likely to be violated by sequential eects and parameter drift (Burbeck & Luce, 1982). Unfortunately, we don't have the tools to deal with this problem. We must assume that the covariances between observations are small enough that they can be neglected.) A random variable can be characterized in a number of ways. One way is to specify parameters of the distribution from which the variable is sampled, such as its mean or variance. Another way is to specify the functional form of the distribution itself. There are several useful functional forms, including the density, distribution, survivor and hazard functions. Each of these dierent functions describes a dierent aspect of the behavior of the random variable, but the relationship between the functions is unique. That is, once a particular density function has been specied, for example, the distribution, survivor and hazard functions are then determined. Random variables from dierent distributions cannot have the same functional form, regardless of which kind of function is examined. Therefore, only one of these functions needs to be specied to completely characterize the behavior of the random variable. I will now discuss each functional form and the relationships between them. For additional information, the reader might consult Luce (Luce, 1986), Townsend and Ashby (1983), or any good textbook on probability theory (e.g., Hoel, Port & Stone, 1971). Density Function The density functions f(t) for an ex-gaussian (the sum of a normal and an exponential random variable) and a Wald random variable are shown in Figure 1, top panel. The x-axis is time, or potential values of RT, and the y-axis is the height of the density function f(t).

7 RT Analysis 7 If RT were a discrete random variable, the height of the function at a point t would give the probability of observing an RT equal to t. However, RT is generally considered to be a continuous random variable, and therefore the value of the density function does not represent probability. Instead, it is the area under the density function that gives measures of probability. For example, suppose that we want to know the probability of observing an RT between values a and b. This probability is measured by integrating the density function between the values a and b: P (a RT b) = Z b a f(t)dt: More generally, any positive function f that has total area of 1 when integrated from negative to positive innity is a density function. For RT data, the shapes of the empirical (estimated) density functions are typically unimodal and positively skewed, like those shown in Figure 1. The majority of the observations are generally fast, but a large proportion of the observations are slower, pulling the positive tail of the density function to the right. There are a number of distributions that have density functions of this form, including the gamma, Wald, and Weibull distributions, but by far the most popular characterization of RT densities is that of the ex-gaussian (e.g., Balota & Spieler, 1999; Heathcote, Popiel & Mewhort, 1991; Ratcli & Murdock, 1976). The ex-gaussian density can capture a very large number of shapes from almost symmetric (like the normal density) to very asymmetric (like the exponential density, which is shown in Figure 2). The density function is a useful way of characterizing RTs because it is an intuitive way to think about how likely dierent RTs are to be observed. The shape of the density function can be useful in discriminating between some classes of random variables, but it should be noted that many unimodal, positively skewed distributions exist, and many are exible enough that they can look very similar to each other. This is demonstrated by the close correspondence between the ex-gaussian and Wald densities in the gure. I will talk again about similarities between distributions at the close of this chapter.

8 RT Analysis 8 Distribution Function The cumulative distribution function (CDF) F (t) of the same ex-gaussian and Wald random variables are shown in Figure 1, center panel. The x-axis gives the possible values of RT, and the y-axis gives the values of the CDF F (t). For a particular point t along the x-axis, the value of F (t) gives the probability of observing an RT less than the value t: F (t) = P (RT < t): The CDF is found from the density function by integration from?1 to the point t: F (t) = Z t?1 f(u)du: Any positive function F (t) that is nondecreasing and has the properties that lim t!?1 F (t) = 0 and lim t!1 F (t) = 1, is a CDF. The CDF gives the percentile ranks of each possible value of RT. In Figure 1, for example, the median RT can been found by drawing a horizontal line from the.50 point on the y-axis and dropping a vertical line to the x-axis at the point of intersection with the desired curve F (t). This relationship between the CDF and the quantiles of the distribution is particularly useful for simulating dierent random variables. To simulate a random variable RT with a known CDF F (t), select a point u on the y-axis at random (using a uniform [0; 1] generator), and then compute the corresponding quantile RT = F?1 (u). Inverting F may require numerical methods, but is typically straightforward. Unlike the density function, the CDF is not as useful in distinguishing between dierent kinds of random variables. Most CDFs are S-shaped; all distributions increase from 0 to 1. Therefore all CDFs tend to look similar regardless of the random variables they describe. For example, in Figure 1, the Wald and ex-gaussian CDFs are practically identical. Survivor and Hazard Functions The survivor and hazard functions are two alternative functions that characterize a random variable. These functions arise frequently in reliability and survival analysis (when machines break or people die), hence the choice of vocabulary.

9 RT Analysis 9 The survivor function F (t) is the probability that the \lifetime" of an object is at least t, that is, the probability that failure occurs after t. In terms of RT, F (t) = P (RT > t) = 1? F (t) : the survivor function is simply one minus the CDF. The hazard function h(t) gives the likelihood that an event will occur in the next small interval dt in time, given that it has not occurred before that point in time. From the denition of a conditional probability, h(t) = lim dt!0 P (t RT t + dtjrt t)=dt = f(t) F (t) : The hazard function for the ex-gaussian and Wald distributions are shown in Figure 1, bottom panel. When F (t) is dierentiable, the hazard function can be expressed as a function of the survivor function: h(t) =? d ln F (t); dt where ln indicates the natural logarithm. The function? ln F (t) is called the log survivor function or integrated hazard function, and it can be useful for distinguishing between dierent distributions. We will later make use of it to answer questions about certain properties of cognitive processes. The density function can also be expressed in terms of the hazard function: Z t f(t) = h(t) exp? h(u)du :?1 Some attention has been paid to hazard function analysis in RT work, because the hazard function can frequently be very useful in discriminating between dierent random variables (Ashby, Tein & Balakrishnan, 1993; Bloxom, 1984; Colonius, 1988; Luce, 1986; Maddox, Ashby & Gottlob, 1998; Thomas, 1971). Although it is very dicult to discriminate between the ex-gaussian and Wald densities shown in the top panel of Figure 1, the dierences between the two variables become clear by examining the hazard functions in the bottom panel. Whereas the ex-gaussian hazard function increases to an asymptote, the Wald hazard function is nonmonotonic, increasing and then decreasing. It has been suggested, therefore, that hazard function estimation may provide a way to

10 RT Analysis 10 identify RT distributions, and hence the process underlying the execution of an RT task (Burbeck & Luce, 1982). If we can determine how RTs are distributed, then we have gone a long way toward isolating the process responsible for generating the RTs. In practice, then, we might try to estimate one or more of these functions in an attempt to identify or rule out various distributional forms for consideration. In the next section, we will talk about problems associated with estimating these functions. The problem of estimation is not a trivial one, however, and I must begin by discussing estimation in general. I will then talk about estimation of parameters, and then estimation of functional forms. Estimation The goal of estimation is to determine from RT data the properties of the distribution from which the data were sampled. We may wish to know only gross properties such as the mean or skewness, or we may hope to determine the exact functional form of the distribution. The most common sort of statistical analysis involves inferences about the mean and variance of the population distribution. Given a particular sample of size n, we can attempt to estimate the central tendency and dispersion of the population, perhaps using the sample mean X and variance s 2. Or, we might try to estimate the shape of the distribution itself, perhaps by constructing a histogram indicating the relative frequency of each observation in the sample. Both of these estimation problems are commonly encountered, and they are not necessarily separate problems. We will discuss each in this section. It is important to remember that any estimate, whether of a parameter or of a function, is a random variable. Associated with it is some degree of variation and some (often unknown) distributional form. The goal of estimation, therefore, is not just estimating values of parameters, but the estimation of those parameters in such a way that the distributions of those estimates have desirable properties. For example, the sample mean X is an estimate of a parameter,, the population mean. The estimate X is a random variable, because its value will change for dierent samples, even when those samples are taken from the same population. There is a distribution associated with X, which, by the Central Limit Theorem, we know to be approximately normal. The mean of this

11 RT Analysis 11 distribution is, the mean of the population from which the sample was drawn. We know something also about the variance of the distribution of X. As long as the sample is iid, the variance of the distribution is 2 =n, the variance of the population from which the sample is drawn divided by the size of the sample n. I will begin by discussing problems of parameter estimation, including the mean and variance of the RT distribution, and also the more explicit problem of estimating the parameters of a theoretical RT distribution. I will then discuss the problem of distribution estimation, including the estimation of the density, distribution, and hazard functions. Throughout the rest of this chapter, I will refer to the data presented in Table 1 to motivate and illustrate the discussion. These data were collected in a simple detection experiment, roughly designed after an experiment presented by Townsend and Nozawa (1995), in which one, two or no dots appeared in two locations on a computer screen. The observer's task was to press one key as soon as a dot was detected, and the other key if no dots were detected. Each dot could be either large or small. Only the target (dot) present data are provided in the table, and error RTs are not included. This experiment is an example of what Townsend and Nozawa have called the double factorial paradigm, which has proved to be very important in examining hypotheses about processing information from multiple sources of information, as we will see later. Properties of Estimators Consider an iid sample fx 1 ; X 2 ; : : : ; X n g of size n from a population with mean and variance 2. The sample mean X = P n i=1 X i =n is an estimate of the parameter. The sample mean has a number of desirable properties as an estimate of : it is (1) unbiased, (2) consistent, (3) ecient, and, if the variable X is normally distributed, (4) a maximum likelihood estimate. If an estimator is unbiased, then its expected value (mean) is equal to the parameter being estimated. For the sample mean, the expected value of X is equal to, the parameter that X estimates. In other words, on average, the sample mean will equal the population mean. More generally, suppose that we wish to estimate a parameter with

12 RT Analysis 12 the estimator ^. The estimator ^ is unbiased if E(^) = ^ = ; where E is the expected value operator, dened as E(X) = Z 1?1 xf(x)dx: If an estimator is also consistent, then as the sample size n grows, the probability that it diers from the estimated parameter shrinks to zero. We sometimes call this property \convergence in probability." For the sample mean X, this property takes the form of the Law of Large Numbers. As n grows very large, the probability that X diers from shrinks to zero. More generally, for a consistent estimator ^, lim P (j^? j ) = 0; n!1 for any > 0, no matter how small. For practical purposes, this means that the accuracy of the estimator ^ can be improved by increasing sample size n. Consistency is therefore a very important property. We might wish to sacrice unbiasedness for consistency, as long as the variance of the estimator decreases fairly rapidly with n. A property closely related to consistency is that of asymptotic unbiasedness, meaning that as n grows, the expected value of the estimator approaches the value of the estimated parameter. Although this may seem just the same as consistency, it is not. It is a stronger property, meaning that the conditions under which it holds are more stringent than those for consistency. Therefore, an estimator may be consistent but asymptotically biased. The third property, eciency, refers to the variance of the estimator. Because the estimate of a parameter is a random variable, we would like for the variance of that variable to be relatively small. So, if the estimator is unbiased, we can be fairly certain that it does not vary too much from the true value of the parameter. The sample mean is ecient, which means that it has a smaller variance relative to any other estimate of that we might choose, such as the median or the mode. The fourth property, maximum likelihood, has to do with the probability of the observed sample. That is, there is some probability associated with having sampled exactly the observations that we obtained: an n-fold joint probability distribution. To use a

13 RT Analysis 13 simplistic perspective, if we observe a particular sample, then it must have been highly probable, and so we should choose the estimates of the population parameters that make this probability as large as possible. If X is normally distributed, then the sample mean is a maximum likelihood estimator of. Note that maximum likelihood estimators are not guaranteed to be unique or to exist at all. I will discuss maximum likelihood estimation in some detail later. Although we have been talking about properties of point estimators, that is, estimators of a single population parameter, it is important to remember that these properties also hold for estimates of functions, such as densities or CDFs. For example, suppose that a density function were to be estimated by a relative frequency histogram of a sample. The height of the histogram at every point is a random variable, and the shape of the histogram will change with every new sample. Therefore, we can characterize empirical estimates of the density, distribution, survivor and hazard functions as unbiased, consistent, ecient, and so forth. Parameter Estimation Mean and Variance By far the most common parameters estimated in an RT analysis are the mean and variance ( and 2 ) of the RT distribution. However, because of the asymmetric shape of the RT distribution, it is important to recognize that these parameters are not necessarily (perhaps not even frequently) the best parameters to characterize the RT distribution. Because of the RT distribution's skewness (see Figure 1), the mean does not represent the most typical or likely RT. It is pulled upward, in the direction of the skew. A similar problem exists with the variance 2. The large upper tail in the RT distribution has the eect of creating \outliers," values in the sample that are much longer than the majority of the observations. Outliers are a problem in all areas of statistical analysis, but the unusual aspect of outliers in RT data is that they potentially derive from the process of interest. That is, they are not necessarily outliers in the sense of contaminations of the data. Because the sample variance is greatly increased by such outliers, the power of the statistical tests to be performed on the data is greatly reduced.

14 RT Analysis 14 As an example, consider the RTs shown in the condition labeled \Left(s)" in Table 1. These are the detection RTs for when a single, small dot appeared in the left position. These data are positively skewed: although most of the observations fall between 300 ms and 600 ms, a small proportion of the sample extends as high as 1029 ms. The sample mean is X = 550:60 ms and the sample standard deviation s = 164:58 ms. Let's change the value of the slowest RT, 1029 ms, to 2060 ms { a value that might typically be observed in such an experiment. We nd now that the sample mean X = 573:50 ms and the sample standard deviation s = 270:42 ms. As a result of changing a single observation the sample mean increased by 23 ms (almost a detectable eect, in some RT experiments) and the sample standard deviation increased by over 60%. We say that neither the mean nor the variance 2 are robust parameters, meaning that very small changes in the shape of the distribution, like the change of a single observation, can produce very large changes in the values of these parameters. In a practical sense, this means that condence intervals constructed for these parameters are likely to be too large and placed incorrectly, reducing the power of hypothesis tests involving the mean RT (Wilcox, 1998). This is particularly a concern for RT data in which not only large degrees of skew are to be expected, but a good number of outliers (either extremely long or short RTs) are to be expected as well. An alternative is to estimate parameters that aren't as sensitive to outliers or skew, such as the median or interquartile range. Using the same example as in the previous paragraph, the original sample has median Md = 485 ms and interquartile range IQR = 169 ms. After doubling the slowest RT to 2060 ms, these estimates are unchanged. Unfortunately, the standard errors of X and s 2 are typically considerably smaller than the standard errors of Md and IQR (Stuart & Ord, 1999). For example, while the standard error of X is = p n, if the sample is taken from a normal population, the standard error of Md is approximately 1:25= p n. Although the sampling distributions of X, s 2, Md and IQR are asymptotically normal, the sample sizes required to approximate normality are much larger for Md and IQR than for X and s 2. Furthermore, whereas the sample mean X is an unbiased estimator of, the sample median Md is a biased estimator of the population median when the population is skewed (Miller, 1988). 2 These factors 2 For even very skewed distributions, median bias is generally less than 10 ms for samples of size 25 or

15 RT Analysis 15 have prevented widespread use of the statistics Md and IQR, despite the fact that they are probably better for characterizing central tendency and dispersion for RT data than are X and s 2. Using Monte Carlo simulations, Ratcli (1993) investigated a number of methods of RT data treatment that reduce or eliminate the eects of outliers on the mean and variance. These included cuto values, above and below which RTs are eliminated from the sample, and transformations such as the inverse and logarithm. For each of these strategies, he computed power and the probability of Type I errors for analyses of variance, with and without outliers mixed into the data. While no method had strong eects on the number of Type I errors, the method chosen had strong eects on power. Using Md as a measure of central tendency generally resulted in lower power than using cutos or transformations, probably because of the greater variance and bias of Md. Fixed cutos (e.g., 2000 or 2500 ms) maintained the highest power. The use of xed cutos, however, is highly dependent on the range of the data in dierent experimental conditions; a 2000 ms cuto in one experiment might eliminate half the observations in another experiment. Thus there is no hard and fast rule that could be used to establish cutos. Because of this problem, it is common to nd examples in which cutos are based on the sample standard deviation. For example, we might eliminate all observations greater than 3 standard deviations above the mean. Unfortunately, Ratcli (1993) found that basing cutos on the standard deviation could have disastrous eects on power, depending on whether the experimental factors had their eects on the fast or slow RTs. Ratcli instead recommended exploring a range of potential cuto values for dierent experimental conditions. However, Ulrich and Miller (1994) have noted that using cutos can introduce asymmetric biases into statistics such as the sample mean, median, and standard deviation, and caution strongly against the use of cutos without consideration of these eects. Van Selst and Jolicoeur (1994) present a procedure that produces a uniform bias across sample sizes, thus minimizing the potential for artifactual dierences between conditions. Unfortunately, even Van Selst and Jolicoeur's method produces a highly biased estimate higher. When one or more groups has sample size less than 25 the median should not be used, as the bias dierence between the groups could introduce an artifactual eect. See Miller, 1988, for more details.

16 RT Analysis 16 (as great as 30 ms too small for the distributions they examined), and this could also result in signicant artifacts, especially in the presence of oor or ceiling eects. The next most powerful approach to minimizing the eects of outliers was the inverse transformation, which Ratcli (1993) recommended as a way to verify the appropriateness of the selected cutos if cutos were used. Transforming RTs to speed, 1/RT, reduces the eect of slow outliers and maintains good power. In the example above taken from Table 1, the mean transformed RT is X = 1:947 10?3 = ms with standard deviation s = 0:466 10?3 = ms. If the slowest RT is doubled, the mean transformed RT is X = 1:936 10?3 = ms with standard deviation s = 0:494 10?3 = ms; there is some eect of the long outlier, but it is greatly reduced relative to the eect on the mean RT. To see how outliers reduce power, consider the two conditions in the Table denoted Both(ss) and Right(s). The nature of the experiment was such that we might have expected that RTs in the Right(s) conditions would be slower than RTs in the Both(ss) condition. This is because, essentially, twice the amount of information was available to the observer in the Both(ss) condition, resulting in a faster accumulation of evidence that targets were present. As predicted, the mean RT for Both(ss) responses is 524:61 ms and the mean RT for Right(s) responses is 563:70 ms. Unfortunately, this dierence (39 ms) does not reach statistical signicance (t(88) = 1:3734, p > :08). If we consider instead the response rates (1:85 10?3 = ms and 2:00 10?3 =ms for Both(ss) and Right(s), respectively), then the rate dierence (?:15 10?3 =ms) is signicant (t(88) = 1:94, p < :05). 3 A large statistical literature has developed over the past few decades addressing robust statistics and how to reduce the eects of outliers or skewed distributions (Barnett & Lewis, 1994). Not many psychology researchers are considering these alternative techniques for their data. Wilcox (1997, 1998) has made a special eort to bring robust analyses to the attention of experimental psychologists, and hopefully, as more and more statistical packages incorporate robust techniques, we will see more interest in these alternatives in the years to come. In the meantime, however, the best approach to dealing with outliers in 3 Note that it is also possible to eliminate signicant eects by transformation. Also note that transformed RTs will not necessarily be useful for testing some kinds of model predictions, e.g., serial stage models in which durations are summed. I am only advocating the use of transformed RTs for null hypothesis testing, not for verifying model predictions.

17 RT Analysis 17 RT data is probably the inverse transformation, which gives the greatest power. Cutos should be avoided, unless one is willing to undertake either Ulrich and Miller's (1994) or Van Selst and Jolicoeur's (1994) procedures. As with any procedures that involve removing data from a sample, when attempting to employ cutos to reduce the eects of outliers, statistical results should be presented for both the complete and trimmed data sets, so that readers are aware of potential artifactual results. The ex-gaussian Parameters Because outliers are not only bothersome but also potentially interesting in the context of RTs, some have argued that the ex-gaussian distribution could be used as a parametric (although atheoretical) estimate of RT distributions (Ratcli, 1979; Ratcli & Murdock, 1976; Heathcote, Popiel & Mewhort, 1991) without excluding any data that are suspiciously slow or fast. The ex-gaussian distribution, the distribution of the sum of a normal and an exponentially distributed variable, has three parameters. The normal component has parameters and 2, the normal mean and variance. The exponential component has parameter, the exponential mean. Figure 2 shows the ex-gaussian density presented in Figure 1, top panel, together with its component normal and exponential densities. (The exponential has been shifted from its minimum, 0, to demonstrate the relation between the tails of the exponential and ex-gaussian densities.) The normal distribution determines the leading edge of the ex-gaussian density, and the exponential distribution determines skewness, or the height of the positive tail. The mean of the ex-gaussian is +, and its variance is Rather than discussing RT means and variances, estimates of the parameters, 2, and could be used to characterize RT data and isolate the eects of experimental variables either in the slow or fast RTs (Heathcote et al., 1991; Hockley, 1982). I estimated these parameters for the data from the Both(ss) and Right(s) condition using maximum likelihood. For Both(ss), ^ = 390:47 ms, ^ = 16:60 ms, and ^ = 134:09 ms. For Right(s), ^ = 458:64 ms, ^ = 47:54 ms, and ^ = 105:06 ms. Several routines are publicly available to assist in performing these computations (Cousineau & Larochelle, 1997; Dawson, 1988; Heathcote, 1996), and MATLAB routines are provided in the Appendix.

18 RT Analysis 18 Given the estimated ex-gaussian parameters, we might want to infer that the presence of two dots decreased and and increased. However, because the distributions of the estimates of these parameters are unknown, inferential statistical procedures are dicult. One approach is to estimate the standard errors of the estimates by \bootstrapping" the samples (Efron, 1979). This is a simple procedure in which bootstrapped samples of size n are obtained from the original sample by selecting n observations from the sample with replacement. The ex-gaussian parameters are then estimated from each bootstrapped sample, and the standard deviation of those estimates is a (pretty good) estimate of the standard error. I simulated 100 bootstrapped samples for each condition and computed the standard errors of ^, ^ and ^ to be 14:80 ms, 13:09 ms, and 22:37 ms, respectively, for the Both(ss) condition, and 25:00 ms, 23:02 ms, and 31:79 ms, respectively, for the Right(s) condition. Given these standard errors, we can argue that is greater for the Right(s) than for the Both(ss) condition. However, we cannot conclude that any dierences exist between or for the two conditions because the variance of the estimates are too large. This nding is consistent with earlier work showing the variance of the estimates in and to be quite large for smaller samples (Van Zandt, 2000). In sum, the ex-gaussian characterization of RTs could potentially be quite useful in skirting problems of skewness and outliers. Unfortunately, the sampling distributions of ^, ^ and ^ cannot be determined explicitly, and their distributions also depend on the underlying (and unknown) RT distribution. It may not be possible, therefore, to argue conclusively about the eects that experimental manipulations have on these parameters. An additional problem arises when one considers that, despite the utility of the ex-gaussian distribution, RTs are not generally distributed as ex-gaussians (Burbeck & Luce, 1982; Luce, 1986; Van Zandt, 2001). Because the ex-gaussian is an atheoretical model of the RT distribution, it is dicult, if not impossible, to attribute psychological meaning to changes in the dierent parameters. As Ratcli (1993) argues, the use of the ex-gaussian parameters to characterize RT distributions might not be very useful in the absence of a model explaining the processes that generated the RTs in the rst place.

19 RT Analysis 19 Nonparametric Function Estimation Some estimation procedures, such as least-squares minimization and maximum likelihood estimation, insure that estimates are unbiased, consistent, ecient, and so forth, given certain constraints on the sample. Unfortunately, these constraints are rarely satised when dealing with RT distributions. Furthermore, when estimating density functions, the extent of bias depends on the true form of the underlying population distribution, which is unknown. Therefore, when estimating RT densities, the extent of error is also unknown. There are a number of issues that bear on the estimation procedure. The issue of primary importance is whether the analysis is model-driven. That is, has a process been specied that states how RTs should be distributed and explains the relationship between the physical parameters of the experiment and the theoretical parameters of the RT distribution? If a model of this degree of precision has been specied, then a parametric procedure will be used to recover the theoretical parameters of the RT distribution. If not, then nonparametric procedures will probably be more appropriate. Quantiles and the CDF The CDF is the easiest of the functional forms to estimate, because an unbiased, consistent, maximum likelihood estimate for the CDF exists in the form of the cumulative relative frequency distribution, or EDF (empirical distribution function). That is, ^F (t) = [number of observations less than or equal to t]=n: (1) The EDF is an estimator of the percentile ranks of the possible RTs that might be observed, and is asymptotically normal at every point. This means that at every point t, ^F (t) will be normally distributed around the value of the true CDF F (t). Because the EDF ^F (t) is an unbiased estimate of the CDF, it should be noted that 1? ^F (t) is therefore an unbiased estimate of the survivor function. We will use the survivor function estimate later in this chapter. It is often useful to compute the estimates ^F?1 (p) of the quantiles of an RT distribution (e.g., Logan, 1992; Van Zandt, 2000; Van Zandt, Colonius & Proctor, 2000). The CDF F (t) can then be estimated by plotting p for a number of estimates t p = ^F?1 (p). To

20 RT Analysis 20 estimate the p th quantile t p, where P (RT t p ) = p, the RTs are ordered from smallest to largest. The simplest way to estimate the p th quantile is to nd the np th observation in the sample, if np is an integer. If np is not an integer, then an average of the [np] th and [np] + 1 th observation is computed, where [np] indicates the integer part of np. Typically the midpoint of [np] and [np] + 1 is used, but other weighting schemes can be used as well (Davis & Steinberg, 1983). As an example, consider the data in conditions Left(s) in Table 1. We will estimate the median, the p = :50 quantile. There are n = 45 observations in this sample, so np = 22:5. The [np] th observation is (counting frequencies from the smallest observed RT) RT 17 = 480 ms, and the [np] + 1 th observation is RT 18 = 485 ms. The average of these two observations is 482:5 ms, therefore ^F?1 (:5) = Md = 482:5 ms. Like the EDF, quantile estimates are asymptotically normal and unbiased. These characteristics of the EDF and quantile estimates makes inferential statistical analyses easy to perform, because the sampling distributions of quantiles and percentile ranks are known exactly (Stuart & Ord, 1999). A widely used method for estimating quantiles of RT distributions is called \Vincentizing" (Ratcli, 1979). To describe this procedure, suppose that q quantiles are to be estimated from a sample of n observations. We will call the estimated quantiles \vincentiles" (Heathcote, Brown & Mewhort, 2000) to distinguish them from the quantiles estimated by the procedure described above. It is typically assumed (Ratcli, 1979) that the vincentiles are, like quantiles, evenly spaced across the data, separating the data into q equally probable bins. Note, however, that the vincentiles are quantile midpoints, so that q + 1 bins are obtained through Vincentizing. For instance, suppose that 10 deciles were computed by the standard quantile estimation procedure described above. These estimates would correspond to the 10 th, 20 th,... percentiles. Now suppose that 10 vincentiles were computed. We would assume that these vincentiles were located at approximately the 5 th, 15 th, 25 th... percentiles { the midpoints of the decile ranges. Thus, the vincentiles separate the sample into q + 1 groups, and the relative frequencies of the middle groups are 1=q, and the relative frequencies of the slowest and fastest groups are 1=2q. Consider the RTs for condition Left(s) shown in Table 1. There are n = 45 observations in this condition, and we will estimate q = 5 vincentiles. To do this, we rst make q copies

21 RT Analysis 21 of each observation. Then, starting from the smallest RTs, we begin averaging the duplicated order statistics in groups of n. The rst vincentile for this sample is therefore (5)(350 ms) + (5)(365 ms) + (5)(375 ms) + (5)(409 ms)+ (15)(422 ms) + (5)(423 ms) + (5)(426 ms) V 1 = 45 = 401:56 ms: The second vincentile is the average of the next n = 45 observations: (5)(428 ms) + (5)(429 ms) + (5)(449 ms)+ (15)(450 ms) + (5)(454 ms) + (10)(457 ms) V 2 = 45 = 447:11 ms: The averaging procedure is continued throughout the remaining observations in the duplicated sample, yielding V 3 = 490:67 ms, V 4 = 578:22 ms, and V 5 = 835:44 ms. The vincentiles V 1, V 2, etc. should estimate the 10 th, 30 th, 50 th, 70 th and 90 th quantiles. It is informative to compare the values of the vincentiles to the estimates of the quantiles found using the standard quantile estimation procedure. For p = :10, [np] = 4 and [np] + 1 = 5. The average of the fourth and fth order statistics (409 and 422 ms) is ^t :1 = 415:5 ms. For p = :30, [np] = 13 and [np] + 1 = 14. The average of the 13 th and 14 th order statistics (450 and 450 ms) is ^t :3 = 450 ms. Continuing in this way, we nd ^t :5 = 482:50, ^t :7 = 567:00 ms, and ^t :9 = 810 ms. Although ^t :3 = 450 ms is close to the value V 2 = 447:11, none of the other vincentiles are as close to the estimated quantiles. Neither do the vincentiles divide the sample into equally frequent groups. It turns out that the only time that vincentiles are estimates of the quantiles to which they are supposed to correspond is when the sample is drawn from a symmetric distribution (Heathcote et al., 2000). Simulations show that the vincentiles do not correspond to known quantiles for nonsymmetric distributions, and therefore they are not useful as a way to estimate CDFs (Van Zandt, 2000). The major attraction of the Vincentizing procedure is that it allows for averaging of RT distributions across subjects in an experiment (Ratcli, 1979). This is useful when small sample sizes prevent accurate estimates of RT distributions for individual subjects

22 RT Analysis 22 (although CDFs and quantiles can be accurately estimated with as few as 50 observations, see Van Zandt, 2000). To average RT distributions across subjects, the vincentiles are computed for each subject's data, and then the vincentiles are averaged. Because each vincentile corresponds (presumably) to a particular percentile rank, the resulting averages can then be used to construct the EDF. However, because the vincentiles do not typically correspond to known percentiles, averaging should be performed using standard quantiles rather than the vincentiles. There seem to be no particular benets to using vincentiles instead of quantiles, and using vincentiles may introduce error, depending on the goals of the analysis. The Density Function Probably the most popular and easiest method for density estimation is the simple histogram. Observations are binned and the relative frequency of the number of observations within each bin are used as a density estimate. Unfortunately, there is no best estimator for the density function as there was for the CDF. There is a large area in statistics devoted to density function estimation, only a little of which we will be able to touch on in this chapter. See Silverman (1986) for a basic and accessible treatment of this problem. The issue of unbiased and consistent estimators for density functions is a tricky one. To illustrate why, let's consider a rather large class of density function estimators called general weight function estimators. The simple histogram is one member of this class. Associated with each of the estimators in this class is a parameter h n, sometimes called a smoothing or bandwidth parameter, which depends on the sample size n. In the case of a simple histogram, h n would be the width of the bins. In general, the larger n is the smaller h n needs to be. Under some fairly general constraints, to be sure that the estimate is asymptotically unbiased, it must be that lim n!1 h n = 0. But even if this holds, to make sure the estimator is consistent, it must also be that lim n!1 nh n = 1. So h n must go to zero as n gets large, but it can't go to zero too quickly. For any particular density estimator, bias will be a function of the sample size n and the true underlying density, which in practice is always unknown. Furthermore, RT densities have a special problem in

23 RT Analysis 23 their skew, which makes some potential estimators unsuitable. It is important to realize that a density estimate is probably biased, the degree of bias will be unobservable, and the bias will not necessarily get smaller with increases in sample size. For this chapter, I present two types of density estimators. The simple and very popular histogram estimators are easy to compute but can be inaccurate. Kernel estimators are more accurate but they are a little more dicult to understand and require more computation. These two types of estimators barely begin to cover the statistical literature on nonparametric density estimators. There are more accurate estimators than the ones presented here. However, it is especially important to remember that density estimation is an exercise in descriptive statistics only. Under no circumstances should parameter estimation (i.e. model tting) depend in some way on a density estimate. Appropriate parameter estimation methods will be discussed shortly. For graphical purposes, the kernel estimator described below should suce. Histogram Estimators Histogram estimators are perhaps the most well-known density estimators. Construction of the histogram estimate is fairly simple, and involves selecting r \bins" with bin boundaries ft 0 ; t 1 ; t 2 ; : : : ; t r g along the time axis. The estimate is a function of the number of observations falling in each bin: ^f(t) = number of observations in bin i nh i ; t i?1 t < t i ; where h i = t i? t i?1 is the width of the i th bin. The bins can be of xed width, or they can vary according to the density of the observations along the time axis. For xed-width estimators, an origin t 0 and a bin width (bandwidth) h n are selected, and the bin boundaries are computed as ft 0 ; t 0 + h n ; t 0 + 2h n ; : : :g. The histogram estimate for the Left(s) data is shown in Figure 3, using t 0 = 200 ms and h n = 50 ms. Unfortunately, there is no automatic way to select t 0 or h n, and no systematic way to adjust h n with sample size to ensure asymptotic unbiasedness or consistency of the estimator. An appropriate t 0 and h n must be selected after inspecting the sample for which the density is to be estimated.

24 RT Analysis 24 For this reason, variable-width histograms that specify the frequency of observations within each bin are frequently used in RT analysis. One such estimator is based on the vincentiles (Ratcli, 1979). For this estimator, the vincentiles are assumed to divide the sample into equally-probable intervals (with the exception of the fastest and slowest bins) as described above. The height of the estimate for each bin is then computed so that the area of the interval is equal to 1/q (the number of vincentiles). I previously investigated the accuracy of histogram estimators, both xed and variable widths, for a number of dierent RT models (Van Zandt, 2000). Unfortunately, variable-width estimators based on quantiles and vincentiles were highly inaccurate and quite variable even for very large sample sizes. As I demonstrated above, the vincentiles do not divide the sample into groups of equal frequency, so the heights of the density estimate computed under this assumption are incorrect. The xed-width histogram performed better than density estimates based on quantiles, but because of the lack of an algorithm to adjust h n with increasing sample size, accuracy did not improve with increases in sample size. Therefore, the histogram estimates should not be relied upon for any serious RT distributional analysis. Kernel Estimators The kernel estimator is also a member of the class of weight function estimators. A \kernel" is simply a function K(x) that is integrated or, in the discrete case, summed, over x. To estimate the density at a particular point t, the kernel is summed over all of the observations T i in the sample. Another way of looking at the kernel estimate is that every point on the estimate is a weighted average of all of the observations in the sample. The general form of the kernel estimate is ^f(t) = 1 nh n nx i=1 Ti? t K : (2) Notice the presence of the bandwidth parameter h n in the denominator of the kernel variable. The larger h n, the smoother the estimate will be: large values of h n will tend to minimize the deviations in the sample from the point t. The kernel K(x) is itself a density function, a positive function that integrates to 1 over all x, and is typically symmetric. The kernel estimate of the density at the point t, then, is found by centering the kernel over the h n

GIST 4302/5302: Spatial Analysis and Modeling

GIST 4302/5302: Spatial Analysis and Modeling GIST 4302/5302: Spatial Analysis and Modeling Basics of Statistics Guofeng Cao www.myweb.ttu.edu/gucao Department of Geosciences Texas Tech University guofeng.cao@ttu.edu Spring 2015 Outline of This Week

More information

Maximum Likelihood Estimation

Maximum Likelihood Estimation Connexions module: m11446 1 Maximum Likelihood Estimation Clayton Scott Robert Nowak This work is produced by The Connexions Project and licensed under the Creative Commons Attribution License Abstract

More information

The Bias-Variance dilemma of the Monte Carlo. method. Technion - Israel Institute of Technology, Technion City, Haifa 32000, Israel

The Bias-Variance dilemma of the Monte Carlo. method. Technion - Israel Institute of Technology, Technion City, Haifa 32000, Israel The Bias-Variance dilemma of the Monte Carlo method Zlochin Mark 1 and Yoram Baram 1 Technion - Israel Institute of Technology, Technion City, Haifa 32000, Israel fzmark,baramg@cs.technion.ac.il Abstract.

More information

Chapter 1 Statistical Inference

Chapter 1 Statistical Inference Chapter 1 Statistical Inference causal inference To infer causality, you need a randomized experiment (or a huge observational study and lots of outside information). inference to populations Generalizations

More information

p(z)

p(z) Chapter Statistics. Introduction This lecture is a quick review of basic statistical concepts; probabilities, mean, variance, covariance, correlation, linear regression, probability density functions and

More information

Department of Mathematical Sciences, Norwegian University of Science and Technology, Trondheim

Department of Mathematical Sciences, Norwegian University of Science and Technology, Trondheim Tests for trend in more than one repairable system. Jan Terje Kvaly Department of Mathematical Sciences, Norwegian University of Science and Technology, Trondheim ABSTRACT: If failure time data from several

More information

Power Calculations for Preclinical Studies Using a K-Sample Rank Test and the Lehmann Alternative Hypothesis

Power Calculations for Preclinical Studies Using a K-Sample Rank Test and the Lehmann Alternative Hypothesis Power Calculations for Preclinical Studies Using a K-Sample Rank Test and the Lehmann Alternative Hypothesis Glenn Heller Department of Epidemiology and Biostatistics, Memorial Sloan-Kettering Cancer Center,

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions

SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions SYSM 6303: Quantitative Introduction to Risk and Uncertainty in Business Lecture 4: Fitting Data to Distributions M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu

More information

Lecture 21. Hypothesis Testing II

Lecture 21. Hypothesis Testing II Lecture 21. Hypothesis Testing II December 7, 2011 In the previous lecture, we dened a few key concepts of hypothesis testing and introduced the framework for parametric hypothesis testing. In the parametric

More information

CHAPTER 4 VARIABILITY ANALYSES. Chapter 3 introduced the mode, median, and mean as tools for summarizing the

CHAPTER 4 VARIABILITY ANALYSES. Chapter 3 introduced the mode, median, and mean as tools for summarizing the CHAPTER 4 VARIABILITY ANALYSES Chapter 3 introduced the mode, median, and mean as tools for summarizing the information provided in an distribution of data. Measures of central tendency are often useful

More information

A mathematical model of social conformity as a function of group size

A mathematical model of social conformity as a function of group size A mathematical model of social conformity as a function of group size To Chin, Yu April 20, 2013 This paper aims to model mathematically the dependence of conformity on majority size with the intention

More information

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances

Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances Advances in Decision Sciences Volume 211, Article ID 74858, 8 pages doi:1.1155/211/74858 Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances David Allingham 1 andj.c.w.rayner

More information

Practice Problems Section Problems

Practice Problems Section Problems Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,

More information

Confidence Intervals. Confidence interval for sample mean. Confidence interval for sample mean. Confidence interval for sample mean

Confidence Intervals. Confidence interval for sample mean. Confidence interval for sample mean. Confidence interval for sample mean Confidence Intervals Confidence interval for sample mean The CLT tells us: as the sample size n increases, the sample mean is approximately Normal with mean and standard deviation Thus, we have a standard

More information

Monte Carlo Studies. The response in a Monte Carlo study is a random variable.

Monte Carlo Studies. The response in a Monte Carlo study is a random variable. Monte Carlo Studies The response in a Monte Carlo study is a random variable. The response in a Monte Carlo study has a variance that comes from the variance of the stochastic elements in the data-generating

More information

Physics 509: Bootstrap and Robust Parameter Estimation

Physics 509: Bootstrap and Robust Parameter Estimation Physics 509: Bootstrap and Robust Parameter Estimation Scott Oser Lecture #20 Physics 509 1 Nonparametric parameter estimation Question: what error estimate should you assign to the slope and intercept

More information

MATH4427 Notebook 4 Fall Semester 2017/2018

MATH4427 Notebook 4 Fall Semester 2017/2018 MATH4427 Notebook 4 Fall Semester 2017/2018 prepared by Professor Jenny Baglivo c Copyright 2009-2018 by Jenny A. Baglivo. All Rights Reserved. 4 MATH4427 Notebook 4 3 4.1 K th Order Statistics and Their

More information

Statistical Inference

Statistical Inference Statistical Inference Bernhard Klingenberg Institute of Statistics Graz University of Technology Steyrergasse 17/IV, 8010 Graz www.statistics.tugraz.at February 12, 2008 Outline Estimation: Review of concepts

More information

PRIME GENERATING LUCAS SEQUENCES

PRIME GENERATING LUCAS SEQUENCES PRIME GENERATING LUCAS SEQUENCES PAUL LIU & RON ESTRIN Science One Program The University of British Columbia Vancouver, Canada April 011 1 PRIME GENERATING LUCAS SEQUENCES Abstract. The distribution of

More information

ADJUSTED POWER ESTIMATES IN. Ji Zhang. Biostatistics and Research Data Systems. Merck Research Laboratories. Rahway, NJ

ADJUSTED POWER ESTIMATES IN. Ji Zhang. Biostatistics and Research Data Systems. Merck Research Laboratories. Rahway, NJ ADJUSTED POWER ESTIMATES IN MONTE CARLO EXPERIMENTS Ji Zhang Biostatistics and Research Data Systems Merck Research Laboratories Rahway, NJ 07065-0914 and Dennis D. Boos Department of Statistics, North

More information

Describing distributions with numbers

Describing distributions with numbers Describing distributions with numbers A large number or numerical methods are available for describing quantitative data sets. Most of these methods measure one of two data characteristics: The central

More information

in the company. Hence, we need to collect a sample which is representative of the entire population. In order for the sample to faithfully represent t

in the company. Hence, we need to collect a sample which is representative of the entire population. In order for the sample to faithfully represent t 10.001: Data Visualization and Elementary Statistical Analysis R. Sureshkumar January 15, 1997 Statistics deals with the collection and the analysis of data in the presence of variability. Variability

More information

A nonparametric two-sample wald test of equality of variances

A nonparametric two-sample wald test of equality of variances University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 211 A nonparametric two-sample wald test of equality of variances David

More information

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS

ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS ACTEX CAS EXAM 3 STUDY GUIDE FOR MATHEMATICAL STATISTICS TABLE OF CONTENTS INTRODUCTORY NOTE NOTES AND PROBLEM SETS Section 1 - Point Estimation 1 Problem Set 1 15 Section 2 - Confidence Intervals and

More information

In Chapter 2, some concepts from the robustness literature were introduced. An important concept was the inuence function. In the present chapter, the

In Chapter 2, some concepts from the robustness literature were introduced. An important concept was the inuence function. In the present chapter, the Chapter 3 Robustness Properties of the Student t Based Pseudo Maximum Likelihood Estimator In Chapter 2, some concepts from the robustness literature were introduced. An important concept was the inuence

More information

Antonietta Mira. University of Pavia, Italy. Abstract. We propose a test based on Bonferroni's measure of skewness.

Antonietta Mira. University of Pavia, Italy. Abstract. We propose a test based on Bonferroni's measure of skewness. Distribution-free test for symmetry based on Bonferroni's Measure Antonietta Mira University of Pavia, Italy Abstract We propose a test based on Bonferroni's measure of skewness. The test detects the asymmetry

More information

STAT 6350 Analysis of Lifetime Data. Probability Plotting

STAT 6350 Analysis of Lifetime Data. Probability Plotting STAT 6350 Analysis of Lifetime Data Probability Plotting Purpose of Probability Plots Probability plots are an important tool for analyzing data and have been particular popular in the analysis of life

More information

Time-varying failure rate for system reliability analysis in large-scale railway risk assessment simulation

Time-varying failure rate for system reliability analysis in large-scale railway risk assessment simulation Time-varying failure rate for system reliability analysis in large-scale railway risk assessment simulation H. Zhang, E. Cutright & T. Giras Center of Rail Safety-Critical Excellence, University of Virginia,

More information

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract

WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of

More information

Confidence intervals for kernel density estimation

Confidence intervals for kernel density estimation Stata User Group - 9th UK meeting - 19/20 May 2003 Confidence intervals for kernel density estimation Carlo Fiorio c.fiorio@lse.ac.uk London School of Economics and STICERD Stata User Group - 9th UK meeting

More information

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY

AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Econometrics Working Paper EWP0401 ISSN 1485-6441 Department of Economics AN EMPIRICAL LIKELIHOOD RATIO TEST FOR NORMALITY Lauren Bin Dong & David E. A. Giles Department of Economics, University of Victoria

More information

A Comparison of Sequential Sampling Models for Two-Choice Reaction Time

A Comparison of Sequential Sampling Models for Two-Choice Reaction Time Psychological Review Copyright 2004 by the American Psychological Association 2004, Vol. 111, No. 2, 333 367 0033-295X/04/$12.00 DOI: 10.1037/0033-295X.111.2.333 A Comparison of Sequential Sampling Models

More information

Probability and Statistics

Probability and Statistics Probability and Statistics Kristel Van Steen, PhD 2 Montefiore Institute - Systems and Modeling GIGA - Bioinformatics ULg kristel.vansteen@ulg.ac.be CHAPTER 4: IT IS ALL ABOUT DATA 4a - 1 CHAPTER 4: IT

More information

IENG581 Design and Analysis of Experiments INTRODUCTION

IENG581 Design and Analysis of Experiments INTRODUCTION Experimental Design IENG581 Design and Analysis of Experiments INTRODUCTION Experiments are performed by investigators in virtually all fields of inquiry, usually to discover something about a particular

More information

BNG 495 Capstone Design. Descriptive Statistics

BNG 495 Capstone Design. Descriptive Statistics BNG 495 Capstone Design Descriptive Statistics Overview The overall goal of this short course in statistics is to provide an introduction to descriptive and inferential statistical methods, with a focus

More information

Intuitive Biostatistics: Choosing a statistical test

Intuitive Biostatistics: Choosing a statistical test pagina 1 van 5 < BACK Intuitive Biostatistics: Choosing a statistical This is chapter 37 of Intuitive Biostatistics (ISBN 0-19-508607-4) by Harvey Motulsky. Copyright 1995 by Oxfd University Press Inc.

More information

SUMMARIZING MEASURED DATA. Gaia Maselli

SUMMARIZING MEASURED DATA. Gaia Maselli SUMMARIZING MEASURED DATA Gaia Maselli maselli@di.uniroma1.it Computer Network Performance 2 Overview Basic concepts Summarizing measured data Summarizing data by a single number Summarizing variability

More information

Estimation of Quantiles

Estimation of Quantiles 9 Estimation of Quantiles The notion of quantiles was introduced in Section 3.2: recall that a quantile x α for an r.v. X is a constant such that P(X x α )=1 α. (9.1) In this chapter we examine quantiles

More information

Recall the Basics of Hypothesis Testing

Recall the Basics of Hypothesis Testing Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE

More information

Learning Objectives for Stat 225

Learning Objectives for Stat 225 Learning Objectives for Stat 225 08/20/12 Introduction to Probability: Get some general ideas about probability, and learn how to use sample space to compute the probability of a specific event. Set Theory:

More information

Lesson Plan. Answer Questions. Summary Statistics. Histograms. The Normal Distribution. Using the Standard Normal Table

Lesson Plan. Answer Questions. Summary Statistics. Histograms. The Normal Distribution. Using the Standard Normal Table Lesson Plan Answer Questions Summary Statistics Histograms The Normal Distribution Using the Standard Normal Table 1 2. Summary Statistics Given a collection of data, one needs to find representations

More information

Distribution Fitting (Censored Data)

Distribution Fitting (Censored Data) Distribution Fitting (Censored Data) Summary... 1 Data Input... 2 Analysis Summary... 3 Analysis Options... 4 Goodness-of-Fit Tests... 6 Frequency Histogram... 8 Comparison of Alternative Distributions...

More information

The Nonparametric Bootstrap

The Nonparametric Bootstrap The Nonparametric Bootstrap The nonparametric bootstrap may involve inferences about a parameter, but we use a nonparametric procedure in approximating the parametric distribution using the ECDF. We use

More information

2.0 Lesson Plan. Answer Questions. Summary Statistics. Histograms. The Normal Distribution. Using the Standard Normal Table

2.0 Lesson Plan. Answer Questions. Summary Statistics. Histograms. The Normal Distribution. Using the Standard Normal Table 2.0 Lesson Plan Answer Questions 1 Summary Statistics Histograms The Normal Distribution Using the Standard Normal Table 2. Summary Statistics Given a collection of data, one needs to find representations

More information

Descriptive Data Summarization

Descriptive Data Summarization Descriptive Data Summarization Descriptive data summarization gives the general characteristics of the data and identify the presence of noise or outliers, which is useful for successful data cleaning

More information

Why is the field of statistics still an active one?

Why is the field of statistics still an active one? Why is the field of statistics still an active one? It s obvious that one needs statistics: to describe experimental data in a compact way, to compare datasets, to ask whether data are consistent with

More information

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity

Outline. Possible Reasons. Nature of Heteroscedasticity. Basic Econometrics in Transportation. Heteroscedasticity 1/25 Outline Basic Econometrics in Transportation Heteroscedasticity What is the nature of heteroscedasticity? What are its consequences? How does one detect it? What are the remedial measures? Amir Samimi

More information

Design and Implementation of CUSUM Exceedance Control Charts for Unknown Location

Design and Implementation of CUSUM Exceedance Control Charts for Unknown Location Design and Implementation of CUSUM Exceedance Control Charts for Unknown Location MARIEN A. GRAHAM Department of Statistics University of Pretoria South Africa marien.graham@up.ac.za S. CHAKRABORTI Department

More information

1 Degree distributions and data

1 Degree distributions and data 1 Degree distributions and data A great deal of effort is often spent trying to identify what functional form best describes the degree distribution of a network, particularly the upper tail of that distribution.

More information

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing

Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing Statistical Inference: Estimation and Confidence Intervals Hypothesis Testing 1 In most statistics problems, we assume that the data have been generated from some unknown probability distribution. We desire

More information

1 Overview. Coefficients of. Correlation, Alienation and Determination. Hervé Abdi Lynne J. Williams

1 Overview. Coefficients of. Correlation, Alienation and Determination. Hervé Abdi Lynne J. Williams In Neil Salkind (Ed.), Encyclopedia of Research Design. Thousand Oaks, CA: Sage. 2010 Coefficients of Correlation, Alienation and Determination Hervé Abdi Lynne J. Williams 1 Overview The coefficient of

More information

Continuous case Discrete case General case. Hazard functions. Patrick Breheny. August 27. Patrick Breheny Survival Data Analysis (BIOS 7210) 1/21

Continuous case Discrete case General case. Hazard functions. Patrick Breheny. August 27. Patrick Breheny Survival Data Analysis (BIOS 7210) 1/21 Hazard functions Patrick Breheny August 27 Patrick Breheny Survival Data Analysis (BIOS 7210) 1/21 Introduction Continuous case Let T be a nonnegative random variable representing the time to an event

More information

Chapter 4. Characterizing Data Numerically: Descriptive Statistics

Chapter 4. Characterizing Data Numerically: Descriptive Statistics Chapter 4 Characterizing Data Numerically: Descriptive Statistics While visual representations of data are very useful, they are only a beginning point from which we may gain more information. Numerical

More information

Stochastic dominance with imprecise information

Stochastic dominance with imprecise information Stochastic dominance with imprecise information Ignacio Montes, Enrique Miranda, Susana Montes University of Oviedo, Dep. of Statistics and Operations Research. Abstract Stochastic dominance, which is

More information

When is a copula constant? A test for changing relationships

When is a copula constant? A test for changing relationships When is a copula constant? A test for changing relationships Fabio Busetti and Andrew Harvey Bank of Italy and University of Cambridge November 2007 usetti and Harvey (Bank of Italy and University of Cambridge)

More information

Comparing Measures of Central Tendency *

Comparing Measures of Central Tendency * OpenStax-CNX module: m11011 1 Comparing Measures of Central Tendency * David Lane This work is produced by OpenStax-CNX and licensed under the Creative Commons Attribution License 1.0 1 Comparing Measures

More information

Inferring from data. Theory of estimators

Inferring from data. Theory of estimators Inferring from data Theory of estimators 1 Estimators Estimator is any function of the data e(x) used to provide an estimate ( a measurement ) of an unknown parameter. Because estimators are functions

More information

Can we do statistical inference in a non-asymptotic way? 1

Can we do statistical inference in a non-asymptotic way? 1 Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.2 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis.

401 Review. 6. Power analysis for one/two-sample hypothesis tests and for correlation analysis. 401 Review Major topics of the course 1. Univariate analysis 2. Bivariate analysis 3. Simple linear regression 4. Linear algebra 5. Multiple regression analysis Major analysis methods 1. Graphical analysis

More information

Statistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation

Statistics - Lecture One. Outline. Charlotte Wickham  1. Basic ideas about estimation Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence

More information

A Note on Bayesian Inference After Multiple Imputation

A Note on Bayesian Inference After Multiple Imputation A Note on Bayesian Inference After Multiple Imputation Xiang Zhou and Jerome P. Reiter Abstract This article is aimed at practitioners who plan to use Bayesian inference on multiplyimputed datasets in

More information

Describing distributions with numbers

Describing distributions with numbers Describing distributions with numbers A large number or numerical methods are available for describing quantitative data sets. Most of these methods measure one of two data characteristics: The central

More information

1-1. Chapter 1. Sampling and Descriptive Statistics by The McGraw-Hill Companies, Inc. All rights reserved.

1-1. Chapter 1. Sampling and Descriptive Statistics by The McGraw-Hill Companies, Inc. All rights reserved. 1-1 Chapter 1 Sampling and Descriptive Statistics 1-2 Why Statistics? Deal with uncertainty in repeated scientific measurements Draw conclusions from data Design valid experiments and draw reliable conclusions

More information

Statistical Analysis of Backtracking on. Inconsistent CSPs? Irina Rish and Daniel Frost.

Statistical Analysis of Backtracking on. Inconsistent CSPs? Irina Rish and Daniel Frost. Statistical Analysis of Backtracking on Inconsistent CSPs Irina Rish and Daniel Frost Department of Information and Computer Science University of California, Irvine, CA 92697-3425 firinar,frostg@ics.uci.edu

More information

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments

Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments Finite Sample Performance of A Minimum Distance Estimator Under Weak Instruments Tak Wai Chau February 20, 2014 Abstract This paper investigates the nite sample performance of a minimum distance estimator

More information

More on Estimation. Maximum Likelihood Estimation.

More on Estimation. Maximum Likelihood Estimation. More on Estimation. In the previous chapter we looked at the properties of estimators and the criteria we could use to choose between types of estimators. Here we examine more closely some very popular

More information

Chapter 9. Non-Parametric Density Function Estimation

Chapter 9. Non-Parametric Density Function Estimation 9-1 Density Estimation Version 1.1 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least

More information

Normal Probability Plot Probability Probability

Normal Probability Plot Probability Probability Modelling multivariate returns Stefano Herzel Department ofeconomics, University of Perugia 1 Catalin Starica Department of Mathematical Statistics, Chalmers University of Technology Reha Tutuncu Department

More information

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages:

Glossary. The ISI glossary of statistical terms provides definitions in a number of different languages: Glossary The ISI glossary of statistical terms provides definitions in a number of different languages: http://isi.cbs.nl/glossary/index.htm Adjusted r 2 Adjusted R squared measures the proportion of the

More information

Introduction. Chapter 1

Introduction. Chapter 1 Chapter 1 Introduction In this book we will be concerned with supervised learning, which is the problem of learning input-output mappings from empirical data (the training dataset). Depending on the characteristics

More information

Lecture 14 - P v.s. NP 1

Lecture 14 - P v.s. NP 1 CME 305: Discrete Mathematics and Algorithms Instructor: Professor Aaron Sidford (sidford@stanford.edu) February 27, 2018 Lecture 14 - P v.s. NP 1 In this lecture we start Unit 3 on NP-hardness and approximation

More information

One-Sample Numerical Data

One-Sample Numerical Data One-Sample Numerical Data quantiles, boxplot, histogram, bootstrap confidence intervals, goodness-of-fit tests University of California, San Diego Instructor: Ery Arias-Castro http://math.ucsd.edu/~eariasca/teaching.html

More information

1. Exploratory Data Analysis

1. Exploratory Data Analysis 1. Exploratory Data Analysis 1.1 Methods of Displaying Data A visual display aids understanding and can highlight features which may be worth exploring more formally. Displays should have impact and be

More information

Hazard Function, Failure Rate, and A Rule of Thumb for Calculating Empirical Hazard Function of Continuous-Time Failure Data

Hazard Function, Failure Rate, and A Rule of Thumb for Calculating Empirical Hazard Function of Continuous-Time Failure Data Hazard Function, Failure Rate, and A Rule of Thumb for Calculating Empirical Hazard Function of Continuous-Time Failure Data Feng-feng Li,2, Gang Xie,2, Yong Sun,2, Lin Ma,2 CRC for Infrastructure and

More information

Joint Estimation of Risk Preferences and Technology: Further Discussion

Joint Estimation of Risk Preferences and Technology: Further Discussion Joint Estimation of Risk Preferences and Technology: Further Discussion Feng Wu Research Associate Gulf Coast Research and Education Center University of Florida Zhengfei Guan Assistant Professor Gulf

More information

Goodness-of-Fit Tests With Right-Censored Data by Edsel A. Pe~na Department of Statistics University of South Carolina Colloquium Talk August 31, 2 Research supported by an NIH Grant 1 1. Practical Problem

More information

4 Testing Hypotheses. 4.1 Tests in the regression setting. 4.2 Non-parametric testing of survival between groups

4 Testing Hypotheses. 4.1 Tests in the regression setting. 4.2 Non-parametric testing of survival between groups 4 Testing Hypotheses The next lectures will look at tests, some in an actuarial setting, and in the last subsection we will also consider tests applied to graduation 4 Tests in the regression setting )

More information

Men. Women. Men. Men. Women. Women

Men. Women. Men. Men. Women. Women Math 203 Topics for second exam Statistics: the science of data Chapter 5: Producing data Statistics is all about drawing conclusions about the opinions/behavior/structure of large populations based on

More information

Probability Distributions

Probability Distributions CONDENSED LESSON 13.1 Probability Distributions In this lesson, you Sketch the graph of the probability distribution for a continuous random variable Find probabilities by finding or approximating areas

More information

AP Statistics Cumulative AP Exam Study Guide

AP Statistics Cumulative AP Exam Study Guide AP Statistics Cumulative AP Eam Study Guide Chapters & 3 - Graphs Statistics the science of collecting, analyzing, and drawing conclusions from data. Descriptive methods of organizing and summarizing statistics

More information

1/sqrt(B) convergence 1/B convergence B

1/sqrt(B) convergence 1/B convergence B The Error Coding Method and PICTs Gareth James and Trevor Hastie Department of Statistics, Stanford University March 29, 1998 Abstract A new family of plug-in classication techniques has recently been

More information

Statistics: Learning models from data

Statistics: Learning models from data DS-GA 1002 Lecture notes 5 October 19, 2015 Statistics: Learning models from data Learning models from data that are assumed to be generated probabilistically from a certain unknown distribution is a crucial

More information

Hypothesis Testing with the Bootstrap. Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods

Hypothesis Testing with the Bootstrap. Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods Hypothesis Testing with the Bootstrap Noa Haas Statistics M.Sc. Seminar, Spring 2017 Bootstrap and Resampling Methods Bootstrap Hypothesis Testing A bootstrap hypothesis test starts with a test statistic

More information

Unit 2. Describing Data: Numerical

Unit 2. Describing Data: Numerical Unit 2 Describing Data: Numerical Describing Data Numerically Describing Data Numerically Central Tendency Arithmetic Mean Median Mode Variation Range Interquartile Range Variance Standard Deviation Coefficient

More information

Estimation and Confidence Intervals for Parameters of a Cumulative Damage Model

Estimation and Confidence Intervals for Parameters of a Cumulative Damage Model United States Department of Agriculture Forest Service Forest Products Laboratory Research Paper FPL-RP-484 Estimation and Confidence Intervals for Parameters of a Cumulative Damage Model Carol L. Link

More information

F78SC2 Notes 2 RJRC. If the interest rate is 5%, we substitute x = 0.05 in the formula. This gives

F78SC2 Notes 2 RJRC. If the interest rate is 5%, we substitute x = 0.05 in the formula. This gives F78SC2 Notes 2 RJRC Algebra It is useful to use letters to represent numbers. We can use the rules of arithmetic to manipulate the formula and just substitute in the numbers at the end. Example: 100 invested

More information

Maximum-Likelihood fitting

Maximum-Likelihood fitting CMP 0b Lecture F. Sigworth Maximum-Likelihood fitting One of the issues I want to address in this lecture is the fitting of distributions dwell times. We want to find the best curve to draw over a histogram,

More information

Range The range is the simplest of the three measures and is defined now.

Range The range is the simplest of the three measures and is defined now. Measures of Variation EXAMPLE A testing lab wishes to test two experimental brands of outdoor paint to see how long each will last before fading. The testing lab makes 6 gallons of each paint to test.

More information

What s New in Econometrics? Lecture 14 Quantile Methods

What s New in Econometrics? Lecture 14 Quantile Methods What s New in Econometrics? Lecture 14 Quantile Methods Jeff Wooldridge NBER Summer Institute, 2007 1. Reminders About Means, Medians, and Quantiles 2. Some Useful Asymptotic Results 3. Quantile Regression

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016

Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 8. For any two events E and F, P (E) = P (E F ) + P (E F c ). Summary of basic probability theory Math 218, Mathematical Statistics D Joyce, Spring 2016 Sample space. A sample space consists of a underlying

More information

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career.

9/2/2010. Wildlife Management is a very quantitative field of study. throughout this course and throughout your career. Introduction to Data and Analysis Wildlife Management is a very quantitative field of study Results from studies will be used throughout this course and throughout your career. Sampling design influences

More information

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood

Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Stat 542: Item Response Theory Modeling Using The Extended Rank Likelihood Jonathan Gruhl March 18, 2010 1 Introduction Researchers commonly apply item response theory (IRT) models to binary and ordinal

More information

Robustness. James H. Steiger. Department of Psychology and Human Development Vanderbilt University. James H. Steiger (Vanderbilt University) 1 / 37

Robustness. James H. Steiger. Department of Psychology and Human Development Vanderbilt University. James H. Steiger (Vanderbilt University) 1 / 37 Robustness James H. Steiger Department of Psychology and Human Development Vanderbilt University James H. Steiger (Vanderbilt University) 1 / 37 Robustness 1 Introduction 2 Robust Parameters and Robust

More information

εx 2 + x 1 = 0. (2) Suppose we try a regular perturbation expansion on it. Setting ε = 0 gives x 1 = 0,

εx 2 + x 1 = 0. (2) Suppose we try a regular perturbation expansion on it. Setting ε = 0 gives x 1 = 0, 4 Rescaling In this section we ll look at one of the reasons that our ε = 0 system might not have enough solutions, and introduce a tool that is fundamental to all perturbation systems. We ll start with

More information

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80

The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 The University of Hong Kong Department of Statistics and Actuarial Science STAT2802 Statistical Models Tutorial Solutions Solutions to Problems 71-80 71. Decide in each case whether the hypothesis is simple

More information

Introduction to Statistical Analysis

Introduction to Statistical Analysis Introduction to Statistical Analysis Changyu Shen Richard A. and Susan F. Smith Center for Outcomes Research in Cardiology Beth Israel Deaconess Medical Center Harvard Medical School Objectives Descriptive

More information

Non-uniform coverage estimators for distance sampling

Non-uniform coverage estimators for distance sampling Abstract Non-uniform coverage estimators for distance sampling CREEM Technical report 2007-01 Eric Rexstad Centre for Research into Ecological and Environmental Modelling Research Unit for Wildlife Population

More information