SIO 221B, Rudnick adapted from Davis 1. 1 x lim. N x 2 n = 1 N. { x} 1 N. N x = 1 N. N x = 1 ( N N x ) x = 0 (3) = 1 x N 2

Size: px
Start display at page:

Download "SIO 221B, Rudnick adapted from Davis 1. 1 x lim. N x 2 n = 1 N. { x} 1 N. N x = 1 N. N x = 1 ( N N x ) x = 0 (3) = 1 x N 2"

Transcription

1 SIO B, Rudnick adapted from Davis VII. Sampling errors We do not have access to the true statistics, so we must compute sample statistics. By this we mean that the number of realizations we average over is necessarily finite. We wish to determine the accuracy of our sample statistics to aid in interpretation. he difference between the true and sampling statistics is called the sampling error. We have already defined the true mean he sample mean is defined as x lim { x} x n () x n () ote the use of curly braces as a notation for the sample mean. he distinguishing factor between () and () is that the sample mean is over a finite number of realizations. Consider the problem of estimating the true mean with the sample mean. Suppose the sample mean were calculated an infinite number of times, then we could consider the mean of the sample mean, and compare it to the true mean. he bias is defined to be the mean difference between the sample mean and the true mean E = { x} x = x n x = hus the sample mean has desirable property of zero bias. x n x = ( x ) x = (3) Consider the variance of the sample mean E = ({ x} x ) = x n x (4) Moving the true mean inside the sum results in E = x n x (5) he quantity in square brackets in (5) is the fluctuation, so E = x n (6) he average squared sum of fluctuations is the sum over all elements of the covariance matrix of the fluctuations E = x x n (7) m m= Assuming the fluctuations are independent with equal variance then

2 SIO B, Rudnick adapted from Davis E = δ nm σ = σ m= So the variance decreases as the inverse of. he square root of E is called the root-meansquare error or the standard error, and decreases as the inverse of the square root of (8) E = σ (9) Recognizing this improvement of error by -/ is key to understanding many measures of statistical error. ote the parallel between sampling error and the random walk. ote that the rms error depends on the true variance. Suppose first that we know the true mean so we can estimate the variance as he mean of this estimate is ˆσ ˆσ = { x } () = x = σ () n so this estimate of variance is unbiased. he variance of the estimate of variance is F = Assuming the fluctuations are independent F = ({ x } σ ) = m= ( )σ 4 + x 4 x x n σ 4 () m σ 4 = ( x 4 σ ) 4 (3) So the error in the variance depends on the fourth moment. his is a problem of closure. o determine the error of any statistic we need to know a higher order statistic. hus errors on an estimated statistic are always less well determined than the original estimate. he reality is we don t know the true mean in order to use (), so we must estimate it using (). Consider the following estimator ˆσ = A ( x { x} ) { } (4) where A is a constant to be determined. he obvious choice is A=, but is it optimal in any sense? Let s investigate the bias in the estimate (4) F = ˆσ σ (5) ow we rewrite the estimate (4) in terms of deviation from the mean to get F = A ( x x ) ({ x} x ) { } σ (6) As the sample mean operator acts only on x, expanding the quadratic yields F = A x { } ({ x} x ) σ (7)

3 SIO B, Rudnick adapted from Davis 3 he second term in the angle brackets of (7) is the variance of the sample mean, which for independent data is given by (8). he first term in the angle brackets is just (-). he result is thus F = Aσ σ (8) If we want the estimate (4) to be unbiased, that is F=, then A = So an unbiased estimate of variance is ˆσ = ( x n { x} ) () Clearly, dividing by - rather than by makes little difference when is large. So any bias in the most obvious estimate of variance is small. Also, there are other estimates of variance that might be optimal in other ways, rather than to eliminate bias. For example, consider the variance of the sample variance (9) F = ( ˆσ σ ) () Expanding the quadratic results, and applying the averaging operator to each term results in F = ˆσ 4 σ ˆσ + σ 4 () We already know the second term from the calculation of bias σ ˆσ = Aσ 4 he first term in () is more challenging to evaluate {( { }) } {( { } x ) } {( { }) } { { } + { x } } ˆσ 4 = A x x = A x x x = A x x = A x x x Since the sample average operates only on x and not on x ˆσ 4 = A = A x ({ x } { x } ) { } { x } x { }, we find (3) (4) (5) { } + { x } 4 Expanding the sample means as sums, and applying the averaging operator to each term, results in

4 SIO B, Rudnick adapted from Davis 4 ˆσ 4 = A x x n x m m= x 3 n x + x m p m= p= 4 n x m x p x q m= p= q= (6) So, we have a number of mean products, and we have to sum over all the elements. We assume, as above, that the data are independent. Because nonzero contributions come only from even moments, we find ˆσ 4 = A ( ( )σ 4 + x ) 4 ( 3 ( )σ 4 + x ) 4 + ( 4 3 ( )σ 4 + x ) 4 (7) he factor of 3 in the last term comes from the fact that there are three possible pairs in the fourth-order product. his can be simplified to σˆ 4 = A ( + 3)σ 4 + ( ) x 4 3 (8) We require the fourth moment. If we assume x to have a normal distribution, then x 4 = 3σ 4, and the result is ( )( +) ˆσ 4 = A σ 4 (9) he variance of the sample mean is found by substituting (3) and (9) into () ( ) ( +) F = A σ 4 Aσ 4 + σ 4 (3) We choose the value of A that minimizes the variance by setting the derivative of (3) with respect to A equal to zero, yielding A = (3) + So for minimum variance (3), A is slightly less than, while for zeros bias (9), A is slightly greater than. Maybe a choice of is not so bad after all. he resulting variance and bias, with A given by (3) is F = σ 4 + ; F = σ + So there is a negative bias, which is understood as the value for A in this case is less than that in the unbiased case (9). For completeness, the variance and bias in the unbiased estimator of variance is, with A given by (9), F = σ 4 ; F = (33) where we have assumed the data to be normally distributed. (3)

5 SIO B, Rudnick adapted from Davis 5 Correlated Observations Real observations are almost always correlated. All the estimates of sampling error we have considered so far have assumed that the realizations are independent. An example of correlated observations occurs when we have closely spaced samples of a variable with very long scales, as daily samples of temperature off the Scripps pier in the presence of ESO s and PDO s. he time series in Figure contains time scales relative long compared to the separation between observations A and B, but comparable to the separation between points A and C. Recall the expression for the sample mean (7), that consists of a sum over all elements in the covariance matrix. Let s relax the assumption of independence, but assume now that the statistics are stationary, so x n x m = σ ρ( n m ) (34) where ρ is correlation. hat is the covariance depends only on the difference in sample index. Assuming that the samples are evenly spaced, the mean square error of the sample mean is E = σ n ρ( n ) (35) n= 3 C x A B t Figure. A serially correlated time series. Observations A and B are closely spaced relative to the dominant time scales. Observations A and C are relatively far apart. hus observation C is more likely than B to be independent of A.

6 SIO B, Rudnick adapted from Davis 6 Understand that the index n in (35) refers to the diagonal of the covariance matrix, the covariance matrix is symmetric, and the covariance matrix is such that each diagonal band is made up of equal values. If the data are uncorrelated, then the only nonzero term in (35) is for n=, and (8) is recovered. he variance of the sample mean can be written E = σ E (36) where the equivalent number of independent samples in serially correlated data E is E = n= n ρ n E is sometimes called the equivalent number of degrees of freedom. If the number of samples is large compared to the value of n where correlation ρ vanishes, then E = n= hese ideas can be extended to continuous time series, which clearly have correlated samples. In this case the sample mean is ( ) ρ n ( ) (37) (38) { x} = x( t)dt (39) he mean square error of the sample mean is E = x ( t ) x ( t ) dt dt (4) Assuming stationary statistics then E = σ he number equivalent degrees of freedom is E = t ρ( t )dt (4) t ρ( t )dt aking the limit as the time series is infinitely long, the mean square error is E = σ (4) ρ( t )dt (43) he integral of the correlation is thus an important quantity, often called the integral time scale. he ratio of the time series length to the integral time scales tells how many independent samples there are in a time series. here is an illuminating relationship between the spectrum and the variance of the sample mean. ote that the spectrum and the covariance are Fourier transform pairs

7 SIO B, Rudnick adapted from Davis 7 ( ) = C t S f ( )e iπ ft dt (44) Here S is the spectrum, C the covariance, and f frequency. Because C = σ ρ, and using (43), the following relationship holds for time series long compared to the integral time scale E = S( ) his somewhat surprising result suggests that the spectrum at zero frequency, that is the lowest frequencies in a continuous time series, determine the error in the sample mean. High frequencies in a continuous time series are effectively averaged out, but the lowest frequencies cause error. A series of independent data have a white spectrum, and that a white spectrum is nonzero at low frequency is key to the effect on the error in the sample mean. Confidence limits he concept of confidence limits is a bit subtle. he idea is to come up with an interval and state the probability that the true value of a statistic is in that interval. Suppose we could calculate a sample mean an infinite number of times, and come up with its true distribution. (Figure ) Given that distribution, we could say that the next sample mean we calculate would fall between lower and upper limits xl and xu with a probability of (45) F {x} (r) x L <x> x U r Figure. he probability density function of the sample mean.

8 SIO B, Rudnick adapted from Davis 8 x U P = F x ( )dr (46) x L { } r After we have calculated the next sample mean, it either does or does not fall within these limits. hat is the probability it is within some interval is either or. Since we claim to know the true mean first before calculating the sample mean, this is not a satisfactory procedure. he question we really want to ask is: What is the range of true means likely to have led to the sample mean we just calculated? he upper limit is the largest true mean for which there is a small ε probability of getting a smaller sample mean than the one observed. he lower limit is the smallest true mean for which there is an ε probability of getting a larger sample mean than the one observed. (Figure 3) If the sample mean we observed is ˆm and the cumulative distribution of the sample mean is D { x} with assumed lower and upper true means of xl and xu, then the following probability statements are appropriate. D { x} ( ˆm x L ) = ε; D { x} ˆm x U ( ) = ε (47).4..8 F {x}.6.4. x L ˆm x U Figure 3. he confidence interval for the sample mean is bounded by the largest and smallest true means that were likely to have led to the observed sample mean. he area underneath the blue curve to the right of ˆm is the probability ε of getting a larger sample mean than the one observed, given a true mean of xl. Similarly, the area under the green curve to the left of ˆm is ε.

9 SIO B, Rudnick adapted from Davis 9 he story we need to tell to have confidence limits is to say we know the distribution, but we don t know the mean. he confidence interval defines as in (47) is said to be at the (-ε)% level. o make more progress, we need an assumed distribution. Because the sample mean is a sum of random variables, its distribution should approach normal with a standard deviation given by (9) as the number of degrees of freedom gets large. he lower and upper limits (47) are then x L = ˆm σ z ε ; x U = ˆm + σ z ε where z ε is the inverse normal cumulative distribution function with zero mean and unit standard deviation evaluated at ε. (demo) ote that the limits are related to the estimated mean as a sum. he chi-squared χ distribution is useful for finding confidence intervals for sample variances. For example, estimates of spectra are often assumed to have a χ distribution. Consider an estimate of the sample variance made by the sum of squares of independent, identically and normally distributed variables ˆσ = σ x n (48) F L ˆσ U Figure 4. he confidence interval for the sample variance. he different widths of the two χ pdfs result because a larger mean χ variable has a larger variance, as the χ pdf must pass through the origin.

10 SIO B, Rudnick adapted from Davis where he variable x = ; x = (49) x n = ˆσ is a χ variable with degrees of freedom. he mean of this variable is and its variance is x n he lower and upper limits are then where χ ;ε x n = x n σ (5) = (5) = x n x m x n + m= = ( ) = σ L = ˆσ χ ; ε ; σ U = ˆσ (5) (53) χ ;ε is the inverse χ cumulative distribution function with degrees of freedom evaluated at ε. (Figure 4) As a χ variable is a sum, the χ distribution approaches normal as gets large. In contrast to the confidence limits for a sample mean, the limits for sample variance are related to the estimate as a ratio.

Analytical Methods. Session 3: Statistics II. UCL Department of Civil, Environmental & Geomatic Engineering. Analytical Methods.

Analytical Methods. Session 3: Statistics II. UCL Department of Civil, Environmental & Geomatic Engineering. Analytical Methods. Analytical Methods Session 3: Statistics II More statistical tests Quite a few more questions that we might want to ask about data that we have. Is one value significantly different to the rest, or to

More information

SIO 211B, Rudnick, adapted from Davis 1

SIO 211B, Rudnick, adapted from Davis 1 SIO 211B, Rudnick, adapted from Davis 1 XVII.Empirical orthogonal functions Often in oceanography we collect large data sets that are time series at a group of locations. Moored current meter arrays do

More information

Economics 583: Econometric Theory I A Primer on Asymptotics

Economics 583: Econometric Theory I A Primer on Asymptotics Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:

More information

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1 Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the

More information

Statistics and Data Analysis

Statistics and Data Analysis Statistics and Data Analysis The Crash Course Physics 226, Fall 2013 "There are three kinds of lies: lies, damned lies, and statistics. Mark Twain, allegedly after Benjamin Disraeli Statistics and Data

More information

Stochastic Processes. A stochastic process is a function of two variables:

Stochastic Processes. A stochastic process is a function of two variables: Stochastic Processes Stochastic: from Greek stochastikos, proceeding by guesswork, literally, skillful in aiming. A stochastic process is simply a collection of random variables labelled by some parameter:

More information

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process

ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random

More information

2008 Winton. Statistical Testing of RNGs

2008 Winton. Statistical Testing of RNGs 1 Statistical Testing of RNGs Criteria for Randomness For a sequence of numbers to be considered a sequence of randomly acquired numbers, it must have two basic statistical properties: Uniformly distributed

More information

White Noise Processes (Section 6.2)

White Noise Processes (Section 6.2) White Noise Processes (Section 6.) Recall that covariance stationary processes are time series, y t, such. E(y t ) = µ for all t. Var(y t ) = σ for all t, σ < 3. Cov(y t,y t-τ ) = γ(τ) for all t and τ

More information

A Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit

A Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit A Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit David Vener Department of Mathematics, MIT May 5, 3 Introduction In 8.366, we discussed the relationship

More information

ECON The Simple Regression Model

ECON The Simple Regression Model ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In

More information

where r n = dn+1 x(t)

where r n = dn+1 x(t) Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution

More information

Estimators as Random Variables

Estimators as Random Variables Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until

More information

ENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University

ENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University ENSC327 Communications Systems 19: Random Processes Jie Liang School of Engineering Science Simon Fraser University 1 Outline Random processes Stationary random processes Autocorrelation of random processes

More information

Lecture 3: Statistical sampling uncertainty

Lecture 3: Statistical sampling uncertainty Lecture 3: Statistical sampling uncertainty c Christopher S. Bretherton Winter 2015 3.1 Central limit theorem (CLT) Let X 1,..., X N be a sequence of N independent identically-distributed (IID) random

More information

Multivariate Time Series

Multivariate Time Series Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form

More information

Time series models in the Frequency domain. The power spectrum, Spectral analysis

Time series models in the Frequency domain. The power spectrum, Spectral analysis ime series models in the Frequency domain he power spectrum, Spectral analysis Relationship between the periodogram and the autocorrelations = + = ( ) ( ˆ α ˆ ) β I Yt cos t + Yt sin t t= t= ( ( ) ) cosλ

More information

Power Spectral Density of Digital Modulation Schemes

Power Spectral Density of Digital Modulation Schemes Digital Communication, Continuation Course Power Spectral Density of Digital Modulation Schemes Mikael Olofsson Emil Björnson Department of Electrical Engineering ISY) Linköping University, SE-581 83 Linköping,

More information

XVI. Objective mapping

XVI. Objective mapping SIO 211B, Rudnick, adapted from Davis 1 XVI. Objective mapping An objective map is the minimum mean-square error estimate of a continuous function of a variable, given discrete data. The interpolation

More information

Probabilities & Statistics Revision

Probabilities & Statistics Revision Probabilities & Statistics Revision Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 January 6, 2017 Christopher Ting QF

More information

Lecture 4: Testing Stuff

Lecture 4: Testing Stuff Lecture 4: esting Stuff. esting Hypotheses usually has three steps a. First specify a Null Hypothesis, usually denoted, which describes a model of H 0 interest. Usually, we express H 0 as a restricted

More information

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit

Statistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit Statistics Lent Term 2015 Prof. Mark Thomson Lecture 2 : The Gaussian Limit Prof. M.A. Thomson Lent Term 2015 29 Lecture Lecture Lecture Lecture 1: Back to basics Introduction, Probability distribution

More information

Statistics for scientists and engineers

Statistics for scientists and engineers Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3

More information

So far we discussed random number generators that need to have the maximum length period.

So far we discussed random number generators that need to have the maximum length period. So far we discussed random number generators that need to have the maximum length period. Even the increment series has the maximum length period yet it is by no means random How do we decide if a generator

More information

Time Series 3. Robert Almgren. Sept. 28, 2009

Time Series 3. Robert Almgren. Sept. 28, 2009 Time Series 3 Robert Almgren Sept. 28, 2009 Last time we discussed two main categories of linear models, and their combination. Here w t denotes a white noise: a stationary process with E w t ) = 0, E

More information

1 Least Squares Estimation - multiple regression.

1 Least Squares Estimation - multiple regression. Introduction to multiple regression. Fall 2010 1 Least Squares Estimation - multiple regression. Let y = {y 1,, y n } be a n 1 vector of dependent variable observations. Let β = {β 0, β 1 } be the 2 1

More information

Fitting a Straight Line to Data

Fitting a Straight Line to Data Fitting a Straight Line to Data Thanks for your patience. Finally we ll take a shot at real data! The data set in question is baryonic Tully-Fisher data from http://astroweb.cwru.edu/sparc/btfr Lelli2016a.mrt,

More information

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1)

Regression #5: Confidence Intervals and Hypothesis Testing (Part 1) Regression #5: Confidence Intervals and Hypothesis Testing (Part 1) Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #5 1 / 24 Introduction What is a confidence interval? To fix ideas, suppose

More information

Consider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y.

Consider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y. ATMO/OPTI 656b Spring 009 Bayesian Retrievals Note: This follows the discussion in Chapter of Rogers (000) As we have seen, the problem with the nadir viewing emission measurements is they do not contain

More information

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!

Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses

More information

The regression model with one stochastic regressor.

The regression model with one stochastic regressor. The regression model with one stochastic regressor. 3150/4150 Lecture 6 Ragnar Nymoen 30 January 2012 We are now on Lecture topic 4 The main goal in this lecture is to extend the results of the regression

More information

Recall the Basics of Hypothesis Testing

Recall the Basics of Hypothesis Testing Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE

More information

A=randn(500,100); mu=mean(a); sigma_a=std(a); std_a=sigma_a/sqrt(500); [std(mu) mean(std_a)] % compare standard deviation of means % vs standard error

A=randn(500,100); mu=mean(a); sigma_a=std(a); std_a=sigma_a/sqrt(500); [std(mu) mean(std_a)] % compare standard deviation of means % vs standard error UCSD SIOC 221A: (Gille) 1 Reading: Bendat and Piersol, Ch. 5.2.1 Lecture 10: Recap Last time we looked at the sinc function, windowing, and detrending with an eye to reducing edge effects in our spectra.

More information

MS&E 226: Small Data

MS&E 226: Small Data MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some

More information

Simple Linear Regression

Simple Linear Regression Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)

More information

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM

Business Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF

More information

Week 9 Generators, duality, change of measure

Week 9 Generators, duality, change of measure Week 9 Generators, duality, change of measure Jonathan Goodman November 18, 013 1 Generators This section describes a common abstract way to describe many of the differential equations related to Markov

More information

Properties of the least squares estimates

Properties of the least squares estimates Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares

More information

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N

Econ 836 Final Exam. 2 w N 2 u N 2. 2 v N 1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.

More information

Taylor Series and Series Convergence (Online)

Taylor Series and Series Convergence (Online) 7in 0in Felder c02_online.te V3 - February 9, 205 9:5 A.M. Page CHAPTER 2 Taylor Series and Series Convergence (Online) 2.8 Asymptotic Epansions In introductory calculus classes the statement this series

More information

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment:

Stochastic Processes: I. consider bowl of worms model for oscilloscope experiment: Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

2.3 Estimating PDFs and PDF Parameters

2.3 Estimating PDFs and PDF Parameters .3 Estimating PDFs and PDF Parameters estimating means - discrete and continuous estimating variance using a known mean estimating variance with an estimated mean estimating a discrete pdf estimating a

More information

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of

covariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of Index* The Statistical Analysis of Time Series by T. W. Anderson Copyright 1971 John Wiley & Sons, Inc. Aliasing, 387-388 Autoregressive {continued) Amplitude, 4, 94 case of first-order, 174 Associated

More information

Solution. The relationship between cartesian coordinates (x, y) and polar coordinates (r, θ) is given by. (x, y) = (r cos θ, r sin θ).

Solution. The relationship between cartesian coordinates (x, y) and polar coordinates (r, θ) is given by. (x, y) = (r cos θ, r sin θ). Problem 1. Let p 1 be the point having polar coordinates r = 1 and θ = π. Let p 2 be the point having polar coordinates r = 1 and θ = π/2. Find the Euclidean distance between p 1 and p 2. The relationship

More information

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis

Chapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis Chapter 12: An introduction to Time Series Analysis Introduction In this chapter, we will discuss forecasting with single-series (univariate) Box-Jenkins models. The common name of the models is Auto-Regressive

More information

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B

Problems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2

More information

Regression Models - Introduction

Regression Models - Introduction Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,

More information

Computational Data Analysis!

Computational Data Analysis! 12.714 Computational Data Analysis! Alan Chave (alan@whoi.edu)! Thomas Herring (tah@mit.edu),! http://geoweb.mit.edu/~tah/12.714! Introduction to Spectral Analysis! Topics Today! Aspects of Time series

More information

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1

M(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1 Math 66/566 - Midterm Solutions NOTE: These solutions are for both the 66 and 566 exam. The problems are the same until questions and 5. 1. The moment generating function of a random variable X is M(t)

More information

Topics in Probability and Statistics

Topics in Probability and Statistics Topics in Probability and tatistics A Fundamental Construction uppose {, P } is a sample space (with probability P), and suppose X : R is a random variable. The distribution of X is the probability P X

More information

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet

Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Please answer all of the questions and show your work. If you think a question is ambiguous, clearly state how you interpret it before

More information

Treatment of Error in Experimental Measurements

Treatment of Error in Experimental Measurements in Experimental Measurements All measurements contain error. An experiment is truly incomplete without an evaluation of the amount of error in the results. In this course, you will learn to use some common

More information

Chapter 3. Introducing Groups

Chapter 3. Introducing Groups Chapter 3 Introducing Groups We need a super-mathematics in which the operations are as unknown as the quantities they operate on, and a super-mathematician who does not know what he is doing when he performs

More information

Multiple Random Variables

Multiple Random Variables Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x

More information

Basic Probability Reference Sheet

Basic Probability Reference Sheet February 27, 2001 Basic Probability Reference Sheet 17.846, 2001 This is intended to be used in addition to, not as a substitute for, a textbook. X is a random variable. This means that X is a variable

More information

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008

2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008 MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.

More information

VIII. Coherence and Transfer Function Applications A. Coherence Function Estimates

VIII. Coherence and Transfer Function Applications A. Coherence Function Estimates VIII. Coherence and Transfer Function Applications A. Coherence Function Estimates Consider the application of these ideas to the specific problem of atmospheric turbulence measurements outlined in Figure

More information

PHYS 6710: Nuclear and Particle Physics II

PHYS 6710: Nuclear and Particle Physics II The ellipse of constant chi-square is described by The error ellipse To describe the tilt of the error ellipse, we need to consider the following: For two parameters a,b, there are new coordinates a',

More information

An example to illustrate frequentist and Bayesian approches

An example to illustrate frequentist and Bayesian approches Frequentist_Bayesian_Eample An eample to illustrate frequentist and Bayesian approches This is a trivial eample that illustrates the fundamentally different points of view of the frequentist and Bayesian

More information

3.3 Estimator quality, confidence sets and bootstrapping

3.3 Estimator quality, confidence sets and bootstrapping Estimator quality, confidence sets and bootstrapping 109 3.3 Estimator quality, confidence sets and bootstrapping A comparison of two estimators is always a matter of comparing their respective distributions.

More information

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley

Review of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate

More information

Signals and Spectra - Review

Signals and Spectra - Review Signals and Spectra - Review SIGNALS DETERMINISTIC No uncertainty w.r.t. the value of a signal at any time Modeled by mathematical epressions RANDOM some degree of uncertainty before the signal occurs

More information

2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES

2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES 2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES 2.0 THEOREM OF WIENER- KHINTCHINE An important technique in the study of deterministic signals consists in using harmonic functions to gain the spectral

More information

Simple Linear Regression for the Climate Data

Simple Linear Regression for the Climate Data Prediction Prediction Interval Temperature 0.2 0.0 0.2 0.4 0.6 0.8 320 340 360 380 CO 2 Simple Linear Regression for the Climate Data What do we do with the data? y i = Temperature of i th Year x i =CO

More information

1 Linear Difference Equations

1 Linear Difference Equations ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with

More information

SIO 210: Data analysis methods L. Talley, Fall Sampling and error 2. Basic statistical concepts 3. Time series analysis

SIO 210: Data analysis methods L. Talley, Fall Sampling and error 2. Basic statistical concepts 3. Time series analysis SIO 210: Data analysis methods L. Talley, Fall 2016 1. Sampling and error 2. Basic statistical concepts 3. Time series analysis 4. Mapping 5. Filtering 6. Space-time data 7. Water mass analysis Reading:

More information

Sequences. We know that the functions can be defined on any subsets of R. As the set of positive integers

Sequences. We know that the functions can be defined on any subsets of R. As the set of positive integers Sequences We know that the functions can be defined on any subsets of R. As the set of positive integers Z + is a subset of R, we can define a function on it in the following manner. f: Z + R f(n) = a

More information

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff

IEOR 165 Lecture 7 1 Bias-Variance Tradeoff IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to

More information

0, otherwise. U = Y 1 Y 2 Hint: Use either the method of distribution functions or a bivariate transformation. (b) Find E(U).

0, otherwise. U = Y 1 Y 2 Hint: Use either the method of distribution functions or a bivariate transformation. (b) Find E(U). 1. Suppose Y U(0, 2) so that the probability density function (pdf) of Y is 1 2, 0 < y < 2 (a) Find the pdf of U = Y 4 + 1. Make sure to note the support. (c) Suppose Y 1, Y 2,..., Y n is an iid sample

More information

If we want to analyze experimental or simulated data we might encounter the following tasks:

If we want to analyze experimental or simulated data we might encounter the following tasks: Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction

More information

Understanding Regressions with Observations Collected at High Frequency over Long Span

Understanding Regressions with Observations Collected at High Frequency over Long Span Understanding Regressions with Observations Collected at High Frequency over Long Span Yoosoon Chang Department of Economics, Indiana University Joon Y. Park Department of Economics, Indiana University

More information

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p

GMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p GMM and SMM Some useful references: 1. Hansen, L. 1982. Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p. 1029-54. 2. Lee, B.S. and B. Ingram. 1991 Simulation estimation

More information

A Primer on Asymptotics

A Primer on Asymptotics A Primer on Asymptotics Eric Zivot Department of Economics University of Washington September 30, 2003 Revised: October 7, 2009 Introduction The two main concepts in asymptotic theory covered in these

More information

Volatility. Gerald P. Dwyer. February Clemson University

Volatility. Gerald P. Dwyer. February Clemson University Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use

More information

Linear Model Under General Variance Structure: Autocorrelation

Linear Model Under General Variance Structure: Autocorrelation Linear Model Under General Variance Structure: Autocorrelation A Definition of Autocorrelation In this section, we consider another special case of the model Y = X β + e, or y t = x t β + e t, t = 1,..,.

More information

Applied Quantitative Methods II

Applied Quantitative Methods II Applied Quantitative Methods II Lecture 4: OLS and Statistics revision Klára Kaĺıšková Klára Kaĺıšková AQM II - Lecture 4 VŠE, SS 2016/17 1 / 68 Outline 1 Econometric analysis Properties of an estimator

More information

Applied Regression Analysis

Applied Regression Analysis Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of

More information

arxiv:hep-ex/ v1 2 Jun 2000

arxiv:hep-ex/ v1 2 Jun 2000 MPI H - V7-000 May 3, 000 Averaging Measurements with Hidden Correlations and Asymmetric Errors Michael Schmelling / MPI for Nuclear Physics Postfach 03980, D-6909 Heidelberg arxiv:hep-ex/0006004v Jun

More information

Stat 704 Data Analysis I Probability Review

Stat 704 Data Analysis I Probability Review 1 / 39 Stat 704 Data Analysis I Probability Review Dr. Yen-Yi Ho Department of Statistics, University of South Carolina A.3 Random Variables 2 / 39 def n: A random variable is defined as a function that

More information

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles

Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036

More information

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME)

WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME) WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME ITAI BENJAMINI, GADY KOZMA, LÁSZLÓ LOVÁSZ, DAN ROMIK, AND GÁBOR TARDOS Abstract. We observe returns of a simple random wal on a finite graph to a fixed node,

More information

Math 016 Lessons Wimayra LUY

Math 016 Lessons Wimayra LUY Math 016 Lessons Wimayra LUY wluy@ccp.edu MATH 016 Lessons LESSON 1 Natural Numbers The set of natural numbers is given by N = {0, 1, 2, 3, 4...}. Natural numbers are used for two main reasons: 1. counting,

More information

Ch 2: Simple Linear Regression

Ch 2: Simple Linear Regression Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component

More information

Multiple Linear Regression

Multiple Linear Regression Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there

More information

Probability Theory and Statistics. Peter Jochumzen

Probability Theory and Statistics. Peter Jochumzen Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................

More information

1 Class Organization. 2 Introduction

1 Class Organization. 2 Introduction Time Series Analysis, Lecture 1, 2018 1 1 Class Organization Course Description Prerequisite Homework and Grading Readings and Lecture Notes Course Website: http://www.nanlifinance.org/teaching.html wechat

More information

Lecture 10: A (Brief) Introduction to Group Theory (See Chapter 3.13 in Boas, 3rd Edition)

Lecture 10: A (Brief) Introduction to Group Theory (See Chapter 3.13 in Boas, 3rd Edition) Lecture 0: A (Brief) Introduction to Group heory (See Chapter 3.3 in Boas, 3rd Edition) Having gained some new experience with matrices, which provide us with representations of groups, and because symmetries

More information

Notes on the Multivariate Normal and Related Topics

Notes on the Multivariate Normal and Related Topics Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions

More information

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)

EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) 1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For

More information

Last Update: April 7, 201 0

Last Update: April 7, 201 0 M ath E S W inter Last Update: April 7, Introduction to Partial Differential Equations Disclaimer: his lecture note tries to provide an alternative approach to the material in Sections.. 5 in the textbook.

More information

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018

Econometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018 Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate

More information

Random Variables and Their Distributions

Random Variables and Their Distributions Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital

More information

Spectral Analysis. Jesús Fernández-Villaverde University of Pennsylvania

Spectral Analysis. Jesús Fernández-Villaverde University of Pennsylvania Spectral Analysis Jesús Fernández-Villaverde University of Pennsylvania 1 Why Spectral Analysis? We want to develop a theory to obtain the business cycle properties of the data. Burns and Mitchell (1946).

More information

F & B Approaches to a simple model

F & B Approaches to a simple model A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys

More information

Lecture 9: Introduction to Kriging

Lecture 9: Introduction to Kriging Lecture 9: Introduction to Kriging Math 586 Beginning remarks Kriging is a commonly used method of interpolation (prediction) for spatial data. The data are a set of observations of some variable(s) of

More information

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA

TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA CHAPTER 6 TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA 6.1. Introduction A time series is a sequence of observations ordered in time. A basic assumption in the time series analysis

More information

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the

Review Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then

More information

Statistical Data Analysis Stat 3: p-values, parameter estimation

Statistical Data Analysis Stat 3: p-values, parameter estimation Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,

More information

DETECTION theory deals primarily with techniques for

DETECTION theory deals primarily with techniques for ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for

More information