SIO 221B, Rudnick adapted from Davis 1. 1 x lim. N x 2 n = 1 N. { x} 1 N. N x = 1 N. N x = 1 ( N N x ) x = 0 (3) = 1 x N 2
|
|
- Phebe Scott
- 5 years ago
- Views:
Transcription
1 SIO B, Rudnick adapted from Davis VII. Sampling errors We do not have access to the true statistics, so we must compute sample statistics. By this we mean that the number of realizations we average over is necessarily finite. We wish to determine the accuracy of our sample statistics to aid in interpretation. he difference between the true and sampling statistics is called the sampling error. We have already defined the true mean he sample mean is defined as x lim { x} x n () x n () ote the use of curly braces as a notation for the sample mean. he distinguishing factor between () and () is that the sample mean is over a finite number of realizations. Consider the problem of estimating the true mean with the sample mean. Suppose the sample mean were calculated an infinite number of times, then we could consider the mean of the sample mean, and compare it to the true mean. he bias is defined to be the mean difference between the sample mean and the true mean E = { x} x = x n x = hus the sample mean has desirable property of zero bias. x n x = ( x ) x = (3) Consider the variance of the sample mean E = ({ x} x ) = x n x (4) Moving the true mean inside the sum results in E = x n x (5) he quantity in square brackets in (5) is the fluctuation, so E = x n (6) he average squared sum of fluctuations is the sum over all elements of the covariance matrix of the fluctuations E = x x n (7) m m= Assuming the fluctuations are independent with equal variance then
2 SIO B, Rudnick adapted from Davis E = δ nm σ = σ m= So the variance decreases as the inverse of. he square root of E is called the root-meansquare error or the standard error, and decreases as the inverse of the square root of (8) E = σ (9) Recognizing this improvement of error by -/ is key to understanding many measures of statistical error. ote the parallel between sampling error and the random walk. ote that the rms error depends on the true variance. Suppose first that we know the true mean so we can estimate the variance as he mean of this estimate is ˆσ ˆσ = { x } () = x = σ () n so this estimate of variance is unbiased. he variance of the estimate of variance is F = Assuming the fluctuations are independent F = ({ x } σ ) = m= ( )σ 4 + x 4 x x n σ 4 () m σ 4 = ( x 4 σ ) 4 (3) So the error in the variance depends on the fourth moment. his is a problem of closure. o determine the error of any statistic we need to know a higher order statistic. hus errors on an estimated statistic are always less well determined than the original estimate. he reality is we don t know the true mean in order to use (), so we must estimate it using (). Consider the following estimator ˆσ = A ( x { x} ) { } (4) where A is a constant to be determined. he obvious choice is A=, but is it optimal in any sense? Let s investigate the bias in the estimate (4) F = ˆσ σ (5) ow we rewrite the estimate (4) in terms of deviation from the mean to get F = A ( x x ) ({ x} x ) { } σ (6) As the sample mean operator acts only on x, expanding the quadratic yields F = A x { } ({ x} x ) σ (7)
3 SIO B, Rudnick adapted from Davis 3 he second term in the angle brackets of (7) is the variance of the sample mean, which for independent data is given by (8). he first term in the angle brackets is just (-). he result is thus F = Aσ σ (8) If we want the estimate (4) to be unbiased, that is F=, then A = So an unbiased estimate of variance is ˆσ = ( x n { x} ) () Clearly, dividing by - rather than by makes little difference when is large. So any bias in the most obvious estimate of variance is small. Also, there are other estimates of variance that might be optimal in other ways, rather than to eliminate bias. For example, consider the variance of the sample variance (9) F = ( ˆσ σ ) () Expanding the quadratic results, and applying the averaging operator to each term results in F = ˆσ 4 σ ˆσ + σ 4 () We already know the second term from the calculation of bias σ ˆσ = Aσ 4 he first term in () is more challenging to evaluate {( { }) } {( { } x ) } {( { }) } { { } + { x } } ˆσ 4 = A x x = A x x x = A x x = A x x x Since the sample average operates only on x and not on x ˆσ 4 = A = A x ({ x } { x } ) { } { x } x { }, we find (3) (4) (5) { } + { x } 4 Expanding the sample means as sums, and applying the averaging operator to each term, results in
4 SIO B, Rudnick adapted from Davis 4 ˆσ 4 = A x x n x m m= x 3 n x + x m p m= p= 4 n x m x p x q m= p= q= (6) So, we have a number of mean products, and we have to sum over all the elements. We assume, as above, that the data are independent. Because nonzero contributions come only from even moments, we find ˆσ 4 = A ( ( )σ 4 + x ) 4 ( 3 ( )σ 4 + x ) 4 + ( 4 3 ( )σ 4 + x ) 4 (7) he factor of 3 in the last term comes from the fact that there are three possible pairs in the fourth-order product. his can be simplified to σˆ 4 = A ( + 3)σ 4 + ( ) x 4 3 (8) We require the fourth moment. If we assume x to have a normal distribution, then x 4 = 3σ 4, and the result is ( )( +) ˆσ 4 = A σ 4 (9) he variance of the sample mean is found by substituting (3) and (9) into () ( ) ( +) F = A σ 4 Aσ 4 + σ 4 (3) We choose the value of A that minimizes the variance by setting the derivative of (3) with respect to A equal to zero, yielding A = (3) + So for minimum variance (3), A is slightly less than, while for zeros bias (9), A is slightly greater than. Maybe a choice of is not so bad after all. he resulting variance and bias, with A given by (3) is F = σ 4 + ; F = σ + So there is a negative bias, which is understood as the value for A in this case is less than that in the unbiased case (9). For completeness, the variance and bias in the unbiased estimator of variance is, with A given by (9), F = σ 4 ; F = (33) where we have assumed the data to be normally distributed. (3)
5 SIO B, Rudnick adapted from Davis 5 Correlated Observations Real observations are almost always correlated. All the estimates of sampling error we have considered so far have assumed that the realizations are independent. An example of correlated observations occurs when we have closely spaced samples of a variable with very long scales, as daily samples of temperature off the Scripps pier in the presence of ESO s and PDO s. he time series in Figure contains time scales relative long compared to the separation between observations A and B, but comparable to the separation between points A and C. Recall the expression for the sample mean (7), that consists of a sum over all elements in the covariance matrix. Let s relax the assumption of independence, but assume now that the statistics are stationary, so x n x m = σ ρ( n m ) (34) where ρ is correlation. hat is the covariance depends only on the difference in sample index. Assuming that the samples are evenly spaced, the mean square error of the sample mean is E = σ n ρ( n ) (35) n= 3 C x A B t Figure. A serially correlated time series. Observations A and B are closely spaced relative to the dominant time scales. Observations A and C are relatively far apart. hus observation C is more likely than B to be independent of A.
6 SIO B, Rudnick adapted from Davis 6 Understand that the index n in (35) refers to the diagonal of the covariance matrix, the covariance matrix is symmetric, and the covariance matrix is such that each diagonal band is made up of equal values. If the data are uncorrelated, then the only nonzero term in (35) is for n=, and (8) is recovered. he variance of the sample mean can be written E = σ E (36) where the equivalent number of independent samples in serially correlated data E is E = n= n ρ n E is sometimes called the equivalent number of degrees of freedom. If the number of samples is large compared to the value of n where correlation ρ vanishes, then E = n= hese ideas can be extended to continuous time series, which clearly have correlated samples. In this case the sample mean is ( ) ρ n ( ) (37) (38) { x} = x( t)dt (39) he mean square error of the sample mean is E = x ( t ) x ( t ) dt dt (4) Assuming stationary statistics then E = σ he number equivalent degrees of freedom is E = t ρ( t )dt (4) t ρ( t )dt aking the limit as the time series is infinitely long, the mean square error is E = σ (4) ρ( t )dt (43) he integral of the correlation is thus an important quantity, often called the integral time scale. he ratio of the time series length to the integral time scales tells how many independent samples there are in a time series. here is an illuminating relationship between the spectrum and the variance of the sample mean. ote that the spectrum and the covariance are Fourier transform pairs
7 SIO B, Rudnick adapted from Davis 7 ( ) = C t S f ( )e iπ ft dt (44) Here S is the spectrum, C the covariance, and f frequency. Because C = σ ρ, and using (43), the following relationship holds for time series long compared to the integral time scale E = S( ) his somewhat surprising result suggests that the spectrum at zero frequency, that is the lowest frequencies in a continuous time series, determine the error in the sample mean. High frequencies in a continuous time series are effectively averaged out, but the lowest frequencies cause error. A series of independent data have a white spectrum, and that a white spectrum is nonzero at low frequency is key to the effect on the error in the sample mean. Confidence limits he concept of confidence limits is a bit subtle. he idea is to come up with an interval and state the probability that the true value of a statistic is in that interval. Suppose we could calculate a sample mean an infinite number of times, and come up with its true distribution. (Figure ) Given that distribution, we could say that the next sample mean we calculate would fall between lower and upper limits xl and xu with a probability of (45) F {x} (r) x L <x> x U r Figure. he probability density function of the sample mean.
8 SIO B, Rudnick adapted from Davis 8 x U P = F x ( )dr (46) x L { } r After we have calculated the next sample mean, it either does or does not fall within these limits. hat is the probability it is within some interval is either or. Since we claim to know the true mean first before calculating the sample mean, this is not a satisfactory procedure. he question we really want to ask is: What is the range of true means likely to have led to the sample mean we just calculated? he upper limit is the largest true mean for which there is a small ε probability of getting a smaller sample mean than the one observed. he lower limit is the smallest true mean for which there is an ε probability of getting a larger sample mean than the one observed. (Figure 3) If the sample mean we observed is ˆm and the cumulative distribution of the sample mean is D { x} with assumed lower and upper true means of xl and xu, then the following probability statements are appropriate. D { x} ( ˆm x L ) = ε; D { x} ˆm x U ( ) = ε (47).4..8 F {x}.6.4. x L ˆm x U Figure 3. he confidence interval for the sample mean is bounded by the largest and smallest true means that were likely to have led to the observed sample mean. he area underneath the blue curve to the right of ˆm is the probability ε of getting a larger sample mean than the one observed, given a true mean of xl. Similarly, the area under the green curve to the left of ˆm is ε.
9 SIO B, Rudnick adapted from Davis 9 he story we need to tell to have confidence limits is to say we know the distribution, but we don t know the mean. he confidence interval defines as in (47) is said to be at the (-ε)% level. o make more progress, we need an assumed distribution. Because the sample mean is a sum of random variables, its distribution should approach normal with a standard deviation given by (9) as the number of degrees of freedom gets large. he lower and upper limits (47) are then x L = ˆm σ z ε ; x U = ˆm + σ z ε where z ε is the inverse normal cumulative distribution function with zero mean and unit standard deviation evaluated at ε. (demo) ote that the limits are related to the estimated mean as a sum. he chi-squared χ distribution is useful for finding confidence intervals for sample variances. For example, estimates of spectra are often assumed to have a χ distribution. Consider an estimate of the sample variance made by the sum of squares of independent, identically and normally distributed variables ˆσ = σ x n (48) F L ˆσ U Figure 4. he confidence interval for the sample variance. he different widths of the two χ pdfs result because a larger mean χ variable has a larger variance, as the χ pdf must pass through the origin.
10 SIO B, Rudnick adapted from Davis where he variable x = ; x = (49) x n = ˆσ is a χ variable with degrees of freedom. he mean of this variable is and its variance is x n he lower and upper limits are then where χ ;ε x n = x n σ (5) = (5) = x n x m x n + m= = ( ) = σ L = ˆσ χ ; ε ; σ U = ˆσ (5) (53) χ ;ε is the inverse χ cumulative distribution function with degrees of freedom evaluated at ε. (Figure 4) As a χ variable is a sum, the χ distribution approaches normal as gets large. In contrast to the confidence limits for a sample mean, the limits for sample variance are related to the estimate as a ratio.
Analytical Methods. Session 3: Statistics II. UCL Department of Civil, Environmental & Geomatic Engineering. Analytical Methods.
Analytical Methods Session 3: Statistics II More statistical tests Quite a few more questions that we might want to ask about data that we have. Is one value significantly different to the rest, or to
More informationSIO 211B, Rudnick, adapted from Davis 1
SIO 211B, Rudnick, adapted from Davis 1 XVII.Empirical orthogonal functions Often in oceanography we collect large data sets that are time series at a group of locations. Moored current meter arrays do
More informationEconomics 583: Econometric Theory I A Primer on Asymptotics
Economics 583: Econometric Theory I A Primer on Asymptotics Eric Zivot January 14, 2013 The two main concepts in asymptotic theory that we will use are Consistency Asymptotic Normality Intuition consistency:
More informationTerminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the
More informationStatistics and Data Analysis
Statistics and Data Analysis The Crash Course Physics 226, Fall 2013 "There are three kinds of lies: lies, damned lies, and statistics. Mark Twain, allegedly after Benjamin Disraeli Statistics and Data
More informationStochastic Processes. A stochastic process is a function of two variables:
Stochastic Processes Stochastic: from Greek stochastikos, proceeding by guesswork, literally, skillful in aiming. A stochastic process is simply a collection of random variables labelled by some parameter:
More informationELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process
Department of Electrical Engineering University of Arkansas ELEG 3143 Probability & Stochastic Process Ch. 6 Stochastic Process Dr. Jingxian Wu wuj@uark.edu OUTLINE 2 Definition of stochastic process (random
More information2008 Winton. Statistical Testing of RNGs
1 Statistical Testing of RNGs Criteria for Randomness For a sequence of numbers to be considered a sequence of randomly acquired numbers, it must have two basic statistical properties: Uniformly distributed
More informationWhite Noise Processes (Section 6.2)
White Noise Processes (Section 6.) Recall that covariance stationary processes are time series, y t, such. E(y t ) = µ for all t. Var(y t ) = σ for all t, σ < 3. Cov(y t,y t-τ ) = γ(τ) for all t and τ
More informationA Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit
A Detailed Look at a Discrete Randomw Walk with Spatially Dependent Moments and Its Continuum Limit David Vener Department of Mathematics, MIT May 5, 3 Introduction In 8.366, we discussed the relationship
More informationECON The Simple Regression Model
ECON 351 - The Simple Regression Model Maggie Jones 1 / 41 The Simple Regression Model Our starting point will be the simple regression model where we look at the relationship between two variables In
More informationwhere r n = dn+1 x(t)
Random Variables Overview Probability Random variables Transforms of pdfs Moments and cumulants Useful distributions Random vectors Linear transformations of random vectors The multivariate normal distribution
More informationEstimators as Random Variables
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maimum likelihood Consistency Confidence intervals Properties of the mean estimator Introduction Up until
More informationENSC327 Communications Systems 19: Random Processes. Jie Liang School of Engineering Science Simon Fraser University
ENSC327 Communications Systems 19: Random Processes Jie Liang School of Engineering Science Simon Fraser University 1 Outline Random processes Stationary random processes Autocorrelation of random processes
More informationLecture 3: Statistical sampling uncertainty
Lecture 3: Statistical sampling uncertainty c Christopher S. Bretherton Winter 2015 3.1 Central limit theorem (CLT) Let X 1,..., X N be a sequence of N independent identically-distributed (IID) random
More informationMultivariate Time Series
Multivariate Time Series Notation: I do not use boldface (or anything else) to distinguish vectors from scalars. Tsay (and many other writers) do. I denote a multivariate stochastic process in the form
More informationTime series models in the Frequency domain. The power spectrum, Spectral analysis
ime series models in the Frequency domain he power spectrum, Spectral analysis Relationship between the periodogram and the autocorrelations = + = ( ) ( ˆ α ˆ ) β I Yt cos t + Yt sin t t= t= ( ( ) ) cosλ
More informationPower Spectral Density of Digital Modulation Schemes
Digital Communication, Continuation Course Power Spectral Density of Digital Modulation Schemes Mikael Olofsson Emil Björnson Department of Electrical Engineering ISY) Linköping University, SE-581 83 Linköping,
More informationXVI. Objective mapping
SIO 211B, Rudnick, adapted from Davis 1 XVI. Objective mapping An objective map is the minimum mean-square error estimate of a continuous function of a variable, given discrete data. The interpolation
More informationProbabilities & Statistics Revision
Probabilities & Statistics Revision Christopher Ting Christopher Ting http://www.mysmu.edu/faculty/christophert/ : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036 January 6, 2017 Christopher Ting QF
More informationLecture 4: Testing Stuff
Lecture 4: esting Stuff. esting Hypotheses usually has three steps a. First specify a Null Hypothesis, usually denoted, which describes a model of H 0 interest. Usually, we express H 0 as a restricted
More informationStatistics. Lent Term 2015 Prof. Mark Thomson. 2: The Gaussian Limit
Statistics Lent Term 2015 Prof. Mark Thomson Lecture 2 : The Gaussian Limit Prof. M.A. Thomson Lent Term 2015 29 Lecture Lecture Lecture Lecture 1: Back to basics Introduction, Probability distribution
More informationStatistics for scientists and engineers
Statistics for scientists and engineers February 0, 006 Contents Introduction. Motivation - why study statistics?................................... Examples..................................................3
More informationSo far we discussed random number generators that need to have the maximum length period.
So far we discussed random number generators that need to have the maximum length period. Even the increment series has the maximum length period yet it is by no means random How do we decide if a generator
More informationTime Series 3. Robert Almgren. Sept. 28, 2009
Time Series 3 Robert Almgren Sept. 28, 2009 Last time we discussed two main categories of linear models, and their combination. Here w t denotes a white noise: a stationary process with E w t ) = 0, E
More information1 Least Squares Estimation - multiple regression.
Introduction to multiple regression. Fall 2010 1 Least Squares Estimation - multiple regression. Let y = {y 1,, y n } be a n 1 vector of dependent variable observations. Let β = {β 0, β 1 } be the 2 1
More informationFitting a Straight Line to Data
Fitting a Straight Line to Data Thanks for your patience. Finally we ll take a shot at real data! The data set in question is baryonic Tully-Fisher data from http://astroweb.cwru.edu/sparc/btfr Lelli2016a.mrt,
More informationRegression #5: Confidence Intervals and Hypothesis Testing (Part 1)
Regression #5: Confidence Intervals and Hypothesis Testing (Part 1) Econ 671 Purdue University Justin L. Tobias (Purdue) Regression #5 1 / 24 Introduction What is a confidence interval? To fix ideas, suppose
More informationConsider the joint probability, P(x,y), shown as the contours in the figure above. P(x) is given by the integral of P(x,y) over all values of y.
ATMO/OPTI 656b Spring 009 Bayesian Retrievals Note: This follows the discussion in Chapter of Rogers (000) As we have seen, the problem with the nadir viewing emission measurements is they do not contain
More informationParameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn!
Parameter estimation! and! forecasting! Cristiano Porciani! AIfA, Uni-Bonn! Questions?! C. Porciani! Estimation & forecasting! 2! Cosmological parameters! A branch of modern cosmological research focuses
More informationThe regression model with one stochastic regressor.
The regression model with one stochastic regressor. 3150/4150 Lecture 6 Ragnar Nymoen 30 January 2012 We are now on Lecture topic 4 The main goal in this lecture is to extend the results of the regression
More informationRecall the Basics of Hypothesis Testing
Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE
More informationA=randn(500,100); mu=mean(a); sigma_a=std(a); std_a=sigma_a/sqrt(500); [std(mu) mean(std_a)] % compare standard deviation of means % vs standard error
UCSD SIOC 221A: (Gille) 1 Reading: Bendat and Piersol, Ch. 5.2.1 Lecture 10: Recap Last time we looked at the sinc function, windowing, and detrending with an eye to reducing edge effects in our spectra.
More informationMS&E 226: Small Data
MS&E 226: Small Data Lecture 12: Frequentist properties of estimators (v4) Ramesh Johari ramesh.johari@stanford.edu 1 / 39 Frequentist inference 2 / 39 Thinking like a frequentist Suppose that for some
More informationSimple Linear Regression
Simple Linear Regression ST 430/514 Recall: A regression model describes how a dependent variable (or response) Y is affected, on average, by one or more independent variables (or factors, or covariates)
More informationBusiness Economics BUSINESS ECONOMICS. PAPER No. : 8, FUNDAMENTALS OF ECONOMETRICS MODULE No. : 3, GAUSS MARKOV THEOREM
Subject Business Economics Paper No and Title Module No and Title Module Tag 8, Fundamentals of Econometrics 3, The gauss Markov theorem BSE_P8_M3 1 TABLE OF CONTENTS 1. INTRODUCTION 2. ASSUMPTIONS OF
More informationWeek 9 Generators, duality, change of measure
Week 9 Generators, duality, change of measure Jonathan Goodman November 18, 013 1 Generators This section describes a common abstract way to describe many of the differential equations related to Markov
More informationProperties of the least squares estimates
Properties of the least squares estimates 2019-01-18 Warmup Let a and b be scalar constants, and X be a scalar random variable. Fill in the blanks E ax + b) = Var ax + b) = Goal Recall that the least squares
More informationEcon 836 Final Exam. 2 w N 2 u N 2. 2 v N
1) [4 points] Let Econ 836 Final Exam Y Xβ+ ε, X w+ u, w N w~ N(, σi ), u N u~ N(, σi ), ε N ε~ Nu ( γσ, I ), where X is a just one column. Let denote the OLS estimator, and define residuals e as e Y X.
More informationTaylor Series and Series Convergence (Online)
7in 0in Felder c02_online.te V3 - February 9, 205 9:5 A.M. Page CHAPTER 2 Taylor Series and Series Convergence (Online) 2.8 Asymptotic Epansions In introductory calculus classes the statement this series
More informationStochastic Processes: I. consider bowl of worms model for oscilloscope experiment:
Stochastic Processes: I consider bowl of worms model for oscilloscope experiment: SAPAscope 2.0 / 0 1 RESET SAPA2e 22, 23 II 1 stochastic process is: Stochastic Processes: II informally: bowl + drawing
More informationPhysics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester
Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability
More information2.3 Estimating PDFs and PDF Parameters
.3 Estimating PDFs and PDF Parameters estimating means - discrete and continuous estimating variance using a known mean estimating variance with an estimated mean estimating a discrete pdf estimating a
More informationcovariance function, 174 probability structure of; Yule-Walker equations, 174 Moving average process, fluctuations, 5-6, 175 probability structure of
Index* The Statistical Analysis of Time Series by T. W. Anderson Copyright 1971 John Wiley & Sons, Inc. Aliasing, 387-388 Autoregressive {continued) Amplitude, 4, 94 case of first-order, 174 Associated
More informationSolution. The relationship between cartesian coordinates (x, y) and polar coordinates (r, θ) is given by. (x, y) = (r cos θ, r sin θ).
Problem 1. Let p 1 be the point having polar coordinates r = 1 and θ = π. Let p 2 be the point having polar coordinates r = 1 and θ = π/2. Find the Euclidean distance between p 1 and p 2. The relationship
More informationChapter 12: An introduction to Time Series Analysis. Chapter 12: An introduction to Time Series Analysis
Chapter 12: An introduction to Time Series Analysis Introduction In this chapter, we will discuss forecasting with single-series (univariate) Box-Jenkins models. The common name of the models is Auto-Regressive
More informationProblems. Suppose both models are fitted to the same data. Show that SS Res, A SS Res, B
Simple Linear Regression 35 Problems 1 Consider a set of data (x i, y i ), i =1, 2,,n, and the following two regression models: y i = β 0 + β 1 x i + ε, (i =1, 2,,n), Model A y i = γ 0 + γ 1 x i + γ 2
More informationRegression Models - Introduction
Regression Models - Introduction In regression models, two types of variables that are studied: A dependent variable, Y, also called response variable. It is modeled as random. An independent variable,
More informationComputational Data Analysis!
12.714 Computational Data Analysis! Alan Chave (alan@whoi.edu)! Thomas Herring (tah@mit.edu),! http://geoweb.mit.edu/~tah/12.714! Introduction to Spectral Analysis! Topics Today! Aspects of Time series
More informationM(t) = 1 t. (1 t), 6 M (0) = 20 P (95. X i 110) i=1
Math 66/566 - Midterm Solutions NOTE: These solutions are for both the 66 and 566 exam. The problems are the same until questions and 5. 1. The moment generating function of a random variable X is M(t)
More informationTopics in Probability and Statistics
Topics in Probability and tatistics A Fundamental Construction uppose {, P } is a sample space (with probability P), and suppose X : R is a random variable. The distribution of X is the probability P X
More informationIntroduction to Econometrics Final Examination Fall 2006 Answer Sheet
Introduction to Econometrics Final Examination Fall 2006 Answer Sheet Please answer all of the questions and show your work. If you think a question is ambiguous, clearly state how you interpret it before
More informationTreatment of Error in Experimental Measurements
in Experimental Measurements All measurements contain error. An experiment is truly incomplete without an evaluation of the amount of error in the results. In this course, you will learn to use some common
More informationChapter 3. Introducing Groups
Chapter 3 Introducing Groups We need a super-mathematics in which the operations are as unknown as the quantities they operate on, and a super-mathematician who does not know what he is doing when he performs
More informationMultiple Random Variables
Multiple Random Variables Joint Probability Density Let X and Y be two random variables. Their joint distribution function is F ( XY x, y) P X x Y y. F XY ( ) 1, < x
More informationBasic Probability Reference Sheet
February 27, 2001 Basic Probability Reference Sheet 17.846, 2001 This is intended to be used in addition to, not as a substitute for, a textbook. X is a random variable. This means that X is a variable
More information2.830J / 6.780J / ESD.63J Control of Manufacturing Processes (SMA 6303) Spring 2008
MIT OpenCourseWare http://ocw.mit.edu 2.830J / 6.780J / ESD.63J Control of Processes (SMA 6303) Spring 2008 For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.
More informationVIII. Coherence and Transfer Function Applications A. Coherence Function Estimates
VIII. Coherence and Transfer Function Applications A. Coherence Function Estimates Consider the application of these ideas to the specific problem of atmospheric turbulence measurements outlined in Figure
More informationPHYS 6710: Nuclear and Particle Physics II
The ellipse of constant chi-square is described by The error ellipse To describe the tilt of the error ellipse, we need to consider the following: For two parameters a,b, there are new coordinates a',
More informationAn example to illustrate frequentist and Bayesian approches
Frequentist_Bayesian_Eample An eample to illustrate frequentist and Bayesian approches This is a trivial eample that illustrates the fundamentally different points of view of the frequentist and Bayesian
More information3.3 Estimator quality, confidence sets and bootstrapping
Estimator quality, confidence sets and bootstrapping 109 3.3 Estimator quality, confidence sets and bootstrapping A comparison of two estimators is always a matter of comparing their respective distributions.
More informationReview of Classical Least Squares. James L. Powell Department of Economics University of California, Berkeley
Review of Classical Least Squares James L. Powell Department of Economics University of California, Berkeley The Classical Linear Model The object of least squares regression methods is to model and estimate
More informationSignals and Spectra - Review
Signals and Spectra - Review SIGNALS DETERMINISTIC No uncertainty w.r.t. the value of a signal at any time Modeled by mathematical epressions RANDOM some degree of uncertainty before the signal occurs
More information2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES
2. SPECTRAL ANALYSIS APPLIED TO STOCHASTIC PROCESSES 2.0 THEOREM OF WIENER- KHINTCHINE An important technique in the study of deterministic signals consists in using harmonic functions to gain the spectral
More informationSimple Linear Regression for the Climate Data
Prediction Prediction Interval Temperature 0.2 0.0 0.2 0.4 0.6 0.8 320 340 360 380 CO 2 Simple Linear Regression for the Climate Data What do we do with the data? y i = Temperature of i th Year x i =CO
More information1 Linear Difference Equations
ARMA Handout Jialin Yu 1 Linear Difference Equations First order systems Let {ε t } t=1 denote an input sequence and {y t} t=1 sequence generated by denote an output y t = φy t 1 + ε t t = 1, 2,... with
More informationSIO 210: Data analysis methods L. Talley, Fall Sampling and error 2. Basic statistical concepts 3. Time series analysis
SIO 210: Data analysis methods L. Talley, Fall 2016 1. Sampling and error 2. Basic statistical concepts 3. Time series analysis 4. Mapping 5. Filtering 6. Space-time data 7. Water mass analysis Reading:
More informationSequences. We know that the functions can be defined on any subsets of R. As the set of positive integers
Sequences We know that the functions can be defined on any subsets of R. As the set of positive integers Z + is a subset of R, we can define a function on it in the following manner. f: Z + R f(n) = a
More informationIEOR 165 Lecture 7 1 Bias-Variance Tradeoff
IEOR 165 Lecture 7 Bias-Variance Tradeoff 1 Bias-Variance Tradeoff Consider the case of parametric regression with β R, and suppose we would like to analyze the error of the estimate ˆβ in comparison to
More information0, otherwise. U = Y 1 Y 2 Hint: Use either the method of distribution functions or a bivariate transformation. (b) Find E(U).
1. Suppose Y U(0, 2) so that the probability density function (pdf) of Y is 1 2, 0 < y < 2 (a) Find the pdf of U = Y 4 + 1. Make sure to note the support. (c) Suppose Y 1, Y 2,..., Y n is an iid sample
More informationIf we want to analyze experimental or simulated data we might encounter the following tasks:
Chapter 1 Introduction If we want to analyze experimental or simulated data we might encounter the following tasks: Characterization of the source of the signal and diagnosis Studying dependencies Prediction
More informationUnderstanding Regressions with Observations Collected at High Frequency over Long Span
Understanding Regressions with Observations Collected at High Frequency over Long Span Yoosoon Chang Department of Economics, Indiana University Joon Y. Park Department of Economics, Indiana University
More informationGMM and SMM. 1. Hansen, L Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p
GMM and SMM Some useful references: 1. Hansen, L. 1982. Large Sample Properties of Generalized Method of Moments Estimators, Econometrica, 50, p. 1029-54. 2. Lee, B.S. and B. Ingram. 1991 Simulation estimation
More informationA Primer on Asymptotics
A Primer on Asymptotics Eric Zivot Department of Economics University of Washington September 30, 2003 Revised: October 7, 2009 Introduction The two main concepts in asymptotic theory covered in these
More informationVolatility. Gerald P. Dwyer. February Clemson University
Volatility Gerald P. Dwyer Clemson University February 2016 Outline 1 Volatility Characteristics of Time Series Heteroskedasticity Simpler Estimation Strategies Exponentially Weighted Moving Average Use
More informationLinear Model Under General Variance Structure: Autocorrelation
Linear Model Under General Variance Structure: Autocorrelation A Definition of Autocorrelation In this section, we consider another special case of the model Y = X β + e, or y t = x t β + e t, t = 1,..,.
More informationApplied Quantitative Methods II
Applied Quantitative Methods II Lecture 4: OLS and Statistics revision Klára Kaĺıšková Klára Kaĺıšková AQM II - Lecture 4 VŠE, SS 2016/17 1 / 68 Outline 1 Econometric analysis Properties of an estimator
More informationApplied Regression Analysis
Applied Regression Analysis Chapter 3 Multiple Linear Regression Hongcheng Li April, 6, 2013 Recall simple linear regression 1 Recall simple linear regression 2 Parameter Estimation 3 Interpretations of
More informationarxiv:hep-ex/ v1 2 Jun 2000
MPI H - V7-000 May 3, 000 Averaging Measurements with Hidden Correlations and Asymmetric Errors Michael Schmelling / MPI for Nuclear Physics Postfach 03980, D-6909 Heidelberg arxiv:hep-ex/0006004v Jun
More informationStat 704 Data Analysis I Probability Review
1 / 39 Stat 704 Data Analysis I Probability Review Dr. Yen-Yi Ho Department of Statistics, University of South Carolina A.3 Random Variables 2 / 39 def n: A random variable is defined as a function that
More informationWeek 5 Quantitative Analysis of Financial Markets Characterizing Cycles
Week 5 Quantitative Analysis of Financial Markets Characterizing Cycles Christopher Ting http://www.mysmu.edu/faculty/christophert/ Christopher Ting : christopherting@smu.edu.sg : 6828 0364 : LKCSB 5036
More informationWAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME)
WAITING FOR A BAT TO FLY BY (IN POLYNOMIAL TIME ITAI BENJAMINI, GADY KOZMA, LÁSZLÓ LOVÁSZ, DAN ROMIK, AND GÁBOR TARDOS Abstract. We observe returns of a simple random wal on a finite graph to a fixed node,
More informationMath 016 Lessons Wimayra LUY
Math 016 Lessons Wimayra LUY wluy@ccp.edu MATH 016 Lessons LESSON 1 Natural Numbers The set of natural numbers is given by N = {0, 1, 2, 3, 4...}. Natural numbers are used for two main reasons: 1. counting,
More informationCh 2: Simple Linear Regression
Ch 2: Simple Linear Regression 1. Simple Linear Regression Model A simple regression model with a single regressor x is y = β 0 + β 1 x + ɛ, where we assume that the error ɛ is independent random component
More informationMultiple Linear Regression
Multiple Linear Regression Simple linear regression tries to fit a simple line between two variables Y and X. If X is linearly related to Y this explains some of the variability in Y. In most cases, there
More informationProbability Theory and Statistics. Peter Jochumzen
Probability Theory and Statistics Peter Jochumzen April 18, 2016 Contents 1 Probability Theory And Statistics 3 1.1 Experiment, Outcome and Event................................ 3 1.2 Probability............................................
More information1 Class Organization. 2 Introduction
Time Series Analysis, Lecture 1, 2018 1 1 Class Organization Course Description Prerequisite Homework and Grading Readings and Lecture Notes Course Website: http://www.nanlifinance.org/teaching.html wechat
More informationLecture 10: A (Brief) Introduction to Group Theory (See Chapter 3.13 in Boas, 3rd Edition)
Lecture 0: A (Brief) Introduction to Group heory (See Chapter 3.3 in Boas, 3rd Edition) Having gained some new experience with matrices, which provide us with representations of groups, and because symmetries
More informationNotes on the Multivariate Normal and Related Topics
Version: July 10, 2013 Notes on the Multivariate Normal and Related Topics Let me refresh your memory about the distinctions between population and sample; parameters and statistics; population distributions
More informationEC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix)
1 EC212: Introduction to Econometrics Review Materials (Wooldridge, Appendix) Taisuke Otsu London School of Economics Summer 2018 A.1. Summation operator (Wooldridge, App. A.1) 2 3 Summation operator For
More informationLast Update: April 7, 201 0
M ath E S W inter Last Update: April 7, Introduction to Partial Differential Equations Disclaimer: his lecture note tries to provide an alternative approach to the material in Sections.. 5 in the textbook.
More informationEconometrics I KS. Module 2: Multivariate Linear Regression. Alexander Ahammer. This version: April 16, 2018
Econometrics I KS Module 2: Multivariate Linear Regression Alexander Ahammer Department of Economics Johannes Kepler University of Linz This version: April 16, 2018 Alexander Ahammer (JKU) Module 2: Multivariate
More informationRandom Variables and Their Distributions
Chapter 3 Random Variables and Their Distributions A random variable (r.v.) is a function that assigns one and only one numerical value to each simple event in an experiment. We will denote r.vs by capital
More informationSpectral Analysis. Jesús Fernández-Villaverde University of Pennsylvania
Spectral Analysis Jesús Fernández-Villaverde University of Pennsylvania 1 Why Spectral Analysis? We want to develop a theory to obtain the business cycle properties of the data. Burns and Mitchell (1946).
More informationF & B Approaches to a simple model
A6523 Signal Modeling, Statistical Inference and Data Mining in Astrophysics Spring 215 http://www.astro.cornell.edu/~cordes/a6523 Lecture 11 Applications: Model comparison Challenges in large-scale surveys
More informationLecture 9: Introduction to Kriging
Lecture 9: Introduction to Kriging Math 586 Beginning remarks Kriging is a commonly used method of interpolation (prediction) for spatial data. The data are a set of observations of some variable(s) of
More informationTIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA
CHAPTER 6 TIME SERIES ANALYSIS AND FORECASTING USING THE STATISTICAL MODEL ARIMA 6.1. Introduction A time series is a sequence of observations ordered in time. A basic assumption in the time series analysis
More informationReview Quiz. 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the
Review Quiz 1. Prove that in a one-dimensional canonical exponential family, the complete and sufficient statistic achieves the Cramér Rao lower bound (CRLB). That is, if where { } and are scalars, then
More informationStatistical Data Analysis Stat 3: p-values, parameter estimation
Statistical Data Analysis Stat 3: p-values, parameter estimation London Postgraduate Lectures on Particle Physics; University of London MSci course PH4515 Glen Cowan Physics Department Royal Holloway,
More informationDETECTION theory deals primarily with techniques for
ADVANCED SIGNAL PROCESSING SE Optimum Detection of Deterministic and Random Signals Stefan Tertinek Graz University of Technology turtle@sbox.tugraz.at Abstract This paper introduces various methods for
More information