Universal, non-asymptotic confidence sets for circular means

Size: px
Start display at page:

Download "Universal, non-asymptotic confidence sets for circular means"

Transcription

1 Universal, non-asymptotic confidence sets for circular means Institut für Mathematik Technische Universität Ilmenau 2nd Conference on Geometric Science of Information October 2015 École polytechnique, Palaiseau Institut für Mathematik, Technische Universität Ilmenau Page 1 / 20

2 Circular / directional data given: Z 1,..., Z n i.i.d. Z random variables on the circle S 1 examples: wind directions directions of a paleomagnetic field (ignoring vertical component) time of admittance to a hospital unit Problem: the circle is not a vector space how to define a mean for circular data? Introduction Institut für Mathematik, Technische Universität Ilmenau Page 2 / 20

3 Circular (or extrinsic) means embed S 1 in C: S 1 = {z C : z = 1} projection: π : C \ {0} S 1, π(z) = z z Euclidean sample mean: Z n := 1 n n i=1 Z i circular (a.k.a. extrinsic) sample mean: ˆµ n = argmin ζ S 1 Z n ζ 2 = π( Z n ) Euclidean population mean: E Z = C Z dp circular population mean: µ = argmin ζ S 1 E Z ζ 2 = π(e Z) unique iff Z n 0 or E Z 0, respectively see e.g. Mardia & Jupp: Directional Statistics (2000). Introduction Institut für Mathematik, Technische Universität Ilmenau Page 3 / 20

4 Aim Construct confidence sets for the circular population mean (set) µ on the circle S 1 Introduction Institut für Mathematik, Technische Universität Ilmenau Page 4 / 20

5 Asymptotics projection π : C \ {0} S 1 is differentiable, so by δ-method: Theorem (central limit theorem) If E Z 0, i.e. if µ is unique, then n Arg(µ 1ˆµ n ) D N (0, E( Im(µ 1 Z) 2) ) E Z 2 where Arg : C \ {0} ( π, π] R (and arbitrary at 0). note: asymptotic variance is small if E Z is large, and/or E ( Im(µ 1 Z) 2), i.e. the variance orthogonal to µ, is small. see e.g. Jammalamadaka & SenGupta: Topics in Circular Statistics (2001). Introduction Institut für Mathematik, Technische Universität Ilmenau Page 5 / 20

6 Asymptotics (cont.) consistent estimation of asympt. variance + Slutsky s lemma gives: Corollary (asymptotic confidence sets) If E Z 0 let δ A = q 1 α 2 n (Im(ˆµ 1 n Z k )) 2 n Z n k=1 where q 1 α denotes the (1 α 2 2 )-quantile of the standard normal distribution N (0, 1). Then C A = {ζ S 1 : Arg(ζ 1ˆµ n ) < δ A } is an asymptotic confidence set of level (1 α). Introduction Institut für Mathematik, Technische Universität Ilmenau Page 6 / 20

7 Why not to use asymptotics needs assumption E Z 0 should be tested, then needs post-test analysis coverage not guaranteed for finite sample sizes see example Introduction Institut für Mathematik, Technische Universität Ilmenau Page 7 / 20

8 Aim Construct confidence sets for the circular population mean (set) µ on the circle S 1 which are universal, i.e. no distributional assumptions (besides i.i.d.) non-asymptotic construction as acceptance region of a corresponding test Introduction Institut für Mathematik, Technische Universität Ilmenau Page 8 / 20

9 Bounding the deviation from the mean Theorem (Hoeffding s inequality) If U 1,..., U n are independent random variables taking values in the bounded interval [a, b] with < a < b < then Ū n = 1 n n k=1 U k with E Ū n = ν fulfills P(Ū n ν t) for any t (0, b ν). [ ( ν a ) ν a+t ( b ν ) ] n b ν t b a ν a+t b ν t = β [a,b] (t, ν) note: β [a,b] (t, ν) exp( 2nt 2 /(b a) 2 ) worse estimate (often called Hoeffding s inequality) β [a,b] (t, ν) = γ has unique solution t [a,b] (γ, ν) for γ (( ) ν a n b a, 1 ) which can easily be computed numerically First construction Institut für Mathematik, Technische Universität Ilmenau Page 9 / 20 see Hoeffding (1963).

10 First construction testing hypothesis E Z = 0: let s 0 = t [ 1,1] ( α 4, 0), exists if α 4 > 2 n 0 First construction Institut für Mathematik, Technische Universität Ilmenau Page 10 / 20

11 First construction testing hypothesis E Z = 0: let s 0 = t [ 1,1] ( α 4, 0), exists if α 4 > 2 n 0 s 0 P(Re Z n s 0 ) α 4 First construction Institut für Mathematik, Technische Universität Ilmenau Page 10 / 20

12 First construction testing hypothesis E Z = 0: let s 0 = t [ 1,1] ( α 4, 0), exists if α 4 > 2 n P(Re Z n s 0 ) α 4 0 s 0 P(Re Z n s 0 ) α 4 First construction Institut für Mathematik, Technische Universität Ilmenau Page 10 / 20

13 First construction testing hypothesis E Z = 0: let s 0 = t [ 1,1] ( α 4, 0), exists if α 4 > 2 n P(Im Z n s 0 ) α 4 s 0 P(Re Z n s 0 ) α 4 0 s 0 P(Re Z n s 0 ) α 4 First construction Institut für Mathematik, Technische Universität Ilmenau Page 10 / 20

14 First construction testing hypothesis E Z = 0: let s 0 = t [ 1,1] ( α 4, 0), exists if α 4 > 2 n P(Im Z n s 0 ) α 4 s 0 P(Re Z n s 0 ) α 4 0 s 0 P(Re Z n s 0 ) α 4 P(Im Z n s 0 ) α 4 First construction Institut für Mathematik, Technische Universität Ilmenau Page 10 / 20

15 First construction testing hypothesis E Z = 0: let s 0 = t [ 1,1] ( α 4, 0), exists if α 4 > 2 n P(Im Z n s 0 ) α 4 s 0 P(Re Z n s 0 ) α 4 0 s 0 P(Re Z n s 0 ) α 4 P(Im Z n s 0 ) α 4 P( Z n 2s 0 ) α First construction Institut für Mathematik, Technische Universität Ilmenau Page 10 / 20

16 First construction (cont.) testing hypothesis E Z = λζ, ζ S 1, λ > 0: ζ λζ 0 First construction Institut für Mathematik, Technische Universität Ilmenau Page 11 / 20

17 First construction (cont.) testing hypothesis E Z = λζ, ζ S 1, λ > 0: ζ λζ 0 P(Re ζ 1 Z n s 0 ) α 4 First construction Institut für Mathematik, Technische Universität Ilmenau Page 11 / 20

18 First construction (cont.) testing hypothesis E Z = λζ, ζ S 1, λ > 0: P(Im ζ 1 Z n s p) 3 8 α ζ s p λζ 0 P(Re ζ 1 Z n s 0 ) α 4 with s p = t [ 1,1] ( 3 8 α, 0) < s 0 First construction Institut für Mathematik, Technische Universität Ilmenau Page 11 / 20

19 First construction (cont.) testing hypothesis E Z = λζ, ζ S 1, λ > 0: P(Im ζ 1 Z n s p) 3 8 α ζ s p λζ 0 P(Re ζ 1 Z n s 0 ) α 4 P(Im ζ 1 Z n s p) 3 8 α with s p = t [ 1,1] ( 3 8 α, 0) < s 0 First construction Institut für Mathematik, Technische Universität Ilmenau Page 11 / 20

20 First construction (cont.) testing hypothesis E Z = λζ, ζ S 1, λ > 0: P(Im ζ 1 Z n s p) 3 8 α ζ s p 0 δ H λζ Z n P(Re ζ 1 Z n s 0 ) α 4 P(Im ζ 1 Z n s p) 3 8 α with s p = t [ 1,1] ( 3 8 α, 0) < s 0; critical angle δ H for Im ζ 1 Zn = s p First construction Institut für Mathematik, Technische Universität Ilmenau Page 11 / 20

21 Confidence set using Hoeffding s inequality Proposition (TH, FK & JW) C H = { S 1 if α 2 n+2 or Z n 2s 0, {ζ S 1 : Arg(ζ 1ˆµ n ) < δ H } otherwise, for δ H = arcsin(s p / Z n ) is a universal, non-asymptotic, level (1 α) confidence set for µ. note: if E Z 0 then P(C H = S 1 ) exp( nτ 2 /8) for some τ > 0 δ H = O P (n 1 2 ) δ H small for Z n large but variance perpendicular to µ not taken into account First construction Institut für Mathematik, Technische Universität Ilmenau Page 12 / 20

22 Estimating the variance varying the first construction: for Y k = Im ζ 1 Z k consider V n = 1 n n k=1 Y 2 k [0, 1] with E V n = Var Y k = σ 2 apply Hoeffding s inequality again to (numerically) get an upper bound on the variance fulfilling σ 2 = V n + t [0,1] ( α 4, 1 σ 2 ) such that P(σ 2 / [0, σ 2 ]) α 4 i.i.d. recall Hoeffding (1963) s Theorem 3: if W 1,..., W n W with W (, 1], E W = 0, Var W = ρ 2 then [ ( P( W n w) 1 + w ) ρ 2 w ( ) ] n w 1 1+ρ 2 ρ 2 1 w. for any w (0, 1) apply to Ȳ n using variance bound σ 2 sharper critical value s v = w( α 4, σ 2 ) perpendicular to µ Second construction Institut für Mathematik, Technische Universität Ilmenau Page 13 / 20

23 Confidence set taking variance into account Proposition (TH, FK & JW) C V = { S 1 if α 2 n+2 or Z n 2s 0, {ζ S 1 : Arg(ζ 1ˆµ n ) < δ V } otherwise, for δ V = arcsin(s v / Z n ) is a universal, non-asymptotic, level (1 α) confidence set for µ. Second construction Institut für Mathematik, Technische Universität Ilmenau Page 14 / 20

24 Simulation 1: two points of equal mass at ±10 50% 0 E Z µ 50% E Z highly concentrated distribution variance perpendicular to µ is maximal nominal coverage 1 α = 95% n = 400 independent draws in each experiment repetitions Simulations and applications Institut für Mathematik, Technische Universität Ilmenau Page 15 / 20

25 Simulation 1: two points of equal mass at ±10 Table: Results based on 10,000 repetitions with n = 400 observations each: average observed δ H, δ V, and δ A (with corresponding standard deviation), as well as frequency (with corresponding standard error) with which µ = 1 was covered by C H (first construction), C V (construction taking variance into account), and C A (asymptotic), respectively; the nominal coverage probability was 1 α = 95%. confidence set mean δ (± s.d.) coverage frequency (± s.e.) C H 8.2 (± ) 100.0% (±0.0%) C V 2.3 (± ) 100.0% (±0.0%) C A 1.0 (± ) 94.8% (±0.2%) C H, C V are conservative variance should be taken into account small standard deviation of δ Simulations and applications Institut für Mathematik, Technische Universität Ilmenau Page 16 / 20

26 Simulation 2: three points placed asymmetrically 1% 5% 0 E Z µ 94% E Z = 0.9 highly concentrated distribution nominal coverage 1 α = 90% n = 100 independent draws in each experiment repetitions Simulations and applications Institut für Mathematik, Technische Universität Ilmenau Page 17 / 20

27 Simulation 2: three points placed asymmetrically Table: Results based on 10,000 repetitions with n = 100 observations each: average observed δ H, δ V, and δ A (with corresponding standard deviation), as well as frequency (with corresponding standard error) with which µ = 1 was covered by C H (first construction), C V (construction taking variance into account), and C A (asymptotic), respectively; the nominal coverage probability was 1 α = 90%. confidence set mean δ (± s.d.) coverage frequency (± s.e.) C H 16.5 (± ) 100.0% (±0.0%) C V 5.0 (± ) 100.0% (±0.0%) C A 0.4 (± ) 62.8% (±0.5%) C H, C V still conservative coverage frequency of C A far from nominal level (90%) Simulations and applications Institut für Mathematik, Technische Universität Ilmenau Page 18 / 20

28 Summary constructed universal, non-asymptotic confidence sets for circular population means used mass concentration inequalities by Hoeffding variance perpendicular to mean can (and should) be taken into account asymptotic confidence sets are smaller but may not provide required coverage and presume uniqueness (requiring pretest?) Outlook: C V has practically useful size ( twice that of C A ) use sharper mass concentration inequalities spheres (etc.) intrinsic means (etc.) Summary Institut für Mathematik, Technische Universität Ilmenau Page 19 / 20

29 Thank you for your attention! Questions? Summary Institut für Mathematik, Technische Universität Ilmenau Page 20 / 20

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini May 30, 2018 1 / 59 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

21.1 Lower bounds on minimax risk for functional estimation

21.1 Lower bounds on minimax risk for functional estimation ECE598: Information-theoretic methods in high-dimensional statistics Spring 016 Lecture 1: Functional estimation & testing Lecturer: Yihong Wu Scribe: Ashok Vardhan, Apr 14, 016 In this chapter, we will

More information

Asymptotic Statistics-VI. Changliang Zou

Asymptotic Statistics-VI. Changliang Zou Asymptotic Statistics-VI Changliang Zou Kolmogorov-Smirnov distance Example (Kolmogorov-Smirnov confidence intervals) We know given α (0, 1), there is a well-defined d = d α,n such that, for any continuous

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 26, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 55 High dimensional

More information

Methods for sparse analysis of high-dimensional data, II

Methods for sparse analysis of high-dimensional data, II Methods for sparse analysis of high-dimensional data, II Rachel Ward May 23, 2011 High dimensional data with low-dimensional structure 300 by 300 pixel images = 90, 000 dimensions 2 / 47 High dimensional

More information

Mod-φ convergence II: dependency graphs

Mod-φ convergence II: dependency graphs Mod-φ convergence II: dependency graphs Valentin Féray (joint work with Pierre-Loïc Méliot and Ashkan Nikeghbali) Institut für Mathematik, Universität Zürich Summer school in Villa Volpi, Lago Maggiore,

More information

δ -method and M-estimation

δ -method and M-estimation Econ 2110, fall 2016, Part IVb Asymptotic Theory: δ -method and M-estimation Maximilian Kasy Department of Economics, Harvard University 1 / 40 Example Suppose we estimate the average effect of class size

More information

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma

Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Multivariate Statistics Random Projections and Johnson-Lindenstrauss Lemma Suppose again we have n sample points x,..., x n R p. The data-point x i R p can be thought of as the i-th row X i of an n p-dimensional

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

large number of i.i.d. observations from P. For concreteness, suppose

large number of i.i.d. observations from P. For concreteness, suppose 1 Subsampling Suppose X i, i = 1,..., n is an i.i.d. sequence of random variables with distribution P. Let θ(p ) be some real-valued parameter of interest, and let ˆθ n = ˆθ n (X 1,..., X n ) be some estimate

More information

Introduction to Statistical Learning Theory

Introduction to Statistical Learning Theory Introduction to Statistical Learning Theory Definition Reminder: We are given m samples {(x i, y i )} m i=1 Dm and a hypothesis space H and we wish to return h H minimizing L D (h) = E[l(h(x), y)]. Problem

More information

Large Sample Properties of Estimators in the Classical Linear Regression Model

Large Sample Properties of Estimators in the Classical Linear Regression Model Large Sample Properties of Estimators in the Classical Linear Regression Model 7 October 004 A. Statement of the classical linear regression model The classical linear regression model can be written in

More information

Lecture 8. October 22, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University.

Lecture 8. October 22, Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University. Lecture 8 Department of Biostatistics Johns Hopkins Bloomberg School of Public Health Johns Hopkins University October 22, 2007 1 2 3 4 5 6 1 Define convergent series 2 Define the Law of Large Numbers

More information

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester

Physics 403. Segev BenZvi. Parameter Estimation, Correlations, and Error Bars. Department of Physics and Astronomy University of Rochester Physics 403 Parameter Estimation, Correlations, and Error Bars Segev BenZvi Department of Physics and Astronomy University of Rochester Table of Contents 1 Review of Last Class Best Estimates and Reliability

More information

Computational Learning Theory. CS534 - Machine Learning

Computational Learning Theory. CS534 - Machine Learning Computational Learning Theory CS534 Machine Learning Introduction Computational learning theory Provides a theoretical analysis of learning Shows when a learning algorithm can be expected to succeed Shows

More information

Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, / 91. Bruce E.

Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, / 91. Bruce E. Forecasting Lecture 3 Structural Breaks Central Bank of Chile October 29-31, 2013 Bruce Hansen (University of Wisconsin) Structural Breaks October 29-31, 2013 1 / 91 Bruce E. Hansen Organization Detection

More information

Long-Run Covariability

Long-Run Covariability Long-Run Covariability Ulrich K. Müller and Mark W. Watson Princeton University October 2016 Motivation Study the long-run covariability/relationship between economic variables great ratios, long-run Phillips

More information

Inference For High Dimensional M-estimates. Fixed Design Results

Inference For High Dimensional M-estimates. Fixed Design Results : Fixed Design Results Lihua Lei Advisors: Peter J. Bickel, Michael I. Jordan joint work with Peter J. Bickel and Noureddine El Karoui Dec. 8, 2016 1/57 Table of Contents 1 Background 2 Main Results and

More information

Math 423/533: The Main Theoretical Topics

Math 423/533: The Main Theoretical Topics Math 423/533: The Main Theoretical Topics Notation sample size n, data index i number of predictors, p (p = 2 for simple linear regression) y i : response for individual i x i = (x i1,..., x ip ) (1 p)

More information

Tail inequalities for additive functionals and empirical processes of. Markov chains

Tail inequalities for additive functionals and empirical processes of. Markov chains Tail inequalities for additive functionals and empirical processes of geometrically ergodic Markov chains University of Warsaw Banff, June 2009 Geometric ergodicity Definition A Markov chain X = (X n )

More information

Lecture 3. Inference about multivariate normal distribution

Lecture 3. Inference about multivariate normal distribution Lecture 3. Inference about multivariate normal distribution 3.1 Point and Interval Estimation Let X 1,..., X n be i.i.d. N p (µ, Σ). We are interested in evaluation of the maximum likelihood estimates

More information

Improving linear quantile regression for

Improving linear quantile regression for Improving linear quantile regression for replicated data arxiv:1901.0369v1 [stat.ap] 16 Jan 2019 Kaushik Jana 1 and Debasis Sengupta 2 1 Imperial College London, UK 2 Indian Statistical Institute, Kolkata,

More information

Technische Universität Dresden Herausgeber: Der Rektor

Technische Universität Dresden Herausgeber: Der Rektor Als Manuskript gedruckt Technische Universität Dresden Herausgeber: Der Rektor The Gradient of the Squared Residual as Error Bound an Application to Karush-Kuhn-Tucker Systems Andreas Fischer MATH-NM-13-2002

More information

Chapter 10: Inferences based on two samples

Chapter 10: Inferences based on two samples November 16 th, 2017 Overview Week 1 Week 2 Week 4 Week 7 Week 10 Week 12 Chapter 1: Descriptive statistics Chapter 6: Statistics and Sampling Distributions Chapter 7: Point Estimation Chapter 8: Confidence

More information

A UNIQUENESS THEOREM FOR MONOGENIC FUNCTIONS

A UNIQUENESS THEOREM FOR MONOGENIC FUNCTIONS Annales Academiæ Scientiarum Fennicæ Series A. I. Mathematica Volumen 8, 993, 05 6 A UNIQUENESS THEOREM FOR MONOGENIC FUNCTIONS Jörg Winkler Technische Universität Berlin, Fachbereich 3, Mathematik Straße

More information

Risk-Minimality and Orthogonality of Martingales

Risk-Minimality and Orthogonality of Martingales Risk-Minimality and Orthogonality of Martingales Martin Schweizer Universität Bonn Institut für Angewandte Mathematik Wegelerstraße 6 D 53 Bonn 1 (Stochastics and Stochastics Reports 3 (199, 123 131 2

More information

Uniformly and Restricted Most Powerful Bayesian Tests

Uniformly and Restricted Most Powerful Bayesian Tests Uniformly and Restricted Most Powerful Bayesian Tests Valen E. Johnson and Scott Goddard Texas A&M University June 6, 2014 Valen E. Johnson and Scott Goddard Texas A&MUniformly University Most Powerful

More information

Characterizing Forecast Uncertainty Prediction Intervals. The estimated AR (and VAR) models generate point forecasts of y t+s, y ˆ

Characterizing Forecast Uncertainty Prediction Intervals. The estimated AR (and VAR) models generate point forecasts of y t+s, y ˆ Characterizing Forecast Uncertainty Prediction Intervals The estimated AR (and VAR) models generate point forecasts of y t+s, y ˆ t + s, t. Under our assumptions the point forecasts are asymtotically unbiased

More information

Statistical Data Analysis

Statistical Data Analysis DS-GA 0 Lecture notes 8 Fall 016 1 Descriptive statistics Statistical Data Analysis In this section we consider the problem of analyzing a set of data. We describe several techniques for visualizing the

More information

Extrinsic Antimean and Bootstrap

Extrinsic Antimean and Bootstrap Extrinsic Antimean and Bootstrap Yunfan Wang Department of Statistics Florida State University April 12, 2018 Yunfan Wang (Florida State University) Extrinsic Antimean and Bootstrap April 12, 2018 1 /

More information

P (A G) dp G P (A G)

P (A G) dp G P (A G) First homework assignment. Due at 12:15 on 22 September 2016. Homework 1. We roll two dices. X is the result of one of them and Z the sum of the results. Find E [X Z. Homework 2. Let X be a r.v.. Assume

More information

Statistical Machine Learning

Statistical Machine Learning Statistical Machine Learning UoC Stats 37700, Winter quarter Lecture 2: Introduction to statistical learning theory. 1 / 22 Goals of statistical learning theory SLT aims at studying the performance of

More information

Eigenvalue (mis)behavior on manifolds

Eigenvalue (mis)behavior on manifolds Bucknell University Lehigh University October 20, 2010 Outline 1 Isoperimetric inequalities 2 3 4 A little history Rayleigh quotients The Original Isoperimetric Inequality The Problem of Queen Dido: maximize

More information

Math 3T03 - Topology

Math 3T03 - Topology Math 3T03 - Topology Sang Woo Park April 5, 2018 Contents 1 Introduction to topology 2 1.1 What is topology?.......................... 2 1.2 Set theory............................... 3 2 Functions 4 3

More information

STAT Sample Problem: General Asymptotic Results

STAT Sample Problem: General Asymptotic Results STAT331 1-Sample Problem: General Asymptotic Results In this unit we will consider the 1-sample problem and prove the consistency and asymptotic normality of the Nelson-Aalen estimator of the cumulative

More information

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D.

Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Web Appendix for Hierarchical Adaptive Regression Kernels for Regression with Functional Predictors by D. B. Woodard, C. Crainiceanu, and D. Ruppert A. EMPIRICAL ESTIMATE OF THE KERNEL MIXTURE Here we

More information

Logarithmic Sobolev inequalities in discrete product spaces: proof by a transportation cost distance

Logarithmic Sobolev inequalities in discrete product spaces: proof by a transportation cost distance Logarithmic Sobolev inequalities in discrete product spaces: proof by a transportation cost distance Katalin Marton Alfréd Rényi Institute of Mathematics of the Hungarian Academy of Sciences Relative entropy

More information

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses.

Review: General Approach to Hypothesis Testing. 1. Define the research question and formulate the appropriate null and alternative hypotheses. 1 Review: Let X 1, X,..., X n denote n independent random variables sampled from some distribution might not be normal!) with mean µ) and standard deviation σ). Then X µ σ n In other words, X is approximately

More information

Econometrics A. Simple linear model (2) Keio University, Faculty of Economics. Simon Clinet (Keio University) Econometrics A October 16, / 11

Econometrics A. Simple linear model (2) Keio University, Faculty of Economics. Simon Clinet (Keio University) Econometrics A October 16, / 11 Econometrics A Keio University, Faculty of Economics Simple linear model (2) Simon Clinet (Keio University) Econometrics A October 16, 2018 1 / 11 Estimation of the noise variance σ 2 In practice σ 2 too

More information

7 Influence Functions

7 Influence Functions 7 Influence Functions The influence function is used to approximate the standard error of a plug-in estimator. The formal definition is as follows. 7.1 Definition. The Gâteaux derivative of T at F in the

More information

Uncertainty. Jayakrishnan Unnikrishnan. CSL June PhD Defense ECE Department

Uncertainty. Jayakrishnan Unnikrishnan. CSL June PhD Defense ECE Department Decision-Making under Statistical Uncertainty Jayakrishnan Unnikrishnan PhD Defense ECE Department University of Illinois at Urbana-Champaign CSL 141 12 June 2010 Statistical Decision-Making Relevant in

More information

Inference for High Dimensional Robust Regression

Inference for High Dimensional Robust Regression Department of Statistics UC Berkeley Stanford-Berkeley Joint Colloquium, 2015 Table of Contents 1 Background 2 Main Results 3 OLS: A Motivating Example Table of Contents 1 Background 2 Main Results 3 OLS:

More information

Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing

Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing Economics 583: Econometric Theory I A Primer on Asymptotics: Hypothesis Testing Eric Zivot October 12, 2011 Hypothesis Testing 1. Specify hypothesis to be tested H 0 : null hypothesis versus. H 1 : alternative

More information

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and

Discriminant Analysis with High Dimensional. von Mises-Fisher distribution and Athens Journal of Sciences December 2014 Discriminant Analysis with High Dimensional von Mises - Fisher Distributions By Mario Romanazzi This paper extends previous work in discriminant analysis with von

More information

UNIVERSITÄT POTSDAM Institut für Mathematik

UNIVERSITÄT POTSDAM Institut für Mathematik UNIVERSITÄT POTSDAM Institut für Mathematik Testing the Acceleration Function in Life Time Models Hannelore Liero Matthias Liero Mathematische Statistik und Wahrscheinlichkeitstheorie Universität Potsdam

More information

Random matrix pencils and level crossings

Random matrix pencils and level crossings Albeverio Fest October 1, 2018 Topics to discuss Basic level crossing problem 1 Basic level crossing problem 2 3 Main references Basic level crossing problem (i) B. Shapiro, M. Tater, On spectral asymptotics

More information

Lecture Notes 1: Vector spaces

Lecture Notes 1: Vector spaces Optimization-based data analysis Fall 2017 Lecture Notes 1: Vector spaces In this chapter we review certain basic concepts of linear algebra, highlighting their application to signal processing. 1 Vector

More information

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing

Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing Quantitative Introduction ro Risk and Uncertainty in Business Module 5: Hypothesis Testing M. Vidyasagar Cecil & Ida Green Chair The University of Texas at Dallas Email: M.Vidyasagar@utdallas.edu October

More information

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf

Lecture 13: Subsampling vs Bootstrap. Dimitris N. Politis, Joseph P. Romano, Michael Wolf Lecture 13: 2011 Bootstrap ) R n x n, θ P)) = τ n ˆθn θ P) Example: ˆθn = X n, τ n = n, θ = EX = µ P) ˆθ = min X n, τ n = n, θ P) = sup{x : F x) 0} ) Define: J n P), the distribution of τ n ˆθ n θ P) under

More information

Maximum Mean Discrepancy

Maximum Mean Discrepancy Maximum Mean Discrepancy Thanks to Karsten Borgwardt, Malte Rasch, Bernhard Schölkopf, Jiayuan Huang, Arthur Gretton Alexander J. Smola Statistical Machine Learning Program Canberra, ACT 0200 Australia

More information

Modern Portfolio Theory with Homogeneous Risk Measures

Modern Portfolio Theory with Homogeneous Risk Measures Modern Portfolio Theory with Homogeneous Risk Measures Dirk Tasche Zentrum Mathematik Technische Universität München http://www.ma.tum.de/stat/ Rotterdam February 8, 2001 Abstract The Modern Portfolio

More information

Generalization theory

Generalization theory Generalization theory Daniel Hsu Columbia TRIPODS Bootcamp 1 Motivation 2 Support vector machines X = R d, Y = { 1, +1}. Return solution ŵ R d to following optimization problem: λ min w R d 2 w 2 2 + 1

More information

Graduate Econometrics I: Asymptotic Theory

Graduate Econometrics I: Asymptotic Theory Graduate Econometrics I: Asymptotic Theory Yves Dominicy Université libre de Bruxelles Solvay Brussels School of Economics and Management ECARES Yves Dominicy Graduate Econometrics I: Asymptotic Theory

More information

STAT 100C: Linear models

STAT 100C: Linear models STAT 100C: Linear models Arash A. Amini June 9, 2018 1 / 56 Table of Contents Multiple linear regression Linear model setup Estimation of β Geometric interpretation Estimation of σ 2 Hat matrix Gram matrix

More information

Orthonormal Systems. Fourier Series

Orthonormal Systems. Fourier Series Yuliya Gorb Orthonormal Systems. Fourier Series October 31 November 3, 2017 Yuliya Gorb Orthonormal Systems (cont.) Let {e α} α A be an orthonormal set of points in an inner product space X. Then {e α}

More information

Lecture 15. Hypothesis testing in the linear model

Lecture 15. Hypothesis testing in the linear model 14. Lecture 15. Hypothesis testing in the linear model Lecture 15. Hypothesis testing in the linear model 1 (1 1) Preliminary lemma 15. Hypothesis testing in the linear model 15.1. Preliminary lemma Lemma

More information

1 Statistical inference for a population mean

1 Statistical inference for a population mean 1 Statistical inference for a population mean 1. Inference for a large sample, known variance Suppose X 1,..., X n represents a large random sample of data from a population with unknown mean µ and known

More information

Modeling Real Estate Data using Quantile Regression

Modeling Real Estate Data using Quantile Regression Modeling Real Estate Data using Semiparametric Quantile Regression Department of Statistics University of Innsbruck September 9th, 2011 Overview 1 Application: 2 3 4 Hedonic regression data for house prices

More information

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing

STAT 135 Lab 5 Bootstrapping and Hypothesis Testing STAT 135 Lab 5 Bootstrapping and Hypothesis Testing Rebecca Barter March 2, 2015 The Bootstrap Bootstrap Suppose that we are interested in estimating a parameter θ from some population with members x 1,...,

More information

Operational Risk and Pareto Lévy Copulas

Operational Risk and Pareto Lévy Copulas Operational Risk and Pareto Lévy Copulas Claudia Klüppelberg Technische Universität München email: cklu@ma.tum.de http://www-m4.ma.tum.de References: - Böcker, K. and Klüppelberg, C. (25) Operational VaR

More information

Lecture 6 Channel Coding over Continuous Channels

Lecture 6 Channel Coding over Continuous Channels Lecture 6 Channel Coding over Continuous Channels I-Hsiang Wang Department of Electrical Engineering National Taiwan University ihwang@ntu.edu.tw November 9, 015 1 / 59 I-Hsiang Wang IT Lecture 6 We have

More information

Bootstrap inference for the finite population total under complex sampling designs

Bootstrap inference for the finite population total under complex sampling designs Bootstrap inference for the finite population total under complex sampling designs Zhonglei Wang (Joint work with Dr. Jae Kwang Kim) Center for Survey Statistics and Methodology Iowa State University Jan.

More information

Inner product spaces. Layers of structure:

Inner product spaces. Layers of structure: Inner product spaces Layers of structure: vector space normed linear space inner product space The abstract definition of an inner product, which we will see very shortly, is simple (and by itself is pretty

More information

A contamination model for approximate stochastic order

A contamination model for approximate stochastic order A contamination model for approximate stochastic order Eustasio del Barrio Universidad de Valladolid. IMUVA. 3rd Workshop on Analysis, Geometry and Probability - Universität Ulm 28th September - 2dn October

More information

MATH 423/ Note that the algebraic operations on the right hand side are vector subtraction and scalar multiplication.

MATH 423/ Note that the algebraic operations on the right hand side are vector subtraction and scalar multiplication. MATH 423/673 1 Curves Definition: The velocity vector of a curve α : I R 3 at time t is the tangent vector to R 3 at α(t), defined by α (t) T α(t) R 3 α α(t + h) α(t) (t) := lim h 0 h Note that the algebraic

More information

Rank-sum Test Based on Order Restricted Randomized Design

Rank-sum Test Based on Order Restricted Randomized Design Rank-sum Test Based on Order Restricted Randomized Design Omer Ozturk and Yiping Sun Abstract One of the main principles in a design of experiment is to use blocking factors whenever it is possible. On

More information

Intrinsic means on the circle: uniqueness, locus and asymptotics

Intrinsic means on the circle: uniqueness, locus and asymptotics Ann Inst Stat Math (215) 67:177 193 DOI 1.17/s1463-13-444-7 Intrinsic means on the circle: uniqueness, locus and asymptotics T. Hotz S. Huckemann Received: 24 August 212 / Revised: 19 November 213 / Published

More information

Robustness and bootstrap techniques in portfolio efficiency tests

Robustness and bootstrap techniques in portfolio efficiency tests Robustness and bootstrap techniques in portfolio efficiency tests Dept. of Probability and Mathematical Statistics, Charles University, Prague, Czech Republic July 8, 2013 Motivation Portfolio selection

More information

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A. 1. Let P be a probability measure on a collection of sets A. (a) For each n N, let H n be a set in A such that H n H n+1. Show that P (H n ) monotonically converges to P ( k=1 H k) as n. (b) For each n

More information

17: INFERENCE FOR MULTIPLE REGRESSION. Inference for Individual Regression Coefficients

17: INFERENCE FOR MULTIPLE REGRESSION. Inference for Individual Regression Coefficients 17: INFERENCE FOR MULTIPLE REGRESSION Inference for Individual Regression Coefficients The results of this section require the assumption that the errors u are normally distributed. Let c i ij denote the

More information

Bandit models: a tutorial

Bandit models: a tutorial Gdt COS, December 3rd, 2015 Multi-Armed Bandit model: general setting K arms: for a {1,..., K}, (X a,t ) t N is a stochastic process. (unknown distributions) Bandit game: a each round t, an agent chooses

More information

Technische Universität Ilmenau Institut für Mathematik

Technische Universität Ilmenau Institut für Mathematik Technische Universität Ilmenau Institut für Mathematik Preprint No. M 18/03 A branch-and-bound based algorithm for nonconvex multiobjective optimization Julia Niebling, Gabriele Eichfelder Februar 018

More information

Topic 6: Non-Spherical Disturbances

Topic 6: Non-Spherical Disturbances Topic 6: Non-Spherical Disturbances Our basic linear regression model is y = Xβ + ε ; ε ~ N[0, σ 2 I n ] Now we ll generalize the specification of the error term in the model: E[ε] = 0 ; E[εε ] = Σ = σ

More information

Estimation of the functional Weibull-tail coefficient

Estimation of the functional Weibull-tail coefficient 1/ 29 Estimation of the functional Weibull-tail coefficient Stéphane Girard Inria Grenoble Rhône-Alpes & LJK, France http://mistis.inrialpes.fr/people/girard/ June 2016 joint work with Laurent Gardes,

More information

Asymmetric least squares estimation and testing

Asymmetric least squares estimation and testing Asymmetric least squares estimation and testing Whitney Newey and James Powell Princeton University and University of Wisconsin-Madison January 27, 2012 Outline ALS estimators Large sample properties Asymptotic

More information

The multi armed-bandit problem

The multi armed-bandit problem The multi armed-bandit problem (with covariates if we have time) Vianney Perchet & Philippe Rigollet LPMA Université Paris Diderot ORFE Princeton University Algorithms and Dynamics for Games and Optimization

More information

Module 3. Function of a Random Variable and its distribution

Module 3. Function of a Random Variable and its distribution Module 3 Function of a Random Variable and its distribution 1. Function of a Random Variable Let Ω, F, be a probability space and let be random variable defined on Ω, F,. Further let h: R R be a given

More information

Exact goodness-of-fit tests for censored data

Exact goodness-of-fit tests for censored data Exact goodness-of-fit tests for censored data Aurea Grané Statistics Department. Universidad Carlos III de Madrid. Abstract The statistic introduced in Fortiana and Grané (23, Journal of the Royal Statistical

More information

Principles of Bayesian Inference

Principles of Bayesian Inference Principles of Bayesian Inference Sudipto Banerjee University of Minnesota July 20th, 2008 1 Bayesian Principles Classical statistics: model parameters are fixed and unknown. A Bayesian thinks of parameters

More information

An algorithm for computing Fréchet means on the sphere

An algorithm for computing Fréchet means on the sphere An algorithm for computing Fréchet means on the sphere Gabriele Eichfelder Thomas Hotz Johannes Wieditz February 26, 2018 Abstract For most optimisation methods an essential assumption is the vector space

More information

Introduction to Bandit Algorithms. Introduction to Bandit Algorithms

Introduction to Bandit Algorithms. Introduction to Bandit Algorithms Stochastic K-Arm Bandit Problem Formulation Consider K arms (actions) each correspond to an unknown distribution {ν k } K k=1 with values bounded in [0, 1]. At each time t, the agent pulls an arm I t {1,...,

More information

Subdivisions of Large Complete Bipartite Graphs and Long Induced Paths in k-connected Graphs Thomas Bohme Institut fur Mathematik Technische Universit

Subdivisions of Large Complete Bipartite Graphs and Long Induced Paths in k-connected Graphs Thomas Bohme Institut fur Mathematik Technische Universit Subdivisions of Large Complete Bipartite Graphs and Long Induced Paths in k-connected Graphs Thomas Bohme Institut fur Mathematik Technische Universitat Ilmenau Ilmenau, Germany Riste Skrekovski z Department

More information

1 of 7 7/16/2009 6:12 AM Virtual Laboratories > 7. Point Estimation > 1 2 3 4 5 6 1. Estimators The Basic Statistical Model As usual, our starting point is a random experiment with an underlying sample

More information

Stratégies bayésiennes et fréquentistes dans un modèle de bandit

Stratégies bayésiennes et fréquentistes dans un modèle de bandit Stratégies bayésiennes et fréquentistes dans un modèle de bandit thèse effectuée à Telecom ParisTech, co-dirigée par Olivier Cappé, Aurélien Garivier et Rémi Munos Journées MAS, Grenoble, 30 août 2016

More information

where x and ȳ are the sample means of x 1,, x n

where x and ȳ are the sample means of x 1,, x n y y Animal Studies of Side Effects Simple Linear Regression Basic Ideas In simple linear regression there is an approximately linear relation between two variables say y = pressure in the pancreas x =

More information

Inference For High Dimensional M-estimates: Fixed Design Results

Inference For High Dimensional M-estimates: Fixed Design Results Inference For High Dimensional M-estimates: Fixed Design Results Lihua Lei, Peter Bickel and Noureddine El Karoui Department of Statistics, UC Berkeley Berkeley-Stanford Econometrics Jamboree, 2017 1/49

More information

Hypothesis Testing. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA

Hypothesis Testing. Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA Hypothesis Testing Robert L. Wolpert Department of Statistical Science Duke University, Durham, NC, USA An Example Mardia et al. (979, p. ) reprint data from Frets (9) giving the length and breadth (in

More information

A PECULIAR COIN-TOSSING MODEL

A PECULIAR COIN-TOSSING MODEL A PECULIAR COIN-TOSSING MODEL EDWARD J. GREEN 1. Coin tossing according to de Finetti A coin is drawn at random from a finite set of coins. Each coin generates an i.i.d. sequence of outcomes (heads or

More information

10-704: Information Processing and Learning Fall Lecture 24: Dec 7

10-704: Information Processing and Learning Fall Lecture 24: Dec 7 0-704: Information Processing and Learning Fall 206 Lecturer: Aarti Singh Lecture 24: Dec 7 Note: These notes are based on scribed notes from Spring5 offering of this course. LaTeX template courtesy of

More information

arxiv: v1 [stat.co] 26 Feb 2018

arxiv: v1 [stat.co] 26 Feb 2018 An algorithm for computing Fréchet means on the sphere Gabriele Eichfelder Thomas Hotz Johannes Wieditz arxiv:1802.09243v1 [stat.co] 26 Feb 2018 February 27, 2018 Abstract For most optimisation methods

More information

STAT 200C: High-dimensional Statistics

STAT 200C: High-dimensional Statistics STAT 200C: High-dimensional Statistics Arash A. Amini April 27, 2018 1 / 80 Classical case: n d. Asymptotic assumption: d is fixed and n. Basic tools: LLN and CLT. High-dimensional setting: n d, e.g. n/d

More information

The Informativeness of k-means for Learning Mixture Models

The Informativeness of k-means for Learning Mixture Models The Informativeness of k-means for Learning Mixture Models Vincent Y. F. Tan (Joint work with Zhaoqiang Liu) National University of Singapore June 18, 2018 1/35 Gaussian distribution For F dimensions,

More information

Kernel Methods. Lecture 4: Maximum Mean Discrepancy Thanks to Karsten Borgwardt, Malte Rasch, Bernhard Schölkopf, Jiayuan Huang, Arthur Gretton

Kernel Methods. Lecture 4: Maximum Mean Discrepancy Thanks to Karsten Borgwardt, Malte Rasch, Bernhard Schölkopf, Jiayuan Huang, Arthur Gretton Kernel Methods Lecture 4: Maximum Mean Discrepancy Thanks to Karsten Borgwardt, Malte Rasch, Bernhard Schölkopf, Jiayuan Huang, Arthur Gretton Alexander J. Smola Statistical Machine Learning Program Canberra,

More information

Marginal Screening and Post-Selection Inference

Marginal Screening and Post-Selection Inference Marginal Screening and Post-Selection Inference Ian McKeague August 13, 2017 Ian McKeague (Columbia University) Marginal Screening August 13, 2017 1 / 29 Outline 1 Background on Marginal Screening 2 2

More information

Research Article Statistical Tests for the Reciprocal of a Normal Mean with a Known Coefficient of Variation

Research Article Statistical Tests for the Reciprocal of a Normal Mean with a Known Coefficient of Variation Probability and Statistics Volume 2015, Article ID 723924, 5 pages http://dx.doi.org/10.1155/2015/723924 Research Article Statistical Tests for the Reciprocal of a Normal Mean with a Known Coefficient

More information

Reliable Inference in Conditions of Extreme Events. Adriana Cornea

Reliable Inference in Conditions of Extreme Events. Adriana Cornea Reliable Inference in Conditions of Extreme Events by Adriana Cornea University of Exeter Business School Department of Economics ExISta Early Career Event October 17, 2012 Outline of the talk Extreme

More information

Complete Surfaces of Constant Gaussian Curvature in Euclidean Space R 3.

Complete Surfaces of Constant Gaussian Curvature in Euclidean Space R 3. Summary of the Thesis in Mathematics by Valentina Monaco Complete Surfaces of Constant Gaussian Curvature in Euclidean Space R 3. Thesis Supervisor Prof. Massimiliano Pontecorvo 19 May, 2011 SUMMARY The

More information

Lecture 22: Variance and Covariance

Lecture 22: Variance and Covariance EE5110 : Probability Foundations for Electrical Engineers July-November 2015 Lecture 22: Variance and Covariance Lecturer: Dr. Krishna Jagannathan Scribes: R.Ravi Kiran In this lecture we will introduce

More information

Course 212: Academic Year Section 1: Metric Spaces

Course 212: Academic Year Section 1: Metric Spaces Course 212: Academic Year 1991-2 Section 1: Metric Spaces D. R. Wilkins Contents 1 Metric Spaces 3 1.1 Distance Functions and Metric Spaces............. 3 1.2 Convergence and Continuity in Metric Spaces.........

More information

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T

The goal of this chapter is to study linear systems of ordinary differential equations: dt,..., dx ) T 1 1 Linear Systems The goal of this chapter is to study linear systems of ordinary differential equations: ẋ = Ax, x(0) = x 0, (1) where x R n, A is an n n matrix and ẋ = dx ( dt = dx1 dt,..., dx ) T n.

More information