Root-Unroot Methods for Nonparametric Density Estimation and Poisson Random-Effects Models
|
|
- Dominick Strickland
- 5 years ago
- Views:
Transcription
1 University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 00 Root-Unroot Methods for Nonparametric Density Estimation and Poisson Random-Effects Models Lwawrence D. Brown University of Pennsylvania Ren Zhang Linda H. Zhao University of Pennsylvania Follow this and additional works at: Part of the Statistics and Probability Commons Recommended Citation Brown, L. D., Zhang, R., & Zhao, L. H. (00). Root-Unroot Methods for Nonparametric Density Estimation and Poisson Random- Effects Models. Retrieved from This paper is posted at ScholarlyCommons. For more information, please contact
2 Root-Unroot Methods for Nonparametric Density Estimation and Poisson Random-Effects Models Disciplines Statistics and Probability This technical report is available at ScholarlyCommons:
3 Root-Unroot Methods for Nonparametric Density Estimation And Poisson random-effects models 1 Lawrence D. Brown University of Pennsylvania Statistics Department Philadelphia, PA (lbrown@wharton.upenn.edu) Joint work with: Ren Zhang Linda Zhao 1. For departmental Statistics seminar 1/16/00. Revision and expansion of invited talk delivered at the Joint Statistical Meetings in Atlanta, GA, August, 001 1
4 Outline of Presentation 0. Preamble (motivation). 1.Description of the root-unroot method for density estimation. (Really, a methodette.).motivation via Poissonization and variance stabilization. 3.Properties of the root-unroot step. 4.Some simulation pictures. 5.Comments about signal-to-noise ratio. 6.An empirical example from a telephone call-in service center. 7.A two-way random-effects analysis of call arrival rates.
5 Motivation Nonparametric regression and nonparametric density estimation are siblings. (Maybe even non-identical twins.) But I like better working with the regression formulation. (I feel a better sense of rapport and understanding!) So my goal is to convert a density problem into a regression one; so I can work with it as such. 3
6 Description of the Methodette 1. DATA: {X 1,, X n } i.i.d. from a density f on [0, 1].. BIN: Create K (equal width) bins. Let N k = # of {X i } in the kth bin, Bin k T k = Center of Bin k. 3. ROOT : Calculate Y k = K N k + 1 n ESTIMATE: Apply your favorite nonparametric regression estimator to {T k, Y k }. This produces an estimate: ĥ(t) of h(t) f(t). Since E( Y ) f(t ) = h(t ). k If helpful, capitalize on k k Var(Y ) k K for all k. 4n 4
7 5. UN-ROOT : Calculate ˆ h(t) and renormalize to be a density. For equal width bins this gives ˆ ˆ (h(t k ) + ) f(t k ) = K K ˆ (h(t ) ) j=1 j +. 5
8 Data Example 1. DATA: Histogram of Data: n = 18 It s accidental that n is a power of ; this is irrelevant to the methodette. This data is a random sample from the two-humps density in Wahba (1983). 6
9 & 3. BIN AND ROOT: SqRt(N+1/4) T Scatterplot of SqRt(N k +1/4) versus T k 7
10 4. ESTIMATE: SqRt(N+1/4) T Smoothing Spline Fit Smoothing Spline Fit, lambda= R-Square 0.94 Sum of Squares Error This fit has about 8 df for error ( 8 ); and consequently about 4 df for the 1/4 model. 8
11 6. UN-ROOT: New Pred% X Estimate of f(x) 9
12 How Well Did We Do? (1) Y X Plot of Estimate and True Density True density = thick curve Estimate = narrow curve 10
13 How Well Did We Do? () Y X Comparison of Estimators Thin curve: Root-unroot estimator Medium curve: Kernel density estimator Heavy curve: True density Both estimators capture the qualitative two-humps feature. Neither estimator provides a tremendously good fit. (More on this later.) Fair and revealing Monte Carlo comparisons are difficult to construct; the comparisons depend much more on the appropriateness of the respective density and regression estimators than on the integrity of direct density estimation versus the root-unroot regression paradigm. More, later. 11
14 Poissonization: Motivation Suppose n = N, a Poisson(λ) random variable. Then N k Poisson(λ k ) with λ k = λ f(t k ) andf(t k ) = Average of f over Bin k. These N k are independent. λ We then wish to make inference aboutf(t ) = k. λ k N k is ancillary to λ k /λ. A natural way to proceed is to estimate λ k ˆ k (nonparametrically) by λ, and then estimate f(t k ) by λˆ λ ˆ. k j 1
15 Variance Stabilization: An asymptotically unbiased, second-order mean-stabilizing, first-order variance stabilizing transformation for the independent counts, N k, is N These variables are approximately normal, with mean = λ k, and variance = 1/4, as well as independent. k Variance stabilization and Poissonization each have extensive statistical pedigrees. Density estimation schemes closely related to the above have also been often proposed. See Fan & Gijbels (1996), and Vidakovic (1999) for two recent versions, each of which is somewhat different from the above Nussbaum (1996), Nussbaum and Klemela (1999), Carter (001) and Brown, Low and Zhang(001) all contain asymptotic equivalence theorems formally justifying (sometimes complicated) versions of the above scheme, under suitable regularity conditions. 13
16 Properties of the Root-Unroot Step Stabilizes the mean: Y lambda { } Mean of ( ) λ E Poiss +c for c = 0 and c = 1/4 Also shown is the line Y = λ. 14
17 Detail from Previous Plot Y lambda { } E Poiss +c for c = 0 and c = 1/4 µ λ = ( λ ) Also shown is the line Y = λ. (Lowest curve is for c = 0. Highest curve is for c = 1/4.) 15
18 Stabilizes the variance: Y λ lambda σ = Var( Poiss λ + c) for c = 0 and c =1/4 The nominal (and limiting) value is 1/4. Upper curve is for c = 0; Other curve is for c = 1/4; CONCLUSION: It is suitable to use the root-unroot method with values of n/k as small as 3 or 4. As n increases it seems suitable to let n/k increase slowly. 16
19 Notes: Apparently, Bartlett (1936) was the first to propose the transformation Y = Poiss as the variance stabilizing transformation. He then proposed using it in a homoscedastic linear model. Anscombe (1948) proposed improving the variance stabilizing properties by instead using Y = Poiss + 3/8. (He credits this result to A. H. L. Johnson.) He apparently thought that this transformation was also optimal with respect to its mean. (See his equation (.1).) In fact this transformation is slightly more biased than our Y = Poiss + 1/4, as is shown by asymptotic analysis or by the plot below. I do not know who should be credited as the first to propose Y = Poiss + 1/4 as the asymptotically unbiased, nearly variance-stabilizing transformation. On the grounds that bias is more important here than variance, we prefer to use Y = Poiss + 1/4. This transformation has such nice properties that it doesn t seem worthwhile to try to make further improvements at the expense of complicating the formula. 17
20 Plots for Anscombe s Transformation Y lambda { } µ λ = ( ) λ E Poiss +c for c = 1/4 and c = 3/8 Also shown is the line Y = λ. Lower curve is for c = 1/4. Higher curve is for c = 3/8. Revealed in the plot is the limiting bias of 1/8 for { } Anscombe s ( 3/8) λ E Poiss +. 18
21 Y λ lambda σ = Var( Poiss λ + c) for c = 1/4 and c = 3/8 The nominal (and limiting) value is 1/4. Upper curve is for c = 1/4; Other curve is for c = 3/8; 19
22 Our transformation does pretty well with higher moments too: Skewness Y lambda* Skewness = λ 3 λ λ 3 E((Poiss - µ ) ) σ Lower curve is for c = 0; Other curve is for c = ¼; The nominal (and limiting) value is 0. 0
23 6 5 4 Y lambda* Kurtosis = The more variable curve is for c = 0; Other curve is for c = ¼; The nominal (and limiting) value is 3. λ 4 λ λ 4 E((Poiss - µ ) ) σ 1. A digression because this is about one Poisson observation, but the principal goal here is density estimation and related inference.. 1
24 Digression 1 : Confidence Intervals For single Poisson variables The transformation N + 1 produces good individual 4 confidence intervals. Here is a plot of the coverage of the nominal 95% interval N 1.96, 1 N for λ, where N~Poiss(λ) A digression because this is about one Poisson observation, but the principal goal here is density estimation and related inference..
25 Y lambda Y = Coverage of 95% root-unroot interval as a function of λ 3
26 A naïve 95% confidence interval could be formed by exploiting the fact that N is the MLE of Var(N) = λ. This interval would be N 1.96 N, N N. ( ) Note that the 95% root-unroot interval is N 1.96 N +, N 1.96 N so long as N Thus, for N 4 the two intervals have the same length, but the root-unroot interval is (always) shifted to the right. Here is a comparison of the coverage of the two intervals: 4
27 Y lambda Y = Coverage of 95% root-unroot interval and of 95% conventional Wald interval as a function of λ End of Digression 5
28 An Asymptotic Comparison Fair and revealing Monte Carlo comparisons are difficult to construct; the comparisons depend much more on the appropriateness of the respective density and regression estimators than on the integrity of direct density estimation versus the root-unroot regression paradigm. Here is an asymptotic comparison to give an idea of the situation. Consider a standard kernel estimator, with kernel W. For the ordinary density problem this is 1 1 X () ( i t f t = W ). n d d i 6
29 For the root-unroot method this uses the formula 1 1 t ˆ( ) ( k t ht = W ) Yk, K k d d and then un-roots according to fˆ() t = ( hˆ()) t. (Asymptotically ( ht ) ˆ( ) dt 1, very nearly, so we not not carry out the renormaliztion step.) Take W = Unif(-1/, 1/) for numerical simplicity, and d = C/ n 1/5. Define the IMSE Risk for an estimator f as 1 R= ( f ( t) f( t)) dt ( f ( tk) f( tk)). K Note that R can be decomposed as R = Bias + Var. 7
30 Then, for the density estimator 4 4/5 C 4 dens = ( ") + Ψ dens 1 1 n R f C +, 4 C C and for the Root-unroot estimator 4 4/5 C f 4 r u~ ( " ) + Ψ R U ( ') 1 1 n R f C +, 4 f C C since ˆ ~ ˆ h h h h 4h so that ( ) ( ) ( prob ) hˆ ( ) ˆ bias ~ 4 h bias ( f ') 4 h ( h") = ( f " ) f h =. Clearly, it is possible that R dens < R r-u (if f is linear); and also that R dens > R r-u (if h = f is linear). 8
31 Here is a table showing the results for some simple situations: Density =.95f+.05 : Ψ dens Ψ R-U R** dens f= sin πx R** R-U 30x (1-x) x(1-x) Table of Bias term, Ψ, for 3 simple density functions. The table also shows the Oracle Risk, R**, where 4/5 5 /5 R** = inf C{ n R} = 4/5 4 Ψ. Conclusion: The root-unroot estimator is very slightly superior to the original density estimator when f is smooth in the region where it s small. It is considerably inferior when f is linear in that region. 9
32 The previous considerations suggest that it might be more sensible to model the problem in terms of h = f; and to judge the quality of an estimator according to its squared Hellinger risk H R = 1 f t f t dt f t K f t. (The above general conclusion remains valid.) ( () ()) ( ( ) ( )) k k 30
33 Simulations The Root-unroot scheme enables reduction of the problem to one effectively having K = m equally spaced, homoscedastic and independent observations. This is exactly suitable for wavelet analyses. Here are plots showing the result of using the root-unroot scheme along with the Block J-S wavelet procedure in Cai and Silverman(1998+). The functions exhibited here are hard-to-fit forms adapted from Donoho and Johnstone (1995). ( Heavisine and Blocks ) 31
34 n = obs, K = 4096 bins, (Basic wavelet is s10.) f(x) binest true density x 3
35 Here is the same plot, also showing a local linear fit estimator. n = obs, K = 4096 bins f(x) binest true density locfit x 33
36 More Heavisine plots obs, K =51 bins (Basic wavelet is s10) f(x) binest true density x obs, 51 intnum, f(x) binest true density locfit x 34
37 Here are Heavisine plots with fewer observations. Note that the quality of fit is much poorer obs, 104 intnum, f(x) binest true density x 6000 obs, 104 intnum f(x) binest true density locfit x 35
38 BLOCKS: n = obs, K = 4096 (s8) f(x) binest true density x n = obs, K = 4096 (s8) f(x) binest true density locfit x Note: Use of smaller K, K = 51, barely affects the plots. 36
39 BLOCKS: n = 6000 obs, K = 104 (s8) f(x) binest true density x n = 6000 obs, K = 104 (s8) f(x) binest true density locfit x 37
40 Signal to noise ratio Question: Why are the plots for n = 6000 so poor? Why aren t the plots for n = nearly perfect? (That s the situation with the nonparametric regression data analyzed in Donoho, et. al. (1994, 1995, +) for these curves.) Answer: Because density estimation is intrinsically much harder than the nonparametric regression situations of those articles. Detail: Here is a scale-free (in Y) Difficulty of Estimation index: Average local S. D. of the error DE =. Average intensity of the signal Where denotes a scale free (in Y) measure of the local complexity of the signal (see below). For a homoscedastic regression problem with signal g and q observations each having S.D. = σ this is 38
41 σ / q 1 DE = =, SNR q ( ( gx ( ) g) ) 1/ where SNR denotes the conventional Signal to Noise ratio, SNR ( ( gx ( ) g ) ) 1/ =. σ A suitable form for is 4 = ( ( f '') ) 1 1 ( ( f f ) ) The root-ed density problem is a nonparametric regression with signal h = f, and there are K observed values of Y k, each having σ = K n. We get 39
42 DE h K n 1 = K 1/ h = 1/ ( hx ( ) h) ( hx ( ) h) n ( ) ( ) h (Note that K cancels out of this expression, as it should so long as n/k is not very small.) This is the DE for the root-ed problem. Unrooting makes the problem twice as hard since ( h+ ε ) h + ε = f + ε; The complexity factor also changes when comparing the rooted and unrooted problems, but the change is small; thus h = 1 ± ( a little) f Hence we should ascribe to the density problem the difficulty 1 DE orig prob =. ( ) 1/ ( hx ( ) h) 1 n 40
43 Here is a table showing DE orig for our density problem as compared to the DE s for the regression setup of Donoho, et. al. in which SNR = 7 and n = DE for Function density regression Heavisine Blocks DE s for n = (DE reg is for SNR=7.) According to this measure, the density problem is from 0 to 30 times harder than the regression one at the same sample size. THUS It needs from 400 to 900 as many observations to provide a comparable quality of estimate. 41
44 Empirical Example From a Bank Call-in Center in Israel The data are records of all telephone calls made during 1999 to a telephone assistance and service center operated by a bank in Israel. Today I ll present just an analysis of the time of day that these calls arrive at the center. We look only at telephone business hours - 7am to midnight on regular weekdays (Sun. through Thurs. in Israel), excluding holidays. The first plot is for REGULAR customers. (n =58,500). The second is for INTERNET customers (n = 14,30), who call a special number seeking assistance with on-line banking. (Differences for the days of the week and the seasons of the year did exist, but were relatively minor. These are ignored in the current analysis, but will be considered in the analysis in the last section of the talk.) An advantage of the root-unroot methodology is that it reduces the problem to a homoscedastic regression in which better-understood tools lead to 95% confidence (variance) interval bands in addition to estimates. The following plots were produced via a root-unroot scheme, using the free-knots nonparametric regression spline methodology of Zhao (1999, rev. 001). 4
45 Arrival time of REGULAR calls 43
46 Arr ival tim e of IN TE RN ET call s Note that the arrival patterns of regular and internet customers are significantly different. (The internet calls arrive more uniformly across the day, but with a statistically significant, noticeable late night local-mode at 10-11pm.) 44
Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Asymptotic Nonequivalence of Nonparametric Experiments When the Smoothness Index is ½ Lawrence D. Brown University
More informationEmpirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss;
BFF4, May 2, 2017 Empirical Bayes Quantile-Prediction aka E-B Prediction under Check-loss; Lawrence D. Brown Wharton School, Univ. of Pennsylvania Joint work with Gourab Mukherjee and Paat Rusmevichientong
More informationBayesian Nonparametric Point Estimation Under a Conjugate Prior
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 5-15-2002 Bayesian Nonparametric Point Estimation Under a Conjugate Prior Xuefeng Li University of Pennsylvania Linda
More informationThe Root-Unroot Algorithm for Density Estimation as Implemented. via Wavelet Block Thresholding
The Root-Unroot Algorithm for Density Estimation as Implemented via Wavelet Block Thresholding Lawrence Brown, Tony Cai, Ren Zhang, Linda Zhao and Harrison Zhou Abstract We propose and implement a density
More informationLawrence D. Brown* and Daniel McCarthy*
Comments on the paper, An adaptive resampling test for detecting the presence of significant predictors by I. W. McKeague and M. Qian Lawrence D. Brown* and Daniel McCarthy* ABSTRACT: This commentary deals
More informationChapter 11 Output Analysis for a Single Model. Banks, Carson, Nelson & Nicol Discrete-Event System Simulation
Chapter 11 Output Analysis for a Single Model Banks, Carson, Nelson & Nicol Discrete-Event System Simulation Purpose Objective: Estimate system performance via simulation If q is the system performance,
More informationL. Brown. Statistics Department, Wharton School University of Pennsylvania
Non-parametric Empirical Bayes and Compound Bayes Estimation of Independent Normal Means Joint work with E. Greenshtein L. Brown Statistics Department, Wharton School University of Pennsylvania lbrown@wharton.upenn.edu
More informationStatistics - Lecture One. Outline. Charlotte Wickham 1. Basic ideas about estimation
Statistics - Lecture One Charlotte Wickham wickham@stat.berkeley.edu http://www.stat.berkeley.edu/~wickham/ Outline 1. Basic ideas about estimation 2. Method of Moments 3. Maximum Likelihood 4. Confidence
More informationSparse Nonparametric Density Estimation in High Dimensions Using the Rodeo
Outline in High Dimensions Using the Rodeo Han Liu 1,2 John Lafferty 2,3 Larry Wasserman 1,2 1 Statistics Department, 2 Machine Learning Department, 3 Computer Science Department, Carnegie Mellon University
More informationConfidence intervals for kernel density estimation
Stata User Group - 9th UK meeting - 19/20 May 2003 Confidence intervals for kernel density estimation Carlo Fiorio c.fiorio@lse.ac.uk London School of Economics and STICERD Stata User Group - 9th UK meeting
More informationWavelet Shrinkage for Nonequispaced Samples
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1998 Wavelet Shrinkage for Nonequispaced Samples T. Tony Cai University of Pennsylvania Lawrence D. Brown University
More informationBoundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta
Boundary Correction Methods in Kernel Density Estimation Tom Alberts C o u(r)a n (t) Institute joint work with R.J. Karunamuni University of Alberta November 29, 2007 Outline Overview of Kernel Density
More informationA nonparametric two-sample wald test of equality of variances
University of Wollongong Research Online Faculty of Informatics - Papers (Archive) Faculty of Engineering and Information Sciences 211 A nonparametric two-sample wald test of equality of variances David
More informationThe Nonparametric Bootstrap
The Nonparametric Bootstrap The nonparametric bootstrap may involve inferences about a parameter, but we use a nonparametric procedure in approximating the parametric distribution using the ECDF. We use
More informationHigh-dimensional regression
High-dimensional regression Advanced Methods for Data Analysis 36-402/36-608) Spring 2014 1 Back to linear regression 1.1 Shortcomings Suppose that we are given outcome measurements y 1,... y n R, and
More informationVariables, distributions, and samples (cont.) Phil 12: Logic and Decision Making Fall 2010 UC San Diego 10/18/2010
Variables, distributions, and samples (cont.) Phil 12: Logic and Decision Making Fall 2010 UC San Diego 10/18/2010 Review Recording observations - Must extract that which is to be analyzed: coding systems,
More informationComputer Science, Informatik 4 Communication and Distributed Systems. Simulation. Discrete-Event System Simulation. Dr.
Simulation Discrete-Event System Simulation Chapter 0 Output Analysis for a Single Model Purpose Objective: Estimate system performance via simulation If θ is the system performance, the precision of the
More informationResearch Article A Nonparametric Two-Sample Wald Test of Equality of Variances
Advances in Decision Sciences Volume 211, Article ID 74858, 8 pages doi:1.1155/211/74858 Research Article A Nonparametric Two-Sample Wald Test of Equality of Variances David Allingham 1 andj.c.w.rayner
More informationMark Gordon Low
Mark Gordon Low lowm@wharton.upenn.edu Address Department of Statistics, The Wharton School University of Pennsylvania 3730 Walnut Street Philadelphia, PA 19104-6340 lowm@wharton.upenn.edu Education Ph.D.
More informationLawrence D. Brown, T. Tony Cai and Anirban DasGupta
Statistical Science 2005, Vol. 20, No. 4, 375 379 DOI 10.1214/088342305000000395 Institute of Mathematical Statistics, 2005 Comment: Fuzzy and Randomized Confidence Intervals and P -Values Lawrence D.
More informationBetter Bootstrap Confidence Intervals
by Bradley Efron University of Washington, Department of Statistics April 12, 2012 An example Suppose we wish to make inference on some parameter θ T (F ) (e.g. θ = E F X ), based on data We might suppose
More informationPractice Problems Section Problems
Practice Problems Section 4-4-3 4-4 4-5 4-6 4-7 4-8 4-10 Supplemental Problems 4-1 to 4-9 4-13, 14, 15, 17, 19, 0 4-3, 34, 36, 38 4-47, 49, 5, 54, 55 4-59, 60, 63 4-66, 68, 69, 70, 74 4-79, 81, 84 4-85,
More informationQuantitative Empirical Methods Exam
Quantitative Empirical Methods Exam Yale Department of Political Science, August 2016 You have seven hours to complete the exam. This exam consists of three parts. Back up your assertions with mathematics
More informationCharacterizing Forecast Uncertainty Prediction Intervals. The estimated AR (and VAR) models generate point forecasts of y t+s, y ˆ
Characterizing Forecast Uncertainty Prediction Intervals The estimated AR (and VAR) models generate point forecasts of y t+s, y ˆ t + s, t. Under our assumptions the point forecasts are asymtotically unbiased
More information12 Statistical Justifications; the Bias-Variance Decomposition
Statistical Justifications; the Bias-Variance Decomposition 65 12 Statistical Justifications; the Bias-Variance Decomposition STATISTICAL JUSTIFICATIONS FOR REGRESSION [So far, I ve talked about regression
More informationChapter 11. Output Analysis for a Single Model Prof. Dr. Mesut Güneş Ch. 11 Output Analysis for a Single Model
Chapter Output Analysis for a Single Model. Contents Types of Simulation Stochastic Nature of Output Data Measures of Performance Output Analysis for Terminating Simulations Output Analysis for Steady-state
More informationDiscussion of Regularization of Wavelets Approximations by A. Antoniadis and J. Fan
Discussion of Regularization of Wavelets Approximations by A. Antoniadis and J. Fan T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania Professors Antoniadis and Fan are
More informationModel Selection and Geometry
Model Selection and Geometry Pascal Massart Université Paris-Sud, Orsay Leipzig, February Purpose of the talk! Concentration of measure plays a fundamental role in the theory of model selection! Model
More informationDiscussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon
Discussion of the paper Inference for Semiparametric Models: Some Questions and an Answer by Bickel and Kwon Jianqing Fan Department of Statistics Chinese University of Hong Kong AND Department of Statistics
More informationinterval forecasting
Interval Forecasting Based on Chapter 7 of the Time Series Forecasting by Chatfield Econometric Forecasting, January 2008 Outline 1 2 3 4 5 Terminology Interval Forecasts Density Forecast Fan Chart Most
More informationOptimal Estimation of a Nonsmooth Functional
Optimal Estimation of a Nonsmooth Functional T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania http://stat.wharton.upenn.edu/ tcai Joint work with Mark Low 1 Question Suppose
More informationStat 451 Lecture Notes Monte Carlo Integration
Stat 451 Lecture Notes 06 12 Monte Carlo Integration Ryan Martin UIC www.math.uic.edu/~rgmartin 1 Based on Chapter 6 in Givens & Hoeting, Chapter 23 in Lange, and Chapters 3 4 in Robert & Casella 2 Updated:
More informationWeb-based Supplementary Material for. Dependence Calibration in Conditional Copulas: A Nonparametric Approach
1 Web-based Supplementary Material for Dependence Calibration in Conditional Copulas: A Nonparametric Approach Elif F. Acar, Radu V. Craiu, and Fang Yao Web Appendix A: Technical Details The score and
More informationBayesian Quadrature: Model-based Approximate Integration. David Duvenaud University of Cambridge
Bayesian Quadrature: Model-based Approimate Integration David Duvenaud University of Cambridge The Quadrature Problem ˆ We want to estimate an integral Z = f ()p()d ˆ Most computational problems in inference
More informationChapter 9. Non-Parametric Density Function Estimation
9-1 Density Estimation Version 1.2 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least
More informationAdditive Isotonic Regression
Additive Isotonic Regression Enno Mammen and Kyusang Yu 11. July 2006 INTRODUCTION: We have i.i.d. random vectors (Y 1, X 1 ),..., (Y n, X n ) with X i = (X1 i,..., X d i ) and we consider the additive
More informationPlugin Confidence Intervals in Discrete Distributions
Plugin Confidence Intervals in Discrete Distributions T. Tony Cai Department of Statistics The Wharton School University of Pennsylvania Philadelphia, PA 19104 Abstract The standard Wald interval is widely
More informationAUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET. Questions AUTOMATIC CONTROL COMMUNICATION SYSTEMS LINKÖPINGS UNIVERSITET
The Problem Identification of Linear and onlinear Dynamical Systems Theme : Curve Fitting Division of Automatic Control Linköping University Sweden Data from Gripen Questions How do the control surface
More informationIntroduction Wavelet shrinage methods have been very successful in nonparametric regression. But so far most of the wavelet regression methods have be
Wavelet Estimation For Samples With Random Uniform Design T. Tony Cai Department of Statistics, Purdue University Lawrence D. Brown Department of Statistics, University of Pennsylvania Abstract We show
More informationIntroduction to Regression
Introduction to Regression Chad M. Schafer May 20, 2015 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures Cross Validation Local Polynomial Regression
More informationDr. Maddah ENMG 617 EM Statistics 10/15/12. Nonparametric Statistics (2) (Goodness of fit tests)
Dr. Maddah ENMG 617 EM Statistics 10/15/12 Nonparametric Statistics (2) (Goodness of fit tests) Introduction Probability models used in decision making (Operations Research) and other fields require fitting
More informationRecall the Basics of Hypothesis Testing
Recall the Basics of Hypothesis Testing The level of significance α, (size of test) is defined as the probability of X falling in w (rejecting H 0 ) when H 0 is true: P(X w H 0 ) = α. H 0 TRUE H 1 TRUE
More informationWEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION. Abstract
Journal of Data Science,17(1). P. 145-160,2019 DOI:10.6339/JDS.201901_17(1).0007 WEIGHTED QUANTILE REGRESSION THEORY AND ITS APPLICATION Wei Xiong *, Maozai Tian 2 1 School of Statistics, University of
More informationMA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems
MA/ST 810 Mathematical-Statistical Modeling and Analysis of Complex Systems Principles of Statistical Inference Recap of statistical models Statistical inference (frequentist) Parametric vs. semiparametric
More informationMallows Cp for Out-of-sample Prediction
Mallows Cp for Out-of-sample Prediction Lawrence D. Brown Statistics Department, Wharton School, University of Pennsylvania lbrown@wharton.upenn.edu WHOA-PSI conference, St. Louis, Oct 1, 2016 Joint work
More informationIntroduction to Econometrics. Review of Probability & Statistics
1 Introduction to Econometrics Review of Probability & Statistics Peerapat Wongchaiwat, Ph.D. wongchaiwat@hotmail.com Introduction 2 What is Econometrics? Econometrics consists of the application of mathematical
More informationA Simulation Study on Confidence Interval Procedures of Some Mean Cumulative Function Estimators
Statistics Preprints Statistics -00 A Simulation Study on Confidence Interval Procedures of Some Mean Cumulative Function Estimators Jianying Zuo Iowa State University, jiyizu@iastate.edu William Q. Meeker
More informationWhat to do if Assumptions are Violated?
What to do if Assumptions are Violated? Abandon simple linear regression for something else (usually more complicated). Some examples of alternative models: weighted least square appropriate model if the
More informationGaussian Processes. Le Song. Machine Learning II: Advanced Topics CSE 8803ML, Spring 2012
Gaussian Processes Le Song Machine Learning II: Advanced Topics CSE 8803ML, Spring 01 Pictorial view of embedding distribution Transform the entire distribution to expected features Feature space Feature
More informationWeek 9 The Central Limit Theorem and Estimation Concepts
Week 9 and Estimation Concepts Week 9 and Estimation Concepts Week 9 Objectives 1 The Law of Large Numbers and the concept of consistency of averages are introduced. The condition of existence of the population
More informationIntroduction to Regression
Introduction to Regression Chad M. Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/100 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression
More informationTerminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1
Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the
More informationAnalysis of Call Center Data
University of Pennsylvania ScholarlyCommons Wharton Research Scholars Wharton School 4-1-2004 Analysis of Call Center Data Yu Chu Cheng University of Pennsylvania Follow this and additional works at: http://repository.upenn.edu/wharton_research_scholars
More informationIntroduction to Regression
Introduction to Regression David E Jones (slides mostly by Chad M Schafer) June 1, 2016 1 / 102 Outline General Concepts of Regression, Bias-Variance Tradeoff Linear Regression Nonparametric Procedures
More informationPart 4: Multi-parameter and normal models
Part 4: Multi-parameter and normal models 1 The normal model Perhaps the most useful (or utilized) probability model for data analysis is the normal distribution There are several reasons for this, e.g.,
More informationModel-free prediction intervals for regression and autoregression. Dimitris N. Politis University of California, San Diego
Model-free prediction intervals for regression and autoregression Dimitris N. Politis University of California, San Diego To explain or to predict? Models are indispensable for exploring/utilizing relationships
More informationThe choice of weights in kernel regression estimation
Biometrika (1990), 77, 2, pp. 377-81 Printed in Great Britain The choice of weights in kernel regression estimation BY THEO GASSER Zentralinstitut fur Seelische Gesundheit, J5, P.O.B. 5970, 6800 Mannheim
More informationINFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION. 1. Introduction
INFERENCE APPROACHES FOR INSTRUMENTAL VARIABLE QUANTILE REGRESSION VICTOR CHERNOZHUKOV CHRISTIAN HANSEN MICHAEL JANSSON Abstract. We consider asymptotic and finite-sample confidence bounds in instrumental
More informationGeneralized Cp (GCp) in a Model Lean Framework
Generalized Cp (GCp) in a Model Lean Framework Linda Zhao University of Pennsylvania Dedicated to Lawrence Brown (1940-2018) September 9th, 2018 WHOA 3 Joint work with Larry Brown, Juhui Cai, Arun Kumar
More informationIntroduction to the regression problem. Luca Martino
Introduction to the regression problem Luca Martino 2017 2018 1 / 30 Approximated outline of the course 1. Very basic introduction to regression 2. Gaussian Processes (GPs) and Relevant Vector Machines
More informationUnderstanding Generalization Error: Bounds and Decompositions
CIS 520: Machine Learning Spring 2018: Lecture 11 Understanding Generalization Error: Bounds and Decompositions Lecturer: Shivani Agarwal Disclaimer: These notes are designed to be a supplement to the
More informationUsing R in Undergraduate Probability and Mathematical Statistics Courses. Amy G. Froelich Department of Statistics Iowa State University
Using R in Undergraduate Probability and Mathematical Statistics Courses Amy G. Froelich Department of Statistics Iowa State University Undergraduate Probability and Mathematical Statistics at Iowa State
More informationUniform Inference for Conditional Factor Models with Instrumental and Idiosyncratic Betas
Uniform Inference for Conditional Factor Models with Instrumental and Idiosyncratic Betas Yuan Xiye Yang Rutgers University Dec 27 Greater NY Econometrics Overview main results Introduction Consider a
More informationSingle Index Quantile Regression for Heteroscedastic Data
Single Index Quantile Regression for Heteroscedastic Data E. Christou M. G. Akritas Department of Statistics The Pennsylvania State University SMAC, November 6, 2015 E. Christou, M. G. Akritas (PSU) SIQR
More informationLecture 4: Multivariate Regression, Part 2
Lecture 4: Multivariate Regression, Part 2 Gauss-Markov Assumptions 1) Linear in Parameters: Y X X X i 0 1 1 2 2 k k 2) Random Sampling: we have a random sample from the population that follows the above
More informationAsymptotic Equivalence and Adaptive Estimation for Robust Nonparametric Regression
Asymptotic Equivalence and Adaptive Estimation for Robust Nonparametric Regression T. Tony Cai 1 and Harrison H. Zhou 2 University of Pennsylvania and Yale University Abstract Asymptotic equivalence theory
More informationAdaptive Wavelet Estimation: A Block Thresholding and Oracle Inequality Approach
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 1999 Adaptive Wavelet Estimation: A Block Thresholding and Oracle Inequality Approach T. Tony Cai University of Pennsylvania
More informationA Classroom Approach to Illustrate Transformation and Bootstrap Confidence Interval Techniques Using the Poisson Distribution
International Journal of Statistics and Probability; Vol. 6, No. 2; March 2017 ISSN 1927-7032 E-ISSN 1927-7040 Published by Canadian Center of Science and Education A Classroom Approach to Illustrate Transformation
More informationMinimax Estimation of a nonlinear functional on a structured high-dimensional model
Minimax Estimation of a nonlinear functional on a structured high-dimensional model Eric Tchetgen Tchetgen Professor of Biostatistics and Epidemiologic Methods, Harvard U. (Minimax ) 1 / 38 Outline Heuristics
More informationFinding an upper limit in the presence of an unknown background
PHYSICAL REVIEW D 66, 032005 2002 Finding an upper limit in the presence of an unnown bacground S. Yellin* Department of Physics, University of California, Santa Barbara, Santa Barbara, California 93106
More informationIntroduction to Regression
Introduction to Regression p. 1/97 Introduction to Regression Chad Schafer cschafer@stat.cmu.edu Carnegie Mellon University Introduction to Regression p. 1/97 Acknowledgement Larry Wasserman, All of Nonparametric
More informationHypothesis Testing. ) the hypothesis that suggests no change from previous experience
Hypothesis Testing Definitions Hypothesis a claim about something Null hypothesis ( H 0 ) the hypothesis that suggests no change from previous experience Alternative hypothesis ( H 1 ) the hypothesis that
More informationSection 8.1: Interval Estimation
Section 8.1: Interval Estimation Discrete-Event Simulation: A First Course c 2006 Pearson Ed., Inc. 0-13-142917-5 Discrete-Event Simulation: A First Course Section 8.1: Interval Estimation 1/ 35 Section
More informationChapter 9. Non-Parametric Density Function Estimation
9-1 Density Estimation Version 1.1 Chapter 9 Non-Parametric Density Function Estimation 9.1. Introduction We have discussed several estimation techniques: method of moments, maximum likelihood, and least
More informationHow Accurate is My Forecast?
How Accurate is My Forecast? Tao Hong, PhD Utilities Business Unit, SAS 15 May 2012 PLEASE STAND BY Today s event will begin at 11:00am EDT The audio portion of the presentation will be heard through your
More informationRobustness to Parametric Assumptions in Missing Data Models
Robustness to Parametric Assumptions in Missing Data Models Bryan Graham NYU Keisuke Hirano University of Arizona April 2011 Motivation Motivation We consider the classic missing data problem. In practice
More informationCan we do statistical inference in a non-asymptotic way? 1
Can we do statistical inference in a non-asymptotic way? 1 Guang Cheng 2 Statistics@Purdue www.science.purdue.edu/bigdata/ ONR Review Meeting@Duke Oct 11, 2017 1 Acknowledge NSF, ONR and Simons Foundation.
More informationBusiness Statistics. Lecture 10: Course Review
Business Statistics Lecture 10: Course Review 1 Descriptive Statistics for Continuous Data Numerical Summaries Location: mean, median Spread or variability: variance, standard deviation, range, percentiles,
More informationTwo Correlated Proportions Non- Inferiority, Superiority, and Equivalence Tests
Chapter 59 Two Correlated Proportions on- Inferiority, Superiority, and Equivalence Tests Introduction This chapter documents three closely related procedures: non-inferiority tests, superiority (by a
More informationSmooth nonparametric estimation of a quantile function under right censoring using beta kernels
Smooth nonparametric estimation of a quantile function under right censoring using beta kernels Chanseok Park 1 Department of Mathematical Sciences, Clemson University, Clemson, SC 29634 Short Title: Smooth
More informationStatistical inference on Lévy processes
Alberto Coca Cabrero University of Cambridge - CCA Supervisors: Dr. Richard Nickl and Professor L.C.G.Rogers Funded by Fundación Mutua Madrileña and EPSRC MASDOC/CCA student workshop 2013 26th March Outline
More informationNonparametric Methods
Nonparametric Methods Michael R. Roberts Department of Finance The Wharton School University of Pennsylvania July 28, 2009 Michael R. Roberts Nonparametric Methods 1/42 Overview Great for data analysis
More informationEconometrics I. Lecture 10: Nonparametric Estimation with Kernels. Paul T. Scott NYU Stern. Fall 2018
Econometrics I Lecture 10: Nonparametric Estimation with Kernels Paul T. Scott NYU Stern Fall 2018 Paul T. Scott NYU Stern Econometrics I Fall 2018 1 / 12 Nonparametric Regression: Intuition Let s get
More informationResearch Statement. Harrison H. Zhou Cornell University
Research Statement Harrison H. Zhou Cornell University My research interests and contributions are in the areas of model selection, asymptotic decision theory, nonparametric function estimation and machine
More informationAdaptive Online Gradient Descent
University of Pennsylvania ScholarlyCommons Statistics Papers Wharton Faculty Research 6-4-2007 Adaptive Online Gradient Descent Peter Bartlett Elad Hazan Alexander Rakhlin University of Pennsylvania Follow
More informationADJUSTED POWER ESTIMATES IN. Ji Zhang. Biostatistics and Research Data Systems. Merck Research Laboratories. Rahway, NJ
ADJUSTED POWER ESTIMATES IN MONTE CARLO EXPERIMENTS Ji Zhang Biostatistics and Research Data Systems Merck Research Laboratories Rahway, NJ 07065-0914 and Dennis D. Boos Department of Statistics, North
More information11 Survival Analysis and Empirical Likelihood
11 Survival Analysis and Empirical Likelihood The first paper of empirical likelihood is actually about confidence intervals with the Kaplan-Meier estimator (Thomas and Grunkmeier 1979), i.e. deals with
More informationDiscussant: Lawrence D Brown* Statistics Department, Wharton, Univ. of Penn.
Discussion of Estimation and Accuracy After Model Selection By Bradley Efron Discussant: Lawrence D Brown* Statistics Department, Wharton, Univ. of Penn. lbrown@wharton.upenn.edu JSM, Boston, Aug. 4, 2014
More informationWhy Do Statisticians Treat Predictors as Fixed? A Conspiracy Theory
Why Do Statisticians Treat Predictors as Fixed? A Conspiracy Theory Andreas Buja joint with the PoSI Group: Richard Berk, Lawrence Brown, Linda Zhao, Kai Zhang Ed George, Mikhail Traskin, Emil Pitkin,
More informationO Combining cross-validation and plug-in methods - for kernel density bandwidth selection O
O Combining cross-validation and plug-in methods - for kernel density selection O Carlos Tenreiro CMUC and DMUC, University of Coimbra PhD Program UC UP February 18, 2011 1 Overview The nonparametric problem
More informationOverall Plan of Simulation and Modeling I. Chapters
Overall Plan of Simulation and Modeling I Chapters Introduction to Simulation Discrete Simulation Analytical Modeling Modeling Paradigms Input Modeling Random Number Generation Output Analysis Continuous
More informationChapter 11. Regression with a Binary Dependent Variable
Chapter 11 Regression with a Binary Dependent Variable 2 Regression with a Binary Dependent Variable (SW Chapter 11) So far the dependent variable (Y) has been continuous: district-wide average test score
More informationLeast Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error Distributions
Journal of Modern Applied Statistical Methods Volume 8 Issue 1 Article 13 5-1-2009 Least Absolute Value vs. Least Squares Estimation and Inference Procedures in Regression Models with Asymmetric Error
More informationMIT Spring 2015
Assessing Goodness Of Fit MIT 8.443 Dr. Kempthorne Spring 205 Outline 2 Poisson Distribution Counts of events that occur at constant rate Counts in disjoint intervals/regions are independent If intervals/regions
More informationCOS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION
COS513: FOUNDATIONS OF PROBABILISTIC MODELS LECTURE 9: LINEAR REGRESSION SEAN GERRISH AND CHONG WANG 1. WAYS OF ORGANIZING MODELS In probabilistic modeling, there are several ways of organizing models:
More informationNonparametric Inference In Functional Data
Nonparametric Inference In Functional Data Zuofeng Shang Purdue University Joint work with Guang Cheng from Purdue Univ. An Example Consider the functional linear model: Y = α + where 1 0 X(t)β(t)dt +
More information12 - Nonparametric Density Estimation
ST 697 Fall 2017 1/49 12 - Nonparametric Density Estimation ST 697 Fall 2017 University of Alabama Density Review ST 697 Fall 2017 2/49 Continuous Random Variables ST 697 Fall 2017 3/49 1.0 0.8 F(x) 0.6
More information41903: Introduction to Nonparametrics
41903: Notes 5 Introduction Nonparametrics fundamentally about fitting flexible models: want model that is flexible enough to accommodate important patterns but not so flexible it overspecializes to specific
More informationNonparametric Principal Components Regression
Int. Statistical Inst.: Proc. 58th World Statistical Congress, 2011, Dublin (Session CPS031) p.4574 Nonparametric Principal Components Regression Barrios, Erniel University of the Philippines Diliman,
More informationSimple and Honest Confidence Intervals in Nonparametric Regression
Simple and Honest Confidence Intervals in Nonparametric Regression Timothy B. Armstrong Yale University Michal Kolesár Princeton University June, 206 Abstract We consider the problem of constructing honest
More information