Active Set Methods for Log-Concave Densities and Nonparametric Tail Inflation

Size: px
Start display at page:

Download "Active Set Methods for Log-Concave Densities and Nonparametric Tail Inflation"

Transcription

1 Active Set Methods for Log-Concave Densities and Nonparametric Tail Inflation Lutz Duembgen (Bern) Aleandre Moesching and Christof Straehl (Bern) Peter McCullagh and Nicholas G. Polson (Chicago) January 2018

2 The general setting Data. Summarized as a distribution P = n w i δ i i=1 with weights w 1, w 2,..., w n > 0 support points 1 < 2 < < n in an open interval X R

3 The general setting Data. Summarized as a distribution P = n w i δ i i=1 with weights w 1, w 2,..., w n > 0 support points 1 < 2 < < n in an open interval X R Model. P estimates distribution P(d) = e θ() M(d) with given measure M on X unknown function θ in given family Θ 1

4 Goal. Estimate θ Θ 1 via MLE θ arg ma θ Θ 1 θ d P

5 Goal. Estimate θ Θ 1 via MLE θ arg ma θ Θ 1 θ d P Modification. Suppose that for a larger function family Θ { } Θ 1 = θ Θ : e θ dm = 1 θ + c Θ for all θ Θ, c R Then ( θ arg ma θ d P θ Θ ) e θ dm

6 Setting 1: Log-concave densities P has log-concave density on X, i.e. M = Lebesgue measure on X Θ = { θ : X [, ) concave and u.s.c. } f() log(f())

7 Setting 2A: Tail inflation P has log-conve density w.r.t. given distribution P 0 on X, i.e. M = P 0 Θ = { θ : X R conve } p 0(), p() θ()

8 Eample. Observe X i = µ i + σ i ε i, 1 i n, with unknown parameters µ i R, σ i 1 and independent r.v.s ε i P 0 := N (0, 1)

9 Eample. Observe X i = µ i + σ i ε i, 1 i n, with unknown parameters µ i R, σ i 1 and independent r.v.s ε i P 0 := N (0, 1) Marginal dist. P := n 1 i=1 L(X i): θ() := log dp dp 0 () = log( 1 n n ) e θ i () i=1 θ i () := log(σ i ) ( µ i) 2 2σi 2 θ i, θ 0 (Artin s theorem)

10 Setting 2B: Tail inflation (McCullagh and Polson 2012) Assume that P has log-conve and isotonic density w.r.t. given distribution P 0 on X, i.e. M = P 0 Θ = { θ : X R conve and isotonic }. p 0(), p() θ()

11 Eample. Observe with independent r.v.s X i = S i ε i, 1 i n, ε i N (0, 1), S i 1.

12 Eample. Observe with independent r.v.s X i = S i ε i, 1 i n, ε i N (0, 1), S i 1. Then Y i := X 2 i P 0 = χ 2 1 = Gamma(1/2, 2) if S i 1 IE Gamma(1/2, 2Si 2) in general

13 Eample. Observe with independent r.v.s X i = S i ε i, 1 i n, ε i N (0, 1), S i 1. Then Y i := X 2 i P 0 = χ 2 1 = Gamma(1/2, 2) if S i 1 IE Gamma(1/2, 2Si 2) in general Marginal dist. P := n 1 i=1 L(Y i): θ := log dp dp 0 is conve and isotonic

14 Eistence and uniqueness of θ Lemma 1 (Log-concavity) In Setting 1, (! θ arg ma θ d P ) e θ() d. θ Θ X

15 Eistence and uniqueness of θ Lemma 1 (Log-concavity) In Setting 1, (! θ arg ma θ d P ) e θ() d. θ Θ X Additional properties: θ piecewise linear on [1, n ] changes of slope only in { 2,..., n 1 } θ on R \ [1, n ]

16 Lemma 2. In Settings 2A and 2B, (! θ arg ma θ d P θ Θ e θ dp 0 ) provided that supp(p 0 ) = X.

17 Lemma 2. In Settings 2A and 2B, (! θ arg ma θ d P θ Θ e θ dp 0 ) provided that supp(p 0 ) = X. Additional properties: θ piecewise linear on X changes of slope only in n 1 i=1 ( i, i+1 ) at most one change of slope in ( i, i+1 ), 1 i < n

18 Active set algorithm In Settings 1 and 2A, θ V Θ with V := { linear splines on X o with kinks on D } { [ 1, n ] in Setting 1 X o := X in Setting 2A { { 2,..., n 1 } in Setting 1 D := X in Setting 2A

19 For such a spline v V set D(v) := { τ D : v (τ ) v (τ +) } (deactivated (equality) constraints)

20 For such a spline v V set D(v) := { τ D : v (τ ) v (τ +) } (deactivated (equality) constraints) For finite set D D define V D := { v V : D(v) D } a linear space with dim(v D ) = 2 + #D

21 Target functional L(θ) := θ d P e θ dm with directional derivatives DL(θ, v) := lim t 0 + L(θ + tv) L(θ) t

22 Target functional L(θ) := θ d P e θ dm with directional derivatives DL(θ, v) := lim t 0 + L(θ + tv) L(θ) t Characterization of θ. θ V Θ equals θ if, and only if, DL(θ, v) 0 whenever θ + tv Θ for some t > 0.

23 Local Search (Shape-constrained Newton). Let θ V Θ. θ new Newton(θ, V D(θ) ) δ DL(θ, θ new θ) while δ > δ o do while L(θ new ) < L(θ) + δ/3 do θ new (θ + θ new )/2 δ δ/2 end while if θ new Θ do t o ma { t (0, 1] : (1 t)θ + tθ new Θ } θ new (1 t o )θ + t o θ new end if θ θ new θ new Newton(θ, V D(θ) ) δ DL(θ, θ new θ) end while

24 Essential properties of local search:

25 Essential properties of local search: L(θ) increases

26 Essential properties of local search: L(θ) increases D(θ) decreases

27 Essential properties of local search: L(θ) increases D(θ) decreases Eventually, θ is locally optimal (appro.): θ = arg ma η V D(θ) L(η)

28 Checking optimality. Let θ V Θ be locally optimal (appro.). Then θ = θ if, and only if, DL(θ, V τ ) 0 for all τ D \ D(θ) where V τ () := { ( τ) + in Setting 1 +( τ) + in Setting 2A

29 Checking optimality. Let θ V Θ be locally optimal (appro.). Then θ = θ if, and only if, DL(θ, V τ ) 0 for all τ D \ D(θ) where V τ () := { ( τ) + in Setting 1 +( τ) + in Setting 2A If not, determine τ(θ) D \ D(θ) such that 0 < DL(θ, V τ(θ) ) ma τ D\D(θ) DL(θ, V τ ) and run a modified local search.

30 Modified local search. θ new Newton(θ, V D(θ) {τ (θ)} ) δ DL(θ, θ new θ) while δ > δ o do while L(θ new ) < L(θ) + δ/3 do θ new (θ + θ new )/2 δ δ/2 end while if θ new Θ do t o ma { t (0, 1] : (1 t)θ + tθ new Θ } θ new (1 t o )θ + t o θ new end if θ θ new θ new Newton(θ, V D(θ) ) δ DL(θ, θ new θ) end while

31 Remarks. After finitely many (modified) local searches algorithm will stop at θ = θ (appro.)

32 Remarks. After finitely many (modified) local searches algorithm will stop at θ = θ (appro.) Replace simple kink functions V τ = ±( τ) + with localized versions to gain numerical precision

33 Remarks. After finitely many (modified) local searches algorithm will stop at θ = θ (appro.) Replace simple kink functions V τ = ±( τ) + with localized versions to gain numerical precision In Settings 2A and 2B, the function τ DL(θ, V τ ) is strictly concave on any interval ( i, i+1 )...

34 Eample for Setting 1. n = 800 observations from Gamma(3, 1) Starting point: LL = phi()

35 LL = , dir. deriv. = phi(), phi.new()

36 LL = , dir. deriv. = phi(), phi.new()

37 LL = , dir. deriv. = phi(), phi.new()

38 Newton proposals in dimensions 7, 7, 7, 7, 7, 7 Local optimum: LL = phi()

39 LL = , dir. deriv. = phi()

40 LL = , dir. deriv. = phi(), phi.new()

41 phi()

42 LL = , dir. deriv. = phi(), phi.new()

43 phi()

44 Newton proposals in dimensions 8, 7, 6, 6 Local optimum: LL = phi()

45 LL = , dir. deriv. = phi()

46 LL = , dir. deriv. = phi(), phi.new()

47 Newton proposals in dimensions 7, 7, 7, 7 Local optimum: LL = phi()

48 LL = , dir. deriv. = phi()

49 LL = , dir. deriv. = phi(), phi.new()

50 phi()

51 LL = , dir. deriv. = phi(), phi.new()

52 phi()

53 LL = , dir. deriv. = phi(), phi.new()

54 Newton proposals in dimensions 8, 7, 6, 6, 6 Local optimum: LL = phi()

55 LL = , dir. deriv. = phi()

56 Local optimum: LL = phi()

57 LL = , dir. deriv. = phi()

58 LL = , dir. deriv. = e-05 phi(), phi.new()

59 Local optimum: LL = phi()

60 After 14 (modified) local searches, 45 Newton proposals:

61 After 14 (modified) local searches, 45 Newton proposals: Global optimum: LL = φ()

62 Directional derivatives for etra knot at τ: DL(φ, V τ ) τ

63 Log-densities: φ()

64 Densities: f()

65 Eample for Setting 2B. n = 1000 observations from P P 0 = Gamma(1.5, 1.2) p 0(), p(), p()

66 θ(), θ()

67 Open questions for Settings 2A-B Suppose we replace discrete P with arbitrary distribution.

68 Open questions for Settings 2A-B Suppose we replace discrete P with arbitrary distribution. Under which conditions on P and P 0 eists a unique ( θ arg ma θ d P ) e θ dp 0? θ Θ

69 Open questions for Settings 2A-B Suppose we replace discrete P with arbitrary distribution. Under which conditions on P and P 0 eists a unique ( θ arg ma θ d P ) e θ dp 0? θ Θ What continuity properties has the mapping P θ?

70 Open questions for Settings 2A-B Suppose we replace discrete P with arbitrary distribution. Under which conditions on P and P 0 eists a unique ( θ arg ma θ d P ) e θ dp 0? θ Θ What continuity properties has the mapping P θ? (Analogous questions for Setting 1 well understood)

71 References Groeneboom, Jongbloed, Wellner (2008). The support reduction algorithm for computing nonparametric function estimates in miture models. Scand. J. Statist. 35 D., Hüsler, Rufibach (2007/2011). Active set and EM algorithms for log-concave densities based on complete and censored data. Technical report, IMSV, University of Bern (ariv: ) D., Mösching, Strähl (2018). Active set algorithms for density estimators under shape-constraints. In preparation McCullagh, Polson (2012). Tail inflation. Biometrika 99 D., Samworth, Schuhmacher (2011). Approimation by log-concave distributions, with applications to regression. Ann. Statist. 39

Maximum likelihood estimation of a log-concave density based on censored data

Maximum likelihood estimation of a log-concave density based on censored data Maximum likelihood estimation of a log-concave density based on censored data Dominic Schuhmacher Institute of Mathematical Statistics and Actuarial Science University of Bern Joint work with Lutz Dümbgen

More information

Nonparametric estimation of log-concave densities

Nonparametric estimation of log-concave densities Nonparametric estimation of log-concave densities Jon A. Wellner University of Washington, Seattle Seminaire, Institut de Mathématiques de Toulouse 5 March 2012 Seminaire, Toulouse Based on joint work

More information

1.4 FOUNDATIONS OF CONSTRAINED OPTIMIZATION

1.4 FOUNDATIONS OF CONSTRAINED OPTIMIZATION Essential Microeconomics -- 4 FOUNDATIONS OF CONSTRAINED OPTIMIZATION Fundamental Theorem of linear Programming 3 Non-linear optimization problems 6 Kuhn-Tucker necessary conditions Sufficient conditions

More information

Chapter 3 : Likelihood function and inference

Chapter 3 : Likelihood function and inference Chapter 3 : Likelihood function and inference 4 Likelihood function and inference The likelihood Information and curvature Sufficiency and ancilarity Maximum likelihood estimation Non-regular models EM

More information

OPTIMISATION CHALLENGES IN MODERN STATISTICS. Co-authors: Y. Chen, M. Cule, R. Gramacy, M. Yuan

OPTIMISATION CHALLENGES IN MODERN STATISTICS. Co-authors: Y. Chen, M. Cule, R. Gramacy, M. Yuan OPTIMISATION CHALLENGES IN MODERN STATISTICS Co-authors: Y. Chen, M. Cule, R. Gramacy, M. Yuan How do optimisation problems arise in Statistics? Let X 1,...,X n be independent and identically distributed

More information

Log Concavity and Density Estimation

Log Concavity and Density Estimation Log Concavity and Density Estimation Lutz Dümbgen, Kaspar Rufibach, André Hüsler (Universität Bern) Geurt Jongbloed (Amsterdam) Nicolai Bissantz (Göttingen) I. A Consulting Case II. III. IV. Log Concave

More information

Helpful Concepts for MTH 261 Final. What are the general strategies for determining the domain of a function?

Helpful Concepts for MTH 261 Final. What are the general strategies for determining the domain of a function? Helpful Concepts for MTH 261 Final What are the general strategies for determining the domain of a function? How do we use the graph of a function to determine its range? How many graphs of basic functions

More information

Nonparametric estimation: s concave and log-concave densities: alternatives to maximum likelihood

Nonparametric estimation: s concave and log-concave densities: alternatives to maximum likelihood Nonparametric estimation: s concave and log-concave densities: alternatives to maximum likelihood Jon A. Wellner University of Washington, Seattle Statistics Seminar, York October 15, 2015 Statistics Seminar,

More information

Nonparametric estimation under Shape Restrictions

Nonparametric estimation under Shape Restrictions Nonparametric estimation under Shape Restrictions Jon A. Wellner University of Washington, Seattle Statistical Seminar, Frejus, France August 30 - September 3, 2010 Outline: Five Lectures on Shape Restrictions

More information

Nonparametric estimation of log-concave densities

Nonparametric estimation of log-concave densities Nonparametric estimation of log-concave densities Jon A. Wellner University of Washington, Seattle Northwestern University November 5, 2010 Conference on Shape Restrictions in Non- and Semi-Parametric

More information

Shape constrained estimation: a brief introduction

Shape constrained estimation: a brief introduction Shape constrained estimation: a brief introduction Jon A. Wellner University of Washington, Seattle Cowles Foundation, Yale November 7, 2016 Econometrics lunch seminar, Yale Based on joint work with: Fadoua

More information

Nonparametric estimation under shape constraints, part 2

Nonparametric estimation under shape constraints, part 2 Nonparametric estimation under shape constraints, part 2 Piet Groeneboom, Delft University August 7, 2013 What to expect? Theory and open problems for interval censoring, case 2. Same for the bivariate

More information

Nonparametric estimation of. s concave and log-concave densities: alternatives to maximum likelihood

Nonparametric estimation of. s concave and log-concave densities: alternatives to maximum likelihood Nonparametric estimation of s concave and log-concave densities: alternatives to maximum likelihood Jon A. Wellner University of Washington, Seattle Cowles Foundation Seminar, Yale University November

More information

Review and continuation from last week Properties of MLEs

Review and continuation from last week Properties of MLEs Review and continuation from last week Properties of MLEs As we have mentioned, MLEs have a nice intuitive property, and as we have seen, they have a certain equivariance property. We will see later that

More information

arxiv:math/ v2 [math.st] 17 Jun 2008

arxiv:math/ v2 [math.st] 17 Jun 2008 The Annals of Statistics 2008, Vol. 36, No. 3, 1064 1089 DOI: 10.1214/009053607000000983 c Institute of Mathematical Statistics, 2008 arxiv:math/0609021v2 [math.st] 17 Jun 2008 CURRENT STATUS DATA WITH

More information

1969 AP Calculus BC: Section I

1969 AP Calculus BC: Section I 969 AP Calculus BC: Section I 9 Minutes No Calculator Note: In this eamination, ln denotes the natural logarithm of (that is, logarithm to the base e).. t The asymptotes of the graph of the parametric

More information

Statistics 3858 : Maximum Likelihood Estimators

Statistics 3858 : Maximum Likelihood Estimators Statistics 3858 : Maximum Likelihood Estimators 1 Method of Maximum Likelihood In this method we construct the so called likelihood function, that is L(θ) = L(θ; X 1, X 2,..., X n ) = f n (X 1, X 2,...,

More information

Geodesic Convexity and Regularized Scatter Estimation

Geodesic Convexity and Regularized Scatter Estimation Geodesic Convexity and Regularized Scatter Estimation Lutz Duembgen (Bern) David Tyler (Rutgers) Klaus Nordhausen (Turku/Vienna), Heike Schuhmacher (Bern) Markus Pauly (Ulm), Thomas Schweizer (Bern) Düsseldorf,

More information

Stat 710: Mathematical Statistics Lecture 12

Stat 710: Mathematical Statistics Lecture 12 Stat 710: Mathematical Statistics Lecture 12 Jun Shao Department of Statistics University of Wisconsin Madison, WI 53706, USA Jun Shao (UW-Madison) Stat 710, Lecture 12 Feb 18, 2009 1 / 11 Lecture 12:

More information

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R.

Ergodic Theorems. Samy Tindel. Purdue University. Probability Theory 2 - MA 539. Taken from Probability: Theory and examples by R. Ergodic Theorems Samy Tindel Purdue University Probability Theory 2 - MA 539 Taken from Probability: Theory and examples by R. Durrett Samy T. Ergodic theorems Probability Theory 1 / 92 Outline 1 Definitions

More information

FULL LIKELIHOOD INFERENCES IN THE COX MODEL

FULL LIKELIHOOD INFERENCES IN THE COX MODEL October 20, 2007 FULL LIKELIHOOD INFERENCES IN THE COX MODEL BY JIAN-JIAN REN 1 AND MAI ZHOU 2 University of Central Florida and University of Kentucky Abstract We use the empirical likelihood approach

More information

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions

Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions International Journal of Control Vol. 00, No. 00, January 2007, 1 10 Stochastic Optimization with Inequality Constraints Using Simultaneous Perturbations and Penalty Functions I-JENG WANG and JAMES C.

More information

The Kuhn-Tucker and Envelope Theorems

The Kuhn-Tucker and Envelope Theorems The Kuhn-Tucker and Envelope Theorems Peter Ireland EC720.01 - Math for Economists Boston College, Department of Economics Fall 2010 The Kuhn-Tucker and envelope theorems can be used to characterize the

More information

The International Journal of Biostatistics

The International Journal of Biostatistics The International Journal of Biostatistics Volume 1, Issue 1 2005 Article 3 Score Statistics for Current Status Data: Comparisons with Likelihood Ratio and Wald Statistics Moulinath Banerjee Jon A. Wellner

More information

Slopes and Rates of Change

Slopes and Rates of Change Slopes and Rates of Change If a particle is moving in a straight line at a constant velocity, then the graph of the function of distance versus time is as follows s s = f(t) t s s t t = average velocity

More information

The Kuhn-Tucker and Envelope Theorems

The Kuhn-Tucker and Envelope Theorems The Kuhn-Tucker and Envelope Theorems Peter Ireland ECON 77200 - Math for Economists Boston College, Department of Economics Fall 207 The Kuhn-Tucker and envelope theorems can be used to characterize the

More information

Computing Confidence Intervals for Log-Concave Densities. by Mahdis Azadbakhsh Supervisors: Hanna Jankowski Xin Gao York University

Computing Confidence Intervals for Log-Concave Densities. by Mahdis Azadbakhsh Supervisors: Hanna Jankowski Xin Gao York University Computing Confidence Intervals for Log-Concave Densities by Mahdis Azadbakhsh Supervisors: Hanna Jankowski Xin Gao York University August 2012 Contents 1 Introduction 1 1.1 Introduction..............................

More information

Topic 5: Non-Linear Relationships and Non-Linear Least Squares

Topic 5: Non-Linear Relationships and Non-Linear Least Squares Topic 5: Non-Linear Relationships and Non-Linear Least Squares Non-linear Relationships Many relationships between variables are non-linear. (Examples) OLS may not work (recall A.1). It may be biased and

More information

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics

Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Chapter 2: Fundamentals of Statistics Lecture 15: Models and statistics Data from one or a series of random experiments are collected. Planning experiments and collecting data (not discussed here). Analysis:

More information

NONPARAMETRIC ESTIMATION OF MULTIVARIATE CONVEX-TRANSFORMED DENSITIES. By Arseni Seregin, and Jon A. Wellner, University of Washington

NONPARAMETRIC ESTIMATION OF MULTIVARIATE CONVEX-TRANSFORMED DENSITIES. By Arseni Seregin, and Jon A. Wellner, University of Washington Submitted to the Annals of Statistics arxiv: math.st/0911.4151v1 NONPARAMETRIC ESTIMATION OF MULTIVARIATE CONVEX-TRANSFORMED DENSITIES By Arseni Seregin, and Jon A. Wellner, University of Washington We

More information

Estimation of Dynamic Regression Models

Estimation of Dynamic Regression Models University of Pavia 2007 Estimation of Dynamic Regression Models Eduardo Rossi University of Pavia Factorization of the density DGP: D t (x t χ t 1, d t ; Ψ) x t represent all the variables in the economy.

More information

Chapter 9. Derivatives. Josef Leydold Mathematical Methods WS 2018/19 9 Derivatives 1 / 51. f x. (x 0, f (x 0 ))

Chapter 9. Derivatives. Josef Leydold Mathematical Methods WS 2018/19 9 Derivatives 1 / 51. f x. (x 0, f (x 0 )) Chapter 9 Derivatives Josef Leydold Mathematical Methods WS 208/9 9 Derivatives / 5 Difference Quotient Let f : R R be some function. The the ratio f = f ( 0 + ) f ( 0 ) = f ( 0) 0 is called difference

More information

Density Estimation. Seungjin Choi

Density Estimation. Seungjin Choi Density Estimation Seungjin Choi Department of Computer Science and Engineering Pohang University of Science and Technology 77 Cheongam-ro, Nam-gu, Pohang 37673, Korea seungjin@postech.ac.kr http://mlg.postech.ac.kr/

More information

Likelihood Based Inference for Monotone Response Models

Likelihood Based Inference for Monotone Response Models Likelihood Based Inference for Monotone Response Models Moulinath Banerjee University of Michigan September 11, 2006 Abstract The behavior of maximum likelihood estimates (MLEs) and the likelihood ratio

More information

Efficient Estimation in Convex Single Index Models 1

Efficient Estimation in Convex Single Index Models 1 1/28 Efficient Estimation in Convex Single Index Models 1 Rohit Patra University of Florida http://arxiv.org/abs/1708.00145 1 Joint work with Arun K. Kuchibhotla (UPenn) and Bodhisattva Sen (Columbia)

More information

Recent progress in log-concave density estimation

Recent progress in log-concave density estimation Submitted to Statistical Science Recent progress in log-concave density estimation Richard J. Samworth University of Cambridge arxiv:1709.03154v1 [stat.me] 10 Sep 2017 Abstract. In recent years, log-concave

More information

Calculus One variable

Calculus One variable Calculus One variable (f ± g) ( 0 ) = f ( 0 ) ± g ( 0 ) (λf) ( 0 ) = λ f ( 0 ) ( (fg) ) ( 0 ) = f ( 0 )g( 0 ) + f( 0 )g ( 0 ) f g (0 ) = f ( 0 )g( 0 ) f( 0 )g ( 0 ) f( 0 ) 2 (f g) ( 0 ) = f (g( 0 )) g

More information

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction

More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order Restriction Sankhyā : The Indian Journal of Statistics 2007, Volume 69, Part 4, pp. 700-716 c 2007, Indian Statistical Institute More Powerful Tests for Homogeneity of Multivariate Normal Mean Vectors under an Order

More information

Hyperbolic Systems of Conservation Laws. in One Space Dimension. II - Solutions to the Cauchy problem. Alberto Bressan

Hyperbolic Systems of Conservation Laws. in One Space Dimension. II - Solutions to the Cauchy problem. Alberto Bressan Hyperbolic Systems of Conservation Laws in One Space Dimension II - Solutions to the Cauchy problem Alberto Bressan Department of Mathematics, Penn State University http://www.math.psu.edu/bressan/ 1 Global

More information

Introduction: exponential family, conjugacy, and sufficiency (9/2/13)

Introduction: exponential family, conjugacy, and sufficiency (9/2/13) STA56: Probabilistic machine learning Introduction: exponential family, conjugacy, and sufficiency 9/2/3 Lecturer: Barbara Engelhardt Scribes: Melissa Dalis, Abhinandan Nath, Abhishek Dubey, Xin Zhou Review

More information

Maximum likelihood: counterexamples, examples, and open problems. Jon A. Wellner. University of Washington. Maximum likelihood: p.

Maximum likelihood: counterexamples, examples, and open problems. Jon A. Wellner. University of Washington. Maximum likelihood: p. Maximum likelihood: counterexamples, examples, and open problems Jon A. Wellner University of Washington Maximum likelihood: p. 1/7 Talk at University of Idaho, Department of Mathematics, September 15,

More information

The Support Reduction Algorithm for Computing Non-Parametric Function Estimates in Mixture Models

The Support Reduction Algorithm for Computing Non-Parametric Function Estimates in Mixture Models Scandinavian Journal of Statistics, Vol. 35: 385 399, 2008 doi: 10.1111/j.1467-9469.2007.00588.x Published by Blackwell Publishing Ltd. The Support Reduction Algorithm for Computing Non-Parametric Function

More information

arxiv:math/ v2 [math.st] 17 Jun 2008

arxiv:math/ v2 [math.st] 17 Jun 2008 The Annals of Statistics 2008, Vol. 36, No. 3, 1031 1063 DOI: 10.1214/009053607000000974 c Institute of Mathematical Statistics, 2008 arxiv:math/0609020v2 [math.st] 17 Jun 2008 CURRENT STATUS DATA WITH

More information

Part I Analysis in Economics

Part I Analysis in Economics Part I Analysis in Economics D 1 1 (Function) A function f from a set A into a set B, denoted by f : A B, is a correspondence that assigns to each element A eactly one element y B We call y the image of

More information

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag.

1. Sets A set is any collection of elements. Examples: - the set of even numbers between zero and the set of colors on the national flag. San Francisco State University Math Review Notes Michael Bar Sets A set is any collection of elements Eamples: a A {,,4,6,8,} - the set of even numbers between zero and b B { red, white, bule} - the set

More information

Nonparametric estimation under Shape Restrictions

Nonparametric estimation under Shape Restrictions Nonparametric estimation under Shape Restrictions Jon A. Wellner University of Washington, Seattle Statistical Seminar, Frejus, France August 30 - September 3, 2010 Outline: Five Lectures on Shape Restrictions

More information

Differential Equations: Homework 2

Differential Equations: Homework 2 Differential Equations: Homework Alvin Lin January 08 - May 08 Section.3 Exercise The direction field for provided x 0. dx = 4x y is shown. Verify that the straight lines y = ±x are solution curves, y

More information

An Introduction to the Gamma Function

An Introduction to the Gamma Function UNIVERSITY OF WARWICK Second Year Essay An Introduction to the Gamma Function Tutor: Dr. Stefan Adams April 4, 29 Contents 1 Introduction 2 2 Conve Functions 3 3 The Gamma Function 7 4 The Bohr-Möllerup

More information

Likelihood Based Inference for Monotone Response Models

Likelihood Based Inference for Monotone Response Models Likelihood Based Inference for Monotone Response Models Moulinath Banerjee University of Michigan September 5, 25 Abstract The behavior of maximum likelihood estimates (MLE s) the likelihood ratio statistic

More information

Bayesian statistics: Inference and decision theory

Bayesian statistics: Inference and decision theory Bayesian statistics: Inference and decision theory Patric Müller und Francesco Antognini Seminar über Statistik FS 28 3.3.28 Contents 1 Introduction and basic definitions 2 2 Bayes Method 4 3 Two optimalities:

More information

Likelihood, MLE & EM for Gaussian Mixture Clustering. Nick Duffield Texas A&M University

Likelihood, MLE & EM for Gaussian Mixture Clustering. Nick Duffield Texas A&M University Likelihood, MLE & EM for Gaussian Mixture Clustering Nick Duffield Texas A&M University Probability vs. Likelihood Probability: predict unknown outcomes based on known parameters: P(x q) Likelihood: estimate

More information

Variational Bayes. A key quantity in Bayesian inference is the marginal likelihood of a set of data D given a model M

Variational Bayes. A key quantity in Bayesian inference is the marginal likelihood of a set of data D given a model M A key quantity in Bayesian inference is the marginal likelihood of a set of data D given a model M PD M = PD θ, MPθ Mdθ Lecture 14 : Variational Bayes where θ are the parameters of the model and Pθ M is

More information

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem

Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Lecture 3: Lagrangian duality and algorithms for the Lagrangian dual problem Michael Patriksson 0-0 The Relaxation Theorem 1 Problem: find f := infimum f(x), x subject to x S, (1a) (1b) where f : R n R

More information

Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52

Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52 Statistics for Applications Chapter 10: Generalized Linear Models (GLMs) 1/52 Linear model A linear model assumes Y X N(µ(X),σ 2 I), And IE(Y X) = µ(x) = X β, 2/52 Components of a linear model The two

More information

Evaluating the Performance of Estimators (Section 7.3)

Evaluating the Performance of Estimators (Section 7.3) Evaluating the Performance of Estimators (Section 7.3) Example: Suppose we observe X 1,..., X n iid N(θ, σ 2 0 ), with σ2 0 known, and wish to estimate θ. Two possible estimators are: ˆθ = X sample mean

More information

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA

PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA PENALIZED LIKELIHOOD PARAMETER ESTIMATION FOR ADDITIVE HAZARD MODELS WITH INTERVAL CENSORED DATA Kasun Rathnayake ; A/Prof Jun Ma Department of Statistics Faculty of Science and Engineering Macquarie University

More information

DIFFERENTIATION. 3.1 Approximate Value and Error (page 151)

DIFFERENTIATION. 3.1 Approximate Value and Error (page 151) CHAPTER APPLICATIONS OF DIFFERENTIATION.1 Approimate Value and Error (page 151) f '( lim 0 f ( f ( f ( f ( f '( or f ( f ( f '( f ( f ( f '( (.) f ( f '( (.) where f ( f ( f ( Eample.1 (page 15): Find

More information

Choosing among models

Choosing among models Eco 515 Fall 2014 Chris Sims Choosing among models September 18, 2014 c 2014 by Christopher A. Sims. This document is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 3.0 Unported

More information

Minimax estimators of the coverage probability of the impermissible error for a location family

Minimax estimators of the coverage probability of the impermissible error for a location family Minimax estimators of the coverage probability of the impermissible error for a location family by Miguel A. Arcones Binghamton University arcones@math.binghamton.edu Talk based on: Arcones, M. A. (2008).

More information

Rejection sampling - Acceptance probability. Review: How to sample from a multivariate normal in R. Review: Rejection sampling. Weighted resampling

Rejection sampling - Acceptance probability. Review: How to sample from a multivariate normal in R. Review: Rejection sampling. Weighted resampling Rejection sampling - Acceptance probability Review: How to sample from a multivariate normal in R Goal: Simulate from N d (µ,σ)? Note: For c to be small, g() must be similar to f(). The art of rejection

More information

asymptotic normality of nonparametric M-estimators with applications to hypothesis testing for panel count data

asymptotic normality of nonparametric M-estimators with applications to hypothesis testing for panel count data asymptotic normality of nonparametric M-estimators with applications to hypothesis testing for panel count data Xingqiu Zhao and Ying Zhang The Hong Kong Polytechnic University and Indiana University Abstract:

More information

Score Statistics for Current Status Data: Comparisons with Likelihood Ratio and Wald Statistics

Score Statistics for Current Status Data: Comparisons with Likelihood Ratio and Wald Statistics Score Statistics for Current Status Data: Comparisons with Likelihood Ratio and Wald Statistics Moulinath Banerjee 1 and Jon A. Wellner 2 1 Department of Statistics, Department of Statistics, 439, West

More information

Statistical Approaches to Learning and Discovery. Week 4: Decision Theory and Risk Minimization. February 3, 2003

Statistical Approaches to Learning and Discovery. Week 4: Decision Theory and Risk Minimization. February 3, 2003 Statistical Approaches to Learning and Discovery Week 4: Decision Theory and Risk Minimization February 3, 2003 Recall From Last Time Bayesian expected loss is ρ(π, a) = E π [L(θ, a)] = L(θ, a) df π (θ)

More information

CHAPTER 1-2: SHADOW PRICES

CHAPTER 1-2: SHADOW PRICES Essential Microeconomics -- CHAPTER -: SHADOW PRICES An intuitive approach: profit maimizing firm with a fied supply of an input Shadow prices 5 Concave maimization problem 7 Constraint qualifications

More information

OPTIMAL RATES OF CONVERGENCE FOR CONVEX SET ESTIMATION FROM SUPPORT FUNCTIONS. BY ADITYANAND GUNTUBOYINA University of Pennsylvania

OPTIMAL RATES OF CONVERGENCE FOR CONVEX SET ESTIMATION FROM SUPPORT FUNCTIONS. BY ADITYANAND GUNTUBOYINA University of Pennsylvania The Annals of Statistics 01, Vol. 40, No. 1, 385 411 DOI: 10.114/11-AOS959 Institute of Mathematical Statistics, 01 OPTIMAL RATES OF CONVERGENCE FOR CONVEX SET ESTIMATION FROM SUPPORT FUNCTIONS BY ADITYANAND

More information

Piecewise Smooth Solutions to the Burgers-Hilbert Equation

Piecewise Smooth Solutions to the Burgers-Hilbert Equation Piecewise Smooth Solutions to the Burgers-Hilbert Equation Alberto Bressan and Tianyou Zhang Department of Mathematics, Penn State University, University Park, Pa 68, USA e-mails: bressan@mathpsuedu, zhang

More information

A Law of the Iterated Logarithm. for Grenander s Estimator

A Law of the Iterated Logarithm. for Grenander s Estimator A Law of the Iterated Logarithm for Grenander s Estimator Jon A. Wellner University of Washington, Seattle Probability Seminar October 24, 2016 Based on joint work with: Lutz Dümbgen and Malcolm Wolff

More information

Optimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes

Optimization. Charles J. Geyer School of Statistics University of Minnesota. Stat 8054 Lecture Notes Optimization Charles J. Geyer School of Statistics University of Minnesota Stat 8054 Lecture Notes 1 One-Dimensional Optimization Look at a graph. Grid search. 2 One-Dimensional Zero Finding Zero finding

More information

1. Basic Model of Labor Supply

1. Basic Model of Labor Supply Static Labor Supply. Basic Model of Labor Supply.. Basic Model In this model, the economic unit is a family. Each faimily maximizes U (L, L 2,.., L m, C, C 2,.., C n ) s.t. V + w i ( L i ) p j C j, C j

More information

4.5.1 The use of 2 log Λ when θ is scalar

4.5.1 The use of 2 log Λ when θ is scalar 4.5. ASYMPTOTIC FORM OF THE G.L.R.T. 97 4.5.1 The use of 2 log Λ when θ is scalar Suppose we wish to test the hypothesis NH : θ = θ where θ is a given value against the alternative AH : θ θ on the basis

More information

Information in Data. Sufficiency, Ancillarity, Minimality, and Completeness

Information in Data. Sufficiency, Ancillarity, Minimality, and Completeness Information in Data Sufficiency, Ancillarity, Minimality, and Completeness Important properties of statistics that determine the usefulness of those statistics in statistical inference. These general properties

More information

The Scalar Conservation Law

The Scalar Conservation Law The Scalar Conservation Law t + f() = 0 = conserved qantity, f() =fl d dt Z b a (t, ) d = Z b a t (t, ) d = Z b a f (t, ) d = f (t, a) f (t, b) = [inflow at a] [otflow at b] f((a)) f((b)) a b Alberto Bressan

More information

Limits and the derivative function. Limits and the derivative function

Limits and the derivative function. Limits and the derivative function The Velocity Problem A particle is moving in a straight line. t is the time that has passed from the start of motion (which corresponds to t = 0) s(t) is the distance from the particle to the initial position

More information

Outline for today. Computation of the likelihood function for GLMMs. Likelihood for generalized linear mixed model

Outline for today. Computation of the likelihood function for GLMMs. Likelihood for generalized linear mixed model Outline for today Computation of the likelihood function for GLMMs asmus Waagepetersen Department of Mathematics Aalborg University Denmark Computation of likelihood function. Gaussian quadrature 2. Monte

More information

Spectral Theory of Orthogonal Polynomials

Spectral Theory of Orthogonal Polynomials Spectral Theory of Orthogonal Barry Simon IBM Professor of Mathematics and Theoretical Physics California Institute of Technology Pasadena, CA, U.S.A. Lecture 2: Szegö Theorem for OPUC for S Spectral Theory

More information

ST495: Survival Analysis: Hypothesis testing and confidence intervals

ST495: Survival Analysis: Hypothesis testing and confidence intervals ST495: Survival Analysis: Hypothesis testing and confidence intervals Eric B. Laber Department of Statistics, North Carolina State University April 3, 2014 I remember that one fateful day when Coach took

More information

Statistical Estimation

Statistical Estimation Statistical Estimation Use data and a model. The plug-in estimators are based on the simple principle of applying the defining functional to the ECDF. Other methods of estimation: minimize residuals from

More information

ST495: Survival Analysis: Maximum likelihood

ST495: Survival Analysis: Maximum likelihood ST495: Survival Analysis: Maximum likelihood Eric B. Laber Department of Statistics, North Carolina State University February 11, 2014 Everything is deception: seeking the minimum of illusion, keeping

More information

Concentration behavior of the penalized least squares estimator

Concentration behavior of the penalized least squares estimator Concentration behavior of the penalized least squares estimator Penalized least squares behavior arxiv:1511.08698v2 [math.st] 19 Oct 2016 Alan Muro and Sara van de Geer {muro,geer}@stat.math.ethz.ch Seminar

More information

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test.

LECTURE 10: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING. The last equality is provided so this can look like a more familiar parametric test. Economics 52 Econometrics Professor N.M. Kiefer LECTURE 1: NEYMAN-PEARSON LEMMA AND ASYMPTOTIC TESTING NEYMAN-PEARSON LEMMA: Lesson: Good tests are based on the likelihood ratio. The proof is easy in the

More information

Economics 205 Exercises

Economics 205 Exercises Economics 05 Eercises Prof. Watson, Fall 006 (Includes eaminations through Fall 003) Part 1: Basic Analysis 1. Using ε and δ, write in formal terms the meaning of lim a f() = c, where f : R R.. Write the

More information

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning

Lecture 3: More on regularization. Bayesian vs maximum likelihood learning Lecture 3: More on regularization. Bayesian vs maximum likelihood learning L2 and L1 regularization for linear estimators A Bayesian interpretation of regularization Bayesian vs maximum likelihood fitting

More information

Invariant HPD credible sets and MAP estimators

Invariant HPD credible sets and MAP estimators Bayesian Analysis (007), Number 4, pp. 681 69 Invariant HPD credible sets and MAP estimators Pierre Druilhet and Jean-Michel Marin Abstract. MAP estimators and HPD credible sets are often criticized in

More information

VARIABLE SELECTION AND INDEPENDENT COMPONENT

VARIABLE SELECTION AND INDEPENDENT COMPONENT VARIABLE SELECTION AND INDEPENDENT COMPONENT ANALYSIS, PLUS TWO ADVERTS Richard Samworth University of Cambridge Joint work with Rajen Shah and Ming Yuan My core research interests A broad range of methodological

More information

Measuring nuisance parameter effects in Bayesian inference

Measuring nuisance parameter effects in Bayesian inference Measuring nuisance parameter effects in Bayesian inference Alastair Young Imperial College London WHOA-PSI-2017 1 / 31 Acknowledgements: Tom DiCiccio, Cornell University; Daniel Garcia Rasines, Imperial

More information

Grade 12 (MCV4UE) AP Calculus Page 1 of 5 Derivative of a Function & Differentiability

Grade 12 (MCV4UE) AP Calculus Page 1 of 5 Derivative of a Function & Differentiability Grade 2 (MCV4UE) AP Calculus Page of 5 The Derivative at a Point f ( a h) f ( a) Recall, lim provides the slope of h0 h the tangent to the graph y f ( at the point, f ( a), and the instantaneous rate of

More information

Answer Key 1973 BC 1969 BC 24. A 14. A 24. C 25. A 26. C 27. C 28. D 29. C 30. D 31. C 13. C 12. D 12. E 3. A 32. B 27. E 34. C 14. D 25. B 26.

Answer Key 1973 BC 1969 BC 24. A 14. A 24. C 25. A 26. C 27. C 28. D 29. C 30. D 31. C 13. C 12. D 12. E 3. A 32. B 27. E 34. C 14. D 25. B 26. Answer Key 969 BC 97 BC. C. E. B. D 5. E 6. B 7. D 8. C 9. D. A. B. E. C. D 5. B 6. B 7. B 8. E 9. C. A. B. E. D. C 5. A 6. C 7. C 8. D 9. C. D. C. B. A. D 5. A 6. B 7. D 8. A 9. D. E. D. B. E. E 5. E.

More information

Differential Calculus

Differential Calculus Differential Calculus. Compute the derivatives of the following functions a() = 4 3 7 + 4 + 5 b() = 3 + + c() = 3 + d() = sin cos e() = sin f() = log g() = tan h() = 3 6e 5 4 i() = + tan 3 j() = e k()

More information

Advanced Quantitative Methods: maximum likelihood

Advanced Quantitative Methods: maximum likelihood Advanced Quantitative Methods: Maximum Likelihood University College Dublin 4 March 2014 1 2 3 4 5 6 Outline 1 2 3 4 5 6 of straight lines y = 1 2 x + 2 dy dx = 1 2 of curves y = x 2 4x + 5 of curves y

More information

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models

On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models On the Behavior of Marginal and Conditional Akaike Information Criteria in Linear Mixed Models Thomas Kneib Department of Mathematics Carl von Ossietzky University Oldenburg Sonja Greven Department of

More information

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017

Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Ph.D. Qualifying Exam Friday Saturday, January 6 7, 2017 Put your solution to each problem on a separate sheet of paper. Problem 1. (5106) Let X 1, X 2,, X n be a sequence of i.i.d. observations from a

More information

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications.

is a Borel subset of S Θ for each c R (Bertsekas and Shreve, 1978, Proposition 7.36) This always holds in practical applications. Stat 811 Lecture Notes The Wald Consistency Theorem Charles J. Geyer April 9, 01 1 Analyticity Assumptions Let { f θ : θ Θ } be a family of subprobability densities 1 with respect to a measure µ on a measurable

More information

Bayesian Methods. David S. Rosenberg. New York University. March 20, 2018

Bayesian Methods. David S. Rosenberg. New York University. March 20, 2018 Bayesian Methods David S. Rosenberg New York University March 20, 2018 David S. Rosenberg (New York University) DS-GA 1003 / CSCI-GA 2567 March 20, 2018 1 / 38 Contents 1 Classical Statistics 2 Bayesian

More information

Constrained Gaussian processes: methodology, theory and applications

Constrained Gaussian processes: methodology, theory and applications Constrained Gaussian processes: methodology, theory and applications Hassan Maatouk hassan.maatouk@univ-rennes2.fr Workshop on Gaussian Processes, November 6-7, 2017, St-Etienne (France) Hassan Maatouk

More information

STAT 512 sp 2018 Summary Sheet

STAT 512 sp 2018 Summary Sheet STAT 5 sp 08 Summary Sheet Karl B. Gregory Spring 08. Transformations of a random variable Let X be a rv with support X and let g be a function mapping X to Y with inverse mapping g (A = {x X : g(x A}

More information

Learning MN Parameters with Approximation. Sargur Srihari

Learning MN Parameters with Approximation. Sargur Srihari Learning MN Parameters with Approximation Sargur srihari@cedar.buffalo.edu 1 Topics Iterative exact learning of MN parameters Difficulty with exact methods Approximate methods Approximate Inference Belief

More information

Maximum likelihood: counterexamples, examples, and open problems

Maximum likelihood: counterexamples, examples, and open problems Maximum likelihood: counterexamples, examples, and open problems Jon A. Wellner University of Washington visiting Vrije Universiteit, Amsterdam Talk at BeNeLuxFra Mathematics Meeting 21 May, 2005 Email:

More information

Lecture 2: Conjugate priors

Lecture 2: Conjugate priors (Spring ʼ) Lecture : Conjugate priors Julia Hockenmaier juliahmr@illinois.edu Siebel Center http://www.cs.uiuc.edu/class/sp/cs98jhm The binomial distribution If p is the probability of heads, the probability

More information

Cubic Splines; Bézier Curves

Cubic Splines; Bézier Curves Cubic Splines; Bézier Curves 1 Cubic Splines piecewise approximation with cubic polynomials conditions on the coefficients of the splines 2 Bézier Curves computer-aided design and manufacturing MCS 471

More information

Bootstrap and Parametric Inference: Successes and Challenges

Bootstrap and Parametric Inference: Successes and Challenges Bootstrap and Parametric Inference: Successes and Challenges G. Alastair Young Department of Mathematics Imperial College London Newton Institute, January 2008 Overview Overview Review key aspects of frequentist

More information