Two examples of the use of fuzzy set theory in statistics. Glen Meeden University of Minnesota.

Similar documents
Charles Geyer University of Minnesota. joint work with. Glen Meeden University of Minnesota.

Fuzzy and Randomized Confidence Intervals and P-values

Fuzzy and Randomized Confidence Intervals and P-values

Fuzzy and Randomized Confidence Intervals and P -values

Lawrence D. Brown, T. Tony Cai and Anirban DasGupta

Stat 5421 Lecture Notes Fuzzy P-Values and Confidence Intervals Charles J. Geyer March 12, Discreteness versus Hypothesis Tests

Hypotheses testing as a fuzzy set estimation problem

Bayesian Inference: Posterior Intervals

Chapter 4 HOMEWORK ASSIGNMENTS. 4.1 Homework #1

Part III. A Decision-Theoretic Approach and Bayesian testing

Data Mining Chapter 4: Data Analysis and Uncertainty Fall 2011 Ming Li Department of Computer Science and Technology Nanjing University

Interval Estimation. Chapter 9

Design of the Fuzzy Rank Tests Package

Chapter 9: Interval Estimation and Confidence Sets Lecture 16: Confidence sets and credible sets

STA 732: Inference. Notes 2. Neyman-Pearsonian Classical Hypothesis Testing B&D 4

STAT 830 Hypothesis Testing

Stat 535 C - Statistical Computing & Monte Carlo Methods. Arnaud Doucet.

5.1 Uniformly Most Accurate Families of Confidence

Definition 3.1 A statistical hypothesis is a statement about the unknown values of the parameters of the population distribution.

Bios 6649: Clinical Trials - Statistical Design and Monitoring

Statistical Inference

STA 732: Inference. Notes 10. Parameter Estimation from a Decision Theoretic Angle. Other resources

Introduction to Bayesian Statistics 1

Uniformly Most Powerful Bayesian Tests and Standards for Statistical Evidence

Confidence Distribution

STAT 830 Hypothesis Testing

Statistics for Particle Physics. Kyle Cranmer. New York University. Kyle Cranmer (NYU) CERN Academic Training, Feb 2-5, 2009

Statistical Inference: Maximum Likelihood and Bayesian Approaches

STAT 499/962 Topics in Statistics Bayesian Inference and Decision Theory Jan 2018, Handout 01

Lecture 2: Statistical Decision Theory (Part I)

Statistical Inference

Bayesian Statistics. Debdeep Pati Florida State University. February 11, 2016

Classical and Bayesian inference

Parameter estimation and forecasting. Cristiano Porciani AIfA, Uni-Bonn

STATS 200: Introduction to Statistical Inference. Lecture 29: Course review

The Bayesian Choice. Christian P. Robert. From Decision-Theoretic Foundations to Computational Implementation. Second Edition.

Notes on Decision Theory and Prediction

A Bayesian solution for a statistical auditing problem

Hypothesis Testing. Part I. James J. Heckman University of Chicago. Econ 312 This draft, April 20, 2006

Bayesian philosophy Bayesian computation Bayesian software. Bayesian Statistics. Petter Mostad. Chalmers. April 6, 2017

Fall 2017 STAT 532 Homework Peter Hoff. 1. Let P be a probability measure on a collection of sets A.

A Very Brief Summary of Bayesian Inference, and Examples

Module 22: Bayesian Methods Lecture 9 A: Default prior selection

Lecture notes on statistical decision theory Econ 2110, fall 2013

On the GLR and UMP tests in the family with support dependent on the parameter

Bayesian performance

Sequential Monitoring of Clinical Trials Session 4 - Bayesian Evaluation of Group Sequential Designs

ST 740: Model Selection

Chapter 9: Hypothesis Testing Sections

Statistics of Small Signals

Introduction to Bayesian Methods

Approximate Bayesian computation for spatial extremes via open-faced sandwich adjustment

Final Exam. 1. (6 points) True/False. Please read the statements carefully, as no partial credit will be given.

University of Texas, MD Anderson Cancer Center

A Very Brief Summary of Statistical Inference, and Examples

Reasoning with Uncertainty

Bayesian Econometrics

Bayesian Learning (II)

On the Bayesianity of Pereira-Stern tests

Statistics: Learning models from data

A NONINFORMATIVE BAYESIAN APPROACH FOR TWO-STAGE CLUSTER SAMPLING

Statistical Inference: Uses, Abuses, and Misconceptions

Frequentist Statistics and Hypothesis Testing Spring

Bayesian Regression (1/31/13)

HPD Intervals / Regions

LECTURE 5 NOTES. n t. t Γ(a)Γ(b) pt+a 1 (1 p) n t+b 1. The marginal density of t is. Γ(t + a)γ(n t + b) Γ(n + a + b)

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

Uniformly and Restricted Most Powerful Bayesian Tests

(1) Introduction to Bayesian statistics

Bayesian learning Probably Approximately Correct Learning

Introduction to Probabilistic Machine Learning

Chapter 4. Bayesian inference. 4.1 Estimation. Point estimates. Interval estimates

Basic Concepts of Inference

Review: Statistical Model

Introductory Econometrics

DS-GA 1002 Lecture notes 11 Fall Bayesian statistics

Introduction: MLE, MAP, Bayesian reasoning (28/8/13)

Statistical Inference

The Purpose of Hypothesis Testing

Parameter Estimation. William H. Jefferys University of Texas at Austin Parameter Estimation 7/26/05 1

Introduction to Imprecise Probability and Imprecise Statistical Methods

EXAMPLE: INFERENCE FOR POISSON SAMPLING

Journeys of an Accidental Statistician

Markov Chain Monte Carlo methods

INTRODUCTION TO BAYESIAN METHODS II

Physics 403. Segev BenZvi. Credible Intervals, Confidence Intervals, and Limits. Department of Physics and Astronomy University of Rochester

Hypothesis Testing - Frequentist

Uncertain Inference and Artificial Intelligence

STAT 425: Introduction to Bayesian Analysis

Beta statistics. Keywords. Bayes theorem. Bayes rule

Math 141. Lecture 10: Confidence Intervals. Albyn Jones 1. jones/courses/ Library 304. Albyn Jones Math 141

PARAMETER ESTIMATION: BAYESIAN APPROACH. These notes summarize the lectures on Bayesian parameter estimation.

Bayesian inference. Fredrik Ronquist and Peter Beerli. October 3, 2007

Fundamental Probability and Statistics

Machine Learning 4771

Chapter 5. Bayesian Statistics

Ridge regression. Patrick Breheny. February 8. Penalized regression Ridge regression Bayesian interpretation

Estimation of reliability parameters from Experimental data (Parte 2) Prof. Enrico Zio

Bayesian Inference. Chapter 1. Introduction and basic concepts

Statistical Decision Theory and Bayesian Analysis Chapter1 Basic Concepts. Outline. Introduction. Introduction(cont.) Basic Elements (cont.

Transcription:

Two examples of the use of fuzzy set theory in statistics Glen Meeden University of Minnesota http://www.stat.umn.edu/~glen/talks 1

Fuzzy set theory Fuzzy set theory was introduced by Zadeh in (1965) as another approach to represent uncertainty. A fuzzy set A is characterized by its membership function. This is a function whose range is contained in the unit interval. At a point the value of this function represents the degree of membership of the point in the set A. 2

What is the average yearly snowfall in the Twin Cities? You might answer somewhere between 20 and 50 inches. A better answer could be the following fuzzy membership function. mf 0.0 0.2 0.4 0.6 0.8 1.0 20 25 30 35 40 45 50 55 inches 3

Fuzzy Sets Indicator function I A of ordinary set A I A (u) 1.0 0.0 x Membership function I B of fuzzy set B I B (u) 1.0 0.0 x 4

Interpreting a fuzzy membership function The value I B (u) is the degree of membership of the point u in the fuzzy set B. Ordinary sets are special case of fuzzy sets called crisp sets. A non-probabilistic measure of uncertainty. Think partial credit! Geyer and Meeden (Statist. Sci., 2005) 5

Buying a used car Consider the set of cars for sale in your area. Let A be the fuzzy set of cars that you could consider owning. For each car you can assign it a value between 0 and 1 which would represent the degree of membership of this particular car in the fuzzy set A. For a given car this depends on its age, condition, style, price and so forth. Here the fuzzy membership function measures the overall attractiveness of a car to you. This value cannot be interpreted as a probability! Can fuzzy set theory be used in statistical inference? 6

Confidence intervals Let X 1,..., X n be iid normal(µ, σ 2 ). Let X = n i=1 X i /n and S 2 = n i=1 (X i X) 2 /(n 1) then ( X t n,0.025 S/ n, X + t n,0.025 S/ n is a 95% confidence interval for µ. ) As a function of µ it has constant coverage probability. 7

Good CI s come from good tests In the normal case, given X 1 = x 1,..., X n = x n, the CI is just the set of µ o s for which the data accepts as true the null hypothesis H : µ = µ o. Optimal properties for CI s comes from the optimal properties of the related tests of hypotheses. 8

Ordinary Confidence Intervals OK for continuous data, but a really bad idea for discrete data. Why? Let X be binomial(n, p) Coverage Probability for a CI (l(x), u(x)) is γ(p) = pr p {l(x) < p < u(x)} = n x=0 I (l(x),u(x)) (p) f p (x) As p moves across the boundary of a possible confidence interval (l(x), u(x)), the coverage probability jumps by f p (x). Ideally, γ is a constant function equal to the nominal confidence coefficient. But that s not possible. 9

Performance of ˆp ± 1.96 ˆp(1 ˆp) n Nominal 95% Wald Interval, n = 30 coverage probability 0.0 0.2 0.4 0.6 0.8 0.0 0.2 0.4 0.6 0.8 1.0 p 10

Randomized Tests Randomized tests defined by a critical function φ(x, α, θ). depend on observed data x, the significance level α and the null hypothesis H 0 : θ = θ 0 Decision is randomized: reject H 0 with probability φ(x, α, θ 0 ). Since probabilities are between zero and one, so is φ(x, α, θ). Classical uniformly most powerful (UMP) and UMP unbiased (UMPU) tests are randomized when data are discrete and will generate non-crisp CI s. 11

Binomial Example 1 φ(x, α, θ) 0.0 0.2 0.4 0.6 0.8 1.0 0.0 0.2 0.4 0.6 0.8 1.0 θ Sample size n = 10, fuzzy confidence interval associated with UMPU test, confidence level 1 α = 0.95. Data x = 0 (solid curve), x = 4 (dashed curve) and x = 9 (dotted curve) 12

This is good If we allow non-crisp fuzzy membership functions to be CI s then our CI s will have exact 95% coverage frequency. Thinking of CI s as fuzzy membership functions discourages users from believing that a 95% CI contains the unknown parameter with probability equal to 0.95. 13

The Bayesian approach X, θ, f(x θ) and π(θ) the prior distribution. Inference is based on the posterior π(θ x) = f(x θ)π(θ)/ f(x θ)π(θ) dθ Bayesian credible intervals have the same problem as CI s for discrete data. Subjective Bayesians will not care since frequentist properties of a procedure are not important to them. 14

Fuzzy and Bayes Typically Bayesians are not interested in the fuzzy notion of uncertainty. An exception are those interested in imprecise probability. See Gert de Cooman (Fuzzy sets and systems, 2005) Didier Dubois has also written about the relationship between a fuzzy membership function and families of priors. How can a Bayesian change their prior or posterior into a fuzzy set membership function which gives a good representation of their uncertainty and can be interpreted like a CI. 15

A Bayesian decision problem θ Θ an interval of real numbers. π(θ) a continuous prior density A A, the class of measurable membership functions on Θ. The loss function depends on four known parameters specified by the statistician: a 1 0, a 2 0, b 1 0 and b 2 0 where at least one of the a i s and one of the b i s must be > 0. L(A, θ) = a 1 {1 I A (θ)}+ a 2 2 {1 I A(θ)} 2 + Θ { b 1 I A (θ)+ b 2 2 (I A(θ)) 2 } dθ 16

The solution The fuzzy set membership A which satisfies is given by I A (θ) = Θ L(A, θ)π(θ) dθ = inf A A Θ L(A, θ)π(θ) dθ 0 for 0 π(θ) < b 1 /(a 1 + a 2 ) (a 1 +a 2 )π(θ) b 1 a 2 π(θ)+b 2 for b 1 /(a 1 + a 2 ) π(θ) (b 1 + b 2 )/a 1 1 for π(θ) > (b 1 + b 2 )/a 1 17

0.0 0.5 1.0 1.5 2.0 2.5 3.0 0.0 0.2 0.4 0.6 0.8 1.0 A fuzzy membership function (solid line) and for 4 different loss functions the 4 priors whose solution is the fmf. 18

Final comments We have seen how to convert a prior or posterior density into a fuzzy membership function. Formally this increases a Bayesian s uncertainty but in some cases it could be a good thing. Why? For a Bayesian the posterior distribution summarizes their information about the parameter given the data. Bayesian credible intervals are a way to summarize the information in the posterior. These intervals have little formal justification in the Bayesian paradigm and seem to be popular because they mimic frequentist CI s. For discrete problems it may more useful to convert the posterior to a fuzzy membership function. Perhaps frequentists and Bayesians should be more interested in fuzzy membership functions as a way to represent uncertainty about a parameter value. 19