Game-Theoretic Probability: Theory and Applications Glenn Shafer

Similar documents
What is risk? What is probability? Game-theoretic answers. Glenn Shafer. For 170 years: objective vs. subjective probability

What is risk? What is probability? Glenn Shafer

Lecture 4 How to Forecast

Defensive Forecasting: 4. Good probabilities mean less than we thought.

arxiv: v1 [math.pr] 26 Mar 2008

Hoeffding s inequality in game-theoretic probability

How to base probability theory on perfect-information games

Good randomized sequential probability forecasting is always possible

On a simple strategy weakly forcing the strong law of large numbers in the bounded forecasting game

Discussion of Dempster by Shafer. Dempster-Shafer is fiducial and so are you.

arxiv: v3 [math.pr] 5 Jun 2011

Defensive forecasting for linear protocols

Imprecise Probability

A survey of the theory of coherent lower previsions

Probability COMP 245 STATISTICS. Dr N A Heard. 1 Sample Spaces and Events Sample Spaces Events Combinations of Events...

1 Gambler s Ruin Problem

Announcements. Midterm Review Session. Extended TA session on Friday 9am to 12 noon.

Lecture 1: Review on Probability and Statistics

Defensive forecasting for optimal prediction with expert advice

EnM Probability and Random Processes

Lecture 4 - Random walk, ruin problems and random processes

Introduction to Imprecise Probability and Imprecise Statistical Methods

The Complexity of Forecast Testing

What is the probability of getting a heads when flipping a coin

Kousha Etessami. U. of Edinburgh, UK. Kousha Etessami (U. of Edinburgh, UK) Discrete Mathematics (Chapter 7) 1 / 13

Game-theoretic probability in continuous time

Basic Probabilistic Reasoning SEG

Learning Methods for Online Prediction Problems. Peter Bartlett Statistics and EECS UC Berkeley

Markov Chains (Part 3)

Simple Counter-terrorism Decision

Predictions as statements and decisions

Bayesian Inference. STA 121: Regression Analysis Artin Armagan

Intro to Probability. Andrei Barbu

Monopoly An Analysis using Markov Chains. Benjamin Bernard

Trajectorial Martingales, Null Sets, Convergence and Integration

Markov Chains on Countable State Space

Lecture 4 An Introduction to Stochastic Processes

Probability, Statistics, and Bayes Theorem Session 3

Introduction to Probability

EC476 Contracts and Organizations, Part III: Lecture 2

Kolmogorov s contributions to the foundations of probability

Mostly calibrated. Yossi Feinberg Nicolas S. Lambert

Grundlagen der Künstlichen Intelligenz

CS 188: Artificial Intelligence Spring Today

The topics in this section concern with the first course objective.

Lecture 2 Sep 5, 2017

A gentle introduction to imprecise probability models

Testing Exchangeability On-Line

Continuum Probability and Sets of Measure Zero

Merging of opinions in game-theoretic probability

Nonparametric Bayesian Methods (Gaussian Processes)

Quiz 1. Name: Instructions: Closed book, notes, and no electronic devices.

APM 421 Probability Theory Probability Notes. Jay Taylor Fall Jay Taylor (ASU) Fall / 62

Department of Agricultural Economics. PhD Qualifier Examination. May 2009

STOR Lecture 4. Axioms of Probability - II

Single Maths B: Introduction to Probability

Serena: I don t think that works because if n is 20 and you do 6 less than that you get 20 6 = 14. I think we should write! 6 > 4

Testable Forecasts. Luciano Pomatto. Caltech

Origins of Probability Theory

Answers to Spring 2014 Microeconomics Prelim

Structural Uncertainty in Health Economic Decision Models

INDEPENDENT NATURAL EXTENSION

The Island Problem Revisited

Theory Field Examination Game Theory (209A) Jan Question 1 (duopoly games with imperfect information)

11.4 Partial Sums of Arithmetic and Geometric Sequences

General Equilibrium and Welfare

LEM. New Results on Betting Strategies, Market Selection, and the Role of Luck. Giulio Bottazzi Daniele Giachini

DEVELOPING A TENNIS MODEL THAT REFLECTS OUTCOMES OF TENNIS MATCHES

Lecture 4. 1 Examples of Mechanism Design Problems

Consider an experiment that may have different outcomes. We are interested to know what is the probability of a particular set of outcomes.

Probability Theory Review

Whaddya know? Bayesian and Frequentist approaches to inverse problems

Motivation. Game Theory 24. Mechanism Design. Setting. Preference relations contain no information about by how much one candidate is preferred.

Chapter 16 focused on decision making in the face of uncertainty about one future

STOCHASTIC MODELS LECTURE 1 MARKOV CHAINS. Nan Chen MSc Program in Financial Engineering The Chinese University of Hong Kong (ShenZhen) Sept.

Topic 3: Introduction to Probability

From Cournot s Principle to Market Efficiency

FIRST ORDER SENTENCES ON G(n, p), ZERO-ONE LAWS, ALMOST SURE AND COMPLETE THEORIES ON SPARSE RANDOM GRAPHS

Self-calibrating Probability Forecasting

Lecture Slides. Elementary Statistics Eleventh Edition. by Mario F. Triola. and the Triola Statistics Series 4.1-1

MATH 56A: STOCHASTIC PROCESSES CHAPTER 1

Universal probability-free conformal prediction

Which property allows the addition of 5 to both sides in Step 1? A. Subtraction property of equality B. Reflexive property of equality

Probability and the Second Law of Thermodynamics

A birthday in St. Petersburg

Best Guaranteed Result Principle and Decision Making in Operations with Stochastic Factors and Uncertainty

Bayesian Games and Mechanism Design Definition of Bayes Equilibrium

Game Theory Fall 2003

The Game of Normal Numbers

Numerical Probabilistic Analysis under Aleatory and Epistemic Uncertainty

T 1. The value function v(x) is the expected net gain when using the optimal stopping time starting at state x:

1,1 1,2 1,3 1,4 1,5 1,6 2,1 2,2 2,3 2,4 2,5 2,6 3,1 3,2 3,3 3,4 3,5 3,6 4,1 4,2 4,3 4,4 4,5 4,6 5,1 5,2 5,3 5,4 5,5 5,6 6,1 6,2 6,3 6,4 6,5 6,6

Probability and Measure

6.041/6.431 Spring 2009 Quiz 1 Wednesday, March 11, 7:30-9:30 PM. SOLUTIONS

Probability, Random Processes and Inference

Lecture Note II-3 Static Games of Incomplete Information. Games of incomplete information. Cournot Competition under Asymmetric Information (cont )

Inference for Stochastic Processes

Learning Theory. Sridhar Mahadevan. University of Massachusetts. p. 1/38

Weighted Majority and the Online Learning Approach

Question Points Score Total: 137

Transcription:

ISIPTA 07 July 17, 2007, Prague Game-Theoretic Probability: Theory and Applications Glenn Shafer Theory To prove a probabilistic prediction, construct a betting strategy that makes you rich if the prediction fails. Application: Testing To test a probabilistic theory, bet against its predictions. Application: Forecasting To make probability forecasts, construct a forecasting strategy that defeats all reasonable betting strategies. Application: New understanding of epistemic probability In repetitive structure, you can use defensive forecasting to get additive probabilities. Judgements of evidential independence are needed to extend reasoning to non-repetitive situations (Dempster-Shafer, etc.). 1

References Probability and Finance: It s Only a Game! Glenn Shafer and Vladimir Vovk, Wiley, 2001. www.probabilityandfinance.com: Chapters from book, reviews, many working papers. Most important working papers at www.probabilityandfinance.com #4 Abridged version in Statistical Science, 21 70 98, 2006. #8 Abridged version in Proceedings of the Tenth International Workshop on Artificial Intelligence and Statistics, http://www.gatsby. ucl.ac.uk/aistats/ 2

Volodya Vovk atop the World Trade Center in 1998. Born 1960. Student of Kolmogorov. Born in Ukraine, educated in Moscow, teaches in London. Volodya is a nickname for the Ukrainian Volodimir and the Russian Vladimir. 3

Part I. Game-theoretic probability Blaise Pascal (1623 1662) Probability is about betting. Antoine Cournot (1801 1877) Events of small probability do not happen. Jean Ville (1910 1988) Pascal + Cournot: If the probabilities are right, you don t get rich. 4

A physically impossible event is one whose probability is infinitely small. This remark alone gives substance an objective and phenomenological value to the mathematical theory of probability. (1843) Antoine Cournot (1801 1877) This is more basic than frequentism. 5

Borel was emphatic: the principle that an event with very small probability will not happen is the only law of chance. Émile Borel 1871 1956 Inventor of measure theory. Impossibility on the human scale: p < 10 6. Impossibility on the terrestrial scale: p < 10 15. Impossibility on the cosmic scale: p < 10 50. Minister of the French navy in 1925. 6

Kol- In his celebrated 1933 book, mogorov wrote: Andrei Kolmogorov 1903 1987 Hailed as the Soviet Euler, Kolmogorov was credited with establishing measure theory as the mathematical foundation for probability. When P(A) very small, we can be practically certain that the event A will not happen on a single trial of the conditions that define it. 7

In 1939, Ville showed that the laws of probability can be derived from this principle: You will not multiply the capital you risk by a large factor. Jean Ville, 1910 1988, on entering the École Normale Supérieure. Ville showed that this principle is equivalent to the principle that events of small probability will not happen. We call both principles Cournot s principle. 8

Suppose you gamble without risking more than your initial capital. Your resulting wealth is a nonnegative random variable X with expected value E(X) equal to your initial capital. Markov s inequality says P ( X E(X) ɛ ) ɛ. You have probability ɛ or less of multiplying your initial capital by 1/ɛ or more. Ville proved what is now called Doob s inequality, which generalizes Markov s inequality to a sequence of bets. 9

The Ville/Vovk perfect-information protocol for probability K 0 = 1. FOR n = 1, 2,..., N: Forecaster announces prices for various payoffs. Skeptic decides which payoffs to buy. Reality determines the payoffs. K n := K n 1 + Skeptic s net gain or loss. Ville showed that any test of Forecaster can be expressed as a betting strategy for Skeptic. Vovk, Takemura, and I showed that Forecaster can beat Skeptic. 10

Part II. Ville/Vovk game-theoretic testing In Ville s theory, Forecaster is a known probability distribution for Reality s move. It always gives conditional probabilities for Reality s next move given her past moves. In Vovk s generalization, (1) Forecaster does not necessarily use a known probability distribution, and (2) he may give less than a probability distribution for Reality s next move. For both reasons, we get upper and lower probabilities instead of probabilities. 11

Ville s strong law of large numbers. (Special case where probability is always 1/2.) K 0 = 1. FOR n = 1, 2,... : Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n 1 2 ). Skeptic wins if (1) K n is never negative and (2) either lim n 1 n ni=1 y i = 1 2 or lim n K n =. Theorem Skeptic has a winning strategy. 12

Ville s strategy K 0 = 1. FOR n = 1, 2,... : Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n 1 2 ). Ville suggested the strategy s n (y 1,..., y n 1 = 4 n + 1 K n 1 It produces the capital ( r n 1 n 1 2 ), where r n 1 := K n = 2 nr n!(n r n )!. (n + 1)! From the assumption that this remains bounded by some constant C, you can easily derive the strong law of large numbers using Stirling s formula. n 1 i=1 y i. 13

Vovk s weak law of large numbers K 0 := 1. FOR n = 1,..., N: Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n ( yn 1 2). Winning: Skeptic wins if K n is never negative and either K N C or 1 N Nn=1 y n 1 2 < ɛ. Theorem. Skeptic has a winning strategy if N C/4ɛ 2. 14

Ville s more general game. Ville started with a probability distribution for P for y 1, y 2,.... The conditional probability for y n = 1 given y 1,..., y n 1 is not necessarily 1/2. K 0 := 1. FOR n = 1, 2,... : Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n ( yn P(y n = 1 y 1,..., y n 1 ) ). Skeptic wins if (1) K n is never negative and (2) either lim n 1 n ni=1 ( yi P(y i = 1 y 1,..., y i 1 ) ) = 0 or lim n K n =. Theorem Skeptic has a winning strategy. 15

Vovk s generalization: Replace P with a forecaster. K 0 := 1. FOR n = 1, 2,... : Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n p n ). Skeptic wins if (1) K n is never negative and (2) either lim n 1 n ni=1 (y i p i ) = 0 or lim n K n =. Theorem Skeptic has a winning strategy. 16

Vovk s weak law of large numbers K 0 := 1. FOR n = 1,..., N: Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n p n ). Winning: Skeptic wins if K n is never negative and either K N C or N 1 Nn=1 (y n p n ) < ɛ. Theorem. Skeptic has a winning strategy if N C/4ɛ 2. 17

Put it in terms of upper probability K 0 := 1. FOR n = 1,..., N: Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n p n ). Theorem. P { 1 N N n=1 (y n p n ) ɛ } 1 4Nɛ 2. 18

Definition of upper price and upper probability K 0 := α. FOR n = 1,..., N: Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. K n := K n 1 + s n (y n p n ). For any real-valued function X on ([0, 1] {0, 1}) N, E X := inf{α Skeptic has a strategy guaranteeing K N X(p 1, y 1,..., p N, y N )} For any subset A ([0, 1] {0, 1}) N, P A := inf{α Skeptic has a strategy guaranteeing K N 1 if A happens and K N 0 otherwise}. E X = E( X) P A = 1 P A 19

Part III. Defensive forecasting Under repetition, good probability forecasting is possible. We call it defensive because it defends against a quasi-universal test. Your probability forecasts will pass this test even if reality plays against you. 20

Why Phil Dawid thought good probability prediction is impossible... FOR n = 1, 2,... Forecaster announces p n [0, 1]. Skeptic announces continuous s n R. Reality announces y n {0, 1}. Skeptic s profit := s n (y n p n ). Reality can make Forecaster uncalibrated by setting y n := { 1 if pn < 0.5 0 if p n 0.5, Skeptic can then make steady money with s n := { 1 if p < 0.5 1 if p 0.5, But if Skeptic is forced to approximate s n by a continuous function of p n, then the continuous function will be zero close to p = 0.5, and Forecaster can set p n equal to this point. 21

Skeptic adopts a continuous strategy S. FOR n = 1, 2,... Reality announces x n X. Forecaster announces p n [0, 1]. Skeptic makes the move s n specified by S. Reality announces y n {0, 1}. Skeptic s profit := s n (y n p n ). Theorem Forecaster can guarantee that Skeptic never makes money. We actually prove a stronger theorem. Instead of making Skeptic announce his entire strategy in advance, only make him reveal his strategy for each round in advance of Forecaster s move. FOR n = 1, 2,... Reality announces x n X. Skeptic announces continuous S n : [0, 1] R. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. Skeptic s profit := S n (p n )(y n p n ). Theorem. Forecaster can guarantee that Skeptic never makes money. 22

FOR n = 1, 2,... Reality announces x n X. Skeptic announces continuous S n : [0, 1] R. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. Skeptic s profit := S n (p n )(y n p n ). Theorem Forecaster can guarantee that Skeptic never makes money. Proof: If S n (p) > 0 for all p, take p n := 1. If S n (p) < 0 for all p, take p n := 0. Otherwise, choose p n so that S n (p n ) = 0. 23

TWO APPROACHES TO FORECASTING FOR n = 1, 2,... Forecaster announces p n [0, 1]. Skeptic announces s n R. Reality announces y n {0, 1}. 1. Start with strategies for Forecaster. Improve by averaging (Bayes, prediction with expert advice). 2. Start with strategies for Skeptic. Improve by averaging (defensive forecasting). 24

We can always give probabilities with good calibration and resolution. FOR n = 1, 2,... Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. There exists a strategy for Forecaster that gives p n with good calibration and resolution. 25

FOR n = 1, 2,... Reality announces x n X. Forecaster announces p n [0, 1]. Reality announces y n {0, 1}. 1. Fix p [0, 1]. Look at n for which p n p. If the frequency of y n = 1 always approximates p, Forecaster is properly calibrated. 2. Fix x X and p [0, 1]. Look at n for which x n x and p n p. If the frequency of y n = 1 always approximates p, Forecaster is properly calibrated and has good resolution. 26

Fundamental idea: Average strategies for Skeptic for a grid of values of p. (The p -strategy makes money if calibration fails for p n close to p.) The derived strategy for Forecaster guarantees good calibration everywhere. Example of a resulting strategy for Skeptic: S n (p) := n 1 i=1 e C(p p i) 2 (y i p i ) Any kernel K(p, p i ) can be used in place of e C(p p i) 2. 27

Skeptic s strategy: S n (p) := n 1 i=1 e C(p p i) 2 (y i p i ) Forecaster s strategy: Choose p n so that n 1 i=1 e C(p n p i ) 2 (y i p i ) = 0. The main contribution to the sum comes from i for which p i is close to p n. So Forecaster chooses p n in the region where the y i p i average close to zero. On each round, choose as p n the probability value where calibration is the best so far. 28

Part IV. Aleatory (objective) vs. epistemic (subjective) From a 1970s perspective: Aleatory probability is the irreducible uncertainty that remains when knowledge is complete. Epistemic probability arises when knowledge is incomplete. New game-theoretic perspective: Under a repetitive structure you can make make good probability forecasts relative to whatever state of knowledge you have. If there is no repetitive structure, your task is to combine evidence rather than to make probability forecasts. 29

Cournotian understanding of Dempster-Shafer Fundamental idea: transferring belief Conditioning Independence Dempster s rule 30

Fundamental idea: transferring belief Variable ω with set of possible values Ω. Random variable X with set of possible values X. We learn a mapping Γ : X 2 Ω with this meaning: If X = x, then ω Γ(x). For A Ω, our belief that ω A is now B(A) = P{x Γ(x) A}. Cournotian judgement of independence: Learning the relationship between X and ω does not affect our inability to beat the probabilities for X. 31

Example: The sometimes reliable witness Joe is reliable with probability 30%. When he is reliable, what he says is true. Otherwise, it may or may not be true. X = {reliable, not reliable} P(reliable) = 0.3 P(not reliable) = 0.7 Did Glenn pay his dues for coffee? Ω = {paid, not paid} Joe says Glenn paid. Γ(reliable) = {paid} Γ(not reliable) = {paid, not paid} New beliefs: B(paid) = 0.3 B(not paid) = 0 Cournotian judgement of independence: Hearing what Joe said does not affect our inability to beat the probabilities concerning his reliability. 32

Art Dempster (born 1929) with his Meng & Shafer hatbox. Retirement dinner at Harvard, May 2005. See http://www.stat.purdue.edu/ chuanhai/projects/ds/ for Art s D-S papers. 33