Experimental Design to Persuade

Similar documents
Australian School of Business Working Paper

A Rothschild-Stiglitz approach to Bayesian persuasion

Definitions and Proofs

A Rothschild-Stiglitz approach to Bayesian persuasion

A Rothschild-Stiglitz approach to Bayesian persuasion

Bayesian Persuasion Online Appendix

Disclosure of Endogenous Information

Optimal information disclosure: A linear programming approach

Persuading Skeptics and Reaffirming Believers

Optimal Signals in Bayesian Persuasion Mechanisms with Ranking

9 A Class of Dynamic Games of Incomplete Information:

Some Notes on Costless Signaling Games

Pre-Communication in a Coordination Game with Incomplete Information

An Example of Conflicts of Interest as Pandering Disincentives

Static Information Design

Reputational Concern with Endogenous Information Acquisition. Haibo Xu Washington University in Saint Louis. December 28, 2011

The Intuitive and Divinity Criterion:

Persuasion Under Costly Lying

Static Information Design

Endogenous Persuasion with Rational Verification

When to Ask for an Update: Timing in Strategic Communication. National University of Singapore June 5, 2018

Bayes Correlated Equilibrium and Comparing Information Structures

Introduction. Introduction

Hierarchical Bayesian Persuasion

Volume 30, Issue 3. Monotone comparative statics with separable objective functions. Christian Ewerhart University of Zurich

Not Only What But also When A Theory of Dynamic Voluntary Disclosure

Costly Expertise. Dino Gerardi and Leeat Yariv yz. Current Version: December, 2007

Graduate Microeconomics II Lecture 5: Cheap Talk. Patrick Legros

Game Theory and Economics of Contracts Lecture 5 Static Single-agent Moral Hazard Model

Reputation and Bounded Memory in Repeated Games with Incomplete Information

Persuading a Pessimist

Discussion of "Persuasion in Global Games with an Application to Stress Testing" by Nicolas Inostroza and Alessandro Pavan

BAYES CORRELATED EQUILIBRIUM AND THE COMPARISON OF INFORMATION STRUCTURES IN GAMES. Dirk Bergemann and Stephen Morris

Cowles Foundation for Research in Economics at Yale University

The Kuhn-Tucker Problem

Introduction. 1 University of Pennsylvania, Wharton Finance Department, Steinberg Hall-Dietrich Hall, 3620

Economics Bulletin, 2012, Vol. 32 No. 1 pp Introduction. 2. The preliminaries

Lecture Notes on Game Theory

Games and Economic Behavior

Coordination and Cheap Talk in a Battle of the Sexes with Private Information

Continuity in Mechanism Design without Transfers 1

Informed Principal in Private-Value Environments

Labor Economics, Lecture 11: Partial Equilibrium Sequential Search

EconS Microeconomic Theory II Midterm Exam #2 - Answer Key

Correlated Equilibrium in Games with Incomplete Information

Persuasion of Heterogenous Audiences and Optimal Media Censorship

Notes on Mechanism Designy

Sender s Small Concern for Credibility and Receiver s Dilemma

Discussion Paper #1541

Deceptive Advertising with Rational Buyers

Competition relative to Incentive Functions in Common Agency

Optimal Information Disclosure

When to Ask for an Update: Timing in Strategic Communication

BAYES CORRELATED EQUILIBRIUM AND THE COMPARISON OF INFORMATION STRUCTURES IN GAMES. Dirk Bergemann and Stephen Morris

Selling Experiments. Dirk Bergemann 1 Alessandro Bonatti 2 Alex Smolin 1. BFI Chicago July 10th, Yale University.

Ex Post Cheap Talk : Value of Information and Value of Signals

Robust Mechanism Design and Robust Implementation

A Nonspeculation Theorem with an Application to Committee Design

Mechanism Design: Basic Concepts

Epsilon Ex Post Implementation

Renegotiation proof mechanism design with imperfect type veri cation

Dynamic Learning and Strategic Communication

Benefits from non-competing persuaders

Online Appendices for Large Matching Markets: Risk, Unraveling, and Conflation

Principal-Agent Games - Equilibria under Asymmetric Information -

Limit pricing models and PBE 1

Dynamic Mechanism Design:

Upstream capacity constraint and the preservation of monopoly power in private bilateral contracting

Using Cheap Talk to Polarize or Unify a Group of Decision Makers

Experimentation, Patents, and Innovation

Competition in Persuasion

Microeconomic Theory (501b) Problem Set 10. Auctions and Moral Hazard Suggested Solution: Tibor Heumann

A Nonspeculation Theorem with an Application to Committee Design

Minimum Wages and Excessive E ort Supply

Blackwell s Informativeness Theorem Using Diagrams

Economics of Open Source: A Dynamic Approach

Learning and Risk Aversion

Delayed Persuasion. First version: December 2015 This version: August Abstract

On the Unique D1 Equilibrium in the Stackelberg Model with Asymmetric Information Janssen, M.C.W.; Maasland, E.

Subgame Perfect Implementation With Almost Perfect. Information

Informational Control and Organizational Design

Competition in Persuasion

Information Design. Dirk Bergemann and Stephen Morris. Johns Hopkins University April 2017

Almost Transferable Utility, Changes in Production Possibilities, and the Nash Bargaining and the Kalai-Smorodinsky Solutions

1 Web Appendix: Equilibrium outcome under collusion (multiple types-multiple contracts)

Some Notes on Adverse Selection

Costly Persuasion. Matthew Gentzkow and Emir Kamenica University of Chicago. December 2013

Sending Information to Interactive Receivers Playing a Generalized Prisoners Dilemma

Implementability, Walrasian Equilibria, and Efficient Matchings

Northwestern University

On the Informed Principal Model with Common Values

A Principal-Agent Model of Sequential Testing

Alvaro Rodrigues-Neto Research School of Economics, Australian National University. ANU Working Papers in Economics and Econometrics # 587

CMS-EMS Center for Mathematical Studies in Economics And Management Science. Discussion Paper #1583. Monotone Persuasion

Internet Appendix for The Labor Market for Directors and Externalities in Corporate Governance

Arm s Length Relationships without Moral Hazard

Decentralized Trading with Private Information

Simultaneous Choice Models: The Sandwich Approach to Nonparametric Analysis

Government 2005: Formal Political Theory I

INFORMATION DESIGN: A UNIFIED PERSPECTIVE. Dirk Bergemann and Stephen Morris. February 2017 Revised March 2017

Transcription:

Experimental Design to Persuade Anton Kolotilin y This Version: September, 2012 Comments Welcome. Abstract A sender chooses ex ante how her information will be disclosed ex post. A receiver obtains public information and information disclosed by the sender. Then he takes one of two actions. The sender wishes to maximize the probability that the receiver takes the desired action. I show that the sender optimally discloses only whether the receiver s utility is above a cuto, which makes the receiver indi erent between the two actions upon learning that his utility is above it. I derive necessary and su cient conditions for the sender s and receiver s welfare to be monotonic in information. Most notably, the sender s welfare increases and the receiver s welfare does not change with the precision of the sender s information. Moreover, the sender s welfare decreases and the receiver s welfare increases with the precision of public information. JEL Classi cation: C44, D81, D82, D83 Keywords: information, disclosure, persuasion, stochastic orders This paper is a revised version of Chapter 1 of my dissertation at MIT. I thank Robert Gibbons and Muhamet Yildiz for their invaluable guidance and advice. I also thank Ricardo Alonso, Abhijit Banerjee, Gabriel Carroll, Glenn Ellison, Péter Es½o, Richard Holden, Bengt Holmström, Sergei Izmalkov, Uliana Loginova, Niko Matouschek, Parag Pathak, Michael Powell, Andrea Prat, Jean Tirole, Juuso Toikka, Alexander Wolitzky, Juan Xandri, Luis Zermeño, and the participants at NES, MIT, and UNSW seminars for helpful comments and suggestions. y University of New South Wales, School of Economics. Email: a.kolotilin@unsw.edu.au. 1

1 Introduction This paper studies the optimal design of information disclosure by a sender who has an interest in a non-contractible action taken by a receiver. The central questions of this paper are: First, how much and what types of information are optimally disclosed by the sender? Second, how do the sender s and receiver s expected utilities depend on the information structure? To answer these questions, I consider the following sender-receiver game. The receiver has a binary action choice: to act or not to act. The sender s utility depends only on the action taken by the receiver, and she prefers the receiver to act. The receiver s utility depends both on his action and on information. The receiver takes an action that maximizes his expected utility given his beliefs. He forms his beliefs based on public information and information disclosed by the sender. The sender chooses ex ante how information will be disclosed to the receiver ex post. Formally, she can publicly choose any conditional distribution of messages given information. I call this distribution a mechanism. The sender chooses the mechanism that maximizes her expected utility the ex ante probability that the receiver will act. No monetary transfers between the sender and receiver are allowed. The drug approval process by the Food and Drug Administration (FDA) is a good example of my model. It takes on average between 8 to 12 years of studies and testing for a pharmaceutical company (manufacturer) to get a new drug approved by the FDA. The FDA requires a manufacturer to submit a research protocol for each test that the manufacturer is planning to undertake. The research protocol includes both tests required by the FDA and tests chosen by the manufacturer. The FDA closely monitors the record keeping and the adherence to the research protocol. If the test results are satisfactory, the manufacturer submits an application that contains all the information obtained during the testing phase. Penalties for fraudulent or misleading claims are imposed. Finally, the FDA either approves the drug or rejects it. 1 To see that my model is a good approximation of the drug approval process, let us reinterpret the manufacturer as the sender, the FDA as the receiver, the FDA s approval decision as the receiver s action, the research protocol as the sender s choice of a mechanism, the results of the required tests as public information, and the results of the remaining tests as a message. Because of the length and complexity of the drug approval process, the manufacturer presumably has little information at the beginning of the process. Furthermore, because of the large cost of the process and large bene ts of approval, the manufacturer 1 The description of the drug approval process is taken from Lipsky and Sharp (2001). 2

has strong incentives to optimally design tests to maximize the probability of the FDA s approval. For example, the manufacturer chooses dosage and characteristics of volunteer patients, such as gender, age, and health condition. Due to the FDA s regulation and close monitoring, the FDA observes both the design and results of all tests. Finally, based on the results of these tests, the FDA approves the drug if its bene ts outweigh its costs and risks. This model of optimal information disclosure describes many real-life examples. Scienti c experiments are often designed by interested parties. Grading policies are chosen by schools that wish to succeed in placing students. Accounting rules are chosen by rms seeking to maximize stock value. For concreteness, however, I motivate the questions and explain the answers using the drug approval process. 2 The manufacturer optimally chooses the test that produces only two possible outcomes: a positive outcome that gives minimal evidence su cient for the FDA s approval or a negative outcome that gives maximal evidence against approving. More speci cally, if the drug s quality is above a cuto, the optimal test produces the positive outcome, which persuades the FDA to approve the drug; otherwise, the test produces the negative outcome, which persuades the FDA to reject it. The cuto quality is determined by the condition that the FDA is indi erent between approving and rejecting the drug when the FDA gets positive news from the manufacturer. This design maximizes the probability of drug approval given the feasibility constraint imposed by Bayesian belief formation. How does the manufacturer s welfare (or equivalently the probability of the drug approval) depend on the information structure? First, if the manufacturer is able to design more informative tests, for example, by having more available volunteers, then its welfare is higher. This is because the manufacturer is free to choose any mechanism, so it has free disposal of information. For example, the manufacturer is free to choose to run tests on a smaller number of volunteers than maximum available. Second, if the FDA is more optimistic about the drug s quality, then the manufacturer s welfare is higher as well. This is because the manufacturer nds it easier to persuade the FDA to approve the drug if the FDA believes that the drug is good in the rst place. Interestingly, these two conditions are not only su cient but also necessary if the manufacturer s welfare is required to be higher for all values of the FDA s opportunity cost of approving the drug. Finally, if public 2 A practitioner should note, however, that my model is not directly applicable, because it abstracts away from some aspects inherent in the drug approval process. For example, in reality, the FDA has some commitment power in its approval decision, the manufacturer cares not only about the probability of approval, and there are constraints on choosing a research protocol. 3

information is more precise and less optimistic, then the manufacturer s welfare is lower. For example, if there are more required tests that the manufacturer must carry out, then the probability that the FDA will approve the drug decreases because the manufacturer has less freedom in designing a research protocol. How does the FDA s welfare (or equivalently the expected quality of approved drugs) depend on the information structure? Surprisingly, the FDA s welfare does not change if the manufacturer is able to design more informative tests. This is because the optimal design ensures that, for each approved drug, the FDA is indi erent between approving and rejecting it. However, the FDA s welfare increases with the precision of public information. That is, the quality of approved drugs increases if the FDA requires the manufacturer to run more tests. Finally, the overall welfare of the manufacturer and FDA is increasing in the precision of potential information of the manufacturer but is not monotonic in the precision of public information. Although the above monotone comparative statics results are intuitive, they do not generally hold in the large existing literature where the sender chooses what information to disclose when she already has her private information. In particular, they do not hold under cheap talk and veri able communication (Green and Stokey (2007) and Ivanov (2010)). The di erence is due to the sender s incentive compatibility constraint on information disclosure, which is absent in my model, because the sender chooses what information to reveal at the ex ante stage. It is straightforward to allow the manufacturer and FDA to have private information at the ex ante stage. The manufacturer obtains ex ante private information when it carries out preclinical trials before submitting a research protocol. By the standard unravelling argument due to Milgrom (1981), in the submitted protocol, the manufacturer discloses all its veri able information, such as recorded experimental results. However, since the manufacturer s and FDA s preferences are su ciently misaligned, the manufacturer discloses none of its unveri able information, such as its private opinion. The FDA obtains its veri able ex ante private information from applications submitted by other manufacturers that tested similar drugs. Again by the unravelling argument, the manufacturer optimally elicits all veri able private information from the FDA and then implements the optimal design of tests as if this private information was public. There is a large literature on strategic communication. In the cheap talk literature (Crawford and Sobel (1982)), the sender sends an unveri able message to the receiver who then takes a non-contractible action, so neither party can commit. In the delegation 4

literature (Holmstrom (1984)), the receiver can commit to how he will use the information transmitted by the sender. In contrast, in Rayo and Segal (2010), Kamenica and Gentzkow (2011), Kolotilin (2012), and this paper, the sender can essentially commit to how she will transmit information to the receiver. Kamenica and Gentzkow (2011) consider a much more general model than mine, with an arbitrary set of actions, and arbitrary utility functions for the sender and receiver. They derive some interesting properties of the optimal mechanism. To completely characterize the optimal mechanism, I impose more structure that still ts many real-life examples well. 3 More importantly, I derive general monotone comparative statics results that relate the sender s and receiver s expected utilities to information. Rayo and Segal (2010) and Kolotilin (2012) study optimal information disclosure when the receiver has private information. Lerner and Tirole (2006), Brocas and Carrillo (2007), and Benoit and Dubra (2011) study information disclosure in environments similar to mine, but in their models, the sender exogenously constrained in choosing a mechanism, so they do not characterize the optimal mechanism. Bergemann and Pesendorfer (2007), and Eso and Szentes (2007) study optimal information disclosure in environments in which monetary transfers are allowed. Finally, Athey and Levin (2001) derive monotone comparative statics results in certain single-person decision problems. The rest of the paper is organized as follows. Section 2 presents the model. Section 3 completely characterizes the optimal information disclosure mechanism and presents monotone comparative statics results. Section 4 extends the model to allow the sender and receiver to have private information at the ex ante stage. Section 5 concludes. All proofs and technical details are relegated to the appendices. 2 Model Consider a communication game between a female sender and a male receiver. The receiver takes a binary action a = 0; 1. Say that the receiver acts if he takes a = 1, and the receiver does not act if he takes a = 0. The sender s utility depends only on a, but the receiver s utility depends both on a and on (s; r), where components s and r denote the sender s type 3 In the context of my model, their results imply that the optimal test produces two outcomes: positive and negative. The former makes the FDA indi erent between approving and rejecting the drug, and the latter makes the FDA know that the true drug s quality is such that rejecting is uniquely optimal. On top of these results, I show that the positive outcome occurs if and only if the drug s quality is above a cuto. Moreover, my proof is more elementary and direct. 5

and public type, respectively. Without loss of generality, the sender s utility is a, and the receiver s utility is u 0 if a = 0 and s if a = 1. 4 Before (s; r) is realized, the sender can commit to a mechanism that sends a message m to the receiver as a (stochastic) function of (s; r); speci cally, the sender chooses the conditional distribution (mjs; r) of m given (s; r). Assume that the set of messages M contains at least two elements m 0 and m 1, the set of sender s types S is [s; s], the set of public types R is an arbitrary set that satis es mild regularity conditions that ensure that all conditional expectations exist. 5 The information (s; r) has some joint distribution. For this distribution, assume that the marginal distribution G (r) of r and the conditional distribution F (sjr) of s given r admit strictly positive densities g (r) and f (sjr). The timing of the communication game is as follows: 1. The sender publicly chooses a mechanism (mjs; r). 2. A triple (m; s; r) is drawn according to, F, and G. 3. The receiver observes (m; r) and takes an action a. 4. Utilities of the sender and receiver are realized. The solution concept used is a Perfect Bayesian Equilibrium (PBE). I view PBEs as identical if they have the same equilibrium mapping from information (s; r) to the receiver s action a. At the third stage, the receiver forms beliefs and acts if and only if the conditional expectation E [sjm; r] of s given (m; r) is at least u 0. (Note that a PBE requires that the receiver takes the sender s preferred action whenever he is indi erent between the two actions.) At the rst stage, the sender chooses an optimal mechanism that maximizes her expected utility, the probability that the receiver acts. Using the revelation principle, restrict attention to mechanisms that send only two messages: m 0 that persuades the receiver not to act and m 1 that persuades the receiver 4 De ning the sender s preferred action as a = 1 and applying an a ne transformation gives that his utility is a. Suppose now that the receiver s utility is u 0 (s; r) if a = 0 and u 1 (s; r) if a = 1. Because the action is binary, only the di erence u 1 (s; r) u 0 (s; r) matters for the receiver s choice of action, so u 0 (s; r) can be normalized to u 0 (or even to 0). Further, for any given r, which is observed both by the sender and the receiver, the sender s type can be transformed according to u 1 (:; r) to ensure that the receiver s utility from acting is s. 5 For example, R is allowed to be a complete separable metric space endowed with the Borel sigma algebra (Theorems 1.4.12 and 4.1.6 in Durrett (1996)). 6

to act. Adopt the convention that (m 1 js; r) denotes the probability of the message m 1 given (s; r). Hereafter, all notions are in the weak sense. For example, increasing means not decreasing and higher means not lower. 3 Analysis 3.1 Optimal Mechanism The optimal mechanism has a simple cuto structure. Theorem 1 The optimal mechanism is given by (m 1 js; r) = ( 1 if s s (r) ; 0 if s < s (r) : If R s sf (sjr) ds u s 0, then s (r) = s; otherwise s (r) < u 0 R s sf (sjr) ds = u s (r) 0. (1) is the unique solution to Clearly, the optimal mechanism is conditioned on each piece of public information r. This implies that it does not matter whether the sender commits to a mechanism before or after the realization of r. I give the intuition for Theorem 1 conditional on some value r. If it is not possible to induce the receiver to always act, then the optimal mechanism induces the receiver to act if and only if his utility is above the cuto. The cuto is such that the receiver is indi erent to act whenever he acts. Intuitively, the optimal mechanism has two de ning features: (i) it makes the receiver indi erent to act whenever he acts; and (ii) it makes the receiver know whether his utility is above the cuto. If the rst feature were violated, then the receiver would strongly prefer to act whenever he acts. Thus, it would be possible to increase the probability that the receiver acts by sending m 1 for a slightly larger set of types s. If the second feature were violated, then it would be possible to construct a mechanism that sends m 1 with the same probability, but to higher types s. This mechanism would violate the rst feature, so it would be possible to increase the probability that the receiver acts. Theorem 1 and subsequent results extend when the distribution of (s; r) does not admit a density, as I show in Appendix C. The only di erence is that the optimal mechanism may randomize over messages at the cuto as the following example shows. Suppose that u 0 = 0, there is no public information, and F is a discrete distribution that assigns probabilities 1 and 2 to 1 and 1. The optimal mechanism sends the message m 3 3 1 if s = 1, and the 7

messages m 1 and m 0 with equal probabilities if s = 1. As a result, the receiver who gets m 1 is indi erent to act and the probability of m 1 is 2. 3 Weaker versions of Theorem 1 appear in the literature. Lerner and Tirole (2006) show that the mechanism from Theorem 1 is optimal in a smaller class of feasible mechanisms in a more speci c setting than mine. Kamenica and Gentzkow (2011) establish Theorem 1 for the above discrete example. For a more general setting than mine, they derive interesting properties of the optimal mechanism. In particular, these properties imply that m 1 makes the receiver indi erent to act and that m 0 can only be sent to types s < u 0. However, they do not imply that that the optimal mechanism has a cuto structure in that m 0 is sent if and only if s < s (r). Moreover, my proof is simpler. 3.2 Comparative Statics without Public Information For simplicity, assume in this section that there is no public information. Theorem 2 presents monotone comparative statics results that relate the sender s and receiver s expected utilities under the optimal mechanism to the distribution of the sender s type. This theorem uses the standard de nitions from the literature on stochastic orders. Let P 1 and P 2 be two distributions. P 2 is higher than P 1 in the increasing convex order if there exists a distribution P such that P 2 rst-order stochastically dominates P and P is a meanpreserving spread of P 1. 6 Theorem 2 Let F 1 and F 2 be two distributions of s that do not depend on r. 1. The sender s expected utility under the optimal mechanism is higher under F 2 than under F 1 for all u 0 if and only if F 2 is higher than F 1 in the increasing convex order. 2. The receiver s expected utility under the optimal mechanism is higher under F 2 than under F 1 for all u 0 if and only if E F2 [s] E F1 [s]. Part 2 holds because the optimal mechanism is as uninformative as possible from the receiver s perspective, as follows from Theorem 1. Indeed, under the optimal mechanism, if the receiver acts, then he either holds the prior beliefs or is indi erent to act. Thus, the receiver s expected utility under the optimal mechanism is max fe [s] ; u 0 g, which is equal to his expected utility under a mechanism that sends the same message regardless of s. Part 1 is more interesting. Suppose for the sake of argument that there is an underlying binary state! that can take only two values! 1 < s and! 2 > s. The receiver s utility is 6 See De nition 1 and Lemma 1 in Appendix A for more de nitions and results on stochastic orders. 8

! if he acts (and u 0 if he does not). The sender s type s is a noisy signal about! and is normalized to the expectation of! given s; speci cally, the posterior Pr (! 2 js) is equal to s! 1! 2! 1. If F is a mean-preserving spread of F 1, then the distribution of posteriors under F is a mean-preserving spread of that under F 1 by the linearity of Pr (! 2 js) in s. This can be interpreted as F being more informative about the state (Blackwell (1953)). Since the sender can choose any mechanism, all mechanisms under F 1 are also feasible under F, which immediately implies that the sender s expected utility is higher under F. If F 2 rst-order stochastically dominates F, then F 2 is more favorable for acting and thus the sender can persuade the receiver to act more often. This intuition shows that F 2 being higher than F 1 in the increasing convex order is su cient for the conclusion that the sender s expected utility is higher under F 2. It turns out that this condition is also necessary if the conclusion is required to hold for all u 0. 7 Based on this intuition, the comparative statics results can be extended beyond this model as long as the sender can choose any mechanism at the ex ante stage. This assumption, however, is critical for the results. Under a cheap talk version of my model, the sender would not be able to disclose any information because she always prefers the receiver to act. Thus, the sender s expected utility would not change as her information becomes more precise. More generally, Green and Stokey (2007) and Ivanov (2010) show that the sender s expected utility may strictly decrease in the precision of her information. This happens because having less precise information may reduce the sender s incentive to misrepresent information. Similarly, under a veri able communication version of my model, the sender would disclose all her information by the unravelling argument due to Milgrom (1981). Thus, by Theorem 1, it is optimal for the sender to know only whether the receiver s utility is above the cuto, which is less informative than knowing the receiver s utility exactly. That is, the sender s utility may strictly decrease as her information becomes more precise. 3.3 Comparative Statics with Public Information This section generalizes the previous section and obtains comparative statics results with respect to public information. To present these comparative statics results, we need to 7 As can be seen from the proof, it is straightforward to write a strong version of Theorem 2 in which the sender s and receiver s expected utilities are strictly higher under F 2. 9

extend the stochastic orders introduced in Section 3.1 to the multidimensional case. 8 The multidimensionality arises because each piece of public information r is associated with a distinct conditional distribution F (:jr) of the sender s type s. Therefore, to compare distributions of r, we essentially need to compare distributions of distributions of s. Theorem 3 presents monotone comparative statics results that relate the sender s and receiver s expected utilities to the distribution of public information. This theorem uses a new stochastic order. P 2 is higher than P 1 in the increasing concave order if there exists P such that P 2 rst-order stochastically dominates P and P 1 is a mean-preserving spread of P. Theorem 3 Let P be the set of distributions on [s; s]. Let r be identi ed with the conditional distribution F of s given r. Let G 1 and G 2 be two distributions of r. 1. Let a partial order on P be the increasing convex order; speci cally, for all r 1 ; r 2 2 P, r 2 is higher than r 1 if F (:jr 2 ) is higher than F (:jr 1 ) in the increasing convex order. The sender s expected utility under the optimal mechanism is higher under G 2 than under G 1 for all u 0 if G 2 is higher than G 1 in the increasing concave order. 2. Let an order on P be the mean order; speci cally, for all r 1 ; r 2 2 P, r 2 is higher than r 1 if E [sjr 2 ] E [sjr 1 ]. The receiver s expected utility under the optimal mechanism is higher under G 2 than under G 1 for all u 0 if and only if G 2 is higher than G 1 in the increasing convex order. Part 1 of Theorem 3 states that the sender s expected utility increases as public information becomes more favorable for acting and less precise. By de nition, if G 2 is higher than G 1 in the increasing concave order, then there exists G such that G 2 rst-order stochastically dominates G and G 1 is a mean-preserving spread of G. First, G 2 puts a higher probability than G on public types r that result in a higher probability that the receiver acts conditional on r by Theorem 2 part 1. Therefore, the unconditional probability that the receiver acts is also higher under G 2. Second, based on the intuition from the previous section, the receiver is less informed about s under G than under G 1, and, therefore, it is easier to persuade him to act under G. Intuitively, since the sender can choose any mechanism, she can use a mechanism that sends two messages sequentially under G to achieve the optimal mechanism under G 1. The rst stage of this mechanism will then make public 8 See De nition 2 and Lemma 2 in Appendix A for de nitions and results on multidimensional stochastic orders. 10

information more precise, from G to G 1, and the second stage can be designed to disclose the optimal amount of information given more precise public information G 1. 9 Part 2 of Theorem 3 states that the receiver s expected utility increases as public information becomes more favorable for acting and more precise. Theorem 1 implies that the receiver has the same expected utility under the optimal mechanism and the mechanism 0 that sends the same message regardless of s. By de nition, if G 2 is higher than G 1 in the increasing convex order, then there exists G such that G 2 rst-order stochastically dominates G and G is a mean-preserving spread of G 1. First, G 2 puts a higher probability on public types r that are more favorable for acting by Theorem 2 part 2. Thus, overall, G 2 is more favorable for acting than G 1. Taking into account that the receiver s utility from not acting is xed, we get that the receiver s expected utility under 0 is higher under G 2 than under G. Second, public information is more precise under G than under G 1. Therefore, under 0, the receiver takes a more appropriate action under G and, thus, the receiver s expected utility is higher under G than under G 1. Similarly to Theorem 2, if the receiver s expected utility is higher under G 2 than under G 1 for all u 0, then G 2 must be higher than G 1 in the increasing convex order. Although the receiver s expected utility increases with the precision of public information, the social welfare does not necessarily increase even if it puts a very small weight on the sender s utility. Indeed, suppose that initially there is no public information. As public information appears, the marginal increase in the receiver s expected utility is 0 by the Envelope Theorem, as noted by Radner and Stiglitz (1984), but the marginal decrease in the sender s expected utility is strictly positive. 4 Extensions 4.1 Receiver s Veri able Private Information In this section, the receiver has veri able private information at the ex ante stage. As usual, veri able information is the information that cannot be lied about but can be concealed. In this case, the sender extracts the receiver s information at no cost and then discloses her information optimally as if the receiver s type was public. Therefore, all results of Section 3 apply. 9 The above conditions on G 2 and G 1 are not only su cient but also necessary for the sender s expected utility to be higher under G 2 for all u 0 if the support of F (:jr) contatins only two elements. However, the conditions are not necessary if the support of F (:jr) contatins more elements. See Appendix D for details. 11

To illustrate this result, assume that the type r is privately known by the receiver rather than publicly known. In other respects, the environment is the same as in Section 2. In particular, players, actions, the information structure, and preferences are the same. In addition, assume that the set of receiver s types R is given by [r; r] and is ordered in such a way that s (r) is strictly increasing in r where s (r) is given by Theorem 1. Similarly to Milgrom (1981), assume that the set of receiver s reports is N (r) = [r; r]. That is, the receiver can report any type that is lower than his true type. Intuitively, the report n can be viewed as the receiver s claim that his true type r is at least n and the receiver s claims are required to be truthful in that r must belong to [n; r]. Now a mechanism sends a message m to the receiver as a (stochastic) function of (s; n). Finally, the timing of the game is as follows: 1. The sender publicly chooses a mechanism (mjs; n). 2. The receiver s type r is drawn according to G. 3. The receiver makes a report n. 4. A pair (m; s) is drawn according to and F. 5. The receiver gets a message m and takes an action a. 6. Utilities are realized. Again, the solution concept used is a PBE. Theorem 4 characterizes the unique PBE. Theorem 4 In the unique PBE, the receiver reports his true type n = r and the sender chooses the optimal mechanism given by Theorem 1. This theorem shows that without loss of generality we can view the receiver s veri able private information as public information. This result is in spirit of the standard unravelling result due to Milgrom (1981) who show that generally all veri able private information is disclosed in an equilibrium. Note that the mechanism and truthful reporting of the receiver constitutes a PBE even if the sender has partial commitment in that she can choose a mechanism only after the receiver s report. However, this PBE is not unique in this new model. For example, there exists a PBE in which the receiver always reports n = 0. 10 4.2 Sender s Ex Ante Private Information In this section, the sender has private information before she chooses a mechanism. As a result, the sender discloses all of her veri able information and none of her unveri able 10 Indeed, suppose that the sender believes that each out-of-equilibrium report n 6= 0 is made by the receiver with type r = n. Note that under such a belief, the sender chooses a mechanism (mjs; n) for any n 6= 0. Thus, the receiver s interim expected utility from reporting n 6= 0 is max fu 0 ; E [sjr]g, which is smaller than that from reporting n = 0. 12

information. Thus, without loss of generality, the sender s veri able information can be viewed as public information, and the sender s ex ante unveri able information can be integrated out. Again, all results of Section 3 apply. In this section, the type r is privately known by the sender rather than publicly known. In other respects, the environment is the same as in Section 2. In addition, assume that R is given by [r; r] and is ordered in such a way that s (r) is strictly decreasing in r. The timing of the game is as follows: 1. The sender s type r is drawn according to G. 2. The sender makes a report n. 3. The sender publicly chooses a mechanism (mjs; n). 4. A pair (m; s) is drawn according to and F. 5. The receiver gets a report n and a message m and takes an action a. 6. Utilities are realized. Again, the solution concept is a PBE. The sender s information is veri able if the set of her reports is N (r) = [r; r], where the report n can be viewed as the receiver s truthful claim that his type is at least n. The sender s information is unveri able if the set of her reports is N = [r; r] regardless of r. Theorem 5 characterizes the unique PBE for both cases of veri able and unveri able information of the sender. Theorem 5 If the sender s ex ante private information is veri able, then in the unique PBE, the sender reports n = r and chooses the optimal mechanism given by Theorem 1. If the sender s ex ante private information is unveri able, then in the unique PBE, the sender reports some xed n regardless of r and chooses the optimal mechanism given by Theorem 1 where F (sjr) is replaced with R F (sjr) dg (r) for all r. R The sender discloses all her veri able private information again due to the standard unravelling argument (Milgrom (1981)). The sender conceals all her unveri able private information, because regardless of her information, she always wants to pretend that she has the best news for the receiver. Note that if the sender could commit to a mechanism before realization of r, then by Theorem 1, the optimal mechanism would be where is de ned in Theorem 5. That is, the full commitment optimum is achieved as the equilibrium outcome if the sender s information is unveri able. This observation is consistent with Theorem 3, which shows that the sender s expected utility decreases with the precision of public information. 13

5 Conclusions In this paper, I have studied optimal information disclosure mechanisms. I have imposed the following key assumptions. First, at the ex ante stage, the sender can publicly choose how her information will be disclosed ex post; speci cally, she can choose any conditional distribution of messages given her information. Second, the receiver has a binary action choice. Third, the sender s utility depends on the receiver s action but does not depend on information. The optimal mechanism has a particularly simple structure. It discloses only whether the sender s utility is above the cuto where the cuto is such that the receiver does not get any rent from learning whether his utility is above or below it. The sender s and receiver s welfare is monotonic in information. The sender s welfare increases with the precision of her potential information and decreases with the precision of public information. The receiver s welfare does not change with the precision of the sender s potential information and increases with the precision of public information. I have also analyzed the extensions where the sender and receiver have private information at the ex ante stage. However, in this paper, I have not explored the possibility of the receiver having private information that cannot be elicited by the sender ex ante. Generically, the receiver does have such private information at least by the time he takes an action. For example, the FDA carries out an independent review after receiving the application from the manufacturer. Moreover, the manufacturer is uncertain about preferences and beliefs of the FDA regarding the safety and e cacy of a new drug. Since the optimal mechanism leaves no rent to the receiver if the receiver is uninformed, as a trivial result, the optimal mechanism is (weakly) more informative if the receiver is privately informed. The detailed analysis of this situation is my central goal in Kolotilin (2012). References Athey, Susan and Jonathan Levin (2001) The Value of Information in Monotone Decision Problems. Mimeo, MIT. Benoit, Jean-Pierre and Juan Dubra (2011) Apparent Overcon dence, Econometrica, 79 (5), pp. 1591 1625. Bergemann, Dirk and Martin Pesendorfer (2007) Information Structures in Optimal Auctions, Journal of Economic Theory, 137 (1), pp. 580 609. 14

Blackwell, David (1953) Equivalent Comparisons of Experiments, Annals of Mathematical Statistics, 24 (2), pp. 265 272. Brocas, Isabelle and Juan D. Carrillo (2007) In uence through Ignorance, RAND Journal of Economics, 38 (4), pp. 931 947. Crawford, Vincent P. and Joel Sobel (1982) Strategic Information Transmission, Econometrica, 50 (6), pp. 1431 1451. Durrett, Richard (1996) Probability: Theory and Examples, Belmont, CA: Duxbury Press. Eso, Peter and Balazs Szentes (2007) Optimal information disclosure in auctions and the handicap auction, Review of Economic Studies, 74 (3), p. 705. Green, Jerry R. and Nancy L. Stokey (2007) A Two-Person Game of Information Transmission, Journal of Economic Theory, 135 (1), pp. 90 104. Holmstrom, Bengt (1984) On the Theory of Delegation, in M Boyer and R. Kihlstrom eds. Bayesian Models in Economic Theory, New York: North-Holland. Ivanov, Maxim (2010) Informational Control and Organization Design, Journal of Economic Theory, 145 (2), pp. 721 751. Kamenica, Emir and Matthew Gentzkow (2011) Bayesian Persuasion, American Economic Review, 101 (6), pp. 2590 2615. Kolotilin, Anton (2012) Optimal Information Disclosure to an Informed Decision-Maker. Mimeo, MIT. Lerner, Josh and Jean Tirole (2006) A Model of Forum Shopping, American Economic Review, 96 (4), pp. 1091 1113. Lipsky, Martin S. and Lisa K. Sharp (2001) From Idea to Market: The Drug Approval Process, Journal of the American Board of Family Medicine, 14 (5), pp. 362 367. Milgrom, Paul (1981) Good News and Bad News: Representation Theorems and Applications, Bell Journal of Economics, 12 (2), pp. 380 391. Radner, Roy and Joseph E. Stiglitz (1984) A Nonconcavity in the Value of Information, in M Boyer and R. Kihlstrom eds. Bayesian Models in Economic Theory, New York: North-Holland. 15

Rayo, Luis and Ilya Segal (2010) Optimal Information Disclosure, Journal of Political Economy, 118 (5), pp. 949 987. Shaked, Moshe and J. George Shanthikumar (2007) Stochastic Orders, New York: Springer. Appendix A: Stochastic Orders De nition 1 presents the unidimensional stochastic orders used in Section 3.2. De nition 1 Let X 1 and X 2 be two random variables with distributions P 1 and P 2 on [x; x]. Say that 1. P 2 rst-order stochastically dominates P 1 (denoted by P 2 st P 1 ) if P 2 (x) P 1 (x) for all x. 2. P 2 is a mean-preserving spread of P 1 (denoted by P 2 cx P 1 ) if there exist two random variables X b 2 and h X b 1, de ned on the same probability space, with distributions P 2 and P 1 such that E bx2 jx b i 1 = X b 1. 3. P 2 is higher than P 1 in the increasing convex order (denoted by P 2 icx P 1 ) if there exists a distribution P such that P 2 st P cx P 1. Lemma 1 gives useful equivalent representations of the above stochastic orders. Lemma 1 Let P 1 and P 2 be two distributions that admit densities on [x; x]. 1. P 2 st P 1 if and only if E [h (X 2 )] E [h (X 1 )] for all increasing functions h. 2. P 2 cx P 1 if and only if E [h (X 2 )] E [h (X 1 )] for all convex functions h. 3. P 2 icx P 1 is equivalent to each of the following conditions: (a) E [h (X 2 )] E [h (X 1 )] for all increasing convex functions h; (b) R x x P 2 (ex) dex R x x P 1 (ex) dex for all x 2 [x; x]; (c) R 1 p P 1 2 (ep) dep R 1 p P 1 1 (ep) dep for all p 2 [0; 1]. Proof. See Shaked and Shanthikumar (2007) Section 1.A.1 for part 1, Section 3.A.1 for part 2, and Section 4.A.1 for part 3. De nition 2 presents the multidimensional stochastic orders used in Section 3.3. 16

De nition 2 Let P be the set of distributions on [x; x] endowed with some partial order P. Let X 1 and X 2 be two random elements with distributions Q 1 and Q 2 on P. Say that 1. Q 2 rst-order stochastically dominates Q 1 (denoted by Q 2 mst Q 1 ) if Pr Q2 (X 2 2 U) Pr Q1 (X 1 2 U) for all measurable increasing sets U P in that P P P 0 and P 0 2 U imply P 2 U. 2. Q 2 is a mean-preserving spread of Q 1 (denoted by Q 2 mcx Q 1 ) if there exist two random elements X b 2 and h X b 1, de ned on the same probability space, with distributions Q 2 and Q 1 such that E bx2 jx b i 1 = X b 1. 3. Q 2 is higher than Q 1 in the increasing convex order (denoted by Q 2 micx Q 1 ) if there exists a distribution Q such that Q 2 mst Q mcx Q 1. 4. Q 2 is higher than Q 1 in the increasing concave order (denoted by Q 2 micv Q 1 ) if there exists a distribution Q such that Q 2 mst Q and Q 1 mcx Q. Lemma 1 gives equivalent representations of the multidimensional stochastic orders. Lemma 2 Let Q 1 and Q 2 be two distributions on P. 1. Q 2 mst Q 1 if and only if E [h (X 2 )] E [h (X 1 )] for all increasing functions h in that h (P 2 ) h (P 1 ) for all P 1 ; P 2 2 P such that P 2 P P 1. 2. Q 2 mcx Q 1 if and only if E [h (X 2 )] E [h (X 1 )] for all convex functions h in that h (P 1 + (1 ) P 2 ) h (P 1 ) + (1 ) h (P 2 ) for all P 1 ; P 2 2 P and all 2 (0; 1). 3. Q 2 micx Q 1 if and only if E [h (X 2 )] E [h (X 1 )] for all increasing convex functions h. 4. Q 2 micv Q 1 if and only if E [h (X 2 )] E [h (X 1 )] for all increasing concave functions h. Proof. See Shaked and Shanthikumar (2007) Section 6.B.1 for part 1, and Section 7.A.1 for parts 2, 3, and 4. 17

Appendix B: Proofs Proof of Theorem 1. subject to The optimal mechanism solves Z f (sjr) g (r) (m 1 js; r) drds maximize (m 1 js;r)2[0;1] Z S SR sf (sjr) (m 1 js; r) ds u 0 for all r 2 R where the objective function is the probability that the receiver acts and the constraint requires that the receiver prefers to act whenever he receives m 1. The Lagrangian for this problem is given by: Z L = (1 + s (r)) f (sjr) g (r) (m 1 js; r) drds; SR where (r) g (r) is a multiplier for the constraint. Since the choice variable (m 1 js; r) 1 belongs to the unit interval, we have (m 1 js; r) = 1 if s and (m (r) 1js; r) = 0 if 1 s < where (r) is 0 if E (r) F [sjr] > u 0 and is such that the constraint is binding if E F [sjr] u 0. Proof of Theorem 2. I start by proving the rst part. Let s i be given by Theorem 1 where F is replaced with F i. If F 2 icx F 1 (see De nition 1), then the sender can induce the receiver to act with a higher probability under F 2 than under F 1 because Z s F 1 2 (F 1(s 1)) sdf 2 (s) = Z 1 F 1(s 1) F 1 2 (ep) dep Z 1 F 1(s 1) F 1 1 (ep) dep = Z s s 1 sdf 1 (s) u 0 ; where the equalities hold by the appropriate change of variables, the rst inequality holds by Lemma 1 part 3 (c), and the last inequality holds by Theorem 1. Conversely, if F 2 icx F 1, then by Lemma 1 part 3 (c), there exists p such that R 1 F 1 p 2 (ep) dep < R 1 F 1 p 1 (ep) dep. Setting u 0 = R 1 sdf F 1 2 (p) 2 (s) and using an analogous argument, we get that the receiver acts with a strictly higher probability under F 1 than under F 2 : Z s F 1 1 (p) sdf 1 (s) = Z 1 p F 1 1 (ep) dep > Z 1 p Z s F2 1 (ep) dep = sdf 2 (s) = u 0 : F 1 2 (p) Now I prove the second part. The receiver s expected utility under F i is max fe Fi [s] ; u 0 g by Theorem 1. Clearly, if E F2 [s] E F1 [s], then max fe F2 [s] ; u 0 g max fe F1 [s] ; u 0 g for all u 0. Conversely, if E F2 [s] < E F1 [s], then max fe F2 [s] ; u 0 g < max fe F1 [s] ; u 0 g for any u 0 2 (E F2 [s] ; E F1 [s]). 18

Proof of Theorem 3. The probability that the receiver acts is R R p (r) dg (r) where the conditional probability p (r) that the receiver acts is given by 1 F (s (r) jr) with s (r) given by Theorem 1. The function p is increasing in r in the increasing convex order by Theorem 2 part 1. Moreover, p is concave in r, as I show in the next paragraph. Therefore, part 1 of the theorem follows by Lemma 2 part 4. For concavity of p, it su ces to show that there exists a mechanism that induces the receiver to act with probability p (r 1 ) + (1 F (sjr 1 ) + (1 ) p (r 2 ) when the distribution of s is ) F (sjr 2 ). Without loss of generality, suppose that s (r 1 ) s (r 2 ). The required mechanism is simply a mechanism that implements 1 and 2 with probabilities and 1. Speci cally, if s s (r 1 ), the receiver gets the message m 1. If s < s (r 2 ), the receiver gets the message m 0. Finally, if s 2 [s (r 2 ) ; s (r 1 )), the receiver gets the messages m 1 and m 0 with probabilities p 1 and 1 p 1 p 1 where (1 ) (F (s (r 1 ) jr 2 ) F (s (r 2 ) jr 2 )) (F (s (r 1 ) jr 1 ) F (s (r 2 ) jr 1 )) + (1 ) (F (s (r 1 ) jr 2 ) F (s (r 2 ) jr 2 )) : The receiver s expected utility under the optimal mechanism is Z Z r U R = max fu 0 ; E [sjr]g dg (r) = max fu 0 ; rg dg (r) = r R r Z r u 0 G (r) dr; where the rst equality holds by Theorem 1, the second by the ordering assumption E [sjr] = r, and the third by integration by parts. Part 2 of the theorem follows immediately by Lemma 1 part 3 (b). Proof of Theorem 4. I start by showing that the described strategies constitute a PBE. If the receiver reports n = r, then his interim expected utility is max fu 0 ; E [sjr]g as follows from Theorem 1. If the receiver reports n < r, then his interim expected utility is again max fu 0 ; E [sjr]g because s (r) is increasing in r. Thus, given the mechanism, it is a best response for the receiver to report his true type n = r. To see that it is optimal for the sender to choose at the rst stage, note that is the optimal mechanism in the relaxed problem where r is publicly known, so gives a higher expected utility to the sender than any other feasible mechanism. To complete the proof, I show that in all PBEs, the sender chooses and the receiver reports n = r. Suppose to get a contradiction that there exists another PBE. In this PBE, the sender s expected utility is strictly less than in the above PBE because is the optimal mechanism in the relaxed problem. Consider a mechanism e that sends the message m 1 if and only if s s (r) + where > 0 is su ciently small. Under this mechanism, the receiver strictly prefers to report his true type r and the sender s expected utility is arbitrarily close to that under. A contradiction. 19

Proof of Theorem 5. Suppose that given the sender s report n, r is distributed according to G n. Given this report, the receiver believes that s is distributed according to F n (s) = R F (sjr) dg R n (r). By sequential rationality, at the third stage, the sender chooses the optimal mechanism n that sends m 1 if and only if s s n where s n is given by Theorem 1 where F (sjr) is replaced with F n (s). I start by considering the case where the sender s information is veri able. In this case, the sender r can make a report n only if r n. Thus, the support of G n does not intersect [r; n). Suppose to get a contradiction that there exists an equilibrium report n such that H n is supported on [r n ; r n ] with r n > n. This means that with a strictly positive probability, the sender r n makes the report n and induces the receiver to act with probability 1 F (s njr n ). If this sender made the report r n instead, then she would induce the receiver to act with a strictly higher probability because s r n < s n, as I show in the next paragraph. Thus, G n assigns probability one to r = n, meaning that the sender reports n = r for all r. The inequality s r n < s n holds because s r n s (r n ) < s n. Suppose to get a contradiction that s n s (r n ), then R s sdf s n (s) = R r n R s sdf (sjr) dg n r n s n (r) < 0, contradicting the n de nition of s n. The equality holds by Fubini s Theorem and the de nition of F n (s). The inequality holds because R s sdf (sjr) < R s sdf (sjr) = 0 for all r < r s n s (r) n as follows from s n < s (r) < 0 which is implied by the supposition s n s (r n ) and the assumption that s (r) is strictly decreasing in r. Noting that the support of G rn does not intersect [r; r n ) and using a similar argument gives s r n s (r n ). Now, I consider the case where the sender s information is unveri able. Suppose to get a contradiction that there exist two equilibrium reports n 1 and n 2 such that s n 1 < s n 2. Then the sender would always prefer to report n 1 regardless of r. A contradiction. Appendix C: Discontinuous Distributions This appendix relaxes the assumption that all distributions are continuous. Instead, assume that G (r) and F (sjr) are arbitrary distributions whose supports are subsets of R and S = [s; s]. Theorem 6, a generalization of Theorem 1, characterizes the optimal mechanism. Theorem 6 The optimal mechanism is given by 8 >< 1 if s > s (r) ; (m 1 js; r) = (r) if s = s (r) ; >: 0 if s < s (r) : (2) 20

If R s sdf (sjr) u s 0, then s (r) = s and (r) = 1; otherwise s (r) 0 and q (r) (r) Pr (s = s (r) jr) 2 [0; Pr (s = s (r) jr)] are the unique solution to Z E [sjm 1 ] = sdf (sjr) + s (r) q (r) = 0: (3) (s (r);s] Proof. By Fubini s Theorem, the optimal mechanism solves Z Z (m 1 js; r) df (sjr) dg (r) subject to maximize (m 1 js;r)2[0;1] Z S R S s (m 1 js; r) df (sjr) u 0 for all r 2 R. We can see that the problem is separable; speci cally, for each r, the optimal mechanism maximizes the inside integral subject to the constraint. Therefore, (m 1 js; r) = e (mjs), where e (mjs) is the optimal mechanism in the model in which r is xed, and the distribution of s is given by F (sjr). Thus, we can omit r from consideration as if it was xed. Now I prove that if R s s sdf (s) < u 0, then the optimal mechanism satis es (2) where (s ; ) solves (3). The remaining parts of Theorem 6 are immediate. Suppose to get a contradiction that there exists a mechanism e that results in a higher probability that the receiver acts: Pr e (m 1 ) > Pr (m 1 ). In the next paragraph, I show that F (sjm 1 ) F e (sjm 1 ) for all s 2 [s; s] with strict inequality for s 2 [s ; s), and, thus, E [sjm 1 ] > E e [sjm 1 ] by the well-known result (a strong version of Lemma 1 part 1). Therefore, E e [sjm 1 ] < u 0 because E [sjm 1 ] = u 0 by (3). The conclusion that E e [sjm 1 ] < u 0 contradicts the assumption that the message m 1 induces the receiver to act. To complete the proof, I show that F (sjm 1 ) F e (sjm 1 ) for all s 2 [s; s] with strict inequality for s 2 [s ; s). The inequality trivially holds for s < s because F (sjm 1 ) = 0 and for s = s because F (sjm 1 ) = F e (sjm 1 ) = 1. m and s by (m; s). Denote the joint distribution of The following sequence of equalities and inequalities proves that 21

F (sjm 1 ) < F e (sjm 1 ) for s 2 [s ; s): 1 F (sjm 1 ) = (m 1 ; s) (m 1 ; s) Pr (m 1 ) F (s) F (s) = Pr (m 1 ) = e (m 1 ; s) e (m1 ; s) Pr (m 1 ) e (m 1 ; s) e (m1 ; s) Pr (m 1 ) > e (m 1 ; s) e (m1 ; s) Pr e (m 1 ) = 1 F e (sjm 1 ) : + e (m 0 ; s) e (m0 ; s) Pr (m 1 ) The rst and last equalities hold by Bayes rule. The second equality holds by (2), which de- nes (m; s). The third equality holds by the consistency condition: (m 1 ; s)+ (m 0 ; s) = F (s) for all mechanisms and all s 2 [s; s]. The rst inequality holds because (m 0 ; :) is a distribution function of s. Pr e (m 1 ) > Pr (m 1 ). The second inequality holds by the assumption that Theorem 2 holds regardless of whether F 1 and F 2 admit densities. The original proof of the second part of Theorem 2 applies to arbitrary F 1 and F 2. To prove the rst part of Theorem 2, one should replace the inverse functions with the quantile functions in Lemma 1 part 3 (c) and in the original proof. Speci cally, for an arbitrary distribution P, the quantile function is de ned as ' (p) inf fx : p P (x)g. If F 2 icx F 1, then the receiver acts with a higher probability under F 2 than under F 1 because Z 1 ' 2 (ep) dep F 1(s 1) q1 Z 1 ' 1 (ep) dep = F 1(s 1) q1 Z (s ;s] sdf 1 (s) + s 1q1 u 0 : 1 Conversely, if F 2 icx F 1, there exists p such that R 1 p ' 2 (ep) dep < R 1 p ' 1 (ep) dep, so the receiver acts with a strictly higher probability under F 1 than under F 2 if u 0 = R 1 p ' 2 (ep) dep. Using similar logic, it is straightforward to extend Theorems 3, 4, and 5 to the case of arbitrary distribution functions. Appendix D: Necessity in Theorem 3 If the support of F (:jr) contains only two elements for all r, say s 1 and s 2 with s 1 < s 2, then part 1 of Theorem 3 can be strengthened. In this case, the sender s expected utility 22