Testing Ambiguity Models through the Measurement of Probabilities for Gains and Losses

Similar documents
Recursive Ambiguity and Machina s Examples

Recursive Ambiguity and Machina s Examples

Uniform Sources of Uncertainty for Subjective Probabilities and

RECURSIVE AMBIGUITY AND MACHINA S EXAMPLES 1. INTRODUCTION

Second-Order Expected Utility

Measurable Ambiguity. with Wolfgang Pesendorfer. August 2009

Ambiguity Models and the Machina Paradoxes

Are Probabilities Used in Markets? 1

An Axiomatic Model of Reference Dependence under Uncertainty. Yosuke Hashidate

MEASURING AMBIGUITY ATTITUDES FOR ALL (NATURAL) EVENTS

Bayesian consistent prior selection

Partial Ambiguity. CHEW Soo Hong, MIAO Bin, and ZHONG Songfa. November 2016

Comparing uncertainty aversion towards different sources

Allais, Ellsberg, and Preferences for Hedging

Ellsberg Paradox: Ambiguity and Complexity Aversions Compared

AMBIGUITY AVERSION WITH THREE OR MORE OUTCOMES

Allais Paradox. The set of prizes is X = {$0, $1, 000, 000, $5, 000, 000}.

Temporal Resolution of Uncertainty and Recursive Models of Ambiguity Aversion. Tomasz Strzalecki Harvard University

Completing the State Space with Subjective States 1

The Ellsberg paradox: A challenge to quantum decision theory?*

Skewed Noise. David Dillenberger 1 Uzi Segal 2. 1 University of Pennsylvania 2 Boston College and WBS

Vector Expected Utility and Attitudes toward Variation

Consideration Sets. Mark Dean. Behavioral Economics G6943 Fall 2015

Sealed Bid Auctions with Ambiguity: Theory and Experiments

We consider choice over uncertain, monetary payoffs and study a general class of preferences. These preferences

Quantum Decision Theory

Skewed Noise. David Dillenberger Uzi Segal. November 25, 2013

The willingness to pay for health improvements under comorbidity ambiguity 1

Endogenizing Prospect Theory s Reference Point. by Ulrich Schmidt and Horst Zank

Sealed bid auctions with ambiguity: Theory and experiments

Rationality and Uncertainty

Discussion Papers in Economics

September 2007, France

A Truth-Serum for Non-Bayesians: Correcting Proper Scoring Rules for Risk Attitudes 1

Essays on Decision Making: Intertemporal Choice and. Uncertainty

Cautious Expected Utility and the Certainty Effect

Common ratio using delay

Is it Possible to Define Subjective Probabilities in Purely Behavioral Terms? A Comment on Epstein-Zhang (2001)

Agrowing body of qualitative evidence shows that loss aversion, a phenomenon formalized in prospect

Pre-probabilistic sophistication. Osamu Kada. July, 2009

Bayesian consistent prior selection

OPTIMISM AND PESSIMISM IN GAMES

Ambiguity Framed. Mark Schneider, Jonathan Leland, and Nathaniel T. Wilcox

Ambiguity Aversion: An Axiomatic Approach Using Second Order Probabilities

Making the Anscombe-Aumann Approach to Ambiguity Suited for Descriptive Applications

Event-Separability in the Ellsberg urn

Risk Sharing in the Small and in the Large

Indifference Pricing with. Uncertainty Averse Preferences

Risk Sharing in the Small and in the Large

An Ambiguity Aversion Model for Decision Making under Ambiguity

A Tailor-Made Test of Intransitive Choice

Great Expectations. Part I: On the Customizability of Generalized Expected Utility*

When does aggregation reduce risk aversion?

Cautious Expected Utility and the Certainty Effect

Small Worlds: Modeling Attitudes towards Sources of Uncertainty. October Abstract

CER-ETH Center of Economic Research at ETH Zurich

Bounded Rationality I: Consideration Sets

The nature of utility is controversial. Whereas decision theory commonly assumes that utility is context

Preference, Choice and Utility

Prospect Theory: An Analysis of Decision Under Risk

Sealed Bid Auctions with Ambiguity: Theory and Experiments

A Theory of Subjective Compound Lotteries

A Bayesian Approach to Uncertainty Aversion

Questions in Decision Theory

A powerful Tool for Analyzing Concave/Convex Utility and Weighting Functions

A Behavioral Characterization of Plausible Priors

Sealed Bid Auctions with Ambiguity: Theory and Experiments

A theory of robust experiments for choice under uncertainty

Eliciting Gul s Theory of Disappointment Aversion by the Tradeoff Method 1 HAN BLEICHRODT. March 2007

Dynamic Consistency and Subjective Beliefs

Incentives in Experiments: Theory and an Experimental Test

Axiomatic Foundations of Multiplier Preferences

Comments on prospect theory

Range-Dependent Utility

SUBJECTIVE INDEPENDENCE AND CONCAVE EXPECTED UTILITY. February 12, 2014

Cautious and Globally Ambiguity Averse

TESTING AND CHARACTERIZING PROPERTIES OF NONADDITIVE MEASURES THROUGH VIOLATIONS OF THE SURE-THING PRINCIPLE

Efficient Allocations under Ambiguity

Ambiguity Aversion Without Asymmetric Information

Ambiguity Aversion and Trade

Expected Utility Framework

Relevance and Symmetry

Objective Imprecise Probabilistic Information, Second Order Beliefs and Ambiguity Aversion: an Axiomatization. Raphaël Giraud

Rational Expectations and Ambiguity: A Comment on Abel (2002)

Ambiguity under Growing Awareness

Decision-making with belief functions

Certainty Equivalent Representation of Binary Gambles That Are Decomposed into Risky and Sure Parts

PIER Working Paper

Quantum Decision Theory and the Ellsberg Paradox

Testing for Ambiguity in Coordination Games

A Behavioral Characterization of Plausible Priors

Ellsberg re-revisited: An experiment disentangling model uncertainty and risk aversion

Bounded Rationality I: Consideration Sets

A SMOOTH MODEL OF DECISION MAKING UNDER AMBIGUITY

Choice under uncertainty

DEPARTMENT OF ECONOMICS DISCUSSION PAPER SERIES

Parametric weighting functions

Bayesian Updating for General Maxmin Expected Utility Preferences

Perceived Ambiguity and Relevant Measures

Cautious Expected Utility and the Certainty Effect

Transcription:

Testing Ambiguity Models through the Measurement of Probabilities for Gains and Losses Aurélien Baillon and Han Bleichrodt Erasmus School of Economics Erasmus University Rotterdam P.O. Box 1738, Rotterdam, 3000 DR, the Netherlands baillon@ese.eur.nl bleichrodt@ese.eur.nl 03 April 2014 Abstract: This paper reports on two experiments that test the descriptive validity of ambiguity models using a natural source of uncertainty (the evolution of stock indices) and both gains and losses. We observed violations of probabilistic sophistication, violations that imply a fourfold pattern of ambiguity attitudes: ambiguity aversion for likely gains and unlikely losses and ambiguity seeking for unlikely gains and likely losses. Our data are most consistent with prospect theory and, to a lesser extent, -maxmin expected utility and Choquet expected utility. Models with uniform ambiguity attitudes are inconsistent with most of the observed behavioral patterns. Keywords: Decision under uncertainty, probabilistic sophistication, ambiguity, experiments. Acknowledgments: We are grateful to Yoram Halevy, Olivier L Haridon, Peter P. Wakker, the editor Robert Porter, and two anonymous reviewers for helpful comments and to Kirsten I.M. Rohde, Uyanga Turmunkh and Tong Wang for their help in collecting the data. The research of Han Bleichrodt and Aurélien Baillon was made possible through grants from the Netherlands Organization for Scientific Research (NWO). 1/40

1. Introduction Many real-life decisions involve uncertainty, where objective probabilities are unavailable. Examples include a physician making a diagnosis, an entrepreneur contemplating whether to introduce a new product, and a politician determining the date of a future election. To analyze such decisions, we must measure people s beliefs about the likelihood of uncertain events. Under subjective expected utility (Savage 1954), the classical model of decision under uncertainty, and the more general model of probabilistic sophistication (Machina and Schmeidler, 1992), beliefs can be measured through the elicitation of matching probabilities (e.g., Raiffa 1968 section 5.3, Spetzler and Staël von Holstein 1975, Holt 2007 section 30.5). A probability p is a matching probability of an event E, if a decision maker is indifferent between receiving x if E occurs and x with probability p. If matching probabilities measure beliefs, they must be additive and independent of the outcome x used to elicit them. In particular, they must be independent of the sign of x, and they must be the same for gains and losses. Probabilistic sophistication is often regarded as a normative requirement (Machina and Schmeidler 1995). Descriptively, probabilistic sophistication was challenged by the Ellsberg paradox (Ellsberg 1961), which showed that people are averse to ambiguity associated with betting on uncertain events, casting doubt on the existence of subjective probabilities. In response to Ellsberg s paradox, new models have been proposed to explain ambiguity aversion. Important examples are maxmin expected utility (Gilboa and Schmeidler 1989) and its generalizations, -maxmin expected utility (Ghirardato, Maccheroni and Marinacci 2004) and the variational model (Maccheroni et al. 2006), Choquet expected utility (Gilboa 1987, Schmeidler 1989) and its generalization in the form of prospect theory (Tversky and Kahneman 1992, Wakker 2010), the smooth ambiguity model (Klibanoff et al. 2005), and vector expected utility (Siniscalchi 2009). It is generally believed that tests of these ambiguity models are difficult because they employ 2/40

concepts that are difficult to observe empirically. Previous tests (Halevy 2007, Machina 2009, Lo 2008, L Haridon and Placido 2010, Hey, Lotito, Maffioletti 2010, Baillon, L Haridon, and Placido 2011) have used complex choices with several non-zero outcomes and artificial Ellsberg-like uncertainties (urns and balls). In this paper, we use elicited matching probabilities to shed light on patterns of violations of probabilistic sophistication and the descriptive validity of some widely used ambiguity models. Our design uses elementary choices involving just one non-zero money amount and naturally occurring ambiguous events based on the performances of stock markets. In two experiments, we found clear violations of probabilistic sophistication: matching probabilities were not additive and differed between gains and losses, and the violations of additivity were sign-dependent. The observed violations suggest that descriptive ambiguity models should allow for less sensitivity to changes in likelihood under ambiguity than under risk and for different ambiguity attitudes for gains and losses. Together, these results imply a fourfold pattern of ambiguity attitudes: ambiguity aversion for likely gains and unlikely losses and ambiguity seeking for unlikely gains and likely losses. Models that can account for this pattern include prospect theory and - maxmin expected utility. Models that assume uniform ambiguity aversion or ambiguity seeking, by contrast, are incompatible with most of the patterns that we observed. 2. Matching probabilities and their properties 2.1. Notation and definitions We consider a decision maker who must make a choice in the face of uncertainty. Uncertainty is modeled using a state space S. The decision maker knows that exactly one state will obtain, but he does not know which one. Subsets E of S are called events, and E c denotes the complement of E. In our experiments, we used a partition of S into three events: E 1, E 2, and E 3, where 3/40

the unions of these events are denoted E 12, E 23, and E 13. Outcomes are money amounts, and more money is preferred to less money. The set of outcomes is equal to R. Positive amounts are gains, and negative amounts are losses. Acts are functions from the state space S to the outcome set R. We only consider binary acts, called bets. An example of a bet is x E 0, paying x if event E occurs and nothing otherwise. We also consider objective probabilities, denoting by x p 0 a bet that pays x with probability p and nothing otherwise. Let denote the decision maker s preference relation over bets. Strict preference ( ) and indifference (~) are defined as usual. For any event E, we can elicit its matching probability p such that x E 0 ~ x p 0. Let m(e,x) denote the matching probability of event E elicited by outcome x. This notation implies that matching probabilities may depend on x and therefore may differ between gains and losses. As shown below, the properties of matching probabilities depend on the assumed ambiguity model. As a benchmark, we assume probabilistic sophistication, under which matching probabilities are additive and sign-independent. 2.2. Benchmark: probabilistic sophistication Subjective expected utility holds if a bet x E 0 is evaluated by P(E)U(x), where U is a strictly increasing utility function 1, and P(E) is the subjective probability of event E. Probabilistic sophistication holds if there exists a probability measure P over S such that, for each bet x E 0, preference is determined entirely by the resulting probability distribution x p 0, where p = P(E). Probabilistic sophistication is more general than subjective expected utility: it keeps the probability measure P from expected utility but allows for different evaluations of probability distributions. 1 Throughout the paper, we set U(0) = 0. This is permitted by the uniqueness properties of U in all the theories that we consider. 4/40

Under probabilistic sophistication, m(e,x) = P(E) for all x, i.e., matching probabilities are independent of the outcome used, are unique for each event, and satisfy additivity (because P does). Additivity was challenged by Ellsberg s (1961) paradox. Ellsberg argued that, when faced with a known urn containing 50 red and 50 black balls and an unknown urn containing 100 balls in an unknown mixture of red and black, people would prefer to bet on the known urn than on the unknown urn, regardless of the winning color. This implies that the matching probabilities of red and black from the unknown urn are both less than 0.5. Their sum is therefore less than 1, violating additivity and, hence, probabilistic sophistication. Ellsberg s paradox, confirmed in many experiments (Camerer and Weber 1992), suggests that people prefer decision situations with known probabilities to decision situations in which probabilities are unknown. The absence of known probabilities is often called ambiguity, and dislike (like) of ambiguity is often called ambiguity aversion (seeking). In response to Ellsberg s paradox, new models have been proposed to explain ambiguity aversion. We describe several of these ambiguity models in the next section. To test for additivity, we define five indices, which are all equal to zero under probabilistic sophistication. The indices are independent: if any four of them are not equal to zero, the remaining one can still be zero (see the online appendix for a proof). The additivity indices provide insight into patterns of deviations from probabilistic sophistication. These patterns, in turn, provide information about the descriptive validity of ambiguity models. As we will show in the next section, the ambiguity models allow for specific deviations of the indices from zero. The first additivity index tests binary complementarity: for any event E, the probability of an event and that of its complement should sum to 1. We compute the indices BC s (E i ) = 1 m(e i,x) m(e c i,x), where s is the sign of x, and i = 1,2,3. Our experiments use only one gain (x = + 10) and one loss (x= 10), and the indices BC s (E i ) are therefore unique. Binary complementarity 5/40

is satisfied when these indices equal 0. Positive values correspond to ambiguity aversion for gains (as in the Ellsberg paradox) and to ambiguity seeking for losses. The next two indices are based on Tversky and Wakker (1995). We define a lower additivity index LA s (E i, E j ) as m(e i,x) + m(e j,x) m(e ij,x), where s is the sign of x, i,j {1,2,3}, and i j. That is, for any disjoint pair of events, we compute the difference between the sum of their matching probabilities and the matching probability of their union. Lower subadditivity [lower superadditivity] (Tversky and Wakker 1995) holds if the LA indices are positive [negative]. While the LA index measures the difference between adding an event to the empty event and adding it to a nonempty event, the upper additivity index measures the difference between removing an event from the universal event and removing it from a proper subset of the universal event: UA s (E i ) = [1 m(e ik,x)] [m(e ij,x) m(e i,x)], where s is the sign of x, i,j,k {1,2,3}, and i, j, and k differ. Because UA s (E i ) = [1 m(e ij,x)] [m(e ik,x) m(e i,x)], the upper additivity indices are unique for a given threefold partition of S. 2 Upper subadditivity [upper superadditivity] holds if the UA index is positive [negative]. If additivity holds, the matching probabilities m(e 1,x), m(e 2,x), and m(e 3,x) should sum to one. The fourth additivity index is based on this property and is defined as TA s = m(e 1,x) + m(e 2,x) + m(e 3,x) 1, where s is the sign of x. Because it involves a three event partition of the state space, we will refer to this index as ternary additivity. If additivity holds, then it should also be true that m(e 12,x) + m(e 13,x) + m(e 23,x) = 2*(m(E 1,x) + m(e 2,x) + m(e 3,x)) = 2. That is, the probability of 2 The notation is perhaps counterintuitive, as the reader may be inclined to call the first index, where event j is removed, UA s (E j ), and the second index, where event k is removed, UA s (E k ). However, this definition would make the indices nonunique. 6/40

the complements of E 1, E 2, and E 3 should sum to 2. We call this property indirect ternary additivity and construct the corresponding index as ITA s = 2 (m(e 12,x) + m(e 13,x) + m(e 23,x)). BC LA UA TA & ITA BC (E) = BC + (E) = 0 LA (Ei,Ej) = LA + (Ei,Ej) = 0 UA (E) = UA + (E) = 0 TA + = TA = ITA = ITA + = 0 Table 1: Predictions of expected utility and probabilistic sophistication 3. Ambiguity models and their predictions 3.1. Maxmin expected utility, -maxmin, and the variational model In maxmin expected utility and -maxmin, ambiguity is captured through a convex set C of priors. The set C is a subset of the set of all possible probability measures P on the state space S. For any event E, let I E denote the interval of probabilities for event E that the decision maker considers possible: I E = {P(E): P C}. For decision under risk, I E consists of a single point, the objective probability of the event. For any given bet f, let EU P (f) denote its expected utility (EU) when probability measure P is used. Maxmin EU (Gilboa and Schmeidler 1989) evaluates a bet f as inf P C EU P (f), the minimum of the expected utilities with respect to the probability measures included in C. Maxmin EU can account for differing matching probabilities for gains and losses if the set C is not a singleton. The indifference x E 0 ~ x p 0 implies that p = m(e,x) = inf(i E ) for gains and that p = m(e,x) = sup(i E ) for losses. The matching probability of the complementary event E c is 1 sup(i E ) for gains and 1 inf(i E ) for losses. Consequently, binary complementarity will be violated both for gains and losses, and m(e,x) + m(e c,x) 0 m(e, x) + m(e c, x) for all events E and gains x. Hence, BC (E) 0 BC + (E). Moreover, BC + (E) = BC (E). Table 2 shows the properties that can be derived for each of the five indices under maxmin EU. Proofs of all results presented in this Section are in the online appendix. We also show in the online appendix that the opposite model, 7/40

maxmax EU, which evaluates an act f as sup P C EU P (f), predicts all reversed inequalities. BC LA UA TA & ITA BC (E) 0 BC + (E) = BC (E) LA (E i,e j ) 0 LA + (E i,e j ) UA (E) 0 UA + (E) TA + = ITA 0 TA = ITA + Table 2: Predictions of maxmin EU A maxmin EU maximizer is extremely pessimistic in the sense that his choices are based on the minimum expected utilities with respect to the probability measures included in C. A less pessimistic model is -maxmin EU, in which the value of a bet is equal to the weighted average of the minimum and the maximum expected utility with respect to the probability measures in C (Ghirardato et al. 2004). Formally, -maxmin EU evaluates a bet f as inf P C EU P (f) + (1 )sup P C EU P (f). Under maxmin EU, the indifference x E 0 ~ x p 0 reveals that p = m(e,x) = inf(i E ) + (1 )sup(i E ) when x is a gain and p = m(e,x) = sup(i E ) + (1 )inf(i E ) when x is a loss. As in maxmin EU, the matching probabilities are sign-dependent; however, unlike maxmin EU, the inequality, m(e,x) + m(e c,x) m(e, x) + m(e c, x), x > 0, need no longer hold. Table 3 shows that the properties of our additivity indices under - maxmin EU depend on. Although it only adds one extra parameter to maxmin EU, -maxmin EU permits much richer preference patterns. For instance, it does not require that LA + (E i,e j ) and LA (E i,e j ) have opposite signs. The reason for this is that under -maxmin, the decision maker can be ambiguity-seeking for some events and ambiguity-averse for others. Maxmin EU, by contrast, only permits ambiguity aversion, and maxmax EU only permits ambiguity seeking. 8/40

BC LA UA TA & ITA α ⅓ TA = ITA + 0 ITA = TA +. BC (E) 0 BC + (E) α ½ BC (E) 0 BC + (E) α ½. LA + (E i,e j ) LA (E i,e j ) α ½ LA + (E i,e j ) LA (E i,e j ) α ½. UA + (E) UA (E) α ½ UA + (E) UA (E) α ½. ⅓ α ½ TA = ITA + ITA = TA +. ½ α ⅔ TA + = ITA ITA + = TA. BC (E) = BC + (E) α ⅔ TA + = ITA 0 ITA + = TA. Table 3: Predictions of -maxmin EU (with 0< <1) A limitation of both maxmin EU and -maxmin is their dichotomous nature: probability measures are either fully included or fully excluded from the set of priors C. A more realistic case is modeled by the variational model (Maccheroni et al. 2006), in which the degree to which probability measures are included is continuous. Let be the set of all probability measures on the state space S. The variational model evaluates the indifference x E 0 ~ x p 0 as min P (P(E)U(x)+c(P)) = pu(x) = m(e,x)u(x), where c: [0, ] is a convex function with infimum zero that reflects the decision maker s ambiguity aversion. Lower values of c indicate greater ambiguity aversion. Hansen and Sargent s (2001) multiplier preferences, axiomatized by Strzalecki (2011), are a special case of the variational model. The variational model is general and, as Table 4 shows, imposes no constraints on lower and upper additivity. It also relaxes the constraints on ternary additivity and indirect ternary additivity that maxmin and -maxmin impose. The variational model still requires that the BC, TA, and ITA indices have opposite signs for gains and losses, which makes it possible to test the model. 9/40

BC LA UA TA & ITA BC (E) 0 BC + (E) - - TA +, ITA 0 TA, ITA + Table 4: Predictions of the variational model 3.2. Choquet expected utility and prospect theory Under Choquet EU (Gilboa 1987, Schmeidler 1989), a bet x E 0 is evaluated as U(x), where is a decision weight that is equal to W(E) if x is a gain and to 1 W(E c ) if x is a loss. U is a strictly increasing utility function, and W is a weighting function that assigns a number between 0 and 1 to each event E, such that W( ) = 0, W(S) = 1, and E F implies W(E) W(F). The decision maker s attitude toward ambiguity is captured through the nonadditivity of W. Subjective expected utility is equal to Choquet EU if W is additive. For decision under risk, where probabilities of events are known, the weighting function W is replaced by a strictly increasing probability weighting function w, which satisfies w(0) = 0 and w(1) = 1. Under Choquet EU, the indifference x E 0 ~ x p 0 implies W(E) = w(p) = w(m(e,x)), if x is a gain, and W(E c ) = w(1 p) = w(1 m(e,x)), if x is a loss. Hence, the matching probabilities will generally be different for gains and losses. Table 5 presents the predictions of Choquet EU for our five additivity indices. The italicized conditions follow from the increasingness of w and W and impose only weak constraints. Unlike the other ambiguity models, Choquet EU predicts that violations of binary complementarity are the same for gains and losses. BC LA UA TA & ITA BC (E) = BC + (E). LA + (E i,e j ) + LA (E i,e j ) 1 Table 5: Predictions of Choquet EU UA + (E) + UA (E) 1 TA + = ITA TA = ITA + and TA + + ITA + 1 10/40

Prospect theory is more general than Choquet EU and allows for different weighting functions W and probability weighting functions w for gains (W +, w + ) and losses (W, w ). Under prospect theory, the indifference x E 0 ~ x p 0 implies that W + (E) = w + (p) if x is a gain and that W (E) = w (p) if x is a loss. Prospect theory imposes few restrictions on the matching probabilities, which can be very general, non-additive, and sign-dependent, and all types of sign-dependent violations are allowed. To test prospect theory, we added the restriction that decision makers are less sensitive to changes in likelihood under ambiguity than under risk (Wakker 2010, Abdellaoui et al., 2011). Formally, a decision maker exhibits ambiguity-generated insensitivity if, for any disjoint events E and F and for s = +,, (i) W s (E) = w s (p) and W s (F) = w s (q) imply that W s (E F) w s (p+q), provided that w s (p+q) is bounded away from 1, and (ii) W s (S E) = w s (1 p) and W s (S F) = w s (1 q) imply that W s ((E F) c ) w s (1 (p+q)), provided that W s (S (E F)) is bounded away from 0 (Tversky and Wakker, 1995). Ambiguity-generated insensitivity, which has been empirically observed (Abdellaoui et al. 2011, Maafi 2011, Baillon et al. 2012), is often associated with prospect theory. As Table 6 shows, it yields clear predictions of the directions in which four of the five additivity conditions are violated. The exception is binary complementarity, for which no testable implications can be derived. BC LA UA TA & ITA - LA + (Ei,Ej) 0 and LA (E i,e j ) 0 UA + (E) 0 and UA (E) 0 0 TA + + ITA + 1 and 0 ITA + TA 1 Table 6: Predictions of prospect theory with ambiguity-generated insensitivity 3.3. The smooth ambiguity model Klibanoff, Marinacci, and Mukerji s (2005) smooth ambiguity model evaluates the indifference x E 0 ~ x p 0 as (P(E)U(x))d = (pu(x)) = 11/40

(m(e,x)u(x)), where is the decision maker s subjective probability over the set of probability measures P on the state space S. The function captures the decision maker s attitudes toward ambiguity, with concavity [convexity] of corresponding to ambiguity aversion [seeking]. The intuition behind the smooth model is that ambiguity, reflected by, is evaluated separately from ambiguity aversion, reflected by. Table 7 shows that if is concave, the smooth model makes the same predictions as the variational model for our additivity indices. If is convex, these predictions are reversed. BC LA UA TA & ITA concave BC (E) 0 BC + (E) convex BC + (E) 0 BC (E) - - Table 7: Predictions of the smooth model concave TA +, ITA 0 TA, ITA + convex TA, ITA + 0 TA +, ITA 3.4. Vector expected utility Siniscalchi s (2009) vector EU model also separates ambiguity and ambiguity aversion. The model resembles Tversky and Kahneman s (1974) anchoring and adjustment heuristic (see Hogarth and Einhorn (1990) for a similar idea). The anchor is expected utility P(E)U(x), which the decision maker adjusts to account for his perception of ambiguity, using adjustment factors ζ i. Ambiguity aversion is captured through a negative and symmetric adjustment function A. Vector EU evaluates x E 0 ~ x p 0 as P(E)U(x) + A((ζ i (E) P(E)U(x)) 0 i<n ) = pu(x) = m(e,x)u(x). Table 8 shows that vector EU, with A negative (positive), makes the same predictions as the smooth model with concave (convex) for our additivity indices. 12/40

BC LA UA TA & ITA A negative BC (E) 0 BC + (E) A positive BC + (E) 0 BC (E) - - Table 8: Predictions of vector EU A negative TA +, ITA 0 TA, ITA + A positive TA, ITA + 0 TA +, ITA 4. Experiments We performed two experiments to test probabilistic sophistication and explore whether the predicted patterns of Tables 2-8 were indeed observed. We first present the designs and results of the two experiments and then, in Section 5, turn to their implications for the ambiguity models. 4.1 First experiment 4.1.1. Subjects and procedures Thirty-seven economics students (13 female) at Erasmus University Rotterdam took part in the experiment. Subjects received a flat-fee of 15 at the start of the experiment. In addition, each subject played out one choice for real. The experiment was computer-run, lasted approximately one hour, including instructions and payment, and was conducted in three group sessions. Subjects, seated in cubicles, were required to switch off their phones, and interaction was forbidden. Two interviewers were present during the group sessions. To prevent the second and third groups from having more information and facing less ambiguity than the first group, we made communication between the subject groups impossible: the first group of subjects was debriefed, while the second group participated in the experiment, and the third group entered the lab to receive instructions. We could not run more than 3 groups and prevent subjects from different groups speaking to each other 13/40

about the experiment. This and our desire to keep the group sizes manageable restricted the number of subjects in the experiment. As events, we used variations of two stock indices, the Dutch AEX index and the Indian SENSEX index, during the 30 minutes following the start of the experiment. We chose a more familiar source of uncertainty (the AEX index) and a more unfamiliar source (the SENSEX index) 3 because previous studies have shown that ambiguity attitudes can depend on the source of uncertainty (Fox and Tversky 1998, Kilka and Weber 2001, Abdellaoui et al. 2011). We defined the events as E 1 = [ 100%, 0.5%), in which the stock index falls by more than 0.5%, E 2 = [ 0.5%, 0.5%], in which the stock index varies by at most 0.5%, E 3 = (0.5%, + ), in which the stock index rises by more than 0.5%, and their complements E 23 = [ 0.5%, + ), E 13 = [ 100%, 0.5%) (0.5%, + ), and E 12 = [ 100%, 0.5%]. Subjects had to choose between two options, an uncertain bet on one of the six events, with either a gain or a loss of 10 if the event occurred, and a risky bet that produced this outcome (a gain or a loss of 10) with a given probability. Matching probabilities were measured in three steps. A first screen asked subjects to choose between the uncertain bet and the risky bet, where the risky bet involved probabilities of 0%, 10%, 20%,..., 90%, and 100% of winning or losing 10. The software enforced monotonicity. If, for example, a 70% probability of winning 10 was preferred to winning 10 if event E occurred, then a preference for the risky bet was automatically selected for probabilities of 80%, 90%, and 100%. If a subject s choice switched from the uncertain bet to the risky bet between probabilities q% and (q+10)%, a second screen then displayed choices between the two bets for probabilities q%, (q+1)%, (q+2)%,, and (q+10)%. Subjects chose between the two bets for each probability, with 3 None of our subjects was Indian, 20 were Dutch. 14/40

monotonicity enforced. If a subject s preference switched between probabilities q% and (q+1)%, we took (q+0.5)% as the matching probability for event E. For gains [losses], the third screen asked subjects to confirm that they indeed preferred the uncertain bet for all probabilities below [above] the elicited matching probability and the risky bet for all probabilities above [below] the elicited matching probability. The aim of this confirmation question was to reduce the impact of response errors. If subjects did not confirm their preferences, measurement of the matching probability was started anew. Examples of the three screens are presented in Figure 1. 15/40

Figure 1: Screenshots of the experiment 16/40

We elicited matching probabilities for the 6 events, E 1, E 2, E 3, E c 1 = E 23, E c 2 = E 13, and E c 3 = E 12, for both sources (AEX and SENSEX) and for a gain and a loss of 10. The order in which these 24 matching probabilities were elicited was random and differed between subjects. At the end of the experiment, we repeated the measurements of m(e 2, 10) and m(e 13, 10) for the SENSEX index to test for consistency. These measurements were the most difficult, involving both an unfamiliar index and losses, and, consequently, were most likely to lead to inconsistencies. To incentivize the experiment, each subject played out one choice for real. We used three boxes with numbered tickets to implement payments. The first box contained 26 tickets numbered 1 to 26 and was used to select one of the matching probability measurements. The second box contained 101 tickets numbered 0 to 100 and was used to select the probability of a gain or a loss in the risky bet. If the subject had chosen the uncertain bet for this probability, we used the actual variation of the stock index to determine the payoff. Otherwise, the subject randomly drew a number from a third box containing 100 tickets numbered 1 to 100 to play out the risky bet. 4 A random incentive system, in which a subset of the experimental questions is played out for real, is widely used in experimental economics. It assumes that subjects isolate each choice and do not view the experiment as a whole. The available evidence generally supports this assumption (see Bardsley et al. 2010, p.269). If subjects were to view the experiment as a whole, then they would face a mixture of bets. Following Schmeidler (1989), this would reduce their exposure to ambiguity, moving them in the direction of 4 To illustrate the procedure, suppose that the first box selected event E 1 for gains and the AEX. Suppose that 10 E1 0 ~ 10 47 0, i.e., the matching probability for event E 1 for gains and the AEX was 47%, and that the subject drew number 60 from the second box. Because the subject preferred 10 60 0 to 10 47 0 and hence to 10 E1 0, he was asked to draw a number from the third box. If this number was 60 or less, the subject received 10; otherwise, he received nothing. 17/40

probabilistic sophistication and expected utility. Our tests of probabilistic sophistication would then be conservative. We could not influence payments because the performances of the two stock indices were, of course, beyond our control. Hence, the common criticism of experiments on ambiguity, that subjects might think experimenters manipulate the experimental protocol to minimize payments, does not apply to our study. To exclude the possibility that subjects thought that we had assigned gains to unlikely events and losses to likely events, we explicitly told them that for every bet on an event, there would also be a bet on its complement. Differences between matching probabilities were tested using analysis of variance (with the Greenhouse-Geisser correction for repeated measures) and paired t-tests. 4.1.3 Consistency The consistency results were mixed. The two measurements of m(e 2, 10) for the SENSEX index did not differ significantly (p = 0.13), but the repeated measurement of m(e 13, 10) for the SENSEX index was lower than the original measurement (p = 0.04). 5 In contrast, the correlation between the original and the repeated measurement of m(e 13, 10) for the SENSEX index was substantial (0.74). 4.1.4. Aggregate results Figure 2 shows the 24 matching probabilities that we measured. Matching probabilities were sign-dependent (p = 0.001) and were higher for gains than for losses, particularly for more likely events. The matching 5 The difference may be due to an order effect, as the consistency tests were always performed at the end of the experiments and were not randomized, unlike the original measurements. 18/40

probabilities were similar for the AEX index and for the unfamiliar SENSEX index (p = 0.26). Figure 2: Mean matching probabilities A: AEX B: SENSEX Mean matching probability 60% Gain Loss Mean matching probability 60% Gain Loss 40% 40% 20% 20% 0% 0% E 1 E 2 E 3 E 12 E 13 E 23 E 1 E 2 E 3 E 12 E 13 E 23 Event Event We will now move to the analysis of the five additivity indices. Figure 3 shows that binary complementarity held for gains but not for losses. For gains, the sum of the probability of an event and the probability of its complement was close to 100% for all events and for both sources, and no significant differences were observed (all p > 0.10). For losses, the probability of an event and the probability of its complement summed to less than 100% for all three events and for both stock indices. For the AEX, there was binary superadditivity for losses for all events (all p < 0.01, one-tailed t-tests). For the SENSEX, there was binary superadditivity for losses for events E 2 (p = 0.03) and E 3 (p = 0.015), but the effect was only marginal for event E 1 (p = 0.06). Previous studies that used judged probabilities and found support for binary complementarity used only gains (Tversky and Koehler 1994, Tversky and Fox 1995, Wu and Gonzalez 1999). Our results confirm binary complementarity for gains but also show that binary complementarity is significantly violated for losses. 19/40

Binary complementarity is consistent with ambiguity neutrality for gains. This does not mean that subjects were always ambiguity neutral for gains. As we discuss below, the ITA indices showed ambiguity aversion for gains. In the Discussion Section, we return to the observed ambiguity neutrality for gains in the tests of binary complementarity and explain why this finding is consistent with a fourfold pattern of attitudes toward ambiguity. The binary complementarity index was lower for losses than for gains (p < 0.05 in all one-tail paired t-tests), but we could not reject the null that BC + = BC (which is predicted by maxmin EU and -maxmin EU). The difference between BC + and BC was marginally significant for event E 3 and the AEX (p = 0.06) and insignificant in all other tests (all p > 0.11). Figure 3: Binary complementarity A: AEX B: SENSEX Mean matching probability 100% P(E) gain P(E c ) gain P(E) loss P(E c ) loss Mean matching probability 100% P(E) gain P(E c ) gain P(E) loss P(E c ) loss 80% 80% 60% 60% 40% 40% 20% 20% 0% E 1 E 2 E 3 0% Event E 1 E 2 E 3 Event Figure 4 clearly shows that lower additivity was violated (all p < 0.01). The sums of the matching probabilities (the solid bars) exceeded the matching probabilities of their unions (the striped bars) in all tests, and all LA-indices were positive, which is consistent with lower subadditivity. The violations of lower additivity were similar for gains and losses. 20/40

Figure 4: Lower subadditivity A: AEX B: SENSEX Mean matching probability Sum gains Union gains Sum losses Union losses Mean matching probability Sum gains Union gains Sum losses Union losses 100% 100% 80% 80% 60% 60% 40% 40% 20% 20% 0% E 12 E 13 E 23 E 12 E 13 E 23 0% Gains Losses E 12 E 13 E 23 E 12 E 13 E 23 Gains Losses Upper additivity was also violated (see Figure 5), and all UA indices were positive, consistent with upper subadditivity (all p < 0.01). Upper subadditivity was stronger for losses than for gains (all p < 0.05). The combination of lower and upper subadditivity suggests that unlikely events are overweighted, that likely events are underweighted and that the underweighting of likely events is stronger for losses. A: AEX Figure 5: Upper subadditivity B: SENSEX Value Gain Loss Value Gain Loss 50% 50% 40% 40% 30% 30% 20% 20% 10% 10% 0% UA(E 1 ) UA(E 2 ) UA(E 3 ) 0% UA(E 1 ) UA(E 2 ) UA(E 3 ) 21/40

Finally, Figure 6 shows that the TA and ITA indices also indicated violations of additivity (all p < 0.01). The TA indices were all positive, which is consistent with ambiguity seeking for unlikely gains and ambiguity aversion for unlikely losses. They were similar for gains and losses, both for the AEX and the SENSEX. The ITA indices were lower for gains than for losses (all p < 0.01). This can be explained by more underweighting of likely events for losses than for gains and agrees with our finding in the analysis of the upper additivity indices. The positive values of the ITA-indices indicate ambiguity aversion for gains and ambiguity seeking for losses. For the AEX, the ITA index was also higher than the TA + and TA indices (both p < 0.05), which we will use to test the ambiguity models. Value 60% Figure 6: Violations of additivity in three-fold partitions A: AEX Value 60% B: SENSEX 50% 50% 40% 40% 30% 30% 20% 20% 10% 10% 0% TA TA ITA ITA Additivity index 0% TA TA ITA ITA Additivity index 4.1.5. Individual results The individual results confirmed the aggregate results. Table 9 lists the proportions of subjects who satisfied additivity, according to each of our five additivity indices. To account for response error, we classified subjects as satisfying additivity according to one of our indices if the absolute value of 22/40

that index was less than 0.10, the median of the absolute differences in the repeated measurements. Because we repeated the most complex questions, this error bound may be too wide, in which case, our tests are conservative and favor additivity. The proportion of subjects who satisfy additivity was small in most tests. Additivity was satisfied by about half of the subjects in the tests of binary complementarity and approximately 25% of the subjects in the other tests. Only one subject satisfied all five additivity conditions. AEX SENSEX BC + =BC =0 E 1 : 49% E 2 : 46% E 3 : 32% E 1 : 49% E 2 : 43% E 3 : 41% LA + =LA =0 E 12 : 27% E 2 : 24% E 3 : 24% E 1 : 24% E 2 : 24% E 3 : 22% UA + =UA =0 E 12 : 22% E 2 : 22% E 3 : 27% E 1 : 27% E 2 : 22% E 3 : 27% TA + =TA =0 16% 16% ITA + =ITA =0 19% 27% Table 9: Proportions of subjects satisfying additivity for each test and both for the AEX and the SENSEX 4.2. Second experiment 4.2.1. Subjects and procedures The second experiment tested the robustness of the results of the first experiment, using a larger sample. Ninety-eight students of Erasmus University Rotterdam participated. To obtain more subjects than in the first experiment, we ran 5 sessions and therefore could not fully avoid the possibility of communication among subjects about the experiment, although this was unlikely. The second experiment was largely similar to the first experiment, and we will only list the differences. We replaced the SENSEX index by the South African TOP40 index because the Indian stock market was closed during part of the experiment. Because we observed quotations with a 15-minutes delay, subjects bet on variations of the stock indices between 15 minutes before and 23/40

15 minutes after the beginnings of the sessions they attended. We used E 1 = [ 100%, 0.2%), E 2 = [ 0.2%, 0.2%], and E 3 = (0.2%, + ) to reduce the likelihood of E 2. 6 Unlike in the first experiment, the two questions that were repeated did not come at the end of the experiment; rather, the order of the questions was fully random. The instructions appeared on the computer screen (they were read aloud in the first experiment), and we added three practice questions and understanding checks. Subjects could only start the experiment if they had correctly answered these checks. 4.2.2 Consistency Consistency was good and better than in the first experiment. The original and repeated measurements of m(e 2, 10) and m(e 13, 10) for the TOP40 index did not differ significantly (both p > 0.11) and were substantially correlated (both > 0.57, p < 0.01). 4.2.3. Aggregate results The results of the second experiment were similar to those of the first. Figure 7 shows that the elicited matching probabilities were sign-dependent (all p < 0.01) and higher for gains than for losses, particularly for more likely events. This result is consistent with greater ambiguity-generated insensitivity for losses. As in the first experiment, we found no evidence of sourcedependence: the matching probabilities were similar for the AEX and the TOP40 (all p > 0.69). 6 In the first experiment, event E 2 was objectively very likely. For instance, between March 5 and April 1, 2014, it occurred 94.6% and 96.5% of the time for the AEX and SENSEX, respectively (based on the variations of the indices between 9 a.m. and 10 a.m., 10 a.m. and 11 a.m until 4 p.m. and 5 p.m.). In the second experiment, we reduced the likelihood of E 2. E 1, E 2, and E 3, as defined in the second experiment, occurred 9.1%, 81.2%, and 9.7% of the time, respectively, between March 5 and April 1, 2014 for the AEX (for the TOP40 these data were not accessible). 24/40

Mean matching probability 60% Gain Loss Figure 7: Mean matching probabilities A: AEX Mean matching probability 60% B: TOP40 Gain Loss 40% 40% 20% 20% 0% E 1 E 2 E 3 E 12 E 13 E 23 Event 0% E 1 E 2 E 3 E 12 E 13 E 23 Event Figure 8 shows the results for binary complementarity. In contrast with the first experiment, binary complementarity did not hold for gains, but we found binary superadditivity. The sum of the probability of an event and that of its complement exceeded 100% for all events for both stock indices (all p < 0.05, except for E 2 in the TOP40, where p = 0.06). For losses, we found binary subadditivity for all three events for both stock indices (all p < 0.01), as in the first experiment. These findings are consistent with ambiguity seeking for both gains and losses. The binary complementarity index for losses exceeded the index for gains (all p < 0.01). We could not reject the null hypothesis that the absolute values of the binary complementarity indices for gains and losses were equal (all p > 0.25). 25/40

Figure 8: Binary complementarity A: AEX B: TOP40 Mean matching probability 100% P(Event) gain P(Complement) gain P(Event) loss P(Complement) loss Mean matching probability 100% P(Event) gain P(Complement) gain P(Event) loss P(Complement) loss 80% 80% 60% 60% 40% 40% 20% 20% 0% E 1 E 2 E 3 0% Event E 1 E 2 E 3 Event Figures 9 and 10 confirm lower and upper subadditivity (all p < 0.01) for both gains and losses, consistent with ambiguity-generated insensitivity. As in the first experiment, the LA indices did not differ between gains and losses, but the UA indices did (all p < 0.01). This suggests more ambiguitygenerated insensitivity for losses than for gains. Figure 9: Lower subadditivity A: AEX B: TOP40 Mean matching probability 100% Sum gains Union gains Sum losses Union losses Mean matching probability 100% Sum gains Union gains Sum losses Union losses 80% 80% 60% 60% 40% 40% 20% 20% 0% E 12 E 13 E 23 E 12 E 13 E 23 Gains Losses 0% E 12 E 13 E 23 E 12 E 13 E 23 Gains Losses 26/40

Figure 10: Upper subadditivity A: AEX B: TOP40 Value 50% Gain Loss Value 50% Gain Loss 40% 40% 30% 30% 20% 20% 10% 10% 0% UA(E 1 ) UA(E 2 ) UA(E 3 ) 0% UA(E 1 ) UA(E 2 ) UA(E 3 ) Figure 11 shows that both direct and indirect ternary additivity were violated (all p < 0.01). For both the AEX and the TOP40, the TA index was lower than the TA + and the ITA indices (all p < 0.01). The ITA + index was lower than the ITA index for both sources (both p < 0.01) and lower than the TA + index for the TOP40 only (p < 0.05). The positive values of the ITAindices are consistent with ambiguity aversion for gains and ambiguity seeking for losses. Figure 11: Violations of additivity in three-fold partitions Value 60% A: AEX Value 60% B: TOP40 50% 50% 40% 40% 30% 30% 20% 20% 10% 10% 0% TA TA ITA ITA Additivity index 0% TA TA ITA ITA Additivity index 27/40

4.2.4. Individual results AEX TOP40 BC + =BC =0 E 1 : 52% E 2 : 45% E 3 : 53% E 1 : 45% E 2 : 37% E 3 : 54% LA + =LA =0 E 12 : 26% E 2 : 22% E 3 : 29% E 1 : 30% E 2 : 28% E 3 : 28% UA + =UA =0 E 12 : 27% E 2 : 21% E 3 : 29% E 1 : 26% E 2 : 30% E 3 : 26% TA + =TA =0 21% 23% ITA + =ITA =0 30% 18% Table 10: Proportions of subjects satisfying additivity for each test and for both the AEX and SENSEX. Second experiment. Table 10 shows the proportions of subjects who satisfied additivity, according to each of the five indices. These proportions were almost identical to those in the first experiment. Few subjects satisfied additivity, although we allowed for a (wide) margin of error of plus or minus 0.10. We used 0.10 to have the same error margin as in the first experiment, even though the median absolute difference between the original and repeated measurement was only 0.07 in the second experiment. Despite this, only 3 subjects satisfied additivity in all five tests (allowing for response error). 5. Predictions of the ambiguity models. Next, we turn to predictions of the ambiguity models. The deviations from additivity show that subjects tended to overweight unlikely events, underweight likely events, and display differing ambiguity attitudes for gains and losses. Taken together, these findings lead to a fourfold pattern of ambiguity attitudes: ambiguity seeking for unlikely gains and likely losses and ambiguity aversion for likely gains and unlikely losses. This result suggests that ambiguity models that permit such a rich pattern of ambiguity attitudes are probably more consistent with our data than more restrictive models. Models that predict uniform ambiguity attitudes indeed performed poorly. Aggregate behavior violated the predictions of maxmin expected utility for BC, BC +, UA, LA +, ITA, and TA +. It also violated the predictions of maxmax EU for UA +, LA, ITA +, and TA and those of the variational 28/40

model for BC, BC +, ITA, and TA +. The signs of BC, ITA, and TA + violate the smooth model with concave, and the smooth model with convex is inconsistent with the signs of ITA + and TA. Allowing to switch from convex (ambiguity seeking) for losses to concave (ambiguity averse) for gains cannot accommodate the signs of the TA indices. Vector EU suffers from the same problems. Choquet EU and -maxmin EU (with between ⅓ and ½) allow for the overweighting of unlikely events and the underweighting of likely events and can explain most of the data. However, these models cannot explain the difference between ITA and TA +. Additionally, Choquet EU wrongly predicts BC (E) = BC + (E). An explanation of the difference between ITA and TA + requires sign-dependent ambiguity attitudes. Prospect theory can explain all the aggregate data and correctly predicts that all the UA, LA, TA, and ITA indices should be positive. If we allow in -maxmin to differ for gains and losses, then -maxmin can also explain all data patterns. Table B.1 in the online appendix shows how many subjects behaved according to each of the predictions of the various ambiguity models. In constructing this table, we again used an error margin of 10% to account for response error. Table 11 shows, for each ambiguity model, how many subjects satisfied all of its predictions. 29/40

Model # Conditions Proportion of subjects Experiment 1 probabilistic sophistication 4 3% 3% maxmin EU 4 3% 3% maxmax EU 4 5% 5% variational model 2 14% 8% α-maxmin EU ⅔ α < 1 4 5% 6% ½ α ⅔ 4 11% 18% ⅓ α ½ 4 8% 20% 0 < α ⅓ 4 5% 10% all cases 4 14% 28% Experiment 2 vector EU & A negative / 2 14% 8% smooth model concave A positive / 2 16% 22% convex Both 2 27% 27% Choquet EU 4 22% 41% prospect theory with ambiguitygenerated insensitivity 3 70% 92% Table 11: Proportion of subjects satisfying all predictions of each model Table 11 confirms that few subjects behaved according to models with uniform ambiguity attitudes. Models that permit both ambiguity aversion and ambiguity seeking performed better, even though they can explain the behavior of only a minority of subjects. The only theory that can explain the choices of most subjects is prospect theory: 70% in the first experiment and 92% in the second experiment behaved according to prospect theory with ambiguity-generated insensitivity. The superior performance of prospect theory illustrates the importance of allowing for sign-dependence in modeling ambiguity attitudes. 30/40

6. Discussion Our subjects violated probabilistic sophistication. Matching probabilities were sign-dependent, additivity did not hold, and violations of additivity differed for gains and losses. Unlikely events were overweighted, and likely events were underweighted. The underweighting of likely events was stronger for losses, indicating greater ambiguity-generated insensitivity for losses. Our findings suggest a fourfold pattern of attitudes toward ambiguity: ambiguity seeking for unlikely gains and likely losses and ambiguity aversion for likely gains and unlikely losses. Attitudes toward ambiguity are richer than uniform ambiguity aversion, which is often assumed in the literature, suggests. As early as 1962, Ellsberg remarked that ambiguity seeking might prevail for unlikely events (Ellsberg, 2001, p.203). Our results corroborate Ellberg s conjecture. The fourfold pattern of ambiguity attitudes can explain why we observed no evidence of ambiguity aversion in tests of binary complementarity for gains. The tests of binary complementarity combined a less likely event with its more likely complement. If the overweighting of the less likely event is approximately offset by the underweighting of the more likely (complementary) event then, on balance, the data will indicate ambiguity neutrality. Because likely events were underweighted more for losses, this argument can also explain why we found binary subadditivity (ambiguity seeking) for losses. In the ITA tests, which employed three likely events, we found, consistent with the underweighting of likely events, clear evidence of ambiguity aversion for gains (and ambiguity seeking for losses). There is a rich and still growing literature on modeling ambiguity attitudes. Unfortunately, empirical evidence on the relative performance of the different ambiguity models is scarce. Our data show that models that predict uniform ambiguity aversion and do not distinguish between gains and losses do not describe ambiguity attitudes well. Models that perform well are 31/40

Choquet EU, -maxmin and, in particular, prospect theory. At the individual level, prospect theory clearly outperforms the other models. Our findings are largely consistent with other tests of ambiguity models. Kothiyal, Spinu, and Wakker (2013) found support for prospect theory over Choquet EU and the multiple priors models in a re-analysis of the data of Hey, Lotito, and Maffioletti (2010). Chew, Miao, and Zhong (2013) report data that are consistent with Choquet EU (and with prospect theory, as they only consider gains) but not with the multiple priors models, the smooth ambiguity model, and vector EU. Their study, however, did not directly test predictions of Choquet EU and prospect theory. Yang and Yao (2012) obtained evidence against the multiple priors models and the smooth ambiguity model. Hayashi and Wada (2011) found evidence against maxmin EU and -maxmin. On the other hand, Cubitt et al. (2013) found support for the smooth model over -maxmin. Finally, Machina (2009) presented several Ellsberg-like examples that were confirmed experimentally by l Haridon and Placido (2010). Baillon, l Haridon, and Placido (2011) showed that these examples violate the maxmin EU, the variational model, Choquet EU, and the smooth model but not vector EU. Our analysis differs from these studies in that we test more models, test precise predictions of each of these models, consider both gains and losses, employ simple choices, and use a natural source of ambiguity. Our analysis includes many but not all ambiguity models. Some recent ambiguity models are too general to generate predictions for our data (e.g., Cerreia-Vioglio et al. 2011, Ergin and Gul 2009). Gajdos et al. (2008) proposed a model in which decision makers receive imprecise information. To apply their model, it is crucial to know the received information set. As the latter was unclear in our experiment, we could not test their model. Other models that were not included in our analysis are those of Neilson (2010) and Chateauneuf and Faro (2009). 32/40