Skip to content
Publicly Available Published by De Gruyter August 26, 2020

Ambiguity and Awareness: A Coherent Multiple Priors Model

  • Simon Grant EMAIL logo , Ani Guerdjikova and John Quiggin

Abstract

Ambiguity in the ordinary language sense means that available information is open to multiple interpretations. We model this by assuming that individuals are unaware of some possibilities relevant to the outcome of their decisions and that multiple probabilities may arise over an individual’s subjective state space depending on which of these possibilities are realized. We formalize a notion of coherent multiple priors and derive a representation result that with full awareness corresponds to the usual unique (Bayesian) prior but with less than full awareness generates multiple priors. When information is received with no change in awareness, each element of the set of priors is updated in the standard Bayesian fashion (that is, full Bayesian updating). An increase in awareness, however, leads to an expansion of the individual’s subjective state and (in general) a contraction in the set of priors under consideration.

JEL Classification: D81

1 Introduction

The idea that choices under uncertainty are subject to problems arising from ambiguity was first put forward by Ellsberg (1961), drawing on the earlier work of Knight (2006). Like Knight, Ellsberg argued that, in many cases, decisionmakers do not, and could not be expected to, act as if they assigned well-defined numerical probabilities to all the possible outcomes of a given choice. His well-known thought experiments illustrating this argument formed the basis of a large subsequent literature both theoretical and empirical.

In most of this literature, the term “ambiguity” has been treated as a synonym for what Knight called “uncertainty” namely the fact that relative likelihoods are not characterized by well-defined numerical probabilities. The standard method of dealing with ambiguity in decision theory is to endow the decisionmaker with multiple priors as in Gilboa and Schmeidler (1989). This approach may be combined with a variety of preference models, notably including the maxmin model of Gilboa and Schmeidler (1989) and the smooth model of Klibanoff, Marinacci, and Mukerji (2005).

For a non-specialist this is puzzling; there is no obvious link to the ordinary meaning (or meanings[1]) of ambiguity as a characteristic of propositions with more than one interpretation. In its normal usage, ambiguity is a linguistic concept, but in decision theory it is typically treated as a property of preferences.

The now-standard usage is quite different from that in Ellsberg’s (1961) original article. Ellsberg treated ambiguity, not as a property of preferences or relative likelihoods, but as a property of the information on which judgments of relative likelihoods might be based.

“Responses from confessed violators [of the expected utility (EU) axioms] indicate that the difference is not to be found in terms of the two factors commonly used to determine a choice situation, the relative desirability of the possible payoffs and the relative likelihood of the events affecting them, but in a third dimension of the problem of choice: the nature of one’s information concerning the relative likelihood of events. What is at issue might be called the ambiguity of this information, a quality depending on the amount, type, reliability and ‘unanimity’ of information, and giving rise to one’s degree of ‘confidence’ in an estimate of relative likelihoods.”

In this paper, we argue that informational ambiguity, in the ordinary language sense that the available information is open to multiple interpretation, may be modeled using concepts from the literature on unawareness. When individuals are unaware of some possibilities relevant to the outcome of their decisions, there are multiple probability distributions that may be applicable, depending on whether or not these possibilities are realized.

To represent this idea, we adopt a syntactic representation, in which the state of the world is characterized by the truth values of a finite set of elementary propositions P. The state space Ω is given by the set of all logically possible combinations of truth values, that is, by the truth table for P. Events in Ω correspond to the truth values of compound propositions in P.

An unboundedly rational decision-maker is aware of all the propositions in P, and in the logical closure of P, and therefore has access to a complete and unambiguous description of the state space Ω . If her unconditional and conditional preferences over acts (mappings from the state space to a set of consequences) conform to expected utility theory, then it is as if the decision-maker can assign a unique subjective probability π to any event E, and update that probability in line with Bayes rule as new information is received.

In contrast, we represent a boundedly rational decision-maker as one who is unaware of at least some propositions in P. For simplicity, consider the case when an agent is aware of a proposition p, but not of a related proposition q . In this situation, the proposition p is ambiguous since it may mean either p q or p ¬ q . From the agent’s viewpoint, her information about p is incomplete, since it is open to multiple interpretations.[2]

In this paper, we formalize this idea to derive a coherent multiple priors (CMP) model. Our goals are twofold. First, we derive a representation theorem for the CMP model and show that, with full awareness, it corresponds to the usual Bayesian model. Second, we consider the problem of updating beliefs. In our setting, updating may arise in response to the receipt of new information or to increased awareness, represented as awareness of new elementary propositions p. When information is received with no change in awareness, each element of the set of priors is updated in the standard Bayesian fashion as in Ghirardato, Maccheroni, and Marinacci (2008). An increase in awareness is represented by an expansion of the state space to which the decision maker has access, and by a corresponding contraction in the set of priors under consideration. As the decisionmaker approaches full awareness, the set of priors contracts to a singleton { π } . Relative to π the set of priors at any time t is made of conditional probabilities, depending on the truth values of propositions of which the decisionmaker is unaware.

1.1 Roadmap

The paper is organized as follows. We begin with a motivating example, involving a policy-maker’s response to an epidemic disease outbreak. This example illustrates the relationship between bounded awareness and ambiguous beliefs and explains the notion of coherence.

We next set up the description of the decision-making problem in both propositional (syntactic) and state-space (semantic) terms. Awareness, information and acts are defined. Under full awareness, individuals have access to a set of propositions P such that the associated truth table 2 P encompasses all relevant distinctions between states of the world. Boundedly aware individuals have access to a more limited set of propositions A which gives rise to a coarser state space S A , as well as an associated ‘complementary’ space of unawareness states, with each ‘unawareness state’ corresponding to an assignment of truth values for the propositions of which the individual is unaware.

Next we consider preferences and ambiguity. We restate the Ghirardato, Maccheroni, and Marinacci (2004) axioms. The crucial result of this section is to show that preferences satisfying the Ghirardato, Maccheroni, and Marinacci (2004) axioms may be derived from the preferences of a fully aware EU-maximizer, by introducing unawareness.

As in Ghirardato, Maccheroni, and Marinacci (2004), the preference over acts on S A has a sub-relation which is incomplete and obeys the Independence axiom. This sub-relation may be represented by a unique, closed convex set of priors: an act is preferred over another by this sub-relation if it yields greater expected utility for every prior in the representation. Hence, a DM who is unaware of some of the propositions is endowed with an ambiguous preference relation which is coherent (in a sense that will be made precise) with the expected-utility preferences of a fully aware DM with the same utility function and the appropriate unique prior.

Our new key assumption, Axiom 6, states that if an act is conditionally preferable to another by each conditional preference relation obtained by conditioning the complete expected utility preference of a fully aware agent on each ‘unawareness state’, then it is unambiguously preferred.

We next consider updating in response to increases in information and awareness. We represent updating as a two-stage process. First, in even periods, the DM gets to know an event (in the part of the state space she is aware of). Second, in odd periods the DM becomes aware of proprositions she had not previously considered.

For changes in information with constant awareness, we show that the preferences we derive display prior-by-prior Bayesian updating, as in Ghirardato, Maccheroni, and Marinacci (2008) and Pires (2002). To convey the intuition for changes in awareness, we first address the simplest case where the individual becomes aware of a single additional proposition. We show that the result is to expand the state space, dividing each existing state into two new states, one in which the newly discovered proposition is true and the other in which it is false. Conversely, any pair of priors conditioned on events that differ only on the truth value of the new proposition is replaced by a convex combination of the two. In the finite setting we have here, the state space doubles in size, while the set of priors halves. This result is shown to hold more generally for any changes in awareness consistent with our model structure.

In Section 7, we discuss important links between the concept of bounded awareness used here and problems arising in econometric and statistical theory associated with concept such as ‘latent variables’ and ‘unobserved heterogeneity’ and the techniques for their estimation as developed by Dempster, Laird, and Rubin (1977). The crucial idea is that the relationship between observed variables of interest may be influenced by unobserved variables. In some cases, these variables are known on theoretical grounds to be relevant but cannot be measured. A variety of classical and Bayesian methods may be applied to this case. More generally, any trained statistician understands that relationships estimated on the basis of a given data set may be rendered unreliable by the omission of relevant variables, without necessarily being aware of which variables might be relevant in a particular case. This may be seen as an example of “awareness of unawareness”.

Section 8 relates our work to the existing literature. Finally, in Section 9, we offer some concluding comments.

2 Illustrative Example

Consider the regrettably topical problem of developing a public health response to an epidemic disease outbreak. The disease has only recently been discovered in the country in question, but has already had severe impacts elsewhere.

The possible options include a low-cost campaign, focusing on basic hygiene measures such as hand-washing and a high-cost response involving putting the entire population into some form of quarantine. To simplify, we will assume that the high-cost option is guaranteed to control the pandemic if applied sufficiently early before the number of cases reaches some critical proportion of the population. The policy-maker (hereafter, PM) has sufficient data to estimate the probability that the number of cases is below the critical level, which we will denote by r. For the purpose of the numerical exercise below, we will set r = 4 / 5 .

The success or failure of the low-cost option depends on a range of factors, only some of which the PM is fully aware of. Consider the proposition

q = “the low-cost option will contain the pandemic” and its negation ¬ q = “the low-cost option will result in uncontained spread”

The PM is aware that the success of the low-cost option will depend on the extent of voluntary compliance, but is not explicitly aware of other relevant factors. To make this more precise, she is aware of the propositions: p 1 = “voluntary compliance will be high”. Hence, the initial set of propositions of which the decision maker is aware is A = { q , p 1 } . This defines the relevant state-space: S A = 2 | A | = { ( q , p 1 ) , ( q , ¬ p 1 ) , ( ¬ q , p 1 ) , ( ¬ q , ¬ p 1 ) }

Given knowledge about the relevant population, the PM can form unambiguous statements about the probability of p 1 being true, that is whether the population is likely to comply with her advice. Given the current state of knowledge about the disease, no further information is available.

The PM understands that there are other factors relevant to q, and therefore that information about q is ambiguous in the sense described by Ellsberg. In these circumstances, some advocates of evidence-based medicine point to randomized controlled trials (RCT) as the ‘gold standard’ for assessing interventions.[3]

Ideally, an RCT would yield precise estimates of the proportions π ( q | p 1 ) and π ( q | ¬ p 1 ) of programs that succeed in high compliance and low-compliance populations respectively. However, there is no guarantee that RCTs conducted on different populations will yield the same results, even if these populations are similar with respect to compliance. On the contrary, it is possible that results differ too much to be consistent with the hypothesis that low-cost interventions in populations with similar compliance probabilities have the same probability of success.

Suppose, thus, that individual RCTs conducted in different countries have yielded probability estimates π k ( q | p 1 ) [ π ¯ ( p 1 ) , π ¯ ( p 1 ) ] and π k ( q | ¬ p 1 ) [ π ¯ ( q | ¬ p 1 ) , π ¯ ( q | ¬ p 1 ) ] and that the differences are too large to be explained by chance variation. As Cowan (2020) observes:

“A positive result for treatment against control in a randomized controlled trial shows you that an intervention worked in one place, at one time for one set of patients but not why and whether to expect it to work again in a different context. Evidence based medicine proponents try to solve this problem by synthesizing the results of RCTs from many different contexts, often to derive some average effect size that makes a treatment expected to work overall or typically. The problem is that, without background knowledge of what determined the effect of an intervention, there is little warrant to be confident that this average effect will apply in new circumstances. Without understanding the mechanism of action, or what we call a theory of change, such inferences rely purely on induction.

“The opposite problem is also present. An intervention that works for some specific people or in some specific circumstances might look unpromising when it is tested in a variety of cases where it does not work. It might not work ‘on average’. But that does not mean it is ineffective when the mechanism is fit to solve a particular problem such as a pandemic situation. Insistence on a narrow notion of evidence will mean missing these interventions in favor of ones that work marginally in a broad range of cases where the answer is not as important or relevant.”

In these circumstances, the PM is cognizant of the fact that there are differences between the populations of the various countries that result in differing rates of success for the low-cost intervention, but is unaware what these differences might be.[4] We may suppose that some of the RCT populations match that of the country in question.

How can the PM reason about the probability that the treatment will be successful for a population with known compliance rates?

One possible answer, noted by Cowen, is to impute the mean probability π ( q | p ) (respectively, π ( q | ¬ p ) ) averaged across RCT studies of populations with high (low) compliance. Given that the country is a member of a particular, yet undefined, sub-class, this answer is wrong with probability 1, although the direction and magnitude of the error cannot be determined.

An alternative might be to set up a higher-order probability model. That is, the PM might impute a subjective probability w k to the proposition ‘the relevant characteristics of my country are most similar to those of the patient population of country k’. This yields, for compliant populations, the success probability π * ( q | p 1 ) = k π k ( q | p 1 ) w k . Some version of this subjective approach would be required of an agent satisfying the Savage axioms. Yet the solution is obviously problematic. While the success rates for each country, π k ( q | p 1 ) , are objectively defensible, the choices of w k are just guesses. The resulting weighted average π * ( q | p 1 ) cannot be justified to the public, or to another policy-maker whose guesses w k are different.

The third response to unawareness is to say that the success probability lies in the interval [ π ¯ ( p 1 ) , π ¯ ( p 1 ) ] . With this response, bounded awareness implies ambiguous probability beliefs.

The idea that probabilistic judgments apply to large populations but not to individual cases (in our case, that of individual countries) may be traced back to Knight. It is eloquently expressed by Mukerjee (2015) as applied to medical research:

“We have invented many rules to understand normalcy - but we still lack a deeper, more uniform understanding of physiology and pathology. This is true for even the most common and extensively studied diseases - cancer, heart disease, and diabetes. … Why do certain immune diseases cluster together in some people, while others have only one variant? Why do patients with some neurological diseases such as Parkinson, have a reduced risk on cancer?

“These ‘outlying’ questions are the Mars problems of medicine … ”

In all these questions there is “awareness” of several different outcomes, but “unawareness” of what the cause of these different outcomes might be, and how they apply in particular cases.

Suppose indeed, that the reality in our epidemic example is more complicated: whether or not a low-cost program succeeds depends not only on the population compliance, p 1 but also on whether household size is small, p 2 . What is more, these two factors are interrelated.

In particular, traditional societies with large families are more inclined to follow directives from established authorities. Suppose no other factors are relevant.

Thus the set of propositions P = ( q , p 1 , p 2 ) defines an eight element “small world” in the sense of Savage (1954).

The following table gives the joint probability of the validity of the relevant propositions, p 1 and p 2 for the population in question.

For later reference, Pr ( p 1 | p 2 ) = 3 / 11 , ( Pr p 1 | ¬ p 2 ) = 9 / 13 .

Given the interplay between p 1 , p 2 and q explained above, we specify the following conditional probabilities of q given the truth realizations of p 1  and p 2 :[5]

This implies that the joint probability distribution of the truth values of q, p 1 and p 2 is:

and the joint (marginal) probability distribution of q, p 1 is:

A fully aware PM can thus state the unconditional probability of q as Pr { q } = 23 / 48 .

In this context, learning the truth value of p 2 generates a partition

1 = { { ( q , p 1 , p 2 ) , ( q , ¬ p 1 , p 2 ) , ( ¬ q , p 1 , p 2 ) , ( ¬ q , ¬ p 1 , p 2 ) } , { ( q , p 1 , ¬ p 2 ) , ( ¬ q , p 1 , ¬ p 2 ) , ( q , ¬ p 1 , ¬ p 2 ) , ( ¬ q , ¬ p 1 , ¬ p 2 ) } }

and the updated conditional probabilities are calculated as usual, e.g.,

Pr { q , p 1 | p 2 } = Pr { ( q , p 1 , p 2 ) } Pr { ( q , p 1 , p 2 ) } + Pr { ( q , ¬ p 1 , p 2 ) } + Pr { ( ¬ q , p 1 , p 2 ) } + Pr { ( ¬ q , ¬ p 1 , p 2 ) } .

Hence, for each realization of the truth value of p 2 , we obtain the probability distribution on the state space S A :

(5) Pr { q , p 1 | p 2 } = 3 / 11 Pr { q , p 1 | ¬ p 2 } = 9 / 26 Pr { ¬ q , p 1 | p 2 } = 0 Pr { ¬ q , p 1 | ¬ p 2 } = 9 / 26 Pr { q , ¬ p 1 | p 2 } = 4 / 11 Pr { q , ¬ p 1 | ¬ p 2 } = 0 { ¬ q , ¬ p 1 | p 2 } = 4 / 11 Pr { ¬ q , ¬ p 1 | ¬ p 2 } = 4 / 13

Now consider a partially aware PM. We suppose that the PM entertains two priors over S A , corresponding to the possible truth values of p 2 . These priors represent the PM’s perception of the ambiguity of the information she has available concerning S A .

We will be particularly concerned with the case of coherent priors. Coherent priors correspond to the conditional probabilities the agent would assign if she were actually aware of the truth values. The underlying idea is that, while the PM cannot articulate p 2 , her beliefs on S A correctly reflect the beliefs she would hold if she were fully aware. That is, each row in (5) defines an interval of probabilities a PM may entertain about the occurrence of a state in S A .

In our example, the probability distribution in the right column in (5) corresponds to a data set from an RCT study conducted in a country with small households whereas the left column would correspond to RCT data generated in a country with large households.

Thus, with k { 1 , 2 } , where p 2 is true for country 1 and ¬ p 2 holds for country 2, we would have

π 1 ( q | p 1 ) = 1 , π 1 ( q | ¬ p 1 ) = 1 2

π 2 ( q | p 1 ) = 1 2 , π 2 ( q | ¬ p 1 ) = 0

Realistically, for most countries, the proportions of households would be between the extreme values, giving rise to a convex set of probabilities.

The range of possible probabilities over S A is given by an interval in [ 0 , 1 ] 4 with extreme points given by the two conditional distributions ( 3 / 11 , 0 , 4 / 11 , 4 / 11 ) and ( 9 / 26 , 9 / 26 , 0 , 4 / 13 ) identified in (5).

Upon learning the truth value of p 1 , that is, the compliance characteristics of the population, and using full Bayesian updating on the set of priors, the beliefs over the value of q reduce to:

where the extreme points of the interval in each case reflect the two distributions conditional on the truth of p 2 presented in Table 2.

The resulting convex set of priors is illustrated in Figure 1. The upper red line shows the set of priors for the pair ( Pr ( p 1 ) , Pr ( q | p 1 ) ) . Point A represents the prior belief derived from the implicit assumption of p 2 . Point B represents the prior belief derived from the implicit assumption of ¬ p 2 .

Figure 1: 
Convex sets of priors for illustrative example.
Figure 1:

Convex sets of priors for illustrative example.

Similarly, the lower blue line represents the set of priors for the pair ( Pr ( p 1 ) , Pr ( q | ¬ p 1 ) ) . Point A represents the prior belief conditional on the implicit assumption of p 2 . Point B represents the prior belief conditional on the implicit assumption of ¬ p 2 .

The information represented by Points A and A is sufficient to represent the prior belief over the state space { ( q , p 1 ) , ( q , ¬ p 1 ) , ( ¬ q , p 1 ) , ( ¬ q , ¬ p 1 ) } derived from the implicit assumption of p 2 . Similarly, the information represented by Points B and B is sufficient to represent the prior belief derived from the implicit assumption of ¬ p 2 .

Now suppose our PM becomes aware of p 2 and assigns the “correct” prior probability 11 / 24 to p 2 being true. (If the available data is representative, this prior could also be inferred from the data, once the truth value of p 2 has been measured for each observation). Now the two priors on S A , ( 3 / 11 , 0 , 4 / 11 , 4 / 11 ) and ( 9 / 26 , 9 / 26 , 0 , 4 / 13 ) , are collapsed into one, namely that given by (4):

Pr ( p 2 ) π 1 ( q | p 1 ) + Pr ( ¬ p 2 ) π 2 ( q | p 1 )

= 11 24 ( 3 11 , 0 , 4 11 , 4 11 ) + 13 24 ( 9 26 , 9 26 , 0 , 4 13 ) = ( 5 16 , 3 16 , 1 6 , 1 3 )

and the interval convex hull shrinks to a single point.

Note that the so-obtained probability distribution will coincide with π if and only if the relative frequency of p 2 observed in the RCTs, coincides with the actual probability of p 2 for the population in question.

In Figure 1, upon becoming aware of p 2 , the pairs ( A , B ) and ( A , B ) are replaced by the single priors C and C respectively, where C = λ A + ( 1 λ ) B with λ = Pr ( p 2 | p 1 ) , and similarly C = λ A + ( 1 λ ) B with λ = Pr ( p 2 | ¬ p 1 ) . The vertical axis values of C and C are the prior conditional probabilities Pr ( q | p 1 ) and Pr ( q | ¬ p 1 ) , taking account of available information about p 2 .[6]

As we will show below, under appropriate conditions on beliefs and preferences, this coherence result holds in general. Becoming aware of some new propositions, while remaining unaware of others, and updating each element of the set of coherent priors according to Bayes rule yields a set of posteriors which is the same as the set of coherent priors obtained by beginning with expanded awareness and information, and then deriving priors as above. Moreover, the derivation above shows that coherence is, in principle, testable. Once the PM is fully aware, a probability distribution over Ω can be elicited, and this is sufficient to derive the set of priors associated with any state of partial awareness and any information set.

Now consider the choice faced by the PM, between the low-cost and high-cost responses. We may consider three possible outcomes: an uncontrolled pandemic, a successful high cost response and a successful low-cost response. We will associate these outcomes with payoffs 0, 1 / 2 , and 1.

The high-cost option α h generates an unambiguous lottery yielding 1 / 2 with probability r = 4 / 5 and 0 with probability 1 r = 1 / 5 . The expected utility of this lottery is 2 / 5 .

Under full awareness, the low cost option α also generates an unambiguous lottery yielding 1 with probability 23/48 and 0 with probability 25/48.

By contrast, a PM unaware of p 2 entertains the two priors ( 3 / 11 , 0 , 4 / 11 , 4 / 11 ) and ( 9 / 26 , 9 / 26 , 0 , 4 / 13 ) .

If the PM is uncertain about the degree of voluntary compliance of the population (the validity of p 1 ), the MMEU of α is determined using the minimal probability of success for the low-cost option, Pr { q } = min { 7 / 11 , 9 / 26 } = 9 / 26 .

With these parameter values

9 / 26 < 2 / 5 < 23 / 48 .

That is, under full awareness the low-cost option α will be preferred to the high cost option, but under partial awareness with MMEU preferences, the high-cost option α h will be preferred. The PM, realizing that there are relevant factors of which she is yet unaware, might thus prefer to choose a costlier and more aggressive course of action, which would however help control the pandemic with a known probability r, rather than facing the worst possible scenario, in which the unknown factor drives the probability of success to its lowest possible value 9 / 26 . This can be interpreted as an instance of the precautionary principle in face of unawareness.

3 Setup

3.1 The State of the World: Propositional and State-space Descriptions

Information and awareness evolve over time. However, we will initially consider an individual with fixed information and awareness, suppressing time subscripts.

The world is described by the truth values of a finite set of elementary propositions P = { p 1 , , p N } . Compound propositions formed by conjunction and negation of elementary propositions are denoted by boldface p .

Individuals have bounded awareness, represented by a set A P of elementary propositions which they can express. Awareness and information are mutually dependent. On the one hand, as will be described in more detail below, the individual’s awareness depends on the information they have. On the other hand, that information must be expressed in terms of propositions p A P expressible by the individual.

This propositional description of the world may be represented equivalently in state space terms more familiar to decision theorists. The state space associated with the truth table for P may be represented by Ω = 2 N with ω a representative element/state.[7] Let Σ denote the field of subsets of Ω with elements E.

Similarly, for an individual with awareness A the state space of which she is aware can be expressed as

S A = 2 A ,

with a generic element s A and field of subsets Σ A with elements E A , referred to as events. Note that each event E A corresponds to the truth of a compound proposition p made up of elementary propositions in P A .

Clearly, each state ω Ω uniquely identifies a state s A S A by its projection to S A denoted by ω S A . Conversely, each s A S A corresponds to an event in Ω , that is, the set of states ω whose projection to S A is exactly s A : E s A = { ω | ω S A = s A } . Such events are measurable with respect to Σ A . More generally, an event E Ω is measurable with respect to Σ A iff

(7) ω E implies ω E for all ω Ω with ω S A = ω S A .

We define a probability as a mapping π : Σ A [ 0 , 1 ] with the usual properties, and define the conditional probability operator π ( | E ) according to Bayes rule

π ( E | E ) = π ( E E ) π ( E )

whenever π ( E ) > 0 .

Note that a probability measure π on Σ induces a unique probability measure π ˜ on Σ A by setting

π ˜ ( E ) = π { ω | ω S A E } .

Let A ¯ = P \ A be the set of propositions of which the individual is unaware. The “complementary state space” of which she is unaware can be expressed as S ¯ A = 2 A ¯ , with generic element s ¯ A . Notice that Ω = S A × S ¯ A .

Furthermore, each (awareness) state s A S A corresponds to the event { s A } × S ¯ A in Ω and each s ¯ A S ¯ A corresponds to the event S A × { s ¯ A } in Ω . That is, awareness of the form described above leads to a “coarsening” of the state space as, for example, in Quiggin (2016), represented by the projection of Ω onto S A . In the case of full awareness, S ¯ A must be a singleton and similarly for S A in the case of maximal unawareness.

Unawareness in this sense may be distinguished from the case of “reduction” or “restriction” of the state space, in which some possible elements of Ω are disregarded or, equivalently, in which some propositions that are possibly true are implicitly assumed to be false. This leads to the possibility of “surprise”, see, for example, Grant and Quiggin (2015).

3.2 Acts

Acts will be represented formally in the usual Anscombe-Aumann framework. To understand how this framework is applicable in situations of differential awareness, it is useful to recall that each state and event in Ω corresponds to the truth value of a (typically compound) proposition.

In order to focus on beliefs rather than preferences over outcomes, we concentrate on a set comprising just two final consequences Z = { 0 , 1 } .[8] Hence, each simple act may be expressed in propositional terms as a bet of the form “If p , receive 1, otherwise receive 0”. The extension to general outcome sets is straightforward but adds complexity without additional insight.

Let Δ ( Z ) denote the set of all lotteries on Z, with a generic element denoted by x [ 0 , 1 ] , yielding the (“good”) outcome 1 with probability x and the (“bad”) outcome 0 with complementary probability 1 x .

An act α maps Ω to Δ ( Z ) . The set of acts available under full awareness is denoted A , and is equal to Ω Δ ( Z ) . Let C denote the set of all constant acts. Let denote the set of “bets (on events)”, that is, α if and only if α ( ω ) { 0 , 1 } for all ω Ω .

The outcomes of acts considered by an individual with limited awareness must be conditional on propositions of which the individual is aware. Hence, for given awareness A P , any act α must be measurable with respect to Σ A , and we denote by A ( A ) the subset of such acts. Formally:

A ( A ) = { α : Ω [ 0 , 1 ] | α ( ω ) = α ( ω ) for all ω , ω with ω S A = ω S A } .

Clearly, an agent with limited awareness cannot perceive such mappings. Hence, the set of acts available to an agent with awareness A is given by:

A A = { α : S A [ 0 , 1 ] } .

Note that there is a one-to-one correspondence between the acts in A ( A ) and those in A A , but that the mappings defining these acts have different domains ( Ω vs. S A ). Analogously, C A denotes the set of constant acts A A and A denotes the set of bets in A A .

Each act in A A induces a mapping of elements of S A into lotteries on Z. In particular, for any s A an act α specifies the probability of obtaining 1, which with slight abuse of notation we shall denote by α ( s A ) .

As is standard, convex mixtures of acts are defined as state-by-state probability mixtures: for α and α A , and λ ( 0 , 1 ) , we define λ α + ( 1 λ ) α by

( λ α + ( 1 λ ) α ) ( ω ) = λ α ( ω ) + ( 1 λ ) α ( ω )

for all ω Ω . Convex mixtures in A A are defined analogously with respect to the state space S A .

Example 1.

Suppose that an agents initial awareness consists of a single proposition, A = { p 1 } , and that the agent later becomes aware of the full set of propositions P = { p 1 , , p N } . In the initial condition of partial awareness, the state space is S { p 1 } = S A = { p 1 , ¬ p 1 } , and an act may be represented as an ordered pair of consequences: ( α ( p 1 ) , α ( ¬ p 1 ) ) . There are only four simple acts available to the agent: the two constant acts ( 0 , 0 ) and ( 1 , 1 ) the bet on p 1 ( 1 , 0 ) and the bet against p 1 , ( 0 , 1 ) . The set of acts A A = [ 0 , 1 ] 2 consists of convex mixtures over these simple acts. Trivially, each of these acts corresponds to an act in Ω measurable with respect to Σ A . For example, the bet on p 1 can be represented as α ( ω ) = 1 if p 1 is true in ω and α ( ω ) = 0 if ¬ p 1 is true in ω. The logic of this simple example is fully general.

4 Preferences and Ambiguity

For each level of awareness A P , we define preferences on A A by . We will denote = P preferences under full awareness. That is, we consider an agent with a family of preference orderings over state spaces S A , one for each A.

4.1 The GMM Approach

We first impose the Ghirardato, Maccheroni, and Marinacci (2004) (hereafter, GMM) axioms: for each A A , we assume the preference relation A satisfies:

Axiom 1

Completeness and transitivity

Axiom 2

Archimedean axiom: for all α, α , α A A , if α A α A α , then there are λ and μ ( 0 , 1 ) such that

λ α + ( 1 λ ) α A α A μ α + ( 1 μ ) α .

Axiom 3

( a ) Certainty independence: for α, α A A and α ¯ C A , α A α iff λ α + ( 1 λ ) α ¯ A λ α + ( 1 λ ) α ¯ for all λ [ 0 , 1 ] .

( b ) Independence under full awareness: if A = P , then for any α, α , α A , α P α iff λ α + ( 1 λ ) α P λ α + ( 1 λ ) α for all λ [ 0,1 ] .

Axiom 4

Monotonicity: for α, α A A if α ( s A ) α ( s A ) for all s A S A , then α A α .

Axiom 5

Non-degeneracy: there are α and α A A such that α A α .

Axiom 2 is purely technical, Axiom 4 is intuitive and uncontroversial, while Axiom 5 ensures the setting and the analysis that follows is not vacuous. Axiom 1 is of more interest. The standard completeness axiom, adopted by GMM, is very demanding, since it requires agents to have complete preferences over all conceivable acts with respect to a state space which may be arbitrarily large. Since we are restricting the axiom to apply to acts expressed in terms of propositions of which the agent is explicitly aware, our Axiom 1 is consistent with reasonable bounds on cognitive capacity. The same reasoning extends to transitivity.

For the case of partial awareness A P , Axiom 3 is less restrictive than the standard independence axiom (as in GMM) except in the special case of full awareness, for which independence does hold. This implies that in our setting partial awareness is the only source of deviations from expected utility.

Definition 1

For α, α A A , α is A -unambiguously preferred to α , if

(8) λ α + ( 1 λ ) α A λ α + ( 1 λ ) α for all λ [ 0 , 1 ] and all α A A .

Whenever (8) is satisfied for two acts α and α , we write α A α′. As in GMM, the relationship A is a preorder satisfying all the Anscombe-Aumann axioms (except completeness), and so admits a representation as given in the Lemma below, GMM, Proposition 5, p. 144. For the case of full awareness, this relationship is complete.

Lemma 1

Axioms A1-A5 imply for each A the existence of a unique convex and (weak*) closed set of priors Π A such that for any two acts α and α A A

(9) α A α i f f s A S A π ( s A ) α ( s A ) s A S A π ( s A ) α ( s A ) for all   π Π A .

In particular, Π P is a singleton { π } .

Remark 1

Just as in Anscombe-Aumann, for the case of full awareness, Axiom 3 ( b ) allows us to define conditional preferences and thus, identify conditional beliefs π * ( | E ) which satisfy Bayes rule.

Since each constant act can be identified with the probability x [ 0 , 1 ] , with which it results in the outcome 1 in every state, we write α x C . As in GMM (pp. 153-4), for α A A , define the interval of possible certainty equivalents:

C E ( α ) = { x [ 0 , 1 ]   | for any y [ 0 , 1 ] ,   α y A α implies y x and   α A α y   implies x y } .

GMM show that x C E * ( α ) iff

min π Π A s A S A π ( s A ) α ( s A ) x max π Π A s A S A π ( s A ) α ( s A ) .

Interpreting constant acts x as utilities, and noting that we have restricted utilities to the unit interval, C E ( α ) corresponds to the set of utilities for α consistent with some π Π A . In the case of full awareness, C E * ( α ) is a singleton.

4.2 Unawareness and Ambiguity

The preferences described in the previous section are normally interpreted in terms of ambiguity. Given our setup, there is a natural interpretation in terms of awareness.

Consider EU preferences over A , , described by the full awareness probability distribution π derived using Axiom 3 ( b ) in the previous section.[9] In this setting, conditional preferences can be defined as usual.

Definition 2

For α ^ , α ^ A , α ^ is preferred to α ^ conditional on event E Σ , denoted α ^ E α ^ iff α ˜ α ˜ holds for some (and thus, by Axiom 3 ( b ) , for all) α ˜ , α ˜ A with α ˜ ( s ) = α ^ ( s ) , α ˜ ( s ) = α ^ ( s ) for all s E and α ˜ ( s ) = α ˜ ( s ) for all s E .

Remark 2

The so-defined conditional preferences under full awareness can be shown to satisfy the same Axioms 15 as above.

For any s ¯ A S ¯ A , that is, for any set of truth values for the propositions outside of a given set A P (which describes a possible level of awareness), π induces a conditional probability distribution π ( | s ¯ A ) over S A . Correspondingly, the induced conditional preferences s ¯ A over A ( A ) , are given by α ^ s ¯ A α ^ iff

s A S A π ( s A | s ¯ A ) α ^ ( s A ) s A S A π ( s A | s ¯ A ) α ^ ( s A ) .

Now consider the family of preferences ( A ) A which satisfy Axioms 1–5. The following axiom relates the preference relation at a level of awareness A to the conditional preferences under full awareness.

Axiom 6

Unanimity: for α, α A A , α A α if and only if for every s ¯ A S ¯ A and every α ^ s ¯ A , α ^ s ¯ A A with α ^ s ¯ A ( s × s ¯ A ) = α ( s ) and α ^ s ¯ A ( s × s ¯ A ) = α ( s ) for each s S A and α ^ s ¯ A ( s × s ˜ A ) = α ^ s ¯ A ( s × s ˜ A ) for all s S A and s ˜ A S ¯ A \ s ¯ A , a ^ s ¯ A a ^ s ¯ A .

Note that this condition can be equivalently rewritten in terms of the conditional preferences in Definition 2. Recall that to each act α A A paying off on the state-space S A , there corresponds an act α ^ A ( A ) paying off on the state-space Ω and measurable with respect to Σ A such that α ^ ( s ˜ ) = α ( s ) for each each s ˜ E s and each s S A . Then Axiom 6 says that for any two acts α, α A A , α is unambiguously preferred to α , α A α , if and only if for α ^ , α ^ A ( A ) with α ^ ( s ˜ ) = α ( s ) and α ^ ( s ˜ ) = α ( s ) for each s ˜ E s and each s S A , α ^ is preferred to a ^ conditionally on any state s ¯ A S ¯ A , α ^ s ¯ A α ^ . Recalling that α ^ and α ^ are measurable with respect to S A , the only effect of s ¯ A is to determine the conditional probability distribution π * ( | s ¯ A ) . So, we are evaluating α and α with respect to a set of probability distributions. Axiom 6 says that if α is preferable with respect to each such distribution, then it must be preferred unambiguously.

This property may be viewed as a version of the sure-thing principle. Unambiguous preferences of the type α A α here correspond to comparisons for which existing evidence is uncontroversial, i.e., for which the truth realizations of the yet unknown propositions and thus of the states in A ¯ all yield the same preference ranking for a given choice. Axiom 6 then requires a certain consistency when the preferences of a decision maker in a state in which he is not yet aware of A ¯ are revised upon becoming fully aware of all propositions in A ¯ . If it is the case that the realizations of the uknown states were irrelevant for the comparison between α and α ex-ante, then this should also be true ex-post for every such possible realization, i.e., conditional on any possible truth value of the propositions in A ¯ . Vice versa, if there is a truth value specification for the propositions in A ¯ for which the comparison between α and α is reversed, representative and accurate empirical evidence would reflect this, even if the decision maker is yet unaware of the cause for such a reversal. Thus, the decision maker would not have unambiguous preference for α versus α at awareness level A.

Since Axiom 6 uses the conditional preferences derived from the full awareness preference relation for each level of awareness A, its effect is to tie all the preference relations A to the fully aware preferences and thereby to each other.

4.3 The Awareness-based Multiple Priors Model

We now establish the claim that unawareness, as we have defined it, naturally lends itself to a multiple priors model to represent preferences.

Definition 3

Multiple priors preferences are awareness-based if for each A P , the set of probabilities Π A derived in Lemma 1 satisfies:

Π A = C H ¯ { π * ( | s ¯ A ) | s ¯ A S ¯ A } ,

where C H ¯ stands for the convex hull of a set, and π is a probability as described in Section 1.

Clearly, for awareness-based preferences, under full awareness, Π is a singleton (recall that in this case S ¯ A is a singleton).

Given these definitions we have the following implication of the axioms, the proof of which is relegated to the Appendix:

Proposition 2

Under axioms A1–A6, the family of preferences ( A ) A are awareness-based.

Notice that we do not expect the converse to hold. As we shall show, in the absence of changes in awareness, the probabilities Π derived as conditional distributions based on unawareness follow Bayesian updating in response to the arrival of new information. Axioms 1–6 are insufficient to ensure this – we need additional properties as discussed by Ghirardato et al. (2008). These additional properties give rise to a coherent multiple priors model.

5 Time, Information, Awareness and Histories

We now consider changes in information as well as awareness over time, and the induced changes in beliefs and preferences. Time t = 0 , 1 , 2 , , T is discrete and finite.

Information is formally modeled by partitions: t denotes a partition of Ω at time t. As with acts, understanding is assisted by considering the interpretation in terms of propositions. Each element of a partition corresponds to the truth value of a compound proposition, and the set of such propositions must be exhaustive and mutually exclusive. Information consists of learning that precisely one of these propositions is true.

The element of t that obtains at time t is denoted f t t . The collection = { t } t = 0 T constitutes a filtration. That is, for each t = 0 , , T 1 , t + 1 is a refinement of t , or equivalently, if f t + 1 t + 1 then f t + 1 f t for some f t t . We write ρ t ( f ) for the element of the partition at time t which contains f Ω (provided such an element exists). That is,

ρ t ( f ) = { f t t | f f t }

In particular, for f t + 1 t + 1 , ρ t ( f t + 1 ) is the immediate predecessor of f t + 1 .

We write σ t ( f t ) = { f ˜ t t | ρ t ( f ˜ t ) = f t } for the set of successors of f t at time t > t . In particular, σ t + 1 ( f t ) is the set of immediate successors of f t .

Without loss of generality, we will assume that non-trivial new information arrives only at odd periods, that is, | σ t + 1 ( f t ) | > 1 only if t = 2 k for some k 0 . For even periods, t = 2 k + 1 for some k 0 , we have t + 1 = t and no new information is revealed.

We assume that no uncertainty is resolved at date t = 0 , that is, 0 = { Ω } and all uncertainty is resolved by date T, so that for each ω Ω , { ω } T .[10]

The explanation above shows that information must be measurable with respect to the awareness of the agent. An agent cannot learn the truth of a proposition of which she is unaware. It is possible to become aware of a proposition because its truth becomes evident. Next, we introduce the joint dynamics of information and awareness.

To do so, we associate with each pair ( t , f ) such that f t , an awareness level A ( t , f ) . We impose the following restrictions on the awareness structure defined by A ( t , f ) :

Definition 4

An awareness structure { A ( t ) : t 2 P \ { } } t = 0 T is admissible relative to an information partition if:

  1. Awareness is non-empty, increasing and can only change in even periods: A ( t , f ) A ( t + 1 , σ t + 1 ( f ) ) for all t and A ( t + 1 , σ t + 1 ( f ) ) = A ( t , f ) if t = 2 k for some k 0 .

  2. Information is measurable with respect to awareness: for an even period t = 2 k for some k 0 and a given f t , each f t + 1 σ t + 1 ( f t ) is measurable with respect to Σ A ( t , f ) as defined in (7).

We will assume that at time T, the decision maker is fully aware: A ( T , f ) = P for all f T .

Remark 3

The fact that for even t = 2 k , σ t + 1 ( f t ) is measurable with respect to the awareness level at t, A ( t , f t ) = A implies that f t itself is measurable with respect to Σ A . We write f t A for the projection of f t to A, whenever measurability is satisfied.

Definition 5

An information structure { t } t = 0 T and an admissible awareness structure

{ A ( t ) : t 2 P \ { } } t = 0 T give rise to a set of histories

= { h = ( t , f , A ) | t { 0 , 1... T } ,   f t ,   A = A ( t , f ) } .

For a given h = ( t , f , A ) , the set of one-step-ahead histories is:

+ 1 h = { h + 1 = ( t + 1 , f , A ) | f σ t + 1 ( f ) , A = A ( t + 1 , f ) }

Definition 6

The awareness-adapted history h A corresponding to h = ( t , f , A ) is given by:

h A = ( t , f A , A )

where f A is the projection of f on Σ A . For t = 2 k , k , and h = ( t , f , A ) , the set of one-step-ahead awareness-adapted histories is:

+ 1 h , A = { h + 1 A | h + 1 + 1 h } .

For t = 2 k , and a history h = ( t , f , A ) , we recall that both f and every f σ t + 1 ( f ) are measurable with respect to Σ A . Thus, for every such h and the corresponding awareness-adapted history h A , the set of one-step-ahead awareness-adapted histories, i.e., histories expressible in terms of propositions in A is well-defined.

Remark 4

For a given information structure and an admissible awareness structure

{ A ( t ) : t 2 P \ { } } t = 0 T , consider two consecutive periods t ^ = 2 k , and t ^ + 1 = 2 k + 1 . Setting ^ 0 = { f } , ^ 1 = σ t + 1 ( f ) , and the corresponding awareness structure to A ^ ( 0 , f ) = A ( t , f ) = A ( t + 1 , f ) = A for all f σ t + 1 ( f ) results in an information structure together with an awareness structure admissible with respect to ^ . For h 0 = ( 0 , f , A ) , the corresponding set of one-step-ahead awareness-adapted histories is given exactly by:

1 h 0 , A = { h 1 A = ( 1 , f , A ) | f σ t + 1 ( f ) }

Note that up to renumbering of the periods, these histories correspond to the one-step-ahead awareness adapted histories starting at h = ( t , f , A ) under the original partition structure and awareness structure A ( t , f ) :

+ 1 h , A = { h + 1 A | h + 1 = ( t + 1 , f , A ) ,   f σ t + 1 ( f ) } .

5.1 Conditional Preferences over Acts

For a given history h = ( t , f , A ) , the acts of which the decision-maker is aware at h are those measurable with respect to her awareness level A, that is,

A A = { α : S A [ 0 , 1 ] } .

As usual, the information f available to the decision maker at history h allows us to define conditional preferences.

In particular, for t = 2 k and h = ( t , f , A ) , let Σ + 1 h , A be the σ-algebra (the set of all subsets) of all one-step-ahead awareness adapted histories. Preferences on A A conditional on an event H Σ + 1 h , A are written H A . Such conditional preferences reflect the fact that for a given awareness level, the arrival of information that the true state is in H will in general change the beliefs of the decision maker and thus, the evaluation of each of the available acts.

In contrast, for t = 2 k + 1 and h = ( t , f , A ) , the direct successor of h is h = ( t + 1 , f , A ) , where A A with strict inclusion being the non-trivial case. The transition between h and h is a change in awareness level, the information f remaining unchanged. Thus, an agent with conditional preferences h A on A A = { α : S A [ 0 , 1 ] } will adjust her preferences to the new awareness level, that is, to h A on A A = { α : S A [ 0 , 1 ] } .

These two adjustment processes will obey different principles. The former will incorporate new information in a deductive, Bayesian way, leaving existing ambiguity unchanged. The latter will incorporate newly learned states into the model by expanding the state-space and simultaneously reducing the level of perceived ambiguity, that is, the set of priors.

We now describe our desideratum, the coherent multiple prior model, which explains how these principles can be combined in a coherent way.

5.2 The Coherent Multiple Prior Model

The following definition of the coherent multiple prior model combines the intuition presented so far. For a probability distribution π Δ ( Ω ) , denote by π h the Bayesian update of π conditional on history h.

Definition 7

For a given information structure { t } t = 0 T and an admissible awareness structure { A ( t ) : t 2 P \ { } } t = 0 T , let be the set of corresponding histories. The beliefs of an agent satisfy the coherent multiple prior model (CMP) if there exists a probability distribution π Δ ( Ω ) and for each h = ( t , f , A ) a set of priors Π h Δ ( S A ) s.t.

(i) for t = 0 , and h 0 = ( 0 , Ω , A 0 = A ( 0 , Ω ) ) ,

Π h 0 = C H ¯ { π * ( | s ¯ A 0 ) | s ¯ A 0 S ¯ A 0 }

(ii) for any t = 2 k , k , and any h + 1 = ( t + 1 , f , A ) + 1 h , with f σ t + 1 ( f )

Π h + 1 = { π h + 1 A | π Π h }

(iii) for any t = 2 k + 1 , k 0 and the direct successor of h, h = ( t + 1 , f , A ) ,

Π h = C H ¯ { π h ( | s ¯ A ) | s ¯ A S ¯ A } .

Part (i) states that for the trivial history, h 0 , the model mimics the static awareness-based multiple prior model presented above. That is, there exists a probability distribution π of the fully aware agent, which, when applied to the situation with partial awareness A 0 results in multiple priors π ( | s ¯ A 0 ) , one for each of the states of which the agent with awareness level A 0 is unaware.

Part (ii) is an analog of the generalized Bayesian updating rule for multiple priors introduced by Ghirardato, Maccheroni, and Marinacci (2008) and is applied to those periods, in which new information arrives.

Finally, Part (iii) is the inductive extension of the static awareness-based multiple prior model: whenever awareness increases, the set of priors is redefined on a larger set of states and the number of priors is reduced to the number of those states S ¯ A , of which the agent is still unaware. In particular, this part of the definition implies that as long as information remains unchanged, the sequence in which awareness is updated does not matter for the final set of priors.

6 Axiomatizing the Coherent Multiple Prior Model

In this section, we provide axiomatic foundations for the coherent multiple prior model. We already provided a condition, Axiom 6, that relates beliefs under full awareness to beliefs under partial awareness and thus, establishes property ( i ) of the coherent multiple prior model. This axiom will now be extended to hold at all conceivable histories, which will allow us to establish property ( i i i ) . In contrast, to obtain generalized Bayesian updating upon arrival of new information, we will make use of the axiomatization of this updating rule in Ghirardato, Maccheroni, and Marinacci (2008).

6.1 Updating with Constant Awareness

In Definition 2, we defined conditional preferences for the case of full awareness. Here, we first extend this definition to arbitrary levels of awareness and provide a characterization of multiple prior updating with partial, but constant awareness. In a second step, we incorporate changes in awareness and characterize beliefs satisfying the coherent multiple prior representation.

For the remainder of the subsection, we will focus on the process of belief updating between two consecutive periods t = 2 k , and t + 1 = 2 k + 1 . As explained in Remark 4, for a given h = ( t , f , A ) , setting ^ 0 = { f } , ^ 1 = σ t + 1 ( f ) , and the corresponding awareness structure to A ^ ( 0 , f ) = A ( t , f ) = A ( t + 1 , f ) = A for all f σ t + 1 ( f ) results in an information structure together with an admissible awareness structure. For h 0 = ( 0 , f , A ) , the corresponding set of awareness-adapted histories is given exactly by:

+ 1 h , A = { h 1 A | h 1 = ( 1 , f , A ) , f σ t + 1 ( f ) }

Let Σ + 1 h , A denote the σ-algebra on such histories and define for each H Σ + 1 h , A the conditional preferences H A on A A . Recall that the unconditional preferences are given by h A .

For a fully aware agent, conditional preferences are defined on the one-step-ahead histories

+ 1 h = { h 1 = ( 1 , f , A = P ) | f σ t + 1 ( f ) }

In fact, for a fully aware agent, awareness can no longer change and thus, only the information partition matters. Thus a history h = ( 1 , f , A = P ) can be uniquely identified by f. The relevant events are given by the algebra generated by ^ 1 , Σ ^ 1 . The conditional preferences of a fully aware agent are defined using Definition 2 for each E Σ ^ 1 .

We are, in a first step, interested in the change in beliefs of an agent who observes and reacts to changes in information while keeping awareness constant at A. In order to ensure that such beliefs are well-defined we require in the following all information events to be non-null, regardless of the relevant awareness level. For an act α A A , let α 1 ( 1 ) (respectively, α 1 ( 0 ) ) denote the event in S A , on which α pays 1 (respectively 0).

Axiom 7

Non-null (information-)events: Let h = ( t , f , A ) and the corresponding ^ 1 and + 1 h , A be defined as above:,

(a) For any H Σ + 1 h , A , for the bets α, α 0 A A defined as α 1 ( 1 ) = H and α 0 1 ( 0 ) = H , we have α H A α 0 .

(b) For any E Σ ^ 1 , for the bets α and α 0 defined as α 1 ( 1 ) = E and α 0 1 ( 0 ) = E we have α E α 0 .

The first part of the axiom ensures that an agent with awareness A at a given history h considers every information event H corresponding to an element of the information partition ^ 1 adapted to her level of awareness, A as non-null. The second part ensures the same property for the agent who observes the same history h, but is aware of all possible propositions. In particular, since T = { { ω } ω Ω } , applying Axiom 7 ( b ) to histories at T 1 ensures that each state ω is considered non-null by the fully aware agent.

First consider the case of a fully-aware individual. As usual, Axiom 3 ( b ) implies Bayesian updating at every possible history h = ( t , f , P ) : a fully aware individual, A = P , with prior probabilities π replaces them by the conditional at history h, π h = π ( | f ) upon learning f. Under Axiom 7 ( b ) , conditional preferences are not-trivial and thus, every history and thus, every f is not null with respect to the preferences . It follows that the Bayesian updating process is always well-defined.

Extending this reasoning to all possible subsets of histories, E measurable with respect to the information partition , Axioms 1–3 ( b ) , 4, 5 and 7 ( b ) imply:

Lemma 3

The class of fully-aware preferences ( E ) E Σ t ,   t { 0... T } has an expected utility representation with a prior given by π and a system of conditional beliefs ( π E ) E which satisfy Bayesian updating, i.e., π E = π ( | E ) for every E Σ t , t { 0... T } .

In particular, the so-defined system of beliefs uniquely identifies, for each history h = ( t , f , A ) , beliefs conditional on events of the type[11] ( h , s ¯ A ) , π h ( | s ¯ A ) .

We next turn to an agent who at history h = ( f , t , A ) has awareness level A. We assume that for each H Σ + 1 h , A , conditional preferences H A over awareness adapted acts A A satisfy the same set of axioms, 1–3 ( a ) , 4, 5, as the unconditional preferences A . Adding 7 ( a ) , implies that for each H, H A is non-degenerate. Unambiguous preferences * H A are defined in analogy to A . Thus, since A remains constant, for any H, we can associate with the preference H A a set of probability distributions Π H on S A such that for α, α A A ,

α H A α iff  s A S A π ( s A ) α ( s A ) s A S A π ( s A ) α ( s A )  for all π Π H .

The next two axioms suggested by Ghirardato, Maccheroni, and Marinacci (2008) impose the desired structure on Π H across all possible measurable events H Σ + 1 h , A .

Axiom 8

Consequentialism: For a given h = ( t , f , A ) and H Σ + 1 h , A , if α and α A A such that α ( h ˜ ) = α ( h ˜ ) for all h ˜ H , α H A α .

Axiom 9

Dynamic consistency of H A : For a given h = ( t , f , A ) and H Σ + 1 h , A , and any acts α and α A A such that α ( h ˜ ) = α ( h ˜ ) for all h ˜ H , α h A α iff α H A α .

Intuitively, Axiom 8, Consequentialism, requires that acts which have identical payoffs on any element of a set H are considered indifferent given H. Notably, this axiom is weaker than the corresponding version in Ghirardato, Maccheroni, and Marinacci (2008), since it is only imposed on acts of which the agent is aware, i.e., on acts measurable with respect to Σ A . Axiom 9 requires unambiguous preferences to be dynamically consistent for a given level of awareness A. No such restriction is imposed across histories with different levels of awareness.

The following corollary follows from the main result in Ghirardato, Maccheroni, and Marinacci (2008):

Corollary 4

If for a given t = 2 k , and a corresponding history h = ( t , f , A ) the family of preferences h A and ( H A ) H Σ + 1 h , A satisfies Axioms 13 ( a ) , 4, 5, 7 ( a ) , 8 and 9, then there exist sets of probability distributions, Π h and ( Π H ) H Σ + 1 h , A such that

(i) unambiguous preferences at h, * h A = * + 1 h , A A are represented by:

α h A α i f f   s A S A π ( s A ) α ( s A ) s A S A π ( s A ) α ( s A ) f o r a l l π Π h ;

(ii) unambiguous preferences conditional on H Σ + 1 h , A are represented by:

α H A α i f f   s A S A π ( s A ) α ( s A ) s A S A π ( s A ) α ( s A ) f o r a l l π Π H ;

(iii) for every H the set of posteriors Π H is given by the generalized Bayesian updating of Π h :

Π H = { π H | π Π h } ,

where π H is the Bayesian update of π conditional on H.

Proposition 5

For any given even period t = 2 k , and a corresponding history h = ( t ; f ; A ) , the coherent multiple prior model satisfies consequentialism. Furthermore, if π h i n t ( Δ ( S A ) ) for all π h Π h , the coherent prior model at h satisfies dynamic consistency.

6.2 Changes in Awareness

We now consider pure changes in awareness. For t = 2 k + 1 , consider a history h = ( t , f , A ) . To simplify the exposition, we first consider the case in which the individual becomes aware of a single new proposition p. Suppose the individual’s beliefs at h can be represented by a set of coherent probability measures Π h Δ ( S A ) generated by the prior probability π Δ ( Ω ) and the information f available at h. Now suppose the individual becomes aware of a proposition p A ¯ at h = ( t + 1 , f , A ) so that A = A { p } .[12] Consider any s A S A corresponding to a set of truth values for all propositions p A P A , and giving rise to a compound proposition p A .

For each such p A , the individual at h considers two possible compound propositions p A ' ¬ p , p A ' ¬ p corresponding to the truth or falsity of p. Noting that s A { 0 , 1 } | A | is a binary number, we may define the states ( s A , 1 ) (for p true) and ( s A , 0 ) (for p false) in S A ' . Similarly, any s ¯ A S ¯ A corresponds to two complementary states ( s ¯ A , 1 ) (for p true) and ( s ¯ A , 0 ) (for p false) in S ¯ A .

We wish to compare the priors of the agent at h with those she would have held at h if she were already aware of p. As we showed in our example in Section 2, under coherence these will coincide.

Proposition 6

Consider an information structure { t } t = 0 T and an admissible awareness structure { A ( t ) : t 2 P \ { } } t = 0 T such that at history h = ( t , f , A ) for t = 2 k + 1 the change in awareness involves a single elementary proposition p: that is, h + 1 = ( t + 1 , f , A = A { p } ) for some p A . The coherent multiple prior model implies

Π h = C H ¯ { π h ( | ( s ¯ A , 0 ) ) , π h ( | ( s ¯ A , 1 ) ) | s ¯ A S ¯ A }

(10) Π h + 1 = C H ¯ { π h ( | s ¯ A ) | s ¯ A S ¯ A }

where π h and π * h + 1 represent the conditional beliefs of the fully aware agent and for each of the priors π * h ( | s ¯ A ) ,

(11) π * h ( | s ¯ A ) = π * h ( s ¯ A , 1 ) π * h + 1 ( s ¯ A ) π * h + 1 ( | ( s ¯ A , 1 ) ) + π * h ( s ¯ A , 0 ) π * h + 1 ( s ¯ A ) π * h + 1 ( | ( s ¯ A , 0 ) )

We next adapt Axiom A6 to the intertemporal setting. Recall that Definition 2 specifies the conditional preferences under full awareness for each subset in Σ and hence, allows us to talk about preferences conditional on the realization of any event in Σ in conjunction with any history h.

Axiom 10

Conditional Unanimity: Let h = ( t , f , A ) . For α, α A A , α * h A α if and only if for every s ¯ A S ¯ A and every α ^ s ¯ A , α ^ s ¯ A A with α ^ s ¯ A ( s × s ¯ A ) = α ( s ) and α ^ s ¯ A ( s × s ¯ A ) = α ( s ) for each s S A and α ^ s ¯ A ( s × s ˜ A ) = α ^ s ¯ A ( s × s ˜ A ) for all s S A and s ˜ A S ¯ A \ s ¯ A , a ^ s ¯ A h a ^ s ¯ A .

Just as Axiom 6, Axiom 10 can be equivalently expressed in terms of conditional preferences to state that an act α is unambiguously preferred to act α at history h and awareness level A, α * h A α , if and only if for every two acts α ^ , α ^ A ( A ) with and α ^ ( s ˜ ) = α ( s ) for each s ˜ E s and each s S A , α ^ is preferred to α ^ conditional on history h and state s ¯ A , α ^ s ¯ A , h α ^ , for all s ¯ A S ¯ A . Axiom 10 thus requires that at any history h = ( t , f , A ) an act α is unambiguously preferred to α if and only if this ranking is preserved for a fully aware agent, conditional on every possible truth value of the propositions not contained in A.

Proposition 7.

Consider an information structure { t } t = 0 T and an admissible awareness structure { A ( t ) : t 2 P \ { } } t = 0 T with a corresponding set of histories . If

(i) for each h, (conditional) preferences under full awareness ( h ) h satisfy Axioms 13 ( b ) , 4, 5 and 7 ( b ) ;

(ii) for each t = 2 k , h = ( t , f , A ) , (conditional) preferences under partial awareness ( H A ) H Σ + 1 h , A satisfy Axioms 13 ( a ) , 4, 5, 7 ( a ) , 8 and 9;

(iii) for each h = ( t , f , A ) , h A satisfies Axiom 10,

then there exists a prior π * on ( Ω , Σ ) and for each h = ( t , f , A ) a set of priors Π h on ( S A , Σ A ) s.t. for any two acts α and α A A and the corresponding α ^ , α ^ A ( A ) with α ^ ( s ˜ ) = α ( s ) and α ^ ( s ˜ ) = α ( s ) for each s ˜ E s and each s S A ,

(12) α * h A α   i f f

α ^ s ¯ A , h α ^   f o r   a l l   s ¯ A S ¯ A   i f f

s A S A π ( s A ) α ( s A ) s A S A π ( s A ) α ( s A )   f o r   a l l   π Π h

and for any two acts α and α A , any h and any s ¯ A ,

(13) α h α i f f   s Ω π * h ( s ) α ( s ) s Ω π * h ( s ) α ( s )   a n d α s ¯ A , h α i f f   s Ω π * h ( s | s ¯ A ) α ( s ) s Ω π * h ( s | s ¯ A ) α ( s )

Furthermore, π * , together with the family of multiple priors ( Π h ) h is a coherent multiple prior model.

Proposition 7 shows that under Axiom 10, Conditional Unanimity, the set of priors the decisionmaker entertains at h, before becoming aware of a certain set of propositions is given exactly by the convex hull of the posteriors conditional on the truth values of these propositions once she has become aware of them at h + 1 . Using the fact that under Axioms 1–5 and 7–9, conditional beliefs are formed via generalized Bayesian updating, we obtain the coherent multiple prior representation of beliefs.

Finally, since for each history h = ( t , f , A ) , the extreme points of the set Π h are derived from conditional probabilities obtained from the (full awareness) prior π * Δ ( Ω ) , conditioning on h and in turn on each s ¯ A in S ¯ A , it follows that π * h Π h , since from the iterative law of expectations, we have:

π * h = s ¯ A S ¯ A π * h ( s ¯ A ) π * h ( | s ¯ A ) .

More generally, for any history h = ( t , f , A ) , with t odd, and with an immediate successor h + 1 = ( t + 1 , f , A ) embodying a pure increase in awareness (that is, A A ), we have from the construction that Π h + 1 Π h . That is, the mapping from Π h to Π h + 1 may be viewed as a contraction with π * h as a fixed point.

7 An Application: Latent Variables and Mixture Models

The interpretation of multiple priors models offered here may be considered in relation to mixture and latent variables used in a variety of statistical settings. Consider a model in which the dataset consists of M observations, with each observation consisting of a variable of interest y, taking the values 0 and 1, and a vector of explanatory variables X.

Begin by considering a standard binary response model

y = { 1 if  f ( X , β , ε ) > 0 0 otherwise

where β is a vector of parameters and ε is a random variable determined by the state of nature and characterized by parameters θ .

The parameters β and θ may be estimated by maximizing the log-likelihood function

( β , θ ) = m log p ( y m | X m , β , θ )

In particular, if β is known, this estimation procedure amounts to Bayesian updating with respect to θ.

To relate this model to the decision theory problems discussed suppose that the variables in X may be partitioned into a vector of exogenous variables X 1 and a vector of control variables X 2 . As before, utility may be given as u ( y ) = y . The agent seeks to choose X 2 to maximize expected utility p ( y = 1 ) , given the observable value of X 1 and the estimated parameters ( β , θ ) .

A latent variable model associates with each m a vector of K unobserved latent variables ( u k m ) k = 1 K , commonly taken to be discrete. We will focus on the case where each u k m is a dummy variable taking the values 0 and 1. Hence, each latent variable may be interpreted as the truth value of a proposition. Consider for simplicity the case K = 1 . In this case, the distribution of ε is determined by parameters θ 0 if u m = 0 , and θ 1 if u m = 1 .

One way to approach this problem would be to formulate two separate models, one assuming u m = 0 , and the other assuming u m = 1 . Now, we have two possible likelihoods, p 0 ( y m | X m , β ) = p ( y m | X m , β , θ 0 ) and p 1 ( y m | X m , β ) = p ( y m | X m , β , θ 1 ) . This corresponds to the simplest case of multiple priors discussed above. Hence, expected utility lies in the interval [ p 0 , p 1 ] and the choice of X 2 might be given by a criterion such as maxmin. In terms of the discussion above, this approach corresponds to the case where the agent is unaware of the latent variable u.

Now consider the case where the agent is aware of u but cannot observe its value, instead assigning a probability p m = p ( u m = 1 | X 1 m ) . Hence, the log-likelihood is given by a weighted sum

( β , θ 0 , θ 1 ) = m log p ( y m | X m , β , θ 0 , θ 1 )

where

p ( y m | X m , β , θ 0 , θ 1 ) = p ( y m | X m , β , θ 1 ) ( 1 p ( u m = 1 | X 1 m ) ) + p ( y m | X m , β , θ 1 ) p ( u m = 1 | X 1 m )

This corresponds to a latent variable model, more precisely a mixture model. A standard iterative algorithm for estimating the parameters of a model of this kind is expectation-maximization or EM, due to Dempster, Laird, and Rubin (1977). The algorithm proceeds with alternating steps: an “expectation” step in which estimates of p m = p ( u m = 1 | X 1 m ) are updated and a maximization step in which the log-likelihood ( β , θ 0 , θ 1 ) is maximized with respect to the parameter vector ( β , θ 0 , θ 1 ) .

There is an obvious analogy with the dynamic process set out above, where the alternating steps involve changes in awareness of previously unconsidered propositions (that is, latent variables) and updating in response to new information. We conjecture that it might be possible to represent a boundedly rational individual as following a real-time dynamic version of the Dempster, Laird, and Rubin (1977) algorithm in which the expectation and maximization steps incorporate changes in awareness and information respectively.

Finally, note that in the case when the agent is aware of u and can observe its value, we have a standard dummy variable.

8 Related Literature

Our paper is related to the growing literature on unawareness, see Schipper (2014) for an introduction. Two main approaches have arisen in this literature: the first relies on explicitly modeling the knowledge and the awareness of the decision maker at each possible state of the world. The second, consists in axiomatizing choice behavior: either for a given awareness structure which is exogenously specified, or for a subjective awareness structure, which can then be deduced from preferences.

Fagin and Halpern (1988) were the first to introduce awareness structures into a model of knowledge. Their approach was then followed by that of Heifetz, Meier, and Schipper (2006). The main characteristic of this approach is the exogenous specification of awareness at each state of the world. Combined with the information structure, which at each state specifies the event the agent is informed about, this gives rise to two knowledge operators: implicit knowledge – that captured by the information partition and explicit knowledge, which necessitates the agent to simultaneously implicitly know the event and be aware of it.[13]

A simple awareness structure, which captures this distinction is provided by Li (2009). In the present paper, we adopt her framework by modeling unawareness of propositions (“questions” in her framework). At each state, the model exogenously defines the subset of propositions the agent is aware of, as well as the event known to have occurred. Li (2009) shows that such structures can be used to generate meaningful “knowledge” and “unawareness” operators. Our structure is simpler in that we require information obtained at each state (node) to be measurable with respect to the agent’s awareness at this node, which need not be the case in Li (2009). This simplification proves convenient for the analysis and the resulting model captures well the phenomena we have in mind. A generalization to information structures non-measurable with respect to the awareness level would involve techniques similar to those used by Li (2009), whereby the information actually revealed to the agent is the finest coarsening measurable with respect to her current awareness level. This complicates the notation without providing meaningful insights into our main results.

The relation between the model of Li (2009) and the awareness structures of Fagin and Halpern (1988) and unawareness structures by Heifetz, Meier, and Schipper (2006) are discussed in Schipper (2014, p. 4). In as far as these models rely on implicit knowledge, Schipper (2014, p. 3) suggests that such awareness structures can be viewed as capturing “features of logical non-omniscience”.

The class of axiomatic models of unawareness can be roughly divided into two: in the first category fall the models, which take the awareness structure as exogenously given and model preferences related to such structures. This comprises the work by Ahn and Ergin (2010), Grant and Quiggin (2013, 2015), Karni and Viero (2013, 2017), Viero (2018), Lehrer and Teper (2014), Alon (2015), Dominiak and Tserenjigmid (2018), Dietrich (2018). These papers study conditions, which relate preferences across different levels of awareness.

In a seminal paper, Karni and Viero (2013) study growing awareness as a consequence of discovering new acts or new consequences.[14] This results in a refinement or in an expansion of the original state space. Karni and Viero (2013) impose a restriction on beliefs – reverse Bayesianism, which implies that relative likelihoods of events considered at lower levels of awareness are preserved when awareness increases.

Dietrich (2018) also considers partial awareness in terms of both coarsening and reduction of both the state-space and the outcome space. He models a subjective expected utility maximizer, who satisfies Savage axioms at all levels of awareness. His axiomatization imposes stringent constraints on utility and beliefs across awareness structures: both utility and probability at lower awareness levels are intimately related to those utilities and probabilities the agent would entertain were she fully aware.[15] While we study ambiguity as arising from partial awareness, our approach is similar both to those of Karni and Viero (2013) and Dietrich (2018) in that beliefs under full awareness uniquely pin down beliefs at lower levels of awareness.[16] The normative aspect of such constraints as captured by the axioms we propose is of interest in problems of inference, such as the latent variable model discussed in Section 7.

While the approach taken by Karni and Viero (2013) and Dietrich (2018) is purely Bayesian, Grant and Quiggin (2015), Karni and Viero (2017) and Viero (2018) introduce the notion of unknown contingencies and associate with them a decision weight and a utility. In a similar spirit, Alon (2015) models awareness of unforeseen contingencies by introducing an “unforeseen event” with an associated weight and an associated “worst-case-contingency”. While in these models, such decision weights and utilities are purely subjective, our coherent multiple prior model imposes restrictions on the way in which a partially aware agent incorporates the possibility of becoming more aware into her decisions. In particular, partial awareness leads to ambiguity, and thus, multiple priors reflecting all possible ways in which the realization of the yet undiscovered states might affect the probabilities of the currently known ones.

The papers by Lehrer and Teper (2014) and Dominiak and Tserenjigmid (2018) advance the idea that the agent’s confidence/perception of ambiguity may depend on her awareness. Both papers consider two preference relations, one corresponding to a larger state-space (higher level of awareness), the other to a coarsening of it. Both papers argue that the agent behaving as an EUM when partially aware might face more ambiguity / be less confident with a larger state-space. This is captured by incompleteness of preferences on the larger state-space in Lehrer and Teper (2014) and by perceived ambiguity on the expanded state-space in Dominiak and Tserenjigmid (2018). In a similar spirit, Grant, Meneghel, and Tourky (2019) characterize a family of beliefs conditional on realizations of a stochastic process such that: ( i ) as new states are observed, there is maximal ambiguity with respect to the probability of these new states; ( i i ) as information about already known states accumulates, the agent updates her priors in a Bayesian way, eventually learning the correct probabilities of each state. In the limit, the expectations taken with respect to the posterior of each of the initial priors are arbitrarily close.

Similarly to these works, we establish a relation between preferences and beliefs with “varying degrees of awareness” and impose consistency requirements across such beliefs. Our point of departure, however is that ambiguity arises due to partial awareness and disappears as the agent becomes fully aware. We thus propose a deductive (forward-looking) rather than an inductive approach in that beliefs at lower levels of awareness already incorporate all possible unforeseen contingencies as multiple priors. Furthermore, we concentrate on the characterization of ambiguous beliefs (multiple priors) for partial awareness and remain agnostic as to how the agent deals with such ambiguity. This allows for a simpler framework with a binary outcome space and fewer restrictions on preferences.

Finally, Ahn and Ergin (2010)’s model of framing effects can be interpreted as a model of preferences for different levels of awareness. Similarly, to the models in the preceding paragraph, their premise is that the agent has non-additive beliefs on the complete state-space, but forms additive beliefs for each subset of this space (each possible frame). These additive beliefs are derived from the non-additive capacity over the complete state-space.

The second class of axiomatic models uses preferences over extended classes of objects (e.g., menus or consumption streams) and proposes behavioral conditions to identify the awareness of the agent. In Epstein, Marinacci, and Seo (2007) the coarse state space is subjective and derived from preferences over menus, which violate indifference to randomization. When choosing over menus, the agent takes into account for each such coarse state, the worst payoff realization, thus exhibiting complete ignorance about the probability of the individual utility realizations.

Preferences on richer domains can provide a distinction between ambiguity and unawareness in an axiomatic framework. In particular, Piermont (2017) relates awareness of unawareness to the agent’s unwillingness to commit to any contingent plan even when delaying the decision is costly. Kochov (2017) interprets awareness of unawareness as the agent recognizing her inability to correctly judge the autocorrelation of payoffs over time. In his framework, awareness of unawareness is revealed whenever a payoff stream with state-contingent outcomes to which the agent assigns identical utility is assigned a utility different from that assigned to the outcomes.

In contrast, our paper seeks to deduce ambiguity from unawareness rather than distinguishing the two concepts. In that sense, it is related to the papers by Mukerji (1997), Ghirardato (2001), Epstein, Marinacci, and Seo (2007) and Billot and Vergopoulos (2018). For example, Mukerji (1997) shows that probability weighting may be derived from a decision-maker’s anticipation that her perception of future contingencies is incomplete. Ghirardato (2001) models ambiguity as a consequence of the coarse perception of state-contingent payoffs. Billot and Vergopoulos (2018) show how ambiguity can arise when the observable state space differs from the relevant state space.

The difference between these papers and our model is two-fold: first, we study changes in awareness and in information and impose consistency requirements across preferences and beliefs at different levels of awareness and information for the same agent. To do so, we rely extensively the model of generalized Bayesian updating for multiple prior preferences as developed by Ghirardato, Maccheroni, and Marinacci (2008).

Second, we constrain the agent to reason only about acts which are measurable with respect to her coarse state space. Thus, ambiguity is not due to the fact that the agent conceives of multiple outcomes related to a given coarse contingency, but to the fact that the probabilities of the contingencies of which the decision maker is aware depend on factors of which she is not. Rather than assuming that such ambiguity leads to full ignorance and to behavior which is guided by the worst possible outcome, in our context, ambiguity is objectively related to the stochastic process which determines the realization of the factors of which the decision maker is unaware.

Finally, there is a well-known formal result, see Ghirardato and Breton (2000) and Gilboa and Schmeidler (1994), showing that the Choquet integral (the standard tool for modeling ambiguity-sensitive preferences) can be represented in an additive way on an extended state-space (the set of { 0 , 1 } -valued capacities). Our construction is different: we do model additive beliefs on a larger state space and non-additive ones on a smaller state-space, but the two spaces are not related as in Ghirardato and Breton (2000) and Gilboa and Schmeidler (1994) and the resulting evaluations of acts in general differ between the two spaces, reflecting the decrease in ambiguity which accompanies growing awareness.

9 Conclusion

Beginning in the late 1970s, alternatives to and generalizations of Expected Utility theory have proliferated in response to behavioral violations of EU predictions and theoretical criticism of the axiomatic foundations of EU. Examples have included probability weighting models for choice under risk (Allais 1953; Kahneman and Tversky 1979; Quiggin 1982; Yaari 1987), ambiguity models for choice under uncertainty (Gilboa and Schmeidler (1989), Schmeidler (1989), Ghirardato, Maccheroni, and Marinacci (2004), Klibanoff, Marinacci, and Mukerji (2005)) and the rapidly growing literature on unawareness (Schipper (2014)).

We reviewed some of the attempts at unification of the two theories in Section 1. Similarly, in this paper, we have shown that the invariant biseparable model of Ghirardato, Maccheroni, and Marinacci (2004) model of choice under ambiguity (which incorporates α-maxmin EU as a special cases), may be derived from the preferences of an EU maximizer with coarse awareness. Updating in response to both new information and refined awareness is well-behaved.

This development raises the possibility of a more general unified theory of EU behavior with bounded awareness that might encompass a wide range of observed behavior as well as being consistent with the fundamental postulate that all humans have bounded cognitive capacity.

Such models can prove useful as a decision-theoretical underpinning of statistical methods related to latent variables.


Article note: We thank Adam Dominiak, Peter Klibanoff, Sujoy Mukerji, Klaus Nehring, and participants at RUD2019 and at the economics seminar at UGA. We are also especially grateful to the guest editor and two anonymous referees for their patience and perseverance in reviewing earlier versions and providing us with many useful comments and suggestions that have considerably improved the clarity of our exposition. This research has been supported by IUF, Labex MME-DII and the French National Research Agency in the framework of the “Investissements d’avenir” program ANR-15-IDEX-02.



Corresponding author: Simon Grant, Research School of Economics, Australian National University, Canberra, ACT, Australia; and School of Economics, University of Queensland, St Lucia, QLD, Australia, E-mail:

Funding source: French National Research Agency

Award Identifier / Grant number: ANR-15-IDEX-02

Appendix

Proof of Proposition 2:

Suppose that the family of preferences ( A ) A satisfies Axioms 1–6. Fix an awareness level A. By Lemma 1, for the so-chosen A, there exists a unique closed and convex set of priors Π A such that the unambiguous preferences induced by A , * A can be represented by (15). Furthermore, by Axiom 6 and by Definition 2 of conditional preferences, we obtain α * A α iff for any s ¯ A S ¯ A , the corresponding acts α ^ s ¯ A and α ^ s ¯ A defined in statement of Axiom 6, satisfy α ^ s ¯ A α ^ s ¯ A , or, equivalently:

(14) s S π ( s ) α ^ s ¯ A ( s ) s S π ( s ) α ^ s ¯ A ( s ) .

Furthermore, since α ^ s ¯ A ( s ) = α ^ s ¯ A ( s ) for every s S A × S ¯ A \ s ¯ A , we have that (14) is equivalent to:

s A S A π * ( s A | s ¯ A ) α ^ s ¯ A ( s A ) s A S A π * ( s A | s ¯ A ) α ^ s ¯ A ( s A )

and since α ^ s ¯ A ( s ) = α ( s ) and α ^ s ¯ A ( s ) = α ( s ) for every s S A × s ¯ A , we have:

(15) s A S A π ( s A | s ¯ A ) α ( s A ) s A S A π ( s A | s ¯ A ) α ( s A ) .

We thus conclude that the requirement that α ^ s ¯ A α ^ s ¯ A for every s ¯ A S ¯ A is satisfied if and only if (15) holds for every s ¯ A S ¯ A , or, equivalently, if and only if

(16) s A S A π ( s A ) α ( s A ) s A S A π ( s A ) α ( s A ) for all π Π ^ = C H ¯ { π ( | s ¯ A ) | s ¯ A S ¯ A } .

Axiom 6 requires (16) to be equivalent to α A α . But from Axioms 1–5 and Lemma 1, we have that α A α iff

s A S A π ( s A ) α ( s A ) s A S A π ( s A ) α ( s A ) for all π Π A ,

where Π A is unique. Thus, we obtain Π ^ = Π A = C H ¯ { π * ( | s ¯ A ) | s ¯ A S ¯ A } as required by the definition of awareness-based beliefs. Since the awareness level A was chosen arbitrarily, this holds for every A and the claim of the proposition obtains.

Proof of Proposition 5:

Assume a representation by an CMP model. Fix h = ( t , f , A ) . To see that Consequentialism holds, note that generalized Bayesian updating implies that conditional on H Σ + 1 h , A , all h ˜ H are assigned 0-probability under all π Π H . Let

S h = { s A S A | s A f }

S H = { s A S A | s A σ t + 1 ( f ) with  ( t + 1 , σ t + 1 ( f ) , A ) H }

be the set of states in S A consistent with histories in H. By the definition of α and α in Axiom 8, we have that for all π Π H ,

s A S H π ( s A ) α ( s A ) = s A S H π ( s A ) α ( s A )  

so that α * H A α , which implies α H A α .

As for dynamic consistency of H * , note that for any acts α and α as defined in Axiom 9 and every π Π h

s A S h π ( s A | S h ) α ( s A ) s A S h π ( s A | S h ) α ( s A )

= π ( S H ) [ s A S h π ( s A | S H ) α ( s A ) s A S h π ( s A | S H ) α ( s A ) ] .

α * h A α holds iff the first difference is positive for all π Π h and, as long as all such π are in the interior of Δ ( S A ) , this is clearly equivalent to the last difference being positive for all π Π H , or to α * H A α .

Proof of Proposition 6:

Notice that

π h + 1 ( s ¯ A ) = π h + 1 ( s ¯ A , 1 ) + π h + 1 ( s ¯ A , 0 )

where π * is the probability on Ω and its arguments are considered as events. Furthermore, by condition ( i i i ) of the definition of CMP,

Π h + 1 = C H ¯ { π h + 1 ( | s ¯ A ) | s ¯ A S ¯ A } = C H ¯ { π h ( | s ¯ A ) | s ¯ A S ¯ A }

where the second equality follows from the fact that no new information arrives between h and h + 1 , and hence π * h = π * h + 1 .

For h 1 = ( t 1 , ρ t 1 ( f ) , A ) , we have by property ( i i i ) of CMP,

Π h 1 = C H ¯ { π * h 1 ( | s ¯ A ) | s ¯ A S ¯ A }

and hence, by property ( i i ) ,

Π h = { π ( | h ) | π Π h 1 } = { π * h ( | s ¯ A ) | s ¯ A S ¯ A }

= C H ¯ { π h ( | s ¯ A ) } = C H ¯ { π h ( | ( s ¯ A , 0 ) ) , π h ( | ( s ¯ A , 1 ) ) | s ¯ A S ¯ A } .

Hence, Bayesian updating of the beliefs of the fully aware agent π * , together with π * h = π * h + 1 implies:

π * h ( | s ¯ A ) = π * h ( s ¯ A , 1 ) π * h + 1 ( s ¯ A ) π * h + 1 ( | ( s ¯ A , 1 ) ) + π * h ( s ¯ A , 0 ) π * h + 1 ( s ¯ A ) π * h + 1 ( | ( s ¯ A , 0 ) ) .

Proof of Proposition 7:

The representation of h follows directly from the conditions in ( i i ) and Lemma 3. These conditions uniquely identify π * and ensure Bayesian updating of fully aware beliefs.

The representation of the conditional * h A preference relations follows from Axioms 1–3 ( a ) , 4, 5, 7 ( a ) stated in condition ( i ) and Lemma 1. These axioms identify the corresponding sets of priors Π h . Axiom 10 relates the partially aware preferences conditional on h to the conditional preferences under full awareness and implies condition (12).

Just as Axiom 6 for the static case, Axiom 10 relates Π h 0 (the partially aware preference with no information) to the conditionals of π ( | s ¯ A 0 ) and by Lemma 2 implies property ( i ) of the CMP model.

Property ( i i ) of the CMP model follows from Axioms 8 and 9, which by Corollary 4, imply generalized Bayesian updating of every prior in Π h .

Finally, the proof of property ( i i i ) of the CMP model is a consequence of Axiom 10 and is shown exactly as in the proof of Proposition 2. Indeed, fix a history h and in the proof of Proposition 2, replace preferences * A by * h A and by h , whereas the set of priors Π A is replaced by Π h and the full awareness probability distribution π * is replaced by π * h . Repeating the arguments in the proof of Proposition 2 gives property (12). Property (13) follows from the fact that the fully aware preference at history h satisfies all axioms of expected utility maximization for the probability distribution π * h , Axioms 1–3 ( b ) , 4, 5 and by Axiom 7 ( b ) all relevant events are non-null.

References

Agotnes, T., and N. Alechina. 2008. “Proceedings of the 11th Conference on Theoretical Aspects of Rationalioty and Knowledge (tark).” In Full and Relative Awareness: A Decidable Logic for Reasoning about Knowledge and Unawareness, 6–14. Belgium: Presses Universitaires De Louvain.10.1145/1324249.1324255Search in Google Scholar

Ahn, D., and H. Ergin. 2010. “Framing Contingencies.” Econometrica 78: 655–95. https://doi.org/10.3982/ecta7019.Search in Google Scholar

Allais, M. 1953. “Le comportement de l’homme rationnel devant le risque: critique des postulats et axiomes de l’école Américaine.” Econometrica 21 (4): 503–46. https://doi.org/10.2307/1907921.Search in Google Scholar

Alon, S. 2015. “Worst-case Expected Utility.” Journal of Mathematical Economics 60: 43–8. https://doi.org/10.1016/j.jmateco.2015.06.017.Search in Google Scholar

Billot, A., and V. Vergopoulos. 2018. “Expected Utility Without Parsimony.” Mathematical Social Sciences 93: 14–21. https://doi.org/10.1016/j.mathsocsci.2017.11.004.Search in Google Scholar

Cowan, N. 2020. Pandemic Responses are Beyond Evidence-based Medicine. Also available at https://notesonliberty.com/2020/03/23/pandemic-responses-are-beyond-evidence-based-medicine/.Search in Google Scholar

Dempster, A. P., N. M. Laird, and D. B. Rubin. 1977. “Maximum Likelihood from Incomplete Data via the EM Algorithm.” Journal of the Royal Statistical Society, Series B 39 (1): 1–38.10.1111/j.2517-6161.1977.tb01600.xSearch in Google Scholar

Dietrich, F. 2018. “Savage’s Theorem Under Changing Awareness.” Journal of Economic Theory 176: 1–54. https://doi.org/10.1016/j.jet.2018.01.015.Search in Google Scholar

Dominiak, A., and G. Tserenjigmid. 2018. “Ambiguity Under Growing Awareness.” SSRN Electronic Journal. https://doi.org/10.2139/ssrn.3247761.Search in Google Scholar

Ellsberg, D. (1961). Risk, Ambiguity, and the Savage Axioms. The Quarterly Journal of Economics 75(4), 643–69.10.2307/1884324Search in Google Scholar

Empson, W. 1930. Seven Types of Ambiguity. Read Books Limited.Search in Google Scholar

Epstein, L., M. Marinacci, and K. Seo. 2007. “Recursive Smooth Ambiguity Preferences.” Theoretical Economics 2: 355–94.Search in Google Scholar

Fagin, R., and J. Y. Halpern. 1988. “Belief, Awareness, and Limited Reasoning.” Artificial Intelligence 34: 39–76. https://doi.org/10.1016/0004-3702(87)90003-8.Search in Google Scholar

Ghirardato, P. 2001. “Coping with Ignorance: Unforeseen Contingencies and Non-additive Uncertainty.” Economic Theory 17: 247–76. https://doi.org/10.1007/pl00004108.Search in Google Scholar

Ghirardato, P., and M. L. Breton. 2000. “Choquet Rationality.” Journal of Economic Theory 90: 277–85. https://doi.org/10.1006/jeth.1999.2599.Search in Google Scholar

Ghirardato, P., F. Maccheroni, and M. Marinacci. 2004. “Differentiating Ambiguity and Ambiguity Attitude.” Journal of Economic Theory 118 (2): 133–73. https://doi.org/10.1016/j.jet.2003.12.004.Search in Google Scholar

Ghirardato, P., F. Maccheroni, and M. Marinacci. 2008. “Revealed Ambiguity and Its Consequences: Updating.” In Advances in Decision Making under Risk and Uncertainty, 3–18. Springer.10.1007/978-3-540-68437-4_1Search in Google Scholar

Gilboa, I., and D. Schmeidler. 1989. “Maxmin Expected Utility with Non-unique Prior.” Journal of Mathematical Economics 18 (2): 141–53. https://doi.org/10.1016/0304-4068(89)90018-9.Search in Google Scholar

Gilboa, I., and D. Schmeidler. 1994. “Maxmin Expected Utility with Non-unique Prior.” Annals of Operations Research 52: 43–65.10.1007/BF02032160Search in Google Scholar

Grant, S., J. Kline, and J. Quiggin. 2018. “Contracting Under Uncertainty: A Principal-agent Model with Ambiguity Averse Parties.” Games and Economic Behavior 109: 582–97. https://doi.org/10.1016/j.geb.2018.02.008.Search in Google Scholar

Grant, S., I. Meneghel, and R. Tourky. 2019. “Learning Under Unawareness.” SSRN Electronic Journal. https://doi.org/10.2139/ssrn.3113983.Search in Google Scholar

Grant, S., and J. Quiggin. 2013. “Inductive Reasoning about Unawareness.” Economic Theory 54: 717–55. https://doi.org/10.1007/s00199-012-0734-y.Search in Google Scholar

Grant, S., and J. Quiggin. 2015. “A Preference Model for Choice Subject to Surprise.” Theory and Decision 79 (2): 167–80. https://doi.org/10.1007/s11238-014-9472-x.Search in Google Scholar

Heifetz, A., M. Meier, and B. C. Schipper. 2006. “Interactive Unawareness.” Journal of Economic Theory 130: 78–94. https://doi.org/10.1016/j.jet.2005.02.007.Search in Google Scholar

Kahneman, D., and A. Tversky. 1979. “Prospect Theory: An Analysis of Decision under Risk.” Econometrica 47 (2): 263–91. https://doi.org/10.2307/1914185.Search in Google Scholar

Karni, E., and M.-L. Viero. 2013. ““Reverse Bayesianism”: A Choice-based Theory of Growing Awareness.” American Economic Review 103 (7): 2790–810. https://doi.org/10.1257/aer.103.7.2790.Search in Google Scholar

Karni, E., and M.-L. Viero. 2015. “Probabilistic Sophistication and Reverse Bayesianism.” Journal of Risk and Uncertainty 50: 189–208. https://doi.org/10.1007/s11166-015-9216-5.Search in Google Scholar

Karni, E., and M.-L. Viero. 2017. “Awareness of Unawareness: A Theory of Decision Making in the Face of Ignorance.” Journal of Economic Theory 168: 301–28. https://doi.org/10.1016/j.jet.2016.12.011.Search in Google Scholar

Klibanoff, P., M. Marinacci, and S. Mukerji. 2005. “A Smooth Model of Decision Making under Ambiguity.” Econometrica 73 (6): 1849–92. https://doi.org/10.1111/j.1468-0262.2005.00640.x.Search in Google Scholar

Knight, F. H. (1921, 2006). Risk, Uncertainty, and Profit. Dover Publications.Search in Google Scholar

Kochov, A. 2017. “A Behavioral Definition of Unforeseen Contingencies.” Journal of Economic Theory 175: 265–90. https://doi.org/10.1016/j.jet.2018.01.018.Search in Google Scholar

Lehrer, E., and R. Teper. 2014. “Extension Rules or What Would the Sage Do?” American Economic Journal: Microeconomics 6: 5–22. https://doi.org/10.1257/mic.6.1.5.Search in Google Scholar

Li, J. 2009. “Information Structures with Unawareness.” Journal of Economic Theory 144: 977–93. https://doi.org/10.1016/j.jet.2008.10.001.Search in Google Scholar

Mukerjee, S. 2015. The Laws of Medicine: Field Notes from An Uncertain Science. Simon Schuster.Search in Google Scholar

Mukerji, S. 1997. “Understanding the Non-additive Probability Decision Model.” Economic Theory 9 (1): 23–46. https://doi.org/10.1007/s001990050108.Search in Google Scholar

Piermont, E. 2017. “Introspective Unawareness and Observable Choice.” Games and Economic Behavior 106: 134–52. https://doi.org/10.1016/j.geb.2017.09.008.Search in Google Scholar

Pires, C. P. 2002. “A rule for Updating Ambiguous Beliefs.” Theory and Decision 53 (2): 137–52. https://doi.org/10.1023/A:1021255808323.10.1023/A:1021255808323Search in Google Scholar

Quiggin, J. 1982. “A Theory of Anticipated Utility.” Journal of Economic Behavior & Organization 3 (4): 323–43. https://doi.org/10.1016/0167-2681(82)90008-7.Search in Google Scholar

Quiggin, J. 2016. “The value of Information and the Value of Awareness.” Theory and Decision 80 (2): 167–85. https://doi.org/10.1007/s11238-015-9496-x.Search in Google Scholar

Savage, L. J. 1954. The Foundations of Statistics. John Wiley & Sons.Search in Google Scholar

Schipper, B. C. 2014. “Unawareness – A Gentle Introduction to both the Literature and the Special Issue.” Mathematical Social Sciences 70: 1–9. https://doi.org/10.1016/j.mathsocsci.2014.03.002.Search in Google Scholar

Schmeidler, D. 1989. “Subjective Probability and Expected Utility without Additivity.” Econometrica 57 (3): 571–87. https://doi.org/10.2307/1911053.Search in Google Scholar

Viero, M.-L. 2018. An Intertemporal Model of Growing Awareness.Search in Google Scholar

Walker, O. 2014. “Unawareness with ”possible” Possible Worlds.” Mathematical Social Sciences 70: 23–33. https://doi.org/10.1016/j.mathsocsci.2014.03.003.Search in Google Scholar

Yaari, M. E. 1987. “The Dual Theory of Choice under Risk.” Econometrica 55 (1): 95–115. https://doi.org/10.2307/1911158.Search in Google Scholar

Received: 2018-12-14
Accepted: 2020-05-09
Published Online: 2020-08-26

© 2020 Walter de Gruyter GmbH, Berlin/Boston

Downloaded on 19.4.2024 from https://www.degruyter.com/document/doi/10.1515/bejte-2018-0185/html
Scroll to top button