5.4 Decision Theory


5.4 Decision Theory

The aim of decision theory is to help agents make rational decisions. Consider an agent that has to choose between a number of acts, such as whether to bet on a horse race and, if so, which horse to bet on. Are there reasonable rules to help the agent decide how to make this choice?

5.4.1 The Basic Framework

There are a number of equivalent ways of formalizing the decision process. Intuitively, a decision situation describes the objective part of the circumstance that an agent faces (i.e., the part that is independent of the tastes and beliefs of the agent). Formally, a decision situation is a tuple DS = (A, W, C), where

  • W is the set of possible worlds (or states of the world),

  • C is the set of consequences, and

  • A is the set of acts (i.e., functions from W to C).

An act a is simple if its range is finite. That is, a is simple if it has only finitely many possible consequences. For simplicity, I assume here that all acts are simple. (Note that this is necessarily the case if either W or C is finite, although I do not require this.) The advantage of using simple acts is that the expected utility of an act can be defined using (plausibilistic) expectation.

A decision problem is essentially a decision situation together with information about the preferences and beliefs of the agent. These tastes and beliefs are represented using a utility function and a plausibility measure, respectively. A utility function maps consequences to utilities. Intuitively, if u is the agent's utility function and c is a consequence, then u(c) measures how happy the agent would be if c occurred. That is, the utility function quantifies the preferences of the agent. Consequence c is at least as good as consequence c iff u(c) u(c). In the literature, the utility function is typically assumed to be real-valued and the plausibility measure is typically taken to be a probability measure. However, assuming that the utility function is real-valued amounts to assuming (among other things) that the agent can totally order all consequences. Moreover, there is the temptation to view a consequence with utility 6 as being "twice as good" as a consequence with utility 3. This may or may not be reasonable. In many cases, agents are uncomfortable about assigning real numbers with such a meaning to utilities. Even if an agent is willing to compare all consequences, she may be willing to use only labels such as "good," "bad," or "outstanding."

Nevertheless, in many simple settings, assuming that people have real-valued utility function and a probability measure does not seem so unreasonable. In the horse race example, the worlds could represent the order in which the horses finished the race. The plausibility measure thus represents the agent's subjective estimate of the likelihood of various orders of finish. The acts correspond to possible bets (perhaps including the act of not betting at all). The consequence of a bet of $10 on Northern Dancer depends on how Northern Dancer finishes in the world. The consequence could be purely monetary (the agent wins $50 in the worlds where Northern Dancer wins the race) but could also include feelings (the agent is dejected if Northern Dancer finishes last, and he also loses $10). Of course, if the consequence includes feelings such as dejection, then it may be more difficult to assign it a numeric utility.

Formally, a (plausibilistic) decision problem is a tuple DP = (DS, ED, Pl, u), where DS = (A, W, C) is a decision situation, ED = (D1, D2, D3, , ) is an expectation domain, Pl : 2W D1 is a plausibility measure, and u : C D2 is a utility function. It is occasionally useful to consider nonplausibilistic decision problems, where there is no plausibility measure. A nonplausibilistic decision problem is a tuple DP = (DS, D, u) where DS = (A, W, C) is a decision situation and u : C D is a utility function.

5.4.2 Decision Rules

I have taken utility and probability as primitives in representing a decision problem. But decision theorists are often reluctant to take utility and probability as primitives. Rather, they take as primitive a preference order on acts, since preferences on acts are observable (by observing agents actions), while utilities and probabilities are not. However, when writing software agents that make decisions on behalf of people, it does not seem so unreasonable to somehow endow the software agent with a representation of the tastes of the human for which it is supposed to be acting. Using utilities is one way of doing this. The agent's beliefs can be represented by a probability measure (or perhaps by some other representation of uncertainty). Alternatively, the software agent can then try to learn something about the world by gathering statistics. In any case, the agent then needs a decision rule for using the utility and plausibility to choose among acts. Formally, a (nonplausibilistic) decision rule is a function mapping (nonplausibilistic) decision problems DP to preference orders on the acts of DP. Many decision rules have been studied in the literature.

Perhaps the best-known decision rule is expected utility maximization. To explain it, note that corresponding to each act a there is a random variable ua on worlds, where ua(w) = u(a(w)). If u is real-valued and the agent's uncertainty is characterized by the probability measure μ, then Eμ(ua) is the expected utility of performing act a. The rule of expected utility maximization orders acts according to their expected utility: a is considered at least as good as a if the expected utility of a is greater than or equal to that of a.

There have been arguments made that a "rational" agent must be an expected utility maximizer. Perhaps the best-known argument is due to Savage. Savage assumed that an agent starts with a preference order A on a rather rich set A of acts: all the simple acts mapping some space W to a set C of consequences. He did not assume that the agents have either a probability on W or a utility function associating with each consequence in C its utility. Nevertheless, he showed that if A satisfies certain postulates, then the agent is acting as if she has a probability measure μ on W and a real-valued utility function u on C, and is maximizing expected utility; that is, a A a iff Eμ(ua) Eμ(ua). Moreover, μ is uniquely determined by A and u is determined up to positive affine transformations. That is, if the same result holds with u replaced by u, then there exist real values a > 0 and b such that u(c) = au(c) + b for all consequences c. (This is because Eμ(aX + b) = aEμ(x) + b and x y iff ax + b ay + b.)

Of course, the real question is whether the postulates that Savage assumes for A are really ones that the preferences of all rational agents should obey. Not surprisingly, this issue has generated a lot of discussion. (See the notes for references.) Two of the postulates are analogues of RAT2 and RAT3 from Section 2.2.1: is transitive and is total. While it seems reasonable to require that an agent's preferences on acts be transitive, in practice transitivity does not always hold (as was already observed in Section 2.2.1). And it is far from clear that it is "irrational" for an agent not to have a total order on the set of all simple acts. Suppose that W ={w1, w2, w3}, and consider the following two acts, a1 and a2:

  • In w1, a1 has the consequence "live in a little pain for three years"; in w2, a1 results in "live for four years in moderate pain"; and in w3, a1 results in "live for two years with moderate pain."

  • In w1, a2 has the consequence "undergo a painful medical operation and live for only one day"; in w2, a2 results in "undergo a painful operation and live a pain-free life for one year"; and in w3, a2 results in "undergo a painful operation and live a pain-free life for five years."

These acts seem hard enough to compare. Now consider what happens if the set of worlds is extended slightly, to allow for acts that involve buying stocks. Then the worlds would also include information about whether the stock price goes up or down, and an act that results in gaining $10,000 in one world and losing $25,000 in another would have to be compared to acts involving consequences of operations. These examples all involve a relatively small set of worlds. Imagine how much harder it is to order acts when the set of possible worlds is much larger.

There are techniques for trying to elicit preferences from an agent. Such techniques are quite important in helping people decide whether or not to undergo a risky medical procedure, for example. However, the process of elicitation may itself affect the preferences. In any case, assuming that preferences can be elicited from an agent is not the same as assuming that the agent "has" preferences.

Many experiments in the psychology literature show that people systematically violate Savage's postulates. While this could be simply viewed as showing that people are irrational, other interpretations are certainly possible. A number of decision rules have been proposed as alternatives to expected utility maximization. Some of them try to model more accurately what people do; others are viewed as more normative. Just to give a flavor of the issues involved, I start by considering two well-known decision rules: maximin and minimax regret.

Maximin orders acts by their worst-case outcomes. Let worstu(a) = min{ua(w) : w W}; worstu(a) is the utility of the worst-case consequence if a is chosen. Maximin prefers a to a if worstu(a) worstu(a). Maximin is a "conservative" rule; the "best" act according to maximin is the one with the best worst-case outcome. The disadvantage of maximin is that it prefers an act that is guaranteed to produce a mediocre outcome to one that is virtually certain to produce an excellent outcome but has a very small chance of producing a bad outcome. Of course, "virtually certain" and "very small chance" do not make sense unless there is some way of comparing the likelihood of two sets. Maximin is a nonplausibilistic decision rule; it does not take uncertainty into account at all.

Minimax regret is based on a different philosophy. It tries to hedge an agent's bets, by doing reasonably well no matter what the actual world is. It is also a nonplausibilistic rule. As a first step to defining it, for each world w, let u*(w) be the sup of the outcomes in world w; that is, u*(w) = supaA ua(w). The regret of a in world w, denoted regretu(a, w), is u*(w) ua(w); that is, the regret of a in w is essentially the difference between the utility of the best possible outcome in w and the utility of performing a in w. Let regretu(a) = supwW regretu(a, w) (where regretu(a) is taken to be if there is no bound on {regretu(a, w) : w W}). For example, if regretu(a) = 2, then in each world w, the utility of performing a in w is guaranteed to be within 2 of the utility of any act the agent could choose, even if she knew that the actual world was w. The decision rule of minimax regret orders acts by their regret and thus takes the "best" act to be the one that minimizes the maximum regret. Intuitively, this rule is trying to minimize the regret that an agent would feel if she discovered what the situation actually was: the "I wish I had done a instead of a" feeling.

As the following example shows, maximin, minimax regret, and expected utility maximization give very different recommendations in general:

Example 5.4.1

start example

Suppose that W ={w1, w2}, = {a1, a2, a3}, and μ is a probability measure on W such that μ(w1) = 1/5 and μ(w2) = 4/5. Let the utility function be described by the following table:

w1

w2

a1

3

3

a2

1

5

a3

2

4

Thus, for example, u(a3(w2)) = 4. It is easy to check that Eμ(ua1) = 3, Eμ(ua2) = 3.8, and Eμ(ua3) = 3.6, so the expected utility maximization rule recommends a2. On the other hand, worstu(a1) = 3, worstu(a2) =1, and worstu(a3) = 2, so the maximin rule recommends a1. Finally, regretu(a1) = 2, regretu(a2) = 4, and regretu(a3) = 1, so the regret minimization rule recommends a3.

Intuitively, maximin "worries" about the possibility that the true world may be

w1, even if it is not all that likely relative to w2, and tries to protect against the eventuality of w1 occurring. Although, a utility of 1 may not be so bad, if all these number are multiplied by 1,000,000—which does not affect the recommendations at all (Exercise 5.35)—it is easy to imagine an executive feeling quite uncomfortable about a loss of $1,000,000, even if such a loss is relatively unlikely and the gain in w2 is $5,000,000. (Recall that essentially the same point came up in the discussion of RAT4 in Section 2.2.1.) On the other hand, if 1is replaced by 1.99 in the original table (which can easily be seen not to affect the recommendations), expected utility maximization starts to seem more reasonable.

end example

Even more decision rules can be generated by using representations of uncertainty other than probability. For example, consider a set probability measures. Define so that aa iff E (ua) E(ua). This can be seen to some extent as a maximin rule; the best act(s) according to are those whose worst-case expectation (according to the probability measures in ) are best. Indeed, if W consists of all probability measures on W, then it is easy to show that EW = Worstu(a); it thus follows that defines the same order on acts as maximin (Exercise 5.36). Thus, can be viewed as a generalization of maximin. If there is no information at all about the probability measure, then the two orders agree. However, can take advantage of partial information about the probability measure. Of course, if there is complete information about the probability measure (i.e., is a singleton), then just reduces to the ordering provided by maximizing expected utility. It is also worth recalling that, as observed in Section 5.3, the ordering on acts induced by EBel for a belief function Bel is the same as that given by

Other orders using can be defined in a similar way:

The order on acts induced by is very conservative; aa iff the best expected outcome according to a is no better than the worst expected outcome according to a. The order induced by is more refined. Clearly EPl (ua) EPl (ua) iff Eμ(ua) Eμ(ua) for all μ . It easily follows that if aa, then a (Exercise 5.37). The converse may not hold. For example, suppose that = {μ, μ′}, and acts a and a are such that Eμ(ua) = 2, Eμ(ua) = 4, Eμ(ua) = 1, and Eμ′(ua) = 3. Then E(ua) = 2, E(ua) = 4, E(ua) = 1, and E(ua) = 3, so a and a are incomparable according to , yet a.

Given this plethora of decision rules, it would be useful to have one common framework in which to make sense of them all. The machinery of expectation domains developed in the previous section helps.

5.4.3 Generalized Expected Utility

There is a sense in which all of the decision rules mentioned in Section 5.4.2 can be viewed as instances of generalized expected utility maximization, that is, utility maximization with respect to plausibilistic expectation, defined using (5.14) for some appropriate choice of expectation domain. The goal of this section is to investigate this claim in more depth.

Let GEU (generalized expected utility) be the decision rule that uses (5.14). That is, if DP = (DS, ED, Pl, u), then GEU(DP) is the preference order such that a a iff EPl, ED(ua) EPl, ED(ua). The following result shows that GEU can represent any preference order on acts:

Theorem 5.4.2

start example

Given a decision situation DS = (A, W, C) and a partial preorder A on A, there exists an expectation domain ED = (D1, D2, D3, , ), a plausibility measure Pl : W D1, and a utility function u : C D2 such that GEU(DP) =A, where DP = (DS, ED, Pl, u) (i.e., EPl, ED(ua) EPl, ED(ua) iff a A a).

end example

Proof See Exercise 5.38.

Theorem 5.4.2 can be viewed as a generalization of Savage's result. It says that, whatever the agent's preference order on acts, the agent can be viewed as maximizing expected plausibilistic utility with respect to some plausibility measure and utility function. Unlike Savage's result, the plausibility measure and utility function are not unique in any sense.

The key idea in the proof of Theorem 5.4.2 is to construct D3 so that its elements are expected utility expressions. The order 3 on D3 then mirrors the preference order A on A. This flexibility in defining the order somewhat undercuts the interest in Theorem 5.4.2. However, just as in the case of plausibility measures, this flexibility has some advantages. In particular, it makes it possible to understand exactly which properties of the plausibility measure and the expectation domain (in particular, and ) are needed to capture various properties of preferences, such as those characterized by Savage's postulates. Doing this is beyond the scope of the book (but see the notes for references). Instead, I now show that GEU is actually a stronger representation tool than this theorem indicates. It turns out that GEU can represent not just orders on actions, but the decision process itself.

Theorem 5.4.2 shows that for any preference order A on the acts of A, there is a decision problem DP whose set of acts is A such that GEU(DP) =A. What I want to show is that, given a decision rule , GEU can represent in the sense that there exists an expectation domain (in particular, a choice of and ) such that GEU (DP) = (DP) for all decision problems DP. Thus, the preference order on acts given by is the same as that derived from GEU.

The actual definition of representation is a little more refined, since I want to be able to talk about a plausibilistic decision rule (like maximizing expected utility) representing a nonplausibilistic decision rule (like maximin). To make this precise, a few definitions are needed.

Two decision problems are congruent if they agree on the tastes and (if they are both plausibilistic decision problems) the beliefs of the agent. Formally, DP1 and DP2 are congruent if they involve the same decision situation, the same utility function, and (where applicable) the same plausibility measure. Note, in particular, that if DP1 and DP2 are both nonplausibilistic decision problems, then they are congruent iff they are identical. However, if they are both plausibilistic, it is possible that they use different notions of and in their expectation domains.

A decision-problem transformation τ maps decision problems to decision problems. Given two decision rules 1 and 2, 1 represents 2 iff there exists a decision-problem transformation τ such that, for all decision problems DP in the domain of 2, the decision problem τ(DP) is in the domain of 1, τ(DP) is congruent to DP, and DR1(τ(DP)) = DR2(DP). Intuitively, this says that 1 and 2 order acts the same way on corresponding decision problems.

I now show that GEU can represent a number of decision rules. Consider maximin with real-valued utilities (i.e., where the domain of maximin consists of all decision problems where the utility is real-valued). Let EDmax = (, {0,1}, , min, ), and let Plmax be the plausibility measure such that Plmax(U) is 0 if X = and 1 otherwise. For the decision problem described in Example 5.4.1,

Note that EPlmax, EDmax(ua) = worstu(ai), for i = 1, 2, 3. Of course, this is not a fluke. If DP = (DS, , u), where DS = (A, W, C), then it is easy to check that EPlmax, EDmax(ua) = worstu(a) for all actions a A (Exercise 5.39).

Take τ(DP) = (DS, EDmax, Plmax, u). Clearly DP and τ(DP) are congruent; the agent's tastes have not changed. Moreover, it is immediate that GEU(τ(DP)) = maximin(DP). Thus, GEU represents maximin. (Although maximin is typically considered only for real-valued utility functions, it actually makes perfect sense as long as utilities are totally ordered. GEU also represents this "generalized" maximin, using essentially the same transformation τ ; see Exercise 5.40.)

Next, I show that GEU can represent minimax regret. For ease of exposition, I consider only decision problems DP = ((A, W, C), , u) with real-valued utilities such that MDP = sup{u*(w) : w W} < . (If MDP =, given the restriction to simple acts, then it is not hard to show that every act has infinite regret; see Exercise 5.41.) Let EDreg = ([0, 1], , , min, ), where x y = y log(x) if x > 0, and x y = 0 if x = 0. Note that = 0 and = 1. Clearly, min is associative and commutative, and r = r log(1) = r for all r . Thus, EDreg is an expectation domain.

For U W, define MU = sup{u*(w) : w U}. Note that MDP = MW. Since, by assumption, MDP < , it follows that MU < for all U W. Let PlDP be the plausibility measure such that PlDP() = 0 and PlM(U) = eMUMDP for U .

For the decision problem described in Example 5.4.1, MDP = 5. An easy computation shows that, for example,

Similar computations show that EPlDP, EDreg(uai) = 5 regretu(ai) for i = 1, 2, 3. More generally, it is easy to show that

for all acts a A (Exercise 5.42). Let τ be the decision-problem transformation such that τ(DP) = (DS, EDreg, PlDP, u). Clearly, higher expected utility corresponds to lower regret, so GEU(τ(DP)) = regret(DP).

Note that, unlike maximin, the plausibility measure in the transformation from DP to τ(DP) for minimax regret depends on DP (more precisely, it depends on MDP). This is unavoidable; see the notes for some discussion.

Finally, given a set of probability measures, consider the decision rules induced by and , (Formally, these rules take as inputs plausibilistic decision problems where the plausibility measure is Pl.) Proposition 5.4.3 shows that all these rules can be represented by GEU.

The following proposition summarizes the situation:

Proposition 5.4.3

start example

GEU represents maximin, minimax regret, and the decision rules induced by and .

end example

Conspicuously absent from the list in Proposition 5.4.3 is the decision rule determined by expected belief (i.e., a1 a2 iff EBel(ua1) EBel(ua2). In fact, this rule cannot be represented by GEU. To understand why, I give a complete characterization of the rules that can be represented.

There is a trivial condition that a decision rule must satisfy in order to be represented by GEU. Say that a decision rule respects utility if the preference order on acts induced by agrees with the utility ordering. More precisely, respects utility if, given as input a decision problem DP with utility function u, if two acts a1 and a2 in DP induce constant utilities d1 and d2 (i.e., if uai (w) = di for all w W, for i = 1, 2), then d1 d2 iff a1 (DP). It is easy to see that all the decision rules I have considered respect utility. In particular, it is easy to see that GEU respects utility, since the expected utility of the constant act with utility d is just d. (This depends on the assumption that d = d.) Thus, GEU cannot possibly represent a decision rule that does not respect utility. (Note that there is no requirement of respecting utility in Theorem 5.4.2, nor does there have to be. Theorem 5.4.2 starts with a preference order, so there is no utility function in the picture. Here, I am starting with a decision rule, so it makes sense to require that it respect utility.)

While respecting utility is a necessary condition for a decision rule to be representable by GEU, it is not sufficient. It is also necessary for the decision rule to treat acts that behave in similar ways similarly. Given a decision problem DP, two acts a1, a2 in DP are indistinguishable, denoted a1 ~DP a2 iff either

  • DP is nonplausibilistic and ua1 = ua2, or

  • DP is plausibilistic, (ua1) = (ua2), and Pl(u1 a1 (d)) = Pl(u1a2 (d)) for all utilities d in the common range of ua1 and ua2, where Pl is the plausibility measure in DP.

In the nonplausibilistic case, two acts are indistinguishable if they induce the same utility random variable; in the nonplausibilistic case, they are indistinguishable if they induce the same plausibilistic "utility lottery."

A decision rule is uniform if it respects indistinguishability. More formally, a decision rule is uniform iff, for all DP in the domain of and all a1, a2, a3 in DP, if a1 ~DP a2 then

Clearly GEU is uniform, so a decision rule that is not uniform cannot be represented by GEU. However, as the following result shows, this is the only condition other than respecting utility that a decision rule must satisfy in order to be represented by GEU:

Theorem 5.4.4

start example

If is a decision rule that respects utility, then is uniform iff can be represented by GEU.

end example

Proof See Exercise 5.43.

Most of the decision rules I have discussed are uniform. However, the decision rule induced by expected belief is not, as the following example shows:

Example 5.4.5

start example

Consider the decision problem where

  • A = {a1, a2};

  • W = {w1, w2, w3};

  • C = {1, 2, 3};

  • u(j) = j, for j = 1, 2, 3;

  • a1(wj) = j and a2(wj) = 3 j, for j = 1, 2, 3; and

  • Bel is the belief function such that Bel({w1, w2}) = Bel(W) = 1, and Bel(U) = 0 if U is not a superset of {w1, w2}.

Clearly a1 ~DP a2, since u1ai (j) is a singleton, so Bel(u1ai (j)) = 0 for i = 1, 2 and j = 1, 2, 3. On the other hand, by definition,

while

It follows that the decision rule that orders acts based on their expected belief is not uniform, and so cannot be represented by GEU.

end example

The alert reader may have noticed an incongruity here. By Theorem 2.4.1, Bel = (Bel)* and, by definition, EBel = EBel. Moreover, the preference order induced by EBel can be represented by GEU. There is no contradiction to Theorem 5.4.4 here. If BEL is the decision rule induced by expected belief and DP is the decision problem in Example 5.4.5, then there is no decision problem τ(DP) = (DS, ED, Bel, u) such that GEU (τ(DP)) = BEL(DP). Nevertheless, GEU ((A, W, C), EDD, PlBel, u) = BEL(DP). (Recall that EDD is defined at the end of Section 5.4.2.) (A, W, C), EDD, PlBel, u) and DP are not congruent; PlBel and Bel are not identical representations of the agent's beliefs. They are related, of course. It is not hard to show that if PlBel (U) PlBel (V, then Bel(U) Bel(V), although the converse does not hold in general (Exercise 5.44).

For a decision rule 1 to represent a decision rule 2, there must be a decision-problem transformation τ such that τ(DP) and DP are congruent and 2(τ(DP)) = 2(DP) for every decision problem DP in the domain of 2. Since τ(DP) and DP are congruent, they agree on the tastes and (if they are both plausibilistic) the beliefs of the agent. I now want to weaken this requirement somewhat, and consider transformations that preserve an important aspect of an agent's tastes and beliefs, while not requiring them to stay the same.

There is a long-standing debate in the decision-theory literature as to whether preferences should be taken to be ordinal or cardinal. If they are ordinal, then all that matters is their order. If they are cardinal, then it should be meaningful to talk about the differences between preferences, that is, how much more an agent prefers one outcome to another. Similarly, if expressions of likelihood are taken to be ordinal, then all that matters is whether one event is more likely than another.

Two utility functions u1 and u2 represent the same ordinal tastes if they are defined on the same set C of consequences and for all c1, c2 C, u1(c1) u1(c2) iff u2(c1) u2(c2). Similarly, two plausibility measures Pl1 and Pl2 represent the same ordinal beliefs if Pl1 and Pl2 are defined on the same domain W and Pl1(U) Pl1(V) iff Pl2(U) Pl2(V) for all U, V W. Finally, two decision problems DP1 and DP2 are similar iff they involve the same decision situations, their utility functions represent the same ordinal tastes, and their plausibility measures represent the same ordinal beliefs. Note that two congruent decision problems are similar, but the converse may not be true.

Decision rule 1 ordinally represents decision rule 2 iff there exists a decision-problem transformation τ such that, for all decision problems DP in the domain of 2, the decision problem τ(DP) is in the domain of 1, DP is similar to τ(DP), and 1(τ(DP)) = 2(DP). Thus, the definition of emulation is just like that of representation, except that τ(DP) is now required only to be similar to DP, not congruent.

I now want to show that GEU can ordinally represent essentially all decision rules. Doing so involves one more subtlety. Up to now, I have assumed that the range of a plausibility measure is partially ordered. To get the result, I need to allow it to be partially preordered. That allows two sets that have equivalent plausibility to act differently when combined using and in the computation of expected utility. So, for this result, I assume that the relation on the range of a plausibility measure is reflexive and transitive, but not necessarily antisymmetric. With this assumption, GEU ordinally represents BEL (Exericse 5.45). In fact, an even more general result holds.

Theorem 5.4.6

start example

If is a decision rule that respects utility, then GEU ordinally represents .

end example

Proof See Exercise 5.46.

Theorem 5.4.6 shows that GEU ordinally represents essentially all decision rules. Thus, there is a sense in which GEU can be viewed as a universal decision rule.

Thinking of decision rules as instances of expected utility maximization gives a new perspective on them. The relationship between various properties of an expectation domain and properties of decision rules can then be studied. To date, there has been no work on this topic, but it seems like a promising line of inquiry.




Reasoning About Uncertainty
Reasoning about Uncertainty
ISBN: 0262582597
EAN: 2147483647
Year: 2005
Pages: 140

flylib.com © 2008-2017.
If you may any questions please contact us: flylib@qtcs.net