3.8 Utility Theory

We have focused on the EMV criterion for making decisions under risk. However, there are many occasions in which people make decisions that would appear to be inconsistent with the EMV criterion. When people buy insurance, the amount of the premium is greater than the expected payout to them from the insurance company because the premium includes the expected payout, the overhead cost, and the profit for the insurance company. A person involved in a lawsuit may choose to settle out of court rather than go to trial even if the expected value of going to trial is greater than the proposed settlement. A person buys a lottery ticket even though the expected return is negative. Casino games of all types have negative expected returns for the player, and yet millions of people play these games. A businessperson may rule out one potential decision because it could bankrupt the firm if things go bad, even though the expected return for this decision is better than that of all other alternatives.

Why do people make decisions that don’t maximize their EMV? They do this because the monetary value is not always a true indicator of the overall value of the result of the decision. The overall worth of a particular outcome is called utility, and rational people make decisions that maximize the expected utility. Although at times the monetary value is a good indicator of utility, there are other times when it is not. This is particularly true when some of the values involve an extremely large payoff or an extremely large loss. Suppose that you are the lucky holder of a lottery ticket. Five minutes from now a fair coin could be flipped, and if it comes up tails, you would win $5 million. If it comes up heads, you would win nothing. Just a moment ago a wealthy person offered you $2 million for your ticket. Let’s assume that you have no doubts about the validity of the offer. The person will give you a certified check for the full amount, and you are absolutely sure the check would be good.

A decision tree for this situation is shown in Figure 3.6. The EMV for rejecting the offer indicates that you should hold on to your ticket, but what would you do? Just think, $2 million for sure instead of a 50% chance at nothing. Suppose you were greedy enough to hold on to the ticket, and then lost. How would you explain that to your friends? Wouldn’t $2 million be enough to be comfortable for a while?

Most people would choose to sell the ticket for $2 million. Most of us, in fact, would probably be willing to settle for a lot less. Just how low we would go is, of course, a matter of personal preference. People have different feelings about seeking or avoiding risk. Using the EMV alone is not always a good way to make these types of decisions.

One way to incorporate your own attitudes toward risk is through utility theory. In the next section, we explore first how to measure utility and then how to use utility measures in decision making.

Measuring Utility and Constructing a Utility Curve

The first step in using utility theory is to assign utility values to each monetary value in a given situation. It is convenient to begin utility assessment by assigning the worst outcome a utility of 0 and the best outcome a utility of 1. Although any values may be used as long as the utility for the best outcome is greater than the utility for the worst outcome, using 0 and 1 has some benefits. Because we have chosen to use 0 and 1, all other outcomes will have a utility value between 0 and 1. In determining the utilities of all outcomes, other than the best or worst outcome, a standard gamble is considered. This gamble is shown in Figure 3.7.

Two branches, Accept Offer and Reject Offer, stem from the square decision node.

Figure 3.6 Your Decision Tree for the Lottery Ticket

Two branches, Alternative 1 and Alternative 2, stem from the square decision node.

Figure 3.7 Standard Gamble for Utility Assessment

In Figure 3.7, p is the probability of obtaining the best outcome, and (1p) is the probability of obtaining the worst outcome. Assessing the utility of any other outcome involves determining the probability (p) that makes you indifferent between alternative 1, which is the gamble between the best and worst outcomes, and alternative 2, which is obtaining the other outcome for sure. When you are indifferent between alternatives 1 and 2, the expected utilities for these two alternatives must be equal. This relationship is shown as

Expected utility of alternative 2=Expected utility of alternative 1Utility of other outcome=(p)(Utility of best outcome, which is 1)=+ (1p)(Utility of the worst outcome, which is 0)Utility of other outcome=(p)(1) + (1p)(0) = p
(3-7)

Now all you have to do is to determine the value of the probability (p) that makes you indifferent between alternatives 1 and 2. In setting the probability, you should be aware that utility assessment is completely subjective. It’s a value set by the decision maker that can’t be measured on an objective scale. Let’s take a look at an example.

Jane Dickson would like to construct a utility curve revealing her preference for money between $0 and $10,000. A utility curve is a graph that plots utility value versus monetary value. She can invest her money either in a bank savings account or in a real estate deal.

If the money is invested in the bank, in 3 years Jane would have $5,000. If she invested in the real estate, after 3 years she could either have nothing or have $10,000. Jane, however, is very conservative. Unless there is an 80% chance of getting $10,000 from the real estate deal, Jane would prefer to have her money in the bank, where it is safe. What Jane has done here is to assess her utility for $5,000. When there is an 80% chance (this means that p is 0.8) of getting $10,000, Jane is indifferent between putting her money in real estate and putting it in the bank. Jane’s utility for $5,000 is thus equal to 0.8, which is the same as the value for p. This utility assessment is shown in Figure 3.8.

Other utility values can be assessed in the same way. For example, what is Jane’s utility for $7,000? What value of p would make Jane indifferent between $7,000 and the gamble that would result in either $10,000 or $0? For Jane, there must be a 90% chance of getting the $10,000. Otherwise, she would prefer the $7,000 for sure. Thus, her utility for $7,000 is 0.90. Jane’s utility for $3,000 can be determined in the same way. If there were a 50% chance of obtaining the $10,000, Jane would be indifferent between having $3,000 for sure and taking the gamble of either winning the $10,000 or getting nothing. Thus, the utility of $3,000 for Jane is 0.5. Of course, this process can be continued until Jane has assessed her utility for as many monetary values as she wants. These assessments, however, are enough to get an idea of Jane’s feelings toward risk. In fact, we can plot these points in a utility curve, as is done in Figure 3.9. In the figure, the assessed utility points of $3,000, $5,000, and $7,000 are shown by dots, and the rest of the curve is inferred from these.

The Utility Outcome decision tree template from the previous figure is replicated here, with actual data inserted into the tree.

Figure 3.8 Utility of $5,000

A line graph with Monetary Value,

Figure 3.9 Utility Curve for Jane Dickson

Jane’s utility curve is typical of a risk avoider. A risk avoider is a decision maker who gets less utility or pleasure from a greater risk and tends to avoid situations in which high losses might occur. As monetary value increases on her utility curve, the utility increases at a slower rate.

Figure 3.10 illustrates that a person who is a risk seeker has an opposite-shaped utility curve. This decision maker gets more utility from a greater risk and higher potential payoff. As monetary value increases on his or her utility curve, the utility increases at an increasing rate. A person who is indifferent to risk has a utility curve that is a straight line. The shape of a person’s utility curve depends on the specific decision being considered, the monetary values involved in the situation, the person’s psychological frame of mind, and how the person feels about the future. It may well be that you have one utility curve for some situations you face and completely different curves for others.

Utility as a Decision-Making Criterion

After a utility curve has been determined, the utility values from the curve are used in making decisions. Monetary outcomes or values are replaced with the appropriate utility values, and then decision analysis is performed as usual. The expected utility for each alternative is computed instead of the EMV. Let’s take a look at an example in which a decision tree is used and expected utility values are computed in selecting the best alternative.

Line graph with Monetary Outcome on the horizontal axis and Utility on the vertical axis.

Figure 3.10 Preferences for Risk

Mark Simkin loves to gamble. He decides to play a game that involves tossing thumbtacks in the air. If the point on the thumbtack is facing up after it lands, Mark wins $10,000. If the point on the thumbtack is down, Mark loses $10,000. Should Mark play the game (alternative 1), or should he not play the game (alternative 2)?

Alternatives 1 and 2 are displayed in the tree shown in Figure 3.11. As can be seen, alternative 1 is to play the game. Mark believes that there is a 45% chance of winning $10,000 and a 55% chance of suffering the $10,000 loss. Alternative 2 is not to gamble. What should Mark do? Of course, this depends on Mark’s utility for money. As stated previously, he likes to gamble. Using the procedure just outlined, Mark was able to construct a utility curve showing his preference for money. Mark has a total of $20,000 to gamble, so he has constructed the utility curve based on a best payoff of $20,000 and a worst payoff of a $20,000 loss. This curve appears in Figure 3.12.

We see that Mark’s utility for –$10,000 is 0.05, his utility for not playing ($0) is 0.15, and his utility for $10,000 is 0.30. These values can now be used in the decision tree. Mark’s objective is to maximize his expected utility, which can be done as follows:

  1. Step 1.

    U($10,000) = 0.05U($0) = 0.15U($10,000) = 0.30
  2. Step 2. Replace monetary values with utility values. Refer to Figure 3.13 . Here are the expected utilities for alternatives 1 and 2:

    Two branches stem from the square decision node.

    Figure 3.11 Decision Facing Mark Simkin

    A line graph showing Monetary Outcome on the horizontal axis; it ranges from negative twenty thousand dollars to positive twenty thousand dollars in increments of ten thousand dollars.

    Figure 3.12 Utility Curve for Mark Simkin

    The decision tree for Mark Simkin playing or not playing the game is replicated from a prior figure.

    Figure 3.13 Using Expected Utilities in Decision Making

    E(alternative 1: play the game)=(0.45)(0.30) + (0.55)(0.05)=0.135 + 0.027 = 0.162E(alternative 2: don't play the game)=0.15

    Therefore, alternative 1 is the best strategy using utility as the decision criterion. If EMV had been used, alternative 2 would have been the best strategy. The utility curve is a risk-seeker utility curve, and the choice of playing the game certainly reflects this preference for risk.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset