Chapter 17Behavioral and Experimental Economics

Introduction

Every model in this book has assumed rational agents — consumers who maximize expected utility, firms that minimize costs, agents with consistent time preferences and correct beliefs. This chapter asks: what if these assumptions are systematically wrong?

Behavioral economics documents predictable deviations from the standard model: people are loss-averse, overweight small probabilities, discount the future inconsistently, and are influenced by framing and context. The question is not whether people are "irrational" — it's whether the deviations are systematic enough to improve our models and inform better policy.

By the end of this chapter, you will be able to:
  1. State the key features of prospect theory and contrast it with expected utility
  2. Identify the Allais and Ellsberg paradoxes as violations of expected utility axioms
  3. Model present bias using the β-δ framework and derive its implications
  4. Evaluate when behavioral biases survive market aggregation
  5. Apply nudge theory and libertarian paternalism to policy design

17.1 Violations of Expected Utility

Expected utility (review). A theory of decision under risk where an agent chooses the lottery that maximizes $EU = \sum p_i u(x_i)$. It requires the independence axiom: preferences between two lotteries should not depend on a common third component. EU is the benchmark against which behavioral deviations are measured.
Independence axiom. If lottery $A$ is preferred to $B$, then a mixture $pA + (1-p)C$ must be preferred to $pB + (1-p)C$ for any lottery $C$ and probability $p$. The Allais paradox demonstrates systematic violations of this axiom.
Allais paradox. The empirical finding (Allais, 1953) that most people prefer a certain \$1M over a risky gamble with higher expected value (certainty effect), yet simultaneously prefer a riskier gamble when both options involve uncertainty. These joint preferences violate the independence axiom of expected utility theory.
Ellsberg paradox. The empirical finding (Ellsberg, 1961) that people prefer gambles with known probabilities over gambles with unknown (ambiguous) probabilities, even when EU theory predicts indifference. This reveals ambiguity aversion — a preference for known risks over unknown risks — which EU cannot accommodate.
Ambiguity aversion. The preference for known probabilities over unknown ones. An ambiguity-averse agent prefers a 50/50 gamble from a known urn over an equivalent gamble from an urn with unknown composition. This violates the Savage axioms underlying subjective expected utility.

Under the axioms of Chapter 10 (completeness, transitivity, continuity) plus the independence axiom, preferences over lotteries can be represented by expected utility:

$$EU = \sum_i p_i \cdot u(x_i)$$ (Eq. 17.1)

The Allais Paradox (1953)

Gamble A: \$1,000,000 with certainty. Gamble B: 89% chance of \$1M; 10% chance of \$1M; 1% chance of \$1. Most people choose A.

Gamble C: 11% chance of \$1M; 89% chance of \$1. Gamble D: 10% chance of \$1M; 90% chance of \$1. Most people choose D.

But $A \succ B$ and $D \succ C$ together violate the independence axiom.

Interactive: Allais Paradox Calculator

Select your preferred gamble in each pair, then see whether your choices are consistent with expected utility theory.

GambleProbabilities & Payoffs
A100% chance of \$1M
B89% × \$1M + 10% × \$1M + 1% × \$1
C11% × \$1M + 89% × \$1
D10% × \$1M + 90% × \$1

Pair 1: A vs B — I prefer:

Pair 2: C vs D — I prefer:

Near risk-neutral (0.05)Very risk-averse (0.95)
Select your preferred gamble in each pair to see the analysis.

Figure 17.A. Expected utility of each gamble under power utility $u(x) = x^{1-r}/(1-r)$. The slider varies the risk aversion parameter $r$. If your choices are A and D (the common Allais pattern), no value of $r$ can rationalize both preferences simultaneously — the independence axiom is violated.

The Ellsberg Paradox (1961)

An urn contains 30 red balls and 60 balls that are either black or yellow (unknown ratio). People prefer known probabilities over unknown ones — revealing ambiguity aversion that EU cannot accommodate.

17.2 Prospect Theory

Kahneman and Tversky (1979) proposed prospect theory as a descriptive alternative to expected utility.

Prospect theory. A descriptive theory of decision under risk (Kahneman and Tversky, 1979) that replaces EU with four key modifications: reference dependence, loss aversion, diminishing sensitivity, and probability weighting. It captures systematic patterns in human choice that EU cannot.
Value function (S-curve). The prospect theory analog of the utility function. It is defined over gains and losses relative to a reference point, concave for gains (risk aversion), convex for losses (risk seeking), and steeper for losses than gains (loss aversion). The kink at the reference point captures the asymmetry between gains and losses.
Loss aversion. The empirical finding that losses loom larger than equivalent gains: $|v(-x)| > v(x)$ for $x > 0$. The loss aversion coefficient $\lambda \approx 2.25$ means losing \$100 feels about 2.25 times worse than gaining \$100 feels good. This explains the endowment effect, the status quo bias, and the disposition effect in finance.
Probability weighting. The distortion of objective probabilities in decision-making: $\pi(p) \neq p$. Small probabilities are overweighted ($\pi(0.01) > 0.01$), explaining lottery purchases. Large probabilities are underweighted ($\pi(0.99) < 0.99$), explaining insurance purchases. At $\delta = 1$, weighting reduces to EU.
Reference dependence. The principle that outcomes are evaluated as gains or losses relative to a reference point, not as final wealth states. The reference point is typically the status quo, but can be expectations, aspirations, or social comparisons. Reference dependence means the same objective outcome can be experienced as a gain or loss depending on context.
Endowment effect. The tendency to value an item more highly once you own it than you would pay to acquire it. In experiments, the selling price (WTA) exceeds the buying price (WTP) by a factor of 2–3, consistent with loss aversion: giving up an owned item is a loss, while acquiring it is a gain.
Framing effect. The phenomenon where the way a choice is presented (framed) affects decisions, even when the objective outcomes are identical. For example, people prefer "90% survival rate" over "10% mortality rate." Framing violates the descriptive invariance principle of rational choice.
Mental accounting. The cognitive process of organizing financial decisions into separate "accounts" (e.g., vacation fund, emergency fund) rather than treating wealth as fungible. Mental accounting leads to violations of standard theory: people may simultaneously hold credit card debt at 18% and savings at 2%, because the accounts are psychologically separate.
$$v(x) = \begin{cases} x^\gamma & \text{if } x \geq 0 \\ -\lambda(-x)^\gamma & \text{if } x < 0 \end{cases}$$ (Eq. 17.2)

with $\gamma \approx 0.88$ and $\lambda \approx 2.25$.

Interactive: Prospect Theory Value Function

The value function is S-shaped: concave for gains (risk aversion), convex for losses (risk seeking), and steeper for losses than gains (loss aversion). Compare to the linear EU value function.

Very curved (0.20)Linear (1.00)
No loss aversion (1.0)Extreme (4.0)
At x = +100: v(100) = 57.5  |  At x = -100: v(-100) = -129.5  |  Ratio |v(-100)/v(100)|: 2.25

Figure 17.1. The prospect theory value function (blue S-curve) versus expected utility (gray straight line). The kink at the origin reflects loss aversion — the slope is steeper on the loss side. Higher $\lambda$ makes losses more painful; lower $\gamma$ increases curvature. Drag sliders to reshape the function.

Probability Weighting

People do not weight outcomes by their true probabilities:

$$\pi(p) = \frac{p^\delta}{(p^\delta + (1-p)^\delta)^{1/\delta}}$$ (Eq. 17.3)

with $\delta \approx 0.65$. Small probabilities are overweighted (explaining lottery ticket purchases); large probabilities are underweighted (explaining insurance against near-certain losses).

Interactive: Probability Weighting Function

Compare the weighted probability $\pi(p)$ to the true probability (the 45-degree line). Where the curve is above the diagonal, people act as if the probability is higher than it really is.

Extreme distortion (0.20)No distortion (1.00)
π(0.01) = 0.066 (6.6x overweight)  |  π(0.50) = 0.42  |  π(0.99) = 0.91 (underweight)

Figure 17.2. The probability weighting function. Above the 45-degree line: overweighting (small probabilities seem larger than they are). Below: underweighting (large probabilities seem smaller). At $\delta = 1$, the curve collapses to the diagonal — no distortion. Drag the slider to explore.

Prospect theory valuation:

$$V = \sum_i \pi(p_i) \cdot v(x_i)$$ (Eq. 17.4)

Applications

Endowment effect: People demand more to sell an owned object than they'd pay to acquire it. Equity premium puzzle: Myopic loss aversion with short evaluation horizons explains the large stock-bond return gap. Insurance and gambling: The same person buys insurance (loss domain, concave) and lottery tickets (overweighted small-probability gains).

Example 17.1 — EU vs Prospect Theory: Certainty Equivalents

A gamble offers a 50% chance of winning \$100 and a 50% chance of losing \$100. Compare evaluations.

Expected Utility (CRRA with $r = 0.5$, $W = 1000$): $EU = 0.5 \cdot u(1200) + 0.5 \cdot u(900) = 0.5 \times 1200^{0.5} + 0.5 \times 900^{0.5} = 0.5(34.64) + 0.5(30.00) = 32.32$. Certainty equivalent: \$12.32^2 = 1044.6$. Net CE gain: \$14.6 > 0$. Take the gamble.

Prospect Theory ($\gamma = 0.88$, $\lambda = 2.25$, $\pi(0.5) = 0.42$):

$V = \pi(0.5) \cdot v(200) + \pi(0.5) \cdot v(-100)$

$= 0.42 \times 200^{0.88} + 0.42 \times (-2.25)(100^{0.88})$

$= 0.42 \times 138.4 + 0.42 \times (-2.25 \times 72.4) = 58.1 - 68.5 = -10.4 < 0$. Reject the gamble.

Key insight: Loss aversion flips the decision. EU says the positive expected value makes this attractive. Prospect theory says the \$100 loss looms larger than the \$100 gain — consistent with the empirical observation that most people reject such gambles.

17.3 Present Bias and Hyperbolic Discounting

Present bias. The tendency to overweight immediate payoffs relative to future payoffs, beyond what exponential discounting implies. A present-biased agent may prefer \$100 today over \$110 tomorrow, but prefer \$110 in 31 days over \$100 in 30 days — a preference reversal that violates time consistency.
Beta-delta discounting. The quasi-hyperbolic model $U_0 = u_0 + \beta\sum_{t=1}^\infty\delta^t u_t$, where $\beta < 1$ captures present bias and $\delta$ captures long-run patience. When $\beta = 1$, the model reduces to standard exponential discounting. Typical estimates: $\beta \approx 0.7$, $\delta \approx 0.95$.
Naive vs sophisticated agent. A naive agent does not anticipate their own future present bias — they plan to act optimally tomorrow but procrastinate when tomorrow arrives. A sophisticated agent correctly predicts their future bias and may seek commitment devices, but may also give up on tasks they know their future self will not complete.
Commitment device. A mechanism that restricts future choices to overcome present bias. Examples: automatic savings plans (you cannot easily withdraw), deadlines (penalties for procrastination), and Odysseus-style pre-commitment. A commitment device has positive value for a present-biased agent who recognizes their bias (sophisticated) but zero value for a time-consistent agent.

The β-δ Model

$$U_0 = u_0 + \beta \sum_{t=1}^\infty \delta^t u_t$$ (Eq. 17.5)

where $\beta < 1$ captures present bias. The discount factor between now and next period is $\beta\delta$, but between any two future periods is just $\delta$. This creates time inconsistency: today you plan to start exercising tomorrow; tomorrow you prefer the day after.

Naive present-biased agents don't recognize their future self-control problem. Sophisticated agents recognize their bias and seek commitment devices.

Interactive: β-δ Discounting Explorer

A task costs 6 utils today but yields 8 utils of benefit arriving in 3 days. A present-biased agent keeps planning to do it "tomorrow" but never does. A sophisticated agent recognizes the pattern.

Severe bias (0.10)No bias (1.00)
Impatient (0.70)Patient (1.00)
Naive agent: Plans to do it on day 2, but procrastinates  |  Sophisticated agent: Does it on day 1 (knows future self will procrastinate)

Figure 17.3. Discounted value of doing the task on each day, as seen from that day (blue) vs. as seen from the day before (orange). The gap is present bias — the task always looks better when it's "tomorrow" than when it's "today." Naive agents keep postponing; sophisticated agents anticipate their future selves' behavior. Drag sliders to explore.

Example 17.2 — Beta-Delta Procrastination

A student must write a paper. Cost of doing it today: $c = 10$ utils. Benefit (received at submission in 7 days): $b = 20$ utils. Parameters: $\beta = 0.6$, $\delta = 0.99$.

Step 1 (Day 1, perspective of Day 1): Do it now: $-10 + \beta\delta^7 \times 20 = -10 + 0.6 \times 0.93 \times 20 = -10 + 11.2 = 1.2 > 0$. Looks worth doing!

Step 2 (Day 1, re-evaluation): Wait until tomorrow: $\beta\delta \times (-10) + \beta\delta^7 \times 20 = 0.6 \times 0.99 \times (-10) + 0.6 \times 0.93 \times 20 = -5.9 + 11.2 = 5.3$. Waiting looks even better! The naive agent delays.

Step 3 (Day 2, from Day 2's perspective): The same calculation repeats: doing it today still has net value \$1.2$, but waiting has \$1.3$. The agent procrastinates again — and again, and again.

Naive outcome: The student never does the paper until the deadline forces action (or misses the deadline entirely).

Sophisticated outcome: Knowing future selves will procrastinate, the sophisticated agent recognizes that "do it tomorrow" means "never." If the deadline binds at Day 7, the sophisticated agent may set an artificial deadline or accept the immediate cost on Day 1.

Example 17.3 — Value of a Commitment Savings Account

An agent earns \$1,000/month and wants to save \$100/month for retirement. Parameters: $\beta = 0.7$, $\delta = 0.95$, $r = 5\%$/year.

Without commitment: Each month, the agent plans to save \$100 but faces the temptation to spend. The immediate utility of spending \$100: $u(200) = 200^{0.5} = 14.1$. The discounted future benefit of saving: $\beta\delta^{12} \times u(200 \times 1.05) = 0.7 \times 0.54 \times 14.5 = 5.5$. Since \$14.1 > 5.5$, the agent spends the \$100 every month.

With commitment device: An illiquid savings account automatically deducts \$100/month. The agent cannot access the money for 12 months. From the perspective of enrollment: $PV(\text{annual savings at } r=5\%) = 200 \times 12 \times 1.05 = 2,520$. The agent's long-run self values this highly.

Value of commitment: The difference between the committed outcome (\$1,520$ saved) and the uncommitted outcome (\$1$ saved) is the value of the commitment device. The agent would pay up to $\beta \times PV - 0 = 0.7 \times 2,520 = 1,764$ in present-biased terms to have the option.

17.4 Experimental Economics

The Ultimatum Game

Setup: Player 1 proposes how to split \$10. Player 2 accepts (both get the amounts) or rejects (both get nothing).

Subgame perfect equilibrium: Player 1 offers \$1.01; Player 2 accepts.

Actual behavior: Modal offer is 40–50%. Offers below 20% are rejected about half the time. People sacrifice real money to punish unfairness — suggesting utility functions include fairness and reciprocity.

Interactive: Ultimatum Game Simulator

You are Player 1. Propose a split of \$10. The computer (Player 2) accepts or rejects based on a fairness threshold. How much do you need to offer to avoid rejection?

Rational (0%)Moderately fairStrict (50%)
Your offer to Player 2: $5.00
\$1 (keep all)\$1 (equal split)\$10 (give all)
Make your first offer...
Rounds played: 0  |  Your total earnings: \$1.00  |  Acceptance rate: --%

Figure 17.4. Your earnings per round. Green bars: accepted offers. Red bars: rejected offers (\$1 for both). The rational strategy is to offer just above the threshold — but in real experiments, people offer much more than the minimum. Play multiple rounds to see the pattern.

17.5 Nudge Theory and Policy Implications

Nudge. A feature of the choice environment that predictably alters behavior without forbidding any options or significantly changing economic incentives (Thaler and Sunstein, 2008). Examples include default enrollment in retirement savings, calorie labels on menus, and organ donation opt-out policies.
Libertarian paternalism. The philosophy that it is legitimate to influence behavior through choice architecture (nudges) while preserving freedom of choice. "Libertarian" because no options are removed; "paternalism" because the design steers people toward choices judged to be in their interest.
Choice architecture. The design of the environment in which choices are made, including defaults, ordering, framing, and simplification. Choice architecture is never neutral — some design always exists — so the question is whether to design it intentionally for good outcomes.
Nudge (Thaler & Sunstein, 2008). A change in the choice architecture that alters behavior in predictable ways without restricting options or significantly changing economic incentives. Libertarian paternalism: influence behavior (paternalism) while preserving freedom of choice (libertarianism).
NudgeBias AddressedOutcome
Default enrollment in 401(k)Procrastination, status quo biasParticipation: ~50% → ~90%
Save More TomorrowPresent biasSavings rates nearly quadruple
Opt-out organ donationStatus quo biasConsent: ~15% → ~85%
Social norms messagingConformity2–4% energy reduction
Simplified financial aid formsComplexity aversion+8pp college enrollment

Interactive: Nudge Default Effect

Two identical programs — same benefits, same freedom to choose. The only difference is the default. With opt-in, people must actively enroll. With opt-out, people must actively leave. Small switching costs create enormous differences in participation.

Zero effort (0)High effort (10)
Opt-in participation: 42%  |  Opt-out participation: 88%  |  Gap from default alone: 46pp

Figure 17.5. Participation rates under opt-in vs. opt-out defaults. At zero switching cost, both converge to the "true preference" rate. As switching cost rises, each default becomes stickier — fewer people switch away from whatever the default is. The policy implication: set the default to the socially beneficial option. Drag the slider to vary switching costs.

Example 17.4 — Designing a Nudge for Retirement Savings

A company with 10,000 employees wants to increase 401(k) participation. Current opt-in rate: 40%. Average contribution rate among participants: 6% of salary.

Step 1 (Diagnosis): The low opt-in rate is consistent with status quo bias and present bias. Employees intend to enroll but procrastinate. The default (not enrolled) is the problem.

Step 2 (Nudge design — auto-enrollment): Change the default to automatic enrollment at 3% contribution rate. Employees can opt out at any time (preserving libertarian criterion).

Step 3 (Predicted effect): With switching cost $e = 3$ on a 0–10 scale: opt-out participation $\approx 90\%$ vs. opt-in $\approx 40\%$. The 50-percentage-point gap is entirely due to the default — the economic incentives are unchanged.

Step 4 (Auto-escalation): Add automatic contribution increase of 1% per year until reaching 10%. Present-biased agents do not opt out of gradual increases because each increment is small.

Step 5 (Evidence): Madrian and Shea (2001) found that auto-enrollment raised 401(k) participation from 37% to 86% at one company. Thaler and Benartzi's "Save More Tomorrow" program increased contribution rates from 3.5% to 13.6% over 40 months.

Behavioral welfare economics. The subfield that addresses how to evaluate welfare when agents have biased preferences (present bias, framing dependence, etc.). The Bernheim-Rangel framework distinguishes "revealed preferences" from "welfare-relevant preferences," allowing policy evaluation even when choices are systematically biased.

17.6 Markets and Behavioral Biases

Arguments that markets correct biases: Arbitrageurs exploit mispricing; competition punishes irrational firms; experience teaches better decisions.

Arguments that biases persist: Limits to arbitrage (short-selling constraints, noise trader risk); some biases are robust to experience (loss aversion among professional traders); market prices can reflect aggregate biases (financial bubbles).

The evidence is mixed. Financial markets are approximately efficient for liquid assets, less so for complex or illiquid ones. Consumer markets show persistent behavioral patterns.

Thread: Maya's Enterprise

Maya added a free cookie with every lemonade. Sales increased 15%. She later removed it. Rational prediction: Customers should be indifferent (if cookie worth \$1.25 and price adjusts). Behavioral prediction: Removing the cookie is a loss, weighted $\lambda \approx 2.25$x. Sales dropped 20% — far more than the 15% gain from introducing it.

Lesson: It is easier to add a benefit than to remove one. Loss aversion means "taking away" is not the mirror image of "giving."

Summary

Key Equations

LabelEquationDescription
Eq. 17.1$EU = \sum p_i u(x_i)$Expected utility
Eq. 17.2$v(x) = x^\gamma$ for gains; $-\lambda(-x)^\gamma$ for lossesProspect theory value function
Eq. 17.3$\pi(p) = \frac{p^\delta}{(p^\delta + (1-p)^\delta)^{1/\delta}}$Probability weighting
Eq. 17.4$V = \sum \pi(p_i) v(x_i)$Prospect theory valuation
Eq. 17.5$U_0 = u_0 + \beta\sum\delta^t u_t$Quasi-hyperbolic discounting

Exercises

Practice

  1. A gamble offers a 50% chance of winning \$100 and a 50% chance of losing \$10. Evaluate under (a) expected value, (b) EU with $u(x) = \sqrt{x}$ and wealth $W = 500$, (c) prospect theory with $\gamma = 0.88$, $\lambda = 2.25$, $\pi(0.5) = 0.42$.
  2. Show that the Allais paradox choices $A \succ B$ and $D \succ C$ violate the independence axiom.
  3. An agent with $\beta = 0.6$, $\delta = 1$ chooses between 10 utils today and 15 utils tomorrow. (a) What does she choose today? (b) What would she have chosen yesterday? (c) Is she time-consistent?
  4. Design a nudge to increase hand-washing compliance in a hospital.

Apply

  1. Using prospect theory, explain why customers resist premium increases (\$10 → \$10) more than they appreciate equivalent decreases (\$10 → \$10). If $\lambda = 2.25$, how much must the decrease be to match the emotional impact of a \$10 increase?
  2. Explain how myopic loss aversion can resolve the equity premium puzzle with reasonable loss aversion ($\lambda = 2.25$).
  3. Evaluate whether nudges are manipulative or welfare-enhancing. Under what conditions do nudges cross the line from "helpful defaults" to "manipulation"?

Challenge

  1. Prove that if an agent satisfies the independence axiom, they cannot exhibit the Allais paradox.
  2. In a three-period β-δ model, find the subgame-perfect equilibrium for a sophisticated agent who must decide when to complete a costly task.
  3. Construct a formal model where mental accounting leads to a specific economic inefficiency and show the inefficiency disappears under EU with fungible wealth.