15 Oct 2015
Suppose a firm’s owner contracts a manager to run the firm. The firm’s profit is a random variable \(\pi \in [\pi_a , \pi_b]\) which the manager can influence through some effort. For simplicity, assume effort is discrete: \(e \in \{e_l,e_h\}\), with \(e_h \gt e_l \ge 0\). The manager’s efforts are such that profits with \(e_h\) first-order stochastically dominate profits with \(e_l\), i.e. \(E \ [\pi \|e_h] \gt E \ [\pi \|e_l]\) and \(f(\pi \|e) \gt 0 ~~ \forall \pi\). The manager’s utility is \(u(w,e) = v(w) - g(e)\) with \(g(e_h) \gt g(e_l)\) and \(v'(w) \gt 0, v''(w) \lt 0\). The manager’s utility strictly concave, so they are risk-averse. Let’s assume the manager’s reservation utility is \(\bar{u} \ge 0\).
The idea is that the manager would prefer to exert less effort to more, but that more effort is likely to produce higher profits than less effort. No manager can control everything - profit is still a random variable and low realizations can occur when the manager exerts more effort, it’s just less likely. The manager’s reservaton utility is their outside offer or what they could produce at home. For clarity, let’s call the owner she and the manager he.
Benchmark case: effort is publicly observable
If the manager’s efforts are publicly observable, then the owner solves
\[\begin{align}
\max_{e \in \{e_l, e_h \}, w( \pi )} ~~& \int_{\pi_a}^{\pi_b} [\pi - w( \pi )] f(\pi | e) ~ d \pi \cr
\text{s.t.} ~~& \int_{\pi_a}^{\pi_b} v(w(\pi )) f(\pi | e) ~ d \pi - g(e) \ge \bar{u}
\end{align}\]
We can think of this as a two-stage decision process: given some level of effort from the manager, the owner wants to pay him the profit-maximizing wage. The constraint is called an “individual rationality (IR) constraint” or a “participation constraint”. It’s a way to factor the manager’s problem into the owner’s decision-making - the owner is constrained to pay the manager a wage that is at least as good as the manager’s outside offer. If the owner doesn’t, the manager won’t work for her. For a fixed level of effort \(e\), the owner’s problem can be simplified to minimizing the manager’s wage subject to the participation constraint.
\[\begin{align}
\min_{w( \pi )} ~~& \int_{\pi_a}^{\pi_b} w( \pi ) f(\pi | e) ~ d \pi \cr
\text{s.t.} ~~& \int_{\pi_a}^{\pi_b} v(w(\pi )) f(\pi | e) ~ d \pi - g(e) \ge \bar{u}
\end{align}\]
A minimization problem can be made into a maximization problem by multiplying the objective by \(-1\). Economists seem to prefer maximization, while statisticians and computer scientists seem to prefer minimization.
Since we integrated over \(\pi\) this wage holds for all realizations of profit. We can solve this as a constrained maximization
\[\begin{align}
\mathcal{L} = & - \int_{\pi_a}^{\pi_b} w( \pi ) f(\pi | e) ~ d \pi + \lambda [\int_{\pi_a}^{\pi_b} v(w(\pi )) f(\pi | e) ~ d \pi - g(e) - \bar{u}] \cr
\text{FOC:} ~~ & - \int_{\pi_a}^{\pi_b} f(\pi | e) ~ d \pi + \lambda \int_{\pi_a}^{\pi_b} v'(w(\pi )) f(\pi | e) ~ d \pi = 0 \cr
\implies ~~ & \frac{1}{v'(w(\pi ))} = \lambda
\end{align}\]
From this we see that for a given level of effort, the optimal wage must be a constant. We can think of this as insurance against profit risk: the optimal contract will insure the risk-averse manager against any realization of profit, since the manager’s effort doesn’t guarantee high profits, just makes it more likely. The optimal wage profile for any level of effort is to pay the manager the inverse wage-utility of the sum of their reservation utility and their marginal cost of effort, or \(w^*_e = v^{-1}(\bar{u} + g(e))\).
The lagrange multiplier \(\lambda\) on the participation constraint is the owner’s shadow price of the manager’s participation: it’s the owner’s (maximum) willingness-to-pay to marginally reduce the manager’s reservation utility. More generally, the lagrange multiplier on a constraint is the improvement to the objective function from marginally relaxing that constraint.
The IR constraint must bind in this problem if the manager works for the owner. Any excess utility the owner gives the manager reduces the owner’s profit, so the owner can always do better by reducing the manager’s wage until the constraint binds.
Now suppose the manager’s effort level is not publicly observable anymore, it’s the manager’s private information. Now the owner solves
\[\begin{align}
\min_{w( \pi )} ~~& \int_{\pi_a}^{\pi_b} w( \pi ) f(\pi | e) ~ d \pi \cr
\text{s.t.} ~~& \int_{\pi_a}^{\pi_b} v(w(\pi )) f(\pi | e) ~ d \pi - g(e) \ge \bar{u} \cr
& e = \text{argmax} ~~ \int_{\pi_a}^{\pi_b} v(w(\pi )) f(\pi | e) ~ d \pi - g(e) \cr
\end{align}\]
The second constraint is called an “incentive compatibility” (IC) constraint. It says that the effort level the owner wants to implement has to be solve the manager’s problem. I think of this as a refinement on the IR constraint - not only should it make sense for the manager to participate, it should be what they would most prefer to do. The manager should have no incentive to deviate. It’s not quite the same as the IR constraint, since it’s possible for the argmax of the manager’s problem under effort is lower than the reservation utility.
Implementing \(e_l\) is easy - the owner just pays the manager the same wage as under observable effort. It satisfies the IR constraint, and at that wage the manager would put in low effort anyway (there is no profitable deviation). Implementing \(e_h\) is more interesting, since the IC constraint becomes important. For this problem, the IC implies
\[\int_{\pi_a}^{\pi_b} v(w(\pi )) f(\pi | e_h ) ~ d \pi - g(e_h ) \ge \int_{\pi_a}^{\pi_b} v(w(\pi )) f(\pi | e_l ) ~ d \pi - g(e_l )\]
So the manager’s expected value of putting in high effort has to be greater than or equal to his expected value of putting in low effort. If not, the manager has an incentive to deviate and since effort is unobservable, he will.
As before, we put it into a lagrangian. From the first-order condition of the owner’s problem,
\[\begin{align}
w(\pi) : ~~ & - \int_{\pi_a}^{\pi_b} f(\pi | e_h ) ~ d \pi + \lambda \int_{\pi_a}^{\pi_b} v'(w(\pi )) f(\pi | e_h ) ~ d\pi \cr + \ & \gamma \int_{\pi_a}^{\pi_b} [f(\pi | e_h )-f(\pi | e_l )] \ v'(w(\pi)) = 0 \cr
\implies ~~ & \frac{1}{v'(w(\pi ))} = \lambda + \gamma \left[ 1 - \frac{f(\pi|e_l )}{f(\pi|e_h )} \right]
\end{align}\]
The condition on the wage has the same component of the IR constraint multiplier - the shadow price of the manager’s participation - and a new component from the IC constraint based on the shadow price of the manager’s cooperation. This component is not a constant - it depends on the realizations of profit through the ratio of conditional densities of profit given effort. In repeated play, this term is a likelihood ratio.
By the same argument as earlier, the IR constraint must bind. By a similar argument, the IC constraint will also bind if high effort is implemented.
The cooperation shadow price is scaled based on the likelihood ratio of observed profits given effort, i.e. based on how likely the observed profits were given the efforts the manager could have exerted. I’m not sure how to do this if effort were continuous; maybe a conditional expectation?
If the manager cooperates and doesn’t shirk, the second component can be nearly as large as the owner’s shadow price on cooperation (because the densities are strictly positive on the support, the likelihood ratio can be small but not 0). However, profits are random, and low realizations will affect the manager’s pay - low realizations are just less likely when the manager is cooperating. I see the way the likelihood ratio enters as an “information penalty”. Without further restrictions, the penalty may cause the wage to be negative. I think this is unrealistic, but my guess is that adding a wage nonnegativity constraint will penalize the manager’s wage, since the owner can’t optimally punish the manager for bad profit realizations. I should try that.
My guess is that there is some connection between this information penalty and how much information is contained in the conditional distributions, or how sensitive profits are to efforts. I think that in the limiting case of the conditional distribution of profits being uniform for any effort level, the second term in the optimal wage would disappear. In that case, the owner would only pay the manager enough to participate, but since the owner has no information to check on the manager’s effort levels, the owner won’t pay the manager anything for exerting effort. In that case, I don’t think an equilibrium where the manager exerts high effort can be sustained.
In the other direction, if the conditional distributions are highly informative about effort levels, then the penalties will be applied much more swiftly and harshly but also much less often if the manager is cooperating.
The information penalty also seems related to the Price of Anarchy, like maybe a “likelihood-PoA” or something. I wonder if that’s a thing.
It would be interesting to see a critical discount rate for the manager’s cooperation in a dynamic version of this, given that he may be penalized even if he cooperates.
Conclusion
This model is like my earlier attempt to model an owner-manager problem combined with the labor market signaling problem. Unlike the earlier problem, here the manager’s behavior is embedded in the owner’s problem as a constraint. This problem also abstracts away from the actual generation of profits - the owner in this problem only sees profits as a random variable, not as actual revenues and costs. This lets us get effort into the problem, which lets us drive a wedge between the owner and the manager. Given the manager’s incentive compatibility constraint, it seems like low realizations of profits could cause the manager to shirk, but I’m not sure if that’s true.
The initial assumptions are somewhat strict, and it makes sense to change them based on the specific question at hand. One example I’ve been thinking of is a mutual fund manager principal-agent problem. The owner would be the investor (the manager’s client), the profits would be the mutual fund’s returns, and effort would be trading activity. Many mutual funds charge commissions for “active management”, but generally portfolio returns are higher under “passive management”. This would require reversing our earlier assumptions about effort and the conditional distribution of profits given effort. In the benchmark case, we can imagine the investor is choosing between an actively managed fund versus a passively managed one. When the trading activity is hidden information, then there would probably be a condition very similar to the incentive compatibility constraint here that would scale the manager’s pay based on observed portfolio returns, with a penalty for excess trading.
What if the principal doesn’t know the conditional distribution of payoffs given the agent’s actions? One way to model that could be as a Bayesian game - the principal has some beliefs over the distribution, and learns from repeated play.
I like this model a lot. The problems aren’t hard to solve, but the way they go together makes the model do interesting things. The initial assumptions are important, but it is reasonable to change them and try things. I think these features make a model fun to play with. I wonder how well its predictions have held up if/when they’ve been put to the data.
View or add comments
12 Oct 2015
Consider an individual with preferences over lotteries that have an expected-utility representation. There are three lotteries this individual can choose from:
\[L_1
\begin{cases}
\begin{align}
200 ~~~ &P(200) = 1 \cr
\end{align}
\end{cases}\]
\[L_2
\begin{cases}
\begin{align}
0 ~~~ &P(0) = 2/3 \cr
200 ~~~ &P(200) = 1/6 \cr
1000 ~~~ &P(1000) = 1/6 \cr
\end{align}
\end{cases}\]
\[L_3
\begin{cases}
\begin{align}
0 ~~~ &P(0) = 1/3 \cr
400 ~~~ &P(400) = 1/3 \cr
1000 ~~~ &P(1000) = 1/3 \cr
\end{align}
\end{cases}\]
The expected utilities of the lotteries are:
\[\begin{align}
EU_{L1} =& (200)(1) = 200 \cr
EU_{L2} =& (0)(1/3) + (100)(1/6) + (1000)(1/6) \cr
=& 200 \cr
EU_{L3} =& (0)(1/3) + (100)(1/3) + (1000)(1/3) \cr
=& 1400/3 \cr
\end{align}\]
We can take a first pass at ordering the three lotteries by expected utility alone. This gives us that \(L_3 \succ L_2 \sim L_1\). To go further, we can use the concept of stochastic dominance.
\(X \succ_{FSD} Y\) (X first-order stochastically dominates Y) if \(F_x (t) \le F_y (t) ~ \forall t \in [a,b]\), where \([a,b]\) is the support of \(F_x\) and \(F_y\). This is equivalent to saying \(X \succ_{FSD} Y\) iff \(E[u(x)] \ge E[u(y)] ~ \forall\) nondecreasing, continuous functions \(u\). We apply this concept when we order the lotteries by expected utility.
Second-order stochastic dominance is a refinement on this. We say \(X \succ_{SSD} Y\) (X second-order stochastically dominates Y) if \(\int_a^w F_x (t) dt \le \int_a^w F_y (t) dt ~ \forall w \in [a,b]\), where \([a,b]\) is the support of \(F_x\) and \(F_y\). This is equivalent to saying \(X \succ_{SSD} Y\) iff \(E[u(x)] \ge E[u(y)] ~ \forall\) nondecreasing, continuous, and concave functions \(u\). We apply this concept when we order the lotteries by expected utility.
If all we have is that the agent is an expected-utility maximizer then all we can do is apply first-order stochastic dominance and say that the agent will prefer \(L_3\) over \(L_1\) and \(L_2\), and be indifferent between \(L_1\) and \(L_2\).
If we know or are willing to assume that the agent is risk-averse - that their utility function \(u\) is concave - then we can apply second-order stochastic dominance and rank the lotteries \(L_3 \succ L_1 \succ L_2\). The risk-averse agent would rather take the safe 200-for-sure than the risky 200-on-average.
View or add comments
11 Oct 2015
Today’s post is a variation on monopoly pricing for discrete demand.
Suppose you’re moving, and you have a piece of used furniture that you need to sell. You’ve got two potential buyers, \(B_1\) and \(B_2\), lined up to buy it. They arrive sequentially, so \(B_2\) won’t have anything to look at if \(B_1\) buys. Assume \(B_i\) is willing to pay \(v_i\) for the furniture, where \(v_i\) is an independent draw from a uniform CDF \(F(v_i) = \frac{v_i - a}{b - a}\) for \(v_i \in [a,b]\). The furniture has no value to you if it isn’t sold - you can’t take it with you - and there is no discounting. There’s no bargaining, either - you get to make a take-it-or-leave-it offer to each buyer.
Let’s start at the end of this game with some numbers: \(a=40, b\=140\).
Pricing for the second buyer
Suppose \(B_1\) rejected your offer, so you’re pricing it for \(B_2\). The game ends if this buyer rejects, so this round is basically a standard discrete demand monopoly pricing problem. Your price solves
\[\max_{p_2} ~ (p_2-c)(1-F(p_2))\]
Giving us
\[p_2 = c + \frac{1-F(p_2 )}{f(p_2 )}\]
In this example, the marginal cost would be the opportunity cost - the value of the furniture to you if you didn’t sell it. We know that’s 0 here because you’re moving and can’t take it with you. Plugging in for this uniform CDF and PDF, the optimal price becomes
\[\begin{align}
p_2 &= \frac{b}{2} \cr
\implies p_2 &= 70 \cr
\end{align}\]
Your expected profits from this pricing scheme are
\[\begin{align}
\pi_2 &= (p_2-c)(1-F(p_2)) \cr
&= (70)(\frac{70}{100}) \cr
&= 49 \cr
\end{align}\]
Note that there’s no discounting here - if there was, we’d be hitting this expected profit with a discount rate, and it would be lower than 49.
Pricing for the first buyer
Now that we know the end of the game, we can go backwards to the first round. You’re solving the same problem for the first buyer, but now you have to keep in mind that if the first person doesn’t buy, you’ve still got the second person coming. The optimal price will solve
\[\max_{p_1} ~ p_1(1-F(p_1)) + F(p_1) \pi_2\]
Giving us
\[p_1 = \frac{189}{2} \gt p_2\]
This is nice and what we would expect: when you have multiple buyers lined up, your price should decrease over time as buyers reject, if only because there are fewer buyers coming later. The last buyer should receive the lowest price.
To round it off, your expected profits in the first stage would be 73.7975. Your expected profits decrease as you move through the buyers, since your price gets lower and you lose the expected value of the future buyers.
Adding discounting wouldn’t change the key features of this problem. Adding bargaining would, since the buyers wouldn’t have to accept your profit-maximizing price. Discounting plus bargaining would change things by giving the person with the higher discount rate (more patience) an advantage. This is what happens in Rubinstein bargaining games.
View or add comments
10 Oct 2015
Consider a worker trying to maximize lifetime consumption and leisure. She needs a job to get stuff to consume, which eats into her leisure time, and gives us an interesting tradeoff to study.
The total time available is normalized to 1 and everything is measured in terms of the consumption good. For any wage \(w_t\), the worker needs to decide how much time to work, \(1-l_t\). Her labor income is \(w_t (1-l_t)\). If she’s unemployed, she gets to spend all her time on leisure, but she gets no income and therefore no consumption. Formally, the worker solves
\[\begin{align}
\max_{c_t,l_t} &~ E_0 \sum_{t=0}^{\infty} \beta^t[\ln c_t + \phi \ln l_t] \cr
\text{s.t.} &~ 0 \le l \le 1 \cr
&~ c_t \le w_t (1-l_t)
\end{align}\]
where \(\beta \in (0,1)\) and \(\phi \gt 0\).
In the beginning of each period, the unemployed worker receives an offer \(w \in [0,B],~B \lt \infty\), drawn randomly from the continuous distribution function \(F(w)\). If the worker accepts the offer, she begins receiving labor income in the same period.
The employed worker can pay a search cost \(z \gt 0\) to receive another offer randomly drawn from the same distribution, \(F(w)\) - she can keep searching on the job. If she accepts the offer, she begins working at the new job in the following period.
To keep things simple, let’s assume there is no exogenous job destruction; once an offer is accepted, the job survives with probability 1. There is no saving or borrowing.
The utility function is natural log in consumption and leisure, so we can rule out 0 and \(\infty\) as solutions. The unemployed worker will always accept their first job offer. The action is in the employed worker searching.
The Bellman equations
To solve this problem, we break the consumer’s decision over infinite periods down into a series of smaller one-period decisions and find an optimal solution to that one equation (dynamic programming). In each period, the workers face the following decisions:
\[\begin{align}
\text{Unemployed:} ~~ V^u (w) & = \max_l \{accept , ~ reject \} \cr
& = \max_l \{ u(w(1-l),l) + \beta V^e (w), u(0,1)\} \cr
\end{align}\]
Clearly, the unemployed will always accept any offer \(w \gt 0\).
\[\begin{align}
\text{Employed:} ~~ V^e (w) & = \max_l \{don't~search , ~ search \} \cr
& = \max_l \left\{ \frac{u(w(1-l),l)}{1-\beta}, \frac{u(w(1-l)-z,l)}{1-\beta} + \frac{\beta}{1-\beta} \int_w^B (V^e (w') - V^e (w)) \ dF(w') \right\} \cr
\end{align}\]
\(w'\) is the next period’s offer. The search integral is a Lebesgue integral over the wage distribution. The worker will obviously not accept an offer at a wage lower than the current one, so we only integrate over the current wage to the upper bound.
Is there a reservation wage?
In this context, “is there a wage \(\bar{w}\) such that the employed worker searches if and only if \(w \le \bar{w}\)”? The answer is yes. To find \(\bar{w}\), let’s assume it exists, in which case it satisfies
\[V^e (Don't~search) = V^e (Search)\]
Then we do some algebra and get that
\[\bar{w}:~ \ln(w(1-l)) - \ln(w(1-l)-z) = \beta \int_{\bar{w}}^B (V^e (w') - V^e (\bar{w})) \ dF(w')\]
The LHS is the search penalty to the marginal utility of consumption (marginal cost of searching), and the RHS is the discounted expected increase in wage from the search (marginal benefit of searching).
What else can we say about the worker’s search behavior?
-
As \(w \to 0\), \(V^e_s (w) \to -\infty\) faster than \(V^e_n (w) \to -\infty\). So at a low enough wage, the search cost \(z\) has a large effect on the worker’s utility and she won’t search.
-
At \(w=B\),
\(\\begin{align}
V^e_s (B) & = (1-\beta)^{-1} (\ln(B(1-l)-z) + \phi \ \ln(l)) \cr
V^e_n (B) & = (1-\beta)^{-1} (\ln(B(1-l)) + \phi \ \ln(l)) \cr
\implies V^e_n (B) & \gt V^e_s (B)
\end{align}\)
So the worker won’t search when she’s already earning the maximum wage, which is pretty intuitive.
So the worker won’t search at a low enough wage, and she won’t search at the highest wage… since the value of searching and not searching are both concave functions, this means they must touch twice if they cross. They could potentially touch only once if they don’t cross. This makes things complicated, since now there are potentially three wages where the condition we used to find the reservation wage holds: in the case where they cross, there’s the wage at which workers start searching and the wage at which the workers stop searching (what we found); in the case where they touch only once, there’s the wage at which workers are indifferent between searching and not searching.
To make life easier, let’s assume \(z\) is small enough that \(V^e_s (w) \lt V^e_n (w)\) occurs on the lower end of the wage scale only at wages below 1 (so negative utilities), and the only place where our reservation wage condition holds that matters is the one we explored. Utility has a cardinal interpretation here, so we can rule out the case of the worker earning so little she can’t afford to search.
Anyway, with this behavior, in the long run we should expect every worker to end up with a wage above their reservation wage, since they’ll keep searching for a better job otherwise.
In a future post, I’ll relax the assumption that \(z\) is very small and sketch out all the cases of the reservation wage, and explore the worker’s labor supply behavior in this model.
Dynamic programming is a powerful way to get a simple solution to an optimization problem. Without it, we would have had to set up and solve an infinite-period optimization. With it, we can just find a consistent decision rule for an arbitrary period. To use DP, we need some assumptions on the functions we’re optimizing. When I revisit this model, maybe I’ll talk about those assumptions as well.
View or add comments
08 Oct 2015
Suppose you are considering purchasing a used car. There are \(N \ge 1\) sellers. The value of seller \(i\)’s car is \(v_i\), given by
\[v_i =
\begin{cases}
1 ~~ \text{with probability} ~~ \beta_i \cr
0 ~~ \text{with probability} ~~ 1- \beta_i \cr
\end{cases}\]
where \(\beta_i = \beta^i, i=1,...,N\) and the \(\beta_i\) are all independent of each other. So as the number of sellers increases, we start with the best seller and only add sellers worse than the current ones. Since we search in a random order, for a given \(N\) this just tells us the distribution of sellers. A seller’s \(\beta_i\) is their private information. This is sort of like looking for a car on Craigslist or some other search; holding \(N\) constant, you search randomly over sellers of uncertain quality.
You’re a savvy shopper though, and before you buy the car you’re going to get it inspected, which costs \(c\). You could also think about this as a search cost, to avoid confusion over inspection-less visits to sellers. Suppose the price of a car is \(1/2\) from every seller, for simplicity. You can inspect as many sellers as you want, but only one at a time.
Based on \(\beta\), \(N\), and \(c\), when will it be optimal for you to search at least once?
Benchmark case: Expected-utility maximizer
An expected utility maximizer would start by figuring out the expected quality of a seller,
\[\begin{align}
\hat{\beta} &= \frac{1}{N} \sum_{i=1}^{N} \beta^i \cr
&= \frac{\beta}{N} \left( \frac{1- \beta^N}{1- \beta } \right)
\end{align}\]
We assume we’re starting from no prior searches so that we don’t have to condition on information and update beliefs, which any optimizing agent ought to be doing. The EU-maximizer wants the expected value of a search net of inspection costs to be positive, or
\[\begin{align}
\hat{\beta}(v-p) - c &\ge 0 \cr
\implies \hat{\beta}(1/2) - c &\ge 0 \cr
\implies c &\le \frac{1}{2}\hat{\beta} \cr
\end{align}\]
We use \(\hat{\beta}(v-p)\) instead of \(\hat{\beta}-p\) because our value from the car is \(1\) for sure - the uncertainty is in whether or not we see it, not in the value itself.
This gives us a bound on whether or not it is optimal to search - an expected-utility maximizer will search if the expected value of the car is greater than the inspection cost.
This makes sense, but there’s another way we can approach this question.
Different behavior: A worst-case utility maximizer
I guess another way to put this is a “minimum-utility maximizer”. I’ve never heard that term before. I think worst-case bounds are more popular in computer science than economics, which is probably why I haven’t heard it called that.
The worst case we could encounter on our first search is to visit seller \(N\) first. A minimum-utility maximizer only wants to search if even in this bad draw is worth their time, or
\[\begin{align}
\beta^N(v-p) - c &\ge 0 \cr
\implies c &\le \frac{1}{2}\beta^N \cr
\end{align}\]
which looks pretty similar to the expected-utility maximizer’s solution, but with \(\beta^N\) instead of \(\hat{\beta}\). This implies a more “conservative” decision rule than the benchmark solution.
Summary
This is a pretty cool problem. I initially disagreed with my professor about the solution - I used the worst-case bound, while he used the expected-value bound. He pointed out that in this case, the worst-case bound is a subset of the expected-value bound (I think that must be true generally), so the worst-case bound ignores a large area where an expected-utility maximizer would find it optimal to search. I agree with his argument, but I still prefer the worst-case bound to the expected-value bound. In writing this post, I’ve been trying to understand why.
I think the worst-case bound is a more general solution to the problem (or a solution to a more general problem) than the expectation-bound. In this problem, we’re assuming we know a lot about the distribution of values - we know pretty much exactly what it looks like always, and so we know we can use the expectation to figure out what to do.
If we relax that assumption to only knowing the supports of the distribution of values, then we can’t take an expectation to maximize expected utility. We can still find a worst-case bound and maximize a minimum utility in cases like this.
In general, the worst-case bound requires less information than the expected-value bound; we can set a worst-case bound even if we have no information about the distribution of sellers (or if the distribution of values is continuous and has no expectation, like a Cauchy) by looking at our own budget constraint. That said, if we have the information there’s no reason to ignore it.
While the minimum-utility maximizer may ignore opportunities that the expected-utility maximizer would take, the minimum-utility maximizer can consistently apply their decision rule in more situations than the expected-utility maximizer can.
Below is a plot of the cost bounds with \(\beta=0.99\) as the number of sellers increases. The worst-case bound is in red, the expected-utility bound is in blue.
The bounds both approach 0 as the number of sellers increases. With \(\beta=0.99\) the two are practically indistinguishable after 20000 or so sellers - which seems like a pretty big market to me, maybe not relative to financial markets but still. With \(\beta=0.5\) (not shown) the two are practically indistinguishable after 600 or so sellers. The lower \(\beta\) is, the faster they approach each other - intuitive from the functions, but I hadn’t expected the worst-case bound to drop quite so fast.
I found a neat quote on Turing’s Invisible Hand about worst-case vs. expected-case more generally:
In a situation that already exists, there is obviously some empirical distribution, and it is hard to argue against taking it into account. In a situation that doesn’t exist, there is yet no empirical distribution, so in order to prepare for it we need to either theoretically guess the future distribution, or prepare for worst case.
Most situations we consider in economics already exist, or are similar to ones that do, so there’s usually a distribution we can work with or a reasonable guess we can make. The connection between EU-maxing/minU-maxing and risk preference is interesting to me; I’d like to go back to the manager’s problem and add uncertainty and different risk preferences in this form between the manager and owner.
I like this problem. At some point, I want to try solving it with a different distribution and running it in R.
View or add comments