When is it optimal to search for a used car?

Suppose you are considering purchasing a used car. There are \(N \ge 1\) sellers. The value of seller \(i\)’s car is \(v_i\), given by

\[v_i = \begin{cases} 1 ~~ \text{with probability} ~~ \beta_i \cr 0 ~~ \text{with probability} ~~ 1- \beta_i \cr \end{cases}\]

where \(\beta_i = \beta^i, i=1,...,N\) and the \(\beta_i\) are all independent of each other. So as the number of sellers increases, we start with the best seller and only add sellers worse than the current ones. Since we search in a random order, for a given \(N\) this just tells us the distribution of sellers. A seller’s \(\beta_i\) is their private information. This is sort of like looking for a car on Craigslist or some other search; holding \(N\) constant, you search randomly over sellers of uncertain quality.

You’re a savvy shopper though, and before you buy the car you’re going to get it inspected, which costs \(c\). You could also think about this as a search cost, to avoid confusion over inspection-less visits to sellers. Suppose the price of a car is \(1/2\) from every seller, for simplicity. You can inspect as many sellers as you want, but only one at a time.

Based on \(\beta\), \(N\), and \(c\), when will it be optimal for you to search at least once?

Benchmark case: Expected-utility maximizer

An expected utility maximizer would start by figuring out the expected quality of a seller,

\[\begin{align} \hat{\beta} &= \frac{1}{N} \sum_{i=1}^{N} \beta^i \cr &= \frac{\beta}{N} \left( \frac{1- \beta^N}{1- \beta } \right) \end{align}\]

We assume we’re starting from no prior searches so that we don’t have to condition on information and update beliefs, which any optimizing agent ought to be doing. The EU-maximizer wants the expected value of a search net of inspection costs to be positive, or

\[\begin{align} \hat{\beta}(v-p) - c &\ge 0 \cr \implies \hat{\beta}(1/2) - c &\ge 0 \cr \implies c &\le \frac{1}{2}\hat{\beta} \cr \end{align}\]

We use \(\hat{\beta}(v-p)\) instead of \(\hat{\beta}-p\) because our value from the car is \(1\) for sure - the uncertainty is in whether or not we see it, not in the value itself.

This gives us a bound on whether or not it is optimal to search - an expected-utility maximizer will search if the expected value of the car is greater than the inspection cost.

This makes sense, but there’s another way we can approach this question.

Different behavior: A worst-case utility maximizer

I guess another way to put this is a “minimum-utility maximizer”. I’ve never heard that term before. I think worst-case bounds are more popular in computer science than economics, which is probably why I haven’t heard it called that.

The worst case we could encounter on our first search is to visit seller \(N\) first. A minimum-utility maximizer only wants to search if even in this bad draw is worth their time, or

\[\begin{align} \beta^N(v-p) - c &\ge 0 \cr \implies c &\le \frac{1}{2}\beta^N \cr \end{align}\]

which looks pretty similar to the expected-utility maximizer’s solution, but with \(\beta^N\) instead of \(\hat{\beta}\). This implies a more “conservative” decision rule than the benchmark solution.

Summary

This is a pretty cool problem. I initially disagreed with my professor about the solution - I used the worst-case bound, while he used the expected-value bound. He pointed out that in this case, the worst-case bound is a subset of the expected-value bound (I think that must be true generally), so the worst-case bound ignores a large area where an expected-utility maximizer would find it optimal to search. I agree with his argument, but I still prefer the worst-case bound to the expected-value bound. In writing this post, I’ve been trying to understand why.

I think the worst-case bound is a more general solution to the problem (or a solution to a more general problem) than the expectation-bound. In this problem, we’re assuming we know a lot about the distribution of values - we know pretty much exactly what it looks like always, and so we know we can use the expectation to figure out what to do.

If we relax that assumption to only knowing the supports of the distribution of values, then we can’t take an expectation to maximize expected utility. We can still find a worst-case bound and maximize a minimum utility in cases like this.

In general, the worst-case bound requires less information than the expected-value bound; we can set a worst-case bound even if we have no information about the distribution of sellers (or if the distribution of values is continuous and has no expectation, like a Cauchy) by looking at our own budget constraint. That said, if we have the information there’s no reason to ignore it.

While the minimum-utility maximizer may ignore opportunities that the expected-utility maximizer would take, the minimum-utility maximizer can consistently apply their decision rule in more situations than the expected-utility maximizer can.

Below is a plot of the cost bounds with \(\beta=0.99\) as the number of sellers increases. The worst-case bound is in red, the expected-utility bound is in blue.

Cost bounds

The bounds both approach 0 as the number of sellers increases. With \(\beta=0.99\) the two are practically indistinguishable after 20000 or so sellers - which seems like a pretty big market to me, maybe not relative to financial markets but still. With \(\beta=0.5\) (not shown) the two are practically indistinguishable after 600 or so sellers. The lower \(\beta\) is, the faster they approach each other - intuitive from the functions, but I hadn’t expected the worst-case bound to drop quite so fast.

I found a neat quote on Turing’s Invisible Hand about worst-case vs. expected-case more generally:

In a situation that already exists, there is obviously some empirical distribution, and it is hard to argue against taking it into account. In a situation that doesn’t exist, there is yet no empirical distribution, so in order to prepare for it we need to either theoretically guess the future distribution, or prepare for worst case.

Most situations we consider in economics already exist, or are similar to ones that do, so there’s usually a distribution we can work with or a reasonable guess we can make. The connection between EU-maxing/minU-maxing and risk preference is interesting to me; I’d like to go back to the manager’s problem and add uncertainty and different risk preferences in this form between the manager and owner.

I like this problem. At some point, I want to try solving it with a different distribution and running it in R.