Paper Review - "Earth Orbit Debris - An Economic Model"

Here I review Adilov, Alexander, and Cunningham’s (hereafter AAC) 2013 paper, “Earth Orbit Debris: An Economic Model”. The paper was published in Environmental and Resource Economics in 2015 as “An Economic Analysis of Earth Orbit Pollution”. The published version seems very similar to the SSRN version, which is the version I read and what I describe here. At some point I intend to write a literature review for my own space debris project; hopefully this writeup will help me with that.

This is the first explicitly mathematical economic analysis of orbital debris production, at least as far as I am aware. Two other, earlier, papers that consider the economics of orbital debris are Weeden and Chow’s “Taking a common-pool resources approach to space sustainability: A framework and potential policies” and Bradley and Wein’s Space debris: Assessing risk and responsibility. Both were published in space journals (Weeden and Chow in Space Policy, Bradley and Wein in Advances in Space Research). While both consider economic aspects of the space debris problem, neither takes a full-on “mathematical economics” approach to modeling the problem. Weeden and Chow focus more on the economics, drawing heavily on Ostrom’s framework for managing CPR resources, while Bradley and Wein focus more on the physics, using their physical model to estimate the net present value of a debris-related collision. I’ll write about both of these papers separately at a later date.

This paper has two main conclusions:

  1. The decentralized equilibrium has firms launching too many satellites relative to what the central planner would launch, and under-investing in debris-mitigation technologies relative to what the central planner would invest. This is because firms don’t internalize the costs of debris to other firms, resulting in a higher launch rate and a lower level of debris mitigation.
  2. Firms can be made to internalize the costs of their actions with a Pigouvian tax on launches, which will bring the decentralized launch rate to the “optimal centralized launch rate”.

Because this paper is fully theoretical with no calibration, the conclusions are qualitative in nature.

The Model

AAC’s model is a Salop circle (I’ve written about this as Circle City) with two periods. In the first period the orbital environment has no debris and firms simultaneously decide whether or not to launch a satellite. In the second period there is debris generated from the first period launches. Some satellites are destroyed by debris, but firms do not launch any new satellites in period 2; the ones with surviving satellites just choose a price and collect revenues.

The Salop framework lets AAC to capture product differentiation in the satellite market - e.g. GEO satellites for communications, LEO constellations for imaging, etc. Within this framework, they assume that firms are identical, and must pay a marginal cost per-satellite, a fixed launch cost, and a fixed per-period “maintenance cost”. As an aside, the Circle City model is apparently not analytically solvable with heterogeneous firms.

The Circle City setup gives them a symmetric Bertrand-Nash equilibrium where firms price “satellite services” with a markup over marginal cost, even when they use free entry to get the zero-profit condition which gives them the number of firms per-period. Because AAC assume 1 firm per satellite, the number of firms entering the market gives them the launch rate.

The satellites are also homogeneous and generate the same amount of debris. AAC assume that each satellite has an equal probability of a catastrophic collision with debris, and that the destroyed satellites disappear.

In section 2.3, AAC consider the social planner’s problem. To do this, they assume a consumer surplus function, and have the social planner maximize the discounted sum of expected profits and consumer surplus.

In section 2.5, AAC allow the firms to choose the debris creation rate \(\phi \in [0, \phi_H]\). The debris creation rate directly impacts the firm’s launch cost \(r = h(\phi)\), where \(h'(\phi) < 0\) for \(\phi<\phi_H, h'(\phi_H)=0, h''(\phi)>0\) and \(h(\phi_H)>0\). In words, the launch cost is decreasing in the amount of debris produced at an increasing rate. The assumption of cost-minimizing behavior implies that firms will choose the highest rate of debris creation, \(\phi_H\), since they don’t see an individual benefit from reducing debris creation rates.

Discussion

The planner’s solution gives AAC an optimum to compare the decentralized outcomes to, which lets them derive the Pigouvian tax on entry (launches) which would result in the socially optimal level of launches and debris creation. The decentralized outcome is suboptimal in the sense that it diverges from the planner’s solution. Bradley and Wein take a different approach to estimating a fee on debris generation, by calculating the discounted expected value of a catastrophic collision and suggesting a fee that would be equal to that damage. This would not serve the same deterrent function as AAC’s tax, but it isn’t intended to - it’s intended to give an estimate of the damage that debris could cause.

The firms in AAC’s model are individually maximizing objective functions which have no allowance for consumer surplus generated by the lost satellites, so it doesn’t surprise me that the firms will behave differently from how the planner would have them behave. More theoretically, I think the setup of the problem violates the First Welfare Theorem: since debris is not priced in any way even though it affects consumer welfare, it is natural that firms should produce “too much” debris.

The assumption of product differentiation through a Salop circle with homogeneous firms seemed clever to me at first, but now it seems weird. It seems to imply that satellite services are only differentiated in terms of consumer taste through the “travel cost” parameter \(t\). This is not true - imaging constellations in LEO provide a fundamentally different service than communications satellites in GEO. Ceteris paribus more telecom satellites in GEO shouldn’t affect the price of images from LEO constellations, and vice versa.

The assumption of a common probability of catastrophic collision with debris seems like a big simplification of the underlying orbital mechanics. In truth, orbital debris seems more like a directional externality to me - it only falls down to Earth, so a satellite already in GEO isn’t affected by debris in LEO, though debris falling to Earth may pose a risk to constellations in LEO. Launching a satellite to GEO could be made more complicated by debris in LEO as well, though these are things I expect the engineers and physicists would work reduce as much as possible; as far as I’m aware this is what conjunction analysis is about.

Additionally, a satellite’s probability of collision depends on its size, which is a parameter firms have some control over. The recent move towards cubesats suggests to me that firms are able to reduce the probability that their satellite will get hit.

Bradley and Wein discuss the directional aspect of debris generation in their paper, where they consider debris generation in a single “spherical altitude shell” (the “Shell of Interest”, or SOI). This mirrors the modeling approach taken in Rossi et al.. Bradley and Wein argue that ignoring debris entering a SOI from a higher shell doesn’t significantly affect the model results, though Rossi et al. include it in their paper and I think Liou et al. include this in their three-dimensional LEO-to-GEO debris evolutionary model as well. These papers also include end-of-life satellites as sources of debris, and allow destroyed satellites to generate debris. Both of these features are absent from AAC’s model, but would probably not change the qualitative conclusions drawn.

In any case, the result of these two assumptions - homogenous satellite services differentiated through travel cost on the circle, and common probabilities of collisions with debris - seems to be a strong simplification of the satellite industry and orbital environment. The complexities introduced by weakening these assumptions might be distractions that don’t change anything, or they might not. I think the orbital debris problem is interesting enough to warrant a paper or two that relax these assumptions. In particular, I would like to see how firms respond to specific orbital shells in LEO filling with debris. My guess is that the firms would either try to go to higher shells (above the debris + more coverage area) or lower shells (maybe they can’t penetrate the debris in a higher shell/debris decays faster at lower altitudes and they decide to play nicer). Higher shells would result in increasingly bad pollution, while lower shells would result in some self-cleaning as the debris decayed.

I think AAC’s general conclusion that there will be too many launches and too much debris relative to the social optimum may hold as presented for LEO, where many firms are looking to enter and firms are clustered in a relatively narrow altitude band (I think most LEO satellites are in between 800-1200 km, or thereabouts). I am skeptical that it will hold in the same way for GEO, where the congestibility of the resource is more apparent and the market mechanisms in place for spectrum allocation have led to a slot allocation mechanism. Macauley 1994 looks at the efficiency of this mechanism; this paper is on my reading list.

In GEO, I expect the result to take the form of too little effort expended on end-of-life graveyard orbit policies. Indeed, though the Inter-Agency Space Debris Coordination Committee (IADC) laid out guidelines on acceptable re-orbiting altitude, it appears that only a third of the GEO satellites that reached end-of-life complied. My guess is that this is because of three reasons:

  1. Some firms don’t expect to launch another satellite to the same or a nearby slot, and so don’t care if they leave a mess (myopic behavior)
  2. Some firms would find it too costly to reduce the fuel allocated for stationkeeping (i.e., the lifetime of the satellite and the expected revenues from it) by allocating more fuel to end-of-life disposal (compliance too costly)
  3. Some firms’ satellites may have been non-responsive to the disposal command due to damage, or the firms did not anticipate the disposal requirement and were unable to send such a signal (error/nothing they could do)

I expect that policy could be effective at targeting 1 and 2, but not 3 (error). The possibility of 3 makes me think that some sort of active debris removal will be necessary even if firms and nations all act to minimize future debris and all existing debris magically disappeared.

Weeden and Chow anticipated the Pigouvian approach AAC espouse, and discuss the significant legal difficulties associated with implementing a Pigouvian tax like AAC describe. Basically, it’s really unlikely that such a tax would be implemented globally, and if it was implemented only in a few countries firms would have an incentive to launch from countries without the tax. Weeden and Chow also suggest the possibility of cooperative orbital management arising, which I think would be really interesting to see formalized in a cooperative game theoretic framework. This possibility is absent from AAC’s discussion of the problem and potential solutions.

The 2-period setting is probably fine for general conclusions, but I wonder how (or if) anything would change if the firms chose to launch each period for \(T\) periods, or whenever their satellites broke down. My guess is that it would depend on whether or not the firms have rational expectations. In general, I think that repeated games allow for better outcomes than one-shot games. Reputational incentives, coordination, things of that nature. I think a repeated game framework is a more realistic economic setting for this problem than a one-shot game.

I think the assumptions on \(h(\phi)\) may be a good approximation, but are probably not uniformly true. It might be cheaper for a firm to avoid some sources of debris, like paint on a satellite. In that case, asking firms to reduce those activities of debris might get good compliance. Other sources of debris, like launching a satellite into an orbit too high for debris to naturally decay quickly, might be harder to get compliance with since higher orbits have more coverage area over the Earth. I think firms might underinvest in debris mitigation technologies in general, but I’m skeptical that they would choose the highest level of debris creation or lowest level of mitigation. Proposition 3 derives this for profit-maximizing firms, but under a specific assumption about the cost of mitigation in a 2-period setting where firms only launch once. Rational expectations with launches over multiple periods in a longer horizon might change things.

In closing, I think that AAC’s paper is a good contribution to the field of space debris. It brings environmental and mathematical economics firmly into the discussion of orbital debris management, though I didn’t find the conclusions or intuitions it presented to be very novel. I also thought some of the assumptions were too strong, and glossed over interesting particulars of the problem. It seems to have published pretty decently and attracted some good attention, though, so there might be a market for more papers on the subject. I hope so.

View or add comments

An app for Coase bargaining

I came across an article today about “how you know you’re on a plane with an economist”. The article’s supposed to be a take on another recent news piece, about an economist getting pulled off a plane for doing math. I have a long flight coming up, and the article got me thinking about the problem of people reclining.

The person sitting in front of me on a long (8+ hours) flight has the right to recline their seat back. Now, I’m not a very tall person and I’m used to sitting for long stretches, so I am not too physically uncomfortable if the person in front of me reclines all the way. But I do like to read, write, and occasionally eat, so the lack of space in front of me is not ideal.

(Some airlines get around this by having seats in a fixed bucket, so that instead of actually reclining you just kinda slide forward in the bucket. I really don’t like these; I think they’re the worst of both worlds. You no longer have much space, and you can’t even meaningfully recline. I’m ignoring this case here.)

The situation with the reclining seat seems like something the Coase theorem ought to apply to: we have a well-defined property right (person in front has right to recline) and an externality (reclining reduces my personal space), so maybe we can bargain our way to a better solution. This is such a straightforward idea, I’m sure it’s been discussed elsewhere at length.

What I would like is an app to facilitate this bargaining. I think the app would do something like this:

  1. Given two participating users, allow the users to define the property right and the externality in question. For example, “\(A\) has the right (to recline his chair), \(B\) is negatively affected by ( \(A\) reclining ) and would like to see if a welfare-improving bargain can be struck.”

  2. Users privately enter their maximum willingness-to-pay and minimum willingness-to-accept to reduce the externality. In the seat example, \(A\) would enter his minimum willingness-to-accept to not recline, and \(B\) would enter her maximum willingness-to-pay for \(A\) to not recline.

  3. If the maximum WTP is greater than or equal to the minimum WTA, the app would suggest a price for \(B\) to pay \(A\) to not recline. The specific price can be determined by some “fair” division, like the Nash Bargaining Solution.

In the example above, the Nash Bargaining Solution would be a payment \(p_B\) from \(B\) to \(A\) that solves

\[\max_{p_B} \ (p_B - WTA_A)(WTP_B - p_B)\]

where \(WTA_A\) and \(WTP_B\) are \(A\) and \(B\)’s minimum willingness-to-accept and maximum willingness-to-pay, respectively. The solution is

\[p_B = \frac{WTP_B + WTA_A}{2}\]

i.e., with the NBS, if \(WTP_B \geq WTA_A\), \(B\) would pay \(A\) his minimum willingness-to-accept to not recline, plus half of her excess willingness-to-pay for \(A\) to not recline. This results in the surplus, \(WTP_B - WTA_A\), being evenly divided between \(A\) and \(B\).

The usual problem people cite with the Coase theorem in practice is transaction costs, which an app could reduce. I’ve described the two-person case above but it could be extended to the \(n\)-person case. I would be willing to pay $5 for an app that did this, possibly more if it was widely used.

Update: srlm made the app! Here it is on Google Play and on Github.

View or add comments

Is this worth a search?

Another simple Shiny app, this time to answer the question, “is this worth searching for?”.

The basic idea is taken from this earlier post about searching for a used car.

For example: Suppose you are thinking about looking for a used car, and you want to see whether it’s worth searching for. You plug in your willingness-to-pay for a car, the cost to you to find and inspect one car, the number of sellers in the region, and the average price they’re asking for. Using the formulas from the post for the expected quality of the average seller and the expected value of one search, we have

\[\begin{align} \hat{\beta} &= \frac{\beta}{N} \left( \frac{1- \beta^N}{1- \beta } \right) \cr EV(\text{search}) &= \hat{\beta}(v-p) - c \end{align}\]

where \(N\) is the number of sellers, \(\beta\) is the probability the average seller has what you want, \(p\) is the average price, and \(c\) is the search cost. If the expected value of a search is greater than 0, it’s worth at least one search.

Using the app, I can see what the probability of a seller having what I want would need to be for a search to be worth it. I can interpret the point on the x axis where \(EV(\text{search})=0\) as the minimum level of belief I would have to hold that the average seller has what I’m looking for for a search to be rational.

View or add comments

A conditional probability app

I wrote my first-ever Shiny app! My first-ever app, too. It’s a simple conditional probability calculator.

Say you have two events, \(A\) and \(B\), which occur over two populations \(N_A\) and \(N_B\). Input the number of occurrences of \(A\) and \(B\) and the size of \(N_A\) and \(N_B\), and the app will apply Bayes’ rule to show you $$P(A B)\(for beliefs you may hold over\)P(B A)$$.

For example: Suppose you want to know the probability that someone who died in the USA in 2010 was 85+ years old. Event A is taken to be the number of deaths in the USA in 2010, event B is taken to be the population that was 85+ years old in the USA in 2010, and the population size is taken to be the average total population of the USA in 2010 (\(N_A = N_B\)). Three Wolfram Alpha queries later we have

\[\begin{align} P(A) &= \frac{2,550,000}{310,000,000} = 0.008 \cr P(B) &= \frac{5,176,00}{310,000,000} = 0.017 \end{align}\]

By Bayes’ rule, we have

\[P(A|B) = P(B|A) \frac{P(A)}{P(B)}\]

I think \(P(A)/P(B)\) is called the base rate.

Using the app, I can see how the probability someone died given they were 85 or older would change depending on the probability of being 85 or older given they died (in the USA in 2010), or $$P(\text{died in USA in 2010 85 or older in USA in 2010})\(as a function of my beliefs over\)P(\text{85 or older in USA in 2010 died in USA in 2010})$$. In frequency terms, I can calculate the proportion of deaths that were 85+ year olds using the number of 85+ year olds, the number of deaths, the total US population in 2010, and my beliefs over the proportion of 85 year olds who died. I used this example as a “comparison event” for the app.
In the picture below, the red line is $$P(\text{died in USA in 2010 85 or older in USA in 2010})\(, and the black line is\)P(A B)\(for a hypothetical event where\)P(A) = 15/100\(and\)P(B) = 50/100$$.

Example

I wrote this app to help me think through conditional probabilities more clearly. Often I hold beliefs over some $$P(B A)\(and use those beliefs to infer something about the corresponding\)P(A B)$$, not necessarily considering the base rate. Hopefully I’ll neglect base rates a little less going forward.
View or add comments

Proof of a law of large numbers

Usually, when we have a sequence of random variables, we can say things about what the average of the sequence does as the sample grows by assuming things about the moments and the sampling process and then applying an appropriate law of large numbers. But how would we prove a law of large numbers?

Laws of large numbers are statements about averages of random variables converging in probability to a constant. In this example, we assume the existence of a finite first moment, assume random sampling with replacement (IID), and show that a sequence of averages of the IID random variables will converge to the first moment.

Let \(\{X_t\}, t=1,2,...\) be a sequence of IID random variables on a probability space \(( \Omega, \mathcal{F}, P )\) with \(E(X_t) = \mu \lt \infty\). The goal is to show that

\[\lim_{n \to \infty} P \left( \omega : \left| \frac{1}{n} \sum_{t=1}^n X_t(\omega) - \mu \right| \gt \epsilon \right) = 0\]

Proof

Begin by defining \(A_n\) and its characteristic function:

\[\begin{align} A_n & \equiv \frac{1}{n} \sum_{t=1}^n X_t \cr \phi_{A_n}(s) & \equiv E(e^{isA_n }) \cr & = \int e^{isA_n } ~ dP \cr \end{align}\]

By the properties of exponential functions and the fact that \(X_t\) is IID, we can write \(\phi_{A_n}(s)\) as

\[\phi_{A_n}(s) = ( \phi_{X_t} \left( \frac{s}{n} \right) )^n\]

We can approximate \(\phi_{X_t}(\frac{s}{n})\) as

\[\begin{align} \phi_{X_t} \left(\frac{s}{n} \right) & = 1 + is \frac{E(X_t)}{n} + \mathcal{O} \left(\frac{s^2}{n^2} \right) \cr & = 1 + is \frac{ \mu }{n} + \mathcal{O} \left(\frac{s^2}{n^2} \right) \cr \mathcal{O} \left(\frac{s^2}{n^2} \right) & \equiv i^2 \int_0^{\frac{x}{n}} \left(\frac{x}{n}-s \right) e^{is} ds \cr \end{align}\]

\(\mathcal{O}(\frac{s^2}{n^2})\) is the remainder from the approximation, the \(\mathcal{O}\) is big O notation that the remainder is bounded.

So now we have

\[\phi_{A_n}(s) \approx \left( 1 + is \frac{\mu}{n} \right)^n\]

The trick is to use a binomial expansion here. In general,

\[\begin{align} \left( 1+ \frac{a}{n} \right)^n & = \sum_{j=0}^n \binom{n}{j} \left( \frac{a}{n} \right)^j \cr \lim_{n \to \infty} \left( 1+ \frac{a}{n} \right)^n & = \sum_{j=0}^{\infty} \left( \frac{a^j}{j!} \right) \cr & = e^a \cr \end{align}\]

So, letting \(a = is \mu\),

\[\begin{align} \phi_{A_n}(s) & = \left( 1 + is \frac{ \mu }{n} + \mathcal{O}(\frac{s^2}{n^2}) \right)^n \cr \lim_{n \to \infty} \phi_{A_n}(s) & = e^{is \mu } = \phi_{ \mu } (s) \cr \end{align}\]

The remainder term \(\mathcal{O} (\frac{s^2}{n^2} ) \to 0\) as \(n \to \infty\).

By the uniqueness of mapping between characteristic and distribution functions of random variables, \(\lim_{n \to \infty} \phi_{A_n}(s) = \phi_{ \mu }(s)\) implies \(\lim_{n \to \infty} F_{A_n} = F_{ \mu }\), where \(F_{A_n}\) and \(F_{ \mu }\) are the distribution functions of \(A_n\) and \(\mu\) respectively.

This is still only convergence in distribution, not convergence in probability. However, \(\mu\) is a constant, and convergence in distribution to the distribution of a constant is the same as convergence in probability to that constant (think of the constant’s distribution as a point mass). So we have the desired result, \(A_n \overset{p}{\to} \mu\).

Discussion

I’m not entirely sure how to interpret the parameter \(s\) for a general characteristic function. The characteristic function is just an exponential Fourier transform, and I would usually interpret that parameter as the “time index” for a given sequence I was Fourier transforming. I’m not sure if that interpretation is justified here, since each random variable is an IID draw from the same distribution. If I wanted to read a “time index” into this, it would be \(t\), which is the draw number. Maybe \(s\) is not very meaningful in this setting, so we can let it be whatever we want?

Anyway, the point of doing it this way is that sometimes a product of exponentials is easier to work with than some arbitrary sum. I should learn more about Fourier transforms and characteristic functions, they seem useful.

View or add comments