Rawlsian Voting Revisited

Like all men in Babylon, I have been proconsul; like all, a slave.

Borges, “The Lottery in Babylon”

A few months ago, I wrote a short post on trying to justify random ballot elections using Rawls’s original position argument. In brief, I interpreted the original position argument as saying that a just society is one that maximizes your expected utility under the assumption that you inhabit a role selected uniformly at random. (Intuitively, a society is fair if you’d be happy entering as a randomly selected person.) I then applied this principle in the context of an election between two candidates AA and BB.1 If nAn_A voters support candidate AA, and nBn_B support candidate BB, what’s a “fair” value for the probability pAp_A of AA winning?

Without loss of generality, let nAnBn_A \ge n_B. The result I presented was basically this: with a straightforward indicator utility function—in which you receive unit payoff if your favored candidate wins, and zero otherwise—it’s not hard to show that the optimal solution is pA=1p_A = 1, which is to say that we should just take a simple majority vote.2 Some more exotic utility functions can give you more interesting results, such as pA=nA/(nA+nB)p_A = n_A / (n_A + n_B), which corresponds to a “random ballot” election, i.e. one in which we should draw a single ballot at random and let it dictate the result.

That’s the conclusion that I came to in the previous post, but I was unsatisfied with the argument, for two reasons. First, even if you grant that expected utility maximization has something interesting to say about this hypothetical election problem, the utility function that I proposed to justify random ballots seemed obviously contrived.3 Second, I did not actually engage with (because I did not know about) an important part of Rawls’s actual argument, which is that expected utility maximization is actually not the most reasonable framework to use when operating behind the veil of ignorance.4 Rawls’s famous book A Theory of Justice (1971), in which he first proposed the original position thought experiment, is actually a landmark rejection of utilitarianism, on the grounds of fairness.

I have spent some time recently mulling over both of these deficiencies in the original post. I think I have a satisfactory response to the first, although I’m still not too sure about the second. I think I will probably have more thoughts on the second later, but I thought that the argument for the first was independently interesting, so I’ve written it up below.


To present a less-contrived argument for random ballots, let’s modify the scenario from the previous post slightly. Instead of modeling each voter’s payoff as a random variable over the space of election outcomes, we instead model each voter’s payoff as a (presumably monotonically increasing, and for convenience differentiable) function f:[0,1]Rf : [0,1] \to \mathbb{R} of the probability of her favored candidate’s victory. That is, we are asserting that a voter is happier if her candidate is more likely to win, which seems like a reasonable assumption.5 Of course, a voter is happiest of all if her candidate wins with certainty, but our model allows for probabilistic outcomes.

In general, for some voter utility function ff, the expected utility6 when entering society from behind the veil of ignorance is

E[Uf]=iCniNf(pi), E[U_f] = \sum_{i \in \mathcal{C}} \frac{n_i}{N} f(p_i),

where C\mathcal{C} is the set of candidates, NN is the total population, and the pip_i’s form a probability distribution. What are some reasonable payoff functions ff that we might consider? Probably the most obvious is the identity function f(x)=xf(x) = x, which actually corresponds to the indicator utility in the setting of the previous post. In the simple case of C={A,B}\mathcal{C} = \{A,B\}, we have already shown in the previous post that E[Uid]E[U_{\textsf{id}}] attains its maximum at pA=1p_A = 1, assuming nA>nBn_A > n_B. That is, the identity payoff suggests we should just pick the candidate with the majority vote. (Or the plurality vote, if C>2|\mathcal{C}| > 2.)

This accords well with how we tend to actually run elections, but there’s something very weird about using the identity function as a utility function, which is that an affine utility function is necessarily risk-neutral. That is, an AA-voter with an identity payoff curve is indifferent between a world where pA=0p_A = 0 or pA=1p_A = 1 with equal chance, and a world with the certain outcome pA=1/2p_A = 1/2. Most people in real life are not risk-neutral! Getting a sure fifty dollars is preferable to a fair coin toss for one hundred dollars; we want to be paid a premium in order to take on risk. 7

Mathematically, we can represent this risk aversion by choosing a concave function ff for our voter payoffs: we are imposing a diminishing marginal utility for incremental gains in our favored candidate’s probability of victory.8 One intuitive way to do this is to say that the AA-voter cares about percentage increases to pAp_A, not absolute increases as in the naively risk-neutral case. Put another way, the utility of the marginal increase in pAp_A is inversely proportional to its current value (as opposed to constant), giving the differential equation f(x)=1/xf'(x) = 1/x. Of course, the solution is f(x)=lnx+Cf(x) = \ln x + C, where we don’t really care about the constant.9 So we have:

E[Ulog]=iCniNlnpi, E[U_{\textsf{log}}] = \sum_{i \in \mathcal{C}} \frac{n_i}{N} \ln p_i,

subject to the constraint that the pip_i’s must sum to unity. We can discover the maximum with a Lagrange multiplier, setting E[Ulog]=λ1\nabla E[U_{\textsf{log}}] = \lambda\vec{1}. This is just E[Ulog]/pi=ni/Npi=λ\partial E[U_{\textsf{log}}] / \partial p_i = n_i / Np_i = \lambda, which (normalizing to get a probability distribution) has solution pi=ni/Np_i = n_i / N. But this is exactly a random ballot election: each candidate should have probability of victory equal to her proportion of the electorate!


I think this is an interesting argument for pure random ballot elections, starting from first principles. Now I don’t think it’s totally convincing, even if you allow the premise that expected utility maximization is reasonable here. In particular, although economists often use logarithmic utility functions because of their analytic convenience, it need not be the case that real people’s preferences follow such a rule. Indeed, taking a different function would yield a different result: consider the function f(x)=3xx2f(x) = 3x-x^2, which is strictly concave and monotonically increasing over [0,1][0,1]. In the simple two-candidate election, E[Uf]E[U_f] is maximized at pA=max{(4nAN)/2N,1}p_A = \max\{(4n_A-N)/2N,1\} (exercise for the reader!). Put plainly, this means that a society with such a utility function would prefer some change of BB winning if support for AA is below a three-quarters supermajority, but above that threshold it would prefer that AA win with certainty.

As I alluded to in the beginning of the post, there are also significant questions about whether maximizing expected utility (i.e. a naive classical utilitarianism) is even the right framework to apply here, with Rawls submitting significant objections himself. I have some half-baked thoughts, but no particularly coherent argument for either side, so I guess that will have to wait until a future post. At this rate, I’ll be impressed if I can get that post out before late 2026.


  1. Perhaps my American bias is showing, in implicitly assuming something like a two-party system. But I think the results here generalize fairly well to the nn-candidate setting; in fact, I think they might actually be stronger, or at least more intuitively plausible, if there are many candidates running.↩︎

  2. You could argue that this is not really well-defined in the case of a tie, i.e. nA=nBn_A = n_B. But I would say that this is actually an excellent argument for the random ballot proposal!↩︎

  3. In fact, the proposed “underdog utility” was not even monotonically increasing, which seems like a desirable property from a well-behaved utility function.↩︎

  4. One way to frame it: I’ve heard the interpretation that Rawls advocates for a maximin decision rule, as opposed to naive expected utility maximization. I have not actually read Rawls, and so I cannot comment.↩︎

  5. I do think there is a slight sleight of hand here: what does it mean for you to receive a payoff based on the probability of an event, rather than the event itself? Sure, you could say that this is tantamount to thinking about your expected payoff from the event (or other properties of the payoff distribution, not just the first moment). But in real life, most of us think about payoff in terms of events that we care about, not the probability distributions of those events. To put it evocatively: if you win the lottery, should you be happy that you realized a huge gain, or sad that you made a negative expectancy decision? I think the strongest argument that I can give for this kind of probability-maximization over looking at raw event payoffs is that repeated games, especially over evolutionary timescales, should push you toward expectancy as a reward function. This strikes me as somewhat analogous to the claim that you should prefer rule utilitarianism over act utilitarianism, although it is not exactly the same thing.↩︎

  6. It is interesting to note that for a fixed population size, expected utility maximization is obviously equivalent to total utility maximization. Of course, you can get really weird results if you allow the population size to vary; it becomes difficult to avoid dilemmas like Parfit’s “repugnant conclusion.”↩︎

  7. Though rather infamously, Sam Bankman-Fried once made the claim that an effective altruist should be risk-neutral with respect to wealth accumulation, because there is no diminishing utility for the marginal dollar allocated to charity. Needless to say, this claim was rather controversial; I liked this blog post by Sarah Constantin explaining why this is probably not correct. I especially enjoyed her analysis of the St. Petersburg paradox, on which Bankman-Fried also held controversial opinions.↩︎

  8. The connection between risk aversion and concavity is by Jensen’s inequality: an agent is risk-averse if the utility assigned to the certain outcome E[X]E[X] is greater than that assigned to the random variable XX itself, i.e. u(E[X])>E[u(X)]u(E[X]) > E[u(X)].↩︎

  9. One might fret that this is not totally rigorous, as lnx\ln x is not defined for x=0x = 0. But to channel the lecturer in my undergrad category theory class: it’s best not to worry too much about such details if you can help it. (I think he was addressing some concerns with sizes in definitions, and assuring us that they could be overcome with Grothendieck universes.↩︎


Comments

Submit a comment

Your comment will be held for moderation. If needed, I'll reach out to the provided email address with moderation updates. Your email will not be publicly displayed.

Note: comments are still in beta. Let me know if anything is broken!