From Noűs 34, 4 (2000): 550-557.

 

Reprinted by permission of the author and Blackwell Publishers.

 

Visit Noűs at:

 

http://www.blackwellpublishing.com/journal.asp?ref=0029-4624

6

Parfit’s Puzzle[1]

Philip Kitcher

 

1       The Basic Plot

In the brilliant final section of Reasons and Persons, Derek Parfit presents a puzzle about how the goodness of states of affairs relates to the quality of the lives led by people in those states.[2] Stripped to barest essentials, the puzzle runs as follows: if the value of a state is obtained simply by aggregating the quan­tity of whatever makes life worth living, then a world in which a significant number of people (say ten billion) enjoy lives of very high quality would be inferior to a world in which a vastly greater number of people have lives that are barely worth living (the Repugnant Conclusion); if the value of a state is measured by the average quality of the lives led in that state, then a world in which a few (even one or two) lead lives of exceptional quality would be su­perior to a world in which billions lead lives whose quality is just slightly lower; neither of these consequences is acceptable, and hence neither aggregation nor averaging captures the relation of the quality of a world to the quality of its constituent lives; so what does? Simplifying the puzzle as I have done ignores the richness of Parfit’s discussion, and, in particular, the ways in which he de­vises scenarios that block obvious and not‑so‑obvious attempts at solution.

Questions of this general type—How do we understand collective utility in terms of the utility of individuals?—have, of course, been systematically studied in a number of contexts, and there are well‑known results.[3] None­theless, as philosophical economists (and economically literate philosophers) are well aware, ‘utility’ is a slippery term, and it is important to think carefully about what it means in the particular context of investigation, so that the problem to be addressed is set up with the right conditions.[4] Parfit’s puzzle is inspired by concerns about overpopulation (381), and thus falls within ‘population axiol­ogy’ (rather than, say, the theory of voting procedures). The pertinent notion of utility can’t therefore be construed in terms of the preferences of existing peo­ple but is, rather, grounded in the quality of the lives led by generations yet unborn.[5]

Parfit declares that he is in search of ‘Theory X’, which we can think of as a function that takes representations of the quality of individual lives into some measure of the overall goodness of the world which contains exactly those lives. In subsequent discussions, various authors have proposed candidates, offered counterexamples to the suggestions of others, and, more recently, started to ar­gue that the problem is insoluble.[6] Some of the literature proceeds relatively informally; other contributions examine principles that are at some remove from Parfit’s actual arguments. What I hope to show is that a natural extension of Parfit’s reasoning, when formalized, establishes an interesting impossibility re­sult. As already noted, this is not Parfit’s own assessment of his discussion (see, for example, 443), but it is, I think, the view to which he is committed.

Before undertaking this project, it’s worth underscoring the importance of Parfit’s puzzle. Foes of consequentialism might conclude that the difficulties only arise because of the hopelessness of maintaining that the good is prior to the right. They would be quite wrong to do so. Most views of moral delibera­tion assume that deliberators are able to assess the value of the states that might be produced by their actions.[7] If they suggest that value is affected by the moral quality of the decisions and actions that produce those states, that is itself a view about the form of the function that assigns value: the function can’t be sensi­tive only to the numbers of individuals who lead lives of particular qualities, but must take into account how the activities that produce and develop those lives accord with prior moral principles. Opponents of consequentialism will thus agree from the start that the two simplest functions (aggregation and av­eraging) will not do, but they’ll still face the challenge of saying what the cor­rect function is. One of the great merits of Parfit’s own elaboration of his puzzle is that he demonstrates how a number of prominent non‑consequentialist ideas don’t help us solve it (see, for example, 422‑25).

1       A Simple Formalism

I’ll start by supposing that each person’s life can be associated with a real num­ber that measures how well it goes. The measure applies to the whole life, so that it’s a timeless truth about the individual p that p’s life has quality Q(p). The better p’s life, the higher the value of Q(p). 0 marks the boundary at which lives become worth living: with respect to people for whom Q(p) < 0, we can truly say of them that it would have been better if they had not existed.[8] I’ll set a scale far the quality of lives by supposing that the best lives we know (from history, contemporary acquaintance, or even from fiction) have quality 100.

There are some serious assumptions here, and it’s worth noting them explic­itly. First is the Single Number Assumption: Even if the quality of a life is de­     termined by many different kinds of factors, it’s possible to weigh these against one another and identify a single numerical measure. Parfit’s diagrammatic rep­resentations of parts of his argument effectively embody this assumption, for he displays the relative qualities of lives by drawing lines that extend more or less far up a vertical axis (see 385, 388, 419 for examples). I’ll return to the Single Number Assumption at the very end of this essay.

Second is the Ratio‑Scale Assumption: If Q is an adequate measure of the quality of lives and if Q* is also an adequate measure of the quality of lives, then there’s a strictly positive number k such that, for all p, Q*(p) = kQ(p). It is not plausible to assume that there are ‘natural units’ of the quality of lives, so that we could talk of a unique number that ‘really’ measures the quality of a life.[9] The Ratio‑Scale Assumption assures us that, nonetheless, we can mean­ingfully speak of the ratio of the qualities of lives (as Parfit does: see 385, 389 for examples), independently of the measurement scale we choose. Now in the context of many aggregation problems about collective ‘utility’, this would be a very strong supposition, but I believe that it is defensible in the present case. Parfit’s puzzle starts from recognizing that we can judge that some lives go ex­tremely well, while others are barely worth living. Given that there’s an invari­ant zero, we know that, if Q and Q* are both adequate representations, then Q*(p) = g(Q(p)).Q(p), where g is some strictly positive function. Making the relatively uncontroversial assumptions that transformations that take adequate representations to adequate representations can be iterated and/or inverted[10], consider the ratio g(Q(p+))/g(Q(p!), where p+ is a person whose life goes exceptionally well and p! is a person whose life is barely worth living. If this ratio is less than 1, then iterating sufficiently often will lead us from an initial representation of the relative qualities of the two lives which is very large to a supposedly adequate representation that makes that ratio as small as you please. If the ratio is greater than 1, then we can consider the inverse transformation and iterate that to produce just the same effect. Either route to the same end spells doom for the supposition that we can talk of one kind of life going vastly better than another. Thus the ratio ought to be 1, so that g(Q(p)) reduces to a positive constant.[11]

My third assumption is the Full Comparability Assumption: For any ade­quate representation of the quality of one person’s life, there’s a unique ade­quate representation of the quality of any other person’s life. Again, this is far stronger than suppositions made in the context of other utility aggregation prob­lems, where interpersonal comparisons are often treated with suspicion.[12] In the case at hand, however, many of the normal sources of concern are absent‑for the task is not to assess how much better state X would be for person p (an ac­tual person with actual preferences) than state Y would be for person q (another actual person with actual preferences)—but to consider the lives that would be lived by possible future people and to evaluate them according to the presence or absence of features we regard as crucially important to our own well‑being. We do make such judgments (and it’s arguable that responsible people ought to make such judgments), so that some interpersonal comparability is required. Parfit himself hedges a bit, by talking of ‘rough or partial comparability’, an idea that we could express by supposing that fixing the quality of one person’s life selected a restricted set of assignments of quality of life to others, but this would endanger the comparisons of lives that are well worth living with those that are barely worth living (so that a situation that counted as Repugnant under one representation might become acceptable under a different redescription).[13] I suspect that there are ways to relax the assumption of Full Comparability that will generate Parfit’s puzzle, but I won’t explore them here. Not only is the treat­ment I’ll offer simpler, but recent discussions of interpersonal comparisons of utility provide at least the beginnings of accounts about how such comparisons might be justified.[14]

With these assumptions explicitly noted, we can proceed by taking there to be a single real‑numbered scale on which the quality of different people’s lives can be measured, and we can forget, for the rest of this essay, about the infinite number of rival scales that would multiply our ‘units’ by a positive real num­ber. Any world can now be represented as an ordered pair of vectors n, Q such that ni (the i‑th component of n) is exactly the number of people in that world whose lives have quality Qi (the i-th component of Q). Call such representa­tions ‘thin descriptions’. A fundamental thesis of consequentialism suggests that only numbers and the measure of the good count in ascribing value. Our prob­lem, then, is to try to find a function that maps thin descriptions of worlds to measures of the value of the worlds so represented.[15]

In fact, this is not quite right, because we can represent any world in infi­nitely many ways, and so worth must be assigned by any of a family of func­tions. Thin descriptions have a number of components that depends on the degree of stratification of the world, that is, on the number of distinct levels of quality that the lives in it achieve. Some worlds are homogeneous: they contain a num­ber of people, all of whose lives have exactly the same measure of whatever makes lives worth living. Parfit mainly considers homogeneous worlds and the simplest heterogeneous worlds, those with just two classes of lives. But there’s no limit to the number of distinct classes there might be. Consequently, we can’t operate with a single function whose arguments are 2k-vectors (for some appropriate number k)—V = V(n1,…,nk,Q1,…,Qk)—since we could always envisage worlds for which there were more divisions among the qualities of lives.

Instead, we have to work with an infinite set of value functions, Vk, each defined on 2k-tuples. When we are comparing a pair of worlds we have to em­ploy the function that accommodates the more heterogeneous, and this means that we must use an inflated thin description of the less heterogeneous world. Suppose, for example, that we want to compare w = <10,100> with w* = <<9,1>, <200, 1>>. We proceed by adopting an inflated representation of w as <<10,0>,<100,1>> and comparing the values of V2 (10,0,100,1) and V2 (9,1,200,1). This trick will work perfectly generally so long as we require that inflated measures reduce to uninflated measures.

More precisely, let’s introduce some obvious principles that our value func­tions ought to satisfy:

 

(1) Let F be any permutation on 1,...,k.

     Then śk V2k(n1,...,ni,...,nk,Q1,...,Qi,...,Qk) =

     V2k(nF1,...,nFi,..,nFk,QF1,...,QFi,...,QFk)

     (2) śk, Q V2k(n1,...,nk,Q1,...,Qk)= V2(k+1)(n1,...,nk,Q1,...,Qk,Q)

 

(1) tells us that we can always reshuffle the order of the arguments as long as we maintain the correspondence between numbers and lives measured by a par­ticular quality. (In other words, the order in which we list the quality strata doesn’t matter.) (2) tells us that we can throw into our representation zero lives of any quality whatsoever without changing the value assigned to the world.

I’m going to exploit the reshuffling principle, (1), to propose a canonical thin representation for a world: we’ll always order the numbers and quality‑measures by proceeding from highest to lowest quality; when we compare two worlds with different stratifications, we’ll insert zeros for the more homogeneous world and associate these with the corresponding quality measures in the comparison world. Hence, instead of <<5,1>, <100, 200>> we’ll prefer <<1,5>,<200,100>> and in comparing this world with <<1,2,1,6>,<300,200,100,20>> we’ll adopt the representation<<0,1,5,0>, <300,200,100,20>>.

A more substantive principle than either of those so far introduced officially commits us to the same number quality principle:

 

     (3) Suppose that śi Qi* $ Qi with the inequality holding strictly in at least one case. Then V2k(n1,...,nk,Q1,...,Qk) < V2k(n1,...,nk,Q1*,...,Qk*).

 

It also seems plausible to advance a disaggregated version of the principle: if a group that’s homogeneous in one world corresponds to two groups in another world, one of which has exactly the same measure of quality and the other of which has a higher measure, then the second world has greater value than the first. In fact, this principle follows from (3) provided that we allow the possi­bility that Qi = Qj when i j, and also suppose that

 

(4) V2k(n1,...,nk,Q1,...,Qk) = V2(k+i)(n1,...,ni!r,...,nk,r,Q1,...,Qk,Qi).

 

We can now see very clearly how Parfit’s arguments against Simple Aggre­gation and against Averaging work. The Simple Aggregation Principle is:

 

(5) V2k(n1,..., nk,Q1,...,Qk) = GniQr

 

(5) accords with (1)‑(4), and thus meets the constraints on value functions so far introduced. As Parfit sees, however, (5) has an implausible consequence. Con­sider two homogeneous worlds, w1 = < 1010,100 > and w2 = <1020,10!5 >. In the second world, there are an enormous number of people whose lives are barely worth living; in the first world, there are many people who lead lives of very high quality. We suppose that w1 is better than w2. Hence, V2(1010,100) should be greater than V2(1020,10!5 ). But (5) takes the farmer value to be 1012 and the latter to be 1015. So (5) yields the wrong result.

The values chosen represent a very large class of cases. The fundamental point is that we have a new constraint.

 

(6) m, n, Q, Q* such that mQ > nQ* and V2(m,Q) < V2(n,Q*)

 

(6) embodies Parfit’s sense of repugnance at his conclusion, and it is incompat­ible with (5). (Plainly, (6) has analogues for other functions in our family which take a larger number of arguments.)

The Averaging Principle is easy to formulate.

 

(7) V2k(n1 ,..., nk,Q1 ,...,Qk) = GniQi /Gni.

 

As with Simple Aggregation, this will accord with (1)–(4). To see why (7) is problematic, we need only compare w3 = <1,100> with w4 = <1,1010,100, 99.99999>. Intuitively, w4 in which a large number of people live extraordi­narily well is preferable to w3 in which only one person lives at the very high­est level attained in w4. Thus we should have V2(1,100) = V4(1,0,100,99.99999) < V4(1,1010,100,99.99999). (7) reverses this inequality, and thus should be rejected. In fact, we can extend the point, recognizing that it’s coun­terintuitive to rank <1,100> ahead of <1010, 99.99999>.[16] But, according to (7), we ought to accept that.

Once again the numbers are representative of a class of values, and the im­portant point is to acknowledge a constraint.

 

(8) m, n, Q, 0  such that 0 > 0 and V2(m,Q) < V2(n,Q!0).

 

A solution to Parfit’s Puzzle must satisfy (1)–(4), (6) and (8). Plainly, there may also be other constraints, as yet unrecognized. Parfit believes that there are. But we can actually satisfy the ones he offers.

2       Paradise Regained—The Compromise View

Simple Aggregation goes astray because, as the numbers increase, aggregate quality matters less than average quality. Averaging is wrong because it gives no importance to numbers at all. One[17] natural response is to compromise, giv­ing weight to aggregation when the numbers are small and increasing the weight of averaging as the population becomes larger. The general form of the func­tion we seek would then be

 

(9) V2k(n1,...,nk,Q1,...,Qk) = N(N)GniQi + (1!N(N))L GniQi/Gni.

 

Here, N can be identified with aggregate numbers (N = Gni) or with aggregate quality (N = GniQi),[18] and L is a constant dependent on the level at which in­creasing numbers cease to count.[19] The function N should be 1 (or close to l) when N is small, and should go to 0 as N becomes large.

There are, of course, infinitely many functions of this general form. To un­derstand the successes of the approach (as well as to identify important constraints on any adequate function) it will help to consider a very simple one.

 

(10) If Gni < N,V2k(n1,...,nk,Q1,...,Qk) = GniQi;

if Gn $ N, V2k(n1,...,nk, Q1,...,Qk) = NGniQi/(Gni).

 

In other words, below a certain population size, N, V is just simple aggrega­tion; at or above that level, it’s N times the average quality of life; thus, above N, V behaves like an averaging function (more precisely it responds to the av­erage quality and multiplies it by a parameter so as to be continuous at the thresh­old population size}.

It’s elementary to verify that (10) satisfies our earlier adequacy conditions, (1)–(4), (6), and (8). The significant point is to see how (10) evades the Repug­nant Conclusion. Let’s focus on the version of the puzzle that Parfit presents in elaborating his ‘Mere Addition Paradox’. Here he describes a sequence of worlds that we can represent as follows:

 

A = <107,0,100,50>

A+ = <107,107,100,50>

New A = <2.107,1010,100,1>

New B = <4.107,2.1010,80,1>

New C =<8.107,4.1010,60,1>

New D =<16.107,8.1010, 40,1>

…………………………..

New Z = <1020,1020,1,1>

 

Recall that (10) favors simple aggregation when the population size is beneath a value N. If 2.1020 < N, then the value assigned to worlds increases as we go down the alphabet (this just reflects the fact that while we are stuck with simple aggregation we are forced to the Repugnant Conclusion). If, however, N < 2.1010, the Repugnant Conclusion will not necessarily obtain. Just where the alphabetically increasing value of worlds ends depends on the value of N. If N < 4.107/3, then A is preferable to A+, and the conclusion is blocked at the start; if 4.107/3 < N < 2.107, then A+ is preferable to A, and also (markedly!) to New A. As N approaches 1010 from below, New A becomes preferable to A+, and, while N remains below 1010, New A is preferable to New B. Plainly, how­ever, if N is above 2.1010, New B will be assigned higher value than New A (since simple aggregation will favor New B). There are values of N that halt the march to the Repugnant Conclusion between New B and New C, others that do so between New C and New D, and so on.

This raises an obvious question: which is the right value of N, and which point in the sequence does it mark out as the right point to halt the assignment of increasing value? Consulting intuitions on this point is of no use. That’s be­cause there’s no definite point at which aggregation gives way to averaging, any more than there’s a definite number of grains of sand at which we achieve a heap.

My formalisation motivates an obvious strategy for coping with this aspect of the puzzle, akin to the method of supervaluations. Let’s suppose that we can exhaustively divide values of N into those that are definitely admissible and those that are definitely inadmissible (allowing for the possibility that admissibility itself is vague would complicate the treatment but wouldn’t change its essen­tial features). Inadmissible values of N consist of those that are too small and those that are too large. Now we can define an ordering of worlds in terms of their values as follows:

 

     (11) w >v w* just in case, for every function V of the form of (10) that uses an admissible value of N, V(w) $ V(w*) with the inequality holding strictly in at least one case.

     (12) Let the admissible functions according with (10) be those for which n2 $ N $ n1. Suppose that V(w) > V(w*) for exactly those admissible func­tions according with (10) for which m2 $ N $ m1. Then d(w, w*), the de­gree to which w is more valuable than w*, is (m2 – m1)/(n2 – n1).

 

(11) allows for the possibility that some worlds aren’t ranked as more valuable than others (because they are ordered differently by different admissible func­tions); (12) allows us to treat this range of incomparable pairs of worlds in terms of a degreed notion of relative value (provided that the pertinent values of N form a connected subinterval of the interval of admissible values).

Functions of form (10) are special cases of the more general form (9), and it seems highly likely that the admissible functions for ranking worlds will make up a much vaster set. For (10) proposes that aggregation stops and averaging begins all at once, whereas there are infinitely many ways of producing a grad­ual shifting of the balance between these considerations (Parfit recognizes this; see 403–4). Hence we ought to generalize to refer to all admissible functions in (11), and to compute degrees in terms of measures of sets of functions in (12).

Do (11) and (12) (or their more general versions) subvert any of our earlier conclusions? No. (1)–(4), (6) and (8) are still satisfied. We avoid the Repug­nant Conclusion (in the particular case of the puzzle we’ve been considering) if all admissible values of N are less than 1020, which is plainly plausible. And, as we intended, we block the march from A to New Z without committing our­selves to the view that there is some definite point at which worlds start getting worse.

So have we solved Parfit’s puzzle? In the form in which he presents it, we have. Building on (9) and (10), and deploying (1l) and (12), we can avoid the Repugnant Conclusion, without generating counterintuitive consequences that doom simple averaging—Parfit’s scenarios for undermining the appeal to aver­aging trade on the idea that the number of lives in the superior world is small (i.e. below the threshold value), and, of course, our version of the Compromise View, to wit (10), uses aggregation for small worlds.[20] Parfit himself thinks that the Compromise View encounters troubles of particular kinds, but, when his sce­narios are formalized, it’s not hard to see that (10) gives the right results.[21]

Yet one might worry that our exhibition of a function that satisfies Parfit’s own constraints is an artefact of particular arbitrary choices we have made about population sizes and the values of the lives that people lead. In setting up the sequence of worlds that is supposed to drive us to the Repugnant Conclusion I followed Parfit’s own account of the numbers. The range of values of N that generate admissible functions of form (10) was a direct response to those par­ticular numbers. One might suppose, however, that the same puzzle could arise with different levels of quality (on the same fixed scale) and/or with different population sizes, provided that the same structure was preserved. Imagine that, instead of the sequence given above we were offered:

 

Diminished A = <105,0,100,50>

Diminished A+ = <105,105,100,50>

Diminished New A = <2.105,108,100,1>

Diminished New B = <4.105,2.108,80,1>

Diminished New C = <8.105,4.108,60,1>

Diminished New D = <16.105,8.108,40,1>

…………………………..

Diminished New Z = <1018,1018,1,1>

 

I’ll assume that this new sequence also gives rise to a Repugnant Conclusion, so that we ought to be able to show that Diminished New Z is inferior to Di­minished A. If all admissible functions of form (10) take N to be less than 1018, this result will follow. Nonetheless, as we move from one version of the prob­lem to another, the region of the sequence in which the Repugnant Conclusion is blocked may vary. For smaller population sizes, we may have to go further down the alphabet to discover the places at which worlds no longer increase in value.

Some people may be prepared to accept this feature of the solution—or even the further idea that there are similar sequences for which the Conclusion is no longer Repugnant (consider the case in which the analogue of A is <1,0,100,1> and the analogue of Z is <104,1010,100,1>). Others may hold that the region of the sequence in which the Conclusion is blocked should remain the same. They endorse a requirement of population‑size invariance:

 

     (13) Let <w1,...,wk> be a sequence of worlds. Let 8wi be the world whose populations at different quality levels are 8 times the corresponding populations in wi (thus, e.g. if w = <N,M,Q,Q*>, 8w is <8N, 8M,Q,Q*>). The relative values of the wi must be the same as those of the 8wi.

 

It’s easy to see that (13) is incompatible with any approach based on (10). For suppose that admissible values of N lie between 1010 and 1012. Let w1 = <104,100>, w2 = <1012,0.1>. Then, for all admissible values of N, V2(w2) > 109, while V2(w1) = 106. When we consider 1010w1 and 1010w2, however, we get V2(1010w1) > 1012 > V2(1010w2), reversing the inequality between V2(w1) and V2(w2).

I’m not sure if one ought to accept the invariance requirement (13). Those whose intuitions accord with (13) can’t develop the Compromise View along the lines of (10). Instead, they need a different way of identifying the point at which aggregation gives way to averaging (or, more generally, the way in which averaging comes to have greater weight in the assignment of value). It can’t be a matter of sheer numbers in the populations, but the distribution of numbers across quality levels has to count.

This idea is hard to formalize in general, but special cases are straightfor­ward. When we consider worlds with just two kinds of people, the average qual­ity of life can be used as an index of the extent to which value accrues by aggregation. Thus

 

     (14) If (N1Q1+N2Q2)/(N1+N2) < R and N1Q1+N2Q2 $ 0, then V4(N1,N2, Q1,Q2) = (N1Q1+N2Q2)(1 ! R+Q#)

where Q# = (N1Q1 +N2Q2)/(N1 +N2);

If (N1Q1+N2Q2)/(N1+N2) < R and N1Q1+N2Q2 < 0, then

V4(N1,N2, Q1, Q2) = (N1Q1+N2Q2)(1+R ! Q#);

If (N1Q1+N2Q2)/(N1+N2) $ R, then V4(N1,N2, Q1,Q2) = N1Q1+N2Q2.

 

It’s not hard to check that (14) will satisfy the basic demands (1)–(4), (6) and (8). We evade the Repugnant Conclusion, since, at some stage in the alphabet­ical sequence, the average quality of life falls below R and the aggregate qual­ity is then multiplied by the factor (1 ! R + Q#), assuming that Q# remains positive. More concretely, suppose that all admissible values of R are greater than or equal to 2; since Q# approaches 1 as we go down the alphabet, the mul­tiplier is either negative or approaches zero.

This is only a sketch of a rather different way of developing the Compro­mise View. As I’ve already noted, I think that the invariance requirement is con­troversial. It does, however, represent a general type of consideration that Parfit discusses, namely the idea that some worlds are less valuable because the dis­tribution of quality within them is unequal. Egalitarian concerns can be intro­duced into an account of the value of worlds either by focusing on acts which cause inequalities (which would involve descriptions that are not acceptable to consequentialists) or by considering features of the distribution, no matter how it is caused. The general challenge for egalitarian consequentialists is to mea­sure the degree of inequality in worlds with arbitrarily many levels of stratifi­cation and to show how the degree of inequality affects the values of these worlds. An approach along the lines of (14) is a very crude way of meeting this challenge in a very simple case. Since I believe that the degree of inequality of a distribution does affect the value of a state, I think that the general problem would have to be solved if we were to find the correct family of value‑assignment functions. But our present task is to see if there are functions meeting Parfit’s constraints. For the moment, we seem to have done precisely that: the Compro­mise View appears to provide a defensible reconstruction of our intuitions.

3       Paradise Relost: Troubles for the Compromise View

But only for the moment. I’ll start with an example that is easily fixed, but that points the way to less tractable scenarios.

The Two Large Hells.[22] Let w3 =  <N,0,!100,!90>, where N is above the threshold for (10); let w4 = <N,M,!100,!90>, where M > 0. Intu­itively w4 is worse than w3—or, at best, no better than w3—since it contains extra people who live appalling lives. According to (10), however, w4 is better than w3, since the size of both worlds is beyond the point at which we start to measure by average quality, and the average quality in w4 is higher (V4(N,0,!100,!90) = !100N < N((!100N !90M)/(N+M)) = V4(N,M,!100, !90)).

To fix this, we need only recognize the asymmetry between the sum of pos­itive contributions to the quality of life and the sum of suffering, (albeit in a way quite different from that considered by Parfit, 407 ff.). (10) should give way to:

 

(10*) If Gni <N or if GniQi <0, V2k(n1,...,nk,Q1,...,Qk) = GniQi;

if Gni $ N and if GniQi $ 0, V2k(n1,...,nk,Q1,...,Qk) = NGniQi/(Gni).

 

It’s not hard to verify that (10*) preserves all the advantages of (10) for the cases considered in the last section, and that it gives the appropriate ranking to w3 and w4.

The Really Absurd Conclusion.[23] Unfortunately, (10*) doesn’t guard against a different scenario in which a small amount of intense suffering figures in an otherwise wonderful world. As before, let N be above the threshold for (10) (and (10*)), and consider the worlds <N,m,100,!100> and <N,M,100,90>. Assuming that m is small in comparison with N, the aggregate quality in the former world will be positive and (10) and (10*) assign the same value. Thus we have: V4(N,m,100,!100) = N(100(N ! m)/(N+m)), V4(N,M,100,90) = N((100N+90M)/(N+M)). Now surely a world containing N people living best quality lives and m people living lives of intense misery is worse than a world in which the fortunate N are joined by some large number M people whose lives, while not quite of best quality are, nonetheless, extremely good. Hence we want V4(N, m,100,‑100) < V4(N,M,100,90). But, given the values computed from (10)/(10*), this inequality can easily be reversed. Assuming, as we have done throughout, that N is at least 1010, (10) and (10*) will lead us to prefer the world <N,10,100,!100> to the world <N,1000,100,90>. This is a really absurd conclusion.

What has gone wrong? Parfit argued, cogently, that averaging fails at small numbers. But averaging breaks down at large numbers too, and it does so dra­matically when we compare different ways of reducing the average quality of life‑by adding a few examples of intense misery and by adding a larger num­ber of high quality lives. Because the latter additions can reduce the average more substantially than the former, averaging will lead us to prefer worlds con­taining pockets of extreme suffering. Once this point has been appreciated, we should reject efforts to justify anything like the Compromise View.

The Numbers Always Count. Consider two worlds containing enough people to lie above the threshold for (10)/(10*). The first world, w5 = <N,0,100,100!0> where 0 n 1; the second, w6 = <1000N,1,100,100!0>. According to (10)/(10*) V4(1000N,1,100,100!0) = N((105N + 100!0)/(1000N+1)) < 100N = V4(N,0,100,100!0) so that w5 >v w6. This means that a world in which a large number of people live best lives is superior to one in which a thousand times as many people lead best lives and one person lives a life of extraordinarily high quality, just slightly below the best. In my judg­ment, this conclusion is counterintuitive, and can only be sustained if one thinks that, for sufficiently large worlds, averaging (multiplied by a threshold value) is the appropriate measure. Once we have seen the Real Absurdity of this idea, we can endorse the intuitive verdict. Averaging works neither for worlds of abys­mally low quality (recall the Two Large Hells) nor for large worlds of very high quality. For all its promise, the Compromise View is unsustainable.

This moral echoes themes in Parfit. Parfit is fond of imagining worlds in which the population of a world divides into two disconnected groups (perhaps on different galaxies), so that there can be no question of an attempt to remedy any differences in the quality of the lives they lead. We can extend the compar­ison of w5 and w6 to worlds in which the quality classes are spatially separated in some fashion that constitutes an insuperable barrier, and invite comparison of w7 = <N,0,100,99> with w8 = <N,M,100,99>. I claim that, no matter how large (or small) M may be, w8 is preferable to w7, on the grounds that the extra lives in w8 are of such high quality that a world in which they were the only lives would be extremely valuable. Surely one cannot maintain that a world con­taining two populations of lives of extraordinary quality is inferior to one that simply contains one such population, unless one is in the grip of a commitment to averaging, which, we’ve already seen, can’t be defended.

If this conclusion seems dubious, imagine replacements for w8 that increase the quality of the slightly suboptimal lives. There’s a range of values between 99 and 100, and there will be some x strictly between 99 and 100 such that <N,M,100,x> is superior to w7 for all values of M > 0. If we consider our­selves as world‑changers within w7, and ask whether it would be worth creat­ing M new people who would live lives just like those of the N people already living except for one slight flaw—a mild attack of dyspepsia perhaps or an un­fortunate sequence of coughs at a symphony concert, then the answer surely ought to be affirmative.

This point is very much in the spirit of some of Parfit’s arguments against averaging in the small. It will be important later when I offer my attempt at an Impossibility Theorem.

4       Critical‑Level Utilitarianism: Success and Failure

Since averaging in the large turns out to have counterintuitive consequences, an obvious way to attack Parfit’s puzzle is to try to refine the ordinary notion of aggregation. Perhaps what goes wrong is that, although lives of positive qual­ity are worth living, there’s a critical level and it’s only when lives reach this level that a world containing them has a positive value.[24] The most straightfor­ward version of this approach, developed by the economists Charles Black­orby, Walter Bossert and David Donaldson, proposes the function

 

(15) V2k(n1,...,nk,Q1,...,Qk) = Gni(Qi ! ")

 

where " is the level (> 0) at which life quality makes a positive impact on the value of a world.[25] To fix ideas, let’s suppose for the moment that " = 20.

A Non‑Repugnant Conclusion. Simple aggregation fell victim to the Repug­nant Conclusion. How does Critical‑Level Utilitarianism handle the alphabeti­cal sequence of worlds? Plainly, there are stages of the sequence at which the value of the lives drop below the threshold—worlds <0,M,100,Q> where Q < 20. For such worlds, the value assigned by (15) will be negative (M(Q!20), in fact). So the original Repugnant Conclusion is blocked.

Nevertheless, there’ll be a last world in the sequence far which the quality of lives is above the threshold—a world w9 = <0,M,100,20+s> where s is small. Suppose that the first world in the sequence is w10 = <n,0,100,20+s>. If M is sufficiently large—M > 80n/s—there’ll be an analogue of the Repug­nant Conclusion, in that a world of many people (w10 with n large) who lead best lives will rank as inferior to a world in which an enormous number of peo­ple have lives just above the critical level (w9). Critical‑Level Utilitarians hold that this conclusion isn’t Repugnant. After all, they suggest, the lives in w9 are well worth living, and there are an enormous number of them.

The Ludicrous Conclusion. But we should now consider two further worlds, w11 = <n,M,100,20!s> and w12 = <0,1,100, 20+s>, where n, M, s are as in w9 and w10. According to (15), V4(n,M,100,20!s) = 80n ! Ms < 0, and V4(0,1,100,20+s) = s > 0, so that w11 is ranked as inferior to w12. This means that a world in which a large number of people (n) lead best lives and in which an enormous number of people (M) live lives just below the critical threshold is inferior to a world in which just one person lives a life just above the critical threshold.[26] But the difference between 20+s and 20!s could be truly minute—a matter of the fact that a few great musical works are missing or traffic conges­tion is slightly more severe. Nevertheless, Critical‑Level Utilitarians are com­mitted to believing that although enough 20+s lives could outweigh a large number of best lives, 20!s lives are so bad that adding the same enormous num­ber of them to a world full of best lives would make that world worse than one containing one life just above the critical threshold (and also worse than the null world in which nobody exists).

To see how bizarre this is, consider two further worlds, w13 = <2n/3,0, !100,20!s> and w14 = <0,M,100,20!s>. (15) assigns values of !80n and !Ms respectively. Since Ms > 80n, it follows that w14 is worse than w13. This means that the lives just below the critical threshold are so bad that an enor­mous number of them would generate a world worse than one in which a large number of people lead lives of intense suffering. This is to be consistent with the claim that a slight increase in the quality of those lives—relief from minor dyspepsia, the rescue of some important musical works, a decrease in traffic congestion—would transform them so that the same enormous number of trans­formed lives would be superior to a large number of best lives. I contend that this is a Ludicrous Conclusion.

It may be useful to consider the nature of the ordering of the worlds we’ve been examining. The following relations hold:

 

w9 >V w10 oV w12 >V w11 oV w13 > w14

 

The ordering asks us to think that an enormous number of 20+s lives is better than a large number of best lives, that both of these worlds are vastly prefera­ble to a world containing a large number of best lives and the enormous num­ber of 20!s lives, and vastly preferable to worlds vastly preferable to worlds containing just the enormous number of 20!s lives. Just focusing on w9 and w14, we see that a tiny change—the loss of a few great works of art, say—can change the world from one that is better than a situation in which a large num­ber of people live best lives to one that is worse than a situation in which that large number of people suffer intensely. No adequate account of the value of worlds can have this as a consequence.

Fuzziness to the Rescue. There’s an obvious reply to my indictment of Critical-­Level Utilitarianism. We generate silly comparisons by supposing that there’s a single critical level. But Critical‑Level Utilitarians should take their cue from Parfit, who is quite clear that talk of levels ought to recognize a band across which ‘the personal value of a life increasingly diverges from the value that this life contributes to the outcome’ (412).[27] In section 3, I offered (11) and (12) as a way of defending the Compromise View against the charge of arbi­trary discontinuities. Something similar can be introduced here to ward off the objections so far presented.

Instead of a single critical level, we now operate with a range of levels. One world is ranked higher than another just in case every choice of a level from the range would yield a substitution instance of (15) in which the former world had higher value than the latter. Thus:

 

(16) w >V wN just in case śx0[",$]  Gw ni(Qi ! x) > GwNniN(QiN!x)

w .V wN just in case x, y 0[",$] Gw ni(Qi!x) > GwNniN(QiN! x) and

Gwni(Qi!y) < GwNniN(QiN!y).

 

For concreteness, let [",$] = [20,25].

The analogue of the Repugnant Conclusion now consists in comparing w15 = <n,0,100,25+s> with w16 = <0,M,100,25+s>, where M > 80n/s. For all x in [20,25], the value ascribed to w16 exceeds that of w15—for the former is M(25+s!x) > Ms, and the latter is n(100!x), which is at most 80n < Ms. If we now compare w11 and w12, we find that they are roughly equivalent (w11 .V w12), so that we avoid the previously troublesome conclusion that ranked w11 behind w12. The best we can manage by way of a surrogate is to compare w11 with w17 = <0,1,100,25+s>. Plainly, for every x in [20,25], we have

 

V4(n,M,100,20!s) < 80n ! Ms < 0 < s < V4(0,1,100,25+s)

 

Thus w17 clearly counts as better than w11. But it’s no longer so clear that this gives rises to counterintuitive consequences, since the difference between the lives in the worlds is no longer 2s but is much greater (5+2s), and so cannot be glossed in terms of the relatively trivial differences we cited earlier.

Why Fuzziness Fails. Our revised version of Critical‑Level Utilitarianism evades the original form of the Ludicrous Conclusion. Yet consider two further worlds, w18 = <0,M,100,20+s> and w19 = <0,M,100,25!s>. w18 no longer counts as superior to w10, the world with n best lives, but it’s viewed as roughly equivalent. If the new approach is to succeed, we must agree that this conclu­sion isn’t Repugnant. But now

 

V4(0,M,100,20+s) ! V4(n,M,100,20!s) = M(20+s!x) !

(n(100!x) + M(20!s!x)) = 2Ms ! 100n+nx > 2Ms ! 80n > 80n.

 

Hence w18 is vastly preferable to w11, despite the fact that the former contains an enormous number of lives whose counterparts in the latter are only slightly worse (being shadowed in any of the trivial ways canvassed earlier), and the latter also contains a large number of best lives. This is the same counterintu­itive result we had earlier.

Furthermore, it’s not hard to show that w19 is roughly comparable to w10, the world with just n best lives and that w19 is vastly preferable to w18. In fact, the difference between the values assigned to w19 and w18 is 5M ! 2s, which is truly enormous in comparison with 80n, the maximum value assigned to w10. We thus have the anomaly of two worlds, each roughly equivalent to a world containing n best lives, but which differ in value by an amount orders of mag­nitude greater than the value assigned to the world with n best lives. However flexible our ideas about rough equivalence among worlds, this seems too great a variation between ‘equivalent’ worlds.[28]

As in the earlier argument, it’s not difficult to derive counterintuitive com­parisons between worlds which Critical‑Level Utilitarianism treats as having neg­ative value—consider, for example, <0,M,100,20!s> and <0,16n/25,100,!100>; the value of the latter is greater than !80n, while the value of the for­mer is at best !Ms; thus a world with an enormous number of 20!s lives is counted as worse than a world with a large number of lives of intense suffer­ing. This result is not only counterintuitive on its own, but also fares poorly when we recognize that the same enormous number of 20+s lives count as roughly equivalent to the same large number of best lives.

I conclude that the inadequacies of Critical‑Level Utilitarianism can’t be rem­edied by considering a range of thresholds. I take this to show that the ap­proach is beyond repair. This conclusion will be reinforced in the next section.

5       Towards a Proof of Impossibility: Assembling Some Principles

The approaches already explored and the scenarios on which they founder sug­gest some conditions that a value‑ascribing function ought to satisfy. In this sec­tion, I’ll try to assemble the principles. Later, I’ll argue that they are jointly unsatisfiable. Although I’ll try to motivate the constraints as I specify them, the attempted proof may provide reasons for questioning one or more of them; hence I’ll probe them further later (section 8).

I begin by supposing that our conditions (1)–(4) are uncontroversial. It will simplify the subsequent discussion if we concentrate (as indeed we have done for most of this essay) on worlds that have at most two quality classes. Hence I’ll formulate our constraints as conditions on the function V4. Further, we’ll be concerned mainly with the behavior of this function for worlds with a very large population, N; in this I follow Parfit, who, of course, formulates the Repugnant Conclusion for a world containing a large number of best lives. We’ll also sup­pose that s is a small number, s n 1.

Start with the moral of the Repugnant Conclusion. This tells us that if  s $ x > 0, then, for any number of people M, no matter how large, a world with M people leading lives of value x is worse than a world of N people leading best lives. In other words:

 

     [NRC] śx if s $ x > 0, then śM > 0, V4(N,0,100,x) > V4(0,M,100,x)

 

Next we have the moral of the anti‑averaging arguments, not only those of­fered by Parfit but also the more general considerations urged in critique of the Compromise View. These tell us that, however many people the world contains leading extremely high quality lives, adding more people whose lives are of slightly lower quality doesn’t make the world worse. In particular, if we com­pare a world with N best lives to one that adds a number of lives of quality just s below best (100!s lives), a world in which the two groups of lives never in­teract, then the second world is at least as good as the first.

 

     [NAV] śx if s $ x > 0, then śM > 0 V4(N,M,100,100!x) $ V4(N,0,100,100!x)

 

Parfit’s own discussion also motivates a Replacement Principle. Suppose that <N,M,100,100!x> is at least as good as <N,0,100,100!x>. Then there are k, y such that 0 < y < x, k > 0, and a world with kN people leading lives of quality 100!y would be better than a world of N people leading best lives. In other words

 

     [REP] If V4(N,M,100,100!x) $ V4(N,0,100,100!x) then y x $ y > 0, k > 0 V4(N,0,100,100!y) < V4(0,kN,100,100!y)

 

More general versions of [NAV] and [REP], in which the specific reference to best lives gives way to lives of ‘sufficiently high quality Q’, seem plausible; but we shan’t need stronger principles and so I won’t try to defend them here.

Together [NAV] and [REP] yield a condition that will serve as the basis step in an inductive argument.

 

[BAS] k > 0 V4(N,0,100,100!s) < V4(0,kN,100,100!s)

 

In my judgment, [BAS] is easily motivated directly by the considerations against averaging, but, in case these don’t seem sufficiently powerful, I’ve decom­posed its rationale into the simple entailment by [NAV] and [REP].

We now need to draw on the discussion of Critical‑Level Utilitarianism to capture the essential point of the scenarios that attacked the idea of a critical level. I imagined instances in which a world of N best lives was equivalent or inferior to a world of M people leading lives of lesser quality Q, and then con­sidered worlds containing only lives of quality Q!s. The intuitive judgment was that Q lives must be so good, if they can outweigh N best lives (provided that there are enough of them), that Q!s lives can’t be so awful that no world con­taining just Q!s lives has positive value. After all, the Q!s lives could be very similar to the Q lives, differing only in some rather trivial discomfiture. How can it be that a world containing the mildly discomfited lives have at most zero value, while a world containing Q lives should be superior to a world with a large number of best lives? So I advance a principle to the effect that there are no critical levels.

 

[NCL] śQ  > s, if M V4(N,0,100,Q) < V4(0,M,100,Q) then n > 0 V4(0,n,100,Q!s) > 0

 

The next principle I’ll articulate takes up a point that emerged in consider­ing the Compromise View. Suppose that Q‑lives have so low a quality that no amount of them can ever outweigh a world of a large number of best lives. Then, adding such lives to a world of high quality would inevitably diminish the qual­ity of the world. This is intended to capture the attraction of the Compromise View, by taking account of the idea that lives that are too low in quality don’t add to the value of an already high quality world. Notice that this is much weaker than the claim urged by the Compromise View that, when the population is big enough, lives that are below the average quality lessen the quality of the world. It only claims that something like this holds when the extra lives aren’t just be­low the average but of such poor quality that they could never add up to out­weigh a large number of best lives. We can think of this as a principle of the diminution of value through addition.

 

[DVA] śQ, if śM > 0 V4(N,0,l00,Q) > V4(0,M,l00,Q) then śn > 0 V4(N,0,100,Q) > V4(N,n,100,Q)

 

The last constraint embodies the idea that the context‑dependence of value is limited. Again, this is inspired by tile discussion of the Compromise View, where we found it hard to see how adding a galaxy of lives that, if it were to exist independently would constitute a valuable world, could lower the value of a world, even a world of high value. I don’t suppose that a life always make the same contribution to the value of a world, whatever the distribution of the qualities of other lives in the world—that would be to formulate some variant on the simple aggregation view—but I do think that there’s a limit to the vari­ant impact that a life can have. So I propose

 

[CIV] śQ, if śM > 0 V4(N,0,l00,Q) > V4(N,M,l00,Q) then śn > 0 V4(0,n,100,Q) < 0

 

We’ll return to the motivation for [CIV] again in section 8, since it seems the most controversial of the constraints I’ve articulated.

For the moment, however, it suffices to note that [DVA] and [CIV] directly entail a principle about the negative value of lives that I’ll use in the proof:

 

[NEG] śQ, if śM > 0 V4(N,0,l00,Q) > V4(0,M,l00,Q) then śn > 0 V4(0,n,100,Q) < 0

 

We’re now ready for the impossibility result.

6       Proof of Impossibility

There’s no way of resolving Parfit’s Puzzle in its full form. For the full form of the puzzle demands the specification of a function jointly satisfying [NRC], [BAS], [NCL] and [NEG]. But this contradicts the

 

Theorem.          There is no function jointly satisfying [NRC], [BAS],

              [NCL] and [NEG].

 

Outline of the Proof Strategy. The proof is by induction. We divide up the in­terval [0,100] into subintervals of length s. We then show that [BAS], [NCL] and [NEG] yield the Repugnant Conclusion, and, specifically, contradict [NRC]. [BAS] tells us that there’s a world of 100!s lives that is preferable to the orig­inal world with its N best lives. Now assume that there’s a world of 100!is lives (i a positive integer) that is preferable to the original world. We can now use [NCL] and [NEG] to show that there’s a world of 100!(i+1)s lives that’s pref­erable to the original world. Proceeding by induction, we show that there’s a world of lives of quality s that is preferable to the original world. This, of course, directly contradicts [NRC].

Proof. We assume that, for some integer K, 100 = Ks.[29] We know from [BAS] that there’s a k > 0 such that V4(N,0,100,100!s) < V4(0,kN,100,100!s). Suppose now that there’s a k > 0 such that V4(N,0,100,100!is) < V4(0,kN,100,100!is) but that there’s no k > 0 such that V4(N,0,100,100!(i+1)s) < V4(0,kN,100,100!(i+1)s). If i+1 = K, then our supposition entails that V4(N,0,100,s) < V4(0,kN,100,s) which contradicts [NRC]. So it must be that i+1 < K, so that 100!(i+1)s > 0.

It follows from [NCL] that there’s some n for which V4(0,n,100,100 ! (i+1)s) > 0. Hence, contraposing [NEG], there’s an M > 0 for which V4(N,0,100,100!(i+1)s) < V4(0,M,100,100!(i+1)s). So it follows that there’s a k > 0 such that

 

V4(0,kN,100,100!(i+1)s) > V4(N,0,100,100!(i+1)s)

 

contradicting our supposition. Therefore we can conclude that if there’s a k > 0 such that V4(N,0,100,100!is) < V4(0,kN,100,100!is) then there’s a k > 0 such that

 

V4(N,0,100,100!(i+1)s) < V4(0,kN,100,100!(i+l)s).

 

Using this result and our earlier consequence of [BAS], we know, by induc­tion, that there’s a k > 0 such that

 

V4(N,0,100,100!(K!1)s) < V4(0,kN,100,100!(K!1)s).

 

(Notice that we can’t apply the argument based on [NEG] any further, since (K!1)s + s = Ks = 100, so that 100 ! (K!1)s! s = 0, and the condition on Q in [NEG] is no longer satisfied.) But 100 ! (K!1)s = Ks ! (K!1)s = s. So we have shown that

 

For some k > 0, V4(N,0,l00,s) < V4(0,kN,100,s).

 

This contradicts [NRC].

Hence, there’s no function V4, jointly satisfying [NEG], [BAS], [NCL] and [NRC].

7       Diagnoses

I envisage two responses to the argument of section 7. One is that Parfit’s Puz­zle is shown to be another instance of the familiar sorites problem, and that what goes wrong is that [NCL] isn’t fully true. Now if [NCL] asserted directly some­thing analogous to the crucial premises in the paradoxes of the bald man and the heap—‘If a man with n hairs isn’t bald, then a man with n!1 hairs isn’t bald’—there might be a basis for saying that [NCL] isn’t fully true. Notice, how­ever, that [NCL] doesn’t assert that if enough Q lives can outweigh N best lives then sufficiently many Q!s lives can outweigh N best lives (provided, of course, that Q > s); that, I agree, could be attacked for being less than fully true if it were offered without support from premises that are fully true. Instead, [NCL] says that if life at level Q is so good that sufficiently many lives at that level could outweigh N best lives, then life at level Q!s can’t be so bad that no com­bination of lives at that level has positive value. In the objections to Critical ­Level Utilitarianism, I’ve tried to show why this counts as fully true: imagine a world of lives at some level below 100, with sufficiently many to outweigh N best lives; now imagine those lives slightly diminished in one of the trivial ways canvassed; the result can’t be a world that is no better than the empty world (i.e. no better than a world of zero value).

Thus efforts to reveal Parfit’s Puzzle as another sorites problem are not go­ing to be successful. The alternative diagnostic strategy is to identify one of the constraints that ought to be rejected. I’ve already argued that [NCL] should be retained. We’ll now consider the other conditions in turn.

[NRC] is extremely well‑motivated by Parfit’s own discussions. I believe that he is completely correct in maintaining that no number of lives barely above subsistence level could compensate for a large number N of best lives. So I’ll focus on the conditions enunciated in the generalization of the Puzzle.

Parfit argues against averaging by considering a scenario—‘Only France Survives’—in which he envisages the world population being reduced to the nation with the highest quality of life. I claim that [NAV] can be defended by considering an expanded version of the same scenario. Suppose that the popu­lation is truly enormous, with a large number of people living best lives and the rest (a vastly greater number) leading lives that are best lives except for some trivial discomfiture. Let the people who live best lives inhabit SuperFrance. Just as the actual world is superior to one in which only France survives, so too the original vast‑population world is preferable to one in which only SuperFrance survives. Hence it isn’t plausible to give up [NAV].

Nor can we abandon [REP]. If there are lives sufficiently close to best that they would increase the value of a world containing a large number of best lives, then enough of the mildly suboptimal lives could surely compensate for the best lives. This is to reiterate the moral of the argument in section 4 that the num­bers always count. We can use the considerations urged there, either to moti­vate [BAS] directly or to defend [REP]. So it isn’t possible to evade the argument by giving up [BAS].

Let’s turn now to [DVA]. Here the guiding intuition is that if Q is so low that no number of Q lives could outweigh a large number of best lives, then adding Q lives to a world with a large number of best lives would inevitably diminish the quality of that world. One of the points that motivates averaging is the thought that bringing new lives of very low quality into a world in which a large number of people were leading high quality lives would make the world worse. How low is ‘very low’? [DVA] proposes, plausibly I think, that any lives that are so mean that enough of them couldn’t outweigh a large number of best lives are low enough to fall below the ‘very low’ threshold. Thus [DVA] rep­resents a kind of compromise between aggregation and averaging. Some lives, it suggests, are sufficiently good that an enormous number of them could out­weigh a large number of best lives—this is a nod towards aggregation. Other lives do not have this property because their quality is too poor, and adding lives like this to a world full of best lives would lower the quality of that world—­this is a gesture in the direction of averaging. In my judgment, [DVA] captures what’s plausible about attempts (like the Compromise View) to combine aggre­gation and averaging.

I’ve saved for last [CIV] because I think it’s the most likely to be ques­tioned. As already noted in section 6, [CIV] embodies a limited context­independence of value. Lives that diminish the value of worlds full of best lives must be sufficiently mean that no amount of them could have positive quality.

[CIV] can be defended by the kinds of scenarios urged against averaging at high numbers at the end of section 4. Suppose that some number of Q lives gives a world positive value. Now take a world with a large number of best lives. Suppose we add—in a newly formed galaxy—the appropriate number of Q lives. The resulting world can be no worse. For it is, effectively, two subworlds, the old high‑quality world plus a world that would, by itself, have positive value. Hence, any Q such that there’s some combination of Q lives with positive value is such that adding that number of Q lives to a large number of best lives doesn’t diminish the value of the world. Contraposing, we get [CIV].

This intuitive claim can be buttressed with a more elaborate argument. The general thought that the impact of lives on worlds depends on the other lives in the world—the context‑dependence of value—supposes that there’s some mea­sure of the distribution of the qualities of the other lives and the impact of the new lives depends on their quality and on this measure. More formally, this type of context‑dependence assumes that

 

     V2k(n1,...,nk!1,0,Q1,....,Qk) > V2k(n1,...,nk!1,nk,Ql,...,Qk) when nk > 0 and V2(nk,Qk) > 0 if N(m(n1,...,nk-1,Q1,....,Qk), nk, Qk) < 0.

 

The simplest distributional measure is the mean, and the most obvious pro­posal would take N to subtract the value of the first argument from Qk, so that the impact of the lives would be negative if the value of Qk fell below the mean. Not only does this violate [NAV] but it’s also at odds with our anti‑averaging arguments.

Let’s concentrate, as we have done throughout, on the function V4. We want to know if it’s possible simultaneously to satisfy the conditions:

 

(a) for some n, V4(0,n,100,Q) > 0

(b) for all n, V4(N,n,100,Q) < V4(N,0,100,Q).

 

If we can, then, of course, we can evade [CIV]. We know, of course, that if Q = 100!s, (a) and (b) can’t hold together. But I see no basis for an inductive argument to the effect that (a) and (b) are unsatisfiable for all values of Q > 0. In this instance, there may be critical levels, between which (a) and (b) are jointly satisfiable. Intuitively, as we diminish Q we continue to satisfy (a) even though the lives fall below the level at which they continue to augment a world origi­nally full of best lives.

The analysis so far suggests that a function V4 that avoids [CIV] should take the following form. We want

 

(17) V4(N,n,100,Q) = 100N + nf(100!Q)

where f(0) = 100, df(x)/dx < 0, and for some Q* f(100!Q*) = 0

(18) V4(0,n,100,Q) = nQ

 

expressing the idea that the top quality lives in a world are given full value, while those of lesser quality have their impact adjusted by a function of their distance from the top lives in the world.[30] Intuitively, to circumvent [CIV] and satisfy both (a) and (b), you have to think that a homogeneous world of Q lives works by something like aggregation but that, in a heterogeneous world, where lives belong to the poorer class, the value of the Q lives is discounted with a larger discount the further they are below the higher value.

Notice, however, that the combination of (17) and (18) leads to immediate peculiarities. For consider w20 = <N,M,100,Q*!s> and w21 = <0,M,100,Q*!s>. (17) assigns w20 a value less than 100N; (18) assigns w21 the value M(Q*!s). If M is sufficiently large w21 will be superior to w20, which is absurd, since the latter world contains an extra N best lives.

Perhaps it’s a mistake to believe that (17) and (18) are the right ways to ar­ticulate the idea that the top quality lives in a world serve as the standard against which other lives in that world are measured and by which their impact is as­sessed. I conclude the discussion by claiming the following: [CIV] can be sup­ported by a powerful intuitive argument (one drawn from the anti‑averaging considerations) and an analysis of ways of circumventing [CIV] shows that the obvious strategies engender absurd consequences. [CIV] is hard to reject.

8       A Non‑Archimedean Way Out?

So where did we go wrong? Maybe at the very beginning, when we started to talk about the quality of lives. There I made the Single Number Assumption, adopting a very simple view of the structure of that kind of good we conceive of as ‘quality of life’.[31] Once that assumption has been made, the Archime­dean property of the reals takes over: for any real number however small, r, and for any real number however big, R, there’s an integer N such that Nr > R. With numbers you can get from something tiny to something vast in tiny steps, provided that you take enough of them. Can you do the same with the quality of lives?

Think about the way in which the impossibility proof works. We start with lives of high quality, and then imagine small changes—a cough at a symphony here, mild indigestion there—arguing that such changes conform to particular algebraic relations. It’s supposed that enough small changes could lead us from best lives to lives that are barely above the point at which lives are worth liv­ing. Parfit’s own elaboration of the Repugnant Conclusion, especially in the se­quence of worlds we considered in section 3, embodies just that supposition (which is one reason why I take the proof of section 7 to distill what’s crucial to his puzzle). But is this right?

We could think of things differently. Lives have many dimensions, and one can’t represent the quality of a life by a single real number.[32] Instead, we should think of the quality of a life as a vector. That doesn’t doom attempts to com­pare the qualities of lives.[33] But maybe it will enable us to tame Parfit’s puzzle.

Suppose then that Q(p), the quality of p’s life, is an ordered pair <I,T>, where I reflects the important things (p’s relationships with others, achieve­ments, appreciation of beauty and so forth) and T measures the trivia (the absence of coughs at symphony concerts, free‑flowing traffic, unprotesting digestion etc.). T is a real number; I’ll leave open for the moment what I might be.

Now one might propose that the impossibility theorem really shows that you can’t satisfy my proposed principles along the T dimension. But that’s no prob­lem, because when you recognize that the result is restricted to this dimension, you see immediately what to do: abandon [NRC]. For it’s not Repugnant to con­clude that a large enough number of lives whose quality is <I*,0> could out weigh a world full of lives whose quality is <I*, 100>—provided that I* reflects flourishing with respect to the important things. Suppose that two people live lives that are full of rewarding relationships, genuine achievements, apprecia­tion of natural and artistic beauties, that one of these is encumbered by minor problems and that the other is free. The encumbered life is worse than the un­encumbered life. Yet the important things go right in both, and this allows us to say that a world with a sufficient number of encumbered lives would be better than a world with a smaller number of unencumbered lives.

But this is only half a solution. If I is a real number then we’ve made no genuine gain, for the argument of section 7 can now be run along the first di­mension. So to escape Parfit’s puzzle, one must propose that measurement of the quality of human lives, along the dimension (or dimensions) that really counts (count) doesn’t conform to the structure of the real numbers. Specifically, you can’t descend by small steps from flourishing lives to lives that are barely worth living.

So how does one exhibit the structure of the crucial dimensions? Perhaps discretely. In the simplest case we’d take all the aspects of our lives that really matter to us and suppose that each corresponds to a dimension on which one of two values—1 or !1, say—is assigned. Lives that score 1 on enough of these dimensions go well, even if they receive low (even zero) values on the contin­uous dimensions that reflect the trivia.

Here’s a formal solution to Parfit’s Puzzle.[34] The quality of a life is a pair <I,T> where I 0 {!1,1} and T is a real number between 0 and 100. Let w be a world containing N lives, for n of which the I-value is 1. Then the value of w is given by

 

V(w) = <2n!N,G1N Ti>.

 

The value-ordering of worlds is generated by comparing the first components: only when these are the same does the second come into play. Thus if V(w) = <m,R> and V(w*) = <m*,R*>, then

 

w >V w* iff m > m*, or m = m* and R > R*.

 

The Repugnant Conclusion is now, blocked because no number of lives that suc­ceed in trivial ways can compensate for lives that are successful in the impor­tant respects.

Yet this may just be a formal solution, one that violates well‑motivated views about the goodness of states. Can it survive detailed confrontation with scenar­ios as imaginative as those that Parfit offers us? Could it be developed to allow a larger number of different dimensions? Would it require consequentialists rad­ically to reconceive their ideas about the good (or, more specifically, about the quality of lives)? I don’t know the answers to these questions. I hope, however, to have shown that Parfit’s Puzzle is deep and important, that Parfit’s arguments are best seen as steps towards an impossibility proof, and that that proof may force us to abandon some attractive assumptions about the structure of the good.



[1]I am extremely grateful to Clark Glymour, Patricia Kitcher, and Isaac Levi for their helpful suggestions about earlier versions. Erik Carlson and Gustav Arrhenius kindly shared their un­published work with me, and offered many insightful comments. My principal debt is to Richard Arneson, who has provided me with two sets of detailed comments; those comments and the con­structive advice offered in numerous conversations have significantly affected the discussion at many points.

[2]‘Future Generations’, Part Four of Reasons and Persons (Oxford: Clarendon Press, 1984). Par­enthetical page references throughout are to this book.

[3]Most prominently in social choice theory: see, for example, Amartya Sen, Collective Choice and Social Welfare (San Francisco: Holden Day, 1970), Yew‑Kwan Ng, Welfare Economics (Lon­don: Macmillan, 1979) and Per‑Olov Johansson, An Introduction to Modern Welfare Economics (Cambridge: Cambridge University Press, 1991), all of which provide overviews of the field. The most celebrated result is the impossibility (or possibility) theorem due to Kenneth Arrow, proved in the early 1950s; for Arrow’s own discussion, the best source is Social Choice and Individual Values (New York: Whey, 1963 (second edition)); Sen has provided elegant formulations, not only in Collective Choice and Social Welfare but also in two more recent essays, ‘Social Choice Theory: A Re‑examination’ and ‘On Weights and Measures’, both reprinted in Sen’s Choice, Welfare, and Measurement (Cambridge, MA: Harvard University Press, 1982); these works of Sen contain im­portant extensions and refinements of Arrow’s theorem.

[4]Sen begins ‘Social Choice Theory: A Re‑examination’ by making this point lucidly and di­rectly: ‘It can be argued that some of the problems in the general theory of social choice arise from a desire to fit essentially different classes of group aggregation problems into one uniform frame­work and from seeking excessive generality’ (Choice, Welfare, and Measurement 158; see also 159, 163‑4, 189, 226, 229‑230). John Broome has diagnosed important ambiguities in the conception of utility, both in ‘Utility’ (Economics and Philosophy, 7, 1991) and more briefly in Weighing Goods (Oxford: Blackwell, 1991), 65‑66. The most extensive taxonomy of instances of aggrega­tion problems that I know is offered by Jon Elster in his ‘Local Justice and Interpersonal Compar­isons’ (in Elster and John E. Roemer [eds.], Interpersonal Comparisons of Well‑Being (Cambridge: Cambridge University Press, 1991)).

[5]In similar fashion, one can come to Parfit’s puzzle by thinking about the kinds of reproductive choices that are becoming available in the wake of increased opportunities for pre‑natal genetic testing. Indeed, this was the route that led me to the puzzle—see the last two chapters of The Lives to Come: The Genetic Revolution and Human Possibilities (New York: Simon & Schuster, 1996).

[6]See, for example, Yew‑Kwan Ng, ‘What should we do about future generations?’ Economics and Philosophy, 5, 1989; Theodore Sider, ‘Might Theory X be a Theory of Diminishing Marginal Value?’ Analysis, 51, 1991; Charles Blackorby, Walter Bossert, and David Donaldson, ‘Critical­ Level Utilitarianism and the Population‑Ethics Dilemma’, Economics and Philosophy, 13, 1997; Erik Carlson ‘Mere Addition and Two Trilemmas of Population Ethics’, Economics and Philosophy, 14, 1998; and Gustav Arrhenius ‘An Impossibility Theorem for Welfarist Axiologies’ (typescript).

[7]For lucid development of this and kindred points, in defense of what he calls ‘teleological’ ethics, see Broome, Weighing Goods, Chapter 1. It’s also worth noting that it’s hard to understand the force of principles of rightness, principles assumed to be independent of human well‑being, just as it’s difficult to see how there are notions of justification or epistemic rationality that are independent of the propensity to generate truth. Nonetheless, there are views for which Parfit’s puz­zle might be completely unworrying, for example, that defended by Robert Nozick in Anarchy, State, and Utopia (New York: Basic Books, 1974).

[8]Parfit takes it for granted that there is a zero level, corresponding to the point ‘where life ceases to be worth living’ (388). In a penetrating and insightful discussion of well‑being from the per­spective of destitution, Partha Dasgupta argues that this is a mistake that generates Parfit’s puzzle (see Chapter 13 of An Inquiry into Well‑Being and Destitution (Oxford: Clarendon Press, 1993)). Dasgupta’s argument rests on supposing that Parfit is trying to attribute a degree of well‑being to the state of not being born (ibid. 383), but this is, I think, a misinterpretation. Parfit is concerned with possible future individuals whose lives vary in quality, and he reasonably supposes that, as that quality diminishes continuously, there comes a point at which the life ceases to be worth liv­ing. Thus I don’t think that the puzzle can be solved in the way that Dasgupta suggests.

[9]Sen takes this point to be obvious. See ‘On Weights and Measures’, Choice, Welfare, and Mea­surement, 230.

[10]In effect, this supposes that the transformations corresponding to the admissible functions g form a group. I know of no treatment of representations of utility that violate the supposition.

[11]The argument I’ve offered is intended to be suggestive rather than completely rigorous. It ac­cords with Sen’s judgment that ratio‑scales are particularly pertinent to ‘normative population theory’; see Choice, Welfare, and Measurement, 229.

[12]There’s a long tradition descending from Robbins’ famous article ‘Interpersonal Compari­sons of Utility: A Comment’ (Economic Journal, 48, 1938) that takes such comparisons to be ‘mean­ingless’ or ‘ethical’ (or both!). The famous Arrow results are based on avoiding judgments about comparability, and, as Sen shows, the weaker the assumptions about comparability, the stronger the invariance condition that serves as the basis for constraining admissible social welfare func­tions (see ‘On Weights and Measures’ for a lucid exposition). In effect, the impossibility result I’ll try to reconstruct from Parfit’s reasoning is at the opposite pole from Arrow’s seminal derivation, and, as we’ll see, it proceeds very differently.

[13]Here’s the worry. Suppose Happy’s life receives the value 100. We now turn to Miserable and suppose that his life is assigned k.100 for some k in a small interval [r, s]. This means we can choose to assign Miserable r.100. That choice will be compatible with giving Happy the value K(r.100) for some K in an interval [R, S]. We can keep making choices in this way, and since the numbers assigned to Miserable are small the ratios of the values of Happy’s life to Miserable’s life can fluctuate wildly. Thus, without full comparability, we may not be able to make the judgments which seem required of us in normative population theory.

[14]See, for example, Ignacio Ortuno‑Ortin arid John E. Roemer, ‘Deducing Interpersonal Com­parisons from Local Expertise’ in Elster and Roemer, Interpersonal Comparisons of Well‑Being. The approach adopted by Ortuno‑Ortin and Roemer is especially well‑suited, I think, to the assess­ment of the quality of future lives. Sen’s discussion in ‘Interpersonal Comparisons of Welfare’ (in Choice, Welfare and Measurement) is also a valuable response to skepticism about interpersonal comparison.

[15]Implicit in this is a commitment to Parfit’s Same Number Quality Claim, which denies that the identities of the individuals are important (see 360). Parfit’s defense of this claim seems entirely successful, and his approach is consonant with the invocation of anonymity in welfare economics.

[16]I’ll explore the relationship between these two counterintuitive consequences further in Sec­tion 8. See, in particular, the discussion of [REP].

[17]A second natural way to address Parfit’s puzzle is to suppose that there’s a critical level, above zero, and that the value of worlds is obtained by aggregating the amounts by which lives exceed the critical level. This approach, Critical‑level Utilitarianism, developed and defended by Charles Blackorby, Walter Bossert and David Donaldson, will occupy us later.

[18]Although these are the most obvious possibilities, they aren’t the only ones, as we’ll see later in this section.

[19]N is a constant relative to a particular chosen scale; as we modify the admissible represen­tations of quality of life (change the scale), it’s important to change the value of N to avoid para­doxical results.

[20]As we’ll see in the next section, averaging can prove counterintuitive in some scenarios in­volving large worlds.

[21]The opening wedge of Parfit’s argument against the Compromise View is the story of the Two Hells. There are two homogeneous worlds, Hell One and Hell Two; in both the sufferings of people are intense; in Hell One the population is small (10 as opposed to 10 million); in Hell Two the suffering is slightly less (lasting for fifty years minus one day instead of fifty years). Parfit claims that Hell Two is worse than Hell One. That is, we want V2(10,!S) > V2(107,!S+0). It’s easy to check that (10) implies just this.

Parfit’s subsequent arguments against the Compromise View depend on convoluted maneuvers for ‘solving’ the problem of the Two Hells, so, since (10) doesn’t face the problem, the critique doesn’t get off the ground. One can show, relatively easily, that (10) is innocent of the counterin­tuitive consequences Parfit reports as the Ridiculous Conclusion and the Absurd Conclusion.

[22]This example modifies Parfit’s attempt to defeat the Compromise View. See note 21.

[23]The examples in the next paragraphs owe much to discussions with Richard Arneson. Gustav Arrhenius uses a similar strategy to argue against various views (including Ng’s Theory XN and the Critical‑Level Utilitarianism of Blackorby, Bossert, and Donaldson); see Arrhenius ‘An Impossi­bility Theorem for Welfarist Axiologies’, and also Carlson ‘Mere Addition and Two Trilemmas of Population Ethics’. I should note that Arrhenius (but not Carlson) seems committed to a strong principle that always rejects what he terms the ‘Sadistic Conclusion’ and which is essential to his attempt at impossibility proof. That principle implies that, for any n, m, x > 0 <n,m,100,x> is preferable to <n,1,100,!x>—a result that seems to me quite counterintuitive when m is enor­mous and x is very close to 0. In fact, I think that nobody who finds the Repugnant Conclusion repugnant should be tempted by Arrhenius’ principle, which means that his impossibility proof can’t be convincing.

[24]Parfit himself considers the possibility that various levels are important, but his discussion occurs in the context of an attempt (which, for reasons sketched in note 21, I take to be misguided) to work out a version of the Compromise View that will avoid the problems of the Two Hells, the Ridiculous Conclusion, and the Absurd Conclusion. Again, I think that some of the considerations offered below are in the spirit of his arguments, although, as I interpret him, they are not the same.

[25]See C. Blackorby, W. Bossert, and D. Donaldson ‘Intertemporal Population Ethics: Critical­Level Utilitarian Principles’, Econometrica 65, 1995. 1303‑1320, and ‘Critical‑Level Utilitarian­ism and the Population‑Ethics Dilemma’.

[26]There’s a broad philosophical worry about the idea of the threshold. From the perspective of the individuals whose lives have quality 20!s, their lives are worth living—they are glad to have been born. From the perspective of the universe—the perspective of the Value of the World they inhabit‑their lives have a negative impact (this despite the fact that they might be causally iso­lated from lives above the threshold). Yet we can’t say that the people in question have made a mistake, that they’ve somehow misassessed their lives, for their own perspective is perfectly jus­tifiable. This is surely perplexing. For present purposes, however. I won’t develop this broad phil­osophical worry since, for the reasons given in the text, I think that the idea of ‘critical levels’ is technically flawed. For further objections to the Critical Level approach, see Erik Carlson ‘Mere Addition and Two Trilemmas of Population Ethics’.

[27]After writing this essay, I learned that Blackorby, Bossert, and Donaldson have explored the possibility of integrating the critical‑level approach with the idea of a band of critical values. See, for example, their essay, ‘Quasi‑orderings and Population Ethics’, Social Choice and Welfare, 13, 1996, 129‑160. In my version, worlds that are ranked differently by different choices of the crucial parameter are counted as roughly equivalent. Readers who wish to do so can simply substitute ‘nei­ther better nor worse’ for my claims about ‘rough equivalence’.

[28]The implausibility doesn’t go away if we talk about worlds which are ‘neither better nor worse’: we have three worlds such that each of the first two is neither better nor worse than the third, even though the difference in value between the first and the second is orders of magnitude greater than the value assigned to the third!

[29]This doesn’t affect the generality of the proof. For, if there were no such integer, we could always choose s* < s such that 100 = Ks* and all our principles would hold with s replaced by s*.

[30]The sort of view considered here is close to that proposed by Theodore Sider in ‘Might Theory X be a theory of diminishing marginal value?’.

[31]Broome offers valuable counsel about the dangers of adopting too simple a view of the struc­ture of the good. See Weighing Goods, Chapter 1, and also 124‑126.

[32]This is effectively the view taken by Isaac Levi in his Hard Choices (Cambridge University Press, 1985). Thus the argument of my paper could be seen as showing the need for the type of approach to possible structures of value for which Levi argues in that book‑supplementing his diagnoses of the Ellsberg and Allais paradoxes.

[33]See, for example, the lucid discussion of how to balance various aspects of the quality of life in Thomas Hurka’s Perfectionism (New York: Oxford University Press, 1993) Chapter 7.

[34]I had originally thought that one could assign a neutral value, 0, to lives that don’t go well in the important respects‑that is, take 1 0 {0, 1}. In response to an oral presentation of these ideas, Jason Alexander pointed out that this would lead to counterintuitive consequences. For example, a world in which there are n people whose lives go well in the important respects and badly in the trivial respects would count as inferior to a world in which the same n people lead similar lives and in which there are further people whose lives fare badly in the important respects but well in the trivial respects. The present approach avoids this implausible consequence, and I’m indebted to Al­exander for his insightful response.