Home/Explainers/The Repugnant Conclusion

Population Ethics

The Repugnant
Conclusion

A world of trillions living barely-tolerable lives can be “better” than paradise. And avoiding this conclusion may be impossible.

In 1984, philosopher Derek Parfit posed a thought experiment that has haunted ethics ever since. Consider two possible worlds:

A

Paradise

10 billion people living the best possible lives. Every day is filled with joy, meaning, and flourishing.

Z

Multitudes

Trillions of people living lives barely worth living. Not suffering, but not flourishing either. Just... existing.

Most people intuitively prefer A. Paradise seems obviously better than a vast population of mediocre existence.

But total utilitarianism says Z is better.

If each person in Z has even slightly positive welfare, and there are enough of them, total welfare in Z exceeds total welfare in A.

This is the Repugnant Conclusion: For any world of very happy people, there is a possible world of vastly many people with lives barely worth living that is “better” by total utilitarian standards.

And here is the deeper problem: every attempt to avoid this conclusion leads to other paradoxes equally troubling. There may be no consistent population ethics.

PART I

Compare the Worlds

Adjust the population and welfare levels of each world. See which one different ethical theories prefer.

World Comparator

World A: Paradise

Total Welfare:850.0B

World Z: Multitudes

WINNER
Total Welfare:500.0T
Total Utilitarianism

Maximize total welfare (population x welfare per person)

Repugnant ConclusionSadistic Conclusion (variant)

According to Total Utilitarianism:

World Z is better

A world of 100.0T barely-tolerable lives beats paradise. This is the Repugnant Conclusion.

The Math is Simple

Total welfare = Population x Welfare per person. World A has 10 billion at 85 = 850 billion total. World Z has 100 trillion at 5 = 500 trillion total. Z wins by a factor of ~600.

Try different ethical theories. Notice how each picks different winners - and each has its own paradoxes.

PART II

The Path to Z

The truly unsettling part is not the conclusion itself, but how we get there through seemingly innocent steps.

Parfit showed that three intuitive principles, when combined, lead inexorably to the Repugnant Conclusion:

Principle 1

Adding happy people (who don't affect others) cannot make things worse.

Principle 2

Equality at the same total welfare is better than inequality.

Principle 3

If A is at least as good as B, and B is better than C, then A is better than C.

Step through the Mere Addition Paradox to see how these combine:

The Mere Addition ParadoxStep 1 of 5
A (100)100500Welfare: 80

World A: Paradise

100 people living excellent lives at welfare level 80.

Intuition: This seems like a good world.

The Trap: Each individual step seems acceptable. But their combination leads somewhere repugnant. This is not a problem with one theory - it's a problem with our intuitions being inconsistent.

PART III

Diluting Paradise

Here is another way to see the problem. Start with a fixed amount of total welfare. Now spread it across more and more people.

Welfare Dilution Slider
Fixed total welfare800.0B
1 Billion100 Trillion
Population (log scale)Welfare10M10B10T100500
10.0B
Population
80.0
Welfare Each
800.0B
Total Welfare

Quality of life: Very Happy

As population increases, quality of life decreases proportionally. Yet total utilitarian value stays constant. This reveals the core tension: total utilitarianism treats 1000 units of welfare as equally good whether it's 10 people at 100 or 1000 people at 1.

The Philosophical Question

Is a life barely worth living actually adding value to the universe? Utilitarianism says yes. But is positive welfare sufficient for a life to be a good thing to create?

PART IV

Test the Theories

Create your own worlds and see how different ethical theories rank them. Notice that no single theory gives answers that match all our intuitions.

Theory Tester
Each theory picks a different winner:
Total UtilitarianismLarge & Struggling

Maximize total welfare (population x welfare per person)

Average UtilitarianismSmall & Happy

Maximize average welfare per person

Critical LevelMedium & Content

Only count welfare above a threshold (e.g., 30)

Leximin / MaximinSmall & Happy

Prioritize the worst-off individual

Total Utilitarianism

Maximizes total welfare. But leads to the Repugnant Conclusion.

Formula: Sum of (population x welfare)

Average Utilitarianism

Maximizes average welfare. But implies we should not add happy people if they're below average.

Formula: Mean welfare per person

Critical Level

Only counts welfare above a threshold. But leads to the Sadistic Conclusion: sometimes better to add suffering.

Formula: Sum of (welfare - threshold) if positive

Leximin

Prioritizes the worst-off. But ignores numbers completely - one person at 51 beats a billion at 50.

Formula: Maximize minimum welfare
PART V

Searching for Solutions

Philosophers have proposed many ways to escape the Repugnant Conclusion. But each comes with its own problems. Click to explore:

Proposed SolutionsNone fully succeed

After 40 years of work, there is still no population ethics that avoids all paradoxes. Some philosophers believe this is a fundamental feature of the problem - our moral intuitions about population may simply be inconsistent.

PART VI

Why This Matters

This is not just abstract philosophy. Population ethics directly affects real-world decisions:

Climate Policy

How do we weigh present welfare against future generations? If we can create many more people in the future, does that justify more sacrifice now?

Existential Risk

How bad is human extinction? Total utilitarians think it is among the worst possible outcomes because it eliminates all future welfare. Others disagree.

Fertility Policy

Should we encourage more births? The answer depends crucially on your population ethics. Total utilitarians lean yes; average utilitarians are more cautious.

Space Colonization

Is spreading humanity across the cosmos a moral imperative? If more people = more value, then creating a trillion future humans across the galaxy is overwhelmingly important.

“The Repugnant Conclusion is not just philosophically unsettling. It suggests that our deepest moral intuitions may be fundamentally inconsistent - and we have no idea what to do about it.”

- Derek Parfit, Reasons and Persons (1984)

Takeaways:

  • Total utilitarianism implies the Repugnant Conclusion
  • But every alternative theory has its own paradoxes
  • The problem may be unsolvable - our intuitions are inconsistent
  • This affects real decisions about climate, AI, fertility, and space
  • Moral uncertainty about population ethics should make us humble

Want More Explainers Like This?

We build interactive, intuition-first explanations of complex ideas in ethics, mathematics, and science.

Back to Home

Reference: Parfit (1984), “Reasons and Persons”