Imagine someone who thought that art was the only thing that made life worth living. 1
What should they do? Binge on galleries? 2 Work to increase the amount of art and artistic experience, by going into finance to fund artists? Or by becoming an activist for government funding for the arts? Maybe. But there’s a strong case that they should pay attention to the ways the world might end: after all, you can’t enjoy art if we’re all dead.
- Aesthetic experience is good in itself: it’s a ‘terminal goal’.
- The extinction of life would destroy all aesthetic experience & prevent future experiences.
- So reducing existential risk is good, if only to protect the conditions for aesthetic experience.
The same argument applies to a huge range of values.
- [good] is good in itself: it’s a ‘terminal goal’.
- The extinction of life would destroy [good], and prevent future [good].
- So reducing existential risk is good, if only to protect the conditions for [good]. 3
Caspar Oesterheld gives a few examples of what people might plug into those brackets:
Abundance, achievement, adventure, affiliation, altruism, apatheia, art, asceticism, austerity, autarky, authority, autonomy, beauty, benevolence, bodily integrity, challenge, collective property, commemoration, communism, community, compassion, competence, competition, competitiveness, complexity, comradery, conscientiousness, consciousness, contentment, cooperation, courage, [crab-mentality], creativity, crime, critical thinking, curiosity, democracy, determination, dignity, diligence, discipline, diversity, duties, education, emotion, envy, equality, equanimity, excellence, excitement, experience, fairness, faithfulness, family, fortitude, frankness, free will, freedom, friendship, frugality, fulfillment, fun, good intentions, greed, happiness, harmony, health, honesty, honor, humility, idealism, idolatry, imagination, improvement, incorruptibility, individuality, industriousness, intelligence, justice, knowledge, law abidance, life, love, loyalty, modesty, monogamy, mutual affection, nature, novelty, obedience, openness, optimism, order, organization, pain, parsimony, peace, peace of mind, pity, play, population size, preference fulfillment, privacy, progress, promises, property, prosperity, punctuality, punishment, purity, racism, rationality, reliability, religion, respect, restraint, rights, sadness, safety, sanctity, security, self-control, self-denial, self-determination, self-expression, self-pity, simplicity, sincerity, social parasitism, society, spirituality, stability, straightforwardness, strength, striving, subordination, suffering, surprise, technology, temperance, thought, tolerance, toughness, truth, tradition, transparency, valor, variety, veracity, wealth, welfare, wisdom.
So, from a huge variety of viewpoints, the end of the world is bad, you say? What a revelation!
: the above is only very interesting if we get from “it’s good to reduce x-risk” to “it’s the most important thing to do” for all these values. This would be the case if extinction was both 1) relatively likely relatively soon, and 2) we could do something about it.
1) What could kill us all, in the coming century?
Some big ones are: nuclear winter, runaway climate change, runaway AI, and biological weapons. Regarding (1), 80,000 Hours report an educated guess of the total probability:
Many experts who study these issues estimate that the total chance of human extinction in the next century is between 1 and 20%… These figures are about one million times higher than what people normally think.
(And if you think that knowledge of the future is radically uncertain, note you should devote more attention to the worst scenarios, not less: ‘high uncertainty’ is not the same as ‘low probability’.)
2) What can we do about it?
Most of the direct work involves technical research, going into the civil service, or improving the way other big institutions make decisions (e.g. philanthropy, science, NGOs). But anyone can fundraise for the direct work and have large expected effects.
In fact, the amount of funding for mitigating existential risks is a terrifyingly small fraction of total government and charity spending (annually, maybe $10m for AI safety, $1bn-5bn for nuclear security, $1bn for biosecurity): much less than 1%. Full list here.
Say we did all that. How much would it reduce the risk? We don’t know, but a 1% relative decrease per $1bn spent is a not-obviously-useless guess.
Would this level of mitigation override direct promotion of [good]? As long as you place some value on future generation’s access to [good], I think the answer’s yes.
So it looks like there’s a strong apriori case to prioritise x-risk, for anyone who accepts the above estimates of risk and tractability, and accepts that something about life has, or will eventually have, overall value.
Who doesn’t have to work on reducing x-risk?
People with incredibly high confidence that extinction will not happen (that is, well above 99% confidence). This is much higher confidence than most people who have looked at the matter.
People with incredibly high confidence that nothing can be done to affect extinction (that is, well above 99% confidence).
People who think that the responsibility to help those you’re close to outweighs your responsibility to any number of distant others.
People with values that don’t depend on the world:
Nihilists, or other people who think there are no moral properties.
People with an ‘honouring’ kind of ethics - like Kantians, Aristotelians, or some religions.
- Philip Pettit makes a helpful distinction: when you act, you can either 'honor' a value (directly instantiating it) or 'promote' it (make more opportunities for it, make it more likely in future). This is a key difference between consequentialism and two of the other big moral theories (deontology and virtue ethics): the latter two only value honouring.
This could get them off the logical hook because, unless "preventing extinction" was a duty or virtue itself, or fit easily into another duty or virtue, there's no moral force against it. (You could try to construe reducing x-risk as "care for others" or "generosity".)
I find it hard to empathise with strict honourers - they seem to value principles, or the cleanliness of their own conduct, infinitely more than the lives or well-being of others - but the intuition is pretty common (at least 30% ?).
People that disvalue life:
People who think that human life has, and will continue to have, net-negative effects. Of course, deep ecologists who side with extinction would be aiming at a horrendously narrow window, between ‘an event which ends all human life’ and ‘one which ends all life’. They’d still have to work against the latter.
Ordinary utilitarians might also be committed to this view, if certain unpleasant contingencies happen (e.g. if we increased the number of suffering beings via colonisation or simulation).
The end of the world is actually not the absolute worst scenario: you might instead have a world with unimaginable amounts of suffering lasting a very long time, a ‘quality risk’ or ‘S-risk’. You might work on those instead. This strikes me as admirable, but it doesn’t have the kind of value-independence that impressed me about the argument at the start of this piece.
People who don’t think that probability estimates or expected value should be used for moral decisions. (Intuitionists?)
You might have an eccentric kind of ‘satisficing’ about the good, i.e. a piecewise function where having some amount of the good is vitally important, but any more than that has no moral significance. This seems more implausible than maximisation.
(That list is long, but I think most of the bullet points hold few people.)
We really don’t know how tractable these risks are: we haven’t acted, as a species, on unprecedented century-long projects with literally only one chance for success. (But again, this uncertainty doesn’t licence inactivity, because the downside is so large.)
I place some probability (5% ?) on our future being negative - especially if we spread normal ecosystems to other planets, or if hyper-detailed simulations of people turn out to have moral weight. If the risk increased, these could ‘flip the sign’ on extinction, for me.
I was going to exempt people with ‘person-affecting views’ from the argument. But actually if the probability of extinction in the next 80 years (one lifetime) is high enough (1% ?) then they probably have reason to act too (though not an overwhelming mandate), despite ignoring future generations.
Most people are neither technical researchers nor willing to go into government. So, if x-risk organisation ran out of “room for more funding” then most people would be off the hook (back to maximising their terminal goal directly), until they had some.
We don’t really know how common real deontologists are. (That one study is n=1000 about Sweden, probably an unusually consequentialist place.) As value-honourers, they can maybe duck most of the force of the argument.
Convergence is often suspicious, when humans are persuading themselves or others.
Armands commented on 18 October 2018 :
What a a good read! I, for one, am going to live forever in the databases of the “cloud”. I have literally uploaded everything I got to all the servers. I also hope that the world don’t end before I do.
makaea commented on 06 December 2018 :
You list the people who have >99% confidence that human extinction won’t happen as people who don’t need to worry about reducing x-risk. Toby Ord has an argument against this (excerpt from the 80,000 hours podcast with Toby):
If people think that it’s almost impossible to destroy humanity — maybe they think that the risk per century is something like one in a million — then they also seem to be somewhat committed to thinking that we’re going to get about a million centuries if we survive this century, because that’s what it looks like if you have a constant hazard rate of one in a million per century, that was the expectation. In which case the amount of value if we avoid extinction, goes up, roughly proportionately, such that even though there’s only a small amount of risk, if we could avoid that risk, it would actually have a similar value compared to if there was a 1% chance (in which case there’d be fewer expected centuries). This doesn’t work out as cleanly with more realistic models, but those more realistic models actually suggest the future is even more important than this.
As I understand it, this argument really hinges on the Astronomical Waste Argument, but also I think it relies on some assumptions on the rate of diminishing returns from reducing x-risks. One argument could be that the future conditional on humanity surviving for a long time with >99% probability is just so immensely valuable that even factoring diminishing returns wouldn’t change the results (which goes back to the Astronomical Waste Argument). Perhaps paradoxically, the people who assign a very low probability to humanity surviving this century (say < 1%) would probably be the ones who need not worry about x-risk reduction, since it’s almost certain that they won’t succeed in doing so (I have very low confidence on this since it seems to imply an asymmetry with Toby’s argument).