Will climate change cause humans to go extinct?

Humanity can still live in a world with climate change – it’s just going to be more work, and many lives and livelihoods are likely to be threatened.

SOURCEThe Conversation

I see a lot of resources talking about near-term human extinction, or the fact that thanks to climate change my generation will see the end of humanity. How likely is an outcome like this? Is there any hope for our futures?

Anonymous, aged 18. London, UK.

This article is part of I Need To Know, a Q&A service for teenagers by The Conversation. Find out how to submit your questions at the end of this article.

The claim that humanity only has just over a decade left due to climate change is based on a misunderstanding. In 2018, a fairly difficult-to-read report by the Intergovernmental Panel on Climate Change (IPCC) warned that humanity needs to cut its carbon dioxide (CO₂) emissions in half by 2030, to avoid global warming of 1.5°C above the levels seen before the industrial revolution.

What this actually means is roughly, “We have about 12 years before fixing climate change becomes really expensive and tough.”

Humanity can still live in a world with climate change – it’s just going to be more work, and many lives and livelihoods are likely to be threatened. But it’s complicated, because this century we are facing many problems at the same time, and we are more dependent on each other than ever.

Under pressure

To get our food, most of us humans depend on global transport, payment and logistics systems. These, in turn, require fuel, electricity, communications and a lot of other things to work properly.

All these systems are connected to each other, so if one starts crashing, the chaos may cause other systems to crash, and before we know it we’ll have massive shortages and conflicts.

It’s hard to calculate the exact risk of this happening, since it has never happened before. Until recently, the world was split into separate regions that were largely independent of each other.

But we do know that climate change puts the whole world under pressure – everywhere, at the same time – making the risk of these systems collapsing more serious.

For example, it’s easier for businesses to handle cybersecurity and energy supply when they don’t also have to cope with natural hazards. Likewise, it’s difficult for governments to maintain infrastructure when politicians are busy dealing with the public’s reactions to food prices, refugees and ecological crises.

Building resilience

Geoengineering to reduce the impact of climate change – for example, by reducing CO₂ levels or pumping reflective particles into the Earth’s atmosphere to deflect the sun’s rays – might work. But if disaster strikes and those operations stop, the effects of climate change can return quickly.

The reasonable thing to do is to work on making our systems more resilient – and there are plenty of opportunities to do this. In practice, this means more local energy production, better backup systems, work on reducing climate change, and being more willing to pay extra for safety.

Disasters and diseases

So what about the other threats humanity is facing? Though natural hazards such as earthquakes, tsunamis, volcanoes and hurricanes can be disastrous, they pose a comparatively small threat to the survival of the human race.

Asteroids? Ideally not. Shutterstock.

Hazards big enough to cause entire species to go extinct are relatively rare. The typical mammalian species survives for about a million years, so the risk is roughly one in a million per year.

Asteroid impacts and supervolcanos do happen, but they are rare enough that we do not have to worry about them. Even so, planning for the day when we need to deflect an asteroid or make do without agriculture for a decade is a smart move.

Pandemics are worse. We know the 1918 flu killed tens of millions of people worldwide. New influenza viruses are popping up all the time, and we should expect to see a big pandemic at least once every 100 years.

Over the past century, we have become better at medicine (which lowers the risk from disease) but we also travel more (which increases the spread of diseases). Natural pandemics are unlikely to wipe out the human race, since there is almost always somebody who is immune. But a bad pandemic might still wreck our global society.

Technology attacks

Bioweapons, which use bacteria, viruses or fungi to harm humans or agriculture, are another issue. Fortunately, they have rarely been used in war, but they might become more dangerous in the near future because advances in biotechnology are making it easier and cheaper to modify organisms and automate lab work.

As this technology becomes more accessible, there’s a growing risk it could be used as a “doomsday device” by nasty regimes, to deter other states from seeking to topple them. Right now, the risk is smallish, but it will surely become larger if we do not figure out better ways to detect pathogens early on, keep an eye on risky biotechnology and do diligent diplomacy to keep governments sane.

Perhaps the biggest risk to humanity right now is nuclear weapons. I would personally guess the risk of a nuclear war (not necessarily world-ending but still horrifying) to be somewhere between one in 100 and one in 1,000 per year. This risk goes up or down, depending on tensions between countries and the competence of the people handling early warning systems.

At the Future of Humanity Institute at the University of Oxford, we do a lot of work on Artificial Intelligence (AI). As with biotechnology, the risk right now is pretty minimal, but it might grow in time as AI become better and smarter, and we think it’s better to be safe than sorry.

Developing tools to ensure AI stays safe and operates in a way that benefits humanity could save money in the long run, and it’s unlikely to make things worse. Again, the probability of an AI disaster is fairly undefined, since it changes depending on how well we prepare for it.

I can’t give a probability of a world-ending disaster that isn’t more or less guesswork. But I do think there’s a big enough risk of such a disaster in our lifetimes that we should work hard to fix the world – whether by making sure governments and AI stay safe and sane, replacing fossil fuels, building backup systems and plans, decentralising key systems and so on. These things are worthwhile, even if the risk is one in a million: the world is precious, and the future we are risking is vast.


If you liked this article, please donate $5 to keep NationofChange online through November.

Previous articleThe human costs of the nationalist agendas of Trump and Brexit
Next articleFunding of human needs must be front and center: Coalition asks presidential candidates to support Pentagon spending cuts
Anders Sandberg’s research at the Future of Humanity Institute centres on societal and ethical issues surrounding human enhancement and new technology, as well as estimating the capabilities and underlying science of future technologies. Topics of particular interest include enhancement of cognition, cognitive biases, technology-enabled collective intelligence, neuroethics and public policy. He has worked on this within the EU project ENHANCE, where he also was responsible for public outreach and online presence. Besides scientific publications in neuroscience, ethics, and future studies, he has also participated in the public debate about human enhancement internationally. Anders also held an AXA Research Fellowship, and is now the senior researcher in the FHI-Amlin collaboration on systemic risk of risk modelling. . Anders has a background in computer science, neuroscience and medical engineering. He obtained his Ph.D in computational neuroscience from Stockholm University, Sweden, for work on neural network modelling of human memory. He has also been the scientific produce for the major neuroscience exhibition “Se Hjärnan!” (“Behold the Brain!”), organized by Swedish Travelling Exhibitions, the Swedish Research Council and the Knowledge Foundation that toured Sweden 2005–2007. He is co-founder and writer for the think tank Eudoxa.