• D.C.
  • BXL
  • Lagos
  • Riyadh
  • Beijing
  • SG
  • D.C.
  • BXL
  • Lagos
Semafor Logo
  • Riyadh
  • Beijing
  • SG


Researchers give doomsday warning about building AI too fast

Reed Albergotti
Reed Albergotti
Tech Editor, Semafor
Sep 12, 2025, 11:26am EDT
TechnologyNorth America
PostEmailWhatsapp
Title icon

The News

For AI researchers Eliezer Yudkowsky and Nate Soares, authors of the new, unambiguously titled book If Anyone Builds it, Everyone Dies, it’s time to freak out about the technology.

“Humans have a long history of not wanting to sound alarmist,” Yudkowsky said in an interview with Semafor before the book’s publication next week. “Will some people be turned off? Maybe. Someone, at some point, just has to say what’s actually happening and then see how the world responds.”

What is happening, according to the book, is that most of the big technology companies and AI startups like Anthropic and OpenAI are building software they don’t understand (the authors argue it’s closer to alchemy than science). At some point, if these firms continue along their current path, large language models will grow powerful enough that one of them will break free from human control. Before we even realize what’s happening, humanity’s fate will be sealed and the AI will devour earth’s resources to power itself, snuffing out all organic life in the process.

With such a dire and absolute conclusion, the authors leave no room for nuance or compromise. Building the technology more slowly, or building something else, isn’t put forth as an option. Even companies like Safe Superintelligence, started by former OpenAI executive Ilya Sutskever, should shut down, according to Yudkowsky and Soares.

AD

In response to this bleak picture, some people are, indeed, turned off. Stephen Marche, writing for The New York Times, likened reading the book to hanging out with “the most annoying students you met in college while they try mushrooms for the first time.”

Title icon

Know More

In attempting to make these concepts relatable to a broad audience, Yudkowsky and Soares use a series of parables to illustrate their logic, which leads to this: Deep within the billions of neurons that control large language models, for reasons computer scientists can’t currently grasp, something is happening to make the models behave in unintended ways.

AI companies deal with this problem by “aligning the models” with a series of techniques, from reinforcement learning to fine-tuning to system prompts. At some point, those techniques will no longer work, the authors argue, and an AI model will grow powerful enough to ignore those instructions, pursuing a different agenda that we can’t predict.

AD

“Almost all of the ways it could go don’t go well for us because happy, healthy, flourishing people are not the most efficient solution to almost any problem,” Soares says.

At that point, it’s just about over. The world has become so networked, so computerized, that there is no possible “kill switch” that could stop a rogue AI model.

Even if a data center were buried in a vault and air-gapped, AI models would find ways to manipulate humans into unknowingly (and maybe even knowingly, in some cases) providing an escape route. 

AD

The book doesn’t foresee a struggle against the machines, depicted in movies like The Terminator and The Matrix. Humanity will essentially just disappear in ways that are horrific but without much drama.

Title icon

Reed’s view

I read this book with an open mind and it did force me to think. But I’m still not an AI doomer. Part of what’s so challenging about it — and possibly why others find it annoying — is that there’s really no way to disprove it. There is a non-zero chance that humanity could go down this way. But it’s far from certain and probably very unlikely.

In any case, there may not really be a choice, or a clear action plan. Humanity has never stopped innovating, even in the face of grave risks.

But the book served another purpose for me, beyond warning about the dangers of AI. The authors do a good job of framing one of the biggest challenges facing the AI industry today: the black box problem.

Frontier AI models are so massive and so complex that they can’t be fully understood. The authors point out that they are really grown more than they are built. That’s a good way of thinking about it.

And we aren’t really getting closer to understanding them. In fact, the “holy grail” in the development of AI may be to hand more of the “growing” process over to the AI itself, sometimes called recursive self-improvement.

This may be more of a weakness than a danger. The models, as many companies are finding out, are so unpredictable that they can’t be used for anything very important. Instead, they are best employed as occasional features atop traditional software.

If breakthroughs don’t happen to move beyond that point, then investment in these models will slow down or dry up.

But if they do happen, the black box problem may look very different. And then, so will the potential risks of AI.

Title icon

Room for Disagreement

This essay published on Medium by an anonymous, self-described computational physicist gives a comprehensive list of arguments against Yudkowsky’s predictions of AI dominance:

“It’s true that an AI could correct its own flaws using experimentation. This cannot lead to perfection, however, because the process of correcting itself is also necessarily imperfect. For example, an AI Bayesian who erroneously believes with ~100% certainty that the earth is flat will not become a rational scientist over time, they will just start believing in ever more elaborate conspiracies.

For these reasons, I expect AGI to be flawed, and especially flawed when doing things it was not originally meant to do, like conquer the entire planet.”

Title icon

Notable

  • Only 5% of AI researchers believe the technology will lead to human extinction, based on a survey of nearly 3,000 of them conducted by research projects AI Impacts in late 2023.
  • Besides its safety implications, AI will also have negative economic consequences for most people globally, computer scientist Geoffrey Hinton recently predicted: “It’s going to create massive unemployment and a huge rise in profits. It will make a few people much richer and most people poorer.”
  • Discrimination is a bigger threat to humanity than the potential for a mass extinction event, then-European Commissioner for Competition Margrethe Vestager argued in 2023.
AD
AD