DoomsdAI
A central figure in the history of artificial intelligence, Eliezer Yudkowsky has been warning everyone for the last 20 years that building systems based on advanced artificial intelligence would end life on Earth. So, he wants to shut it all down

NEW YORK: The first time I met Eliezer Yudkowsky, he said there was a 99.5% chance that AI was going to kill me. I didn’t take it personally. Yudkowsky (46) is the founder of the Machine Intelligence Research Institute, a Berkeley-based non-profit that studies risks from advanced artificial intelligence (AI).
For the last 20 years, he has been Silicon Valley’s version of a doomsday preacher — telling anyone who will listen that building powerful AI systems is a terrible idea, one that will end in disaster.
That is also the message of Yudkowsky’s new book, ‘If Anyone Builds It, Everyone Dies’. The book, co-written with MIRI’s president, Nate Soares, is a distilled, mass-market version of the case they have been making to AI insiders for years. Their goal is to stop the development of AI — and the stakes, they say, are existential.
“If any company or group, anywhere on the planet, builds an artificial superintelligence using anything remotely like current techniques, based on anything remotely like the present understanding of AI, then everyone, everywhere on Earth, will die,” they write.
This kind of blunt doomsaying has gotten Yudkowsky dismissed by some as an extremist or a crank. But he is a central figure in modern AI history, and his influence on the industry is undeniable.
Silicon Valley’s influencer
He was among the first to warn of risks from powerful AI systems, and many AI leaders, including OpenAI’s Sam Altman and Elon Musk, have cited his ideas. Altman has said Yudkowsky was “critical in the decision to start OpenAI”, and suggested that he might deserve a Nobel Peace Prize.
Google, too, owes some of its AI success to Yudkowsky. In 2010, he introduced the founders of DeepMind — a London start-up that was trying to build advanced AI systems — to Peter Thiel, the venture capitalist. Thiel became DeepMind’s first major investor, before Google acquired the company in 2014. Today, DeepMind’s co-founder Demis Hassabis oversees Google’s AI efforts.
In addition to his work on AI safety — a field he more or less invented — Yudkowsky is the intellectual force behind Rationalism, a loosely organised movement (or a religion, depending on whom you ask) that pursues self-improvement through rigorous reasoning. Today, Silicon Valley tech companies are full of young Rationalists, many of whom grew up reading Yudkowsky’s writing online.
I’m not a Rationalist, and my view of AI is considerably more moderate than Yudkowsky’s. I don’t, for instance, think we should bomb data centres if rogue nations threaten to develop superhuman AI in violation of international agreements, a view he has espoused. But in recent months, I’ve sat down with him several times to better understand his views.
At first, he resisted being profiled. Ideas, not personalities, are what he thinks rational people should care about. Eventually, he agreed, in part because he hopes that by sharing his fears about AI, he might persuade others to join the cause of saving humanity. “To have the world turn back from super-intelligent AI, and we get to not die in the immediate future,” he told me. “That’s all I presently want out of life.”
Yudkowsky’s origins
Yudkowsky grew up in an Orthodox Jewish family in Chicago. He dropped out of school after eighth grade because of chronic health issues, and never returned. Instead, he devoured science fiction books, taught himself computer science and started hanging out online with a group of far-out futurists known as the Extropians.
He was enchanted by the idea of the singularity — a hypothetical future point when AI would surpass human intelligence. And he wanted to build an artificial general intelligence, or AGI, an AI system capable of doing everything the human brain can.
“He seemed to think AGI was coming soon,” said Ben Goertzel, an AI researcher who met Yudkowsky as a teenager. “He also seemed to think he was the only person on the planet smart enough to create AGI.”
He moved to the Bay Area in 2005 to pursue what he called ‘friendly AI’ that would be aligned with human values, and would care about human well-being. But the more Yudkowsky learned, the more he came to believe that building friendly AI would be difficult, if not impossible.
One reason is what he calls ‘orthogonality’ — a notion that intelligence and benevolence are separate traits, and that an AI system would not automatically get friendlier as it got smarter.
Another is what he calls ‘instrumental convergence’ — the idea that a powerful, goal-directed AI system could adopt strategies that end up harming humans. A well-known example is the ‘paper clip maximiser’, a thought experiment popularised by philosopher Nick Bostrom, based on what Yudkowsky claims is a misunderstanding of an idea of his, in which an AI is told to maximise paper clip production and destroys humanity to gather more raw materials.
He also worried about what he called an ‘intelligence explosion’ — a sudden, drastic spike in AI capabilities that could lead to the rapid emergence of superintelligence.
At the time, these were abstract, theoretical arguments hashed out among internet futurists. Nothing remotely like today’s AI systems existed, and the idea of a rogue, super-intelligent AI was too far-fetched for serious scientists to worry about.
But over time, as AI capabilities improved, Yudkowsky’s ideas found a wider audience.
In 2010, he started writing ‘Harry Potter and the Methods of Rationality’, a serialised work of Harry Potter fan fiction that he hoped would introduce more people to the core concepts of Rationalism. The book eventually sprawled to more than 6 lakh words — longer than ‘War and Peace’.
Brevity is not Yudkowsky’s strong suit — another of his works, a BDSM-themed Dungeons & Dragons fan fiction that contains his views of decision theory, clocks in at 1.8 million words.
Despite its length, ‘Harry Potter and the Methods of Rationality’ was a cult hit, and introduced legions of young people to Yudkowsky’s worldview. Even today, I routinely meet employees of top AI companies who tell me, somewhat sheepishly, that reading the book inspired their career choice.
Some young Rationalists went to work for MIRI, Yudkowsky’s organisation. Others fanned out across the tech industry, taking jobs at companies like OpenAI and Google. But nothing they did slowed the pace of AI progress, or allayed any of Yudkowsky’s fears about how powerful AI would turn out.
In 2022, Yudkowsky announced — in what some interpreted as an April Fools joke — that he and MIRI were pivoting to a new strategy he called ‘death with dignity’. Humanity was doomed to die, he said, and instead of continuing to fight a losing battle to align AI with human values, he was shifting his focus to helping people accept their fate.
“It’s obvious at this point that humanity isn’t going to solve the alignment problem, or even try very hard, or even go out with much of a fight,” he wrote.
Are we really doomed?
These are, it should be said, extreme views even by the standards of AI pessimists. And during our most recent conversation, I raised some objections to Yudkowsky’s claims. Haven’t researchers made strides in areas like mechanistic interpretability — the field that studies the inner workings of AI models — that may give us better ways of controlling powerful AI systems?
“The course of events over the last 25 years has not been such as to invalidate any of these underlying theories,” he replied. “Imagine going up to a physicist and saying, ‘Have any of the recent discoveries in physics changed your mind about rocks falling off cliffs?’”
What about the more immediate harms that AI poses — such as job loss, and people falling into delusional spirals while talking to chatbots? Shouldn’t he be at least as focused on those as on doomsday scenarios?
Yudkowsky acknowledged that some of these harms were real, but scoffed at the idea that he should focus more on them.
“It’s like saying that Leo Szilard, the person who first conceived of the nuclear chain reactions behind nuclear weapons, ought to have spent all of his time and energy worrying about the current harms of the Radium Girls,” a group of young women who developed radiation poisoning while painting watch dials in factories in the 1920s.
Are there any AI companies he’s rooting for? Any approaches less likely to lead us to doom? “Among the crazed mad scientists driving headlong toward disaster, every last one of which should be shut down, OpenAI’s management is noticeably worse than the pack, and some of Anthropic’s employees are noticeably better than the pack,” he said. “None of this makes a difference, and all of them should be treated the same way by the law.”
And what about the good things that AI can do? Wouldn’t shutting down AI development also mean delaying cures for diseases, AI tutors for students and other benefits? “We totally acknowledge the good effects,” he replied. “Yes, these things could be great tutors, and useful in drug discovery. Is that worth exterminating all life on Earth? No.”
Does he worry about his followers committing acts of violence, going on hunger strikes or carrying out other extreme acts in order to stop AI?
“Humanity seems more likely to perish of doing too little here than too much,” he said. “Our society refusing to have a conversation about a threat to all life on Earth, because somebody else might possibly take similar words and mash them together and then say or do something stupid, would be a foolish way for humanity to go extinct.”
One last battle
Even among his fans, Yudkowsky is a divisive figure. He can be arrogant and abrasive, and some of his followers wish he were a more polished spokesperson. He has also had to adjust to writing for a mainstream audience, rather than for Rationalists who will wade through thousands of dense, jargon-packed pages.
“He wrote 300% of the book,” his co-author, Soares, quipped. “I wrote another negative 200%.”
He has adjusted his image in preparation for his book tour. He shaved his beard down from its former, rabbinical length and replaced his signature golden top hat with a muted newsboy cap. The new hat, he said dryly, was “a result of observer feedback”.
Several months before his new book’s release, a fan suggested he take one of his self-published works, an erotic fantasy novel, off Amazon to avoid alienating potential readers. He did, but grumbled to me that “it’s not actually even all that horny, by my standards”.
In 2023, he started dating Gretta Duleba, a relationship therapist, and moved to Washington state, far from the Bay Area tech bubble. To his friends, he seems happier now, and less inclined to throw in the towel on humanity’s existence.
Even the way he talks about doom has changed. He once confidently predicted, with mathematical precision, how long it would take for superhuman AI to be developed. But he now balks at those exercises.
“What is this obsession with timelines?” he asked. “People used to trade timelines the way that they traded astrological signs. Now, they’re trading probabilities of everybody dying the way they used to trade timelines. If the probability is quite large and you don’t know when it’s going to happen, deal with it. Stop making up these numbers.”
No slowing down
I’m not persuaded by the more extreme parts of Yudkowsky’s arguments. I don’t think AI alignment is a lost cause, and I worry less about ‘Terminator’-style takeovers than about the more mundane ways AI could steer us toward disaster. My p(doom) — a rough, vibes-based measure of how probable I feel AI catastrophe is — hovers somewhere between 5% and 10%, making me a tame moderate by comparison.
I also think there is essentially no chance of stopping AI development through international treaties and nuclear-weapon-level controls on AI chips, as Yudkowsky and Soares argue is necessary. The Trump administration is set on accelerating AI progress, not slowing it down, and ‘doomer’ has become a pejorative in Washington.
Even under a different White House administration, hundreds of millions of people would be using AI products like ChatGPT every day, with no clear signs of impending doom. And absent some obvious catastrophe, AI’s benefits would seem too obvious, and the risks too abstract, to hit the kill switch now.
But I also know that Yudkowsky and Soares have been thinking about AI risks far longer than most, and that there are still many reasons to worry about AI. For starters, AI companies still don’t really understand how large language models work, or how to control their behaviour.
Their brand of ‘doomsaying’ isn’t popular these days. But in a world of mealy-mouthed pablum about ‘maximising the benefits and minimising the risks’ of AI, maybe they deserve some credit for putting their cards on the table.
“If we get an effective international treaty shutting AI down, and the book had something to do with it, I’ll call the book a success,” Yudkowsky told me. “Anything other than that is a sad little consolation prize on the way to death.”
The New York Times