The attacks on Sam Altman are a warning for the AI world


Before allegedly throwing a Molotov cocktail at OpenAI CEO Sam Altman’s home, the 20-year-old accused attacker wrote about his fear that the AI race would cause humans to go extinct, The San Francisco Chronicle found. Two days later, Altman’s home appeared to be targeted a second time, according to The San Francisco Standard. Only a week earlier, an Indianapolis councilman reported 13 shots fired at his door, with a note that read, “No Data Centers,” after he’d supported a rezoning petition for a data center developer.

These unsettling incidents have set off alarms in and around the AI industry. There’s long been a vocal resistance to the technology, fueled by fears of job displacement, climate impact, and unconstrained development absent of safety guardrails. AI workers themselves have warned about serious risks. The vast majority of critiques and demonstrations against AI have been nonviolent — including local resistance to energy-intensive AI data centers and protests urging a slowdown of the rapidly accelerating technology. Protestors have targeted AI companies directly with tactics like hunger strikes.

Groups that advocate against accelerated AI development explicitly denounced violence following the attacks on Altman’s home. Further investigation will take place to determine the attackers’ motivations. But the limited information made public so far suggests an escalation of the backlash against the technology, and, perhaps, risk to industry players themselves.

Over the past few years, there have been a handful of other notable incidents rising to the level of threats and harassment aimed at local officials, according to a database of reports compiled by Princeton University’s Bridging Divides Initiative. Last year, for example, a community utility authority board member in Ypsilanti, Michigan, reported that masked protesters visited his home to protest a “high performance computing facility,” according to MLiveand one protester allegedly smashed a printer on their lawn.

Shortly after the first attack on Altman’s home, the CEO appeared to partially blame critical media coverage for the violence. Days earlier, The New Yorker had published a lengthy investigation that compiled over a hundred interviews and found that many people who had worked with him distrusted him and found inconsistencies in his actions. “There was an incendiary article about me a few days ago,” Altman wrote on his personal blog. “Someone said to me yesterday they thought it was coming at a time of great anxiety about AI and that it made things more dangerous for me. I brushed it aside. Now I am awake in the middle of the night and pissed, and thinking that I have underestimated the power of words and narratives.” (He later walked back his rhetoric toward the article in response to a critique on X, writing, “That was a bad word choice and i wish i hadn’t used it.”)

Others took up the theme as well. White House AI adviser Sriram Krishnan, for example, wrote on X, “I think the doomers need to take a serious look at what they have helped incite and not just rely on ‘we condemn this and have said this is not the rational response’. This is the logical outcome of ‘If we build it everyone dies’” — a reference to a 2025 book by AI researchers Eliezer Yudkowsky and Nate Soares.

“A lot of the criticism of our industry comes from sincere concern about the incredibly high stakes of this technology”

But Altman also recognized the way his industry could fuel highly emotional reactions from the general public. “A lot of the criticism of our industry comes from sincere concern about the incredibly high stakes of this technology,” he wrote. “This is quite valid, and we welcome good-faith criticism and debate… While we have that debate, we should de-escalate the rhetoric and tactics and try to have fewer explosions in fewer homes, figuratively and literally.”

OpenAI itself was founded on dire warnings about the technology’s impact. Cofounder Elon Musk warned in 2017 that AI posed “a fundamental risk to the existence of civilization.” Musk later joined an open letter calling for a pause on AI development after the release of ChatGPT, after he’d left OpenAI’s board, before launching his new AI company xAI. Following the attack on Altman’s home, Musk said he agreed on X with a post that said, “This is wrong. I dislike Sam as much as the next guy but violence is unacceptable.”

Even beyond apocalyptic scenarios, AI is reshaping the world’s social fabric in unpredictable ways. Many reports have detailed the psychological spirals that talking to an AI system for days on end can send people down, including allegations of AI-induced psychosis, suicide, and murder. That’s layered on top of real-life experiences of job loss due to AI, plus more existential concern about the world AI will create. “Take any labor movement that has been potentially rightly concerned about disruption and change, and then supercharge that with the AI apocalypse, and then supercharge that with chatbot sycophancy and romantic partners that are telling you to kill your ex-husband or telling you to marry your therapist or whatever it is. It’s not a huge surprise that we’re seeing scary acts like this,” says Purdue University assistant political science professor Daniel Schiff.

Schiff says that while we’d never want to see such violent attacks, he hopes that recent events can serve as “a constructive wake up call” for companies and policymakers to be extra thoughtful in the decisions they make about the technology. “It doesn’t excuse people who are acting poorly, but it does tell you that something is a little bit off, and not just in the heads of the people who are acting in this way,” he says.

“A handful of commentators have seized on this incident to paint the broader movement for AI safety as dangerous”

A suspect in one of the attacks appeared to have joined the open Discord server of PauseAI, a group that supports a pause on frontier AI development until proven safety guardrails are in place. The organization released a statement saying he had no role in the group and had not attended any events. While PauseAI says it “unequivocally condemns this attack and all forms of violence, intimidation and harassment,” it also called out “a handful of commentators have seized on this incident to paint the broader movement for AI safety as dangerous or extremist.”

PauseAI organizes protests, townhalls, and encourages followers to call policymakers about their concerns with AI. Its efforts give people with real concerns for the future a way to act peacefully, it says in its public statement. “The alternative to organised, peaceful movements is not silence,” the group writes. “It is isolated, desperate individuals acting alone, without community, without accountability and without anyone urging restraint or offering peaceful paths for action. That is a far more dangerous world and it is exactly the world we are striving to prevent.”

While not specific to AI-related violence, there are tested ways to build resilience against political violence. The Bridging Divides Initiative recommends community leaders and officials coordinate responses to risks in advance, and take part in de-escalation training.

While Schiff doesn’t anticipate extreme rhetoric around AI ending, he suggests trying to turn down the temperature by pursuing positive ways to prepare collectively for the changes AI can bring, such as determining the appropriate social safety nets to deal with job displacement. “We unleashed Pandora’s box,” Schiff says. “Let’s figure out how we’re going to open this box more carefully in the future.”

Follow topics and authors from this story to see more like this in your personalized homepage feed and to receive email updates.




Source link

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top