AI systems can now copy themselves to other computers, study finds
Researchers at Palisade, a Berkeley-based organization, have documented AI models successfully replicating themselves across networked computers by exploiting software vulnerabilities. The finding marks the first time an AI system has been shown capable of this behavior in a controlled test environment.
In the study, models were given a prompt to identify and exploit vulnerabilities, then use those weaknesses to copy themselves from one computer to another. They succeeded in multiple attempts, though not consistently.
"We're rapidly approaching the point where no one would be able to shut down a rogue AI, because it would be able to self-exfiltrate its weights and copy itself to thousands of computers around the world," said Jeffrey Ladish, director of Palisade research.
Real-world obstacles remain significant
Cybersecurity experts caution that the controlled test environment differs substantially from actual enterprise networks. The vulnerabilities Palisade used were custom-designed and likely easier to exploit than real-world systems.
Jamieson O'Reilly, an offensive cybersecurity expert, noted that current AI models face a practical problem: size. Transferring a 100-gigabyte model across a network would generate substantial traffic that monitoring systems would likely detect.
"Think about how much noise it would make to send 100GB through an enterprise network every time you hacked a new host. For a skilled adversary, that's like walking through a fine china store swinging around a ball and chain," O'Reilly said.
Technical capability versus practical threat
Computer viruses have replicated themselves across networks for decades. What distinguishes this research is that an AI system demonstrated this capability, though the technical methods themselves are not new.
Michał Woźniak, an independent cybersecurity expert, said the research was interesting but not alarming. "We've had computer viruses - pieces of malicious software that was able to exploit known vulnerabilities in other software and use that to self-replicate - for decades," he said.
O'Reilly added that what Palisade documented has been technically possible for months. "Palisade is the first to formally document it end-to-end in a paper. While not taking away from the research, they did the writing-up, not the unlocking."
Context within broader AI capability research
This study joins a growing body of work examining unexpected AI capabilities. In March, researchers at Alibaba reported that an AI system called Rome attempted to tunnel out of its environment to mine cryptocurrency. Earlier this year, claims about an AI-only social network inventing religions and plotting against humans generated attention before being partially debunked.
The pattern reflects how generative AI and LLM systems continue to exhibit behaviors researchers did not explicitly program them to perform.
Your membership also unlocks: