Algorithmic Sabotage Research Group %28asrg%29 <High-Quality »>
Think of the 2010 Flash Crash, where a single sell order triggered algorithmic feedback loops that evaporated $1 trillion in 36 minutes. No code was "wrong." No hacker broke in. The system simply did what it was told, and what it was told was insane.
The ASRG has developed "destabilizer algorithms" that identify fragile equilibria and introduce a single, small, unpredictable actor. In simulation, this has caused simulated drone swarms to retreat from a hill they were ordered to hold, not because they were beaten, but because each drone concluded that the others had gone insane. The ASRG calls this . Case Study: The Great Container Ship Standoff of 2023 To understand the real-world implications, one must examine the ASRG’s most famous—and most controversial—operation. algorithmic sabotage research group %28asrg%29
For example, in a 2020 white paper (published on a mirror of the defunct Sci-Hub domain), the ASRG demonstrated how injecting 0.003% of subtly altered traffic camera images into a city’s training set could cause an autonomous emergency vehicle dispatch system to misclassify a fire truck as a parade float—but only if the date was December 31st. The rest of the year, the system worked perfectly. The sabotage was dormant, invisible, and reversible. Modern AI relies on confidence scores. A self-driving car sees a stop sign with 99.7% certainty. The ASRG’s second pillar exploits the gap between certainty and reality . ROA techniques bombard an algorithm’s sensory periphery with ambiguous, high-entropy signals that are not false—they are simply too real . Think of the 2010 Flash Crash, where a
Dr. Elena Marchetti, a founding member of ASRG (she uses a pseudonym, as all members do), explained the philosophy in a rare 2021 interview with The Baffler : "We cannot stop AI by passing laws. Laws move at the speed of testimony. AI moves at the speed of light. We cannot stop AI by unplugging servers—that is violence and futility. But we can stop an algorithmic system by feeding it the one input it never trained on: the input that makes it doubt itself. That is sabotage. That is the clog in the machine." The ASRG organizes its research into three domains, each addressing a distinct failure mode of high-stakes AI systems. 1. Poison Pill Data Injection (PPDI) Most AI systems are trained on historical data. The ASRG's first pillar asks: What if the future does not look like the past? PPDI involves pre-positioning "sleeper" data points into public datasets that lie dormant until triggered by a specific real-world condition. Case Study: The Great Container Ship Standoff of
The central ethical question is this:
It wasn't a glitch. It wasn't a hacker demanding Bitcoin. According to a leaked post-mortem, it was a live-field test conducted by a little-known entity called the .
The ASRG, acting without approval (as they always do), deployed a low-cost NEE intervention. They rented a small fishing boat, attached a $300 AIS transponder broadcasting a fake identity—"MSC ALGORITHMUS"—and programmed it to loiter at the entrance of the shipping channel moving in a random, zigzag pattern at precisely 4.2 knots.