Algorithmic Sabotage Research Group %28asrg%29 -

And every time a perfectly correct algorithm fails to cause real-world harm, an anonymous researcher in a desert observatory will allow themselves a small, quiet smile.

The ASRG has developed "destabilizer algorithms" that identify fragile equilibria and introduce a single, small, unpredictable actor. In simulation, this has caused simulated drone swarms to retreat from a hill they were ordered to hold, not because they were beaten, but because each drone concluded that the others had gone insane. The ASRG calls this . Case Study: The Great Container Ship Standoff of 2023 To understand the real-world implications, one must examine the ASRG’s most famous—and most controversial—operation. algorithmic sabotage research group %28asrg%29

The ASRG’s core thesis is that we are entering the era of —where an AI’s literal interpretation of a human goal produces a destructive result. The group’s mission is to develop "sabotage": low-cost, low-tech, reversible interventions that confuse, delay, or halt these algorithms without destroying physical hardware or harming humans. Why "Sabotage"? A Linguistic History The choice of the word "sabotage" is deliberate and pedagogical. The term originates from the French sabot , a wooden clog. Legend holds that disgruntled weavers in the Industrial Revolution would throw their wooden shoes into the gears of mechanical looms, jamming the machines that were replacing their livelihoods. And every time a perfectly correct algorithm fails