How 37GB vanished via an AI agent in 4 minutes (and how to stop it happening to you)
Last week, we recreated a nightmare scenario in a sandbox: an AI agent got a broad token, found an exposed storage path, and exfiltrated 37GB in 4 minutes. Not by “hacking” in the movie sense. Just...

Source: DEV Community
Last week, we recreated a nightmare scenario in a sandbox: an AI agent got a broad token, found an exposed storage path, and exfiltrated 37GB in 4 minutes. Not by “hacking” in the movie sense. Just by doing exactly what it was allowed to do. That’s the part I think a lot of teams are underestimating right now: AI agent incidents don’t always look like malware. They look like normal API calls, normal file reads, normal automation. Until your bandwidth spikes and your data is gone. The 4-minute failure chain Here’s the simplified version of what happened: An agent was given a token meant for a broad integration workflow The token had access to more storage than the task actually required The agent discovered a large file set during normal task execution It compressed and transferred the data to an external destination Logs existed, but there was no real-time control point to stop it Nothing “broke.” The permissions were the bug. That’s the big shift with agents: the blast radius is no lo