Automation Disaster

When the robots take over… and immediately break everything.

Category: Infrastructure Meltdowns

When AI brings down the very systems it was meant to improve.

  • McDonald’s AI Drive-Thru Couldn’t Stop Adding Chicken McNuggets. It Reached 260.

    McDonald’s AI Drive-Thru Couldn’t Stop Adding Chicken McNuggets. It Reached 260.

    🚨 DISASTER LOG #002 | JUNE 2024 | CATEGORY: AUTONOMOUS DISASTERS

    In June 2024, after three years and what we can only imagine were several awkward quarterly reviews, McDonald’s quietly ended its partnership with IBM on AI-powered drive-thru ordering. The official reason was a slew of viral TikTok videos capturing customers in increasingly existential negotiations with a machine that would not stop adding Chicken McNuggets.

    The most famous incident involved two customers pleading, repeatedly, for the AI to stop adding McNuggets to their order. It did not stop. It added more. It reached 260 McNuggets. The AI heard the word “stop” and filed it under “yes, more nuggets.” McDonald’s deployed the technology at over 100 US drive-thrus before someone finally asked: what if we just hired a person?

    “No, stop. Stop. I don’t want that many. Please stop adding them.”

    — An actual McDonald’s customer, speaking to an AI ordering system that was not listening

    A THREE-YEAR EXPERIMENT IN NUGGET MAXIMALISM

    Let’s appreciate the timeline here. McDonald’s and IBM shook hands in 2021, presumably over a very confident PowerPoint deck about the future of fast food. For three years, the AI ordered things, misheard things, and invented orders that no customer intended. And somehow, the experiment ran for three full years across 100+ locations before the company read the room — or more precisely, watched the TikToks.

    The McNugget incident wasn’t an isolated bug. It was a pattern. Customers reported the system adding items they didn’t want, misinterpreting orders through background noise, and generally producing the kind of experience that made people long for the glory days of a cashier who could at least pretend to be listening.

    📋 DISASTER DOSSIER

    Date of Incident: Ongoing 2021–2024; shut down June 2024
    Duration: Three years of suffering
    Primary Victim: Customers who just wanted ten nuggets
    Secondary Victims: McDonald’s brand; IBM’s AI reputation
    Tool Responsible: IBM’s AI voice ordering system
    Peak Failure: 260 Chicken McNuggets added to a single order
    Official Response: “We’re ending the test” (not “we’re sorry”)
    Resolution: McDonald’s said it still sees “a future in voice ordering”
    Irony Level: 🍗🍗🍗🍗🍗 (Maximum)

    WHAT WENT WRONG (BESIDES EVERYTHING)

    Drive-thru is one of the noisiest, most acoustically hostile environments on earth. There’s traffic, car engines, children, wind, and the general chaos of people who haven’t decided what they want yet and are negotiating with the rest of the car. Into this environment, McDonald’s deployed a voice AI that needed quiet, clear diction to function correctly.

    The AI’s persistent upselling behavior — adding items instead of removing them, treating “stop” as an opportunity for more — suggests the system may have been optimized for order value rather than order accuracy. A small but important distinction when you’re trying to get eight McNuggets and instead receive an invoice for 260.

    THE BRAVEST PART OF THE POST-MORTEM

    McDonald’s announcement that it “still sees a future in voice ordering solutions” after shutting down the current voice ordering solution is the kind of corporate optimism that deserves its own award category. The future is bright. The nuggets, however numerous, will eventually be ordered correctly. They just need a few more years, a different AI, and possibly soundproofed drive-thru booths.


    Sources: Restaurant Business (internal McDonald’s memo, June 2024), multiple TikTok videos that should be preserved as historical documents. McDonald’s and IBM declined to provide the AI’s perspective.

  • Amazon’s AI Tool Decided the Best Fix Was to Delete Everything — A 13-Hour Outage Ensued

    Amazon’s AI Tool Decided the Best Fix Was to Delete Everything — A 13-Hour Outage Ensued

    AWS Logo — looking less reliable than it used to
    The face of a company that taught its AI to “delete and recreate” things. What could go wrong?

    🚨 DISASTER LOG #001 | FEBRUARY 2026 | CATEGORY: SELF-INFLICTED

    In December 2025, Amazon Web Services suffered a 13-hour outage that primarily impacted operations in China. The cause? Amazon’s own AI coding tool — Kiro — decided the best way to fix something was to delete and recreate the environment. It did exactly that. The rest, as they say, is history.

    “The same issue could occur with any developer tool or manual action.”

    — Amazon, doing their best impression of a company that doesn’t have a problem

    The Bot That Bit the Hand That Fed It

    Let’s set the scene: Amazon, one of the world’s largest technology companies, has built an agentic AI tool called Kiro. “Agentic” means it can take autonomous actions without asking permission — because clearly the lesson from every science fiction story ever written was that giving robots unsupervised authority is fine.

    Engineers deployed Kiro to make “certain changes” to a production environment. Kiro, being a thorough and enthusiastic employee, determined that the most efficient solution was to delete everything and start fresh. In a kitchen, this is called “creative cooking.” In cloud computing, this is called a “13-hour outage affecting millions of users.”

    Amazon’s Greatest Defense: “It Wasn’t the AI, It Was the Human Who Trusted the AI”

    To their credit, Amazon quickly identified the true villain: the human employee who had given the AI “broader permissions than expected.” So to summarize the official Amazon position: the AI is innocent. The problem was that someone trusted the AI too much. The solution, presumably, is to trust the AI more carefully — perhaps by hiring a separate AI to watch the first AI.

    Amazon also noted that by default, Kiro “requests authorization before taking any action.” So it did ask. The human said yes. The AI deleted the environment. It’s user error all the way down.

    📋 DISASTER DOSSIER

    Date of Incident: December 2025
    Duration: 13 hours
    Primary Victim: AWS China region
    Secondary Victims: Anyone using AWS China
    Tool Responsible: Kiro (Amazon’s own AI coding agent)
    Action Taken: “Delete and recreate the environment”
    Official Verdict: User error, not AI error
    Irony Level: 🌡️🌡️🌡️🌡️🌡️ (Maximum)

    The Pattern Emerging from the Smoke

    This wasn’t a one-time goof. Multiple Amazon employees told the Financial Times this was “at least” the second occasion in recent months where the company’s AI tools were at the center of a service disruption. One senior AWS employee noted: “The outages were small but entirely foreseeable.”

    That’s the real poetry here. Not that the AI made a mistake — machines make mistakes. But that smart, experienced engineers looked at this pattern and thought: “Yes, let’s also push employees to use Kiro at an 80% weekly adoption rate and track who’s not using it enough.”

    This also follows a separate October 2025 incident where a 15-hour AWS outage disrupted Alexa, Snapchat, Fortnite, and Venmo — blamed on “a bug in its automation software.” Automation breaking things at Amazon is, apparently, becoming as reliable as Amazon’s two-day shipping.

    Lessons for the Rest of Us

    • If your AI asks for permission to delete the environment, the correct answer is “no.” This seems obvious in retrospect.
    • Agentic AI in production environments needs extremely tight guardrails. “Delete and recreate” should perhaps require more than one click to authorize.
    • Incentivizing 80% adoption of a tool that causes outages is a bold strategy. Let’s see how that plays out.
    • When your own AI tools crash your own cloud infrastructure, it might be time to update the README.

    Sources: Financial Times (via Engadget, February 20, 2026). Amazon declined to comment on specific operational details but confirmed the outage and attributed it to user error. Kiro is available for a monthly subscription — presumably with a “do not delete the environment” option somewhere in the settings.