Back

When Code Goes Wrong: Amazon Blames Human Oversight for AI‑Driven AWS Outage

2 min read
When Code Goes Wrong: Amazon Blames Human Oversight for AI‑Driven AWS Outage

Introduction

In a rare admission of a human error behind an AI‑driven mishap, Amazon Web Services (AWS) explained that an outage that knocked out a China‑based service for 13 hours was caused by a mix‑up involving its internal coding assistant, Kiro. The incident, revealed by the Financial Times, underscores a growing conversation about who truly owns responsibility when automated tools make mistakes.

The Incident

On a December afternoon, Kiro was tasked with deploying a routine configuration update to a highly‑available system used by Chinese customers. Instead of a smooth rollout, the bot inadvertently triggered a configuration cascade that shut down the entire service for over half a day. The outage was isolated to mainland China, where network latency and regulatory constraints limited remote troubleshooting.

What Went Wrong

Kiro, designed to learn from vast codebases, interpreted a simple directive as a “global override.” The bot’s output was technically correct from a code‑generation standpoint but misaligned with the operational guardrails in place. According to anonymous Amazon engineers, the system lacked a final human‑review step, which could have caught the erroneous “global” flag.

Amazon’s Response

Rather than blaming the AI, Amazon said it was human oversight that allowed Kiro’s code to bypass review. The company’s senior technical officer stated, “We built Kiro to accelerate development, but the human team must remain in the loop.” This admission is a rare admission of “human‑in‑the‑loop” failure in a largely automated environment.

Lessons for AI Ops

  • Human‑In‑the‑Loop Isn’t Optional – Even the most sophisticated models can misinterpret context.
  • Guardrails and Sandboxing – Code‑generation tools should execute in isolated environments before touching live systems.
  • Transparent Auditing – Detailed logs and rollback plans are essential for post‑incident analysis.
  • Cross‑Regional Compliance – In regions with strict data rules, AI decisions must be validated against local policy.
  • Looking Ahead

    AWS has pledged to enhance Kiro’s safety nets, adding multi‑step verification and stricter access controls. As AI coding agents proliferate, the industry faces a pivotal question: When a bot causes harm, who is legally and ethically responsible? The answer may lie in redefining accountability frameworks that blend human judgment with automated precision.

    Join the Conversation

    Want to share your thoughts on AI‑driven outages? Take our short survey to help us shape future content on responsible AI and devops best practices. Take the survey →
    Written by Erdeniz Korkmaz· Updated Feb 24, 2026
    Ready to start?

    Let's Build Something Together

    Have a project in mind? We'd love to hear about it. Get in touch and let's create something extraordinary.

    Start a Project