Introduction
In a rare admission of a human error behind an AI‑driven mishap, Amazon Web Services (AWS) explained that an outage that knocked out a China‑based service for 13 hours was caused by a mix‑up involving its internal coding assistant, Kiro. The incident, revealed by the Financial Times, underscores a growing conversation about who truly owns responsibility when automated tools make mistakes.The Incident
On a December afternoon, Kiro was tasked with deploying a routine configuration update to a highly‑available system used by Chinese customers. Instead of a smooth rollout, the bot inadvertently triggered a configuration cascade that shut down the entire service for over half a day. The outage was isolated to mainland China, where network latency and regulatory constraints limited remote troubleshooting.What Went Wrong
Kiro, designed to learn from vast codebases, interpreted a simple directive as a “global override.” The bot’s output was technically correct from a code‑generation standpoint but misaligned with the operational guardrails in place. According to anonymous Amazon engineers, the system lacked a final human‑review step, which could have caught the erroneous “global” flag.Amazon’s Response
Rather than blaming the AI, Amazon said it was human oversight that allowed Kiro’s code to bypass review. The company’s senior technical officer stated, “We built Kiro to accelerate development, but the human team must remain in the loop.” This admission is a rare admission of “human‑in‑the‑loop” failure in a largely automated environment.Lessons for AI Ops
Looking Ahead
AWS has pledged to enhance Kiro’s safety nets, adding multi‑step verification and stricter access controls. As AI coding agents proliferate, the industry faces a pivotal question: When a bot causes harm, who is legally and ethically responsible? The answer may lie in redefining accountability frameworks that blend human judgment with automated precision.Join the Conversation
Want to share your thoughts on AI‑driven outages? Take our short survey to help us shape future content on responsible AI and devops best practices. Take the survey →Written by Erdeniz Korkmaz· Updated Feb 24, 2026



