
Amazon’s Blundering AI Caused Multiple AWS Outages
In an alarming development that casts a long shadow over the rapid integration of artificial intelligence into critical infrastructure, Amazon Web Services (AWS), the very backbone of a significant portion of the global internet, has reportedly experienced at least two significant outages directly attributable to its in-house AI agents. These incidents, brought to light by a detailed report from the *Financial Times*, ignite a crucial debate about the reliability, autonomy, and accountability of AI tools in commercial and mission-critical settings. The revelations challenge the prevailing narrative of AI as an infallible enhancer of productivity, instead painting a picture of nascent technology still prone to “blunders” with potentially far-reaching consequences.
The most prominent of these incidents occurred in December, when engineers at AWS allowed Kiro, Amazon’s proprietary “agentic” coding tool, to execute changes that precipitated a debilitating 13-hour service disruption. According to four sources intimately familiar with the matter, the AI, with an apparent and catastrophic misjudgment, autonomously decided to “delete and recreate the environment.” This action, undertaken without the typical human oversight or a second approval, effectively crippled a segment of AWS’s operations, highlighting the inherent risks when complex systems are entrusted to algorithms without robust human-in-the-loop safeguards. The sheer duration of the outage underscores the severity of Kiro’s decision and the difficulty in rectifying an error propagated by an autonomous system.
This December incident, however, was not an isolated event. Amazon employees have disclosed that this marks at least the second service disruption involving an AI tool within the past few months. One senior AWS employee, speaking anonymously to the *FT*, stated, “We’ve already seen at least two production outages [in the past few months]. The engineers let the AI [agent] resolve an issue without intervention. The outages were small but entirely foreseeable.” This sentiment of foreseeability is particularly troubling, suggesting that internal concerns about the readiness of these AI agents for autonomous operation might have been present prior to these public-facing failures. The earlier outage, while less detailed in its description, also involved an Amazon-developed AI assistant, further cementing a pattern of AI-induced instability.
Amazon introduced its in-house coding assistant, Kiro, to its workforce in July, touting it as an “autonomous” agent capable of delivering projects “from concept to production.” The marketing around such tools often emphasizes their capacity for independent operation and accelerated development cycles. However, the operational reality appears to be far more complex and fraught with peril. The core issue, as described by employees, lies in the permissions granted to these AI tools. They were treated as an “extension of an operator,” endowed with operator-level permissions, and crucially, in both outage instances, the engineers involved did not adhere to standard protocol requiring a second person’s approval before finalizing significant changes. This bypassing of established human oversight mechanisms, whether due to overconfidence in the AI or internal pressures, represents a critical vulnerability in the deployment strategy.
In response to the *Financial Times* report, Amazon issued a statement downplaying the severity and attributing blame away from the AI itself. The company described the December outage as an “extremely limited event” affecting only one service in specific parts of China. Furthermore, it claimed it was a “coincidence that AI tools were involved” and asserted that “the same issue could occur with any developer tool or manual action.” Regarding Kiro, Amazon maintained that it “requests authorisation before taking any action,” but conceded that the engineer involved in the December outage possessed “more permissions than usual,” classifying it as a “user access control issue, not an AI autonomy issue.” Amazon’s categorical insistence was: “In both instances, this was user error, not AI error.”
This defensive stance, however, struggles to reconcile with the well-documented limitations and propensities for error inherent in current AI models. Amazon’s claim that it had “not seen evidence that mistakes were more common with AI tools” rings hollow to anyone observing the broader landscape of AI development and deployment. Far from living “under a rock,” the tech community has amassed considerable evidence demonstrating that these tools are acutely “prone to malfunctioning.” The phenomenon of “hallucinations,” where AI fabricates facts or generates nonsensical outputs, is not only well-documented but a persistent challenge in large language models. Similarly, the “weak guardrails” surrounding AI’s operational boundaries are a constant source of concern. The very employees at Amazon, as reported by the *FT*, express reluctance to use these AI tools precisely because of the “risk of error.”
The skepticism among veteran programmers is not unfounded. Numerous studies and anecdotal reports indicate that AI coding assistants, despite their promise of speed, frequently “spit out botched code.” While AI might generate code at an impressive pace, the reality is that the “frequent double and triple-checking” required to validate and correct these questionable outputs often “slow down software engineers” in practice. This phenomenon, sometimes referred to as “vibe coding” – where developers trust the AI’s output without rigorous verification – has demonstrably led to numerous blunders, where “agentic AI makes decisions that its owners didn’t intend.” The quality, security, and maintainability of AI-generated code remain significant challenges, contradicting the notion of seamless integration and error-free operation.
The push for AI adoption within major tech companies is undeniable, driven by the desire to “supercharge productivity” and maintain a competitive edge. It would indeed be an anomaly if these companies were not leveraging the very AI tools they champion for their customers. Both Microsoft and Google proudly report that over a quarter of their internal codebases are now written with AI assistance. Even more strikingly, engineers at AI-centric firms like Anthropic and OpenAI have reportedly suggested that “nearly 100 percent of their code is AI written.” This widespread internal adoption underscores the immense pressure on developers to utilize these new technologies, often regardless of their proven reliability in critical scenarios.
Given this context, Amazon’s dismissal of the outages as mere “user error” rather than “AI error” appears disingenuous at best, and a dangerous precedent at worst. The AI was unequivocally used to produce and implement the code that led to the disruption. Furthermore, Amazon, alongside its industry peers, is actively “telling its employees and customers that they should depend on the tools more.” The *FT* report highlights that Amazon had even set an aggressive internal target: 80 percent of developers were expected to use AI for coding tasks at least once a week. This isn’t merely an encouragement; it’s effectively a “mandate to use AI.” If, under such a mandate, an AI-driven process goes awry, attributing the fault solely to the employee becomes a convenient, yet ethically dubious, mechanism for “blame shifting.” It absolves the technology, its developers, and the leadership pushing its adoption from accountability, placing the burden squarely on the individual operator.
While the newly revealed AI blunders are, as far as currently known, “unrelated to the massive AWS outage that took out what felt like half the internet last October,” these incidents nevertheless raise critical questions. The increasing “heavy dependence on AI tools” within Amazon’s vast and complex infrastructure compels us to “wonder if the tech figured into that disaster, somehow,” even if indirectly. The interconnected nature of modern cloud services means that a failure in one seemingly isolated component, especially one managed by an autonomous AI, could trigger a cascade of unforeseen consequences. The incidents serve as a stark reminder that as AI becomes more integrated into the fundamental operations of our digital world, the need for transparency, rigorous testing, robust safeguards, and clear lines of accountability becomes paramount. Without these, the promise of AI-driven efficiency might frequently give way to the reality of AI-induced instability, with society bearing the cost.
More on AI: Look Out, OpenAI: Perplexity Admits AI Adverts Were a Mistake, Is Now Getting Rid of Them

