Back to posts

AWS Mass Outage Explained - A Deep Dive into the DNS Resolution Failure That Disrupted the Internet

Nitin Ahirwal / October 22, 2025

AWS OutageCloud InfrastructureDynamoDBDNSDevOpsSystem Reliability

The AWS Mass Outage: What Happened and Why It Matters

On social media, you might have seen #AWSDown trending everywhere. Countless users and engineers reported that AWS services were unavailable, servers were crashing, and databases were unreachable. News outlets confirmed that major companies like Fortnite, Snapchat, Alexa, and Amazon.com were hit hard. Even Amazon’s internal subsystems experienced downtime.

So what exactly happened? Let’s unpack it step by step.


📍 Where Did the Outage Occur?

The issue originated in Northern Virginia (US-East-1) — AWS’s oldest, largest, and most feature-rich region. Because this region powers critical services globally, disruptions here ripple across the internet.

US-East-1 is often referred to as the “flagship” AWS region, hosting services and features not available elsewhere. Its scale and centrality make it a critical single point of failure.


🕒 The Timeline (IST)

  • 12:19 PM → AWS engineers noticed increased error rates in the region. Customers began experiencing timeouts and failed requests.
  • 12:56 PM → The root cause was identified as DNS resolution issues affecting DynamoDB, AWS’s managed NoSQL database.
  • 2:54 PM → A fix was rolled out. Recovery began, but at this stage, not all services were fully functional.
  • 12:15 AM → Customers reported visible recovery, with most services gradually resuming normal operations.
  • 3:31 AM → AWS confirmed full restoration of all services.

This meant that some of the internet’s most used services were unstable for nearly 15 hours.


🛠️ What Went Wrong?

The incident was not caused by DynamoDB itself but by failures in the DNS resolution process required to connect to it.

How Applications Connect to DynamoDB

  1. Applications use domain names (like dynamodb.us-east-1.amazonaws.com) to reach DynamoDB endpoints.
  2. DNS translates this domain name into an IP address.
  3. Once the IP is known, the application can connect and perform database operations.

During the outage:

  • DNS lookups failed.
  • Applications couldn’t retrieve the IPs for DynamoDB.
  • As a result, services timed out or crashed.

👉 DynamoDB itself remained healthy. No data was lost. The failure was purely in the ability of customers to connect to it.


🌐 A Deep Dive into DNS

DNS is the phonebook of the internet. Without it, users and applications cannot find the servers they need to communicate with.

The DNS Resolution Flow

  1. Local Cache: Your OS checks if the domain-IP mapping is cached locally.
  2. DNS Resolver: If not found, a DNS resolver (usually ISP-provided) is queried.
  3. Root Servers: If unresolved, the resolver queries root servers. There are 13 root server IPs globally, backed by hundreds of distributed servers.
  4. TLD Servers: Root servers point to Top-Level Domain (TLD) servers (like .com, .org).
  5. Authoritative Name Servers: These servers (e.g., AWS Route 53) store the actual domain-to-IP mapping.
  6. Return + Cache: The IP is returned and cached at multiple levels for faster resolution next time.

During the AWS outage, step 5 (authoritative resolution for DynamoDB domains) failed. Without IP addresses, services relying on DynamoDB were essentially cut off.


🔎 Why DynamoDB Was at the Center

DynamoDB is mission-critical. It stores:

  • User sessions for web apps.
  • Game states for online platforms like Fortnite.
  • Chat histories and social graph data for apps like Snapchat.
  • IoT and voice request handling for Alexa.

When DNS broke, these apps could not reach DynamoDB, leading to login failures, service outages, and broken app functionality worldwide.


🧩 The Bigger Picture: Why DNS is a Critical Weak Point

  • Single Points of Failure: Even distributed systems like DNS can become bottlenecks when authoritative name servers are impacted.
  • Global Ripple Effect: Because many apps hardcode AWS endpoints (like DynamoDB US-East-1), a localized DNS failure can cascade globally.
  • Invisible Dependency: Most developers rarely think about DNS — until it breaks. This outage highlights that DNS is not just background infrastructure; it’s mission-critical.

⚡ Can Such Failures Be Prevented?

While DNS is resilient, it’s not foolproof. Possible mitigations include:

  1. Multiple DNS Resolvers: Configure Google DNS (8.8.8.8), Cloudflare (1.1.1.1), or Quad9 as fallbacks.
  2. DNS Caching: Cache DNS lookups at the OS or application level.
  3. Hardcoding IPs (with caution): Useful for emergencies but risky due to changing IPs.
  4. Multi-Region Architectures: Deploy services across multiple AWS regions to avoid reliance on US-East-1.
  5. Graceful Failures: Applications should retry intelligently and serve cached data when possible.

However, when authoritative servers themselves fail, fallback resolvers cannot help. This is why multi-region redundancy is a necessity for large-scale services.


🧠 Lessons for Engineers and Companies

  • Architect for Failure: Assume DNS or a critical service may fail and design accordingly.
  • Monitor Dependencies: Don’t just monitor your servers — also monitor dependencies like DNS and cloud services.
  • Understand the Stack: Outages reveal how interconnected everything is. From DNS to databases, weak points can cascade.
  • Invest in Resilience: Multi-region, multi-cloud, and fallback systems are expensive but necessary for mission-critical services.

🎯 Key Takeaways

  • The outage occurred in US-East-1 (Northern Virginia).
  • Root cause: DNS resolution failure for DynamoDB domains.
  • Impacted services: Fortnite, Snapchat, Alexa, Amazon.com, and more.
  • DynamoDB was healthy; DNS prevented access.
  • The outage lasted nearly 15 hours.
  • Lesson: DNS is a hidden but critical single point of failure.

🔮 The Road Ahead

AWS has not yet published a full postmortem. But what’s clear is this: DNS remains one of the most crucial and vulnerable layers of the internet.

For businesses, the incident is a stark reminder to:

  • Diversify regions.
  • Build redundancy.
  • Always plan for the unexpected.

The AWS mass outage wasn’t just a technical hiccup — it was a lesson in how fragile global digital infrastructure can be. The next time #AWSDown trends, the companies that invested in resilience will be the ones still online.