On October 20, 2025, Amazon Web Services (AWS) — the world’s largest cloud provider — suffered a major outage in its US-EAST-1 region (Northern Virginia) that disrupted services globally for nearly 24 hours. The event underscored the critical dependency of modern internet infrastructure on a single cloud provider and reignited discussions about resilience, redundancy, and multi-cloud strategies.
Incident Overview
Event: Increased Error Rates and Latencies
Region: US-EAST-1 (N. Virginia)
Duration: October 19, 11:49 PM – October 20, 3:01 PM (PDT)
Severity: Disrupted
Primary Root Cause: DNS Resolution Failure in DynamoDB Endpoint
Affected Services: Over 140 AWS services including EC2, Lambda, S3, DynamoDB, CloudWatch, Redshift, and more.
Timeline and Root Cause Analysis
The outage began late on October 19, 2025, when engineers detected increased error rates across multiple AWS services. Initial investigations pointed to Amazon DynamoDB, a core database service powering numerous internal and customer applications. By 12:26 AM PDT, AWS identified that the issue stemmed from a faulty DNS update which disrupted endpoint resolution — effectively breaking the “phonebook” that directs services to their destinations.
The DNS failure triggered a cascade of dependent system errors:
- EC2 instance launches stalled due to DynamoDB dependencies.
- Network Load Balancer health checks failed, causing connectivity loss across services like Lambda, SQS, and CloudWatch.
- IAM updates and DynamoDB Global Tables also suffered delays due to reliance on the impacted region.
AWS engineers applied mitigations in parallel: flushing DNS caches, throttling EC2 instance launches, and gradually restoring network connectivity. By 2:24 AM PDT, the primary DNS issue was resolved, but network and EC2 subsystem issues lingered into the morning. The Network Load Balancer health subsystem was fully recovered by 9:38 AM PDT, with final service normalization at 3:01 PM PDT.
Impact Scope
The impact was extensive, affecting both enterprise services and popular consumer platforms worldwide. More than 140 AWS services were impaired, including:
- Compute & Networking: EC2, ECS, EKS, Elastic Load Balancing
- Data & Storage: DynamoDB, S3, RDS, Redshift, ElastiCache
- Serverless: Lambda, EventBridge, SQS, Step Functions
- Security & Management: IAM, AWS Organizations, CloudTrail, Config
- Developer Tools: CodeBuild, Amplify, AppSync, CloudFormation
The outage’s reach went beyond AWS customers. Global platforms such as Snapchat, Fortnite, Roblox, Coinbase, Venmo, and even Amazon’s own Prime Video and Ring services experienced disruptions. Financial institutions like Lloyds and Halifax reported login issues, and government portals temporarily went offline. With AWS holding roughly 33% of global cloud infrastructure market share, the event’s ripple effect was unprecedented.
Lessons in Cloud Dependence
This incident demonstrates a key challenge in modern cloud architecture: single-region dependency. Despite AWS’s multi-Availability Zone design, many global systems remain regionally anchored — particularly to US-EAST-1, which hosts numerous control plane and global API endpoints.
While no cyberattack was involved, the event revealed how an internal configuration error in a single foundational service (DNS in this case) can propagate across dependent systems, crippling global operations.
RELIANOID’s Perspective: Achieving True High Availability with GSLB
At RELIANOID, we believe that resilience in cloud environments must go beyond redundancy within a single provider. Our Global Server Load Balancing (GSLB) solution ensures continuous availability even when a major cloud provider or region experiences an outage.
How RELIANOID GSLB Helps Prevent Such Outages
- Multi-Cloud and Multi-Region Continuity: GSLB intelligently distributes traffic across independent regions or providers (e.g., AWS, Azure, GCP, on-premise), ensuring service continuity during regional or provider-level failures.
- Real-Time Health Monitoring: Continuous endpoint checks allow automatic rerouting of traffic to healthy nodes, minimizing downtime during events like DNS or API endpoint failures.
- Intelligent DNS Load Balancing: RELIANOID’s DNS-based GSLB dynamically resolves client requests to optimal data centers, mitigating risks tied to DNS misconfiguration or propagation delays.
- Seamless Failover and Recovery: With policies like weighted round robin, latency-based routing, and geolocation awareness, GSLB maintains service consistency and minimizes disruption even in complex multi-region deployments.
Implementing GSLB as part of a broader high-availability strategy decouples business-critical applications from the operational dependencies of a single provider. Whether an issue stems from DNS resolution, network health checks, or internal API failures, GSLB provides a transparent mechanism for automatic failover and continuous user experience.
Conclusion
The AWS US-EAST-1 outage of October 2025 serves as a powerful reminder: even the most advanced cloud infrastructures can fail. True resilience requires architectural independence, proactive failover mechanisms, and intelligent global load balancing.
RELIANOID’s GSLB delivers this resilience — helping organizations ensure uptime, reliability, and trust, regardless of where the next disruption originates.
Learn more about GSLB and high-availability strategies.