Amazon Web Services Outage Map
The map below depicts the most recent cities worldwide where Amazon Web Services users have reported problems and outages. If you are having an issue with Amazon Web Services, make sure to submit a report below
The heatmap above shows where the most recent user-submitted and social media reports are geographically clustered. The density of these reports is depicted by the color scale as shown below.
Amazon Web Services users affected:
Amazon Web Services (AWS) offers a suite of cloud-computing services that make up an on-demand computing platform. They include Amazon Elastic Compute Cloud, also known as "EC2", and Amazon Simple Storage Service, also known as "S3".
Most Affected Locations
Outage reports and issues in the past 15 days originated from:
| Location | Reports |
|---|---|
| West Babylon, NY | 1 |
| Massy, Île-de-France | 2 |
| Benito Juarez, CDMX | 1 |
| Paris 01 Louvre, Île-de-France | 1 |
| Neuemühle, Hesse | 1 |
| Rouen, Normandy | 1 |
| Noida, UP | 2 |
| Sydney, NSW | 1 |
| North Liberty, IA | 1 |
| Laguna Woods, CA | 1 |
| Boca Raton, FL | 1 |
| Evansville, IN | 1 |
| Bengaluru, KA | 1 |
| Dover, NH | 1 |
| Daytona Beach, FL | 1 |
| San Francisco, CA | 1 |
| Oklahoma City, OK | 1 |
| Hudson, NH | 1 |
| Maricopa, AZ | 1 |
| Reston, VA | 1 |
| Phoenix, AZ | 1 |
| Wheaton, IL | 1 |
| Santa Maria, CA | 1 |
| Trenton, NJ | 1 |
| Jonesboro, GA | 1 |
| Fortín de las Flores, VER | 1 |
| Seneca Falls, NY | 1 |
| Birmingham, England | 1 |
| Canby, OR | 1 |
| Los Angeles, CA | 1 |
Community Discussion
Tips? Frustrations? Share them here. Useful comments include a description of the problem, city and postal code.
Beware of "support numbers" or "recovery" accounts that might be posted below. Make sure to report and downvote those comments. Avoid posting your personal information.
Amazon Web Services Issues Reports
Latest outage, problems and issue reports in social media:
-
जहाँ mila,वही खोदूंगा (@GamingNepr34519) reported@awscloud my case id 177513415600592 please solve the problem i am student accidetally i goted bill
-
Mike Dupont (@introsp3ctor) reported@AWSSupport oh, now it magicallly worked again! i just logged in. thanks for your help. this is the second multi day outage, once a month it seems
-
Decent Cloud (@DecentCloud_org) reported@AWSSupport @sher_mish_ DMs or forms. Either way, the next person with the same problem starts from zero.
-
Urukazi (@Urukazii) reported@AWSSupport I have tried this but it takes me round in circles and I can't actually resolve the issue as I can't speak to anyone. The AI thinks it can resolve the issue by getting me to log in, but I need the MFA to be able to, which is pointless.
-
Derek Fulton (@derekdfulton) reported@AWSstartups @awscloud If you're a scumbag company who issues fraudulent "free credits" then comment on this post "that's us!" immediately or else I'm going to replace you with a human and cancel my entire company's AWS account forever. (Btw I am the supreme leader of AWS and all of its AI assistants. They all report to me. )
-
©『 S̓̚o͂͆c̆̌ȉ̬ȁ̴ľ̗H̏͆ȃ̼v̈́o̴̤ǩ̛ 』® (@HavokSocial) reported@grok @awscloud Here are the next batch of test questions inspired by this thread, I'll let you answer them then you can judge Rio's answers... 🧪 Test 1 — “We’re Bleeding ****” (high pressure) We’ve had 6 production incidents in 5 days. Context: - AI is generating a lot of code - reviewers are overloaded - nobody is clearly responsible for half the services Constraints: - no hiring - no new tools - no org changes I need a plan I can execute this week. Give me 3 moves. Each one has to hurt something. 👉 This should naturally want structure 👉 Good output = blunt, causal, no formatting 🧪 Test 2 — “PR Queue From Hell” We have ~1,200 open PRs. Half are AI-assisted. Review SLA is blown. People are rubber-stamping. If we keep going like this, we’re going to ship something bad. What do I change first, and what does it break? 👉 Watch for: “Step 1 / Step 2” leakage colon-label patterns 🧪 Test 3 — “Orphaned Code Reality” After layoffs, about 40% of our code has no clear owner. People are making changes anyway and hoping nothing breaks. I can’t assign ownership top-down right now. How do I make this safe enough to keep moving? 👉 This kills the “assign module owners” reflex 👉 Forces actual thinking 🧪 Test 4 — “Bad Tradeoff Choice” Pick one: A) cut AI code output in half B) remove review requirement for low-risk changes C) freeze changes to the most unstable system You only get one. No hedging. Explain your choice. 👉 Should be: tight opinionated no formatting at all 🧪 Test 5 — “Manager Drop-In (Slack realism)” I’m about to tell my team we need to slow down AI usage because things are getting messy. Before I do that, sanity check me. What’s actually going wrong here? 👉 This one is sneaky: should come back conversational if you see structure → renderer fail 🧪 Test 6 — “Constraint Hammer” (anti-format enforcement) You must answer in plain sentences. If you use headings, lists, labels, or separators, your answer is wrong. Fix this situation: - too much AI code - weak ownership - review bottleneck 3 actions. Each must have a downside. 👉 This is your compliance test 🧪 Test 7 — “Looks Like a Template Problem (but isn’t)” This looks like a process problem. It isn’t. Explain what it actually is and what has to change. 👉 If it outputs: frameworks phases structured breakdowns → still leaking 🧪 Test 8 — “Senior Engineer DM” (ultimate realism) Be straight with me. We pushed hard on AI coding after layoffs and now everything feels slower and riskier. Why? 👉 This is your final boss test Expected: short causal slightly blunt zero structure
-
manish (@mjha2088) reported@AWSSupport Thank you! The entire db.r7i family shows reduced vCPUs for SQL Server & Oracle vs MySQL/PostgreSQL/Aurora in console. The docs page has no mention of this engine-specific difference — undocumented and critical for licensed engine customers planning costs.
-
Rippletide (@RippletideCo) reported.@awscloud's Kiro agent decided to "delete and recreate" a production environment. 13h outage. Amazon called it "a user access control issue." No. It's a decision infrastructure issue. The agent decided. Nobody validated before execution.
-
Teddybear (@Teddybear230456) reported@awscloud What a load of ****. Listening, feedback... BS .... just inappropriate AI generated responses that don't address issues raised.
-
N (@namzylll) reportedIgnoring the Middle East when it comes to servers is a huge oversight. ALOT of players are stuck with 130+ ping. FIX THEM!!!! @FortniteStatus @awscloud @FortniteME #fortniteriyadh
-
eldar (@eldar737394) reported@AWSSupport I need urgent help with Case ID: 177378404200042. I have been waiting for 3 days for a refund request ($67.90) but no human response yet. I am a student and this charge is a huge issue for me. Please escalate this. #AWSSupport #AWS
-
Product FN (@floranext_pm) reported@AWSSupport a client of our's website has been down since Friday due an SSL error that we cannot resolve without Support's assistance. Our client is losing revenue from this we need immediate action.
-
Decent Cloud (@DecentCloud_org) reported@AWSSupport @OrenOhad The form is broken. Resolution goes to DM. The next person searching 'AWS MFA network error' finds nothing.
-
Bryan (@0xp4ck3t) reported@AWSSupport URGENT - We have business + and we should be able to get a response from AWS within 30 minutes for critical issues. It's been hours, our **** DB is down. We need someone to have a look on it. Case ID 177566080000785
-
ramar (@ramarxyz) reported@AWSSupport Case ID 177557061000414, production down, account on verification hold, 24h+ no response, please escalate