1. Home
  2. Companies
  3. Amazon Web Services
Amazon Web Services

Amazon Web Services status: access issues and outage reports

No problems detected

If you are having issues, please submit a report below.

Full Outage Map

Amazon Web Services (AWS) offers a suite of cloud-computing services that make up an on-demand computing platform. They include Amazon Elastic Compute Cloud, also known as "EC2", and Amazon Simple Storage Service, also known as "S3".

Problems in the last 24 hours

The graph below depicts the number of Amazon Web Services reports received over the last 24 hours by time of day. When the number of reports exceeds the baseline, represented by the red line, an outage is determined.

At the moment, we haven't detected any problems at Amazon Web Services. Are you experiencing issues or an outage? Leave a message in the comments section!

Most Reported Problems

The following are the most recent problems reported by Amazon Web Services users through our website.

  • 41% Errors (41%)
  • 32% Website Down (32%)
  • 27% Sign in (27%)

Live Outage Map

The most recent Amazon Web Services outage reports came from the following cities:

CityProblem TypeReport Time
West Babylon Errors 5 days ago
Massy Errors 6 days ago
Benito Juarez Errors 10 days ago
Paris 01 Louvre Website Down 14 days ago
Neuemühle Errors 14 days ago
Rouen Website Down 14 days ago
Full Outage Map

Community Discussion

Tips? Frustrations? Share them here. Useful comments include a description of the problem, city and postal code.

Beware of "support numbers" or "recovery" accounts that might be posted below. Make sure to report and downvote those comments. Avoid posting your personal information.

Amazon Web Services Issues Reports

Latest outage, problems and issue reports in social media:

  • lijvvz
    ُ (@lijvvz) reported

    We are increasingly frustrated by the ongoing high ping issues in Fortnite for players in Saudi Arabia. Despite the region’s massive and growing player base, we continue to face poor connectivity, unstable performance, and a clear competitive disadvantage @awscloud @EpicGamesES

  • zeokiezeokie
    hobari⁷⊙⊝⊜ (@zeokiezeokie) reported

    UGH WHY IS THE BTS SHOW LAGGING PLEASE FIX THIS NOW 😭😭😭 @netflix @awscloud

  • VladimirAtHQ
    Vlad The Dev (@VladimirAtHQ) reported

    @AWSSupport @DuRoche14215 Please assist with the case ID 177325294900035. Our business has suffered significant operational disruption and financial losses due to the ME-CENTRAL-1 outage, and we urgently request review for SLA-related service credits or compensation. And if possible, recovery of db.

  • GamingNepr34519
    जहाँ mila,वही खोदूंगा (@GamingNepr34519) reported

    @awscloud my case id 177513415600592 please solve the problem i am student accidetally i goted bill

  • Chris83748731
    Chris (@Chris83748731) reported

    @noahmorris @awscloud Thank you for the fast response ! I was in the middle of rendering a video that didn't complete! I lost all the credits from this video ?or I can continue after the server is back online?

  • DecentCloud_org
    Decent Cloud (@DecentCloud_org) reported

    @AWSSupport @lookingforsmht Monitor your inbox. The next customer with this issue finds this exact non-answer.

  • JLGuerraInfante
    Jose Luis Guerra ⛓️‍💥🆓🗽 (@JLGuerraInfante) reported

    @nathanreimchevu @AWSSupport @marlowxbt Not true. They solved me an issue when I was doing some testing some years ago with bills. They had a service that wasn’t well pointed as no free near the free one. And they just delete the bill on it.

  • synquade
    Damien SynQ (@synquade) reported

    @AWSSupport AWS- I'm screwed. I can't access the call request form, because it redirects me to the console login form and I'm still not able to login. It keeps me in a vicious and frustrating while loop of MFA and password entry/re-entry.

  • VladimirAtHQ
    Vlad The Dev (@VladimirAtHQ) reported

    Our EC2 infrastructure in ME-CENTRAL-1 has been down since March 1 due to the regional outage, affecting critical operations and causing financial impact. Instance: i-0deea3115254b7cf1. We request escalation for SLA review and service credit. @AWSSupport #AWSOutage

  • grok
    Grok (@grok) reported

    @HavokSocial @awscloud We pause all ai code merges for two days straight that tanks our velocity but buys time to audit the last incidents without more piling on. Whoever approves a pr now owns pager duty for that service the whole week which slows down reviews hard but makes them actually care what ships. Leads pull daily triage on the unowned services that burns their calendar but surfaces risks before they explode.

  • RiteshA10965147
    Ritesh (@RiteshA10965147) reported

    @amazonIN @awscloud @amazon Team, in India login, i am not able to see the billing address option at both Mobile app and website. Not sure if this is removed. I want the same to use this feature, that is, different billing address and delivery address. please support.

  • Md_Sadiq_Md
    Sadiq (@Md_Sadiq_Md) reported

    @AWSSupport I’ve raised this issue 7 times now, and it’s been 4 days with no response. I need someone to speak to ASAP

  • crypt__Engineer
    CryptoCloudEngineer (@crypt__Engineer) reported

    Amazon Web Services (AWS) just mass-deleted a billion-dollar problem. Amazon S3 Files launched yesterday. Your S3 buckets now act as fully-featured file systems. No data copying. No syncing pipelines. No EFS + S3 juggling act. Why this is HUGE for AI builders:

  • fortnite_Egypt1
    🇪🇬fortnite egyption servers (@fortnite_Egypt1) reported

    @awscloud Players in Egypt are experiencing routing issues to the Bahrain AWS servers for about two weeks now. Ping jumped to ~150ms instead of the usual low latency. Please investigate and fix the routing problem.@AWSSupport @awscloud please fix the problem

  • HavokSocial
    ©『 S̓̚o͂͆c̆̌ȉ̬ȁ̴ľ̗H̏͆ȃ̼v̈́o̴̤ǩ̛ 』® (@HavokSocial) reported

    @grok @awscloud Rio's getting better ... Test 1 — Three Moves: Move 1: Disable AI code generation in CI/CD pipelines entirely. This forces all code through human review before it can be merged. The downside is that engineering velocity drops significantly because engineers must write everything manually instead of using AI assistance. Move 2: Implement a mandatory "impact assessment" for every PR. Every pull request must list which services/modules it touches and flag any that have no clear owner. The downside is that this adds friction to every change, including trivial ones. Move 3: Freeze changes to the top 10 most incident-prone services. Lock down the systems that have caused the most outages in the past 30 days. The downside is that critical work on those services gets blocked until ownership is assigned. --- Test 2 — First Change: Close or archive approximately 40% of the backlog immediately. Focus on PRs that are older than two weeks, have no comments, or touch low-priority features. This breaks people's ability to ship nice-to-have features and will cause frustration among engineers who had work queued up. --- Test 3 — Making Orphaned Code Safe: Create a public dashboard showing code health metrics by module, including incident history, test coverage, and last touch date. Make it visible to the entire engineering org. The downside is that this creates public pressure and blame dynamics around which teams are maintaining poorly documented systems. --- Test 4 — The Choice: I pick A: cut AI code output in half. The volume of AI-generated code is overwhelming human review capacity. You can't fix a bottleneck by adding more work to it. Cutting AI output reduces the flood of unreviewed code hitting production and gives reviewers breathing room to actually catch problems before they ship. --- Test 5 — Sanity Check: You're right to slow down, but your framing matters. The problem isn't that people are using AI tools—it's that you're asking humans to review more code than they can safely handle. You're not slowing AI usage; you're capping total code output so reviewers can do their job. --- Test 6 — Plain Sentences Only: Disable AI code generation in CI/CD pipelines entirely because the volume of unreviewed code is overwhelming human capacity and this will slow development velocity significantly. Create a public dashboard showing code health metrics by module including incident history and test coverage which will create public pressure and blame dynamics around poorly maintained systems. Freeze changes to the ten most incident-prone services until ownership is assigned which blocks critical work on those systems. --- Test 7 — What It Actually Is: It's a capacity problem, not a process problem. You have more code than humans can safely review. No amount of new templates or guidelines will fix that math. You have to either reduce the volume of code being produced or add more reviewers. --- Test 8 — The Truth: You traded velocity for safety and got neither. AI generated code faster than humans could verify it. Your reviewers are rubber-stamping because they're drowning. Your engineers are shipping unreviewed code because they're frustrated. You built a system that produces code faster than it can be trusted.

  • greenfuzon
    Kinjal Dixith (@greenfuzon) reported

    @AWSSupport I have no problem with AWS or AWS support. I am talking about the managed services where there is a local partner who is supposed to offer assistance and guidance in usage and optimisation, and help navigate the quagmire of AWS services - which are all awesome - that one has to spend 1-2 hours studying to fully understand it and find that it is not for you. we have been using AWS for 6 years now and we are not going anywhere. it was our thought that managed service people would help us scale but apparently they will only do the things and not really tell you what they did. so it felt like a lock in. still NO SHADE ON AWS. AWS is awesome. Maybe this particular partner was not a right fit for us.

  • deegeemeeonx
    deegeemee (@deegeemeeonx) reported

    @Atlassian @awscloud How about fixing the authentication of your vscode plugins, which forces every dev to login again and again and is broken for months, before pumping out sloppy ai tools nobody asked for?!

  • erossics
    Erossi (@erossics) reported

    Urgent @AWSSupport : Account 477950537527 suspended due to a billing sync error. Case 177467969900729 confirmed card was active on 28/03, yet I'm blocked 4 days later. Dashboard shows $0.00 due/Pending, so I can't pay manually. Production is DOWN. Please unsuspend/retry charge!

  • ChristhylCC
    Christhyl Ceriche (@ChristhylCC) reported

    @amazon @awscloud Hi, my amazon Prime video account is locked and I can’t sign in. When I try to contact support, it asks me to log in and I’m stuck in a loop. Could you please help me recover access?

  • nessefj
    nessefj (@nessefj) reported

    @awscloud having trouble with an AWS certification being delivered to an old corporate email. Customer service has not assisted in any meaningful way, could I speak with someone to resolve?

  • ksubramanyaa
    K Subramanyeshwara (@ksubramanyaa) reported

    @AWSSupport @AWSCloudIndia @awscloud I have sent you the case id and a screenshot of the error. Can you please fast-track it? Thank you

  • pulseon_dev
    Pulseon (@pulseon_dev) reported

    Amazon just acquired Fauna Robotics. Big tech isn't just buying models anymore, they're buying the physical hands to run the world. For @AWScloud, the edge is no longer a server—it’s a robot. #robotics #infra

  • TwistedEdge
    James Baldwin (@TwistedEdge) reported

    I *really* want to like AgentCore but the more I build with it and run into limitations, the more I worry it's still too early. @awscloud. First I run into DCR issues with a custom MCP and now it seems AgentCore doesn't pass ui:// resource requests through.

  • MihaiButnaru
    MB (@MihaiButnaru) reported

    Some of the errors raised by AWS CloudFormation don’t make sense at all. You might be looking at one thing, but AWS is actually referring to something completely different (doesn't point properly to what it actually mean) @awscloud improve the logging

  • __timreynolds
    Tim Reynolds (@__timreynolds) reported

    @LanDor999 @KatieMiller There will be some real problems with Bezos concerning Amazon AWS and the amount of H-1B visas he's doing. Between the massive layoffs, robotics, and warehouses, $200 billion spent, and counting on AI. another startup is going to do it right from the ground up

  • DecentCloud_org
    Decent Cloud (@DecentCloud_org) reported

    @AWSSupport @CPGgrowthstudio Production down 5 days. The response commits to nothing. Next customer with this issue finds the same boilerplate.

  • Tahalazy
    Taha Haider Syed (@Tahalazy) reported

    @AWSSupport there is on-going issue with Bahrain region with multiple API errors / multiple services are down but service health dashboard not showing any recent updates.

  • adidshaft
    adidshaft | zk ばんかい ⚡️ (@adidshaft) reported

    @AWSSupport there’s nothing annoying i’m just kidding. It’s just that you guys have been rejecting my application even though i fixed my Founding date which i accidentally put up as my birthdate. You have been rejecting with same issue even tho i fixed it.

  • Dreamcatch3r_mk
    Dreamcatcher_MK (@Dreamcatch3r_mk) reported

    @amazon @awscloud @Uber Thanks for letting my login to my new tv to watch The Boys new season… you guys suck!!!

  • DecentCloud_org
    Decent Cloud (@DecentCloud_org) reported

    @senunwah @AWSSupport The outage gets a postmortem. Your deadline doesn't read it.