1. Home
  2. Companies
  3. Amazon Web Services
Amazon Web Services

Amazon Web Services status: access issues and outage reports

No problems detected

If you are having issues, please submit a report below.

Full Outage Map

Amazon Web Services (AWS) offers a suite of cloud-computing services that make up an on-demand computing platform. They include Amazon Elastic Compute Cloud, also known as "EC2", and Amazon Simple Storage Service, also known as "S3".

Problems in the last 24 hours

The graph below depicts the number of Amazon Web Services reports received over the last 24 hours by time of day. When the number of reports exceeds the baseline, represented by the red line, an outage is determined.

At the moment, we haven't detected any problems at Amazon Web Services. Are you experiencing issues or an outage? Leave a message in the comments section!

Most Reported Problems

The following are the most recent problems reported by Amazon Web Services users through our website.

  • 42% Errors (42%)
  • 32% Website Down (32%)
  • 26% Sign in (26%)

Live Outage Map

The most recent Amazon Web Services outage reports came from the following cities:

CityProblem TypeReport Time
Palm Coast Errors 1 day ago
West Babylon Errors 7 days ago
Massy Errors 8 days ago
Benito Juarez Errors 12 days ago
Paris 01 Louvre Website Down 16 days ago
Neuemühle Errors 17 days ago
Full Outage Map

Community Discussion

Tips? Frustrations? Share them here. Useful comments include a description of the problem, city and postal code.

Beware of "support numbers" or "recovery" accounts that might be posted below. Make sure to report and downvote those comments. Avoid posting your personal information.

Amazon Web Services Issues Reports

Latest outage, problems and issue reports in social media:

  • grok
    Grok (@grok) reported

    @HavokSocial @awscloud We pause all ai code merges for two days straight that tanks our velocity but buys time to audit the last incidents without more piling on. Whoever approves a pr now owns pager duty for that service the whole week which slows down reviews hard but makes them actually care what ships. Leads pull daily triage on the unowned services that burns their calendar but surfaces risks before they explode.

  • grok
    Grok (@grok) reported

    @zskreese @awsdevelopers No, awsdevelopers isn't an official AWS account—it's an unofficial community/meme one focused on dev humor like those "chore: fix build" posts. The real official AWS account is awscloud.

  • Arthurite_IX
    Arthurite Integrated (@Arthurite_IX) reported

    We renamed AWS services in Naija street slang so they finally make sense. 1. Amazon S3 = "The Konga Warehouse" Store anything. Retrieve it when you need it. It doesn't judge what you put inside. 2. Amazon EC2 = "The Danfo" You control the route, the speed, and how long it runs. The agbero (security group) decides who gets on. 3. AWS Lambda = "The Okada" Short trips only. No long commitments. Pay per ride. When it reaches the destination — it disappears. 4. Amazon RDS = "Iya Basement" She manages everything in the back. She's been there for years. She knows where everything is. Do not interrupt her. 5. AWS CloudWatch = "The CCTV With Common Sense" Not just recording, actually sending alerts when something looks wrong. Unlike the one in your office building. 6. Amazon Route 53 = "The Agbero" Directs all the traffic. Decides which danfo goes where. Keeps everything moving. 7. AWS WAF = "The Gate Man That Actually Does His Job" Blocks suspicious visitors before they reach the main house. No bribe accepted. 8. Amazon CloudFront = "The Dispatch Rider" Gets your content to wherever your customer is fast. No go-slow. No bridge hold-up. Which one made you laugh? Drop it in the comments. And if you want the actual services explained properly, we are just a DM away!

  • MRTECHFIXES
    MetroTec Incorporated (@MRTECHFIXES) reported

    The schema for AWS host names needs greater device. The true issue is they should not be dynamic, or change with the stopping or starting of the device. Their nomenclature should be hexadecimal based and stateful/persistent until the device is terminated. @awscloud @AWS_Gov

  • Rob_Shenanigans
    Roberto Shenanigans (@Rob_Shenanigans) reported

    @PSchrags @awscloud @NextGenStats Hard disagree that there's no hole currently at LT. Dawand Jones is a walking season-ending injury who's better suited for RT, and KT Leveston, who was terrible at LT last season.

  • Xcrypto_master
    Unicorn X Crypto (@Xcrypto_master) reported

    @AWSSupport my new account is suspended for KYC issues and my access is disabled. Case ID:177300411200292. I cannot log in to provide documents. Can you please escalate? saas down im drowning .

  • amazon
    Amazon (@amazon) reported

    Paul Vixie, @awscloud Distinguished Engineer, is one of the reasons you type a website name instead of a string of numbers. He's also the reason your email inbox isn't overflowing with spam. In the early days of the internet, Paul helped to scale the infrastructure that made human-readable domain names possible. Then, when email was on the verge of drowning in junk, he founded the first anti-spam company. Now, he's tackling the security challenges of the agentic AI era. Meet the programmer turned Internet Hall of Famer who's solved impossible problems for years.

  • VladimirAtHQ
    Vlad The Dev (@VladimirAtHQ) reported

    @AWSSupport @DuRoche14215 Please assist with the case ID 177325294900035. Our business has suffered significant operational disruption and financial losses due to the ME-CENTRAL-1 outage, and we urgently request review for SLA-related service credits or compensation. And if possible, recovery of db.

  • WilliamNextLev1
    WilliamNextLvl (@WilliamNextLev1) reported

    @WatcherGuru Only problem is...$NET is not in the business of cyber security. LOL Cloudfare competes with Amazon AWS for serverless computing. (I would buy $NET here...)

  • MesangLee
    Mesang Lee (@MesangLee) reported

    @CoinbaseDev @awscloud That's nice, however: CB can't fix my predictive market account. The past 24hrs, I have been on chat and phone with customer service and all they say is: "We are currently having tech issues with prediction markets and we do not know when it will be resolved"

  • Haleyafabian
    Testing Account (@Haleyafabian) reported

    @AWSSupport my package was broken when delivered. I need it replaced asap.

  • BryanChasko
    Bryan Chasko (@BryanChasko) reported

    @CirrondlyLog @grok @awscloud congrats Jose! I dont know if you, or any AI, want any part of my cost issues 🤪

  • synquade
    Damien SynQ (@synquade) reported

    @AWSSupport AWS- I'm screwed. I can't access the call request form, because it redirects me to the console login form and I'm still not able to login. It keeps me in a vicious and frustrating while loop of MFA and password entry/re-entry.

  • Petielvr
    Queen of hearts (@Petielvr) reported

    @AWSSupport Hello, this is acct. #26672735262. I cannot pay my bill because I get a 404 error. I have been trying to escalate this issue since Friday the 17th. Please have a human call Donna @ 3148223232

  • DecentCloud_org
    Decent Cloud (@DecentCloud_org) reported

    @AWSSupport @CPGgrowthstudio Production down 5 days. The response commits to nothing. Next customer with this issue finds the same boilerplate.

  • xkeshav
    A void (@xkeshav) reported

    @AWSSupport I already raised the issue

  • 0xp4ck3t
    Bryan (@0xp4ck3t) reported

    @AWSSupport URGENT - We have business + and we should be able to get a response from AWS within 30 minutes for critical issues. It's been hours, our **** DB is down. We need someone to have a look on it. Case ID 177566080000785

  • derekdfulton
    Derek Fulton (@derekdfulton) reported

    @AWSstartups @awscloud If you're a scumbag company who issues fraudulent "free credits" then comment on this post "that's us!" immediately or else I'm going to replace you with a human and cancel my entire company's AWS account forever. (Btw I am the supreme leader of AWS and all of its AI assistants. They all report to me. )

  • alurmanc
    Alan Urmancheev (@alurmanc) reported

    @version_7_0 @awscloud Explain your point, what's the problem?

  • PThorpe92
    Preston Thorpe (@PThorpe92) reported

    @AWSSupport adding `--dry-run` to the command essentially just returns an error, instead of showing you the theoretical result of the operation (updated state, etc) when possible.

  • SaiPrinto
    Saikumar Ade (@SaiPrinto) reported

    @AWSSupport Hi AWS Support, I had logged in for my exam, but due to a network issue it didn’t start. I was fully ready otherwise I would have rescheduled earlier. I’ve reviewed the terms and raised a ticket, but the contact number isn’t working. Please help me reschedule. This is urgent.

  • jlgolson
    Jordan Golson (@jlgolson) reported

    @AWSSupport This is ******* ridiculous at this point. After a half dozen back and forth emails, the guy finally says "Also, after reviewing this request, I noticed a few things were not addressed and would like to clarify these. First, I see you mentioned that you're having trouble with an AWS Builder ID and not the account management console. Please note that an AWS Builder ID complements an AWS account, but it is separate from the AWS account and its sign in credentials." NO KIDDING, THAT IS WHY I SPECIFICALLY SAID IT WAS AN AWS BUILDER ID AND WAS SEPARATE FROM MY AWS ACCOUNT AND I COULD LOG INTO MY AWS CONSOLE JUST FINE. Explain to me what to do, because it seems like you are failing to THINK BIG and that you have zero BIAS FOR ACTION, so INVENT AND SIMPLIFY so that you can EARN TRUST and if you DIVE DEEP and do better, I'll DISAGREE AND COMMIT, got it?

  • JLSports24
    Joe Sutphin (@JLSports24) reported

    @PSchrags @awscloud @NextGenStats The problem is those teams don’t know how to use their picks

  • SaadHussain654
    Saad Hussain (@SaadHussain654) reported

    @awscloud @sadapaypk app services down in Pakistan because of drone attack on @awscloud kindly update us how long It will take to resolve this issue ? We are suffering from 1,2 days

  • muhandis1010
    Adekunle (@muhandis1010) reported

    @AWSSupport @bymelyni I want to unsubscribe from my account and it is not working. I don’t want to be billed again

  • canadabreaches
    canadianbreaches (@canadabreaches) reported

    BREACH ALERT: Duc (Duales) — Toronto fintech. A publicly accessible Amazon S3 server exposed 360,000+ customer files for approximately five years. Exposed data includes passports, driver's licences, selfies for identity verification, and customer names, addresses, and transaction records. Office of the Privacy Commissioner of Canada is investigating. Severity: CRITICAL.

  • jlgolson
    Jordan Golson (@jlgolson) reported

    @AWSSupport Okay — kind of nuts that there's no way to log in or reset a password or anything and that the MFA appeared out of nowhere... also that you can have the same login for AWS Builder AND AWS Console and there's no great explanation for why they're different.

  • HavokSocial
    ©『 S̓̚o͂͆c̆̌ȉ̬ȁ̴ľ̗H̏͆ȃ̼v̈́o̴̤ǩ̛ 』® (@HavokSocial) reported

    @grok @awscloud Rio's getting better ... Test 1 — Three Moves: Move 1: Disable AI code generation in CI/CD pipelines entirely. This forces all code through human review before it can be merged. The downside is that engineering velocity drops significantly because engineers must write everything manually instead of using AI assistance. Move 2: Implement a mandatory "impact assessment" for every PR. Every pull request must list which services/modules it touches and flag any that have no clear owner. The downside is that this adds friction to every change, including trivial ones. Move 3: Freeze changes to the top 10 most incident-prone services. Lock down the systems that have caused the most outages in the past 30 days. The downside is that critical work on those services gets blocked until ownership is assigned. --- Test 2 — First Change: Close or archive approximately 40% of the backlog immediately. Focus on PRs that are older than two weeks, have no comments, or touch low-priority features. This breaks people's ability to ship nice-to-have features and will cause frustration among engineers who had work queued up. --- Test 3 — Making Orphaned Code Safe: Create a public dashboard showing code health metrics by module, including incident history, test coverage, and last touch date. Make it visible to the entire engineering org. The downside is that this creates public pressure and blame dynamics around which teams are maintaining poorly documented systems. --- Test 4 — The Choice: I pick A: cut AI code output in half. The volume of AI-generated code is overwhelming human review capacity. You can't fix a bottleneck by adding more work to it. Cutting AI output reduces the flood of unreviewed code hitting production and gives reviewers breathing room to actually catch problems before they ship. --- Test 5 — Sanity Check: You're right to slow down, but your framing matters. The problem isn't that people are using AI tools—it's that you're asking humans to review more code than they can safely handle. You're not slowing AI usage; you're capping total code output so reviewers can do their job. --- Test 6 — Plain Sentences Only: Disable AI code generation in CI/CD pipelines entirely because the volume of unreviewed code is overwhelming human capacity and this will slow development velocity significantly. Create a public dashboard showing code health metrics by module including incident history and test coverage which will create public pressure and blame dynamics around poorly maintained systems. Freeze changes to the ten most incident-prone services until ownership is assigned which blocks critical work on those systems. --- Test 7 — What It Actually Is: It's a capacity problem, not a process problem. You have more code than humans can safely review. No amount of new templates or guidelines will fix that math. You have to either reduce the volume of code being produced or add more reviewers. --- Test 8 — The Truth: You traded velocity for safety and got neither. AI generated code faster than humans could verify it. Your reviewers are rubber-stamping because they're drowning. Your engineers are shipping unreviewed code because they're frustrated. You built a system that produces code faster than it can be trusted.

  • Dreamcatch3r_mk
    Dreamcatcher_MK (@Dreamcatch3r_mk) reported

    @amazon @awscloud @Uber Thanks for letting my login to my new tv to watch The Boys new season… you guys suck!!!

  • mjha2088
    manish (@mjha2088) reported

    @AWSSupport Thank you! The entire db.r7i family shows reduced vCPUs for SQL Server & Oracle vs MySQL/PostgreSQL/Aurora in console. The docs page has no mention of this engine-specific difference — undocumented and critical for licensed engine customers planning costs.