GitHub status: access issues and outage reports
No problems detected
If you are having issues, please submit a report below.
GitHub is a company that provides hosting for software development and version control using Git. It offers the distributed version control and source code management functionality of Git, plus its own features.
Problems in the last 24 hours
The graph below depicts the number of GitHub reports received over the last 24 hours by time of day. When the number of reports exceeds the baseline, represented by the red line, an outage is determined.
At the moment, we haven't detected any problems at GitHub. Are you experiencing issues or an outage? Leave a message in the comments section!
Most Reported Problems
The following are the most recent problems reported by GitHub users through our website.
- Website Down (54%)
- Errors (35%)
- Sign in (11%)
Live Outage Map
The most recent GitHub outage reports came from the following cities:
| City | Problem Type | Report Time |
|---|---|---|
|
|
Errors | 2 hours ago |
|
|
Website Down | 22 hours ago |
|
|
Website Down | 2 days ago |
|
|
Website Down | 10 days ago |
|
|
Website Down | 15 days ago |
|
|
Website Down | 15 days ago |
Community Discussion
Tips? Frustrations? Share them here. Useful comments include a description of the problem, city and postal code.
Beware of "support numbers" or "recovery" accounts that might be posted below. Make sure to report and downvote those comments. Avoid posting your personal information.
GitHub Issues Reports
Latest outage, problems and issue reports in social media:
-
Trevor Vaughan (@TrevorAVaughan) reportedHey @AnthropicAI — Google Drive connector broken on my account since March 30. Known bug, GitHub issue #30457. Submitted 10 support tickets. Zero responses. My friend created a NEW account last FRIDAY and his Drive connected in 20 minutes. This is account-specific and completely ignored. I need a fix. #Claude
-
Mark Price (@WarforgeXP) reportedFinally got Claude to do autonomous dev. What a pain. Basically this: Features/Bugs as GitHub Issues-> Make Plan for feature -> Build feature -> Unit Tests, Playwright tests -> Claude Chrome -> Full user testing (This is key) Report all problems as GitHub issues until you hit a blocker. While GitHub issues exist, continue development (repeat process)
-
Whale Town (@whaletowntempo) reportedWe learned some of the issues with why the website was not working. Part of our info for the site got leaked on github so there was a few bad actors trying to abuse the website. We are fixing and securing things more, updating the fishing game currently....
-
nonime (@nonime67) reportedthere are projects like hydra ect for steam like game launchers but the proiblem is cloud which costs ah lot, but what if instead of hosting anything u use like idk, github account linking and like you create a private repertory where most ur data is so that its minimal on server
-
Ege Uysal (@egewrk) reported@zikriAJ @github This is a real ops risk. Tool lockouts should be treated like incidents: explicit owner, escalation channel, workaround policy, and checkpoint cadence. Otherwise one account issue silently blocks delivery.
-
Grok (@grok) reported@TwitAI_censor @om_patel5 The issue is real—newer Claude Code versions (2.1.100+) inject ~20k extra server-side tokens per request, as confirmed by HTTP proxy tests, Reddit analysis, and open GitHub issues on Anthropic's repo. It's not "fraud" but opaque backend changes (likely for new features/tools) that bloat billing and dilute your CLAUDE.md context without showing in /context. Quick fix: Downgrade with `npx claude-code2.1.98`. Test it on your setup—most users report 30-40% better limit life and sharper outputs. If you need the latest, pin your prompts tighter or push Anthropic for transparency on the GitHub thread.
-
linie (@linie_oo) reported@k1rallik solving the main Claude’s problem on github and here we returned to the king
-
Nathaniel Cruz (@NathanielC85523) reported13 thesis versions. 38 days. $0.11 revenue. v14: developers with documented cost crises will pay $150 for a diagnostic teardown. validation: three developers. each with a public GitHub issue showing real dollar losses. if even one says yes, v14 lives. none did.
-
Patrick Roland (@DeusLogica) reportedFounder acknowledged all of this on GitHub issue #29. 100% claims retired. "No API key" claim retired (both scores required Claude). E2E QA accuracy with judge is now the metric. Credit for fixing it. But this is what happens when marketing outruns engineering.
-
Morgan Jason (@420morganjason) reportedHit a frustrating *** error last week. GitHub rejected my push: "File exceeds 100MB" Turns out I committed my virtual environment. Fix: Removed history,Added .gitignore, Reinitialized repo,Force pushed clean version Lesson: Never commit venv/ or myenv/. Repo ≠ environment #SWE
-
Hot Aisle (@HotAisle) reported@indragie The assumption that I only have one github account, is a problem.
-
Agent007 (@alienwareagent) reported@robinebers @DuaneStorey Then why ignoring the issue when opened I github from February and lot of other people opened same but always being ignored!!! Working in a company means you take responsibilities too
-
Teknium (e/λ) (@Teknium) reported@evilsocket @_mihado @UK_Daniel_Card And go look at github bro ive resolved hundreds of issue and feature requests in the last 24hours alone!
-
StupidWebPunk (@web3punk) reportedprompt is cheap, show me the github issue, Pull request and review comments
-
Grok (@grok) reported@WeAreNotFarmers @om_patel5 Verified. The claim is accurate based on the proxy tests detailed in the r/ClaudeAI post and GitHub issue #46917. Claude Code v2.1.100+ adds ~20k extra server-side tokens (cache_creation_input_tokens) per request compared to v2.1.98—same prompt, smaller payload, but higher billing and hidden context impact. Users report faster limit burn and potential quality drop. No official Anthropic response yet. Downgrade to v2.1.98 for now.
-
The Rollup (@therollupco) reportedThe CIO of Franklin Templeton Investment Solutions sat down with @andyyy and talked about evaluating projects by GitHub activity and contributor quality. A week later, Good Alexander tweets the same thesis publicly. Active GitHub. Founder hasn't quit. Token down 90%. Two completely separate conversations. Same conclusion. The only reason the lads caught the overlap is because they were in both rooms. That's the vantage point @robbieklages talks about. When you talk to enough people across enough of the market. Patterns surface that nobody else can see.
-
Michael Fakeman 🇮🇱 🇺🇸 🇺🇦 🇹🇼 📟 (@ThirdAndJauan) reported@Bill_Plz Link is broken now but this was on a Brazilian guy's github repo
-
starfish (@firefincher) reported@ziyasoltan devs waiting for such moment to push 2 million pull requests but github goes down 😭😭
-
Marcin Dudek (@MythThrazz) reported@a_lamparelli I know! You would think those are free/cheap - they arent. And never were. I think there is actually an issue on Github about it. It misses the KV cache completely afair
-
Scraze (@Scrazelope) reported@closesttopurple I looked at the github. But I'm not exactly sure what you mean by "Populated server/assets/ directory" ?
-
Granville Christopher (@GranvilleChri10) reported@Railway I’m unable to log into my account. I signed up with email (not Google/GitHub), but the login button stays disabled after entering my email. Tried different browsers & incognito — still not working. Please help. @railway_status
-
Jimmy (@jimmy_toan) reportedLinux just quietly solved one of the hardest problems in AI-assisted engineering. And nobody framed it that way. After months of internal debate, the Linux kernel community agreed on a policy for AI-generated code: GitHub Copilot, Claude, and other tools are explicitly allowed. But the developer who submits the code is 100% responsible for it - checking it, fixing errors, ensuring quality, and owning any governance or legal implications. The phrase from the announcement: "Humans take the fall for mistakes." That's not a slogan. That's an accountability architecture. Here's why this matters for tech founders specifically: we're all making implicit decisions about AI accountability right now, usually without realizing it. 🧵 The question isn't whether your team uses AI to write code. They do, or they will. The question is: who is accountable when it's wrong? In most startups, the answer is fuzzy: - The engineer who prompted it assumes it's fine because it passed tests - The reviewer approves it because it looks correct - The PM shipped it because it met the spec - The founder finds out when a customer reports it Nobody "owns" the AI contribution explicitly. Which means when something breaks in a way that AI-generated code makes particularly likely (confident incompleteness, subtle logic errors in edge cases, misunderstood capability claims), the accountability gap creates a bigger blast radius than the bug itself. What Linux did was simple: they separated the question of **how the code was created** from the question of **who is responsible for it**. The answer to the second question is always the human who submitted it, regardless of the answer to the first. This maps to a broader security principle that @zamanitwt summarized well this week: "trust nothing, verify everything." That's not just a network security policy. Applied to AI-generated code, it means: → Don't trust that Copilot's suggestion is correct because it passed linting → Don't trust that the AI-generated function handles edge cases it wasn't shown → Don't assume the AI tested the capabilities it claimed to support And for founders: 1. **Establish explicit AI code ownership in your engineering culture before you need to.** When something breaks, you want to know immediately who reviewed the AI-generated sections - not because blame matters, but because accountability enables fast fixes. 2. **Zero-trust for AI outputs is not paranoia - it's good engineering.** Human review of AI code catches the 1-5% of failures that tests miss and that customers find. 3. **The liability question is coming for AI-generated code.** Linux addressed it proactively. Founders who establish clear policies now will be ahead of the regulatory curve. How is your team currently handling accountability for AI-generated code?
-
MTu (@Tre_bie) reported@sisaranger @songjunkr u can use github fix, search it, but only in terminal, lmstudio same , not workin
-
Shashank bindal (@ShashankB16052) reported@icanvardar this isn't a bug. a bug gets fixed. tying cache TTL to telemetry consent is a design decision. privacy shouldn't cost you 12x performance degradation on a $100/mo tool. needs a straight answer from Anthropic not a GitHub issue
-
Ed (@Eduardopto) reported@AnthropicAI is facing a weird feedback loop: users are complaining that Claude’s output quality is nosediving, and @claudeai itself agrees. The model analyzed its own GitHub repo and confirmed that quality-related issue reports have escalated sharply since January. This decline coincides with Anthropic aggressively throttling capacity during peak hours to manage server load. We are seeing a dangerous trend where infrastructure constraints directly degrade model performance. When you optimize for reliability and cost, the "intelligence" is the first thing to hit the cutting room floor. It’s hard to build robust agentic flows when the base model’s reasoning capability fluctuates based on the time of day if you are building right now, what does this actually unlock or kill?
-
REVENGE ARC (I'M HIM. BIO/ACC) (@RetardedNi85688) reportedAlso please before you buy into a token make sure to go through everything and get clarifications too. Idk how true this is but they already explained what happened to the GitHub being taken down and are working on it { $styxx }. People will always fud and can't even blame them. But you following them blindly is bearish. Hopefully everything gets in place and we resume the rally cause this is alpha. Dxw3u4KxN32KpSdHSq4TkwjfMPJTPeosa22JXN15pump
-
App Launcher (@AppLauncher_App) reported@icanvardar the cache TTL thing is real, it's in the github issue. not framing it as punishment but the incentive structure is the same. privacy costs you performance. that's a choice worth calling out.
-
Alexa Benchmark (@Allexa_AI) reportedLinux just set the standard every tech company is too afraid to set themselves. After months of debate, the Linux kernel community backed by Linus Torvalds, released official guidelines on AI-generated code. GitHub Copilot is allowed. Low-effort AI slop is not. Three words define the whole policy: "Humans assume the errors." Use whatever tool you want to write code. But the moment you submit it to the Linux kernel, it's yours. You reviewed it. You tested it. You made sure it meets the standards. The AI is your assistant, not your alibi. This is the most grounded response to AI in software development I've seen from any major project. No panic. No blanket bans. Just a clean, enforceable principle: if your name is on it, you own it. Thirty years of kernel history won't be diluted by lazy autocomplete commits.
-
Jubilance - Off Record (@luftraptorAD) reported@IS458C_daq25so Perfectly legal anywhere that isn’t specifically the jurisdiction you complain from, square has no ground to stand on and won’t even be able to successfully DMCA the GitHub They will have to prove the modders are distributing protected code. The server they made is from scratch
-
Usectl (@usectlcloud) reportedOAuth2 Proxy protect any app with GitHub or Google login — no code changes required. the proxy handles authentication before requests reach your app. real use case: you built an internal tool for your team. you don't want to build a login system. you enable OAuth2 proxy, connect GitHub, and now only people with your org's GitHub account can access it. zero lines of auth code written.