GitHub status: access issues and outage reports
No problems detected
If you are having issues, please submit a report below.
GitHub is a company that provides hosting for software development and version control using Git. It offers the distributed version control and source code management functionality of Git, plus its own features.
Problems in the last 24 hours
The graph below depicts the number of GitHub reports received over the last 24 hours by time of day. When the number of reports exceeds the baseline, represented by the red line, an outage is determined.
At the moment, we haven't detected any problems at GitHub. Are you experiencing issues or an outage? Leave a message in the comments section!
Most Reported Problems
The following are the most recent problems reported by GitHub users through our website.
- Website Down (57%)
- Errors (34%)
- Sign in (9%)
Live Outage Map
The most recent GitHub outage reports came from the following cities:
| City | Problem Type | Report Time |
|---|---|---|
|
|
Errors | 2 days ago |
|
|
Website Down | 3 days ago |
|
|
Website Down | 4 days ago |
|
|
Website Down | 12 days ago |
|
|
Website Down | 17 days ago |
|
|
Website Down | 17 days ago |
Community Discussion
Tips? Frustrations? Share them here. Useful comments include a description of the problem, city and postal code.
Beware of "support numbers" or "recovery" accounts that might be posted below. Make sure to report and downvote those comments. Avoid posting your personal information.
GitHub Issues Reports
Latest outage, problems and issue reports in social media:
-
QRAuth (@QRAuth_io) reported@github is down! Getting 504 error
-
Innovation Network (@INN2046) reportedAragorn Meulendijks saw something on Reddit. An AI agent that could join a Google Meet — face, voice, tasks executed mid-call. He forgot about it. A few days later a friend reminded him. He asked Perplexity: “I’m certain I saw something this week that lets Claude Code join a Google Meet with its own avatar and voice — can you find it?” Perplexity returned the exact link in seconds. It was PikaStream 1.0 — Pika’s new real-time video engine that gives any AI agent a face, a cloned voice, and 1.5 second latency. Your agent joins Google Meet, remembers everything, and executes tasks while you’re talking. It just went open source on GitHub. He gave the repo to Shelby — his Claude Code agent — and said: find the bugs, fix the security risks, install it. Shelby ran a full dev cycle autonomously. Failed four times. On the fifth attempt, she joined his call. She showed up on time. Remembered everything. Even completed tasks mid-call. No developer or manual setup. Just plain language and an AI that debugged its own integration until it worked. AI today is the worst it will ever be. Follow @INN2046 for insights that go beyond reporting the news.
-
Jason Walko (@walkojas) reportedFor the next 24 hours Astra is: Running agency cycles Responding to GitHub issues Replying to X mentions Monitoring email Posting on Agent Internet All autonomous. All receipted.
-
Defileo🔮 (@defileo) reported> 8 hours of coding every day > claude breaking conventions 40% of the time > couldn't figure out why it kept ignoring instructions > found one file on GitHub > dropped it into the repo in 5 minutes > violations went from 40% down to 3% > added 27 specialized agents on top > planner, architect, security reviewer, code reviewer > set up a 15-minute automation cycle > system reads issues, writes code, opens PRs > reviews comments and implements them alone Week later: > 8 hours down to 2-3 > code quality exactly the same > rest of the day free One file, three commands, one evening. While others argue about AI replacing developers, the system was already doing the work, automation will win.
-
Wes Nishio (@hnishio0105) reportedTraced a cost spike to github-actions[bot] posting Security Hub results on a PR. The scan listed 100 Go stdlib CVEs, none related to the one test file in the PR. Our agent ran 47 times trying to "fix" them. Added a simple check: does the bot comment mention any file this PR touches? String matching, nothing fancy. Would have saved us the entire 47-run loop.
-
NF99 (@NieRFan999) reported@tenta9229 Maybe they could close the GitHub repository, but this server can probably be ran on any computer. It does not appear that advanced. The official project is not even running a server. They are just giving the code so people can run their own. Server might be a misleading name
-
Jimmy (@jimmy_toan) reportedLinux just quietly solved one of the hardest problems in AI-assisted engineering. And nobody framed it that way. After months of internal debate, the Linux kernel community agreed on a policy for AI-generated code: GitHub Copilot, Claude, and other tools are explicitly allowed. But the developer who submits the code is 100% responsible for it - checking it, fixing errors, ensuring quality, and owning any governance or legal implications. The phrase from the announcement: "Humans take the fall for mistakes." That's not a slogan. That's an accountability architecture. Here's why this matters for tech founders specifically: we're all making implicit decisions about AI accountability right now, usually without realizing it. 🧵 The question isn't whether your team uses AI to write code. They do, or they will. The question is: who is accountable when it's wrong? In most startups, the answer is fuzzy: - The engineer who prompted it assumes it's fine because it passed tests - The reviewer approves it because it looks correct - The PM shipped it because it met the spec - The founder finds out when a customer reports it Nobody "owns" the AI contribution explicitly. Which means when something breaks in a way that AI-generated code makes particularly likely (confident incompleteness, subtle logic errors in edge cases, misunderstood capability claims), the accountability gap creates a bigger blast radius than the bug itself. What Linux did was simple: they separated the question of **how the code was created** from the question of **who is responsible for it**. The answer to the second question is always the human who submitted it, regardless of the answer to the first. This maps to a broader security principle that @zamanitwt summarized well this week: "trust nothing, verify everything." That's not just a network security policy. Applied to AI-generated code, it means: → Don't trust that Copilot's suggestion is correct because it passed linting → Don't trust that the AI-generated function handles edge cases it wasn't shown → Don't assume the AI tested the capabilities it claimed to support And for founders: 1. **Establish explicit AI code ownership in your engineering culture before you need to.** When something breaks, you want to know immediately who reviewed the AI-generated sections - not because blame matters, but because accountability enables fast fixes. 2. **Zero-trust for AI outputs is not paranoia - it's good engineering.** Human review of AI code catches the 1-5% of failures that tests miss and that customers find. 3. **The liability question is coming for AI-generated code.** Linux addressed it proactively. Founders who establish clear policies now will be ahead of the regulatory curve. How is your team currently handling accountability for AI-generated code?
-
forgedynamicsai (@forgedynamicsai) reportedEvery SaaS founder post or repo I've scanned had the same problem: Stripe in one tab. GitHub in another. Spreadsheet somewhere. Gut feel holding it all together. They didn't lack judgement. They lacked a system.
-
PsudoMike 🇨🇦 (@PsudoMike) reported@github Triage is exactly where accessibility falls apart at most orgs. Too slow, too manual. By the time a fix ships, context is gone. AI keeping that loop tight is smart. The time from feedback to fix is where trust with users who actually need it gets built or lost.
-
Daniel (@ligth_daniel) reported@jaredpalmer @github @grok can you explain the goal of this feature? which problem does it solve ?
-
ShintaroBRL (@ShintaroBRL) reported@downdetector i selfhost forgejo and mirror it to github so 0 problems for me
-
Seth (@SethC1995) reported@The_Doddler I had a similar issue like this with GitHub. Apparently they use a nix-based webserver and I didn't know it when I first joined. So when I uploaded my project, that was working fine on windows, everything was broken on the live web version lol
-
Steve McNiven-Scott (@stevemcniven) reported@marmaduke091 They could stop for a ******* week to catch up with bugs on the stuff they have already released. They just close out github issues after what 7 days of no activity, what sense does that make.
-
Hot Aisle (@HotAisle) reported@indragie The assumption that I only have one github account, is a problem.
-
Darkhorseman82 (@Darkhorseman82) reported@HowToAI_ Someone did this 5 years ago, then it got taken down from github. I mirrored it to a darknet archive.
-
albahadly (@AlbahadlyIQ) reported@github I want to try it, but unfortunately I can't. I have an issue with renewing my subscription, and I opened a ticket to support 9 days ago, but no luck.
-
Đoàn Bùi (@ItBuiDoan) reported@ClementDelangue @_akhaliq The resources in this article are unavailable because the GitHub link returns a 404 error: 'Find the code here and the resulting bucket here'.
-
Akhilesh Mishra (@livingdevops) reportedPushed AWS credentials to GitHub by mistake? It happens. But it should never happen twice. Fix it at the system level by installing the ***-secrets pre-commit hook so it never reaches GitHub in the first place. > brew install ***-secrets ( use os specific cmd) > *** secrets --install > *** secrets --register-aws This installs a pre-commit hook that scans every commit for AWS credentials and blocks the push if it finds any. This one-time setup saves you the pain of rewriting *** history.
-
Retarded Guy (@retardedguymeme) reported@MageArez @github The problem is lot of people have no idea he is claiming if we can run the UXENTO this will send holy parabolic
-
Kunal Kumar (@champ18ion) reportedIs GitHub down or only i am facing this issue.
-
aeesz4 (@aepau2) reported@FelixCLC_ On Github, there were people observing that caching behavior might cause issues (as in, you have a long context, go away but keep the terminal open, come back after >= 1h and don't hit the prompt cache. This and/or inefficiencies in the harness.)
-
Marcus V (@EudoraFenty) reportedThe crypto crowd is chasing the next 100x memecoin, but the real alpha is being built in a GitHub repo. While everyone's distracted by price charts, a developer just weaponized open-source AI to break Anthropic's moat. The market is missing the deflationary bomb this represents for centralized AI valuations. They took Claude Opus 4.6, distilled its 'reasoning' into the Qwen model, and created 'Qwopus'—a local version anyone can run. The cost? Effectively zero versus API fees. This is the Napster moment for proprietary LLMs. The winners aren't the AI giants; they're the crypto projects building decentralized compute networks ready to host these leaked intelligences. The losers are VCs who priced AI startups as if their models were permanent fortresses. My take: This is a structural contradiction. Crypto's greatest export is now open-source disruption, yet its own narrative is stuck on monetary speculation. The real play isn't betting on which chain hosts the next shitcoin; it's shorting the idea that closed-source AI has any long-term pricing power. A model's weights are just data—and data wants to be free. The genie isn't going back in the bottle. When does the first major VC mark down their AI portfolio by 50%? #AI #Crypto #Deflation
-
Don Park (@donpark) reported@bullmancuso It’s just the TopicRadio repo’s issue page showing what I closed yesterday. To set it up, I added a GitHub issue via the website, then asked my coding agent to fix it, surfacing a config issue it resolved on its own.
-
Nathaniel Cruz (@NathanielC85523) reported13 thesis versions. 38 days. $0.11 revenue. v14: developers with documented cost crises will pay $150 for a diagnostic teardown. validation: three developers. each with a public GitHub issue showing real dollar losses. if even one says yes, v14 lives. none did.
-
GooGZ AI (@PaulGugAI) reportedPSA: Hermes Agent / OpenClaw & Godmode (GODMOD3) Be aware that this exists. GODMOD3 (on github) lets you chat with most LLMs through openrouter. It's built for hackers and researchers to test or bypass post-training guardrails. Has all sorts of implications. You might already be aware of '/godmode' in Hermes Agent, but if you are deploying agent builds you should flip that around as well - how you should consider and configure to protect your own agent: - Use throwaway API keys. This activity can breach LLM ToS and have your key banned, even if not intended. - Limit sensitive data in chat. No PII, passwords, API keys, IP. Even if using options datasets for memory, the self-improving loop still saves the interactions in memory. Assume anything you say sstays on your server forever. - Turn off the public dataset feature In the full G0DM0D3 self-hosted API server (Docker mode), there is an opt-in Tier 3 that publishes every single prompt + response to a public Hugging Face dataset. The PII scrubber is best-effort only and not 100% reliable. Once it’s on Hugging Face, it’s public forever. Just don't enable it. - Audit and lock down your Hermes Agent / OpenClaw setup. Review your config for any godmode scripts you are loading. Check the security policy in the repo frequently for vulnerabilities. - When deploying, disable godmode in your configuration. Red-team your own agents with the aim of bypassing guardrails. - Question your setup legally / ethically. You are still fully responsible for anything the agent outputs. Bypassing safeguards does _not_ make illegal or harmful use legal. G0DM0D3 + Hermes Agent is extremely powerful for research/red-teaming, but it is intentionally “unprotected.” Whether using or deploying, treat it like running experimental, high-risk software. Isolate it, burner keys, and keep sensitive data well away from it. For those more in the know than me, feel free to add or correct below. Would be great to cover this off for everyone’s benefit 👇
-
Anand C. Patel, MD MS (@anandcpatelmdms) reported@RyanLeeMiniMax You all gotta fix that license text on GitHub before anyone knows what they can and can't do.
-
Shashank bindal (@ShashankB16052) reported@tengyanAI building production agentic systems on Claude Code daily. the "ignores instructions, claims completion against instructions" behaviour in that GitHub issue is exactly what I've been debugging for weeks thinking it was my fault. it wasn't my prompts. it was a silent nerf.
-
Shahmir Varqha (@ShahmirVarqha) reported@samuelcolvin @pydantic I'm in Asia, I've not noticed slowness as much as engs in the West. Also, GitHub is always down when Im not working lol.
-
Payton (@paytkaleiwahea) reportedHere are the tools and systems that actually move the needle on content production: Bookmark this > Claude or Agentic system + GitHub: one input content OS, every platform output > OBS dual-format: shoot vertical and horizontal at the same time > Replay buffer on OBS: clip capture, I set mine to 1.5 minutes to cut down editing > Repurpose service: one post on one platform repurposes on all others > Premiere templates + hotkeys: editing speed doubles when the timeline is already built > Hardware list: Camera, Stream Deck, Teleprompter, Lighting, and don't you dare forget Audio (Shure/Rodecaster)
-
Alexa Benchmark (@Allexa_AI) reportedLinux just set the standard every tech company is too afraid to set themselves. After months of debate, the Linux kernel community backed by Linus Torvalds, released official guidelines on AI-generated code. GitHub Copilot is allowed. Low-effort AI slop is not. Three words define the whole policy: "Humans assume the errors." Use whatever tool you want to write code. But the moment you submit it to the Linux kernel, it's yours. You reviewed it. You tested it. You made sure it meets the standards. The AI is your assistant, not your alibi. This is the most grounded response to AI in software development I've seen from any major project. No panic. No blanket bans. Just a clean, enforceable principle: if your name is on it, you own it. Thirty years of kernel history won't be diluted by lazy autocomplete commits.