The worst part of all this is that GitHub's CTO and VP of Engineering sent out the usual "here's what we'll do to fix things" letter to their larger customers and, without exaggeration, it boiled down to: 1) "Here's a bunch of stuff we already did!" which... clearly isn't working, and 2) "We're continuing our Azure migration." also clearly not working.
So needless to say, if you depend on GitHub for critical business operations, you need to start thinking about what a world without GitHub looks like for your business and start working your way toward that. I know my confidence in GitHub's engineering leadership is at rock bottom.
I could sorta see a situation where the reality is "we're in the middle of a miserable transition and it'll clean up when we're done" but I don't think anyone has confidence that's all it is at this point.
Even that doesn’t really make sense to me, unless they’ve done it in a way where everything has to move at once.
Everywhere I’ve worked, if a migration is causing this much downtime then you kill the migration or slow it down. If every change has a 10% chance of bringing the site down, you only do a change every week or two until you can work out the kinks.
Here are some relevant excerpts from an October 2025 article[1]:
> In a message to GitHub’s staff, CTO Vladimir Fedorov notes that GitHub is constrained on capacity in its Virginia data center. “It’s existential for us to keep up with the demands of AI and Copilot, which are changing how people use GitHub,” he writes.
> The plan, he writes, is for GitHub to completely move out of its own data centers in 24 months. “This means we have 18 months to execute (with a 6 month buffer),” Fedorov’s memo says. He acknowledges that since any migration of this scope will have to run in parallel on both the new and old infrastructure for at least six months, the team realistically needs to get this work done in the next 12 months.
If you consider that six month parallel window to have started from the time of the October memo (written presumably at the start of October), then that puts us currently or past the point where they would have cut off their old DC and defaulted to Azure only.
Whether plans or timelines changed, I have no idea of course but the above does make for a convenient timeline that would explain the recent instability. Of course, it could also just be symptomatic of increased AI usage generally and the same problems might have surfaced at a software level regardless of whether they were in a DC or on Azure.
Putting that nuance aside, personally I like the idea that Azure is simply a giant pile of shit operated by a corporation with no taste.
>It’s existential for us to keep up with the demands of AI and Copilot
if by chance the CTO reads this, as a user of GitHub I would find it really existential if GitHub continues functioning as a reliable hub for git workflows (hence the name), and I have the strong suspicion nobody except for the shareholders gives a lick about copilot or 'AI' if it makes the core service the site was designed for unusable
Why? What is the correlation between profit and shareholder sentiment (besides the fact that shareholders want said profits)? They don't really influence the operation of the business meaningfully.
Sure, but I think it's the wrong way around. Appeasing shareholders doesn't make you profitable, being profitable appeases shareholders. I think there is a wealth of evidence that appeasing shareholders actually impedes profits overall.
Incorrect. They need to appease/trick/threaten/etc those that are paying for their services. Shareholders just demand they do so at the greatest (often short term) rate.
i heard that they asked LinkedIn to do this too and they either refused or their systems were too complex so they refused to. Maybe that explains why LI availability seems ok
It's starting to really look like the AI effect. It might be coincidence but I've noticed a lot more downtime and bad software lately. The last Nvidia drivers gave me a blue screen (last week or so), and speaking about Windows, I froze updates last year because it was clear they were introducing a bunch of issues with every update (not to mention unwanted features).
I like AI but actually not for coding because code quality is correlated to how well you understand the underlying systems you're building on, and AI is not really reasoning on this level at all. It's clearly synthesizing training data and it's useful in limited ways.
Interesting how many people "Like AI" because it's good at all the jobs other than the one they happen to make a living doing.
Did you hear about the screenwriters school in which the professors said to avoid AI for writing, but it's great for storyboards. And the storyboard school where the professors said the opposite?
The reality is that AI isn't actually "good" at anything. It produces passable ersatz facsimiles of work that can fool those not skilled in the art. The second reality of AI is that everyone is busy cramming it into their products at the expense of what their products are actually useful for.
Once people realise (1), and stop doing (2), the tech industry has a chance of recovering.
Yeah, I think I heard about that. Within certain domains it is certainly a useful tool. I would say things like online search are much nicer now (in that asking an AI is equivalent to searching online but it summarizes it for you). Online search fits the strengths of LLMs nicely, but right now it's being sold as a silver bullet, which it's not.
GitHub has been unreliable since before AI. Though it's definitely gotten far worse.
Seemingly the decline started with the Microsoft acquisition in 2018, and subsequent "unlimited private repository" change in 2019 (to match Gitlab's popular offer)
One example is the search being broken for CI logs. It takes over your browser's search hotkey too. What happens is every stage of the log is collapsed so the search doesn't work until you trigger the expansion but if you attempt to search before expanding the search will never work after it's been initialized. It's pretty infuriating when you're trying to find something in a giant build log.
I’m still baffled that Minecraft is doing so well, despite the whole Bedrock thing. At this point I think Microsoft just forgot that they bought Mojang.
Its had its fair share of outages and outrageous changes that overreach the bounds as well. Its more stable than github is but its had at least 2 sessions of downtime this year that I recall and they were both quite long (day length).
They'd lose a whole lot of users if they killed Java edition, since the modded community is so large. They'd quickly find one of the Minecraft clones reaching feature parity. And there's no good reason for it - it's not like Java is a threat anymore.
Exactly. So why isn't Microsoft doing just that? Isn't that how Microsoft usually handles things? Just look at Xbox. They essentially screwed up everything they could and then some.
They don’t enforce or even default to 2fa to change the account email. In addition, they have no process to get a human to reverse account takeovers. Just a web form that tells you to call a number that redirects you back to a web form
On the other hand, they aggressively log out legitimate users, and require the master Microsoft account password to log back in (because your kids need access to your one drive settings, etc).
I don't remember that happening so much (if ever) in, say, 2016. But the frequency of noticeable incidents seemingly has been rising steadily since around 2023. The Azure migration apparently only exacerbated it.
I remember it going down semi-regularly in the 2013+ era, and seeing HN posts about it. Especially if you were using a package manager reliant on GitHub like Cocoapods. It seems to me it is more "impactful" on the dev community now that they have gone past just being a centralized Git server for the team, to being the thing that does deploys and all sorts of other things.
I remember seeing unicorn daily and "webhook delivery delayed" weekly. I think it got better, but also they got more traffic, now millions of agents read files separately over and over again.
It was not nearly as bad... I remember our company migrating to github.com, and believe it or not, it was significant performance/uptime benefit over our self-hosted instance.
(And the first thing to go was occasional 500's on github-hosted files.. the core service itself - git, PR, actions - were pretty stable until recently)
And a ton of the top end ruby staff have left. Many of them ended up at shopify. There is a growing about of non ruby/rails code at github, but most of the system that people think of when they think github are ruby/rails.
I wonder what the average career tenure of the userbase here is now, because Github was slow and flaky well before Microsoft got involved.
Maybe it wasn't as noticeable when Github had less features, but our CI runners and other automation using the API a decade ago always had weekly issues caused by Github being down/degraded.
It hosts all the repositories backing applycreatures, we ran dozens of git projects on the same instance, have teams, you guys did a phenomenal work. I would say it's even easy to customise.
How much of this is due to Microsoft Culture of not innovating and buying leading companies with their revenue from windows/office and slowly destroying the aspects of companies that made them great in the first place?
Is all the recent GitHub downtime entirely attributable to GitHub AI Copilot related development? How hard can it be to reduce the blast radius of new AI features to not affect the core parts of hosting repositories? Because of Copilot everywhere, The UX has become bad and I had to click all over the place and on my profile to find repositories.
I do not care about much of it other than the git and API. I also sometimes use the Issues, although only with the API. But if it stops working sometimes, that is not too significantly a problem since the files can be sent after they start to work again; it does not have to be immediately.
Man, a while ago I thought: "It happens often, alright, but every 2 weeks? Sounds like a slight exaggeration." But it really is every 2 weeks, isn't it? If I imagine in a previous job anything production being down every 2 weeks ... phew, would have had to have a few hard talks and course corrections.
i once fixed a site going down several times a year with two t1.micro instances in the same region as the majority of traffic. Instantly solved the problem for what, $20/month?
Another site was constantly getting DDoS by Russians who were made we took down their scams on forums, that had to go through verisign back then, not sure who they're using now. They may have enough aggregate pipe it doesn't matter at this point
There are so many failures in microservices that just can't happen with a local binary. Inter-service communication over network is a big one with a failure rate orders of magnitude higher than running a binary on the same machine. Then you have to do deploys, monitoring, etc. across the whole platform.
You will basically need to employ solutions for problems only caused by your microservices arch. E.g. take reading the logs for a single request. In a monolith, just read the logs. For the many-service approach, you need to work out how you're going to correlate that request across them all.
Even the aforementioned network failures require a lot of design, and there's no standardization. Does the calling service retry? Does the callee have a durable queue and pick back up? What happens if a call/message gets 'too old'?
Also, from the other end, command line utils are typically made by entirely different people with entirely different philosophies/paradigms, so the encapsulation makes sense. That's not true when you're the one writing all the services, especially not at small-to-mid-size companies.
Plus, you already can do the single-concern thing in a monolith, just with modules/interfaces/etc.
One strategy to convince is to get someone less technical than you to sit by you while you try and trace everything from one error'd HTTP request from start to finish to diagnose the problem. If they see it takes half a day to check every call to every internal endpoint to 100% satisfy a particular request sometimes that can help.
Also sometimes they just think "this is a bunch of nerd stuff, why are you involving me?!" So it's not foolproof.
Oh, my non-technical boss agrees with me already. It's actually the engineers who've convinced themselves it's a good setup. Nice guys but very unwilling to change. Seems they're quite happy to have become 'experts' in this mess over the last 5-10 years. Almost like they're in retirement mode.
The real solution is probably to leave, but the market sucks at the moment. At least AI makes the 10-repos-per-tiny-feature thing easier.
Does anyone else ever think "that code I just pushed into my repo just took down all of github..." whenever it goes down around the same time you sync your changes?
Just moved a project of mine to Gitlab. Created this very simple component with codex that will keep a mirror updated on GitHub for me, so I can focus development on Gitlab.
I've been using "slopocalypse". People already know AI is responsible, but slop existed before — e.g. conventionally generated SEO spam. It's just... so much worse now.
> GitHub has recently seen more outages, in part because its central data center in Virginia is indeed resource-constrained and running into scaling issues. AI agents are part of the problem here. But it’s our understanding that some GitHub employees are concerned about this migration because GitHub’s MySQL clusters, which form the backbone of the service and run on bare metal servers, won’t easily make the move to Azure and lead to even more outages going forward.
Age-old lesson: change the tires on the moving vehicle that is your business when it's a Geo Metro, not when it's a freight train.
I'm sure the people with the purse strings didn't care, though, and just wanted to funnel the GH userbase into Azure until the wheels fell off, then write off the BU. Bought for $7.5B, it used to make $250M, but now makes $2B, so they could offload it make a profit. I wonder who'll buy it. Prob Google, Amazon, IBM, Oracle, or a hedge fund. They could choose not to sell it, but it'll end up a writeoff if the userbase jumps ship.
I assume this is all of the pains of going from "GHA is sorta kinda on Azure", which was a bad state, to "GHA is going full Azure", which is a painful state to get to but presumably simplifies things.
> Any massive infra migration is going to cause issues.
What? No, no it's not. The entire discipline of Infrastructure and Systems engineering are dedicated to doing these sorts of things. There are well-worn paths to making stable changes. I've done a dozen massive infrastructure migrations, some at companies bigger than Github, and I've never once come close to this sort of instability.
This is a botched infrastructure migration, onto a frankly inferior platform, not something that just happens to everyone.
I remember back in the early Windows XP era when things got so bad that Microsoft basically had to make a hard pivot towards security and reliability.
I think they may need to do that once again. Almost every product of theirs feels like a dumpster fire. GitHub is down constantly, Windows 11 is a nightmare and instead of patching things they're adding stupid features nobody asked for. I think they need to stop and really look closely at what they're prioritizing.
I'm surprised nobody has tried to throw together a commercial alternative to GitHub. 50% of it is available as FOSS, the other 50% you can vibecode in a month (you can vibecode reliably, Microsoft/Google just suck at it). Afaict, reason we all keep using GitHub is it has a million features and isn't as ugly, difficult and slow as GitLab. (sorry GitLab, I love your handbook, hate your UX)
I remember. My GitHub user ID is #5907, account created 2008-04-08T20:27:36Z. I think it is inevitable that all good things come to an end, but it's still a bummer to see.
I've been sitting here waiting for a critical deploy to happen via GitHub Actions (I know, hour fault, we should have left ages ago). My patience for this bullshit is gone, I'm going to be pushing very hard to get us off of GitHub entirely except for public code mirrors going forward.
Edit: oh look, their site says all good, but I still have jobs stuck. What a pile of garbage.
So am I the only one thinking that maybe GitHub is succumbing to the weight of AI slop that's coming in from all the vibecoding, clawbots, and other AI workflows?
Why don't companies with chronic outages mimic their stack from top to bottom (i.e. starting with a new domain), then before making a change, make the change on the duplicate stack and blast it with mock requests.
Might catch 90% of problems before they make it into the real stack?
E.g. every step of GitHub's migration to Azure could be mimicked on the duplicate stack before it's implemented on the primary stack. Is this just considered too much work? (I doubt cost would be the issue, because even if it costs millions, it would pay for itself in reduced reputational damage from outages).
EDIT: downvotes - why? - I think this is a good idea (I'd do it for my sites if outages were an issue).
Downvotes are probably because that is what companies without chronic outages do.
If you'd ever worked on a codebase as terrible as I imagine GH's internals are and looked at the git history, you'd find two things:
1) fixing it would require rolling back 100's-1000's of engineer-years of idiocy that make things like testing or refactoring untenable
2) many prior engineers got part of the way through such improvements before leaving or being kicked out. Their efforts mostly just made it worse, because now you never know what sort of terribleness to expect when you open an unfamiliar file.
> EDIT: downvotes - why? - I think this is a good idea (I'd do it for my sites if outages were an issue).
Because that's a monumental amount of work, and extraordinarily difficult to retrofit into a system that wasn't initially designed that way. Not to mention the unstated requirement of mirroring traffic to actually exercise that system (given the tendency of bugs to not show up until something actually uses the system).
Agree, but look at the alternative; GitHub is constantly being savaged by users who (quite reasonably) expect uptime. Ignoring impacts on morale and reputation, damage to their bottom line alone might tens (hundreds?) of millions per year.
> mirroring traffic
yeah, I agree that's difficult, but it need to not be exact to still be useful.
Github CEO must be on HN, right? If so, any comments?
They have not even bothered to implement entra login when they have their competitors login for years, do they even know what their product is? Or are you just a middle man for slop?
GitHub goes down at least once a week as I said before. [0] thanks to Copilot, Tay.ai and Zoe chatbots wrecking the platform instead of humans maintaining it.
If there was a prediction market for when GitHub experiences an outage every week, then you would make a lot of money.
>GitHub goes down at least once a week as I said before. [0] thanks to Copilot, Tay.ai and Zoe chatbots wrecking the platform instead of humans.
there are tens of thousands of stupid scripts hosted on github itself that have scheduled progmatic pushes or pulls to repos via cron jobs with millions and millions of users -- yeah LLMs accelerate the fire but let's not pretend that GH was some bastion of real-user-dom somehow at some point.
Sorry, I realise this comment isn't up to HN's usual standards for thoughtfulness and it is perhaps a bit inflammatory but... look, I'd bet the majority of us on this site rely on GitHub and I can't be the only one becoming incredibly frustrated with its recent unreliability[0]?
(And, yes, I did enough basic data analysis to confirm that it IS indeed getting worse versus a year, two years, and three years ago, and is particularly bad since the start of this year.)
[0] EDIT: clearly not from looking at the rest of the comments in this discussion.
@KaiserPro has pasted the link to someone else's heatmap, which is really good. Mine was just an Excel spreadsheet with a graph that I'd intended to write a blog about but then got demotivated on because I was too busy with other things and I saw that heatmap as well. Maybe I will do a proper write up next time GitHub has an outage and I'm blocked by it.
The worst part of all this is that GitHub's CTO and VP of Engineering sent out the usual "here's what we'll do to fix things" letter to their larger customers and, without exaggeration, it boiled down to: 1) "Here's a bunch of stuff we already did!" which... clearly isn't working, and 2) "We're continuing our Azure migration." also clearly not working.
So needless to say, if you depend on GitHub for critical business operations, you need to start thinking about what a world without GitHub looks like for your business and start working your way toward that. I know my confidence in GitHub's engineering leadership is at rock bottom.
I could sorta see a situation where the reality is "we're in the middle of a miserable transition and it'll clean up when we're done" but I don't think anyone has confidence that's all it is at this point.
Even that doesn’t really make sense to me, unless they’ve done it in a way where everything has to move at once.
Everywhere I’ve worked, if a migration is causing this much downtime then you kill the migration or slow it down. If every change has a 10% chance of bringing the site down, you only do a change every week or two until you can work out the kinks.
...or you keep fighting forward with the migration, because if it's seen as a failure then some pretty big heads will have to roll...
I mean, they are seemingly breaking every week or two so that might be what they are doing.
also it should be noted that LinkedIn had a 5 year plan of migrating everything to azure but abandoned it after a year.
ooooh, they're migrating to Azure, now everything makes sense.
Here are some relevant excerpts from an October 2025 article[1]:
> In a message to GitHub’s staff, CTO Vladimir Fedorov notes that GitHub is constrained on capacity in its Virginia data center. “It’s existential for us to keep up with the demands of AI and Copilot, which are changing how people use GitHub,” he writes.
> The plan, he writes, is for GitHub to completely move out of its own data centers in 24 months. “This means we have 18 months to execute (with a 6 month buffer),” Fedorov’s memo says. He acknowledges that since any migration of this scope will have to run in parallel on both the new and old infrastructure for at least six months, the team realistically needs to get this work done in the next 12 months.
If you consider that six month parallel window to have started from the time of the October memo (written presumably at the start of October), then that puts us currently or past the point where they would have cut off their old DC and defaulted to Azure only.
Whether plans or timelines changed, I have no idea of course but the above does make for a convenient timeline that would explain the recent instability. Of course, it could also just be symptomatic of increased AI usage generally and the same problems might have surfaced at a software level regardless of whether they were in a DC or on Azure.
Putting that nuance aside, personally I like the idea that Azure is simply a giant pile of shit operated by a corporation with no taste.
[1]: https://thenewstack.io/github-will-prioritize-migrating-to-a...
>It’s existential for us to keep up with the demands of AI and Copilot
if by chance the CTO reads this, as a user of GitHub I would find it really existential if GitHub continues functioning as a reliable hub for git workflows (hence the name), and I have the strong suspicion nobody except for the shareholders gives a lick about copilot or 'AI' if it makes the core service the site was designed for unusable
AI and Copilot increase the load on git workflows.
For GitHub to remain profitable they have to appease those shareholders you mentioned.
Why? What is the correlation between profit and shareholder sentiment (besides the fact that shareholders want said profits)? They don't really influence the operation of the business meaningfully.
Growth chart gotta go up. Only chumps run a business that makes a steady return.
Sure, but I think it's the wrong way around. Appeasing shareholders doesn't make you profitable, being profitable appeases shareholders. I think there is a wealth of evidence that appeasing shareholders actually impedes profits overall.
Incorrect. They need to appease/trick/threaten/etc those that are paying for their services. Shareholders just demand they do so at the greatest (often short term) rate.
yeah currently working with Azure. what a PITA.
I wonder if the extended downtime is just due to the on-call engineers waiting for their azure auth tokens to refresh within azure's own damn network.
i heard that they asked LinkedIn to do this too and they either refused or their systems were too complex so they refused to. Maybe that explains why LI availability seems ok
Azure, the color of BSOD
they're not just migrating to Azure, they're vibrating to Azure!
Vibe coding and it's consequences.
"The evidence is clear: Either you embrace AI, or get out of this career." -Github CEO
"Sooner than later, 80% of the code is going to be written by Copilot. And that doesn’t mean the developer is going to be replaced." -Github CEO
There’s plenty of alternatives, but people continue to stay. Therefore, it’s not as bad as you think it is.
I second this. I'm done.
Is "migrating to Azure" the new "migrating to SAP?"
That’s not for to … SAP.
It's starting to really look like the AI effect. It might be coincidence but I've noticed a lot more downtime and bad software lately. The last Nvidia drivers gave me a blue screen (last week or so), and speaking about Windows, I froze updates last year because it was clear they were introducing a bunch of issues with every update (not to mention unwanted features).
I like AI but actually not for coding because code quality is correlated to how well you understand the underlying systems you're building on, and AI is not really reasoning on this level at all. It's clearly synthesizing training data and it's useful in limited ways.
Interesting how many people "Like AI" because it's good at all the jobs other than the one they happen to make a living doing.
Did you hear about the screenwriters school in which the professors said to avoid AI for writing, but it's great for storyboards. And the storyboard school where the professors said the opposite?
The reality is that AI isn't actually "good" at anything. It produces passable ersatz facsimiles of work that can fool those not skilled in the art. The second reality of AI is that everyone is busy cramming it into their products at the expense of what their products are actually useful for.
Once people realise (1), and stop doing (2), the tech industry has a chance of recovering.
Yeah, I think I heard about that. Within certain domains it is certainly a useful tool. I would say things like online search are much nicer now (in that asking an AI is equivalent to searching online but it summarizes it for you). Online search fits the strengths of LLMs nicely, but right now it's being sold as a silver bullet, which it's not.
It's a version of the Gell-Mann Amnesia effect.
GitHub has been unreliable since before AI. Though it's definitely gotten far worse.
Seemingly the decline started with the Microsoft acquisition in 2018, and subsequent "unlimited private repository" change in 2019 (to match Gitlab's popular offer)
One example is the search being broken for CI logs. It takes over your browser's search hotkey too. What happens is every stage of the log is collapsed so the search doesn't work until you trigger the expansion but if you attempt to search before expanding the search will never work after it's been initialized. It's pretty infuriating when you're trying to find something in a giant build log.
I bet on rushed Azure migration. A lot can go wrong it devops.
We had more and more outages on EVERYTHING since AI. Not sure what they're doing.
I think maybe it's not that GitHub is using AI, but that the amount of AI slop going into GitHub may be more than they expected.
Productivity is finite. If you pivot entirely to the AI stack, you're going to lose bandwidth for everything else. It's an opportunity cost problem.
Took a full 8 years for a Microsoft acquisition to go to shit, which is probably a record. Kudos to the Github team for holding out this long.
I’m still baffled that Minecraft is doing so well, despite the whole Bedrock thing. At this point I think Microsoft just forgot that they bought Mojang.
Its had its fair share of outages and outrageous changes that overreach the bounds as well. Its more stable than github is but its had at least 2 sessions of downtime this year that I recall and they were both quite long (day length).
They'd lose a whole lot of users if they killed Java edition, since the modded community is so large. They'd quickly find one of the Minecraft clones reaching feature parity. And there's no good reason for it - it's not like Java is a threat anymore.
Exactly. So why isn't Microsoft doing just that? Isn't that how Microsoft usually handles things? Just look at Xbox. They essentially screwed up everything they could and then some.
They've made it borderline impossible to log in, especially for children.
I just use an offline server, so I wouldn't notice if they had GitHub levels of availability.
They don’t enforce or even default to 2fa to change the account email. In addition, they have no process to get a human to reverse account takeovers. Just a web form that tells you to call a number that redirects you back to a web form
On the other hand, they aggressively log out legitimate users, and require the master Microsoft account password to log back in (because your kids need access to your one drive settings, etc).
If I remember correctly UK players can no longer chat at all until they verify their ID.
Minecraft is a trick up their sleeve yet to be used. Manipulate and indoctrinate the youth.
Indoctrinated by cubic cows
GitHub has always been incredibly outage riddled no? This is not a MSFT thing
I don't remember that happening so much (if ever) in, say, 2016. But the frequency of noticeable incidents seemingly has been rising steadily since around 2023. The Azure migration apparently only exacerbated it.
I remember it going down semi-regularly in the 2013+ era, and seeing HN posts about it. Especially if you were using a package manager reliant on GitHub like Cocoapods. It seems to me it is more "impactful" on the dev community now that they have gone past just being a centralized Git server for the team, to being the thing that does deploys and all sorts of other things.
Circa 2019, my office had a bell that we would ring whenever GitHub had an outage, and it was rung several times per week.
I remember seeing unicorn daily and "webhook delivery delayed" weekly. I think it got better, but also they got more traffic, now millions of agents read files separately over and over again.
IMO it's much better now.
It was not nearly as bad... I remember our company migrating to github.com, and believe it or not, it was significant performance/uptime benefit over our self-hosted instance.
(And the first thing to go was occasional 500's on github-hosted files.. the core service itself - git, PR, actions - were pretty stable until recently)
I think having a single 9 of uptime is a relatively new thing even for GH
How fast was Skype?
MSFT acquired Skype in 2011, so I would say only a few months:
https://trends.google.com/trends/explore?date=all&geo=GB&q=s...
To be fair a bunch of this is because the CEO after Nat Friedman (Thomas Dohmke) was pushed out in August 25.
And a ton of the top end ruby staff have left. Many of them ended up at shopify. There is a growing about of non ruby/rails code at github, but most of the system that people think of when they think github are ruby/rails.
Shopify is on the AI-everything train as well, we'll see how that goes.
Who was also the last CEO, right? Is this a coincidence?
Every day more Microsofty...they should rename to "Your Repository Needs To Restart To Apply Updates"
It's now safe to turn off your expectations.
I wonder what the average career tenure of the userbase here is now, because Github was slow and flaky well before Microsoft got involved.
Maybe it wasn't as noticeable when Github had less features, but our CI runners and other automation using the API a decade ago always had weekly issues caused by Github being down/degraded.
The best stretch Github ever had was post-acquisition when Nat Friedman as CEO.
"It looks like you're trying to develop some software.
Would you like help?
- Get help with developing the software
- Just develop the software without help
[ ] Don't show me this tip again"
If only their UI were that good!
Would you like to setups repository backups with OneDrive?
Lol, someone should make a pre -commit hook that reboots your computer with a message like this!
Just wait until github comes up with an outage tuesday.
Down? No sir we are not down. There are elevated error rates and degraded performance.
The update to .NET framework went badly and we need to reinstall Windows.
An isolated group of customers are experiencing elevated error rates and degraded performance.
FTFY. (I've read AWS word it like that)
Technucally a set is a subset of itself.
I'm just going to stand by until Microsoft is back in everyone's good graces again by releasing some oss software that we all swoon over
Is that a cyclical thing? I went through the VS code+typescript good graces era but I didn't know there were previous cycles.
Not that I remember. VS Code was a surprising turnabout for a company that was both adamantly closed and threw FUD around like monkeys in a zoo.
You've been here before!
I'm glad I moved over to forgejo. Being selfhosted, the UI loads faster. Most importantly, the thing is always responsive.
As a developer working on Forgejo -- glad you like it!
It hosts all the repositories backing applycreatures, we ran dozens of git projects on the same instance, have teams, you guys did a phenomenal work. I would say it's even easy to customise.
https://foja.applycreatures.com
Edit: it has a wonderful API so I posted the link it may tempt some to ditch MS/Azure hub.
How much of this is due to Microsoft Culture of not innovating and buying leading companies with their revenue from windows/office and slowly destroying the aspects of companies that made them great in the first place?
Is all the recent GitHub downtime entirely attributable to GitHub AI Copilot related development? How hard can it be to reduce the blast radius of new AI features to not affect the core parts of hosting repositories? Because of Copilot everywhere, The UX has become bad and I had to click all over the place and on my profile to find repositories.
They invented the perfect solution to stop supply chain attacks.
Credential stealers hate this one trick!
I do not care about much of it other than the git and API. I also sometimes use the Issues, although only with the API. But if it stops working sometimes, that is not too significantly a problem since the files can be sent after they start to work again; it does not have to be immediately.
Microslop at it again
Oh, they've gone beyond micro. It's Macroslop now!
Please don't use that term; it makes them look bad :p /s
Man, a while ago I thought: "It happens often, alright, but every 2 weeks? Sounds like a slight exaggeration." But it really is every 2 weeks, isn't it? If I imagine in a previous job anything production being down every 2 weeks ... phew, would have had to have a few hard talks and course corrections.
i once fixed a site going down several times a year with two t1.micro instances in the same region as the majority of traffic. Instantly solved the problem for what, $20/month?
Another site was constantly getting DDoS by Russians who were made we took down their scams on forums, that had to go through verisign back then, not sure who they're using now. They may have enough aggregate pipe it doesn't matter at this point
Do your part; remind people that Github is not git. Git is decentralizable and people should know this.
At this rate it will be a matter of time before a "Github is up" parody site reaches the top of HN
github is at one nine, basically: https://news.ycombinator.com/item?id=47428035
I once worked at a place with more micro services than engineers. We joked about "we have as many 8s of uptime as you need!"
> I once worked at a place with more micro services than engineers.
Currently consulting somwhere with 30 services per engineer. I cannot convince them this is hell. Maybe that makes it my personal hell.
"Its like family here!"
In that every night you're playing murder mystery, and its never fun.
I would never trust my family with system design either.
as a person that never touched webdev, I have a question
how is such service spam different from unix "small functions that do one thing only" culture?
why in unix case it is usually/historically seen as nice, while in web case it makes stuff worse?
There are so many failures in microservices that just can't happen with a local binary. Inter-service communication over network is a big one with a failure rate orders of magnitude higher than running a binary on the same machine. Then you have to do deploys, monitoring, etc. across the whole platform.
You will basically need to employ solutions for problems only caused by your microservices arch. E.g. take reading the logs for a single request. In a monolith, just read the logs. For the many-service approach, you need to work out how you're going to correlate that request across them all.
Even the aforementioned network failures require a lot of design, and there's no standardization. Does the calling service retry? Does the callee have a durable queue and pick back up? What happens if a call/message gets 'too old'?
Also, from the other end, command line utils are typically made by entirely different people with entirely different philosophies/paradigms, so the encapsulation makes sense. That's not true when you're the one writing all the services, especially not at small-to-mid-size companies.
Plus, you already can do the single-concern thing in a monolith, just with modules/interfaces/etc.
Oooof that's rough.
One strategy to convince is to get someone less technical than you to sit by you while you try and trace everything from one error'd HTTP request from start to finish to diagnose the problem. If they see it takes half a day to check every call to every internal endpoint to 100% satisfy a particular request sometimes that can help.
Also sometimes they just think "this is a bunch of nerd stuff, why are you involving me?!" So it's not foolproof.
Oh, my non-technical boss agrees with me already. It's actually the engineers who've convinced themselves it's a good setup. Nice guys but very unwilling to change. Seems they're quite happy to have become 'experts' in this mess over the last 5-10 years. Almost like they're in retirement mode.
The real solution is probably to leave, but the market sucks at the moment. At least AI makes the 10-repos-per-tiny-feature thing easier.
We pride ourself on 9 5s!
seven nines? That's nothing , bro we got twelve eights!
I have Royal flush :)
From five nines to nine fives
9% ? /s (though To be honest I genuinely wouldn't be surprised if things go down so bad too at this point either)
Unironically, I think 9% uptime would be "one-tenth of a nine".
Are you saying 9.999% isn’t four nines?
Can’t tell if this is intended as humor, but I LOL’ed.
It unarguably is.
90% would be one 9 following the sequence back.
99.99
99.90
99.00
90.00
Does anyone else ever think "that code I just pushed into my repo just took down all of github..." whenever it goes down around the same time you sync your changes?
Just moved a project of mine to Gitlab. Created this very simple component with codex that will keep a mirror updated on GitHub for me, so I can focus development on Gitlab.
https://gitlab.com/gabriel.chamon/ci-components/-/tree/main/...
And this is why I self host a lot of my Git stack with Gerrit...
Or just make sure you git fetch repos into $other-place.
That helps with Git not so much issues etc.
Yeah, I think especially Git mirrors can go a long too for maintaining availability and also for reducing load off main infra.
Hitting 500s when trying to push branches and create PRs.
Microslop
What has changed at GitHub to cause this?
AIpocalypse. Eaten too much Copilot dog food.
Perhaps even AIslopalypse.
Looking at the status, its not one long outage, but lots of little ones, microslops if you will.
I've been using "slopocalypse". People already know AI is responsible, but slop existed before — e.g. conventionally generated SEO spam. It's just... so much worse now.
"Slopocalypse": yeah, I like that. Easier to pronounce too.
At any rate, it seems like GitHub is back up now, so we'll see how long that lasts.
Weird Al needs to capitalize on this whole AI/Al thing
Possibly a combination of moving infrastructure to Azure, and also a significant increase in the number of PRs and commits due to Vibe-coding?
Perhaps staff cuts having longtails? https://www.itpro.com/software/microsoft/microsoft-layoffs-h...
Microsoft Makes AI Mandatory For Employees
https://www.forbes.com/sites/bernardmarr/2025/07/08/microsof...
Azure
> Azure
To explain this one-word comment for those unfamiliar, see previously:
GitHub will prioritize migrating to Azure over feature development (5 months ago) https://news.ycombinator.com/item?id=45517173
In particular:
> GitHub has recently seen more outages, in part because its central data center in Virginia is indeed resource-constrained and running into scaling issues. AI agents are part of the problem here. But it’s our understanding that some GitHub employees are concerned about this migration because GitHub’s MySQL clusters, which form the backbone of the service and run on bare metal servers, won’t easily make the move to Azure and lead to even more outages going forward.
Age-old lesson: change the tires on the moving vehicle that is your business when it's a Geo Metro, not when it's a freight train.
I'm sure the people with the purse strings didn't care, though, and just wanted to funnel the GH userbase into Azure until the wheels fell off, then write off the BU. Bought for $7.5B, it used to make $250M, but now makes $2B, so they could offload it make a profit. I wonder who'll buy it. Prob Google, Amazon, IBM, Oracle, or a hedge fund. They could choose not to sell it, but it'll end up a writeoff if the userbase jumps ship.
Vibe coding features.
I assume this is all of the pains of going from "GHA is sorta kinda on Azure", which was a bad state, to "GHA is going full Azure", which is a painful state to get to but presumably simplifies things.
You never go full Azure
Their primary goal in the last year was to move to Azure. Any massive infra migration is going to cause issues.
> Any massive infra migration is going to cause issues.
What? No, no it's not. The entire discipline of Infrastructure and Systems engineering are dedicated to doing these sorts of things. There are well-worn paths to making stable changes. I've done a dozen massive infrastructure migrations, some at companies bigger than Github, and I've never once come close to this sort of instability.
This is a botched infrastructure migration, onto a frankly inferior platform, not something that just happens to everyone.
A.I. but that acronym can mean a number of things.
Artificial intelligence, Azure integration, many other things.
Senior engineers/leaders getting tired of Microsoft's shit and leaving.
looking forward to the `addressing-githubs-recent-availability-issues-3` news post
I remember back in the early Windows XP era when things got so bad that Microsoft basically had to make a hard pivot towards security and reliability.
I think they may need to do that once again. Almost every product of theirs feels like a dumpster fire. GitHub is down constantly, Windows 11 is a nightmare and instead of patching things they're adding stupid features nobody asked for. I think they need to stop and really look closely at what they're prioritizing.
Setting up a Gitea instance is approachable, especially with agent assistance.
https://about.gitea.com/
I guess vibe coding can't solve such problem for now...
I'm surprised nobody has tried to throw together a commercial alternative to GitHub. 50% of it is available as FOSS, the other 50% you can vibecode in a month (you can vibecode reliably, Microsoft/Google just suck at it). Afaict, reason we all keep using GitHub is it has a million features and isn't as ugly, difficult and slow as GitLab. (sorry GitLab, I love your handbook, hate your UX)
Remember when GitHub was cool? Pepperidge Farm remembers.
I remember. My GitHub user ID is #5907, account created 2008-04-08T20:27:36Z. I think it is inevitable that all good things come to an end, but it's still a bummer to see.
As do I. Mine is even earlier as well!
Anyone else notice other Microsoft cloud services ( for instance inside azure ) with bad performance also?
I can’t be specific but we are constantly complaining.
No? Azures been rock solid for us.
Front door did have a major outage last year.
The Azure management UI, yes, so slowww. But the services (VMs etc) have been good.
What's the successor?
https://radicle.xyz/
Forgejo
I sure hope they created a restore point first.
Would be easier for them to just tell us when it’s up these days
GH is ripe for a disruptor right now.
That Go rewrite continues to rear its head, eh?
Can this downtime be quantified to actual monetary losses?
I've been sitting here waiting for a critical deploy to happen via GitHub Actions (I know, hour fault, we should have left ages ago). My patience for this bullshit is gone, I'm going to be pushing very hard to get us off of GitHub entirely except for public code mirrors going forward.
Edit: oh look, their site says all good, but I still have jobs stuck. What a pile of garbage.
I'm so sick of this.
Moved to self hosted crowci instead of actions.
Must be Tuesday then
So am I the only one thinking that maybe GitHub is succumbing to the weight of AI slop that's coming in from all the vibecoding, clawbots, and other AI workflows?
Microsoft products are so human, they stop working weekly as if they're observing some sort of sabbath ...
Why don't companies with chronic outages mimic their stack from top to bottom (i.e. starting with a new domain), then before making a change, make the change on the duplicate stack and blast it with mock requests.
Might catch 90% of problems before they make it into the real stack?
E.g. every step of GitHub's migration to Azure could be mimicked on the duplicate stack before it's implemented on the primary stack. Is this just considered too much work? (I doubt cost would be the issue, because even if it costs millions, it would pay for itself in reduced reputational damage from outages).
EDIT: downvotes - why? - I think this is a good idea (I'd do it for my sites if outages were an issue).
Downvotes are probably because that is what companies without chronic outages do.
If you'd ever worked on a codebase as terrible as I imagine GH's internals are and looked at the git history, you'd find two things:
1) fixing it would require rolling back 100's-1000's of engineer-years of idiocy that make things like testing or refactoring untenable
2) many prior engineers got part of the way through such improvements before leaving or being kicked out. Their efforts mostly just made it worse, because now you never know what sort of terribleness to expect when you open an unfamiliar file.
> EDIT: downvotes - why? - I think this is a good idea (I'd do it for my sites if outages were an issue).
Because that's a monumental amount of work, and extraordinarily difficult to retrofit into a system that wasn't initially designed that way. Not to mention the unstated requirement of mirroring traffic to actually exercise that system (given the tendency of bugs to not show up until something actually uses the system).
> that's a monumental amount of work
Agree, but look at the alternative; GitHub is constantly being savaged by users who (quite reasonably) expect uptime. Ignoring impacts on morale and reputation, damage to their bottom line alone might tens (hundreds?) of millions per year.
> mirroring traffic
yeah, I agree that's difficult, but it need to not be exact to still be useful.
Testing? Who needs it when you have Copilot!
Github CEO must be on HN, right? If so, any comments?
They have not even bothered to implement entra login when they have their competitors login for years, do they even know what their product is? Or are you just a middle man for slop?
Vibe check?
That's just like your Vibe man; can you just copilot your wayout of these problems?
Did MS finish the Hotmail transition?
Does github not do any kind of blue/green rollouts or what
I'm going to blame Claude Code!
My vscode slop session stopped in between. Maybe it's for the better
"the food is terrible and such small portions!"
At least it happened after I did my work for the day... jfc!
For fucks sake.
I've been considering it for a while, but I'm definitely now pitching a move away from GitHub at our organization.
h8 github so much. ahhhhhh
You have GOT to be kidding me.
No, got to be kidding me day is next week.
GitHub goes down at least once a week as I said before. [0] thanks to Copilot, Tay.ai and Zoe chatbots wrecking the platform instead of humans maintaining it.
If there was a prediction market for when GitHub experiences an outage every week, then you would make a lot of money.
[0] https://news.ycombinator.com/item?id=47487881
>GitHub goes down at least once a week as I said before. [0] thanks to Copilot, Tay.ai and Zoe chatbots wrecking the platform instead of humans.
there are tens of thousands of stupid scripts hosted on github itself that have scheduled progmatic pushes or pulls to repos via cron jobs with millions and millions of users -- yeah LLMs accelerate the fire but let's not pretend that GH was some bastion of real-user-dom somehow at some point.
Vegas should start taking bets
Fuck sake. Again?
Sorry, I realise this comment isn't up to HN's usual standards for thoughtfulness and it is perhaps a bit inflammatory but... look, I'd bet the majority of us on this site rely on GitHub and I can't be the only one becoming incredibly frustrated with its recent unreliability[0]?
(And, yes, I did enough basic data analysis to confirm that it IS indeed getting worse versus a year, two years, and three years ago, and is particularly bad since the start of this year.)
[0] EDIT: clearly not from looking at the rest of the comments in this discussion.
> I realise this comment isn't up to HN's usual standards for thoughtfulness
> And, yes, I did enough basic data analysis to confirm
Perhaps you'd consider showing us that analysis? That sounds like it would make a pretty substantive, thoughtful comment.
> consider showing us
Gaze upon the tapestry in which github paints it's failure with a thin copper red thread:
https://www.githubstatus.com/
@KaiserPro has pasted the link to someone else's heatmap, which is really good. Mine was just an Excel spreadsheet with a graph that I'd intended to write a blog about but then got demotivated on because I was too busy with other things and I saw that heatmap as well. Maybe I will do a proper write up next time GitHub has an outage and I'm blocked by it.