I’m certain I’m up there in the 1% of users, or close to it, that are writing software daily in terms of consistent prolonged volume of work and work that is actually used by others over the past nearly 20 years based on user activity statistics I’ve collected.
I, too, am a fairly, but not immediate early user of GitHub. Despite GitHub’s poor metrics, I am still shipping, because writing software doesn’t require GitHub.
Hashimoto’s comments sound disturbed and I hope he finds some peace, but if he wasn’t who he was and you read these comments, you’d think this person had a problem. So, I think he does.
> I, too, am a fairly, but not immediate early user of GitHub. Despite GitHub’s poor metrics, I am still shipping, because writing software doesn’t require GitHub.
GitHub’s downtime is a problem for issue tracking, PR merging, contributing and reviewing PRs, and more.
Your exact point was already pre-addressed in the blog post because it’s so predictable that some would completely miss the point. GitHub’s downtime isn’t about stopping you from writing code on your own machine.
Nothing is pissing me off more than GitHub's stability going down the tubes RIGHT as work is migrating everything, and I mean everything, from CircleCI to GH.
The wildest thing is that Azure Repos/Pipelines was better than this.
Their one caveat is also that they are still migrating it to Azure infra, so it's possible that's still in a one foot in one foot out kinda scenario, from what I've heard. But, this isn't inspiring confidence.
Could be. But 99% of the repos are static garbage with no PR nor actions.
They mentioned they have some elasticsearch reindexing going to, I would guess they needed to regard or move stuff and something didn't work well. But if I understood it right they mentioned the PRs ES index which they didn't shared proof increased as the number of repos.
It might be anything. It seems they lost huge chunks due to layoffs and structural changes and MS which has the reverse golden Midas touch.
This is just pure speculation but also now there is no reason for MS to keep GH working. They absorbed all code they wanted. Now they can let it burn. Would be even better for them if that happened
> Could be. But 99% of the repos are static garbage with no PR nor actions.
But the 1% of repos that do have PRs and actions are likely going to be seeing enormous increases in volumes
I have been a part of two very large companies with self hosted gits and I've seen enough to be confident that this is an incredibly hard thing to manage
Serious question, have you been part of an org that had to scale orders of magnitude very quickly?
Anyone who has been part of that journey knows how painful it really is. A lot of times the systems to fail at all levels, and you have to redesign it from the first principles.
> Serious question, have you been part of an org that had to scale orders of magnitude very quickly?
I have, but it depends what you mean.
Scenario 1: e-commerce SaaS (think: Amazon but whitelabel, and before CPUs even had AES instructions); Christmas was "fun".
Scenario 2: Video Games. The first day is the worst day when it comes to scale. Everything has to be flawless from day 0 and you get no warning as to what can go wrong.
Yet, somehow, I managed to make highly reliable systems.
In scenario 1; I had an existing system that had to scale up and down with load, this was before there was cloud and hardware had a 3-4 month lead time, so most of the effort was around optimising existing code, increasing job timeouts and "quenching" sources that were expensive. We used to also do so 'magic' when it came to serving requests that had session token or shopping cart cookie.
In scenario 2; we have a clean-room implementation and no legacy, which is a blessing but also a curse, there's no possibility to sample real usage: but you also don't need to worry about making breaking changes that are for the better. With legacy you have to figure out how to migrate to the new behaviour gradually.
So, pro's and con's... but it's not like handling huge load hasn't been done before, computers are faster than they ever have been and while my personal opinion is that operational knowledge is dying (due to general distain for people who actually used to run systems that scale: not just write hopeful "eventually consistent" yaml that they call deterministic) - the systems that do exist today hold your hand much better than they did for me 20 years ago.
And I ran 1% of web traffic with an ops team of 5 back then. So, idk what's going on here.
EDIT: Likely people are flagging me because I sound arrogant (or I hurt their feelings by talking bad about YAML-ops), but all I am doing is answering the question presented based on my experience.
I wouldn't be surprised. Have you not noticed the sheer volume of slop being posted everywhere these days? Almost all of that is hosted on Github. And some of those repos have insane commit frequencies.
They say it is at least one order of magnitude[1]; "our plan to increase GitHub’s capacity by 10X in October 2025 .. By February 2026, it was clear that we needed to design for a future that requires 30X today’s scale."
Azure repos are kinda fine. It's really basic and there is nothing to break. I actually really really like their ticketing thingy for the same reason. It has the necessary stuff and the management types can't add a million of fields to it and annoy me with reporting, burndown charts or what not.
It has an annoying bug where approving PR's from the cli won't delete branches when you squash commit, while clicking the button in the UI does it perfectly fine. It's been a bug for a while (as in several years), and if you find something like that, don't expect it to ever be fixed. As a whole it's not a bad tool though.
As you say it's limited, but that can be both good and bad.
I'm on the other side of the fence. We're just about done migrating from GitHub to GitLab (self-hosted) and it's been refreshing to DGAF about any of the GH outages I read about.
Similar boat myself too, finished moving all important stuff from GitHub to self-hosted Forgejo with cross-platform builds. Not only do I avoid all the downtime stuff, but E2E builds also takes ~20% of the completion time it used to take, since now my runners have dedicated hardware hosted at home.
To maintain a fair comparison, GitHub has supported self-hosted runners for several years (maybe that doesn’t work for your specific usage, for whatever reason).
> To maintain a fair comparison, GitHub has supported self-hosted runners for several years
Yeah, tried that first, as I didn't want to move to Forgejo, I just wanted to keep working when I wanted to work.
The GitHub runner on Linux seemed fine, but the ones for macOS and Windows seemingly did something that made them a hell lot slower than even running VMs and then executing stuff inside those. I'm not sure what the runner is doing, if there is some built-in sandboxing or what not for those platforms, but it wasn't feasible to rely on for me as the builds took way too long time.
We were on self-hosted Gitlab but after a merger were forced to Github. Navigation feels painful in comparison and basic features such as commit graph are now behind more expensive tiers.
Interesting! I worked with Gitlab and I also thought it was quite clunky. If it was not for the stability issues GitHub is fine. Any other alternatives to GH or GL?
I think you’ve got it backwards. GitHub is by far the market leader for hosted repositories and maybe for CI too. This is like asking “Why are companies interested in using AWS?”
When one firm is so dominant for so long, the question is more like “Why shouldn’t we just use GitHub like 80% of software companies do?”
The issues they’ve had are almost all very recent. Very few companies have reevaluated that decision, because moving a big and well-integrated part of infrastructure is a huge project that delivers no value to the business. Speculating that you’ll have fewer development-slowing outages is not the most convincing when asking for the budget to do this. Plus, self-hosted isn’t necessarily going to have better uptime - mistakes happen.
I think before Actions, it would have been a lot easier to migrate off GH though. You’d just need to change a lot of repo URLs and find a way to set up webhooks from the new place to poke CI. Now with Actions, a lot lives in GH and in a proprietary flavor that doesn’t just ‘lift and shift.’
I don’t know why you would even really need hosted git or why you’d be affected by its downtime. Git is decentralized by design. One node going down should not stop development. You don’t need a “central hub” to keep working.
I guess it’s all the other non-git stuff like issue tracking and other (unfortunately) centralized products on GitHub that causes disruption when they go down.
Weird how GitHub built itself around a distributed VC system and then made all its other services centralized.
Yes, you want to run automated builds, unit test, end to end test, UI tests, make it easy for testers to deploy specific versions / tags to internal server. Also kick off builds for iOS on mac computers. We use Teamcity for that.
Tracking of issues, feature and epics. Maybe also knowledge base / wiki. We use Jira.
> Honest question, why are companies interested in hosting on github?
Mostly boils down to marketing and easier to establish a community. Almost every developer has an account there, leading to network effects being much larger, so if you're a new FOSS project, finding contributors and getting your project in front of other's eyes is much easier when you're on GitHub compared to your own Forgejo instance.
With that said, I'd question if chasing "most external one-time contributors" or GitHub stars is the right way to actually run a FOSS project, personally I'd avoid thinking about those vanity-numbers as much as possible and focus on the project, code and contributors themselves.
But, I've literally heard those two arguments for "why GitHub" countless of times over the years.
Usually, at large enough corporations, it's one of two things. Some random project gets open sourced, and it ends up on Github(see, for example, Salesforce) - or, more commonly, some subsidiary or acquisition had github and has either refused to migrate to the internal source system or the hassle of migration isn't worth it.
Go with the flow, don't rock the boat and use what developers already know, are probably the most cited reasons I've heard.
I've tried so many times in the past to argue for self-hosted setup that you fully control if you can afford it, things just get so much smoother and if you're a software development company, you probably want to own the software development workflow E2E so you can actually ship as fast as you want.
Why do you care about github? It’s Just another corporation doing what they know best: harvesting money. The software ecosystem can live without github just fine
With Ghostty being the latest project to leave GitHub, it does make me wonder who will leave next.
I don't expect everybody and their nan to leave GitHub by next wednesday and spin up their own Forgejo server, but I do think GitHub should be worried that people are finally looking to move away from them.
The entrenchment here is ridiculous. The average software engineer does not care at all about their vcs or their forge and their knowledge of both is extremely shallow.
For the people who want to do their work and get on with their life, it does not really matter that much.
Exactly. I run my own Gitea server, but put my stuff on Github, because that's where the people are. Self-hosting an MP3 is not the same as being on Spotify.
Zero to one nines of uptime. Shoving copilot in everyone’s faces rather than focusing on quality. Threatening to charge people for self-hosted runners.
> GitHub saw a 58% year-over-year increase in the number of incidents, reaching 109 reported cases:
> 17 of them were classified as major, leading to over 100 hours of total disruption.
> April stood out as the most turbulent month, with incidents accumulating to 330 hours and 6 minutes.
It's probably due to a more recent change, IE, focusing on features over stability. Or, it could be that there was some turnover in ops and someone who was a hawk about stability isn't there.
If I were to bet, there's probably a product manager or other leader who's just gung-ho on new features and loosing track of who their customers are and what their needs are.
That can happen many times during a buyout. Some company buys a thing. The problem then is ownership of the thing. Who in the new company is going to own the 'make sure it stays good' problem. Sometimes with a buy out the people who were doing that may even stay at the company. But it is a matter of motivation. MS has a real serious problem. You can see the gaps where they have glued together at least 10 companies together and called it microsoft. They have a huge reputational risk issue. Where something breaking in the xbox div can have a negative impact on the tools division. Also the other way around. They lack focus on many items. They have needed a 'service pack 2' stop the presses moment and fix this mount everest of tech debt.
Definitely not, I remember some 4 years ago some random bug in a github-supported github-action and a comment in an issue saying: "I heard the team responsible for this action was laid off, don't expect a fix". This was shortly after the microsoft acquisition.
But the vibe coding BS probably made it 10 times worse.
They started with a hands off approach and then went hands on, I’m not sure but that ‘hands on’ timing is likely to happen shortly after the usual acquisition vesting period of 3 years when the old guard starts to leave.
> But the vibe coding BS probably made it 10 times worse.
Yup, keep seeing this in various companies. Teams that were effective and did solid engineering now are more effective and does even better engineering. Teams that were effectively already just "boilerplate monkies" now produce a lot more code than before, but the quality is the same so effectively they're worse at contributing now than before, and take more shortcuts, not less.
From my point of view, agents are amplifiers, so if you usually build spaghetti projects, agents just help you do that faster, not avoid the spaghetti altogether. If you usually build well-designed stuff, they can help you put that together faster.
I'm not sure it's specific to vibe coding so much as the AI feature add rush. Every SAAS company is throwing more shit at the wall than I've ever seen, to the point where I'm actively avoiding some software because I don't want yet another new feature release pop-up when I log in.
Add in them being extremely high scale and critical infrastructure and it's easy to see where things can go wrong, vibe added code or not. I think we'd all prefer they have long slow roadmaps but clearly leadership thinks they're in a fight with the other AI companies to release the newest and bestest every day.
Agreed. In general the amount and variety of bugs introduced since everyone started vibing is worrying. It is probably a national security concern but I guess so is the economy tanking due to failed AI investments. Guess we will see
Perhaps they can’t help themselves out of habit, it is their nature.
The original red dog team that started azure is long gone and the general success of the cloud papers over all levels of incompetence so that the incompetence is now entrenched and unable to do better.
Cloud service providers have this unfortunate property where poor designs will make more money which makes it hard to maintain a culture of excellence. I tried to push a design change that would result in a 10x throughput for a certain product and was told that a 90% drop in usage is the last thing they want. I self host my own stuff with GitLab, so far not a single unplanned outage in 6 years.
Titles can be genrated as well, just tell the LLM whether your readers love drama and witch hunts or are pseudo intellectuals. It'll come up with the correct framing.
M$ bought GH almost a decad ago. Instability is more due to mainstream traffic, combined (+ AI automation pushes on pointless repos) than some slow motion evil mastermind plot. Hanlon's razor applies.
I feel like I’m out of the loop, or maybe I’m just not a super GitHub power user, but GitHub does pretty much what I expect and I haven’t had issues with it. All my git commands for GitHub just work and PRs and code reviews are the same as it’s always been.
Can someone explain what exactly is so bad now that leaving it entirely to use some new platform, even spinning up your own servers, is a reasonable alternative?
It's more if you use it for things beyond traditional dev work. GitHub Actions have become very unstable plus someone using it at this level where people are trying to download/ file issues/ send code up 24/7 would feel the pain of every outage, not just those that happen during one's working hours.
> and PRs and code reviews are the same as it’s always been
When they changed the PR view to not display all the changes at once, was the moment I said "I really need to find something else", not only is the platform very unreliable (at least from Spain), but most product changes they do are making the platform less efficient for me as a developer to use.
> Can someone explain what exactly is so bad now that leaving it entirely to use some new platform, even spinning up your own servers, is a reasonable alternative?
It always was, but network-effect of GitHub been large. But seemingly not infinite, at one point people start favoring "Being able to access platform" over "people can star my repository" it seems.
When using it every day (and especially when using Github Actions), there's something broken or half-broken nearly every day.
Most random errors in Github Actions (e.g. jobs just randomly failing or getting stuck and requiring a manual restart, or just being plain slow) also never show up on the Github Status page. The Github Actions VMs are also so slow that I'm seriously pondering setting up a cheap throw-away laptop at home as runner, that would easily be 10x faster. But then we're at playing IT admin at home :/
I’m a relatively casual user of GitHub and even I’ve run into availability issues when pushing up changes. Your comment makes it sound like you don’t use GitHub much at all or maybe are in some time zone or AZ that’s somehow insulated.
something reliably breaks 7 am PST (sometimes earlier) if you're using anything more than the git command line and sometimes (not too often, true) even the git protocol breaks.
TLDR; don't use their SaaS offering, but probably better, yes, though who knows for how long.
I don't use their SaaS offering, but I've been using the self-hosted versions (mostly in CE flavours, but occasionally paid) since the days of the weird black and white fox, when gitlab looked very bootstrap-y. (The logo in question for the curious, but you can't unsee it: https://upload.wikimedia.org/wikipedia/commons/0/0a/Gitlab_l... )
Anyway, since LLMs for coding became a thing, coupled with the realities of running a business post-IPO, it's been a slow-ish downward trend for the self-hoster, as their offering gets more and more bloat that's likely easier to manage at scale for Gitlab, but stands in stark contrast to what it once was.
Little things are pilling up; components left for dead (we now have both TODO - which is an abysmal mess, and "assigned work items" - WHY?!), issue boards that remain messy, advertisements creeping into the CE version, increasingly wild hardware requirements... and some recent changes to their documentation that strike me as a dark pattern; very much a recognition that either you're an enterprise running your own paid GitLab, with some kind of support, or you're a SaaS user and don't GAF about the ops docs.
The transition to websockets was annoying. Mostly because it kinda-doesn't work and there's no decent polling fallback, which results in time wasted hitting refresh, in 2026, when everything worked fine from 201x-2025.
I've kept my eye out for alternatives, but Gitlab's CI/CD, and the self hosted runners, is still my preferred flavour hands down and continues to be the reason I stick around.
Overall, it's a much slower decline, but like all stock-market-centric companies, you can feel the writing on the wall. Nevertheless, we're in the middle of a Gitlab migration from one cloud provider to anther because we still haven't found something better. :/
I wonder if there's a place for something like matrix, but for repositories (or maybe matrix protocol can handle that?). A world where we have selfhosted, saas, etc, but all interlinked and searchable? Say I find a project on gitlab, I want to contribute to, I checkout to my personal server (or someone elses hosted), and raise a pr back to original repo.
I know it doesn't answer your question, just thinking aloud really.
This would be best, though it's a big ask, when everyone has gone from self hosting plus a sprinkling of cloud services, to only cloud services and no remembrance of how to self host.
I used to run a git server for all my main projects, and mirrored public ones on GitHub. Then the convenience of GitHub lured me in, to the point I shut down my private git server 5 years ago.
Let's be honest there's an order of magnitude or more higher throughput volume of PR jitter and new repo bloat which makes this look like a viral digital native at scale.. couple that with being owned by one of the most scale immature companies on the planet ... of course it's a problem.
Get these folks off Azure and Cosmos DB (or whatever MSFT forces them to use) to something real and maybe you'd have a shot
It isn't surprising at all, Microsoft is doing a PE firm playbook with what they buy. You don't need to look much far, let's think about its biggest acquisition to date, Blizzard.
Blizzcon canceled. All of its IP barely got any love.
See what players think about the latest World of Warcraft patch. It's absolutely shit and broken. People say they fired the entire QA department since a few years back and since then the quality has just gone down.
They buy those businesses because they have nothing to do with that free cash flow, and for accounting reasons it makes sense to have them.
They didn't buy those businesses to develop it further and make it worth more.
Github will just become ever more irrelevant.
The key issue is that the US governments let those huge monopolies exist, and then use their money to buy other businesses and enshiftify them.
Unless that changes in the US, this will continue happening.
The Steam userbase would appear to disagree, with the recent reviews being mostly negative reviews (and the user reviews for Overwatch have hovered between mixed and negative for years now). And this doesn't appear to be from review bombing by some specific subset of players, the language breakdown shows reviews ranging from mixed to negative in all major language groups (English, Russian, Chinese, etc.).
> See what players think about the latest World of Warcraft patch. It's absolutely shit and broken
Crazy to me that the loot tables are still broken for some players/characters, they've tried to fix it several times now, and it's still not working - Since (some) endgame gear can only be obtained this way they've effectively soft locked those players/character out of the endgame.
Context: Some players are always receiving the same drops i.e. a belt. Rather than a varied loot table that gives them a chance to get items they need.
I’m certain I’m up there in the 1% of users, or close to it, that are writing software daily in terms of consistent prolonged volume of work and work that is actually used by others over the past nearly 20 years based on user activity statistics I’ve collected.
I, too, am a fairly, but not immediate early user of GitHub. Despite GitHub’s poor metrics, I am still shipping, because writing software doesn’t require GitHub.
Hashimoto’s comments sound disturbed and I hope he finds some peace, but if he wasn’t who he was and you read these comments, you’d think this person had a problem. So, I think he does.
„I don’t use any of GitHub’s non-git features, so if you use them you have a problem”
Who are you quoting?
> I, too, am a fairly, but not immediate early user of GitHub. Despite GitHub’s poor metrics, I am still shipping, because writing software doesn’t require GitHub.
GitHub’s downtime is a problem for issue tracking, PR merging, contributing and reviewing PRs, and more.
Your exact point was already pre-addressed in the blog post because it’s so predictable that some would completely miss the point. GitHub’s downtime isn’t about stopping you from writing code on your own machine.
Nothing is pissing me off more than GitHub's stability going down the tubes RIGHT as work is migrating everything, and I mean everything, from CircleCI to GH.
The wildest thing is that Azure Repos/Pipelines was better than this.
Their one caveat is also that they are still migrating it to Azure infra, so it's possible that's still in a one foot in one foot out kinda scenario, from what I've heard. But, this isn't inspiring confidence.
They're claiming a huge increase in traffic due to vibe coded projects. It might just be an excuse, but it certainly seems plausible to me.
Could be. But 99% of the repos are static garbage with no PR nor actions.
They mentioned they have some elasticsearch reindexing going to, I would guess they needed to regard or move stuff and something didn't work well. But if I understood it right they mentioned the PRs ES index which they didn't shared proof increased as the number of repos.
It might be anything. It seems they lost huge chunks due to layoffs and structural changes and MS which has the reverse golden Midas touch.
This is just pure speculation but also now there is no reason for MS to keep GH working. They absorbed all code they wanted. Now they can let it burn. Would be even better for them if that happened
> Could be. But 99% of the repos are static garbage with no PR nor actions.
But the 1% of repos that do have PRs and actions are likely going to be seeing enormous increases in volumes
I have been a part of two very large companies with self hosted gits and I've seen enough to be confident that this is an incredibly hard thing to manage
Serious question, have you been part of an org that had to scale orders of magnitude very quickly?
Anyone who has been part of that journey knows how painful it really is. A lot of times the systems to fail at all levels, and you have to redesign it from the first principles.
> Serious question, have you been part of an org that had to scale orders of magnitude very quickly?
I have, but it depends what you mean.
Scenario 1: e-commerce SaaS (think: Amazon but whitelabel, and before CPUs even had AES instructions); Christmas was "fun".
Scenario 2: Video Games. The first day is the worst day when it comes to scale. Everything has to be flawless from day 0 and you get no warning as to what can go wrong.
Yet, somehow, I managed to make highly reliable systems.
In scenario 1; I had an existing system that had to scale up and down with load, this was before there was cloud and hardware had a 3-4 month lead time, so most of the effort was around optimising existing code, increasing job timeouts and "quenching" sources that were expensive. We used to also do so 'magic' when it came to serving requests that had session token or shopping cart cookie.
In scenario 2; we have a clean-room implementation and no legacy, which is a blessing but also a curse, there's no possibility to sample real usage: but you also don't need to worry about making breaking changes that are for the better. With legacy you have to figure out how to migrate to the new behaviour gradually.
So, pro's and con's... but it's not like handling huge load hasn't been done before, computers are faster than they ever have been and while my personal opinion is that operational knowledge is dying (due to general distain for people who actually used to run systems that scale: not just write hopeful "eventually consistent" yaml that they call deterministic) - the systems that do exist today hold your hand much better than they did for me 20 years ago.
And I ran 1% of web traffic with an ops team of 5 back then. So, idk what's going on here.
EDIT: Likely people are flagging me because I sound arrogant (or I hurt their feelings by talking bad about YAML-ops), but all I am doing is answering the question presented based on my experience.
Is GitHub scaling by orders of magnitude though? That would be an insane increase at this stage of their lifecycle.
I wouldn't be surprised. Have you not noticed the sheer volume of slop being posted everywhere these days? Almost all of that is hosted on Github. And some of those repos have insane commit frequencies.
They say it is at least one order of magnitude[1]; "our plan to increase GitHub’s capacity by 10X in October 2025 .. By February 2026, it was clear that we needed to design for a future that requires 30X today’s scale."
[1] https://github.blog/news-insights/company-news/an-update-on-...
Note the lack of concrete numbers on how much they have scaled. Somebody may have just asked an LLM for projections.
If they're suffering the onslaught of ai slop, it's possible.
> you have to redesign it from the first principles
And that start by layoffing your best engineers, I guess
Yep, definitely more traffic and also more new Github repos being created, with a pretty huge spike the last 2 months [1]
[1] https://bloomberry.com/data/github/
I’d be shocked if this wasn’t the reason.
You can cancel the migration, no need for sunk cost fallacy.
"You" can't necessarily do anything (you would be making a lot of assumptions about the influence this person has over the decision making process).
"Someone" can cancel the migration. "Someone" just won't.
https://en.wikipedia.org/wiki/Generic_you
Azure repos are kinda fine. It's really basic and there is nothing to break. I actually really really like their ticketing thingy for the same reason. It has the necessary stuff and the management types can't add a million of fields to it and annoy me with reporting, burndown charts or what not.
Yea, I have Azure DevOps with free action minutes and I’ve started using it a ton more since it avoids all GH outages.
It has an annoying bug where approving PR's from the cli won't delete branches when you squash commit, while clicking the button in the UI does it perfectly fine. It's been a bug for a while (as in several years), and if you find something like that, don't expect it to ever be fixed. As a whole it's not a bad tool though.
As you say it's limited, but that can be both good and bad.
i might be connecting unrelated dots, yet when i read "migration to Azure" this came back to me
https://news.ycombinator.com/item?id=47616242 https://isolveproblems.substack.com/p/how-microsoft-vaporize...
If you’re making this change now, I wonder how the technical leadership evaluated GitHub and its competitors.. and then still landed on GH.
What made it better than e.g GitLab?
I'm on the other side of the fence. We're just about done migrating from GitHub to GitLab (self-hosted) and it's been refreshing to DGAF about any of the GH outages I read about.
Similar boat myself too, finished moving all important stuff from GitHub to self-hosted Forgejo with cross-platform builds. Not only do I avoid all the downtime stuff, but E2E builds also takes ~20% of the completion time it used to take, since now my runners have dedicated hardware hosted at home.
To maintain a fair comparison, GitHub has supported self-hosted runners for several years (maybe that doesn’t work for your specific usage, for whatever reason).
> To maintain a fair comparison, GitHub has supported self-hosted runners for several years
Yeah, tried that first, as I didn't want to move to Forgejo, I just wanted to keep working when I wanted to work.
The GitHub runner on Linux seemed fine, but the ones for macOS and Windows seemingly did something that made them a hell lot slower than even running VMs and then executing stuff inside those. I'm not sure what the runner is doing, if there is some built-in sandboxing or what not for those platforms, but it wasn't feasible to rely on for me as the builds took way too long time.
We were on self-hosted Gitlab but after a merger were forced to Github. Navigation feels painful in comparison and basic features such as commit graph are now behind more expensive tiers.
Interesting! I worked with Gitlab and I also thought it was quite clunky. If it was not for the stability issues GitHub is fine. Any other alternatives to GH or GL?
Self-hosting with open source code:
- SourceHut: https://sr.ht/~sircmpwn/sourcehut/
- Forgejo (used by Codeberg, etc.): https://forgejo.org/
SourceHut never really clicked for me. It doesn't give me anything useful that I don't already have in a bare git repo through a ssh.
Forgejo, on the other hand, is a drop-in replacement for GitHub.
We switched from Bit bucket to Gerrit internally and it was a steep learning curve for the des but it's fine.
At a customer we're implementing GitHub Actions and even on our Dev environment there are so many hickups with GitHub.
Demo of Gerrit here: https://gerrithub.io/q/status:open+-is:wip
Jira / Bitbucket / Teamcity.
Might be pricy though.
Having used Teamcity for CI I cannot think of a more clunky and hard to use system (compared to GHA, which is what we migrated to).
Artifacts - C'mon Wit Da Git Down
https://www.youtube.com/watch?v=Js_Y_q-IkYo
Mee too. We just did a very similar migration at work it's incredibly frustrating, I've got all my CI ported over and now this.
MSFT should just create slophub.com they'd make money im sure.
Honest question, why are companies interested in hosting on github?
As a private person I use it too as a free hoster, but from work I mainly know self hosted instances of jenkins and TeamCity.
I think you’ve got it backwards. GitHub is by far the market leader for hosted repositories and maybe for CI too. This is like asking “Why are companies interested in using AWS?”
When one firm is so dominant for so long, the question is more like “Why shouldn’t we just use GitHub like 80% of software companies do?”
The issues they’ve had are almost all very recent. Very few companies have reevaluated that decision, because moving a big and well-integrated part of infrastructure is a huge project that delivers no value to the business. Speculating that you’ll have fewer development-slowing outages is not the most convincing when asking for the budget to do this. Plus, self-hosted isn’t necessarily going to have better uptime - mistakes happen.
I think before Actions, it would have been a lot easier to migrate off GH though. You’d just need to change a lot of repo URLs and find a way to set up webhooks from the new place to poke CI. Now with Actions, a lot lives in GH and in a proprietary flavor that doesn’t just ‘lift and shift.’
> I think you’ve got it backwards. GitHub is by far the market leader for hosted repositories
Maybe, but I never heard about any company using github for internal projects in my real life. For me it was always to go to for open source projects.
Then again it's not a topic that often comes up in my developer circles.
I don’t know why you would even really need hosted git or why you’d be affected by its downtime. Git is decentralized by design. One node going down should not stop development. You don’t need a “central hub” to keep working.
I guess it’s all the other non-git stuff like issue tracking and other (unfortunately) centralized products on GitHub that causes disruption when they go down.
Weird how GitHub built itself around a distributed VC system and then made all its other services centralized.
> I guess it’s all the other non-git stuff
Yes, you want to run automated builds, unit test, end to end test, UI tests, make it easy for testers to deploy specific versions / tags to internal server. Also kick off builds for iOS on mac computers. We use Teamcity for that.
Tracking of issues, feature and epics. Maybe also knowledge base / wiki. We use Jira.
And pull requests. Bitbucket.
Most developers have experience using GitHub. The UI and concepts are familiar. The friction for adopting features like Actions is relatively low.
> The UI and concepts are familiar.
I guess, but it's not like you can't learn how to create a pullrequest on bitbucket or how to create an issue on jira as well within a work day?
That seems like the smallest thing when switching to a new company.
> The friction for adopting features like Actions is relatively low.
Yeah, I know almost nothing about the CI integration and actions when it comes to Github. Will look into it. Thank you.
At one point it was also used as signaling that a company was “modern.”
> Honest question, why are companies interested in hosting on github?
Mostly boils down to marketing and easier to establish a community. Almost every developer has an account there, leading to network effects being much larger, so if you're a new FOSS project, finding contributors and getting your project in front of other's eyes is much easier when you're on GitHub compared to your own Forgejo instance.
With that said, I'd question if chasing "most external one-time contributors" or GitHub stars is the right way to actually run a FOSS project, personally I'd avoid thinking about those vanity-numbers as much as possible and focus on the project, code and contributors themselves.
But, I've literally heard those two arguments for "why GitHub" countless of times over the years.
Oh FOSS projects I totally understand. It's where I go to too.
But closed source companies surly don't need to establish a community?
Usually, at large enough corporations, it's one of two things. Some random project gets open sourced, and it ends up on Github(see, for example, Salesforce) - or, more commonly, some subsidiary or acquisition had github and has either refused to migrate to the internal source system or the hassle of migration isn't worth it.
Go with the flow, don't rock the boat and use what developers already know, are probably the most cited reasons I've heard.
I've tried so many times in the past to argue for self-hosted setup that you fully control if you can afford it, things just get so much smoother and if you're a software development company, you probably want to own the software development workflow E2E so you can actually ship as fast as you want.
Onboarding construct workers is super easy.
Why do you care about github? It’s Just another corporation doing what they know best: harvesting money. The software ecosystem can live without github just fine
Github uptime down to 86% according to https://mrshu.github.io/github-statuses/ (not my website)
[Dupe] https://news.ycombinator.com/item?id=47939579
With Ghostty being the latest project to leave GitHub, it does make me wonder who will leave next.
I don't expect everybody and their nan to leave GitHub by next wednesday and spin up their own Forgejo server, but I do think GitHub should be worried that people are finally looking to move away from them.
The entrenchment here is ridiculous. The average software engineer does not care at all about their vcs or their forge and their knowledge of both is extremely shallow.
For the people who want to do their work and get on with their life, it does not really matter that much.
Exactly. I run my own Gitea server, but put my stuff on Github, because that's where the people are. Self-hosting an MP3 is not the same as being on Spotify.
I'm a bit out of the loop, why are people moving away from Github?
Zero to one nines of uptime. Shoving copilot in everyone’s faces rather than focusing on quality. Threatening to charge people for self-hosted runners.
Mainly this https://mrshu.github.io/github-statuses/ But also product decisions that don't seem to be aligned with the preferences of users... see https://mitchellh.com/writing/ghostty-leaving-github
> GitHub saw a 58% year-over-year increase in the number of incidents, reaching 109 reported cases: > 17 of them were classified as major, leading to over 100 hours of total disruption. > April stood out as the most turbulent month, with incidents accumulating to 330 hours and 6 minutes.
https://gitprotect.io/blog/devops-threats-unwrapped-mid-year...
That's from 2025 but it continued getting even worse this year.
do we have already an HN user creating "who-left-gh.net"? (domain is free)
Created a repository to track all the latest github projects (ironically, my project is on github xD)
https://github.com/SerJaimeLannister/who-left-gh/
Currently I know 3 projects, Ghostty, Bookstack-app, Hardenedbsd who have seemed to move away from Github from my understanding.
The Zig project has also left GitHub relatively recently
https://ziglang.org/news/migrating-from-github-to-codeberg/
Thanks to both you and mi_lk, I have added zig to the list.
arguably Zig started the trend just about the time GH really gone shit (this year)
Is it me, or did get issues get a lot worse with the transfer to MSFT?
The purchase wasn't a year ago, it was 8 years ago. In that time how much has it grown? 10x? 100x? More?
It's probably due to a more recent change, IE, focusing on features over stability. Or, it could be that there was some turnover in ops and someone who was a hawk about stability isn't there.
If I were to bet, there's probably a product manager or other leader who's just gung-ho on new features and loosing track of who their customers are and what their needs are.
Yes, of course, but also more recently under the new CoreAI unit: https://www.businessinsider.com/microsoft-ai-coding-rivals-o...
That can happen many times during a buyout. Some company buys a thing. The problem then is ownership of the thing. Who in the new company is going to own the 'make sure it stays good' problem. Sometimes with a buy out the people who were doing that may even stay at the company. But it is a matter of motivation. MS has a real serious problem. You can see the gaps where they have glued together at least 10 companies together and called it microsoft. They have a huge reputational risk issue. Where something breaking in the xbox div can have a negative impact on the tools division. Also the other way around. They lack focus on many items. They have needed a 'service pack 2' stop the presses moment and fix this mount everest of tech debt.
I think is more related to vibe coding
Definitely not, I remember some 4 years ago some random bug in a github-supported github-action and a comment in an issue saying: "I heard the team responsible for this action was laid off, don't expect a fix". This was shortly after the microsoft acquisition.
But the vibe coding BS probably made it 10 times worse.
> I remember some 4 years ago ... This was shortly after the microsoft acquisition.
The acquisition was 8 years ago.
They started with a hands off approach and then went hands on, I’m not sure but that ‘hands on’ timing is likely to happen shortly after the usual acquisition vesting period of 3 years when the old guard starts to leave.
> But the vibe coding BS probably made it 10 times worse.
Yup, keep seeing this in various companies. Teams that were effective and did solid engineering now are more effective and does even better engineering. Teams that were effectively already just "boilerplate monkies" now produce a lot more code than before, but the quality is the same so effectively they're worse at contributing now than before, and take more shortcuts, not less.
From my point of view, agents are amplifiers, so if you usually build spaghetti projects, agents just help you do that faster, not avoid the spaghetti altogether. If you usually build well-designed stuff, they can help you put that together faster.
I'm not sure it's specific to vibe coding so much as the AI feature add rush. Every SAAS company is throwing more shit at the wall than I've ever seen, to the point where I'm actively avoiding some software because I don't want yet another new feature release pop-up when I log in.
Add in them being extremely high scale and critical infrastructure and it's easy to see where things can go wrong, vibe added code or not. I think we'd all prefer they have long slow roadmaps but clearly leadership thinks they're in a fight with the other AI companies to release the newest and bestest every day.
Agreed. In general the amount and variety of bugs introduced since everyone started vibing is worrying. It is probably a national security concern but I guess so is the economy tanking due to failed AI investments. Guess we will see
In terms of at Microsoft's end, or in general with the amount of new repos and pushes / commits from other people vibe-coding?
GitHub actions sucked and fell over itself long before vibe coding became mainstream.
Even after decades, the policy is the same:
Embrace, extend, and extinguish.
Which was a policy that increased their market dominance for their existing dominate products.
What exactly are they extinguishing GitHub to the benefit of? Azure Repos?
Perhaps they can’t help themselves out of habit, it is their nature.
The original red dog team that started azure is long gone and the general success of the cloud papers over all levels of incompetence so that the incompetence is now entrenched and unable to do better.
Cloud service providers have this unfortunate property where poor designs will make more money which makes it hard to maintain a culture of excellence. I tried to push a design change that would result in a 10x throughput for a certain product and was told that a 90% drop in usage is the last thing they want. I self host my own stuff with GitLab, so far not a single unplanned outage in 6 years.
It is absolutely not just you
Is it just me, or [thing that has been repeated a billion times every day on this and every other website]
It certainly seems like low effort engagement farming.
just like hashicorp, its careers site redirects to IBM site with no filter to find hashicorp positions
You gotta admire journalists.
Such a one punch sentence that distills the message with a little bit of dramatic flair.
got damn, anyone got recommendations on how to write like a journalist ?
Take a sentence for a title, replace one of the verbs with SLAM, LAMBAST and CLAPS BACK, you're hired.
The rest of the article can be AI generated, don't fret about it.
Don't forget the "..., and no one/nothing is <verb>ing" clause.
Titles can be genrated as well, just tell the LLM whether your readers love drama and witch hunts or are pseudo intellectuals. It'll come up with the correct framing.
Have you tried asking an LLM?
I imagine they'll ruin VS Code and NPM next.
Are we watching live the first company that implodes due to Vibe Coding?
everyone knew M$ would ruin github
the fake surprise is so fake
M$ bought GH almost a decad ago. Instability is more due to mainstream traffic, combined (+ AI automation pushes on pointless repos) than some slow motion evil mastermind plot. Hanlon's razor applies.
No one forced them to migrate GitHub to switch to React. And the atrocious sleep bug in Actions long predates the AI gold rush.
Microslop turned GitHub to gitslop.
I haven't seen anyone faking surprise, or even genuinely surprised, what are you referring to?
I feel like I’m out of the loop, or maybe I’m just not a super GitHub power user, but GitHub does pretty much what I expect and I haven’t had issues with it. All my git commands for GitHub just work and PRs and code reviews are the same as it’s always been.
Can someone explain what exactly is so bad now that leaving it entirely to use some new platform, even spinning up your own servers, is a reasonable alternative?
Their stability and reliability has deteriorated significantly.
So much so that they stopped posting uptime metrics for a while on their status page and an independent 3rd party created a website just for this:
https://mrshu.github.io/github-statuses/ (not my website)
According to that website, which unsurprisingly reports a lower number than what Github themselves claim, Github uptime is down to ~86%.
And if you work in the space, you know how terrible that is, but even more so for such a critical piece of infrastructure.
It's more if you use it for things beyond traditional dev work. GitHub Actions have become very unstable plus someone using it at this level where people are trying to download/ file issues/ send code up 24/7 would feel the pain of every outage, not just those that happen during one's working hours.
> and PRs and code reviews are the same as it’s always been
When they changed the PR view to not display all the changes at once, was the moment I said "I really need to find something else", not only is the platform very unreliable (at least from Spain), but most product changes they do are making the platform less efficient for me as a developer to use.
> Can someone explain what exactly is so bad now that leaving it entirely to use some new platform, even spinning up your own servers, is a reasonable alternative?
It always was, but network-effect of GitHub been large. But seemingly not infinite, at one point people start favoring "Being able to access platform" over "people can star my repository" it seems.
When using it every day (and especially when using Github Actions), there's something broken or half-broken nearly every day.
Most random errors in Github Actions (e.g. jobs just randomly failing or getting stuck and requiring a manual restart, or just being plain slow) also never show up on the Github Status page. The Github Actions VMs are also so slow that I'm seriously pondering setting up a cheap throw-away laptop at home as runner, that would easily be 10x faster. But then we're at playing IT admin at home :/
I’m a relatively casual user of GitHub and even I’ve run into availability issues when pushing up changes. Your comment makes it sound like you don’t use GitHub much at all or maybe are in some time zone or AZ that’s somehow insulated.
something reliably breaks 7 am PST (sometimes earlier) if you're using anything more than the git command line and sometimes (not too often, true) even the git protocol breaks.
They have less uptime than my home NAS, this is the #1 thing that is wrong with it.
And the most recent bug is "we added random code to some PR and some PR became invisible" which is a wtf bug which should be impossible to exist.
Is Gitlab doing better at this point? Or where do they stand?
GitLab has been doing better than GitHub for a decade or more.
TLDR; don't use their SaaS offering, but probably better, yes, though who knows for how long.
I don't use their SaaS offering, but I've been using the self-hosted versions (mostly in CE flavours, but occasionally paid) since the days of the weird black and white fox, when gitlab looked very bootstrap-y. (The logo in question for the curious, but you can't unsee it: https://upload.wikimedia.org/wikipedia/commons/0/0a/Gitlab_l... )
Anyway, since LLMs for coding became a thing, coupled with the realities of running a business post-IPO, it's been a slow-ish downward trend for the self-hoster, as their offering gets more and more bloat that's likely easier to manage at scale for Gitlab, but stands in stark contrast to what it once was.
Little things are pilling up; components left for dead (we now have both TODO - which is an abysmal mess, and "assigned work items" - WHY?!), issue boards that remain messy, advertisements creeping into the CE version, increasingly wild hardware requirements... and some recent changes to their documentation that strike me as a dark pattern; very much a recognition that either you're an enterprise running your own paid GitLab, with some kind of support, or you're a SaaS user and don't GAF about the ops docs.
The transition to websockets was annoying. Mostly because it kinda-doesn't work and there's no decent polling fallback, which results in time wasted hitting refresh, in 2026, when everything worked fine from 201x-2025.
I've kept my eye out for alternatives, but Gitlab's CI/CD, and the self hosted runners, is still my preferred flavour hands down and continues to be the reason I stick around.
Overall, it's a much slower decline, but like all stock-market-centric companies, you can feel the writing on the wall. Nevertheless, we're in the middle of a Gitlab migration from one cloud provider to anther because we still haven't found something better. :/
so where should people move to instead
I wonder if there's a place for something like matrix, but for repositories (or maybe matrix protocol can handle that?). A world where we have selfhosted, saas, etc, but all interlinked and searchable? Say I find a project on gitlab, I want to contribute to, I checkout to my personal server (or someone elses hosted), and raise a pr back to original repo. I know it doesn't answer your question, just thinking aloud really.
https://tangled.org/ is close to what you are describing
tangled, forgejo, radicle, sourcehut - to name a few
Different places. Stop centralising the web. Embrace diversity.
This would be best, though it's a big ask, when everyone has gone from self hosting plus a sprinkling of cloud services, to only cloud services and no remembrance of how to self host.
I used to run a git server for all my main projects, and mirrored public ones on GitHub. Then the convenience of GitHub lured me in, to the point I shut down my private git server 5 years ago.
Now I kinda regret that decision.
As long as it's protected by Cloudflare, that is.
[dupe] Discussion on source: https://news.ycombinator.com/item?id=47939579
Let's be honest there's an order of magnitude or more higher throughput volume of PR jitter and new repo bloat which makes this look like a viral digital native at scale.. couple that with being owned by one of the most scale immature companies on the planet ... of course it's a problem.
Get these folks off Azure and Cosmos DB (or whatever MSFT forces them to use) to something real and maybe you'd have a shot
It isn't surprising at all, Microsoft is doing a PE firm playbook with what they buy. You don't need to look much far, let's think about its biggest acquisition to date, Blizzard.
Blizzcon canceled. All of its IP barely got any love.
See what players think about the latest World of Warcraft patch. It's absolutely shit and broken. People say they fired the entire QA department since a few years back and since then the quality has just gone down.
They buy those businesses because they have nothing to do with that free cash flow, and for accounting reasons it makes sense to have them.
They didn't buy those businesses to develop it further and make it worth more.
Github will just become ever more irrelevant.
The key issue is that the US governments let those huge monopolies exist, and then use their money to buy other businesses and enshiftify them.
Unless that changes in the US, this will continue happening.
Just FYI Blizzcon is returning in September.
https://blizzcon.com/en-us/event/
I don’t think this is true across Blizzard. Overwatch is the best it’s ever been.
The Steam userbase would appear to disagree, with the recent reviews being mostly negative reviews (and the user reviews for Overwatch have hovered between mixed and negative for years now). And this doesn't appear to be from review bombing by some specific subset of players, the language breakdown shows reviews ranging from mixed to negative in all major language groups (English, Russian, Chinese, etc.).
> See what players think about the latest World of Warcraft patch. It's absolutely shit and broken
Crazy to me that the loot tables are still broken for some players/characters, they've tried to fix it several times now, and it's still not working - Since (some) endgame gear can only be obtained this way they've effectively soft locked those players/character out of the endgame.
Context: Some players are always receiving the same drops i.e. a belt. Rather than a varied loot table that gives them a chance to get items they need.
Exactly why does it keep happening, why is the default strategy find golden goose, kill goose, look for new geese
The thing is that they didn't buy Blizzard, they bought Activision. They were interested in CoD numbers.
I think Diablo Immortal was likely the biggest success Blizz provided there