TheNewsIsHere 7 hours ago

I don’t think the publication date (May 8, as I type this) on the GitHub blog article is the same date this change became effective.

From a long-term, clean network I have been consistently seeing these “whoa there!” secondary rate limit errors for over a month when browsing more than 2-3 files in a repo.

My experience has been that once they’ve throttled your IP under this policy, you cannot even reach a login page to authenticate. The docs direct you to file a ticket (if you’re a paying customer, which I am) if you consistently get that error.

I was never able to file a ticket when this happened because their rate limiter also applies to one of the required backend services that the ticketing system calls from the browser. Clearly they don’t test that experience end to end.

gnabgib 5 days ago

60 req/hour for unauthenticated users

5000 req/hour for authenticated - personal

15000 req/hour for authenticated - enterprise org

According to https://docs.github.com/en/rest/using-the-rest-api/rate-limi...

I bump into this just browsing a repo's code (unauth).. seems like it's one of the side effects of the AI rush.

  • mijoharas 4 hours ago

    Why would the changelog update not include this? it's the most salient piece of information.

    I thought I was just misreading it and failing to see where they stated what the new rate limits were, since that's what anyone would care about when reading it.

PaulDavisThe1st 3 hours ago

Several people in the comments seem to be blaming Github for taking this step for no apparent reason.

Those of us who self-host git repos know that this is not true. Over at ardour.org, we've passed the 1M-unique-IP's banned due to AI trawlers sucking our repository 1 commit at a time. It was killing our server before we put fail2ban to work.

I'm not arguing that the specific steps Github have taken are the right ones. They might be, they might not, but they do help to address the problem. Our choice for now has been based on noticing that the trawlers are always fetching commits, so we tweaked things such that the overall http-facing git repo works, but you cannot access commit-based URLs. If you want that, you need to use our github mirror :)

jorams 4 hours ago

> These changes will apply to operations like cloning repositories over HTTPS, anonymously interacting with our REST APIs, and downloading files from raw.githubusercontent.com.

Or randomly when clicking through a repository file tree. The first time I hit a rate limit was when I was skimming through a repository on my phone, and about the 5th file I clicked I was denied and locked out. Not for a few seconds either, it lasted long enough that I gave up on waiting then refreshing every ~10 seconds.

thih9 an hour ago

What does “secondary” stand for here in the error message?

> You have exceeded a secondary rate limit.

Edit and self-answer:

> In addition to primary rate limits, GitHub enforces secondary rate limits

(…)

> These secondary rate limits are subject to change without notice. You may also encounter a secondary rate limit for undisclosed reasons.

https://docs.github.com/en/rest/using-the-rest-api/rate-limi...

pogue 14 hours ago

I assume they're trying to keep ai bots from strip mining the whole place.

Or maybe your IP/browser is questionable.

  • globie 4 hours ago

    What's being strip mined is the openness of the Internet, and AI isn't the one closing up shop. Github was created to collaborate on and share source code. The company in the best position to maximize access to free and open software is now just a dragon guarding other people's coins.

    The future is a .txt file of John Carmack pointing out how efficient software used to be, locked behind a repeating WAF captcha, forever.

  • voidnap 13 hours ago

    I encountered this on github last week. Very agressive rate limiting. My browser and IP is very ordinary.

    Since Microsoft is struggling to make ends meet, maybe they could throw a captcha or proof of work like Anubis by xe iaso.

    They already disabled code search for unauthenticated users. Its totally plausible they will disable code browsing as well.

    • kstrauser 5 hours ago

      That hit me, too. I thought it was an accidental bug and didn’t realize it was actually malice.

  • confusing3478 13 hours ago

    > Or maybe your IP/browser is questionable.

    I'm using Firefox and Brave on Linux from a residential internet provider in Europe and the 429 error triggers consistantly on both browsers. Not sure I would consider my setup questionable considering their target audience.

    • grodriguez100 13 hours ago

      I’m browsing from an iPhone in Europe right now and can browse source code just fine without being logged in.

  • tostr 14 hours ago

    *other ai bots, ms will obviously mine anything on there.

    Personally, I like sourcehut (sr.ht)

    • immibis 5 hours ago

      Same way Reddit sells all its content to Google, then stops everyone else from getting it. Same way Stack Overflow sells all its content to Google, then stops everyone else from getting it.

      (Joke's on Reddit, though, because Reddit content became pretty worthless since they did this, and everything before they did this was already publicly archived)

  • croes 8 hours ago

    Other bots or MS bots too?

Zdh4DYsGvdjJ 13 hours ago
  • croes 8 hours ago

    Doesn‘t make it any better.

    Collateral damage of AI I guess

    • formerly_proven 6 hours ago

      It's even more hilarious because this time it's Microsoft/Github getting hit by it. (It's funny because MS themselves are such a bad actor when it comes to AIAIAI).

      • fragmede 6 hours ago

        This is the same Microsoft that owns LinkedIn which got sued by HiQ which is where the ruling came from that is making sites login required.

        • immibis 5 hours ago

          Wow! Website terms of use actually meant something in a court of law!

          • fragmede 4 hours ago

            that wasn't what the case was about, so not really.

jarofgreen 13 hours ago

Also https://github.com/orgs/community/discussions/157887 "Persistent HTTP 429 Rate Limiting on *.githubusercontent.com Triggered by Accept-Language: zh-CN Header" but the comments show examples with no language headers.

I encountered this too once, but thought it was a glitch. Worrying if they can't sort it.

Euphorbium 13 hours ago

I remember getting this error a few months ago, this does not seem like a temporary glitch. They dont want llm makers to slurp all the data.

  • new_user_final 13 hours ago

    Isn't git clone faster than browsing web?

    • PaulDavisThe1st 4 hours ago

      Yep. But AI trawlers don't use it. Ask them why.

trallnag 3 hours ago

Good that tools like Homebrew that heavily rely on GitHub usually support environment variables like GITHUB_TOKEN

stevekemp 3 hours ago

Once again people post in the "community", but nobody official replies; these discussion-pages are just users shouting into the void.

xnx 6 hours ago

It sucks that we've collectively surrendered the urls to our content to centralized services that can change their terms at any time without any control. Content can always be moved, but moving the entire audience associated with a url is much harder.

  • turblety 5 hours ago

    Gitea [1] is honestly awesome and lightweight. I've been running my own for years, and since they've put Actions in a while ago (with GitHub compatibility) it does everything I need it to. It doesn't have all the AI stuff in it (but for some that's a positive :P)

    1. https://about.gitea.com/

    • kstrauser 5 hours ago

      Gitea’s been great, but I think a lot of its development has moved to Forgejo: https://forgejo.org/

      That’s what I run on my personal server now.

      • TheNewsIsHere 4 hours ago

        I’ve almost completed the move of my business from GitHub’s corporate offering to self-hosted Forgejo.

        Almost went with Gitea, but the ownership structure is murky, feature development seems to have plateaued, and they haven’t even figured out how to host their own code. It’s still all on GitHub.

        I’ve been impressed by Forgejo. It’s so much faster than Github to perform operations, I can actually backup my entire corpus of data in a format that’s restorable/usable, and there aren’t useless (AI) upsells cluttering my UX.

        • kstrauser 4 hours ago

          I agree with every word of that.

          For listeners at home wondering why you'd want that at all:

          I want a centralized Git repo where I can sync config files from my various machines. I have a VPS so I just create a .git directory and start using SSH to push/pull against it. Everything works!

          But then, my buddy wants to see some of my config files. Hmm. I can create an SSH user for him and then set the permissions on that .git to give him read-only access. Fine. That works.

          Until he improves some of them. Hey, can I give him a read-write repo he can push a branch to? Um, sure, give me a bit to think this through...

          And one of his coworkers thinks this is fascinating and wants to look, too. Do I create an SSH account for this person I don't know well at all?

          At this point, I've done more work than just installing something like Forgejo and letting my friend and his FOAF create accounts on it. There's a nice UI for configuring their permissions. They don't have SSH access directly into my server. It's all the convenience of something like GitHub, except entirely under my control and I don't have to pay for private repos.

watermelon0 13 hours ago

Time for Mozilla (and other open-source projects) to move repositories to sourcehut/Codeberg or self-hosted Gitlab/Forgejo?

  • gsich 7 hours ago

    Not Mozilla.

radicality 13 hours ago

Just tried it on chrome incognito on iOS and do hit this 429 rate limit :S That sucks, it’s already bad enough when GitHub started enforcing login to even do a simple search.