Mitigating SourceHut's partial outage caused by aggressive crawlers

135 points by fratti


lilac

I run an entirely smaller git service for just me, my wife, and some friends. It’s nothing special, and yet the scrapers have set their sights on me too. I don’t know what to do about them except block entire swaths of IP addresses once every few days when I just start getting hit from a different CIDR range.

I have no idea how frustrating it must be to deal with it at this scale, where there are thousands if not millions of links shared all around the public web to your service, and paying users who would likely be a lot less happy than my friends when I tell them I’m disabling web access for a few hours to get them to leave temporarily again.

This shit is exhausting. Good luck y’all. 💕