a lot of projects are having their self hosted git repositories flooded by scrapers from ai companies blasting 50-100 terabytes of bandwidth and crawling every endpoint like git blame which can be computationally expensive, so they have a rate limiter up that uses PoW to verify Also shows a cat girl
this scraper bullshit is insane. I've had my site get bombarded multiple times with a gazillion random requests to literally every possible URL in existence to the point of just freezing the entire server and having to force reboot it.
I get complaints from users because I limit php-fpm to 50 threads and these scrapers easily do 100 rps to non-cached content… and since they use some kind of distributed approach, it's thousands of IPs so my throttle controls cannot kick in properly
every girl i have met for the past twenty-five years of being online loves to look at cute pictures of anime girls smiling. no idea where this guy found his girlfriend
Comments
hosting multiple instances of MediaWiki is pain