Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I don't see how an AI crawler is different from any others.

The simplest approach is to count the UA as risky or flag multiple 404 errors or HEAD requests, and block on that. Those are rules we already have out of the box.

It's open source, there's no pain in writing specific rules for rate limiting, thus my question.

Plus, we have developed a dashboard for manually choosing UA blocks based on name, but we're still not sure if this is something that would be really helpful for website operators.



>It's open source, there's no pain in writing specific rules for rate limiting, thus my question.

Depends on the goal.

Author wants his instance not to get killed. Request rate limiting may achieve that easily in a way transparent to normal users.


> count the UA as risky

It's trivial to spoof UAs unfortunately.


It depends. If you want to stop OAI-SearchBot/1.3, UA will be enough.


Why would you need tirreno if you just want to stop OAI's bot though?

OAI's is just an example that's easy to explain.

I believe that if something is publicly available, it shouldn't be overprotected in most cases.

However, there are many advanced cases, such as crawlers that collect data for platform impersonation (for scams) or custom phishing attacks, or account brute-force attacks. In those cases, I use tirreno to understand traffic through different dimensions.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: