

Your training is complete young Padawan go hence from this place and create AI slop for the masses...And just as suddenly, the number of crawlers have dropped over night.
Mine spiked and went back to normal as well, but much lower numbers than yours. Maybe they got what they needed and left! Or thought "nothing to see here - move along"And just as suddenly, the number of crawlers have dropped over night. It has under 500 guests right now, compared to 30K+ the last few days.
Just weird.

Why not list China under challenge and not dealing with their ASNs? 99% of those bots wont solve the challenge and will move on to other domainI'm quite sure i'm going to ban the other China ISP as the amount of bandwidth is still enormous.
You seem to have missed that @ES Dev Team does not use Cloudflare and also does not want to. Cloudflare is not a given - many people do not want to use it, i.e. for privacy reasons.Why not list China under challenge and not dealing with their ASNs? 99% of those bots wont solve the challenge and will move on to other domain
What i thought.I might have to move to cloudflare for a while until i finish my system
Cloudflare is not a given - many people do not want to use it, i.e. for privacy reasons.
What i thought.
![]()
A lot of companies intentionally do not utilize Cloudflare (or other like-typed providers) for varying reasons. Be it legal, privacy, datamining concerns, company decision, network stack 'issues', existing infrastructure already handles the needs, or even financial burden. There is nothing wrong with not using Cloudflare. I don't understand the tin-foil statement gif.What i thought.
![]()
As for me, I finished implementing Anubis this morning on just about all websites that I manage (not all of these are Xenforo or forum related), and I've literally shut the door on all of these abusive LLM/AI Scrapers. The only virtual machines on my side of things are now just the load balancers/ingest servers and reverse proxy servers delegating, checking and terminating requests at Anubis. Nothing bot-like is making it to the actual websites, it is actual users and friendly search indexers hitting the backend now.
The one particular site in question does not utilize XenForo, as it is a game combat parser website, with an easy two to three million subpages (probably more, I've never really counted the true totals) of raw numerical data and charts spanning a bit more than a decade now. In its heyday, the site would bring in a couple thousand users per day. Fast forward into 2024 and 2025, those users have all went their own ways in life, and the site began taking in thousands of random URL connections per minute from abusive AI bots at random intervals of the day and week (previously discussed ASN's in previous posts). Thus, raking the server over the coals, and at a couple points, exhausting available bandwidth. Now, with Anubis in front of it, it has culled off the bots right at the door. Nothing AI-scraper-like is even hitting the actual site. Prior methods of ASN and mass CIDR blocks were indeed filtering out a lot of it, but not all of it. The Anubis firewall integration absolutely stopped all AI-bot-like access. I guess they don't want to pay for maximum throttle CPU time on their piddly virtual machines and such.That's awesome to hear. what do your average guest counts look like before/after?
Keep us up to date, haven't heard of anyone else where running anubis.
Sounds pretty good! Do you have any idea about the amount of false positives or distracting normal users by comfort issues with your solution?As i write this, I am seeing about two to five requests per second.
You can configure it to do just about anything on the security checks, with who is and will not be checked, or how stringent the security check will be against clients on X or Y CIDR blocks. Configuration file is a top-down-FIFO style with what is checked/actioned. If a rule declaration is not satisfied, it goes down the list to the next rule.Sounds pretty good! Do you have any idea about the amount of false positives or distracting normal users by comfort issues with your solution?
bots:
- import: (data)/crawlers/_allow-good.yaml
- name: public-paths
action: ALLOW
path_regex: ^/(?:\.well-known/.*|favicon\.ico|robots\.txt)$
- name: whitelist
action: ALLOW
remote_addresses:
- 10.0.0.0/24
- name: challenge-all
action: CHALLENGE
path_regex: .*
challenge:
algorithm: fast
difficulty: 4
store:
backend: valkey
parameters:
url: "redis://10.0.0.250:6379/42"
path_regex is where the powerhouse configuration is at. For example, on a XenForo forum, you can make it activate a check when the client goes to view a thread or forum rather everything that is XenForo. It's worth noting that it does set a cookie on the client so that repeated checks do not happen - that might collide with GDPR 'stuff' as the cookie is forced onto the client.botPolicies.yaml file in the data/ folder, so you can get a clearer idea of what's happening under the hood. I've yet to really deep dive and make Anubis be more paranoid against known bad actor IP ranges.As for user sentiment: The only complaint that I have encountered thus far is "the check took about 40 seconds to complete on my mobile, but was done in a 20ish seconds on my laptop". However, that was the same day that I implemented it and was using an elevated challenge difficulty level of 6. A level of 4 seems to be about the sweet spot. No complaints since then.Sounds pretty good! Do you have any idea about the amount of false positives or distracting normal users by comfort issues with your solution?
We use essential cookies to make this site work, and optional cookies to enhance your experience.