![Re: Re[2]: Correctly dealing with bots and scrapers.](/storage/media/uploads/article/12852/main/6878c2aa69286-nanog-img.png)

nanog mailing list archives
From: "Compton, Rich via NANOG" <nanog () lists nanog org>
Date: Wed, 16 Jul 2025 21:01:53 +0000
I don’t have experience with it, but I was pointed to this project recently which may be of interest to you. “Anubis is a Web AI Firewall Utility that weighs the soul of your connection<https://en.wikipedia.org/wiki/Weighing_of_souls> using one or more challenges in order to protect upstream resources from scraper bots. “ https://github.com/TecharoHQ/anubis From: Ryland Kremeier via NANOG <nanog () lists nanog org> Date: Wednesday, July 16, 2025 at 2:08 PM To: North American Network Operators Group <nanog () lists nanog org> Cc: Ryland Kremeier <rkremeier () barryelectric com> Subject: Re[2]: Correctly dealing with bots and scrapers. I've thought about using client-side tokens with behavioral analysis and a points system, but haven't implemented it in one of my sites to test it's reliability just yet. But basically issue a signed token to each client and watch that client using js instantiated on each page of the site. Give actions (or lack thereof) scores based on likelyhood to be a bot and fail2ban them past a certain threshold. look at mouse movement, scroll depth, keypress timings, page transitions, dwell time, js events, start/end routes, etc. - so if a client is on a page for under 100ms, it accumulates 'bot' points, say +10 pts - no EventListener JS events +40 pts - client joins site at a nested page and drills down +5 pts (low because people could bookmark a specific page that triggers this) - etc. then decay these points based on the opposite effects happening. There's no guaranteed solution and this would require careful tweaking of the ban threshold and whatnot with real testing to not accidentally block your users, but I feel it could mitigate a lot of bot situations. For banned users just redirect them to a page with a little unban form they can quickly fill out or something. -- Ryland ------ Original Message ------ From "Tom Beecher via NANOG" <nanog () lists nanog org> To "North American Network Operators Group" <nanog () lists nanog org> Cc "Tom Beecher" <beecher () beecher cc> Date 7/16/2025 2:43:51 PM Subject Re: Correctly dealing with bots and scrapers.
As Chris states, broad IP based blocking is unlikely to be very effective , and likely more problematic down the line anyway. For the slightly more 'honorable' crawlers, they'll respect robots.txt, and you can block their UAs there. Fail2ban is a very good option right now. It will be even better if nepenthes eventually integrates with it. Then you can have some real fun. On Wed, Jul 16, 2025 at 3:39 PM Andrew Latham via NANOG < nanog () lists nanog org> wrote:Chris Spot on, and I am getting the feeling this is where the value to a geo-ip service comes to play that offers defined "eyeball networks" to allow. On Wed, Jul 16, 2025 at 12:57 PM Chris Adams via NANOG <nanog () lists nanog org> wrote: >> Once upon a time, Marco Moock <mm () dorfdsl de> said: > > Place a link to a file that is hidden to normal people. Exclude the > > directory via robots.txt. > > > > Then use fail2ban to block all IP addresses that poll the file. > > The problem with a lot of the "AI" scrapers is that they're apparently > using botnets and will often only make a single request from a given IP > address, so reactive blocking doesn't work (and can cause other issues, > like trying to block 100,000 IPs, which fail2ban for example doesn't > really handle well). > -- > Chris Adams <cma () cmadams net> > _______________________________________________ > NANOG mailing list > https://urldefense.com/v3/__https://lists.nanog.org/archives/list/nanog () lists nanog org/message/AFJF4UQJZW6ALTY6SA7OHBN2AZC72SZQ/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-NgPkVX1Vg$<https://urldefense.com/v3/__https:/lists.nanog.org/archives/list/nanog () lists nanog org/message/AFJF4UQJZW6ALTY6SA7OHBN2AZC72SZQ/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-NgPkVX1Vg$> -- - Andrew "lathama" Latham - _______________________________________________ NANOG mailing list https://urldefense.com/v3/__https://lists.nanog.org/archives/list/nanog () lists nanog org/message/DHUYTBIXFMWE2KWC5NKCR7AJIWPYUL4E/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-NiX3zC-RA$<https://urldefense.com/v3/__https:/lists.nanog.org/archives/list/nanog () lists nanog org/message/DHUYTBIXFMWE2KWC5NKCR7AJIWPYUL4E/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-NiX3zC-RA$>_______________________________________________ NANOG mailing list https://urldefense.com/v3/__https://lists.nanog.org/archives/list/nanog () lists nanog org/message/ECB77Z6SVDZZ6SZ5YGWP4YJ5HVX6KRQE/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-Ng-8Wl2SQ$<https://urldefense.com/v3/__https:/lists.nanog.org/archives/list/nanog () lists nanog org/message/ECB77Z6SVDZZ6SZ5YGWP4YJ5HVX6KRQE/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-Ng-8Wl2SQ$>
_______________________________________________ NANOG mailing list https://urldefense.com/v3/__https://lists.nanog.org/archives/list/nanog () lists nanog org/message/P5QTTOVSM3J73OSNMDW2MALXA7BIXZQV/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-Ng4AKhkpw$<https://urldefense.com/v3/__https:/lists.nanog.org/archives/list/nanog () lists nanog org/message/P5QTTOVSM3J73OSNMDW2MALXA7BIXZQV/__;!!CQl3mcHX2A!HHKMxs_ccAQOJOwC_YKtrF7G1WneCnCa-eVa1fOtm7gtXNEQmCfoGSNzwbBbBHoF5dI02OqwcRW4-Ng4AKhkpw$> _______________________________________________ NANOG mailing list https://lists.nanog.org/archives/list/nanog () lists nanog org/message/PNUTJVLETF4JYFM7ECJKKK5XFNQ5OCLA/
Current thread:
- Correctly dealing with bots and scrapers. Andrew Latham via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Marco Moock via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Chris Adams via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Andrew Latham via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Tom Beecher via NANOG (Jul 16)
- Re[2]: Correctly dealing with bots and scrapers. Ryland Kremeier via NANOG (Jul 16)
- Re: Re[2]: Correctly dealing with bots and scrapers. Compton, Rich via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Chris Adams via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Marco Moock via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. William Herrin via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Marco Moock via NANOG (Jul 16)
- Re: Correctly dealing with bots and scrapers. Constantine A. Murenin via NANOG (Jul 16)
Related Articles
Stay Informed
Get the best articles every day for FREE. Cancel anytime.