With all the buzz about AI models using publicly available content without permission, I've been wondering if there's any effective way to prevent these models from scraping or training on content from my website. Is there a technical solution or standard that can actually block this?
4 Answers
Anubis uses a proof of work algorithm to fend off those pesky AI web scrapers. It seems like a worthwhile tool to consider if you want to add some extra protection.
Honestly, you just have to accept that anything you post online is essentially public now. Embracing this reality may actually lead to better security practices.
Not really. You could try blocking user agents, but they can easily change if the operators are determined. IP blocking might work, but it also requires constant monitoring and updating to keep up with changes. You’ll likely just end up dealing with new scrapers as they pop up, so it's a pretty tough battle to win.
Check out Anubis! It can at least make things difficult for scrapers. There’s no foolproof way to block them entirely, but this might help mitigate some of their efforts.
Related Questions
Cloudflare Origin SSL Certificate Setup Guide
How To Effectively Monetize A Site With Ads