To stop AIs scraping your website for content, add this to your robots.txt file on your website.

Thanks to Neil Clarke for most of these.

User-agent: CCBot
Disallow: /

User-agent: ChatGPT-User
Disallow: /

User-agent: GPTBot
Disallow: /

User-agent: Google-Extended
Disallow: /

User-agent: Omgilibot
Disallow: /

User-Agent: FacebookBot
Disallow: /

@patricksamphire I’m using Cloudflare with “bot fight mode” to block most of the ones that ignore the robots.txt. What you might try is the reverse, add the bots you want to index your site with the “Allow” directive, then at the bottom set a wildcard disallow.
I also have my sites geo locked to only the US and Canada as that is where my target audience is located, but Cloudflare will allow you to allow or block any country.