In one sense, I like the idea of human.json creating a "web of humanity" ( https://codeberg.org/robida/human.json )
But the reason I won't be joining in is:
- AI scrapers are ruthless and well known for ignoring robots.txt restrictions.
- The high value web content is the human-written content, to avoid model collapse risks (training LLMs on LLM content).
- So providing a network of human content will be like gold dust for AI companies. I bet they're rubbing their hands with glee.
