The reason is that alt text is scraped in order to train multimodal text + image models.
But for anyone who actually listens to the blind community, it doesn't take long to learn that these models are pretty important to a lot of people for accessibility. And regardless, alt text is directly helpful. I'm happy to put my displeasure with big tech on hold for things that genuinely improve people's lives.
@dilmandila @hosford42 @ChristineMalec @ClimateJenny
Another approach would be to write clear, descriptive alt text but use something like Nightshade to poison your image so that "AI" can't make sense of it:
I still think that's working against the public good. If the models being trained on this data are used for accessibility by the blind community, and we are intentionally working to lower their accuracy, we are hurting the blind community. We should fight the AI vendors on a front where there aren't innocent bystanders who will be hit.
Sure, that's poisoning the image itself, not for the purpose of ruining models that could generate alt text, but ruining models that could generate othe images of the same style. I say go for it.
@hosford42 Is there a way to do one but not the other, though? Do alt text-generating models and generative diffusion models use different datasets or get their data differently? 🤔 Ideally there would be NO non-consensual scraping of any websites, and models that purport to have pro-social uses shouldn't be resorting to it.
I think the core problem here is that the laws need to catch up to the reality -- and be enforced, too. Someday we will live in a world where we each have irrevocable ownership of our own data, and can choose to donate it all or in part to those causes we deem just. I have no idea how long it will take us to get there, though. Probably not our lifetimes, from the look of it.