For the first time in my life on Masto, I read alt text that made me wish the person hadn’t bothered. After a wordy post, they wrote in the alt text field something like, “I don’t really want to write alt text and train the AI for free, but okay, person holding a computer.” I’m going to assume they didn’t think about what a slap in the face this was to me, and possibly to anyone else who relies on alt text. 3 Times I started an acerbic reply and deleted it, but I’m still seething, so I’m asking anyone who is so disingenuous to just not bother; you’ll create less hard feelings that way.
@ChristineMalec That’s weird. Why do they think it has anything to do with training AI? Maybe I’m just ignorant, but I thought alt text was for humans.
@ClimateJenny No idea, goodness knows the body of their post had enough verbage in it.

@ChristineMalec @ClimateJenny

The reason is that alt text is scraped in order to train multimodal text + image models.

But for anyone who actually listens to the blind community, it doesn't take long to learn that these models are pretty important to a lot of people for accessibility. And regardless, alt text is directly helpful. I'm happy to put my displeasure with big tech on hold for things that genuinely improve people's lives.

@hosford42 @ChristineMalec @ClimateJenny In that case, I think writing subjective alt text would be a solution for those who fear it trains AI. Something like, "I took this pretty flower near sunset, there's a bird in the background spoiling my perfect view of the sunset and the flower is..." I like writing my alt text like that, not thinking of AI but because I suck at describing images accurately and so put in my feelings, which I suspect would be worthless to AI?

@dilmandila

Maybe, IDK. It depends on what they're training the model for, and how they're using it. But do we really want to make it unusable for training the model, when those models are then used for accessibility as well?

@ChristineMalec @ClimateJenny

@hosford42 @ChristineMalec @ClimateJenny Good question. I find it useful for some tedious work like creating sub-titles, some vfx tasks (I use offline models that run on my PC rather than on cloud). It has some good and practical uses, but it has stirred a lot of mistrust. Perhaps after this toxic hype goes away and the dust settles, and if we get tools that run on local devices rather than environmentally unfriendly Clouds, well, people would adopt them more widely, and contribute data freely.

@dilmandila
I think alt text should focus on what you want to communicate with the image, not some "objective" description of what's in it.

I like it when people do that even though I'm seeing, sometimes it helps me when I find it cryptic why people posted the image.
@hosford42 @ChristineMalec @ClimateJenny

@dilmandila @hosford42 @ChristineMalec @ClimateJenny

Another approach would be to write clear, descriptive alt text but use something like Nightshade to poison your image so that "AI" can't make sense of it:

https://nightshade.cs.uchicago.edu/whatis.html

@CppGuy

I still think that's working against the public good. If the models being trained on this data are used for accessibility by the blind community, and we are intentionally working to lower their accuracy, we are hurting the blind community. We should fight the AI vendors on a front where there aren't innocent bystanders who will be hit.

@dilmandila @ChristineMalec @ClimateJenny

@hosford42 @CppGuy @dilmandila @ChristineMalec @ClimateJenny I wish there was a way to poison it for some models and not for others. I'm all for alt text generators if they're trained on less data than the typical LLM and are used explicitly for the purposes of accessibility. Unfortunately, the alt is almost definitely also being scraped by LLMs and image generators. But the way to solve that problem is absolutely not to neglect those who depend on alt text, increasing the need for generators

@hosford42 @CppGuy @dilmandila @ChristineMalec @ClimateJenny

You've hit the nail on the head here.

When we fight big tech, which we should, we shouldn't be fighting to destroy accessibility. like you said, it hurts innocent bystanders.

People often don't realize what IS accessibility, and so I wrote a zine to break it down and try to teach folks, so they don't do harmful stuff like poisoning datasets used specifically for improving accessibility. https://reshapingreality.org/2025/06/17/accessibility-zine-completed-and-ready-for-download/

@hosford42 @dilmandila @ChristineMalec @ClimateJenny I think calling image poisoning harmful to accessibility is a bridge too far, though. Should artists hand over their online art for free to tech companies for the sake of accessibility? Surely there's a way to aid accessibility without giving everything to scrapers indiscriminately. (For clarity I'm not defending alt text tampering, which I agree is a jerk move.)

@ljwrites

Sure, that's poisoning the image itself, not for the purpose of ruining models that could generate alt text, but ruining models that could generate othe images of the same style. I say go for it.

@dilmandila @ChristineMalec @ClimateJenny

@hosford42 Is there a way to do one but not the other, though? Do alt text-generating models and generative diffusion models use different datasets or get their data differently? 🤔 Ideally there would be NO non-consensual scraping of any websites, and models that purport to have pro-social uses shouldn't be resorting to it.

@dilmandila @ChristineMalec @ClimateJenny

@ljwrites

I think the core problem here is that the laws need to catch up to the reality -- and be enforced, too. Someday we will live in a world where we each have irrevocable ownership of our own data, and can choose to donate it all or in part to those causes we deem just. I have no idea how long it will take us to get there, though. Probably not our lifetimes, from the look of it.

@dilmandila @ChristineMalec @ClimateJenny