https://smallsheds.garden/blog/2026/on-the-acceptance-of-genai/
None of these are true if you run your own LLMs on your own hardware, using FLOSS models.
But the #MastodonHOA has deemed all AI to be abhorrent as a blanket decision.
And frankly, if you exist in a capitalist society, and you're not an owner, there is 100% chance you are exploited. The capitalist system requires it.
"Trained on stolen data". Its at best a copyright violation. And I view things like Anna's Archive and Libgen to be internationally renowned Public Libraries.
"Massaged by people in global majority countries" - yes, people work in capitalism. And guess what... You're exploited.
"Trained in environmentally harmful data centers". This assumes that training is always needed, and its not. You can train once, and run X times. Again, you're stretching to make local LLM look horrible.
And really, the rest of these are poor excuses. I won't use poop smear(anthropic), or OpenAI, or other SaaS token companies. I run local, and does not have those things you claim.
Except for the copyright issue. But again, I dont have that much respect for current US copyright.
Its at best a copyright violation
This may be true for published and public data... but that's not the only data that goes into these things. Any data that comes from breaches, users private cameras, and anything else stored with an expectation of privacy is much worse than a copyright violation.
And yes, that is a big issue with the SaaS token vendors. Claude, OpenAI, MS, and the rest do use whatever user data they can get. I am not arguing their horrific behavior.
I'm talking about locally running Qwen, or Deepseek, or other FLOSS models.
That local LLM running on my machine only sees and uses data I provide. And a control-c in the relevant console window kills the LLM.
What folks do not realize is this is #Leibniz's ultimate dream, of being able to do #calculus with words, sentences, and more. He tried to do single word-vectors, but even that had to wait for Word2Vec in 2012.
@Epic_Null @crankylinuxuser @tante “local” models are as reliant on illegal data acquisition, because they depend on the larger mainstream models to reach any level of tolerable performance. Whether it’s for training, fine tuning, distillation, or another method, that dependency means anything that goes into the development of the nonlocal model is also a requirement for the development of the local versions.
Deepseek and Qwen are no exception.
@Epic_Null @crankylinuxuser @tante
Data wants to be free. This argument simply doesn't work for those of us that have always been open data, anti copyright.
@Epic_Null @crankylinuxuser @tante
yup, so you better e2e encrypt that sort of thing
I don't care about LLMs being trained on things I want everyone to have access to because in order for everyone to have access to those things, they have to be available in a way that LLMs have access.
I'd prefer the frontier LLMs companies collapse into a black hole of capitalism but that's just because I hate corpos, not LLMs.
@komali_2 @crankylinuxuser @tante I will admit that I reserve the right to be interested in AI once the bubble bursts and it's no longer being shoved into literally everything and forcefed to everyone.
Until then though, I am a hard out.
Check out https://github.com/brontoguana/krasis and Qwen3-coder-next . 80B param model.
Also check out safetensor variants of https://huggingface.co/collections/huihui-ai/qwen35-abliterated . Gets rid of the pesky chinese laws and moral bullshit.

Krasis is a Hybrid LLM runtime which focuses on efficient running of larger models on consumer grade VRAM limited hardware - brontoguana/krasis
I dont think you understand.
I'm running a local LLM, 80b parameter, on my local machine. It doesnt even have a webserver, and I run OpenWebUI on the same machine, and use a 127.0.0.1 address to connect.
I also run an abliterated model. Means it will do anything I ask. No corpo models are abliterated.
@ratsnakegames @Epic_Null @crankylinuxuser @tante
water wants to flow downhill
@komali_2 For some reason, nobody ever brings up the other part of the quote:
On the one hand you have—the point you’re making Woz—is that information sort of wants to be expensive because it is so valuable—the right information in the right place just changes your life. On the other hand, information almost wants to be free because the costs of getting it out is getting lower and lower all of the time. So you have these two things fighting against each other.
@clayote "value" is an overinflated term. Instrumental value? Allocation value? Preference? Attachment?
On that note arguments around capitalistic value probably aren't interesting to anarchists. By all means, debate the number of dollarydoos that should be exchanged for, lol, bits on a disk
@komali_2 If you're not interested in the monetary value of data, then you're not interested in what Stewart Brand meant when he said it wants to be free, but, well, death of the author and all that
Do you value privacy at all? If so, then you might want to find some solidarity with people who've ended up sharing more than they intended to on the open web, such as a lot of the queers and sex workers on Tumblr, who objected to ArchiveTeam scraping their blogs. Inasmuch as inanimate data can be said to "want" things, the fact that their writing is now available to any interested fascist in power in the USA is what that writing wants, but it's not what the authors want.
If your anarchism has more loyalty to the rights of data than those of the people who produced it, it's shit.
I sympathize with the pain of anyone facing violence at the hands of fascists, and my commitment to put my body in the way of that violence remains the same, and remains proven. I've done it before, and I'll keep doing it.
That said this is why we've been telling people for the last 20 years that it's not a good idea to put PID info next to the kinds of things christo fascists target. We've been warning about *exactly this outcome*.
@clayote privacy delivered by law or contract is just privacy for corporations, not for people.
True privacy can only exist through encryption.
Encrypt things you want private, everything else should be free data available to anyone, including frontier AI companies who I am very much looking forward to the catastrophic collapse of.
@komali_2 Sure. Fine. It would be better if they hadn't put that info on the public internet in the first place. They did so, either out of naïvete, or because the blogging tools available to them didn't offer the grain of privacy control they wanted, and they made the pragmatic decision to risk exposure to the wrong people, in order to be read by the right people.
Now that their data is being abused by Anthropic, they're trying to do something to limit the harm, and are using the tools available to them, which are not necessarily the tools that they want. You, as an anarchist, should support them in that effort, and that means supporting them in getting their copyright enforced -- whether or not you think copyright should exist, in the abstract.
@clayote I support them attacking corpos however they please, but I argue that using copyright to do so will be at best ineffective, at worst long term harmful.
Copyright is a tool belonging to the Capital class. Its "protections" of normal people are a part of the mythology of it so that we tolerate something absurd, the idea that only certain people (read: companies) are allowed to do things with information, stories, characters.
It won't work because corpos can just ignore it, worse case.
@komali_2 There are lots of ways to use the law other than suing your enemy under that same law. For instance, the Writer's Guild of America got it written into their standard contracts that the studios can't train AI on the scripts produced under those contracts.
Case law that results from litigating those contracts is still copyright law.
@clayote I'm happy for them, but
1. LLMs are still being trained on those scripts because there's no way to catch them doing so, and if they get caught they'll get away with it by arguing it was oopsie accident here's .00001% of our profit as a fine
2. Their rights will continue being degraded in a constant battle against exploitation
3. Corpos will still foolishly try to replace them with LLMs regardless
My point is that LLMs are a symptom of a far greater problem
@komali_2 That's an agreeable point
You replied to a post that says:
Any data that comes from breaches, users private cameras, and anything else stored with an expectation of privacy is much worse than a copyright violation.
And your reply said:
Data wants to be free. This argument simply doesn't work for those of us that have always been open data, anti copyright.
I think practically everyone who reads your reply, including people like me who turn out to agree with you, will get the impression that you're uninterested in mitigating present harms to actual people's privacy.
I'm unhappy that the State has captured all the bodies that regulate privacy, I'd like them to function independently, and any data that comes from breaches, users private cameras, and anything else stored with an expectation of privacy is much worse than a copyright violation.