Saying that you do not want GenAI in the #books you read, the things you watch or the games you play is an understandable and NORMAL position. Maybe they have ethical concerns, maybe they love their artist homies. Maybe they don't like the bland garbage that AI generates. Stop framing this like an horde of neoluddites is starting the Butlerian Jihad (would be fun doh) just because they do not want to follow a romance autor who has a computer shitting novels instead of writing them herself.

Look, we do not have a lot of ways to avoid that shit in the workplace lately. But people like to select their own fun.

Books about AI are fun.

Books made by AI are not.

And please spare me with the "both sides" argument. Because one of them is trying to force feed things to the other. And one of them has all the money and resources and the other has not. This is not people taking sides. This is people trying for the boot to stop pressing against their face.

You keep mentioning fear of the machines.

Not about that.

You keep mentioning grammar checking and transcription functionality.

Not about that.

Look, we get it. AI sounds hot. We read the same science fiction books and it is nice to think that maybe one day LLM technology can be leveraged against oppression. News about fake open models are fun in a sense because every time one of those pops up, some idiot is going to lose millions yadda yadda. But you are missing a very important point here: a permission structure is being built around us, and stopping it is absolutely crucial
Every time you do a "both sides" stuff between "AI hypers and deniers" you are basically telling me that the person worried about the destruction of their life, their job and the environment has the level of delusion of a person like Peter Thiel, an eldritch horror in a vessel made of flesh that thinks humanity, umm, should not exist.
I think you're misrepresenting the "both sides" people. Sure, some are just like the "just asking questions" people, but the LLM/AI discussion has been derailed.

The LLM peddlers are hyping up their machines as general AI that is perpetually 2 months away from achieving sentience. It probably started as naive optimism and has devolved into a need to keep stocks up to justify insane bubble-like valuations.

Simultaneously, people against it insist it's nothing more than a T9 dictionary without any merit or use. That is a false dichotomy and IMO disingenuous as it gives the LLM peddlers an easy counter: LLMs are more than prediction machines, even if they are not the super-intelligences the other side claims.

LLMs (and similar generative neural networks) are useful for certain things. Photoshop is augmented with tools for cleaning up images. If you pay "a real artist" to make an image "without AI," you can be 100% certain they have used "AI" as part of their process. Textual LLMs can be useful to get a starting point or reference for research, or to massage text.

"Both sides" is not saying that both the people that underplay and those that overplay the abilities of LLMs are right. It's saying they are both wrong and that LLMs can have uses even if they are not general thinking machines.

Refusing opinions other than "LLMs are useless for everything" allows people to (rightfully) ignore actual good points about ethics in training.

@michael @berniethewordsmith What are LLMs, then, if not merely next-token prediction? Obviously they're more sophisticated than a markov chain, but if they're more than prediction, *what* more?

Also you're flat-out wrong about artists. A great many of them hate AI enough that they won't use it on principle, and avoid tools that might sneak it into their workflow without their consent.

Technically, LLMs of course just generate text one token at a time. But it's an over-simplification that somebody just loaded it up with reddit, project Gutenberg, and PirateBay and use that to generate text similar to what it has seen. As you say, an advanced Markov generator.

LLMs do get trained on all that more or less ethically obtained text, but that's just step 1. Next is fine-tuning and last reinforcement learning. Reinforcement is what makes the difference: the model is trained not just to produce text it has seen, but text that makes the trainer happy.

Part of what makes the trainer happy is that the answer is correct (or reinforces their existing bias). Sometimes that's more or less verbatim quoting a reddit shitpost, sometimes that's picking out points from an academic paper, and sometimes that's just random gibberish formatted so it looks right.

But at the core, an LLM has 3 things: a model of what language looks like, a non-trivial percentage of human knowledge in the form of the internet, and an incentive to provide answers that made the trainer happy.

We do not know what that model looks like, except in very simple cases. It works surprisingly well for what it is but is not intelligent.

People didn't hate Google Translate or image search before it was rebranded as AI. It uses simpler versions of LLMs. Artists often make use of Photoshop features like "select subject" or "context-aware fill," which are run by neural networks. There is much more to AI than just generative AI and parts of it are useful.

The problem is not LLMs (a priori, ignoring the ethics of how they are trained), it's the AI companies having to hype them up. IMO, the correct response is not claiming they are not useful for anything – they obviously are to a lot of people – it's to challenge what they are used for.