Machine translations are often brought up as a gotcha whenever I criticize LLMs. It's worth pointing out two things: Machine translations existed decades before LLMs, and yes, machine translations are useful. However: I would never in my life read a machine translated book. Understanding what a social media post is talking about in rough terms? Sure. Literature? Absolutely not. Hell, have you ever seen machine translated subtitles? It's absolute garbage.
I have the impression that primarily anglophone people don't read as much translated literature, because so much good literature already exists in their language, so this issue may not be as familiar within that demographic. As someone who did not grow up anglophone, I can tell you there is a world of difference between a good and a bad translation even when done by humans. Machine translations are not even on the scale.
From what I've observed, people who claim that LLMs can replace artists don't understand art, people who claim that they can replace musicians don't understand music, people who claim that they can replace writers don't understand literature, and people who claim they can replace translators don't rely on translations. If I had a button that would erase LLMs from the world but it would take machine translations away (which is a false dichotomy anyway), I would absolutely still press it.
Technology is not inevitable. We've decided not to have asbestos in our walls, lead in our pipes, or carginogenic chemicals in our food. (If you're going to argue that it's not everywhere, where would you rather live?) We could just not do LLMs. It's allowed.
@Gargron It is a technology that humanity has been seeking for a long time. At least since the 1950s, with Turing and his colleagues.
@df No, this is marketing. OpenAI, Google, Anthropic &co want you to believe that what they're doing is artificial intelligence. My professional opinion is that LLMs are a dead end technology to creating actual intelligence. And if any of those companies did create actual intelligence for the purposes they pursue, it would be slavery, for which I cannot advocate.
@Gargron LLMs are not exclusively a product of large corporations or just marketing. Much of the research and development also takes place in open source and academic communities. The codes for these LLMs are public and can be audited or run locally. Furthermore, I argue that serious ethical reflection is necessary, but prohibition is not the way forward.
@df
Consciously not using something ≠ prohibition
Edit: Also, who cares who worked/ envisioned or works on this now? If you think about LLMs enough, you will likely see enough good arguments about the resource waste, centralization of power and multiplication of slop which describe LLMs. We lived without it before and we can live without it in future times.
@df @Gargron Academics may study LLMs out in the open, but I don't think academia has been able to produce LLMs whose outputs are sufficiently marketable compared to the current commercially available ones. Because the first "L" ("large") is - in our current, limited understanding - crucial for the verisimilitude of the synthetic text, and only corporations (and governments, but they mostly haven't gotten to this yet) have the scale to get large enough for that so far.

@joshuagrochow and the lack of moral compass or publicly stated ethicals standards that would allow university employees to steal large enough sets. small sets of text are read and understood by humans who can, far more efficiently, apply appropriate prior written and other formats of source material to a specific use case.

programming a calculator only makes reasonable sense if the computation requires enough repetition to warrant the resources used in building it, or it's a closed set without novelty... like, for example, a numerical calculator. ; )

edited for typos and clarity: it was killing me, apologies for the notification disruption.

@df @Gargron

@Gargron they'll never create intelligence because intelligence requires will and they do not understand will. they dont even posses one of their own: their own behaviour is driven by feelings and shaped by a commercial playbook. there is zero chance they will ever create intelligence.

@df @Gargron

Transformers are neural networks.

LLMs are transformers wrapped in some Python scripting.

Every neural network can be accurately represented as an Excel sheet, even if it ends up having billions of cells.

Since it's just addition and multiplication, the model is fully deterministic. Same input, same output. Not intelligent.

It's Python code that does probabilistic sampling of the output. It's just a few lines of well-understood math plus a dice roll. Again, not intelligent.

@df @Gargron To be clear, “Python” is a placeholder language, it can be Rust, or it can be a GPU shader, and it changes nothing.

@patrys @df @Gargron does determinism imply non-intelligence?
If you hooked up the computer to a Geiger counter for true random noise and used that to modulate the output, would that have any bearing on its intelligence?

Or from the other side, what makes you think our brains are non deterministic, and why does that make us more intelligent than if the exact same history and sense-data always produced the same response?

@FishFace @df @Gargron If it’s deterministic, it can be unrolled into a giant lookup table. Did we kill phone books because they were on the verge of achieving AGI?

To me, intelligence implies a lot of things, like being able to form higher-order abstractions, learn, and thus remember things (no, being passed your “memories” as part of every prompt does not count). It also implies being curious.

@patrys @df @Gargron given that the lookup table would generally be infinite, I don't even see what that would have to do with anything. What about the Geiger counter?

I don't think those things are really needed for human-like intelligence, and something like curiosity can easily be simulated by a rules-based system.

@FishFace @df @Gargron No, you got it wrong. The model itself can be unrolled into a finite lookup table. The only random part is which word you take from the few options in the resulting row.
@patrys a computable function can generally produce infinitely many different outputs. You're still not saying why a non-deterministic part affects intelligence.
@FishFace It generates one token at a time, which makes it impossible to formulate higher-order abstractions that are not already baked into the weight matrix. I said it in another answer, not being able to learn disqualifies it as intelligence.

@patrys LLMs are intelligent only in the sense of pattern recognition; that is, they possess logical intelligence. However, some psychologists argue that there are multiple intelligences that cannot be reduced to logic, nor are LLMs capable of possessing them. See psychologist Howard Gardner.

@FishFace @Gargron

@df @FishFace @Gargron This pattern recognition is an artifact of the training process, not something that occurs at inference time. It’s like having termites dig some tunnels in an earth mound, then removing the termites, pouring aluminum into the mound, and attributing the resulting intricate shapes to the intelligence of the mound. The patterns it carries are from human artifacts used as input for the model before its weights settled.

@patrys Undoubtedly, LLMs in this regard end up being mirrors of who we are, reflecting our biases, our prejudices, and our worldviews. That is why they are not innocuous tools and why #ethics and regulation of #AI are necessary.

@FishFace @Gargron

@FishFace @patrys @df @Gargron

"Or from the other side, what makes you think our brains are non deterministic"

Us having free will/being non-deterministic is pretty much the base assumption we all operate on to even be able to function as humans. That of course doesn't mean that it's automatically true, but it makes the question of why do you think your brain is non-deterministic a no-brainer to answer: because we can't help but perceive ourselves as such.

@frog_reborn @FishFace @df @Gargron The very fact that you can read and mid-sentence learn something that changes your perception of the world means that you have brain plasticity that no neural network possesses. It’s deterministic AND rigid because training and inference happen separately.
@patrys you're talking about differences between brains and neural networks that exist, but still not arguing the philosophical point about why that is relevant to intelligence.
@df @Gargron Turing did not dream of spending the entire energy budget of the world at the time so people could generate letters from a few bullet points and the recipients could summarise them to different bullet points.

@Gargron

LLMs are Shannon 1948 as far as the theory goes (building on Markov, but adding computer technology). With some compression techniques.

But I think you're talking about something else entirely, not purely syntactical.

@df @Gargron

A small section of humanity. Not everyone.

@df @Gargron No, it is a fake, an emulation of what we have been seeking but not the real thing.
@Gargron while all your examples are 100% valid, I seriously question whether we would be able to manage to do that today. With the utter shambles most democracies are in currently, multi-national Corporations can run roughshod on environmental protection, worker safety, child protection and just about everything that past generations fought hard for.

@DJGummikuh

imagine for a moment, the billionaires have been beheaded and the yachts sunk into the sea. the value in the output of workers 100% reinvested into local communities. all of it. none for colonial masters far away. the 20 hour work weeks and all human workers hands full of the satisfaction their efforts are meaningful... no more busy work for shareholders to skim value out of. only meaningful work. custom artisanal everything. housewares repaired by local handicrafters. clothes sewn and tailored to each body. homes and townhomes and communal living spaces built and maintained by cooperative owners. neighboring towns and regions and nations translating with loving care between the communities of meaning... interconnected with care. 💜

@Gargron

@melioristicmarie @DJGummikuh @Gargron That's a dreamy vision. Thank you. I love it.
@mason
hanging on to the hope we can survive as a species and get to the good stuff of loving each other up, bigly. ;
@DJGummikuh @Gargron

@melioristicmarie

And that lasts 1-2 generations before new people who don't understand the problems that lead their parents to create the paradise chafe under their constraints and begin changing the system to something its originators wouldn't like, this creating conflict, diversity of thought, and continuing the cycle of history.

See: reality.

@TheServitor hm. so you do not believe in evolution then?

you ignore the myriad plants and beasties with behavior change documented in the geological record? or the changes in human written language documented over millennia? or dr. martin luther king's statement on the arch of the moral universe? or even the historical shifts in graphic novels and comic book tales over 80 or so years?

weird.

nihilism has never had much sway with me. i accept humans believe such a framework applies beyond the self and assert omniscient determinism universally. however, i do not abide such a concept as the universal human. we are varied beasties and i have faith in evolution of arrangements of living matter as well as of patterned practices.

see: the data.

@TheServitor of course, the u.s. and israeli "leadership" could destroy the species in the next week and whatever survives the nuclear fallout would have a very different future. humans are weird.
@DJGummikuh @Gargron I'm sure the oldschool labor activism methods would still work.
@DJGummikuh @Gargron I would argue that LLMs shouldn’t require special regulation, just no exemptions from ordinary regulation. A lawyer who files a specious brief full of fake citations should be sanctioned the same regardless of whether an LLM wrote it.

@Gargron would you know if you've seen a good outcome of an LLM? You'd somehow be able to identify when the LLM got it right?

I assure you you've experienced good LLM output and don't even know it. Because that's what good LLM output looks like. Indistinguishable from human output.

Your examples are perhaps false equivalencies. Take asbestos. We didn't abolish insulation. We developed better, safer insulation. We didn't stop dying food colors, we just developed safer dyes etc.

@Gargron ultimately LLMs like any other software is a tool. It's all about how a human uses them.

Lets take photoshop as an example. Humans generate vast amounts of garbage photoshopped images. Ever been to deviant art?

And yet the same tool is used by professionals all day every day to create stuff we like and enjoy.

The same applies to LLM use, and back to my first reply. What you lament is low quality output a human shared. Meanwhile the tool gets used masterfully to great effect elsewhere

@Tekchip
maybe these toots are slop output? philosophically average, without comprehension of qualitative significance.

there is no value in the average. it is only in the deviations that standards exist. and these toots... are valorizing mediocrity. sad.

@Gargron

@melioristicmarie @Gargron okay, sure, maybe my toots are slop. So...what? I should be deleted as the OP implied of LLMs? How about all those folks making mediocre art over on deviant art? Disappear them too?

That's a bit of a moral quandary isn't it? How do we get the good stuff without the bad stuff?

You stand to make the world perfect if you can figure that out. I look forward to it.

@Tekchip

human, is fine. perfection is a scam sold by ponzi schemers who have no useful skill. second sons of the british empire looking for some purpose that makes daddy approve of their existence.

maybe... just maybe... talk to a human and ask them how you can help them, with your actual meat space body. then maybe you could find some meaning in life instead of trying to get techbros to think you are pretty.

@Gargron

@Tekchip my walls are full of art by humans that some would call terrible... who the fuck cares? they have love and craft and pain and power from the hands and soul of a human creator. they are beautiful. i fucking love bad art.

slop generation is the nothingness.

just write your toot from your heart, fuck the machine. being human is fine.
@Gargron

@ClipHead @melioristicmarie @Gargron which this?

"there is no value in the average."

or

"my walls are full of art by humans that some would call terrible... who the fuck cares?"

Can't have it both ways.

@Tekchip
so... is this a slop account? am i tooting with cheapgpt?

are you a human playing with toys you do not comprehend?

dear dogs, may i have the confidence of a mediocre "white" man.

so... l.l.m.s tokenize english text... and then calculate an average.

humans making shitty art is qualitatively perfection in comparison to word salad from a calculator. when you enter this into wannabe deep seek... i will be waiting with bated breath for the token response. ; )

@ClipHead @Gargron

@melioristicmarie @ClipHead @Gargron lol are you an LLM or just don't care to review my profile? Shoot even do a google search. I'm easy to find. Seems like you've lost the plot.

@Tekchip
There's no point in explaining, if you don't get "this", tbh.

@melioristicmarie @Gargron

@ClipHead sigh. i agree. there are more porous space to beat my head against. 💜

@Tekchip @Gargron

@melioristicmarie @Tekchip

There's also the problem of your essentialist thinking that decides only terrible human beings could find any value in LLMs, because use of LLMs is proof of same. QED. It's like thinking poor people must be morally bankrupt. It's a non-sequitur.

Anti-LLM posts quickly turns to deep pronouncements about the personality and motives of people who do things you don't like, not an honest discussion of the harms of the tech, because it's a purity test, not a position.

@TheServitor another cheapgpt word salad... bankrupt of meaning.

quality... qualitative difference.

those who boost calculator speak are emotionally deficient. as a society, in overdeveloped regions running the rat race of shareholder value, snake oil salesman is the job that the most vulnerable fall into, the men who have lost the plot likely because disruptive masculinity has violently warped their soul; the norm, the average, the most common... is what those deficient in humanity, lacking sufficient love and care and sorrow and compassion and support and acceptance and meaningful connections to other living beings... equate to the ever pure whiteness of power. the diagram so empty of illustration as the vastness of white space between concepts is preferred over the density of interdependent contexts.

existence is complex, and the snake oil salesmen chose to call it complicated and unnecessary, to erase the necessity of that which is incomprehensible... to sell a share in bullshit machines.

@Tekchip

@Tekchip @Gargron photoshop doesn't require stealing all the reachable content in the internet (and then claim it's fair use to make barely average commercial derivatives from it).

@cygnathreadbare @Gargron yeah, that's a garbage way this technology has been developed. Unfortunately if we threw away every technology built on the back of people doing bad things we wouldn't have much technology, unfortunately.

I don't fault lamenting how it's come to be and even how it's used broadly. But claiming it's useless because some folks use it poorly isn't really accurate indicator of the technologies usefulness.

@Tekchip @Gargron I worry that you are only looking at the local output: what the screen shows you. What are the externalized costs? The hidden costs?
@Tekchip @Gargron The technology doesn't yet do much of what is claimed for it; and it is already expensive in terms of externalized costs: memory, energy, water. It really looks like the future of LLMs depends on mass acceptance of the "what-if" scenario - those hoped-for advances where it works better, uses less energy, and somehow doesn't wipe out thousands of middle- to low-level jobs.

@Tekchip @Gargron

Let me ask you this: It's your birthday.
5 of your friends met some days before and wrote a song for you. It's not really good, the text doesn't even rhyme...but they did this for you and they had fun.
They enjoyed the act of creating.

5 other friends wrote a prompt and pressed a button to generate a song.

Which song will you remember?

@ClipHead @Gargron it depends. Was the song written by prompt also delivered by my friends? If yes, then I'd enjoy it just as much.

Is it any less valid than a mass reproduced pre-written card that a friend, who I know is busy, still made the time to buy for me?

@Tekchip
No, they just gave you the song.

They had the possibility to meet and write a song, but chose not to.

Are you making excuses now for "fake" songs...or fake friends?

@Gargron

@Tekchip @Gargron the tiny potential for very rare good outcomes are not worth the constant poisoning of humanity's collective information corpus.

For every "good" generated content there are dozens of thousands of terrible slop that are difficult to separate from genuine useful information or material when doing research or code reviews, etc.

Not to mention that these "good" outcomes are much costlier to humanity than creating by hand, with no benefit.

@Kiloku @Gargron the problem is you want to assume they are rare outcomes. I don't believe they are. Unfortunately that's where we're at an impasse. It's literally impossible to measure the good outcomes.

I agree the environmental outcome is terrible. I don't like that part. What we can look forward to is the technology improving. General computers used to use WAY more power than they do now. The same is going to happen with LLM technology. Hopefully sooner than later. Folks are working on it.

@Tekchip @Gargron I *know* they are rare.
@Kiloku @Gargron Please let the rest of us know how to tell when we've seen a good LLM output. Seriously, if we can all tell the good and the bad then we can start gathering some data to have an even more rational conversation.
@Tekchip @Gargron (also, most of what "AI" boosters *think* is good generated content is actually laughably bad to anyone who knows the subject matter of the content it generates. I'm certain you've shared something that you thought was indistinguishable from human created content that other people knew and saw a bunch of problems with as soon as they examined it further than a cursory glance)