Machine translations are often brought up as a gotcha whenever I criticize LLMs. It's worth pointing out two things: Machine translations existed decades before LLMs, and yes, machine translations are useful. However: I would never in my life read a machine translated book. Understanding what a social media post is talking about in rough terms? Sure. Literature? Absolutely not. Hell, have you ever seen machine translated subtitles? It's absolute garbage.
I have the impression that primarily anglophone people don't read as much translated literature, because so much good literature already exists in their language, so this issue may not be as familiar within that demographic. As someone who did not grow up anglophone, I can tell you there is a world of difference between a good and a bad translation even when done by humans. Machine translations are not even on the scale.
From what I've observed, people who claim that LLMs can replace artists don't understand art, people who claim that they can replace musicians don't understand music, people who claim that they can replace writers don't understand literature, and people who claim they can replace translators don't rely on translations. If I had a button that would erase LLMs from the world but it would take machine translations away (which is a false dichotomy anyway), I would absolutely still press it.
Technology is not inevitable. We've decided not to have asbestos in our walls, lead in our pipes, or carginogenic chemicals in our food. (If you're going to argue that it's not everywhere, where would you rather live?) We could just not do LLMs. It's allowed.
@Gargron It is a technology that humanity has been seeking for a long time. At least since the 1950s, with Turing and his colleagues.
@df No, this is marketing. OpenAI, Google, Anthropic &co want you to believe that what they're doing is artificial intelligence. My professional opinion is that LLMs are a dead end technology to creating actual intelligence. And if any of those companies did create actual intelligence for the purposes they pursue, it would be slavery, for which I cannot advocate.
@Gargron LLMs are not exclusively a product of large corporations or just marketing. Much of the research and development also takes place in open source and academic communities. The codes for these LLMs are public and can be audited or run locally. Furthermore, I argue that serious ethical reflection is necessary, but prohibition is not the way forward.
@df @Gargron Academics may study LLMs out in the open, but I don't think academia has been able to produce LLMs whose outputs are sufficiently marketable compared to the current commercially available ones. Because the first "L" ("large") is - in our current, limited understanding - crucial for the verisimilitude of the synthetic text, and only corporations (and governments, but they mostly haven't gotten to this yet) have the scale to get large enough for that so far.

@joshuagrochow and the lack of moral compass or publicly stated ethicals standards that would allow university employees to steal large enough sets. small sets of text are read and understood by humans who can, far more efficiently, apply appropriate prior written and other formats of source material to a specific use case.

programming a calculator only makes reasonable sense if the computation requires enough repetition to warrant the resources used in building it, or it's a closed set without novelty... like, for example, a numerical calculator. ; )

edited for typos and clarity: it was killing me, apologies for the notification disruption.

@df @Gargron