The funny thing about LLMs is theyโ€™re not good for knowledge work because they sometimes make up stuff that doesnโ€™t exist and theyโ€™re not good for creative work because they sometimes make up stuff that does exist.
Turns out this was a banger of a toot. Do we do soundclouds on #mastodon?
@dlevenstein
Legally obligated to link to the funkwhale instance with one track uploaded
@dlevenstein Yeah, what's your ... castapod, funkwhale, peertube or something? Patreon? Favorite charitable cause?
@dlevenstein and then LLM ppl will say "RAG fixes this," but then the problem becomes RAG-LLMs are not good for knowledge work because it's impossible to tell when they are making stuff up, and they're not good for creative work because it's impossible to tell when they are making stuff up.
@dlevenstein that is such a smart encapsulation of the problem that it implies a useful solution. It might be possible to have a dial or switch to tell these programs whether you want more facts or more fiction, more accurate portrayals of prior work or a novel creation.
@guacamayan I mean that would solve it, sure, but there's no reason to think it can possibly be built. "Fact" and "fiction" aren't concepts that LLMs understand โ€” all they know is which words in which order are shaped most like the text in their training corpus. That's it. It's like asking a lottery machine to preferentially choose odd or even numbers โ€” it can't be done, because the balls are mechanically identical.
@andrewt @guacamayan I think it's absolutely crazy that people are using AI for things like writing contracts.
@guacamayan @dlevenstein maybe what we need is masters level knowledge of how to write better queries? Assuming we could get accurate or creative answers based on asking the talking data base more rigorously defined questions?

@dlevenstein

It's usually good for coming up with marketing copy. That's what my wife uses it for....

...as even human written marketing copy is rarely accurate.

@dlevenstein AI can't even draw a correct Rubik's cube in a solved state. A regular 3x3x3 Rubik's cube, where each side has a different colour. It doesn't seem to be capable. I gave it a few tries:

https://social.linux.pizza/@realestninja/111624206179182110

Nikhil ๐Ÿง (@[email protected])

Attached: 1 image I didn't expect AI to have such difficulties drawing a Rubik's Cube. This picture is the position that I want AI to draw the cube: Basically a (solved) cube with one side rotated by 45ยฐ. So I tried a couple of times but noticed that the AI doesn't even manage to draw a regular Rubik's Cube where there is no rotation but with correct colors. I'll be uploading some pics next. #ai #rubikscube #chatgpt (1/x)

Linux.Pizza
@dlevenstein Just like Trofim Lysenko, who published many original ideas and many true ideas, the only two minor flaws being that his original ideas weren't true and his true ideas weren't original.

@dlevenstein I honestly think the only real uses for creative work LLMs will essentially be "inspiration generators," e.g. providing a place to get started, similar to doing a mood board or using an old text prompt list or looking through stock photos etc.

And IMO that would be defensible, because (as an example) you don't have to credit a photo of say New York or Tokyo for inspiring a sci-fi story set in a large metropolis on a distant star many millennia from now

@dlevenstein It's however a perfect recipe for innovating.
@dlevenstein This toot is incredibly funny if youโ€™re more familiar with LLM as denoting a person who has a Masterโ€™s Degree in Law than anything to do with AI.
@dlevenstein Every single thing they produce has to be reviewed by someone competent to review it, yup. That still leaves a large amount of usefulness, but also creates an endless series of fails.
@dlevenstein And as we all know, humans never do either of those things . . .
@dlevenstein But more seriously, you should not be treating these things as semi-autonomous agents working alone. It's more like asking someone to do the research for you, and it's also inherent in that process that the findings may be faulty or biased, and that's exactly how it should be treated.

Clearly LLMs are not on par with competent human experts but nor is there any human who can parse, collate and even to some extent synthesise from such vast amounts of data at such speed.
@dlevenstein The problem with LLMs is that they never provide the source. (because they literally can't)

@dlevenstein that's of no interest to users or investors who have already confused the map with the territory

this is like the rise of movies, but instead of the illusion of motion it's the illusion of mind

@dlevenstein to be fair, human workers have the same problem, but have the ability to plan ahead and double check before declaring success.

@dlevenstein But they're very helpfull when learning yourself a foreign language as they will answer you any grammar related question in great detail because THAT is their core strength!

Also never getting tired in exercising dialogues with virtual barkeepers and such ๐Ÿ˜

@fxk8y I wouldn't trust grammar explanations-- it's likely making up things that arent true even there ๐Ÿ˜ฌ being able to use grammar doesnt mean understanding it even for humans, and llms just create statistically probable stences

@kip The problem is that googling such stuff is nearly impossible in the first place, but after asking how the concept at play is called I was able to quickly verify that the answers were true.

Did this some dozen times until I gained some trust but I do the checks now and then when the answer seems odd or Duolingo rejects the sentence (happened only once so far).

Sadly Duolingo doesn't explain any grammar (at least not in their norwegian course) so LLMs are really helpfull in for example checking if something is a general rule or just valid for the few words I saw

@dlevenstein Nice!

Another one I like: "why would I bother reading something that nobody could be bothered to write?"

@dlevenstein "My colleagues are not good at knowledge work because they are *sometimes* wrong."
@mypalmike @dlevenstein Yeah, it is no better than a colleague who can sometimes be brilliant, but sometimes a disaster with the same level of certainty, and you have to dig deep to find out which is which.
If you don't have such a colleague and feel like you should, then LLM is your option!

@ReinisLazda @dlevenstein I largely agree.

Used as a tool (as I do every day), the brilliant stuff is very much worth it. Very few people are able to distill expertise as clearly as an LLM. I'm ok with having to fact check, although it's not often that I ask anything purely factual.

Asking an LLM something like, "Summarize (some fairly deep subtopic of a subject)" tends to yield results fast superior to any other method, including asking an expert.

@dlevenstein In short, they behave just like humans?
@dlevenstein Makes me think they might be a natural replacement for some kinds of business consulting โ€ฆ. ๐Ÿค”