The constant mental vigilance in a generative world is exhausting.

"I asked Claude to do $thing and it did this!"

No it didn't. No you didn't. Probably none of that happened.

And somehow, being unwilling to admit the thing is just making stuff up is annoying and unnecessary, not the damn model.

@mttaggart

I can't get people to understand that the "hallucination" problem is unsolvable because "hallucination" is how it works. That's all it does. Next tokens based on the whole previous series of tokens that represent "the conversation" being had between prompts and responses combined with the hidden prompts that give the thing its flavor. The fact that it is "right" isn't part of it. That's why they never say, "I don't know". They don't know anything. They are literally making it up every single time. It's why they are so expensive and why they are ruining the environment. There is no recall, no memory, no "knowing". As I've seen it said elsewhere, "there is no 'there' there". It's worse than the Chinese Room thought experiment because at least that produces correct responses. This creates the illusion of a correct response. We are killing the earth and building an inescapable surveillance state around technology that will never get any better than it is right now.

@jrdepriest The one that gets me is the "reasoning" models. They're just making up more text to fluff the context! No thought is happening, nor can it! It's maddening.

@mttaggart

It's the same "model" your know-it-all uncle uses every Thanksgiving: bloviation.

@mttaggart @jrdepriest it's just working backwards to "explain" it's bullshit answer, or so I heard.

if so that's a straight up con.

@dannotdaniel @jrdepriest Two different things.

"Reasoning" models literally talk to themselves to build out context for prompts. This leads to incredibly high token usage, and also some truly insane, neurotic-looking behavior as the model talks itself into corners.

But yes, if you ask a model "why" it did or said anything, you'll just get nonsense. There's no introspection happening there at all. Researchers can't pinpoint how some of these inferences are made, but the model itself can?

@mttaggart @jrdepriest all right well that's at least what they advertise I guess

I was impressed with the "reasoning" when they rolled it out but I've barely used it, and I've definitely encountered the situation where the thing just doesn't shut up

@mttaggart @jrdepriest neatly explains why so many managerial types are fans of it though - it sounds JUST like them

👀

@jrdepriest @mttaggart it's intelligence theatre, the appearance of words that resemble intellect. I can't describe how much it frustrates me that so many are happy to accept the illusion
@jrdepriest @mttaggart hallucination is an awful term for it; it implies a form of perception that is being undermined, when no such perception exists. A philosophy professor of mine refers to model output as "bullshit" in that it does not distinct between truth and falsehood, only seeking to accurately reproduce langauage patterns.

@fireye @jrdepriest @mttaggart

In reality, its high dimensional vector calculus, over the trained material and your words in your context.

All this LLM workings is just vector calculus, that Leibniz devised in the 1700s. We only had the compute to do it in 2012.

Its not thinking, yet. Its not intelligence. Its a stochastic parrot trained on TBs of data, following Leibniz' dream of word-calculus.

We still dont know how consciousness works, or to create a thinking machine.

@jrdepriest @mttaggart Y'all really think it's *that* different from how humans work? People make overconfident statements about shit they don't really have 'knowledge' of all the time (see this toot/this thread/social media et al.). Sure, it's nOn-DetERMinIsTic—so are we, that's kind of the point.

Doesn't mean it's not extremely useful in certain contexts.

@zero_gravitas @jrdepriest You're arguing a point about utility I did not make. There's no doubt generative text can be "useful" in some contexts.

But the process matters here.

If you think the model's generation process is indistinguishable from human thought, we won't agree and that's that. I do think it's quite different from how humans work. If you do see a distinction, then the distinction is material when users impute the properties of one to the other.

Does the model evaluate the truth of a statement before it returns the output? Can it? If the answer is no, we have a divergence.

@mttaggart @jrdepriest Fair. For the record, I don't think it's indistinguishable from human thought. People seem more focused though on how it diverges from programatic input/output—which is I think the wrong framing all together. I agree that it's a bad idea to anthropomorphize, though that's certainly not a problem limited to how people talk about large language models.

@zero_gravitas @mttaggart

Yes. Is it that different.

As has been pointed out, LLMs operate using Linear Algebra Matrix Math to make predictions for "next token". That is all. Any appearance of bullshitting, bloviation, lying, confabulation, hallucinations, etc are just us anthropomorphizing math. I am guilty of it because it is much easier to talk about these systems when you do it. I'm also guilty of anthropmorphizing my car when it won't start and I beg it to please start this one time, I'll get the good gas next time. Neither system has any understanding of what is being asked of it; one of them just happens to have been trained on enough human language that it can spit out something to make it look like it does.

Is language the same as intelligence The AI industry desperately needs it to be

I had a good article that summarized historical metaphors for how the human mind works. They tend to crop up around whatever the hottest new technology is. "Mind as Steam Engine" was popular for a while. And we've all heard about keeping the humors in balance. I cant' find the article now.

There are entire books written about the idea. Our minds are not simply in our brains. Our minds are our entire body, every cell and nerve impulse, every sense and all the information that is processed back and forth. All of that is our "mind". A machine cannot create that or simulate that. LLMs aren't even built to do that. They are built to produce something that looks like our language as if we are having a conversation.

Large language mistake

Neuroscience indicates language is distinct from thought, raising questions about whether AI large language models are a viable path to artificial general intelligence.

The Verge