In retrospect, the flaw in the Turing Test was always using human discernment as the model to measure against.

LLMs easily pass the Turing Test, not because they’re sentient, but because we’re dumbasses who interpret the most dopamine-supplying response to ourselves as the most intelligently human.

@Catvalente I think the Turing Test was intended as a necessary condition required to demonstrate artificial intelligence, but not necessarily a sufficient condition.

To be fair, there are probably people who cannot pass a Turing Test.

@Jvmguy @Catvalente Someone recently put it as (roughly paraphrasing, because I can't remember their exact words) - "The thing that makes LLMs seem intelligent isn't the technology, it's peoples ability to see faces in toast".
@StryderNotavi @Jvmguy @Catvalente Yes! I've been making this same analogy (anthropomorphizing Markov models vs. pareidolia). They're just approximate information retrieval systems -- astonishingly good ones, but still just IR; they have no capacity for reasoning, even the so-called "reasoning" ones.
@wollman @StryderNotavi @Jvmguy @Catvalente
Brains—ours and everybody else’s—are pattern recognizing machines. LLMs are pattern creation machines. The problem with brains is they’re so thirsty for patterns they will create them out of thin air even—especially—if there’s nothing really there. We’ve created machines that feed us patterns that may or may not be there. Marriage made in hell.

@wollman @StryderNotavi @Jvmguy @Catvalente

Exactly. We’re mistaking pattern mimicry for reasoning — like seeing faces in clouds and calling them sentient.

The danger isn’t that these systems “decide” anything. It’s that we hand them power because they sound like they do.

Surveillance. Targeting. Soon, weapons.
Not intelligence — just simulated judgment wrapped in UI.

Anthropomorphism is a bug — a potentially deadly one.

@StryderNotavi @Jvmguy @Catvalente @Tooden
Or in the shower hair that is stuck bizarrely to the side of the shower all the time.
@GeePawHill @StryderNotavi @Jvmguy @Catvalente I've changed my name to JasonGPT for commercial purposes
@jasongorman @StryderNotavi @Jvmguy @Catvalente Dear JasonGPT, can you tell me the last 5 digits of the decimal expansion of pi?
Jason Gorman (@jasongorman@mastodon.cloud)

What makes LLMs work isn't deep neural networks or attention mechanisms or vector databases or anything like that. What makes LLMs work is our tendency to see faces on toast.

mastodon.cloud

@StryderNotavi Maybe you saw it here: “People see faces in the side of a mountain or a piece of toast, and in the same way perceive the output of LLMs, mistaking cogent-sounding statistical probability with intelligence.”

https://lwn.net/Articles/973406/

Debian dismisses AI-contributions policy [LWN.net]

@davel Looks like a few people have made the observation, but @GeePawHill rings true here - https://mastodon.cloud/@jasongorman/114595098303670564

That's the post I remember.

Jason Gorman (@jasongorman@mastodon.cloud)

What makes LLMs work isn't deep neural networks or attention mechanisms or vector databases or anything like that. What makes LLMs work is our tendency to see faces on toast.

mastodon.cloud
@StryderNotavi @Jvmguy It’s also them being programmed to speak in the first person and refer to their selfhood, combined with how many interactions these days are in text form to begin with, making the difference more difficult to parse out
@StryderNotavi @Jvmguy @Catvalente ai use statistics to give answers camouflaged by doubtless and cunning paraphrases
That's the same technique we use with our bosses or customers, when we are not sure of the answer. And that's what bosses or customers like and accept
So, it's an old trick
Jason Gorman (@jasongorman@mastodon.cloud)

What makes LLMs work isn't deep neural networks or attention mechanisms or vector databases or anything like that. What makes LLMs work is our tendency to see faces on toast.

mastodon.cloud

@Jvmguy @Catvalente well, Turing test was created in the period in psychology that behaviorism dominated it, with an iron fist.

The idea was that any talk of emotion (any at all) was subjective. Babies don’t get upset and cry. They make loud noises in periods of not eating.

Also consciousness, itself, is really not measurable.

So for someone that wanted to operationalize the whole thing at the time “if it talks like a duck and walks like a duck… “ was probably the only option.

@Jvmguy @Catvalente

The Turing test was not intended as a test, but as a rhetorical ploy.

It was a sidenote by Turing to illustrate the concept to his peers.

Thus, it's no surprise, it does not measure what people think it measures.

@Jvmguy @Catvalente humans have been able and willing to anthropomorphize even our pets so I suppose it should come as no surprise we do that with something that actually talks back. I find the determination in the minds of some to replace peoples roles with LLM technologies to be based on a lack of understanding and likely over run by greed and an inherent distrust and dislike for those that know more about a thing than they do.
@Catvalente @jon yes. Driven by greed.
I do believe there is a remote possibility that LLMs are a step towards true AI. But the way the business world has grabbed onto it makes no sense at all.
At this stage, it still belongs in a lab. Not out in the world. Even if the technology was real, there’s too many social and ethical questions that need to be answered before putting us all at risk.
@Jvmguy @Catvalente very much so, I agree with your thoughts. I use AI but have always fire walled it off from doing any damage. It has accelerated my learning and ability to create solutions in code. I don't have to remember 'syntax' of things to do with coding as much as before, so I can focus on solving the problem, not memorizing things like I'm some kind of encyclopedia. But outside of my 'home or office lab' no, it is not fit for purpose to start governing our lives. Regrettably this is not an opinion held by technocrats and venture capitalists just now.

@jon @Catvalente It’s being put into use whether we want it or not. Given that, I guess I’ll hope for the best, but I will not be surprised at all if it leads to the collapse of civilization entirely.

An exciting time to be alive.

@Jvmguy @Catvalente Some are saying AI is as big as Gutenberg's System of Movable Type and that certainly shook things up a bit if I read my history right. What I doubt is those taking control in this have all our best interests in mind, mostly rather, their continued growth and profit. So yes, that could spiral out of control. Like you say, its here, want it or not.
@Catvalente speak for yourself dummy! I'm smart and I've got a certificate on its way from Trump university to prove it.
:-D
@Catvalente yes, the Turing Test is much more a measure of our own perception than it is of any sentience.
@Catvalente It should have been a sign when people started reporting seeing messages from aliens in the hiss of TV static.
@Catvalente Some might say we humans put the "dope" in "dopamine". 😆

I will always maintain that the Turing test is not a serious piece of computer science, but a product of Turing living in a society with a lot of expected thoughts and behaviors. The argument runs "if some subject can respond to a human's questions such that that the human cannot tell that they are different, then it is not truly different." That makes no sense about a computer, but makes perfect sense for someone deep in the closet to try and find comfort in it.

@Catvalente

@dogfox wow this is incredibly profound and broke my heart a little. Thank you
@Catvalente That's a pretty insightful take on the Turing Test!
@Catvalente
in short: the Turing test doesn’t measure machine intelligence. It measures human gullibility.

@Catvalente

Honestly, I can't see the Turing test as more than a metaphor to get people over the hump of rejecting the very idea of machine rationality.

I mean, one of his "pass" criteria was the use of psychic powers by a computer. This is not science, it's marketing.

@Catvalente @RnDanger we have a slew of biases. I think there’s like 249 or so cognitive biases?

@Catvalente @RnDanger

If that’s true that’s like 249 possible biases not including the implicit learned ones. So calculating that diversity that we don’t even realize exists. Sounds like an adventure in experience and self I guess.

@Catvalente In the future it's not going to be enough to say "passes a Turing Test". We'll have to say "passes a Turing Test judged by randomly selected Americans" or "passes a Turing Test judged by cybersecurity experts" or whatever.
@Catvalente Also, how easily people are fooled is likely to change. Like, I think of the early days of cinema, and how audiences who weren't acclimated to the new tech purportedly fainted because they were viscerally convinced that the train barrelling towards the screen was going to run into them. A generation that's grown up with LLMs will be better than us at spotting their tells.
@Catvalente I really dislike this post. Not because you’re wrong, but because you’re right. 🙁
@Catvalente Interesting. AI is probably pretty good in determining if I am sentient, human, so if it was to be deceptive it might be good at fooling me that it is not sentient.
@Catvalente the future of coding podcast episode on the turing test paper is EXCELLENT https://futureofcoding.org/episodes/076
76 • Computing Machinery and Intelligence by Alan Turing (feat. Felienne Hermans)

Are you looking for the real computer revolution? Join the club! Future of Coding is a podcast and community of toolmakers, researchers, and creators working together to reimagine computing.

Future of Coding

@Catvalente Also, the Turing test was never about a machine *being* intelligent - it's about a machine *convincing a human* that it is intelligent.

Which, as we now know far better than Alan Turing could have ever imagined, is tragically easy to do.

@Catvalente people have given way too much credence to the Turing test, it is a case of faulty logic.

If intelligent beings talk a certain way, then talking a certain way shows you are an intelligent being?

Genius artists can replicate a view realistically. Cameras can replicate a view realistically so cameras are genius artists?

Intelligent people are good at chess, AlphaZero is good at chess, so AlphaZero is an intelligent person?

Causation doesn't work like that

@Catvalente

I always thonk that Turing test was hypothetical like Schrodinger cat or Russel's teapot or trolley problem

@Catvalente @Gargron
Hmm 🤔
If I cross the streams, the Turing test should be done in part by the most intelligent bears 🐻.

While they are away the Pope can use the woods.

🤷🏻‍♂️

@Catvalente it was never intended as a test of sentience or intelligence, but a challenge to our reflexive dismissal of the ability for computers to think.

From wikipedia:

Turing did not explicitly state that the Turing test could be used as a measure of "intelligence", or any other human quality. He wanted to provide a clear and understandable alternative to the word "think", which he could then use to reply to criticisms of the possibility of "thinking machines" and to suggest ways that research might move forward.

LLMs do think and possess limited intelligence (but not human intelligence, think more like cockroach intelligence).

@Catvalente terry Winograd wrote about this (as did hofstadter) two decades ago. You are absolutely right.

See https://berryvilleiml.com/bibliography/ for pointers.

@Catvalente I know a tenured PhD preparing to jump ship to an AI firm and he brought up the Turing test with such fervor that I had to change the subject.
It's not a flaw in the test. Nor is is it an error of perception.

After the Chinese Room thought experiment, "AI" was split into two, "weak AI" and "strong AI." Weak AI seems intelligent (passes the Turing test), while strong AI is self-aware. We have no meaningful test of strong AI at the moment.

Such line of argument neglects the useful argument. People confuse weak AI for strong AI, and current LLMs for weak AI. Since strong AI is universally useful, so also must LLMs be. But it is perfectly possible to be below weak AI and be useful (quick sort, despite being very dumb, is very useful). It just means that LLMs and weak AI are not universally useful.

The discussion should not be whether AI is useful or not, or whether LLMs are weak AI or strong AI, but which cases LLMs are useful for. They have already proven them selves to be excellent for transcription of spoken text and pretty good at translations, but have other serious limits (and costs) that should also not be ignored.
@Catvalente Now, it there was a chat bot that could tell with 100% accuracy whether it was talking to a human or another chat bot, that would be useful.
@Catvalente I've been always saying that Turing's Test only works when the examiner is Turing.
@Catvalente Similarly, if someone invokes "common sense" to justify their argument, the heuristic is, their argument is most likely flawed.
Wandering Star (@pawsplay@dice.camp)

Attached: 1 image

Dice.camp

@Catvalente

Perhaps our mistake is in thinking that we are sentient.

@Catvalente
Turing test was always a thought experiment, but your solid point does make me think- have we actually tried the mental acuity tests?

I genuinely don't know if they would work, but the tests where they ask a number of things about current events, remember these 4 words, could you talk about this subject, repeat these things back to me...

I actually am not sure.