Computers have been beating the best human Go players since 2016. The Go world champion retired in part because AI is “an entity that cannot be defeated.”

But a human just trounced one of the world’s best Go AIs 14 games to 1: https://arstechnica.com/information-technology/2023/02/man-beats-machine-at-go-in-human-victory-over-ai

I think this news story is more interesting than it might first appear (without knowing details, so grain of salt). It isn’t just a gaming curiosity; it points to a fundamental flaw with “deep learning” approaches in general.
1/

Man beats machine at Go in human victory over AI

Amateur exploited weakness in systems that have otherwise dominated grandmasters.

Ars Technica

The Go AI was trained by feeding it a huge number of Go games. It built a model based on w̶h̶a̶t̶ ̶h̶u̶m̶a̶n̶s̶ ̶d̶o̶. CORRECTION: KataGo is trained by playing against itself; the model input is “past AI games.”

The human beat it by doing something so obvious a human or sensible AI would never do it — so it wasn’t in the training data, so the AI didn't counter it.

(Basically, the human forms a conspicuous giant capture ring while distracting the AI with tactical battles the AI knows how to counter.)

#ai
2/

Why is this interesting?

The recent eye-popping advances in AI have come from models that scan huge datasets, either generated by humans (e.g. “all competitive Go games” or “all the text we could find on the web”) or by computer (“the AI plays itself a billion times”), and imitating the patterns in that dataset — with no underlying model of meaning, no experience to check against, no underlying theory formation, just parroting.
3/

These systems produce such striking results, it raises the question of whether our brains are any different. Are we also just fancy pattern mimics?

Would an LLM gain human-like intelligence if only we had more processing power?

This result suggests that no, there’s something else our brains do. The tactic the AI missed would be painfully obvious to a human player. There’s still something our brains do — theorizing, generalizing, reasoning through the unfamiliar — that these AIs don’t.
4/

As usual on Ars, there are actually some good comments.

This person wisely reminds us that AI has been littered with bold predictions that human-like AI is just around the corner — and every time, we realized that we’d failed to understand what the hard part even was. The Marvin Minsky quote here is eye-popping:
https://arstechnica.com/information-technology/2023/02/man-beats-machine-at-go-in-human-victory-over-ai/?comments=1&post=41644794
5/

Man beats machine at Go in human victory over AI

Amateur exploited weakness in systems that have otherwise dominated grandmasters.

Ars Technica

But does this matter? Sure, the AI did a faceplant on some bizarro strategy that would never fool a competent human. So what? It’s just a board game.

OK, what if it’s a self-driving car?

Have you ever encountered a traffic situation that was just totally bizarre, but had a common-sense solution like “wait” or “just go around?” What would an AI do in that situation?

Think of the reports of self-driving Teslas suddenly swerving or accelerating straight into an obvious crash.
6/

Or as this comment remarks: what if it’s an autonomous killbot? (arguably a superset of the previous item, I know) https://arstechnica.com/information-technology/2023/02/man-beats-machine-at-go-in-human-victory-over-ai/?comments=1&post=41644720

Several comments point out the parallels to the recent story about Marines defeating an AI with Jim-Carrey-style nonsense antics that bore no relationship to the training data: https://arstechnica.com/information-technology/2023/02/man-beats-machine-at-go-in-human-victory-over-ai/?comments=1&post=41644757

The linked article: https://taskandpurpose.com/news/marines-ai-paul-scharre/
7/

Man beats machine at Go in human victory over AI

Amateur exploited weakness in systems that have otherwise dominated grandmasters.

Ars Technica

This is probably what’s going on with the hilarious ChatGPT faceplants making the rounds on social media.

People try to fool GPT with esoteric questions, but those are easy for it: if anybody anywhere on the web already answered the question, no problem — and making it esoteric just narrows the search space.

But give it a three-digit addition problem, and there’s no single specific example to match. And GPT can’t turn all those examples into a generalized theory of how to do addition.
8/

What’s the moral here?

1. Beware the AI hype.

2. Know your AI history, at least a little.

3. If AI ever does become intelligent (whatever that means), it’s going to be •weird• for a long time first.

4. Beware your own heuristics about what “intelligence” looks like, because •that• is a place where our brains are as easily fooled as the Go AI was. Our brains didn’t evolve around things like GPT, and we’re easily bamboozled by them.
/end

Correction from some replies:

Commenters correctly pointed out that the specific AIs in question (KataGo and Leela Zero) did not use human play as training data, only AI play. Thanks! Edited posts to fix.

Correction of some replies:

None of the variants of AlphaGo are the AI in question.

From @Iguanadelmar: Yes, interesting indeed: https://kolektiva.social/@Iguanadelmar/109892672212437525

The point of my thread was that deep learning has weird blind spots that a human would be unlikely to develop, and that these blinds spots are intrinsic to the approach.

The additional insight I take from the fact that it was computer analysis that found the AI’s weakness is that our human metacognition makes it hard for us to find deep learning’s blind spots. Again, we’re too easily bamboozled.

iguana (@[email protected])

@[email protected] interesting to note that the tactic the human used to defeat the AI was suggested by a different AI that tried out millions of tactics to find the AI Go champion's weakness..

kolektiva.social
From @psu_13: This paper gets at the underlying problems of that now-embarrassing Minsky quote. It gets at many of the things I was talking about upthread, and much more beside. Thanks for the recommendation!
https://pgh.social/@psu_13/109892784391376542
psu_13 (@[email protected])

@[email protected] Why AI is Harder Than We Think https://arxiv.org/abs/2104.12871

pgh.social

From @Globaltom, a very good example of a human handling a novel problem not in the training data:
https://mstdn.ca/@Globaltom/109892880540137200

I tend to think that Roomba had the right idea with respect to AI: build stupid-ish machines that assist humans by handling the easy 80% of dreary tasks, and complement humans instead of trying to mimic humans.

Globaltom (@[email protected])

@[email protected] I think about the first time I encountered a pilot vehicle (roadworks in rural Alberta). A sign explained how it worked; I was fine. You can't use words to tell a self-driving car how to deal with something novel...

Mastodon Canada

From @mykl: https://infosec.exchange/@mykl/109893061731112567

The mistake in that line of reasoning is imagining that “intelligence” (whatever that means) is a linear continuum. Computers are already far “smarter” than humans in specific ways. (Can you sum a billion numbers in one second?)

Part of my point is that what we call AI won’t progress toward human intelligence like going up a staircase. It will be •weird• the whole way, and will thwart our intuitions.

Mastadon Juan :donor: :pizza_pineapple: :1password: :apple_inc: :loading: (@[email protected])

@[email protected] Ok, hear me out... what if these systems are, at this moment in time, actually as smart as the *average* human. We may be in for trouble if they get smarter, and do not gain some things like compassion, tolerance, and other attributes that humans were supposed to have.

Infosec Exchange

This is a good clarification from @ekwessel: https://mstdn.party/@ekwessel/109893444487949870

Please don’t read my thread to mean “Machines can’t handle novel situations,” or “Machines can’t generalize,” or “Machines can’t identify underlying patterns.” All those statements are false! There is no such bright line to be found here.

My argument is a fuzzier one: software is surprising; it’s hard to know what the hard part is; we should expect to see ML impress us but then faceplant on the seemingly obvious.

Erik Wessel (@[email protected])

@[email protected] Great thread! I wanted to point out a small caveat. It is often said ML methods *can’t* generalize outside of their training domain. While this is often the case, it *isn’t always true*: this paper shows simple examples where strong generalization happens from sparse addition examples. It is true big ML models struggle to generalize, but people should know that strong generalization isn’t *impossible*, it just might be hard in practice. https://arxiv.org/abs/2201.02177

Mastodon Party
Erik Haugen (@[email protected])

@[email protected] maybe also: 5. Understand the tool you're using and its limitations.

Mastodon 🐘

Aha! Here is the paper on the research behind this story (ht ‪@leftpaddotpy‬):

https://arxiv.org/abs/2211.00241

The researchers’ abstract puts it quite crisply: “Our results demonstrate that even superhuman AI systems may harbor surprising failure modes.”

Adversarial Policies Beat Superhuman Go AIs

We attack the state-of-the-art Go-playing AI system KataGo by training adversarial policies against it, achieving a >97% win rate against KataGo running at superhuman settings. Our adversaries do not win by playing Go well. Instead, they trick KataGo into making serious blunders. Our attack transfers zero-shot to other superhuman Go-playing AIs, and is comprehensible to the extent that human experts can implement it without algorithmic assistance to consistently beat superhuman AIs. The core vulnerability uncovered by our attack persists even in KataGo agents adversarially trained to defend against our attack. Our results demonstrate that even superhuman AI systems may harbor surprising failure modes. Example games are available https://goattack.far.ai/.

arXiv.org

Many replies take one of these two viewpoints:

“Ha! Humans win after all!”

“Whatever, brains are just neurons and so is this, therefore human-like AI is still forthcoming.”

Both rather miss the point: “AI systems harbor surprising failure modes.” Yes, humans will patch the problem, and AIs will keep winning at Go.

But those failure modes that •would have been easily detectable by humans• indicate intrinsic gaps in the •structure• of these “deep learning” systems, not just their scale.

The lesson here isn’t about who wins at Go. It’s about the danger of extrapolating from narrow domains (like winning at Go) to open-ended problems (like driving a car) or general human-like intelligence.

Software is surprising.

AI is surprising.

“Intelligence” — whatever that means — is surprising.

We keep realizing we’ve failed to understand what the hard part even is. We are fools if we do not expect that from the outset.

@inthehands also, corollary: we shouldn’t be surprised by the oddball failure modes that humans have, either.
@steve Exactly so. There are two difficult layers, both at the heart of engineering: (1) prepare for the unexpected; seek it out relentlessly; mitigate after and mitigate before; build in redundancy; and (2) intuitions about human failure modes create blind spots for us when reasoning about AI; no failure mode is too weird, too stupid, or too obvious to consider.
@inthehands Fun to speculate what bizarre failure modes our human intelligence might have too. We’re lucky adversarial training would be too slow to be practical.
@inthehands i dunno. Humans stepping on the rake of easily detectable failure modes seems to be the story of this century
@inthehands unrelated, I’m totally stealing your use of •bullets• for emphasis. Much more clear than stars or underscores if Markdown isn’t being used.