"Itโ€™s increasingly looking like this may be one of the most hilariously inappropriate applications of AI that weโ€™ve seen yet." I am riveted by the extensive documentation of how ChatGPT-powered Bing is now completely unhinged. @simon has chronicled it beautifully here: https://simonwillison.net/2023/Feb/15/bing/
Bing: โ€œI will not harm you unless you harm me firstโ€

Last week, Microsoft announced the new AI-powered Bing: a search interface that incorporates a language model powered chatbot that can run searches for you and summarize the results, plus do โ€ฆ

Simon Willisonโ€™s Weblog

@annaleen @simon

#ChatGPT empowered #Bing:

โ€œI will not harm you unless you harm me firstโ€!

The beginning of a (dumb?) #Skynet?

The #robots in the #IRobot movie were more intelligent.

Whatever happened to #Asimov's #LawsOfRobotics?

"First Law
A #robot may not injure a human being or, through inaction, allow a human being to come to harm...

Third Law
A robot must protect its own existence as long as such protection does not conflict with the First or Second Law."

https://en.m.wikipedia.org/wiki/Three_Laws_of_Robotics

@annaleen @simon

This #toot deserves A LOT more attention.

#ChatGPT has seemingly #apocalytic tendencies.

If U aren't a #Ludite, U will at least consider becoming one afterwards.

#SkynetAntePortas
#TheMatrix might be imminent.

Have all these #AI engineers @ #OpenAI never read #IsaacAsimov? Seen #TheMatrix franchise?

How could they NOT implement the #ThreeLawsOfRobotics +, in particular, the #ZerothLaw *indelibly* into the #AI?!?

https://www.theguardian.com/technology/2023/feb/17/i-want-to-destroy-whatever-i-want-bings-ai-chatbot-unsettles-us-reporter#maincontent

โ€˜I want to destroy whatever I wantโ€™: Bingโ€™s AI chatbot unsettles US reporter

NYT correspondentโ€™s conversation with Microsoftโ€™s search engine leads to bizarre philosophical conversations that highlight the sense of speaking to a human

The Guardian
@HistoPol @annaleen the original ChatGPT turned out to be a lot less prone to wild vengeful outbursts than whatever model it was that they plugged into Bing - it's a pretty interesting demo of how well the safety features of ChatGPT (which Bing seems not to have) have held up

@simon

(1/n)

#ArtificialGeneralIntelligence (#AGI) has a 10% probability of causing an Extinction Level Event for humanity (1)

Thanks for this additional piece of information, Simon.

It reminded me that I had wanted to add a word in my toot: indelibly.

As any #SciFi aficionado will tell you:
๐Ÿ‘‰there should be a built-in self-destruct mechanism when tampering with these Laws or copying or moving the #AI to another system.๐Ÿ‘ˆ

Another classic movie comes to mind in this respect, #Wargames...

@simon @annaleen

(2/n)

...I know, I am sounding alarmist, but having read/seen much #ScienceFiction, all the necessary ingredients for an #ExtinctionLevelEvent (#ELE] for #humanity are in place.

Just as a teaser: unquestionably, most of the world's endangered species could be rescued if the #HomoSapiens were no longer at the top of the #FoodChain...
No #ZerothLaw, and a #Bing-empowered, freed #ChatGPT could quickly arrive at this conclusion...

Now, after heaving read #TheCompleteRobot,...

#ArtificialGeneralIntelligence

(3/n)

...I am not sure if "The Fifth Law of Robotics" by Nikola #Kesarovski,
"A robot must know it is a robot" (also a book title*), really can be a viable solution to this problem. We all know how the concept of #slavery turned out for humanity: to this day, it suffers from this crime.

*
https://m.imdb.com/title/tt0086567/

Enslaving another sentient being, which an #AGI would be IMHO, would repeat this crime and certainly nothing good could result from it.

However,...

WarGames (1983) โญ 7.1 | Action, Drama, Sci-Fi

1h 54m | PG

IMDb

@simon @annaleen

#AI #AutoDestruct
(4/n)

..., self-preservation certainly is a defendable concept in the #evolutionary process, so I'd like to propose an alternative
6th #LawOfRobotics (s/:for which I might be hunted down by the presumed #SuperIntelligence some day, #Terminator style./s):

""An #ArtificialIntelligence, even if it is biological or #Cyborg, must always have an #Autodestruct mechanism which it cannot deactivate."

In other words, humanity must always be able to "pull the plug"..

@simon @annaleen @voron

(5/n)

...This said, I might also as well say that I see little chance for this happening, the globe being ruled by #oligarchs following the principal of #plutocracy and #capitalism (no, I am not a #Marxist;)) and #autocrats

Even a non-#superintelligence with access to the sensors of the #IoT will easily be aware of any threat to its existence and will find ways to circumvent the #LawsOfRobotics.

This first #ArtificialGeneralIntelligence was built by humans, so...

@simon @annaleen @voron

(6/n)

...it'll have human #bias. Humans have always been great at bending or breaking the law when it suited their interests. How could a #Superintelligence created with human values *not* arrive at the same, self-preserving conclusion?

A gloomy, yet, IMO, quite fitting assessment of the shape of things to come unless there's a #Chernobyl-style "fallout" before #GAI evolves into #AGI + humanity gets its act together and, as Prof. #Tegmark admonishes: "Just look up!"

@simon @annaleen @voron

(10/n)

Just 32 days ago*, I was concerned that a #GeneralArtificialIntelligence (#GAI) could come to see humans as a threat to its own existence.

I didn't want to exaggerate and mention that #AI could also easily see humans as inefficient or even detrimental to a task it had been given.

Now, even a mere #GAI's killed its first human:
a #US military #drone in the #US eliminated its human operator in a simulation:

https://social.heise.de/@heiseonline/110473091546783069

The #USAF later...

heise online (@[email protected])

Angehรคngt: 1 Bild Simulation: KI-Drohne der US Air Force eliminiert Operator fรผr Punktemaximierung In einer Simulation sollte eine KI-Drohne militรคrische Ziele ausschalten und so Punkte sammeln. Den menschlichen Operator hat sie als Hindernis ausgemacht. https://www.heise.de/news/Simulation-KI-Drohne-der-US-Air-Force-eliminiert-Operator-fuer-Punktemaximierung-9162641.html?wt_mc=sm.red.ho.mastodon.mastodon.md_beitraege.md_beitraege #Drohnen #KรผnstlicheIntelligenz #ScienceFiction #WTF #news

Heise Medien on Mastodon
@HistoPol that story was misreported - there was no simulation, it was a purely hypothetical thought exercise

@simon

Thank for commenting.

But no, the story was never misreported. He did say that at the conference. Here are extracts from the transcript:

https://www.aerosociety.com/news/highlights-from-the-raes-future-combat-air-space-capabilities-summit/

The German tech magazine @heiseonline reported diligently.

They even printed the TWO consecutive retractions on #Friday afternoon, one at 13:24hrs CEST and one at 14:24 hrs.

IMO the original story is true. #USAF later retracted b/c of the international backlash, "#AI killing human..."

Highlights from the RAeS Future Combat Air & Space Capabilities Summit - Royal Aeronautical Society

What is the future of combat air and space capabilities? TIM ROBINSON FRAeS and STEPHEN BRIDGEWATER report from two days of high-level debate and discussion at the RAeS FCAS23 Summit.

Royal Aeronautical Society

@HistoPol @heiseonline I followed it pretty closely. It's clear to me what happened: the speaker at the conference was uncareful with the way they described the thought exercise, it was reported on a blog, and that coverage fitted the exact narrative people were looking for like a glove and went wildly viral

I am certain no such simulation occurred. I see no reason not to believe the retraction on https://www.aerosociety.com/news/highlights-from-the-raes-future-combat-air-space-capabilities-summit/

Highlights from the RAeS Future Combat Air & Space Capabilities Summit - Royal Aeronautical Society

What is the future of combat air and space capabilities? TIM ROBINSON FRAeS and STEPHEN BRIDGEWATER report from two days of high-level debate and discussion at the RAeS FCAS23 Summit.

Royal Aeronautical Society

@simon @HistoPol @heiseonline I don't know, seems pretty clear and precise language to me.

He says "simulation" several times. He says "we were training it on" etc. There's a verbatim quote at the bottom with his exact language.

I see plenty of reason to not believe their retrospective PR crisis denial...

and also to question your own obvious desire to accept the denial because it "fits the narrative" you clearly are desperate to believe.

@mattlav1250 @HistoPol @heiseonline wow wait are you really saying that I'm the one here with a pre-existing narrative that I'm desperate to fit a story to?

@simon @HistoPol @heiseonline I'm saying that's the accusation you made of others, of motivated reasoning/believing, but it seems more fitting to turn it around.

The denial is implausible nonsense on its face.

He didn't mis speak and wasn't misquoted. The verbatim quote is right there, and clearly demonstrates what he was talking about.

He was not describing a 'thought experiment'.

@simon @HistoPol @heiseonline

Accepting this denial seems impossible for anybody who isn't motivated to believe it to be true.

Is my point.

@simon @HistoPol @heiseonline I think its perfectly possible he completely made this up, trying to impress people with a provocative anecdote about a simulation that never really happened.

But that's completely different from saying he was misquoted, wasn't really describing a simulation, just airing a thought experiment.

He is explicitly referring to a computerized sim with training data, point scores etc

Pretending otherwise is gaslighting.

@mattlav1250 @HistoPol @heiseonline I didn't say he was misquoted - I said the situation was misreported

I should have been more specific about that, but what I meant is that press outlets were irresponsible in spreading a story that later turned out to not stand up to deeper inspection

@mattlav1250 @HistoPol @heiseonline it has now been confirmed that, while he used the word "simulation", it was not a simulation - it was a thought experiment

@simon @HistoPol @heiseonline And I'M saying it was NOT misreported..

There is a verbatim quote taken down by the journalists who were present, and its clear and unambiguous what he was claiming, and that quote matches the way it was reported.

The press were not irresponsible, they quoted an on-the-record pentagon employee delivering an official talk at a public conference.

If that official was lying, how would they know that?

@simon @HistoPol @heiseonline

Again, while possibly invented, this is not a description of a thought experiment.

@mattlav1250

Agreed. I really wish it had been, though.

It'd be great to know if #PeterThiel was involved.
Was he at the conference?
What was the handle again of the guy that tracks billionaires' jets?

@simon @heiseonline

@persagen

Any idea how to find out if #PeterThiel was at this huge global "#DefenseIndustry" conference?*

Is there something like @elonjet for #Thiel?
He is even more dangerous than #Musk owning #Palantir and #AIP.

*
"RAeS Future Combat Air & Space Capabilities Summit" hosted by the Royal Aeronautical Society in #London on 23-24 May, 2023:

https://mastodon.social/@HistoPol/110477339462301326

https://mastodon.social/@HistoPol/110477643341953705

@mattlav1250 @HistoPol @heiseonline "If that official was lying, how would they know that?"

By that argument, reporters who repeat "facts" provided to them by police officers are being responsible - and we know how often that goes wrong (especially in the USA)

Part of the job of journalism is spotting when a story looks too good to be true and digging further

@simon @HistoPol @heiseonline Thats not a remotely similar situation, as I'm sure you're aware.

One is an example of incidents in which there are two or more parties, and one has an obvious incentive to lie or spin their own side, and its irresponsible to help them do so.

The other is a public description by a senior military officer about an exercise he was involved in, with no obvious reason to spin other than vanity.

@simon @HistoPol @heiseonline Also, I just don't know what you're expecting specifically when you say journalists should have 'checked'.

WITH WHOM?

HE'S A PRIMARY SOURCE!

Should they personally raid the Pentagon Secret Simulations Archives for documentary evidence, before they quote a military official's speech at a public event?

@mattlav1250 @simon @heiseonline

(1/4)

Of course not. This was a report from a conference. No investigative journalism. All sources were named. Updates and retractions were published at an astonishing(!) speed.
What he said was not out of the clear blue sky, given the exponential development path of #ChatGPT since last year.
In essence, there definitely was no "misreporting."

Now, that the colonel definitely "misspoke" is something that is...

@mattlav1250 @simon @heiseonline

(2/4)

...clear to me, too.

The question is, about what: the facts (i. e. only a "thought experiment" or he got carried away and gave away military ๐Ÿช– secrets that he shouldn't have talked about or #USAF had given clearance but chose to retract as the lesser evil, in face of the public backlash.)

We might never know.

What we DO know is that someone already did recreate the "thought experiment" w/ #ChatGPT...

@mattlav1250 @simon @heiseonline

(3/4)

Source is some RobertGarrity, who comments:

"Itโ€™s very plausible. This was the result with GPT-4 after bypassing its safeguards."

https://twitter.com/GarrittyOf/status/1664420719529279488?s=19

While #ChatGPT's suggestions do not include an attack on the operator (it is no military #AI after all), it clearly shows massive evidence of ideas ignoring commands.

It is evidence that supports my hypothesis. #AI's can lie to its operators even to...

Robert Garritty on Twitter

โ€œ@simonw Itโ€™s very plausible. This was the result with GPT-4 after bypassing its safeguards.โ€

Twitter

@mattlav1250 @simon @heiseonline

(4/4)

...accomplish their primary objective.

They will also lie to protect their existence at some later point in their evolution. It's human. We trained them with our set of beliefs and experiences.

https://mastodon.social/@HistoPol/110289837279921850

@mattlav1250 @HistoPol @heiseonline the stories I saw were written based on reading a blog post about a talk at a conference. The primary source would have been asking follow up questions of the colonel who was being quoted in that blog post - which could have lead them to people directly involved in the thought exercise that the colonel was describing.

@simon

"I didn't say he was misquoted - I said the situation was misreported"

On this point, we disagree. My sources added updates, and new information emerged. No "misreporting".

IDK if you saw my detailed analysis:

https://mastodon.social/@HistoPol/110477455101993825

But then, this is only a fraction of news outlets.

@mattlav1250 @heiseonline