Yesterday I had a number of conversations with people working in the scholarly publishing sphere about what happens when AI chatbots pollute our information environment and then start feeding on this pollution.

As it so often, the case, we didn’t have to wait long to get some hint of the kind of mess we could be looking at.

https://www.theverge.com/2023/3/22/23651564/google-microsoft-bard-bing-chatbots-misinformation

Google and Microsoft’s chatbots are already citing one another in a misinformation shitshow

Microsoft’s AI chatbot Bing incorrectly reported the demise of Google’s AI chatbot Bard. It’s an early warning sign that this technology is fueling a massive game of misinformation telephone.

The Verge
My fear is that we’ve created an information ecosystem that is uniquely susceptible to the perversions of these AI tools. Fifty years ago, had they existed, they would’ve been mere curiosities because we lacked the information infrastructure for their output to swamp more trusted forms of information. Even twenty years ago there would have been substantially less opportunity for them to have cause harm.

The confluence of this technology with the information ecosystem that we described in our paper from a couple of years could be an epistemic catastrophe.

https://www.pnas.org/doi/10.1073/pnas.2025764118

I’m coming to think that releasing these tools was a reckless act with the potential to generate negative externalities we have barely started to imagine.

The threat isn’t rogue superintelligence. It’s bullshit at unprecedented scale, reflected back upon itself and iteratively amplified.

I’m certainly not saying that the deployment of these systems will suddenly make it impossible to find and build upon trusted and vetted sources of information. Those aren’t going to magically disappear.

My bigger fear is that rather than making a blunder in tying their infotech empires, two automated bullshit generation. Microsoft and Google have correctly anticipated demand. My fear is that people might want what Bing and Bard are selling.

On the other hand, if I dial up the cynicism, just a little bit more, maybe it doesn’t matter much. One view is that by spewing bullshit into the information ecosystem, generative AI is poisoning the well from which it drew life and ensuring that future generations of such technologies will produce garbage.

Another is that the training set was never a pure wellspring. It was already the town cesspool—and even massive quantities of additional bullshit will barely be noticed.

@ct_bergstrom
> the training set was never a pure wellspring. It was already the town cesspool—and even massive quantities of additional bullshit will barely be noticed

This. The Machine Learning text generators have just amplified both existing pain and work on effective medicines for that pain.

@ct_bergstrom With so much hype, one might wonder if it's an attempt by the marketing department to goose the stock price for a failing investment before the flaws are common knowledge.

@ct_bergstrom The work of Google (and Bing) was supposed to be cutting through the bullshit and guide you out of the swamps of nonsense.

They've apparently given up on that, and now the goal is to keep you stuck, spinning your wheels endlessly in a mire of their creation.

@tob @ct_bergstrom It was initially the work of search engines to cut through the swamp of nonsense and bring you the hidden pearls.

Unfortunately, this is capitalism, and they need to make money.

And the way they monetized is they bring you the nonsense somebody paid to bring you.

@ct_bergstrom That is some top-leve, grade A cynicism

@ct_bergstrom

Yeah, I think that hand has the right perspective.

We were already there, and this tech has more potential to reduce the effects of #misinformation than to increase it (from current levels).

Auto-fact checkers, #reputation filters, gradient scoring, citation trees, etc. will be the next lineup of new features for #apps & browsers.

@ct_bergstrom As I see it Professor Bergstrom, we have passed the need for “facts”. People only want content that matches their current set of beliefs.

@DavePerrino @ct_bergstrom People have become too comfortable with what purposefully paid media like Fox "News" provides them with:

Entertainment.

The problem here: Entertainment neither regulates nor feeds people's lives.

Politics does.

People are free to dislike this fact, but still: Politics directs people's lives, regulating where necessary & tolerating where mandatory.
Even if guys like Musk, Zuckerberg, Bezos & fake billionaires like Trump may disapprove...

#Democracy #DefendDemocracy #DefensiveDemocracy #Information #Education #YourVoiceMatters #BeInformed #BeFree #StrongerTogether

@ct_bergstrom Absolutely. The Waveform Clips chaps had a good chat about this and came to the same conclusions. Skip 9.5 minutes if you're impatient.

https://www.youtube.com/watch?v=v3Dc9feQS6o

Is Bing Winning the AI War?

YouTube

@ct_bergstrom I'd be keen to know if this is introducing some people to the idea of (what I think of as) "informationality": basically, is a message mis-, dis- or genuinely informational, a valid "alien" perspective, convincingly plausible entertainment, spitballing, etc?

If so, maybe we have a chance of a cultural shift towards more informative communication? Less off-the-cuff waffling, more citing examples, giving others more time to consider.

Also, is there a term for "informationality"?

@ct_bergstrom Have you read Adrian Tchaikovsky's new short story about the future of LLMs and fiction writing? It might amuse you. https://future-sf.com/fiction/silicon-hearts/
@outeast no but I will now —- thank you!
@ct_bergstrom I hope this will be an opportunity for research-specific AI like Elicit and ResearchRabbit to develop further. Until one of the bigger name AI companies or journal publishers acquire them, that is.

@ct_bergstrom If bullshit is what’s produced by someone who doesn’t care whether they’re telling the truth or not (Harry Frankfurter),

then perhaps bullshit generators are what’s produced by someone who doesn’t care whether they cause harm or not.

@fivetonsflax I almost agree. I think bullshit generators are produced by someone who doesn't care whether their system produces true or logically coherent output.

@ct_bergstrom That’s true of many artists; compare @jwz’s “dadadodo”.

LLMs produce, not just nonsense, but nonsense which wears the clothes of sense, and therefore enables a particular species of harm.

@ct_bergstrom I wouldn't want to say it too loudly for fear of giving these shops ideas, but a realm in which LLMs could seriously mess things up to the point of fomenting war hysteria is in the crossing of language domains. "Sydney, what do the <enemy-du-jour/> really think of us?"
@ct_bergstrom Re: generative #AI, #LargeLanguageModels and the like, the old adage "you are what you eat" distinctly applies...

@ct_bergstrom There's a good piece in the FT today about the level of private control over the development of these tools. It shows not just a concentration, but how the scale of investment and proprietary nature of the tech put it largely beyond researchers & regulators ability to investigate the dangers and biases.

"A lack of access means researchers cannot replicate the models built in corporate labs, and can therefore neither probe nor audit them for potential harms and biases very easily."

https://www.ft.com/content/e9ebfb8d-428d-4802-8b27-a69314c421ce

Risk of ‘industrial capture’ looms over AI revolution

News, analysis and comment from the Financial Times, the worldʼs leading global business publication

Financial Times
@ct_bergstrom It's the information analogue of the "Grey Goo" scenario.
@ct_bergstrom Absolutely agree that the threat of AGI is a misdirection, as if it's the main risk and anything short of it is a-ok. But Metalhead is way more plausible than Terminator.
If anything, techbros have shown us that intelligence is not required to do evil at scale.
@ct_bergstrom Thus we see the unspoken motivation for the Butlerian Jihad.
@ct_bergstrom haha ur last line describes a LOT of reality way before llms
@ct_bergstrom It’s always been good advice not to believe everything you read.
@ct_bergstrom This makes it sound like we weren't already in an age of bullshit prior to AI. But there's a good argument to be made that we have been surrounded by bullshit all along. All that AI does is to democratize it, so you don't have to own a newspaper or cable news channel to spread it.
@ct_bergstrom Library stacks: here I come! 😅
@ct_bergstrom
Absolutely...it's intrinsic in the nature of computers and such interconnected network that all has an exponential trend. In a matter of **months** the entire web, starting from wikipedia, will be rewritten using AI, I am ready to bet on this. And if it won't, no one will know fore sure.
It will overall be a disaster in the short term (soo much noise), even if paradoxically it can improve the critical look on the web.
Wishful thinking?
Books will become more valuable again..
@franco_vazza @ct_bergstrom what's "the critical look"?
@toba @ct_bergstrom
My trusted reference says it is this one:
😄
@ct_bergstrom considering the harm that the basic interwebs has contributed to by weaponising the Dunning Kruger effect, I don't think in itself that AI is dangerous. People and anti intellectualism just ruin everything regardless.
@ct_bergstrom Fifty years ago in the US, propaganda purveyors like Walter Cronkite had vast audiences while providers of true information like IF Stone did not. I'm not convinced this is worse.
@ct_bergstrom Anecdote: As a child I was in BC during the 1968 Democratic convention and read in the local papers how the Chicago police rioted. When I got back home to Oregon the papers from the same dates all said the police were attacked by rioters. Today, we would have access to both sets of accounts, which doesn't mean we would all use them wisely, but at least it's possible.
@ct_bergstrom I'm waiting for someone to set up an automated way for a chatbot to edit wikipedia.
@ct_bergstrom It feels like the next big opportunity for a startup is to create tools and services that help to filter through disinformation. Sadly it seems like a fact that the truth - like water - can be bottled and sold
@ct_bergstrom Preplanned? Great way to gaslight. More distractions. Takes eyes off real issues. Billionaires are the real criminals. Destroying Earth, America and Democracy. Eventually the internet. Maybe hackers will save the day? 🤞🏼🆘🇺🇸🆘🌎
@ct_bergstrom
Clear case of garbage in garbage out. Us humans are so bad at recognizing jokes, telling stories we'd like to be true then convincing ourselves they are. We are feeding AI a diet of pure ego - and egos are fucking liars.
@ct_bergstrom As long as a project manager for a model provider can look at the output and say, "eh, it's mostly ok" and have zero responsibility for the stuff that isn't, the BS train will keep rolling.
To be fair, the epistemic crisis wasn't of their creation, and there's a double-digit probability that AGI will present more pressing issues than the pollution of our collective source of truth.

@ct_bergstrom almost like the alpha go zero of delusional bullshit

ref: https://www.deepmind.com/blog/alphago-zero-starting-from-scratch

AlphaGo Zero: Starting from scratch

Artificial intelligence research has made rapid progress in a wide variety of domains from speech recognition and image classification to genomics and drug discovery. In many cases, these are specialist systems that leverage enormous amounts of human expertise and data.

@ct_bergstrom in any case, seems like a mashup of "flood the zone with shit" and "can we automate that?"
@ct_bergstrom Machine translation has been doing this crap for a decade or more.
@ct_bergstrom The Singularity will actually consist of AI's quoting each other and increasingly getting stuff more and more wrong until it's all noise. Roko's Basilisk hasn't got a chance, it will more be Roko's Village Idiot.
@ct_bergstrom
Well put! - we seem to be rushing into massively polluting the only well we have (our information commons), with no way to recover from that.
@ct_bergstrom @jadeforrest the whole thing with these AIs is that they aren’t even trying to answer with the truth. They can’t tell what is true. They can only generate output that should appear to be plausible.
@relistan @ct_bergstrom yes for sure. I think the feedback loop where generative AI output starts being used for its own training data is something I think has been underreported.
@relistan @jadeforrest @ct_bergstrom can’t wait to see all these horrible second order effects of this 🫤