Meta. OpenAI. Google.

Your AI chatbot is not *hallucinating*.

It's bullshitting.

It's bullshitting, because that's what you designed it to do. You designed it to generate seemingly authoritative text "with a blatant disregard for truth and logical coherence," i.e., to bullshit.

@ct_bergstrom perfect for writing hollow but good sounding corporate pr flack pieces. Or the trader joe's fearless flyer product descriptions.
@bikerglen @ct_bergstrom Our university is going through a "strategic planning process" so I asked it to write up a strategic plan and guided it along with a few of our institution's general parameters. The output text read like I expect the eventual true product will read... and like every other university strategic plan that I've read before with a few institution-specific spice words. (We could save a lot of money on consultant fees this way.)
@dezene @bikerglen @ct_bergstrom Maybe the insight is that you don't need the plan document at all.
@tob @bikerglen @ct_bergstrom The only benefit that I can see from these exercises are that they keep senior administrators and their small empires busy, and not inventing new forms for faculty and staff to fill out.
@dezene @bikerglen @ct_bergstrom You definitely don't want them automating their processes using AI.
@dezene @tob @bikerglen @ct_bergstrom If anything good comes out of this, it may be a better understanding of the mechanics of, also human, BS production. And for some folks, developing a better eye for spotting it.
@martinvermeer @dezene @tob @bikerglen @ct_bergstrom Is there an algorithm for measuring the extent to which a human text, say a campus strategic plan, *deviates* from a series of AI attempts? Then we might grade ourselves on originality.

@jimproctor @martinvermeer @dezene @bikerglen @ct_bergstrom IMO, this is the (awkward) way forward for these LLM-based attempts at "AI". Take the LLM generated content, run it through some sort of meta-analysis, and then present that to the user.

I mean, in a way they're already doing that to suppress the nazi stuff.

In your case, this hypothetical tool would be capable of doing a meta-analysis of your document and the LLM-generated document on the same subject.

@jimproctor @martinvermeer @dezene @bikerglen @ct_bergstrom Of course, you can't ask the LLM to do the meta-analysis. So we're relying on an as-yet non-existent technology to make the LLM useful in a practical way.

At which point, why are we using the LLM?

@tob @jimproctor @martinvermeer @bikerglen @ct_bergstrom

Really all I want is an AI that could retrieve, fill, and send forms to the next approval stage with a few simple typed or spoken commands.

@ct_bergstrom In other words, computer generated mansplaining.
@kyozou @ct_bergstrom The creation resembles its creators.
@ct_bergstrom Disagree. They're designed to mimic what a human would write. If they end up bullshitting it's because the models aren't good enough, not because that's what they're designed to do.
@ct_bergstrom @moultano It’s because that’s what a human with no ethical compass does when they either don’t have the facts or can’t be bothered to stick to them. AI has no ethical compass and no concept of facts, so I’d say it’s modelling a certain type of human very accurately.

@pjie2 I would not even put it that morally judgemental.

As long as you have no clue, you don’t know that you’re bullshitting.

I have quite a few old articles of mine that I changed over and over again, adding corrections, updates, or clarifications when I learned more about a topic.

But those LLM's do not know whether I have the expertise on a specific topic that my writing is actually well-founded.

And I found that the more I understood the less I wrote unasked.

@ct_bergstrom @moultano

@moultano Humans have an underlying knowledge model. They have beliefs about the world, and choose whether to represent those beliefs accurately or inaccurately using language.

LLMs do not have an underlying knowledge model, they don't have a concept of what is true or false in the world. They just string together words they don't "understand" in ways that are likely to seem credible.

It's not a matter of making better LLMs; it'll take a fundamentally different type of model.

@ct_bergstrom LLMs represent whether they "believe something to be true" in a way that you can extract unsupervised. Not disagreeing that their world model isn't good enough to be used without auxiliary retrieval, but there's some evidence they have one. https://arxiv.org/abs/2212.03827
Discovering Latent Knowledge in Language Models Without Supervision

Existing techniques for training language models can be misaligned with the truth: if we train models with imitation learning, they may reproduce errors that humans make; if we train them to generate text that humans rate highly, they may output errors that human evaluators can't detect. We propose circumventing this issue by directly finding latent knowledge inside the internal activations of a language model in a purely unsupervised way. Specifically, we introduce a method for accurately answering yes-no questions given only unlabeled model activations. It works by finding a direction in activation space that satisfies logical consistency properties, such as that a statement and its negation have opposite truth values. We show that despite using no supervision and no model outputs, our method can recover diverse knowledge represented in large language models: across 6 models and 10 question-answering datasets, it outperforms zero-shot accuracy by 4\% on average. We also find that it cuts prompt sensitivity in half and continues to maintain high accuracy even when models are prompted to generate incorrect answers. Our results provide an initial step toward discovering what language models know, distinct from what they say, even when we don't have access to explicit ground truth labels.

arXiv.org
@ct_bergstrom Another way of disentangling things. For the question you're asking, does the answer exist on the web? If it does, then the problem can't be with the "design" (I. e. the training regime) but rather the power of the model.
@moultano @ct_bergstrom The answer may exist on the web along with its negation.
@moultano @ct_bergstrom I wouldn't use the world 'believe'
@moultano @ct_bergstrom They consume language and then produce language. Their "beliefs" can be about the structure of the English language (when generating text in English), like that adjectives that describe color always go after adjectives that describe size: "the little red hen", not "the red little hen". But they don't have a model of the external world.
@not2b @moultano Precisely. Their "beliefs" have no anchor point outside the world of text.
@ct_bergstrom @not2b Huge fractions of what I know have no anchor point outside of text, like nearly all science and math.
@moultano @ct_bergstrom @not2b Not a whole lot of anchoring for history either. But there are already multimodal models, like Flamingo. If text really has to be grounded in sense experience we will presumably see that research path take the lead and produce better textual prediction.
@moultano @ct_bergstrom @not2b if multimodal training doesn’t make a model much better at predicting text, then at some point we’ll need to revise our priors and consider the possibility that a functional world model can largely be inferred from text
@TedUnderwood @ct_bergstrom @not2b I think it's plausible that a multimodal model might eventually benefit, but the bandwidth advantage of text over video is just too great. You'd need enough video frames to have cause and effect, physics, plot, object permanence.
@moultano @ct_bergstrom @not2b I agree. Eventually sense experience will help, but language alone is providing more of a world model than lots of us would have expected. And if that’s true, then the people dismissing predict-the-next-word as inherently just bullshit generation are prob sneering too hastily.

@TedUnderwood @moultano @ct_bergstrom @not2b

People who understand the technology of large language models aren't dismissing it as "inherently just bullshit generation", but they are warning that its output smoothly mixes both fact and falsehood with no distinction or care.

And they warn that the quantity and impact of this #bullshit could likely surpass that of #politics, #consumerism, and other forms of rampant #disinformation for which we humans have demonstrated we are poorly prepared.

@TedUnderwood @moultano @ct_bergstrom @not2b This connects with Angus Fletcher's article in Narrative (why computer AI will never do what we imagine it can), where he says narrative capacity derives from 500 million years of evolutionary practice at "flailing a flagellum or other primitive limb (...) In response to positive and negative reinforcement."
@Ben_Carver @moultano @ct_bergstrom @not2b I remember that essay. Surprised but grateful that people are lining up to make falsifiable predictions about this stuff.
@ct_bergstrom @not2b @moultano
So true.
Why are babies so eager to learn language? Because they perceive, have needs, and want to interact.
@ct_bergstrom @not2b @moultano
Actually, it must be clear to everyone that current so-called AI cannot have any idea of the actual world without at least the perception of the external world part.
@nicolegoebel
It does have access to parts of the external world. Text. The text we feed it is causally connected to the rest of the world, it didn't just appear in a vacuum. It is seeing a small slice of the universe but it can infer things from that slice. We only see a small slice of reality too, it's no different.
@gnomekat text is second hand perception. Try this analogy from image generating AI: feeding images won't make it understand that humans have bones under their skin.
@gnomekat which doesn't even mean, that you can even call probabilistic treatment of second hand perception like text is anything but what it really is. Don't get me wrong. I think it's nevertheless useful tool and a part of the puzzle. But the hype is built on misunderstanding. And I think this is dangerous.
@ct_bergstrom @not2b @moultano What I keep saying about these toys is that they don't spontaneously anticipate counter-arguments because they don't think, and so have no awareness of an audience. That's kind of where my interest in them ends.
@not2b @moultano @ct_bergstrom Except there is no single English and you need to have a back story of your life and education and produce the relevant text. As far as I can see the current AI generated text reads like a machine trying to mimic a pompous ass deliberately trying to cause offence.
The media have discovered AI. The BS brigade have moved in.

@ct_bergstrom @moultano I'm starting to have doubts about the idea tha LLMs are "stochastic parrots" that can't generalize after watching a short talk from Francois Charton of Meta at NeurIPS 2022.

TL;DR - he trained a small LLM to learn how to diagonalize matrices using only triplets of the similarity transform. No hallucinations were observed.

The talk was "Leveraging Maths to Understand Transformers" https://neurips.cc/virtual/2022/workshop/50015#wse-detail-63846

@ct_bergstrom @moultano Ugh, not small LLM - small transformer based language model.

@ct_bergstrom @moultano @marshall_0i

without doubt LLMs are not like humans, but this statement assumes we know what it means a) for humans to “understand” and b) what processes that rests on, so that we can compare which of these are and not present in LLMs, and c) seems to assume “understand” (whatever it is) is all or none.

As a cognitive scientist, I personally wouldn’t subscribe to any of the three.

see e.g., https://www.journals.uchicago.edu/doi/abs/10.1086/443791

@UlrikeHahn @ct_bergstrom @moultano To me the benefit of doubt goes in the other direction. Whatever it is that we do internally with manipulation of representations about external-world states and properties (or what is readily modeled as manipulation of representations), there is at present little reason to think that LLMs are doing that. The evidence is, if anything, that they don't, given the sorts of crazy "inferences" they sometimes seem to make.

@marshall_0i @ct_bergstrom @moultano

I don’t get that argument: they are incredibly powerful in what they do, well beyond what anything in NLP did 20 years ago. People now spend days trying to find, and then excitedly communicate things that ‘break’ them.
I think it’s a fundamental mistake to conflate whether they should be let loose on an unsuspecting public as a readily available tool (no!) or used in search (no!) with how they relate to human cognition.

1/

@marshall_0i @ct_bergstrom @moultano 2/ so much of the latter discourse seems to proceed *as if we understood human cognition*, which, as a cognitive scientist, I very much feel we don’t.

Top notch cognitive scientists consider them to be interesting and worthy of consideration as models of cognition. That should maybe give people some pause for thought here. I’ve linked to these two recent talks before:

by Ellie Pavlick
https://youtu.be/1_wTMxdVgOI

by Noah Goodman
https://youtu.be/dYXxkS4rrAs

1/3 Pavlick

YouTube

@marshall_0i @ct_bergstrom @moultano 3/ I think they are demonstrably the closest anything has come to generating surface behaviours that look like human performance when it comes to generating and responding to language. So taking them seriously in that context makes sense to me.

Their scientific value and their suitability as online tools ready for current deployment are two separate matters.

@ct_bergstrom @moultano then surely hallucinating is a better description than bullshitting
Playful Technology Limited ~ The Future of Natural Language Processing

NLP systems need knowledge and logic

@moultano @ct_bergstrom
Every loony needs good handlers they say, be it certain Tech Bros or useful half-idiot politicians. And now an "AI" that's supposed to reel in Billions, but is always one step away from praising fascism, declaring the moon is made of cheese, stating 2+2 is 5 or... stealing the watermarks off photos.
So they hire the cheap labor of thousands of "Mechanical Turks" in developing nations to keep the "AI" from going off the rails.
Pure SciFi!
@moultano @ct_bergstrom it's also not like Bullshit originated with AI and was unheard of before.
@moultano @ct_bergstrom They just forgot to include the parts of humans where they strive towards a coherent identity and with having core values. (I mean... some humans really struggle with that as well, but it's also a pretty important 'thing' in humans.)
@ct_bergstrom yeah, i wish more people understood this. she is basically you in HS trying to write a 5 paragraph essay about Antigone after only skimming the cliffnotes
@ct_bergstrom annoyingly the chatbots don’t HAVE to behave this way. They are missing relatively straightforward steps.
- After its main response for the user, it needs to parse it like a prompt from the user.
- It needs to identify statements of fact, names, dates, math, geography, news, etc.
- then classic search engine lookup, history reference lookup, calculator, maps, etc query will rcan generate a confidence score, for each part. Label accordingly as the final output.

@mickdarling @ct_bergstrom

That's fairly clever, and it could be implemented as a browser plugin.

@mickdarling @ct_bergstrom that would result in high BS scores or hidden BS scores since the "if it's on the Internet it is true" doesn't apply anymore (did it ever?). They're selling the idea that LLM helps, BS scores wouldn't help selling it.

@jt_rebelo @ct_bergstrom

All the search engines have old school tools for validating certain content on the web. Various sites are rated for higher accuracy than others and that has been tweaked and modified over the last two decades. Thats not the hard part.

The hard part is just as you say, they are trying to convince people ‘This one new tool is the answer!’ and slapping a 59% confidence metric on a statement about the JWST doesn’t look nearly impressive enough.

@mickdarling @ct_bergstrom yes, I know that every search engine has it already (with better or worse results) and agree that is the easy part (a sisyphus effort, nonetheless). My problem with the hard part is that the confidence metric will always be opaque to us (unless some opensource code is used and outed) and might even be an AI/LLM own assessment. They're as fallible as we are on that (because we were the ones that programmed them and the ones that fed them information).

@jt_rebelo @ct_bergstrom

This is how Google could maintain its Search supremacy, and its golden goose. They list all the sites that back up their LLMs output with a page of links, or more like the sidebar thing that Bing is using.

I would love a Open Source rating system, but you are right it is hard to deobfuscate their inner workings.

My company, Tomorrowish, built a rating system for Tweets more than a decade ago, and though it gave good results, finding out WHY was very tough.

@jt_rebelo @ct_bergstrom And that was our own internal tool we had trouble understanding!

Admittedly, we were a tiny team running a startup on a shoestring, and constantly trying to adjust to pull in some more revenue, SO, we didn't have a lot of time to investigate rigorously something that already worked.
🤷

I think a Google, or Microsoft level company could pull it off. with their personnel and resources.