Earlier I posted about using ChatGPT's propensity to fabricate citations entirely as a short-term strategy for detecting journal submissions and classroom assignments that had been written by machine.

I've been playing with the system for the last couple of hours, and as best as I can tell, ChatGPT now does a much better job than it did when first released at only citing papers that actually exist.

They're not perfect—for example, DOIs can be wrong and some are fabricated—but most are not.

If I'm not just imagining things, it raises an interesting question.

While this constitutes an "improvement" in the technology in some manner of speaking, it's unclear whether this is a desirable development. It strikes me as a step that makes detecting more difficult without confering any significant epistemic improvement.

In other words, if the system has really been adjusted to avoid citing fake papers, this constitutes a deliberate choice to create more persuasive bullshit.

Also to be clear, the system is most definitely bullshitting if not outright lying.

Take, for example, this methods section ChatGPT just generated in response to my prompt "Write an scientific review paper with references about whether sunscreen ingredients are carcinogenic."

Unlike many factual claims about background, this is straight up false.

The system decidedly DID NOT conduct a comprehensive literature search on specifically those sources using specifically these terms.

What I'm seeing here is bullshit in the precise technical sense of generating content designed to be persuasive yet produced without regard for truth or logical coherence.

Moreover this bullshit generation is clearly aided and abetting by ongoing development at OpenAI.

In general I've been pretty pleased with OpenAI's approach compared to e.g. what we saw with Galactica, but this is disturbing.

If they are making piecemeal changes to features such as citation behavior, designed to help the system cover its "tells", then I'd think that documenting these changes publicly as @emilymbender has requested seems like the bare minimum as far as social responsibility is concerned.

@ct_bergstrom @emilymbender I would guess they won’t document such changes because a) it would help other companies developing competing systems, and b) it would make clear how the system is just a bullshit generator that needs continual tweaking and checking of its output to paper over that fact.

@ct_bergstrom @emilymbender

stuff like this really makes the "hallucination" description seem flawed, and more like we're talking about a corrupt bureaucracy that's simply papering over evidence of transgressions, rather than addressing the core issue

(that AIs have no intrinsic tether to reality the way humans do, and so AIs have nothing particularly anchoring them to anything that's *real* more than, say, anything that's complete and utter bullshit)

@ali @ct_bergstrom @emilymbender "corrupt bureaucracy motivated only by the formsn looking right" here is a really excellent thought tool for working with opaque models trained on obscure-to-us data.

"What is the least Real Work such a model would have to do to present the appropriate papers? Assume the bureaucrat will do that."

@ali @ct_bergstrom @emilymbender

The CYA games that OpenAI is playing, like penalizing imaginary citations and recognizing transparently racist leading questions, are really just attempts to nudge the Path of Least Work away from the most obvious shortcuts.

But it will almost always be easier to bullshit, or, in the context of this analogy, make a minor error on internal paperwork that allows the bureaucrat to shortcut doing real work

@ali @ct_bergstrom @emilymbender

It's a little weird to see the Chinese Room reappear, but with a different question entirely -- we don't care if the bureaucracy "is thinking"; we care if it's "serving the people".

Functionaries who follow all the rules with the least possible effort _might_ be "serving the people" but, in a corrupt bureaucracy or a bullshitter, it's optimized towards "least effort" instead

@ct_bergstrom the gold standard for these #AI is to generate output that *sounds like something a human would say*. Period. There's no need to actually perform those literature searches, it's just something a human would say. There's no capacity to assess truth or fact. These aren't the objective. It's like they were trained on speeches by Donald Trump. "I know words. I have the best words."
@ct_bergstrom I can see why they might want to make the system not output fake references if they are conceiving of it as an information access system, but then this just underscores the point that "general text synthesis machine" is a bad UI. If making it a better search engine means it's also better at making fake scientific papers, that's a fundamentally flawed design.

@emilymbender @ct_bergstrom

I agree.

It wouldn’t make sense for them to block references to nonexistent materials but keep references to existing sources that weren’t actually referenced by the system.

It’s not an information access system.

@ct_bergstrom I don’t see any evidence that they are making piecemeal changes to “features” such as citation behavior.