An LLM "creates textual claims, and then predicts the citations that might be associated with similar text. Obviously, this practice violates all norms of scholarly citation.

At best, LLMs gesticulate toward the shoulders of giants."

@emilymbender , Jevin West, and I contributed to this perspective piece in PNAS. We took a skeptical position; others are very much enthusiasts. Before you pillory me for some random quote in this article, we strongly disagree with some of the claims in the other perspectives.

https://www.pnas.org/doi/10.1073/pnas.2401227121

@ct_bergstrom @emilymbender

Carl, I really appreciate your portion of the article.  The following quotes from you and your co-writers pinpoint my concerns.  That such misconstructions are being projected across the net and quoted w no context around their generation appears to me likely to  trend towards chaotic dumbing down of our collective understandings. 

From the Article above:
(https://www.pnas.org/doi/10.1073/pnas.2401227121)

"LLMs are simply models of word form distributions extracted from text—not models of the information that people might get from reading that text (45). "

"And most importantly, when someone uses an LLM to generate a literature review, the claims generated are not directly derived from the manuscripts cited. Rather, the machine creates textual claims, and then predicts the citations that might be associated with similar text. Obviously, this practice violates all norms of scholarly citation. At best, LLMs gesticulate toward the shoulders of giants. "

"Automatically generating something that looks like a manuscript is very different from the iterative process of actually writing a manuscript. Yet the output can be difficult to distinguish, particularly in a cursory read or by inexpert readers. "

"This false dichotomy between communication and investigation reflects a fundamental misunderstanding of the nature of science (56) that devalues the communicative aspects of science and ignores the role of writing in the process of formulating, organizing, and refining ideas."

This last being, from my view, one of the most critical dangers to new discovery and creative thought that mis-use  of AI represents. 

Thank you for clearly pointing this out.

@oldoldcojote @emilymbender thank you so much for the kind words and much credit to Emily for helping me see this all in the first place.

@ct_bergstrom @emilymbender

I have a transcript snack for you both;
"A Conversation with Perplexity AI"
(15 minute read)

A shocking transcript of a January 2025 interaction with a new AI enhanced search engine.
https://gchrd.org/docs/PerplexityDiscussesLies.pdf

#AI #DeepSeek #lies #Democracy

@GCHRD @ct_bergstrom I don't waste any of my time reading synthetic text.

@emilymbender @ct_bergstrom

Ah well, this human wrote;
"Defending Truth, the Next Level for Democracy"
The public has a right to truth and good governance requires that it be delivered.
https://gchrd.org/docs/PublicTrust.pdf

Exploring the scope and methods of attacks against public discourse, with a detailed discussion of the solutions at hand.
(15 minute read) Edited January 2025

Your 15 minutes and cogent comments are appreciated. 🙂