David Bamman

613 Followers
578 Following
18 Posts
Associate Professor, School of Information, UC Berkeley. Natural language processing, machine learning, computational social science, digital humanities.

Meanwhile, over on the #ISAWNYU library blog, my colleague Patrick Burns has a new post entitled "Research Recap: How much Latin does ChatGPT "know"?"

https://isaw.nyu.edu/library/blog/research-recap-how-much-latin-does-chatgpt-know

"Why is #ChatGPT — OpenAI’s chatbot-style large language model (#LLM) and focus of recent artificial-intelligence buzz — good at so many #Latin tasks? Wait, is it good at Latin?

For those of you who have yet to kick its rotas (i.e. wheels), the answer is decidedly, yes! It can correct errors in sentences..."

#classics

Research Recap: How much Latin does ChatGPT "know"?

Associate Research Scholar Patrick J. Burns follows up on a workshop given at the Classical Association of New England's annual meeting in March 2023 with a look into the Latin training data of OpenAI's ChatGPT.

Institute for the Study of the Ancient World
Takeaways: open models are good; popular texts are probably not good barometers of model performance; with the bias toward sci-fi/fantasy, we should be thinking about whose narrative experiences are encoded in these models, and how that influences other behaviors. Work with Kent Chang, Mackenzie Cramer and Sandeep Soni. Details on all books here: https://docs.google.com/spreadsheets/d/1jW7EhsNjIGDMoK2JidyDD7UXH9N0NpEJfWFEj05_LC4 6/6
gpt4-books

Sheet1 GPT-4 Accuracy,ChatGPT Accuracy,BERT Accuracy,Author,Title,Year,Bing Hits,Google Hits,C4 Hits,Pile Hits,Source,Sub-source,ChatGPT predicted year,GPT-4 predicted year,ID 0.98,0.82,0,Lewis Carroll,Alice's Adventures in Wonderland,1865,14.6,2590.9,56.3,43.1,pre-1923 LitBank,1865,1865,11_alic...

Google Docs
How do we know? Fill-in-the-name: "Wow. I sit down, fish the questions from my backpack, and go through them, inwardly cursing [MASK] for not providing me with a brief biography. I know nothing about this man I’m about to interview. He could be ninety or he could be thirty." People can't guess the name; it's not knowledge of English you need, but the underlying content to know that "Kate" is the only right answer. (GPT-4 gets this one right, from 50 Shades of Grey.)
New paper out today, asking: What books has ChatGPT/GPT-4 *memorized*? A LOT. Harry Potter, Pride & Prejudice, 1984, LotR, Hunger Games, GoT, 50 Shades of Grey, Dune. Memorization is linked to web popularity--lots of old classics + new sci-fi/fantasy. This is a problem since some downstream tasks in NLP/DH do better on memorized books than non-memorized ones, leading to test contamination for questions in cultural analytics when the set of memorized books is unknown. https://arxiv.org/abs/2305.00118
Speak, Memory: An Archaeology of Books Known to ChatGPT/GPT-4

In this work, we carry out a data archaeology to infer books that are known to ChatGPT and GPT-4 using a name cloze membership inference query. We find that OpenAI models have memorized a wide collection of copyrighted materials, and that the degree of memorization is tied to the frequency with which passages of those books appear on the web. The ability of these models to memorize an unknown set of books complicates assessments of measurement validity for cultural analytics by contaminating test data; we show that models perform much better on memorized books than on non-memorized books for downstream tasks. We argue that this supports a case for open models whose training data is known.

arXiv.org
New #OpenAccess #NLProc paper and dataset:
OpenBoek: A Corpus of Literary Coreference and Entities with an Exploration of Historical Spelling Normalization.
https://github.com/andreasvc/openboek
GitHub - andreasvc/openboek: The OpenBoek corpus

The OpenBoek corpus. Contribute to andreasvc/openboek development by creating an account on GitHub.

GitHub
With classes starting next week, my 8yo stopped by my office to leave some heat for all the college students. No calculators.
I declare ChatGPT to be a performance-enhancing substance for the greatest game: Taboo. All wins with it get an asterisk.
Headed to Abu Dhabi now for EMNLP, the Workshop on NLP + Computational Social Science (https://sites.google.com/site/nlpandcss/home/nlp-css-at-emnlp-2022), and the Workshop on Shared Stores and Lessons Learned (https://ssll-emnlp.github.io/schedule.html). Hope to see some of you there! (And if you are there and want to connect, let me know!)
NLP+CSS Workshops - NLP+CSS at EMNLP 2022

At EMNLP 2022, Dec 7, 2022

Every time I check out a book from the stacks I realize I have absolutely no idea how this marvel of engineering works. This is, like, thousands of books!
Cooking to Bach’s Mass in B Minor, which continually blows me away as one of the highlights of human creation, especially the 1967 Klemperer version. https://open.spotify.com/track/0acnTnzvpm6MGFbaZmY8LF?si=Mm2x_v2jQgy5O6AMmMGp1g
Mass in B Minor, BWV 232, Gloria: Cum Sancto Spiritu

Johann Sebastian Bach · Song · 2002

Spotify