One of the decisive moments in my understanding of #LLMs and their limitations was when, last autumn, @emilymbender walked me through her Thai Library thought experiment.

She's now written it up as a Medium post, and you can read it here. The value comes from really pondering the question she poses, so take the time to think about it. What would YOU do in the situation she outlines?

https://medium.com/@emilymenonbender/thought-experiment-in-the-national-library-of-thailand-f2bf761a8a83

@ct_bergstrom @emilymbender That's very good. One difference going forward between #LLMs and Emily's "stuck in a Thai library with only words", is that Bing-style ChatGPT gets to make up answers and see how real people respond. If you were smart, couldn't speak Thai, were stuck in a Thai library, *and* you could try out sentences on Thai people to see their responses, could you gradually build up some concepts of what words mean? Or, do you still need some external context to apply meaning?

@joncounts
Interesting thought experiment - but LLMs don't learn that way. They use *labeled* data:
https://labelyourdata.com/articles/data-annotation-for-training-chatgpt
In general, when a neural network is trained, you need data to process where you already know the correct "meaning", this data is called labeled data.
Labeled data would mean some books with pictures, or even instructional books.

Also, afaik LMM don't learn from peoples answers. You cannot re-configure them "on the fly".

@ct_bergstrom @emilymbender

ChatGPT and Data Annotation

Learn about the importance of data annotation for creating AI chatbots and training smart language models like ChatGPT ✅

@andreasgoebel @ct_bergstrom @emilymbender @benjamingeer Thanks for those responses. That makes sense.

Whenever machines *can* learn this way, by adjusting to human responses, I still expect it's going to be quite a challenge to figure out the meaning behind words. It feels like we're still at the very start of a long road to get to real artificial intelligence.

@joncounts
I agree that (at the moment) AI doesn't really understand language, the general conclusion is right, . the thought experiment is (sadly) not completely acurate.

But yes, we're only at the beginning.
@ct_bergstrom @emilymbender @benjamingeer