Could LLMs develop general intelligent behavior – which might endanger humans? 🤖 According to a new study of UKP Lab & BathNLP: Probably not – at least not in the near future. (1/🧵)

Read the full press release by @TU Darmstadt at @idw_online: https://nachrichten.idw-online.de/2024/08/12/independent-complex-thinking-not-yet-possible-after-all-study-led-by-tu-shows-limitations-of-chatgpt-co

Independent, complex thinking not (yet) possible after all: Study led by TU shows limitations of ChatGPT & co.

The 2024 study, authored by Sheng Lu, Irina Bigoulaeva, Rachneet Sachdeva, Harish Tayyar Madabushi and Iryna Gurevych (BathNLP Lab | Ubiquitous Knowledge Processing (UKP) Lab), was just presented at #ACL2024NLP. It found no evidence of emergent abilities in LLMs that go beyond in-context learning.(2/🧵)

ArXiv: https://arxiv.org/abs/2309.01809

Are Emergent Abilities in Large Language Models just In-Context Learning?

Large language models, comprising billions of parameters and pre-trained on extensive web-scale corpora, have been claimed to acquire certain capabilities without having been specifically trained on them. These capabilities, referred to as "emergent abilities," have been a driving force in discussions regarding the potentials and risks of language models. A key challenge in evaluating emergent abilities is that they are confounded by model competencies that arise through alternative prompting techniques, including in-context learning, which is the ability of models to complete a task based on a few examples. We present a novel theory that explains emergent abilities, taking into account their potential confounding factors, and rigorously substantiate this theory through over 1000 experiments. Our findings suggest that purported emergent abilities are not truly emergent, but result from a combination of in-context learning, model memory, and linguistic knowledge. Our work is a foundational step in explaining language model performance, providing a template for their efficient use and clarifying the paradox of their ability to excel in some instances while faltering in others. Thus, we demonstrate that their capabilities should not be overestimated.

arXiv.org

»[O]ur results do not mean that AI is not a threat at all« emphasized Iryna Gurevych. »[But future research should] focus on other risks posed by the models, such as their potential to be used to generate fake news.« (3/🧵)

Read the full press release here: https://nachrichten.idw-online.de/2024/08/12/independent-complex-thinking-not-yet-possible-after-all-study-led-by-tu-shows-limitations-of-chatgpt-co

#ACL2024NLP

Independent, complex thinking not (yet) possible after all: Study led by TU shows limitations of ChatGPT & co.

Doch (noch) kein selbstständiges, komplexes Denken möglich

KI-Modelle wie ChatGPT sind laut einer neuen Studie unter führender Beteiligung der TU Darmstadt offenbar weniger selbstständig lernfähig als bisher angenommen. Es gebe keine Hinweise darauf, dass die sogenannten Large Language Models (LLMs) anfingen, ein allgemeines „intelligentes“ Verhalten zu entwickeln, das ihnen etwa ein planvolles oder intuitives Vorgehen oder komplexes Denken ermögliche, heißt es in der Untersuchung. Die Studie wird im August auf der Jahrestagung der renommierten Association for Computational Linguistics (ACL) in Bangkok vorgestellt, der größten internationalen Konferenz zur Automatischen Sprachverarbeitung.

TU Darmstadt