Study finds LLMs can identify their own mistakes

It turns out that LLMs encode quite a bit of knowledge about the truthfulness of their answers, even when they give the wrong one.

VentureBeat

@nic221
Interpretation of internal representation seems to be a major problem with generative AI. #aihype

https://news.mit.edu/2024/generative-ai-lacks-coherent-world-understanding-1105

Despite its impressive output, generative AI doesn’t have a coherent understanding of the world

Large language models can achieve incredible performance on some tasks without having internalized a coherent model of the world or the rules that govern it, MIT researchers find. This means these models are likely to fail unexpectedly if they are deployed in situations where the environment or task slightly changes.

MIT News | Massachusetts Institute of Technology