Apple did the research; LLMs cannot do formal reasoning. Results change by as much as 10% if something as basic as the names change.
https://garymarcus.substack.com/p/llms-dont-do-formal-reasoning-and
Apple did the research; LLMs cannot do formal reasoning. Results change by as much as 10% if something as basic as the names change.
https://garymarcus.substack.com/p/llms-dont-do-formal-reasoning-and
@ShadowJonathan it’s really weird that some people are pushing LLM as something that can reason, while its architecture is Key-Value storage with sophisticated probabilistic query and value encoding mechanisms.
LLMs just don’t have enough layers for anything besides queries, so it can’t have any relational capabilities that allow to make multi step decisions.
Also tokenization hides a lot of structure of the language from encoding process, which adds additional source of errors.
I’m sure we can build something that can reason at some point, but it requires very diffirent and more complex architecture.