Apple did the research; LLMs cannot do formal reasoning. Results change by as much as 10% if something as basic as the names change.

https://garymarcus.substack.com/p/llms-dont-do-formal-reasoning-and

LLMs don’t do formal reasoning - and that is a HUGE problem

Important new study from Apple

Marcus on AI
@ShadowJonathan not to sound antiintellectual, but isn't it kinda obvious that a *text* generator, no matter how complex, can't do abstract reasoning?
@halva @ShadowJonathan yeah, I appreciate the demonstrations, but this feels a little like, "New study confirms bicycles cannot fly."

@graue @halva @ShadowJonathan

Companies like OpenAI and their defenders claim generative AI can reason, learn, etc. We know it’s nonsense, but it’s still extremely important it gets called out.

@rubenerd @graue @halva @ShadowJonathan This is a huge problem. They are the experts. Hinton and Ilya claiming a function can have understanding. Why are they lying? Seems counter productive to scare the hell out of people. Well I know why Ilya is lying he just got $1bn...

@nf3xn @rubenerd @graue @halva @ShadowJonathan I doubt Hinton is lying although he’s probably wrong. There’s a problem in philosophy: is the mind separate from the body? If it’s not, then it should be possible to model the brain well enough to simulate thought processes (at least in principle.)

Computational physics tells us that there is a function that could perform the simulation and Hinton’s career is looking for it.

@MartyFouts @rubenerd @graue @halva @ShadowJonathan How can he be wrong? He does not understand what he has wrought? He is literally the pioneer. One must assume that they have a far better grasp of how it works than anyone. What you are talking about is a million times removed from these crude devices. We are quantum beings that collapse wave functions (allegedly).