Over just a few months, ChatGPT went from accurately answering a simple math problem 98% of the time to just 2%, study finds

https://lemmy.one/post/1385551

Over just a few months, ChatGPT went from accurately answering a simple math problem 98% of the time to just 2%, study finds - Lemmy.one

Can we discuss how it’s possible that the paid model got worse and the free one got better? Is it because the free one is being trained on a larger pool of users or what?

Has it ever been good at mathematical/logical problems? It seems it’s good at text-based problems like imitating a writing style or even writing code, but if you ask it a logic puzzle like “if two cars take 3 hours to reach NYC, how long will 5 cars take?” it often fails completely.

Humans are capable of both understanding language and logical thought, I’m not sure if the latter will ever be easy for the LLMs to do, and perhaps older Symbolic approaches to AI might perform better in this space.