Is it just me? Am I using this wrong or am I asking questions that are too hard?
Here’s an example of a hallucination that happened while explaining away another hallucination I called it out on. I rarely have experiences other than these.
Hallucinations remain common, and the strongest mitigation is a combination of search use & "extended thinking". Even then, models may over-privilege low quality sources.
May I ask which version of ChatGPT this is from (paid or free; Thinking mode enabled or not)? Here's the result with 5.2 + Thinking mode, in case it's useful for comparison:
https://chatgpt.com/share/694cb369-b3d4-800b-8abc-e29d565910d8
@scottjenson @mwichary the problem is whatever the percentage if you have to know the field to detect which answers are wrong (or how they are missing key context or steps etc) then most users of such systems will either assume that the wrong answers are right
Or will eventually distrust all the answers even the real/correct ones and be unable to figure out what the right answers are
This is already happening as online search degrades and content spreads so evaluating reliable info is hard
@scottjenson @mwichary I know that for topics I know at an expert level AI’s answers (as found in places like Google’s ai generated summaries that you can’t easily escape when using google search) but also from places like FB’s ai generated content about stuff posted to meta properties etc are almost always deeply flawed and contain mistakes and hallucinations.
But when it is something I’m less expert on its far harder even for me (with decades of search expertise) to find the accurate info now