The point isn’t that some models are better than others. The point is that yet again it’s an example that LLMs are not thinking machines and you can’t trust anything from them and people are burning the world to run a glorified auto complete.
Most people get their info from forums and blog posts. Unless you limit yourself to nothing but peer reviewed papers, you probably do some kind of calculation on the legitimacy of whatever source you are perusing.