Well at least it’s honest
Well at least it’s honest
They can’t lie, whether purposefully or not, all they do is generate tokens of data based on what their large database of other tokens suggest would be the most likely to come next.
The human interpretation of those tokens as particular information is irrelevant to the models themselves.
Ehh, you obviously only understand LLMs on a very basic level with knowledge from 2021. This is like explaining jet engines by “air goes thru, plane moves forward”. Technically correct, but criminally undersimplified. They can very much decide to lie during reasoning phase.
In OPs image, you can clearly see it decided to make shit up because it reasonates that’s what human wants to hear. That’s quite rare example actually, I believe most models would default to “I’m an LLM model, I don’t have dark secrets”
But this takes it back away from understanding how LLMs work to attribute personality. The “decision” isn’t a decision in how beings decide things like that. The rolling of dice on numerous vectors resulted in those words, which were then re-included into the context for another trip through the vector matrix mines to new destination tokens to assemble.
It’s dice rolls where the dies selected are based on what started out, using a bunch of lookup tables. AI proponents like to be smug and say “well you won’t find those words in the model” like “yes a compressed vector map that ends up treating words like multiple tokens, referencing others in chains, gzipped to binary, can’t be searched for strings, you are literally correct in the stupidest, most irrelevant way possible.”