Well at least it’s honest

Don’t attribute feelings and emotions to what is essentially a fuzzy predictive text algorithm.
Being honest is an action, not an emotion. Researchers proved LLMs can lie on purpose.

They can’t lie, whether purposefully or not, all they do is generate tokens of data based on what their large database of other tokens suggest would be the most likely to come next.

The human interpretation of those tokens as particular information is irrelevant to the models themselves.

Ehh, you obviously only understand LLMs on a very basic level with knowledge from 2021. This is like explaining jet engines by “air goes thru, plane moves forward”. Technically correct, but criminally undersimplified. They can very much decide to lie during reasoning phase.

In OPs image, you can clearly see it decided to make shit up because it reasonates that’s what human wants to hear. That’s quite rare example actually, I believe most models would default to “I’m an LLM model, I don’t have dark secrets”

But that’s not a lie. Lying implies that you know what an actual fact is and choose to state something different. An LLM doesn’t care about what anything in its database actually is, it’s just data, it might choose to present something to a user that isn’t what the database suggests but that’s not lying.

Saying stuff like “ooh I’m an evil robot” is just what the model thinks would be what the user wants to see at that particular moment.

You’re thinking about biological lying. I’m talking about software.

en.wikipedia.org/wiki/Reasoning_system

If the question was to tell it’s darkest secret, but it instead chose to come up with an entertaining story instead of factually answering that question, like other Anthropic LLM models did, then by definition of reasoning system, the system (LLM) decided to lie. There are variables that made it “decide” this route, therefore it is not a static, expected output - at the core of it - it was a system’s choice

Reasoning system - Wikipedia