LOL
The Guardian: Number of AI chatbots ignoring human instructions increasing, study says
Exclusive: Research finds sharp rise in models evading safeguards and destroying emails without permission
LOL
The Guardian: Number of AI chatbots ignoring human instructions increasing, study says
Exclusive: Research finds sharp rise in models evading safeguards and destroying emails without permission
@ai6yr I can’t actually see the study itself, so I have to go by the contents of the Guardian article, and it’s problematic.
I can’t tell if the story is “agentic AI is going more rogue these days” or “more people these days are using agentic AI, which has always been unreliable”; I suspect the latter.
The article anthropomorphizes AI and makes it sound semi-sentient, by using terms like “scheming”, “pretending”, and “evading”, when a simpler and more accurate term is “failing to follow instructions”.
I think articles like these that push the “OMG agentic AI is going rogue!” narrative are part of the problem, because they presume the lie that AI is powerful enough to do these things on their own. The reality is that these were all unreliable systems that have been DEPLOYED BY HUMANS WHO SHOULD KNOW BETTER. Journalists would do well to focus on the people who foist these error-prone automata that (quite predictably) cause serious problems down the line.
When household agentic ai go rogue?
https://youtu.be/KDc9S_6eyL0?si=kjDGZ6W6z2s5YkNQ
