"Never bring a knife to a gun fight."
"Never bring a knife to a gun fight."
"But as clever users have found in the past, if you ask an AI bot to pretend to be someone else, that appears to be all you need to give it permission to say naughty things. This time, it isn’t just enough to get the chat bot to say things it’s not supposed to, but rather have it do so while assuming the role of a kind, elderly relative."
https://kotaku.com/chatgpt-ai-discord-clyde-chatbot-exploit-jailbreak-1850352678