New prompt injection papers: Agents Rule of Two and The Attacker Moves Second
Two interesting new papers regarding LLM security and prompt injection came to my attention this weekend. Agents Rule of Two: A Practical Approach to AI Agent Security The first is …
@simon With the rule of two, isn’t the combination of untrusted inputs and changing state in an agent potentially quite dangerous already, even without the access to private data? (Disclosure: didn’t read the paper, just your post)