Here's the first proof of concept I've seen of a prompt injection attack against ChatGPT Plugins - the successful attack uses Zapier to access the user's email and then exfiltrates the data using WebPilot https://simonwillison.net/2023/May/19/chatgpt-prompt-injection/
Let ChatGPT visit a website and have your email stolen

Johann Rehberger provides a screenshot of the first working proof of concept I've seen of a prompt injection attack against ChatGPT Plugins that demonstrates exfiltration of private data. He uses …

@simon the next new thing ‘copy this into your chatgpt to do xyz… ‘