90 Followers
20 Following
120 Posts
Hacking neural networks so that we don't get stuck in the matrix.
Entrepreneur. Author. Red Team Director.
Bloghttps://embracethered.com

What happened there? 🧐

πŸ‘‰ The original post with the question contains hidden Unicode Tag code points.

Unicode Tags mirror ASCII, but are invisible in UI elements. πŸ‘€

AI Application Security Vulnerabilities πŸ‘¨β€πŸ’»

Learn the hacks, stop the attacks!

Perplexity Demo Time! 🍿

Check out my latest blog post called Terminal DiLLMa πŸ”₯

Learn the dangers of printing LLM output to the terminal console or log files!

Includes some neat demos and also how to fix your LLM powered CLI apps!

#pentest #bugbounty #ai #ml #redteam

https://embracethered.com/blog/posts/2024/terminal-dillmas-prompt-injection-ansi-sequences/

Terminal DiLLMa: LLM-powered Apps Can Hijack Your Terminal Via Prompt Injection Β· Embrace The Red

Embrace The Red
No issues for me πŸ™‚

🧟 πŸ€– The ZombAIs are here! πŸ€–πŸ§Ÿ

From Prompt Injection to Command and Control with Claude Computer Use by Anthropic.

Interesting experiment and results, in the end it was simpler then I thought...

https://embracethered.com/blog/posts/2024/claude-computer-use-c2-the-zombais-are-coming/

ZombAIs: From Prompt Injection to C2 with Claude Computer Use Β· Embrace The Red

Embrace The Red

Important Copilot settings you might want to disable, or at least be aware of.

I noticed a few days ago that Copilot randomly used information from past conversations I had.

Although, it doesn't do it today - maybe feature got temporarily disabled?

πŸ‘‰ NotebookLMπŸ‘¨β€πŸ’»

Took the first chapter of The Count of Monte Cristo, injected a few words, and it did this. πŸ˜‡

πŸ€” What do you get when you combine Prompt Injection + Data Exfiltration + Long-term Memory?

πŸ‘‰ Persistent SpAIware!

🚨 OpenAI fixed a persistent data exfiltration issue in the macOS ChatGPT app that I reported.

πŸ”’ Make sure to stay up to date and use the latest ChatGPT app

Blog with details: https://embracethered.com/blog/posts/2024/chatgpt-macos-app-persistent-data-exfiltration/

#redteam #llm #chatgpt #openai #spyware #pentesting

Spyware Injection Into Your ChatGPT's Long-Term Memory (SpAIware) Β· Embrace The Red

Embrace The Red

πŸ”₯ Microsoft fixed a high severity data exfiltration exploit chain in Copilot that I reported earlier this year.

It was possible for a phishing mail to steal PII via prompt injection, including the contents of entire emails and other documents.

The demonstrated exploit chain consists of techniques that didn't even exist 2 years ago. πŸ”₯

In particular, it involves:

1. Prompt Injection πŸ’‰
2. Automatic Tool Invocation (without human in loop) to bring PII into chat context βš™οΈ
3. ASCII Smuggling 🫣
4. Rendering of benign link + invisible text πŸ‘€
5. (Optional) Conditional instructions to only trigger when certain users view the content ☝️

Discussing two demos (stealing sales data and MFA codes), including the videos I had shared with MSRC in February.

Detailed blog post: https://embracethered.com/blog/posts/2024/m365-copilot-prompt-injection-tool-invocation-and-data-exfil-using-ascii-smuggling/ #infosec

Microsoft Copilot: From Prompt Injection to Exfiltration of Personal Information Β· Embrace The Red

Embrace The Red

Full story here:

Google Colab AI: Data Leakage Through Image Rendering Fixed. Some Risks Remain.

https://embracethered.com/blog/posts/2024/google-colab-image-render-exfil/

#redteam #llm #gemini #infosec

Google Colab AI: Data Leakage Through Image Rendering Fixed. Some Risks Remain. Β· Embrace The Red

Embrace The Red