Our AI chatbot passed every prompt injection test we threw at it. Then we just asked it nicely for customer data, and it happily obliged.

New from our ASMOC team, how a vibe-coded website with LLM became a high-risk finding on a client's attack surface.

https://blog.blacklanternsecurity.com/p/artificial-foolishness-the-hidden

Artificial Foolishness: The Hidden Dangers of External-Facing LLMs

The rise of AI opens more doors to attackers

Black Lantern Security (BLSOPS)