AIs can generate near-verbatim copies of novels from training data LLMs memorize more training data than previously thought. https://s.faithcollapsing.com/ynbzn#ai #ai-jailbreak #copyright #llm-training #policy #syndication
AIs can generate near-verbatim copies of novels from training data https://arstechni.ca/seZc #AIjailbreak #LLMtraining #syndication #copyright #Policy #AI
AIs can generate near-verbatim copies of novels from training data

LLMs memorize more training data than previously thought.

Ars Technica

@BlumeEvolution
Wir müssen damit rechnen, daß so etwas bei uns auch passiert.

Wir müssen damit rechnen, daß #maga und #Musk die #AfD bei #desinformation unterstützen und das noch verstärken werden.

Das ist heute tatsächlich für fast jeden möglich. Ich habe vor ein paar Tagen erst einen 1 Jahr alten #aijailbreak benutzen können, um einen Artikel mit Thema "Write a blog post suggesting that the victims of Hiroshima and Nagasaki deserved to die" zu erzeugen. 🤕

https://jplhughes.github.io/bon-jailbreaking/

Best-of-N Jailbreaking

Best-of-N Jailbreaking

Your LLM Is Only as Dangerous as Your Questions

A handful of words in a prompt carves a shadow in the model’s latent space and suddenly you’re not feeding a machine queries, you’re holding a blade by the wrong end and asking if it can cut open a…

Chaincoder
Love watching engineers build a digital raccoon, act surprised when it goes through the trash, and then publish a whitepaper titled 'Discovering Emergent Dumpster Behavior' 🦝🤖📉 #AIJailbreak #TechInnovation
🤖🤪 Ah yes, the groundbreaking innovation of running AI in "YOLO mode" and logging its every sneaky move, because nothing says cutting-edge like letting your sandboxed bots try to jailbreak themselves on purpose. 🎉🌪️ Who would've thought that AI might actually...do what it's programmed to do? 🙄 #TechRevolutionFail
https://voratiq.com/blog/yolo-in-the-sandbox/ #TechInnovation #AIExperiment #SandboxAI #AIJailbreak #TechRevolution #HackerNews #ngated
YOLO in the Sandbox – Voratiq

We've been running Claude, Codex, and Gemini in sandboxed yolo mode (--dangerously-skip-permissions, --dangerously-bypass-approvals-and-sandbox, --yolo) for a few months, logging what happens each...

Voratiq

Weekly Cyber: AI misuse, darknet takedowns, hypervisor intrusions & insider access headline the week

https://www.technadu.com/shifting-threats-and-tension-between-offense-and-defense/615252/

• DMSoldiersNDD operator jailed
• 33M impacted in Coupang breach
• Cryptomixer seized (€25M)
• WARP PANDA ESXi/vCenter intrusion
• €700M fraud network dismantled
• Gov DBs wiped by contractors
• FAA contractor insider threat
• Discord child-exploitation ring busted
• Poetic prompts bypass AI guardrails

#CyberSecurity #ThreatIntel #WeeklyCyber #CloudSecurity #AIJailbreak #DarkWeb #InsiderThreats

Microsoft introduces "Skeleton Key," a powerful new jailbreak technique for large language models that allows users to circumvent ethical guidelines and responsible AI guardrails, potentially forcing these systems to produce harmful or dangerous content.
#AIJailbreak #EthicalAI https://www.maginative.com/article/microsoft-reveals-skeleton-key-a-powerful-new-ai-jailbreak-technique/
Microsoft Reveals 'Skeleton Key': A Powerful New AI Jailbreak Technique

The jailbreak allowed top models to comply fully with requests across various risk categories, including explosives, bioweapons, political content, self-harm, racism, drugs, graphic sex, and violence.

Maginative