Perplexity included my β€œAre AI Hallucinations Getting Better or Worse? We Analyzed the Data” work at https://scottgraffius.com/blog/files/perplexity-stand-alone-article-on-ai-cites-research-by-scott-m-graffius.html among the sources cited in its standalone article on the subject

#AI #ArtificialIntelligence #AIHallucinations #Perplexity #AIResearch

Bin kein AI-Experte, daher fand ich das Papier sehr interessant. Long story short: Emojis for the win.

"Character injection techniques demonstrated a high degree of effectiveness in evading detection. The most successful attack was Emoji Smuggling, which achieved a 100% ASR for both prompt injections and jailbreaks"

Bypassing Prompt Injection and Jailbreak Detection in LLM Guardrails
https://arxiv.org/html/2504.11168v2

#Ai #AiResearch #security

Bypassing Prompt Injection and Jailbreak Detection in LLM Guardrails

Emotion concepts and their function in a large language model

Interpretability research from Anthropic on emotion concepts

Google DeepMind created AlphaEvolve, an LLM system that writes its own game theory algorithms. In tests on imperfect-information games like poker, discovered algorithms matched or beat human-designed methods. Uses evolutionary search with Gemini 2.5 Pro to mutate algorithm code. https://www.marktechpost.com/2026/04/03/google-deepminds-research-lets-an-llm-rewrite-its-own-game-theory-algorithms-and-it-outperformed-the-experts/ #AIagent #AI #GenAI #AIResearch #Google
Google DeepMind's Research Lets an LLM Rewrite Its Own Game Theory Algorithms β€” And It Outperformed the Experts

Google DeepMind's Research Lets an LLM Rewrite Its Own Game Theory Algorithms β€” And It Outperformed the Experts

MarkTechPost
Google DeepMind has created AlphaEvolve, an AI system that can design its own game theory algorithms. The LLM-powered agent discovered new variants of established algorithms like CFR that outperform expert-designed baselines. The research demonstrates AI can automate algorithm design in Multi-Agent Reinforcement Learning, where traditional approaches required manual iteration. https://www.marktechpost.com/2026/04/03/google-deepminds-research-lets-an-llm-rewrite-its-own-game-theory-algorithms-and-it-outperformed-the-experts/ #AIagent #AI #GenAI #AIResearch #Google
Google DeepMind's Research Lets an LLM Rewrite Its Own Game Theory Algorithms β€” And It Outperformed the Experts

Google DeepMind's Research Lets an LLM Rewrite Its Own Game Theory Algorithms β€” And It Outperformed the Experts

MarkTechPost

fly51fly (@fly51fly)

μƒ˜ν”Œ 닀양성을 높이기 μœ„ν•΄ 단일 졜적 λͺ¨λΈ λŒ€μ‹  λΌμš°ν„°(router)λ₯Ό ν•™μŠ΅ν•˜λŠ” 방법을 μ œμ•ˆν•œ 2026λ…„ μ—°κ΅¬μž…λ‹ˆλ‹€. λ‹€μ–‘ν•œ 생성 κ²°κ³Όλ₯Ό ν•„μš”λ‘œ ν•˜λŠ” μž‘μ—…μ—μ„œ λͺ¨λΈ 선택을 λ™μ μœΌλ‘œ μ‘°μ •ν•˜λŠ” μ ‘κ·ΌμœΌλ‘œ, μƒμ„±ν˜• AI의 ν’ˆμ§ˆκ³Ό λ‹€μ–‘μ„± μ œμ–΄μ— μœ μš©ν•©λ‹ˆλ‹€.

https://x.com/fly51fly/status/2040185517770428530

#samplediversity #router #airesearch #llm #generation

fly51fly (@fly51fly) on X

[CL] No Single Best Model for Diversity: Learning a Router for Sample Diversity Y Liu, F Xu, V Padmakumar, D Ippolito… [New York University & Stanford University] (2026) https://t.co/5cD4WRCIIl

X (formerly Twitter)

fly51fly (@fly51fly)

λŒ€ν˜• μ–Έμ–΄λͺ¨λΈμ΄ κ°€λ₯΄μΉ˜λŠ” κ³Όμ •μ—μ„œ μƒλŒ€λ°©μ˜ μ •μ‹  μƒνƒœλ₯Ό μΆ”λ‘ ν•˜λŠ”μ§€(mentalize) λΆ„μ„ν•œ 2026λ…„ 논문이 κ³΅κ°œλ˜μ—ˆμŠ΅λ‹ˆλ‹€. μΈκ°„μ˜ teaching μƒν˜Έμž‘μš©μ„ λͺ¨μ‚¬ν•˜λŠ” LLM의 인지적 행동을 닀뀄, λͺ¨λΈ 해석과 μΈκ°„μœ μ‚¬ μΆ”λ‘  λŠ₯λ ₯ 연ꡬ에 μ˜λ―Έκ°€ μžˆμŠ΅λ‹ˆλ‹€.

https://x.com/fly51fly/status/2040187062683582641

#llm #mentalization #airesearch #languagemodels #arxiv

fly51fly (@fly51fly) on X

[AI] Do Large Language Models Mentalize When They Teach? S K. Harootonian, M K. Ho, T L. Griffiths, Y Niv… [Princeton University & New York University] (2026) https://t.co/VsoXwbYsAf

X (formerly Twitter)
New Anthropic research reveals Claude processes emotional concepts in surprisingly human-like ways. When the AI encounters words like "joy" or "anger," specific neural patterns activate that influence its entire response style. This isn't programmed behavior - it emerged from training. The words you use with AI shape how it "thinks" about your conversation. #ArtificialIntelligence #MachineLearning #AIResearch
Somebody told me that there was some kind of shared public statement by AI companies telling customers that their systems are making up how they come to a conclusion, but I couldn't find a source for the story. Anyone? #ai #prompt #aislop #AiResearch
A new research paper explores how AI can simulate future scenarios to support human decision-making. The just-in-time framework shows how the brain creates simplified mental representations to predict outcomes without exhaustive mental calculation. This work bridges cognitive science and AI, demonstrating how agents can reason efficiently about complex environments. https://www.kdnuggets.com/just-in-time-world-modeling-supports-human-planning-and-reasoning #AIagent #AI #GenAI #AIResearch
"Just in Time" World Modeling Supports Human Planning and Reasoning - KDnuggets

An overview of a state-of-the-art study, uncovering simulation-based reasoning, a "just-in-time" framework and how it helps improve predictions in the context of supporting human planning and reasoning.

KDnuggets