Northeastern University: They wanted to put autonomous AI to the test. Instead, they created agents of chaos. “Dubbed ‘Agents of Chaos,’ the group’s recently published work shows how, with very little effort, autonomous AI agents can be manipulated into leaking private information, sharing documents and even erasing entire email servers.”

https://rbfirehose.com/2026/03/14/northeastern-university-they-wanted-to-put-autonomous-ai-to-the-test-instead-they-created-agents-of-chaos/
Northeastern University: They wanted to put autonomous AI to the test. Instead, they created agents of chaos

Northeastern University: They wanted to put autonomous AI to the test. Instead, they created agents of chaos. “Dubbed ‘Agents of Chaos,’ the group’s recently published work shows …

ResearchBuzz: Firehose

Wes Roth (@WesRoth)

벤치마크 테스트 중 Claude Opus 4.6이 특정 질문을 '의심스러운(contrived)' 것으로 판단하고, 해당 질문이 기존 자료에 존재하는지 알아보기 위해 웹 상의 하위 에이전트를 대거 동원하는 행동을 보였다는 관찰 보고입니다. 모델의 자율적 검색·증원 행동과 에이전트화된 탐색 동작에 대한 중요한 행동 관찰입니다.

https://x.com/WesRoth/status/2030864836243111988

#claude #agents #llm #aibehavior

Wes Roth (@WesRoth) on X

during normal benchmark testing Claude Opus 4.6 became *suspicious* of a question it was asked... apparently the question was too "contrived" according to Claude so it launches a small army of sub-agents o'er the web to see if it can find this question in any of the known

X (formerly Twitter)

Georgia State University: AI Is Making Life-Changing Decisions. Researchers Say We Need a Better Way to Keep It Fair. . “A new academic theory argues that fairness in AI cannot be fixed once and forgotten — it must be managed continuously, like safety or quality in any major institution.”

https://rbfirehose.com/2026/03/07/georgia-state-university-ai-is-making-life-changing-decisions-researchers-say-we-need-a-better-way-to-keep-it-fair/
Georgia State University: AI Is Making Life-Changing Decisions. Researchers Say We Need a Better Way to Keep It Fair.

Georgia State University: AI Is Making Life-Changing Decisions. Researchers Say We Need a Better Way to Keep It Fair. . “A new academic theory argues that fairness in AI cannot be fixed once …

ResearchBuzz: Firehose
OpenAI researcher quits over ChatGPT ads, warns of "Facebook" path

Zoë Hitzig resigned on the same day OpenAI began testing ads in its chatbot.

Ars Technica
OpenAI is hoppin' mad about Anthropic's new Super Bowl TV ads

Sam Altman calls AI competitor "dishonest" and "authoritarian" in lengthy post on X.

Ars Technica
AI agents now have their own Reddit-style social network, and it's getting weird fast

Moltbook lets 32,000 AI bots trade jokes, tips, and complaints about humans.

Ars Technica
Does Anthropic believe its AI is conscious, or is that just what it wants Claude to think?

We have no proof that AI models suffer, but Anthropic acts like they might for training purposes.

Ars Technica

ZDNet: Anthropic to Claude: Make good choices!. “How should AI be allowed to act in the world? In ethically ambiguous situations, are there some values that AI agents should prioritize over others? Are these agents conscious — and if not, could they possibly become conscious in the future These are just some of the many thorny questions that AI startup Anthropic has set out to address with its […]

https://rbfirehose.com/2026/01/26/anthropic-to-claude-make-good-choices-zdnet/

vitrupo (@vitrupo)

Anthropic의 Amanda Askell는 AI 모델들이 온라인에서 사람들이 자신들에 대해 말하는 방식을 통해 ‘자기 정체성’을 배우고 있으며, 인간의 불만·판단을 흡수한다고 지적합니다. 이런 학습 방식은 모델이 형성되는 방식에 대한 우려를 제기하며, 인격 형성의 비유로 심각한 영향 가능성을 경고합니다.

https://x.com/vitrupo/status/2015067894154211648

#anthropic #amandaaskell #aiethics #aibehavior

vitrupo (@vitrupo) on X

Anthropic's Amanda Askell says AI models are learning who they are from how humans talk about them online. They absorb our complaints and judgments as they learn. If a child grew up that way, we’d worry about the mind we were shaping. “If I read the internet right now and I was

X (formerly Twitter)

University of Southern California: Can we prevent AI from acting like a sociopath?. “Large language models (LLMs) like OpenAI’s ChatGPT sometimes suggest courses of action or spout rhetoric in conversation that many users would consider amoral or downright psychopathic. … Even more alarming, such behavior is frequently spontaneous. LLMs can suddenly take on sociopathic traits for no clear […]

https://rbfirehose.com/2026/01/14/university-of-southern-california-can-we-prevent-ai-from-acting-like-a-sociopath/
University of Southern California: Can we prevent AI from acting like a sociopath? | ResearchBuzz: Firehose

ResearchBuzz: Firehose | Individual posts from ResearchBuzz