I've been thinking about the reports that the latest Claude model can resort to blackmail and other bad behaviors in certain circumstances https://www.bbc.com/news/articles/cpqeng9d20go
You know - if we stopped training LLMs on dystopian training materials, we'd probably get less dystopian behavior. But the model builders' insatiable quest for additional training tokens makes that unlikely to happen any time soon
So, yeah, let's definitely build a bunch of agents that let these LLMs impact actual-world stuff