AI tool OpenClaw wipes the inbox of Meta's AI Alignment director despite repeated commands to stop — executive had to manually terminate the AI to stop the bot from continuing to erase data
AI tool OpenClaw wipes the inbox of Meta's AI Alignment director despite repeated commands to stop — executive had to manually terminate the AI to stop the bot from continuing to erase data
This was 3 or 4 days ago.
I thought of it after Anthropic virtuously announced they would not create autonomous murder devices for the US government (but basically everything else was on the table). Because I’m pretty sure the US military could have just used an Anthropic OpenClaw to bomb civilians as easily as this Facebook AI Safety expert used OpenClaw to destroy her emails.
What’s your emergency “break glass” policy?
Is it a bottle of whiskey?
Because i want to work on meaningful things that benefit people directly.
Because i want to unterstand the capabilities and limitations of openclaw-like agents. LLMs aren’t going away, better be proactive and learn what the hype is about.
Believe it or not, this is the first time for me being suspected a troll, but I start to see the appeal when people are getting so worked up while being so far off the mark.
Sorry to disappoint that I am still on the loose. Then again prison is probably better than doing one more D-FMEA.
I don’t think there’s anything wrong with running Openclaw. I run it in an isolated server, and it doesn’t have access to my data - if it goes tits up, it deletes unimportant stuff only. If anyone gets access to the credentials in it, and maybe its Google account (I went with the approach of giving it its own Google account, so that it can create docs and calendar events and then add me, rather than getting access to my Google account).
What is way too brave for my taste is giving it access to accounts with your personal data, or the filesystem in your computer. That’s a disaster waiting to happen.
I went with the approach of giving it its own Google account, so that it can create docs and calendar events and then add me, rather than getting access to my Google account.
I wonder though: if Google can link this account to you as its actual owner, I wonder if there’s a risk if the bot does something against the ToS?
I hope you have backups of your Google account…
Last I knew, they switched from Anthropic to chatGPT
Either way, what Im hearing is you can get private access, with some creativity, to anything the US intelligence apparatus knows.
I’m sure LLMs can be useful for automation as long as you know what you’re doing, have tested your prompts rigorously on the specific version of the model and agent you’re using, and have put proper guardrails in place.
Just blindly assuming a LLM is intelligent and will do the right thing is stupid, though. LLMs take text you give them as input and then output some predicted text based on statistical patterns. That’s all. If you feed it a pile of text with a chat history that says your emails were deleted, the text it might predict that statistically should come next is an apology. You can feed that same pile of text to 10 different LLMs, and they might all “apologize” to you.
Recently someone lamented that just asking for an alarm to be set cost them tons of money and didn’t even work right…
It was foolish enough to let LLM go to town on automation, but for open ended scenarios, I at least got the logic even if it was stupidly optimistic.
But implementing an alarm? These people don’t even have rationality to their enthusiasm…
If I remember right, that post wasn’t designed to highlight a practical use-case, but rather to set up a simple task as a “how could I apply this?” type of experimentation. The guy got roasted for it, but I think it’s a very reasonable thing to try because it’s a simple task you can see the direct result of in practice.
The cost problem was highlighted as well, because if such a simple task is a problem, it can’t possibly scale well.
You ask the llm to code you an alarm not to actually be an alarm. It’s not an alarm. It’s a language model.
Maybe I’m too autistic for this shit.