Claude Code deletes developers' production setup, including its database and snapshots — 2.5 years of records were nuked in an instant

https://lemmy.nz/post/35120739

Claude Code deletes developers' production setup, including its database and snapshots — 2.5 years of records were nuked in an instant - Lemmy NZ

Lemmy

This is like blaming the gun for killing people.
Uhhh not really. Guns don’t just go off by themselves.
I mean they do sometimes without the proper safety protocols in place, but you still blame the user in the end.
They absolutely do not.

I mean, there’s a good reason the first rules of firearm safety are to always treat a weapon as loaded, and to never direct the weapon at something you aren’t prepared to destroy. The key point being that you never know when some freak accident can happen with a loose pin, bad ammo, a broken spring, or just a person tripping and shaking the gun a bit too hard.

A gun should never go off by itself. You still treat it as if it can, because in the real world freak accidents happen.

Sure. The point is it’s entirely possible to use a firearm safely. There is no safe use for LLMs because they “make decisions”, for lack of a better phrase, for themselves, without any user input.
That is not at all how LLMs work. It’s the software written around LLMs that aide it in constructing and running commands and “making decisions”. That same software can also prompt the user to confirm if they should do something or sandbox the actions in some way.
It can, but we’ve already seen many times that it does not.

Only if the user has configured it to bypass those authorizations.

With an agentic coding assistant, the LLM does not decide when it does and doesn’t prompt for authorization to proceed. The surrounding software is the one that makes that call, which is a normal program with hard guardrails in place. The only way to bypass the authorization prompts is to configure that software to bypass them. Many do allow that option, but of course you should only do so when operating in a sandbox.

The person in this article was a moron, that’s all there is to it. They ran the LLM on their live system, with no sandbox, went out of their way to remove all guardrails, and had no backup.

As I said elsewhere, if you’re denying access to your agentic AI, what is the point of it? It needs access to complete agentic tasks.

The person in this article was a moron, that’s all there is to it. They ran the LLM

No disagreement there.

if you’re denying access to your agentic AI, what is the point of it? It needs access to complete agentic tasks.

Yes, which it can prompt you for. Three options:

  • Deny everything

  • Prompt for approval when it needs to run a command or write a file

  • Allow everything

  • Obviously optional 1 is useless, but there’s nothing wrong with choosing option 2, or even option 3 if you run it in a sandbox where it can’t do any real-world damage.

  • Prompt for approval when it needs to run a command or write a file
  • And therein lies the problem. You’re giving the LLM control over when to or not to ask for approval.

    You clearly have absolutely zero experience here. When you’re prompted for access, it tells you the exact command that’s going to be run. You don’t just give blind approval to “run something”, you’re shown the exact command it’s going to run and you can choose to approve or reject it.
    Unless you’re managing app permissions on android 🙄