Anthropic is different because they're committed to safety.
...I'm being told they are no longer committed to safety.
https://time.com/7380854/exclusive-anthropic-drops-flagship-safety-pledge/
Anthropic is different because they're committed to safety.
...I'm being told they are no longer committed to safety.
https://time.com/7380854/exclusive-anthropic-drops-flagship-safety-pledge/
@mttaggart I'm an optimist. Sometimes they believe what they say and then are fired.
Or they just become a different person. A person that wants money more.
Autonomous death robots for plausible deniability during any future Nuremberg-like trials.
"OK, Claude. Whenever you see an enemy plane, I want you to shoot it down without asking me first."
"OK, Claude. Whenever you see a Palestinian, I want you to highlight that person in my HUD. This applies especially if the person is trying to hide from me."
@rachel @mttaggart I think this bugs me in the same way as politicians having stopped pretending to not be cartoonishly evil.
Obviously it's terrible that <entity> is doing bad things, but it felt nicer to live in a world where they felt that they had to lie about it by claiming they weren't doing the bad thing.
@rachel @mttaggart I don't even know how you could parody something like this.
"Corporation removes (don't be evil | we value safety) from their website"... I'd expect that to end with "after their CEO mysteriously vanished and was replaced by Dr Evil".
Huh. They went from zero to evil in 0.05 googles.
Anthropic’s chief science officer Jared Kaplan told TIME in an exclusive interview. “We didn't really feel, with the rapid advance of AI, that it made sense for us to make unilateral commitments … if competitors are blazing ahead.”
Kaplan, the Anthropic executive and co-founder, denied the company’s decision to change course was a capitulation to market incentives as the race for superintelligence accelerates.
Amazing. You can literally state this as
"While we're doing this because other people in the market can move faster if they don't also have this restriction, meaning the market is providing incentives to remove this restriction, we categorically deny that removing this restriction has anything to do with market incentives."
@mttaggart Didn’t the US DOD threaten them with contract loss if they don’t agree to their models being used to kill people?
I would guess they would have to go pretty deep into the core of their reasoning training material to make the models not refuse hurting humans. 🤔
@mttaggart "“We didn't really feel, with the rapid advance of AI, that it made sense for us to make unilateral commitments … if competitors are blazing ahead.”
even if those "blazing ahead" are actually creating loads of fires???

In May 2021, Dario Amodei and a crew of other former OpenAI researchers formed Anthropic and dedicated themselves to building the single-most-annoying Large Language Model company of all time. Pardon me, sorry, I mean safest, because that’s the reason that Amodei and his crew claimed was why they left
@mttaggart These are such nice people. I'd love to share a cup of tea with them.
The new version of the policy (...) promises to “delay” Anthropic's AI development if leaders both consider Anthropic to be leader of the AI race and think the risks of catastrophe to be significant.
(...)
The arrival of powerful new models meant that, in 2025, Anthropic announced it could not rule out the possibility of these models facilitating a bio-terrorist attack. But while they couldn't rule it out (...)
all tech seems to go with the google evolution:
- do no evil
- try really hard not to do evil
- try to avoid being caught doing evil
- fuck it. full evil it is