An "AI" company saying "we have this super magic new model that is so powerful that we cannot release it" is just marketing. It's the old OpenAI scam.
Anthropic needs to keep the hype going to ride it to the IPO. That's all this is.
@tante
Smoke and mirrors. Its perfectly fine to fake the demo...
@tante One of the lead stories on BBC 6Music news this morning FFS

@tante

"The annihilation of Iran seems to have been averted for now. Also, here's a press-release from an AI company presented as Actual News"

@tante That's, honestly, only partially true. It might, indeed, by even the primary driver.

(Plus pre-empting competitors, marketing, as well as explaining the delays with scaling the model to release capacity without coming out and saying "the US' deranged president's fascist wars impede our business operations.")

But what we see in Open Source wrt security reports _does_ suggest that the newer models have real security implications that need mitigating and somewhat coordinated release.

@tante Sure, Anthropic is hyping. But it does contain a kernel of truth.

@larsmb Sure. Vulnerabilities can be detected through using pattern recognition. Question is: How good are these systems actually?

Like how high is the chance to find a relevent hidden security issue in a larger code base for 1000 USD, 10K USD, 1 M USD? How many false positives that you need to check are generated? Right now it's unclear how the economics work.
It's Anthropic
a) claiming that you can pay them to generate attacks (monetize the attackers)
b) claiming that by paying them you can defend yourself and your code base (I mean if you care, many are already just having claude commit to `main` so security doesn't matter anymore.

@tante Part of the red teaming limited release official goals at least seems to be to apply it to the more critical projects ahead of release to identify the high risk vulnerabilities before the malicious actors can easily replicate said feat.

For security vulns, the cost asymmetry - generating a plausible exploit vs validating it - *does* work in favor of the attacker, so this is possibly the *one* claim that has some merit in the staged release.

@tante People *still* fall for this? jfc.
@peter_sc @tante Have you met people? We aren't very good at this, evolutionarily our brains barely understand edible berry colors and lions = bad.
@tante historically speaking, letting the public test the new models causes the magic to evaporate immediately

@tante

It seems that Sam Altman has had the “revolutionary idea” of a four-day week to allow workers to benefit from AI™.

It's amazing that the labour movement hasn't thought of this!

No doubt he will go on to explain how he'll get corporations to do this instead of just firing people.
</sarcasm>

Caution: link is to marketing version: be prepared to lol.

https://opentools.ai/news/sam-altman-champions-four-day-workweek-a-revolution-in-work-culture

@blogdiva

Sam Altman Champions Four-Day Workweek: A Revolution in Work Culture!

Sam Altman, CEO of OpenAI, is making waves with his proposal for a four-day workweek (32 hours) to spread the benefits of AI productivity to workers worldwide. This revolutionary idea aims to redefine work life, balancing AI advancements with workforce welfare. The proposal encourages companies and governments to incentivize and experiment with this new model, harnessing AI's potential to transform economies.

OpenTools
@tante also: probability it will come out that claude is double digit percentage mechanical turk for harder queries?...
@tante I am skeptical as well, but it came with CVE’s in code that was old and proven so… not all of it is marketing?