Passing thought on the Claude code leak, and how messy and wasteful the code apparently is (as many of us suspected):

There’s been a lot of focus on the energy & environmental costs of running LLMs. There are also energy & environmental costs to deploying LLM-generated code.

Similarly, there’s a lot of focus on the kickbacks cloud vendors who fund AI get from LLMs renting their servers. There may be a similar kickback from computationally inefficient LLM-generated software renting their servers.

1/

There’s a lot of software out there where either (1) the users are captive users, or (2) actual outcomes don’t matter, and the important thing is to check the box, to have officially pretended to build the thing.

That’s the sort of software where development costs are especially painful for the MBAs, and where pushing the frontiers of the “fast build, low quality” quadrant for may be a killer market — even if it’s just fast and not so cheap.

2/

Highly optimized code is of course extremely expensive to develop — both to build and maintain. But now even just-average-performance code is the more costly alternative.

And sure, grinding out a vibe-coded LLM horror so you can check that business box may leave you with code that’s not only especially buggy but RAM-hungry and CPU-hungry. And sure that’s expensive to deploy. But hey: your cloud spend is already preposterous, right? And high deployment costs are more predictable than high development costs…right?

3/

I can easily imagine scenarios where huge swaths of guts-of-the-business software increase their resource needs in production by an order of magnitude, •and• where businesses are perfectly happy to pay that cost.

That’s a story with horrifying environmental costs.

4/

A whole lot of our present moment boils down to resource cost externalities — externalities that in many society has intentionally created.

We made carbon emission and water use way too cheap, and we pay dearly for it every day.

/end

@inthehands

Yep. Also factor in who gets a new job after claiming victory.

@inthehands
question, as a tech literate non-tech-person: what characteristics does Claude code have that makes it more energy intensive, and why do you think it has those characteristics?
forgive me if this is an annoying question, I’m trying to keep up with AI development, it’s really an uphill battle haha

@cerulean_corvid @inthehands for one example, borrowed from a thread on here, to produce json, it puts the LLM into a loop to try to produce some json, each time through it validates the output, and back into the loop if it's not valid.

If the LLM produces invalid json multiple times then one operation becomes many operations under the hood, all of which are draining your account.

And it has to do this because the LLM itself has no idea what JSON is, it's just producing sonething that probabilistically could be json.

@inthehands There’s also a financial benefit for Claude (and other agents) to take the long way around to solve a problem because they use more tokens.
@inthehands And because they’re non-deterministic, even when you provide the same context, they will often produce different solutions, so they’re always reinventing the wheel.
@inthehands ah yes, the good old economic perverse incentive. This is endemic across the use of AI.

@inthehands

Lately; I had to ask if a group proposing a machine learning system for a particular research task was actually going to do that, which might save ongoing computing costs; or just stick a chatbot interface and layers of text generator between the users and existing code, which would steeply increase costs and add new failure modes.

So. Yeah.

@inthehands

Most of my professional and recreational computing work has been focused on problems that are CPU or memory bound. (These are problems where the solution, or quality of the solution, is limited by your computer's power, and improvement is possible only with better code and careful resource management.)

I'm currently tackling problems where "AI" is often applied, but the computers are tiny and far from home.

Perhaps this is why the "make money by wasting resources" angle never occurred to me.

Now I want to throw up.

Wasting resources for one sided gain is almost a textbook definition of "war".