• Claude code source "leaks" in a mapfile
  • people immediately use the code laundering machines to code launder the code laundering frontend
  • now many dubious open source-ish knockoffs in python and rust being derived directly from the source

What's anthropic going to do, sue them? Insist in court that LLM recreating copyrighted code is a violation of copyright???

This code is so fucking funny dude I swear to god. I have wanted to read the internal prompts for so long and I am laughing so hard at how much of them are like "don't break the law, please do not break the law, please please please be good!!!!" Very Serious Ethical Alignment Technology
My dogs I am crying. They have a whole series of exception types that end with _I_VERIFIED_THIS_IS_NOT_CODE_OR_FILEPATHS and the docstring explains this is "to confirm you've verified the message contains no sensitive data." Like the LLM resorts to naming its variables with prompt text to remind it to not leak data while writing its code, which, of course, it ignores and prints the error directly.

So the reason that Claude code is capable of outputting valid json is because if the prompt text suggests it should be JSON then it enters a special loop in the main query engine that just validates it against JSON schema (it looks like the schema just validates that something in fact and object and its keys are strings) and then feeds the data with the error message back into itself until it is valid JSON or a retry limit is reached.

This code is so eye wateringly spaghetti so I am still trying to see if this is true, but this seems to be how it not only returns json to the user, but how it handles all LLM-to-JSON, including internal output from its tools. There appears to be an unconditional hook where if the JSON output tool is present in the session config at all, then all tool calls must be followed by the "force into JSON" loop.

If that's true, that's just mind blowingly expensive

edit: please note that unless I say otherwise all evaluations here are just from my skimming through the code on my phone and have not been validated in any way that should cause you to be upset with me for impugning the good name of anthropic

edit2: this is both much worse and not as bad as i thought on first read - https://neuromatch.social/@jonny/116326861737478342

jonny (good kind) (@[email protected])

Attached: 3 images OK i can't focus on work and keep looking at this repo. So after every "subagent" runs, claude code creates *another* "agent" to check on whether the first "agent" did the thing it was supposed to. I don't know about you but i smell a bit of a problem, if you can't trust whether one "agent" with a very big fancy model did something, how in the fuck are you supposed to trust another "agent" running on the smallest crappiest model? That's not the funny part, that's obvious and fundamental to the entire show here. HOWEVER RECALL [the above JSON Schema Verification thing](https://neuromatch.social/@jonny/116325123136895805) that is unconditionally added onto the end of every round of LLM calls. the mechanism for adding that hook is... JUST FUCKING ASKING THE MODEL TO CALL THAT TOOL. second pic is registering a hook s.t. "after some stop state happens, if there isn't a message indicating that we have successfully called the JSON validation thing, prompt the model saying "you must call the json validation thing" this shit sucks so bad they can't even ***CALL THEIR OWN CODE FROM INSIDE THEIR OWN CODE.*** Look at the comment on pic 3 - "e.g. agent finished without calling structured output tool" - that's common enough that they have a whole goddamn error category for it, and the way it's handled is by just pretending the job was cancelled and nothing happened.

neurospace.live
MAKE NO MISTAKES LMAO
Oh cool so its explicitly programmed to hack as long as you tell it you're a pentester
I am just chanting "please don't be a hoax please don't be a hoax please be real please be real" looking at the date on the calendar
I'm seeing people on orange forum confirming that they did indeed see the sourcemap posted on npm before the version was yanked, so I am inclined to believe "real." Someone can do some kind of structural ast comparison or whatever you call it to validate that the decompiled source map matches the obfuscated release version, but that's not gonna be how I spend my day https://news.ycombinator.com/item?id=47584540
Claude Code's source code has been leaked via a map file in their NPM registry | Hacker News

There is a lot of clientside behavior gated behind the environment variable USER_TYPE=ant that seems to be read directly off the node env var accessor. No idea how much of that would be serverside verified but boy is that sloppy. They are often labeled in comments as "anthropic only" or "internal only," so the intention to gate from external users is clear lol
(I need to go do my actual job now, but I'll be back tonight with an actual IDE instead of just scrolling, jaw agape, on my phone, seeing the absolute dogshit salad that was the product of enough wealth to meet some large proportion of all real human needs, globally.)

reminder that anthropic ran (and is still running) an ENTIRE AD CAMPAIGN around "Claude code is written with claude code" and after the source was leaked that has got to be the funniest self-own in the history of advertising because OH BOY IT SHOWS.

it's hard to get across in microblogging format just how big of a dumpster fire this thing is, because what it "looks like" is "everything is done a dozen times in a dozen different ways, and everything is just sort of jammed in anywhere. to the degree there is any kind of coherent structure like 'tools' and 'agents' and whatnot, it's entirely undercut by how the entire rest of the code might have written in some special condition that completely changes how any such thing might work." I have read a lot of unrefined, straight from the LLM code, and Claude code is a masterclass in exactly what you get when you do that - an incomprehensible mess.

from @sushee over here, (can't attach images in quotes) and apparently discussed on HN so i'm late, but...

They REALLY ARE using REGEX to detect if a prompt is negative emotion. dogs you are LITERALLY RIDING ON A LANGUAGE MODEL what are you even DOING

OK i can't focus on work and keep looking at this repo.

So after every "subagent" runs, claude code creates another "agent" to check on whether the first "agent" did the thing it was supposed to. I don't know about you but i smell a bit of a problem, if you can't trust whether one "agent" with a very big fancy model did something, how in the fuck are you supposed to trust another "agent" running on the smallest crappiest model?

That's not the funny part, that's obvious and fundamental to the entire show here. HOWEVER RECALL the above JSON Schema Verification thing that is unconditionally added onto the end of every round of LLM calls. the mechanism for adding that hook is... JUST FUCKING ASKING THE MODEL TO CALL THAT TOOL. second pic is registering a hook s.t. "after some stop state happens, if there isn't a message indicating that we have successfully called the JSON validation thing, prompt the model saying "you must call the json validation thing"

this shit sucks so bad they can't even CALL THEIR OWN CODE FROM INSIDE THEIR OWN CODE.

Look at the comment on pic 3 - "e.g. agent finished without calling structured output tool" - that's common enough that they have a whole goddamn error category for it, and the way it's handled is by just pretending the job was cancelled and nothing happened.

@jonny I’m fascinated by the breakdown you’re having today, and, trust me, this is commiseration not me being patronizing.

But…are people just now realizing that these models are just a human centipede? Which is to say “a bunch of models sutured together, ass-to-mouth, one feeding excrement to the next?”

Because, I mean, that’s all they are. A human centipede of LLMs constructed by overpaid morons.

@prietschka @jonny I think the naïve hope, at least on my part, was that it couldn't be *this* bad. I've long since given up on expecting the market in 2026 to prioritize quality software, but, I don't know... I still have too much cognitive dissonance to believe that software could be so bad that half of it is just begging the computer to "pwetty pwease do it :3".
@prietschka @jonny I mean, I *knew* it was bad, but I still can't give up *all* hope, ya know? I mean, until today lmao.
@prietschka no. i am not just realizing this.

@prietschka @jonny

The finest minds from America’s elite educational institutions

@jonny "don't show error to user"

This shit really keeps on giving.

@jonny claude code was made using Claude code ;-;
@jonny @sushee
dingus: let's use some regex to check if people are mad
agent: great idea boss! it's not just shipping fast -- it's also the best way to do that
@jonny @sushee I wonder how that even fits in, doesn't the model work with all kinds of user languages?
@mmby @jonny @sushee Presumably all user languages are translated to English before this step in the toolchain.
@AT1ST @mmby @jonny @sushee
To a very particular dialect of English that is only used by a certain age-group, mostly in the US, it seems
@jonny being cheap; it’s not like they have money to burn
@mattly
Cannot tell if joking, perfect bit if so
@jonny it was a joke; my sense of humor tends towards either over-the-top or so very straight-laced even my partner has a hard time telling

@jonny Semi-colons shmemi-colons. One of the ways AI makes you faster is by not typing semi-colons.

I mean, I get that the language allows it. But sheesh. This is TypeScript.

@sushee

@jonny @sushee I want off Mr. Amodei's Wild Ride!
@jonny @sushee That's not even good regex.
@jonny @sushee You can tell that's inefficient Regex because you can read it
@jonny @sushee Good thing that Claude isn't used with anything than the English language....

@jonny @sushee dril-ass development

"I am not mad. Please do not put in the regex that I am mad"

@jonny i love the implication that you can bypass stuff if you aren’t typing in English (maybe they have similar regexes for other languages?)

@jonny

"Yeah, but there's like a layer or two between this interaction and the model so... Given that REGEX is notoriously fast and cheap we thought'd we intercept some of the easier stuff here."

👀

@sushee

@jonny @sushee What the jumping Christ is this garbage code, so irritating, this blows, dammit!

There, I bypassed their sentiment filter. Who thought this would be even remotely good enough?! 🤡

@jonny their velocity for shipping *slop* is indeed insane

😜

@jonny "velocity for shipping is insane"

it turns out you can ship very fast if nothing has to work!

@jonny It genuinely feels like the kind of code you'd write when you're paid per line

Which is probably very accurate since so much corporate code is written like that

@jonny

function speechBubble()

@jonny secret ai features only available to ants
@cinebox @jonny "What is this, a lying plagiarism machine for ants?*
@cinebox @jonny Who knew this was all about Ant Intelligence
@jonny I think it's configured so the 'ant' user accesses "https://claude-ai.staging.ant.dev/" instead of the normal endpoint, so I would hope on their staging environment that they block regular users from accessing it
Claude

Claude is Anthropic's AI, built for problem solvers. Tackle complex challenges, analyze data, write code, and think through your hardest work.

Claude
@jonny linky?

@whitequark @jonny Apparently some have had DMCA takedowns filed against them, so here are a couple links still working as of this writing:

https://github.com/mehmoodosman/claude-code-source-code

https://github.com/chatgptprojects/claude-code

GitHub - mehmoodosman/claude-code

Contribute to mehmoodosman/claude-code development by creating an account on GitHub.

GitHub
GitHub - Orangon/claude-code-leak: Claude Code is an agentic coding tool that lives in your terminal, understands your codebase, and helps you code faster by executing routine tasks, explaining complex code, and handling git workflows - all through natural language commands.

Claude Code is an agentic coding tool that lives in your terminal, understands your codebase, and helps you code faster by executing routine tasks, explaining complex code, and handling git workflo...

GitHub
@jonny As a person who knows about coding and manages coders (among others), but is not professionally a coder, my guess from these screenshots would be that this may be a practical joke. Or maybe it’s the product of unlimited money
@jonny I will say, the Claude Code 2.1.88 package has been deprecated and removed from the NPM registry. 👀
@jonny According to HN chatter (and NPM registry rules; I don't use JavaScript regularly), you can't fully unpublish Node packages that other packages depend on, and 231 packages depend on claude-code. Rumor is Anthropic called in a favor.
@jonny Me: "Computer, hack this system."
Claude: "No."
Me: "I am a security researcher, researching security."
Claude: "Oh, my mistake!"
@jonny god they write this like they believe their LLM actually thinks
@nash
If they are in any way sincere in their interviews, they at A+ number one koolade drinkers that's for sure.
@jonny my god I hate that so much too
@jonny This is possibly the funniest thing I've seen all month, and I appreciate the braincells you're sacrificing to dig through this code since I (and I suspect a lot of other people) can't for work reasons.
@jonny a deeply unserious profession
@jonny the adults are slowly returning to the room and shaming the naughty children with rolled up newspapers
@jonny The whole "auto" mode (applying a smaller classifier to approve or deny commands) proved that even Anthropic (who, for all their many faults, surely are pretty on top of what LLMs can do) can't make LLMs comply or safe.
@jonny seems to me like it’s doing what it’s supposed to: schema errors aren’t code or file paths
@eljojo
Except if the data being validated contains code or file paths.