I remember cynically joking last year that all the "advanced" AIs like Claude were just the same kind of black-box LLMs but with a bunch of regexes glued onto them
now that the source has leaked: HAHAHA
I remember cynically joking last year that all the "advanced" AIs like Claude were just the same kind of black-box LLMs but with a bunch of regexes glued onto them
now that the source has leaked: HAHAHA
@foone @VegaHarmonia wait what, not even a proper text embedding model which runs through that???
also why would they be using sentiment analysis, for what, to tell angry devs to get a cup of water?
@foone @VegaHarmonia oh wait nvm it can be implicit negative feedback submissions when they detect the human being angry at the AI
hm
@foone that's...
Rofl
So the reason that Claude code is capable of outputting valid json is because if the prompt text suggests it should be JSON then it enters a special loop in the main query engine that just validates it against JSON schema (it looks like the schema just validates that something in fact and object and its keys are strings) and then feeds the data with the error message back into itself until it is valid JSON or a retry limit is reached. This code is so eye wateringly spaghetti so I am still trying to see if this is true, but this seems to be how it not only returns json to the user, but how it handles *all* LLM-to-JSON, including internal output from its tools. There appears to be an unconditional hook where if the JSON output tool is present in the session config at all, then all tool calls must be followed by the "force into JSON" loop. If that's true, that's just *mind blowingly expensive* edit: please note that unless I say otherwise all evaluations here are just from my skimming through the code on my phone and have not been validated in any way that should cause you to be upset with me for impugning the good name of anthropic edit2: this is both much worse and not as bad as i thought on first read - https://neuromatch.social/@jonny/116326861737478342
I stopped scripting bots back in... 2016, so anything done with my dev accounts isn't on me. Lmfao.
ID theft is a magical thing.
So is the statute of limitations, lmfao.
@foone honestly
from a description i got of claude's planning mode (that @iamada gave), i basically went "okay so its a bunch of prompts which shits out a few files in a specific way and then later on pastes them in the context window in a smart way so it can focus on one task at a time"
this is what ive been doing for myself for the last 6 years, its not anything special if an AI does it, and it can have an even worse track record at global recall than i do, lmao
@ShadowJonathan @foone app defined markdown-based (executive) memory cache
but like, for a recursive predictive statistical model
@freya they fucked up an NPM package and leaked all the source (well, the code, the models aren't there)
https://www.theregister.com/2026/03/31/anthropic_claude_code_source_code/
@foone Ha! This reminds me: At work today, I needed to analyze data in an Excel file.
I asked our team’s Excel guru. He took a look, told me “the data is crap, go ask an LLM rather than build a pivot table.”
I did. The LLM spat out a Python script to analyze the data.
So I guess it worked - just not the way I thought it would.
@foone wait I thought everyone knew this? Like these tools are just packages of pre-made prompts and loops to reprocess things until it works with a pretty UI...
I guess the code is interesting as it tells you how it works and stuff more exactly which is definitely fascinating but I hope nobody genuinely thought it was something more advanced than scaffolding on the main chat models
512,000 lines of sophisticated agentic architecture, multi-agent orchestration, custom React renderers, and constrained decoding...
...there's a big regex full of profanity checking if you're angry.
#Anthropic coded that section for my alone when I #vibecode 💀
Good way to save tokens, you gotta admit