time 2 implant this brain into an openclaw and give it full access to my email
mostly because i don't want to retain any of my email
so i installed it into the openclaw meme thing. and it's not like, doing the stuff it claims it is doing.
like it is hallucinating things like "i updated SOUL.md with xyz"
i seriously do not think this stuff is real now
and you tell me people legitimately are using this software.
how?
is it really magically better when you hook up claude?
ok, i incorporated the feedback of some of the ML researchers who follow me, and dropped the openclaw-as-IRC-bot idea. it just isn't feasible.
instead, i've written a very simple vector database in Elixir, and a very simple IRC client in Elixir.
it can remember things about people in the vector database, those factoids are spliced into the system prompt.
the last 10 messages are also spliced into the system prompt
and then the new message is the user-supplied prompt.
no sliding context window.
The key is to realise that the average is so low β we can't all be experts at everything, so we are bad at most things β that a model performing slightly above average at one of the tasks we aren't good at means a majority of users will perceive its outcomes as positively better than what they could do themselves.
To any expert, the model falls very short, as it performs well below its own ability.
@linear oh this isn't a serious thing, I just wanted to connect an LLM to IRC trained on all of my (anonymized and sanitized) IRC logs, as a friend is going through a midlife crisis and is dealing with it by playing with IRC stuff. The goal in using openclaw was that perhaps it could maintain a better narrative.
I suspect I will solve this goal by just writing a shitty IRC bot in Python that bridges the two worlds together with a decent enough system prompt for it to "understand" (to the extent that it can understand anyway) what the input is.
@ariadne You need climate destroying approach to get a model that can pattern match sufficiently well for people with no self awarenessβa surprisingly huge percentage of populationβto mistake it for intelligence.
Models still collapse then, but collapse is esoteric enough to be framed as βbad prompt engineeringβ.
@jfkimmes i built an LLM from scratch with transformers kinda loosely following the scripts the qwen people released
the LLM is basically trained on ~30ish GB of mostly furry smut and public Linux IRC logs.
*nods sagely*
@jfkimmes this does explain something: it seems to be able to invoke tools when it is planning, but then those tools do not get invoked in the final step.
so it uses tools to read files when planning, then fails to use tools when executing.
what a fascinating conundrum.