@tante as I noted, “skills are just macros”, but I guess they’re a little cleverer than that.
They work-around limited context space by being part of your prompt that only gets loaded if the skill-triggering metadata indicates the “skill” may be relevant to the problem at hand.
@tante I know very little about LLMs and even less about product architecture at these companies but that has not stopped me from making devastatingly accurate predictions about a bunch of things, so here's my guess:
the major challenge of current models is context rot, which is obvious and catastrophic when applying them to any real-world problems. "skills" are one of a set of tools to play a little shell game to shorten the prompt by hiding the prompt until heuristics determine it's needed
we have to burn that shit to the ground before it eats the world
"So, imagine an intelligent computer that does whatever you tell it to!"
"An *intelligent computer!* Wow. So I can just talk into this microphone and-"
"Well, no. You've gotta type it out."
"Oh, okay. So I just type it out here and it'll do whatever I want?"
"Well...no. I mean, if you ask it *correctly*"
"So...I've gotta be really clear about what I want? Like a genie!"
"Actually you have to write extensive code then input that or it won't do anything useful."
So... A compiler?
Thing is, they don't "disregard instructions" at all. "Instructions" aren't part of their universe. It's completely orthogonal.
"Instructions" are just more tokens of context to complete. It is purely 100% a game of linguistics.
Prompts, skills, system messages, are all context, the pattern to be extended.
What trips us up is that it *sounds* like it is listening, processing, and doing things with concepts.
But only insofar as those concepts are encoded into next tokens.
@pseudonym @tante I know in the lower level it works like that, but models are still trained to react to instruction-like sequences so that it would be meaningful for humans to think in terms of giving instructions and getting the model to follow them.
It's a leaky abstraction, even more so that some other abstractions. But still one that in my view exists and what people build on. Or at least attempt to.