the first ai agent worm
the first ai agent worm
What? They’re just computer programs. Almost all computers have high quality entropy sources that can generate truly random numbers. LLMs’ whole thing is basically turning sequences of random numbers into sequences of less random stuff that makes sense. They have a built-in dial for nondeterminism, and it’s almost never at zero.
I feel like I’m missing your meaning because the literal interpretation is nonsense.
Yes and no. The models themselves are just a big pile of floating point numbers that represent a compression of the dataset they were trained on. The patterns in that dataset will absolutely dominate the output of the model even if you tweak the inference parameters. Try it. Ask it ten times to make list of 20-30 random words. Each time a new context. The alignment between each of those lists will be uncanny. Hell, you’ll even see repeats within the list. Size of the model matters here with the small ones (especially quantized ones) having less patterns or bigger semantic gravity wells. But even the big boys will give you the same slop patterns that are mostly fixed. Unless you are specifically introducing more entropy into the prompt, you can mostly treat a fixed prompt as a function with a somewhat deterministic output (within a given bounds).
This means that the claims in the OP are simply not true. At least not without some caveats and specific work arounds to make it true
At least not without some caveats and specific work arounds to make it true
Luckily hackers are terrible at doing that, otherwise we might be in trouble.
Ask it ten times to make list of 20-30 random words
This is true on ootb models but not the universal rule. You could adjust the temperature all the way up and get something way more random, probably to the point of incoherence.
The trick is balancing that with keeping the model doing something useful. If you’re clever you could leverage /dev/random as a tool to manually inject randomness while keeping the result deterministic.