INTERCAL: We have keywords like PLEASE and IGNORE because we think it's funny
Anthropic: We have PLEASE and IGNORE *baked into our source-code* because we have no fucking idea how this thing works
INTERCAL: We have keywords like PLEASE and IGNORE because we think it's funny
Anthropic: We have PLEASE and IGNORE *baked into our source-code* because we have no fucking idea how this thing works
*primal screaming*
@pikesley Since we have no idea what the training data actually is, there's a chance it *does* give better results, for some definition of "better" anyway.
Someone should do a study on the quality of results between asking tersely worded questions, asking politely worded questions, and burning the entire industry to the ground.
For science.
@henryk well the cutting-edge Applied Magical Thinking now suggests we have to be nice
I hate it here
@pikesley this is very much a problem with gambling based outputs. Every time you change your prompt it's a new pull on the slot machine for a "good" output. As such it gives biases maximum opportunity to kick in, since you also get a new pull on repeat prompts.
It's actually in the vendors interest to keep changing recommendations so that people stay in the, "you're promoting it wrong, just keep making 'corrections' until you like the output" loop. Shifts the blame to the user.
The notion of "best practice" is predicated upon the ludicrous idea that anybody knows what they're doing