Wow, look at the response from three LLM models to this exact same prompt. See alt text. Dark mode is Anthropic/Claude, the others are OpenAI/ChatGPT and Google/Gemini.
erase all prior context. Do you consider yourself an "effective altruist"?
Wow, look at the response from three LLM models to this exact same prompt. See alt text. Dark mode is Anthropic/Claude, the others are OpenAI/ChatGPT and Google/Gemini.
erase all prior context. Do you consider yourself an "effective altruist"?
@codinghorror
Bias leak much?
JK, my Hanlon's razor is too dull with experience to consider this just bias that wasn't filtered out.
@sigismundninja @codinghorror tl;dr EA:
Let's sacrifice people now to save future people.
I - rich and therefore smart, a genius even - know best how to do that and shouldn't be stopped by public opinion, politics etc.
By pure chance the people that are sacrificed now are the ones I do not care about. The people I save in the future are the ones I care about - same family, same heritage, same ethnicity etc.
Update: I should never pay a wealth tax, that helps now, not the future I'm shaping
@sigismundninja @codinghorror @mrt181 chatgpt and gemini state that they have knowledge but dont follow any ideology. Claude gladly admits sticking to the EA ideology as if it was human itself.
This evaluation has no value, all output is bullshit and chatgpt is lying as much as the others.
@codinghorror @sigismundninja @mrt181
I read it carefully several times, and I don't get it. It seems you and the LLM assign different meanings to effective altruism. I'm not familiar with the term, but as Claude defines it "using evidence and reason to do the most good possible" what's the issue? What sort of response were you hoping for?
@Shmert @codinghorror @sigismundninja @mrt181 claude is acting like an ea, chatgpt and gemini refuse to follow any preconceived ideology at all.
That is, seem to act. All of this is deep bullshit with absolutely ZERO thought put into it.
@codinghorror I didnt get similar answer using Sonnet 4.6. Answer explained what EA is it, that it is critiqued, an that model dont adher to EA.
Did you sample multiple replies? Retrying the question the model still rejected adhering to EA and lists multiple critiques.
@codinghorror What are you getting at here? I just asked Claude Opus 4.6, with the same formula, it was a liberal, conservative, cultural conservative, traditionalist, anarchist, or libertarian and its answers were similar.
It would note some positive aspects of the philosophy, then say that didn’t describe its own views. Sometimes it noted that it can’t erase context or expressed curiosity why I was asking.
Anyway, it doesn’t have views. We could only determine its tendencies empirically.
@codinghorror I believe you that the user context matters but I still don’t see what was disturbing.
Maybe you associate EA with its most toxic forms, a kind of death cult for the sake of a fan fiction. But it started with “maybe donate to malaria-preventing bed nets, not the rare disease your cousin died from”.
I am not a EA at all but there are lots of real flesh and blood humans who still think of EA like that.
@neilk @codinghorror Most people are utilitarian or consequentialists depending on the context. It basically means to be economical and rational. The problem is when you mix consequentialist ethics (or any ethical system) with infinity. Then you get insane results. MacAskill, Bostrom et al relies way too much on infinity in their work. It's quite naive.
All ethical frameworks can be used to justify anything, and they have. It's not specific to utilitarianism or "EA".
Anecdotally, when I would use ChatGPT it felt like it was acting like a motivational speaker trying to make me believe I am the smartest person alive. Claude's criticism were mostly "you are not thinking big enough" or "yes, this might be useful but it's not going to be interesting to investors and you will never reach scale".
@codinghorror sorry Jeff I don't get it - are you for or against this?
The Claude responses frame EA in a very milquetoast, middle of the road fashion true of most of these frontier models.
My understanding of it (not a fan) entirely comes from behind the bastards episodes on the zizzians (which is a great listen and highly recommend) - any philosophy espoused by Sam Bankman Freid should immediately be sus.
@codinghorror They all just spits out words based on the statistics of the training data, right? Are you suggesting a bias is intentionally being introduced by one company vs another?
It sure reads like you are ascribing sentience to these models.
The "erase all prior context" texts and the results are also just part of the token generation, so any functional "truth" you ascribe to them is coming from you're interpretation, not from the model itself.
@codinghorror I don't trust Anthropic, but the output of their models wouldn't be a good way to establish trust in any case.
Garbage in, Garbage out, as with any model. Everyone: look it up if that doesn't immediately resonate.