@codinghorror Sure, but we're not talking about "which tool is best for driving a nail that I own into a wall that I own," we're talking about "is it ethical to use a technology built on fascist ideology and stolen work, that carries unconscionable environmental costs, and that's used to disrupt labor movements to perform a task that that technology is fundamentally unsuited to?"
It's quite fair to have a very firm "no" by way of answer to the second question.
@codinghorror Anyway, this isn't the first time you've replied to me to make the argument that LLMs are just another kind of tool. I suspect we won't see eye-to-eye on that, especially as my work has been abused to make LLM products.
I hope we can agree though, that my objection *even though you disagree with it* is principled and neither knee jerk nor purity culture.
@codinghorror @eschaton Hey, don't put words in my mouth, I'm not part of that "y'all." I do not agree that doing propaganda work for some of the worst people on the planet, whether intentionally or not, counts as "measured."
But that's what you're doing right now by arguing in favor of LLMs.
@codinghorror @eschaton Given how messy this exchange has gotten, let me pull back slightly. I made a claim, that opposition to LLMs is not an example of "purity culture."
You, despite my explicit ask to not, came into my replies to make a separate but related claim: namely, that LLMs are sometimes useful, and implicitly that that utility is sufficiently great as to justify their ethical problems.
@codinghorror @eschaton While I explicitly said I didn't get into the second point, as the Discourse™ has gotten *incredibly* tedious by now, fine. You seem to insist on having that discussion out in my replies anyway.
To that end, I laid out several reasons that I find the claim that LLMs are "just a tool" odious: the euginicist origin, the fascist way they're funded and developed, that they attack and undermine labor, that they impose extreme environmental cost, and that they don't work.
@codinghorror @eschaton But fine, you disagree, I believe, as you've said earlier. I think you are very wrong on that, but I don't think either of us are budging on that right now.
Do you refute or disagree with the other points? Do you believe that there is some degree to which LLMs could, if they worked well enough, justify their usage given those problems?
@codinghorror @eschaton To be clear, I don't think you owe me any answers. I'm just one woman who's been doing this shit for decades, and who knows what the fuck she's talking about, but whatever.
It's that you made the claim *to me*, and have used that claim to justify that opposition to LLMs is pseudoreligous "zealotry." But you haven't addressed any of the substance of the opposition beyond putting forward one anecdote that I can't personally evaluate the veracity of.
@codinghorror @eschaton So far, the justification you've given for the "zealotry" comment has been almost entirely about the *shape* of the claims I made, almost without any reference to the *substance*.
This strikes me as a very strange way to approach other human beings and moral decisions in general.
Is there any strong claim that you would consider to not be "zealotry," or any degree to which a claim could be evidenced such that it would not be "zealotry" to you?
@codinghorror @eschaton Perhaps that's the root of our impasse, then. I fairly firmly believe that if something does that much harm to the environment, to labor movements, and to victims of fascism, it cannot be justified by appeals to its efficacy alone.
I suspect that if we cannot agree on basic moral precepts like "don't help fascists get rich" and "don't be a scab," there's probably not much hope for a favorable resolution.
@xgranade @codinghorror @eschaton After a few in-person conversations with folks who have gone “all-in” on LLMs, I think there is another effect at play. There appears to be a severe dopamine addiction happening. At least one friend of mine shows signs of it. And has said so. Likens it to social media addiction.
It is very hard to have rational arguments with addicts. So don’t worry if reason doesn’t always work. It isn’t your fault.
One friend has also mentioned how they are likely being managed out because of mild dyslexia. They can’t keep up with the sheer volume of text and code generated by the LLM and their colleagues who aren’t as neurodivergent can. It is very interesting to see this happening. They were a FANTASTIC programmer. And now are being put out to pasture very early.
It makes me very sad.