"You should really try Claude/WhateverLLM before criticizing"
is the new
"But it contains electrolytes"
"You should really try Claude/WhateverLLM before criticizing"
is the new
"But it contains electrolytes"
Better question:
How many neurons does it take to be "slop"?
I introduce a 3 neuron example. A PID loop, from control systems theory.
It has a training phase in which it 'learns' the control from known inputs. And it has a execution phase, of which it applies the learned inputs.
Even my 3d printers use PID for the nozzle and bed. My oven in the kitchen does so as well.
Is all learning software "evil"? If no, where's the cutoff?
Thats kind of the point.
Theres intermediate learning software like K-Nearest-Neighbors that also are trained on classified (properly annotated) data, and then can provide percentage responses on trained data. We see this with tools like Merlin birdsong identification.
I even made a 10 position classifier with the MYO myoelectrical armband back in 2016. No GPU needed. Modest CPU and ram was needed, something that easily a RPi 2 could do.
Point being is this whole debate is being forced into a binary with folks saying "this is amazing" to "horrific garbage". Maybe LLMs could be made more useful if they output % confidence and citations accurately?
But again, I'm not going to dismiss, nor am I going to trust everything. Both actions are foolish.
@crankylinuxuser @ploum you make this discussion a technical one, where it is political. A tool does not exists in a vacuum, it serves interests of some group, controling it means power etc...
The "slop" debate maybe would have not happened if big tech companies had not massively stolen content to train LLM. If they had not then sold them as "tools to help creativity" to the very same people it stole from.
It is always the same story :
- attempt at domination by big companies (yay capitalism)
- attempts to resist domination by some folks
Well, it is both technological AND political.
Copyright is also highly political, and has changed many times from its inception. And how its implemented currently only favors the large entities (eg: disney etm).
I am not opposed in gobbling up as much content and training something with the sum of human knowledge. Now this thing needs to also be free.
And enter US vs Chinese AI. You can run DeepSeek R2 yourself. Same with Qwen. They are fully open source including the current state of the art. I'm even running a 30B Qwen at home on my desktop.
FLOSS is how we avoid the hypercapitalist bullshit, including the inevitable rug-pull by the US AI vendors. Anthropic already did that with "unlimited but not really" but didn't suffer.
But also, moving to "political problem" versus "technical problem" is also just moving the goalposts. The real political problem with all of these things is how structural capitalism further striates worker vs owner. Nedd Ludd discussed this centuries ago, as did Marx.
For the time being, yes, I will run my own AI on my equipment, under my control.
@crankylinuxuser @ploum you were answering to a post that made a political comment comparing the use of "slop" with a movie that is famously a humorous political commentary (and often used as such). I'm not moving any goalpost here.
You are free to use LLM if you like. The term "slop" is not technical : it does not describe the output of all LLMs (and what you said about PID was kinda demonstrating that). It describe a usage that some group find negative. If you use the tool ethically and correctly (whatever this means, we are still figuring it out), what you do with it is not slop (i think, but that is a personnal opinion).
@crankylinuxuser @whiteshoulders @ploum
A SOTA GPU hour (H200) starts at $3.72 and ends at $10.60. Lets make $5 for simplicity. If the model spends like 3 minutes thinking(in a single prompt) you you get 0.25$ of money wasted (imagine no profit).
The people with Claude Code do like 50 prompts a day(being optimistic), and this is $12.5 of money.
Claude Max 20x is $200. If the guy prompts 28 days a month it means 350$ of GPU. -150$ in losses
@Pibert @whiteshoulders @ploum
The math doesn't pan out if you're using your own machines. Then its just cost of electricity.
Its what I'm doing. And it works pretty darn well. I can run thinking models, tool using models, image models. And data stays local.
@crankylinuxuser @whiteshoulders @ploum
1/2
it’s the cost of electricity and hardware depreciation. I picked the lowest and highest price of GPU hour of the H200.
But like an poorly optimized 8 x H200 can generate 1 million tokens in 30 seconds being optimized.(deepseek v3)
If it’s like deepseek v3 costs $0.42 1M tokens. So it’s $50.4 an hour and you wasted $15.92 in GPU hour(Deep infra). Thinking in the API manner you get a ton of money.