AI’s Memorization Crisis | Large language models don’t “learn”—they copy. And that could change everything for the tech industry.
AI’s Memorization Crisis | Large language models don’t “learn”—they copy. And that could change everything for the tech industry.
but we can reasonably assume that Stable Diffusion can render the image on the right partly because it has stored visual elements from the image on the left.
No, you cannot reasonably assume that. It absolutely did not store the visual elements. What it did, was store some floating point values related to some keywords that the source image had pre-classified. When training, it will increase or decrease those floating point values a small amount when it encounters further images that use those same keywords.
What the examples demonstrate is a lack of diversity in the training set for those very specific keywords. There’s a reason why they chose Stable Diffusion 1.4 and not Stable Diffusion 2.0 (or later versions)… Because they drastically improved the model after that. These sorts of problems (with not-diverse-enough training data) are considered flaws by the very AI researchers creating the models. It’s exactly the type of thing they don’t want to happen!
The article seems to be implying that this is a common problem that happens constantly and that the companies creating these AI models just don’t give a fuck. This is false. It’s flaws like this that leave your model open to attack (and letting competitors figure out your weights; not that it matters with Stable Diffusion since that version is open source), not just copyright lawsuits!
Here’s the part I don’t get: Clearly nobody is distributing copyrighted images by asking AI to do its best to recreate them. When you do this, you end up with severely shitty hack images that nobody wants to look at. Basically, if no one is actually using these images except to say, “aha! My academic research uncovered this tiny flaw in your model that represents an obscure area of AI research!” why TF should anyone care?
They shouldn’t! The only reason why articles like this get any attention at all is because it’s rage bait for AI haters. People who severely hate generative AI will grasp at anything to justify their position. Why? I don’t get it. If you don’t like it, just say you don’t like it! Why do you need to point to absolutely, ridiculously obscure shit like finding a flaw in Stable Diffusion 1.4 (from years ago, before 99% of the world had even heard of generative image AI)?
Generative AI is just the latest way of giving instructions to computers. That’s it! That’s all it is.
Nobody gave a shit about this kind of thing when Star Trek was pretending to do generative AI in the Holodeck. Now that we’ve got he pre-alpha version of that very thing, a lot of extremely vocal haters are freaking TF out.
Do you want the cool shit from Star Trek’s imaginary future or not? This is literally what computer scientists have been dreaming of for decades. It’s here! Have some fun with it!
Generative AI uses up less power/water than streaming YouTube or Netflix (yes, it’s true). So if you’re about to say it’s bad for the environment, I expect you’re just as vocal about streaming video, yeah?
Edit: It’s interesting how this snippet always gets downvoted without explanation. Let’s not be like the crazies. Acknowledge the facts even if you don’t like the technology.
Source for the claim on using less water than YouTube or Netflix (or even walking, for that matter)
Using chatbots emits the same tiny amounts of CO2 as other normal things we do online, and way less than most offline things we do. Even when you include “hidden costs” like training, the emissions from making hardware, energy used in cooling, and AI chips idling between prompts, the carbon cost of an average chatbot prompt adds up to less than 1/150,000th of the average American’s daily emissions. Water is similar. Everything we do uses a lot of water. Most electricity is generated using water, and most of the way AI “uses” water is actually just in generating its electricity. The average American’s daily water footprint is ~800,000 times as much as the full cost of an AI prompt. The actual amount of water used per prompt in data centers themselves is vanishingly small.
Because chatbot prompts use so little energy and water, if you’re sitting and reading the full responses they generate, it’s very likely that you’re using way less energy and water than you otherwise would in your daily life. It takes ~1000 prompts to raise your emissions by 1%. If you sat at your computer all day, sending and reading 1000 prompts in a row, you wouldn’t be doing more energy intensive things like driving, or using physical objects you own that wear out, need to be replaced, and cost emissions and water to make. Every second you spend walking outside wears out your sneakers just a little bit, to the point that they eventually need to be replaced. Sneakers cost water to make. My best guess is that every second of walking uses as much water in expectation as ~7 chatbot prompts. So sitting inside at your computer saves that water too. It seems like it’s near impossible to raise your personal emissions and water footprint at all using chatbots, because using all day on something that ends up causing 1% of your normal emissions is exactly like spending all day on an activity that costs only 1% of the money you normally spend.
There are no other situations, anywhere, where we worry about amounts of energy and water this small. I can’t find any other places where people have gotten worried about things they do that use such tiny amounts of energy. Chatbot energy and water use being a problem is a really bizarre meme that has taken hold, I think mostly because people are surprised that chatbots are being used by so many people that on net their total energy and water use is noticeable. Being “mindful” with your chatbot usage is kind of like filling a large pot of water to boil to make food, and before boiling it, taking a pipet and removing tiny drops of the water from the pot at a time to “only use the water you need” or stopping your shower a tenth of a second early for the sake of the climate. You do not need to be “mindful” with your chatbot usage for the same reason you don’t need to be “mindful” about those additional droplets of water you boil.
The usage metrics for energy that support his argument came from Sam Altman. And this is also terrible reasoning because it doesn’t really matter that the prompts are only 0.3ml. Since the rest comes from generating a response.
However, that 2 mL of water is mostly the water used in the normal power plants the data center draws from. The prompt itself only uses about 0.3 mL, so if you’re mainly worried about the water data centers use per prompt, you use about 300,000 times as much every day in your normal life.
I will take a look at the original article.
But again I’m going to restate that the article you posted uses tech oligarchs as primary sources. Which just on the face of it looks like green washing.