“CSAM generated by AI is still CSAM,” DOJ says after rare arrest

https://lemmy.world/post/15665099

“CSAM generated by AI is still CSAM,” DOJ says after rare arrest - Lemmy.World

Then we should be able to charge AI (the developers moreso) for the same disgusting crime, and shut AI down.

…no

That’d be like outlawing hammers because someone figured out they make a great murder weapon.

Just because you can use a tool for crime, doesn’t mean that tool was designed/intended for crime.

It would be more like outlawing ivory grand pianos because they require dead elephants to make - the AI models under question here were trained on abuse.

A person (the arrested software engineer from the article) acquired a tool (a copy of Stable Diffusion, available on github) and used it to commit crime (trained it to generate CSAM + used to to generate CSAM).

That has nothing to do with the developer of the AI, and everything to do with the person using it.

I stand by my analogy.

Unfortunately the developer trained it on some CSAM which I think means they’re not free of guilt - we really need to rebuild these models from the ground up to be free of that taint.

Reading that article:

Given it’s public dataset not owned or maintained by the developers of Stable Diffusion; I wouldn’t consider that their fault either.

I think it’s reasonable to expect a dataset like that should have had screening measures to prevent that kind of data being imported in the first place. It shouldn’t be on users (here meaning the devs of Stable Diffusion) of that data to ensure there’s no illegal content within the billions of images in a public dataset.

That’s a different story now that users have been informed of the content within this particular data, but I don’t think it should have been assumed to be their responsibility from the beginning.