I'm not that concerned about like rogue AI destroying humanity but I'm scared shitless of the human costs of a bunch of business geeks trying to shoehorn AI into every space of human endeavor as a hedge against workers (who they despise)
Something that really strikes me about people posting like "Hollywood is OVER!!!" over some weird uncanny valley nightmare video is - I genuinely do not believe these people understand that when directors or writers or illustrators have creative visions they want control over how those ideas are actualized. That the process of changing an idea into a piece of art is where the art COMES FROM
It reminds me a lot of when I took film classes and there was always a kid who was really dismissive of close readings of films - they'd be like, "am I supposed to believe that it's supposed to MEAN something that there's oranges in the GODFATHER" and its like, yeah, dude! Nothing on screen in a movie is an accident! A bunch of people have been paid to think about everything you see in a movie!
@mark it blew my mind when I figured this out about comics (yes I am not smart), every single thing in that panel is there because some folks wrote/drew/inked/coloured it and they did it all *on purpose* in order to tell the story they wanted to tell (or, y’know, make a fart gag, both are valid); without any of that you’d be looking at a blank page, a comic is literally a bunch of folks choosing what to put on a page
@cawhitworth I don't think it's a "not smart" thing, it's just.... every comic is such an incredible labor that it's actually difficult to evaluate them as such. I think we kind of mentally gloss over stuff like this, especially if it's not in our field of endeavour, because otherwise it becomes stupefying!

@mark this was a gradual realization that came to me over many years. My lifelong enthusiasm for the creative process of motion pictures had to fight through the layers of snottiness that critical and Internet culture paved over me.

Now I’m routinely shocked by people who treat massive creative labors involving thousands of people as so shallow as to be devoid of any literary merit.

@mark there are certainly major motion pictures that I find misbegotten—PROMETHEUS springs immediately to mind—but there are still incredible works of art and moments of drama as part of that whole. TRANSFORMERS, a franchise many would find misbegotten, has intricate depths of detail and history.
@mark also getting involved in live theater, though I’ve let that lapse, had incredibly positive effects on myself as a person and my appreciation of all performing arts.
@mark (resorting to cgi to open wesley snipes' eyes in blade did not mean anything; my larger point being that everything important is a huge group project and the group members are constantly in conflict instead of producing results with hive mind purpose)
@mark Counter argument: Tommy Wiseau. Sometimes things just happen.
@mark The sheer number of stories I've written (including a certain ongoing one I will NOT shut up about) that started one way and evolved as I worked with them, learned more about the characters/setting/ideas, experienced more things, and had weird epiphanies in the bathroom...well, it's practically all of them, y'know? That's part of the thing! You can't (and shouldn't) skip it! This is something anyone who tries to see a project to completion will learn quickly!
@Ferrovore @mark and, indeed, if that sort of art-having-a-spirit-of-its-own aspect ISN'T there, it inevitably ends up being lifeless or meaningless. Dorothy Sayers is correct about the connections between Idea, Energy, and Power.

@mark The story emerging around "AI" has so many parallels to the history of the Luddites.

As @pluralistic says, "It’s not what technology does that matters, but who it does it for and who it does it to."

@mark @pluralistic The issue with AI isn't that it will automate some processes previously done by human creators. It's that it will do so using the output of those creators as "training data" without those creators ever being compensated.

And it will do so to further enrich people that are already rich at the expense of creators, performers, and audiences.

@MadMadMadMadRN @mark @pluralistic

The training process only ever uses existing work as a point of reference to determine if the computer is making correct predictions. Not even a single pixel of the original work is ever stored in the resulting model. It's very unclear why anyone would be entitled to compensation for that. It's like saying someone needs to be compensated for measurement. The goal is not to copy the original work, and that's the only thing IP protects.

@srcrist @mark @pluralistic

Other people's work is an essential input to "AI". Without the data, the "AI" is useless. People who generated that data should be compensated for that.

There are already apps that render photos in the style of particular artists without paying those artists at all. This is not an academic problem we might encounter down the line. This is happening right now.

@MadMadMadMadRN @mark @pluralistic

This is a fundamental misunderstanding of the training process and the models themselves and how they work. As long as people have a shared concept of what makes a thing look "right," you could train an AI model on literally random noise as long as human beings provided it feedback like a rorschach test to say "sure, that sorta looks like a couch."

Existing work is simply a shortcut for this process. They are extant examples of what things look like.

@srcrist @mark @pluralistic

But companies like OpenAI don't buy a bunch of couches and hire a bunch of people to take pictures and label those couches and upload those images to its servers. They just scrape the web for other people's work without compensating or even notifying them. They then fed those images into proprietary, closed systems that they sell to customers. At best, that's fencing the commons. To me, it's more like theft.

Calling it "measurement" doesn't change anything.

@MadMadMadMadRN @mark @pluralistic

Again, though, copyright does not protect an author from USE, it protects an author's right to COPY the image. I think it is an uphill argument that calculating probabilistic weights about what is or is not noise in an image, from millions of images of various concepts, is somehow infringing on an author's right to consent for replication. I do not see how that is anything other than a derivative work.

You may object to OpenAI, but I don't see the harm here.

@srcrist @mark @pluralistic I'm not going to pretend to be a copyright expert, but I'm pretty sure it protects use as well. That's why artists have to pay to cover a song written by someone else. It's why people can't make money by selling fan fiction of works not in the public domain.

But beyond the issue of copyright, the fact remains that private, for-profit companies are taking work produced by others without their consent or even knowledge and using it to make proprietary products.

@srcrist @mark @pluralistic Like, if I fed all of the code for Windows into an AI as training data and told it to make an operating system, do you think Microsoft would be cool with that? Do you think they would say I am just "measuring" their code to see what operating system code looks like?

@MadMadMadMadRN @mark @pluralistic

To be clear: if the algorithm did not duplicate any of Window's code, you have every right in the world to do exactly this. And this is even settled law. Emulators reverse engineer computer code every day to replicate it, and they are completely legal as long as they do not copy the code itself.

@srcrist @mark @pluralistic How do you "replicate" code without "duplicating" it?

@MadMadMadMadRN @mark @pluralistic

You can replicate the function of code without replicating the actual text and syntax of the code. And this is legal. You cannot copyright an algorithm, and that's all a function in code is. So two people who understand the algorithm conceptually are free to write two different functions, in code, that lead to the same outcomes.

@srcrist @MadMadMadMadRN @mark @pluralistic
The problem is that settled law for copyright is quite different depending upon the domain.

E.g. covering music can be trigger copyright, in some cases actually a couple of notes in a particular order, and the toll booth of the music industry is there to collect.

Fictional characters can be copyright protected.

https://en.wikipedia.org/wiki/Copyright_protection_for_fictional_characters

And so on. And because of WIPO, the perversions of 100+ countries copyright laws are suddenly relevant.

Copyright protection for fictional characters - Wikipedia

@yacc143 @MadMadMadMadRN @mark @pluralistic

It is for this exact reason that I would argue that the appropriate point of regulation with respect to computer generated images is at the point of output, and not training.

No amount of training makes Stable Diffusion or Midjourney capable of doing anything that Photoshop cannot already be made to do by a user. And the real issue here is in the output, as it is with any other image.

@srcrist @MadMadMadMadRN @mark @pluralistic I do admit that's my first instinctive PoV too, it's nothing more than a Photoshop/Gimp plugin.

@MadMadMadMadRN @mark @pluralistic

Your example is a replication. A cover is a rendition of the same song. You certainly do not have to pay an author to listen to a bunch of songs and compile a list of the songs with a particular chord progression, though.

All of this is an oversimplification in general, and IP law can be nuanced. But I don't think it's a very good argument to make that the sort of things "AI" training does are copyright infringement. If that's the case, so is a file hash.

@MadMadMadMadRN @mark @pluralistic

And, again, "take" implies that someone has been deprived of something. That framing is, in and of itself, wrong. They did not "take" anything. They ran an analysis algorithm on something and then the original was discarded.

@srcrist @mark @pluralistic
Than why does pirated software violate copyright?
@MadMadMadMadRN Because it is a copy? Pirated software is not some disambiguation of a copyrighted work. It's a copy of the copyrighted work. Your confusion of these two concepts is at the core of your misunderstanding.

@MadMadMadMadRN @mark @pluralistic

Your concerns about using AI to replicate copyrighted works are fine, but that's an output issue that has nothing to do with training the model. A model can replicate a style without ever having seen the work. It isn't able to replicate their style because it was trained on their work, it's able to replicate their style because that's just how diffusion works.

@srcrist @mark @pluralistic Honestly, I really don't care that much about copyright per se. I think a lot of these terms like "copy" or "use" or "rendition" are not immutable, universal constants but social constructs created by humans with tons of ambiguity and contradictions.

My concern is that OpenAI are using technology that would not be possible without the work of countless creators as inputs. That technology will be deployed in a way that makes the life of those creators worse.

@MadMadMadMadRN @mark @pluralistic

I mean, I guess it's fine, ultimately, to just be like "I don't care about these things, I think it should be illegal if it isn't already." But that's just an ethical opinion. The law very much cares about those terms.

And, again, those countless creators are only relevant in the fact that a concept must exist, yes, for a neural network to be trained. But the value is only in the aggregate weights for many, many examples of concepts. Not any specific work.

@srcrist @mark @pluralistic I never claimed to be issuing a legal opinion. You're the one who brought up copyright.

I guess I just don't see how "analysis in the service of making a proprietary product that people will have to pay to use" is fundamentally different from "use".

I don't want to criminalize the analysis of creative work in all circumstances. Like fair use protects certain types of copying and use, I think there could be "fair analysis" that protects certain types of analysis.

@MadMadMadMadRN @mark @pluralistic

I'm not criticizing you, to be clear, I'm just pointing out the distinction. All of this is effectively inextricable from issues of IP law anyway.

I would argue that existing concepts of fair use already permit the use of copyrighted images for training AI, generally, and there are a lot of legally tested parallel examples that support that.

There may be issues with OpenAI or StabilityAI's process specifically, but not at a 30,000 ft view, imo.

@srcrist @mark @pluralistic You clearly know more about copyright law than I do, so you're probably right. Though I have to admit I don't see how a computer program could analyze text, images, etc. without making a copy of them. But ,again, you know more than I do about how AI works.

@srcrist @mark @pluralistic For me, it's the is-ought distinction. Is it legal for ChatGPT to use the output of creators to make a system off of which OpenAI makes a lot of money without compensating those creators in any way? Maybe. Probably.

Ought it to be legal? To me, no.

@MadMadMadMadRN @mark @pluralistic

It is true that your computer even makes a copy of an image when you load a website. It does require a copy of the image in RAM in order to display it. But it's pretty well established that if you publish an image to be viewed publicly, the public's need to replicate the data in order to display it is not an infringement on your rights as an author. This is similar, imo.

It should also be noted that training uses destroyed copies of images, not originals.

@srcrist @MadMadMadMadRN @pluralistic but the process of converting the idea of "a vase of sunflowers" to Van Gogh's "Sunflowers" - through a combination of technical ability and artistic decision - is the VALUE
@mark @srcrist @pluralistic Right. And, at the end of the day, copyright and IP are fundamentally supposed to be a means to the end of ensuring that creators are somehow compensated for the value they create for others. This is both for moral reasons (it's unjust when people are not compensated for the value they create for others) and practical reasons (if people are not compensated for things they create that generate value for others, they will stop generating such things).

@MadMadMadMadRN @mark @pluralistic

To be clear: copyright ALREADY protects the author from an AI being used to duplicate their work. That protection already exists. That's not actually the conversation here. What you are talking about is extending and reinterpreting copyright to mean new things, so that infringement happens at the point where the work is simply used for analysis...and I think that's highly problematic.

@mark @MadMadMadMadRN @pluralistic

I'm not sure that I'm following your point here. There is certainly value in being able to take a photo or a concept and reinterpret it as similar to a given artists style. But that is a matter of output.

Setting aside that it's not clear that "render my sister in the style of this artist" is actually infringement to begin with; to the extent that it is, it's a factor of what the user is telling the machine to do, not training. Photoshop can do that.

@MadMadMadMadRN @mark @pluralistic well not only. It will also enable lots of artists with little budget to create, and probably earn a decent sum of their creation (well the latter is already in place)
@MiloWinterBurn @mark @pluralistic
No, it won't. Things like Dall-E and ChatGPT are free to use now. But they are owned by huge tech companies that exist to make their shareholders wealthy. As with so many other tech products, once creators are hooked, the companies that own these AI products will enshittify them by charging more for an ever worse product. Also, why would anyone pay an artist for art generated by an AI when they can just go to the source?

@MadMadMadMadRN @mark @pluralistic because the tool isn't the art.

And - the OS models are chasing them big companies, in some areas they are not that far behind (eg stable diffusion). I think there is a decent chance for this to play out with not that much enshittification, since that drives OS innovation.

There is no platform power here, and it remains to be seen how advantageous big models are over medium sized models that we can run on our own machines or with a couple of € of cloud gpus

@MiloWinterBurn @mark @pluralistic Maybe. I hope so. My concern is that these tools will do to creators what Pages did to graphic designers. It won't completely eradicate jobs, but it will make them far more precarious and poorly compensated.

@MadMadMadMadRN @mark @pluralistic well sure, lots of jobs are precious now or soon. Times are a changing. In the end humanity will gain from this, at least overal.

Some people will need to up skill/reskill, or well, lots. I think it's time for UBI. But I give it a couple of years for that to sink in.

AI plus MRI yields the ability to recognize what the mind is hearing

System can also reconstruct speech a person imagines.

Ars Technica

@mark

i was talking to someone about this recently and they mentioned that even when art is bad, it's often interesting, just since it tells you something about the mind of the person who decided to make it.

whereas, when ai "art" is bad, it's almost always bad in a boring way, since it's either misconfigured or bad in the same way everything that uses the same training set is

@mark I think this is where the cognitive dissonance lives — people who don’t do art or make stuff think that coming up with a good idea is 99% of the project, and executing it is just menial labor.

But in fact, ideas are cheaper and decidedly more plentiful than dirt. Realizing the idea in specific, skilled way is what gives it value and interest.

@mark (Thus the constant flow since time immemorial of self-important doofuses who think that authors would LOVE to write a whole book based on their idea and split the income.)