zlacker

[return to "We’ve filed a law­suit chal­leng­ing Sta­ble Dif­fu­sion"]
1. dr_dsh+12[view] [source] 2023-01-14 07:17:25
>>zacwes+(OP)
“Sta­ble Dif­fu­sion con­tains unau­tho­rized copies of mil­lions—and pos­si­bly bil­lions—of copy­righted images.”

That’s going to be hard to argue. Where are the copies?

“Hav­ing copied the five bil­lion images—with­out the con­sent of the orig­i­nal artists—Sta­ble Dif­fu­sion relies on a math­e­mat­i­cal process called dif­fu­sion to store com­pressed copies of these train­ing images, which in turn are recom­bined to derive other images. It is, in short, a 21st-cen­tury col­lage tool.“

“Diffu­sion is a way for an AI pro­gram to fig­ure out how to recon­struct a copy of the train­ing data through denois­ing. Because this is so, in copy­right terms it’s no dif­fer­ent from an MP3 or JPEG—a way of stor­ing a com­pressed copy of cer­tain dig­i­tal data.”

The examples of training diffusion (eg, reconstructing a picture out of noise) will be core to their argument in court. Certainly during training the goal is to reconstruct original images out of noise. But, do they exist in SD as copies? Idk

◧◩
2. yazadd+X3[view] [source] 2023-01-14 07:43:18
>>dr_dsh+12
> That’s going to be hard to argue. Where are the copies?

In fairness, Diffusion is arguably a very complex entropy coding similar to Arithmetic/Huffman coding.

Given that copyright is protectable even on compressed/encrypted files, it seems fair that the “container of compressed bytes” (in this case the Diffusion model) does “contain” the original images no differently than a compressed folder of images contains the original images.

A lawyer/researcher would likely win this case if they re-create 90%ish of a single input image from the diffusion model with text input.

◧◩◪
3. visarg+D4[view] [source] 2023-01-14 07:50:34
>>yazadd+X3
> 90%ish of a single input image

Oh, one image is enough to apply copyright as if it were a patent, to ban a process that makes original works most of the time?

The article authors say it works as a "collage tool" trying to minimise the composition and layout of the image as unimportant elements. At the same time forgetting that SD is changing textures as well, so it's a collage minus textures and composition?

Is there anything left to complain about? unless, by draw of luck, both layout and textures are very similar to a training image. But ensuring no close duplications are allowed should suffice.

Copyright should apply one by one, not in bulk. Each work they complain about should be judged on its own merits.

◧◩◪◨
4. manhol+a6[view] [source] 2023-01-14 08:08:51
>>visarg+D4
But they are not original works, they are wholly derived works of the training data set. Take that data set away and the algorithm is unable to produce a single original pixel.

The fact that the derivation involves millions of works as opposed to a single one is immaterial for the copyright issue.

◧◩◪◨⬒
5. smegge+ef1[view] [source] 2023-01-14 18:37:21
>>manhol+a6
how is that any different from new human artist that study other artists work to learn a style or technique. In fact it used to be that the preferred way for painters to learn was to repeatedly copy paintings of masters.
◧◩◪◨⬒⬓
6. manhol+dC2[view] [source] 2023-01-15 09:05:56
>>smegge+ef1
What you and many other in the thread seem to be oblivious about is that algorithms are not people. Yes, it may come as a shock to autistic engineers, but the fact that a machine can do something to what a person does does not warant it equal protection under the law.

Copyright, and laws in general, exists to protect the human members of society not some abstract representation of them.

◧◩◪◨⬒⬓⬔
7. weknow+Ws4[view] [source] 2023-01-16 00:04:33
>>manhol+dC2
It seems like you're using "autistic" as an insult here. If that's not your intention you might want to edit this comment to use different verbage.
◧◩◪◨⬒⬓⬔⧯
8. manhol+ee5[view] [source] 2023-01-16 07:47:59
>>weknow+Ws4
What do you mean, autism is well established as a personality trait that diminishes empathy and the ability to understand other people's desires and emotions, while having a strong affinity to things, for example machines and algorithms.

Legislation is driven by people who are, on aggregate, not autistic. So it's entirely appropriate to presume that a person not understanding how that process works is indeed autistic, especially if they suggest machines are subjects of law by analogy with human beings.

It's not that autists are bad people, they are just outliers in the political spectrum, as you can see from the complete disconnect of up-voted AI-related comments on Hacker News, where autistic engineers are clearly over-represented, versus just about any venue where other professionals, such as painters or musicians, congregate. Just try to suggest to them that a corporation has the right to use their work for free and profit from it while leaving them unemployed, because the algorithm the corporation uses to exploit them is in some abstract sense similar to how their brain works. That position is so for out on the spectrum that presuming a personality peculiarity of the emitter is the absolutely most charitable interpretation.

[go to top]