zlacker

[return to "We’ve filed a law­suit chal­leng­ing Sta­ble Dif­fu­sion"]
1. dr_dsh+12[view] [source] 2023-01-14 07:17:25
>>zacwes+(OP)
“Sta­ble Dif­fu­sion con­tains unau­tho­rized copies of mil­lions—and pos­si­bly bil­lions—of copy­righted images.”

That’s going to be hard to argue. Where are the copies?

“Hav­ing copied the five bil­lion images—with­out the con­sent of the orig­i­nal artists—Sta­ble Dif­fu­sion relies on a math­e­mat­i­cal process called dif­fu­sion to store com­pressed copies of these train­ing images, which in turn are recom­bined to derive other images. It is, in short, a 21st-cen­tury col­lage tool.“

“Diffu­sion is a way for an AI pro­gram to fig­ure out how to recon­struct a copy of the train­ing data through denois­ing. Because this is so, in copy­right terms it’s no dif­fer­ent from an MP3 or JPEG—a way of stor­ing a com­pressed copy of cer­tain dig­i­tal data.”

The examples of training diffusion (eg, reconstructing a picture out of noise) will be core to their argument in court. Certainly during training the goal is to reconstruct original images out of noise. But, do they exist in SD as copies? Idk

◧◩
2. groest+J7[view] [source] 2023-01-14 08:23:52
>>dr_dsh+12
> It is, in short, a 21st-cen­tury col­lage tool.

Interesting that they mention collages. IANAL but it was my impression that collages are derivative work if they incorporate many different pieces and only small parts of the original. Their compression argument seems more convincing.

◧◩◪
3. Fillig+LG[view] [source] 2023-01-14 14:30:12
>>groest+J7
Compression down to two bytes per image?

You run into the pigeonhole argument. That level of compression can only work if there are less than seventy thousand different images in existence, total.

Certainly there’s a deep theoretical equivalent between intelligence and compression, but this scenario isn’t what anyone means by “compression” normally.

◧◩◪◨
4. Xelyne+FZ[view] [source] 2023-01-14 16:58:18
>>Fillig+LG
When gzip turns my 10k character ASCII text file into a a 2kb archive, has it "compressed each character down to a fifth of a byte per character"? No, thats a misunderstanding of compression.

Just like gzip, training stable diffusion certainly removes a lot of data, but without understanding the effect of that transformation of the entropy of the data it's meaningless to say thing like "two bytes per image" because(like gzip) you need the whole encoded dataset to recover the image.

It's compressing many images into 10GB of data, not a single image into two bytes. This is directly analogous to what people usually mean by "compression"

[go to top]