zlacker

[parent] [thread] 95 comments
1. cardan+(OP)[view] [source] 2022-12-15 12:15:07
I don't see the point. There is a copyright (and in that regard most of these images are fine) and then there is trademark which they might violate.

Regardless, the human generating and publishing these images is obviously responsible to ensure they are not violating any IP property. So they might get sued by Disney. I don't get why the AI companies would be effected in any way. Disney is not suing Blender if I render an image of Mickey Mouse with it.

Though I am sure that artists might find an likely ally in Disney against the "AI"'s when they tell them about their idea of making art-styles copyright-able Being able to monopolize art styles would be indeed a dream come true for those huge corporations.

replies(6): >>palata+v1 >>Tepix+V1 >>wokwok+92 >>xg15+94 >>yellow+Le >>dredmo+zi
2. palata+v1[view] [source] 2022-12-15 12:25:46
>>cardan+(OP)
Well Disney would probably sue Blender if there was a "generate Mickey Mouse model" button in it. It's not a totally fair comparison.
replies(3): >>subw00+h2 >>idleha+Ka >>poulpy+7e
3. Tepix+V1[view] [source] 2022-12-15 12:27:50
>>cardan+(OP)
It boils down to this: Do you need permission if you train your AI model with copyrighted things or not?
replies(5): >>gt565k+Q2 >>ben_w+m3 >>cardan+Z3 >>residu+Q5 >>jefftk+Sb
4. wokwok+92[view] [source] 2022-12-15 12:29:51
>>cardan+(OP)
Are you being deliberately obtuse?

It’s blatantly obvious that regardless of if it will work or not, they’re trying to get companies with enough money to file law suits to make a move and do so.

> I don’t see the point.

…or you don’t agree with the intent?

I’m fine with that, if so, but you’d to be deliberately trying very hard not to understand what they’re trying to do.

Quite obviously they’re hoping, similar to software that lets you download videos from YouTube, that tools that enable things are bad, not neutral.

Agree / disagree? Who cares. I can’t believe anyone who “doesn’t get it” is being earnest in their response.

Will it make any difference? Well, it may or may not, but there’s a fair precedent of it happening, and bluntly, no one is immune to law suits.

replies(1): >>dredmo+Xi
◧◩
5. subw00+h2[view] [source] [discussion] 2022-12-15 12:30:19
>>palata+v1
I'm sure it's easy to write an addon for that.
replies(2): >>gl-pro+D8 >>Siempr+R8
◧◩
6. gt565k+Q2[view] [source] [discussion] 2022-12-15 12:33:14
>>Tepix+V1
Ehhh that’s like saying an artist who studies other art pieces and then creates something using combined techniques and styles from those set pieces is what ???? Now liable ???
replies(6): >>Tepix+y3 >>Taywee+j5 >>TaupeR+d7 >>Double+T9 >>alt227+2s >>bakugo+ts
◧◩
7. ben_w+m3[view] [source] [discussion] 2022-12-15 12:35:52
>>Tepix+V1
If you do need permission, is Page Rank a copyright infringing AI, or just a sparkling matrix multiplication derived entirely from everyone else's work?
replies(2): >>people+99 >>Lalaba+U9
◧◩◪
8. Tepix+y3[view] [source] [discussion] 2022-12-15 12:37:08
>>gt565k+Q2
That's like saying creating a thing that looks at one artists artwork and then copies her unique style ad infinitum may need permission first.
replies(2): >>pigsty+84 >>Gigach+d5
◧◩
9. cardan+Z3[view] [source] [discussion] 2022-12-15 12:39:24
>>Tepix+V1
As a human, I can use whatever I want for reference for my drawings. Including copyrighted material.

Now, as for training "AI" models, who knows. You can argue it is the same thing a human is doing or you could argue it a new, different quality and should be under different rules. Regardless, the current copyright laws were written before "AI" models were in widespread use so whatever is allowed or not is more of a historic accident.

So the discussion needs to be about the intention of copyright laws and what SHOULD be.

replies(2): >>vgathe+l5 >>people+L9
◧◩◪◨
10. pigsty+84[view] [source] [discussion] 2022-12-15 12:39:55
>>Tepix+y3
Copying an artist’s style is very much not considered copyright infringement and is how artists learn.

Copying a work itself can be copyright infringement if it’s very close to the original to the point people may think they’re the same work.

11. xg15+94[view] [source] 2022-12-15 12:39:56
>>cardan+(OP)
If thouse mouse images are generated, that implies that Disney content is already part of the training data and models.

So in effect, they are pitting Disney's understanding of copyright (maximally strict) against that of the AI companies (maximally loose).

Even if it's technically the responsibility of the user not to publish generated images that contain copyrighted content, I can't imagine that Disney is very happy with a situation where everyone can download Stable Diffusion and generate their own arbitrary artwork of Disney characters in a few minutes.

So that strategy might actually work. I wish them good luck and will restock my popcorn reserves just in case :)

The problem I see though is that both sides are billion dollar companies - and there is probably a lot of interest in AI tech within Disney themselves. So it might just as well happen that both sides find some kind of agreement that's beneficial for both of them and leaves the artists holding the bag.

replies(4): >>astran+q5 >>taeric+P6 >>wnkrsh+k9 >>jamesh+un
◧◩◪◨
12. Gigach+d5[view] [source] [discussion] 2022-12-15 12:45:35
>>Tepix+y3
You don’t need permission. Style is not an owned thing.
◧◩◪
13. Taywee+j5[view] [source] [discussion] 2022-12-15 12:45:50
>>gt565k+Q2
An AI is not a person. Automated transformation does not remove the original copyright, otherwise decompilers would as well. That the process is similar to a real person is not actually important, because it's still an automated transformation by a computer program.

We might be able to argue that the computer program taking art as input and automatically generating art as output is the exact same as an artist some time after general intelligence is reached, until then, it's still a machine transformation and should be treated as such.

AI shouldn't be a legal avenue for copyright laundering.

replies(3): >>CyanBi+18 >>idleha+gb >>jefftk+vc
◧◩◪
14. vgathe+l5[view] [source] [discussion] 2022-12-15 12:45:58
>>cardan+Z3
This would be a fairly novel law as it would legislate not just the release of an AI but the training as well? That would imply legislating what linear algebra is legal and illegal to do, no?

And practically speaking, putting aside whether a government should even be able to legislate such things, enforcing such a law would be near impossible without wild privacy violations.

replies(2): >>CyanBi+J8 >>manimi+Q9
◧◩
15. astran+q5[view] [source] [discussion] 2022-12-15 12:46:18
>>xg15+94
> If thouse mouse images are generated, that implies that Disney content is already part of the training data and models.

It doesn't mean that. You could "find" Mickey in the latent space of any model using textual inversion and an hour of GPU time. He's just a few shapes.

(Main example: the most popular artist StableDiffusion 1 users like to imitate is not in the StableDiffusion training images. His name just happens to work in prompts by coincidence.)

replies(2): >>Taywee+8a >>mcv+Fo
◧◩
16. residu+Q5[view] [source] [discussion] 2022-12-15 12:47:57
>>Tepix+V1
I would argue if people are allowed to see your art for free, so should AI models.
replies(3): >>people+xa >>dotanc+pe >>bakugo+2t
◧◩
17. taeric+P6[view] [source] [discussion] 2022-12-15 12:53:07
>>xg15+94
This is a bit silly, though? Search Google images for Mickey Mouse, is the results page a possible liability for Google? Why not?

Go to a baker and commission a Mickey Mouse cake. Is that a violation if the bakery didn't advertise it? (To note, a bakery can't advertise it due to trademark, not copyright. Right?)

For that matter, any privately commissioned art? Is that really what artists want to lock away?

replies(4): >>logifa+98 >>crote+89 >>wongar+ga >>sigmoi+0b
◧◩◪
18. TaupeR+d7[view] [source] [discussion] 2022-12-15 12:55:35
>>gt565k+Q2
Not at all, for many reasons.

1) the artist is not literally copying the copyrighted pixel data into their "system" for training

2) An individual artist is not a multi billion dollar company with a computer system that spits out art rapidly using copyrighted pixel data. A categorical difference.

replies(3): >>brushf+h8 >>endorp+r8 >>astran+x8
◧◩◪◨
19. CyanBi+18[view] [source] [discussion] 2022-12-15 12:59:03
>>Taywee+j5
Except the machine is not automatically generating an input

> automatically generating art as output

The user is navigating the latent space to obtain said output, I don't know if that's transformative or not, but it is an important distinction

If the program were wholy automated as in it had a random number/words generator added to it and no navigation of the latent space by users happened, then yeah I would agree, but that's not the case at least so far as ml algos like midjourney or stable diffusion are concerned

replies(3): >>Taywee+y9 >>Retric+G9 >>jamesd+ia
◧◩◪
20. logifa+98[view] [source] [discussion] 2022-12-15 13:00:02
>>taeric+P6
> Search Google images for Mickey Mouse, is the results page a possible liability for Google?

In 2018[0], didn't Getty force Google to change how Google Images presented results, following a lawsuit in 2016[1]?

[0] https://arstechnica.com/gadgets/2018/02/internet-rages-after... [1] https://arstechnica.com/tech-policy/2016/04/google-eu-antitr...

◧◩◪◨
21. brushf+h8[view] [source] [discussion] 2022-12-15 13:01:06
>>TaupeR+d7
Those reasons don't make sense to me.

On 1, human artists are copying copyrighted pixel data into their system for training. That system is the brain. It's organic RAM.

On 2, money shouldn't make a difference. Jim Carrey should still be allowed to paint even though he's rich.

If Jim uses Photoshop instead of brushes, he can spit out the style ideas he's copied and transformed in his brain more rapidly - but he should still be allowed to do it.

replies(3): >>astran+db >>Taywee+gc >>Alexan+8r
◧◩◪◨
22. endorp+r8[view] [source] [discussion] 2022-12-15 13:02:13
>>TaupeR+d7
Have to disagree with point 1, often this is what artists are doing. More strictly in the music part (literally playing others songs), less strictly in the drawing part. But copying, incorporating and developing are some of the core foundations of art.
◧◩◪◨
23. astran+x8[view] [source] [discussion] 2022-12-15 13:02:29
>>TaupeR+d7
Diffusion models don't copy the pixels you show them. You cannot generally tell which training images inspired which output images.

(That's as opposed to a large language model, which does memorize text.)

Also, you can train it to imitate an artist's style just by showing it textual descriptions of the style. It doesn't have to see any images.

replies(1): >>mejuto+Vv
◧◩◪
24. gl-pro+D8[view] [source] [discussion] 2022-12-15 13:03:10
>>subw00+h2
Then the author of that addon would be liable. Not blender.
◧◩◪◨
25. CyanBi+J8[view] [source] [discussion] 2022-12-15 13:03:34
>>vgathe+l5
> That would imply legislating what linear algebra is legal and illegal to do, no?

No, it would just legislate what images are and which ones are not on the training data to be parsed, artists want a copyright which makes their images unusable for machine learning derivative works.

The trick here is that eventually the algorithms will get good enough that it won't be necessary for said images to even be on the training data in the first place, but we can imagine that artists would be OK with that

replies(2): >>astran+Q8 >>stale2+mu1
◧◩◪◨⬒
26. astran+Q8[view] [source] [discussion] 2022-12-15 13:04:34
>>CyanBi+J8
> The trick here is that eventually the algorithms will get good enough that it won't be necessary for said images to even be on the training data in the first place, but we can imagine that artists would be OK with that

They shouldn't be OK with that and they probably aren't. That's a much worse problem for them!

The reason they're complaining about copyright is most likely coping because this is what they're actually concerned about.

◧◩◪
27. Siempr+R8[view] [source] [discussion] 2022-12-15 13:04:37
>>subw00+h2
Try it and see if it is blender or you as the addon creator that gets sued.
◧◩◪
28. crote+89[view] [source] [discussion] 2022-12-15 13:06:25
>>taeric+P6
> Is the results page a possible liability for Google?

Absolutely. Google previously had a direct link to the full-size image, but it has removed this due to potential legal issues. See [0].

> Is that a violation if the bakery didn't advertise it?

According to Disney, it is. See [1].

> Any privately commissioned art?

Not any art, no. Only that which uses IP/material they do not have a license to.

[0]: https://www.ghacks.net/2018/02/12/say-goodbye-to-the-view-im...

[1]: https://en.wikipedia.org/wiki/Cake_copyright#Copyright_of_ar...

replies(1): >>taeric+BJ
◧◩◪
29. people+99[view] [source] [discussion] 2022-12-15 13:06:41
>>ben_w+m3
Page Rank doesn't reproduce any content claiming it's new.

You can however disallow Google from indexing your content using robots.txt a met tag in the HTML or an HTTP header.

Or you can ask Google to remove it from their indexes.

Your content will disappear from then on.

You can't un-train what's already been trained.

You can't disallow scraping for training.

The damage is already done and it's irreversible.

It's like trying to unbomb Hiroshima.

replies(1): >>CyanBi+la
◧◩
30. wnkrsh+k9[view] [source] [discussion] 2022-12-15 13:07:10
>>xg15+94
You can search the LAION5B CLIP-space and you find a lot of mickey in it, lots of fan art between photos of actual merch. If you search with a high aesthetic score, you'll find lots of actual Disney illustrations etc. in the neighbourhood. [0]

[0] https://rom1504.github.io/clip-retrieval/

replies(1): >>xg15+rf
◧◩◪◨⬒
31. Taywee+y9[view] [source] [discussion] 2022-12-15 13:08:18
>>CyanBi+18
That's still automated in the same way that a compiler is automated. A compiler doesn't remove the copyright, neither does a decompiler. This isn't different enough to have different copyright rules. There are more layers to the transformation, but it's still a program with input and output. I'm not sure what you mean by "navigation of latent space". It's generating a model from copyrighted input and then using that model and more input to generate output. It's a machine transformation in more steps.
◧◩◪◨⬒
32. Retric+G9[view] [source] [discussion] 2022-12-15 13:09:09
>>CyanBi+18
The output is probably irrelevant here, the model itself is a derivative work from a copyright standpoint.

Going painting > raw photo (derivative work), raw photo > jpg (derivative work), jpg > model (derivative work), model > image (derivative work). At best you can make a fair use argument at that last step, but that falls apart if the resulting images harm the market for the original work.

replies(2): >>Peteri+8o >>strken+jx
◧◩◪
33. people+L9[view] [source] [discussion] 2022-12-15 13:09:28
>>cardan+Z3
> As a human

you have rights.

AIs don't.

Because they don't have will.

It's like arresting a gun for killing people.

So, as a human, the individual(s) training the AI or using the AI to reproduce copyrighted material, are responsible for the copyright infringement, unless explicitly authorized by the author(s).

◧◩◪◨
34. manimi+Q9[view] [source] [discussion] 2022-12-15 13:09:44
>>vgathe+l5
I am not allowed to print $100 bills with my general-purpose printer. Many printing and copy machines come with built-in safeguards to prevent users from even trying.

It's quite possible to apply the same kind of protections to generative models. (I hope this does not happen, but it is fully possible.)

replies(1): >>bootsm+Dh
◧◩◪
35. Double+T9[view] [source] [discussion] 2022-12-15 13:09:55
>>gt565k+Q2
That's still the question that it boils down to, even if the answer is a "No".
◧◩◪
36. Lalaba+U9[view] [source] [discussion] 2022-12-15 13:09:59
>>ben_w+m3
The output of Pagerank for a given page is not another new page, that's curiously close in style and execution but laundered of IP concerns.

A tool that catalogues attributed links can't really be evaluated the same way as pastiche machine.

You'd be much closer using the example of Google's first page answer snippets, that are pulled out of a site's content with minimal attribution.

◧◩◪
37. Taywee+8a[view] [source] [discussion] 2022-12-15 13:11:29
>>astran+q5
If you can find a copyrighted work in that model that wasn't put there with permission, then why would that model and its output not violate the copyright?
replies(2): >>astran+ma >>mcv+kq
◧◩◪
38. wongar+ga[view] [source] [discussion] 2022-12-15 13:11:59
>>taeric+P6
The right to citation is already part of the 1886 Berne Convention, a precedent that enables services like Google images.

The matters of the baker and the privately comissioned art are more complicated. The artist and baker hold copyrigh for their creation, but their products are also derived from copyrighted work, so Disney also has rights here [1]. This is just usually not enforced by copyright holders because who in their right mind would punish free marketing.

1: https://en.wikipedia.org/wiki/Derivative_work

◧◩◪◨⬒
39. jamesd+ia[view] [source] [discussion] 2022-12-15 13:12:09
>>CyanBi+18
finally, a good use for a blockchain, decentralized defeating of copyright
◧◩◪◨
40. CyanBi+la[view] [source] [discussion] 2022-12-15 13:12:28
>>people+99
That's actually interesting, adding Metadata to the images as a check for allowing or disallowing ai usage

That might be a good way to go about it

replies(1): >>ben_w+bj
◧◩◪◨
41. astran+ma[view] [source] [discussion] 2022-12-15 13:12:30
>>Taywee+8a
https://en.wikipedia.org/wiki/The_Library_of_Babel

A latent space that contains every image contains every copyrighted image. But the concept of sRGB is not copyrighted by Disney just yet.

replies(1): >>Taywee+Lc
◧◩◪
42. people+xa[view] [source] [discussion] 2022-12-15 13:13:11
>>residu+Q5
people are allowed to take a walk in the park, so why cars or tanks or bulldozers are not?
replies(1): >>residu+Wc
◧◩
43. idleha+Ka[view] [source] [discussion] 2022-12-15 13:14:17
>>palata+v1
These AI models are closer to Google in that regard, yes, you can instruct them to generate a Mickey Mouse image, but you can instruct them to generate any kind of image, just like you can search for anything on Google, including Mickey Mouse. When using these models you are essentially performing a search in the model weights.
replies(1): >>thih9+k21
◧◩◪
44. sigmoi+0b[view] [source] [discussion] 2022-12-15 13:15:32
>>taeric+P6
>is the results page a possible liability for Google?

That's actually a tricky question and lengthy court battles were held over this in both the US and Europe. In the end, all courts decided that the image result page is questionable when it comes to copyright, but generally covered by fair use. The question is how far fair use goes when people are using the data in derivative work. Google specifically added licensing info about images to further cover their back, but this whole fair use stuff gets really murky when you have automatic scrapers using google images to train AIs who in turn create art for sale eventually. There's a lot of actors in that process that profit indirectly from the provided images. This will probably once again fall back to the courts sooner or later.

replies(1): >>red_tr+Xh
◧◩◪◨⬒
45. astran+db[view] [source] [discussion] 2022-12-15 13:16:33
>>brushf+h8
> On 1, human artists are copying copyrighted pixel data into their system for training. That system is the brain. It's organic RAM.

They probably aren't doing that. Studying the production methods and WIPs is more useful for a human. (ML models basically guess how to make images until they produce one that "looks like" something you show it.)

replies(1): >>Mezzie+Ri1
◧◩◪◨
46. idleha+gb[view] [source] [discussion] 2022-12-15 13:16:44
>>Taywee+j5
Now we are in Ship of Theseus territory. If I downsample an image and convert it into a tiny delta in the model weights, from which the original image can never be recovered, is that infringement?
◧◩
47. jefftk+Sb[view] [source] [discussion] 2022-12-15 13:19:16
>>Tepix+V1
Which is also what the GitHub co-pilot suit is about: https://githubcopilotlitigation.com

If you have views on whether they'll win, the prediction market is currently at 49%: https://manifold.markets/JeffKaufman/will-the-github-copilot...

◧◩◪◨⬒
48. Taywee+gc[view] [source] [discussion] 2022-12-15 13:21:15
>>brushf+h8
A human can grow and learn based on their own experiences separate from their art image input. They'll sometimes get creative and develop their own unique style. Through all analogies, the AI is still a program with input and output. Point 1 doesn't fit for the same reason it doesn't work for any compiler. Until AI can innovate itself and hold its own copyright, it's still a machine transformation.
◧◩◪◨
49. jefftk+vc[view] [source] [discussion] 2022-12-15 13:22:15
>>Taywee+j5
> Automated transformation does not remove the original copyright

Automated transformation is not guaranteed to remove the original copyright, and for simple transformations it won't, but it's an open question (no legal precedent, different lawyers interpreting the law differently) whether what these models are doing is so transformative that their output (when used normally, not trying to reproduce a specific input image) passes the fair use criteria.

◧◩◪◨⬒
50. Taywee+Lc[view] [source] [discussion] 2022-12-15 13:23:50
>>astran+ma
Sure, but this isn't philosophy. An AI model that contains every image is a copyright derivative of all those images and so is the output generated from it. It's not an abstract concept or a human brain. It's a pile of real binary data generated from real input.
replies(1): >>astran+tf
◧◩◪◨
51. residu+Wc[view] [source] [discussion] 2022-12-15 13:24:52
>>people+xa
A bulldozer destroys the park and other people's ability to enjoy it -- active, destructive. Passively training a model on an artwork does not change the art in the slightest -- passive, non-destructive

Mind you, this is not talking about the usage rights of images generated from such a model, that's a completely different story and a legal one.

replies(1): >>6P58r3+bf
◧◩
52. poulpy+7e[view] [source] [discussion] 2022-12-15 13:30:29
>>palata+v1
But you can already make mickey mouse models, and people do it all the time https://www.youtube.com/watch?v=CqVXoGCTfuk&ab_channel=Ashle...
◧◩◪
53. dotanc+pe[view] [source] [discussion] 2022-12-15 13:32:20
>>residu+Q5
Bad argument. Being allowed to see art and being allowed to copy art are two different things. Being allowed to _copy_ is a reserved _right_, that's the root of the word copyright.
replies(2): >>concor+1h >>Curiou+EN
54. yellow+Le[view] [source] 2022-12-15 13:33:47
>>cardan+(OP)
> I don't get why the AI companies would be effected in any way.

It doesn't necessarily matter if they're affected. My thought when seeing this is that they want some legal precedent to be set which determines that this is not fair use.

◧◩◪◨⬒
55. 6P58r3+bf[view] [source] [discussion] 2022-12-15 13:36:17
>>residu+Wc
> A bulldozer destroys the park and other people's ability to enjoy it

hear hear...

> Passively training a model on an artwork does not change the art in the slightest

copyright holders, I mean individual authors, people who actually produced the content being used, disagree.

They say AI is like a bulldozer destroying the park to them.

Which technically is true, it's a machine that someone (some interested party maybe?) is trying to disguise as a human, doing human stuff.

But it's not.

> passive, non-destructive

Passive, non-destructive, in this context means

- passive: people send the images to you, you don't go looking for them

- non-destructive: people authorized you, otherwise it's destructive of their rights.

◧◩◪
56. xg15+rf[view] [source] [discussion] 2022-12-15 13:38:04
>>wnkrsh+k9
Yes, and probably the copyrighted art of lots of other artists as well. That's the entire point.
◧◩◪◨⬒⬓
57. astran+tf[view] [source] [discussion] 2022-12-15 13:38:11
>>Taywee+Lc
StableDiffusion is 4GB which is approximately two bytes per training image. That's not very derivative, it's actual generalization.

"Mickey" does work as a prompt, but if they took that word out of the text encoder he'd still be there in the latent space, and it's not hard to find a way to construct him out of a few circles and a pair of red shorts.

◧◩◪◨
58. concor+1h[view] [source] [discussion] 2022-12-15 13:48:19
>>dotanc+pe
Except they aren't copying it, but instead drawing inspiration from it. Which all humans have done forever.
replies(2): >>Alexan+Us >>mejuto+bx
◧◩◪◨⬒
59. bootsm+Dh[view] [source] [discussion] 2022-12-15 13:50:37
>>manimi+Q9
Entirely different scales apply here. You can hardcode a printer the 7 different bills each country puts out no problem, but you cannot hardcode the billions of "original" art pieces that the model is supposed to check against during training, its just infeasible.
replies(1): >>Curiou+EM
◧◩◪◨
60. red_tr+Xh[view] [source] [discussion] 2022-12-15 13:52:16
>>sigmoi+0b
Europe has no concept of Fair Use. How did the courts argue there?
replies(2): >>FinnKu+uk >>sigmoi+Or
61. dredmo+zi[view] [source] 2022-12-15 13:55:09
>>cardan+(OP)
Among the goals seems to be a bit of well-poisoning. Artists have done this previously by creating art saying, say, "This site sells STOLEN artwork, do NOT by from them", and encouraging followers to reply with "I want this on a t-shirt", which had previously been used by rip-off sites to pirate artwork. See:

<https://waxy.org/2019/12/how-artists-on-twitter-tricked-spam...>

If art streams are tree-spiked with copyrighted or trademarked works, then AI generators might be a bit more gun-shy about training with abandon on such threads.

It's a form of monkeywrenching.

<https://en.wikipedia.org/wiki/Tree_spiking>

<https://en.wikipedia.org/wiki/Sabotage#As_environmental_acti...>

replies(1): >>gwd+aN
◧◩
62. dredmo+Xi[view] [source] [discussion] 2022-12-15 13:56:49
>>wokwok+92
When disagreeing, please reply to the argument instead of calling names. "That is idiotic; 1 + 1 is 2, not 3" can be shortened to "1 + 1 is 2, not 3."

<https://news.ycombinator.com/newsguidelines.html>

◧◩◪◨⬒
63. ben_w+bj[view] [source] [discussion] 2022-12-15 13:57:50
>>CyanBi+la
If you can make the metadata survive cropping, format shifts, and screenshots.

Can probably do all that well-enough (probably doesn't need to be perfect) by leaning on FAANG, with or without legislation.

But: opt-in by default, or opt-out by default?

◧◩◪◨⬒
64. FinnKu+uk[view] [source] [discussion] 2022-12-15 14:03:51
>>red_tr+Xh
Not a lawyer, but from how I understand it the German courts argued that if you don't use any technology to prevent web crawlers from accessing the pictures on your website you need to accept that they are used for preview images (what the Google picture search technically is) as this is a usual use case.

-> here is the actual judgement though: https://juris.bundesgerichtshof.de/cgi-bin/rechtsprechung/do...

◧◩
65. jamesh+un[view] [source] [discussion] 2022-12-15 14:16:29
>>xg15+94
There's nothing wrong with the model knowing what Mickey Mouse looks like.

There are noninfringing usecases for generating images containing Mickey Mouse - not least, Disney themselves produce thousands of images containing the mouse's likeness every year; but also parody usecases exist.

But even if you are just using SD to generate images, if we want to make sure to avoid treading on Disney's toes, the AI would need to know what Mickey Mouse looks like in order to avoid infringing trademark, too. You can feed it negative weights already if you want to get 'cartoon mouse' but not have it look like Mickey.

The AI draws what you tell it to draw. You get to choose whether or not to publish the result (the AI doesn't automatically share its results with the world). You have the ultimate liability and credit for any images so produced.

replies(1): >>xg15+bq
◧◩◪◨⬒⬓
66. Peteri+8o[view] [source] [discussion] 2022-12-15 14:19:35
>>Retric+G9
It's not clear at all whether the model is a derivative work from a copyright standpoint. Maybe they are, may be they are not - it's definitely not settled, the law isn't very explicit and as far as I know, there is no reasonable precedent yet - and arguably that would be one of the key issues decided (and set as precedent) in these first court battles. I also wouldn't be surprised if it eventually doesn't matter what current law says as the major tech companies may lobby passing a law to explicitly define the rules of the game; I mean if Disney could lobby multiple copyright laws to protect their interests, then the ML-heavy tech companies, being much larger and more wealthy than Disney, can do it as well.

But currently, first, there is a reasonable argument that the model weights may be not copyrightable at all - it doesn't really fit the criteria of what copyright law protects, no creativity was used in making them, etc, in which case it can't be a derivative work and is effectively outside the scope of copyright law. Second, there is a reasonable argument that the model is a collection of facts about copyrighted works, equivalent to early (pre-computer) statistical ngram language models of copyrighted books used in e.g. lexicography - for which we have solid old legal precedent that creating such models are not derivative works (again, as a collection of facts isn't copyrightable) and thus can be done against the wishes of the authors.

Fair use criteria comes into play as conditions when it is permissible to violate the exclusive rights of the authors. However, if the model is not legally considered a derivative work according to copyright law criteria, then fair use conditions don't matter because in that case copyright law does not assert that making them is somehow restricted.

Note that in this case the resulting image might still be considered derivative work of an original image, even if the "tool-in-the-middle" is not derivative work.

replies(1): >>Retric+XM
◧◩◪
67. mcv+Fo[view] [source] [discussion] 2022-12-15 14:22:23
>>astran+q5
How do you get that coincidence? To be able to accurately respond to the cue of an artist's name, it has to know the artist, doesn't it?

In any case, in the example images here, the AI clearly knew who Mickey is and used that to generate Mickey Mouse images. Mickey has got to be in the training data.

replies(1): >>esrauc+ib1
◧◩◪
68. xg15+bq[view] [source] [discussion] 2022-12-15 14:28:26
>>jamesh+un
Not a lawyer (and certainly no disney lawyer), but my understanding was that copyright is specifically concerned with how an image is created, less so that it is created. Which is why you can copyright certain recordings that only consist of silence. It just prevents you from using this record to base your own record of silence on, it doesn't generally block you from recording silence.

In the same way, making the model deliberately unable to generate Micky Mouse images would be much more far-reaching than just removing Micky imagery from the trainset.

replies(1): >>jamesh+6s
◧◩◪◨
69. mcv+kq[view] [source] [discussion] 2022-12-15 14:29:03
>>Taywee+8a
The idea behind that is probably that any artist learns from seeing other artists' copyrighted art, even if they're not allowed to reproduce it. This is easily seen from the fact that art goes through fashions; artists copy styles and ideas from each other and expand on that.

Of course that probably means that those copyrighted images exist in some encoded form in the data or neural network of the AI, and also in our brain. Is that legal? With humans it's unavoidable, but that doesn't have to mean that it's also legal for AI. But even if those copyrighted images exist in some form in our brains, we know not to reproduce them and pass them off as original. The AI does that. Maybe it needs a feedback mechanism to ensure its generated images don't look too much like copyrighted images from its data set. Maybe art-AI necessarily also has to become a bit of a legal-AI.

◧◩◪◨⬒
70. Alexan+8r[view] [source] [discussion] 2022-12-15 14:32:18
>>brushf+h8
I think the parent's point about (2) wasn't about money, but category. A human is a human and has rights, an AI model is a tool and does not have rights. The two would not be treated equally under the law in any other circumstances, so why would you equate them when discussing copyright?
◧◩◪◨⬒
71. sigmoi+Or[view] [source] [discussion] 2022-12-15 14:34:55
>>red_tr+Xh
Fair use is just a limitation of copyright in case of public interest. Europe has very similar exclusions, even though they are spelled out more concretely. But they don't make this particular issue any less opaque.
◧◩◪
72. alt227+2s[view] [source] [discussion] 2022-12-15 14:35:49
>>gt565k+Q2
Depends if the artist creates something new which looks exactly like one of the things he has studied.
◧◩◪◨
73. jamesh+6s[view] [source] [discussion] 2022-12-15 14:36:01
>>xg15+bq
Most Mickey Mouse image usage problems will be trademark infringement not copyright.

Copyright infringement does generally require you to have been aware of the work you were copying. So for sure there's an issue with using AI to generate art where you could use the tool to generate you an image, which you think looks original, because you are unaware of a similar original work, so you could not be guilty of copyright infringement - but if the AI model was trained on a dataset that includes an original copyrighted work that is similar, obviously it seems like someone has infringed something there.

But that's not what we're talking about in the case of mickey mouse imagery, is it? You're not asking for images of 'utterly original uncopyrighted untrademarked cartoon mouse with big ears' and then unknowingly publishing a mouse picture that the evil AI copied from Disney without your knowledge.

replies(1): >>xg15+Vx
◧◩◪
74. bakugo+ts[view] [source] [discussion] 2022-12-15 14:37:41
>>gt565k+Q2
No, it's not the same thing at all, in fact it's entirely unrelated.

Say it with me: Computer algorithms are NOT people. They should NOT have the same rights as people.

◧◩◪◨⬒
75. Alexan+Us[view] [source] [discussion] 2022-12-15 14:38:59
>>concor+1h
This falls apart for 2 reasons. First, I don't think there's any technical definition of "inspiration" that applies to a deeply nested model of numerical weights. It's a machine. A hammer does not draw inspiration from nails that have been hammered in before. Second an AI is not a human under the law and there's no reason to think that an activity that would be considered "transformative" (e.g. learning then painting something similar) when done by a human would still be considered such if performed by an AI.
◧◩◪
76. bakugo+2t[view] [source] [discussion] 2022-12-15 14:39:47
>>residu+Q5
AI models are not people.
◧◩◪◨⬒
77. mejuto+Vv[view] [source] [discussion] 2022-12-15 14:49:32
>>astran+x8
> Also, you can train it to imitate an artist's style just by showing it textual descriptions of the style. It doesn't have to see any images.

And the weights. The weights it has learned come originally from the images.

◧◩◪◨⬒
78. mejuto+bx[view] [source] [discussion] 2022-12-15 14:53:50
>>concor+1h
Following your logic: if AI is like humans why don't we tax its work?
replies(1): >>Curiou+1N
◧◩◪◨⬒⬓
79. strken+jx[view] [source] [discussion] 2022-12-15 14:54:18
>>Retric+G9
The question for me is whether "jpg > model" is derivative or transformative. It's not clear it would be derivative.
replies(1): >>Retric+OM
◧◩◪◨⬒
80. xg15+Vx[view] [source] [discussion] 2022-12-15 14:56:10
>>jamesh+6s
> But that's not what we're talking about in the case of mickey mouse imagery, is it? You're not asking for images of 'utterly original uncopyrighted untrademarked cartoon mouse with big ears' and then unknowingly publishing a mouse picture that the evil AI copied from Disney without your knowledge.

I think this is exactly the problem that many artists have with imagine generators. Yes, we could all easily identify if a generated artwork contained popular Disney characters - but that's because it's Disney, owners of some of the most well-known IP in the world. The same isn't true for small artists: There is a real risk that a model reproduces parts of a lesser known copyrighted work and the user doesn't realise it.

I think this is what artists are protesting: Their works have been used as training data and will now be parts of countless generated images, all with no permission and no compensation.

replies(2): >>jamesh+y61 >>TchoBe+aG1
◧◩◪◨
81. taeric+BJ[view] [source] [discussion] 2022-12-15 15:38:04
>>crote+89
I started to go down the rabbit hole of commissioned fan art. To say that that is a quagmire is an understatement. :(
replies(1): >>Macha+iG1
◧◩◪◨⬒⬓
82. Curiou+EM[view] [source] [discussion] 2022-12-15 15:49:59
>>bootsm+Dh
Not exactly true. Given an image, you can find the closest point in the latent space that image corresponds to. It is totally feasible to do this with every image in the training set, and if that point in the latent space is too close to the training image, just add it to a set of "disallowed" latent points. This wouldn't fly for local generation, as the process would take a long time and generate a multi gigabyte (maybe even terabyte) "disallowed" database, but for online image generators it's not insane.
◧◩◪◨⬒⬓⬔
83. Retric+OM[view] [source] [discussion] 2022-12-15 15:50:52
>>strken+jx
You seem to be confused, transformative works are still derivative works. Being sufficiently transformative can allow for a fair use exception but you may need a court case to prove something is sufficiently transformative to qualify.
replies(1): >>strken+Ia2
◧◩◪◨⬒⬓⬔
84. Retric+XM[view] [source] [discussion] 2022-12-15 15:51:31
>>Peteri+8o
You seem to be confused as to nomenclature, transformative works are still derivative works. Being sufficiently transformative can allow for a fair use exception, the distinction is important because you can’t tell if something is sufficiently transformative without a court case.

Also, a jpg seemingly fits your definition as “no creativity was used in making them, etc” but clearly they embody the original works creativity. Similarly, a model can’t be trained on random data it needs to extract information from it’s training data to be useful.

The specific choice of algorithm used to extract information doesn’t change if something is derivative.

◧◩◪◨⬒⬓
85. Curiou+1N[view] [source] [discussion] 2022-12-15 15:51:49
>>mejuto+bx
If an AI ever gets paid for the work it does, I'm sure we will.
◧◩
86. gwd+aN[view] [source] [discussion] 2022-12-15 15:52:09
>>dredmo+zi
Not sure about Stable Diffusion / Metawhatsit, but OpenAI's training set is already curated to make sure it avoids violence and pornography; and in any case, the whole thing relies on humans to come up with descriptions. Not clear how this sort of thing would "spike the well" in that sense.
◧◩◪◨
87. Curiou+EN[view] [source] [discussion] 2022-12-15 15:53:52
>>dotanc+pe
Bad argument. Copying art is not the crime, distributing the copied art is the crime. The Disney Gestapo can't send storm troopers to your house if your kid draws a perfect rendition of Mickey, but they can if your kid draws a bunch of perfect renditions and sells them online.
◧◩◪
88. thih9+k21[view] [source] [discussion] 2022-12-15 16:55:26
>>idleha+Ka
Google Image results have a note that says: “Images may be subject to copyright”.
◧◩◪◨⬒⬓
89. jamesh+y61[view] [source] [discussion] 2022-12-15 17:12:57
>>xg15+Vx
Right.

So Disney don’t need to worry about AI art tools - so ‘attacking’ them with such tools does nothing.

◧◩◪◨
90. esrauc+ib1[view] [source] [discussion] 2022-12-15 17:34:07
>>mcv+Fo
For other artist cases the corpus can include many images that includes a description with phrases like "inspired by Banksy". Then the model can learn to generate images in the style of Banksy without having any copyrighted images by Banksy in the training set.

The Mickey Mouse case though is obviously bs, the training data definitely does just have tons of infringing examples of Mickey Mouse, it didn't somehow reinvent the exact image of him from first principles.

◧◩◪◨⬒⬓
91. Mezzie+Ri1[view] [source] [discussion] 2022-12-15 18:10:50
>>astran+db
They do sometimes, or at least they used to. I have some (very limited) visual art training, and one of the things I/we did in class was manually mash up already existing works. In my case I smushed the Persistence of Memory and the Arnolfini portrait. It was pretty clear copycat; the work was divided into squares and I poorly replicated the Arnolfini Portrait from square to square.
◧◩◪◨⬒
92. stale2+mu1[view] [source] [discussion] 2022-12-15 19:04:06
>>CyanBi+J8
> but we can imagine that artists would be OK with that

No they won't. If AI art was just as good as it is today, but didn't use copyrighted images in the training set, people would absolutely still be finding some other thing to complain about.

Artists just don't want the tech to exist entirely.

◧◩◪◨⬒⬓
93. TchoBe+aG1[view] [source] [discussion] 2022-12-15 19:57:36
>>xg15+Vx
This already happens all the time in the current status quo with no need for AI.
◧◩◪◨⬒
94. Macha+iG1[view] [source] [discussion] 2022-12-15 19:58:17
>>taeric+BJ
I mean, isn't most of that "It's trademark infringement, but it is both financially tedious and a PR disaster to go after any but the most prominent cases"

Which is why e.g. Bethesda is not going to slap you for your Mr House or Pip-Boy fanart, but will slap the projects that recreate Fallout 3 in engine X.

replies(1): >>spooki+ZY2
◧◩◪◨⬒⬓⬔⧯
95. strken+Ia2[view] [source] [discussion] 2022-12-15 22:26:00
>>Retric+OM
Sorry, yes.
◧◩◪◨⬒⬓
96. spooki+ZY2[view] [source] [discussion] 2022-12-16 03:45:19
>>Macha+iG1
The tables turn when it's not just some fans doing it, which takes time and effort. AI generated images can be pumped out by the thousands, and big companies are behind these services. See the problem?
[go to top]