zlacker

[parent] [thread] 1 comments
1. bobbru+(OP)[view] [source] 2023-01-15 00:55:31
That is not true. The dataset is needed, the same way that examples are used by a person learning to draw. But the dataset alone is not capable of producing images not derived from any part of it (and there are many examples of SD results that seem so far to be wholly original), so you can’t reduce stable diffusion to being only derived from the dataset. It may “remember” and generate parts of images in the dataset - but that is a bug, not a feature. With enough prompt tweaking, it may even generate a fairly good copy of pre-existing work - which was what the prompt requested, so responsibility should lie on the prompt writer, not on SD.

But the fact that it often generates new content, that didn’t exist before, or at least doesn’t breach the limits of fair use, goes against the argument made in the lawsuit.

replies(1): >>manhol+LA
2. manhol+LA[view] [source] 2023-01-15 09:01:43
>>bobbru+(OP)
The model can generate original images, yes, and those images might be fair use. But it can also generate near verbatim copies of the source works or substantial parts thereof, so the model itself is not fair use, it's a wholly derivative work.

For example, if a publish a music remix tool with a massive database of existing music, creators might use to create collages that are original and fall under fair use. But the tool itself is not and requires permission from the rights owners.

[go to top]