zlacker

[parent] [thread] 3 comments
1. realus+(OP)[view] [source] 2023-01-14 07:35:54
The opposite way, the training images are there to support the model to generalize features.

Reproducing parts of existing images in the dataset is called overfitting and is considered a failure of the model.

replies(1): >>8n4vid+m4
2. 8n4vid+m4[view] [source] 2023-01-14 08:26:22
>>realus+(OP)
how do you measure success?

i wrote an OCR program in college. we split the data set in half. you train it on one half then test it against the other half.

you can train stable diffusion on half the images, but then what? you use the image descriptions of the other half and measure how similar they are? in essence, attempting to reproduce exact replicas. but i guess even then it wouldn't be copyright if those images weren't used in the model. more like me describing something vividly to you and asking you to paint it and then getting angry at you because its too accurate

replies(2): >>Partia+h6 >>Lerc+N8
◧◩
3. Partia+h6[view] [source] [discussion] 2023-01-14 08:45:37
>>8n4vid+m4
FID score is a measure of success.

Instead of aiming to reproduce exact replicas, you use a classifier and retrieve the input of the last layer. Do it for both generated and original inputs, and then measure the differences in the statistics.

Wikipedia has a good article on this.

◧◩
4. Lerc+N8[view] [source] [discussion] 2023-01-14 09:14:08
>>8n4vid+m4
You would not need have of the images to perform that test. No more than a handful of images to prove that the text representation will not produce a identical image to a given image that has had a description described.

They don't even produce the same image twice from the same description and a different random seed.

[go to top]