zlacker

[parent] [thread] 4 comments
1. astran+(OP)[view] [source] 2023-01-14 09:33:32
Usually judges would care more about whether the bytes came from than how many of them there are.

Since SD is trained by gradient updating against several different images at the same time, it of course never copies any image bits straight into it. Since it's a latent-diffusion model, actual "image"ness is limited to the image encoder (VAE), so any fractional bits would be in there if you want to look.

The text encoder (LAION OpenCLIP) does have bits from elsewhere copied straight into it to build the tokens list.

https://huggingface.co/stabilityai/stable-diffusion-2-1/raw/...

replies(2): >>synu+E3 >>derang+hS
2. synu+E3[view] [source] 2023-01-14 10:14:21
>>astran+(OP)
The important distinction then is using another program or device to analyze the bits but without copying them, that takes its own new impression? Like using a camera?
replies(1): >>astran+35
◧◩
3. astran+35[view] [source] [discussion] 2023-01-14 10:27:31
>>synu+E3
Well, theoretically more like a vague memory of it or taking notes on it.
4. derang+hS[view] [source] 2023-01-14 17:45:48
>>astran+(OP)
“any fractional bits would be in there if you want to look.”

What do you mean by this in the context of generating images via prompt? “Fractional bits” don’t make sense and it’s more misleading if anything. Regardless, a model violating criteria for being within fair use will always be judged by the outputs it generates rather than its composing bytes (which can be independent)

replies(1): >>astran+1T1
◧◩
5. astran+1T1[view] [source] [discussion] 2023-01-15 01:58:29
>>derang+hS
Fractional bits makes perfect sense. Do you know how arithmetic coders work?
[go to top]