zlacker

[return to "Send: Open-source fork of Firefox Send"]
1. Arubis+3s[view] [source] 2024-10-19 17:24:31
>>leonry+(OP)
Best of luck to the author! My understanding is that anything that makes large file sharing easy and anonymous rapidly gets flooded with CSAM and ends up shuttering themselves for the good of all. Would love to see a non-invasive yet effective way to prevent such an incursion.
◧◩
2. jart+cA1[view] [source] 2024-10-20 06:56:01
>>Arubis+3s
If governments and big tech want to help, they should upload one of their CSAM detection models to Hugging Face, so system administrators can just block it. Ideally I should be able to run a command `iscsam 123.jpg` and it prints a number like 0.9 to indicate 90% confidence that it is. No one else but them can do it, since there's obviously no legal way to train such a model. Even though we know that governments have already done it. If they won't give service operators the tools to keep abuse off their communications systems, then operators shouldn't be held accountable for what people do with them.
◧◩◪
3. miki12+PN1[view] [source] 2024-10-20 10:02:43
>>jart+cA1
This would potentially let somebody create a "reverse" model, so I don't think that's a good idea.

Imagine an image generation model whose loss function is essentially "make this other model classify your image as CSAM."

I'm not entirely convinced whether it would create actual CSAM instead of adversarial examples, but we've seen other models of various kinds "reversed" in a similar vein, so I think there's quite a bit of risk there.

◧◩◪◨
4. jart+7Q1[view] [source] 2024-10-20 10:38:36
>>miki12+PN1
Are you saying someone will use it to create a CSAM generator? It'd be like turning smoke detectors into a nuclear bomb. If someone that smart wants this, then there are easier ways for them to do it. Analyzing the detector could let you tune normal images in an adversarial way that'll cause them to be detected as CSAM by a specific release of a specific model. So long as you're not using the model to automate swatting, that's not going to amount to much more than a DEFCON talk about annoying people.
◧◩◪◨⬒
5. throwa+x02[view] [source] 2024-10-20 13:01:51
>>jart+7Q1
I think the point is generating an image that looks normal but causes the model to false positive and the unsuspecting person then gets reported
[go to top]