Imagine an image generation model whose loss function is essentially "make this other model classify your image as CSAM."
I'm not entirely convinced whether it would create actual CSAM instead of adversarial examples, but we've seen other models of various kinds "reversed" in a similar vein, so I think there's quite a bit of risk there.