If I don't want to see profanity, I should be able to set my filter to exclude profane comments. If I don't want to see nudity, I can set that filter too. Just like movies get a certain rating (G, PG, R, etc.), we should be able to properly label data.
Which inspires another really weird, super uncomfortable thought. If the CSAM producers had cheap, reliable methods of creating their awful content without the use of real people, would that reduce the harm done?
I can't remember the last time I felt so conflicted just asking a question, but there it is.
This is basically what he describes in the article as a form of moderation: "If you wanted to get fancy, you could have a bunch of filters - harassing content, sexually explicit content, conspiracy theories - and let people toggle which ones they wanted to see vs. avoid."
If your concern is with the labels themselves being used to convey a (possibly offensive) message, I think you could just have a way for people to hide specific labels and never see them again. Or maybe a way to label the labels as subjective, or just delete ones that are obvious flamebait.
People love to misuse tools meant for good, on Reddit I've been on the receiving end of the "reddit cares" self-harm notification because of some barely spicy comments.