Try submitting a URL from the following domains, and it will be automatically flagged (but you can't see it's flagged unless you log out):
- archive.is
- watcher.guru
- stacker.news
- zerohedge.com
- freebeacon.com
- thefederalist.com
- breitbart.comEdit: about 67k sites are banned on HN. Here's a random selection of 10 of them:
vodlockertv.com
biggboss.org
infoocode.com
newyorkpersonalinjuryattorneyblog.com
moringajuice.wordpress.com
surrogacymumbai.com
maximizedlivingdrlabrecque.com
radio.com
gossipcare.com
tecteem.comI don't think that makes sense. The supposed spammers can just try looking up whether their submissions show up or not when not logged in.
Even Cory Doctorow made this case in "Como is Infosec" [1].
The only problem with Cory's argument is, he points people to the SC Principles [2]. The SCP contain exceptions for not notifying about "spam, phishing or malware." But anything can be considered spam, and transparency-with-exceptions has always been platforms' position. They've always argued they can secretly remove content when it amounts to "spam." Nobody has challenged them on that point. The reality is, platforms that use secretive moderation lend themselves to spammers.
[1] https://doctorow.medium.com/como-is-infosec-307f87004563
I once had the domain 'moronsinahurry' registered, though not with this group in mind...
No research has been done about whether shadow moderation is good or bad for discourse. It was simply adopted by the entire internet because it's perceived as "easier." Indeed, for platforms and advertisers, it certainly is an easier way to control messaging. It fools good-faith users all the time. I've shared examples of that elsewhere in this thread.
[0] https://deer-run.com/users/hal/sysadmin/greet_pause.html
The internet has run on secrets for 40 years. That doesn't make it right. Now that everyone and their mother is online, it's time to consider the harms that secrets create.
Another commenter argued "Increasing cost of attacks is an effective defense strategy."
I argued it is not, and you said adding a delay can cut out bad stuff. Delays are certainly relevant to the main post, but that's not what I was referring to. And I certainly don't argue against using secrets for personal security! Securitizing public discourse, however, is another matter.
Can you elaborate on GreetPause? Was it to prevent a DDOS? I don't understand why bad requests couldn't just be rejected.
[1] >>37130143
https://www.revsys.com/tidbits/greet_pause-a-new-anti-spam-f...
I get several thousand SPAM attempts per day: I estimate that this one technique kills a large fraction of them. And look how old the feature is...
I don't consider GreetPause to be a form of shadow moderation because the sender knows the commands were rejected. The issue with shadow moderation on platforms is that the system shows you one thing while showing others something else.
Legally speaking, I have no problem with shadow moderation. I only argue it's morally wrong and bad for discourse. It discourages trust and encourages the growth of echo chambers and black-and-white thinking.