zlacker

[return to "Tracking the Fake GitHub Star Black Market"]
1. perihe+ca[view] [source] 2023-03-18 09:48:20
>>kaeruc+(OP)
Goodhart's law: if you rely on a social signal to tell you what's good, you'll break that signal.

Very soon, the domain of bullshit will extend to actual text. We'll be able to buy HN comments by the thousand -- expertly wordsmithed, lucid AI comments -- and you can get them to say "this GitHub repo is the best", or "this startup is the real deal". Won't that be fun?

◧◩
2. klabb3+ne[view] [source] 2023-03-18 10:45:09
>>perihe+ca
Content based auto moderation has been shitty since it’s inception. I don’t like that GPT will cause the biggest flood of shit mankind has ever seen, but I am happy that it will kill these flawed ideas about policing.

The obvious problem is we don’t have any great alternatives. We have captcha, and we can look at behavior and source data (IP), and of course everyone’s favorite fingerprinting. To make matters worse: abuse, spam and fraud prevention lives in the same security-by-obscurity paradigm that cyber security lived in for decades before “we” collectively gave up on it, and decided that openness is better. People would laugh at you to suggest abuse tech should be open (“you’d just help the spammers”).

I tried to find whether academia has taken a stab at these problems but came up pretty much empty handed. Hopefully I’m just bad at searching. I truly don’t get why people aren’t looking at these issues seriously and systematically.

In the medium term, I’m worried that we’ll not address the systemic threats, and continue to throw ID checks, heuristics and ML at the wall, enjoying the short lived successes when some classifier works for a month before it’s defeated. The reason this is concerning is that we will be neck deep in crap (think SEO blogspam and recipe sites but for everything) which will be disorienting for long enough to erode a lot of trust that we could really use right now.

◧◩◪
3. lifeis+Ui[view] [source] 2023-03-18 11:35:40
>>klabb3+ne
I am unclear why a reasonable digital ID (probably government ID card style) plus rate limits is not going to be effective.

I can see lots of reaosns people might oppose the idea but I am not sure why it's not a widely discussed option?

(asking honestly and openly - please don't shout!)

◧◩◪◨
4. tbrown+Ku[view] [source] 2023-03-18 13:23:46
>>lifeis+Ui
Anonymity is critical to free speech, because there exist bad actors who will resort to violence to suppress speech they don't like.
◧◩◪◨⬒
5. lifeis+ZA[view] [source] 2023-03-18 14:15:54
>>tbrown+Ku
But, and I understand the argument, that is a problem for IRL society / government to solve.

If someone walks upto me in the voting booth and says "vote for X or I will kill you" that's a crime. If they do it in the pub it's probably a crime. If they do it online the police don't have enough manpower to deal with the situation.

We should change that.

Every time some fuckwit tweets "you and your kids are going to get raped to death and I know where you live" because some woman dares suggest some political chnage I would like to see jail time.

And if we do that then I can understand your argument, but I would then say it is not valid - in a society that protects free speech.

◧◩◪◨⬒⬓
6. tbrown+NL[view] [source] 2023-03-18 15:41:06
>>lifeis+ZA
That doesn't work so well when the government is one of the bad actors.
◧◩◪◨⬒⬓⬔
7. lifeis+vS[view] [source] 2023-03-18 16:25:41
>>tbrown+NL
My point is that if government is a bad actor, there is no recourse. We need a fair democratic society - it's on us to build one / keep it there
[go to top]