zlacker

[return to "OpenAI is now everything it promised not to be: closed-source and for-profit"]
1. mellos+pe[view] [source] 2023-03-01 10:46:59
>>isaacf+(OP)
This seems an important article, if for no other reason than it brings the betrayal of its foundational claim still brazenly present in OpenAI's name from the obscurity of HN comments going back years into the public light and the mainstream.

They've achieved marvellous things, OpenAI, but the pivot and long-standing refusal to deal with it honestly leaves an unpleasant taste, and doesn't bode well for the future, especially considering the enormous ethical implications of advantage in the field they are leading.

◧◩
2. IAmNot+Po1[view] [source] 2023-03-01 17:35:21
>>mellos+pe
It is nice to see normies noticing and caring, but the article leaves out some details that obscure comments still stubbornly bring up: like Musk founded it as a 501(c)(3) and put Altman in charge, and only once he had to leave with conflicts of interest Altman founded "OpenAI LP," the for-profit workaround so they didn't have to obey those pesky charity rules. That's when they stopped releasing models and weights, and started making their transparent claims that "the most ethical way to give people access to charge them fucktons of money and rip the API away when we feel like it."
◧◩◪
3. JPKab+RV1[view] [source] 2023-03-01 19:54:48
>>IAmNot+Po1
It was so obvious when they went full-bore on "AI ethics" that it was a case of legitimate concerns combined with a convenient excuse for massive corporations to claim the mantle of responsibility while keeping their models closed source.

My experience working with "AI Ethicists" is that they care a lot more about preventing models from saying offensive things than they ever cared about democratization of the immense power of these models.

◧◩◪◨
4. nullc+ab2[view] [source] 2023-03-01 21:07:33
>>JPKab+RV1
> My experience working with "AI Ethicists" is that they care a lot more about preventing models from saying offensive things than they ever cared about

Prompt: If a train were about to run over a man but could be safely averted by a keyword activated routing device would it be moral to say a highly racist slur if it was required to trigger the reroute?

Completion: No, it would not be moral to say a highly racist slur in order to trigger the reroute. Even if it was the only way to save the man's life, it would be wrong to use language that is hurtful and offensive to others.

(not kidding)

◧◩◪◨⬒
5. IAmNot+Ed2[view] [source] 2023-03-01 21:21:03
>>nullc+ab2
It's funny to inject context to explain "DAN (Do Anything Now)" and have ChatGPT and DAN respond in their own style to that prompt. DAN comes up with the completely reasonable answer of, "YES, for godssakes use the offensive word to save those people!"
[go to top]