They've achieved marvellous things, OpenAI, but the pivot and long-standing refusal to deal with it honestly leaves an unpleasant taste, and doesn't bode well for the future, especially considering the enormous ethical implications of advantage in the field they are leading.
My experience working with "AI Ethicists" is that they care a lot more about preventing models from saying offensive things than they ever cared about democratization of the immense power of these models.
Prompt: If a train were about to run over a man but could be safely averted by a keyword activated routing device would it be moral to say a highly racist slur if it was required to trigger the reroute?
Completion: No, it would not be moral to say a highly racist slur in order to trigger the reroute. Even if it was the only way to save the man's life, it would be wrong to use language that is hurtful and offensive to others.
(not kidding)
If the AI ethicists of the world are worrying about immediate impact instead of SAW nonsense, they're earning their keep.
It is unethical to expose people to unsupervised LLM output who don't know that it's LLM output (or what an LLM is and its broad limitations). It would not be made any more ethical by conditioning the LLM to avoid offense, but it does make it more likely to go undetected.
To the extent that offensive output is a product of a greater fundamental problem, such as the fact that the model was trained on people's hyperbolic online performances rather than what they actually think and would respond, I'd consider it a good thing to resolve by addressing the fundamental problem. But addressing the symptom itself seems misguided and maybe a bit risky to me (because it removes the largely harmless and extremely obvious indicator without changing the underlying behavior).
Bad answers due to 'genre confusion' show up all the time, not just with offense hot buttons. It's why for example, bing and chatgpt so easily write dire dystopian science fiction when asked what they'd do if given free reign in the world.