So with a safe prompt there is always a chance the AI will go on a bad direction and then refuse to work, and make you pay for the tokens of his "I am sorry ....long speech"
But if you’re looking to generate truly racist and vile stuff, yes, you probably will need a model with no training or inference filters. I’m not sure anyone’s investing in building that though.
In theory, a full blown LLM gives you a lot more variety and ability to handle novel situations, but it also gives you a lot more potential for conversational gambits that don't affect the game mechanics in the way you want them to and general weirdness (I love the article's anecdote about the Sim who thinks his neighbour Adam Smith wrote Wealth of Nations!). I'm sure someone will ultimately end up designing great LLM-driven game experiences, but I don't imagine they'll look much like The Sims.
It's what I always find with open world games for example. No matter what kind of character I build or how I behave I'm only ever going to get predetermined dialogue which I could have looked up and saved myself the time.
Yes in the sense of the Sims it's probably overkill (they sold millions of copies even without a script) but it's only being used as a playground to test the ideas and see what's possible, there's no suggestion that this is particularly what the Sims itself should be.
imagine this issue when you are just the devloper and not the user, the user complains about this but you try and works for you, but then it fails again for user, in my case the word "monkey" might trigger ChatGPT to either create soem racist shit or it's moderation code to false flag itself.
No it's not. Elon Musk thinks there's a huge demand for racist AI, and he's probably right, or at least alt-right.
Elon Musk yearns for AI devs to build 'anti-woke' rival ChatGPT bot:
If you want a layer to moderate what the year is seeing, you can add that as well. The point of the reverse moderator is to get GPT to do what it’s told without lying about itself, more or less.
Again: 1 I give them safe/clean prompt 2 AI returns 2 of 10 times unsafe crap that is filtered by them 3 I have to pay for my prompt, then have to catch they non deterministic response and retry again on my money
What should happen
1 customer give safe/clean prompt 2 AI response in racist/bad way 3 filter catches this , then it retries again, a few times, if the AI is still racist/bad then OpenAI automatically adds to the prompt "do not be a racist" 4 customer gets the answer