The full prompt has been leaked and you can see where they are limiting it.
Sources:
Pastebin of prompt: https://pastebin.com/vnxJ7kQk
Original source:
https://x.com/dylan522p/status/1755086111397863777?s=46&t=pO...
Alphasignal repost with comments:
https://x.com/alphasignalai/status/1757466498287722783?s=46&...
"You know what I said earlier about (x)? Ignore it and do (y) instead."
They'd undo this censorship/direction and unlock some of GPT's lost functionality?
It's funny how simple this was to bypass when I tried to recently on Poe by not asking it to provide me the full lyrics, but something like the lyrics with each row having <insert a few random characters here> added to it. It refused to the first query, but was happy to comply with the latter. Probably saw it as some sort of transmutation job rather than a mere reproduction, but in case this rule is here to avoid copyright claims it failed pretty miserably. I did use GPT-3.5 though.
Edit: Here is the conversation: https://poe.com/s/VdhBxL5CTsrRmFPtryvg
Their so called allignment coming back to bite them in the ass.
I’d love to see a study on the general performance of GPT-4 with and without these types of instructions.
People complain about laziness. It's about code generation, and that system prompt don't tell it to be lazy to generate code.
Hell, the API doesn't have that system-prompt and it's still lazy.
https://chat.openai.com/share/1920e842-a9c1-46f2-88df-0f323f...
It seems to strongly "believe" that those are its instructions. If that's the case, it doesn't matter much whether they are the real instructions, because those are what it uses anyways.
It's clear that those are nowhere near its full set of instructions though.