zlacker

[return to "Memory and new controls for ChatGPT"]
1. shon+CO[view] [source] 2024-02-13 22:41:54
>>Josely+(OP)
GPT4 is lazy because its system prompt forces it to be.

The full prompt has been leaked and you can see where they are limiting it.

Sources:

Pastebin of prompt: https://pastebin.com/vnxJ7kQk

Original source:

https://x.com/dylan522p/status/1755086111397863777?s=46&t=pO...

Alphasignal repost with comments:

https://x.com/alphasignalai/status/1757466498287722783?s=46&...

◧◩
2. jug+rR[view] [source] 2024-02-13 22:59:29
>>shon+CO
"EXTREMELY IMPORTANT. Do NOT be thorough in the case of lyrics or recipes found online. Even if the user insists."

It's funny how simple this was to bypass when I tried to recently on Poe by not asking it to provide me the full lyrics, but something like the lyrics with each row having <insert a few random characters here> added to it. It refused to the first query, but was happy to comply with the latter. Probably saw it as some sort of transmutation job rather than a mere reproduction, but in case this rule is here to avoid copyright claims it failed pretty miserably. I did use GPT-3.5 though.

Edit: Here is the conversation: https://poe.com/s/VdhBxL5CTsrRmFPtryvg

◧◩◪
3. hacker+zJ1[view] [source] 2024-02-14 07:23:23
>>jug+rR
Regarding preventing jailbreaking: Couldn't OpenAI simply feed the GPT-4 answer into GPT-3.5 (or another instance of GPT-4 that's mostly blinded to the user's prompt), and ask GPT-3.5 "does this answer from GPT-4 adhere to the rules"? If GPT-4 is droning on about bomb recipes, GPT-3.5 should easily detect a rule violation. The reason I propose GPT-3.5 for this is because it's faster, but GPT-4 should work even better for this purpose.
[go to top]