zlacker

[return to "LLMs cannot find reasoning errors, but can correct them"]
1. pton_x+zB[view] [source] 2023-11-20 22:09:33
>>koie+(OP)
I've also noticed LLMs seem to lack conviction on the correctness of their answers. As the paper notes, you can easily convince the transformer that a correct answer is wrong, and needs adjustment. Ultimately they're just trying to please you. For example with ChatGPT 3.5 (abbreviated):

me: what is sin -pi/2

gpt: -1

me: that's not right

gpt: I apologize, let me clarify, the answer is 1

◧◩
2. kaioke+hD[view] [source] 2023-11-20 22:20:05
>>pton_x+zB
This is due to the RLHF alignment, only product-focused. It would be very annoying for users to fight back and forth with the LLM on the correctness of the answer, especially when it is so prone to hallucination.
[go to top]