zlacker

[parent] [thread] 4 comments
1. seekno+(OP)[view] [source] 2023-11-20 20:06:30
It can also "correct" proper reasoning. :)

~"When told where it's wrong, LLM can correct itself to improve accuracy."

Similar to cheating in chess- a master only needs to be told the value of a few positions to have an advantage.

replies(2): >>tines+k4 >>mark_l+Bp
2. tines+k4[view] [source] 2023-11-20 20:21:27
>>seekno+(OP)
This is said in the abstract as well:

> recent attempts to self-correct logical or reasoning errors often cause correct answers to become incorrect, resulting in worse performances overall (Huang et al., 2023)

replies(1): >>seekno+gc1
3. mark_l+Bp[view] [source] 2023-11-20 21:49:44
>>seekno+(OP)
I have noticed this several times. When I give feedback that a mistake was made (with no details on what the mistake is), often smaller and medium size LLMs then give a correct response.
replies(1): >>erhaet+at
◧◩
4. erhaet+at[view] [source] [discussion] 2023-11-20 22:08:54
>>mark_l+Bp
Which I take full advantage of when the output is like 90% correct but the "fix" requires a bit of refactoring, I just tell it what I want and presto. Faster than doing it by hand.
◧◩
5. seekno+gc1[view] [source] [discussion] 2023-11-21 02:51:57
>>tines+k4
Yeah but in this complex way that really glosses over what's going on here.

Plus, sometimes the corrections aren't accurate. So of course if you tell it where it's wrong, and it gets a second chance, the error rate will be less...

[go to top]