zlacker

[parent] [thread] 4 comments
1. einpok+(OP)[view] [source] 2023-11-20 20:49:27
No, they can't "correct reasoning errors", and that's a clickbait title.
replies(2): >>swatco+h4 >>ming03+tg
2. swatco+h4[view] [source] 2023-11-20 21:06:10
>>einpok+(OP)
They can produce text that is more sound than prior text that appeared earlier in the same input, when interim text indicates that something in the earlier block was unsound. (Sometimes)

It's the same pattern you'd see in a pedagological article about correcting reasoning errors, except that it's able to generate some share of the article content on its own.

With more layers of post-processing behind a curtain, you might be able to build an assembly over this behavior that looked convincingly like it was correcting reasoning errors on its own.

So... yes and no.

3. ming03+tg[view] [source] 2023-11-20 22:00:09
>>einpok+(OP)
If you look at the paper, they only claim LLM can correct the errors if the mistake location is given. And the mistake finding part is not yet solved.
replies(1): >>einpok+Gy
◧◩
4. einpok+Gy[view] [source] [discussion] 2023-11-20 23:42:45
>>ming03+tg
They don't correct errors even then. They just generate something which sounds like what one might say in a conversation when constrained not to express the error. If there's essentially just one option for that, its the correct one - but then it's like telling someone that the answer to a yes/no question is not the one they generated. Not much "error correction" to do then.
replies(1): >>cmrdpo+1k1
◧◩◪
5. cmrdpo+1k1[view] [source] [discussion] 2023-11-21 05:11:05
>>einpok+Gy
Yep. Where you can see them really get tripped up is if there's multiple "levers" or points for potential contradiction. Lots of dependent clauses, chains of predicates that all have to line up for something to make sense. When they get one item wrong, they don't "see" the consequences for the others. And if you get them to correct one, they'll then often turn around and mess up the others.

Because at no point is the "mind" involved doing a step by step reduction of the problem. It doesn't do formal reasoning.

Humans usually don't either, but they can almost all do a form of it when required to. Either under the assistance of a teacher, or in extremis when they need to. We've all had the experience of being flustered, taking a deep breath, and then "working through" something. After spending time with GPT, etc it becomes clear they're not doing that.

It's not that reasoning comes intrinsic to all human thoughts -- we're far lazier than that -- but when we need to, we can usually do it.

[go to top]