zlacker

[parent] [thread] 2 comments
1. Muffin+(OP)[view] [source] 2023-12-21 01:42:08
> fails at math of course

what did OpenAI do for the LLM to know "if given a math question, write Python for it, and run the code in order to get result" instead of trying to do the math itself?

replies(1): >>Me1000+c1
2. Me1000+c1[view] [source] 2023-12-21 01:57:23
>>Muffin+(OP)
It trained the model with a lot of data to write code instead (probably sandwiched between some special tokens like [run-python]. The LLM runner then takes the code, runs it in a sandbox, and feeds the output back into the prompt and lets GPT continue inferencing. But TL;DR: it trained the model to write code for math problems instead of trying to solve them itself.
replies(1): >>averev+zp
◧◩
3. averev+zp[view] [source] [discussion] 2023-12-21 06:35:24
>>Me1000+c1
It also has some training on problem decomposition. Many smaller models fail before writing the code, they fail when parsing the question.

You can ask them to serialized a problem in prolog, and see exactly when their understanding breaks - this is open hermes 2.5: https://pastebin.com/raw/kr62Hybq

[go to top]