zlacker

[parent] [thread] 1 comments
1. mhink+(OP)[view] [source] 2023-04-27 23:18:04
> Do you fine-tune the LLM to adhere to a format that the planner understands?

This one. It looks like they're using GPT3 to translate the natural-language problem context and goal into a format called PDDL (planning domain definition language), then feeding the result into a separate program that generates a plan based on the context and goal.

With that in mind, the thing they're really testing here is how well GPT3 can translate the natural-language prompt into PDDL, evaluated on the basis of whether the generated PDDL can actually solve the problem and how long the resulting solution takes.

Naturally, I could be wrong but that's at least what it looks like.

replies(1): >>YeGobl+au1
2. YeGobl+au1[view] [source] 2023-04-28 14:10:08
>>mhink+(OP)
There is no fine-tuning. They simply use prompt engineering. See Section 3 for "Method", they have a short, easy-to grok motivating example, I don't think you need to be an expert in planning to see what they do.

To summarise, they assume a human expert can provide a domain description, specifying all actions that can be taken at each situation, and their effects. Then it looks like they include that domain description to the prompt, along with an example of the kind of planning task they want it to solve, and get the LLM to generate PDDL in the context of the prompt.

[go to top]