What is least-to-most prompting?
Least-to-most prompting is a technique to elicit complex reasoning from large language models
An AI model that takes in some text and predicts how the text is most likely to continue.
Source: Zhou et al., Least-to-most prompting enables complex reasoning in large language models (2023)
Both of these stages are implemented with few-shot prompting instead of using additional training or fine-tuning
Fine-tuning is the process of adapting a pre-trained ML model for more specific tasks or to display more specific behaviors.
Least-to-most prompting was proposed as an alternative to chain-of-thought prompting. In chain-of-thought prompting
A technique which makes a language model generate intermediate reasoning steps in its output.
Least-to-most prompting mitigates this problem by using multiple prompts recursively. Subproblems are solved sequentially based on the results of simpler subproblems. The recursive approach helps the model progressively work up to solutions of more complex problems than the examples, addressing the problem of easy-to-hard generalization.
Zhou et al. found that least-to-most prompting surpasses standard prompting and chain-of-thought prompting in the following tasks:
-
Symbolic manipulation: Take the last letters of each word from a list and concatenate them, e.g., “robot, love, stamp” becomes “tep”.
-
Compositional generalization: Translate natural language into sequential action commands, e.g., “look thrice after jump” becomes “JUMP, LOOK, LOOK, LOOK”.
-
Mathematical reasoning: Math word problems, e.g., “Elsa has 5 apples. Anna has 2 more apples than Elsa. How many apples do they have together?”
Least-to-most prompting works best when tasks are effectively decomposed. This means that the individual subtasks must be simple enough for the model to solve given the solutions of prerequisite subtasks, and that the subtasks must collectively work up to a solution of the original problem. For this to occur, the examples given in the decomposition prompt must properly illustrate the subtask structure. Since the decomposed structure can vary for tasks across domains, new decomposition prompts with appropriate examples must be created for each type of task being solved.