Paper here: https://arxiv.org/pdf/2305.10601.pdf
Large language models are increasingly being used for general problem solving, but they are still limited by their token-level, left-to-right generation process.
The authors propose a new Tree of Thoughts (ToT) framework to address this. ToT frames problem solving as search through a tree of "thoughts", where each thought is an intermediate step towards the solution.
This allows the language model to:
- Generate and explore multiple potential thoughts at each step
- Evaluate the progress of different thoughts using self-evaluation prompts
- Perform lookahead and backtracking to make global decisions
The authors propose 3 tasks to test ToT: Game of 24, Creative Writing and Crosswords.
ToT significantly outperforms standard input-output and chain-of-thought prompting baselines on all 3 tasks. This shows the benefits of ToT's ability to explore, evaluate and search through different reasoning paths.
The key benefits of ToT are:
- Generality: It generalizes existing prompting methods
- Modularity: The components can be varied independently
- Adaptability: It can accommodate different problem properties and resource constraints
- Convenience: It only requires a pre-trained language model
The ToT framework shows how deliberate search through a tree of "thoughts" can help large language models solve problems that require planning and search, beyond their standard left-to-right generation process.
Code to be (isn't yet) released at: https://github.com/ysymyth/tree-of-thought-llm
// via: https://kagi.com/summarizer/index.html?url=https%3A%2F%2Farx...