Skip to content

PatrickAllenCooper/tree-of-thoughts

 
 

Repository files navigation

Multi-Modality

Tree of Thoughts Banner

Discord Twitter LinkedIn Facebook Reddit Hacker News Pinterest WhatsApp

Paper link Author's implementation

Introduction

Tree of Thoughts (ToT) is a powerful and flexible algorithm that significantly advances model reasoning by up to 70%. This plug-and-play version allows you to connect your own models and experience superintelligence!

🔥 Updates

  • Langchain TOT
  • MonteCarlo
  • A* Search
  • Best First Search

Coming soon!

  • Iterative Depth Search
  • Any search algorithms you like? Open an issue 😊

Basic Prompts

No complex implementations, just pass in one of these prompts to your model. Head over to prompts.txt.

"Three experts with exceptional logical thinking skills are collaboratively answering a question using the tree of thoughts method. Each expert will share their thought process in detail, taking into account the previous thoughts of others and admitting any errors. They will iteratively refine and expand upon each other's ideas, giving credit where it's due. The process continues until a conclusive answer is found. Organize the entire response in a markdown table format. The question is..."

Getting Started

Method 1

Clone this repository:

git clone https://github.com/kyegomez/tree-of-thoughts
cd tree-of-thoughts
python3 -m pip install -r requirements.txt
cd tree_of_thoughts

Set OpenAI key in an environment file:

  1. Create a file called .env.
  2. Get your OpenAI key and input it inside the .env file as OPENAI_API_KEY='SK-YOUR KEY'.

Then go to montecarlo_example.py and fill in your API key!

For much improved performance, provide custom prompt shots in the generate thoughts and generate states.

In the examples folder, we have other examples for Hugging Face Transformers + Hugging Face Pipelines.

Method 2

Alternatively, you can use pip to install Tree of Thoughts:

pip install tree-of-thoughts

Create a Python script (e.g., example.py) and import the necessary classes:

import os
from tree_of_thoughts import OpenAILanguageModel, MonteCarloTreeofThoughts

api_model = "gpt-3.5-turbo"
model = OpenAILanguageModel(api_key='api key', api_model=api_model)

# Initialize the MonteCarloTreeofThoughts class with the model
tree_of_thoughts = MonteCarloTreeofThoughts(model)

# To reproduce the same results from the Tree of Thoughts paper, or even better,
# craft a one-shot chain of thought prompt for your task below:

initial_prompt =  """
Input: 2 8 8 14
Possible next steps:
2 + 8 = 10 (left: 8 10 14)
8 / 2 = 4 (left: 4 8 14)
14 + 2 = 16 (left: 8 8 16)
2 * 8 = 16 (left: 8 14 16)
8 - 2 = 6 (left: 6 8 14)
14 - 8 = 6 (left: 2 6 8)
14 /  2 = 7 (left: 7 8 8)
14 - 2 = 12 (left: 8 8 12)
Input: use 4 numbers and basic arithmetic operations (+-*/) to obtain 24 in 1 equation


Possible next steps:
"""
num_thoughts = 1
max_steps = 3
max_states = 4
pruning_threshold = 0.5

solution = tree_of_thoughts.solve(
    initial_prompt=initial_prompt,
    num_thoughts=num_thoughts, 
    max_steps=max_steps, 
    max_states=max_states, 
    pruning_threshold=pruning_threshold,
)

print(f"Solution: {solution}")

Or integrate your own custom language model:

class CustomLanguageModel(AbstractLanguageModel):
    def __init__(self, model):
        self.model = model

    def generate_thoughts(self, state, k):
        # implement the thought generation logic using self.model
        pass

    def evaluate_states(self, states):
        # implement state evaluation logic using self.model
        pass

Run the example script.

🌟 Features:

  • General problem-solving framework for language models
  • Supports both breadth-first search (BFS) and depth-first search (DFS) algorithms
  • Easy integration with popular language models like OpenAI and Hugging Face
  • Extensible and adaptable to different problem properties and resource constraints

Algorithmic Pseudocode

  1. Define the thought decomposition based on the problem properties.
  2. Create a thought generator function G(pθ, s, k) with two strategies: a. Sample i.i.d. thoughts from a CoT prompt. b. Propose thoughts sequentially using a "propose prompt".
  3. Create a state evaluator function V(pθ, S) with two strategies: a. Value each state independently. b. Vote across states.
  4. Choose a search algorithm (BFS or DFS) based on the tree structure.
  5. Implement the chosen search algorithm.
  6. Execute the chosen search algorithm with the input problem, thought generator, state evaluator, and other required parameters.

Usage Examples

OpenAI API

To use Tree of Thoughts with OpenAI's API, create a custom model class that inherits from AbstractLanguageModel and implements the required methods using OpenAI's API. Then, create an instance of the TreeOfThoughts class with the custom model and the desired search algorithm ('BFS' or 'DFS').

Hugging Face Transformers

To run Hugging Face Transformers with Tree of Thoughts:

git clone https://github.com/kyegomez/tree-of-thoughts
cd tree-of-thoughts
python3 huggingfaceExample.py
from tree_of_thoughts import HuggingLanguageModel

model_name = "gpt2"
model_tokenizer = "your tokenizer"

huggingface_model = HuggingLanguageModel(model_name, model_tokenizer)
class HuggingLanguageModel(AbstractLanguageModel):
    def __init__(self, model_name):
        self.model = AutoModelForCausalLM.from_pretrained(model_name)
        self.tokenizer = AutoTokenizer.from_pretrained(model_name)

    def generate_thoughts(self, state, k):
        state_text = ' '.join(state)
        prompt = f"Given the current state of reasoning: '{state_text}', generate {k} coherent thoughts to achieve the reasoning process:"

        inputs = self.tokenizer(prompt, return_tensors="pt")
        outputs = self.model.generate(**inputs, num_return_sequences=k)
        thoughts = [self.tokenizer.decode(output, skip_special_tokens=True) for output in outputs]

        return thoughts

    def evaluate_states(self, states, initial_prompt):
        state_values = {}
        for state in states:
            state_text = ' '.join(state)
            prompt = f"Given the current state of reasoning: '{state_text}', pessimistically evaluate its value as a float between 0 and 1 based on its potential to achieve {initial_prompt}"

            inputs = self.tokenizer(prompt, return_tensors="pt")
            outputs = self.model.generate(**inputs, num_return_sequences=1)
            value_text = self.tokenizer.decode(outputs[0], skip_special_tokens=True)

            try:
                value = float(value_text)
            except ValueError:
                value = 0  # Assign a default value if the conversion fails

            state_values[state] = value

        return state_values

Contributing

This algorithm is still in its infancy, but its potential remains unimaginable. Let's advance the reasoning of AI together under this banner.

Share With Your Network

You can easily share this repository by clicking on the following buttons:

Twitter LinkedIn

For Instagram, while it doesn't directly support sharing web links, you can share the screenshot of our project and the link in your caption or bio. You can download the project screenshot by clicking the image below:

Tree of Thoughts

We greatly appreciate any help in spreading the word about our project. Thank you for your support!

Roadmap

  • Resilient Prompting: Teach model how to think rather than what to think.
  • Add pruning threshold management for precise bad state cutoff.
  • Evaluating each thought as soon as it's generated, then evaluating a chain of thoughts or the state of thoughts by averaging out the values of each thought evaluation.
  • Add Traversal method, which will encapsulate the run of either DFS or BFS under the hood so that the issue of different args is solved (from @ivanzhovannik).
  • Add Delay between generating solutions and generating values.
  • Dynamic and adaptive parameters like max steps, num_thoughts, max_states, and value threshold that shift depending on the complexity of the user objective.
  • Add Rejected reasoning metadata (thought, state, reasoning_on_state) into generate solutions.
  • And more! Feel free to suggest any other ideas. This algorithm is very young, and its potential is limitless.
  • Chain of Thought Hub Evaluation tests!

Documentation:

Search Algorithms in Tree of Thoughts

The Tree of Thoughts library supports a variety of search algorithms that can be employed for different problem-solving contexts. Here's a brief overview of each search algorithm along with their primary benefits and use-cases.

1. Breadth-First Search (BFS)

BFS explores all the nodes at the present depth before going on to the nodes at the next depth level. It is an excellent choice when the depth of the tree is relatively small, and solutions are spread out evenly.

Benefits:

  • It guarantees to find the shallowest goal, i.e., the solution with fewer steps.
  • It is a simple and straightforward algorithm for traversing trees or graphs.

Use-cases:

  • Ideal for problems where the depth of the tree/graph is not very large.
  • Useful when the goal is close to the root.

2. Depth-First Search (DFS)

DFS explores as far as possible along each branch before backing up. It is suitable when the tree depth is significant, and solutions are located deep in the tree.

Benefits:

  • It uses less memory compared to BFS as it needs to store only a single path from the root to a leaf node, along with remaining unexplored sibling nodes for each node on the path.
  • It can explore deeper solutions that are not accessible with BFS.

Use-cases:

  • It is often used in simulations due to its more aggressive (deeper) search.
  • Ideal for searching through a big search space.

3. Best-First Search

Best-First Search uses an evaluation function to decide which adjacent node is most promising and then explores. It is suitable for problems where we have some heuristic information about the distance from the current state to the goal.

Benefits:

  • It can provide a more efficient solution by using heuristics.
  • It does not explore unnecessary paths, thus saving resources.

Use-cases:

  • Suitable for a large dataset where the goal node's location is unknown.
  • Ideal for problems where some heuristic can guide the search to the goal.

4. A* Search

A* Search finds the least-cost path from the given initial node to one goal node (out of one or more possible goals). It uses a best-first search and finds the least-cost path to a goal.

Benefits:

  • It is complete, optimal, optimally efficient, and uses heuristics to guide itself.
  • A* balances between BFS and DFS and avoids expanding paths that are already expensive.

Use-cases:

  • Widely used in pathfinding and graph traversal, the process of plotting an efficiently directed path between multiple points.
  • Suitable for games, mapping apps, and routing paths for vehicles where we need an optimal solution.

5. Monte Carlo Tree Search (MCTS)

MCTS uses random sampling of the search space and uses the results to guide the search. It is best when the search space is vast and cannot be completely traversed.

Benefits:

  • It can provide good solutions for extremely complex problems with a large search space, where traditional methods fail.
  • It uses statistical analysis of the results for decision making, which can handle the uncertainty and variability in the problem.

Use-cases:

  • Suitable for "perfect information" games, which are games where players have complete knowledge of all events and states.
  • Also useful in real-time video games and other domains where the decision-making time is limited.

input_problem (str):

The initial problem statement or prompt for which the Tree of Thoughts algorithm will generate a solution.

num_thoughts (int, default=5):

The number of thoughts to generate at each state. A higher value of k will result in more thoughts being generated, potentially leading to a more diverse set of solutions. However, increasing k may also increase the computational complexity and time required to find a solution.

max_steps (int, default=3):

The maximum depth of the search tree. A higher value of T allows the algorithm to explore deeper states, potentially leading to better solutions. However, increasing T may also increase the computational complexity and time required to find a solution.

max_states (int, default=5):

The branching factor of the search tree, which determines the maximum number of child nodes for each parent node. A higher value of b allows the algorithm to explore more states, potentially leading to better solutions. However, increasing b may also increase the computational complexity and time required to find a solution.

value_threshold (float, default=0.5):

The value threshold for pruning states. States with a value below this threshold will be discarded, reducing the search space. A higher value of vth will result in a more aggressive pruning strategy, potentially speeding up the search process. However, setting vth too high may cause the algorithm to discard promising states, leading to suboptimal solutions.

LangchainTOT class

LangchainTOT is the main class you'll interact with. It acts as a wrapper for the Large Language Model and allows you to set a problem description, add thoughts, and check the validity of these thoughts using a specified checker.

Initialization

You initialize a LangchainTOT object with an optional problem description and a checker class:

problem_description = """
3,*,*,2|1,*,3,*|*,1,*,3|4,*,*,1
- This is a 4x4 Sudoku puzzle.
- The * represents a cell to be filled.
- The | character separates rows.
- At each step, replace one or more * with digits 1-4.
- There must be no duplicate digits in any row, column or 2x2 subgrid.
- Keep the known digits from previous valid thoughts in place.
- Each thought can be a partial or the final solution.
""".strip()

langchain_tot = LangchainTOT(problem_description=problem_description, checker_class=lambda: my_checker)

If you want to change the problem description or checker class later, you can use the set_problem_description and set_checker_class methods.

Adding thoughts

Once you have your LangchainTOT object, you can add thoughts to it. A thought is a string representing a possible solution or step towards a solution. You can add thoughts using the add_thought method:

langchain_tot.add_thought("3,*,*,2|1,*,3,*|*,1,*,3|4,*,*,1")

Checking thoughts

Once you've added one or more thoughts, you can check their validity using the check_thoughts method:

print(langchain_tot.check_thoughts())

This method will return a ThoughtValidity value representing whether the latest thought is a final valid solution (VALID_FINAL), an intermediate valid step (VALID_INTERMEDIATE), or invalid (INVALID).

MyChecker class

MyChecker is a class for creating custom checkers. It inherits from ToTChecker and must implement the evaluate method.

Initialization

You initialize a MyChecker object with a validation function:

my_checker = MyChecker(validate_fn=lambda p, t: validate_sudoku(p, t, sudoku_solution))

Custom validation function

The validation function is a callable that takes a problem description and a tuple of thoughts and returns a ThoughtValidity value. It defines how to check the validity of thoughts for a specific problem.

In this code, we're using the validate_sudoku function as our validation function. This function takes a problem description, a tuple of thoughts, and a solution string, and returns the validity of the last thought based on whether it matches the solution and whether it's a possible step towards the solution.

The validate_sudoku function is specific to 4x4 Sudoku puzzles and you'll need to create your own validation function for different types of problems.

Acknowledgements

Thanks to: Shunyu Yao Princeton University, Dian Yu Google DeepMind, Jeffrey Zhao, Google DeepMind, Izhak Shafran Google DeepMind, Thomas L. Griffiths, Princeton University, Yuan Cao Google DeepMind, Karthik Narasimha, Princeton University for sharing this amazing work with the world!

And, thanks to Phil Wang or Lucidrains for inspiring me to devote myself to open source AI Research

About

Plug in and Play Implementation of Tree of Thoughts: Deliberate Problem Solving with Large Language Models that Elevates Model Reasoning by atleast 70%

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 96.1%
  • Jupyter Notebook 3.1%
  • Shell 0.8%