search
HomeTechnology peripheralsAIThe Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.

In order for large language models (LLM) to give full play to their capabilities, effective prompt design is essential. For this reason, the emerging field of prompt engineering has even emerged.

Among various prompt design solutions, Chain of Thinking (CoT) has attracted the attention of many researchers and users with its powerful reasoning capabilities. Based on its improved CoT-SC and further Thinking Tree (ToT) ) has also received a lot of attention.

Recently, a research team from ETH Zurich, Cledar and Warsaw University of Technology proposed a further idea: Maps of Mind (GoT). The ability to think from chains to trees to graphs and build a reasoning process for LLM has been continuously improved, and researchers have also proven this through experiments. They also released their own implementation of the GoT framework.

The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.

Research paper: https://arxiv.org/pdf/2308.09687v2.pdf

Official implementation: https://github.com/spcl /graph-of-thoughts

Paper Overview

Large-scale language models are becoming the dominant technology in the artificial intelligence world. The models that have developed rapidly in recent years are mainly based on decoder-only Transformer variants, such as GPT, PaLM or LLaMA.

When solving different LLM tasks, prompt engineering design is a method that can efficiently utilize resources. Simply put, a description of the task is included in the input sent to the LLM. If the task can be described in a suitable form, LLM can solve it with the help of its autoregressive token-based mechanism for generating text. Such prompts may contain example tasks with answers (few-shot prompt design, also known as contextual learning (ICL)), or they may not contain example tasks at all (zero-shot prompt design). Research and applications in recent years have shown that, This mechanism can be used to solve many types of tasks involving mathematics, common sense, or symbolic reasoning.

Chain of Thought (CoT) is a method used to design prompts, that is, in addition to the input of the task, the prompt In addition to output and output, it also includes intermediate steps of reasoning (intermediate thinking). Research shows that CoT can greatly improve the ability of LLM, enabling it to solve some difficult problems without any model updates.

Some researchers have also improved CoT proposes a method of using CoT to achieve self-consistency (CoT-SC); this solution is to generate multiple CoTs and then select the best result.

Recently, some researchers have further proposed thinking Tree (ToT), which models the LLM inference process through trees. This allows the model to use different thinking paths and can provide new functions, such as backtracking the inference process based on bad results. Unfortunately Unfortunately, since the ToT method imposes a strict tree structure on the thinking process, it will greatly limit prompt's reasoning ability. For more details, please refer to the article on this site "Thinking, thinking, thinking without stopping, Thinking Tree ToT" Military training "LLM" .

This research team from ETH Zurich, Cledar and Warsaw University of Technology believes that if the thinking of LLM can be constructed into an arbitrary graph structure, then it can provide prompt capabilities bring significant improvements. They said the idea was inspired by a variety of phenomena, such as the way humans reason, the structure of the brain, and the way algorithms execute.

Humans don’t behave like CoTs when thinking Following only one chain of thinking, instead of trying many different paths like ToT, will form a more complex network of thinking. For example, a person may first explore one chain of thinking, then backtrack and explore another, and then You may realize that an idea from the previous chain can be combined with the current chain to learn from each other and arrive at a new solution. Similarly, the brain will form complex networks, showing graph-like patterns, such as cyclic patterns. Algorithms The pattern of the network will also be revealed during execution, which can often be expressed as a directed acyclic graph.

The researchers said that if this corresponding graph-enabled transformation is used in LLM thinking, it is expected to create A powerful way to design prompts, but this transformation cannot be naturally expressed through CoT or ToT.

Then they observed: If the reasoning process of LLM is modeled as a graph, then it can be naturally Achieving these and many other thinking transformations. Based on this observation, they proposed GoT/Graph of Thoughts, a method that can enhance the capabilities of LLM through network-form reasoning.

In GoT, an LLM thought is modeled as a vertex, and the dependencies between vertices are modeled as edges. Using GoT, arbitrary ideas can be aggregated by constructing vertices with more than one input edge. Overall, the graph abstraction method used by GoT can seamlessly generalize CoT and ToT to more complex thinking models, and this process does not require updating the model.

However, to actually implement GoT, there are some design challenges that need to be solved. For example, what is the best graph structure for different tasks? What is the best approach to convergent thinking in order to maximize accuracy and minimize cost?

To answer these questions and more, these researchers designed a modular architecture for implementing GoT. The design has two major highlights.

First, it can achieve fine-grained control of each thinking. This gives users full control over conversations with LLM and the use of advanced thought transformations, such as combining the two most promising thoughts from an ongoing inference to get a new one.

The second is that this architecture is designed with scalability in mind - it can be seamlessly extended for new thinking transformations, reasoning modes (i.e., mind maps) and LLM models. This allows users to use GoT to quickly prototype new design ideas for prompt while experimenting with different models such as GPT-3.5, GPT-4 or Llama-2.

The researchers also demonstrated some use cases of GoT (sorting, keyword counting of summaries, set operations, document merging), and they also detailed how to implement them using a graph-based paradigm. They experimentally evaluate GoT, demonstrating its advantages over other state-of-the-art methods.

Researchers say that overall, GoT is particularly suitable for tasks that can be naturally decomposed into smaller subtasks, and these subtasks can be solved separately and then merged into a final solution. In this regard, GoT performs better than other solutions. For example, on the sorting task, GoT is better than CoT and ToT by about 70% and 62% respectively, while the cost is more than 31% lower than ToT.

Table 1 gives a qualitative comparison between GoT and other prompt design solutions. GoT is the only solution that can implement any graph-based thinking transformation (such as aggregation) within a prompt, thereby encompassing all previous solutions.

The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.

They also have another contribution, which is to propose a new evaluation metric - the volume of a thought, which can be used to evaluate prompt design strategies. The goal of using this metric, the researchers say, is to better understand the differences between prompt design options.

For a given thought v, the capacity of v refers to the number of LLM thoughts, based on which the user can use directed edges to obtain v. Intuitively, these are all the LLM ideas that are expected to contribute to v.

The author has shown through research that by integrating thinking transformation technologies such as aggregation, GoT can make the thinking capacity significantly larger than other solutions.

GoT Framework

The following is a detailed introduction to the GoT framework. The schematic diagram is shown in Figure 1, which also provides schematic diagrams of other prompt design strategies.

The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.

In mathematical form, GoT can be modeled as a tuple (G, T, E, R), where G is the LLM inference process (i.e. all the LLM thoughts and their relations), T is the possible thought transformations, E is the evaluator function used to obtain the thought score, and R is the ranking function used to select the most relevant thought.

Inference process

Here, the inference process is modeled as a directed graph G = (V, E), where V is a set of vertices, E ⊆ V × V is a set of edges. G is directed, so edges are subsets of ordered vertex pairs E ⊆ V × V . A vertex contains a solution to the current problem, whether it is the initial, intermediate, or final problem. The exact form of this thinking depends on the use case; it might be a piece of text (in a writing task) or a sequence of values ​​(in a sorting task). The directed edge (t_1, t_2) represents the way the thought t_2 is constructed by using t_1 as "direct input", i.e. by explicitly instructing the LLM to use t_1 to generate t_2.

In some use cases, graph nodes belong to different categories. For example, in a writing task, some vertices model the plan for writing a text segment, while other nodes model the actual text segment. In this case, GoT adopts a heterogeneous graph G = (V, E, c) to model LLM inference, where c maps vertices V to their respective classes C (in the above case, C = {plan, par} ). This way, any vertex v can model different aspects of reasoning.

So G is associated with the LLM inference process. To facilitate this process, the user can use Thought Shift on G. An example of this kind of transformation: fuse the thought with the highest score so far into a new one. Another example is looping a thought in order to strengthen it. Note that these transformations strictly extend the set of transformations available in CoT, CoT-SC or ToT.

Thinking transformation

Thanks to the use of graph-based models for reasoning, GoT can achieve new thinking transformations. Researchers call this a graph-enabled transformation. For example, in a writing task multiple input articles can be combined into a coherent summary. When sorting, multiple sorted numeric subarrays can be combined into a final sorted array. Figure 2 gives an example of aggregation and generation.

The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.

Mathematically speaking, each such transformation can be modeled as T (G, p_θ), where G = (V, E) is the reflection of the current inference State diagram, p_θ is the LLM used. T usually modifies G by adding new vertices and their incoming edges. Then we have G′ = T (G, p_θ) = (V′, E′), where V′ = (V ∪ {V^ }) \ {V^−} and E′ = (E ∪ {E^ }) \{E^−}. V^ and E^ are new vertices and edges injected into G, which model new thinking and their dependencies respectively.

To maximize the expressive power of GoT, users can also delete thoughts by specifying the corresponding vertices and edges to be deleted (V^− and E^− respectively). Here, it is the user's responsibility to ensure that the sets V^ , E^ , V^− and E^− have consistent transformations (for example, the user will not try to delete non-existent vertices). This enables seamless integration of prompt solutions, where the user can remove non-improved parts of the reasoning to save space in the context.

The specific form of T and how it affects G depends on the specific transformation. The following will first introduce in detail the thinking transformations enabled by the main graphs, and then describe how GoT includes the transformations of previous solutions. Unless otherwise stated, V^− = E^− = ∅.

Aggregation Transformation: Users can use GoT to aggregate any thoughts into new thoughts to learn from each other's strengths. Here's a look at the basic form that only creates a new vertex: V^ = {v^ } and E^ = {(v_1, v^ ), ...,(v_k, v^ )}, where v_1, ..., v_k is the k thoughts being merged. More generally, this enables the aggregation of reasoning paths, i.e. longer chains of thoughts, rather than just individual thoughts. Using a graph model, an aggregation transformation can be easily implemented: by adding outgoing edges from vertices v_1, ..., v_k that model the last thought in several chains, to point to a single thought v^ that combines these chains.

Refining transformation: Another thinking transformation is to refine the current thinking v by modifying the content: V^ = {} and E^ = {(v, v)}. This loop in the diagram represents an iterative version of thinking that has the same connections as the original thought.

Generate transformation: Finally, the user can also generate one or more new thoughts based on an existing single thought v. Included in this category are similar reasoning steps from earlier schemes such as ToT or CoT-SC. Mathematically speaking, there Is the answer good enough? The score is modeled as a general function E (v, G, p_θ), where v is the thought being evaluated. In order to make E as general as possible, the whole process of reasoning (G) is also used in E, since in some assessment scenarios the scores may be related to other thinking.

GoT can also be ranked. The researchers used the function R (G, p_θ, h) to model, where h specifies the number of top-ranked thoughts in G to be returned by R. While the exact form of R depends on the use case, a simple yet effective approach that is most commonly used is to return the h thoughts with the highest scores, i.e. v_1, ..., v_h = R (G, p_θ, h). The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.
The exact form of E and R depends on the use case.

System architecture and expansion capabilities

GoT consists of a set of interactive modules, see Figure 3 (blue part). These modules are Prompter (prepare messages for LLM), Parser (parser, extract information in LLM replies), Scoring module (validate LLM replies and score), Controller (controller, coordinate the entire reasoning process and decide how to proceed reasoning). Controller contains two other important components: graph of operations (GoO) and graph reasoning state (GRS). A GoO is a static structure that specifies the graph decomposition for a given task, i.e. it specifies the transformations applied to LLM thinking and their order and dependencies. The GRS is a dynamic structure that maintains the state of the ongoing LLM inference process (the history of its thinking and its states).

Use case examples

The researcher describes some use cases of GoT, including sorting, set operations, keyword counting, and document merging ; Figure 4 below is an example of graph decomposition in GoT's sorting use case. We will not introduce the use cases in detail here, please refer to the original paper for details.

The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.
The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.

Latency vs. Capacity Tradeoff

Latency (how long it takes to reach a given final thought in the mind map The trade-off between hop count) and capacity is also very important, and researchers have shown that GoT is better than previous prompt designs on this trade-off. This paper defines a new metric - thought capacity, which is the number of previous LLM thoughts that can influence a given thought t. Mathematically, the capacity of thought t is the number of thoughts that have paths between t and t in the thought map. The researchers assumed that the cost of outputting a single thought is O (1) and fixed the total cost of each prompt solution as Θ(n).

The structure of various schemes is as follows. CoT-SC consists of k independent chains originating from a single starting thought. ToT is a complete k-ary tree. In GoT, a complete k-ary tree is added to its leaf nodes, with a "mirror" k-ary tree - the size is the same but the edges are reversed.

See Table 2 for detailed analysis. CoT has a larger capacity, up to N, but it also has a high latency cost of N. CoT-SC reduces latency by a factor of k (corresponding to its branching factor), but at the same time its capacity is reduced by a factor of k. The latency of ToT is log_k N, but the capacity is also low. GoT is the only solution that can achieve low latency log_k N and high capacity N. GoT is able to do this because it leverages thought aggregation, allowing it to arrive at the final thought from any other intermediate thoughts in the graph decomposition.

The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.

Evaluation

The researchers demonstrated the advantages of GoT over other solutions through experiments. The key comparison is between GoT and ToT, because the performance of ToT is already better than other solutions. Of course, they still did some experiments with IO, CoT and CoT-SC.

Figures 5 (sorting), 6 (set intersection), 7 (keyword counting), and 8 (document merging) show the experimental results.

The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.
The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.
The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.
The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.
# Overall, after experimental evaluation On all benchmarks, the output quality of GoT is better than that of ToT, and it also achieves lower inference cost.

The above is the detailed content of The Thought Chain CoT evolved into the Thought Map GoT, and a prompt engineering technology better than the Thought Tree was born.. For more information, please follow other related articles on the PHP Chinese website!

Statement
This article is reproduced at:机器之心. If there is any infringement, please contact admin@php.cn delete
How to Run LLM Locally Using LM Studio? - Analytics VidhyaHow to Run LLM Locally Using LM Studio? - Analytics VidhyaApr 19, 2025 am 11:38 AM

Running large language models at home with ease: LM Studio User Guide In recent years, advances in software and hardware have made it possible to run large language models (LLMs) on personal computers. LM Studio is an excellent tool to make this process easy and convenient. This article will dive into how to run LLM locally using LM Studio, covering key steps, potential challenges, and the benefits of having LLM locally. Whether you are a tech enthusiast or are curious about the latest AI technologies, this guide will provide valuable insights and practical tips. Let's get started! Overview Understand the basic requirements for running LLM locally. Set up LM Studi on your computer

Guy Peri Helps Flavor McCormick's Future Through Data TransformationGuy Peri Helps Flavor McCormick's Future Through Data TransformationApr 19, 2025 am 11:35 AM

Guy Peri is McCormick’s Chief Information and Digital Officer. Though only seven months into his role, Peri is rapidly advancing a comprehensive transformation of the company’s digital capabilities. His career-long focus on data and analytics informs

What is the Chain of Emotion in Prompt Engineering? - Analytics VidhyaWhat is the Chain of Emotion in Prompt Engineering? - Analytics VidhyaApr 19, 2025 am 11:33 AM

Introduction Artificial intelligence (AI) is evolving to understand not just words, but also emotions, responding with a human touch. This sophisticated interaction is crucial in the rapidly advancing field of AI and natural language processing. Th

12 Best AI Tools for Data Science Workflow - Analytics Vidhya12 Best AI Tools for Data Science Workflow - Analytics VidhyaApr 19, 2025 am 11:31 AM

Introduction In today's data-centric world, leveraging advanced AI technologies is crucial for businesses seeking a competitive edge and enhanced efficiency. A range of powerful tools empowers data scientists, analysts, and developers to build, depl

AV Byte: OpenAI's GPT-4o Mini and Other AI InnovationsAV Byte: OpenAI's GPT-4o Mini and Other AI InnovationsApr 19, 2025 am 11:30 AM

This week's AI landscape exploded with groundbreaking releases from industry giants like OpenAI, Mistral AI, NVIDIA, DeepSeek, and Hugging Face. These new models promise increased power, affordability, and accessibility, fueled by advancements in tr

Perplexity's Android App Is Infested With Security Flaws, Report FindsPerplexity's Android App Is Infested With Security Flaws, Report FindsApr 19, 2025 am 11:24 AM

But the company’s Android app, which offers not only search capabilities but also acts as an AI assistant, is riddled with a host of security issues that could expose its users to data theft, account takeovers and impersonation attacks from malicious

Everyone's Getting Better At Using AI: Thoughts On Vibe CodingEveryone's Getting Better At Using AI: Thoughts On Vibe CodingApr 19, 2025 am 11:17 AM

You can look at what’s happening in conferences and at trade shows. You can ask engineers what they’re doing, or consult with a CEO. Everywhere you look, things are changing at breakneck speed. Engineers, and Non-Engineers What’s the difference be

Rocket Launch Simulation and Analysis using RocketPy - Analytics VidhyaRocket Launch Simulation and Analysis using RocketPy - Analytics VidhyaApr 19, 2025 am 11:12 AM

Simulate Rocket Launches with RocketPy: A Comprehensive Guide This article guides you through simulating high-power rocket launches using RocketPy, a powerful Python library. We'll cover everything from defining rocket components to analyzing simula

See all articles

Hot AI Tools

Undresser.AI Undress

Undresser.AI Undress

AI-powered app for creating realistic nude photos

AI Clothes Remover

AI Clothes Remover

Online AI tool for removing clothes from photos.

Undress AI Tool

Undress AI Tool

Undress images for free

Clothoff.io

Clothoff.io

AI clothes remover

Video Face Swap

Video Face Swap

Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Tools

SublimeText3 Linux new version

SublimeText3 Linux new version

SublimeText3 Linux latest version

Dreamweaver Mac version

Dreamweaver Mac version

Visual web development tools

ZendStudio 13.5.1 Mac

ZendStudio 13.5.1 Mac

Powerful PHP integrated development environment

SecLists

SecLists

SecLists is the ultimate security tester's companion. It is a collection of various types of lists that are frequently used during security assessments, all in one place. SecLists helps make security testing more efficient and productive by conveniently providing all the lists a security tester might need. List types include usernames, passwords, URLs, fuzzing payloads, sensitive data patterns, web shells, and more. The tester can simply pull this repository onto a new test machine and he will have access to every type of list he needs.

SublimeText3 Mac version

SublimeText3 Mac version

God-level code editing software (SublimeText3)