"Query-Aware Learnable Graph Pooling Tokens as Prompt for LLMs"
Below podcast on this paper is generated with Google's Illuminate.
https://arxiv.org/abs/2501.17549
This paper addresses limitations in how LLMs process graph data for question answering. Existing methods either lack scalability by treating each graph node as a separate input token, or lose information by compressing the entire graph into a single token. The paper introduces a new method to improve graph question answering performance.
This paper proposes Learnable Graph Pooling Tokens (LGPT) and Early Query Fusion. LGPT uses a set of learnable tokens to represent graph information for LLMs. Early Query Fusion incorporates the question context into the graph representation process itself.
-----
๐ LGPT uses learnable tokens to represent graphs, moving beyond single-vector summaries. This allows LLMs to capture more nuanced graph information, improving performance by 4.13% on GraphQA.
๐ Early Query Fusion strategically incorporates query context before graph encoding. This approach focuses graph representation on task-relevant information, unlike late fusion methods.
๐ LGPT offers a practical way to handle graph complexity for LLMs. Using a fixed, small set of tokens ensures scalability, bypassing limitations of node-level graph prompting.
----------
Methods Explored in this Paper ๐ง:
โ The paper introduces Learnable Graph Pooling Tokens (LGPT). LGPT uses learnable parameters as tokens to represent graph information for LLMs.
โ LGPT connects these learnable tokens to all nodes in the graph. Message passing through a Graph Neural Network (GNN) aggregates graph information into these tokens.
โ This method aims to balance detailed node information with global graph context, avoiding information loss seen in single-vector graph representations.
โ The paper also explores Early Query Fusion. This technique integrates the query information before encoding the graph.
โ A virtual query node is created using a text encoder and connected to all graph nodes. GNNs then process the graph incorporating this query node, creating query-aware graph embeddings.
-----
Key Insights ๐ก:
โ Node-level projection of graphs into LLMs is not scalable due to increasing graph size.
โ Graph-level projection into a single vector for LLMs leads to information loss.
โ LGPT offers a balanced approach, reducing information loss compared to single vector methods and improving scalability versus node-level projection.
โ Early Query Fusion enhances graph representation by incorporating query context earlier in the encoding process, leading to more effective graph embeddings.
โ Combining LGPT with Early Query Fusion achieves further performance improvements.
-----
Results ๐:
โ Achieves a 4.13% average performance improvement on the GraphQA benchmark compared to G-Retriever without training the LLM.
โ Early Query Fusion alone improves average performance by 2.88%.
โ LGPT alone improves average performance by 3.87%.


