Researchers are exploring the integration of language descriptions into graphs, known as text-attributed graphs (TAGs), to enhance model encoding capabilities.
Graph structure learning (GSL) is a crucial technique for improving data utility, and it is highly relevant to efficient TAG learning.
The challenge is to define a reasonable optimization objective for GSL in the era of large language models (LLMs) and design an efficient model architecture for LLM integration.
The proposed Large Language and Tree Assistant (LLaTA) leverages tree-based LLM in-context learning to enhance the understanding of topology and text in order to generate improved graph structures.