Abstract: Representation learning on text-attributed graphs (TAGs) is vital for real-world applications, as they combine semantic textual and contextual structural information. Research in this field generally consist of two main perspectives: local-level encoding and global-level aggregating, respectively refer to textual node information unification ($e.g.$, using Language Models) and structure-augmented modeling ($e.g.$, using Graph Neural Networks). Most existing works focus on combining different information levels but overlook the interconnections, $i.e.$, the contextual textual information among nodes, which provides semantic insights to bridge local and global levels. In this paper, we propose GraphBridge, a $multi-granularity integration$ framework that bridges local and global perspectives by leveraging contextual textual information, enhancing fine-grained understanding of TAGs. Besides, to tackle scalability and efficiency challenges, we introduce a graph-aware token reduction module. Extensive experiments across various models and datasets show that our method achieves state-of-the-art performance, while our graph-aware token reduction module significantly enhances efficiency and solves scalability issues. Codes are available at https://anonymous.4open.science/r/GraphBridge-13E0
Paper Type: Long
Research Area: Machine Learning for NLP
Research Area Keywords: graph-based methods
Contribution Types: Publicly available software and/or pre-trained models
Languages Studied: English
Submission Number: 1989
Loading