BIGSAGE: unsupervised inductive representation learning of graph via bi-attended sampling and global-biased aggregating
Abstract: Different kinds of representation learning techniques on graph have shown significant effect in downstream machine learning tasks. Recently, in order to inductively learn representations for graph structures that is unobservable during training, a general framework with sampling and aggregating (GraphSAGE) was proposed by Hamilton and Ying and had been proved more efficient than transductive methods on fileds like transfer learning or evolving dataset. However, GraphSAGE is uncapable of selective neighbor sampling and lack of memory of known nodes that've been trained. To address these problems, we present an unsupervised method that samples neighborhood information attended by co-occurring structures and optimizes a trainable global bias as a representation expectation for each node in the given graph. Experiments show that our approach outperforms the state-of-the-art inductive and unsupervised methods for representation learning on graphs.
Keywords: network embedding, unsupervised learning, inductive learning
TL;DR: For unsupervised and inductive network embedding, we propose a novel approach to explore most relevant neighbors and preserve previously learnt knowledge of nodes by utilizing bi-attention architecture and introducing global bias, respectively
Data: [PPI](https://paperswithcode.com/dataset/ppi), [Pubmed](https://paperswithcode.com/dataset/pubmed), [Reddit](https://paperswithcode.com/dataset/reddit)
4 Replies
Loading