Rethinking link prediction: A multi-scale graph masked autoencoder
In link prediction tasks, Graph Self-Supervised Learning (GSSL) has enormous potential for tackling the fundamental issue of sparse labels in real-world graph data. However, existing mainstream methods based on contrastive and generative approaches do not simultaneously consider hierarchical informa...
Gespeichert in:
| Veröffentlicht in: | Neurocomputing (Amsterdam) Jg. 664; S. 132115 |
|---|---|
| Hauptverfasser: | , , , , , |
| Format: | Journal Article |
| Sprache: | Englisch |
| Veröffentlicht: |
Elsevier B.V
01.02.2026
|
| Schlagworte: | |
| ISSN: | 0925-2312 |
| Online-Zugang: | Volltext |
| Tags: |
Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
|
| Zusammenfassung: | In link prediction tasks, Graph Self-Supervised Learning (GSSL) has enormous potential for tackling the fundamental issue of sparse labels in real-world graph data. However, existing mainstream methods based on contrastive and generative approaches do not simultaneously consider hierarchical information from both local neighborhood information and global structural information. Therefore, we revisit these two approaches from a novel perspective and propose a multi-scale graph masked autoencoder (MS-GMAE) to enable the model to learn stable representations that capture aggregated local neighborhood and global structural information, thereby achieving precise predictions of both local and global links in link prediction tasks. Specifically, we propose a Cross-Correlation Decoder and a latent global representation prediction strategy. The Cross-Correlation Decoder utilizes multi-layer embedding representations of two nodes to decode and predict links between nodes, enabling the model to accurately perceive local neighborhoods. Latent Global Representation Prediction forces the reconstruction of embeddings that capture global structural information in the latent space, enhancing the model’s ability to capture global structural information. Additionally, we reconstruct stable contextual representations for nodes, enabling the model to learn more stable graph representation information. Results show that compared to 17 state-of-the-art baselines across five datasets, our model achieves state-of-the-art performance on four datasets, with the highest performance improvement reaching 9.83 %.
•A multi-scale autoencoder learns hierarchical features for link prediction.•Cross-correlation decoding captures local patterns, mitigating overfitting.•Predicting latent global representations captures long-range dependencies.•Experimental results demonstrate superior performance over 17 baseline models. |
|---|---|
| ISSN: | 0925-2312 |
| DOI: | 10.1016/j.neucom.2025.132115 |