PaperPlayer biorxiv bioinformatics

selfRL: Two-Level Self-Supervised Transformer Representation Learning for Link Prediction of Heterogeneous Biomedical Networks


Listen Later

Link to bioRxiv paper:
http://biorxiv.org/cgi/content/short/2020.10.20.347153v1?rss=1
Authors: Wang, X., Yang, Y., Liao, X., Li, K., Li, F., Peng, S.
Abstract:
Predicting potential links in heterogeneous biomedical networks (HBNs) can greatly benefit various important biomedical problem. However, the self-supervised representation learning for link prediction in HBNs has been slightly explored in previous researches. Therefore, this study proposes a two-level self-supervised representation learning, namely selfRL, for link prediction in heterogeneous biomedical networks. The meta path detection-based self-supervised learning task is proposed to learn representation vectors that can capture the global-level structure and semantic feature in HBNs. The vertex entity mask-based self-supervised learning mechanism is designed to enhance local association of vertices. Finally, the representations from two tasks are concatenated to generate high-quality representation vectors. The results of link prediction on six datasets show selfRL outperforms 25 state-of-the-art methods. In particular, selfRL reveals great performance with results close to 1 in terms of AUC and AUPR on the NeoDTI-net dataset. In summary, selfRL provides a general framework that develops self-supervised learning tasks with unlabeled data to obtain promising representations for improving link prediction.
Copy rights belong to original authors. Visit the link for more info
...more
View all episodesView all episodes
Download on the App Store

PaperPlayer biorxiv bioinformaticsBy Multimodal LLC