DDHH: A Decentralized Deep Learning Framework for Large-scale Heterogeneous Networks

被引:3
|
作者
Imran, Mubashir [1 ]
Yin, Hongzhi [1 ]
Chen, Tong [1 ]
Huang, Zi [1 ]
Zhang, Xiangliang [2 ]
Zheng, Kai [3 ]
机构
[1] Univ Queensland, Brisbane, Qld, Australia
[2] King Abdullah Univ Sci & Technol, Jeddah, Saudi Arabia
[3] Univ Elect Sci & Technol China, Chengdu, Peoples R China
关键词
D O I
10.1109/ICDE51399.2021.00196
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Learning vector representations (i.e., embeddings) of nodes for graph-structured information network has attracted vast interest from both industry and academia. Most real-world networks exhibit a complex and heterogeneous format, enclosing high-order relationships and rich semantic information among nodes. However, existing heterogeneous network embedding (HNE) frameworks are commonly designed in a centralized fashion, i.e., all the data storage and learning process take place on a single machine. Hence, those HNE methods show severe performance bottlenecks when handling large-scale networks due to high consumption on memory, storage, and running time. In light of this, to cope with large-scale HNE tasks with strong efficiency and effectiveness guarantee, we propose Decentralized Deep Heterogeneous Hypergraph (DDHH) embedding framework in this paper. In DDHH, we innovatively formulate a large heterogeneous network as a hypergraph, where its hyperedges can connect a set of semantically similar nodes. Our framework then intelligently partitions the heterogeneous network using the identified hyperedges. Then, each resulted subnetwork is assigned to a distributed worker, which employs the deep information maximization theorem to locally learn node embeddings from the partition received. We further devise a novel embedding alignment scheme to precisely project independently learned node embeddings from all subnetworks onto a public vector space, thus allowing for downstream tasks. As shown from our experimental results, DDHH significantly improves the efficiency and accuracy of existing HNE models, and can easily scale up to large-scale heterogeneous networks.
引用
收藏
页码:2033 / 2038
页数:6
相关论文
共 50 条
  • [31] Decentralized iterative learning control for switched large-scale systems
    Gu, Panpan
    Tian, Senping
    Liu, Qian
    TRANSACTIONS OF THE INSTITUTE OF MEASUREMENT AND CONTROL, 2019, 41 (04) : 1045 - 1056
  • [32] Unsupervised Embedding Learning for Large-Scale Heterogeneous Networks Based on Metapath Graph Sampling
    Zhong, Hongwei
    Wang, Mingyang
    Zhang, Xinyue
    ENTROPY, 2023, 25 (02)
  • [33] Large-scale transport simulation by deep learning
    Jie Pan
    Nature Computational Science, 2021, 1 : 306 - 306
  • [34] Tractable large-scale deep reinforcement learning
    Sarang, Nima
    Poullis, Charalambos
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 232
  • [35] Large-scale transport simulation by deep learning
    Pan, Jie
    NATURE COMPUTATIONAL SCIENCE, 2021, 1 (05): : 306 - 306
  • [36] The three pillars of large-scale deep learning
    Hoefler, Torsten
    2021 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2021, : 908 - 908
  • [37] Learning Deep Representation with Large-scale Attributes
    Ouyang, Wanli
    Li, Hongyang
    Zeng, Xingyu
    Wang, Xiaogang
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 1895 - 1903
  • [38] Large-scale Pollen Recognition with Deep Learning
    de Geus, Andre R.
    Barcelos, Celia A. Z.
    Batista, Marcos A.
    da Silva, Sergio F.
    2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,
  • [39] Deep Learning on Large-scale Muticore Clusters
    Sakiyama, Kazumasa
    Kato, Shinpei
    Ishikawa, Yutaka
    Hori, Atsushi
    Monrroy, Abraham
    2018 30TH INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE AND HIGH PERFORMANCE COMPUTING (SBAC-PAD 2018), 2018, : 314 - 321
  • [40] Large-Scale Embedding Learning in Heterogeneous Event Data
    Gui, Huan
    Liu, Jialu
    Tao, Fangbo
    Jiang, Meng
    Norick, Brandon
    Han, Jiawei
    2016 IEEE 16TH INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2016, : 907 - 912