4.5 Article

Lime: Low-Cost and Incremental Learning for Dynamic Heterogeneous Information Networks

Journal

IEEE TRANSACTIONS ON COMPUTERS
Volume 71, Issue 3, Pages 628-642

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TC.2021.3057082

Keywords

Task analysis; Social networking (online); Heuristic algorithms; Computational modeling; Optimization; Semantics; Recurrent neural networks; Network representation learning; heterogeneous information networks; incremental learning; memory optimization

Funding

  1. National Science Foundation of China [62002007, U20B2053]
  2. Key Research and Development Project of Hebei Province [20310101D]
  3. U.K. EPSRC [EP/T01461X/1, EP/T021985/1]
  4. UK Royal Society International Collaboration Grant
  5. NERC [NE/P017134/1, SKLSDE-2020ZX-12]
  6. Australian Research Council [DP200103494]
  7. NSF ONR [N00014-18-1-2009]
  8. NSF [III-1763325, III-1909323, SaTC-1930941]
  9. CAAI-Huawei MindSpore Open Fund
  10. Australian Research Council [DP200103494] Funding Source: Australian Research Council

Ask authors/readers for more resources

Understanding the interconnected relationships of large-scale information networks like social, scholar and Internet of Things networks is crucial for tasks like recommendation and fraud detection. Lime, a new approach for modeling dynamic and heterogeneous information networks, significantly reduces memory resources and computational time by utilizing semantic relationships among network nodes and employing optimization strategies with recursive neural networks. By efficiently adapting to evolving networks through incremental learning, Lime demonstrates comparable performance for network representation tasks with greatly reduced processing time and memory footprint.
Understanding the interconnected relationships of large-scale information networks like social, scholar and Internet of Things networks is vital for tasks like recommendation and fraud detection. The vast majority of the real-world networks are inherently heterogeneous and dynamic, containing many different types of nodes and edges and can change drastically over time. The dynamicity and heterogeneity make it extremely challenging to reason about the network structure. Unfortunately, existing approaches are inadequate in modeling real-life dynamical networks as they either have strong assumption of a given stochastic process or fail to capture the heterogeneity of network structure, and they all require extensive computational resources. We introduce Lime, a better approach for modeling dynamic and heterogeneous information networks. Lime is designed to extract high-quality network representation with significantly lower memory resources and computational time over the state-of-the-arts. Unlike prior work that uses a vector to encode each network node, we exploit the semantic relationships among network nodes to encode multiple nodes with similar semantics in shared vectors. By using many fewer node vectors, our approach significantly reduces the required memory space for encoding large-scale networks. To effectively trade information sharing for reduced memory footprint, we employ the recursive neural network (RsNN) with carefully designed optimization strategies to explore the node semantics in a novel cuboid space. We then go further by showing, for the first time, how an effective incremental learning approach can be developed - with the help of RsNN, our cuboid structure, and a set of novel optimization techniques - to allow a learning framework to quickly and efficiently adapt to a constantly evolving network. We evaluate Lime by applying it to three representative network-based tasks, node classification, node clustering and anomaly detection, performing on three large-scale datasets. We compare Lime against eleven prior state-of-the-art approaches for learning network representation. Our extensive experiments demonstrate that Lime not only reduces the memory footprint by over 80 percent and the processing time over 2x when learning network representation but also delivers comparable performance for downstream processing tasks. We show that our incremental learning method can boost the learning time by up to 20x without compromising the quality of the learned network representation.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available