3.8 Proceedings Paper

High-order Proximity Preserving Information Network Hashing

Publisher

ASSOC COMPUTING MACHINERY
DOI: 10.1145/3219819.3220034

Keywords

Information Network Hashing; Matrix Factorization; Hamming Subspace Learning

Funding

  1. National Natural Science Foundation of China [61502077, 61631005, 61532018, 61502324]
  2. Fundamental Research Funds for the Central Universities [ZYGX2016J087]
  3. Australian Research Council [FT130100746, DP180100106, LP150100671]
  4. National Research Foundation
  5. Prime Minister's Office, Singapore under its Campus for Research Excellence and Technological Enterprise (CREATE) programme
  6. Alibaba Innovative Research program

Ask authors/readers for more resources

Information network embedding is an effective way for efficient graph analytics. However, it still faces with computational challenges in problems such as link prediction and node recommendation, particularly with increasing scale of networks. Hashing is a promising approach for accelerating these problems by orders of magnitude. However, no prior studies have been focused on seeking binary codes for information networks to preserve high-order proximity. Since matrix factorization (MF) unifies and outperforms several well-known embedding methods with high-order proximity preserved, we propose a MF-based Information Network Hashing (INH-MF) algorithm, to learn binary codes which can preserve high-order proximity. We also suggest Hamming subspace learning, which only updates partial binary codes each time, to scale up INH-MF. We finally evaluate INH-MF on four real-world information network datasets with respect to the tasks of node classification and node recommendation. The results demonstrate that INH-MF can perform significantly better than competing learning to hash baselines in both tasks, and surprisingly outperforms network embedding methods, including DeepWalk, LINE and NetMF, in the task of node recommendation. The source code of INH-MF is available online(1).

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

3.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available