Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Tag2Vec: Learning Tag Representations in Tag Networks (1905.03041v2)

Published 19 Apr 2019 in cs.SI, cs.LG, physics.soc-ph, and stat.ML

Abstract: Network embedding is a method to learn low-dimensional representation vectors for nodes in complex networks. In real networks, nodes may have multiple tags but existing methods ignore the abundant semantic and hierarchical information of tags. This information is useful to many network applications and usually very stable. In this paper, we propose a tag representation learning model, Tag2Vec, which mixes nodes and tags into a hybrid network. Firstly, for tag networks, we define semantic distance as the proximity between tags and design a novel strategy, parameterized random walk, to generate context with semantic and hierarchical information of tags adaptively. Then, we propose hyperbolic Skip-gram model to express the complex hierarchical structure better with lower output dimensions. We evaluate our model on the NBER U.S. patent dataset and WordNet dataset. The results show that our model can learn tag representations with rich semantic information and it outperforms other baselines.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (6)
  1. Junshan Wang (10 papers)
  2. Zhicong Lu (41 papers)
  3. Guojie Song (39 papers)
  4. Yue Fan (46 papers)
  5. Lun Du (50 papers)
  6. Wei Lin (207 papers)
Citations (18)