Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
80 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Cross-relation Cross-bag Attention for Distantly-supervised Relation Extraction (1812.10604v1)

Published 27 Dec 2018 in cs.CL and cs.LG

Abstract: Distant supervision leverages knowledge bases to automatically label instances, thus allowing us to train relation extractor without human annotations. However, the generated training data typically contain massive noise, and may result in poor performances with the vanilla supervised learning. In this paper, we propose to conduct multi-instance learning with a novel Cross-relation Cross-bag Selective Attention (C$2$SA), which leads to noise-robust training for distant supervised relation extractor. Specifically, we employ the sentence-level selective attention to reduce the effect of noisy or mismatched sentences, while the correlation among relations were captured to improve the quality of attention weights. Moreover, instead of treating all entity-pairs equally, we try to pay more attention to entity-pairs with a higher quality. Similarly, we adopt the selective attention mechanism to achieve this goal. Experiments with two types of relation extractor demonstrate the superiority of the proposed approach over the state-of-the-art, while further ablation studies verify our intuitions and demonstrate the effectiveness of our proposed two techniques.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (8)
  1. Yujin Yuan (6 papers)
  2. Liyuan Liu (49 papers)
  3. Siliang Tang (116 papers)
  4. Zhongfei Zhang (44 papers)
  5. Yueting Zhuang (164 papers)
  6. Shiliang Pu (106 papers)
  7. Fei Wu (317 papers)
  8. Xiang Ren (194 papers)
Citations (63)