Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Deep Bidirectional Transformers for Relation Extraction without Supervision (1911.00313v1)

Published 1 Nov 2019 in cs.LG and stat.ML

Abstract: We present a novel framework to deal with relation extraction tasks in cases where there is complete lack of supervision, either in the form of gold annotations, or relations from a knowledge base. Our approach leverages syntactic parsing and pre-trained word embeddings to extract few but precise relations,which are then used to annotate a larger cor-pus, in a manner identical to distant supervision. The resulting data set is employed to fine tune a pre-trained BERT model in order to perform relation extraction. Empirical evaluation on four data sets from the biomedical domain shows that our method significantly outperforms two simple baselines for unsupervised relation extraction and, even if not using any supervision at all, achieves slightly worse results than the state-of-the-art in three out of four data sets. Importantly, we show that it is possible to successfully fine tune a large pre-trained LLM with noisy data, as op-posed to previous works that rely on gold data for fine tuning.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Yannis Papanikolaou (10 papers)
  2. Ian Roberts (13 papers)
  3. Andrea Pierleoni (8 papers)
Citations (21)

Summary

We haven't generated a summary for this paper yet.