Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
GPT-4o
Gemini 2.5 Pro Pro
o3 Pro
GPT-4.1 Pro
DeepSeek R1 via Azure Pro
2000 character limit reached

GraphER: A Structure-aware Text-to-Graph Model for Entity and Relation Extraction (2404.12491v1)

Published 18 Apr 2024 in cs.CL and cs.AI

Abstract: Information extraction (IE) is an important task in NLP, involving the extraction of named entities and their relationships from unstructured text. In this paper, we propose a novel approach to this task by formulating it as graph structure learning (GSL). By formulating IE as GSL, we enhance the model's ability to dynamically refine and optimize the graph structure during the extraction process. This formulation allows for better interaction and structure-informed decisions for entity and relation prediction, in contrast to previous models that have separate or untied predictions for these tasks. When compared against state-of-the-art baselines on joint entity and relation extraction benchmarks, our model, GraphER, achieves competitive results.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (74)
  1. Uri Alon and Eran Yahav. 2021. On the bottleneck of graph neural networks and its practical implications.
  2. Fastus: A finite-state processor for information extraction from real-world text. In International Joint Conference on Artificial Intelligence.
  3. Nguyen Bach and Sameer Badaskar. 2007. A review of relation extraction.
  4. SciBERT: A pretrained language model for scientific text. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 3615–3620, Hong Kong, China. Association for Computational Linguistics.
  5. Sergey Brin. 1999. Extracting patterns and relations from the world wide web. In The World Wide Web and Databases, pages 172–183, Berlin, Heidelberg. Springer Berlin Heidelberg.
  6. Xavier Carreras and Lluís Màrquez. 2004. Introduction to the CoNLL-2004 shared task: Semantic role labeling. In Proceedings of the Eighth Conference on Computational Natural Language Learning (CoNLL-2004) at HLT-NAACL 2004, pages 89–97, Boston, Massachusetts, USA. Association for Computational Linguistics.
  7. Rich Caruana. 1997. Multitask learning. Machine Learning, 28:41–75.
  8. Measuring and relieving the over-smoothing problem for graph neural networks from the topological view.
  9. Jason P. C. Chiu and Eric Nichols. 2015. Named entity recognition with bidirectional lstm-cnns. Transactions of the Association for Computational Linguistics, 4:357–370.
  10. Bert: Pre-training of deep bidirectional transformers for language understanding. ArXiv, abs/1810.04805.
  11. Kalpit Dixit and Yaser Al-Onaizan. 2019. Span-level model for relation extraction. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pages 5308–5314, Florence, Italy. Association for Computational Linguistics.
  12. Markus Eberts and Adrian Ulges. 2019. Span-based joint entity and relation extraction with transformer pre-training. ArXiv, abs/1909.07755.
  13. LasUIE: Unifying information extraction with latent adaptive structure-aware generative language model. In Advances in Neural Information Processing Systems.
  14. Matthias Fey and Jan Eric Lenssen. 2019. Fast graph representation learning with pytorch geometric.
  15. Learning discrete structures for graph neural networks. In Proceedings of the 36th International Conference on Machine Learning.
  16. Exploring structure-adaptive graph learning for robust semi-supervised classification.
  17. Grammar-constrained decoding for structured NLP tasks without finetuning. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 10932–10952, Singapore. Association for Computational Linguistics.
  18. Neural message passing for quantum chemistry. In Proceedings of the 34th International Conference on Machine Learning, volume 70 of Proceedings of Machine Learning Research, pages 1263–1272. PMLR.
  19. Optimal transport for unsupervised hallucination detection in neural machine translation. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 13766–13784, Toronto, Canada. Association for Computational Linguistics.
  20. Table filling multi-task recurrent neural network for joint entity and relation extraction. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, pages 2537–2547, Osaka, Japan. The COLING 2016 Organizing Committee.
  21. Inductive representation learning on large graphs. In Neural Information Processing Systems.
  22. Is information extraction solved by chatgpt? an analysis of performance, evaluation criteria, robustness and errors. ArXiv, abs/2305.14450.
  23. Span-based joint entity and relation extraction with attention-based span-specific and contextual semantic representations. In Proceedings of the 28th International Conference on Computational Linguistics, pages 88–99, Barcelona, Spain (Online). International Committee on Computational Linguistics.
  24. Jing Jiang and ChengXiang Zhai. 2007. A systematic exploration of the feature space for relation extraction. In Human Language Technologies 2007: The Conference of the North American Chapter of the Association for Computational Linguistics; Proceedings of the Main Conference, pages 113–120, Rochester, New York. Association for Computational Linguistics.
  25. Graph structure learning for robust graph neural networks.
  26. Pure transformers are powerful graph learners. In Advances in Neural Information Processing Systems.
  27. Neural relational inference for interacting systems. arXiv preprint arXiv:1802.04687.
  28. Thomas Kipf and Max Welling. 2016. Semi-supervised classification with graph convolutional networks. ArXiv, abs/1609.02907.
  29. Albert: A lite bert for self-supervised learning of language representations. ArXiv, abs/1909.11942.
  30. Adaptive graph convolutional neural networks.
  31. Gslb: The graph structure learning benchmark.
  32. Neural relation extraction with selective attention over instances. In Annual Meeting of the Association for Computational Linguistics.
  33. Autoregressive structured prediction with language models. In Findings of the Association for Computational Linguistics: EMNLP 2022, pages 993–1005, Abu Dhabi, United Arab Emirates. Association for Computational Linguistics.
  34. Unified structure generation for universal information extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 5755–5772, Dublin, Ireland. Association for Computational Linguistics.
  35. Multi-task identification of entities, relations, and coreference for scientific knowledge graph construction. In Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pages 3219–3232, Brussels, Belgium. Association for Computational Linguistics.
  36. Learning to drop: Robust graph neural network via topological denoising. In Proceedings of the 14th ACM International Conference on Web Search and Data Mining, WSDM ’21, page 779–787, New York, NY, USA. Association for Computing Machinery.
  37. Joint entity and relation extraction based on table labeling using convolutional neural networks. In Proceedings of the Sixth Workshop on Structured Prediction for NLP, pages 11–21, Dublin, Ireland. Association for Computational Linguistics.
  38. SelfCheckGPT: Zero-resource black-box hallucination detection for generative large language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 9004–9017, Singapore. Association for Computational Linguistics.
  39. David Nadeau and Satoshi Sekine. 2007. A survey of named entity recognition and classification. Lingvisticae Investigationes, 30:3–26.
  40. OpenAI. 2023. Gpt-4 technical report.
  41. Graph edit networks. In International Conference on Learning Representations.
  42. Structured prediction as translation between augmented natural languages. In International Conference on Learning Representations.
  43. Deep contextualized word representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long Papers), pages 2227–2237, New Orleans, Louisiana. Association for Computational Linguistics.
  44. Learning and inference over constrained output. In International Joint Conference on Artificial Intelligence.
  45. A novel global feature-oriented relational triple extraction model based on table filling. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 2646–2656, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  46. Ellen Riloff. 1993. Automatically constructing a dictionary for information extraction tasks. In Proceedings of the Eleventh National Conference on Artificial Intelligence, AAAI’93, page 811–816. AAAI Press.
  47. Benjamin Rosenfeld and Ronen Feldman. 2007. Using corpus statistics on entities to improve semi-supervised relation extraction from the web. In Proceedings of the 45th Annual Meeting of the Association of Computational Linguistics, pages 600–607, Prague, Czech Republic. Association for Computational Linguistics.
  48. Dan Roth and Wen-tau Yih. 2004. A linear programming formulation for global inference in natural language tasks. In Proceedings of the Eighth Conference on Computational Natural Language Learning (CoNLL-2004) at HLT-NAACL 2004, pages 1–8, Boston, Massachusetts, USA. Association for Computational Linguistics.
  49. Joint entity and relation extraction with set prediction networks. IEEE transactions on neural networks and learning systems, PP.
  50. Progressive multi-task learning with controlled information flow for joint entity and relation extraction. Proceedings of the AAAI Conference on Artificial Intelligence, 35(15):13851–13859.
  51. Graph structure learning with variational information bottleneck.
  52. Attention is all you need. In Advances in Neural Information Processing Systems, volume 30. Curran Associates, Inc.
  53. Pointer graph networks.
  54. Graph attention networks. In International Conference on Learning Representations.
  55. Entity, relation, and event extraction with contextualized span representations. In Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP), pages 5784–5789, Hong Kong, China. Association for Computational Linguistics.
  56. Revisiting relation extraction in the era of large language models. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 15566–15589, Toronto, Canada. Association for Computational Linguistics.
  57. Ace 2005 multilingual training corpus.
  58. Jue Wang and Wei Lu. 2020. Two are better than one: Joint entity and relation extraction with table-sequence encoders. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1706–1721, Online. Association for Computational Linguistics.
  59. UniRE: A unified label space for entity relation extraction. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pages 220–231, Online. Association for Computational Linguistics.
  60. UTC-IE: A unified token-pair classification architecture for information extraction. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4096–4122, Toronto, Canada. Association for Computational Linguistics.
  61. A partition filter network for joint entity and relation extraction. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 185–197, Online and Punta Cana, Dominican Republic. Association for Computational Linguistics.
  62. Packed levitated marker for entity and relation extraction. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 4904–4917, Dublin, Ireland. Association for Computational Linguistics.
  63. Graph-revised convolutional network.
  64. Global span selection for named entity recognition. In Proceedings of the Workshop on Unimodal and Multimodal Induction of Linguistic Structures (UM-IoS), pages 11–17, Abu Dhabi, United Arab Emirates (Hybrid). Association for Computational Linguistics.
  65. Named entity recognition as structured span prediction. In Proceedings of the Workshop on Unimodal and Multimodal Induction of Linguistic Structures (UM-IoS), pages 1–10, Abu Dhabi, United Arab Emirates (Hybrid). Association for Computational Linguistics.
  66. An autoregressive text-to-graph framework for joint entity and relation extraction. Proceedings of the AAAI Conference on Artificial Intelligence, 38(17):19477–19487.
  67. Kernel methods for relation extraction. In Proceedings of the 2002 Conference on Empirical Methods in Natural Language Processing (EMNLP 2002), pages 71–78. Association for Computational Linguistics.
  68. Kernel methods for relation extraction. In Journal of machine learning research.
  69. Xiang Zhang and Marinka Zitnik. 2020. Gnnguard: Defending graph neural networks against adversarial attacks.
  70. Heterogeneous graph structure learning for graph neural networks. Proceedings of the AAAI Conference on Artificial Intelligence, 35(5):4697–4705.
  71. Robust graph representation learning via neural sparsification. In Proceedings of the 37th International Conference on Machine Learning, volume 119 of Proceedings of Machine Learning Research, pages 11458–11468. PMLR.
  72. Zexuan Zhong and Danqi Chen. 2021. A frustratingly easy approach for entity and relation extraction. In Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 50–61, Online. Association for Computational Linguistics.
  73. Opengsl: A comprehensive benchmark for graph structure learning.
  74. A survey on graph structure learning: Progress and opportunities.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-up Questions

We haven't generated follow-up questions for this paper yet.