Papers
Topics
Authors
Recent
Assistant
AI Research Assistant
Well-researched responses based on relevant abstracts and paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses.
Gemini 2.5 Flash
Gemini 2.5 Flash 75 tok/s
Gemini 2.5 Pro 40 tok/s Pro
GPT-5 Medium 36 tok/s Pro
GPT-5 High 27 tok/s Pro
GPT-4o 97 tok/s Pro
Kimi K2 196 tok/s Pro
GPT OSS 120B 455 tok/s Pro
Claude Sonnet 4.5 34 tok/s Pro
2000 character limit reached

Enhancing Language Models for Financial Relation Extraction with Named Entities and Part-of-Speech (2405.06665v1)

Published 2 May 2024 in cs.CL, cs.IR, and cs.LG

Abstract: The Financial Relation Extraction (FinRE) task involves identifying the entities and their relation, given a piece of financial statement/text. To solve this FinRE problem, we propose a simple but effective strategy that improves the performance of pre-trained LLMs by augmenting them with Named Entity Recognition (NER) and Part-Of-Speech (POS), as well as different approaches to combine these information. Experiments on a financial relations dataset show promising results and highlights the benefits of incorporating NER and POS in existing models. Our dataset and codes are available at https://github.com/kwanhui/FinRelExtract.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Unsupervised cross-lingual representation learning at scale. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp.  8440–8451, 2020.
  2. BERT: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pp.  4171–4186, 2019.
  3. A survey on personalized itinerary recommendation: From optimisation to deep learning. Applied Soft Computing, pp.  111200, 2024.
  4. Kpi-bert: A joint named entity recognition and relation extraction model for financial reports. In Proceedings of the 26th International Conference on Pattern Recognition (ICPR), pp.  606–612. IEEE, 2022.
  5. A french corpus and annotation schema for named entity recognition and relation extraction of financial news. In Proceedings of the Twelfth Language Resources and Evaluation Conference, pp.  2293–2299, 2020.
  6. Spanbert: Improving pre-training by representing and predicting spans. Transactions of the association for computational linguistics, 8:64–77, 2020.
  7. Refind: Relation extraction financial dataset. arXiv preprint arXiv:2305.18322, 2023.
  8. Albert: A lite bert for self-supervised learning of language representations. In Proceedings of the International Conference on Learning Representations, 2019.
  9. A transformer-based framework for poi-level social post geolocation. In European Conference on Information Retrieval, pp.  588–604. Springer, 2023.
  10. Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
  11. Deep neural approaches to relation triplets extraction: A comprehensive survey. Cognitive Computation, 13:1215–1232, 2021.
  12. A simple approach to financial relation classification with pre-trained language models. In Proceedings of the 4th Workshop on Knowledge Discovery from Unstructured Data in Financial Services (KDF), 2023.
  13. Distilbert, a distilled version of bert: smaller, faster, cheaper and lighter. arXiv preprint arXiv:1910.01108, 2019.
  14. Finred: A dataset for relation extraction in financial domain. In Companion Proceedings of the Web Conference 2022, pp.  595–597, 2022.
  15. imetre: Incorporating markers of entity types for relation extraction. arXiv preprint arXiv:2307.00132, 2023.
  16. Deep neural network-based relation extraction: an overview. Neural Computing and Applications, pp.  1–21, 2022.
  17. Enriching pre-trained language model with entity information for relation classification. In Proceedings of the 28th ACM international conference on information and knowledge management, pp.  2361–2364, 2019.
  18. Finbert: A pretrained language model for financial communications. arXiv preprint arXiv:2006.08097, 2020.
  19. Xlnet: Generalized autoregressive pretraining for language understanding. Advances in neural information processing systems, 32, 2019.
  20. Relation classification via recurrent neural network. arXiv preprint arXiv:1508.01006, 2015.
  21. Position-aware attention and supervised data improve slot filling. In Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing, pp.  35–45, 2017.

Summary

We haven't generated a summary for this paper yet.

Lightbulb Streamline Icon: https://streamlinehq.com

Continue Learning

We haven't generated follow-up questions for this paper yet.

Authors (2)

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 2 posts and received 0 likes.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube