Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
158 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
45 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Using Zero-shot Prompting in the Automatic Creation and Expansion of Topic Taxonomies for Tagging Retail Banking Transactions (2401.06790v2)

Published 8 Jan 2024 in cs.CL and cs.AI

Abstract: This work presents an unsupervised method for automatically constructing and expanding topic taxonomies using instruction-based fine-tuned LLMs. We apply topic modeling and keyword extraction techniques to create initial topic taxonomies and LLMs to post-process the resulting terms and create a hierarchy. To expand an existing taxonomy with new terms, we use zero-shot prompting to find out where to add new nodes, which, to our knowledge, is the first work to present such an approach to taxonomy tasks. We use the resulting taxonomies to assign tags that characterize merchants from a retail bank dataset. To evaluate our work, we asked 12 volunteers to answer a two-part form in which we first assessed the quality of the taxonomies created and then the tags assigned to merchants based on that taxonomy. The evaluation revealed a coherence rate exceeding 90% for the chosen taxonomies. The taxonomies' expansion with LLMs also showed exciting results for parent node prediction, with an f1-score above 70% in our taxonomies.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (20)
  1. Latent dirichlet allocation. Journal of machine Learning research, 3(Jan):993–1022.
  2. Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901.
  3. Hierarchical classification of financial transactions through context-fusion of transformer-based embeddings and taxonomy-aware attention layer. In Anais do II Brazilian Workshop on Artificial Intelligence in Finance, pages 13–24. SBC.
  4. Yake! keyword extraction from single documents using multiple local features. Information Sciences, 509:257–289.
  5. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805.
  6. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
  7. Taxocom: Topic taxonomy completion with hierarchical discovery of novel topic clusters. In Proceedings of the ACM Web Conference 2022, pages 2819–2829.
  8. Unleashing infinite-length input capacity for large-scale language models with self-controlled memory system. arXiv preprint arXiv:2304.13343.
  9. Russe’2020: Findings of the first taxonomy enrichment task for the russian language. arXiv preprint arXiv:2005.11176.
  10. OpenAI. 2023. Gpt-4 technical report.
  11. Training language models to follow instructions with human feedback.
  12. Octavian Popescu and Carlo Strapparava. 2015. Semeval 2015, task 7: Diachronic text evaluation. In Proceedings of the 9th International Workshop on Semantic Evaluation (SemEval 2015), pages 870–878.
  13. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9.
  14. Learning syntactic patterns for automatic hypernym discovery. Advances in neural information processing systems, 17.
  15. Low-resource taxonomy enrichment with pretrained language models. In Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pages 2747–2758.
  16. Llama: Open and efficient foundation language models. arXiv preprint arXiv:2302.13971.
  17. Attention is all you need. Advances in neural information processing systems, 30.
  18. Making use of external company data to improve the classification of bank transactions. In Advanced Data Mining and Applications: 13th International Conference, ADMA 2017, Singapore, November 5–6, 2017, Proceedings 13, pages 767–780. Springer.
  19. Rethinking lda: Why priors matter. Advances in neural information processing systems, 22.
  20. Taxogen: Unsupervised topic taxonomy construction by adaptive term embedding and clustering. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, pages 2701–2709.

Summary

We haven't generated a summary for this paper yet.