Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
173 tokens/sec
GPT-4o
7 tokens/sec
Gemini 2.5 Pro Pro
46 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

GuideWalk: A Novel Graph-Based Word Embedding for Enhanced Text Classification (2404.18942v2)

Published 25 Apr 2024 in cs.CL, cs.AI, cs.LG, and cs.SI

Abstract: One of the prime problems of computer science and machine learning is to extract information efficiently from large-scale, heterogeneous data. Text data, with its syntax, semantics, and even hidden information content, possesses an exceptional place among the data types in concern. The processing of the text data requires embedding, a method of translating the content of the text to numeric vectors. A correct embedding algorithm is the starting point for obtaining the full information content of the text data. In this work, a new text embedding approach, namely the Guided Transition Probability Matrix (GTPM) model is proposed. The model uses the graph structure of sentences to capture different types of information from text data, such as syntactic, semantic, and hidden content. Using random walks on a weighted word graph, GTPM calculates transition probabilities to derive text embedding vectors. The proposed method is tested with real-world data sets and eight well-known and successful embedding algorithms. GTPM shows significantly better classification performance for binary and multi-class datasets than well-known algorithms. Additionally, the proposed method demonstrates superior robustness, maintaining performance with limited (only $10\%$) training data, showing an $8\%$ decline compared to $15-20\%$ for baseline methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (40)
  1. Bart: Denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 7871–7880, 2020.
  2. Multi-granularity hierarchical attention fusion networks for reading comprehension and question answering. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1705–1714, 2018.
  3. Benchmarking large language models for news summarization. Transactions of the Association for Computational Linguistics, 12:39–57, 2024.
  4. A voice-based digital assistant for intelligent prompting of evidence-based practices during icu rounds. Journal of Biomedical Informatics, 146:104483, 2023.
  5. A local explainability technique for graph neural topic models. Human-Centric Intelligent Systems, pages 1–24, 2024.
  6. A review of natural language processing in contact centre automation. Pattern Analysis and Applications, 26(3):823–846, 2023.
  7. A systematic literature review on phishing email detection using natural language processing techniques. IEEE Access, 10:65703–65727, 2022.
  8. Contextual analysis of social media: The promise and challenge of eliciting context in social media posts with natural language processing. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, pages 337–342, 2020.
  9. Research trends in tourism and hospitality from 1991 to 2020: an integrated approach of corpus linguistics and bibliometrics. Journal of Hospitality and Tourism Insights, 6(2):509–529, 2023.
  10. Natural language processing versus rule-based text analysis: Comparing bert score and readability indices to predict crowdfunding outcomes. Journal of Business Venturing Insights, 16:e00276, 2021.
  11. A survey on classification techniques for opinion mining and sentiment analysis. Artificial intelligence review, 52(3):1495–1545, 2019.
  12. A survey on text classification: From traditional to deep learning. ACM Transactions on Intelligent Systems and Technology (TIST), 13(2):1–41, 2022.
  13. Christopher Bishop. Pattern recognition and machine learning. Springer google schola, 2:5–43, 2006.
  14. Machine learning: Trends, perspectives, and prospects. Science, 349(6245):255–260, 2015.
  15. Understanding bag-of-words model: a statistical framework. International journal of machine learning and cybernetics, 1:43–52, 2010.
  16. Akiko Aizawa. An information-theoretic perspective of tf–idf measures. Information Processing & Management, 39(1):45–65, 2003.
  17. Syntactic n-grams as machine learning features for natural language processing. Expert Systems with Applications, 41(3):853–860, 2014.
  18. Iqbal H Sarker. Machine learning: Algorithms, real-world applications and research directions. SN computer science, 2(3):160, 2021.
  19. Time-evolving text classification with deep neural networks. In IJCAI, volume 18, pages 2241–2247, 2018.
  20. Grounded theory: A practical guide. Sage, 2022.
  21. Imelda T Coyne. Sampling in qualitative research. purposeful and theoretical sampling; merging or clear boundaries? Journal of advanced nursing, 26(3):623–630, 1997.
  22. Qualitative research: A guide to design and implementation. John Wiley & Sons, 2015.
  23. Tpm: Transition probability matrix-graph structural feature based embedding. Kybernetika, 59(2):234–253, 2023.
  24. Deepwalk: Online learning of social representations. In Proceedings of the 20th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 701–710, 2014.
  25. node2vec: Scalable feature learning for networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pages 855–864, 2016.
  26. Discriminative deep random walk for network classification. In Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 1004–1013, 2016.
  27. Harp: Hierarchical representation learning for networks. In Proceedings of the AAAI conference on artificial intelligence, volume 32, 2018.
  28. Albert-László Barabási. Network science. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 371(1987):20120375, 2013.
  29. David MW Powers. Applications and explanations of zipf’s law. In Proceedings of the Joint Conferences on New Methods in Language Processing and Computational Natural Language Learning, pages 151–160, 1998.
  30. Francois Chollet. Deep learning with Python. Simon and Schuster, 2021.
  31. Anonymous walk embeddings. In International Conference on Machine Learning, pages 2186–2195. PMLR, 2018.
  32. Introduction to information retrieval, volume 39. Cambridge University Press Cambridge, 2008.
  33. Yoon Kim. Convolutional neural networks for sentence classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 1746–1751. Association for Computational Linguistics, 2014.
  34. Baseline needs more love: On simple word-embedding-based models and associated pooling mechanisms. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, 2018.
  35. Graph convolutional networks for text classification. In Proceedings of the AAAI conference on artificial intelligence, volume 33, pages 7370–7377, 2019.
  36. Every document owns its structure: Inductive text classification via graph neural networks. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pages 334–339, 2020.
  37. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171–4186, 2019.
  38. Vgcn-bert: augmenting bert with graph embedding for text classification. In Advances in Information Retrieval: 42nd European Conference on IR Research, ECIR 2020, Lisbon, Portugal, April 14–17, 2020, Proceedings, Part I 42, pages 369–382. Springer, 2020.
  39. Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
  40. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com