Harnessing Large Language Models Over Transformer Models for Detecting Bengali Depressive Social Media Text: A Comprehensive Study (2401.07310v1)
Abstract: In an era where the silent struggle of underdiagnosed depression pervades globally, our research delves into the crucial link between mental health and social media. This work focuses on early detection of depression, particularly in extroverted social media users, using LLMs such as GPT 3.5, GPT 4 and our proposed GPT 3.5 fine-tuned model DepGPT, as well as advanced Deep learning models(LSTM, Bi-LSTM, GRU, BiGRU) and Transformer models(BERT, BanglaBERT, SahajBERT, BanglaBERT-Base). The study categorized Reddit and X datasets into "Depressive" and "Non-Depressive" segments, translated into Bengali by native speakers with expertise in mental health, resulting in the creation of the Bengali Social Media Depressive Dataset (BSMDD). Our work provides full architecture details for each model and a methodical way to assess their performance in Bengali depressive text categorization using zero-shot and few-shot learning techniques. Our work demonstrates the superiority of SahajBERT and Bi-LSTM with FastText embeddings in their respective domains also tackles explainability issues with transformer models and emphasizes the effectiveness of LLMs, especially DepGPT, demonstrating flexibility and competence in a range of learning contexts. According to the experiment results, the proposed model, DepGPT, outperformed not only Alpaca Lora 7B in zero-shot and few-shot scenarios but also every other model, achieving a near-perfect accuracy of 0.9796 and an F1-score of 0.9804, high recall, and exceptional precision. Although competitive, GPT-3.5 Turbo and Alpaca Lora 7B show relatively poorer effectiveness in zero-shot and few-shot situations. The work emphasizes the effectiveness and flexibility of LLMs in a variety of linguistic circumstances, providing insightful information about the complex field of depression detection models.
- Justifying arabic text sentiment analysis using explainable ai (xai): Lasik surgeries case study. Information, 13(11):536, 2022.
- Charu C. Aggarwal. Neural networks and deep learning, 2013. URL https://link.springer.com/book/10.1007/978-3-319-94463-0. Accessed: 2023-12-18.
- An attention-based approach to detect emotion from tweets. In 2020 3rd International Conference on Computer and Informatics Engineering (IC2IE), pp. 182–187. IEEE, 2020.
- Mega: Multilingual evaluation of generative ai. arXiv.org, 03 2023. URL https://arxiv.org/abs/2303.12528.
- Uzh_pandas at simpletext@ clef-2023: Alpaca lora 7b and lens model selection for scientific literature simplification. arXiv.org, 2023.
- Zero-and few-shot prompting with llms: A comparative study with fine-tuned models for bangla sentiment analysis. arXiv e-prints, pp. arXiv–2308, 2023.
- bangla stemmer. bangla-stemmer, 12 2023. URL https://pypi.org/project/bangla-stemmer/. Accessed: 2023-12-18.
- A deep learning based sentiment analysis on bang-lish disclosure. In 2021 National Computing Colleges Conference (NCCC), pp. 1–6. IEEE, 2021.
- Banglabert: Language model pretraining and benchmarks for low-resource language understanding evaluation in bangla. arXiv preprint arXiv:2101.00204, 2021.
- Bangla text sentiment analysis using supervised machine learning with extended lexicon dictionary. Natural Language Processing Research, 1(3-4):34–45, 2021.
- Detecting depression in social media posts using machine learning. In K. C. Santosh and Ravindra S. Hegadi (eds.), Recent Trends in Image Processing and Pattern Recognition, pp. 716–725, Singapore, 2019. Springer Singapore. ISBN 978-981-13-9187-3.
- bnltk. bnltk, 12 2023. URL https://pypi.org/project/bnltk/. Accessed: 2023-12-18.
- Deep learning-based depression detection from social media: Comparative evaluation of ml and transformer techniques. Electronics, 12(21), 2023. ISSN 2079-9292. doi: 10.3390/electronics12214396. URL https://www.mdpi.com/2079-9292/12/21/4396. Accessed: 2023-12-18.
- Language models are few-shot learners. Advances in neural information processing systems, 33:1877–1901, 2020. Accessed: 2023-12-18.
- A lexicon-based approach to examine depression detection in social media: the case of twitter and university community. Humanities and Social Sciences Communications, 9, 09 2022. doi: 10.1057/s41599-022-01313-2.
- Large language models for text classification: From zero-shot learning to fine-tuning. Open Science Foundation, 2023.
- Llm-empowered chatbots for psychiatrist and patient simulation: Application and evaluation. arXiv preprint arXiv:2305.13614, 2023.
- Learning phrase representations using rnn encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078, 2014.
- Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv preprint arXiv:1412.3555, 2014.
- Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555, 2020.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
- Enhancing psychological counseling with large language model: A multifaceted decision-support system for non-professionals. arXiv preprint arXiv:2308.15192, 2023.
- Predicting depression via social media. In Proceedings of the Association for the Advancement of Artificial Intelligence (AAAI). Association for the Advancement of Artificial Intelligence, 07 2013. URL https://www.semanticscholar.org/paper/Predicting-Depression-via-Social-Media-Choudhury-Gamon/e992de7c53579bd1f23c80be17ca3f6692593ea6.
- Sentiment analysis of twitter data using machine learning approaches and semantic analysis. In 2014 Seventh international conference on contemporary computing (IC3), pp. 437–442. IEEE, 2014.
- GBDResults. Institute for health metrics and evaluation, 2023. URL https://vizhub.healthdata.org/gbd-results/. Accessed: 2023-12-18.
- Framewise phoneme classification with bidirectional lstm and other neural network architectures. Neural networks, 18(5-6):602–610, 2005.
- A transformer based approach to detect suicidal ideation using pre-trained language models. In 2020 23rd international conference on computer and information technology (ICCIT), pp. 1–5. IEEE, 2020.
- Machine learning-based sentiment analysis for twitter accounts. Mathematical and computational applications, 23(1):11, 2018.
- Md. Arid Hasan. Zero- and few-shot prompting with llms: A comparative study with fine-tuned models for bangla sentiment analysis, August 2023. URL https://arxiv.org/abs/2308.10783. Accessed: 2023-12-18.
- Long short-term memory. Neural computation, 9(8):1735–1780, 1997.
- Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685, 2021.
- Analyze detection depression in social media twitter using bidirectional encoder representations from transformers. Journal of Information System Research (JOSH), 3(4):476–482, July 2022. doi: 10.47065/josh.v3i4.1885. URL https://doi.org/10.47065/josh.v3i4.1885. Accessed: 2023-12-18.
- Rethinking large language models in mental health applications. arXiv preprint arXiv:2311.11267, 2023.
- Benllmeval: A comprehensive evaluation into the potentials and pitfalls of large language models on bengali nlp. arXiv preprint arXiv:2309.13173, 2023.
- Albert: A lite bert for self-supervised learning of language representations. arXiv preprint arXiv:1909.11942, 2019.
- Ehsan Latif. Fine-tuning chatgpt for automatic scoring, October 2023. URL https://arxiv.org/abs/2310.10072. Accessed: 2023-12-18.
- Anonymity vs. familiarity: Self-disclosure and privacy in social virtual reality. In Proceedings of the 26th ACM Symposium on Virtual Reality Software and Technology, VRST ’20, New York, NY, USA, 2020. Association for Computing Machinery. ISBN 9781450376198. doi: 10.1145/3385956.3418967. URL https://doi.org/10.1145/3385956.3418967.
- MayoClinic2022. Depression (major depressive disorder), 2022. URL https://www.mayoclinic.org/diseases-conditions/depression/symptoms-causes/syc-20356007. Accessed: 2023-6-26.
- Depressed people detection from bangla social media status using lstm and cnn approach. Journal of Engineering Advancements, 2(01):41–47, 2021.
- NLTK. Nltk :: nltk.tokenize package, 12 2023. URL https://www.nltk.org/api/nltk.tokenize.html. Accessed: 2023-12-18.
- OpenAI. Gpt-4 technical report, March 2023. URL https://arxiv.org/abs/2303.08774. Accessed: 2023-12-18.
- World Health Organization. Depressive disorder (depression), March 2023. URL https://www.who.int/news-room/fact-sheets/detail/depression.
- Reddit. Reddit dataset: r/depression and r/suicidewatch, March 2023. URL https://www.kaggle.com/datasets/xavrig/reddit-dataset-rdepression-and-rsuicidewatch. Accessed: 2023-12-18.
- sahajBERT. bangla-stemmer, 12 2023. URL https://huggingface.co/neuropark/sahajBERT. Accessed: 2023-12-18.
- Sagor Sarker. Banglabert: Bengali mask language model for bengali language understading. textsIGitHub, 2020.
- Sentiment140. Sentiment140 dataset with 1.6 million tweets, March 2023. URL https://www.kaggle.com/datasets/kazanova/sentiment140. Accessed: 2023-12-18.
- Early depression detection from social network using deep learning techniques. In 2020 IEEE Region 10 Symposium (TENSYMP), pp. 823–826. IEEE, 2020a.
- A hybridized feature extraction approach to suicidal ideation detection from social media post. In 2020 IEEE Region 10 Symposium (TENSYMP), pp. 985–988. IEEE, 2020b.
- Depression analysis from social media data in bangla language using long short term memory (lstm) recurrent neural network technique. In 2019 International Conference on Computer, Communication, Chemical, Materials and Electronic Engineering (IC4ME2), pp. 1–4. IEEE, 2019.
- Mental-llm: Leveraging large language models for mental health prediction via online text data. arXiv preprint arXiv:2307.14385, 2023.
- Mentalllama: Interpretable mental health analysis on social media with large language models. arXiv preprint arXiv:2309.13567, 2023.
- Junjie Ye. A comprehensive capability analysis of gpt-3 and gpt-3.5 series models, March 2023. URL https://arxiv.org/abs/2303.10420. Accessed: 2023-12-18.
- Ahmadul Karim Chowdhury (1 paper)
- Md. Saidur Rahman Sujon (1 paper)
- Md. Shirajus Salekin Shafi (1 paper)
- Tasin Ahmmad (1 paper)
- Sifat Ahmed (1 paper)
- Khan Md Hasib (8 papers)
- Faisal Muhammad Shah (14 papers)