SOS-1K: A Fine-grained Suicide Risk Classification Dataset for Chinese Social Media Analysis (2404.12659v1)
Abstract: In the social media, users frequently express personal emotions, a subset of which may indicate potential suicidal tendencies. The implicit and varied forms of expression in internet language complicate accurate and rapid identification of suicidal intent on social media, thus creating challenges for timely intervention efforts. The development of deep learning models for suicide risk detection is a promising solution, but there is a notable lack of relevant datasets, especially in the Chinese context. To address this gap, this study presents a Chinese social media dataset designed for fine-grained suicide risk classification, focusing on indicators such as expressions of suicide intent, methods of suicide, and urgency of timing. Seven pre-trained models were evaluated in two tasks: high and low suicide risk, and fine-grained suicide risk classification on a level of 0 to 10. In our experiments, deep learning models show good performance in distinguishing between high and low suicide risk, with the best model achieving an F1 score of 88.39%. However, the results for fine-grained suicide risk classification were still unsatisfactory, with an weighted F1 score of 50.89%. To address the issues of data imbalance and limited dataset size, we investigated both traditional and advanced, LLM based data augmentation techniques, demonstrating that data augmentation can enhance model performance by up to 4.65% points in F1-score. Notably, the Chinese MentalBERT model, which was pre-trained on psychological domain data, shows superior performance in both tasks. This study provides valuable insights for automatic identification of suicidal individuals, facilitating timely psychological intervention on social media platforms. The source code and data are publicly available.
- W. H. Organization et al., “World mental health report: transforming mental health for all,” 2022.
- B. Keles, N. McCrae, and A. Grealish, “A systematic review: the influence of social media on depression, anxiety and psychological distress in adolescents,” International journal of adolescence and youth, vol. 25, no. 1, pp. 79–93, 2020.
- P. Chen, Y. Qian, Z. Huang, C. Zhao, Z. Liu, and B. Yang, “Negative emotional characteristics of weibo “tree hole”’ users,” Zhongguo Xinliweisheng Zazhi, vol. 5, pp. 437–444, 2020.
- B. X. Yang, L. Xia, L. Liu, W. Nie, Q. Liu, X. Y. Li, M. Q. Ao, X. Q. Wang, Y. D. Xie, Z. Liu et al., “A suicide monitoring and crisis intervention strategy based on knowledge graph technology for “tree hole” microblog users in China,” Frontiers in psychology, vol. 12, p. 674481, 2021.
- Z. Yimeng and L. Kun, “AI gives potential suicides pause for thought,” China Daily, 2020.
- G. Coppersmith, R. Leary, P. Crutchley, and A. Fine, “Natural language processing of social media as screening for suicide risk,” Biomedical informatics insights, vol. 10, p. 1178222618792860, 2018.
- G. Fu, C. Song, J. Li, Y. Ma, P. Chen, R. Wang, B. X. Yang, and Z. Huang, “Distant supervision for mental health management in social media: suicide risk classification system development study,” Journal of medical internet research, vol. 23, no. 8, p. e26119, 2021.
- S. Renjith, A. Abraham, S. B. Jyothi, L. Chandran, and J. Thomson, “An ensemble deep learning technique for detecting suicidal ideation from posts in social media platforms,” Journal of King Saud University-Computer and Information Sciences, vol. 34, no. 10, pp. 9564–9575, 2022.
- J. Gorai and D. K. Shaw, “A BERT-encoded ensembled CNN model for suicide risk identification in social media posts,” Neural Computing and Applications, pp. 1–16, 2024.
- R. Wang, B. X. Yang, Y. Ma, P. Wang, Q. Yu, X. Zong, Z. Huang, S. Ma, L. Hu, K. Hwang et al., “Medical-level suicide risk analysis: A novel standard and evaluation model,” IEEE Internet of Things Journal, vol. 8, no. 23, pp. 16 825–16 834, 2021.
- C. Shorten, T. M. Khoshgoftaar, and B. Furht, “Text data augmentation for deep learning,” Journal of big Data, vol. 8, no. 1, p. 101, 2021.
- J. Chen, D. Tam, C. Raffel, M. Bansal, and D. Yang, “An Empirical Survey of Data Augmentation for Limited Data Learning in NLP,” Transactions of the Association for Computational Linguistics, vol. 11, pp. 191–211, 03 2023. [Online]. Available: https://doi.org/10.1162/tacl_a_00542
- W. X. Zhao, K. Zhou, J. Li, T. Tang, X. Wang, Y. Hou, Y. Min, B. Zhang, J. Zhang, Z. Dong et al., “A survey of large language models,” arXiv preprint arXiv:2303.18223, 2023.
- J. Kaddour, J. Harris, M. Mozes, H. Bradley, R. Raileanu, and R. McHardy, “Challenges and applications of large language models,” arXiv preprint arXiv:2307.10169, 2023.
- T. He, G. Fu, Y. Yu, F. Wang, J. Li, Q. Zhao, C. Song, H. Qi, D. Luo, H. Zou et al., “Towards a psychological generalist AI: A survey of current applications of large language models and future prospects,” arXiv preprint arXiv:2312.04578, 2023.
- J. M. Liu, D. Li, H. Cao, T. Ren, Z. Liao, and J. Wu, “Chatcounselor: A large language models for mental health support,” arXiv preprint arXiv:2309.15461, 2023.
- G. Fu, Q. Zhao, J. Li, D. Luo, C. Song, W. Zhai, S. Liu, F. Wang, Y. Wang, L. Cheng et al., “Enhancing psychological counseling with large language model: A multifaceted decision-support system for non-professionals,” arXiv preprint arXiv:2308.15192, 2023.
- W. Zhou, L. C. Prater, E. V. Goldstein, S. J. Mooney et al., “Identifying rare circumstances preceding female firearm suicides: validating a large language model approach,” JMIR mental health, vol. 10, no. 1, p. e49359, 2023.
- J. W. Ayers, A. Poliak, M. Dredze, E. C. Leas, Z. Zhu, J. B. Kelley, D. J. Faix, A. M. Goodman, C. A. Longhurst, M. Hogarth et al., “Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum,” JAMA internal medicine, vol. 183, no. 6, pp. 589–596, 2023.
- H. Qi, Q. Zhao, J. Li, C. Song, W. Zhai, L. Dan, S. Liu, Y. J. Yu, F. Wang, H. Zou et al., “Supervised learning and large language model benchmarks on mental health datasets: Cognitive distortions and suicidal risks in chinese social media,” 2023.
- W. Zhai, H. Qi, Q. Zhao, J. Li, Z. Wang, H. Wang, B. X. Yang, and G. Fu, “Chinese MentalBERT: Domain-adaptive pre-training on social media for chinese mental health text analysis,” arXiv preprint arXiv:2402.09151, 2024.
- Y. Xiao, K. Bi, P. S.-F. Yip, J. Cerel, T. T. Brown, Y. Peng, J. Pathak, and J. J. Mann, “Decoding suicide decedent profiles and signs of suicidal intent using latent class analysis,” JAMA psychiatry, 2024.
- Z. Huang, Q. Hu, J. Gu, J. Yang, Y. Feng, and G. Wang, “Web-based intelligent agents for suicide monitoring and early warning,” China Digital Medicine, vol. 14, no. 3, pp. 2–6, 2019.
- J. D. M.-W. C. Kenton and L. K. Toutanova, “BERT: Pre-training of deep bidirectional transformers for language understanding,” in Proceedings of naacL-HLT, vol. 1, 2019, p. 2.
- A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” Advances in neural information processing systems, vol. 30, 2017.
- Y. Liu, M. Ott, N. Goyal, J. Du, M. Joshi, D. Chen, O. Levy, M. Lewis, L. Zettlemoyer, and V. Stoyanov, “RoBERTa: A robustly optimized bert pretraining approach,” arXiv preprint arXiv:1907.11692, 2019.
- K. Clark, M.-T. Luong, Q. V. Le, and C. D. Manning, “ELECTRA: Pre-training text encoders as discriminators rather than generators,” arXiv preprint arXiv:2003.10555, 2020.
- Y. Cui, W. Che, T. Liu, B. Qin, and Z. Yang, “Pre-training with whole word masking for Chinese BERT,” IEEE/ACM Transactions on Audio, Speech, and Language Processing, vol. 29, pp. 3504–3514, 2021.
- J. Wei, X. Ren, X. Li, W. Huang, Y. Liao, Y. Wang, J. Lin, X. Jiang, X. Chen, and Q. Liu, “Nezha: Neural contextualized representation for chinese language understanding,” arXiv preprint arXiv:1909.00204, 2019.
- Y. Sun, S. Wang, S. Feng, S. Ding, C. Pang, J. Shang, J. Liu, X. Chen, Y. Zhao, Y. Lu et al., “ERNIE 3.0: Large-scale knowledge enhanced pre-training for language understanding and generation,” arXiv preprint arXiv:2107.02137, 2021.
- J. Chen, D. Tam, C. Raffel, M. Bansal, and D. Yang, “An empirical survey of data augmentation for limited data learning in NLP,” Transactions of the Association for Computational Linguistics, vol. 11, pp. 191–211, 2023.
- O. Kolomiyets, S. Bethard, and M.-F. Moens, “Model-portability experiments for textual temporal analysis,” in Proceedings of the 49th annual meeting of the association for computational linguistics: human language technologies, vol. 2. ACL; East Stroudsburg, PA, 2011, pp. 271–276.
- M. Aiken and M. Park, “The efficacy of round-trip translation for MT evaluation,” Translation Journal, vol. 14, no. 1, pp. 1–10, 2010.
- G. Salton, A. Wong, and C.-S. Yang, “A vector space model for automatic indexing,” Communications of the ACM, vol. 18, no. 11, pp. 613–620, 1975.
- H. Y. X. Hai Liang Wang. (2017) Chinese synonym toolkit: Synonyms. [Online]. Available: https://github.com/chatopera/Synonyms
- Baidu, Inc., “Baidu translate api,” https://api.fanyi.baidu.com/, accessed: 2024-02-21.
- J. Achiam, S. Adler, S. Agarwal, L. Ahmad, I. Akkaya, F. L. Aleman, D. Almeida, J. Altenschmidt, S. Altman, S. Anadkat et al., “GPT-4 technical report,” arXiv preprint arXiv:2303.08774, 2023.
- ZouFan, “Sina weibo ”Zoufan” comment,” 2023. [Online]. Available: https://www.weibo.com/xiaofan116?is_all=1
- A. Paszke, S. Gross, F. Massa, A. Lerer, J. Bradbury, G. Chanan, T. Killeen, Z. Lin, N. Gimelshein, L. Antiga et al., “PyTorch: An imperative style, high-performance deep learning library,” Advances in neural information processing systems, vol. 32, 2019.
- D. P. Kingma and J. Ba, “Adam: A method for stochastic optimization,” in 3rd International Conference on Learning Representations, ICLR 2015 - Conference Track Proceedings, San Diego, CA, USA, 2015.