Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Cordyceps@LT-EDI: Depression Detection with Reddit and Self-training (2310.01418v1)

Published 24 Sep 2023 in cs.CL and cs.AI

Abstract: Depression is debilitating, and not uncommon. Indeed, studies of excessive social media users show correlations with depression, ADHD, and other mental health concerns. Given that there is a large number of people with excessive social media usage, then there is a significant population of potentially undiagnosed users and posts that they create. In this paper, we propose a depression severity detection system using a semi-supervised learning technique to predict if a post is from a user who is experiencing severe, moderate, or low (non-diagnostic) levels of depression. Namely, we use a trained model to classify a large number of unlabelled social media posts from Reddit, then use these generated labels to train a more powerful classifier. We demonstrate our framework on Detecting Signs of Depression from Social Media Text - LT-EDI@RANLP 2023 shared task, where our framework ranks 3rd overall.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. BERT: pre-training of deep bidirectional transformers for language understanding. CoRR, abs/1810.04805.
  2. A meta-analysis of correlations between depression and first person singular pronoun use. Journal of Research in Personality, 68:63–68.
  3. Facebook language predicts depression in medical records. Proceedings of the National Academy of Sciences of the United States of America, 115:11203 – 11208.
  4. Detecting depression and mental illness on social media: an integrative review. Current Opinion in Behavioral Sciences, 18:43–49.
  5. Depression detection from social networks data based on machine learning and deep learning techniques: An interrogative survey. IEEE Transactions on Computational Social Systems.
  6. Zaheer Hussain and Mark D. Griffiths. 2019. The associations between problematic social networking site use and sleep quality, attention-deficit hyperactivity disorder, depression, anxiety and stress. International Journal of Mental Health and Addiction, 19:686 – 700.
  7. MentalBERT: Publicly Available Pretrained Language Models for Mental Healthcare. In Proceedings of LREC.
  8. Domain-specific continued pretraining of language models for capturing long context in mental health. arXiv preprint arXiv:2304.10447.
  9. The nature of clinical depression: Symptoms, syndromes, and behavior analysis. The Behavior Analyst, 31:1–21.
  10. Roberta: A robustly optimized BERT pretraining approach. CoRR, abs/1907.11692.
  11. Roberta: A robustly optimized bert pretraining approach. ArXiv, abs/1907.11692.
  12. Natural language processing reveals vulnerable mental health support groups and heightened health anxiety on reddit during covid-19: Observational study. Journal of medical Internet research, 22(10):e22635.
  13. Michael C. Meinzer and Andrea Chronis-Tuscano. 2017. Adhd and the development of depression: Commentary on the prevalence, proposed mechanisms, and promising interventions. Current Developmental Disorders Reports, 4:1–4.
  14. Inna Loginovna Pirina and Çagri Çöltekin. 2018. Identifying depression on reddit: The effect of training data. In Conference on Empirical Methods in Natural Language Processing.
  15. Rafal Poswiata and Michal Perelkiewicz. 2022. Opi@lt-edi-acl2022: Detecting signs of depression from social media text using roberta pre-trained language models. In LTEDI.
  16. Forecasting the onset and course of mental illness with twitter data. Scientific Reports, 7.
  17. Adhd and depression: investigating a causal explanation. Psychological Medicine, 51:1890 – 1897.
  18. Better self-training for image classification through self-supervision. ArXiv, abs/2109.00778.
  19. Overview of the second shared task on detecting signs of depression from social media text. In Proceedings of the Third Workshop on Language Technology for Equality, Diversity and Inclusion, Varna, Bulgaria. Recent Advances in Natural Language Processing.
  20. H. J. Scudder. 1965. Probability of error of some adaptive pattern-recognition machines. IEEE Trans. Inf. Theory, 11:363–371.
  21. Katrine Bønneland Tølbøll. 2019. Linguistic features in depression: a meta-analysis.
  22. Self-training with noisy student improves imagenet classification. 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pages 10684–10695.
  23. Semi-supervised approach to monitoring clinical depressive symptoms in social media. Proceedings of the 2017 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining 2017.
  24. Natural language processing applied to mental illness detection: a narrative review. NPJ Digital Medicine, 5.
  25. Rethinking pre-training and self-training. ArXiv, abs/2006.06882.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (1)
  1. Dean Ninalga (4 papers)