Deep Representations of First-person Pronouns for Prediction of Depression Symptom Severity (2310.03232v1)
Abstract: Prior work has shown that analyzing the use of first-person singular pronouns can provide insight into individuals' mental status, especially depression symptom severity. These findings were generated by counting frequencies of first-person singular pronouns in text data. However, counting doesn't capture how these pronouns are used. Recent advances in neural language modeling have leveraged methods generating contextual embeddings. In this study, we sought to utilize the embeddings of first-person pronouns obtained from contextualized language representation models to capture ways these pronouns are used, to analyze mental status. De-identified text messages sent during online psychotherapy with weekly assessment of depression severity were used for evaluation. Results indicate the advantage of contextualized first-person pronoun embeddings over standard classification token embeddings and frequency-based pronoun analysis results in predicting depression symptom severity. This suggests contextual representations of first-person pronouns can enhance the predictive utility of language used by people with depression symptoms.
- The state of mental health in America 2022. 2021;.
- The impact of Covid-19 experiences and associated stress on anxiety, depression, and functional impairment in American adults. Cognitive Therapy and Research. 2020;44:1043–1051.
- Changes in the global burden of depression from 1990 to 2017: Findings from the Global Burden of Disease study. Journal of psychiatric research. 2020;126:134–140.
- County-level estimates of mental health professional shortage in the United States. Psychiatric services. 2009;60(10):1323–1328.
- Ben-Zeev D. The digital mental health genie is out of the bottle. Psychiatric Services. 2020;71(12):1212–1213.
- Effectiveness of text messaging interventions for the management of depression: A systematic review and meta-analysis. Journal of telemedicine and telecare. 2019;25(9):513–523.
- Ingram RE, Smith TW. Depression and internal versus external focus of attention. Cognitive Therapy and Research. 1984;8:139–151.
- Pyszczynski T, Greenberg J. Self-regulatory perseveration and the depressive self-focusing style: a self-awareness theory of reactive depression. Psychological bulletin. 1987;102(1):122.
- Language use of depressed and depression-vulnerable college students. Cognition & Emotion. 2004;18(8):1121–1133.
- Holtzman NS, et al. A meta-analysis of correlations between depression and first person singular pronoun use. Journal of Research in Personality. 2017;68:63–68.
- Tausczik YR, Pennebaker JW. The psychological meaning of words: LIWC and computerized text analysis methods. Journal of language and social psychology. 2010;29(1):24–54.
- Quantifying Mental Health Signals in Twitter. In: Proceedings of the Workshop on Computational Linguistics and Clinical Psychology: From Linguistic Signal to Clinical Reality. Baltimore, Maryland, USA: Association for Computational Linguistics; 2014. p. 51–60.
- Behavioral activation and depression symptomatology: longitudinal assessment of linguistic indicators in text-based therapy sessions. Journal of Medical Internet Research. 2021;23(7):e28244.
- Tracking Mental Health Risks and Coping Strategies in Healthcare Workers’ Online Conversations Across the COVID-19 Pandemic. In: Proceedings of the Eighth Workshop on Computational Linguistics and Clinical Psychology; 2022. p. 76–88.
- BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. CoRR. 2018;abs/1810.04805. Available from: http://arxiv.org/abs/1810.04805.
- Domain-Specific Contextualized Embedding: A Systematic Literature Review. In: 2020 12th International Conference on Information Technology and Electrical Engineering (ICITEE); 2020. p. 162–167.
- Petersen E, Potts C. Lexical Semantics with Large Language Models: A Case Study of English “break”. In: Findings of the Association for Computational Linguistics: EACL 2023; 2023. p. 490–511.
- Mentalbert: Publicly available pretrained language models for mental healthcare. arXiv preprint arXiv:211015621. 2021;.
- The PHQ-9: validity of a brief depression severity measure. Journal of general internal medicine. 2001;16(9):606–613.
- Screening and case finding for major depressive disorder using the Patient Health Questionnaire (PHQ-9): a meta-analysis. General hospital psychiatry. 2015;37(6):567–576.
- Protocol for a sequential, multiple assignment, randomised trial to test the effectiveness of message-based psychotherapy for depression compared with telepsychotherapy. BMJ open. 2021;11(11):e046958.
- BertForSequenceClassification;. Available from: https://huggingface.co/docs/transformers/v4.26.1/en/model_doc/bert#transformers.BertForSequenceClassification.
- What does BERT learn about the structure of language? In: ACL 2019-57th Annual Meeting of the Association for Computational Linguistics; 2019. .
- Transformers: State-of-the-art natural language processing. In: Proceedings of the 2020 conference on empirical methods in natural language processing: system demonstrations; 2020. p. 38–45.
- Kingma DP, Ba J. Adam: A method for stochastic optimization. arXiv preprint arXiv:14126980. 2014;.
- The development and psychometric properties of LIWC-22. Austin, TX: University of Texas at Austin. 2022;p. 1–47.
- Scikit-learn: Machine learning in Python. the Journal of machine Learning research. 2011;12:2825–2830.
- Liu DC, Nocedal J. On the limited memory BFGS method for large scale optimization. Mathematical programming. 1989;45(1-3):503–528.
- Comparing emotion feature extraction approaches for predicting depression and anxiety. In: Proceedings of the eighth workshop on computational linguistics and clinical psychology; 2022. p. 105–115.
- Staner L. Comorbidity of insomnia and depression. Sleep medicine reviews. 2010;14(1):35–46.
- Kendall MG. A new measure of rank correlation. Biometrika. 1938;30(1/2):81–93.
- Test review: patient health questionnaire–9 (PHQ-9). Rehabilitation Counseling Bulletin. 2014;57(4):246–248.
- Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:221105100. 2022;.
- Opt: Open pre-trained transformer language models. arXiv preprint arXiv:220501068. 2022;.
- Misra V. Black box attacks on transformer language models. In: ICLR 2019 Debugging Machine Learning Models Workshop; 2019. .
- Does BERT Pretrained on Clinical Notes Reveal Sensitive Data? arXiv preprint arXiv:210407762. 2021;.
- BioBERT: a pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2020;36(4):1234–1240.
- Williams N. The GAD-7 questionnaire. Occupational medicine. 2014;64(3):224–224.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.