Rethinking Large Language Models in Mental Health Applications (2311.11267v2)
Abstract: LLMs have become valuable assets in mental health, showing promise in both classification tasks and counseling applications. This paper offers a perspective on using LLMs in mental health applications. It discusses the instability of generative models for prediction and the potential for generating hallucinatory outputs, underscoring the need for ongoing audits and evaluations to maintain their reliability and dependability. The paper also distinguishes between the often interchangeable terms explainability'' and
interpretability'', advocating for developing inherently interpretable methods instead of relying on potentially hallucinated self-explanations generated by LLMs. Despite the advancements in LLMs, human counselors' empathetic understanding, nuanced interpretation, and contextual awareness remain irreplaceable in the sensitive and complex realm of mental health counseling. The use of LLMs should be approached with a judicious and considerate mindset, viewing them as tools that complement human expertise rather than seeking to replace it.
- Large-scale analysis of counseling conversations: An application of natural language processing to mental health. Transactions of the Association for Computational Linguistics, 4:463–476, 2016.
- Will affective computing emerge from foundation models and general artificial intelligence? a first evaluation of ChatGPT. IEEE Intelligent Systems, 38(2):15–23, 2023.
- Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862, 2022.
- MindWatch: A smart cloud-based AI solution for suicide ideation detection leveraging large language models. medRxiv, 2023. doi: 10.1101/2023.09.25.23296062. URL https://www.medrxiv.org/content/early/2023/09/26/2023.09.25.23296062.
- Is attention explanation? an introduction to the debate. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 3889–3900, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.acl-long.269. URL https://aclanthology.org/2022.acl-long.269.
- Language models are few-shot learners. Advances in Neural Information Processing Systems, 33:1877–1901, 2020.
- Ethical dilemmas, mental health, artificial intelligence, and LLM-based chatbots. In International Work-Conference on Bioinformatics and Biomedical Engineering, pages 313–326. Springer, 2023.
- Seven pillars for the future of artificial intelligence. IEEE Intelligent Systems, 38(6), 2023.
- EmpHi: Generating empathetic responses with human-like intents. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 1063–1074, 2022.
- Why can GPT learn in-context? language models secretly perform gradient descent as meta-optimizers. In A. Rogers, J. Boyd-Graber, and N. Okazaki, editors, Findings of the Association for Computational Linguistics: ACL 2023, pages 4005–4019, Toronto, Canada, July 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.findings-acl.247. URL https://aclanthology.org/2023.findings-acl.247.
- Chatbots and mental health: Insights into the safety of generative AI. Journal of Consumer Psychology, 2022.
- Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR mental health, 4(2):e7785, 2017.
- Enhancing psychological counseling with large language model: A multifaceted decision-support system for non-professionals, 2023.
- Large language model AI chatbots require approval as medical devices. Nature Medicine, pages 1–3, 2023.
- On the state of social media data for mental health research. In Proceedings of the Seventh Workshop on Computational Linguistics and Clinical Psychology: Improving Access, pages 15–24. ACL, 2021.
- T. F. Heston. Evaluating risk progression in mental health chatbots using escalating prompts. medRxiv, 2023. doi: 10.1101/2023.09.10.23295321. URL https://www.medrxiv.org/content/early/2023/09/12/2023.09.10.23295321.
- S. Ji. Towards intention understanding in suicidal risk assessment with natural language processing. In Findings of EMNLP, pages 4028–4038. Association for Computational Linguistics, 2022. URL https://aclanthology.org/2022.findings-emnlp.297.
- MentalBERT: Publicly Available Pretrained Language Models for Mental Healthcare. In Proceedings of LREC, pages 7184–7190, Marseille, France, 2022. European Language Resources Association. URL https://aclanthology.org/2022.lrec-1.778.
- Domain-specific continued pretraining of language models for capturing long context in mental health. arXiv preprint arXiv:2304.10447, 2023. URL https://arxiv.org/abs/2304.10447.
- Explainable artificial intelligence for mental health through transparency and interpretability for understandability. npj Digital Medicine, 6(1):6, 2023.
- The PHQ-9: validity of a brief depression severity measure. Journal of General Internal Medicine, 16(9):606–613, 2001.
- Psy-LLM: Scaling up global mental health psychological services with AI-based large language models, 2023.
- LCA: Loss change allocation for neural network training. Advances in Neural Information Processing Systems, 32, 2019.
- Understanding client reactions in online mental health counseling. In Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 10358–10376, 2023.
- Towards an online empathetic chatbot with emotion causes. In Proceedings of the 44th International ACM SIGIR Conference on Research and Development in Information Retrieval, pages 2041–2045, 2021.
- ChatCounselor: A large language models for mental health support. arXiv preprint arXiv:2309.15461, 2023a.
- Lost in the middle: How language models use long contexts. arXiv preprint arXiv:2307.03172, 2023b.
- Rethinking the role of demonstrations: What makes in-context learning work? In Y. Goldberg, Z. Kozareva, and Y. Zhang, editors, Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, pages 11048–11064, Abu Dhabi, United Arab Emirates, Dec. 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.emnlp-main.759. URL https://aclanthology.org/2022.emnlp-main.759.
- Help me heal: A reinforced polite and empathetic mental health and legal counseling dialogue system for crime victims. Proceedings of the AAAI Conference on Artificial Intelligence, 37(12):14408–14416, Jun. 2023. doi: 10.1609/aaai.v37i12.26685. URL https://ojs.aaai.org/index.php/AAAI/article/view/26685.
- Auditing large language models: a three-layered approach. AI and Ethics, pages 1–31, 2023.
- Benchmarking for public health surveillance tasks on social media with a domain-specific pretrained language model. In Proceedings of NLP Power! The First Workshop on Efficient Benchmarking in NLP, pages 22–31, Dublin, Ireland, May 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.nlppower-1.3. URL https://aclanthology.org/2022.nlppower-1.3.
- U. Pavalanathan and M. De Choudhury. Identity management and mental health discourse in social media. In WWW, pages 315–321. ACM, 2015.
- SMILE: Single-turn to multi-turn inclusive language expansion via ChatGPT for mental health support. arXiv preprint arXiv:2305.00450, 2023a.
- A benchmark for understanding dialogue safety in mental health support. In CCF International Conference on Natural Language Processing and Chinese Computing, pages 1–13. Springer, 2023b.
- C. Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5):206–215, 2019.
- Towards motivational and empathetic response generation in online mental health support. In Proceedings of the 45th international ACM SIGIR conference on research and development in information retrieval, pages 2650–2656, 2022.
- A review of the explainability and safety of conversational agents for mental health to identify avenues for improvement. Frontiers in Artificial Intelligence, 6, 2023.
- Towards facilitating empathic conversations in online mental health support: A reinforcement learning approach. In Proceedings of the Web Conference, pages 194–205, 2021.
- Large language models can be easily distracted by irrelevant context. In International Conference on Machine Learning, pages 31210–31227. PMLR, 2023.
- Language models don’t always say what they think: Unfaithful explanations in chain-of-thought prompting. arXiv preprint arXiv:2305.04388, 2023.
- PsychBERT: a mental health language model for social media mental health behavioral analysis. In 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pages 1077–1082. IEEE, 2021.
- Transformers learn in-context by gradient descent. In International Conference on Machine Learning, pages 35151–35174. PMLR, 2023.
- Interactive natural language processing. arXiv preprint arXiv:2305.13246, 2023.
- Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
- X. Wu and L. R. Varshney. A meta-learning perspective on transformers for causal language modeling. arXiv preprint arXiv:2310.05884, 2023.
- Mental-LLM: Leveraging large language models for mental health prediction via online text data. arXiv preprint arXiv:2307.14385, 2023.
- Towards interpretable mental health analysis with large language models. In Proceedings of EMNLP, 2023a. URL https://arxiv.org/abs/2304.03347.
- MentalLLaMA: Interpretable mental health analysis on social media with large language models. arXiv preprint arXiv:2309.13567, 2023b.
- Natural language processing applied to mental illness detection: A narrative review. npj Digital Medicine, 5, 2022.
- Explainability for large language models: A survey. arXiv preprint arXiv:2309.01029, 2023.
- Shaoxiong Ji (39 papers)
- Tianlin Zhang (17 papers)
- Kailai Yang (22 papers)
- Sophia Ananiadou (72 papers)
- Erik Cambria (136 papers)