A Cross Attention Approach to Diagnostic Explainability using Clinical Practice Guidelines for Depression (2311.13852v4)
Abstract: The lack of explainability using relevant clinical knowledge hinders the adoption of Artificial Intelligence-powered analysis of unstructured clinical dialogue. A wealth of relevant, untapped Mental Health (MH) data is available in online communities, providing the opportunity to address the explainability problem with substantial potential impact as a screening tool for both online and offline applications. We develop a method to enhance attention in popular transformer models and generate clinician-understandable explanations for classification by incorporating external clinical knowledge. Inspired by how clinicians rely on their expertise when interacting with patients, we leverage relevant clinical knowledge to model patient inputs, providing meaningful explanations for classification. This will save manual review time and engender trust. We develop such a system in the context of MH using clinical practice guidelines (CPG) for diagnosing depression, a mental health disorder of global concern. We propose an application-specific LLM called ProcesS knowledge-infused cross ATtention (PSAT), which incorporates CPGs when computing attention. Through rigorous evaluation on three expert-curated datasets related to depression, we demonstrate application-relevant explainability of PSAT. PSAT also surpasses the performance of nine baseline models and can provide explanations where other baselines fall short. We transform a CPG resource focused on depression, such as the Patient Health Questionnaire (e.g. PHQ-9) and related questions, into a machine-readable ontology using SNOMED-CT. With this resource, PSAT enhances the ability of models like GPT-3.5 to generate application-relevant explanations.
- Transfer learning for eating disorder sentiment analysis, 2022.
- Beck depression inventory. Harcourt Brace Jovanovich New York:, 1987.
- Longformer: The long-document transformer. arXiv:2004.05150, 2020.
- Post-hoc explanations fail to achieve their purpose in adversarial contexts. arXiv preprint arXiv:2201.10295, 2022.
- The use of mmr, diversity-based reranking for reordering documents and producing summaries. In Proceedings of the 21st annual international ACM SIGIR conference on Research and development in information retrieval, pages 335–336, 1998.
- The advantages of the matthews correlation coefficient (mcc) over f1 score and accuracy in binary classification evaluation. BMC genomics, 21(1):1–13, 2020.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
- A roadmap for a rigorous science of interpretability. arXiv preprint arXiv:1702.08608, 2:1, 2017.
- Expanding explainability: Towards social transparency in ai systems. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, pages 1–19, 2021.
- Michael B First. Structured clinical interview for the dsm (scid). The encyclopedia of clinical psychology, pages 1–6, 2014.
- Knowledge-aware assessment of severity of suicide risk for early intervention. In The world wide web conference, pages 514–525, 2019.
- Maarten Grootendorst. Keybert: Minimal keyword extraction with bert., 2020.
- Depression detection on social media with the aid of machine learning platform: A comprehensive survey. 2021 8th International Conference on Computing for Sustainable Global Development (INDIACom), pages 658–662, 2021.
- Attention is not explanation. arXiv preprint arXiv:1902.10186, 2019.
- Mentalbert: Publicly available pretrained language models for mental healthcare. In International Conference on Language Resources and Evaluation, 2021.
- Learning rich representation of keyphrases from text. In Findings of the Association for Computational Linguistics: NAACL 2022, pages 891–906, Seattle, United States, July 2022. Association for Computational Linguistics.
- Conceptnet—a practical commonsense reasoning tool-kit. BT technology journal, 22(4):211–226, 2004.
- Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
- A test collection for research on depression and language use. In International Conference of the Cross-Language Evaluation Forum for European Languages, pages 28–39. Springer, 2016.
- A unified approach to interpreting model predictions. Advances in neural information processing systems, 30, 2017.
- George A Miller. Wordnet: a lexical database for english. Communications of the ACM, 38(11):39–41, 1995.
- The columbia–suicide severity rating scale: initial validity and internal consistency findings from three multisite studies with adolescents and adults. American journal of psychiatry, 168(12):1266–1277, 2011.
- ” why should i trust you?” explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1135–1144, 2016.
- Proknow: Process knowledge for safety constrained and explainable question generation for mental health diagnostic assistance, 2022.
- Cynthia Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5):206–215, 2019.
- Patternrank: Leveraging pretrained language models and part of speech for unsupervised keyphrase extraction. In Proceedings of the 14th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management - KDIR, pages 243–248. INSTICC, SciTePress, 2022.
- Rachel E Smith. Urban dictionary: youth slanguage and the redefining of definition: What’s up with meep and other words in the urban dictionary. English Today, 27(4):43–48, 2011.
- Understanding the capabilities, limitations, and societal impact of large language models. arXiv preprint arXiv:2102.02503, 2021.
- Explainable artificial intelligence (xai) post-hoc explainability methods: Risks and limitations in non-discrimination law. AI and Ethics, pages 1–12, 2022.
- K-adapter: Infusing knowledge into pre-trained models with adapters. In Findings, 2020.
- Extracting discriminative keyphrases with learned semantic hierarchies. In Proceedings of COLING 2016, the 26th International Conference on Computational Linguistics: Technical Papers, pages 932–942, 2016.
- Attention is not not explanation. arXiv preprint arXiv:1908.04626, 2019.
- Nerys Williams. The gad-7 questionnaire. Occupational medicine, 64(3):224–224, 2014.
- Explaining models of mental health via clinically grounded auxiliary tasks. CLPsych 2022, page 30, 2022.
- Sumit Dalal (2 papers)
- Deepa Tilwani (5 papers)
- Manas Gaur (59 papers)
- Sarika Jain (6 papers)
- Valerie Shalin (8 papers)
- Kaushik Roy (265 papers)
- Amit Sheth (127 papers)