From Query Tools to Causal Architects: Harnessing Large Language Models for Advanced Causal Discovery from Data (2306.16902v1)
Abstract: LLMs exhibit exceptional abilities for causal analysis between concepts in numerous societally impactful domains, including medicine, science, and law. Recent research on LLM performance in various causal discovery and inference tasks has given rise to a new ladder in the classical three-stage framework of causality. In this paper, we advance the current research of LLM-driven causal discovery by proposing a novel framework that combines knowledge-based LLM causal analysis with data-driven causal structure learning. To make LLM more than a query tool and to leverage its power in discovering natural and new laws of causality, we integrate the valuable LLM expertise on existing causal mechanisms into statistical analysis of objective data to build a novel and practical baseline for causal structure learning. We introduce a universal set of prompts designed to extract causal graphs from given variables and assess the influence of LLM prior causality on recovering causal structures from data. We demonstrate the significant enhancement of LLM expertise on the quality of recovered causal structures from data, while also identifying critical challenges and issues, along with potential approaches to address them. As a pioneering study, this paper aims to emphasize the new frontier that LLMs are opening for classical causal discovery and inference, and to encourage the widespread adoption of LLM capabilities in data-driven causal analysis.
- 2022. Quality evaluation of triples in knowledge graph by incorporating internal with external consistency. IEEE Transactions on Neural Networks and Learning Systems.
- 2020. Language models are few-shot learners. Advances in neural information processing systems 33:1877–1901.
- 2016. Learning bayesian networks with ancestral constraints. Advances in Neural Information Processing Systems 29.
- 2023. Mitigating prior errors in causal structure learning: Towards llm driven prior knowledge. arXiv preprint arXiv:2306.07032.
- 2004. Large-sample learning of bayesian networks is np-hard. Journal of Machine Learning Research 5:1287–1330.
- 2023. The impact of prior knowledge on causal structure learning. Knowledge and Information Systems 1–50.
- 2004. Discovery of meaningful associations in genomic data using partial correlation coefficients. Bioinformatics 20(18):3565–3574.
- 2021. Crass: A novel data set and benchmark to test counterfactual reasoning of large language models. arXiv preprint arXiv:2112.11941.
- 2022. Machine intuition: Uncovering human-like intuitive decision-making in gpt-3.5. arXiv preprint arXiv:2212.05206.
- 1995. Learning bayesian networks: a unification for discrete and gaussian domains. In Proceedings of the 11th Conference on Uncertainty in Artificial Intelligence, 274–284.
- 2008. Nonlinear causal discovery with additive noise models. Advances in neural information processing systems 21.
- 2023. Can large language models infer causation from correlation? arXiv preprint arXiv:2306.05836.
- 2023. Causal reasoning and large language models: Opening a new frontier for causality. arXiv preprint arXiv:2305.00050.
- 2018. Bayesian network structure learning with side constraints. In International Conference on Probabilistic Graphical Models, 225–236. PMLR.
- 2023. Llm-eval: Unified multi-dimensional automatic evaluation for open-domain conversations with large language models. arXiv preprint arXiv:2305.13711.
- 2023. Swiftsage: A generative agent with fast and slow thinking for complex interactive tasks. arXiv preprint arXiv:2305.17390.
- 2023. Evaluating the logical reasoning ability of chatgpt and gpt-4. arXiv preprint arXiv:2304.03439.
- 2023. Can large language models build causal graphs? arXiv preprint arXiv:2303.05279.
- Marcus, G. 2022. How come gpt can seem so brilliant one minute and so breathtakingly dumb the next? Substack newsletter. The Road to AI We Can Trust.
- 2023. Capabilities of gpt-4 on medical challenge problems. arXiv preprint arXiv:2303.13375.
- 2006. Causal discovery with prior information. In AI 2006: Advances in Artificial Intelligence: 19th Australian Joint Conference on Artificial Intelligence, Hobart, Australia, December 4-8, 2006. Proceedings 19, 1162–1167. Springer.
- 2023. Instruction tuning with GPT-4. arXiv preprint arXiv:2304.03277.
- 2000. Causation, prediction, and search. MIT press.
- 2012. Ordering-based search: A simple and effective algorithm for learning bayesian networks. arXiv preprint arXiv:1207.1429.
- 2006. The Max-Min Hill-Climbing Bayesian network structure learning algorithm. Machine Learning 65:31–78.
- 2023. Causal-discovery performance of chatgpt in the context of neuropathic pain diagnosis. arXiv preprint arXiv:2301.13819.
- 2021. Knowledge graph quality control: A survey. Fundamental Research 1(5):607–626.
- 2022a. Knowledge verification from data. IEEE Transactions on Neural Networks and Learning Systems 1–15.
- 2022b. Generalization bounds for estimating causal effects of continuous treatments. In Advances in Neural Information Processing Systems, volume 35, 8605–8617.
- 2022. Can foundation models talk causality? arXiv preprint arXiv:2206.10591.
- 2023. Customizing general-purpose foundation models for medical report generation. arXiv preprint arXiv:2306.05642.
- 2023. Response length perception and sequence scheduling: An LLM-Empowered LLM inference pipeline. arXiv preprint arXiv:2305.13144.
- Taiyu Ban (6 papers)
- Lyvzhou Chen (1 paper)
- Xiangyu Wang (79 papers)
- Huanhuan Chen (42 papers)