Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
119 tokens/sec
GPT-4o
56 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

ALCM: Autonomous LLM-Augmented Causal Discovery Framework (2405.01744v1)

Published 2 May 2024 in cs.LG, cs.AI, cs.CL, and stat.ME

Abstract: To perform effective causal inference in high-dimensional datasets, initiating the process with causal discovery is imperative, wherein a causal graph is generated based on observational data. However, obtaining a complete and accurate causal graph poses a formidable challenge, recognized as an NP-hard problem. Recently, the advent of LLMs has ushered in a new era, indicating their emergent capabilities and widespread applicability in facilitating causal reasoning across diverse domains, such as medicine, finance, and science. The expansive knowledge base of LLMs holds the potential to elevate the field of causal reasoning by offering interpretability, making inferences, generalizability, and uncovering novel causal structures. In this paper, we introduce a new framework, named Autonomous LLM-Augmented Causal Discovery Framework (ALCM), to synergize data-driven causal discovery algorithms and LLMs, automating the generation of a more resilient, accurate, and explicable causal graph. The ALCM consists of three integral components: causal structure learning, causal wrapper, and LLM-driven causal refiner. These components autonomously collaborate within a dynamic environment to address causal discovery questions and deliver plausible causal graphs. We evaluate the ALCM framework by implementing two demonstrations on seven well-known datasets. Experimental results demonstrate that ALCM outperforms existing LLM methods and conventional data-driven causal reasoning mechanisms. This study not only shows the effectiveness of the ALCM but also underscores new research directions in leveraging the causal reasoning capabilities of LLMs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (50)
  1. Conversational health agents: A personalized llm-powered agent framework. arXiv preprint arXiv:2310.02374, 2023.
  2. Causal modelling agents: Causal graph discovery through synergising metadata-and data-driven reasoning. In The Twelfth International Conference on Learning Representations, 2023.
  3. Sam Altman. OpenAI — openai.com. https://openai.com/. [Accessed 04-03-2024].
  4. Zero-shot causal graph extrapolation from text via llms. arXiv preprint arXiv:2312.14670, 2023.
  5. Cause and effect: Can large language models truly understand causality? arXiv preprint arXiv:2402.18139, 2024.
  6. Causal structure learning supervised by large language model. arXiv preprint arXiv:2311.11689, 2023.
  7. From query tools to causal architects: Harnessing large language models for advanced causal discovery from data. arXiv preprint arXiv:2306.16902, 2023.
  8. Towards llm-guided causal explainability for black-box text classifiers. researchgate.net, 2024.
  9. Is knowledge all large language models needed for causal reasoning? arXiv preprint arXiv:2401.00139, 2023.
  10. Mitigating prior errors in causal structure learning: Towards llm driven prior knowledge. arXiv preprint arXiv:2306.07032, 2023.
  11. Selective greedy equivalence search: Finding optimal bayesian networks using a polynomial number of score evaluations. arXiv preprint arXiv:1506.02113, 2015.
  12. The impact of prior knowledge on causal structure learning. Knowledge and Information Systems, pages 1–50, 2023.
  13. Review of causal discovery methods based on graphical models. Frontiers in genetics, 10:524, 2019.
  14. A survey on causal discovery methods for iid and time series data. Transactions on Machine Learning Research, 2023.
  15. Can large language models truly understand prompts? a case study with negated prompts. In Transfer Learning for Natural Language Processing Workshop, pages 52–62. PMLR, 2023.
  16. Large language model for causal decision making. arXiv preprint arXiv:2312.17122, 2023.
  17. Can large language models infer causation from correlation? arXiv preprint arXiv:2306.05836, 2023.
  18. Efficient causal graph discovery using large language models. arXiv preprint arXiv:2402.01207, 2024.
  19. Structural agnostic modeling: Adversarial learning of causal graphs. The Journal of Machine Learning Research, 23(1):9831–9892, 2022.
  20. Causal reasoning and large language models: Opening a new frontier for causality. arXiv preprint arXiv:2305.00050, 2023.
  21. Health-llm: Large language models for health prediction via wearable sensor data. arXiv preprint arXiv:2401.06866, 2024.
  22. Passive learning of active causal strategies in agents and language models. arXiv preprint arXiv:2305.16183, 2023.
  23. Discovery of the hidden world with large language models. arXiv preprint arXiv:2402.03941, 2024.
  24. Evaluating the logical reasoning ability of chatgpt and gpt-4. arXiv preprint arXiv:2304.03439, 2023.
  25. Causal discovery with language models as imperfect experts. arXiv preprint arXiv:2307.02390, 2023.
  26. Can large language models build causal graphs? arXiv preprint arXiv:2303.05279, 2023.
  27. Capabilities of gpt-4 on medical challenge problems. arXiv preprint arXiv:2303.13375, 2023.
  28. Tobacco smoke and involuntary smoking, volume 83. Iarc, 2004.
  29. R OpenAI. Gpt-4 technical report. arxiv 2303.08774. View in Article, 2:13, 2023.
  30. Answering causal questions with augmented llms. openreview.net, 2023.
  31. Judea Pearl. Causal inference in statistics: An overview. Statistics Survey, 2009.
  32. Judea Pearl. The causal mediation formula—a guide to the assessment of pathways and mechanisms. Prevention science, 13:426–436, 2012.
  33. Joseph D Ramsey. Scaling up greedy causal search for continuous variables. arXiv preprint arXiv:1507.07749, 2015.
  34. Beware of the simulated dag! causal discovery benchmarks may be easy to game. Advances in Neural Information Processing Systems, 34:27772–27784, 2021.
  35. A meta-reinforcement learning algorithm for causal discovery. In Conference on Causal Learning and Reasoning, pages 602–619. PMLR, 2023.
  36. Marco Scutari. bnlearn - Bayesian Network Repository — bnlearn.com. https://www.bnlearn.com/bnrepository/. [Accessed 13-02-2024].
  37. A linear non-gaussian acyclic model for causal discovery. Journal of Machine Learning Research, 7(10), 2006.
  38. Causation, prediction, and search. MIT press, 2000.
  39. Integrating large language models in causal discovery: A statistical causal approach. arXiv preprint arXiv:2402.01454, 2024.
  40. The max-min hill-climbing bayesian network structure learning algorithm. Machine learning, 65:31–78, 2006.
  41. Causal-discovery performance of chatgpt in the context of neuropathic pain diagnosis, jan 2023.
  42. Neuropathic pain diagnosis simulator for causal discovery algorithm evaluation. Advances in Neural Information Processing Systems, 32, 2019.
  43. Causal inference using llm-guided discovery. arXiv preprint arXiv:2310.15117, 2023.
  44. Exploring the reasoning abilities of multimodal large language models (mllms): A comprehensive survey on emerging trends in multimodal reasoning. arXiv preprint arXiv:2401.06805, 2024.
  45. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837, 2022.
  46. Can foundation models talk causality? arXiv preprint arXiv:2206.10591, 2022.
  47. A* lasso for learning a sparse bayesian network structure for continuous variables. Advances in neural information processing systems, 26, 2013.
  48. A survey on causal discovery: theory and practice. International Journal of Approximate Reasoning, 151:101–129, 2022.
  49. Causal graph discovery with retrieval-augmented generation based large language models. arXiv preprint arXiv:2402.15301, 2024.
  50. Dags with no tears: Continuous optimization for structure learning. Advances in neural information processing systems, 31, 2018.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Elahe Khatibi (7 papers)
  2. Mahyar Abbasian (9 papers)
  3. Zhongqi Yang (10 papers)
  4. Iman Azimi (20 papers)
  5. Amir M. Rahmani (48 papers)
Citations (2)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets