Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
51 tokens/sec
GPT-4o
60 tokens/sec
Gemini 2.5 Pro Pro
44 tokens/sec
o3 Pro
8 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Are LLMs Rational Investors? A Study on Detecting and Reducing the Financial Bias in LLMs (2402.12713v2)

Published 20 Feb 2024 in cs.CL

Abstract: LLMs are increasingly adopted in financial analysis for interpreting complex market data and trends. However, their use is challenged by intrinsic biases (e.g., risk-preference bias) and a superficial understanding of market intricacies, necessitating a thorough assessment of their financial insight. To address these issues, we introduce Financial Bias Indicators (FBI), a framework with components like Bias Unveiler, Bias Detective, Bias Tracker, and Bias Antidote to identify, detect, analyze, and eliminate irrational biases in LLMs. By combining behavioral finance principles with bias examination, we evaluate 23 leading LLMs and propose a de-biasing method based on financial causal knowledge. Results show varying degrees of financial irrationality among models, influenced by their design and training. Models trained specifically on financial datasets may exhibit more irrationality, and even larger financial LLMs (FinLLMs) can show more bias than smaller, general models. We utilize four prompt-based methods incorporating causal debiasing, effectively reducing financial biases in these models. This work enhances the understanding of LLMs' bias in financial applications, laying the foundation for developing more reliable and rational financial analysis tools.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. Qwen technical report. arXiv preprint arXiv:2309.16609.
  2. Nicholas Barberis and Richard Thaler. 2003. A survey of behavioral finance. Handbook of the Economics of Finance, 1:1053–1128.
  3. Finqa: A dataset of numerical reasoning over financial data. arXiv preprint arXiv:2109.00122.
  4. Buy tesla, sell ford: Assessing implicit stock market preference in pre-trained language models. In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers), pages 100–105.
  5. Chatlaw: Open-source legal large language model with integrated external knowledge bases.
  6. Can large language models beat wall street? unveiling the potential of ai in stock selection. arXiv preprint arXiv:2401.03737.
  7. Bias and fairness in large language models: A survey.
  8. Maarten Grootendorst. 2022. Bertopic: Neural topic modeling with a class-based tf-idf procedure. arXiv preprint arXiv:2203.05794.
  9. Opiniongpt: Modelling explicit biases in instruction-tuned llms.
  10. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
  11. Stereomap: Quantifying the awareness of human-like stereotypes in large language models. arXiv preprint arXiv:2310.13673.
  12. Scaling laws for neural language models. arXiv preprint arXiv:2001.08361.
  13. Alejandro Lopez-Lira and Yuehua Tang. 2023. Can chatgpt forecast stock price movements? return predictability and large language models. arXiv preprint arXiv:2304.07619.
  14. Stereoset: Measuring stereotypical bias in pretrained language models.
  15. OpenAI. 2023. Gpt-4 technical report.
  16. Chatgpt-based investment portfolio selection. In Operations Research Forum, volume 4, page 91. Springer.
  17. The self-perception and political biases of chatgpt. Human Behavior and Emerging Technologies, 2024.
  18. Mathematics for economists, volume 7. Norton New York.
  19. Analysis of variance (anova). Chemometrics and intelligent laboratory systems, 6(4):259–272.
  20. Llama 2: Open foundation and fine-tuned chat models. arXiv preprint arXiv:2307.09288.
  21. Huatuo: Tuning llama model with chinese medical knowledge.
  22. Xia Wang and Min Wu. 2011. The quality of financial reporting in china: An examination from an accounting restatement perspective. China journal of accounting research, 4(4):167–196.
  23. Chain-of-thought prompting elicits reasoning in large language models. Advances in Neural Information Processing Systems, 35:24824–24837.
  24. Bloom: A 176b-parameter open-access multilingual language model. arXiv preprint arXiv:2211.05100.
  25. Bloomberggpt: A large language model for finance. arXiv preprint arXiv:2303.17564.
  26. Measuring consistency in text-based financial forecasting models. arXiv preprint arXiv:2305.08524.
  27. Glm-130b: An open bilingual pre-trained model. arXiv preprint arXiv:2210.02414.
  28. Xuanyuan 2.0: A large chinese financial chat model with hundreds of billions parameters.
  29. A survey of large language models.
  30. Large language models are not robust multiple choice selectors. arXiv e-prints, pages arXiv–2309.
  31. Trade the event: Corporate events detection for news-based event-driven trading. arXiv preprint arXiv:2105.12825.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (10)
  1. Yuhang Zhou (52 papers)
  2. Yuchen Ni (4 papers)
  3. Xiang Liu (475 papers)
  4. Jian Zhang (542 papers)
  5. Sen Liu (35 papers)
  6. Guangnan Ye (17 papers)
  7. Hongfeng Chai (7 papers)
  8. Yunhui Gan (1 paper)
  9. Zhangyue Yin (27 papers)
  10. Xipeng Qiu (257 papers)
Citations (3)
X Twitter Logo Streamline Icon: https://streamlinehq.com