Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
41 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
41 tokens/sec
o3 Pro
7 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Multimodal Gen-AI for Fundamental Investment Research (2401.06164v1)

Published 24 Dec 2023 in q-fin.GN and cs.LG

Abstract: This report outlines a transformative initiative in the financial investment industry, where the conventional decision-making process, laden with labor-intensive tasks such as sifting through voluminous documents, is being reimagined. Leveraging LLMs, our experiments aim to automate information summarization and investment idea generation. We seek to evaluate the effectiveness of fine-tuning methods on a base model (Llama2) to achieve specific application-level goals, including providing insights into the impact of events on companies and sectors, understanding market condition relationships, generating investor-aligned investment ideas, and formatting results with stock recommendations and detailed explanations. Through state-of-the-art generative modeling techniques, the ultimate objective is to develop an AI agent prototype, liberating human investors from repetitive tasks and allowing a focus on high-level strategic thinking. The project encompasses a diverse corpus dataset, including research reports, investment memos, market news, and extensive time-series market data. We conducted three experiments applying unsupervised and supervised LoRA fine-tuning on the llama2_7b_hf_chat as the base model, as well as instruction fine-tuning on the GPT3.5 model. Statistical and human evaluations both show that the fine-tuned versions perform better in solving text modeling, summarization, reasoning, and finance domain questions, demonstrating a pivotal step towards enhancing decision-making processes in the financial domain. Code implementation for the project can be found on GitHub: https://github.com/Firenze11/finance_lm.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (10)
  1. Llm4ts: Two-stage fine-tuning for time-series forecasting with pre-trained llms, 2023.
  2. Think you have solved question answering? try arc, the ai2 reasoning challenge. arXiv preprint arXiv:1803.05457, 2018.
  3. A framework for few-shot language model evaluation, 12 2023. URL https://zenodo.org/records/10256836.
  4. Large language models are zero-shot time series forecasters, 2023.
  5. B. Hlavaty and R. Smith. JPM_DeepFin_with_Microso_2023-09-21_4517766, 2023.
  6. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685, 2021.
  7. C.-Y. Lin. Rouge: A package for automatic evaluation of summaries. In Text summarization branches out, pages 74–81, 2004.
  8. Scrolls: Standardized comparison over long language sequences. arXiv preprint arXiv:2201.03533, 2022.
  9. S. Wu and O. Irsoy. BloombergGPT: A Large Language Model for Finance, 2023.
  10. Temporal data meets llm – explainable financial time series forecasting, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Lezhi Li (7 papers)
  2. Ting-Yu Chang (5 papers)
  3. Hai Wang (98 papers)
Citations (2)
Github Logo Streamline Icon: https://streamlinehq.com