Papers
Topics
Authors
Recent
2000 character limit reached

From Deep Learning to LLMs: A survey of AI in Quantitative Investment (2503.21422v1)

Published 27 Mar 2025 in q-fin.CP, cs.LG, q-fin.ST, q-fin.TR, and cs.AI

Abstract: Quantitative investment (quant) is an emerging, technology-driven approach in asset management, increasingy shaped by advancements in artificial intelligence. Recent advances in deep learning and LLMs for quant finance have improved predictive modeling and enabled agent-based automation, suggesting a potential paradigm shift in this field. In this survey, taking alpha strategy as a representative example, we explore how AI contributes to the quantitative investment pipeline. We first examine the early stage of quant research, centered on human-crafted features and traditional statistical models with an established alpha pipeline. We then discuss the rise of deep learning, which enabled scalable modeling across the entire pipeline from data processing to order execution. Building on this, we highlight the emerging role of LLMs in extending AI beyond prediction, empowering autonomous agents to process unstructured data, generate alphas, and support self-iterative workflows.

Summary

  • The paper demonstrates a comprehensive survey of AI progression, documenting the shift from deep learning to LLM-powered autonomous pipelines in quantitative investment.
  • It details methodological innovations in data processing, predictive modeling, portfolio optimization, and order execution using deep learning and reinforcement learning.
  • It highlights the integration of LLMs for sentiment extraction, causal reasoning, and multi-modal analysis, addressing both performance gains and practical challenges.

Survey of AI Progression in Quantitative Investment: From Deep Learning to LLM-Driven Autonomy

This essay comprehensively reviews the conceptual and practical advancements resulting from the integration of Deep Learning (DL) and LLMs in the domain of quantitative investment. Focused on alpha strategy as the canonical use case, it addresses methodological innovations across the investment pipeline, covering data handling, predictive modeling, portfolio construction, and execution, and discusses the shift from human-in-the-loop workflows to AI-autonomous pipelines. The following sections examine the current methodological spectrum, empirical results, and prospective research directions.

Alpha Strategy and the Quantitative Investment Pipeline

The alpha strategy framework serves as a fundamental organizing principle in systematic asset management, aiming to extract idiosyncratic returns beyond benchmark performance while managing exogenous risk. In practice, the pipeline for executing alpha strategies encompasses:

  1. Data Processing—Standardization and feature extraction from raw multi-modal, high dimensional, and often irregularly sampled financial data.
  2. Model Prediction—Forecasting asset-specific returns and risk features by leveraging representations extracted from observed data.
  3. Portfolio Optimization—Solving allocation problems over high-cardinality investment universes under constraints related to expected return, volatility, and transaction cost.
  4. Order Execution—Translating theoretical allocations to market orders with market impact-aware execution logic and feedback-based risk control. Figure 1

    Figure 1: A typical pipeline of quantitative investment.

Integration across these stages is crucial for minimizing error propagation and maximizing overall system performance. Feedback from execution and risk analysis informs adaptive revisions upstream.

Evolution: From Traditional Methods to Deep Learning

The field’s trajectory (Figure 2) has evolved from traditional statistical modeling—characterized by handcrafted feature engineering and fixed functional forms—to modern deep learning models operating on raw or weakly processed data, allowing the capture of nonlinear dependencies and high-order interactions in both temporal and cross-sectional dimensions. Figure 2

Figure 2: The evolutionary process of Alpha investment across different stages.

Deep Learning Architectures in the Alpha Pipeline

Temporal, Spatial, and Spatiotemporal Modeling

  • Temporal Modeling: LSTM, GRU, TCN, and Transformer-based architectures efficiently encode temporal dependence, with Transformer-based methods increasingly supplanting RNNs, particularly for long-horizon dependencies [vaswani2017attention].
  • Spatial Modeling: Cross-sectional dependencies among assets are formalized using GNNs, graph-attention mechanisms, and knowledge graphs where both explicit and learned relational graphs are jointly optimized.
  • Spatiotemporal Integration: Decoupled and coupled architectures aggregate temporal and spatial signals, with coupled models providing a higher degree of cross-interaction at the cost of increased model complexity and resource requirements (see also Figure 3). Figure 4

    Figure 4: Temporal Patterns.

Data Modality and Simulation

Numerical, relational, and alternative data (such as textual news or ESG signals) are now routinely fused (see Figure 5). The generation of simulation data through rule-based multi-agent systems, deep generative models (GAN, VAE, Diffusion) serves robust model training and systematic stress testing [kannan2024review]. Figure 6

Figure 6

Figure 6

Figure 6: Candlestick chart—a primary visualization for financial time series.

Portfolio and Execution Optimization

Deep learning has advanced both plug-in and end-to-end solutions for portfolio optimization. Deep reinforcement learning (RL) agents now directly optimize non-differentiable financial objectives (annualized return, Sharpe, max drawdown), circumventing intermediary predictive targets. Notably, RL-based allocation policies demonstrate improved empirical risk-adjusted returns compared to traditional plug-in estimators. For order execution, RL-based agents achieve superior performance over Almgren-Chriss baselines when minimizing transaction cost and slippage under realistic market frictions [donnelly2022optimal].

LLMs in Quantitative Finance

The advent of foundation LLMs (e.g., GPT-family, BloombergGPT [wu2023bloomberggpt], FinGPT [yang2023fingpt]), pretrained on vast financial corpora, has introduced multi-modality, autonomous reasoning, and agents with tool integration capability into the investment process (see Figure 7). Their involvement spans: Figure 7

Figure 7: Architecture overview of LLM-based quant agents. It has three parts: using data to predict price trends, optimizing asset choices, and making trades.

LLM-Based Predictors

Sentiment Extraction and Causal Representation

  • Embedding-based LLM classifiers (e.g., FinBERT [araci2019finbert], FinLlama [konstantinidis2024finllama]) outperform lexicon and shallow ML baselines for sentiment extraction, directly improving factor quality for both single-asset and cross-sectional strategies.
  • Prompted LLMs (ChatGPT, GPT-4) yield return-predictive sentiment and event signals from structured and unstructured news, with backtests demonstrating significant out-of-sample alpha across equities and FX [lopezlira2023chatgpt, zhang2023unveiling].
  • Causal attention models (e.g., CMIN [luo2023cmin]) integrate LLM-extracted sentiment with inferred causal relationships, improving both interpretability and predictive power for event-driven trading.

Time-Series Forecasting

  • LLMs reprogrammed for time series forecasting (e.g., TIME-LLM [jin2024timellm], S2IP-LLM [pan2024s2ip]) match or exceed conventional statistical models and supervised learning approaches in zero-shot and few-shot settings [gruver2023zeroshot]. They further enable multi-modal integration (e.g., textual, numerical, audio features as in RiskLabs [cao2024risklabs]) and generate human-interpretable rationale for predictions, critical for regulated financial applications.

LLM Agents: Autonomy, Memory, and Multi-Agent Systems

  • Modular, multi-component agent frameworks (e.g., FinMem [yu2024finmem], FinAgent [zhang2024FinAgent], FinVision [fatemi2024finvision]) operationalize workflow automation, integrating layered memory, character-driven profiling, dynamic tool use, and multi-source perception.
  • Collaborative multi-agent systems (e.g., FinRobot [yang2024finrobot], FINCON [yu2025fincon], TradingAgents [xiao2024tradingagents]) replicate institutional structures (managers, analysts, traders) for robust portfolio management and risk control, producing superior cumulative return, Sharpe ratio, and drawdown performance relative to deep RL and classical baseline agents.
  • Public benchmarks (InvestorBench [li2024investorbench]) show proprietary LLM-based agents outperforming open-source and prior domain-specific models in volatile market conditions, especially for multi-asset, multi-modal tasks.

LLMs for Direct Factor and Knowledge-Driven Generation

  • Factor mining with LLMs (Alpha-GPT [wang2023alphagpt], GPT’s Idea of Stock Factors [cheng2024gpt]) automates the generation of alpha signals by interactive code synthesis, iterative domain-driven refinement, and cross-modal reasoning, surpassing symbolic and ML-only approaches in both factor novelty and diversity.
  • Recent proposals apply LLM-based neuro-symbolic search, coupling reasoning and optimization, to encode domain priors and improve robustness and interpretability in feature construction.

Empirical Strengths, Limitations, and Open Problems

Empirical Evidence:

  • DL and RL-based pipelines achieve strong risk-adjusted returns (Sharpe ratios > 2.0 in certain multi-asset RL and agent systems) and robust outperformance in diverse markets and stress scenarios.
  • Agent frameworks integrating LLMs and multi-modal reasoning report improvements in cumulative return (CR > 80% in single-asset, CR > 100% in portfolio-level backtests) and dramatically reduced maximum drawdown under adversarial market shifts.

Limitations:

  • LLM predictors face a latent misalignment between linguistic sentiment and realized market response, exacerbated by the prevalence of confounding economic drivers not fully represented in text.
  • Latency and computational cost of LLM inference restrict deployment to low-frequency, batch, or advisory settings; hybrid architectures integrating low-latency ML modules for execution-critical tasks are necessary in high-frequency domains.
  • LLM agent systems lag institutional standards in portfolio optimization. Risk modeling, transaction cost analysis, and real-time execution remain governed by classical/ML modules invoked by the agent rather than natively reasoned in natural language.
  • End-to-end, fully differentiable training spanning the investment pipeline with LLM-integrated modules has limited research and lacks scalable industrial implementations. Figure 8

    Figure 8: The overall framework of this paper.

Implications and Future Research Directions

  1. Hybrid Architectures—Continued development of neuro-symbolic and hybrid agent architectures with explicit partitioning of tasks between natural language reasoning (problem decomposition, hypothesis generation, reporting) and numerical/ML modules (risk, allocation, execution).
  2. Market-Aware Sentiment and Causal Reasoning—LLM fine-tuning and reinforcement learning based on asset price reactions to news, iterative self-critique, and structural estimation should be prioritized for closing the gap between linguistic and economic sentiment.
  3. Real-Time, Multi-Agent Systems—Building composable multi-agent orchestration platform with adaptive role assignments and feedback-driven learning. Exploiting explicit communication protocols to synthesize insights across agents handling technical, fundamental, and alternative data.
  4. AutoML and Continual Learning—Automating hyperparameter tuning, feature search, and meta-learning to maintain model and pipeline robustness under regime change, while integrating continual learning strategies to mitigate catastrophic forgetting.
  5. Interpretability and Auditing—Task-specific explanation systems should link agent outputs to time-localized, signal-traceable rationales, facilitating compliance and model risk management.
  6. Evaluation and Benchmarking—Expansion of public benchmarks, simulation environments, and shared datasets that reflect realistic capital, liquidity, and information constraints is essential for empirical progress.

Conclusion

This survey establishes that the AI-driven transformation of quantitative investment is characterized by rapidly expanding methodological diversity and practical sophistication. Deep learning models have significantly expanded the function approximation and feature extraction capacity within the traditional pipeline, enhancing alpha generation and portfolio construction. LLMs, and their integration into agent-based autonomous workflows, enable new levels of reasoning, multi-modal data fusion, and decision automation, though practical and methodological challenges persist at the intersection of language, numeracy, and market microstructure.

The full realization of agent-based, AI-autonomous investment workflows will require advances in real-time multi-agent reasoning, hybrid neuro-symbolic optimization, pipeline integration, and empirical benchmarking under realistic constraints. These developments will define the next phase of research in both the academic and institutional application of AI in financial markets.

Dice Question Streamline Icon: https://streamlinehq.com

Open Problems

We found no open problems mentioned in this paper.

List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

X Twitter Logo Streamline Icon: https://streamlinehq.com

Tweets

This paper has been mentioned in 10 tweets and received 230 likes.

Upgrade to Pro to view all of the tweets about this paper: