Dice Question Streamline Icon: https://streamlinehq.com

Residual pre-training bias and hidden prompt effects in LLM-based trading evaluation

Ascertain whether using evaluation periods outside an LLM’s knowledge cut-off suffices to eliminate pre-training-induced biases in LLM-based trading agents, and determine the extent to which undisclosed system prompts in proprietary APIs modify model outputs in ways that could affect trading decisions.

Information Square Streamline Icon: https://streamlinehq.com

Background

The authors note that pre-training data may bias LLM-driven trading even when evaluations use data outside the models’ knowledge cut-offs. They further point out that non-transparent system prompts in some APIs may alter model behavior in ways evaluators cannot observe.

To mitigate these confounders, the paper employs Simudyne Horizon to generate synthetic price paths and news cycles, but the text explicitly states uncertainty about fully discounting pre-training biases and undisclosed system prompt effects.

References

Assuming that all LLMs have been trained on a significant proportion of the internet up to their cut-off date, we cannot preclude the possibility of a biasing factor from their pre-training data affecting their trading performance in the equity markets. Additionally, it is not clear that we can entirely discount this even when using data explicitly outside of their training data knowledge cut-off as we have done in \autoref{sec:model_disc_results}. Moreover, for those LLMs that do not expose their reasoning (via their API), we cannot be sure that system prompts that are unseen have been designed to augment their responses in non-obvious ways.

To Trade or Not to Trade: An Agentic Approach to Estimating Market Risk Improves Trading Decisions (2507.08584 - Emmanoulopoulos et al., 11 Jul 2025) in Section: Context-Aware Backtesting with Simudyne Horizon