Residual pre-training bias and hidden prompt effects in LLM-based trading evaluation
Ascertain whether using evaluation periods outside an LLM’s knowledge cut-off suffices to eliminate pre-training-induced biases in LLM-based trading agents, and determine the extent to which undisclosed system prompts in proprietary APIs modify model outputs in ways that could affect trading decisions.
References
Assuming that all LLMs have been trained on a significant proportion of the internet up to their cut-off date, we cannot preclude the possibility of a biasing factor from their pre-training data affecting their trading performance in the equity markets. Additionally, it is not clear that we can entirely discount this even when using data explicitly outside of their training data knowledge cut-off as we have done in \autoref{sec:model_disc_results}. Moreover, for those LLMs that do not expose their reasoning (via their API), we cannot be sure that system prompts that are unseen have been designed to augment their responses in non-obvious ways.