The paper explores the potential of large-scale models and supervised learning to achieve grandmaster-level chess performance without search algorithms.
Using action-values from Stockfish 16, the researchers trained a 270 million parameter transformer model to predict winning probabilities from chess positions.
The model achieved a Lichess blitz Elo of 2895, surpassing AlphaZero and GPT-3.5-turbo-instruct in some metrics without using domain-specific enhancements or search techniques.
The study shows the impact of model and dataset scale on chess AI performance, suggesting a future where complex reasoning could be captured by neural predictors.
The breakthroughs in AI over the past few years have been driven by the application of large-scale models and massive datasets. This paper investigates the application of such paradigms to the game of chess, traditionally dominated by engines using deep search algorithms and large databases of heuristics. Through the lens of supervised learning, the authors of the paper hypothesize that it is possible to achieve strong chess performance purely from learned action-values, without the traditional explicit search algorithms.
The team constructed a dataset by extracting and annotating 10 million games from Lichess with action-values sourced from the elite chess engine Stockfish 16. A transformer model with 270 million parameters was then trained to predict win probabilities for given chess board positions. The researchers systematically studied the effects of varying model sizes and dataset scales to understand their impact on generalization and chess performance.
The model's prowess is commendable, with a Lichess blitz Elo of 2895, which situates it in the grandmaster echelon. It also outperforms AlphaZero's policy and value networks—sans the Monte Carlo Tree Search—and GPT-3.5-turbo-instruct over a select set of metrics, solidifying the notion of neural predictors as strong generalizers. Key contributions include:
A neural predictor trained on millions of annotated board states capturing Stockfish 16's expertise.
A grandmaster-level policy derived from this predictor, capable of tackling challenging chess puzzles, without any domain-specific enhancements or search algorithms.
Demonstrated significance of scale for achieving robust generalization and performance in chess, accompanied by a detailed analysis through rigorous ablations and scaling studies.
While achieving overwhelming success, the study acknowledges the limitations of lacking historical context which traditional engines use for strategic planning, exposing minor weaknesses. These were pragmatically mitigated through workarounds that, however, do not imply a search algorithm. Furthermore, the study clarifies the scale needed for similar neural predictors to bridge the gap between current model performance and oracle-powered engines like Stockfish 16.
The work underlines a shift in AI research, suggesting that highly complex algorithmic reasoning—once deemed exclusive to sophisticated search-based systems—can be distilled into transformer models using supervised learning. As such, transformers are not simply pattern recognition systems, but powerful approximators capable of substituting intricate algorithmic processes. The implications for future research are profound, promising further shifts in how AI is applied across varied cognitive domains.
JAX: composable transformations of Python+NumPy programs, 2018. http://github.com/google/jax.
N. Carlini. Playing chess with LLMs. https://nicholas.carlini.com/writing/2023/chess-llm.html
The DeepMind JAX Ecosystem, 2020. http://github.com/google-deepmind.
Haiku: Sonnet for JAX, 2020. http://github.com/deepmind/dm-haiku.
Justaz. Exact ratings for everyone on lichess. https://lichess.org/@/justaz/blog/exact-ratings-for-everyone-on-lichess/klIoAEAU
Stockfish, 2008. https://stockfishchess.org.