Dice Question Streamline Icon: https://streamlinehq.com

CPU–GPU Conjecture

Prove or refute the CPU–GPU conjecture that for any deep learning task outcome (e.g., model performance on classification or regression), a model trained on a central processing unit (CPU) can achieve the same result in comparable training time as a model trained on a graphical processing unit (GPU), provided that the CPU-trained model is designed with sufficiently strong mathematical insight.

Information Square Streamline Icon: https://streamlinehq.com

Background

The paper motivates this conjecture by emphasizing that deep learning training can be computationally expensive and that better mathematical understanding of decision boundaries—such as modeling them by low-degree algebraic hypersurfaces—may lead to more efficient models. The authors suggest that principled, insight-driven model design could offset the raw computational advantages typically attributed to GPUs.

They argue that widespread reliance on trial-and-error hyperparameter searches and brute-force compute may be reducible through mathematically informed architectures and training strategies, raising the question of whether GPU performance advantages are fundamentally necessary for achieving state-of-the-art results.

References

This leads to the following conjecture:

CPU-GPU conjecture. Any result achieved by a model trained with graphical processing units (GPU) can be achieved equivalently by a model trained with a central processing unit (CPU) in comparable time, provided that the CPU trained model incorporates sufficient mathematical insight.

The algebra and the geometry aspect of Deep learning (2510.18862 - Aristide, 21 Oct 2025) in Section 1. Introduction