Papers
Topics
Authors
Recent
2000 character limit reached

TabPFN-2.5: Advancing the State of the Art in Tabular Foundation Models (2511.08667v1)

Published 11 Nov 2025 in cs.LG and stat.ML

Abstract: The first tabular foundation model, TabPFN, and its successor TabPFNv2 have impacted tabular AI substantially, with dozens of methods building on it and hundreds of applications across different use cases. This report introduces TabPFN-2.5, the next generation of our tabular foundation model, built for datasets with up to 50,000 data points and 2,000 features, a 20x increase in data cells compared to TabPFNv2. TabPFN-2.5 is now the leading method for the industry standard benchmark TabArena (which contains datasets with up to 100,000 training data points), substantially outperforming tuned tree-based models and matching the accuracy of AutoGluon 1.4, a complex four-hour tuned ensemble that even includes the previous TabPFNv2. Remarkably, default TabPFN-2.5 has a 100% win rate against default XGBoost on small to medium-sized classification datasets (<=10,000 data points, 500 features) and a 87% win rate on larger datasets up to 100K samples and 2K features (85% for regression). For production use cases, we introduce a new distillation engine that converts TabPFN-2.5 into a compact MLP or tree ensemble, preserving most of its accuracy while delivering orders-of-magnitude lower latency and plug-and-play deployment. This new release will immediately strengthen the performance of the many applications and methods already built on the TabPFN ecosystem.

Summary

  • The paper introduces TabPFN-2.5, demonstrating significant scalability and accuracy improvements for tabular data with up to 50,000 samples and 2,000 features.
  • It leverages advanced architectural updates, including deeper network layers and a distillation engine, to boost training speed and inference efficiency.
  • Performance evaluations on the TabArena-lite benchmark show that TabPFN-2.5 consistently outperforms models like AutoGluon and XGBoost while reducing the need for hyperparameter tuning.

TabPFN-2.5: Advancing Tabular Foundation Models

Introduction

The paper "TabPFN-2.5: Advancing the State of the Art in Tabular Foundation Models" (2511.08667) introduces TabPFN-2.5, the next iteration in the series of Tabular Foundation Models (TFMs) with a significant leap in capability. Established on the footing of the previous models, TabPFN and TabPFNv2, TabPFN-2.5 extends the model's ability to handle datasets of up to 50,000 samples and 2,000 features, with a focus on outperforming established models such as AutoGluon and XGBoost.

TabPFN-2.5 underscores considerable uptakes in scalability, speed, and accuracy, further promoting TFMs as viable competitors to classical models in various tabular data applications, including large-scale machine learning tasks. This model eliminates the extensive hyperparameter tuning requirement typical of models like gradient-boosted trees, thanks to a pre-training phase on synthetically generated data, which engenders robust, out-of-the-box performance.

Technological Advancements

TabPFN-2.5 capitalizes on several architectural and methodological advancements:

  • Scalability Enhancements: Improvements to the architecture, such as increased network depth (18 for regression, 24 for classification), and optimized feature group sizes for faster training and inference, support the model's functionality at larger scales.
  • Improved Inference Speed: The introduction of a distillation engine, enabling the transformation of TabPFN-2.5 into a compact Multi-Layer Perceptron (MLP) or tree ensemble, maintains accuracy while significantly reducing latency during deployment.
  • Enhanced Preprocessing: Advanced feature transformations and scaling methods improve robustness, adaptability, and overall performance. Figure 1

    Figure 1: Summary of TabPFN model variants, showcasing improvements in scalability and performance from TabPFNv1 to TabPFN-2.5.

Performance Evaluation

TabPFN-2.5 is subjected to a rigorous evaluation on the TabArena-lite benchmark, which encompasses challenging datasets with up to 100,000 samples. The new model demonstrates a substantial edge not only over previous iterations but also over contemporaries like AutoGluon and XGBoost:

  • Benchmark Dominance: TabPFN-2.5 demonstrates a consistent performance edge, surpassing traditional and ensemble models on both classification and regression tasks.
  • Efficient Computation: Figures illustrate that TabPFN-2.5 manages to achieve these outcomes while offering reduced computational time and resources, particularly notable on high-performance GPU setups. Figure 2

    Figure 2: TabPFN-2.5 performance on the standard TabArena-lite benchmark, illustrating its superior performance over previous models.

Practical Implications

TabPFN-2.5's practical import indicates its readiness for integration into high-stakes, data-intensive environments. With its ability to deliver fast, accurate predictions on substantial datasets without dependence on exhaustive tuning, it stands poised to redefine baseline expectations for tabular data processing. The model's implications reach across domains—ranging from healthcare for diagnostic support, to financial services for risk assessment, manifesting strong multi-domain adaptability. Figure 3

Figure 3: TabArena-Lite results on diverse datasets demonstrating the model's robust performance across various scales.

Future Directions

The advancements in TabPFN-2.5 illuminate potential avenues for exploration. Enhancing dataset capacity towards models that can handle millions of rows remains a critical milestone. Furthermore, integrating domain-specific adaptabilities, such as time-series analysis and more refined multimodal data processing, previews a promising trajectory in the evolution of TFMs.

Conclusion

TabPFN-2.5 capably advances the state of tabular foundation models, providing sophisticated solutions to prevalent challenges in tabular machine learning. Its release not only boosts performance in existing deployments but also sets a course for future iterations to further stretch the limits of scale and application versatility. The developments encapsulated in TabPFN-2.5 underscore a decisive step toward next-generation AI tools that adeptly bridge research with real-world demands.

Whiteboard

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Collections

Sign up for free to add this paper to one or more collections.