PyTorch Frame: A Modular Framework for Multi-Modal Tabular Learning (2404.00776v2)
Abstract: We present PyTorch Frame, a PyTorch-based framework for deep learning over multi-modal tabular data. PyTorch Frame makes tabular deep learning easy by providing a PyTorch-based data structure to handle complex tabular data, introducing a model abstraction to enable modular implementation of tabular models, and allowing external foundation models to be incorporated to handle complex columns (e.g., LLMs for text columns). We demonstrate the usefulness of PyTorch Frame by implementing diverse tabular models in a modular way, successfully applying these models to complex multi-modal tabular data, and integrating our framework with PyTorch Geometric, a PyTorch library for Graph Neural Networks (GNNs), to perform end-to-end learning over relational databases.
- Dnf-net: A neural architecture for tabular data. In International Conference on Learning Representations (ICLR), 2021.
- Optuna: A next-generation hyperparameter optimization framework. In ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), pp. 2623–2631, 2019.
- TabNet: Attentive interpretable tabular learning. In AAAI Conference on Artificial Intelligence, 2021.
- Blake, C. L. Uci repository of machine learning databases. http://www. ics. uci. edu/~ mlearn/MLRepository. html, 1998.
- Language models are few-shot learners. volume 33, pp. 1877–1901, 2020.
- Excelformer: A neural network surpassing gbdts on tabular data. arXiv preprint arXiv:2301.02819, 2023a.
- Learning to simulate complex physics with graph networks. In International Conference on Machine Learning (ICML), 2023b.
- XGBoost: A scalable tree boosting system. In ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), pp. 785–794, 2016.
- Electra: Pre-training text encoders as discriminators rather than generators. arXiv preprint arXiv:2003.10555, 2020.
- Codd, E. F. A relational model of data for large shared data banks. Communications of the ACM, 13(6):377–387, 1970.
- Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805, 2018.
- Fast graph representation learning with PyTorch Geometric. arXiv preprint arXiv:1903.02428, 2019.
- Relational deep learning: Graph representation learning on relational databases. arXiv preprint arXiv:2312.04615, 2023.
- Neural message passing for quantum chemistry. In International Conference on Machine Learning (ICML), pp. 1273–1272, 2017.
- Revisiting deep learning models for tabular data. In Advances in Neural Information Processing Systems (NeurIPS), 2021.
- On embeddings for numerical features in tabular deep learning. In Advances in Neural Information Processing Systems (NeurIPS), 2022.
- Tabr: Tabular deep learning meets nearest neighbors. In International Conference on Learning Representations (ICLR), 2024.
- Why do tree-based models still outperform deep learning on typical tabular data? volume 35, pp. 507–520, 2022.
- Array programming with numpy. Nature, 585(7825):357–362, 2020.
- Deep residual learning for image recognition. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770–778, 2016.
- Open graph benchmark: Datasets for machine learning on graphs. In Advances in Neural Information Processing Systems (NeurIPS), 2020.
- Learning backward compatible embeddings. In ACM SIGKDD Conference on Knowledge Discovery and Data Mining (KDD), pp. 3018–3028, 2022.
- TabTransformer: Tabular data modeling using contextual embeddings. arXiv preprint arXiv:2012.06678, 2020.
- Boost then convolve: Gradient boosting meets graph neural networks. arXiv preprint arXiv:2101.08543, 2021.
- Lightgbm: A highly efficient gradient boosting decision tree. volume 30, 2017.
- Semi-supervised classification with graph convolutional networks. In International Conference on Learning Representations (ICLR), 2017.
- Roberta: A robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692, 2019.
- Text and code embeddings by contrastive pre-training. arXiv preprint arXiv:2201.10005, 2022.
- WaveNet: A generative model for raw audio. arXiv preprint arXiv:1609.03499, 2016.
- Neural oblivious decision ensembles for deep learning on tabular data. In International Conference on Learning Representations (ICLR), 2020.
- Catboost: unbiased boosting with categorical features. volume 31, 2018.
- Sentence-bert: Sentence embeddings using siamese bert-networks. 11 2019. URL https://arxiv.org/abs/1908.10084.
- Benchmarking multimodal automl for tabular data with text fields. arXiv preprint arXiv:2111.02705, 2021.
- Masked label prediction: Unified message passing model for semi-supervised classification. arXiv preprint arXiv:2009.03509, 2020.
- Tabular data: Deep learning is not all you need. Information Fusion, 81:84–90, 2022.
- Saint: Improved neural networks for tabular data via row attention and contrastive pre-training. arXiv preprint arXiv:2106.01342, 2021.
- Attention is all you need. arXiv preprint arXiv:1706.03762, 2017.
- Xtab: Cross-table pretraining for tabular transformers. arXiv preprint arXiv:2305.06090, 2023.