Deep-learning typically does not outperform tree-based models on tabular data. Often this may be explained by the small size of such datasets. For images, sound, text, the solution has been pretrained models, leading to foundation models, adapted and reused for many tasks. I will discuss the challenges to bring these ideas to tabular learning, and the progress that we have made, leading to a first tabular foundation model: the CARTE tabular model.
For machine learning on tables, CARTE outperforms the best baselines, including combining XGBoost with large language models