TabSTAR: A Foundation Tabular Model With Semantically Target-Aware Representations

Technion
Main figure

The TabSTAR architecture illustrated. The model processes numerical features, textual features, and all possible target values for classification.

Abstract

While deep learning has achieved remarkable success across many domains, it has historically underperformed on tabular learning tasks, which remain dominated by gradient boosting decision trees (GBDTs). However, recent advancements are paving the way for Tabular Foundation Models, which can leverage real-world knowledge and generalize across diverse datasets, particularly when the data contains free-text. Although incorporating language model capabilities into tabular tasks has been explored, most existing methods utilize static, target-agnostic textual representations, limiting their effectiveness. We introduce TabSTAR: a Foundation Tabular Model with Semantically Target-Aware Representations. TabSTAR is designed to enable transfer learning on tabular data with textual features, with an architecture free of dataset-specific parameters. It unfreezes a pretrained text encoder and takes as input target tokens, which provide the model with the context needed to learn task-specific embeddings. TabSTAR achieves state-of-the-art performance for both medium- and large-sized datasets across known benchmarks of classification tasks with text features, and its pretraining phase exhibits scaling laws in the number of datasets, offering a pathway for further performance improvements.

Get Started with TabSTAR

TabSTAR makes training tabular models effortless. Just install the package, load your data, and let TabSTAR do the work — from preprocessing to prediction.

TabSTAR will be available as a pip package in the next few days. Join our waiting list to be notified when it's ready!

We'll notify you when TabSTAR is available for installation via pip.

TabSTAR Results

Up to 10K examples

Classification performance on small datasets

Above 10K examples

Classification performance on small datasets

TabSTAR performance comparison on tabular classification benchmarks with textual features. X-axis indicates normalized scores (higher is better). Results shown separately for datasets up to 10,000 examples (left) and above 10,000 examples (right). Error bars represent 95% confidence intervals across multiple dataset splits.

BibTeX

@article{arazi2025tabstarf,
      title = {TabSTAR: A Foundation Tabular Model With Semantically Target-Aware Representations},
      author = {Alan Arazi and Eilam Shapira and Roi Reichart},
      journal = {arXiv preprint arXiv:2505.18125},
      year = {2025},
    }