Submission Type: Short paper (4 pages)
Keywords: Agents, Decision Trees, Interpretability
TL;DR: We use LLMs as agents to build decision trees for small tabular datasets, this approach outperforms heuristic decision tree construction and provides an interpretable and steerable alternative to more performant SoTA methods
Abstract: Tabular foundation models are becoming increasingly popular for low-resource tabular problems. These models make up for small training datasets by pretraining on large volumes of synthetic data. The prior knowledge obtained via pretraining provides the exceptional performance, but the resulting model becomes a black box that is difficult to interpret and costly to inference.
In this work, we explore an alternative strategy: using reasoning-capable LLMs to induce decision trees for small tabular datasets in agentic setup.
We design a minimal set of tools for constructing, analyzing and manipulating decision trees.
By using these tools, LLMs combine their prior knowledge with learning from data to create a lightweight decision tree that outperforms traditional CART on low-resource tabular problems.
While a single decision tree does not outperform state-of-the-art black box models, it comes with a human-readable reasoning trace that can be checked for biases and data leaks. Furthermore, the reasoning-based LLM's creation process allows for additional human input: correcting biases or incorporating domain-specific intuition that is not captured in the data.
Submission Number: 28
Loading