Identify the mechanism by which instruction-tuning without tabular exposure improves tabular classification performance
Ascertain the precise mechanism by which instruction-tuning of Large Language Models (e.g., Llama-3-8B fine-tuned on the Alpaca instruction-following dataset without any tabular data) yields substantial improvements on tabular classification tasks, clarifying whether and how general instruction-following capabilities translate into effective handling of serialized table inputs.
References
While the precise mechanism remains unclear, we hypothesize that instruction-tuning equips the model with general capabilities for comprehending task descriptions and following input-output mappings—skills that may prove sufficient for many tabular classification tasks without requiring explicit tabular exposure.
— The Illusion of Generalization: Re-examining Tabular Language Model Evaluation
(2602.04031 - Gorla et al., 3 Feb 2026) in Section: Instruction-Following, Not Tabular Knowledge, Drives Performance (Classification: Instruction-Tuning Dominates)