|国家预印本平台
首页|Distribution Shift Aware Neural Tabular Learning

Distribution Shift Aware Neural Tabular Learning

Distribution Shift Aware Neural Tabular Learning

来源:Arxiv_logoArxiv
英文摘要

Tabular learning transforms raw features into optimized spaces for downstream tasks, but its effectiveness deteriorates under distribution shifts between training and testing data. We formalize this challenge as the Distribution Shift Tabular Learning (DSTL) problem and propose a novel Shift-Aware Feature Transformation (SAFT) framework to address it. SAFT reframes tabular learning from a discrete search task into a continuous representation-generation paradigm, enabling differentiable optimization over transformed feature sets. SAFT integrates three mechanisms to ensure robustness: (i) shift-resistant representation via embedding decorrelation and sample reweighting, (ii) flatness-aware generation through suboptimal embedding averaging, and (iii) normalization-based alignment between training and test distributions. Extensive experiments show that SAFT consistently outperforms prior tabular learning methods in terms of robustness, effectiveness, and generalization ability under diverse real-world distribution shifts.

Wangyang Ying、Nanxu Gong、Dongjie Wang、Xinyuan Wang、Arun Vignesh Malarkkan、Vivek Gupta、Chandan K. Reddy、Yanjie Fu

计算技术、计算机技术

Wangyang Ying,Nanxu Gong,Dongjie Wang,Xinyuan Wang,Arun Vignesh Malarkkan,Vivek Gupta,Chandan K. Reddy,Yanjie Fu.Distribution Shift Aware Neural Tabular Learning[EB/OL].(2025-08-27)[2025-09-05].https://arxiv.org/abs/2508.19486.点此复制

评论