Should Bias Always be Eliminated? A Principled Framework to Use Data Bias for OOD Generation
Should Bias Always be Eliminated? A Principled Framework to Use Data Bias for OOD Generation
Most existing methods for adapting models to out-of-distribution (OOD) domains rely on invariant representation learning to eliminate the influence of biased features. However, should bias always be eliminated -- and if not, when should it be retained, and how can it be leveraged? To address these questions, we first present a theoretical analysis that explores the conditions under which biased features can be identified and effectively utilized. Building on this theoretical foundation, we introduce a novel framework that strategically leverages bias to complement invariant representations during inference. The framework comprises two key components that leverage bias in both direct and indirect ways: (1) using invariance as guidance to extract predictive ingredients from bias, and (2) exploiting identified bias to estimate the environmental condition and then use it to explore appropriate bias-aware predictors to alleviate environment gaps. We validate our approach through experiments on both synthetic datasets and standard domain generalization benchmarks. Results consistently demonstrate that our method outperforms existing approaches, underscoring its robustness and adaptability.
Yan Li、Guangyi Chen、Yunlong Deng、Zijian Li、Zeyu Tang、Anpeng Wu、Kun Zhang
计算技术、计算机技术
Yan Li,Guangyi Chen,Yunlong Deng,Zijian Li,Zeyu Tang,Anpeng Wu,Kun Zhang.Should Bias Always be Eliminated? A Principled Framework to Use Data Bias for OOD Generation[EB/OL].(2025-07-22)[2025-08-10].https://arxiv.org/abs/2507.17001.点此复制
评论