|国家预印本平台
首页|FairReason: Balancing Reasoning and Social Bias in MLLMs

FairReason: Balancing Reasoning and Social Bias in MLLMs

FairReason: Balancing Reasoning and Social Bias in MLLMs

来源:Arxiv_logoArxiv
英文摘要

Multimodal Large Language Models (MLLMs) already achieve state-of-the-art results across a wide range of tasks and modalities. To push their reasoning ability further, recent studies explore advanced prompting schemes and post-training fine-tuning. Although these techniques improve logical accuracy, they frequently leave the models' outputs burdened with pronounced social biases. Clarifying how reasoning gains interact with bias mitigation-and whether the two objectives inherently trade off-therefore remains an open and pressing research problem. Our study begins by benchmarking three bias-mitigation strategies-supervised fine-uning (SFT), knowledge distillation (KD), and rule-based reinforcement learning (RL)-under identical conditions, establishing their baseline strengths and weaknesses. Building on these results, we vary the proportion of debias-focused and reasoning-centric samples within each paradigm to chart the reasoning-versus-bias trade-off. Our sweeps reveal a consistent sweet spot: a roughly 1:4 mix trained with reinforcement learning cuts stereotype scores by 10% while retaining 88% of the model's original reasoning accuracy, offering concrete guidance for balancing fairness and capability in MLLMs.

Zhenyu Pan、Yutong Zhang、Jianshu Zhang、Haoran Lu、Haozheng Luo、Yuwei Han、Philip S. Yu、Manling Li、Han Liu

计算技术、计算机技术

Zhenyu Pan,Yutong Zhang,Jianshu Zhang,Haoran Lu,Haozheng Luo,Yuwei Han,Philip S. Yu,Manling Li,Han Liu.FairReason: Balancing Reasoning and Social Bias in MLLMs[EB/OL].(2025-07-30)[2025-08-07].https://arxiv.org/abs/2507.23067.点此复制

评论