|国家预印本平台
首页|Social Debiasing for Fair Multi-modal LLMs

Social Debiasing for Fair Multi-modal LLMs

Social Debiasing for Fair Multi-modal LLMs

来源:Arxiv_logoArxiv
英文摘要

Multi-modal Large Language Models (MLLMs) have dramatically advanced the research field and delivered powerful vision-language understanding capabilities. However, these models often inherit deep-rooted social biases from their training data, leading to uncomfortable responses with respect to attributes such as race and gender. This paper addresses the issue of social biases in MLLMs by i) introducing a comprehensive counterfactual dataset with multiple social concepts (CMSC), which complements existing datasets by providing 18 diverse and balanced social concepts; and ii) proposing a counter-stereotype debiasing (CSD) strategy that mitigates social biases in MLLMs by leveraging the opposites of prevalent stereotypes. CSD incorporates both a novel bias-aware data sampling method and a loss rescaling method, enabling the model to effectively reduce biases. We conduct extensive experiments with four prevalent MLLM architectures. The results demonstrate the advantage of the CMSC dataset and the edge of CSD strategy in reducing social biases compared to existing competing methods, without compromising the overall performance on general multi-modal reasoning benchmarks.

Harry Cheng、Qingpei Guo、Yangyang Guo、Ming Yang、Weili Guan、Liqiang Nie、Tian Gan

计算技术、计算机技术

Harry Cheng,Qingpei Guo,Yangyang Guo,Ming Yang,Weili Guan,Liqiang Nie,Tian Gan.Social Debiasing for Fair Multi-modal LLMs[EB/OL].(2025-08-20)[2025-09-02].https://arxiv.org/abs/2408.06569.点此复制

评论