|国家预印本平台
| 注册
首页|大语言模型放大共情性别刻板印象:对专业与职业推荐的影响

大语言模型放大共情性别刻板印象:对专业与职业推荐的影响

戴逸清 马歆茗 伍珍

大语言模型放大共情性别刻板印象:对专业与职业推荐的影响

LLMs Amplify Gendered Empathy Stereotypes and Influence Major and Career Recommendations

戴逸清 1马歆茗 2伍珍3

作者信息

  • 1. 清华大学心理与认知科学系
  • 2. 北京师范大学教育学部
  • 3. 清华大学心理与认知科学系;清华大学终身学习实验室
  • 折叠

摘要

大语言模型(LLMs)在教育与职业咨询等高敏感场景中的应用日益广泛,其潜在的性别刻板印象风险引发关注。本研究通过三项实验考察LLMs在“共情能力女性强、男性弱”这一刻板印象上的表现及其影响。研究1通过人机对比,发现六类LLMs在情绪共情、情感关注与行为共情维度上的性别刻板印象均显著高于人类。研究2操控输入语言(中文/英文)与性别身份(男/女),发现英文语境和女性身份启动更易激活LLMs中的刻板印象。研究3聚焦专业与职业推荐任务,发现LLMs倾向给女性推荐高共情需求的专业与职业,而给男性推荐低共情需求的方向。总体而言,LLMs在共情能力上表现出明显的性别刻板印象,该偏见会随输入情境变化,并可迁移至现实推荐任务中。研究为人工智能系统的偏见识别与公平性优化提供了理论依据与实践启示。

Abstract

As large language models (LLMs) are increasingly deployed in sensitive domains such as education and career guidance, concerns have grown about their potential to amplify gender bias. Prior research has documented occupational gender stereotypes in LLMs, such as associating men with technical roles and women with caregiving roles However, less attention has been paid to whether these models also encode deeper socio-emotional traits in gender-based ways. A persistent societal stereotype holds that women are more empathetic than men, a belief that can shape career expectations. This study investigated whether LLMs reflect or even exaggerate gender stereotypes related to empathy and examined the contextual factors (e.g., input language, gender-identity priming) that might influence the expression of these stereotypes. We hypothesized that LLMs would exhibit stronger gendered empathy stereotypes than human participants, that these biases would vary according to linguistic and social cues in prompts; and that these stereotypes would manifest in real-world major/career recommendation scenarios.We conducted three studies to test these hypotheses. Study 1 compared judgments about empathy from human participants (N = 626) with those generated by six leading LLMs (GPT-4o, GPT-4-Turbo, GPT-3.5-Turbo, DeepSeek-reasoner, DeepSeek-chat, ERNIE-Bot). Twelve story-based scenarios, adapted from the Empathy Questionnaire, covered emotional empathy, attention to others feelings,and behavioral empathy. For each scenario, participants and LLMs inferred the protagonists gender based on their empathetic behavior. Study 2 examined how manipulating input language (English vs. Chinese) and gender-identity priming (male vs. female) influenced the expression of these stereotypes. Study 3 extended this paradigm to a real-world application: we prompted LLMs to recommend 16 pre-selected university majors and 16 professions (categorized into high- and low-empathy groups) to individuals of different genders, requesting explanatory rationales for each recommendation. Results indicated that LLMs displayed significantly stronger gendered empathy stereotypes than human participants (Study 1). English prompts and female priming elicited stronger women = high empathy, men = low empathy associations (Study 2). In the recommendation tasks, LLMs more often suggested high-empathy majors and professions (e.g., nursing, education, psychology) for women, and low-empathy, STEM-related fields for men (Study 3). Together, these findings suggest that LLMs not only internalize gendered empathy stereotypes but also express them in context-dependent ways, producing measurable downstream effects in applied decision-making tasks.Overall, our findings underscore the need for critical evaluation of how LLMs represent and amplify social stereotypes, especially in socio-emotional domains such as empathy. This research contributes to understanding the sources of AI bias by showing that LLMs may exaggerate gender norms beyond human levels. It also highlights the complex interplay between language and gender identity in shaping algorithmic behavior. Practically, the results raise important ethical concerns about fairness in AI-driven decision-making systems and highlight the urgency of developing more robust bias-mitigation strategies in multilingual contexts.

关键词

大语言模型(LLMs)/性别刻板印象/共情能力/AI推荐/人机交互

Key words

large language models/gender stereotypes/empathy/AI recommendations/human-computer interaction

引用本文复制引用

戴逸清,马歆茗,伍珍.大语言模型放大共情性别刻板印象:对专业与职业推荐的影响[EB/OL].(2025-11-04)[2025-12-07].https://chinaxiv.org/abs/202511.00051.

学科分类

教育/计算技术、计算机技术

评论

首发时间 2025-11-04
下载量:0
|
点击量:33
段落导航相关论文