Chain of Thought Still Thinks Fast: APriCoT Helps with Thinking Slow
Chain of Thought Still Thinks Fast: APriCoT Helps with Thinking Slow
Language models are known to absorb biases from their training data, leading to predictions driven by statistical regularities rather than semantic relevance. We investigate the impact of these biases on answer choice preferences in the Massive Multi-Task Language Understanding (MMLU) task. Our findings show that these biases are predictive of model preference and mirror human test-taking strategies even when chain of thought (CoT) reasoning is used. To address this issue, we introduce Counterfactual Prompting with Agnostically Primed CoT (APriCoT). We demonstrate that while Counterfactual Prompting with CoT alone is insufficient to mitigate bias, APriCoT effectively reduces the influence of base-rate probabilities while improving overall accuracy. Our results suggest that mitigating bias requires a slow thinking process which CoT alone may not provide as it tends to reinforce fast thinking model bias under some prompting methodologies. APriCoT is a step toward developing more robust and fair language models that can think slow.
Kyle Moore、Jesse Roberts、Thao Pham、Douglas Fisher
计算技术、计算机技术
Kyle Moore,Jesse Roberts,Thao Pham,Douglas Fisher.Chain of Thought Still Thinks Fast: APriCoT Helps with Thinking Slow[EB/OL].(2025-08-11)[2025-08-24].https://arxiv.org/abs/2408.08651.点此复制
评论