Visual moral inference and communication
Visual moral inference and communication
Humans can make moral inferences from multiple sources of input. In contrast, automated moral inference in artificial intelligence typically relies on language models with textual input. However, morality is conveyed through modalities beyond language. We present a computational framework that supports moral inference from natural images, demonstrated in two related tasks: 1) inferring human moral judgment toward visual images and 2) analyzing patterns in moral content communicated via images from public news. We find that models based on text alone cannot capture the fine-grained human moral judgment toward visual stimuli, but language-vision fusion models offer better precision in visual moral inference. Furthermore, applications of our framework to news data reveal implicit biases in news categories and geopolitical discussions. Our work creates avenues for automating visual moral inference and discovering patterns of visual moral communication in public media.
Warren Zhu、Aida Ramezani、Yang Xu
计算技术、计算机技术
Warren Zhu,Aida Ramezani,Yang Xu.Visual moral inference and communication[EB/OL].(2025-04-11)[2025-04-26].https://arxiv.org/abs/2504.11473.点此复制
评论