|国家预印本平台
首页|Deep Reinforcement Learning Xiangqi Player with Monte Carlo Tree Search

Deep Reinforcement Learning Xiangqi Player with Monte Carlo Tree Search

Deep Reinforcement Learning Xiangqi Player with Monte Carlo Tree Search

来源:Arxiv_logoArxiv
英文摘要

This paper presents a Deep Reinforcement Learning (DRL) system for Xiangqi (Chinese Chess) that integrates neural networks with Monte Carlo Tree Search (MCTS) to enable strategic self-play and self-improvement. Addressing the underexplored complexity of Xiangqi, including its unique board layout, piece movement constraints, and victory conditions, our approach combines policy-value networks with MCTS to simulate move consequences and refine decision-making. By overcoming challenges such as Xiangqi's high branching factor and asymmetrical piece dynamics, our work advances AI capabilities in culturally significant strategy games while providing insights for adapting DRL-MCTS frameworks to domain-specific rule systems.

Berk Yilmaz、Junyu Hu、Jinsong Liu

计算技术、计算机技术

Berk Yilmaz,Junyu Hu,Jinsong Liu.Deep Reinforcement Learning Xiangqi Player with Monte Carlo Tree Search[EB/OL].(2025-06-18)[2025-07-25].https://arxiv.org/abs/2506.15880.点此复制

评论