|国家预印本平台
首页|集装箱码头堆场龙门吊调度强化学习法

集装箱码头堆场龙门吊调度强化学习法

Reinforcement learning method for yard crane scheduling at container terminals

中文摘要英文摘要

研究单台龙门吊最优作业顺序问题,目标是集卡等待时间最小。设计了基于Q学习算法(一种强化学习算法)的求解算法,获得龙门吊在不同状态下的调度策略,提出了应用Q学习算法求解龙门吊最优调度时系统状态、动作规则、学习步长与折扣因子的选择方法。最后,采用数据仿真验证了Q学习算法的有效性,结果表明,当集卡到达间隔较长,以及集卡到达堆场位置较集中时,Q学习算法的结果优于先到先服务、同方向移动、选择最近的集卡等规则。

his paper studies the problem of scheduling a yard crane to perform a given set of loading/unloading jobs with different ready times. The objective is to minimize the sum of job waiting times. Meanwhile a Q-learning based algorithm is designed to develop a decision-making policy on selecting the appropriate dispatching rule at different states. Also, the settings used to apply Q-learning algorithms to the yard crane scheduling, such as states, action, learning step size, and discount rate etc are presented. Lastly, simulation studies are performed to investigate the effect of the proposed approaches. The results show that the Q-learning algorithm can performs better than other dispatching rules such as FCFS(first come first served), UN(uni-directional travel), and NT(nearest truck first served) when the time between yard trailer arrivals to the container yard is long and the distribution of containers’ arrival locations deviates from a uniform distribution.

张璐、曾庆成、杨忠振

水路运输工程自动化技术、自动化技术设备

集装箱码头强化学习龙门吊调度Q学习算法

ontainer terminalReinforcement learningYard crane schedulingQ-learning Algorithms

张璐,曾庆成,杨忠振.集装箱码头堆场龙门吊调度强化学习法[EB/OL].(2009-03-30)[2025-08-16].http://www.paper.edu.cn/releasepaper/content/200903-1107.点此复制

评论