|国家预印本平台
首页|ReinboT: Amplifying Robot Visual-Language Manipulation with Reinforcement Learning

ReinboT: Amplifying Robot Visual-Language Manipulation with Reinforcement Learning

ReinboT: Amplifying Robot Visual-Language Manipulation with Reinforcement Learning

来源:Arxiv_logoArxiv
英文摘要

Vision-Language-Action (VLA) models have shown great potential in general robotic decision-making tasks via imitation learning. However, the variable quality of training data often constrains the performance of these models. On the other hand, offline Reinforcement Learning (RL) excels at learning robust policy models from mixed-quality data. In this paper, we introduce Reinforced robot GPT (ReinboT), a novel end-to-end VLA model that integrates the RL principle of maximizing cumulative reward. ReinboT achieves a deeper understanding of the data quality distribution by predicting dense returns that capture the nuances of manipulation tasks. The dense return prediction capability enables the robot to generate more robust decision-making actions, oriented towards maximizing future benefits. Extensive experiments show that ReinboT achieves state-of-the-art performance on the CALVIN mixed-quality dataset and exhibits superior few-shot learning and out-of-distribution generalization capabilities in real-world tasks.

Hongyin Zhang、Zifeng Zhuang、Han Zhao、Pengxiang Ding、Hongchao Lu、Donglin Wang

计算技术、计算机技术自动化技术、自动化技术设备

Hongyin Zhang,Zifeng Zhuang,Han Zhao,Pengxiang Ding,Hongchao Lu,Donglin Wang.ReinboT: Amplifying Robot Visual-Language Manipulation with Reinforcement Learning[EB/OL].(2025-05-12)[2025-06-07].https://arxiv.org/abs/2505.07395.点此复制

评论