|国家预印本平台
首页|Seeing, Saying, Solving: An LLM-to-TL Framework for Cooperative Robots

Seeing, Saying, Solving: An LLM-to-TL Framework for Cooperative Robots

Seeing, Saying, Solving: An LLM-to-TL Framework for Cooperative Robots

来源:Arxiv_logoArxiv
英文摘要

Increased robot deployment, such as in warehousing, has revealed a need for seamless collaboration among heterogeneous robot teams to resolve unforeseen conflicts. To address this challenge, we propose a novel, decentralized framework for robots to request and provide help. The framework begins with robots detecting conflicts using a Vision Language Model (VLM), then reasoning over whether help is needed. If so, it crafts and broadcasts a natural language (NL) help request using a Large Language Model (LLM). Potential helper robots reason over the request and offer help (if able), along with information about impact to their current tasks. Helper reasoning is implemented via an LLM grounded in Signal Temporal Logic (STL) using a Backus-Naur Form (BNF) grammar to guarantee syntactically valid NL-to-STL translations, which are then solved as a Mixed Integer Linear Program (MILP). Finally, the requester robot chooses a helper by reasoning over impact on the overall system. We evaluate our system via experiments considering different strategies for choosing a helper, and find that a requester robot can minimize overall time impact on the system by considering multiple help offers versus simple heuristics (e.g., selecting the nearest robot to help).

Dan BW Choe、Sundhar Vinodh Sangeetha、Steven Emanuel、Chih-Yuan Chiu、Samuel Coogan、Shreyas Kousik

自动化技术、自动化技术设备计算技术、计算机技术

Dan BW Choe,Sundhar Vinodh Sangeetha,Steven Emanuel,Chih-Yuan Chiu,Samuel Coogan,Shreyas Kousik.Seeing, Saying, Solving: An LLM-to-TL Framework for Cooperative Robots[EB/OL].(2025-05-19)[2025-06-30].https://arxiv.org/abs/2505.13376.点此复制

评论