LinguaFluid: Language Guided Fluid Control via Semantic Rewards in Reinforcement Learning
LinguaFluid: Language Guided Fluid Control via Semantic Rewards in Reinforcement Learning
In the domain of scientific machine learning, designing effective reward functions remains a challenge in reinforcement learning (RL), particularly in environments where task goals are difficult to specify numerically. Reward functions in existing work are predominantly based on heuristics, manual engineering, or task-specific tuning. In this work, we introduce a semantically aligned reinforcement learning method where rewards are computed by aligning the current state with a target semantic instruction using a Sentence-Bidirectional Encoder Representations from Transformers (SBERT). Instead of relying on manually defined reward functions, the policy receives feedback based on the reward, which is a cosine similarity between the goal textual description and the statement description in the episode. We evaluated our approach in several environments and showed that semantic reward can guide learning to achieve competitive control behavior, even in the absence of hand-crafted reward functions. Our study demonstrates a correlation between the language embedding space and the conventional Euclidean space. This framework opens new horizons for aligning agent behavior with natural language goals and lays the groundwork for a more seamless integration of larger language models (LLMs) and fluid control applications.
Aoming Liang、Chi Cheng、Dashuai Chen、Boai Sun、Dixia Fan
语言学计算技术、计算机技术
Aoming Liang,Chi Cheng,Dashuai Chen,Boai Sun,Dixia Fan.LinguaFluid: Language Guided Fluid Control via Semantic Rewards in Reinforcement Learning[EB/OL].(2025-08-14)[2025-08-24].https://arxiv.org/abs/2508.05977.点此复制
评论