|国家预印本平台
首页|A Study of Value-Aware Eigenoptions

A Study of Value-Aware Eigenoptions

A Study of Value-Aware Eigenoptions

来源:Arxiv_logoArxiv
英文摘要

Options, which impose an inductive bias toward temporal and hierarchical structure, offer a powerful framework for reinforcement learning (RL). While effective in sequential decision-making, they are often handcrafted rather than learned. Among approaches for discovering options, eigenoptions have shown strong performance in exploration, but their role in credit assignment remains underexplored. In this paper, we investigate whether eigenoptions can accelerate credit assignment in model-free RL, evaluating them in tabular and pixel-based gridworlds. We find that pre-specified eigenoptions aid not only exploration but also credit assignment, whereas online discovery can bias the agent's experience too strongly and hinder learning. In the context of deep RL, we also propose a method for learning option-values under non-linear function approximation, highlighting the impact of termination conditions on performance. Our findings reveal both the promise and complexity of using eigenoptions, and options more broadly, to simultaneously support credit assignment and exploration in reinforcement learning.

Harshil Kotamreddy、Marlos C. Machado

计算技术、计算机技术

Harshil Kotamreddy,Marlos C. Machado.A Study of Value-Aware Eigenoptions[EB/OL].(2025-07-12)[2025-07-25].https://arxiv.org/abs/2507.09127.点此复制

评论