HMPC-assisted Adversarial Inverse Reinforcement Learning for Smart Home Energy Management
HMPC-assisted Adversarial Inverse Reinforcement Learning for Smart Home Energy Management
This letter proposes an Adversarial Inverse Reinforcement Learning (AIRL)-based energy management method for a smart home, which incorporates an implicit thermal dynamics model. In the proposed method, historical optimal decisions are first generated using a neural network-assisted Hierarchical Model Predictive Control (HMPC) framework. These decisions are then used as expert demonstrations in the AIRL module, which aims to train a discriminator to distinguish expert demonstrations from transitions generated by a reinforcement learning agent policy, while simultaneously updating the agent policy that can produce transitions to confuse the discriminator. The proposed HMPC-AIRL method eliminates the need for explicit thermal dynamics models, prior or predictive knowledge of uncertain parameters, or manually designed reward functions. Simulation results based on real-world traces demonstrate the effectiveness and data efficiency of the proposed method.
Jiadong He、Liang Yu、Zhiqiang Chen、Dawei Qiu、Dong Yue、Goran Strbac、Meng Zhang、Yujian Ye、Yi Wang
热工量测、热工自动控制自动化技术、自动化技术设备计算技术、计算机技术
Jiadong He,Liang Yu,Zhiqiang Chen,Dawei Qiu,Dong Yue,Goran Strbac,Meng Zhang,Yujian Ye,Yi Wang.HMPC-assisted Adversarial Inverse Reinforcement Learning for Smart Home Energy Management[EB/OL].(2025-06-01)[2025-06-16].https://arxiv.org/abs/2506.00898.点此复制
评论