跳到主要导航 跳到搜索 跳到主要内容

Cooperative modular reinforcement learning for large discrete action space problem

  • Fangzhu Ming
  • , Feng Gao
  • , Kun Liu
  • , Chengmei Zhao
  • Xi'an Jiaotong University

科研成果: 期刊稿件文章同行评审

20 引用 (Scopus)

摘要

Deep reinforcement learning (DRL) has achieved remarkable results on high-dimension state tasks. However, it suffers in hard convergence and low sample efficiency when solving large discrete action space problems. To meet these challenges, we develop a cooperative modular reinforcement learning (CMRL) method to distributedly solve the problems with a large discrete action space. A general yet effective task decomposition method is proposed to decompose the complex decision task in a large action space into multiple decision sub-tasks in small action subsets, using a rule-based action division method. The CMRL method consisting of multiple Critic networks is proposed to settle the multiple sub-tasks, where each Critic network learns a decomposed value function to obtain the local optimal action in a sub-task. The global optimal action is cooperatively chosen by all local optimal actions. Moreover, we propose a new parallel training mechanism, which trains multiple Critic networks with different models and multi-data in parallel. Mathematical properties are proposed to analyze the rationality and superiority of CMRL. Four different simulation experiments are conducted to verify the generality and effectiveness of CMRL for large action space problems. The results show that CMRL has superior performance on training efficiency compared with classical and latest DRL methods while maintaining the accuracy of the solution.

源语言英语
页(从-至)281-296
页数16
期刊Neural Networks
161
DOI
出版状态已出版 - 4月 2023

学术指纹

探究 'Cooperative modular reinforcement learning for large discrete action space problem' 的科研主题。它们共同构成独一无二的指纹。

引用此