Hierarchical multi-agent reinforcement learning for cooperative tasks with sparse rewards in continuous domain

Research output: Contribution to journalArticlepeer-review

4 Scopus citations

Abstract

The sparse reward problem has long been one of the most challenging topics in the application of reinforcement learning (RL), especially in complex multi-agent systems. In this paper, a hierarchical multi-agent RL architecture is developed to address the sparse reward problem of cooperative tasks in continuous domain. The proposed architecture is divided into two levels: the higher-level meta-agent implements state transitions on a larger time scale to alleviate the sparse reward problem, which receives global observation as spatial information and formulates sub-goals for the lower-level agents; the lower-level agent receives local observation and sub-goal and completes the cooperative tasks. In addition, to improve the stability of the higher-level policy, a channel is built to transmit the lower-level policy to the meta-agent as temporal information, and then a two-stream structure is adopted in the actor-critic networks of the meta-agent to process spatial and temporal information. Simulation experiments on different tasks demonstrate that the proposed algorithm effectively alleviates the sparse reward problem, so as to learn desired cooperative policies.

Original languageEnglish
Pages (from-to)273-287
Number of pages15
JournalNeural Computing and Applications
Volume36
Issue number1
DOIs
StatePublished - Jan 2024
Externally publishedYes

Keywords

  • Cooperative multi-agent systems
  • Hierarchical framework
  • Reinforcement learning
  • Sparse reward
  • Two-stream structure

Fingerprint

Dive into the research topics of 'Hierarchical multi-agent reinforcement learning for cooperative tasks with sparse rewards in continuous domain'. Together they form a unique fingerprint.

Cite this