Multi-Task Multi-Agent Reinforcement Learning With Task-Entity Transformers and Value Decomposition Training

  • Yuanheng Zhu
  • , Shangjing Huang
  • , Binbin Zuo
  • , Dongbin Zhao
  • , Changyin Sun

Research output: Contribution to journalArticlepeer-review

8 Scopus citations

Abstract

Multi-task multi-agent reinforcement learning aims to control multiple agents to perform well on multiple tasks. It encounters three core challenges: the varying number of agents and entities, the disparities in cooperative behaviors among different tasks, and the training imbalance caused by varying task difficulty levels. To address these issues, we propose a novel framework named Task-Entity Transformer Qmix (TETQmix), which employs pretrained language models for task encoding, utilizes proposed Task-Entity Transformer to handle observations across various tasks, and adjusts task learning weights to achieve balanced multi-task training. Task-Entity Transformer not only enables handling multi-task scenarios with varying numbers of agents and entities, but also leverages cross-attention modules to integrate observation and task embeddings, so that each agent can obtain individual values and decisions for multiple tasks. We then utilize a transformer-based mixer to monotonically combine the individual values, and train the whole network’s parameters using temporal-difference errors. To facilitate multi-task training, we define task regret as the difference between the current-stage return and the candidate best one, and adjust the learning weight of each task based on its task regret. Experiments are conducted on both simulated multi-particle environments and real-world multi-robot systems. Compared with existing baselines, our method not only is superior in multi-task learning efficiency, but also shows promising transfer ability on unseen tasks.

Original languageEnglish
Pages (from-to)9164-9177
Number of pages14
JournalIEEE Transactions on Automation Science and Engineering
Volume22
DOIs
StatePublished - 2025
Externally publishedYes

Keywords

  • Multi-agent systems
  • Multi-task learning
  • Pretrained language model
  • Reinforcement learning
  • Transformer

Fingerprint

Dive into the research topics of 'Multi-Task Multi-Agent Reinforcement Learning With Task-Entity Transformers and Value Decomposition Training'. Together they form a unique fingerprint.

Cite this