TY - GEN
T1 - Attention Based Large Scale Multi-agent Reinforcement Learning
AU - Wang, Xiaoqiang
AU - Ke, Liangjun
AU - Zhang, Gewei
AU - Zhu, Dapeng
N1 - Publisher Copyright:
© 2022 IEEE.
PY - 2022
Y1 - 2022
N2 - Learning in large scale Multi-Agent Reinforcement Learning is fundamentally difficult due to the curse of dimensionality. In homogeneous multi-agent setting, mean field theory provides an effective way of scaling MARL to environments with many agents by abstracting other agents to a virtual mean agent, which assumes the impact of each player on the outcome is equal and infinitesimal. However, in some real scenarios, it is only several neighboring agents that affect the decision-making of an agent, need not all other agents. In addition, different neighboring agents may have different degrees of influence on the decision-making of an agent. In this paper, not restricted to homogeneous setting, we propose Adaptive Mean Field Multi-Agent Reinforcement Learning (AMF-MARL), which is based on the attention mechanism and can be used to deal with many agent scenarios in which there may be different influence relationships among agents. Specifically, we firstly derive the mean field approximation with adaptive weight. Then, we propose the Adaptive Mean Field Q-learning (AMF-Q) approach, and describe how to obtain the adaptive weight. Finally, we conduct experiment to study the learning effectiveness of proposed approach.
AB - Learning in large scale Multi-Agent Reinforcement Learning is fundamentally difficult due to the curse of dimensionality. In homogeneous multi-agent setting, mean field theory provides an effective way of scaling MARL to environments with many agents by abstracting other agents to a virtual mean agent, which assumes the impact of each player on the outcome is equal and infinitesimal. However, in some real scenarios, it is only several neighboring agents that affect the decision-making of an agent, need not all other agents. In addition, different neighboring agents may have different degrees of influence on the decision-making of an agent. In this paper, not restricted to homogeneous setting, we propose Adaptive Mean Field Multi-Agent Reinforcement Learning (AMF-MARL), which is based on the attention mechanism and can be used to deal with many agent scenarios in which there may be different influence relationships among agents. Specifically, we firstly derive the mean field approximation with adaptive weight. Then, we propose the Adaptive Mean Field Q-learning (AMF-Q) approach, and describe how to obtain the adaptive weight. Finally, we conduct experiment to study the learning effectiveness of proposed approach.
KW - adaptive mean field approximation
KW - large scale
KW - multi-agent reinforcement learning
UR - https://www.scopus.com/pages/publications/85134881149
U2 - 10.1109/ICAIBD55127.2022.9820093
DO - 10.1109/ICAIBD55127.2022.9820093
M3 - 会议稿件
AN - SCOPUS:85134881149
T3 - 2022 IEEE 5th International Conference on Artificial Intelligence and Big Data, ICAIBD 2022
SP - 112
EP - 117
BT - 2022 IEEE 5th International Conference on Artificial Intelligence and Big Data, ICAIBD 2022
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 5th IEEE International Conference on Artificial Intelligence and Big Data, ICAIBD 2022
Y2 - 27 May 2022 through 30 May 2022
ER -