JOURNAL ARTICLE

Attention Based Large Scale Multi-agent Reinforcement Learning

Abstract

Learning in large scale Multi-Agent Reinforcement Learning is fundamentally difficult due to the curse of dimensionality. In homogeneous multi-agent setting, mean field theory provides an effective way of scaling MARL to environments with many agents by abstracting other agents to a virtual mean agent, which assumes the impact of each player on the outcome is equal and infinitesimal. However, in some real scenarios, it is only several neighboring agents that affect the decision-making of an agent, need not all other agents. In addition, different neighboring agents may have different degrees of influence on the decision-making of an agent. In this paper, not restricted to homogeneous setting, we propose Adaptive Mean Field Multi-Agent Reinforcement Learning (AMF-MARL), which is based on the attention mechanism and can be used to deal with many agent scenarios in which there may be different influence relationships among agents. Specifically, we firstly derive the mean field approximation with adaptive weight. Then, we propose the Adaptive Mean Field Q-learning (AMF-Q) approach, and describe how to obtain the adaptive weight. Finally, we conduct experiment to study the learning effectiveness of proposed approach.

Keywords:
Reinforcement learning Curse of dimensionality Computer science Artificial intelligence Field (mathematics) Scaling Scale (ratio) Machine learning Mathematics

Metrics

6
Cited By
0.98
FWCI (Field Weighted Citation Impact)
31
Refs
0.74
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Reinforcement Learning in Robotics
Physical Sciences →  Computer Science →  Artificial Intelligence
Adaptive Dynamic Programming Control
Physical Sciences →  Computer Science →  Computational Theory and Mathematics
Complex Systems and Time Series Analysis
Social Sciences →  Economics, Econometrics and Finance →  Economics and Econometrics
© 2026 ScienceGate Book Chapters — All rights reserved.