Multi-agent deep reinforcement learning has demonstrated significant potential as a promising framework for developing autonomous agents capable of operating within complex, multi-agent environments in a wide range of domains, like robotics, traffic management, and video games. Centralized training with decentralized execution has emerged as the predominant training paradigm, demonstrating significant effectiveness in learning complex policies. However, its reliance on a centralized learner necessitates that agents acquire policies offline and subsequently execute them online. This constraint motivates the exploration of decentralized training methodologies. Despite their greater flexibility, decentralized approaches often face critical challenges, like: slower convergence rates, higher instability and lower performance compared to centralized methods. Therefore, this paper proposes three neighbor-based decentralized training strategies based on the Deep-Q Learning algorithm and investigates their effectiveness as a viable alternative to centralized training. We evaluate experience sharing, k-nearest neighbor averaging, and k-nearest neighbor consensus methods in a cooperative multi-agent environment and compare their performance against centralized training and totally decentralized training. Our results show that neighbor-based methods can achieve comparable performance to centralized training while offering improved scalability and communication efficiency.

Malucelli, N., Domini, D., Aguzzi, G., Viroli, M. (2025). Neighbor-Based Decentralized Training Strategies for Multi-Agent Reinforcement Learning. 1601 Broadway, 10th Floor, NEW YORK, NY, UNITED STATES : Association for Computing Machinery [10.1145/3672608.3707923].

Neighbor-Based Decentralized Training Strategies for Multi-Agent Reinforcement Learning

Domini D.;Aguzzi G.;Viroli M.
2025

Abstract

Multi-agent deep reinforcement learning has demonstrated significant potential as a promising framework for developing autonomous agents capable of operating within complex, multi-agent environments in a wide range of domains, like robotics, traffic management, and video games. Centralized training with decentralized execution has emerged as the predominant training paradigm, demonstrating significant effectiveness in learning complex policies. However, its reliance on a centralized learner necessitates that agents acquire policies offline and subsequently execute them online. This constraint motivates the exploration of decentralized training methodologies. Despite their greater flexibility, decentralized approaches often face critical challenges, like: slower convergence rates, higher instability and lower performance compared to centralized methods. Therefore, this paper proposes three neighbor-based decentralized training strategies based on the Deep-Q Learning algorithm and investigates their effectiveness as a viable alternative to centralized training. We evaluate experience sharing, k-nearest neighbor averaging, and k-nearest neighbor consensus methods in a cooperative multi-agent environment and compare their performance against centralized training and totally decentralized training. Our results show that neighbor-based methods can achieve comparable performance to centralized training while offering improved scalability and communication efficiency.
2025
SAC '25: Proceedings of the 40th ACM/SIGAPP Symposium on Applied Computing
1250
1257
Malucelli, N., Domini, D., Aguzzi, G., Viroli, M. (2025). Neighbor-Based Decentralized Training Strategies for Multi-Agent Reinforcement Learning. 1601 Broadway, 10th Floor, NEW YORK, NY, UNITED STATES : Association for Computing Machinery [10.1145/3672608.3707923].
Malucelli, N.; Domini, D.; Aguzzi, G.; Viroli, M.
File in questo prodotto:
File Dimensione Formato  
3672608.3707923.pdf

accesso aperto

Tipo: Versione (PDF) editoriale / Version Of Record
Licenza: Licenza per Accesso Aperto. Creative Commons Attribuzione - Condividi allo stesso modo (CCBYSA)
Dimensione 1.26 MB
Formato Adobe PDF
1.26 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/1026169
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? 1
social impact