Multi Agent Reinforcement Learning (MARL) is an emerging field in machine learning where multiple agents learn, simultaneously and in a shared environment, how to optimise a global or local reward signal. MARL has gained significant interest in recent years due to its successful applications in various domains, such as robotics, IoT, and traffic control. Cooperative Many Agent Reinforcement Learning (CMARL) is a relevant subclass of MARL, where thousands of agents work together to achieve a common coordination goal. In this paper, we introduce ScaRLib, a Scala framework relying on state-of-the-art deep learning libraries to support the development of CMARL systems. The framework supports the specification of centralised training and decentralised execution, and it is designed to be easily extensible, allowing to add new algorithms, new types of environments, and new coordination toolchains. This paper describes the main structure and features of ScaRLib and includes basic demonstrations that showcase binding with one such toolchain: ScaFi programming framework and Alchemist simulator can be exploited to enable learning of field-based coordination policies for large-scale systems.

ScaRLib: A Framework for Cooperative Many Agent Deep Reinforcement Learning in Scala / Davide Domini; Filippo Cavallari; Gianluca Aguzzi; Mirko Viroli. - ELETTRONICO. - 13908:(2023), pp. 52-70. (Intervento presentato al convegno 25th IFIP WG 6.1 International Conference on Coordination Models and Language, COORDINATION 2023, held as part of the 18th International Federated Conference on Distributed Computing Techniques, DisCoTec 2023 tenutosi a Lisbon nel 2023) [10.1007/978-3-031-35361-1_3].

ScaRLib: A Framework for Cooperative Many Agent Deep Reinforcement Learning in Scala

Davide Domini;Filippo Cavallari;Gianluca Aguzzi;Mirko Viroli
2023

Abstract

Multi Agent Reinforcement Learning (MARL) is an emerging field in machine learning where multiple agents learn, simultaneously and in a shared environment, how to optimise a global or local reward signal. MARL has gained significant interest in recent years due to its successful applications in various domains, such as robotics, IoT, and traffic control. Cooperative Many Agent Reinforcement Learning (CMARL) is a relevant subclass of MARL, where thousands of agents work together to achieve a common coordination goal. In this paper, we introduce ScaRLib, a Scala framework relying on state-of-the-art deep learning libraries to support the development of CMARL systems. The framework supports the specification of centralised training and decentralised execution, and it is designed to be easily extensible, allowing to add new algorithms, new types of environments, and new coordination toolchains. This paper describes the main structure and features of ScaRLib and includes basic demonstrations that showcase binding with one such toolchain: ScaFi programming framework and Alchemist simulator can be exploited to enable learning of field-based coordination policies for large-scale systems.
2023
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
52
70
ScaRLib: A Framework for Cooperative Many Agent Deep Reinforcement Learning in Scala / Davide Domini; Filippo Cavallari; Gianluca Aguzzi; Mirko Viroli. - ELETTRONICO. - 13908:(2023), pp. 52-70. (Intervento presentato al convegno 25th IFIP WG 6.1 International Conference on Coordination Models and Language, COORDINATION 2023, held as part of the 18th International Federated Conference on Distributed Computing Techniques, DisCoTec 2023 tenutosi a Lisbon nel 2023) [10.1007/978-3-031-35361-1_3].
Davide Domini; Filippo Cavallari; Gianluca Aguzzi; Mirko Viroli
File in questo prodotto:
Eventuali allegati, non sono esposti

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11585/933273
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact