Abstract

In artificial multi-agent systems, the ability to learn collaborative policies is predicated upon the agents’ communication skills: they must be able to encode the information received from the environment and learn how to share it with other agents as required by the task at hand. We present a deep reinforcement learning approach, Connectivity Driven Communication (CDC), that facilitates the emergence of multi-agent collaborative behaviour only through experience. The agents are modelled as nodes of a weighted graph whose state-dependent edges encode pair-wise messages that can be exchanged. We introduce a graph-dependent attention mechanisms that controls how the agents’ incoming messages are weighted. This mechanism takes into full account the current state of the system as represented by the graph, and builds upon a diffusion process that captures how the information flows on the graph. The graph topology is not assumed to be known a priori, but depends dynamically on the agents’ observations, and is learnt concurrently with the attention mechanism and policy in an end-to-end fashion. Our empirical results show that CDC is able to learn effective collaborative policies and can over-perform competing learning algorithms on cooperative navigation tasks.

Details

Title
Learning multi-agent coordination through connectivity-driven communication
Author
Pesce, Emanuele 1   VIAFID ORCID Logo  ; Montana, Giovanni 2 

 University of Warwick, WMG, Coventry, UK (GRID:grid.7372.1) (ISNI:0000 0000 8809 1613) 
 University of Warwick, Department of Statistics, Coventry, UK (GRID:grid.7372.1) (ISNI:0000 0000 8809 1613); University of Warwick, WMG, Coventry, UK (GRID:grid.7372.1) (ISNI:0000 0000 8809 1613); Alan Turing Institute, London, UK (GRID:grid.499548.d) (ISNI:0000 0004 5903 3632) 
Pages
483-514
Publication year
2023
Publication date
Feb 2023
Publisher
Springer Nature B.V.
ISSN
08856125
e-ISSN
15730565
Source type
Scholarly Journal
Language of publication
English
ProQuest document ID
2771811110
Copyright
© The Author(s) 2022. This work is published under http://creativecommons.org/licenses/by/4.0/ (the “License”). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.