Skip to content Skip to navigation
University of Warwick
  • Study
  • |
  • Research
  • |
  • Business
  • |
  • Alumni
  • |
  • News
  • |
  • About

University of Warwick
Publications service & WRAP

Highlight your research

  • WRAP
    • Home
    • Search WRAP
    • Browse by Warwick Author
    • Browse WRAP by Year
    • Browse WRAP by Subject
    • Browse WRAP by Department
    • Browse WRAP by Funder
    • Browse Theses by Department
  • Publications Service
    • Home
    • Search Publications Service
    • Browse by Warwick Author
    • Browse Publications service by Year
    • Browse Publications service by Subject
    • Browse Publications service by Department
    • Browse Publications service by Funder
  • Help & Advice
University of Warwick

The Library

  • Login
  • Admin

Learning multi-agent coordination through connectivity-driven communication

Tools
- Tools
+ Tools

Pesce, Emanuele and Montana, Giovanni (2023) Learning multi-agent coordination through connectivity-driven communication. Machine Learning, 112 . pp. 483-514. doi:10.1007/s10994-022-06286-6 ISSN 2632-2153.

[img]
Preview
PDF
s10994-022-06286-6.pdf - Published Version - Requires a PDF viewer.
Available under License Creative Commons Attribution 4.0.

Download (3313Kb) | Preview
Official URL: https://doi.org/10.1007/s10994-022-06286-6

Request Changes to record.

Abstract

In artificial multi-agent systems, the ability to learn collaborative policies is predicated upon the agents’ communication skills: they must be able to encode the information received from the environment and learn how to share it with other agents as required by the task at hand. We present a deep reinforcement learning approach, Connectivity Driven Communication (CDC), that facilitates the emergence of multi-agent collaborative behaviour only through experience. The agents are modelled as nodes of a weighted graph whose state-dependent edges encode pair-wise messages that can be exchanged. We introduce a graph-dependent attention mechanisms that controls how the agents’ incoming messages are weighted. This mechanism takes into full account the current state of the system as represented by the graph, and builds upon a diffusion process that captures how the information flows on the graph. The graph topology is not assumed to be known a priori, but depends dynamically on the agents’ observations, and is learnt concurrently with the attention mechanism and policy in an end-to-end fashion. Our empirical results show that CDC is able to learn effective collaborative policies and can over-perform competing learning algorithms on cooperative navigation tasks.

Item Type: Journal Article
Subjects: Q Science > Q Science (General)
Q Science > QA Mathematics > QA76 Electronic computers. Computer science. Computer software
Divisions: Faculty of Science, Engineering and Medicine > Science > Statistics
Faculty of Science, Engineering and Medicine > Engineering > WMG (Formerly the Warwick Manufacturing Group)
Library of Congress Subject Headings (LCSH): Reinforcement learning, Multiagent systems, Neural networks (Computer science)
Journal or Publication Title: Machine Learning
Publisher: Springer
ISSN: 2632-2153
Official Date: February 2023
Dates:
DateEvent
February 2023Published
29 December 2022Available
23 November 2022Accepted
Volume: 112
Page Range: pp. 483-514
DOI: 10.1007/s10994-022-06286-6
Status: Peer Reviewed
Publication Status: Published
Access rights to Published version: Open Access (Creative Commons)
Date of first compliant deposit: 6 February 2023
Date of first compliant Open Access: 6 February 2023
RIOXX Funder/Project Grant:
Project/Grant IDRIOXX Funder NameFunder ID
EP/V024868/1[EPSRC] Engineering and Physical Sciences Research Councilhttp://dx.doi.org/10.13039/501100000266

Request changes or add full text files to a record

Repository staff actions (login required)

View Item View Item

Downloads

Downloads per month over past year

View more statistics

twitter

Email us: wrap@warwick.ac.uk
Contact Details
About Us