Effective Communications: A Joint Learning and Communication Framework for Multi-Agent Reinforcement Learning Over Noisy Channels
File(s)TKPRG_JSAC21.pdf (857.92 KB)
Accepted version
Author(s)
Tung, Tze-Yang
Kobus, Szymon
Roig, Joan Pujol
Gunduz, Deniz
Type
Journal Article
Abstract
We propose a novel formulation of the “effectiveness
problem” in communications, put forth by Shannon and Weaver
in their seminal work “The Mathematical Theory of Communication”, by considering multiple agents communicating over
a noisy channel in order to achieve better coordination and
cooperation in a multi-agent reinforcement learning (MARL)
framework. Specifically, we consider a multi-agent partially
observable Markov decision process (MA-POMDP), in which the
agents, in addition to interacting with the environment, can also
communicate with each other over a noisy communication channel. The noisy communication channel is considered explicitly
as part of the dynamics of the environment, and the message
each agent sends is part of the action that the agent can take.
As a result, the agents learn not only to collaborate with each
other but also to communicate “effectively” over a noisy channel.
This framework generalizes both the traditional communication
problem, where the main goal is to convey a message reliably over
a noisy channel, and the “learning to communicate” framework
that has received recent attention in the MARL literature, where
the underlying communication channels are assumed to be errorfree. We show via examples that the joint policy learned using the
proposed framework is superior to that where the communication
is considered separately from the underlying MA-POMDP. This
is a very powerful framework, which has many real world
applications, from autonomous vehicle planning to drone swarm
control, and opens up the rich toolbox of deep reinforcement
learning for the design of multi-user communication systems.
problem” in communications, put forth by Shannon and Weaver
in their seminal work “The Mathematical Theory of Communication”, by considering multiple agents communicating over
a noisy channel in order to achieve better coordination and
cooperation in a multi-agent reinforcement learning (MARL)
framework. Specifically, we consider a multi-agent partially
observable Markov decision process (MA-POMDP), in which the
agents, in addition to interacting with the environment, can also
communicate with each other over a noisy communication channel. The noisy communication channel is considered explicitly
as part of the dynamics of the environment, and the message
each agent sends is part of the action that the agent can take.
As a result, the agents learn not only to collaborate with each
other but also to communicate “effectively” over a noisy channel.
This framework generalizes both the traditional communication
problem, where the main goal is to convey a message reliably over
a noisy channel, and the “learning to communicate” framework
that has received recent attention in the MARL literature, where
the underlying communication channels are assumed to be errorfree. We show via examples that the joint policy learned using the
proposed framework is superior to that where the communication
is considered separately from the underlying MA-POMDP. This
is a very powerful framework, which has many real world
applications, from autonomous vehicle planning to drone swarm
control, and opens up the rich toolbox of deep reinforcement
learning for the design of multi-user communication systems.
Date Issued
2021-08-01
Date Acceptance
2021-08-01
Citation
IEEE Journal on Selected Areas in Communications, 2021, 39 (8), pp.2590-2603
ISSN
0733-8716
Publisher
Institute of Electrical and Electronics Engineers
Start Page
2590
End Page
2603
Journal / Book Title
IEEE Journal on Selected Areas in Communications
Volume
39
Issue
8
Copyright Statement
© 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
Sponsor
Commission of the European Communities
Engineering & Physical Science Research Council (EPSRC)
Identifier
http://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcApp=PARTNER_APP&SrcAuth=LinksAMR&KeyUT=WOS:000673624000024&DestLinkType=FullRecord&DestApp=ALL_WOS&UsrCustomerID=1ba7043ffcc86c417c072aa74d649202
Grant Number
677854
EP/T023600/1
Subjects
Science & Technology
Technology
Engineering, Electrical & Electronic
Telecommunications
Engineering
Noise measurement
Protocols
Channel coding
Semantics
Reinforcement learning
Modulation
Wireless communication
Learning to communicate
reinforcement learning (RL)
multi-agent systems
joint source-channel coding
error correction coding
Publication Status
Published