Repository logo
  • Log In
    Log in via Symplectic to deposit your publication(s).
Repository logo
  • Communities & Collections
  • Research Outputs
  • Statistics
  • Log In
    Log in via Symplectic to deposit your publication(s).
  1. Home
  2. Faculty of Engineering
  3. Faculty of Engineering
  4. Dif-MAML: Decentralized multi-agent meta-learning
 
  • Details
Dif-MAML: Decentralized multi-agent meta-learning
File(s)
Dif-MAML_Decentralized_Multi-Agent_Meta-Learning.pdf (3.44 MB)
Published version
Author(s)
Kayaalp, Mert
Vlaski, Stefan
Sayed, Ali
Type
Journal Article
Abstract
The objective of meta-learning is to exploit knowledge obtained from observed tasks to improve adaptation to unseen tasks. Meta-learners are able to generalize better when they are trained with a larger number of observed tasks and with a larger amount of data per task. Given the amount of resources that are needed, it is generally difficult to expect the tasks, their respective data, and the necessary computational capacity to be available at a single central location. It is more natural to encounter situations where these resources are spread across several agents connected by some graph topology. The formalism of meta-learning is actually well-suited for this decentralized setting, where the learner benefits from information and computational power spread across the agents. Motivated by this observation, we propose a cooperative fully-decentralized multi-agent meta-learning algorithm, referred to as Diffusion-based MAML or Dif-MAML. Decentralized optimization algorithms are superior to centralized implementations in terms of scalability, robustness, avoidance of communication bottlenecks, and privacy guarantees. The work provides a detailed theoretical analysis to show that the proposed strategy allows a collection of agents to attain agreement at a linear rate and to converge to a stationary point of the aggregate MAML objective even in non-convex environments. Simulation results illustrate the theoretical findings and the superior performance relative to the traditional non-cooperative setting.
Date Issued
2022-01-01
Date Acceptance
2021-12-10
Citation
IEEE Open Journal of Signal Processing, 2022, 3, pp.71-93
URI
http://hdl.handle.net/10044/1/103743
URL
https://ieeexplore.ieee.org/document/9669064
DOI
https://www.dx.doi.org/10.1109/OJSP.2021.3140000
ISSN
2644-1322
Publisher
IEEE
Start Page
71
End Page
93
Journal / Book Title
IEEE Open Journal of Signal Processing
Volume
3
Copyright Statement
© 2022 The Author(s). This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
License URL
http://creativecommons.org/licenses/by/4.0/
Identifier
https://www.webofscience.com/api/gateway?GWVersion=2&SrcApp=PARTNER_APP&SrcAuth=LinksAMR&KeyUT=WOS:000752007300001&DestLinkType=FullRecord&DestApp=ALL_WOS&UsrCustomerID=a2bf6146997ec60c407a63945d4e92bb
Subjects
Adaptation models
BEHAVIOR
Biological system modeling
Data models
Decentralized optimization
diffusion algorithm
distributed learning
Engineering
Engineering, Electrical & Electronic
learning to learn
meta-learning
multi-agent systems
networked agents
NETWORKS
Optimization
Science & Technology
Signal processing
Signal processing algorithms
Task analysis
Technology
Publication Status
Published
Date Publish Online
2022-01-04
About
Spiral Depositing with Spiral Publishing with Spiral Symplectic
Contact us
Open access team Report an issue
Other Services
Scholarly Communications Library Services
logo

Imperial College London

South Kensington Campus

London SW7 2AZ, UK

tel: +44 (0)20 7589 5111

Accessibility Modern slavery statement Cookie Policy

Built with DSpace-CRIS software - Extension maintained and optimized by 4Science

  • Cookie settings
  • Privacy policy
  • End User Agreement
  • Send Feedback