Distributed learning in non-convex environments-Part I: agreement at a linear rate
File(s)nonconvex.pdf (826.28 KB)
Accepted version
Author(s)
Vlaski, Stefan
Sayed, Ali H
Type
Journal Article
Abstract
Driven by the need to solve increasingly complex optimization problems in signal processing and machine learning, there has been increasing interest in understanding the behavior of gradient-descent algorithms in non-convex environments. Most available works on distributed non-convex optimization problems focus on the deterministic setting where exact gradients are available at each agent. In this work and its Part II, we consider stochastic cost functions, where exact gradients are replaced by stochastic approximations and the resulting gradient noise persistently seeps into the dynamics of the algorithm. We establish that the diffusion learning strategy continues to yield meaningful estimates non-convex scenarios in the sense that the iterates by the individual agents will cluster in a small region around the network centroid. We use this insight to motivate a short-term model for network evolution over a finite-horizon. In Part II of this work, we leverage this model to establish descent of the diffusion strategy through saddle points in O(1/μ) steps, where μ denotes the step-size, and the return of approximately second-order stationary points in a polynomial number of iterations.
Date Issued
2021-01-12
Date Acceptance
2021-01-05
Citation
IEEE Transactions on Signal Processing, 2021, 69, pp.1242-1256
ISSN
1053-587X
Publisher
Institute of Electrical and Electronics Engineers
Start Page
1242
End Page
1256
Journal / Book Title
IEEE Transactions on Signal Processing
Volume
69
Copyright Statement
© 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
Identifier
http://gateway.webofknowledge.com/gateway/Gateway.cgi?GWVersion=2&SrcApp=PARTNER_APP&SrcAuth=LinksAMR&KeyUT=WOS:000622094600008&DestLinkType=FullRecord&DestApp=ALL_WOS&UsrCustomerID=1ba7043ffcc86c417c072aa74d649202
Subjects
Science & Technology
Technology
Engineering, Electrical & Electronic
Engineering
Stochastic optimization
adaptation
non-convex cost
gradient noise
stationary points
distributed optimization
diffusion learning
Publication Status
Published
Date Publish Online
2021-01-12