{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,10,30]],"date-time":"2024-10-30T04:26:14Z","timestamp":1730262374264,"version":"3.28.0"},"reference-count":35,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,11,1]],"date-time":"2020-11-01T00:00:00Z","timestamp":1604188800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,11,1]],"date-time":"2020-11-01T00:00:00Z","timestamp":1604188800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,11,1]],"date-time":"2020-11-01T00:00:00Z","timestamp":1604188800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,11,1]]},"DOI":"10.1109\/ieeeconf51394.2020.9443421","type":"proceedings-article","created":{"date-parts":[[2021,6,3]],"date-time":"2021-06-03T21:32:35Z","timestamp":1622755955000},"page":"915-922","source":"Crossref","is-referenced-by-count":2,"title":["Second-Order Guarantees in Federated Learning"],"prefix":"10.1109","author":[{"given":"Stefan","family":"Vlaski","sequence":"first","affiliation":[]},{"given":"Elsa","family":"Rizk","sequence":"additional","affiliation":[]},{"given":"Ali H.","family":"Sayed","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"article-title":"Annealing for distributed global optimization","year":"2019","author":"swenson","key":"ref33"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1137\/18M121784X"},{"key":"ref31","first-page":"1067","article-title":"Gradient descent can take exponential time to escape saddle points","author":"du","year":"2017","journal-title":"Proc of International Conference on Neural Information Processing"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1137\/0329055"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2014.2385046"},{"key":"ref34","doi-asserted-by":"crossref","first-page":"211","DOI":"10.1561\/0400000042","article-title":"The algorithmic foundations of differential privacy","volume":"9","author":"dwork","year":"2014","journal-title":"Foundations and Trends\ufffd in Theoretical Computer Science"},{"key":"ref10","article-title":"Distributed learning in non-convex environments &#x2013; Part II: Polynomial escape from saddle-points","author":"vlaski","year":"2019","journal-title":"IEEE Transactions on Signal Processing"},{"journal-title":"Parallel and Distributed Computation Numerical Methods","year":"1997","author":"bertsekas","key":"ref11"},{"key":"ref12","first-page":"873","article-title":"Distributed delayed stochastic optimization","volume":"24","author":"agarwal","year":"2011","journal-title":"Advances in neural information processing systems"},{"key":"ref13","article-title":"Local SGD converges fast and communicates little","author":"stich","year":"2019","journal-title":"Proc Conference on Learning Representations"},{"article-title":"First analysis of local GD on heterogeneous data","year":"2019","author":"khaled","key":"ref14"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.1137\/S1052623495287022"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2008.2009515"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1561\/2200000051"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/JPROC.2014.2306253"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/TAC.2011.2161027"},{"key":"ref28","first-page":"3219","article-title":"On the convergence properties of a k-step averaging stochastic gradient descent algorithm for non-convex optimization","author":"zhou","year":"2018","journal-title":"Proc International Joint Conference on Artificial Intelligence"},{"key":"ref4","first-page":"797","article-title":"Escaping from saddle points&#x2014;online stochastic gradient for tensor decomposition","author":"ge","year":"2015","journal-title":"Proc Conf Learn Theory"},{"article-title":"Cooperative SGD: A unified framework for the design and analysis of communication-efficient SGD algorithms","year":"2018","author":"wang","key":"ref27"},{"key":"ref3","first-page":"586","article-title":"Deep learning without poor local minima","author":"kawaguchi","year":"2016","journal-title":"Advances in neural information processing systems"},{"article-title":"Stochastic gradient descent escapes saddle points efficiently","year":"2019","author":"jin","key":"ref6"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33015693"},{"key":"ref5","first-page":"1155","article-title":"Escaping saddles with stochastic gradients","author":"daneshmand","year":"2018","journal-title":"Proc International Conference on Machine Learning"},{"article-title":"Second-order guarantees of stochastic gradient descent in non-convex optimization","year":"2019","author":"vlaski","key":"ref8"},{"key":"ref7","first-page":"1192","article-title":"Sharp analysis for nonconvex SGD escaping from saddle points","author":"fang","year":"2019","journal-title":"Proc Conf Learn Theory"},{"key":"ref2","first-page":"192","article-title":"The Loss Surfaces of Multilayer Networks","author":"choromanska","year":"2015","journal-title":"Proc International Conference on Artificial Intelligence and Statistics"},{"article-title":"Distributed learning in non-convex environments &#x2013; Part I: Agreement at a Linear rate","year":"2021","author":"vlaski","key":"ref9"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.4310\/CIS.2020.v20.n3.a4"},{"key":"ref20","first-page":"1273","article-title":"Communication-efficient learning of deep networks from decentralized data","volume":"54","author":"mcmahan","year":"2017","journal-title":"Proc International Conference on Artificial Intelligence and Statistics"},{"key":"ref22","first-page":"429","volume":"2","author":"li","year":"2020","journal-title":"Proceedings of Machine Learning and Systems"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/IEEECONF44664.2019.9049023"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/SPAWC48557.2020.9154327"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICC40277.2020.9148862"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2020.2966273"},{"key":"ref25","first-page":"4424","article-title":"Federated multi-task learning","volume":"30","author":"smith","year":"2017","journal-title":"Advances in neural information processing systems"}],"event":{"name":"2020 54th Asilomar Conference on Signals, Systems, and Computers","start":{"date-parts":[[2020,11,1]]},"location":"Pacific Grove, CA, USA","end":{"date-parts":[[2020,11,4]]}},"container-title":["2020 54th Asilomar Conference on Signals, Systems, and Computers"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9443248\/9443268\/09443421.pdf?arnumber=9443421","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,6,28]],"date-time":"2022-06-28T17:51:52Z","timestamp":1656438712000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9443421\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,11,1]]},"references-count":35,"URL":"https:\/\/doi.org\/10.1109\/ieeeconf51394.2020.9443421","relation":{},"subject":[],"published":{"date-parts":[[2020,11,1]]}}}