{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T22:51:42Z","timestamp":1740178302560,"version":"3.37.3"},"reference-count":31,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"2","license":[{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2023,4,1]],"date-time":"2023-04-01T00:00:00Z","timestamp":1680307200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Suzhou Foreign Experts Project","award":["E290010201"],"award-info":[{"award-number":["E290010201"]}]},{"DOI":"10.13039\/501100005151","name":"Chinese Academy of Sciences Key Project","doi-asserted-by":"publisher","award":["E21Z010101"],"award-info":[{"award-number":["E21Z010101"]}],"id":[{"id":"10.13039\/501100005151","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Innovation Workstation of Suzhou Institute of Nano-Tech and Nano-Bionics","award":["E010210101"],"award-info":[{"award-number":["E010210101"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61876155"],"award-info":[{"award-number":["61876155"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Jiangsu Science and Technology Program","award":["BE2020006-4"],"award-info":[{"award-number":["BE2020006-4"]}]},{"DOI":"10.13039\/501100000266","name":"Engineering and Physical Sciences Research Council","doi-asserted-by":"publisher","award":["EP\/M026981\/1","EP\/T021063\/1","EP\/T024917\/1"],"award-info":[{"award-number":["EP\/M026981\/1","EP\/T021063\/1","EP\/T024917\/1"]}],"id":[{"id":"10.13039\/501100000266","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Emerg. Top. Comput. Intell."],"published-print":{"date-parts":[[2023,4]]},"DOI":"10.1109\/tetci.2022.3171797","type":"journal-article","created":{"date-parts":[[2022,5,23]],"date-time":"2022-05-23T20:44:00Z","timestamp":1653338640000},"page":"565-577","source":"Crossref","is-referenced-by-count":0,"title":["Towards Faster Training Algorithms Exploiting Bandit Sampling From Convex to Strongly Convex Conditions"],"prefix":"10.1109","volume":"7","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-5311-1482","authenticated-orcid":false,"given":"Yangfan","family":"Zhou","sequence":"first","affiliation":[{"name":"School of Nano-Tech and Nano-Bionics, University of Science and Technology of China, Hefei City, Anhui Province, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3034-9639","authenticated-orcid":false,"given":"Kaizhu","family":"Huang","sequence":"additional","affiliation":[{"name":"Data Science Research Center, Duke Kunshan University, Kunshan, Jiangsu, China"}]},{"given":"Cheng","family":"Cheng","sequence":"additional","affiliation":[{"name":"Suzhou Institute of Nano-Tech and Nano-Bionics (SINANO), Chinese Academy of Sciences, Suzhou City, Jiangsu Province, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6082-4541","authenticated-orcid":false,"given":"Xuguang","family":"Wang","sequence":"additional","affiliation":[{"name":"Suzhou Institute of Nano-Tech and Nano-Bionics (SINANO), Chinese Academy of Sciences, Suzhou City, Jiangsu Province, China"}]},{"given":"Amir","family":"Hussain","sequence":"additional","affiliation":[{"name":"School of Computing, Edinburgh Napier University, Edinburgh, U.K."}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4083-4731","authenticated-orcid":false,"given":"Xin","family":"Liu","sequence":"additional","affiliation":[{"name":"School of Nano-Tech and Nano-Bionics, University of Science and Technology of China, Hefei City, Anhui Province, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1007\/s12559-019-09677-5"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1007\/s12559-021-09985-9"},{"key":"ref3","article-title":"Adam: A method for stochastic optimization","volume-title":"Proc. 3rd Int. Conf. Learn. Representations","author":"Jimmy","year":"2015"},{"key":"ref4","article-title":"On the convergence of adam and beyond","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Reddi","year":"2018"},{"key":"ref5","article-title":"Adaptive gradient methods with dynamic bound of learning rate","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Luo","year":"2019"},{"key":"ref6","first-page":"18795","article-title":"AdaBelief optimizer: Adapting stepsizes by the belief in observed gradients","volume-title":"Proc. Neural Inf. Process. Syst.","author":"Zhuang","year":"2020"},{"key":"ref7","first-page":"1579","article-title":"Fast kernel classifiers with online and active learning","volume":"6","author":"Bordes","year":"2005","journal-title":"J. Mach. Learn. Res."},{"key":"ref8","first-page":"1454","article-title":"Importance sampling tree for large-scale empirical expectation","volume-title":"Proc. 33rd Int. Conf. Mach. Learn.","author":"Canevet","year":"2016"},{"key":"ref9","first-page":"674","article-title":"Stochastic dual coordinate ascent with adaptive probabilities","volume-title":"Proc. 32nd Int. Conf. Mach. Learn.","author":"Csiba","year":"2015"},{"key":"ref10","first-page":"2525","article-title":"Not all samples are created equal: Deep learning with importance sampling","volume-title":"Proc. 35th Int. Conf. Mach. Learn.","author":"Katharopoulos","year":"2018"},{"key":"ref11","first-page":"5393","article-title":"Adam with bandit sampling for deep learning","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Liu","year":"2020"},{"key":"ref12","first-page":"9597","article-title":"Lookahead optimizer: $k$ steps forward, 1 step back","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Zhang","year":"2019"},{"key":"ref13","article-title":"Decoupled weight decay regularization","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Ilya","year":"2019"},{"key":"ref14","first-page":"404","article-title":"Dissecting adam: The sign, magnitude and variance of stochastic gradients","volume-title":"Proc. 35th Int. Conf. Mach. Learn.","author":"Balles","year":"2018"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2020\/452"},{"key":"ref16","article-title":"SAdam: A variant of adam for strongly convex functions","volume-title":"Proc. Int. Conf. Learn. Representations","author":"Wang","year":"2020"},{"key":"ref17","first-page":"1585","article-title":"An efficient algorithm for generalized linear bandit: Online stochastic gradient descent and thompson sampling","volume-title":"Proc. 24th Int. Conf. Artif. Intell. Statist.","volume":"130","author":"Ding","year":"2021"},{"article-title":"Stochastic optimization with bandit sampling","year":"2017","author":"Farnood","key":"ref18"},{"key":"ref19","first-page":"1017","article-title":"Stochastic gradient descent, weighted sampling, and the randomized Kaczmarz algorithm","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","author":"Needell","year":"2014"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1007\/s11590-015-0916-1"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2021.3079112"},{"key":"ref22","first-page":"3407","article-title":"Clustered sampling: Low-variance and improved representativity for clients selection in federated learning","volume-title":"Proc. 38th Int. Conf. Mach. Learn.","author":"Fraboni","year":"2021"},{"key":"ref23","doi-asserted-by":"crossref","first-page":"235","DOI":"10.1023\/A:1013689704352","article-title":"Finite-time analysis of the multiarmed bandit problem","volume":"47","author":"Auer","year":"2002","journal-title":"Mach. Learn."},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"article-title":"Learning multiple layers of features from tiny images","year":"2009","author":"Krizhevsky","key":"ref25"},{"key":"ref26","article-title":"Web download. philadelphia: Linguistic data consortium","author":"Marcus","year":"1995","journal-title":"Treebank-2 LDC95T7"},{"article-title":"Very deep convolutional networks for large-scale image recognition","year":"2014","author":"Karen","key":"ref27"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.243"},{"key":"ref30","first-page":"2574","article-title":"Adaptive sampling probabilities for non-smooth optimization","volume-title":"Proc. 34th Int. Conf. Mach. Learn.","author":"Hongseok","year":"2017"},{"key":"ref31","article-title":"Lecture 6D - A separate, adaptive learning rate for each connection","volume-title":"Slides Lecture Neural Netw. Mach. Learn.","author":"Hinton","year":"2012"}],"container-title":["IEEE Transactions on Emerging Topics in Computational Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/7433297\/10081151\/09780212.pdf?arnumber=9780212","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T23:09:00Z","timestamp":1705964940000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9780212\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,4]]},"references-count":31,"journal-issue":{"issue":"2"},"URL":"https:\/\/doi.org\/10.1109\/tetci.2022.3171797","relation":{},"ISSN":["2471-285X"],"issn-type":[{"type":"electronic","value":"2471-285X"}],"subject":[],"published":{"date-parts":[[2023,4]]}}}