{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,29]],"date-time":"2025-10-29T06:25:30Z","timestamp":1761719130621,"version":"3.37.3"},"reference-count":70,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"10","license":[{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2018AAA0100400"],"award-info":[{"award-number":["2018AAA0100400"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61876090","61936005","U21B2049"],"award-info":[{"award-number":["61876090","61936005","U21B2049"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Inform. Theory"],"published-print":{"date-parts":[[2022,10]]},"DOI":"10.1109\/tit.2022.3170434","type":"journal-article","created":{"date-parts":[[2022,4,28]],"date-time":"2022-04-28T20:33:14Z","timestamp":1651177994000},"page":"6663-6681","source":"Crossref","is-referenced-by-count":1,"title":["Stability and Risk Bounds of Iterative Hard Thresholding"],"prefix":"10.1109","volume":"68","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7151-8806","authenticated-orcid":false,"given":"Xiao-Tong","family":"Yuan","sequence":"first","affiliation":[{"name":"Cognitive Computing Laboratory, Baidu Research, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5979-8868","authenticated-orcid":false,"given":"Ping","family":"Li","sequence":"additional","affiliation":[{"name":"Cognitive Computing Laboratory, Baidu Research, Washington, WA, USA"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1214\/009053606000000768"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.3150\/bj\/1106314846"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1016\/j.jeconom.2018.05.001"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1093\/ectj\/utaa017"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1017\/S0266466609990636"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1561\/2200000015"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2006.871582"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1201\/b18401"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1137\/S0097539792240406"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1016\/j.acha.2009.04.002"},{"key":"ref11","first-page":"685","article-title":"On iterative hard thresholding methods for high-dimensional M-estimation","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NIPS)","author":"Jain"},{"key":"ref12","article-title":"Training skinny deep neural networks with iterative hard thresholding methods","volume-title":"arXiv:1607.05423","author":"Jin","year":"2016"},{"key":"ref13","first-page":"1","article-title":"Gradient hard thresholding pursuit","volume":"18","author":"Yuan","year":"2018","journal-title":"J. Mach. Learn. Res."},{"key":"ref14","first-page":"1988","article-title":"Efficient stochastic gradient hard thresholding","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NeurIPS)","author":"Zhou"},{"issue":"1","key":"ref15","first-page":"807","article-title":"Greedy sparsity-constrained optimization","volume":"14","author":"Bahmani","year":"2013","journal-title":"J. Mach. Learn. Res."},{"key":"ref16","first-page":"3787","article-title":"Nearly non-expansive bounds for Mahalanobis hard thresholding","volume-title":"Proc. 32nd Conf. Learn. Theory (COLT)","author":"Yuan"},{"key":"ref17","first-page":"127","article-title":"Gradient hard thresholding pursuit for sparsity-constrained optimization","volume-title":"Proc. 31th Int. Conf. Mach. Learn. (ICML)","author":"Yuan"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1109\/TSP.2017.2706181"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58529-7_40"},{"key":"ref20","first-page":"1346","article-title":"Statistical learning with a nuisance component","volume-title":"Proc. 32nd Conf. Learn. Theory (COLT)","author":"Foster"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1214\/009053607000000929"},{"key":"ref22","first-page":"499","article-title":"Stability and generalization","volume":"2","author":"Bousquet","year":"2002","journal-title":"J. Mach. Learn. Res."},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1007\/s10444-004-7634-z"},{"key":"ref24","first-page":"2635","article-title":"Learnability, stability and uniform convergence","volume":"11","author":"Shalev-Shwartz","year":"2010","journal-title":"J. Mach. Learn. Res."},{"key":"ref25","first-page":"744","article-title":"Stability and generalization of learning algorithms that converge to global optima","volume-title":"Proc. 35th Int. Conf. Mach. Learn. (ICML)","author":"Charles"},{"key":"ref26","first-page":"1225","article-title":"Train faster, generalize better: Stability of stochastic gradient descent","volume-title":"Proc. 33rd Int. Conf. Mach. Learn. (ICML)","author":"Hardt"},{"key":"ref27","first-page":"2820","article-title":"Data-dependent stability of stochastic gradient descent","volume-title":"Proc. 35th Int. Conf. Mach. Learn. (ICML)","author":"Kuzborskij"},{"key":"ref28","first-page":"610","article-title":"Sharper bounds for uniformly stable algorithms","volume-title":"Proc. 33rd Conf. Learn. Theory (COLT)","author":"Bousquet"},{"key":"ref29","first-page":"9770","article-title":"Generalization bounds for uniformly stable algorithms","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NeurIPS)","author":"Feldman"},{"key":"ref30","first-page":"1270","article-title":"High probability generalization bounds for uniformly stable algorithms with nearly optimal rate","volume-title":"Proc. 32nd Conf. Learn. Theory (COLT)","author":"Feldman"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1214\/009053605000000282"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1214\/23-aos2258"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511546921"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1017\/9781108627771"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1090\/bull\/1546"},{"key":"ref36","article-title":"18.s997: High dimensional statistics","volume-title":"Lecture Notes","author":"Rigollet","year":"2015"},{"key":"ref37","first-page":"3558","article-title":"Exact recovery of hard thresholding pursuit","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NIPS)","author":"Yuan"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1111\/j.2517-6161.1996.tb02080.x"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1214\/17-AOS1670"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1214\/12-AOS1018"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1214\/07-AOS582"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2009.2016018"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1198\/016214501753382273"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1214\/13-AOS1198"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.1214\/09-AOS729"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1214\/12-STS399"},{"key":"ref47","article-title":"Assumptionless consistency of the lasso","volume-title":"arXiv:1303.5817","author":"Chatterjee","year":"2013"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1214\/17-AOS1637"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.2018.2884963"},{"issue":"23","key":"ref50","first-page":"671","article-title":"Structured sparsity and generalization","volume":"13","author":"Maurer","year":"2012","journal-title":"J. Mach. Learn. Res."},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1137\/100806278"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1145\/1553374.1553417"},{"issue":"1","key":"ref53","first-page":"7650","article-title":"A tight bound of hard thresholding","volume":"18","author":"Shen","year":"2017","journal-title":"J. Mach. Learn. Res."},{"key":"ref54","first-page":"3115","article-title":"On the iteration complexity of support recovery via hard thresholding pursuit","volume-title":"Proc. 34th Int. Conf. Mach. Learn. (ICML)","author":"Shen"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1214\/aos\/1176344196"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/TIT.1979.1056032"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1162\/089976603321780326"},{"key":"ref58","first-page":"1","article-title":"Stochastic convex optimization","volume-title":"Proc. 22nd Conf. Learn. Theory (COLT)","author":"Shalev-Shwartz"},{"key":"ref59","first-page":"2199","article-title":"Smoothness, low noise and fast rates","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NIPS)","author":"Srebro"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1214\/009053606000001019"},{"key":"ref61","article-title":"Smoothing fast iterative hard thresholding algorithm for \u2113\u2080 regularized nonsmooth convex regression problem","volume-title":"arXiv:2104.13107","author":"Wu","year":"2021"},{"key":"ref62","doi-asserted-by":"publisher","DOI":"10.1214\/12-AOS1032"},{"key":"ref63","first-page":"921","article-title":"Lower bounds on the performance of polynomial-time algorithms for sparse linear regression","volume-title":"Proc. 27th Conf. Learn. Theory (COLT)","author":"Zhang"},{"key":"ref64","doi-asserted-by":"publisher","DOI":"10.1214\/12-STS400"},{"article-title":"The lottery ticket hypothesis: Finding sparse, trainable neural networks","volume-title":"Proc. 7th Int. Conf. Learn. Represent. (ICLR)","author":"Frankle","key":"ref65"},{"article-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and Huffman coding","volume-title":"Proc. 4th Int. Conf. Learn. Represent. (ICLR)","author":"Han","key":"ref66"},{"key":"ref67","article-title":"Optimization for deep learning: Theory and algorithms","volume-title":"arXiv:1912.08957","author":"Sun","year":"2019"},{"key":"ref68","first-page":"6155","article-title":"Learning and generalization in overparameterized neural networks, going beyond two layers","volume-title":"Proc. Adv. Neural Inf. Process. Syst. (NeurIPS)","author":"Allen-Zhu"},{"key":"ref69","first-page":"4951","article-title":"Overparameterized nonlinear learning: Gradient descent takes the shortest path?","volume-title":"Proc. 36th Int. Conf. Mach. Learn. (ICML)","author":"Oymak"},{"key":"ref70","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-55566-4_10"}],"container-title":["IEEE Transactions on Information Theory"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/18\/9893495\/09764881.pdf?arnumber=9764881","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,22]],"date-time":"2024-01-22T22:48:08Z","timestamp":1705963688000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9764881\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10]]},"references-count":70,"journal-issue":{"issue":"10"},"URL":"https:\/\/doi.org\/10.1109\/tit.2022.3170434","relation":{},"ISSN":["0018-9448","1557-9654"],"issn-type":[{"type":"print","value":"0018-9448"},{"type":"electronic","value":"1557-9654"}],"subject":[],"published":{"date-parts":[[2022,10]]}}}