{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T18:45:15Z","timestamp":1772909115534,"version":"3.50.1"},"reference-count":57,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"6","license":[{"start":{"date-parts":[[2025,6,1]],"date-time":"2025-06-01T00:00:00Z","timestamp":1748736000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/legalcode"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Artif. Intell."],"published-print":{"date-parts":[[2025,6]]},"DOI":"10.1109\/tai.2025.3534745","type":"journal-article","created":{"date-parts":[[2025,1,28]],"date-time":"2025-01-28T13:55:45Z","timestamp":1738072545000},"page":"1664-1678","source":"Crossref","is-referenced-by-count":2,"title":["COLT: Cyclic Overlapping Lottery Tickets for Faster Pruning of Convolutional Neural Networks"],"prefix":"10.1109","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4665-0530","authenticated-orcid":false,"given":"Md. Ismail","family":"Hossain","sequence":"first","affiliation":[{"name":"Department of Electrical and Computer Engineering, North South University, Dhaka, Bangladesh"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6201-3729","authenticated-orcid":false,"given":"Mohammed","family":"Rakib","sequence":"additional","affiliation":[{"name":"Department of Electrical and Computer Engineering, North South University, Dhaka, Bangladesh"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0246-8999","authenticated-orcid":false,"given":"M. M. Lutfe","family":"Elahi","sequence":"additional","affiliation":[{"name":"Department of Electrical and Computer Engineering, North South University, Dhaka, Bangladesh"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7661-3570","authenticated-orcid":false,"given":"Nabeel","family":"Mohammed","sequence":"additional","affiliation":[{"name":"Department of Electrical and Computer Engineering, North South University, Dhaka, Bangladesh"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7169-0318","authenticated-orcid":false,"given":"Shafin","family":"Rahman","sequence":"additional","affiliation":[{"name":"Department of Electrical and Computer Engineering, North South University, Dhaka, Bangladesh"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1080\/09540098908915626"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/72.80236"},{"key":"ref3","first-page":"1135","article-title":"Learning both weights and connections for efficient neural networks","volume":"1","author":"Han","year":"2015","journal-title":"Proc. 28th Int. Conf. Neural Inf. Process. Syst."},{"key":"ref4","article-title":"Pruning filters for efficient convnets","author":"Li","year":"2017","journal-title":"Proc. Int. Conf. Learn. Represent."},{"key":"ref5","first-page":"598","article-title":"Optimal brain damage","volume":"2","author":"LeCun","year":"1989","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref6","article-title":"Second order derivatives for network pruning: optimal brain surgeon","volume":"5","author":"Hassibi","year":"1992","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref7","article-title":"Distilling the knowledge in a neural network","author":"Hinton","year":"2015"},{"key":"ref8","article-title":"Pruning convolutional neural networks for resource efficient inference","author":"Molchanov","year":"2017","journal-title":"Proc. 5th Int. Conf. Learn. Represent., ICLR 2017"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.643"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.541"},{"key":"ref11","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2019.04.021"},{"key":"ref12","first-page":"1387","article-title":"Dynamic network surgery for efficient dnns","volume":"29","author":"Guo","year":"2016","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref13","first-page":"2498","article-title":"Variational dropout sparsifies deep neural networks","author":"Molchanov","year":"2017","journal-title":"ICML"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2021.108056"},{"key":"ref15","article-title":"To prune, or not to prune: Exploring the efficacy of pruning for model compression","author":"Zhu","year":"2018","journal-title":"Proc. 6th Int. Conf. Learn. Represent (ICLR)"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103511"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2018.06.008"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-19778-9_24"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2021.104192"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2022.103445"},{"key":"ref21","article-title":"The lottery ticket hypothesis: Finding sparse, trainable neural networks","author":"Frankle","journal-title":"Proc. 7th Int. Conf. Learn. Represent. (ICLR)"},{"key":"ref22","first-page":"4932","article-title":"One ticket to win them all: Generalizing lottery ticket initializations across datasets and optimizers","author":"Morcos","year":"2019","journal-title":"NeurIPS"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"ref25","article-title":"Learning multiple layers of features from tiny images","author":"Krizhevsky","year":"2009"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-015-0816-y"},{"key":"ref27","article-title":"Lottery ticket preserves weight correlation: Is it desirable or not?","author":"Liu","year":"2021","journal-title":"Proc. Int. Conf. Mach. Learn."},{"key":"ref28","article-title":"The state of sparsity in deep neural networks","author":"Gale","year":"2019"},{"key":"ref29","article-title":"SNIP: Single-shot network pruning based on connection sensitivity","author":"Lee","year":"2019","journal-title":"Proc. Int. Conf. Learn. Represent."},{"key":"ref30","first-page":"6158","article-title":"Learning and generalization in overparameterized neural networks, going beyond two layers","author":"Allen-Zhu","year":"2019","journal-title":"NeurIPS"},{"key":"ref31","first-page":"242","article-title":"A convergence theory for deep learning via over-parameterization","volume":"97","author":"Allen-Zhu","year":"2019","journal-title":"Proc. 36th Int. Conf. Mach. Learn., Ser. Proc. Mach. Learn. Res."},{"key":"ref32","first-page":"1329","article-title":"On the power of over-parametrization in neural networks with quadratic activation","volume":"80","author":"Du","year":"2018","journal-title":"Proc. 35th Int. Conf. Mach. Learn., ser. Proc. Mach. Learn. Res."},{"key":"ref33","article-title":"Gradient descent provably optimizes over-parameterized neural networks","author":"Du","year":"2019","journal-title":"7th Int. Conf. Learn. Represent. (ICLR)"},{"key":"ref34","article-title":"In search of the real inductive bias: On the role of implicit regularization in deep learning.","author":"Neyshabur","year":"2015","journal-title":"ICLR (Workshop)"},{"key":"ref35","article-title":"The role of over-parametrization in generalization of neural networks","author":"Neyshabur","year":"2019","journal-title":"Proc. Int. Conf. Learn. Represent."},{"key":"ref36","article-title":"3D point cloud network pruning: When some weights do not matter","author":"Biswas","year":"2024","journal-title":"Proc. 35th Brit. Mach. Vision Conf. (BMVC)"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/IJCNN.2019.8852405"},{"key":"ref38","first-page":"12 749","article-title":"Sanity checks for lottery tickets: Does your winning ticket really win the jackpot?","volume":"34","author":"Ma","year":"2021","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00277"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01424-7_27"},{"key":"ref41","first-page":"3320","article-title":"How transferable are features in deep neural networks?","volume":"27","author":"Yosinski","year":"2014","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00907"},{"key":"ref43","first-page":"3597","article-title":"Deconstructing lottery tickets: Zeros, signs, and the supermask","volume":"32","author":"Zhou","year":"2019","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref44","first-page":"6377","article-title":"Pruning neural networks without any data by iteratively conserving synaptic flow","author":"Tanaka","year":"2020","journal-title":"Proc. 34th Int. Conf. Neural Inf. Process. Syst., ser. NIPS\u201920"},{"key":"ref45","article-title":"Tiny ImageNet Visual Recognition Challenge","author":"Le","year":"2015"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-009-0275-4"},{"key":"ref47","article-title":"Very deep convolutional networks for large-scale image recognition","volume":"2015","author":"Simonyan","journal-title":"CoRR"},{"key":"ref48","first-page":"249","article-title":"Understanding the difficulty of training deep feedforward neural networks","author":"Glorot","year":"2010","journal-title":"AISTATS"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2577031"},{"key":"ref50","article-title":"Very deep convolutional networks for large-scale image recognition","author":"Simonyan","year":"2015","journal-title":"3rd Int. Conf. Learn. Represent. (ICLR)"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00748"},{"key":"ref52","first-page":"3259","article-title":"Linear mode connectivity and the lottery ticket hypothesis","author":"Frankle","year":"2020","journal-title":"Proc. Int. Conf. Mach. Learn."},{"key":"ref53","first-page":"20852","article-title":"The generalization-stability tradeoff in neural network pruning","volume":"33","author":"Bartoldson","year":"2020","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"key":"ref54","article-title":"What do compressed deep neural networks forget?","author":"Hooker","year":"2019"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.1088\/1742-6596\/1168\/2\/022022"},{"key":"ref56","first-page":"2972","article-title":"Implicit regularization for optimal sparse recovery","volume":"32","author":"Vaskevicius","year":"2019","journal-title":"Proc. Adv. Neural Inf. Process. Syst."},{"issue":"241","key":"ref57","first-page":"1","article-title":"Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks","volume":"22","author":"Hoefler","year":"2021","journal-title":"J. Mach. Learn. Res."}],"container-title":["IEEE Transactions on Artificial Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/9078688\/11021002\/10855806.pdf?arnumber=10855806","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,23]],"date-time":"2025-08-23T01:09:50Z","timestamp":1755911390000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10855806\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6]]},"references-count":57,"journal-issue":{"issue":"6"},"URL":"https:\/\/doi.org\/10.1109\/tai.2025.3534745","relation":{},"ISSN":["2691-4581"],"issn-type":[{"value":"2691-4581","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,6]]}}}