{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,14]],"date-time":"2025-10-14T07:11:27Z","timestamp":1760425887081,"version":"3.28.0"},"reference-count":52,"publisher":"IEEE","license":[{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2020,7,1]],"date-time":"2020-07-01T00:00:00Z","timestamp":1593561600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020,7]]},"DOI":"10.1109\/ijcnn48605.2020.9207467","type":"proceedings-article","created":{"date-parts":[[2020,9,30]],"date-time":"2020-09-30T00:40:33Z","timestamp":1601426433000},"page":"1-10","source":"Crossref","is-referenced-by-count":2,"title":["Budgeted Subset Selection for Fine-tuning Deep Learning Architectures in Resource-Constrained Applications"],"prefix":"10.1109","author":[{"given":"Subhankar","family":"Banerjee","sequence":"first","affiliation":[]},{"given":"Shayok","family":"Chakraborty","sequence":"additional","affiliation":[]}],"member":"263","reference":[{"key":"ref39","article-title":"Learning multiple layers of features from tiny images","author":"krizhevsky","year":"2009","journal-title":"Technical Report"},{"key":"ref38","article-title":"Reading digits in natural images with unsupervised feature learning","author":"netzer","year":"2011","journal-title":"Proc NIPS Workshop on Deep Learning and Unsupervised Feature Learning"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2015.2389848"},{"key":"ref32","doi-asserted-by":"crossref","DOI":"10.1007\/978-3-031-01560-1","author":"settles","year":"2012","journal-title":"Active Learning"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1109\/ICIP.2019.8803096"},{"key":"ref30","article-title":"Accurate, large minibatch sgd: Training imagenet in 1 hour","author":"goyal","year":"2017","journal-title":"arXiv 1706 02677"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref36","article-title":"Truncated power method for sparse eigenvalue problems","volume":"14","author":"yuan","year":"2013","journal-title":"Journal of Machine Learning Research (JMLR)"},{"journal-title":"The Random Projection Method","year":"2004","author":"vempala","key":"ref35"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICDM.2015.140"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.97"},{"article-title":"Ai and compute","year":"2018","author":"amodei","key":"ref27"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.634"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2577031"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"article-title":"Mobilenets: Efficient convolutional neural networks for mobile vision applications","year":"2017","author":"howard","key":"ref20"},{"key":"ref22","article-title":"Fast training of convolutional networks through ffts","author":"mathieu","year":"2013","journal-title":"Arxiv 1312 5851"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2015.327"},{"key":"ref24","article-title":"Batch normalization: Accelerating deep network training by reducing internal covariate shift","author":"ioffe","year":"2015","journal-title":"arXiv 1502 03167"},{"key":"ref23","article-title":"Efficient training of very deep neural networks for supervised hashing","author":"zhang","year":"2015","journal-title":"IEEE Conference on Computer Vision and Pattern Recognition (CVPR)"},{"key":"ref26","article-title":"Horovod: Fast and easy distributed deep learning in TensorFlow","author":"sergeev","year":"2018","journal-title":"arXiv 1802 05799"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/SMC.2017.8122736"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00456"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/MSP.2010.936019"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2016.2541146"},{"key":"ref10","article-title":"Prioritized experience replay","author":"schaul","year":"2016","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref11","article-title":"Not all samples are created equal: Deep learning with importance sampling","author":"katharopoulos","year":"2018","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"ref12","article-title":"Training deep models faster with robust, approximate importance sampling","author":"johnson","year":"2018","journal-title":"Neural Information Processing Systems (NIPS)"},{"key":"ref13","article-title":"Distilling the knowledge in a neural network","author":"hinton","year":"2014","journal-title":"Neural Information Processing Systems (NIPS) Deep Learning Workshop"},{"key":"ref14","article-title":"Quantized neural networks: Training neural networks with low precision weights and activations","volume":"18","author":"hubara","year":"2017","journal-title":"Journal of Machine Learning Research"},{"key":"ref15","article-title":"Training wide residual networks for deployment using a single bit for each weight","author":"mcdonnell","year":"2018","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref16","article-title":"Speeding up convolutional neural networks with low rank expansions","author":"jaderberg","year":"2014","journal-title":"British Machine Vision Conference (BMVC)"},{"key":"ref17","article-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding","author":"han","year":"2016","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01237-3_12"},{"article-title":"Network in network","year":"2013","author":"lin","key":"ref19"},{"key":"ref4","article-title":"ImageNet classification with deep convolutional neural networks","author":"krizhevsky","year":"2012","journal-title":"Neural Information Processing Systems (NIPS)"},{"key":"ref3","article-title":"SegNet: A deep convolutional encoder-decoder architecture for image segmentation","author":"badrinarayanan","year":"2015","journal-title":"IEEE Conference on Computer Vision and Pattern Recognition (CVPR)"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"article-title":"Very deep convolutional networks for large-scale image recognition","year":"2014","author":"simonyan","key":"ref5"},{"key":"ref8","article-title":"Variance reduction in SGD by distributed importance sampling","author":"alain","year":"2015","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref7","article-title":"Large scale distributed deep networks","author":"dean","year":"2012","journal-title":"Advances of Neural Information Processing Systems (NIPS)"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2017.156"},{"key":"ref9","article-title":"Online batch selection for faster training of neural networks","author":"loshchilov","year":"2016","journal-title":"International Conference on Learning Representations (ICLR)"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1007\/BF01588971"},{"key":"ref45","article-title":"Diverse sequential subset selection for supervised video summarization","author":"gong","year":"2014","journal-title":"Neural Information Processing Systems (NIPS)"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2015.7298928"},{"key":"ref47","article-title":"Submodularity in data subset selection and active learning","author":"wei","year":"2015","journal-title":"International Conference on Machine Learning (ICML)"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.1017\/CBO9780511814075"},{"article-title":"Freezeout: Accelerate training by progressively freezing layers","year":"2017","author":"brock","key":"ref41"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1561\/2200000044"},{"key":"ref43","doi-asserted-by":"publisher","DOI":"10.1016\/j.tcs.2009.06.018"}],"event":{"name":"2020 International Joint Conference on Neural Networks (IJCNN)","start":{"date-parts":[[2020,7,19]]},"location":"Glasgow, United Kingdom","end":{"date-parts":[[2020,7,24]]}},"container-title":["2020 International Joint Conference on Neural Networks (IJCNN)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/9200848\/9206590\/09207467.pdf?arnumber=9207467","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,10,8]],"date-time":"2023-10-08T17:39:09Z","timestamp":1696786749000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9207467\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,7]]},"references-count":52,"URL":"https:\/\/doi.org\/10.1109\/ijcnn48605.2020.9207467","relation":{},"subject":[],"published":{"date-parts":[[2020,7]]}}}