{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T18:04:54Z","timestamp":1775066694206,"version":"3.50.1"},"reference-count":114,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2020,10,12]],"date-time":"2020-10-12T00:00:00Z","timestamp":1602460800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,10,12]],"date-time":"2020-10-12T00:00:00Z","timestamp":1602460800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100004962","name":"Conseil R\u00e9gional d\u2019Auvergne","doi-asserted-by":"publisher","award":["1145"],"award-info":[{"award-number":["1145"]}],"id":[{"id":"10.13039\/501100004962","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Sign Process Syst"],"published-print":{"date-parts":[[2021,8]]},"DOI":"10.1007\/s11265-020-01596-1","type":"journal-article","created":{"date-parts":[[2020,10,12]],"date-time":"2020-10-12T06:02:41Z","timestamp":1602482561000},"page":"863-878","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":72,"title":["Deep Model Compression and Architecture Optimization for Embedded Systems: A Survey"],"prefix":"10.1007","volume":"93","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-0545-0345","authenticated-orcid":false,"given":"Anthony","family":"Berthelier","sequence":"first","affiliation":[]},{"given":"Thierry","family":"Chateau","sequence":"additional","affiliation":[]},{"given":"Stefan","family":"Duffner","sequence":"additional","affiliation":[]},{"given":"Christophe","family":"Garcia","sequence":"additional","affiliation":[]},{"given":"Christophe","family":"Blanc","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,10,12]]},"reference":[{"key":"1596_CR1","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E. (2012). Imagenet classification with deep convolutional neural networks, NIPS."},{"key":"1596_CR2","doi-asserted-by":"crossref","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G. (2015). Deep learning. Nature, 521, 436\u2013444.","journal-title":"Nature"},{"key":"1596_CR3","unstructured":"Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale Image recognition. arXiv:1409.1556v6."},{"key":"1596_CR4","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D., Vanhoucke, V., Rabinovich, A. (2015). Going deeper with convolutions. In IEEE conference on computer vision and pattern recognition (pp. 1\u20139).","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"1596_CR5","doi-asserted-by":"crossref","unstructured":"He, K., & Sun, J. (2015). Convolutional neural networks at constrained time cost, 5353\u20135360.","DOI":"10.1109\/CVPR.2015.7299173"},{"key":"1596_CR6","doi-asserted-by":"crossref","unstructured":"Chuangxia, H., Hanfeng, K., Xiaohong, C., Fenghua, W. (2013). An lmi approach for dynamics of switched cellular neural networks with mixed delays, Abstract and Applied Analysis.","DOI":"10.1155\/2013\/870486"},{"key":"1596_CR7","doi-asserted-by":"crossref","unstructured":"Chuangxia, H., Jie, C., Peng, W. (2016). Attractor and boundedness of switched stochastic cohen-grossberg neural networks, Discrete Dynamics in Nature and Society.","DOI":"10.1155\/2016\/4958217"},{"key":"1596_CR8","doi-asserted-by":"crossref","unstructured":"Cheng, Y., Wang, D., Zhou, P., Zhang, T., Member, S. (2018). A Survey of Model Compression and Acceleration for Deep Neural Networks, IEEE Signal Processing Magazine.","DOI":"10.1109\/MSP.2017.2765695"},{"issue":"1","key":"1596_CR9","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1631\/FITEE.1700789","volume":"19","author":"J Cheng","year":"2018","unstructured":"Cheng, J., Wang, P., Li, G., Hu, Q., Lu, H. (2018). Recent Advances in Efficient Computation of Deep Convolutional Neural Networks. Frontiers of Information Technology & Electronic Engineering, 19(1), 64\u201377.","journal-title":"Frontiers of Information Technology & Electronic Engineering"},{"key":"1596_CR10","unstructured":"Dauphin, Y.N., & Bengio, Y. (2013). Big neural networks waste capacity."},{"key":"1596_CR11","unstructured":"Ba, J., & Caruana, R. (2014). Do deep nets really need to be deep?, NIPS, 2654\u20132662."},{"key":"1596_CR12","doi-asserted-by":"crossref","unstructured":"Bucilu\u00c7\u017e, C., Caruana, R., Niculescu-Mizil, A. (2006). Model compression, ACM, 535\u2013541.","DOI":"10.1145\/1150402.1150464"},{"key":"1596_CR13","first-page":"1","volume":"14","author":"G Hinton","year":"2014","unstructured":"Hinton, G., Vinyals, O., Dean, J. (2014). Distilling the Knowledge in a Neural Network. NIPS 2014 Deep Learning Workshop, 14, 1\u20139.","journal-title":"NIPS 2014 Deep Learning Workshop"},{"key":"1596_CR14","doi-asserted-by":"crossref","unstructured":"Chen, Y., Wang, N., Zhang, Z. (2017). Darkrank: Accelerating deep metric learning via cross sample similarities transfer.","DOI":"10.1609\/aaai.v32i1.11783"},{"key":"1596_CR15","unstructured":"Huang, Z., & Wang, N. (2017). Like what you like: Knowledge distill via neuron selectivity transfer."},{"key":"1596_CR16","doi-asserted-by":"crossref","unstructured":"Aguilar, G., Ling, Y., Zhang, Y., Yao, B., Fan, X., Guo, C. (2020). Knowledge distillation from internal representations.","DOI":"10.1609\/aaai.v34i05.6229"},{"key":"1596_CR17","unstructured":"Lee, H., Hwang, S.J., Shin, J. (2020). Self-supervised label augmentation via input transformations, ICML."},{"key":"1596_CR18","unstructured":"M\u00fcller, R., Kornblith, S., Hinton, G. (2019). When does label smoothing help?. In Advances in Neural Information Processing Systems."},{"key":"1596_CR19","doi-asserted-by":"crossref","unstructured":"Weinberger, K., Dasgupta, A., Langford, J., Smola, A., Attenberg, J. (2009). Feature hashing for large scale multitask learning. In Proceedings of the 26th annual international conference on machine learning (pp. 1113\u20131120).","DOI":"10.1145\/1553374.1553516"},{"key":"1596_CR20","unstructured":"Chen, W., Wilson, J., Tyree, S., Weinberger, K., Chen, Y. (2015). Compressing neural networks with the hashing trick, 2285\u20132294."},{"key":"1596_CR21","doi-asserted-by":"crossref","unstructured":"Spring, R., & Shrivastava, A. (2017). Scalable and sustainable deep learning via randomized hashing, ACM, 445\u2013454.","DOI":"10.1145\/3097983.3098035"},{"key":"1596_CR22","unstructured":"Ba, J., & Frey, B. (2013). Adaptive dropout for training deep neural networks, 3084\u20133092."},{"key":"1596_CR23","unstructured":"Gionis, A., Indyk, P., Motwani, R. (1999). Similarity Search in High Dimensions via Hashing (pp. 518\u2013529)."},{"key":"1596_CR24","doi-asserted-by":"crossref","unstructured":"Shinde, R., Goel, A., Gupta, P., Dutta, D. (2010). Similarity search and locality sensitive hashing using ternary content addressable memories. In Proceedings of the 2010 ACM SIGMOD International Conference on Management of data (pp. 375\u2013386).","DOI":"10.1145\/1807167.1807209"},{"key":"1596_CR25","doi-asserted-by":"crossref","unstructured":"Sundaram, N., Turmukhametova, A., Satish, N., Mostak, T., Indyk, P., Madden, S., Dubey, P. (2013). Streaming similarity search over one billion tweets using parallel locality-sensitive hashing. In Proceedings of the VLDB Endowment (pp. 1930\u20131941).","DOI":"10.14778\/2556549.2556574"},{"key":"1596_CR26","doi-asserted-by":"crossref","unstructured":"Huang, Q., Feng, J., Zhang, Y., Fang, Q., Ng, W. (2015). Query-aware locality-sensitive hashing for approximate nearest neighbor search. In Proceedings of the VLDB Endowment (pp. 1\u201312).","DOI":"10.14778\/2850469.2850470"},{"key":"1596_CR27","unstructured":"Shrivastava, A., & Li, P. (2014). Asymmetric lsh (alsh) for sublinear time maximum inner product search (mips)."},{"key":"1596_CR28","unstructured":"Cun, Y.L., Denker, J.S., Solla, S. (1990). Optimal Brain Damage, Advances in Neural Information Processing Systems. arXiv:1011.1669v3."},{"key":"1596_CR29","unstructured":"Hassibi, B., & Stork, D.G. (1993). Second order derivatives for network pruning: Optimal brain surgeon."},{"key":"1596_CR30","unstructured":"Han, S., Pool, J., Tran, J., Dally, W. (2015). Learning both weights and connections for efficient neural network. In Advances in neural information processing systems (pp. 1135\u20131143)."},{"key":"1596_CR31","unstructured":"Molchanov, P., Tyree, S., Karras, T., Aila, T., Kautz, J. (2017). Pruning convolutional neural networks for resource efficient transfer learning, ICLR."},{"issue":"3","key":"1596_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3005348","volume":"13","author":"S Anwar","year":"2017","unstructured":"Anwar, S., Hwang, K., Sung, W. (2017). Structured pruning of deep convolutional neural networks. ACM Journal on Emerging Technologies in Computing Systems, 13(3), 1\u201318.","journal-title":"ACM Journal on Emerging Technologies in Computing Systems"},{"key":"1596_CR33","doi-asserted-by":"crossref","unstructured":"Zhou, H., Alvarez, J.M., Porikli, F. (2016). Less Is More: Towards Compact CNNs. In Computer Vision \u2013 ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part IV Springer International Publishing Cham (pp. 662\u2013677).","DOI":"10.1007\/978-3-319-46493-0_40"},{"key":"1596_CR34","unstructured":"Alvarez, J.M., & Salzmann, M. (2016). Learning the number of neurons in deep networks, 2270\u20132278."},{"key":"1596_CR35","doi-asserted-by":"crossref","unstructured":"Lebedev, V., & Lempitsky, V. (2016). Fast convnets using group-wise brain damage.","DOI":"10.1109\/CVPR.2016.280"},{"key":"1596_CR36","unstructured":"Wen, W., Wu, C., Wang, Y., Chen, Y., Li, H. (2016). Learning structured sparsity in deep neural networks. NIPS, 2082\u20132090."},{"key":"1596_CR37","unstructured":"Li, H., Kadav, A., Durdanovic, I., Samet, H., Graf, H.P. (2017). Pruning filters for efficient ConvNets. arXiv:1608.08710."},{"key":"1596_CR38","doi-asserted-by":"crossref","unstructured":"Luo, J.-H., Wu, J., Lin, W. (2017). Thinet: A filter level pruning method for deep neural network compression, ICCV.","DOI":"10.1109\/ICCV.2017.541"},{"key":"1596_CR39","doi-asserted-by":"crossref","unstructured":"Liu, Z., Li, J., Shen, Z., Huang, G., Yan, S., Zhang, C. (2017). Learning Efficient Convolutional Networks through Network Slimming. arXiv:1708.06519.","DOI":"10.1109\/ICCV.2017.298"},{"key":"1596_CR40","unstructured":"Han, S., Mao, H., Dally, W.J. (2016). Deep Compression - Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding. arXiv:1510.00149."},{"key":"1596_CR41","doi-asserted-by":"crossref","unstructured":"Yu, R., Li, A., Chen, C.-F., Lai, J.-H., Morariu, V.I., Han, X., Gao, M., Lin, C.-Y., Davis, L.S. (2018). NISP: pruning networks using neuron importance score propagation, CVPR.","DOI":"10.1109\/CVPR.2018.00958"},{"key":"1596_CR42","unstructured":"Zhuang, Z., Tan, M., Zhuang, B., Liu, J., Guo, Y., Wu, Q., Huang, J., Zhu, J. (2018). Discrimination-aware channel pruning for deep neural networks. In Advances in Neural Information Processing Systems 31 (pp. 875\u2013886)."},{"key":"1596_CR43","doi-asserted-by":"crossref","unstructured":"He, Y., Liu, P., Wang, Z., Hu, Z., Yang, Y. (2019). Filter pruning via geometric median for deep convolutional neural networks acceleration, CVPR.","DOI":"10.1109\/CVPR.2019.00447"},{"key":"1596_CR44","doi-asserted-by":"crossref","unstructured":"Lin, S., Ji, R., Yan, C., Zhang, B., Cao, L., Ye, Q., Huang, F., Doermann, D.S. (2019). Towards optimal structured CNN pruning via generative adversarial learning.","DOI":"10.1109\/CVPR.2019.00290"},{"key":"1596_CR45","doi-asserted-by":"crossref","unstructured":"Sainath, T.N., Kingsbury, B., Sindhwani, V., Arisoy, E., Ramabhadran, B. (2013). Low-rank matrix factorization for Deep Neural Network training with high-dimensional output targets. In IEEE International Conference on Acoustics, Speech and Signal Processing (pp. 6655\u20136659).","DOI":"10.1109\/ICASSP.2013.6638949"},{"issue":"3","key":"1596_CR46","doi-asserted-by":"publisher","first-page":"455","DOI":"10.1137\/07070111X","volume":"51","author":"TG Kolda","year":"2009","unstructured":"Kolda, T.G, & Bader, B.W. (2009). Tensor decompositions and applications. SIAM Review, 51 (3), 455\u2013500.","journal-title":"SIAM Review"},{"key":"1596_CR47","unstructured":"Cheng, Y., Wang, D., Zhou, P., Zhang, T. (2017). A survey of model compression and acceleration for deep neural networks, arXiv:1710.09282."},{"key":"1596_CR48","unstructured":"Lin, J., Rao, Y., Lu, J., Zhou, J. (2017). Runtime neural pruning."},{"key":"1596_CR49","unstructured":"Gong, Y., Liu, L., Yang, M., Bourdev, L. (2014). Compressing deep convolutional networks using vector quantization."},{"key":"1596_CR50","doi-asserted-by":"crossref","unstructured":"LeCun, Y., Bottou, L., Bengio, Y., Haffner, P. (1998). Gradient-based learning applied to document recognition. In Proceedings of the IEEE (pp. 2278\u20132324).","DOI":"10.1109\/5.726791"},{"key":"1596_CR51","unstructured":"Choi, Y., El-Khamy, M., Lee, J. (2017). Towards the limit of network quantization, ICLR."},{"key":"1596_CR52","unstructured":"Kingma, D., & Ba, J. (2014). Adam: A method for stochastic optimization."},{"issue":"7","key":"1596_CR53","first-page":"2121","volume":"12","author":"J Duchi","year":"2011","unstructured":"Duchi, J., Hazan, E., Singer, Y. (2011). Adaptive subgradient methods for online learning and stochastic optimization. Journal of Machine Learning Research, 12(7), 2121\u20132159.","journal-title":"Journal of Machine Learning Research"},{"key":"1596_CR54","unstructured":"Zeiler, M.D. (2012). ADADELTA: An Adaptive Learning Rate Method 6. arXiv:1212.5701."},{"key":"1596_CR55","unstructured":"Hinton, G.E., Srivastava, N., Swersky, K. (2012). Lecture 6a- overview of mini-batch gradient descent."},{"key":"1596_CR56","unstructured":"Abadi, M. (2015). TensorFlow: Large-scale machine learning on heterogeneous systems. Software available from tensorflow.org."},{"key":"1596_CR57","unstructured":"Ahmad, J., Beers, J., Ciurus, M., Critz, R., Katz, M., Pereira, A., Pringle, M., Rames, J. (2017). ios 11 by tutorials: Learning the new ios apis with swift 4 1 Razeware LLC."},{"key":"1596_CR58","doi-asserted-by":"crossref","unstructured":"Farabet, C., Martini, B., Corda, B., Akselrod, P., Culurciello, E., Lecun, Y. (2011). NeuFlow: A runtime reconfigurable dataflow processor for vision IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops.","DOI":"10.1109\/CVPRW.2011.5981829"},{"key":"1596_CR59","doi-asserted-by":"crossref","unstructured":"Gokhale, V., Jin, J., Dundar, A., Martini, B., Culurciello, E. (2014). A 240 G-ops\/s mobile coprocessor for deep neural networks. IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 696\u2013701.","DOI":"10.1109\/CVPRW.2014.106"},{"key":"1596_CR60","unstructured":"Vanhoucke, V., Senior, A., Mao, M.Z. (2011). Improving the speed of neural networks on cpus. In Deep Learning and Unsupervised Feature Learning Workshop, NIPS 2011."},{"key":"1596_CR61","first-page":"171","volume":"18","author":"A Iwata","year":"1989","unstructured":"Iwata, A., Yoshida, Y., Matsuda, S., Sato, Y., Suzumura, Y. (1989). An artificial neural network accelerator using general purpose 24 bit floating point digital signal processors. Proc. IJCNN, 18, 171\u2013175.","journal-title":"Proc. IJCNN"},{"key":"1596_CR62","doi-asserted-by":"crossref","unstructured":"Hammerstrom, D. (1990). A VLSI architecture for high-performance, low-cost, on-chip learning. In IJCNN International Joint Conference on Neural Networks (pp. 537\u2013544).","DOI":"10.1109\/IJCNN.1990.137621"},{"key":"1596_CR63","doi-asserted-by":"publisher","first-page":"281","DOI":"10.1109\/12.210171","volume":"42","author":"JL Holt","year":"1993","unstructured":"Holt, J.L., & Hwang, J.N. (1993). Finite precision error analysis of neural network hardware implementations. IEEE Transactions on Computers, 42, 281\u2013290.","journal-title":"IEEE Transactions on Computers"},{"key":"1596_CR64","unstructured":"Gupta, S., Agrawal, A., Gopalakrishnan, K., Narayanan, P. (2015). Deep Learning with Limited Numerical Precision. In International Conference on Machine Learning. arXiv:1502.02551(pp. 1737\u20131746)."},{"key":"1596_CR65","unstructured":"Courbariaux, M., Bengio, Y., David, J.-P. (2014). Training deep neural networks with low precision multiplications, ICLR."},{"key":"1596_CR66","doi-asserted-by":"crossref","unstructured":"Williamson, D. (1991). Dynamically scaled fixed point arithmetic. In IEEE Pacific Rim Conference on Communications, Computers and Signal Processing Conference Proceedings (pp. 315\u2013318).","DOI":"10.1109\/PACRIM.1991.160742"},{"key":"1596_CR67","doi-asserted-by":"crossref","unstructured":"Mamalet, F., Roux, S., Garcia, C. (2007). Real-time video convolutional face finder on embedded platforms, Eurasip Journal on Embedded Systems.","DOI":"10.1186\/1687-3963-2007-021724"},{"key":"1596_CR68","doi-asserted-by":"crossref","unstructured":"Roux, S., Mamalet, F., Garcia, C., Duffner, S. (2007). An embedded robust facial feature detector. In Proceedings of the 2007 IEEE Signal Processing Society Workshop, MLSP (pp. 170\u2013175).","DOI":"10.1109\/MLSP.2007.4414301"},{"key":"1596_CR69","unstructured":"Courbariaux, M., Hubara, I., Soudry, D., El-Yaniv, R., Bengio, Y. (2016). Binarized neural networks: Training deep neural networks with weights and activations constrained to+\u20091 or-1."},{"key":"1596_CR70","unstructured":"Lin, X., Zhao, C., Pan, W. (2017). Towards accurate binary convolutional neural network."},{"key":"1596_CR71","unstructured":"Srivastava, N. (2013). Improving Neural Networks with Dropout."},{"key":"1596_CR72","unstructured":"Srivastava, N., Hinton, G.E., Krizhevsky, A., Sutskever, I., Salakhutdinov, R. (2014). Dropout: a simple way to prevent neural networks from overfitting."},{"key":"1596_CR73","unstructured":"Courbariaux, M., Bengio, Y., David, J.-P. (2015). Binaryconnect: Training deep neural networks with binary weights during propagations."},{"key":"1596_CR74","doi-asserted-by":"crossref","unstructured":"Rastegari, M., Ordonez, V., Redmon, J., Farhadi, A. (2016). Xnor-net: Imagenet classification using binary convolutional neural networks Springer.","DOI":"10.1007\/978-3-319-46493-0_32"},{"key":"1596_CR75","doi-asserted-by":"crossref","unstructured":"Newell, A., Yang, K., Deng, J. (2016). Stacked hourglass networks for human pose estimation Springer.","DOI":"10.1007\/978-3-319-46484-8_29"},{"key":"1596_CR76","doi-asserted-by":"crossref","unstructured":"Bulat, A., & Tzimiropoulos, G. (2017). Binarized convolutional landmark localizers for human pose estimation and face alignment with limited resources.","DOI":"10.1109\/ICCV.2017.400"},{"key":"1596_CR77","doi-asserted-by":"crossref","unstructured":"Deng, L., Jiao, P., Pei, J., Wu, Z., Li, G. (2017). Gated xnor networks: Deep neural networks with ternary weights and activations under a unified discretization framework.","DOI":"10.1016\/j.neunet.2018.01.010"},{"key":"1596_CR78","unstructured":"LeCun, Y. (1989). Generalization and network design strategies."},{"key":"1596_CR79","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J. (2016). Deep residual learning for image recognition.","DOI":"10.1109\/CVPR.2016.90"},{"key":"1596_CR80","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Weinberger, K.Q., Maaten, L.VD. (2017). Densely connected convolutional networks.","DOI":"10.1109\/CVPR.2017.243"},{"key":"1596_CR81","unstructured":"Iandola, F.N, Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K. (2016). Squeezenet: Alexnet-level accuracy with 50x fewer parameters and<\u20090.5 mb model size."},{"key":"1596_CR82","unstructured":"Nanfack, G., Elhassouny, A., Thami, R. O.H. (2017). Squeeze-segnet: A new fast deep convolutional neural network for semantic segmentation."},{"key":"1596_CR83","doi-asserted-by":"crossref","unstructured":"Badrinarayanan, V., Kendall, A., Cipolla, R. (2017). SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation, IEEE Transactions on Pattern Analysis and Machine Intelligence.","DOI":"10.1109\/TPAMI.2016.2644615"},{"key":"1596_CR84","doi-asserted-by":"crossref","unstructured":"Brostow, G., Shotton, J., Fauqueur, J., Cipolla, R. (2008). Segmentation and recognition using structure from motion point clouds.","DOI":"10.1007\/978-3-540-88682-2_5"},{"key":"1596_CR85","doi-asserted-by":"crossref","unstructured":"Mamalet, F., & Garcia, C. (2012). Simplifying convnets for fast learning.","DOI":"10.1007\/978-3-642-33266-1_8"},{"key":"1596_CR86","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., Adam, H. (2017). Mobilenets: Efficient convolutional neural networks for mobile vision applications."},{"key":"1596_CR87","doi-asserted-by":"crossref","unstructured":"Chollet, F. (2016). Xception: Deep learning with depthwise separable convolutions. arXiv:1610.02357.","DOI":"10.1109\/CVPR.2017.195"},{"key":"1596_CR88","unstructured":"Sifre, L., & Stephane, M. (2014). Rigid-Motion Scattering For Image Classification."},{"key":"1596_CR89","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J. (2017). Shufflenet: An extremely efficient convolutional neural network for mobile devices.","DOI":"10.1109\/CVPR.2018.00716"},{"key":"1596_CR90","doi-asserted-by":"crossref","unstructured":"Rosenblatt, F. (1962). Perceptrons and the Theory of Brain Mechanics.","DOI":"10.21236\/AD0256582"},{"issue":"1","key":"1596_CR91","doi-asserted-by":"publisher","first-page":"59","DOI":"10.1007\/BF00337288","volume":"43","author":"T Kohonen","year":"1982","unstructured":"Kohonen, T. (1982). Self-organized formation of topologically correct feature maps. Biological Cybernetics, 43(1), 59\u201369.","journal-title":"Biological Cybernetics"},{"issue":"1117","key":"1596_CR92","first-page":"431","volume":"194","author":"DJ Willshaw","year":"1976","unstructured":"Willshaw, D.J., & Von Der Malsburg, C. (1976). How patterned neural connections can be set up by self-organization. Proceedings of the Royal Society of London. Series B, Biological Sciences, 194(1117), 431\u2013445. http:\/\/www.jstor.org\/stable\/77138 .","journal-title":"Proceedings of the Royal Society of London. Series B, Biological Sciences"},{"key":"1596_CR93","doi-asserted-by":"crossref","unstructured":"Martinetz, T.M., Berkovich, S.G., Schulten, K.J. (1993). Neural-Gas Network for Vector Quantization and its Application to Time-Series Prediction, 4, 4.","DOI":"10.1109\/72.238311"},{"key":"1596_CR94","first-page":"625","volume":"7","author":"B Fritzke","year":"1995","unstructured":"Fritzke, B. (1995). A Growing Neural Gas Learns Topologies. Advances in Neural Information Processing Systems, 7, 625\u2013632.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"1596_CR95","unstructured":"Fritzke, B. (1994). Supervised Learning with Growing Cell Structures."},{"issue":"1","key":"1596_CR96","first-page":"2","volume":"1","author":"B Fritzke","year":"1994","unstructured":"Fritzke, B., & Bochum, R-. (1994). Fast learning with incremental RBF Networks 1 Introduction 2 Model description. Processing, 1(1), 2\u20135.","journal-title":"Processing"},{"issue":"9","key":"1596_CR97","doi-asserted-by":"publisher","first-page":"1441","DOI":"10.1016\/0893-6080(94)90091-4","volume":"7","author":"B Fritzke","year":"1994","unstructured":"Fritzke, B. (1994). Growing cell structures-A self-organizing network for unsupervised and supervised learning. Neural Networks, 7(9), 1441\u20131460.","journal-title":"Neural Networks"},{"key":"1596_CR98","unstructured":"Montana, D.J., & Davis, L. (1989). Training feedforward neural networks using genetic algorithms. In Proceedings of the International Joint Conference on Artificial Intelligence (pp. 762\u2013767)."},{"issue":"1","key":"1596_CR99","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1007\/s12065-007-0002-4","volume":"1","author":"D Floreano","year":"2008","unstructured":"Floreano, D., D\u00fcrr, P., Mattiussi, C. (2008). Neuroevolution: from architectures to learning. Evolutionary Intelligence, 1(1), 47\u201362.","journal-title":"Evolutionary Intelligence"},{"key":"1596_CR100","doi-asserted-by":"crossref","unstructured":"Stanley, K.O., & Miikkulainen, R. (2002). Evolving neural networks through augmenting topologies, Evolutionary Computation.","DOI":"10.1162\/106365602320169811"},{"issue":"1","key":"1596_CR101","doi-asserted-by":"publisher","first-page":"67","DOI":"10.1007\/BF01411376","volume":"1","author":"NJ Radcliffe","year":"1993","unstructured":"Radcliffe, N.J. (1993). Genetic set recombination and its application to neural network topology optimisation. Neural Computing & Applications, 1(1), 67\u201390.","journal-title":"Neural Computing & Applications"},{"key":"1596_CR102","unstructured":"Thierens, D. (1996). Non-redundant genetic coding of neural networks."},{"key":"1596_CR103","unstructured":"Miikkulainen, R., Liang, J., Meyerson, E., Rawal, A., Fink, D., Francon, O., Raju, B., Shahrzad, H., Navruzyan, A., Duffy, N., Hodjat, B. (2017). Evolving Deep Neural Networks."},{"key":"1596_CR104","unstructured":"Elsken, T., Metzen, J.H., Hutter, F. (2018). Simple and efficient architecture search for convolutional neural networks. https:\/\/openreview.net\/forum?id=SySaJ0xCZ,."},{"key":"1596_CR105","doi-asserted-by":"crossref","unstructured":"Cai, H., Chen, T., Zhang, W., Yu, Y., Wang, J. (2018). Efficient Architecture Search by Network Transformation.","DOI":"10.1609\/aaai.v32i1.11709"},{"key":"1596_CR106","doi-asserted-by":"crossref","unstructured":"Jin, H., Song, Q., & Hu, X. (2018). Efficient Neural Architecture Search with Network Morphism.","DOI":"10.1145\/3292500.3330648"},{"key":"1596_CR107","unstructured":"Cai, H., Yang, J., Zhang, W., Han, S., Yu, Y. (2018). Path-level network transformation for efficient architecture search. In Proceedings of the 35th International Conference on Machine Learning (pp. 678\u2013687)."},{"key":"1596_CR108","unstructured":"Saxena, S., & Verbeek, J. (2016). Convolutional Neural Fabrics."},{"key":"1596_CR109","unstructured":"Pham, H., Guan, M., Zoph, B., Le, Q., Dean, J. (2018). Efficient neural architecture search via parameters sharing. In Proceedings of the 35th International Conference on Machine Learning, Vol. 80."},{"key":"1596_CR110","doi-asserted-by":"crossref","unstructured":"Veniat, T., & Denoyer, L. (2018). Learning time\/memory-efficient deep architectures with budgeted super networks. In Conference on Computer Vision and Pattern Recognition (pp. 3492\u20133500).","DOI":"10.1109\/CVPR.2018.00368"},{"key":"1596_CR111","doi-asserted-by":"crossref","unstructured":"Zoph, B., Yuret, D., May, J., Knight, K. (2016). Transfer Learning for Low-Resource Neural Machine Translation.","DOI":"10.18653\/v1\/D16-1163"},{"key":"1596_CR112","doi-asserted-by":"crossref","unstructured":"Tan, M., Chen, B., Pang, R., Vasudevan, V., Le, Q.V. (2019). MnasNet: Platform-Aware Neural Architecture Search for Mobile.","DOI":"10.1109\/CVPR.2019.00293"},{"key":"1596_CR113","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C. (2018). Inverted Residuals and Linear Bottlenecks: Mobile Networks for Classification, Detection and Segmentation.","DOI":"10.1109\/CVPR.2018.00474"},{"key":"1596_CR114","unstructured":"Frankle, J., & Carbin, M. (2019). The lottery ticket hypothesis: Finding sparse, trainable neural networks. In ICLR."}],"container-title":["Journal of Signal Processing Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11265-020-01596-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11265-020-01596-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11265-020-01596-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,22]],"date-time":"2022-11-22T23:18:24Z","timestamp":1669159104000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11265-020-01596-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020,10,12]]},"references-count":114,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2021,8]]}},"alternative-id":["1596"],"URL":"https:\/\/doi.org\/10.1007\/s11265-020-01596-1","relation":{},"ISSN":["1939-8018","1939-8115"],"issn-type":[{"value":"1939-8018","type":"print"},{"value":"1939-8115","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020,10,12]]},"assertion":[{"value":"16 April 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 August 2020","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 September 2020","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 October 2020","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}