{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,4]],"date-time":"2025-11-04T16:19:45Z","timestamp":1762273185502},"reference-count":59,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2023,12,18]],"date-time":"2023-12-18T00:00:00Z","timestamp":1702857600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,12,18]],"date-time":"2023-12-18T00:00:00Z","timestamp":1702857600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Science and Technology Project of Xi\u2019an City","award":["22GXFW0086"],"award-info":[{"award-number":["22GXFW0086"]}]},{"name":"Science and Technology Project of Beilin District in Xi\u2019an City","award":["GX2243"],"award-info":[{"award-number":["GX2243"]}]},{"name":"School-Enterprise Collaborative Innovation Fund for Graduate Students of Xi\u2019an University of Technology","award":["310\/252062108"],"award-info":[{"award-number":["310\/252062108"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1007\/s11760-023-02832-4","type":"journal-article","created":{"date-parts":[[2023,12,18]],"date-time":"2023-12-18T08:02:16Z","timestamp":1702886536000},"page":"2019-2035","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["EC-RFERNet: an edge computing-oriented real-time facial expression recognition network"],"prefix":"10.1007","volume":"18","author":[{"given":"Qiang","family":"Sun","sequence":"first","affiliation":[]},{"given":"Yuan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Dongxu","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Jing","family":"Wen","sequence":"additional","affiliation":[]},{"given":"Jiaojiao","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yonglu","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,18]]},"reference":[{"issue":"3","key":"2832_CR1","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2020","unstructured":"Li, S., Deng, W.: Deep facial expression recognition: a survey. IEEE Trans. Affect. Comput. 13(3), 1195\u20131215 (2020)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"2","key":"2832_CR2","doi-asserted-by":"publisher","first-page":"524","DOI":"10.1109\/TAFFC.2018.2890471","volume":"12","author":"PV Rouast","year":"2021","unstructured":"Rouast, P.V., Adam, M.T.P., Chong, R.: Deep learning for human affect recognition: Insights and new developments. IEEE Trans. Affect. Comput. 12(2), 524\u2013543 (2021)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"5","key":"2832_CR3","doi-asserted-by":"publisher","first-page":"637","DOI":"10.1109\/JIOT.2016.2579198","volume":"3","author":"W Shi","year":"2016","unstructured":"Shi, W., Cao, J., Zhang, Q., Li, Y., Xu, L.: Edge computing: vision and challenges. IEEE Int. Things J. 3(5), 637\u2013646 (2016)","journal-title":"IEEE Int. Things J."},{"key":"2832_CR4","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. CoRR, https:\/\/arxiv.org\/abs\/1409.1556(2014)"},{"key":"2832_CR5","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: The 2016 conference on computer vision and pattern recognition (CVPR), pp 770\u2013778 (2016).","DOI":"10.1109\/CVPR.2016.90"},{"issue":"6","key":"2832_CR6","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks. Commun. ACM 60(6), 84\u201390 (2017)","journal-title":"Commun. ACM"},{"key":"2832_CR7","doi-asserted-by":"publisher","first-page":"1081","DOI":"10.1007\/s11760-021-02057-3","volume":"16","author":"FR Mohammadian","year":"2022","unstructured":"Mohammadian, F.R., Mahlouji, M., Shahidinejad, A.: Real-time face detection using circular sliding of the gabor energy and neural networks. SIViP 16, 1081\u20131089 (2022)","journal-title":"SIViP"},{"key":"2832_CR8","first-page":"2503912","volume":"71","author":"X Wang","year":"2022","unstructured":"Wang, X., Zhang, L., Huang, W.B., Wang, S.Y., Wu, H., He, J., Song, A.G.: Deep convolutional networks with tunable speed\u2013accuracy tradeoff for human activity recognition using wearables. IEEE Trans. Instrum. Meas. 71, 2503912 (2022)","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2832_CR9","doi-asserted-by":"publisher","first-page":"101172","DOI":"10.1109\/ACCESS.2021.3095844","volume":"9","author":"NE Zarif","year":"2021","unstructured":"Zarif, N.E., Montazeri, L., Leduc-Primeau, F., Sawan, M.: Mobile-optimized facial expression recognition techniques. IEEE Access 9, 101172\u2013101185 (2021)","journal-title":"IEEE Access"},{"key":"2832_CR10","doi-asserted-by":"publisher","first-page":"76178","DOI":"10.1109\/ACCESS.2021.3082641","volume":"9","author":"J Yang","year":"2021","unstructured":"Yang, J., Qian, T., Zhang, F., Khan, S.U.: Real-time facial expression recognition based on edge computing. IEEE Access 9, 76178\u201376190 (2021)","journal-title":"IEEE Access"},{"issue":"6","key":"2832_CR11","doi-asserted-by":"publisher","first-page":"4435","DOI":"10.1016\/j.aej.2021.09.066","volume":"61","author":"YH Nan","year":"2022","unstructured":"Nan, Y.H., Ju, J.G., Hua, Q.Y., Zhang, H.M., Wang, B.: A-MobileNet: an approach of facial expression recognition. Alex. Eng. J. 61(6), 4435\u20134444 (2022)","journal-title":"Alex. Eng. J."},{"issue":"4","key":"2832_CR12","doi-asserted-by":"publisher","first-page":"184","DOI":"10.1049\/el.2018.7871","volume":"55","author":"H Ma","year":"2019","unstructured":"Ma, H., Celik, T.: FER-Net: Facial expression recognition using densely connected convolutional network. Electron. Lett. 55(4), 184\u2013186 (2019)","journal-title":"Electron. Lett."},{"key":"2832_CR13","doi-asserted-by":"crossref","unstructured":"Chen, J. R., Kao, S., He, H., Zhuo, W. P., Wen, S., Lee, C., Chan, S. G.: Run, don\u2019t walk: chasing higher flops for faster neural networks. In: The IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12021\u201312031 (2023)","DOI":"10.1109\/CVPR52729.2023.01157"},{"key":"2832_CR14","doi-asserted-by":"crossref","unstructured":"Chao, P., Kao, C.-Y., Ruan, Y., Huang, C.-H., Lin, Y.-L.: HarDNet: a low memory traffic network. In: The 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 3551\u20133560 (2019)","DOI":"10.1109\/ICCV.2019.00365"},{"issue":"8","key":"2832_CR15","doi-asserted-by":"publisher","first-page":"2011","DOI":"10.1109\/TPAMI.2019.2913372","volume":"42","author":"J Hu","year":"2020","unstructured":"Hu, J., Shen, L., Albanie, S., Sun, G., Wu, E.: Squeeze-and-excitation networks. IEEE Trans. Pattern Anal. Mach. Intell. 42(8), 2011\u20132023 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2832_CR16","unstructured":"Howard, A. G., Zhu, M., Chen, B.: MobileNets: efficient convolutional neural networks for mobile vision applications. CoRR, https:\/\/arxiv.org\/abs\/1704.04861(2017)"},{"key":"2832_CR17","doi-asserted-by":"crossref","unstructured":"Lu, X., Zhang, H., Zhang, Q., Han, X.: A lightweight network for expression recognition based on adaptive mixed residual connections. In: The 2022 5th International Conference on Pattern Recognition and Artificial Intelligence (PRAI), pp 311\u2013315 (2022).","DOI":"10.1109\/PRAI55851.2022.9904198"},{"key":"2832_CR18","doi-asserted-by":"crossref","unstructured":"Zhou, L., Li, S., Wang, Y., Liu, J.: SDNET: Lightweight facial expression recognition for sample disequilibrium. In: The 2022\u20132022 IEEE International Conference on Acoustics, Speech and Signal Processing, pp. 2415\u20132419 (2022)","DOI":"10.1109\/ICASSP43922.2022.9746695"},{"key":"2832_CR19","doi-asserted-by":"crossref","unstructured":"Fang, B., Chen, G., He, J.: Ghost-based convolutional neural network for effective facial expression recognition. In: The 2022 International Conference on Machine Learning and Knowledge Engineering, pp. 121\u2013124 (2022)","DOI":"10.1109\/MLKE55170.2022.00029"},{"key":"2832_CR20","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: GhostNet: more features from cheap operations. In The 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1577\u20131586 (2020)","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"2832_CR21","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L. -C.: MobileNetV2: inverted residuals and linear bottlenecks. In: The 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"2832_CR22","doi-asserted-by":"publisher","first-page":"108469","DOI":"10.1016\/j.compeleceng.2022.108469","volume":"104","author":"Q Sun","year":"2022","unstructured":"Sun, Q., Liang, L., Dang, X.H., Chen, Y.: Deep learning-based dimensional emotion recognition combining the attention mechanism and global second-order feature representations. Comput. Electr. Eng. 104, 108469 (2022)","journal-title":"Comput. Electr. Eng."},{"key":"2832_CR23","doi-asserted-by":"crossref","unstructured":"Zhang, F., Li, Q., Ren, Y., Xu, H., Song, Y., Liu, S.: An expression recognition method on robots based on MobilenetV2-SSD. In: The 2019 6th International Conference on Systems and Informatics (ICSAI), pp. 118\u2013122 (2019)","DOI":"10.1109\/ICSAI48974.2019.9010173"},{"issue":"3","key":"2832_CR24","doi-asserted-by":"publisher","first-page":"2245","DOI":"10.1007\/s10462-021-10059-3","volume":"55","author":"Z Huang","year":"2022","unstructured":"Huang, Z., Yang, S., Zhou, M.C., Gong, Z., Abusorrah, A., Lin, C., Huang, Z.: Making accurate object detection at the edge: review and new approach. Artif. Intell. Rev. 55(3), 2245\u20132274 (2022)","journal-title":"Artif. Intell. Rev."},{"key":"2832_CR25","doi-asserted-by":"crossref","unstructured":"Passalis, N., Raitoharju, J., Tefas, A., Gabbouj, M.: Adaptive inference using hierarchical convolutional bag-of-features for low-power embedded platforms. In: The 2019 IEEE International Conference on Image Processing, pp 3048\u20133052 (2019)","DOI":"10.1109\/ICIP.2019.8803283"},{"key":"2832_CR26","doi-asserted-by":"publisher","first-page":"134065","DOI":"10.1109\/ACCESS.2021.3113337","volume":"9","author":"MNA Wahab","year":"2021","unstructured":"Wahab, M.N.A., Nazir, A., Zhen, A.T., Mohd Noor, M.H., Akbar, M.F., Mohamed, A.S.A.: Efficientnet-lite and hybrid cnn-knn implementation for facial expression recognition on raspberry Pi. IEEE Access 9, 134065\u2013134080 (2021)","journal-title":"IEEE Access"},{"key":"2832_CR27","unstructured":"Tan, M. X., Le, Q.: Efficientnet: rethinking model scaling for convolutional neural networks. In: The 36th International Conference on Machine Learning, Long Beach, 97, pp. 6105\u20136114 (2019)"},{"issue":"4","key":"2832_CR28","doi-asserted-by":"publisher","first-page":"1087","DOI":"10.3390\/s20041087","volume":"20","author":"MN Riaz","year":"2020","unstructured":"Riaz, M.N., Shen, Y., Sohail, M., Guo, M.: eXnet: an efficient approach for emotion recognition in the wild. Sensors 20(4), 1087 (2020)","journal-title":"Sensors"},{"issue":"1","key":"2832_CR29","first-page":"1929","volume":"15","author":"S Nitish","year":"2014","unstructured":"Nitish, S., Geoffrey, H., Alex, K., Ilya, S., Ruslan, S.: Dropout: a simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 15(1), 1929\u20131958 (2014)","journal-title":"J. Mach. Learn. Res."},{"key":"2832_CR30","unstructured":"Viola, P., Jones, M.: Rapid object detection using a boosted cascade of simple features. In: The 2001 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (2001)"},{"key":"2832_CR31","doi-asserted-by":"crossref","unstructured":"Li, S., Deng W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: The 2017 IEEE Conference on Computer Vision and Pattern Recognition, pp. 2584\u20132593 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"issue":"1","key":"2832_CR32","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2019","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: AffectNet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10(1), 18\u201331 (2019)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"2832_CR33","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J. F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended Cohn-Kanade dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: The 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp 94\u2013101 (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"2832_CR34","doi-asserted-by":"crossref","unstructured":"Dhall, A., Goecke, R., Lucey, S., Gedeon, T.: Static facial expression analysis in tough conditions: data, evaluation protocol, and benchmark. In: The 2011 IEEE International Conference on Computer Vision Workshops, pp. 2106\u20132112 (2011)","DOI":"10.1109\/ICCVW.2011.6130508"},{"key":"2832_CR35","unstructured":"Wang, R. J., Li, X., Ling, C. X.: Pelee: A real-time object detection system on mobile devices. In: The 32nd International Conference on Neural Information Processing Systems, pp. 1967\u20131976 (2018)"},{"key":"2832_CR36","unstructured":"Tan, M. X., Le, Q. V.: MixConv: Mixed depthwise convolutional kernels. In: The 30th British Machine Vision Conference (2019)"},{"key":"2832_CR37","doi-asserted-by":"publisher","first-page":"1507","DOI":"10.1007\/s11760-021-01883-9","volume":"15","author":"H Ma","year":"2021","unstructured":"Ma, H., Celik, T., Li, H.C.: Lightweight attention convolutional neural network through network slimming for robust facial expression recognition. SIViP 15, 1507\u20131515 (2021)","journal-title":"SIViP"},{"key":"2832_CR38","doi-asserted-by":"publisher","first-page":"18635","DOI":"10.1007\/s11042-022-14066-6","volume":"82","author":"H Huo","year":"2023","unstructured":"Huo, H., Yu, Y., Liu, Z.: Facial expression recognition based on improved depthwise separable convolutional network. Multimed. Tools Appl. 82, 18635\u201318652 (2023)","journal-title":"Multimed. Tools Appl."},{"key":"2832_CR39","doi-asserted-by":"publisher","first-page":"1083","DOI":"10.1007\/s00371-021-02069-7","volume":"38","author":"S Saurav","year":"2022","unstructured":"Saurav, S., Gidde, P., Saini, R., Singh, S.: Dual integrated convolutional neural network for real-time facial expression recognition in the wild. Vis. Comput. 38, 1083\u20131096 (2022)","journal-title":"Vis. Comput."},{"key":"2832_CR40","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.neucom.2019.05.005","volume":"355","author":"J Shao","year":"2019","unstructured":"Shao, J., Qian, Y.S.: Three convolutional neural network models for facial expression recognition in the wild. Neurocomputing 355, 82\u201392 (2019)","journal-title":"Neurocomputing"},{"key":"2832_CR41","doi-asserted-by":"publisher","first-page":"4245","DOI":"10.1007\/s11760-023-02657-1","volume":"17","author":"LP Ji","year":"2023","unstructured":"Ji, L.P., Wu, S.L., Gu, X.P.: A facial expression recognition algorithm incorporating SVM and explainable residual neural network. Signal Image Video Process 17, 4245\u20134254 (2023)","journal-title":"Signal Image Video Process."},{"issue":"11","key":"2832_CR42","doi-asserted-by":"publisher","first-page":"2914","DOI":"10.1109\/TMM.2020.2966858","volume":"22","author":"G Wen","year":"2020","unstructured":"Wen, G., Chang, T., Li, H., Jiang, L.: Dynamic objectives learning for facial expression recognition. IEEE Trans. Multimedia 22(11), 2914\u20132925 (2020)","journal-title":"IEEE Trans. Multimedia"},{"issue":"5","key":"2832_CR43","doi-asserted-by":"publisher","first-page":"2439","DOI":"10.1109\/TIP.2018.2886767","volume":"28","author":"Y Li","year":"2018","unstructured":"Li, Y., Zeng, J., Shan, S., Chen, X.: Occlusion aware facial expression recognition using CNN with attention mechanism. IEEE Trans. Image Process. 28(5), 2439\u20132450 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"2832_CR44","doi-asserted-by":"publisher","first-page":"199","DOI":"10.3390\/biomimetics8020199","volume":"8","author":"Z Wen","year":"2021","unstructured":"Wen, Z., Lin, W., Wang, T., Xu, G.: Distract your attention: multi-head cross attention network for facial expression recognition. Biomimetics 8(2), 199 (2021)","journal-title":"Biomimetics"},{"key":"2832_CR45","doi-asserted-by":"publisher","first-page":"108737","DOI":"10.1016\/j.patcog.2022.108737","volume":"129","author":"J Zhang","year":"2022","unstructured":"Zhang, J., Yu, H.: Improving the facial expression recognition and its interpretability via generating expression pattern-map. Pattern Recogn. 129, 108737 (2022)","journal-title":"Pattern Recogn."},{"issue":"5","key":"2832_CR46","doi-asserted-by":"publisher","first-page":"666","DOI":"10.1111\/coin.12586","volume":"39","author":"S Saurav","year":"2023","unstructured":"Saurav, S., Saini, R, Singh, S.: A dual-channel ensembled deep convolutional neural network for facial expression recognition in the wild. Comput. Intell. 39(5), 666\u2013706 (2023)","journal-title":"Comput. Intell."},{"key":"2832_CR47","doi-asserted-by":"publisher","first-page":"177","DOI":"10.1016\/j.patcog.2019.03.019","volume":"92","author":"SY Xie","year":"2019","unstructured":"Xie, S.Y., Hu, H.F., Wu, Y.B.: Deep multi-path convolutional neural network joint with salient region attention for facial expression recognition. Pattern Recogn. 92, 177\u2013191 (2019)","journal-title":"Pattern Recogn."},{"key":"2832_CR48","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1016\/j.neucom.2018.03.034","volume":"296","author":"WY Sun","year":"2018","unstructured":"Sun, W.Y., Zhao, H.T., Jin, Z.: A visual attention-based ROI detection method for facial expression recognition. Neurocomputing 296, 12\u201322 (2018)","journal-title":"Neurocomputing"},{"key":"2832_CR49","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1007\/s00371-018-1585-8","volume":"36","author":"I Gogi\u0107","year":"2020","unstructured":"Gogi\u0107, I., Manhart, M., Pand\u017ei\u0107, I.S., Ahlberg, J.: Fast facial expression recognition using local binary features and shallow neural networks. Vis. Comput. 36, 97\u2013112 (2020)","journal-title":"Vis. Comput."},{"key":"2832_CR50","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang, K., Peng, X., Yang, J., Meng, D., Qiao, Y.: Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans. Image Process. 29, 4057\u20134069 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"2832_CR51","doi-asserted-by":"publisher","first-page":"4595","DOI":"10.1007\/s00521-021-06613-3","volume":"34","author":"S Saurav","year":"2022","unstructured":"Saurav, S., Saini, K.A., Saini, R., Singh, S.: Deep learning inspired intelligent embedded system for haptic rendering of facial emotions to the blind. Neural Comput. Appl. 34, 4595\u20134623 (2022)","journal-title":"Neural Comput. Appl."},{"issue":"3","key":"2832_CR52","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3587038","volume":"22","author":"YR Wu","year":"2023","unstructured":"Wu, Y.R., Zhang, L.L., Gu, Z.H., Lu, H., Wan, S.H.: Edge-AI-driven framework with efficient mobile network design for facial expression recognition. ACM Trans Embed. Comput. Syst. 22(3), 1\u201317 (2023)","journal-title":"ACM Trans Embed. Comput. Syst."},{"key":"2832_CR53","unstructured":"Landola, F. N., Han, S., Moskewicz, M. W., Ashraf, K., Dally, W. J., Keutzer, K.: squeezenet: alexnet-level accuracy with 50x fewer parameters and< 0.5 MB model size. In: The 5th International Conference on Learning Representations (2016)"},{"key":"2832_CR54","doi-asserted-by":"crossref","unstructured":"Gholami, A., Kwon, K., Wu, B., Tai, Z., Yue, X. Y., Jin, P., Zhao, S., Keutzer, K.: SqueezeNext: hardware-aware neural network design. In: The 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 1719\u20131728 (2018)","DOI":"10.1109\/CVPRW.2018.00215"},{"key":"2832_CR55","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chen, B., Wang, W., Chen, L., Tan, M., Chuet, G.: Searching for MobileNetV3. In: The 2019 IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"2832_CR56","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: ShuffleNet: An extremely efficient convolutional neural network for mobile devices. In: The 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6848\u20136856 (2018)","DOI":"10.1109\/CVPR.2018.00716"},{"key":"2832_CR57","doi-asserted-by":"crossref","unstructured":"Ma, N. N., Zhang, X. Y., Zheng, H. T., Sun, J.: ShuffleNet V2: Practical guidelines for efficient architecture design. In: The 2018 European conference on computer vision (ECCV), pp. 116\u2013131 (2018)","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"2832_CR58","doi-asserted-by":"crossref","unstructured":"Zhang, T., Qi, G. -J., Xiao, B., Wang, J.: Interleaved group convolutions. In: The 2017 IEEE International Conference on Computer Vision (ICCV), pp. 4383\u20134392 (2017)","DOI":"10.1109\/ICCV.2017.469"},{"key":"2832_CR59","unstructured":"Sun, K., Li, M. J., Liu, D., Wang, J. D.: IGCV3: Interleaved low-rank group convolutions for efficient deep neural networks. CoRR, (2018) http:\/\/arxiv.org\/abs\/1806.00178"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02832-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-023-02832-4\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-023-02832-4.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,19]],"date-time":"2024-03-19T20:14:23Z","timestamp":1710879263000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-023-02832-4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,18]]},"references-count":59,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["2832"],"URL":"https:\/\/doi.org\/10.1007\/s11760-023-02832-4","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,18]]},"assertion":[{"value":"6 September 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 September 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 October 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 December 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. The authors declare they have no financial interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Informed consent was obtained from all the participants in this study.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}