{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,3]],"date-time":"2025-12-03T18:10:10Z","timestamp":1764785410490,"version":"3.37.3"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"8-9","license":[{"start":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T00:00:00Z","timestamp":1717372800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T00:00:00Z","timestamp":1717372800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,9]]},"DOI":"10.1007\/s11760-024-03283-1","type":"journal-article","created":{"date-parts":[[2024,6,3]],"date-time":"2024-06-03T04:01:28Z","timestamp":1717387288000},"page":"5939-5956","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["FGENet: a lightweight facial expression recognition algorithm based on FasterNet"],"prefix":"10.1007","volume":"18","author":[{"given":"Miaomiao","family":"Sun","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5845-3606","authenticated-orcid":false,"given":"Chunman","family":"Yan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,3]]},"reference":[{"key":"3283_CR1","doi-asserted-by":"crossref","unstructured":"Mehrabian, A.: Communication without words[M]. In: Communication Theory, pp. 193\u2013200. Routledge (2017)","DOI":"10.4324\/9781315080918-15"},{"key":"3283_CR2","doi-asserted-by":"publisher","first-page":"215874","DOI":"10.1109\/ACCESS.2020.3041173","volume":"8","author":"TD Lalitharatne","year":"2020","unstructured":"Lalitharatne, T.D., Tan, Y., Leong, F., et al.: Facial expression rendering in medical training simulators: current status and future directions[J]. IEEE Access 8, 215874\u2013215891 (2020)","journal-title":"IEEE Access"},{"key":"3283_CR3","doi-asserted-by":"crossref","unstructured":"Kawamura, R., Murase, K.: Concentration estimation in E-learning based on learner\u2019s facial reaction to teacher\u2019s action[C]. In: Proceedings of the 25th International Conference on Intelligent User Interfaces Companion, pp. 103\u2013104 (2020)","DOI":"10.1145\/3379336.3381487"},{"issue":"8","key":"3283_CR4","doi-asserted-by":"publisher","first-page":"2956","DOI":"10.3390\/app10082956","volume":"10","author":"CM Kim","year":"2020","unstructured":"Kim, C.M., Hong, E.J., Chung, K., et al.: Driver facial expression analysis using LFA-CRNN-based feature extraction for health-risk decisions[J]. Appl. Sci. 10(8), 2956 (2020)","journal-title":"Appl. Sci."},{"key":"3283_CR5","doi-asserted-by":"crossref","unstructured":"Lundqvist, D., Flykt, A., Hman, A.: The karolinska directed emotional faces\u2014KDEF, CD ROM from Department of Clinical Neuroscience. Psychol. Sec. [J] (1998)","DOI":"10.1037\/t27732-000"},{"key":"3283_CR6","doi-asserted-by":"publisher","unstructured":"Kim, S., An, G.H., Kang, S.J.: Facial expression recognition system using machine learning[C]. In: International SoC Design Conference (2017). https:\/\/doi.org\/10.1109\/ISOCC.2017.8368887.","DOI":"10.1109\/ISOCC.2017.8368887"},{"key":"3283_CR7","doi-asserted-by":"crossref","unstructured":"Yue, C., Liang, J., Qu, B., et al.: Sparse representation feature for facial expression recognition[C]. In: Proceedings of ELM-2017. pp. 12\u201321. Springer International Publishing (2019)","DOI":"10.1007\/978-3-030-01520-6_2"},{"key":"3283_CR8","doi-asserted-by":"crossref","unstructured":"Abdulrahman, M., Gwadabe, T.R., Abdu, F.J., et al.: Gabor wavelet transform based facial expression recognition using PCA and LBP[C]. In: 2014 22nd Signal Processing And Communications Applications Conference (SIU), pp. 2265\u20132268, IEEE (2014)","DOI":"10.1109\/SIU.2014.6830717"},{"issue":"6","key":"3283_CR9","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1145\/3065386","volume":"60","author":"A Krizhevsky","year":"2017","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: ImageNet classification with deep convolutional neural networks[J]. Commun. ACM 60(6), 84\u201390 (2017)","journal-title":"Commun. ACM"},{"key":"3283_CR10","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., et al.: Deep residual learning for image recognition[C]. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778, (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"3283_CR11","unstructured":"Howard, A.G., Zhu, M., Chen, B., et al.: Mobilenets: efficient convolutional neural networks for mobile vision applications[J]. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"3283_CR12","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., et al.: Mobilenetv2: Inverted residuals and linear bottlenecks[C]. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"3283_CR13","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chu, G., et al.: Searching for mobilenetv3[C]. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"issue":"6","key":"3283_CR14","doi-asserted-by":"publisher","first-page":"063002","DOI":"10.1117\/1.JEI.30.6.063002","volume":"30","author":"Y Kong","year":"2021","unstructured":"Kong, Y., Ren, Z., Zhang, K., et al.: Lightweight facial expression recognition method based on attention mechanism and key region fusion[J]. J. Electron. Imaging 30(6), 063002\u2013063002 (2021)","journal-title":"J. Electron. Imaging"},{"issue":"6","key":"3283_CR15","doi-asserted-by":"publisher","first-page":"4435","DOI":"10.1016\/j.aej.2021.09.066","volume":"61","author":"Y Nan","year":"2022","unstructured":"Nan, Y., Ju, J., Hua, Q., et al.: A-MobileNet: an approach of facial expression recognition[J]. Alex. Eng. J. 61(6), 4435\u20134444 (2022)","journal-title":"Alex. Eng. J."},{"key":"3283_CR16","doi-asserted-by":"crossref","unstructured":"Chen, J., Kao, S., He, H., et al.: Run, don\u2019t walk: chasing higher FLOPS for faster neural networks[C]. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12021\u201312031 (2023)","DOI":"10.1109\/CVPR52729.2023.01157"},{"key":"3283_CR17","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks[C]. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"3283_CR18","doi-asserted-by":"crossref","unstructured":"Wang, Q., Wu, B., Zhu, P., et al.: ECA-Net: efficient channel attention for deep convolutional neural networks[C]. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11534\u201311542 (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"3283_CR19","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., et al.: Cbam: convolutional block attention module[C]. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"3283_CR20","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design[C]. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13713\u201313722 (2021)","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"3283_CR21","doi-asserted-by":"crossref","unstructured":"Szegedy, C., Vanhoucke, V., Ioffe, S., et al.: Rethinking the inception architecture for computer vision[C]. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2818\u20132826 (2016)","DOI":"10.1109\/CVPR.2016.308"},{"key":"3283_CR22","unstructured":"Li, H., Li, J., Wei, H., et al.: Slim-neck by GSConv: a better design paradigm of detector architectures for autonomous vehicles[J]. arXiv preprint arXiv:2206.02424 (2022)"},{"key":"3283_CR23","doi-asserted-by":"crossref","unstructured":"Han, K., Wang, Y., Tian, Q., et al.: Ghostnet: More features from cheap operations[C]. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1580\u20131589 (2020)","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"3283_CR24","unstructured":"Goodfellow, I.J., Erhan, D., Carrier, P.L., et al.: Challenges in representation learning: a report on three machine learning contests[C]. In: Neural Information Processing: 20th International Conference, ICONIP 2013, Daegu, Korea, November 3\u20137, 2013. Proceedings, Part III 20. Springer berlin heidelberg, pp. 117\u2013124 (2013)"},{"key":"3283_CR25","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., et al.: The extended cohn-kanade dataset (ck+): a complete dataset for action unit and emotion-specified expression[C]. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops, pp. 94\u2013101, IEEE (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"key":"3283_CR26","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.P.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild[C]. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2852\u20132861 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"3283_CR27","doi-asserted-by":"crossref","unstructured":"Sidhom, O., Ghazouani, H., Barhoumi, W.: Three-phases hybrid feature selection for facial expression recognition[J]. J. Supercomput. 1\u201335 (2023)","DOI":"10.1007\/s11227-023-05758-3"},{"issue":"9","key":"3283_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.17485\/ijst\/2017\/v10i9\/108944","volume":"10","author":"M Goyani","year":"2017","unstructured":"Goyani, M., Patel, N.: Multi-level haar wavelet based facial expression recognition using logistic regression[J]. Indian J. Sci. Technol. 10(9), 1\u20139 (2017)","journal-title":"Indian J. Sci. Technol."},{"key":"3283_CR29","doi-asserted-by":"crossref","unstructured":"Sang, D.V., Ha, P.T.: Discriminative deep feature learning for facial emotion recognition[C]. In: 2018 1st International Conference on Multimedia Analysis and Pattern Recognition (MAPR), pp. 1\u20136, IEEE (2018)","DOI":"10.1109\/MAPR.2018.8337514"},{"issue":"9","key":"3283_CR30","doi-asserted-by":"publisher","first-page":"6499","DOI":"10.1007\/s00521-022-08005-7","volume":"35","author":"M Mukhopadhyay","year":"2023","unstructured":"Mukhopadhyay, M., Dey, A., Kahali, S.: A deep-learning-based facial expression recognition method using textural features[J]. Neural Comput. Appl. 35(9), 6499\u20136514 (2023)","journal-title":"Neural Comput. Appl."},{"key":"3283_CR31","doi-asserted-by":"publisher","first-page":"1244","DOI":"10.1109\/ACCESS.2022.3233362","volume":"11","author":"Y He","year":"2022","unstructured":"He, Y.: Facial expression recognition using multi-branch attention convolutional neural network[J]. IEEE Access 11, 1244\u20131253 (2022)","journal-title":"IEEE Access"},{"key":"3283_CR32","doi-asserted-by":"publisher","first-page":"35","DOI":"10.1016\/j.ins.2020.02.047","volume":"522","author":"X Sun","year":"2020","unstructured":"Sun, X., Xia, P., Zhang, L., et al.: A ROI-guided deep architecture for robust facial expressions recognition[J]. Inf. Sci. 522, 35\u201348 (2020)","journal-title":"Inf. Sci."},{"issue":"9","key":"3283_CR33","doi-asserted-by":"publisher","first-page":"3046","DOI":"10.3390\/s21093046","volume":"21","author":"S Minaee","year":"2021","unstructured":"Minaee, S., Minaei, M., Abdolrashidi, A.: Deep-emotion: Facial expression recognition using attentional convolutional network[J]. Sensors 21(9), 3046 (2021)","journal-title":"Sensors"},{"key":"3283_CR34","doi-asserted-by":"publisher","first-page":"7183","DOI":"10.1109\/ACCESS.2020.2964298","volume":"8","author":"X Sun","year":"2020","unstructured":"Sun, X., Zheng, S., Fu, H.: ROI-attention vectorized CNN model for static facial expression recognition[J]. IEEE Access 8, 7183\u20137194 (2020)","journal-title":"IEEE Access"},{"key":"3283_CR35","doi-asserted-by":"publisher","first-page":"847","DOI":"10.1007\/s11571-021-09761-3","volume":"16","author":"X Fan","year":"2022","unstructured":"Fan, X., Jiang, M., Shahid, A.R., et al.: Hierarchical scale convolutional neural network for facial expression recognition. Cogn. Neurodyn. 16, 847\u2013858 (2022). https:\/\/doi.org\/10.1007\/s11571-021-09761-3","journal-title":"Cogn. Neurodyn."},{"key":"3283_CR36","doi-asserted-by":"publisher","first-page":"96964","DOI":"10.1109\/ACCESS.2021.3095191","volume":"9","author":"X Fan","year":"2021","unstructured":"Fan, X., Jiang, M., Yan, H.: A deep learning based light-weight face mask detector with residual context attention and Gaussian heatmap to fight against COVID-19[J]. Ieee Access 9, 96964\u201396974 (2021)","journal-title":"Ieee Access"},{"key":"3283_CR37","doi-asserted-by":"crossref","unstructured":"Zeng, G., Zhou, J., Jia, X., et al.: Hand-crafted feature guided deep learning for facial expression recognition[C]. In: 2018 13th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2018), pp. 423\u2013430, IEEE (2018)","DOI":"10.1109\/FG.2018.00068"},{"issue":"8","key":"3283_CR38","doi-asserted-by":"publisher","first-page":"4245","DOI":"10.1007\/s11760-023-02657-1","volume":"17","author":"L Ji","year":"2023","unstructured":"Ji, L., Wu, S., Gu, X.: A facial expression recognition algorithm incorporating SVM and explainable residual neural network[J]. SIViP 17(8), 4245\u20134254 (2023)","journal-title":"SIViP"},{"key":"3283_CR39","first-page":"317","volume":"12","author":"S Zhao","year":"2018","unstructured":"Zhao, S., Cai, H., Liu, H., et al.: Feature selection mechanism in CNNs for facial expression recognition[C]. BMVC. 12, 317 (2018)","journal-title":"BMVC."},{"key":"3283_CR40","unstructured":"Mehta, S., Rastegari, M.: Separable self-attention for mobile vision transformers[J]. arXiv preprint arXiv:2206.02680 (2022)"},{"key":"3283_CR41","doi-asserted-by":"crossref","unstructured":"Maaz, M., Shaker, A., Cholakkal, H., et al.: Edgenext: efficiently amalgamated cnn-transformer architecture for mobile vision applications[C]. In: European conference on computer vision, pp. 3\u201320, Springer, Cham (2022)","DOI":"10.1007\/978-3-031-25082-8_1"},{"key":"3283_CR42","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions[C]. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251\u20131258 (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"3283_CR43","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.T., et al.: Shufflenet v2: practical guidelines for efficient cnn architecture design[C]. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 116\u2013131 (2018)","DOI":"10.1007\/978-3-030-01264-9_8"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03283-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03283-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03283-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,29]],"date-time":"2024-07-29T19:10:55Z","timestamp":1722280255000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03283-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,3]]},"references-count":43,"journal-issue":{"issue":"8-9","published-print":{"date-parts":[[2024,9]]}},"alternative-id":["3283"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03283-1","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2024,6,3]]},"assertion":[{"value":"9 February 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 April 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 May 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 June 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there are no conflicts of interest related to this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}