{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,5]],"date-time":"2026-04-05T05:34:55Z","timestamp":1775367295934,"version":"3.50.1"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2024,2,25]],"date-time":"2024-02-25T00:00:00Z","timestamp":1708819200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,2,25]],"date-time":"2024-02-25T00:00:00Z","timestamp":1708819200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62071161"],"award-info":[{"award-number":["62071161"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62071161"],"award-info":[{"award-number":["62071161"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2024,6]]},"DOI":"10.1007\/s11760-024-03020-8","type":"journal-article","created":{"date-parts":[[2024,2,25]],"date-time":"2024-02-25T07:01:55Z","timestamp":1708844515000},"page":"3553-3563","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["A lightweight facial expression recognition model for automated engagement detection"],"prefix":"10.1007","volume":"18","author":[{"given":"Zibin","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Yinbei","family":"Li","sequence":"additional","affiliation":[]},{"given":"Jiaqiang","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Yuliang","family":"Ma","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,25]]},"reference":[{"key":"3020_CR1","volume-title":"Engaging Students for Success-2008 Australasian Survey of Student Engagement","author":"H Coates","year":"2009","unstructured":"Coates, H.: Engaging Students for Success-2008 Australasian Survey of Student Engagement. Australian Council for Educational Research, Victoria, Australia (2009)"},{"issue":"4","key":"3020_CR2","first-page":"216","volume":"5","author":"S Gunuc","year":"2014","unstructured":"Gunuc, S.: The relationships between student engagement and their academic achievement. Int. J. New Trends Educ. Their Implic. 5(4), 216\u2013231 (2014)","journal-title":"Int. J. New Trends Educ. Their Implic."},{"issue":"1","key":"3020_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/1472-6920-13-33","volume":"13","author":"MJ Casuso-Holgado","year":"2013","unstructured":"Casuso-Holgado, M.J., Cuesta-Vargas, A.I., Moreno-Morales, N., Labajos-Manzanares, M.T., Bar\u00f3n-L\u00f3pez, F.J., Vega-Cuesta, M.: The association between academic engagement and achievement in health sciences students. BMC Med. Educ. 13(1), 1\u20137 (2013)","journal-title":"BMC Med. Educ."},{"issue":"1","key":"3020_CR4","first-page":"1","volume":"11","author":"V Trowler","year":"2010","unstructured":"Trowler, V.: Student engagement literature review. High. Educ. Acad. 11(1), 1\u201315 (2010)","journal-title":"High. Educ. Acad."},{"issue":"2","key":"3020_CR5","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1037\/0003-066X.36.2.129","volume":"36","author":"GH Bower","year":"1981","unstructured":"Bower, G.H.: Mood and memory. Am. Psychol. 36(2), 129 (1981)","journal-title":"Am. Psychol."},{"issue":"1","key":"3020_CR6","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TAFFC.2014.2316163","volume":"5","author":"J Whitehill","year":"2014","unstructured":"Whitehill, J., Serpell, Z., Lin, Y.-C., Foster, A., Movellan, J.R.: The faces of engagement: automatic recognition of student engagement from facial expressions. IEEE Trans. Affect. Comput. 5(1), 86\u201398 (2014)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"4","key":"3020_CR7","first-page":"1084","volume":"24","author":"A Mehrabian","year":"1968","unstructured":"Mehrabian, A.: Communication without words. Univ. East Lond. 24(4), 1084\u20135 (1968)","journal-title":"Univ. East Lond."},{"key":"3020_CR8","doi-asserted-by":"crossref","unstructured":"Ekman, P., Friesen,W.: Facial action coding system: A technique for the measurement of facial movement. In Environmental Psychology & Nonverbal Behavior. Consulting Psychologists Press: Palo Alto, CA, USA (1978)","DOI":"10.1037\/t27734-000"},{"key":"3020_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/2193-1801-2-455","volume":"2","author":"M Sathik","year":"2013","unstructured":"Sathik, M., Jonathan, S.G.: Effect of facial expressions on student\u2019s comprehension recognition in virtual educational environments. Springerplus 2, 1\u20139 (2013)","journal-title":"Springerplus"},{"issue":"1","key":"3020_CR10","first-page":"99","volume":"33","author":"K Altuwairqi","year":"2021","unstructured":"Altuwairqi, K., Jarraya, S.K., Allinjawi, A., Hammami, M.: A new emotion-based affective model to detect student\u2019s engagement. J. King Saud Univ. Comput. Inf. Sci. 33(1), 99\u2013109 (2021)","journal-title":"J. King Saud Univ. Comput. Inf. Sci."},{"key":"3020_CR11","doi-asserted-by":"publisher","first-page":"1011","DOI":"10.1007\/s12671-016-0548-2","volume":"7","author":"J Fish","year":"2016","unstructured":"Fish, J., Brimson, J., Lynch, S.: Mindfulness interventions delivered by technology without facilitator involvement: what research exists and what are the clinical outcomes? Mindfulness 7, 1011\u20131023 (2016)","journal-title":"Mindfulness"},{"issue":"2","key":"3020_CR12","doi-asserted-by":"publisher","first-page":"320","DOI":"10.1111\/bjet.12235","volume":"47","author":"KF Hew","year":"2016","unstructured":"Hew, K.F.: Promoting engagement in online courses: What strategies can we learn from three highly rated MOOCS. Br. J. Educ. Technol. 47(2), 320\u2013341 (2016)","journal-title":"Br. J. Educ. Technol."},{"key":"3020_CR13","doi-asserted-by":"crossref","unstructured":"Aneja, D., Colburn, A., Faigin, G., Shapiro, L., Mones, B.: Modeling stylized character expressions via deep learning. In: ACCV (2016)","DOI":"10.1007\/978-3-319-54184-6_9"},{"key":"3020_CR14","doi-asserted-by":"crossref","unstructured":"Mollahosseini, A., Chan, D., Mahoor, M.H.: Going deeper in facial expression recognition using deep neural networks. IEEE (2016)","DOI":"10.1109\/WACV.2016.7477450"},{"key":"3020_CR15","unstructured":"Grafsgaard, J., Wiggins, J.B., Boyer, K.E., Wiebe, E.N., Lester, J.: Automatically recognizing facial expression: predicting engagement and frustration. In: Educational Data Mining 2013 (2013)"},{"key":"3020_CR16","doi-asserted-by":"crossref","unstructured":"Dubbaka, A., Gopalan, A.: Detecting learner engagement in MOOCS using automatic facial expression recognition. In: 2020 IEEE Global Engineering Education Conference (EDUCON) (2020)","DOI":"10.1109\/EDUCON45650.2020.9125149"},{"key":"3020_CR17","doi-asserted-by":"crossref","unstructured":"Shen, J., Yang, H., Li, J., Cheng, Z.: Assessing learning engagement based on facial expression recognition in MOOC\u2019s scenario. Multimedia Syst. 28, 469\u2013478 (2022)","DOI":"10.1007\/s00530-021-00854-x"},{"key":"3020_CR18","doi-asserted-by":"publisher","first-page":"6609","DOI":"10.1007\/s10489-020-02139-8","volume":"51","author":"J Liao","year":"2021","unstructured":"Liao, J., Liang, Y., Pan, J.: Deep facial spatiotemporal network for engagement prediction in online learning. Appl. Intell. 51, 6609\u20136621 (2021)","journal-title":"Appl. Intell."},{"issue":"8","key":"3020_CR19","doi-asserted-by":"publisher","first-page":"11365","DOI":"10.1007\/s11042-022-13558-9","volume":"82","author":"S Gupta","year":"2023","unstructured":"Gupta, S., Kumar, P., Tekchandani, R.K.: Facial emotion recognition based real-time learner engagement detection system in online learning context using deep learning models. Multimed. Tools Appl. 82(8), 11365\u201311394 (2023)","journal-title":"Multimed. Tools Appl."},{"key":"3020_CR20","unstructured":"Hewitt, C., Gunes, H.: CNN-based facial affect analysis on mobile devices. arXiv preprint arXiv:1807.08775 (2018)"},{"key":"3020_CR21","doi-asserted-by":"crossref","unstructured":"Barros, P., Churamani, N., Sciutti, A.: The facechannel: a light-weight deep neural network for facial expression recognition. In: 2020 15th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2020), pp. 652\u2013656 (2020). IEEE","DOI":"10.1109\/FG47880.2020.00070"},{"key":"3020_CR22","unstructured":"Ferro-P\u00e9rez, R., Mitre-Hernandez, H.: Resmonet: a residual mobile-based network for facial emotion recognition in resource-limited systems (2020)"},{"key":"3020_CR23","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Liu, Q., Zhou, F.: Robust lightweight facial expression recognition network with label distribution training. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 3510\u20133519 (2021)","DOI":"10.1609\/aaai.v35i4.16465"},{"key":"3020_CR24","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1251\u20131258 (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"3020_CR25","unstructured":"Iandola, F.N., Han, S., Moskewicz, M.W., Ashraf, K., Dally, W.J., Keutzer, K.: Squeezenet: Alexnet-level accuracy with 50x fewer parameters and$$<$$ 0.5 mb model size. arXiv preprint arXiv:1602.07360 (2016)"},{"key":"3020_CR26","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., Adam, H.: Mobilenets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)"},{"key":"3020_CR27","doi-asserted-by":"crossref","unstructured":"Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., Chen, L.-C.: Mobilenetv2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"3020_CR28","doi-asserted-by":"crossref","unstructured":"Howard, A., Sandler, M., Chu, G., Chen, L.-C., Chen, B., Tan, M., Wang, W., Zhu, Y., Pang, R., Vasudevan, V.: Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"3020_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zhou, X., Lin, M., Sun, J.: Shufflenet: an extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6848\u20136856 (2018)","DOI":"10.1109\/CVPR.2018.00716"},{"key":"3020_CR30","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.-T., Sun, J.: Shufflenet v2: practical guidelines for efficient CNN architecture design. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 116\u2013131 (2018)","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"3020_CR31","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"3020_CR32","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"3020_CR33","doi-asserted-by":"crossref","unstructured":"Huang, G., Liu, Z., Van Der\u00a0Maaten, L., Weinberger, K.Q.: Densely connected convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4700\u20134708 (2017)","DOI":"10.1109\/CVPR.2017.243"},{"key":"3020_CR34","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"3020_CR35","doi-asserted-by":"crossref","unstructured":"Zeng, J., Shan, S., Chen, X.: Facial expression recognition with inconsistently annotated datasets. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 222\u2013237 (2018)","DOI":"10.1007\/978-3-030-01261-8_14"},{"key":"3020_CR36","unstructured":"Li, Y., Lu, Y., Li, J., Lu, G.: Separate loss for basic and compound facial expression recognition in the wild. In: Asian Conference on Machine Learning, pp. 897\u2013911 (2019). PMLR"},{"issue":"5","key":"3020_CR37","doi-asserted-by":"publisher","first-page":"2439","DOI":"10.1109\/TIP.2018.2886767","volume":"28","author":"Y Li","year":"2018","unstructured":"Li, Y., Zeng, J., Shan, S., Chen, X.: Occlusion aware facial expression recognition using CNN with attention mechanism. IEEE Trans. Image Process. 28(5), 2439\u20132450 (2018)","journal-title":"IEEE Trans. Image Process."},{"key":"3020_CR38","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang, K., Peng, X., Yang, J., Meng, D., Qiao, Y.: Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans. Image Process. 29, 4057\u20134069 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"3020_CR39","doi-asserted-by":"crossref","unstructured":"Chen, S., Wang, J., Chen, Y., Shi, Z., Geng, X., Rui, Y.: Label distribution learning on auxiliary label space graphs for facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13984\u201313993 (2020)","DOI":"10.1109\/CVPR42600.2020.01400"},{"key":"3020_CR40","doi-asserted-by":"crossref","unstructured":"Farzaneh, A.H., Qi, X.: Discriminant distribution-agnostic loss for facial expression recognition in the wild. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 406\u2013407 (2020)","DOI":"10.1109\/CVPRW50498.2020.00211"},{"key":"3020_CR41","doi-asserted-by":"crossref","unstructured":"Wang, K., Peng, X., Yang, J., Lu, S., Qiao, Y.: Suppressing uncertainties for large-scale facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6897\u20136906 (2020)","DOI":"10.1109\/CVPR42600.2020.00693"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03020-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-024-03020-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-024-03020-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T19:46:46Z","timestamp":1731440806000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-024-03020-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2,25]]},"references-count":41,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["3020"],"URL":"https:\/\/doi.org\/10.1007\/s11760-024-03020-8","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,2,25]]},"assertion":[{"value":"14 August 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 December 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 January 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 February 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests that might be perceived to influence the results and\/or discussion reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}