{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,28]],"date-time":"2026-01-28T20:27:22Z","timestamp":1769632042567,"version":"3.49.0"},"reference-count":36,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2023,7,15]],"date-time":"2023-07-15T00:00:00Z","timestamp":1689379200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,7,15]],"date-time":"2023-07-15T00:00:00Z","timestamp":1689379200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012431","name":"Jiangsu Agricultural Science and Technology Independent Innovation Fund","doi-asserted-by":"publisher","award":["chx2022012"],"award-info":[{"award-number":["chx2022012"]}],"id":[{"id":"10.13039\/501100012431","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-16115-0","type":"journal-article","created":{"date-parts":[[2023,7,15]],"date-time":"2023-07-15T02:01:47Z","timestamp":1689386507000},"page":"16525-16542","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["A novel multi-scale facial expression recognition algorithm based on improved Res2Net for classroom scenes"],"prefix":"10.1007","volume":"83","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5217-0381","authenticated-orcid":false,"given":"Meihua","family":"Gu","sequence":"first","affiliation":[]},{"given":"Jing","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Yalu","family":"Chu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,7,15]]},"reference":[{"key":"16115_CR1","doi-asserted-by":"publisher","unstructured":"Dimitrios K, Viktoriia S, Stefanos Z. (2021) Distribution matching for heterogeneous multi-task learning: a large-scale face study. Proceedings of the IEEE Conference on Computer vision and Pattern Recognition(CVPR), https:\/\/doi.org\/10.48550\/arXiv.2015.03790","DOI":"10.48550\/arXiv.2015.03790"},{"issue":"2","key":"16115_CR2","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1109\/tpami.2019.2938758","volume":"43","author":"SH Gao","year":"2019","unstructured":"Gao SH, Cheng MM, Zhao K et al (2019) Res2net: a new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell 43(2):652\u2013662. https:\/\/doi.org\/10.1109\/tpami.2019.2938758","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"16115_CR3","doi-asserted-by":"publisher","first-page":"393","DOI":"10.11992\/tis.202107028","volume":"17","author":"T Gao","year":"2021","unstructured":"Gao T, Zhaochen Y, Ting C et al (2021) Deep multi-scale fusion attention residual face expression recognition network[J]. J Intell Syst 17(2):393\u2013401. https:\/\/doi.org\/10.11992\/tis.202107028","journal-title":"J Intell Syst"},{"issue":"18","key":"16115_CR4","doi-asserted-by":"publisher","first-page":"25321","DOI":"10.1007\/s11042-019-7651-z","volume":"78","author":"SK Gupta","year":"2019","unstructured":"Gupta SK, Ashwin TS, Guddeti RMR (2019) Students' affective content analysis in smart classroom environment using deep learning techniques. Multimed Tools Appl 78(18):25321\u201325348. https:\/\/doi.org\/10.1007\/s11042-019-7651-z","journal-title":"Multimed Tools Appl"},{"key":"16115_CR5","doi-asserted-by":"publisher","unstructured":"Hou Q, Zhou D, Feng J. (2021) Coordinate attention for efficient mobile network design. Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition(CVPR), pp 13713-13722. https:\/\/doi.org\/10.1109\/cvpr46437.2021.01350","DOI":"10.1109\/cvpr46437.2021.01350"},{"key":"16115_CR6","doi-asserted-by":"publisher","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-Excitation networks. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), pp 7132\u20137141. https:\/\/doi.org\/10.1109\/cvpr.2018.00745","DOI":"10.1109\/cvpr.2018.00745"},{"key":"16115_CR7","doi-asserted-by":"publisher","first-page":"231","DOI":"10.1016\/j.neucom.2018.12.037","volume":"333","author":"Y Ji","year":"2019","unstructured":"Ji Y, Hu Y, Yang Y et al (2019) Cross-domain facial expression recognition via an intra-category common feature and inter-category distinction feature fusion network. Neurocomput 333:231\u2013239. https:\/\/doi.org\/10.1016\/j.neucom.2018.12.037","journal-title":"Neurocomput"},{"key":"16115_CR8","doi-asserted-by":"publisher","unstructured":"Li, D (2021) Research on facial expression recognition based on capsule network. Southwest University. https:\/\/doi.org\/10.27684\/d.cnki.gxndx.2021.003154","DOI":"10.27684\/d.cnki.gxndx.2021.003154"},{"issue":"01","key":"16115_CR9","doi-asserted-by":"publisher","first-page":"356","DOI":"10.1109\/tip.2018.2868382","volume":"28","author":"S Li","year":"2019","unstructured":"Li S, Deng W (2019) Reliable crowdsourcing and deep locality preserving learning for unconstrained facial expression recognition. IEEE Trans Image Process 28(01):356\u2013370. https:\/\/doi.org\/10.1109\/tip.2018.2868382","journal-title":"IEEE Trans Image Process"},{"key":"16115_CR10","doi-asserted-by":"publisher","unstructured":"Li S, Deng W, Du J P (2017) Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. Proceedings of the IEEE conference on computer vision and pattern recognition(CVPR), pp 2852-2861. https:\/\/doi.org\/10.1109\/cvpr.2017.277","DOI":"10.1109\/cvpr.2017.277"},{"issue":"05","key":"16115_CR11","doi-asserted-by":"publisher","first-page":"2439","DOI":"10.1109\/tip.2018.2886767","volume":"28","author":"Y Li","year":"2018","unstructured":"Li Y, Zeng J, Shan S et al (2018) Occlusion aware facial expression recognition using cnn with attention mechanism. IEEE Trans Image Process 28(05):2439\u20132450. https:\/\/doi.org\/10.1109\/tip.2018.2886767","journal-title":"IEEE Trans Image Process"},{"key":"16115_CR12","doi-asserted-by":"publisher","unstructured":"Lin TY, Goyal P, Girshick R et al (2017) Focal loss for dense object detection. Proceed IEEE Int Conf Comput Vis:2980\u20132988. https:\/\/doi.org\/10.1109\/iccv.2017.324","DOI":"10.1109\/iccv.2017.324"},{"key":"16115_CR13","doi-asserted-by":"publisher","unstructured":"Lucey, P, Cohn, JF, Kanade, T, et al (2010) The extended cohn-kanade dataset (CK+): A complete dataset for action unit and emotion-specified expression. In 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition-workshops, IEEE, pp 94\u2013101. https:\/\/doi.org\/10.1109\/cvprw.2010.5543262","DOI":"10.1109\/cvprw.2010.5543262"},{"key":"16115_CR14","doi-asserted-by":"publisher","unstructured":"Lyons, M, Akamatsu, S, Kamachi, M, Gyoba, J (1998) Coding facial expressions with gabor wavelets. In Proceedings Third IEEE international conference on automatic face and gesture recognition, IEEE, pp. 200\u2013205. https:\/\/doi.org\/10.1109\/AFGR.1998.670949","DOI":"10.1109\/AFGR.1998.670949"},{"issue":"9","key":"16115_CR15","doi-asserted-by":"publisher","first-page":"3046","DOI":"10.3390\/s21093046","volume":"21","author":"S Minaee","year":"2021","unstructured":"Minaee S, Minaei M, Abdolrashidi A (2021) Deep-emotion: facial expression recognition using attentional convolutional network. Sensors. 21(9):3046. https:\/\/doi.org\/10.3390\/s21093046","journal-title":"Sensors."},{"key":"16115_CR16","doi-asserted-by":"publisher","unstructured":"Li Peng (2020) Research on an end-to-end student emotion recognition system to assist university classroom teaching. University of Electronic Science and Technology. https:\/\/doi.org\/10.27005\/d.cnki.gdzku.2020.003411","DOI":"10.27005\/d.cnki.gdzku.2020.003411"},{"key":"16115_CR17","doi-asserted-by":"publisher","unstructured":"Radlak K, Smolka B (2016) High dimensional local binary patterns for facial expression recognition in the wild. Mediterranean Electrotechnical Conference(MELECON), pp. 1\u20135. https:\/\/doi.org\/10.1109\/melcon.2016.7495381","DOI":"10.1109\/melcon.2016.7495381"},{"issue":"03","key":"16115_CR18","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1016\/j.jbtep.2005.05.002","volume":"36","author":"B Renneberg","year":"2005","unstructured":"Renneberg B, Heyn K, Gebhard R et al (2005) Facial expression of emotions in borderline personality disorder and depression. J Behav Ther Exp Psychiatry 36(03):183\u2013196. https:\/\/doi.org\/10.1016\/j.jbtep.2005.05.002","journal-title":"J Behav Ther Exp Psychiatry"},{"key":"16115_CR19","doi-asserted-by":"publisher","unstructured":"Ruan D, Yan Y, Lai S, et al (2021) Feature decomposition and reconstruction learning for effective facial expression recognition. IEEE\/CVF conference on computer vision and pattern recognition(CVPR), pp 7660-7669. https:\/\/doi.org\/10.1109\/cvpr46437.2021.00757","DOI":"10.1109\/cvpr46437.2021.00757"},{"key":"16115_CR20","doi-asserted-by":"publisher","unstructured":"Selvaraju R R, Cogswell M, Das A, et al (2017) Grade-cam: visual explanations from deep networks via gradient-based localization. Proceedings of the IEEE international conference on computer vision(ICCV), pp 618-626. https:\/\/doi.org\/10.1109\/iccv.2017.74","DOI":"10.1109\/iccv.2017.74"},{"issue":"03","key":"16115_CR21","doi-asserted-by":"publisher","first-page":"459","DOI":"10.1016\/j.jvcir.2017.10.008","volume":"49","author":"A Sherly Alphonse","year":"2017","unstructured":"Sherly Alphonse A, Dharma D (2017) A novel monogenic directional pattern and pseudo-Voigt kernel for facilitating the identification of facial emotions. J Vis Commun Image Represent 49(03):459\u2013470. https:\/\/doi.org\/10.1016\/j.jvcir.2017.10.008","journal-title":"J Vis Commun Image Represent"},{"key":"16115_CR22","doi-asserted-by":"publisher","unstructured":"Song Y, Gao S, Zeng H, et al (2021) Multi-scale depth-separable expression recognition with embedded attention mechanism. J Beijing Univ Aeronaut Astronaut https:\/\/doi.org\/10.13700\/j.bh.1001-5965.2021.0114","DOI":"10.13700\/j.bh.1001-5965.2021.0114"},{"key":"16115_CR23","doi-asserted-by":"publisher","unstructured":"Stewart A, Bosch N, Chen H, et al (2017) Face forward: detecting mind wandering from video during narrative film comprehension. International conference on artificial intelligence, pp 359-370. https:\/\/doi.org\/10.1007\/978-3-319-61425-0_30","DOI":"10.1007\/978-3-319-61425-0_30"},{"issue":"12","key":"16115_CR24","doi-asserted-by":"publisher","first-page":"299","DOI":"10.19678\/j.issn.1000-3428.0060133","volume":"47","author":"C Su","year":"2021","unstructured":"Su C, Wang L, Lan VJ (2021) A fine-grained expression recognition model based on multi-scale hierarchical bilinear pooling network. Comput Eng 47(12):299\u2013307. https:\/\/doi.org\/10.19678\/j.issn.1000-3428.0060133","journal-title":"Comput Eng"},{"issue":"230","key":"16115_CR25","doi-asserted-by":"publisher","first-page":"397","DOI":"10.1016\/j.neucom.2016.12.043","volume":"100","author":"Y Sun","year":"2017","unstructured":"Sun Y, Wen G (2017) Cognitive facial expression recognition with constrained dimensionality reduction. Neurocomput 100(230):397\u2013408. https:\/\/doi.org\/10.1016\/j.neucom.2016.12.043","journal-title":"Neurocomput"},{"issue":"01","key":"16115_CR26","doi-asserted-by":"publisher","first-page":"12","DOI":"10.1016\/j.neucom.2018.03.034","volume":"296","author":"W Sun","year":"2018","unstructured":"Sun W, Zhao H, Jin Z (2018) A visual attention based ROI detection method for facial expression recognition. Neurocomputing 296(01):12\u201322. https:\/\/doi.org\/10.1016\/j.neucom.2018.03.034","journal-title":"Neurocomputing"},{"key":"16115_CR27","doi-asserted-by":"publisher","unstructured":"Tan M, Pang R, Le Q V (2020) Efficientdet: scalable and efficient object detection. Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition(CVPR), pp 10781-10790. https:\/\/doi.org\/10.1109\/cvpr42600.2020.01079","DOI":"10.1109\/cvpr42600.2020.01079"},{"key":"16115_CR28","first-page":"2579","volume":"9","author":"L van der Maaten","year":"2008","unstructured":"van der Maaten L, Hinton G (2008) Visualizing Data using t-SNE. J Mach Learn Res 9:2579\u20132605","journal-title":"J Mach Learn Res"},{"key":"16115_CR29","doi-asserted-by":"publisher","unstructured":"Vemulapalli R, Agarwala A (2019) A compact embedding for facial expression similarity. Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition(CVPR), pp 5683-5692. https:\/\/doi.org\/10.1109\/cvpr.2019.00583","DOI":"10.1109\/cvpr.2019.00583"},{"key":"16115_CR30","doi-asserted-by":"publisher","unstructured":"Wang K, Peng X, Yang J, et al (2020) Suppressing uncertainties for large-scale facial expression recognition. IEEE\/CVF conference on computer vision and pattern recognition(CVPR), pp 6897-6906. https:\/\/doi.org\/10.1109\/cvpr42600.2020.00693","DOI":"10.1109\/cvpr42600.2020.00693"},{"issue":"01","key":"16115_CR31","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/taffc.2014.2316163","volume":"5","author":"J Whitehill","year":"2014","unstructured":"Whitehill J, Serpell Z, Lin YC et al (2014) The faces of engagement: automatic recognition of student engagement from facial expressions. IEEE Trans Affect Comput 5(01):86\u201398. https:\/\/doi.org\/10.1109\/taffc.2014.2316163","journal-title":"IEEE Trans Affect Comput"},{"key":"16115_CR32","doi-asserted-by":"publisher","unstructured":"Yao L, Wan Y, Ni H, Xu B; (2021) Action unit classification for facial expression recognition using active learning and SVM . Multimed Tools Appl, https:\/\/doi.org\/10.1007\/s11042-021-10836-w","DOI":"10.1007\/s11042-021-10836-w"},{"key":"16115_CR33","doi-asserted-by":"publisher","unstructured":"Yongqiang LV (2021) Research on face expression recognition in natural scenes. Huazhong Normal University. https:\/\/doi.org\/10.27159\/d.cnki.ghzsu.2021.002034","DOI":"10.27159\/d.cnki.ghzsu.2021.002034"},{"key":"16115_CR34","doi-asserted-by":"publisher","unstructured":"Yu Z (2018) Emotion recognition based on small resolution faces and its application in information-based teaching. Shanghai Jiaotong University. https:\/\/doi.org\/10.27307\/d.cnki.gsjtu.2018.004755","DOI":"10.27307\/d.cnki.gsjtu.2018.004755"},{"issue":"01","key":"16115_CR35","doi-asserted-by":"publisher","first-page":"182","DOI":"10.19304\/j.issn1000-7180.2021.0799","volume":"58","author":"P Zhang","year":"2022","unstructured":"Zhang P, Kong W, Teng J (2022) Face expression recognition based on multi-scale feature attention mechanism. Comput Eng Appl 58(01):182\u2013189. https:\/\/doi.org\/10.19304\/j.issn1000-7180.2021.0799","journal-title":"Comput Eng Appl"},{"key":"16115_CR36","doi-asserted-by":"publisher","unstructured":"Zhu R, Sang G, Zhao Q (2016) Discriminative feature adaptation for cross-domain facial expression recognition. 2016 international conference on biometrics (ICB), IEEE, pp 1-7. https:\/\/doi.org\/10.1109\/icb.2016.7550085","DOI":"10.1109\/icb.2016.7550085"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16115-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-16115-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16115-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,31]],"date-time":"2024-01-31T08:35:07Z","timestamp":1706690107000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-16115-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,7,15]]},"references-count":36,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2024,2]]}},"alternative-id":["16115"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-16115-0","relation":{},"ISSN":["1573-7721"],"issn-type":[{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,7,15]]},"assertion":[{"value":"5 October 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 May 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 June 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 July 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"Informed consent\u115f","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Compliance with ethical standards"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}},{"value":"Code of Ethics for Socio-Economic Research and Declaration of Helsinki.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"The ethics agreement"}}]}}