{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T06:05:29Z","timestamp":1743055529377,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":25,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819615346"},{"type":"electronic","value":"9789819615353"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-1535-3_13","type":"book-chapter","created":{"date-parts":[[2025,2,13]],"date-time":"2025-02-13T12:24:18Z","timestamp":1739449458000},"page":"114-125","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Facial Expression Recognition Using Convolutional Neural Network with Residual Units and Attention Mechanism"],"prefix":"10.1007","author":[{"given":"Moe Moe","family":"Htay","sequence":"first","affiliation":[]},{"given":"Zin Mar","family":"Win","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,14]]},"reference":[{"issue":"1","key":"13_CR1","doi-asserted-by":"publisher","first-page":"55","DOI":"10.1186\/s13634-023-01019-w","volume":"2023","author":"H Zhong","year":"2023","unstructured":"Zhong, H., Han, T., Xia, W., Tian, Y., Wu, L.: Research on real-time teachers\u2019 facial expression recognition based on YOLOv5 and attention mechanisms. EURASIP J. Adv. Sign. Process. 2023(1), 55 (2023)","journal-title":"EURASIP J. Adv. Sign. Process."},{"issue":"6","key":"13_CR2","doi-asserted-by":"publisher","first-page":"1815","DOI":"10.1109\/TPAMI.2019.2958341","volume":"43","author":"T Hassan","year":"2019","unstructured":"Hassan, T., et al.: Automatic detection of pain from facial expressions: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 43(6), 1815\u20131831 (2019)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"13_CR3","doi-asserted-by":"publisher","first-page":"118727","DOI":"10.1109\/ACCESS.2019.2936663","volume":"7","author":"W Deng","year":"2019","unstructured":"Deng, W., Wu, R.: Real-time driver-drowsiness detection system using facial features. IEEE Access 7, 118727\u2013118738 (2019)","journal-title":"IEEE Access"},{"key":"13_CR4","doi-asserted-by":"crossref","unstructured":"Doiphode, B.S., Sapkal, S.D.: Classifying facial expression using support vector machine based on bidirectional local binary pattern histogram feature descriptor. In: 2018 Second International Conference on Intelligent Computing and Control Systems (ICICCS), pp. 1890\u20131895. IEEE, Madurai, India (2018)","DOI":"10.1109\/ICCONS.2018.8662980"},{"key":"13_CR5","doi-asserted-by":"publisher","first-page":"29882","DOI":"10.1109\/ACCESS.2019.2899024","volume":"7","author":"M Hu","year":"2019","unstructured":"Hu, M., Zheng, Y., Yang, C., Wang, X., He, L., Ren, F.: Facial expression recognition using fusion features based on center-symmetric local octonary pattern. IEEE Access 7, 29882\u201329890 (2019)","journal-title":"IEEE Access"},{"key":"13_CR6","doi-asserted-by":"publisher","first-page":"12158","DOI":"10.1109\/ACCESS.2021.3051403","volume":"9","author":"J Liu","year":"2021","unstructured":"Liu, J., Wang, H., Feng, Y.: An end-to-end deep model with discriminative facial features for facial expression recognition. IEEE access 9, 12158\u201312166 (2021)","journal-title":"IEEE access"},{"issue":"28","key":"13_CR7","doi-asserted-by":"publisher","DOI":"10.4108\/eetiot.v7i28.685","volume":"7","author":"W Xu","year":"2022","unstructured":"Xu, W., Cloutier, R.S.: A facial expression recognizer using modified ResNet-152. EAI Endorsed Trans Internet Things 7(28), e5 (2022)","journal-title":"EAI Endorsed Trans Internet Things"},{"key":"13_CR8","doi-asserted-by":"crossref","unstructured":"G\u00f3mez-Sirvent, J.L., L\u00f3pez De La Rosa, F., L\u00f3pez, M.T., Fern\u00e1ndez-Caballero, A.: facial expression recognition in the wild for low-resolution images using voting residual network. Electronics 12(18), 3837 (2023)","DOI":"10.3390\/electronics12183837"},{"key":"13_CR9","doi-asserted-by":"publisher","first-page":"70865","DOI":"10.1109\/ACCESS.2023.3294099","volume":"11","author":"DH Lee","year":"2023","unstructured":"Lee, D.H., Yoo, J.H.: CNN learning strategy for recognizing facial expressions. IEEE Access 11, 70865\u201370872 (2023)","journal-title":"IEEE Access"},{"key":"13_CR10","doi-asserted-by":"crossref","unstructured":"Agarwal, R., Mittal, N., Madasu, H., Rao, M.V.S.: Meerut institute of engineering and technology, and engineering college.: convolutional neural network based facial expression recognition using image filtering techniques. Int. J. Intell. Eng. Syst. 14(5), 78\u201387 (2021)","DOI":"10.22266\/ijies2021.1031.08"},{"key":"13_CR11","doi-asserted-by":"crossref","unstructured":"Hu, P., Kang, Q., Zeng, C., Dong, F., Chen, K.: Facial expression recognition based on improved residual network. In: Chen, C.H., Scapellato, A., Barbiero, A., Korzun, D. G., (eds.) Advances in Transdisciplinary Engineering. IOS Press (2022)","DOI":"10.3233\/ATDE221052"},{"key":"13_CR12","doi-asserted-by":"publisher","first-page":"24609","DOI":"10.1109\/ACCESS.2024.3365521","volume":"12","author":"H Qi","year":"2024","unstructured":"Qi, H., Zhang, X., Shi, Y., Qi, X.: A novel attention residual network expression recognition method. IEEE Access 12, 24609\u201324620 (2024)","journal-title":"IEEE Access"},{"issue":"1","key":"13_CR13","first-page":"539","volume":"135","author":"T Shen","year":"2023","unstructured":"Shen, T., Xu, H.: Facial expression recognition based on multi-channel attention residual network\u201d. Comput. Model. Eng. Sci. 135(1), 539\u2013560 (2023)","journal-title":"Comput. Model. Eng. Sci."},{"key":"13_CR14","first-page":"1","volume":"2021","author":"M Chen","year":"2021","unstructured":"Chen, M., Cheng, J., Zhang, Z., Li, Y., Zhang, Y.: Facial expression recognition method combined with attention mechanism. Mob. Inf. Syst. 2021, 1\u201310 (2021)","journal-title":"Mob. Inf. Syst."},{"key":"13_CR15","doi-asserted-by":"publisher","first-page":"7383","DOI":"10.1109\/ACCESS.2020.2963913","volume":"8","author":"Y Gan","year":"2020","unstructured":"Gan, Y., Chen, J., Yang, Z., Xu, L.: Multiple attention network for facial expression recognition. IEEE Access 8, 7383\u20137393 (2020)","journal-title":"IEEE Access"},{"issue":"1","key":"13_CR16","doi-asserted-by":"publisher","first-page":"415","DOI":"10.32604\/csse.2022.019749","volume":"41","author":"K Prabhu","year":"2022","unstructured":"Prabhu, K., SathishKumar, S., Sivachitra, M., Dineshkumar, S., Sathiyabama, P.: Facial expression recognition using enhanced convolution neural network with attention mechanism. Comput. Syst. Sci. Eng. 41(1), 415\u2013426 (2022)","journal-title":"Comput. Syst. Sci. Eng."},{"issue":"10","key":"13_CR17","doi-asserted-by":"publisher","first-page":"1499","DOI":"10.1109\/LSP.2016.2603342","volume":"23","author":"K Zhang","year":"2016","unstructured":"Zhang, K., Zhang, Z., Li, Z., Qiao, Y.: Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process. Lett. 23(10), 1499\u20131503 (2016)","journal-title":"IEEE Signal Process. Lett."},{"key":"13_CR18","doi-asserted-by":"publisher","first-page":"86821","DOI":"10.1109\/ACCESS.2019.2923530","volume":"7","author":"X Tong","year":"2019","unstructured":"Tong, X., Sun, S., Fu, M.: Data augmentation and second-order pooling for facial expression recognition. IEEE Access 7, 86821\u201386828 (2019)","journal-title":"IEEE Access"},{"key":"13_CR19","unstructured":"He, K., Zhang, X., Ren, S., and Sun, J.: Identity mappings in deep residual networks. arXiv, Jul. 25, 2016. Accessed 17 May 2024"},{"issue":"5","key":"13_CR20","doi-asserted-by":"publisher","first-page":"1169","DOI":"10.53106\/160792642023092405015","volume":"24","author":"W Wang","year":"2023","unstructured":"Wang, W., Zhang, H.: Multiscale convolutional attention-based residual network expression recognition. J. Internet Technol. 24(5), 1169\u20131175 (2023)","journal-title":"J. Internet Technol."},{"key":"13_CR21","doi-asserted-by":"crossref","unstructured":"Chen, L., et al.: SCA-CNN: spatial and channel-wise attention in convolutional networks for image captioning. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6298\u20136306. IEEE, Honolulu, HI (2017)","DOI":"10.1109\/CVPR.2017.667"},{"issue":"3","key":"13_CR22","doi-asserted-by":"publisher","first-page":"575","DOI":"10.1016\/S0733-8619(02)00103-2","volume":"21","author":"SP Vecera","year":"2003","unstructured":"Vecera, S.P., Rizzo, M.: Spatial attention: normal processes and their breakdown. Neurol. Clin. 21(3), 575\u2013607 (2003)","journal-title":"Neurol. Clin."},{"key":"13_CR23","doi-asserted-by":"crossref","unstructured":"Lucey, P., Cohn, J.F., Kanade, T., Saragih, J., Ambadar, Z., Matthews, I.: The extended Cohn-Kanade Dataset (CK+): a complete dataset for action unit and emotion-specified expression. In: 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops, pp. 94\u2013101. IEEE, San Francisco, CA, USA (2010)","DOI":"10.1109\/CVPRW.2010.5543262"},{"issue":"1","key":"13_CR24","doi-asserted-by":"publisher","first-page":"226","DOI":"10.1109\/TAFFC.2019.2946540","volume":"13","author":"HD Nguyen","year":"2022","unstructured":"Nguyen, H.D., Kim, S.H., Lee, G.S., Yang, H.J., Na, I.S., Kim, S.H.: Facial expression recognition using a temporal ensemble of multi-level convolutional neural networks. IEEE Trans. Affect. Comput. 13(1), 226\u2013237 (2022)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"7","key":"13_CR25","doi-asserted-by":"publisher","first-page":"2005","DOI":"10.1049\/ipr2.12743","volume":"17","author":"W Zhang","year":"2023","unstructured":"Zhang, W., Zhang, X., Tang, Y.: Facial expression recognition based on improved residual network. IET Image Process. 17(7), 2005\u20132014 (2023)","journal-title":"IET Image Process."}],"container-title":["Lecture Notes in Electrical Engineering","Genetic and Evolutionary Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-1535-3_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,2,13]],"date-time":"2025-02-13T12:24:46Z","timestamp":1739449486000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-1535-3_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819615346","9789819615353"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-1535-3_13","relation":{},"ISSN":["1876-1100","1876-1119"],"issn-type":[{"type":"print","value":"1876-1100"},{"type":"electronic","value":"1876-1119"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"14 February 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICGEC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Genetic and Evolutionary Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Miyazaki","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 August 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icgec2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icgec24.github.io\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}