{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,16]],"date-time":"2025-05-16T11:24:37Z","timestamp":1747394677296,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":23,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819601219"},{"type":"electronic","value":"9789819601226"}],"license":[{"start":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T00:00:00Z","timestamp":1731369600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,12]],"date-time":"2024-11-12T00:00:00Z","timestamp":1731369600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-0122-6_30","type":"book-chapter","created":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T18:23:42Z","timestamp":1731781422000},"page":"347-359","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["A Facial Expression Recognition Model Based on\u00a0a\u00a0Hybrid Attention Mechanism with\u00a0Multiple Information Spaces and\u00a0Channels"],"prefix":"10.1007","author":[{"given":"Weizhi","family":"Xie","sequence":"first","affiliation":[]},{"given":"Yifeng","family":"Yao","sequence":"additional","affiliation":[]},{"given":"Pengcheng","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,12]]},"reference":[{"key":"30_CR1","doi-asserted-by":"crossref","unstructured":"Altaha, M.A., Jarraya, I., Hamdani, T.M., Alimi, A.M.: Facial expression recognition based on Arcface features and Tinysiamese network. In: 2023 International Conference on Cyberworlds (CW), pp. 24\u201331. IEEE (2023)","DOI":"10.1109\/CW58918.2023.00014"},{"key":"30_CR2","doi-asserted-by":"crossref","unstructured":"Borgalli, M.R.A., Surve, S.: Deep learning for facial emotion recognition using custom CNN architecture. J. Phys. Conf. Seri. 2236, 012004 (2022)","DOI":"10.1088\/1742-6596\/2236\/1\/012004"},{"key":"30_CR3","doi-asserted-by":"crossref","unstructured":"Gao, Y., Xie, Y., Hu, Z.Z., Chen, T., Lin, L.: Adaptive global-local representation learning and selection for cross-domain facial expression recognition. IEEE Trans. Multimedia (2024)","DOI":"10.1109\/TMM.2024.3355637"},{"key":"30_CR4","doi-asserted-by":"publisher","first-page":"58","DOI":"10.1016\/j.patrec.2021.01.029","volume":"145","author":"D Gera","year":"2021","unstructured":"Gera, D., Balasubramanian, S.: Landmark guidance independent spatio-channel attention and complementary context information based facial expression recognition. Pattern Recogn. Lett. 145, 58\u201366 (2021)","journal-title":"Pattern Recogn. Lett."},{"key":"30_CR5","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1007\/978-3-642-42051-1_16","volume-title":"Neural Information Processing","author":"IJ Goodfellow","year":"2013","unstructured":"Goodfellow, I.J., et al.: Challenges in representation learning: a report on three machine learning contests. In: Lee, M., Hirose, A., Hou, Z.-G., Kil, R.M. (eds.) ICONIP 2013. LNCS, vol. 8228, pp. 117\u2013124. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-3-642-42051-1_16"},{"issue":"2","key":"30_CR6","doi-asserted-by":"publisher","first-page":"5207","DOI":"10.1007\/s11042-023-15297-x","volume":"83","author":"Z Guo","year":"2024","unstructured":"Guo, Z., Liu, Y., Liu, X., Pan, Z., Liu, S., Fan, Y.: Ltval: label transfer virtual adversarial learning framework for source-free facial expression recognition. Multimedia Tools Appl. 83(2), 5207\u20135228 (2024)","journal-title":"Multimedia Tools Appl."},{"key":"30_CR7","doi-asserted-by":"crossref","unstructured":"Guo, Z., Wei, B., Liu, J., Liu, X., Zhang, Z., Wang, Y.: USTST: unsupervised self-training similarity transfer for cross-domain facial expression recognition. Multimedia Tools Appl. 1\u201321 (2023)","DOI":"10.1007\/s11042-023-17317-2"},{"issue":"6","key":"30_CR8","doi-asserted-by":"publisher","first-page":"2731","DOI":"10.1007\/s11760-023-02490-6","volume":"17","author":"R Helaly","year":"2023","unstructured":"Helaly, R., Messaoud, S., Bouaafia, S., Hajjaji, M.A., Mtibaa, A.: Dtl-i-resnet18: facial emotion recognition based on deep transfer learning and improved resnet18. SIViP 17(6), 2731\u20132744 (2023)","journal-title":"SIViP"},{"key":"30_CR9","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"30_CR10","doi-asserted-by":"crossref","unstructured":"Jeong, J.Y., Hong, Y.G., Kim, D., Jeong, J.W., Jung, Y., Kim, S.H.: Classification of facial expression in-the-wild based on ensemble of multi-head cross attention networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2353\u20132358 (2022)","DOI":"10.1109\/CVPRW56347.2022.00262"},{"key":"30_CR11","doi-asserted-by":"crossref","unstructured":"Le\u00a0Ngwe, J., Lim, K.M., Lee, C.P., Ong, T.S., Alqahtani, A.: Patt-lite: lightweight patch and attention mobilenet for challenging facial expression recognition. IEEE Access (2024)","DOI":"10.1109\/ACCESS.2024.3407108"},{"issue":"3","key":"30_CR12","doi-asserted-by":"publisher","first-page":"1195","DOI":"10.1109\/TAFFC.2020.2981446","volume":"13","author":"S Li","year":"2020","unstructured":"Li, S., Deng, W.: Deep facial expression recognition: a survey. IEEE Trans. Affect. Comput. 13(3), 1195\u20131215 (2020)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"30_CR13","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2852\u20132861 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"30_CR14","doi-asserted-by":"crossref","unstructured":"Misra, D., Nalamada, T., Arasanipalai, A.U., Hou, Q.: Rotate to attend: convolutional triplet attention module. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 3139\u20133148 (2021)","DOI":"10.1109\/WACV48630.2021.00318"},{"key":"30_CR15","doi-asserted-by":"crossref","unstructured":"Ouyang, D., et al.: Efficient multi-scale attention module with cross-spatial learning. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.\u00a01\u20135. IEEE (2023)","DOI":"10.1109\/ICASSP49357.2023.10096516"},{"key":"30_CR16","doi-asserted-by":"crossref","unstructured":"Park, G., Han, C., Yoon, W., Kim, D.: MHSAN: multi-head self-attention network for visual semantic embedding. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1518\u20131526 (2020)","DOI":"10.1109\/WACV45572.2020.9093548"},{"issue":"9","key":"30_CR17","doi-asserted-by":"publisher","first-page":"4366","DOI":"10.3390\/app12094366","volume":"12","author":"X Peng","year":"2022","unstructured":"Peng, X., Gu, Y., Zhang, P.: Au-guided unsupervised domain-adaptive facial expression recognition. Appl. Sci. 12(9), 4366 (2022)","journal-title":"Appl. Sci."},{"key":"30_CR18","doi-asserted-by":"crossref","unstructured":"Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: ECA-Net: efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11534\u201311542 (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"issue":"2","key":"30_CR19","doi-asserted-by":"publisher","first-page":"199","DOI":"10.3390\/biomimetics8020199","volume":"8","author":"Z Wen","year":"2023","unstructured":"Wen, Z., Lin, W., Wang, T., Xu, G.: Distract your attention: multi-head cross attention network for facial expression recognition. Biomimetics 8(2), 199 (2023)","journal-title":"Biomimetics"},{"key":"30_CR20","doi-asserted-by":"crossref","unstructured":"Xiang, J., Zhu, G.: Joint face detection and facial expression recognition with MTCNN. In: 2017 4th International Conference on Information Science and Control Engineering (ICISCE), pp. 424\u2013427. IEEE (2017)","DOI":"10.1109\/ICISCE.2017.95"},{"key":"30_CR21","doi-asserted-by":"crossref","unstructured":"Xie, Y., Gao, Y., Lin, J., Chen, T.: Learning consistent global-local representation for cross-domain facial expression recognition. In: 2022 26th International Conference on Pattern Recognition (ICPR), pp. 2489\u20132495. IEEE (2022)","DOI":"10.1109\/ICPR56361.2022.9956069"},{"key":"30_CR22","unstructured":"Yang, L., Zhang, R.Y., Li, L., Xie, X.: SIMAM: a simple, parameter-free attention module for convolutional neural networks. In: International Conference on Machine Learning, pp. 11863\u201311874. PMLR (2021)"},{"key":"30_CR23","doi-asserted-by":"crossref","unstructured":"Zhang, L., Wang, X., Du, J.: Driver facial emotion monitoring method based on deep learning. In: 2023 5th International Conference on Frontiers Technology of Information and Computer (ICFTIC), pp. 356\u2013359. IEEE (2023)","DOI":"10.1109\/ICFTIC59930.2023.10456333"}],"container-title":["Lecture Notes in Computer Science","PRICAI 2024: Trends in Artificial Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-0122-6_30","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,16]],"date-time":"2024-11-16T19:21:25Z","timestamp":1731784885000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-0122-6_30"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,12]]},"ISBN":["9789819601219","9789819601226"],"references-count":23,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-0122-6_30","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,12]]},"assertion":[{"value":"12 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRICAI","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Pacific Rim International Conference on Artificial Intelligence","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kyoto","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Japan","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 November 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"25 November 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"pricai2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.pricai.org\/2024\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}