{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T02:42:59Z","timestamp":1768272179120,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":39,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556755","type":"print"},{"value":"9789819556762","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5676-2_37","type":"book-chapter","created":{"date-parts":[[2026,1,12]],"date-time":"2026-01-12T20:32:01Z","timestamp":1768249921000},"page":"551-564","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["DDANet: A Dual-Path Direction-Aware Network with\u00a0Cross-Direction Attention for\u00a0Facial Expression Recognition"],"prefix":"10.1007","author":[{"given":"Cheng","family":"Li","sequence":"first","affiliation":[]},{"given":"Xianzhong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Jiayang","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,13]]},"reference":[{"key":"37_CR1","doi-asserted-by":"crossref","unstructured":"Barsoum, E., Zhang, C., Ferrer, C.C., Zhang, Z.: Training deep networks for facial expression recognition with crowd-sourced label distribution. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction, pp. 279\u2013283 (2016)","DOI":"10.1145\/2993148.2993165"},{"issue":"8","key":"37_CR2","doi-asserted-by":"publisher","first-page":"3848","DOI":"10.1109\/TCSVT.2023.3234312","volume":"33","author":"D Chen","year":"2023","unstructured":"Chen, D., Wen, G., Li, H., Chen, R., Li, C.: Multi-relations aware network for in-the-wild facial expression recognition. IEEE Trans. Circuits Syst. Video Technol. 33(8), 3848\u20133859 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"37_CR3","doi-asserted-by":"crossref","unstructured":"Dalal, N., Triggs, B.: Histograms of oriented gradients for human detection. In: 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2005), vol.\u00a01, pp. 886\u2013893. IEEE (2005)","DOI":"10.1109\/CVPR.2005.177"},{"key":"37_CR4","first-page":"1140","volume":"35","author":"MH Guo","year":"2022","unstructured":"Guo, M.H., Lu, C.Z., Hou, Q., Liu, Z., Cheng, M.M., Hu, S.M.: SegNeXt: rethinking convolutional attention design for semantic segmentation. Adv. Neural. Inf. Process. Syst. 35, 1140\u20131156 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"37_CR5","doi-asserted-by":"publisher","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"37_CR6","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13713\u201313722 (2021)","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"37_CR7","doi-asserted-by":"crossref","unstructured":"Huang, Y., Peng, J., Cai, Z., Guo, J., Chen, G., Tan, S.: Facial expression recognition with age-group expression feature learning. In: 2024 International Joint Conference on Neural Networks (IJCNN), pp.\u00a01\u20138. IEEE (2024)","DOI":"10.1109\/IJCNN60899.2024.10649944"},{"key":"37_CR8","doi-asserted-by":"crossref","unstructured":"Hwang, H., Kim, S., Park, W.J., Seo, J., Ko, K., Yeo, H.: Vision transformer equipped with neural resizer on facial expression recognition task. In: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2614\u20132618. IEEE (2022)","DOI":"10.1109\/ICASSP43922.2022.9746526"},{"issue":"10","key":"37_CR9","doi-asserted-by":"publisher","first-page":"3729","DOI":"10.3390\/s22103729","volume":"22","author":"S Kim","year":"2022","unstructured":"Kim, S., Nam, J., Ko, B.C.: Facial expression recognition based on squeeze vision transformer. Sensors 22(10), 3729 (2022)","journal-title":"Sensors"},{"key":"37_CR10","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2852\u20132861 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"37_CR11","doi-asserted-by":"crossref","unstructured":"Liu, H., et al.: MMATrans: muscle movement aware representation learning for facial expression recognition via transformers. IEEE Trans. Ind. Inform. (2024)","DOI":"10.1109\/TII.2024.3431640"},{"issue":"9","key":"37_CR12","doi-asserted-by":"publisher","first-page":"6253","DOI":"10.1109\/TCSVT.2022.3165321","volume":"32","author":"H Liu","year":"2022","unstructured":"Liu, H., Cai, H., Lin, Q., Li, X., Xiao, H.: Adaptive multilayer perceptual attention network for facial expression recognition. IEEE Trans. Circuits Syst. Video Technol. 32(9), 6253\u20136266 (2022). https:\/\/doi.org\/10.1109\/TCSVT.2022.3165321","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"37_CR13","doi-asserted-by":"crossref","unstructured":"Liu, W., Shi, X., Liu, X.: Flipping consistent and counterfactual attention network for facial expression recognition. In: ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2665\u20132669. IEEE (2024)","DOI":"10.1109\/ICASSP48485.2024.10448412"},{"key":"37_CR14","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","volume":"60","author":"DG Lowe","year":"2004","unstructured":"Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vision 60, 91\u2013110 (2004)","journal-title":"Int. J. Comput. Vision"},{"issue":"2","key":"37_CR15","doi-asserted-by":"publisher","first-page":"593","DOI":"10.1109\/TAFFC.2023.3285231","volume":"15","author":"F Ma","year":"2024","unstructured":"Ma, F., Sun, B., Li, S.: Transformer-augmented network with online label correction for facial expression recognition. IEEE Trans. Affect. Comput. 15(2), 593\u2013605 (2024). https:\/\/doi.org\/10.1109\/TAFFC.2023.3285231","journal-title":"IEEE Trans. Affect. Comput."},{"key":"37_CR16","doi-asserted-by":"crossref","unstructured":"Ma, X., Ma, Y.: Relation-aware network for facial expression recognition. In: 2023 IEEE 17th International Conference on Automatic Face and Gesture Recognition (FG), pp.\u00a01\u20137. IEEE (2023)","DOI":"10.1109\/FG57933.2023.10042525"},{"key":"37_CR17","unstructured":"Van\u00a0der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11) (2008)"},{"issue":"1","key":"37_CR18","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2017","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: AffectNet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10(1), 18\u201331 (2017)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"37_CR19","unstructured":"Ning, M., Salah, A.A., Ertugrul, I.O.: Representation learning and identity adversarial training for facial behavior understanding. arXiv preprint arXiv:2407.11243 (2024)"},{"key":"37_CR20","doi-asserted-by":"crossref","unstructured":"Olivares-Mercado, J., Toscano-Medina, K., Sanchez-Perez, G., Portillo-Portillo, J., Perez-Meana, H., Benitez-Garcia, G.: Analysis of hand-crafted and learned feature extraction methods for real-time facial expression recognition. In: 2019 7th International Workshop on Biometrics and Forensics (IWBF), pp.\u00a01\u20136. IEEE (2019)","DOI":"10.1109\/IWBF.2019.8739178"},{"key":"37_CR21","doi-asserted-by":"crossref","unstructured":"Savchenko, A.V.: Facial expression and attributes recognition based on multi-task learning of lightweight neural networks. In: 2021 IEEE 19th International Symposium on Intelligent Systems and Informatics (SISY), pp. 119\u2013124. IEEE (2021)","DOI":"10.1109\/SISY52375.2021.9582508"},{"key":"37_CR22","doi-asserted-by":"crossref","unstructured":"She, J., Hu, Y., Shi, H., Wang, J., Shen, Q., Mei, T.: Dive into ambiguity: latent distribution mining and pairwise uncertainty estimation for facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6248\u20136257 (2021)","DOI":"10.1109\/CVPR46437.2021.00618"},{"key":"37_CR23","unstructured":"Shi, J., Zhu, S., Liang, Z.: Learning to amend facial expression representation via de-albino and affinity. arXiv preprint arXiv:2103.10189 (2021)"},{"key":"37_CR24","doi-asserted-by":"crossref","unstructured":"Wang, C., Chen, L., Wang, L., Li, Z., Lv, X.: QCS: feature refining from quadruplet cross similarity for facial expression recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a039, pp. 7563\u20137572 (2025)","DOI":"10.1609\/aaai.v39i7.32814"},{"key":"37_CR25","doi-asserted-by":"publisher","first-page":"4057","DOI":"10.1109\/TIP.2019.2956143","volume":"29","author":"K Wang","year":"2020","unstructured":"Wang, K., Peng, X., Yang, J., Meng, D., Qiao, Y.: Region attention networks for pose and occlusion robust facial expression recognition. IEEE Trans. Image Process. 29, 4057\u20134069 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"37_CR26","doi-asserted-by":"crossref","unstructured":"Wang, L., Kang, X., Nakagawa, S., Ren, F.: CenterMatch: a center matching method for semi-supervised facial expression recognition. In: Chinese Conference on Pattern Recognition and Computer Vision (PRCV), pp. 371\u2013383. Springer (2023)","DOI":"10.1007\/978-981-99-8537-1_30"},{"key":"37_CR27","unstructured":"Wang, Y., et al.: A survey on facial expression recognition of static and dynamic emotions. arXiv preprint arXiv:2408.15777 (2024)"},{"issue":"2","key":"37_CR28","doi-asserted-by":"publisher","first-page":"199","DOI":"10.3390\/biomimetics8020199","volume":"8","author":"Z Wen","year":"2023","unstructured":"Wen, Z., Lin, W., Wang, T., Xu, G.: Distract your attention: multi-head cross attention network for facial expression recognition. Biomimetics 8(2), 199 (2023)","journal-title":"Biomimetics"},{"key":"37_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01261-8_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Wu","year":"2018","unstructured":"Wu, Y., He, K.: Group normalization. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 3\u201319. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01261-8_1"},{"key":"37_CR30","doi-asserted-by":"publisher","first-page":"50","DOI":"10.1016\/j.neucom.2018.07.028","volume":"317","author":"Z Yu","year":"2018","unstructured":"Yu, Z., Liu, G., Liu, Q., Deng, J.: Spatio-temporal convolutional features with nested LSTM for facial expression recognition. Neurocomputing 317, 50\u201357 (2018)","journal-title":"Neurocomputing"},{"key":"37_CR31","doi-asserted-by":"crossref","unstructured":"Yu, Z., Zhang, C.: Image based static facial expression recognition with multiple deep network learning. In: Proceedings of the 2015 ACM on International Conference on Multimodal Interaction, pp. 435\u2013442 (2015)","DOI":"10.1145\/2818346.2830595"},{"issue":"3","key":"37_CR32","doi-asserted-by":"publisher","first-page":"839","DOI":"10.1109\/TCYB.2017.2788081","volume":"49","author":"T Zhang","year":"2018","unstructured":"Zhang, T., Zheng, W., Cui, Z., Zong, Y., Li, Y.: Spatial-temporal recurrent neural network for emotion recognition. IEEE Trans. Cybern. 49(3), 839\u2013847 (2018)","journal-title":"IEEE Trans. Cybern."},{"key":"37_CR33","first-page":"17616","volume":"34","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Wang, C., Deng, W.: Relative uncertainty learning for facial expression recognition. Adv. Neural. Inf. Process. Syst. 34, 17616\u201317627 (2021)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"37_CR34","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Wang, C., Ling, X., Deng, W.: Learn from all: erasing attention consistency for noisy label facial expression recognition. In: European Conference on Computer Vision, pp. 418\u2013434. Springer (2022)","DOI":"10.1007\/978-3-031-19809-0_24"},{"key":"37_CR35","doi-asserted-by":"crossref","unstructured":"Zhao, Z., Liu, Q., Zhou, F.: Robust lightweight facial expression recognition network with label distribution training. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a035, pp. 3510\u20133519 (2021)","DOI":"10.1609\/aaai.v35i4.16465"},{"key":"37_CR36","doi-asserted-by":"crossref","unstructured":"Zheng, C., Mendieta, M., Chen, C.: Poster: a pyramid cross-fusion transformer network for facial expression recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3146\u20133155 (2023)","DOI":"10.1109\/ICCVW60793.2023.00339"},{"key":"37_CR37","doi-asserted-by":"crossref","unstructured":"Zhi, R., Flierl, M., Ruan, Q., Kleijn, W.B.: Graph-preserving sparse nonnegative matrix factorization with application to facial expression recognition. IEEE Trans. Syst. Man Cybernet. Part B (Cybernet.) 41(1), 38\u201352 (2010)","DOI":"10.1109\/TSMCB.2010.2044788"},{"key":"37_CR38","doi-asserted-by":"crossref","unstructured":"Zhou, B., Khosla, A., Lapedriza, A., Oliva, A., Torralba, A.: Learning deep features for discriminative localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921\u20132929 (2016)","DOI":"10.1109\/CVPR.2016.319"},{"key":"37_CR39","doi-asserted-by":"crossref","unstructured":"Zhou, H., Huang, S., Zhang, F., Xu, C.: CEPrompt: cross-modal emotion-aware prompting for facial expression recognition. IEEE Trans. Circ. Syst. Video Technol. (2024)","DOI":"10.1109\/TCSVT.2024.3424777"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5676-2_37","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,12]],"date-time":"2026-01-12T20:32:06Z","timestamp":1768249926000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5676-2_37"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556755","9789819556762"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5676-2_37","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"13 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}