{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T02:59:23Z","timestamp":1768273163586,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":47,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819556755","type":"print"},{"value":"9789819556762","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5676-2_29","type":"book-chapter","created":{"date-parts":[[2026,1,12]],"date-time":"2026-01-12T20:32:31Z","timestamp":1768249951000},"page":"427-439","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["LLMamba-Net: A Lightweight Network Integrating Linear Mamba for\u00a0Facial Expression Recognition"],"prefix":"10.1007","author":[{"given":"Kaidi","family":"Hu","sequence":"first","affiliation":[]},{"given":"Xiaoyu","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Guojiao","family":"Zhao","sequence":"additional","affiliation":[]},{"given":"Ruigang","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,13]]},"reference":[{"key":"29_CR1","doi-asserted-by":"crossref","unstructured":"Chen, P., Wang, Z., Mao, S., Hui, X., Hu, Y.: Dual-branch residual disentangled adversarial learning network for facial expression recognition. IEEE Signal Process. Lett. 31, 1840\u20131844 (2024)","DOI":"10.1109\/LSP.2024.3390987"},{"key":"29_CR2","doi-asserted-by":"crossref","unstructured":"Chen, S., Liu, Y., Gao, X., Han, Z.: MobileFaceNets: efficient CNNs for accurate real-time face verification on mobile devices. In: Chinese Conference on Biometric Recognition, pp. 428\u2013438. Springer (2018)","DOI":"10.1007\/978-3-319-97909-0_46"},{"key":"29_CR3","doi-asserted-by":"publisher","first-page":"1732","DOI":"10.1109\/LSP.2023.3336257","volume":"30","author":"Q Dong","year":"2023","unstructured":"Dong, Q., Ren, W., Gao, Y., Jiang, W., Liu, H.: Multi-scale attention learning network for facial expression recognition. IEEE Signal Process. Lett. 30, 1732\u20131736 (2023)","journal-title":"IEEE Signal Process. Lett."},{"key":"29_CR4","unstructured":"Dosovitskiy, A., et\u00a0al.: An image is worth 16x16 words: transformers for image recognition at scale. In: International Conference on Learning Representations (2020)"},{"key":"29_CR5","doi-asserted-by":"crossref","unstructured":"Farzaneh, A.H., Qi, X.: Facial expression recognition in the wild via deep attentive center loss. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2402\u20132411 (2021)","DOI":"10.1109\/WACV48630.2021.00245"},{"key":"29_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2024.129323","volume":"622","author":"Z Fei","year":"2025","unstructured":"Fei, Z., Zhang, B., Zhou, W., Li, X., Zhang, Y., Fei, M.: Global multi-scale extraction and local mixed multi-head attention for facial expression recognition in the wild. Neurocomputing 622, 129323 (2025)","journal-title":"Neurocomputing"},{"key":"29_CR7","unstructured":"Foret, P., Kleiner, A., Mobahi, H., Neyshabur, B.: Sharpness-aware minimization for efficiently improving generalization. In: International Conference on Learning Representations (2021)"},{"key":"29_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2024.105183","volume":"149","author":"C Ghadai","year":"2024","unstructured":"Ghadai, C., Patra, D., Okade, M.: A novel facial expression recognition model based on harnessing complementary features in multi-scale network with attention fusion. Image Vis. Comput. 149, 105183 (2024)","journal-title":"Image Vis. Comput."},{"key":"29_CR9","unstructured":"Gu, A., Dao, T.: Mamba: linear-time sequence modeling with selective state spaces. In: First Conference on Language Modeling (COLM) (2024)"},{"key":"29_CR10","doi-asserted-by":"crossref","unstructured":"Guo, L., Jin, L., Ma, G., Xu, X.: Compact selective transformer based on information entropy for facial expression recognition in the wild. In: 2023 IEEE International Conference on Image Processing (ICIP), pp. 2345\u20132349. IEEE (2023)","DOI":"10.1109\/ICIP49359.2023.10222376"},{"key":"29_CR11","doi-asserted-by":"crossref","unstructured":"Guo, Y., Zhang, L., Hu, Y., He, X., Gao, J.: MS-Celeb-1M: a dataset and benchmark for large-scale face recognition. In: Computer Vision\u2013ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11-14, 2016, Proceedings, Part III 14, pp. 87\u2013102. Springer (2016)","DOI":"10.1007\/978-3-319-46487-9_6"},{"key":"29_CR12","first-page":"127181","volume":"37","author":"D Han","year":"2025","unstructured":"Han, D., et al.: Demystify mamba in vision: a linear attention perspective. Adv. Neural. Inf. Process. Syst. 37, 127181\u2013127203 (2025)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"29_CR13","doi-asserted-by":"publisher","first-page":"220","DOI":"10.1016\/j.cag.2023.12.015","volume":"118","author":"K Hu","year":"2024","unstructured":"Hu, K., Xie, Z., Hu, Q.: Dual-resolution transformer combined with multi-layer separable convolution fusion network for real-time semantic segmentation. Comput. Graph. 118, 220\u2013232 (2024)","journal-title":"Comput. Graph."},{"key":"29_CR14","doi-asserted-by":"publisher","DOI":"10.1016\/j.imavis.2024.105053","volume":"146","author":"K Hu","year":"2024","unstructured":"Hu, K., Xie, Z., Hu, Q.: Lightweight convolutional neural networks with context broadcast transformer for real-time semantic segmentation. Image Vis. Comput. 146, 105053 (2024)","journal-title":"Image Vis. Comput."},{"key":"29_CR15","doi-asserted-by":"crossref","unstructured":"Hyeon-Woo, N., Yu-Ji, K., Heo, B., Han, D., Oh, S.J., Oh, T.H.: Scratching visual transformer\u2019s back with uniform attention. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5807\u20135818 (2023)","DOI":"10.1109\/ICCV51070.2023.00534"},{"key":"29_CR16","doi-asserted-by":"crossref","unstructured":"Jiang, S., Xing, X., Liu, F., Xu, X., Wang, L., Guo, K.: CSE-GResNet: a simple and highly efficient network for facial expression recognition. IEEE Trans. Affect. Comput. 16(3), 1732\u20131746 (2025)","DOI":"10.1109\/TAFFC.2025.3535811"},{"issue":"10","key":"29_CR17","doi-asserted-by":"publisher","first-page":"3729","DOI":"10.3390\/s22103729","volume":"22","author":"S Kim","year":"2022","unstructured":"Kim, S., Nam, J., Ko, B.C.: Facial expression recognition based on squeeze vision transformer. Sensors 22(10), 3729 (2022)","journal-title":"Sensors"},{"key":"29_CR18","doi-asserted-by":"crossref","unstructured":"Lee, J., Kim, S., Kim, S., Park, J., Sohn, K.: Context-aware emotion recognition networks. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10143\u201310152 (2019)","DOI":"10.1109\/ICCV.2019.01024"},{"key":"29_CR19","doi-asserted-by":"publisher","first-page":"6977","DOI":"10.1109\/TIP.2020.2996086","volume":"29","author":"J Lee","year":"2020","unstructured":"Lee, J., Kim, S., Kim, S., Sohn, K.: Multi-modal recurrent attention networks for facial expression recognition. IEEE Trans. Image Process. 29, 6977\u20136991 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"29_CR20","doi-asserted-by":"crossref","unstructured":"Li, S., Deng, W., Du, J.: Reliable crowdsourcing and deep locality-preserving learning for expression recognition in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2852\u20132861 (2017)","DOI":"10.1109\/CVPR.2017.277"},{"key":"29_CR21","doi-asserted-by":"publisher","first-page":"781","DOI":"10.1016\/j.ins.2022.11.068","volume":"619","author":"C Liu","year":"2023","unstructured":"Liu, C., Hirota, K., Dai, Y.: Patch attention convolutional vision transformer for facial expression recognition with occlusion. Inf. Sci. 619, 781\u2013794 (2023)","journal-title":"Inf. Sci."},{"key":"29_CR22","doi-asserted-by":"crossref","unstructured":"Liu, W., Shi, X., Liu, X.: Flipping consistent and counterfactual attention network for facial expression recognition. In: ICASSP 2024-2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 2665\u20132669. IEEE (2024)","DOI":"10.1109\/ICASSP48485.2024.10448412"},{"issue":"2","key":"29_CR23","doi-asserted-by":"publisher","first-page":"1236","DOI":"10.1109\/TAFFC.2021.3122146","volume":"14","author":"F Ma","year":"2021","unstructured":"Ma, F., Sun, B., Li, S.: Facial expression recognition with visual transformers and attentional selective fusion. IEEE Trans. Affect. Comput. 14(2), 1236\u20131248 (2021)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"2","key":"29_CR24","doi-asserted-by":"publisher","first-page":"593","DOI":"10.1109\/TAFFC.2023.3285231","volume":"15","author":"F Ma","year":"2023","unstructured":"Ma, F., Sun, B., Li, S.: Transformer-augmented network with online label correction for facial expression recognition. IEEE Trans. Affect. Comput. 15(2), 593\u2013605 (2023)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"29_CR25","doi-asserted-by":"crossref","unstructured":"Ma, N., Zhang, X., Zheng, H.T., Sun, J.: ShuffleNet V2: practical guidelines for efficient CNN architecture design. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 116\u2013131 (2018)","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"29_CR26","doi-asserted-by":"crossref","unstructured":"Mao, J., et al.: POSTER++: a simpler and stronger facial expression recognition network. Pattern Recogn. 157, 110951 (2024)","DOI":"10.1016\/j.patcog.2024.110951"},{"issue":"1","key":"29_CR27","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1109\/TAFFC.2017.2740923","volume":"10","author":"A Mollahosseini","year":"2017","unstructured":"Mollahosseini, A., Hasani, B., Mahoor, M.H.: AffectNet: a database for facial expression, valence, and arousal computing in the wild. IEEE Trans. Affect. Comput. 10(1), 18\u201331 (2017)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"4","key":"29_CR28","doi-asserted-by":"publisher","first-page":"2223","DOI":"10.1109\/TCSVT.2023.3304724","volume":"34","author":"L Qin","year":"2023","unstructured":"Qin, L., et al.: SwinFace: a multi-task transformer for face recognition, expression recognition, age estimation and attribute estimation. IEEE Trans. Circuits Syst. Video Technol. 34(4), 2223\u20132234 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"29_CR29","first-page":"10353","volume":"35","author":"Y Rao","year":"2022","unstructured":"Rao, Y., Zhao, W., Tang, Y., Zhou, J., Lim, S.N., Lu, J.: HorNet: efficient high-order spatial interactions with recursive gated convolutions. Adv. Neural. Inf. Process. Syst. 35, 10353\u201310366 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"29_CR30","doi-asserted-by":"crossref","unstructured":"She, J., Hu, Y., Shi, H., Wang, J., Shen, Q., Mei, T.: Dive into ambiguity: latent distribution mining and pairwise uncertainty estimation for facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6248\u20136257 (2021)","DOI":"10.1109\/CVPR46437.2021.00618"},{"key":"29_CR31","doi-asserted-by":"crossref","unstructured":"Shen, Z.Q., Tang, Y.Y., Yan, J.F., Li, Y., Zhao, G.Y.: MSRFormer: hybrid scale self-attention and local fast convolution transformer for facial expression recognition. In: ICASSP 2025-2025 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp.\u00a01\u20135. IEEE (2025)","DOI":"10.1109\/ICASSP49660.2025.10887738"},{"key":"29_CR32","doi-asserted-by":"crossref","unstructured":"Stergiou, A., Poppe, R., Kalliatakis, G.: Refining activation downsampling with softPool. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 10357\u201310366 (2021)","DOI":"10.1109\/ICCV48922.2021.01019"},{"key":"29_CR33","doi-asserted-by":"publisher","first-page":"337","DOI":"10.1016\/j.neunet.2023.11.033","volume":"170","author":"H Tao","year":"2024","unstructured":"Tao, H., Duan, Q.: Hierarchical attention network with progressive feature fusion for facial expression recognition. Neural Netw. 170, 337\u2013348 (2024)","journal-title":"Neural Netw."},{"key":"29_CR34","doi-asserted-by":"crossref","unstructured":"Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: Eca-net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11534\u201311542 (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"29_CR35","doi-asserted-by":"crossref","unstructured":"Wang, Y., Li, Y., Wang, G., Liu, X.: PlainUSR: chasing faster convnet for efficient super-resolution. In: Proceedings of the Asian Conference on Computer Vision, pp. 4262\u20134279 (2024)","DOI":"10.1007\/978-981-96-0911-6_15"},{"issue":"3","key":"29_CR36","doi-asserted-by":"publisher","first-page":"2035","DOI":"10.1007\/s00371-023-02900-3","volume":"40","author":"H Xia","year":"2024","unstructured":"Xia, H., Lu, L., Song, S.: Feature fusion of multi-granularity and multi-scale for facial expression recognition. Vis. Comput. 40(3), 2035\u20132047 (2024)","journal-title":"Vis. Comput."},{"key":"29_CR37","doi-asserted-by":"crossref","unstructured":"Xu, J., Li, Y., Yang, G., He, L., Luo, K.: Multiscale facial expression recognition based on dynamic global and static local attention. IEEE Trans. Affect. Comput. 16(2), 683\u2013696 (2024)","DOI":"10.1109\/TAFFC.2024.3458464"},{"key":"29_CR38","doi-asserted-by":"crossref","unstructured":"Xue, F., Wang, Q., Guo, G.: Transfer: learning relation-aware facial expression representations with transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3601\u20133610 (2021)","DOI":"10.1109\/ICCV48922.2021.00358"},{"issue":"4","key":"29_CR39","doi-asserted-by":"publisher","first-page":"3244","DOI":"10.1109\/TAFFC.2022.3226473","volume":"14","author":"F Xue","year":"2022","unstructured":"Xue, F., Wang, Q., Tan, Z., Ma, Z., Guo, G.: Vision transformer with attentive pooling for robust facial expression recognition. IEEE Trans. Affect. Comput. 14(4), 3244\u20133256 (2022)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"29_CR40","doi-asserted-by":"crossref","unstructured":"Zeng, D., Lin, Z., Yan, X., Liu, Y., Wang, F., Tang, B.: Face2Exp: combating data biases for facial expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20291\u201320300 (2022)","DOI":"10.1109\/CVPR52688.2022.01965"},{"issue":"17","key":"29_CR41","doi-asserted-by":"publisher","first-page":"3595","DOI":"10.3390\/electronics12173595","volume":"12","author":"S Zhang","year":"2023","unstructured":"Zhang, S., Zhang, Y., Zhang, Y., Wang, Y., Song, Z.: A dual-direction attention mixed feature network for facial expression recognition. Electronics 12(17), 3595 (2023)","journal-title":"Electronics"},{"key":"29_CR42","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Tian, X., Zhang, Z., Xu, X.: Lightweight multi-level information fusion network for facial expression recognition. In: International Conference on Multimedia Modeling, pp. 151\u2013163. Springer (2023)","DOI":"10.1007\/978-3-031-27818-1_13"},{"key":"29_CR43","doi-asserted-by":"publisher","first-page":"1498","DOI":"10.1109\/TIP.2023.3243853","volume":"32","author":"Y Zhang","year":"2023","unstructured":"Zhang, Y., Li, W., Sun, W., Tao, R., Du, Q.: Single-source domain expansion network for cross-scene hyperspectral image classification. IEEE Trans. Image Process. 32, 1498\u20131512 (2023)","journal-title":"IEEE Trans. Image Process."},{"key":"29_CR44","doi-asserted-by":"publisher","first-page":"370","DOI":"10.1016\/j.ins.2023.02.056","volume":"630","author":"Z Zhang","year":"2023","unstructured":"Zhang, Z., Tian, X., Zhang, Y., Guo, K., Xu, X.: Enhanced discriminative global-local feature learning with priority for facial expression recognition. Inf. Sci. 630, 370\u2013384 (2023)","journal-title":"Inf. Sci."},{"key":"29_CR45","doi-asserted-by":"publisher","first-page":"6544","DOI":"10.1109\/TIP.2021.3093397","volume":"30","author":"Z Zhao","year":"2021","unstructured":"Zhao, Z., Liu, Q., Wang, S.: Learning deep global multi-scale and local attention features for facial expression recognition in the wild. IEEE Trans. Image Process. 30, 6544\u20136556 (2021)","journal-title":"IEEE Trans. Image Process."},{"key":"29_CR46","doi-asserted-by":"crossref","unstructured":"Zheng, C., Mendieta, M., Chen, C.: POSTER: a pyramid cross-fusion transformer network for facial expression recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3146\u20133155 (2023)","DOI":"10.1109\/ICCVW60793.2023.00339"},{"key":"29_CR47","unstructured":"Zhu, L., Liao, B., Zhang, Q., Wang, X., Liu, W., Wang, X.: Vision Mamba: efficient visual representation learning with bidirectional state space model. In: Proceedings of the 41st International Conference on Machine Learning. ICML\u201924, JMLR (2024)"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5676-2_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,12]],"date-time":"2026-01-12T20:32:35Z","timestamp":1768249955000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5676-2_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819556755","9789819556762"],"references-count":47,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5676-2_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"13 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}