{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,15]],"date-time":"2026-04-15T19:31:16Z","timestamp":1776281476457,"version":"3.50.1"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,5,16]],"date-time":"2024-05-16T00:00:00Z","timestamp":1715817600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,16]],"date-time":"2024-05-16T00:00:00Z","timestamp":1715817600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62276118"],"award-info":[{"award-number":["62276118"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62276118"],"award-info":[{"award-number":["62276118"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,6]]},"DOI":"10.1007\/s00530-024-01352-6","type":"journal-article","created":{"date-parts":[[2024,5,16]],"date-time":"2024-05-16T10:02:14Z","timestamp":1715853734000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":15,"title":["Dual-stream network with cross-layer attention and similarity constraint for micro-expression recognition"],"prefix":"10.1007","volume":"30","author":[{"given":"Gang","family":"Wang","sequence":"first","affiliation":[]},{"given":"Shucheng","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,16]]},"reference":[{"issue":"2","key":"1352_CR1","first-page":"5","volume":"1","author":"P Ekman","year":"2009","unstructured":"Ekman, P.: Lie catching and microexpressions. Philos. Decep. 1(2), 5 (2009)","journal-title":"Philos. Decep."},{"issue":"4","key":"1352_CR2","doi-asserted-by":"publisher","first-page":"217","DOI":"10.1007\/s10919-013-0159-8","volume":"37","author":"W-J Yan","year":"2013","unstructured":"Yan, W.-J., Wu, Q., Liang, J., Chen, Y.-H., Fu, X.: How fast are the leaked facial expressions: the duration of micro-expressions. J. Nonverbal Behav. 37(4), 217\u2013230 (2013)","journal-title":"J. Nonverbal Behav."},{"issue":"4","key":"1352_CR3","doi-asserted-by":"publisher","first-page":"2028","DOI":"10.1109\/TAFFC.2022.3205170","volume":"13","author":"Y Li","year":"2022","unstructured":"Li, Y., Wei, J., Liu, Y., Kauttonen, J., Zhao, G.: Deep learning for micro-expression recognition: a survey. IEEE Trans. Affect. Comput. 13(4), 2028\u20132046 (2022)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"9","key":"1352_CR4","first-page":"5826","volume":"44","author":"X Ben","year":"2021","unstructured":"Ben, X., Ren, Y., Zhang, J., Wang, S.-J., Kpalma, K., Meng, W., Liu, Y.-J.: Video-based facial micro-expression analysis: a survey of datasets, features and algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 44(9), 5826\u20135846 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"1352_CR5","first-page":"2782","volume":"45","author":"J Li","year":"2022","unstructured":"Li, J., Dong, Z., Lu, S., Wang, S.-J., Yan, W.-J., Ma, Y., Liu, Y., Huang, C., Fu, X.: Cas (me) 3: a third generation facial spontaneous micro-expression database with depth information and high ecological validity. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 2782\u20132800 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1352_CR6","unstructured":"Frank, M., Herbasz, M., Sinuk, K., Keller, A., Nolan, C.: I see how you feel: training laypeople and professionals to recognize fleeting emotions. In: The Annual Meeting of the International Communication Association. Sheraton New York, New York City, pp. 1\u201335 (2009)"},{"key":"1352_CR7","volume-title":"What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS)","author":"EL Rosenberg","year":"2020","unstructured":"Rosenberg, E.L., Ekman, P.: What the Face Reveals: Basic and Applied Studies of Spontaneous Expression Using the Facial Action Coding System (FACS). Oxford University Press, Oxford (2020)"},{"key":"1352_CR8","doi-asserted-by":"crossref","unstructured":"Liong, S.-T., Gan, Y.S., See, J., Khor, H.-Q., Huang, Y.-C.: Shallow triple stream three-dimensional cnn (ststnet) for micro-expression recognition. In: 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pp. 1\u20135 (2019). IEEE","DOI":"10.1109\/FG.2019.8756567"},{"key":"1352_CR9","doi-asserted-by":"crossref","unstructured":"Zhou, L., Mao, Q., Xue, L.: Dual-inception network for cross-database micro-expression recognition. In: 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pp. 1\u20135 (2019). IEEE","DOI":"10.1109\/FG.2019.8756579"},{"issue":"4","key":"1352_CR10","doi-asserted-by":"publisher","first-page":"1967","DOI":"10.1007\/s00530-023-01080-3","volume":"29","author":"G Wang","year":"2023","unstructured":"Wang, G., Huang, S., Tao, Z.: Shallow multi-branch attention convolutional neural network for micro-expression recognition. Multimedia Syst. 29(4), 1967\u20131980 (2023)","journal-title":"Multimedia Syst."},{"issue":"6","key":"1352_CR11","doi-asserted-by":"publisher","first-page":"3863","DOI":"10.1007\/s00530-023-01164-0","volume":"29","author":"H Zhou","year":"2023","unstructured":"Zhou, H., Huang, S., Xu, Y.: Inceptr: micro-expression recognition integrating inception-cbam and vision transformer. Multimedia Syst. 29(6), 3863\u20133876 (2023)","journal-title":"Multimedia Syst."},{"key":"1352_CR12","doi-asserted-by":"crossref","unstructured":"Zhai, Z., Zhao, J., Long, C., Xu, W., He, S., Zhao, H.: Feature representation learning with adaptive displacement generation and transformer fusion for micro-expression recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 22086\u201322095 (2023)","DOI":"10.1109\/CVPR52729.2023.02115"},{"issue":"2","key":"1352_CR13","doi-asserted-by":"publisher","first-page":"585","DOI":"10.1007\/s00371-023-02803-3","volume":"40","author":"Y Gan","year":"2024","unstructured":"Gan, Y., Lien, S.-E., Chiang, Y.-C., Liong, S.-T.: Laenet for micro-expression recognition. Vis. Comput. 40(2), 585\u2013599 (2024)","journal-title":"Vis. Comput."},{"key":"1352_CR14","doi-asserted-by":"publisher","first-page":"354","DOI":"10.1016\/j.neucom.2020.06.005","volume":"410","author":"C Wang","year":"2020","unstructured":"Wang, C., Peng, M., Bi, T., Chen, T.: Micro-attention for micro-expression recognition. Neurocomputing 410, 354\u2013362 (2020)","journal-title":"Neurocomputing"},{"key":"1352_CR15","doi-asserted-by":"publisher","first-page":"8590","DOI":"10.1109\/TIP.2020.3018222","volume":"29","author":"Z Xia","year":"2020","unstructured":"Xia, Z., Peng, W., Khor, H.-Q., Feng, X., Zhao, G.: Revealing the invisible with model and data shrinking for composite-database micro-expression recognition. IEEE Trans. Image Process. 29, 8590\u20138605 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"1352_CR16","first-page":"1541","volume":"33","author":"D Liu","year":"2021","unstructured":"Liu, D., Liang, Z., Sun, Y.: Micro-expression recognition method based on spatial attention mechanism and optical flow features. J. Comput. Aid. Des. Comput. Gr. 33, 1541\u20131552 (2021)","journal-title":"J. Comput. Aid. Des. Comput. Gr."},{"key":"1352_CR17","doi-asserted-by":"crossref","unstructured":"Chaudhry, R., Ravichandran, A., Hager, G., Vidal, R.: Histograms of oriented optical flow and binet-cauchy kernels on nonlinear dynamical systems for the recognition of human actions. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1932\u20131939 (2009)","DOI":"10.1109\/CVPRW.2009.5206821"},{"key":"1352_CR18","doi-asserted-by":"crossref","unstructured":"Pfister, T., Li, X., Zhao, G., Pietik\u00e4inen, M.: Recognising spontaneous facial micro-expressions. In: 2011 International Conference on Computer Vision, pp. 1449\u20131456 (2011)","DOI":"10.1109\/ICCV.2011.6126401"},{"issue":"03","key":"1352_CR19","doi-asserted-by":"publisher","first-page":"394","DOI":"10.1109\/TAFFC.2017.2723386","volume":"10","author":"S Happy","year":"2019","unstructured":"Happy, S., Routray, A.: Fuzzy histogram of optical flow orientations for micro-expression recognition. IEEE Trans. Affect. Comput. 10(03), 394\u2013406 (2019)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"01","key":"1352_CR20","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1109\/TAFFC.2017.2713359","volume":"10","author":"X Huang","year":"2019","unstructured":"Huang, X., Wang, S.-J., Liu, X., Zhao, G., Feng, X., Pietikainen, M.: Discriminative spatiotemporal local binary pattern with revisited integral projection for spontaneous facial micro-expression recognition. IEEE Trans. Affect. Comput. 10(01), 32\u201347 (2019)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1352_CR21","doi-asserted-by":"publisher","first-page":"110149","DOI":"10.1109\/ACCESS.2022.3214808","volume":"10","author":"G Wang","year":"2022","unstructured":"Wang, G., Huang, S., Dong, Z.: Haphazard cuboids feature extraction for micro-expression recognition. IEEE Access 10, 110149\u2013110162 (2022)","journal-title":"IEEE Access"},{"key":"1352_CR22","doi-asserted-by":"publisher","first-page":"22","DOI":"10.1016\/j.neucom.2021.12.088","volume":"479","author":"J Wei","year":"2022","unstructured":"Wei, J., Lu, G., Yan, J., Zong, Y.: Learning two groups of discriminative features for micro-expression recognition. Neurocomputing 479, 22\u201336 (2022)","journal-title":"Neurocomputing"},{"issue":"4","key":"1352_CR23","doi-asserted-by":"publisher","first-page":"299","DOI":"10.1109\/TAFFC.2015.2485205","volume":"7","author":"Y-J Liu","year":"2015","unstructured":"Liu, Y.-J., Zhang, J.-K., Yan, W.-J., Wang, S.-J., Zhao, G., Fu, X.: A main directional mean optical flow feature for spontaneous micro-expression recognition. IEEE Trans. Affect. Comput. 7(4), 299\u2013310 (2015)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1352_CR24","doi-asserted-by":"crossref","unstructured":"Dong, Z., Wang, G., Lu, S., Yan, W.-J., Wang, S.-J.: A brief guide: Code for spontaneous expressions and micro-expressions in videos. In: Proceedings of the 1st Workshop on Facial Micro-Expression: Advanced Techniques for Facial Expressions Generation and Spotting, pp. 31\u201337 (2021)","DOI":"10.1145\/3476100.3484464"},{"key":"1352_CR25","doi-asserted-by":"crossref","unstructured":"Dong, Z., Wang, G., Lu, S., Li, J., Yan, W., Wang, S.-J.: Spontaneous facial expressions and micro-expressions coding: from brain to face. Front. Psychol. 5808 (2022)","DOI":"10.3389\/fpsyg.2021.784834"},{"key":"1352_CR26","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1016\/j.image.2017.11.006","volume":"62","author":"S-T Liong","year":"2018","unstructured":"Liong, S.-T., See, J., Wong, K., Phan, R.C.-W.: Less is more: micro-expression recognition from video using apex frame. Signal Process. Image Commun. 62, 82\u201392 (2018)","journal-title":"Signal Process. Image Commun."},{"key":"1352_CR27","doi-asserted-by":"publisher","first-page":"1745","DOI":"10.3389\/fpsyg.2017.01745","volume":"8","author":"M Peng","year":"2017","unstructured":"Peng, M., Wang, C., Chen, T., Liu, G., Fu, X.: Dual temporal scale convolutional neural network for micro-expression recognition. Front. Psychol. 8, 1745 (2017)","journal-title":"Front. Psychol."},{"key":"1352_CR28","doi-asserted-by":"crossref","unstructured":"Khor, H.-Q., See, J., Liong, S.-T., Phan, R.C., Lin, W.: Dual-stream shallow networks for facial micro-expression recognition. In: 2019 IEEE International Conference on Image Processing (ICIP), pp. 36\u201340 (2019). IEEE","DOI":"10.1109\/ICIP.2019.8802965"},{"issue":"10","key":"1352_CR29","doi-asserted-by":"publisher","first-page":"3476","DOI":"10.1109\/TPAMI.2020.2985708","volume":"43","author":"J Gao","year":"2020","unstructured":"Gao, J., Zhang, T., Xu, C.: Learning to model relationships for zero-shot video classification. IEEE Trans. Pattern Anal. Mach. Intell. 43(10), 3476\u20133491 (2020)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"3","key":"1352_CR30","doi-asserted-by":"publisher","first-page":"1646","DOI":"10.1109\/TCSVT.2021.3075470","volume":"32","author":"J Gao","year":"2021","unstructured":"Gao, J., Xu, C.: Learning video moment retrieval without a single annotated video. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1646\u20131657 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"1352_CR31","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1016\/j.image.2019.02.005","volume":"74","author":"YS Gan","year":"2019","unstructured":"Gan, Y.S., Liong, S.-T., Yau, W.-C., Huang, Y.-C., Tan, L.-K.: Off-apexnet on micro-expression recognition system. Signal Process. Image Commun. 74, 129\u2013139 (2019)","journal-title":"Signal Process. Image Commun."},{"key":"1352_CR32","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.neucom.2020.10.082","volume":"427","author":"X Nie","year":"2021","unstructured":"Nie, X., Takalkar, M.A., Duan, M., Zhang, H., Xu, M.: Geme: dual-stream multi-task gender-based micro-expression recognition. Neurocomputing 427, 13\u201328 (2021)","journal-title":"Neurocomputing"},{"issue":"3","key":"1352_CR33","doi-asserted-by":"publisher","first-page":"460","DOI":"10.3390\/e25030460","volume":"25","author":"H Zhou","year":"2023","unstructured":"Zhou, H., Huang, S., Li, J., Wang, S.-J.: Dual-atme: dual-branch attention network for micro-expression recognition. Entropy 25(3), 460 (2023)","journal-title":"Entropy"},{"key":"1352_CR34","doi-asserted-by":"crossref","unstructured":"Hu, Y., Gao, J., Dong, J., Fan, B., Liu, H.: Exploring rich semantics for open-set action recognition. IEEE Trans. Multimedia (2023)","DOI":"10.1109\/TMM.2023.3333206"},{"key":"1352_CR35","doi-asserted-by":"crossref","unstructured":"Gao, J., Chen, M., Xu, C.: Vectorized evidential learning for weakly-supervised temporal action localization. IEEE Trans. Pattern Anal. Mach. Intell. (2023)","DOI":"10.1109\/CVPR52729.2023.01416"},{"key":"1352_CR36","doi-asserted-by":"crossref","unstructured":"Van\u00a0Quang, N., Chun, J., Tokuyama, T.: Capsulenet for micro-expression recognition. In: 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pp. 1\u20137 (2019). IEEE","DOI":"10.1109\/FG.2019.8756544"},{"issue":"1","key":"1352_CR37","doi-asserted-by":"publisher","first-page":"86041","DOI":"10.1371\/journal.pone.0086041","volume":"9","author":"W-J Yan","year":"2014","unstructured":"Yan, W.-J., Li, X., Wang, S.-J., Zhao, G., Liu, Y.-J., Chen, Y.-H., Fu, X.: Casme ii: an improved spontaneous micro-expression database and the baseline evaluation. PLoS One 9(1), 86041 (2014)","journal-title":"PLoS One"},{"issue":"01","key":"1352_CR38","doi-asserted-by":"publisher","first-page":"116","DOI":"10.1109\/TAFFC.2016.2573832","volume":"9","author":"AK Davison","year":"2018","unstructured":"Davison, A.K., Lansley, C., Costen, N., Tan, K., Yap, M.H.: Samm: a spontaneous micro-facial movement dataset. IEEE Trans. Affect. Comput. 9(01), 116\u2013129 (2018)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1352_CR39","doi-asserted-by":"crossref","unstructured":"Li, X., Pfister, T., Huang, X., Zhao, G., Pietik\u00e4inen, M.: A spontaneous micro-expression database: Inducement, collection and baseline. In: 2013 10th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition (fg), pp. 1\u20136 (2013). IEEE","DOI":"10.1109\/FG.2013.6553717"},{"key":"1352_CR40","doi-asserted-by":"crossref","unstructured":"Liong, S.-T., See, J., Wong, K., Le\u00a0Ngo, A.C., Oh, Y.-H., Phan, R.: Automatic apex frame spotting in micro-expression database. In: 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), pp. 665\u2013669 (2015). IEEE","DOI":"10.1109\/ACPR.2015.7486586"},{"key":"1352_CR41","doi-asserted-by":"crossref","unstructured":"Hou, Q., Zhou, D., Feng, J.: Coordinate attention for efficient mobile network design. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13713\u201313722 (2021)","DOI":"10.1109\/CVPR46437.2021.01350"},{"key":"1352_CR42","doi-asserted-by":"crossref","unstructured":"See, J., Yap, M.H., Li, J., Hong, X., Wang, S.-J.: Megc 2019\u2014the second facial micro-expressions grand challenge. In: 2019 14th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2019), pp. 1\u20135 (2019). IEEE","DOI":"10.1109\/FG.2019.8756611"},{"key":"1352_CR43","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. In: International Conference on Learning Representations (2018)"},{"issue":"05","key":"1352_CR44","first-page":"1075","volume":"38","author":"T Hong","year":"2022","unstructured":"Hong, T., Longjiao, F.S.Z., Hongmei, L.: Micro-expression recognition based on optical flow method and pseudo three-dimensional residual network. J. Signal Process. 38(05), 1075\u20131087 (2022)","journal-title":"J. Signal Process."},{"issue":"3","key":"1352_CR45","doi-asserted-by":"publisher","first-page":"1593","DOI":"10.1007\/s00530-023-01068-z","volume":"29","author":"X Shu","year":"2023","unstructured":"Shu, X., Li, J., Shi, L., Huang, S.: Res-capsnet: an improved capsule network for micro-expression recognition. Multimedia Syst. 29(3), 1593\u20131601 (2023)","journal-title":"Multimedia Syst."}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01352-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01352-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01352-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,7,5]],"date-time":"2024-07-05T17:19:23Z","timestamp":1720199963000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01352-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,16]]},"references-count":45,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,6]]}},"alternative-id":["1352"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01352-6","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,5,16]]},"assertion":[{"value":"3 April 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 May 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 May 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no Conflict of interest to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"147"}}