{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T00:05:22Z","timestamp":1770336322889,"version":"3.49.0"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T00:00:00Z","timestamp":1768003200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T00:00:00Z","timestamp":1768003200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100017691","name":"Guangxi Key Research and Development Program","doi-asserted-by":"publisher","award":["No. Gui Ke AB24010160"],"award-info":[{"award-number":["No. Gui Ke AB24010160"]}],"id":[{"id":"10.13039\/501100017691","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Innovation Project of GUET Graduate Education","award":["No. 2024YCXB10"],"award-info":[{"award-number":["No. 2024YCXB10"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["No. 62362015"],"award-info":[{"award-number":["No. 62362015"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Open Project Program of Guangxi Key Laboratory of Digital Infrastructure","award":["No. GXDIOP2024013"],"award-info":[{"award-number":["No. GXDIOP2024013"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["World Wide Web"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s11280-025-01401-8","type":"journal-article","created":{"date-parts":[[2026,1,10]],"date-time":"2026-01-10T13:38:49Z","timestamp":1768052329000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Confidence-guided dynamic sequential fusion for multimodal sentiment analysis"],"prefix":"10.1007","volume":"29","author":[{"given":"Rui","family":"Xia","sequence":"first","affiliation":[]},{"given":"Yuming","family":"Lin","sequence":"additional","affiliation":[]},{"given":"Guanyu","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Lihua","family":"He","sequence":"additional","affiliation":[]},{"given":"You","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,10]]},"reference":[{"key":"1401_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107134","volume":"226","author":"M Birjali","year":"2021","unstructured":"Birjali, M., Kasri, M., Beni-Hssane, A.: A comprehensive survey on sentiment analysis: Approaches, challenges and trends. Knowl.-Based Syst. 226, 107134 (2021)","journal-title":"Knowl.-Based Syst."},{"issue":"7","key":"1401_CR2","doi-asserted-by":"publisher","first-page":"5731","DOI":"10.1007\/s10462-022-10144-1","volume":"55","author":"M Wankhade","year":"2022","unstructured":"Wankhade, M., Rao, A.C.S., Kulkarni, C.: A survey on sentiment analysis methods, applications, and challenges. Artif. Intell. Rev. 55(7), 5731\u20135780 (2022)","journal-title":"Artif. Intell. Rev."},{"key":"1401_CR3","doi-asserted-by":"publisher","first-page":"306","DOI":"10.1016\/j.inffus.2023.02.028","volume":"95","author":"L Zhu","year":"2023","unstructured":"Zhu, L., Zhu, Z., Zhang, C., Xu, Y., Kong, X.: Multimodal sentiment analysis based on fusion methods: A survey. Information Fusion, vol. 95, 306\u2013325 (2023)","journal-title":"Inform. Fusion"},{"key":"1401_CR4","doi-asserted-by":"crossref","unstructured":"Zadeh, A., Chen, M., Poria, S., Cambria, E., Morency, L.-P.: Tensor fusion network for multimodal sentiment analysis. arXiv:1707.07250 (2017)","DOI":"10.18653\/v1\/D17-1115"},{"key":"1401_CR5","doi-asserted-by":"crossref","unstructured":"Liu, Z., Shen, Y., Lakshminarasimhan, V.B., Liang, P.P., Zadeh, A., Morency, L.-P.: Efficient low-rank multimodal fusion with modality-specific factors. arXiv:1806.00064 (2018)","DOI":"10.18653\/v1\/P18-1209"},{"key":"1401_CR6","doi-asserted-by":"crossref","unstructured":"Koromilas, P., Nicolaou, M.A., Giannakopoulos, T., Panagakis, Y.: Mmatr: A lightweight approach for multimodal sentiment analysis based on tensor methods. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1\u20135 (2023). IEEE","DOI":"10.1109\/ICASSP49357.2023.10097030"},{"key":"1401_CR7","doi-asserted-by":"crossref","unstructured":"Tsai, Y.-H.H., Bai, S., Liang, P.P., Kolter, J.Z., Morency, L.-P., Salakhutdinov, R.: Multimodal transformer for unaligned multimodal language sequences. In: Proceedings of the Conference. Association for Computational Linguistics. Meeting, vol. 2019, p. 6558 (2019). NIH Public Access","DOI":"10.18653\/v1\/P19-1656"},{"issue":"4","key":"1401_CR8","doi-asserted-by":"publisher","first-page":"3149","DOI":"10.1109\/TAFFC.2023.3265653","volume":"14","author":"H Cheng","year":"2023","unstructured":"Cheng, H., Yang, Z., Zhang, X., Yang, Y.: Multimodal sentiment analysis based on attentional temporal convolutional network and multi-layer feature fusion. IEEE Trans. Affect. Comput. 14(4), 3149\u20133163 (2023)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1401_CR9","doi-asserted-by":"crossref","unstructured":"Yu, Y., Zhao, M., Qi, S.-A., Sun, F., Wang, B., Guo, W., Wang, X., Yang, L., Niu, D.: Conki: Contrastive knowledge injection for multimodal sentiment analysis. In: Findings of the Association for Computational Linguistics: ACL 2023, pp. 13610\u201313624 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.860"},{"key":"1401_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.111346","volume":"285","author":"J Huang","year":"2024","unstructured":"Huang, J., Zhou, J., Tang, Z., Lin, J., Chen, C.Y.-C.: Tmbl: Transformer-based multimodal binding learning model for multimodal sentiment analysis. Knowl.-Based Syst. 285, 111346 (2024)","journal-title":"Knowl.-Based Syst."},{"key":"1401_CR11","doi-asserted-by":"crossref","unstructured":"Lin, R., Hu, H.: Multi-task momentum distillation for multimodal sentiment analysis. IEEE Trans Affect Comput. 15(2), 549\u2013565 (2023)","DOI":"10.1109\/TAFFC.2023.3282410"},{"key":"1401_CR12","doi-asserted-by":"crossref","unstructured":"Han, W., Chen, H., Gelbukh, A., Zadeh, A., Morency, L.-P., Poria, S.: Bi-bimodal modality fusion for correlation-controlled multimodal sentiment analysis. In: Proceedings of the 2021 International Conference on Multimodal Interaction, pp. 6\u201315 (2021)","DOI":"10.1145\/3462244.3479919"},{"key":"1401_CR13","unstructured":"Li, Z., Zhou, Y., Zhang, W., Liu, Y., Yang, C., Lian, Z., Hu, S.: Amoa: Global acoustic feature enhanced modal-order-aware network for multimodal sentiment analysis. In: Proceedings of the 29th International Conference on Computational Linguistics, pp. 7136\u20137146 (2022)"},{"key":"1401_CR14","doi-asserted-by":"crossref","unstructured":"Zhang, H., Wang, Y., Yin, G., Liu, K., Liu, Y., Yu, T.: Learning language-guided adaptive hyper-modality representation for multimodal sentiment analysis. In: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pp. 756\u2013767 (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.49"},{"key":"1401_CR15","doi-asserted-by":"crossref","unstructured":"Wang, P., Zhou, Q., Wu, Y., Chen, T., Hu, J.: Dlf: Disentangled-language-focused multimodal sentiment analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 39, pp. 21180\u201321188 (2025)","DOI":"10.1609\/aaai.v39i20.35416"},{"key":"1401_CR16","doi-asserted-by":"crossref","unstructured":"Wu, S., He, D., Wang, X., Wang, L., Dang, J.: Enriching multimodal sentiment analysis through textual emotional descriptions of visual-audio content. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 39, pp. 1601\u20131609 (2025)","DOI":"10.1609\/aaai.v39i2.32152"},{"key":"1401_CR17","doi-asserted-by":"publisher","first-page":"101891","DOI":"10.1016\/j.inffus.2023.101891","volume":"99","author":"Z Li","year":"2023","unstructured":"Li, Z., Guo, Q., Pan, Y., Ding, W., Yu, J., Zhang, Y., Liu, W., Chen, H., Wang, H., Xie, Y.: Multi-level correlation mining framework with self-supervised label generation for multimodal sentiment analysis. Information Fusion, vol. 99, 101891 (2023)","journal-title":"Inform. Fusion"},{"key":"1401_CR18","doi-asserted-by":"crossref","unstructured":"Liu, S., Luo, Z., Fu, W.: Fcdnet: fuzzy cognition-based dynamic fusion network for multimodal sentiment analysis. IEEE Trans. Fuzzy Syst. 33(1), 3\u201314 (2024)","DOI":"10.1109\/TFUZZ.2024.3407739"},{"key":"1401_CR19","doi-asserted-by":"crossref","unstructured":"Feng, X., Lin, Y., He, L., Li, Y., Chang, L., Zhou, Y.: Knowledge-guided dynamic modality attention fusion framework for multimodal sentiment analysis. In: Findings of the Association for Computational Linguistics: EMNLP 2024, pp. 14755\u201314766 (2024)","DOI":"10.18653\/v1\/2024.findings-emnlp.865"},{"key":"1401_CR20","doi-asserted-by":"crossref","unstructured":"Mao, H., Zhang, B., Xu, H., Yuan, Z., Liu, Y.: Robust-msa: Understanding the impact of modality noise on multimodal sentiment analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, pp. 16458\u201316460 (2023)","DOI":"10.1609\/aaai.v37i13.27078"},{"key":"1401_CR21","doi-asserted-by":"crossref","unstructured":"Neumann, M., Vu, N.T.: Investigations on audiovisual emotion recognition in noisy conditions. In: 2021 IEEE Spoken Language Technology Workshop (SLT), pp. 358\u2013364 (2021). IEEE","DOI":"10.1109\/SLT48900.2021.9383588"},{"key":"1401_CR22","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2023.111206","volume":"152","author":"A Pandey","year":"2024","unstructured":"Pandey, A., Vishwakarma, D.K.: Progress, achievements, and challenges in multimodal sentiment analysis using deep learning: A survey. Appl. Soft Comput. 152, 111206 (2024)","journal-title":"Appl. Soft Comput."},{"key":"1401_CR23","doi-asserted-by":"crossref","unstructured":"Tomar, M., Tiwari, A., Saha, T., Saha, S.: Your tone speaks louder than your face! modality order infused multi-modal sarcasm detection. In: Proceedings of the 31st ACM International Conference on Multimedia, pp. 3926\u20133933 (2023)","DOI":"10.1145\/3581783.3612528"},{"key":"1401_CR24","doi-asserted-by":"crossref","unstructured":"Hazarika, D., Zimmermann, R., Poria, S.: Misa: Modality-invariant and-specific representations for multimodal sentiment analysis. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 1122\u20131131 (2020)","DOI":"10.1145\/3394171.3413678"},{"key":"1401_CR25","doi-asserted-by":"crossref","unstructured":"Sun, H., Wang, H., Liu, J., Chen, Y.-W., Lin, L.: Cubemlp: An mlp-based model for multimodal sentiment analysis and depression estimation. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 3722\u20133729 (2022)","DOI":"10.1145\/3503161.3548025"},{"issue":"11","key":"1401_CR26","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1145\/219717.219748","volume":"38","author":"GA Miller","year":"1995","unstructured":"Miller, G.A.: Wordnet: A lexical database for English. Commun. ACM\u00a038(11), 39\u201341 (1995)","journal-title":"Commun. ACM"},{"issue":"10","key":"1401_CR27","doi-asserted-by":"publisher","first-page":"10098","DOI":"10.1109\/TKDE.2023.3250499","volume":"35","author":"Q Zhong","year":"2023","unstructured":"Zhong, Q., Ding, L., Liu, J., Du, B., Jin, H., Tao, D.: Knowledge graph augmented network towards multiview representation learning for aspect-based sentiment analysis. IEEE Trans. Knowl. Data Eng. 35(10), 10098\u201310111 (2023)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"issue":"12","key":"1401_CR28","doi-asserted-by":"publisher","first-page":"2724","DOI":"10.1109\/TKDE.2017.2754499","volume":"29","author":"Q Wang","year":"2017","unstructured":"Wang, Q., Mao, Z., Wang, B., Guo, L.: Knowledge graph embedding: A survey of approaches and applications. IEEE Trans. Knowl. Data Eng. 29(12), 2724\u20132743 (2017)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"1401_CR29","first-page":"2420","volume":"33","author":"VC Dinh","year":"2020","unstructured":"Dinh, V.C., Ho, L.S.: Consistent feature selection for analytic deep neural networks. Adv. Neural. Inf. Process. Syst. 33, 2420\u20132431 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"issue":"177","key":"1401_CR30","first-page":"1","volume":"20","author":"A Fisher","year":"2019","unstructured":"Fisher, A., Rudin, C., Dominici, F.: All models are wrong, but many are useful: Learning a variable\u2019s importance by studying an entire class of prediction models simultaneously. J. Mach. Learn. Res. 20(177), 1\u201381 (2019)","journal-title":"J. Mach. Learn. Res."},{"key":"1401_CR31","doi-asserted-by":"crossref","unstructured":"Wang, C., Zhang, Q., Dong, J., Fang, H., Schaefer, G., Liu, R., Yi, P.: A sequential mixing fusion network for enhanced feature representations in multimodal sentiment analysis. Knowl.-Based Syst. 113638 (2025)","DOI":"10.1016\/j.knosys.2025.113638"},{"key":"1401_CR32","doi-asserted-by":"crossref","unstructured":"Hu, H., Guo, W., Liu, Y., Kan, M.-Y.: Adaptive multi-modalities fusion in sequential recommendation systems. In: Proceedings of the 32nd ACM International Conference on Information and Knowledge Management, pp. 843\u2013853 (2023)","DOI":"10.1145\/3583780.3614775"},{"issue":"3","key":"1401_CR33","doi-asserted-by":"publisher","first-page":"1358","DOI":"10.1109\/TAFFC.2023.3340924","volume":"15","author":"Q Zhu","year":"2023","unstructured":"Zhu, Q., Zheng, C., Zhang, Z., Shao, W., Zhang, D.: Dynamic confidence-aware multi-modal emotion recognition. IEEE Trans. Affect. Comput. 15(3), 1358\u20131370 (2023)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"1401_CR34","doi-asserted-by":"crossref","unstructured":"Jing, Y., Zhao, X.: Dq-former: Querying transformer with dynamic modality priority for cognitive-aligned multimodal emotion recognition in conversation. In: Proceedings of the 32nd ACM International Conference on Multimedia, pp. 4795\u20134804 (2024)","DOI":"10.1145\/3664647.3681599"},{"key":"1401_CR35","unstructured":"Cao, B., Xia, Y., Ding, Y., Zhang, C., Hu, Q.: Predictive dynamic fusion. In: Proceedings of the 41st International Conference on Machine Learning, pp. 5608\u20135628 (2024)"},{"key":"1401_CR36","unstructured":"Zadeh, A., Zellers, R., Pincus, E., Morency, L.-P.: Mosi: multimodal corpus of sentiment intensity and subjectivity analysis in online opinion videos. arXiv:1606.06259 (2016)"},{"key":"1401_CR37","doi-asserted-by":"crossref","unstructured":"Zadeh, A.B., Liang, P.P., Poria, S., Cambria, E., Morency, L.-P.: Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pp. 2236\u20132246 (2018)","DOI":"10.18653\/v1\/P18-1208"},{"key":"1401_CR38","doi-asserted-by":"crossref","unstructured":"Yu, W., Xu, H., Yuan, Z., Wu, J.: Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 35, pp. 10790\u201310797 (2021)","DOI":"10.1609\/aaai.v35i12.17289"},{"key":"1401_CR39","doi-asserted-by":"crossref","unstructured":"Han, W., Chen, H., Poria, S.: Improving multimodal fusion with hierarchical mutual information maximization for multimodal sentiment analysis. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp. 9180\u20139192 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.723"},{"key":"1401_CR40","doi-asserted-by":"crossref","unstructured":"Mao, H., Yuan, Z., Xu, H., Yu, W., Liu, Y., Gao, K.: M-sena: An integrated platform for multimodal sentiment analysis. In: Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics: System Demonstrations, pp. 204\u2013213 (2022)","DOI":"10.18653\/v1\/2022.acl-demo.20"},{"key":"1401_CR41","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.109259","volume":"136","author":"D Wang","year":"2023","unstructured":"Wang, D., Guo, X., Tian, Y., Liu, J., He, L., Luo, X.: Tetfn: A text enhanced transformer fusion network for multimodal sentiment analysis. Pattern Recogn. 136, 109259 (2023)","journal-title":"Pattern Recogn."}],"container-title":["World Wide Web"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11280-025-01401-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11280-025-01401-8","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11280-025-01401-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,5]],"date-time":"2026-02-05T11:21:19Z","timestamp":1770290479000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11280-025-01401-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,10]]},"references-count":41,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["1401"],"URL":"https:\/\/doi.org\/10.1007\/s11280-025-01401-8","relation":{},"ISSN":["1386-145X","1573-1413"],"issn-type":[{"value":"1386-145X","type":"print"},{"value":"1573-1413","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,10]]},"assertion":[{"value":"24 July 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 November 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 December 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"10 January 2026","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}}],"article-number":"12"}}