{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T09:02:36Z","timestamp":1763629356435,"version":"3.45.0"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"17","license":[{"start":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T00:00:00Z","timestamp":1763596800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T00:00:00Z","timestamp":1763596800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Supercomput"],"DOI":"10.1007\/s11227-025-08071-3","type":"journal-article","created":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T08:58:49Z","timestamp":1763629129000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["PMAAN: a priority-guided multi-scale adaptive attention network for multimodal sentiment analysis"],"prefix":"10.1007","volume":"81","author":[{"given":"Fei","family":"Xu","sequence":"first","affiliation":[]},{"given":"Shuo","family":"An","sequence":"additional","affiliation":[]},{"given":"Daipeng","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Xintong","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,20]]},"reference":[{"key":"8071_CR1","doi-asserted-by":"publisher","first-page":"306","DOI":"10.1016\/j.inffus.2023.02.028","volume":"95","author":"L Zhu","year":"2023","unstructured":"Zhu L, Zhu Z, Zhang C, Xu Y, Kong X (2023) Multimodal sentiment analysis based on fusion methods: a survey. Inf Fusion 95:306\u2013325. https:\/\/doi.org\/10.1016\/j.inffus.2023.02.028","journal-title":"Inf Fusion"},{"issue":"2","key":"8071_CR2","doi-asserted-by":"publisher","first-page":"423","DOI":"10.1109\/TPAMI.2018.2798607","volume":"41","author":"T Baltru\u0161aitis","year":"2018","unstructured":"Baltru\u0161aitis T, Ahuja C, Morency L-P (2018) Multimodal machine learning: a survey and taxonomy. IEEE Trans Pattern Anal Mach Intell 41(2):423\u2013443","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8071_CR3","doi-asserted-by":"publisher","unstructured":"Wang K, Yin Q, Wang W, Wu S, Wang L (2016) A comprehensive survey on cross-modal retrieval. CoRR abs\/1607.06215. https:\/\/doi.org\/10.48550\/ARXIV.1607.06215","DOI":"10.48550\/ARXIV.1607.06215"},{"issue":"12","key":"8071_CR4","doi-asserted-by":"publisher","first-page":"4289","DOI":"10.1109\/JBHI.2021.3076762","volume":"25","author":"R Wang","year":"2021","unstructured":"Wang R, Hao Y, Yu Q, Chen M, Humar I, Fortino G (2021) Depression analysis and recognition based on functional near-infrared spectroscopy. IEEE J Biomed Health Inform 25(12):4289\u20134299","journal-title":"IEEE J Biomed Health Inform"},{"key":"8071_CR5","doi-asserted-by":"publisher","first-page":"1785","DOI":"10.1109\/TMM.2020.3003648","volume":"23","author":"W Guo","year":"2020","unstructured":"Guo W, Zhang Y, Cai X, Meng L, Yang J, Yuan X (2020) LD-MAN: layout-driven multimodal attention network for online news sentiment recognition. IEEE Trans Multimed 23:1785\u20131798","journal-title":"IEEE Trans Multimed"},{"key":"8071_CR6","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101958","volume":"100","author":"C Zhu","year":"2023","unstructured":"Zhu C, Chen M, Zhang S, Sun C, Liang H, Liu Y, Chen J (2023) SKEAFN: sentiment knowledge enhanced attention fusion network for multimodal sentiment analysis. Inf Fusion 100:101958","journal-title":"Inf Fusion"},{"key":"8071_CR7","doi-asserted-by":"crossref","unstructured":"Yang Y, Dong X, Qiang Y (2024) Clgsi: a multimodal sentiment analysis framework based on contrastive learning guided by sentiment intensity. In: Findings of the association for computational linguistics: NAACL 2024, pp 2099\u20132110","DOI":"10.18653\/v1\/2024.findings-naacl.135"},{"key":"8071_CR8","doi-asserted-by":"publisher","unstructured":"Yu W, Xu H, Yuan Z, Wu J (2021) Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 10790\u201310797. https:\/\/doi.org\/10.1609\/aaai.v35i12.17289","DOI":"10.1609\/aaai.v35i12.17289"},{"key":"8071_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2023.122454","volume":"239","author":"A Ghorbanali","year":"2024","unstructured":"Ghorbanali A, Sohrabi M (2024) Capsule network-based deep ensemble transfer learning for multimodal sentiment analysis. Expert Syst Appl 239:122454","journal-title":"Expert Syst Appl"},{"key":"8071_CR10","doi-asserted-by":"publisher","unstructured":"Sun Z, Sarma P, Sethares W, Liang Y (2020) Learning relationships between text, audio, and video via deep canonical correlation for multimodal language analysis. In: Proceedings of the AAAI Conference on Artificial Intelligence, pp 8992\u20138999. https:\/\/doi.org\/10.1609\/aaai.v34i05.6431","DOI":"10.1609\/aaai.v34i05.6431"},{"key":"8071_CR11","unstructured":"Tsai Y-H, Liang PP, Zadeh A et al (201) Learning factorized multimodal representations. In: 7th International Conference on Learning Representations, ICLR 2019. arXiv:1806.06176"},{"key":"8071_CR12","doi-asserted-by":"crossref","unstructured":"Hazarika D, Zimmermann R, Poria S (2020) Misa: modality-invariant and -specific representations for multimodal sentiment analysis. In: Proceedings of the 28th ACM International Conference on Multimedia, pp 1122\u20131131","DOI":"10.1145\/3394171.3413678"},{"key":"8071_CR13","doi-asserted-by":"crossref","unstructured":"Han W, Chen H, Poria S (2021) Improving multimodal fusion with hierarchical mutual information maximization for multimodal sentiment analysis. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp 9180\u20139192","DOI":"10.18653\/v1\/2021.emnlp-main.723"},{"key":"8071_CR14","doi-asserted-by":"crossref","unstructured":"Ayodele OO, Harun NH, Yusoff N (2023) Fuzzy layered convolution neutral network for feature level fusion based on multimodal sentiment classification. Emerg Adv Integr Technol 3(2)","DOI":"10.30880\/emait.2022.03.02.007"},{"key":"8071_CR15","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2023.110494","volume":"144","author":"A Aslam","year":"2023","unstructured":"Aslam A, Sargano AB, Habib Z (2023) Attention-based multimodal sentiment analysis and emotion recognition using deep neural networks. Appl Soft Comput 144:110494","journal-title":"Appl Soft Comput"},{"key":"8071_CR16","doi-asserted-by":"publisher","first-page":"2178","DOI":"10.1109\/LSP.2022.3216500","volume":"29","author":"D Wei","year":"2022","unstructured":"Wei D, Liu Y, Zhu X, Liu J, Zeng X (2022) Msaf: multimodal supervise-attention enhanced fusion for video anomaly detection. IEEE Signal Process Lett 29:2178\u20132182","journal-title":"IEEE Signal Process Lett"},{"key":"8071_CR17","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2025.107222","volume":"185","author":"J Hou","year":"2025","unstructured":"Hou J, Omar N, Tiun S, Saad S, He Q (2025) TF-BERT: tensor-based fusion BERT for multimodal sentiment analysis. Neural Netw 185:107222. https:\/\/doi.org\/10.1016\/j.neunet.2025.107222","journal-title":"Neural Netw"},{"key":"8071_CR18","doi-asserted-by":"publisher","DOI":"10.1088\/1361-6579\/ad5bbc","volume":"45","author":"X Wan","year":"2024","unstructured":"Wan X, Wang Y, Wang Z, Tang Y, Liu B (2024) Joint low-rank tensor fusion and cross-modal attention for multimodal physiological signals based emotion recognition. Physiol Meas 45:075003","journal-title":"Physiol Meas"},{"key":"8071_CR19","doi-asserted-by":"publisher","unstructured":"Wang Q, Xiang X, Zhao J (2022) ML-TFN: multi layers tensor fusion network for affective video content analysis. In: Neural Computing for Advanced Applications: 3th International Conference, NCAA 2022, Jinan, China, July 8\u201310, 2022, Proceedings, part I. Springer, pp 184\u2013196. https:\/\/doi.org\/10.1007\/978-981-19-6142-7_14","DOI":"10.1007\/978-981-19-6142-7_14"},{"key":"8071_CR20","doi-asserted-by":"publisher","unstructured":"Tsai Y-HH, Bai S, Liang PP, Kolter JZ, Morency L-P, Salakhutdinov R (2019) Multimodal transformer for unaligned multimodal language sequences. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, ACL 2019, pp 6558\u20136569. https:\/\/doi.org\/10.18653\/v1\/P19-1656","DOI":"10.18653\/v1\/P19-1656"},{"key":"8071_CR21","doi-asserted-by":"publisher","unstructured":"Waligora P, Aslam H, Zeeshan MO, Belharbi S, Koerich AL, Pedersoli M, Bacon S, Granger E (2024) Joint multimodal transformer for emotion recognition in the wild. In: 2024 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024. IEEE, pp 4625\u20134635. https:\/\/doi.org\/10.1109\/CVPRW63382.2024.00465","DOI":"10.1109\/CVPRW63382.2024.00465"},{"key":"8071_CR22","doi-asserted-by":"publisher","unstructured":"Yaghoubi E, Tran TK, Borza D, Frintrop S (2024) Attention-based fusion of intra- and intermodal dynamics in multimodal sentiment analysis. In: Proceedings of the 2024 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops), pp 273\u2013278. https:\/\/doi.org\/10.1109\/PerComWorkshops59983.2024.10502594","DOI":"10.1109\/PerComWorkshops59983.2024.10502594"},{"issue":"3","key":"8071_CR23","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2024.103675","volume":"61","author":"L Wang","year":"2024","unstructured":"Wang L, Peng J, Zheng C, Zhao T, Zhu L (2024) A cross modal hierarchical fusion multimodal sentiment analysis method based on multi-task learning. Inf Process Manag 61(3):103675","journal-title":"Inf Process Manag"},{"key":"8071_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2025.110262","volume":"146","author":"H Wang","year":"2025","unstructured":"Wang H, Du Q, Xiang Y (2025) Image\u2013text sentiment analysis based on hierarchical interaction fusion and contrast learning enhanced. Eng Appl Artif Intell 146:110262","journal-title":"Eng Appl Artif Intell"},{"issue":"6","key":"8071_CR25","doi-asserted-by":"publisher","first-page":"2230","DOI":"10.1093\/comjnl\/bxae002","volume":"67","author":"Y Lei","year":"2024","unstructured":"Lei Y, Qu K, Zhao Y, Han Q, Wang X (2024) Multimodal sentiment analysis based on composite hierarchical fusion. Comput J 67(6):2230\u20132245. https:\/\/doi.org\/10.1093\/comjnl\/bxae002","journal-title":"Comput J"},{"key":"8071_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2024.112220","volume":"300","author":"J Hou","year":"2024","unstructured":"Hou J, Omar N, Tiun S, Saad S, He Q (2024) TCHFN: multimodal sentiment analysis based on text-centric hierarchical fusion network. Knowl Based Syst 300:112220","journal-title":"Knowl Based Syst"},{"issue":"3","key":"8071_CR27","doi-asserted-by":"publisher","first-page":"945","DOI":"10.1007\/s10844-025-00923-x","volume":"63","author":"Y Li","year":"2025","unstructured":"Li Y, Liu A, Lu Y (2025) Multi-level language interaction transformer for multimodal sentiment analysis. J Intell Inf Syst 63(3):945\u2013964. https:\/\/doi.org\/10.1007\/s10844-025-00923-x","journal-title":"J Intell Inf Syst"},{"issue":"18","key":"8071_CR28","doi-asserted-by":"publisher","first-page":"R762","DOI":"10.1016\/j.cub.2005.08.058","volume":"15","author":"NP Holmes","year":"2005","unstructured":"Holmes NP, Spence C (2005) Multisensory integration: space, time and superadditivity. Curr Biol 15(18):R762\u2013R764. https:\/\/doi.org\/10.1016\/j.cub.2005.08.058","journal-title":"Curr Biol"},{"issue":"2","key":"8071_CR29","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1109\/TPAMI.2019.2938758","volume":"43","author":"SH Gao","year":"2021","unstructured":"Gao SH, Cheng MM, Zhao K, Zhang XY, Yang MH, Torr P (2021) Res2Net: a new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell 43(2):652\u2013662","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"6","key":"8071_CR30","doi-asserted-by":"publisher","first-page":"82","DOI":"10.1109\/MIS.2016.94","volume":"31","author":"A Zadeh","year":"2016","unstructured":"Zadeh A, Zellers R, Pincus E, Morency L-P (2016) Multimodal sentiment intensity analysis in videos: facial gestures and verbal messages. IEEE Intell Syst 31(6):82\u201388","journal-title":"IEEE Intell Syst"},{"key":"8071_CR31","doi-asserted-by":"publisher","unstructured":"Zadeh AB, Liang PP, Poria S, et al (2018) Multimodal language analysis in the wild: Cmu-mosei dataset and interpretable dynamic fusion graph. In: Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics, Melbourne, Australia, pp 2236\u20132246. https:\/\/doi.org\/10.18653\/v1\/P18-1208","DOI":"10.18653\/v1\/P18-1208"},{"key":"8071_CR32","doi-asserted-by":"publisher","first-page":"424","DOI":"10.1016\/j.inffus.2022.09.025","volume":"91","author":"A Gandhi","year":"2023","unstructured":"Gandhi A, Adhvaryu K, Poria S, Cambria E, Hussain A (2023) Multimodal sentiment analysis: a systematic review of history, datasets, multimodal fusion methods, applications, challenges and future directions. Inf Fus 91:424\u2013444. https:\/\/doi.org\/10.1016\/j.inffus.2022.09.025","journal-title":"Inf Fus"},{"key":"8071_CR33","unstructured":"Fang Z, He A, Yu Q, Gao B, Ding W, Zhang T, Ma L (2022) FAF: a novel multimodal emotion recognition approach integrating face, body and text. CoRR abs\/2211.15425"},{"issue":"2","key":"8071_CR34","doi-asserted-by":"publisher","first-page":"277","DOI":"10.1007\/s11042-009-0344-2","volume":"49","author":"M Mansoorizadeh","year":"2010","unstructured":"Mansoorizadeh M, Moghaddam Charkari N (2010) Multimodal information fusion application to human emotion recognition from face and speech. Multimed Tools Appl 49(2):277\u2013297","journal-title":"Multimed Tools Appl"},{"issue":"9","key":"8071_CR35","doi-asserted-by":"publisher","DOI":"10.1145\/3652149","volume":"56","author":"U Singh","year":"2024","unstructured":"Singh U, Abhishek K, Azad HK (2024) A survey of cutting-edge multimodal sentiment analysis. ACM Comput Surv 56(9):227:1\u2013227:38. https:\/\/doi.org\/10.1145\/3652149","journal-title":"ACM Comput Surv"},{"key":"8071_CR36","doi-asserted-by":"publisher","unstructured":"Shutova E, Kiela D, Maillard J (2016) Black holes and white rabbits: metaphor identification with visual features. In: Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (NAACL-HLT). ACL, pp 160\u2013170. https:\/\/doi.org\/10.18653\/v1\/N16-1020","DOI":"10.18653\/v1\/N16-1020"},{"key":"8071_CR37","doi-asserted-by":"crossref","unstructured":"Vora A, Paunwala CN, Paunwala M (2014) Improved weight assignment approach for multimodal fusion. In: 2014 IEEE International Conference on Circuits, Systems, Communication and Information Technology Applications (CSCITA). IEEE, pp 70\u201374","DOI":"10.1109\/CSCITA.2014.6839237"},{"key":"8071_CR38","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1007\/s10489-024-06113-6","volume":"55","author":"S Wang","year":"2025","unstructured":"Wang S, Ratnavelu K, Shibghatullah AS (2025) UEFN: efficient uncertainty estimation fusion network for reliable multimodal sentiment analysis. Appl Intell 55:171","journal-title":"Appl Intell"},{"key":"8071_CR39","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101973","volume":"101","author":"Z Liu","year":"2024","unstructured":"Liu Z, Zhou B, Chu D, Sun Y, Meng L (2024) Modality translation-based multimodal sentiment analysis under uncertain missing modalities. Inf Fus 101:101973","journal-title":"Inf Fus"},{"key":"8071_CR40","doi-asserted-by":"publisher","first-page":"5335","DOI":"10.1038\/s41598-024-54872-6","volume":"14","author":"J Du","year":"2024","unstructured":"Du J, Jin J, Zhuang J, Zhang C (2024) Hierarchical graph contrastive learning of local and global presentation for multimodal sentiment analysis. Sci Rep 14:5335. https:\/\/doi.org\/10.1038\/s41598-024-54872-6","journal-title":"Sci Rep"},{"key":"8071_CR41","doi-asserted-by":"publisher","first-page":"63291","DOI":"10.1007\/s11042-023-18032-8","volume":"83","author":"X Miao","year":"2024","unstructured":"Miao X, Zhang X, Zhang H (2024) Low-rank tensor fusion and self-supervised multi-task multimodal sentiment analysis. Multimed Tools Appl 83:63291\u201363308","journal-title":"Multimed Tools Appl"},{"key":"8071_CR42","doi-asserted-by":"crossref","unstructured":"Sun L, Zhao C, Li X, Bai C, Pan J (2024) Multimodal sentiment analysis via low-rank tensor attention network with unimodal self-supervised learning. In: 2024 International Joint Conference on Neural Networks (IJCNN). IEEE, pp 1\u20138","DOI":"10.1109\/IJCNN60899.2024.10651112"},{"key":"8071_CR43","doi-asserted-by":"publisher","first-page":"296","DOI":"10.1016\/j.inffus.2022.07.006","volume":"88","author":"F Zhang","year":"2022","unstructured":"Zhang F, Li X-C, Lim CP, Hua Q, Dong C-R, Zhai J-H (2022) Deep emotional arousal network for multimodal sentiment analysis and emotion recognition. Inf Fus 88:296\u2013304","journal-title":"Inf Fus"},{"key":"8071_CR44","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2023.111346","volume":"285","author":"J Huang","year":"2024","unstructured":"Huang J, Zhou J, Tang Z, Lin J, Chen CYC (2024) Tmbl: transformer-based multimodal binding learning model for multimodal sentiment analysis. Knowl Based Syst 285:111346","journal-title":"Knowl Based Syst"},{"key":"8071_CR45","doi-asserted-by":"publisher","unstructured":"Zadeh A, Chen M, Poria S, Cambria E, Morency L-P (2017) Tensor fusion network for multimodal sentiment analysis. In: Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing (EMNLP). ACL, pp 447\u2013457. https:\/\/doi.org\/10.18653\/v1\/D17-1115","DOI":"10.18653\/v1\/D17-1115"},{"key":"8071_CR46","unstructured":"Vaswani A, Shazeer N, Parmar N, et al (2017) Attention is all you need. In: Advances in neural information processing systems 30. Curran Associates."}],"container-title":["The Journal of Supercomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-08071-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11227-025-08071-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-08071-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,20]],"date-time":"2025-11-20T08:58:55Z","timestamp":1763629135000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11227-025-08071-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,20]]},"references-count":46,"journal-issue":{"issue":"17","published-online":{"date-parts":[[2025,11]]}},"alternative-id":["8071"],"URL":"https:\/\/doi.org\/10.1007\/s11227-025-08071-3","relation":{},"ISSN":["1573-0484"],"issn-type":[{"value":"1573-0484","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,20]]},"assertion":[{"value":"4 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 November 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 November 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"1583"}}