{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T17:14:21Z","timestamp":1772644461110,"version":"3.50.1"},"reference-count":47,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T00:00:00Z","timestamp":1768953600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"},{"start":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T00:00:00Z","timestamp":1772582400000},"content-version":"vor","delay-in-days":42,"URL":"https:\/\/creativecommons.org\/licenses\/by-nc-nd\/4.0"}],"funder":[{"name":"Guangdong Provincial Department of Education Youth Innovative Talent Project","award":["2025KQNCX149"],"award-info":[{"award-number":["2025KQNCX149"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J. King Saud Univ. Comput. Inf. Sci."],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s44443-026-00472-5","type":"journal-article","created":{"date-parts":[[2026,1,21]],"date-time":"2026-01-21T14:46:00Z","timestamp":1769006760000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["SemCap: Sentiment-aware semantic captioning for multimodal aspect-based sentiment analysis"],"prefix":"10.1007","volume":"38","author":[{"given":"Kexin","family":"Jiang","sequence":"first","affiliation":[]},{"given":"Xiaoqin","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Xiangxiang","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Qin","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,21]]},"reference":[{"key":"472_CR1","doi-asserted-by":"crossref","unstructured":"Bao X, Jiang X, Wang Z, Zhang Y, Zhou G (2023) Opinion tree parsing for aspect-based sentiment analysis. In: Findings of the association for computational linguistics: ACL 2023. pp 7971\u20137984. https:\/\/aclanthology.org\/2023.findings-acl.505\/","DOI":"10.18653\/v1\/2023.findings-acl.505"},{"key":"472_CR2","doi-asserted-by":"crossref","unstructured":"Cao J, Li J, Yang Z, Zhou R (2024) Enhanced multimodal aspect-based sentiment analysis by LLM-generated rationales. In: International conference on neural information processing. Springer, pp 228\u2013243","DOI":"10.1007\/978-981-96-6599-0_16"},{"issue":"2","key":"472_CR3","doi-asserted-by":"publisher","first-page":"1389","DOI":"10.1007\/s13042-024-02342-w","volume":"16","author":"Y Chen","year":"2025","unstructured":"Chen Y, Xiong H, Li K, Mai W, Xue Y, Cai Q, Li F (2025) Relevance-aware visual entity filter network for multimodal aspect-based sentiment analysis. Int J Mach Learn Cybern 16(2):1389\u20131402","journal-title":"Int J Mach Learn Cybern"},{"key":"472_CR4","doi-asserted-by":"crossref","unstructured":"Chen G, Tian Y, Song Y (2020) Joint aspect extraction and sentiment analysis with directional graph convolutional networks. In: Proceedings of the 28th international conference on computational linguistics. pp 272\u2013279","DOI":"10.18653\/v1\/2020.coling-main.24"},{"key":"472_CR5","doi-asserted-by":"crossref","unstructured":"He X (2025) CLAMP: contrastive learning with adaptive multi-loss and progressive fusion for multimodal aspect-based sentiment analysis. arXiv:2507.16854","DOI":"10.2139\/ssrn.5822513"},{"key":"472_CR6","unstructured":"He P, Gao J, Chen W (2022) DeBERTaV3: improving DeBERTa using ELECTRA-style pre-training with gradient-disentangled embedding sharing. In: International Conference on Learning Representations (ICLR)"},{"key":"472_CR7","doi-asserted-by":"publisher","unstructured":"Hua Y, Denny P, Wicker J, Taskova K (2024) A systematic review of aspect-based sentiment analysis: Domains, methods, and trends. Artif Intell Rev 57(11). https:\/\/doi.org\/10.1007\/s10462-024-10906-z","DOI":"10.1007\/s10462-024-10906-z"},{"key":"472_CR8","doi-asserted-by":"crossref","unstructured":"Hu Q, Jin G, Zhao Y, Cui R, Huang Z (2025) A knowledge-enhanced network for multimodal aspect-based sentiment classification. In: International conference on advanced data mining and applications. Springer, pp 377\u2013384","DOI":"10.1007\/978-981-95-3453-1_26"},{"key":"472_CR9","doi-asserted-by":"publisher","unstructured":"Hu M, Peng Y, Huang Z, Li D, Lv Y (2019) Open-domain targeted sentiment analysis via span-based extraction and classification. In: Proceedings of the 57th annual meeting of the association for computational linguistics. Association for Computational Linguistics, Florence, Italy, pp 537\u2013546. https:\/\/doi.org\/10.18653\/v1\/P19-1051, https:\/\/aclanthology.org\/P19-1051","DOI":"10.18653\/v1\/P19-1051"},{"key":"472_CR10","doi-asserted-by":"crossref","unstructured":"Ju X, Zhang D, Xiao R, Li J, Li S, Zhang M, Zhou G (2021) Joint multimodal aspect-sentiment analysis with auxiliary cross-modal relation detection. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing (EMNLP). pp 4395\u20134405","DOI":"10.18653\/v1\/2021.emnlp-main.360"},{"key":"472_CR11","doi-asserted-by":"publisher","first-page":"109311","DOI":"10.1016\/j.patcog.2023.109311","volume":"137","author":"T Lai","year":"2023","unstructured":"Lai T, Sadri A, Lin S, Li Z, Chen R, Wang H (2023) Efficient sampling using feature matching and variable minimal structure size. Pattern Recogn 137:109311","journal-title":"Pattern Recogn"},{"key":"472_CR12","unstructured":"Lawan A, Yunusa H (2025) GateMABSA: aspect-image gated fusion for multimodal aspect-based sentiment analysis. arXiv:2509.25037"},{"issue":"4","key":"472_CR13","doi-asserted-by":"publisher","first-page":"78","DOI":"10.1007\/s10462-023-10685-z","volume":"57","author":"Y Li","year":"2024","unstructured":"Li Y, Ding H, Lin Y, Feng X, Chang L (2024) Multi-level textual-visual alignment and fusion network for multimodal aspect-based sentiment analysis. Artif Intell Rev 57(4):78","journal-title":"Artif Intell Rev"},{"issue":"8","key":"472_CR14","doi-asserted-by":"publisher","first-page":"2907","DOI":"10.1007\/s11263-024-02023-9","volume":"132","author":"S Lin","year":"2024","unstructured":"Lin S, Huang F, Lai T, Lai J, Wang H, Weng J (2024) Robust heterogeneous model fitting for multi-source image correspondences. Int J Comput Vision 132(8):2907\u20132928","journal-title":"Int J Comput Vision"},{"key":"472_CR15","doi-asserted-by":"publisher","first-page":"3031","DOI":"10.1109\/TIP.2024.3391002","volume":"33","author":"S Lin","year":"2024","unstructured":"Lin S, Chen X, Xiao G, Wang H, Huang F, Weng J (2024) Multi-stage network with geometric semantic attention for two-view correspondence learning. IEEE Trans Image Process 33:3031\u20133046","journal-title":"IEEE Trans Image Process"},{"key":"472_CR16","doi-asserted-by":"crossref","unstructured":"Ling Y, Yu J, Xia R (2022) Vision-language pre-training for multimodal aspect-based sentiment analysis. In: Proceedings of the 60th annual meeting of the association for computational linguistics (vol 1: Long Papers). pp 2149\u20132159","DOI":"10.18653\/v1\/2022.acl-long.152"},{"key":"472_CR17","doi-asserted-by":"crossref","unstructured":"Lin S, Lo M, Chen H, Liang Y, Wu Q (2025) MGCA-Net: multi-graph contextual attention network for two-view correspondence learning. In: Proceedings of the thirty-fourth international joint conference on artificial intelligence. pp 1539\u20131547","DOI":"10.24963\/ijcai.2025\/172"},{"key":"472_CR18","doi-asserted-by":"publisher","unstructured":"Li Y, Su Z, Chen K, Jiang K (2024) Aspect-based sentiment analysis via knowledge enhancement. In: 2024 International Joint Conference on Neural Networks (IJCNN). IEEE, pp 1\u20138. https:\/\/doi.org\/10.1109\/IJCNN60899.2024.10650800","DOI":"10.1109\/IJCNN60899.2024.10650800"},{"issue":"2","key":"472_CR19","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1007\/s11063-025-11737-x","volume":"57","author":"Z Liu","year":"2025","unstructured":"Liu Z, Lin J, Chen Y, Dong Y (2025) Multimodal aspect-based sentiment analysis with external knowledge and multi-granularity image-text features. Neural Process Lett 57(2):25","journal-title":"Neural Process Lett"},{"key":"472_CR20","doi-asserted-by":"crossref","unstructured":"Liu H, He L, Liang, J (2024) Joint modal circular complementary attention for multimodal aspect-based sentiment analysis. In: 2024 IEEE International Conference on Multimedia and Expo Workshops (ICMEW). pp 1\u20136","DOI":"10.1109\/ICMEW63481.2024.10645483"},{"key":"472_CR21","unstructured":"Liu X, Li R, Ye S, Zhang G, Wang X (2025) Multimodal aspect-based sentiment analysis under conditional relation. In: Proceedings of the 31st international conference on computational linguistics (COLING). pp 313\u2013323"},{"issue":"12","key":"472_CR22","doi-asserted-by":"publisher","first-page":"8787","DOI":"10.1109\/TKDE.2023.3345022","volume":"36","author":"J Mu","year":"2024","unstructured":"Mu J, Nie F, Wang W, Xu J, Zhang J, Liu H (2024) MOCOLNet: a momentum contrastive learning network for multimodal aspect-level sentiment analysis. IEEE Trans Knowl Data Eng 36(12):8787\u20138800","journal-title":"IEEE Trans Knowl Data Eng"},{"key":"472_CR23","doi-asserted-by":"publisher","first-page":"111206","DOI":"10.1016\/j.asoc.2023.111206","volume":"152","author":"A Pandey","year":"2024","unstructured":"Pandey A, Vishwakarma DK (2024) Progress, achievements, and challenges in multimodal sentiment analysis using deep learning: a survey. Appl Soft Comput 152:111206","journal-title":"Appl Soft Comput"},{"key":"472_CR24","unstructured":"Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, Sastry G, Askell A, Mishkin P, Clark J, Krueger G, Sutskever I (2021) Learning transferable visual models from natural language supervision. In: International conference on machine learning (ICML)"},{"key":"472_CR25","doi-asserted-by":"publisher","first-page":"101587","DOI":"10.1016\/j.csl.2023.101587","volume":"85","author":"Y Wan","year":"2024","unstructured":"Wan Y, Chen Y, Lin J, Zhong J, Dong C (2024) A knowledge-augmented heterogeneous graph convolutional network for aspect-level multimodal sentiment analysis. Comput Speech Lang 85:101587","journal-title":"Comput Speech Lang"},{"issue":"10","key":"472_CR26","doi-asserted-by":"publisher","first-page":"827","DOI":"10.1007\/s11760-025-04387-y","volume":"19","author":"Y Wang","year":"2025","unstructured":"Wang Y, Zhao X, Zhong Z (2025) Dual-stream multi-level interaction network for aspect-based multimodal sentiment analysis. SIViP 19(10):827","journal-title":"SIViP"},{"key":"472_CR27","doi-asserted-by":"publisher","first-page":"903","DOI":"10.1007\/s41060-023-00497-3","volume":"20","author":"S Wang","year":"2025","unstructured":"Wang S, Cai G, Lv G (2025) Aspect-level multimodal sentiment analysis based on co-attention fusion. Int J Data Sci Analytics 20:903\u2013916","journal-title":"Int J Data Sci Analytics"},{"key":"472_CR28","doi-asserted-by":"crossref","unstructured":"Wu H, Cheng S, Wang J, Li S, Chi L (2020) Multimodal aspect extraction with region-aware alignment network. In: Natural language processing and Chinese computing. Springer, Cham, pp 145\u2013156","DOI":"10.1007\/978-3-030-60450-9_12"},{"key":"472_CR29","doi-asserted-by":"publisher","unstructured":"Wu Z, Zheng C, Cai Y, Chen J, Leung H-F, Li Q (2020) Multimodal representation with embedded visual guiding objects for named entity recognition in social media posts. In: Proceedings of the 28th ACM international conference on multimedia. Association for Computing Machinery, New York, NY, USA, pp 1038\u20131046. https:\/\/doi.org\/10.1145\/3394171.3413650","DOI":"10.1145\/3394171.3413650"},{"key":"472_CR30","doi-asserted-by":"publisher","first-page":"102304","DOI":"10.1016\/j.inffus.2024.102304","volume":"106","author":"L Xiao","year":"2024","unstructured":"Xiao L, Wu X, Xu J (2024) Atlantis: aesthetic-oriented multiple granularities fusion network for joint multimodal aspect-based sentiment analysis. Information Fusion 106:102304. https:\/\/doi.org\/10.1016\/j.inffus.2024.102304","journal-title":"Information Fusion"},{"key":"472_CR31","doi-asserted-by":"crossref","unstructured":"Xiao L, Mao R, Zhao S, Lin Q, Jia Y, He L, Cambria E (2025) Exploring cognitive and aesthetic causality for multimodal aspect-based sentiment analysis. IEEE Transactions on Affective Computing","DOI":"10.1109\/TAFFC.2025.3565506"},{"key":"472_CR32","doi-asserted-by":"crossref","unstructured":"Yan H, Dai J, Ji T, Qiu X, Zhang Z (2021) A unified generative framework for aspect-based sentiment analysis. In: Proceedings of the 59th annual meeting of the Association for Computational Linguistics (ACL). pp 2416\u20132429","DOI":"10.18653\/v1\/2021.acl-long.188"},{"key":"472_CR33","doi-asserted-by":"publisher","first-page":"127222","DOI":"10.1016\/j.neucom.2023.127222","volume":"573","author":"J Yang","year":"2024","unstructured":"Yang J, Xu M et al (2024) AMIFN: aspect-guided multi-view interactions and fusion network for multimodal aspect-based sentiment analysis. Neurocomputing 573:127222","journal-title":"Neurocomputing"},{"issue":"5","key":"472_CR34","doi-asserted-by":"publisher","first-page":"103038","DOI":"10.1016\/j.ipm.2022.103038","volume":"59","author":"L Yang","year":"2022","unstructured":"Yang L, Na J-C, Yu J (2022) Cross-modal multitask transformer for end-to-end multimodal aspect-based sentiment analysis. Inf Process Manag 59(5):103038. https:\/\/doi.org\/10.1016\/j.ipm.2022.103038","journal-title":"Inf Process Manag"},{"key":"472_CR35","doi-asserted-by":"publisher","first-page":"111724","DOI":"10.1016\/j.knosys.2024.111724","volume":"293","author":"J Yang","year":"2024","unstructured":"Yang J, Xiao Y, Du X (2024) Multi-grained fusion network with self-distillation for aspect-based multimodal sentiment analysis. Knowl-Based Syst 293:111724","journal-title":"Knowl-Based Syst"},{"key":"472_CR36","doi-asserted-by":"crossref","unstructured":"Yang X, Feng S, Wang D et al (2023) Few-shot joint multimodal aspect-sentiment analysis based on generative multimodal prompt. In: Findings of the association for computational linguistics: ACL 2023. pp 11575\u201311589","DOI":"10.18653\/v1\/2023.findings-acl.735"},{"key":"472_CR37","doi-asserted-by":"crossref","unstructured":"Yu J, Jiang J (2019) Adapting BERT for target-oriented multimodal sentiment classification. In: Proceedings of the 28th international joint conference on artificial intelligence, IJCAI 2019. International Joint Conferences on Artificial Intelligence. pp 5408\u20135414","DOI":"10.24963\/ijcai.2019\/751"},{"key":"472_CR38","doi-asserted-by":"publisher","unstructured":"Yu J, Jiang J, Yang L, Xia R (2020) Improving multimodal named entity recognition via entity span detection with unified multimodal transformer. In: Proceedings of the 58th annual meeting of the association for computational linguistics. Association for Computational Linguistics, pp 3342\u20133352. https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.306, https:\/\/aclanthology.org\/2020.acl-main.306\/","DOI":"10.18653\/v1\/2020.acl-main.306"},{"key":"472_CR39","doi-asserted-by":"publisher","first-page":"112331","DOI":"10.1016\/j.knosys.2024.112331","volume":"301","author":"J Zhang","year":"2024","unstructured":"Zhang J, Qu J, Liu J, Wang Z (2024) MCPL: Multi-model co-guided progressive learning for multimodal aspect-based sentiment analysis. Knowl-Based Syst 301:112331","journal-title":"Knowl-Based Syst"},{"key":"472_CR40","doi-asserted-by":"publisher","first-page":"102552","DOI":"10.1016\/j.inffus.2024.102552","volume":"112","author":"T Zhao","year":"2024","unstructured":"Zhao T, Meng L-A, Song D (2024) Multimodal aspect-based sentiment analysis: a survey of tasks, methods, challenges and future directions. Inf Fusion 112:102552","journal-title":"Inf Fusion"},{"key":"472_CR41","doi-asserted-by":"crossref","unstructured":"Zhao F, Li, C, Wu Z, Ouyang Y, Zhang J, Dai X (2023) M2DF: Multi-grained multi-curriculum denoising framework for multimodal aspect-based sentiment analysis. In: Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP). pp 9057\u20139070","DOI":"10.18653\/v1\/2023.emnlp-main.561"},{"key":"472_CR42","doi-asserted-by":"publisher","unstructured":"Zhou T, Shen Y, Chen K, Cao Q (2023) Hierarchical dual graph convolutional network for aspect-based sentiment analysis. Knowl-Based Syst 276:110740. https:\/\/doi.org\/10.1016\/j.knosys.2023.110740","DOI":"10.1016\/j.knosys.2023.110740"},{"key":"472_CR43","doi-asserted-by":"crossref","unstructured":"Zhou R, Guo W, Liu X, Yu S, Zhang Y, Yuan X (2023) AoM: detecting aspect-oriented information for multimodal aspect-based sentiment analysis. In: Findings of the association for computational linguistics: ACL 2023. pp 8184\u20138196","DOI":"10.18653\/v1\/2023.findings-acl.519"},{"key":"472_CR44","doi-asserted-by":"crossref","unstructured":"Zhou L, Pu Y, Zhao Z, Feng J, Xu D, Gu J (2025) Multimodal progressive contrastive learning for sentiment analysis. Neurocomputing, 132033","DOI":"10.1016\/j.neucom.2025.132033"},{"key":"472_CR45","doi-asserted-by":"crossref","unstructured":"Zhou J, Yang H, He Y, Mou H, Yang J (2023) A unified one-step solution for aspect sentiment quad prediction. In: Findings of the association for computational linguistics: ACL 2023. pp 12249\u201312265. https:\/\/aclanthology.org\/2023.findings-acl.777\/","DOI":"10.18653\/v1\/2023.findings-acl.777"},{"key":"472_CR46","doi-asserted-by":"crossref","unstructured":"Zhu L, Sun H, Gao Q, Liu Y, He L (2025) Aspect enhancement and text simplification in multimodal aspect-based sentiment analysis for multi-aspect and multi-sentiment scenarios. In: Proceedings of the AAAI conference on artificial intelligence, vol 39. pp 1683\u20131691","DOI":"10.1609\/aaai.v39i2.32161"},{"key":"472_CR47","doi-asserted-by":"publisher","first-page":"125818","DOI":"10.1016\/j.eswa.2024.125818","volume":"264","author":"W Zou","year":"2025","unstructured":"Zou W, Sun X, Wu W, Lu Q, Zhao X, Bo Q, Yan J (2025) TCMT: target-oriented cross modal transformer for multimodal aspect-based sentiment analysis. Expert Syst Appl 264:125818","journal-title":"Expert Syst Appl"}],"container-title":["Journal of King Saud University Computer and Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s44443-026-00472-5","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00472-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s44443-026-00472-5.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T13:20:15Z","timestamp":1772630415000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s44443-026-00472-5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,21]]},"references-count":47,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["472"],"URL":"https:\/\/doi.org\/10.1007\/s44443-026-00472-5","relation":{},"ISSN":["1319-1578","2213-1248"],"issn-type":[{"value":"1319-1578","type":"print"},{"value":"2213-1248","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,21]]},"assertion":[{"value":"28 November 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 January 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"This study does not involve any human participants, personal data, or experiments on animals. All datasets utilized in this research are publicly available and have been widely used in previous academic studies. Therefore, no ethical approval is required.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical Approval"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"81"}}