{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:29:51Z","timestamp":1757618991148,"version":"3.44.0"},"reference-count":54,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T00:00:00Z","timestamp":1752451200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T00:00:00Z","timestamp":1752451200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Supercomput"],"DOI":"10.1007\/s11227-025-07613-z","type":"journal-article","created":{"date-parts":[[2025,7,14]],"date-time":"2025-07-14T15:43:15Z","timestamp":1752507795000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Dual attention-based graph convolutional neural network for multimodal sentiment analysis"],"prefix":"10.1007","volume":"81","author":[{"given":"Na","family":"Qu","sequence":"first","affiliation":[]},{"given":"Long","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Shengwei","family":"Tian","sequence":"additional","affiliation":[]},{"given":"Pusen","family":"Xia","sequence":"additional","affiliation":[]},{"given":"Chaoyue","family":"Wu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,14]]},"reference":[{"key":"7613_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.asoc.2020.106743","volume":"97","author":"A Khatua","year":"2020","unstructured":"Khatua A, Khatua A, Cambria E (2020) Predicting political sentiments of voters from twitter in multi-party contexts. Appl Soft Comput 97:106743","journal-title":"Appl Soft Comput"},{"issue":"5","key":"7613_CR2","first-page":"5105","volume":"35","author":"X Xue","year":"2022","unstructured":"Xue X, Zhang C, Niu Z, Wu X (2022) Multi-level attention map network for multimodal sentiment analysis. IEEE Trans Knowl Data Eng 35(5):5105\u20135118","journal-title":"IEEE Trans Knowl Data Eng"},{"issue":"1","key":"7613_CR3","doi-asserted-by":"publisher","first-page":"185","DOI":"10.1109\/TKDE.2017.2756658","volume":"30","author":"W Zhao","year":"2017","unstructured":"Zhao W, Guan Z, Chen L, He X, Cai D, Wang B, Wang Q (2017) Weakly-supervised deep embedding for product review sentiment analysis. IEEE Trans Knowl Data Eng 30(1):185\u2013197","journal-title":"IEEE Trans Knowl Data Eng"},{"issue":"10","key":"7613_CR4","doi-asserted-by":"publisher","first-page":"6729","DOI":"10.1109\/TPAMI.2021.3094362","volume":"44","author":"S Zhao","year":"2021","unstructured":"Zhao S, Yao X, Yang J, Jia G, Ding G, Chua T-S, Schuller BW, Keutzer K (2021) Affective image content analysis: two decades review and new perspectives. IEEE Trans Pattern Anal Mach Intell 44(10):6729\u20136751","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"7613_CR5","doi-asserted-by":"crossref","unstructured":"Kaur R, Kautish S (2022) Multimodal sentiment analysis: a survey and comparison. Research anthology on implementing sentiment analysis across multiple disciplines, 1846\u20131870","DOI":"10.4018\/978-1-6684-6303-1.ch098"},{"issue":"1","key":"7613_CR6","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1109\/TCSS.2022.3154442","volume":"10","author":"L Ansari","year":"2022","unstructured":"Ansari L, Ji S, Chen Q, Cambria E (2022) Ensemble hybrid learning methods for automated depression detection. IEEE Trans Comput Soc Syst 10(1):211\u2013219","journal-title":"IEEE Trans Comput Soc Syst"},{"issue":"10","key":"7613_CR7","doi-asserted-by":"publisher","first-page":"2733","DOI":"10.1109\/JBHI.2020.3001216","volume":"24","author":"H Jelodar","year":"2020","unstructured":"Jelodar H, Wang Y, Orji R, Huang S (2020) Deep sentiment classification and topic discovery on novel coronavirus or covid-19 online discussions: NLP using LSTM recurrent neural network approach. IEEE J Biomed Health Inform 24(10):2733\u20132742","journal-title":"IEEE J Biomed Health Inform"},{"key":"7613_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2024.102552","volume":"112","author":"T Zhao","year":"2024","unstructured":"Zhao T, Meng L-A, Song D (2024) Multimodal aspect-based sentiment analysis: a survey of tasks, methods, challenges and future directions. Inf Fusion 112:102552","journal-title":"Inf Fusion"},{"issue":"2","key":"7613_CR9","first-page":"264","volume":"57","author":"RM Zhao","year":"2020","unstructured":"Zhao RM, Xiong X, Ju SG, Li ZZ, Xie C (2020) Implicit sentiment analysis for Chinese texts based on a hybrid neural network. J Sichuan Univ (Nat Sci Edit) 57(2):264\u2013270","journal-title":"J Sichuan Univ (Nat Sci Edit)"},{"key":"7613_CR10","unstructured":"Song K (2018) Research on image sentiment analysis based on deep learning. PhD thesis, Ph. D., University of Science and Technology of China"},{"key":"7613_CR11","doi-asserted-by":"crossref","unstructured":"Pengfei X, Houpan Z, Weidong Z (2020) Pad 3-D speech emotion recognition based on feature fusion. In: Journal of Physics: Conference Series, vol 1616. IOP Publishing, p 012106","DOI":"10.1088\/1742-6596\/1616\/1\/012106"},{"key":"7613_CR12","doi-asserted-by":"crossref","unstructured":"Ju X, Zhang D, Xiao R, Li J, Li S, Zhang M, Zhou G (2021) Joint multi-modal aspect-sentiment analysis with auxiliary cross-modal relation detection. In: Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, pp 4395\u20134405","DOI":"10.18653\/v1\/2021.emnlp-main.360"},{"key":"7613_CR13","doi-asserted-by":"crossref","unstructured":"Sun L, Wang J, Su Y, Weng F, Sun Y, Zheng Z, Chen Y (2020) RIVA: a pre-trained tweet multimodal model based on text-image relation for multimodal NER. In: Proceedings of the 28th International Conference on Computational Linguistics, pp 1852\u20131862","DOI":"10.18653\/v1\/2020.coling-main.168"},{"key":"7613_CR14","doi-asserted-by":"publisher","unstructured":"Chen X, Zhang N, Li L, Yao Y, Deng S, Tan C, Huang F, Si L, Chen H (2022) Good visual guidance makes A better extractor: Hierarchical visual prefix for multimodal entity and relation extraction. CoRR abs\/2205.03521 https:\/\/doi.org\/10.48550\/ARXIV.2205.03521","DOI":"10.48550\/ARXIV.2205.03521"},{"key":"7613_CR15","doi-asserted-by":"crossref","unstructured":"Yang L, Yu J, Zhang C, Na J-C (2021) Fine-grained sentiment analysis of political tweets with entity-aware multimodal network. In: Diversity, Divergence, Dialogue: 16th International Conference, iConference 2021, Beijing, China, March 17\u201331, 2021, Proceedings, Part I 16, Springer, pp 411\u2013420","DOI":"10.1007\/978-3-030-71292-1_31"},{"key":"7613_CR16","doi-asserted-by":"crossref","unstructured":"Khan Z, Fu Y (2021) Exploiting BERT for multimodal target sentiment classification through input space translation. In: Proceedings of the 29th ACM International Conference on Multimedia, pp 3034\u20133042","DOI":"10.1145\/3474085.3475692"},{"key":"7613_CR17","doi-asserted-by":"crossref","unstructured":"Nojavanasghari B, Gopinath D, Koushik J, Baltru\u0161aitis T, Morency L-P (2016) Deep multimodal fusion for persuasiveness prediction. In: Proceedings of the 18th ACM International Conference on Multimodal Interaction, pp 284\u2013288","DOI":"10.1145\/2993148.2993176"},{"key":"7613_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101973","volume":"101","author":"Z Liu","year":"2024","unstructured":"Liu Z, Zhou B, Chu D, Sun Y, Meng L (2024) Modality translation-based multimodal sentiment analysis under uncertain missing modalities. Inf Fusion 101:101973","journal-title":"Inf Fusion"},{"issue":"5","key":"7613_CR19","doi-asserted-by":"publisher","first-page":"493","DOI":"10.1016\/j.cviu.2012.10.009","volume":"117","author":"N Liu","year":"2013","unstructured":"Liu N, Dellandr\u00e9a E, Chen L, Zhu C, Zhang Y, Bichot C-E, Bres S, Tellez B (2013) Multimodal recognition of visual concepts using histograms of textual concepts and selective weighted late fusion scheme. Comput Vis Image Underst 117(5):493\u2013512","journal-title":"Comput Vis Image Underst"},{"issue":"2","key":"7613_CR20","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.103193","volume":"60","author":"D Chen","year":"2023","unstructured":"Chen D, Su W, Wu P, Hua B (2023) Joint multimodal sentiment analysis based on information relevance. Inf Process Manag 60(2):103193","journal-title":"Inf Process Manag"},{"key":"7613_CR21","doi-asserted-by":"crossref","unstructured":"Poria S, Chaturvedi I, Cambria E, Hussain A (2016) Convolutional MKL based multimodal emotion recognition and sentiment analysis. In: 2016 IEEE 16th International Conference on Data Mining (ICDM), IEEE, pp 439\u2013448","DOI":"10.1109\/ICDM.2016.0055"},{"key":"7613_CR22","first-page":"371","volume":"33","author":"N Xu","year":"2019","unstructured":"Xu N, Mao W, Chen G (2019) Multi-interactive memory network for aspect based multimodal sentiment analysis. Proc AAAI Conf Artif Intell 33:371\u2013378","journal-title":"Proc AAAI Conf Artif Intell"},{"issue":"24","key":"7613_CR23","doi-asserted-by":"publisher","first-page":"5954","DOI":"10.1002\/cpe.5954","volume":"32","author":"M Cao","year":"2020","unstructured":"Cao M, Zhu Y, Gao W, Li M, Wang S (2020) Various syncretic co-attention network for multimodal sentiment analysis. Concur Comput: Pract Exp 32(24):5954","journal-title":"Concur Comput: Pract Exp"},{"key":"7613_CR24","doi-asserted-by":"crossref","unstructured":"Cai C, He Y, Sun L, Lian Z, Liu B, Tao J, Xu M, Wang K (2021) Multimodal sentiment analysis based on recurrent neural network and multimodal attention. In: Proceedings of the 2nd on Multimodal Sentiment Analysis Challenge, pp 61\u201367","DOI":"10.1145\/3475957.3484454"},{"issue":"5","key":"7613_CR25","doi-asserted-by":"publisher","DOI":"10.1016\/j.ipm.2022.103038","volume":"59","author":"L Yang","year":"2022","unstructured":"Yang L, Na J-C, Yu J (2022) Cross-modal multitask transformer for end-to-end multimodal aspect-based sentiment analysis. Inf Process Manag 59(5):103038","journal-title":"Inf Process Manag"},{"key":"7613_CR26","unstructured":"Xu K, Ba J, Kiros R, Cho K, Courville A, Salakhudinov R, Zemel R, Bengio Y (2015) Show, attend and tell: Neural image caption generation with visual attention. In: International Conference on Machine Learning, PMLR, pp 2048\u20132057"},{"key":"7613_CR27","doi-asserted-by":"crossref","unstructured":"Braytee A, Yang AS-C, Anaissi A, Chaturvedi K, Prasad M (2024) A novel dual-pipeline based attention mechanism for multimodal social sentiment analysis. In: Companion Proceedings of the ACM Web Conference 2024, pp 1816\u20131822","DOI":"10.1145\/3589335.3651967"},{"key":"7613_CR28","doi-asserted-by":"crossref","unstructured":"Lan Y-T, Liu W, Lu B-L (2020) Multimodal emotion recognition using deep generalized canonical correlation analysis with an attention mechanism. In: 2020 International Joint Conference on Neural Networks (IJCNN), IEEE, pp 1\u20136","DOI":"10.1109\/IJCNN48605.2020.9207625"},{"key":"7613_CR29","doi-asserted-by":"publisher","first-page":"4028","DOI":"10.1109\/TMM.2023.3321435","volume":"26","author":"D Wang","year":"2023","unstructured":"Wang D, Tian C, Liang X, Zhao L, He L, Wang Q (2023) Dual-perspective fusion network for aspect-based multimodal sentiment analysis. IEEE Trans Multimed 26:4028\u20134038","journal-title":"IEEE Trans Multimed"},{"key":"7613_CR30","doi-asserted-by":"publisher","first-page":"429","DOI":"10.1109\/TASLP.2019.2957872","volume":"28","author":"J Yu","year":"2019","unstructured":"Yu J, Jiang J, Xia R (2019) Entity-sensitive attention and fusion network for entity-level multimodal sentiment classification. IEEE\/ACM Trans Audio, Speech, Lang Process 28:429\u2013439","journal-title":"IEEE\/ACM Trans Audio, Speech, Lang Process"},{"issue":"9","key":"7613_CR31","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3477141","volume":"54","author":"S Abadal","year":"2021","unstructured":"Abadal S, Jain A, Guirado R, L\u00f3pez-Alonso J, Alarc\u00f3n E (2021) Computing graph neural networks: a survey from algorithms to accelerators. ACM Comput Surv (CSUR) 54(9):1\u201338","journal-title":"ACM Comput Surv (CSUR)"},{"key":"7613_CR32","doi-asserted-by":"publisher","DOI":"10.1016\/J.KNOSYS.2022.110080","volume":"259","author":"L Wang","year":"2023","unstructured":"Wang L, Song Z, Zhang X, Wang C, Zhang G, Zhu L, Li J, Liu H (2023) SAT-GCN: self-attention graph convolutional network-based 3D object detection for autonomous driving. Knowl Based Syst 259:110080. https:\/\/doi.org\/10.1016\/J.KNOSYS.2022.110080","journal-title":"Knowl Based Syst"},{"key":"7613_CR33","doi-asserted-by":"crossref","unstructured":"Marcheggiani D, Bastings J, Titov I (2018) Exploiting semantics in neural machine translation with graph convolutional networks. In: Walker MA, Ji H, Stent A (eds) Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT, New Orleans, Louisiana, USA, June 1\u20136, 2018, Vol 2 (Short Papers). pp. 486\u2013492. https:\/\/doi.org\/10.18653\/V1\/N18-2078","DOI":"10.18653\/v1\/N18-2078"},{"key":"7613_CR34","doi-asserted-by":"crossref","unstructured":"Wang J, Liu Z, Sheng V, Song Y, Qiu C (2021) Saliencybert: Recurrent attention network for target-oriented multimodal sentiment classification. In: Pattern Recognition and Computer Vision: 4th Chinese Conference, PRCV 2021, Beijing, China, October 29\u2013November 1, 2021, Proceedings, Part III 4, Springer, pp 3\u201315","DOI":"10.1007\/978-3-030-88010-1_1"},{"key":"7613_CR35","unstructured":"Zhao F, Wu Z, Long S, Dai X, Huang S, Chen J (2022) Learning from adjective-noun pairs: A knowledge-enhanced framework for target-oriented multimodal sentiment classification. In: Proceedings of the 29th International Conference on Computational Linguistics, pp 6784\u20136794"},{"issue":"12","key":"7613_CR36","first-page":"3683","volume":"40","author":"L Li","year":"2023","unstructured":"Li L, Li P (2023) Aspect-level multimodal sentiment analysis based on interaction graph neural network. Appl Res Comput 40(12):3683\u20133689","journal-title":"Appl Res Comput"},{"key":"7613_CR37","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2023.127222","volume":"573","author":"J Yang","year":"2024","unstructured":"Yang J, Xu M, Xiao Y, Du X (2024) AMIFN: aspect-guided multi-view interactions and fusion network for multimodal aspect-based sentiment analysis. Neurocomputing 573:127222","journal-title":"Neurocomputing"},{"key":"7613_CR38","doi-asserted-by":"crossref","unstructured":"Li R, Chen H, Feng F, Ma Z, Wang X, Hovy E (2021) Dual graph convolutional networks for aspect-based sentiment analysis. In: Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers), pp 6319\u20136329","DOI":"10.18653\/v1\/2021.acl-long.494"},{"key":"7613_CR39","doi-asserted-by":"publisher","unstructured":"Zhang Y, Jia A, Wang B, Zhang P, Zhao D, Li P, Hou Y, Jin X, Song D, Qin J (2024) M3GAT: A multi-modal, multi-task interactive graph attention network for conversational sentiment analysis and emotion recognition, vol 42. pp 1\u201332. https:\/\/doi.org\/10.1145\/3593583","DOI":"10.1145\/3593583"},{"key":"7613_CR40","doi-asserted-by":"publisher","unstructured":"You L, Peng J, Jin H, Claramunt C, Zeng H, Zhang Z (2024) DRGAT: dual-relational graph attention networks for aspect-based sentiment classification, vol 668. p 120531. https:\/\/doi.org\/10.1016\/J.INS.2024.120531","DOI":"10.1016\/J.INS.2024.120531"},{"key":"7613_CR41","doi-asserted-by":"publisher","first-page":"12039","DOI":"10.1109\/ACCESS.2024.3354844","volume":"12","author":"H Zhao","year":"2024","unstructured":"Zhao H, Yang M, Bai X, Liu H (2024) A survey on multimodal aspect-based sentiment analysis. IEEE Access 12:12039\u201312052","journal-title":"IEEE Access"},{"key":"7613_CR42","doi-asserted-by":"publisher","unstructured":"Lewis M, Liu Y, Goyal N, Ghazvininejad M, Mohamed A, Levy O, Stoyanov V, Zettlemoyer L (2020) BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Jurafsky D, Chai J, Schluter N, Tetreault JR (eds) Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, July 5-10, 2020, pp 7871\u20137880. https:\/\/doi.org\/10.18653\/V1\/2020.ACL-MAIN.703","DOI":"10.18653\/V1\/2020.ACL-MAIN.703"},{"key":"7613_CR43","unstructured":"Chen T, Borth D, Darrell T, Chang S (2014) Deepsentibank: Visual sentiment concept classification with deep convolutional neural networks. CoRR abs\/1410.8586"},{"key":"7613_CR44","doi-asserted-by":"publisher","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2016, Las Vegas, NV, USA, June 27-30, 2016, pp 770\u2013778. https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"7613_CR45","doi-asserted-by":"crossref","unstructured":"Ma Y, Peng H, Cambria E (2018) Targeted aspect-based sentiment analysis via embedding commonsense knowledge into an attentive LSTM. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32","DOI":"10.1609\/aaai.v32i1.12048"},{"key":"7613_CR46","doi-asserted-by":"publisher","unstructured":"Yu J, Jiang J (2019) Adapting BERT for target-oriented multimodal sentiment classification. In: Kraus S (ed) Proceedings of the 28th International Joint Conference on Artificial Intelligence, IJCAI 2019, Macao, China, August 10-16, 2019, pp 5408\u20135414. https:\/\/doi.org\/10.24963\/IJCAI.2019\/751","DOI":"10.24963\/IJCAI.2019\/751"},{"key":"7613_CR47","doi-asserted-by":"publisher","first-page":"1904","DOI":"10.7717\/peerj-cs.1904","volume":"10","author":"T Zhang","year":"2024","unstructured":"Zhang T, Zhou G, Lu J, Li Z, Wu H, Liu S (2024) Text-image semantic relevance identification for aspect-based multimodal sentiment analysis. PeerJ Comput Sci 10:1904","journal-title":"PeerJ Comput Sci"},{"key":"7613_CR48","unstructured":"Devlin J, Chang M-W, Lee K, Toutanova K (2019) Bert: Pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Vol 1 (long and Short Papers). pp 4171\u20134186"},{"key":"7613_CR49","doi-asserted-by":"crossref","unstructured":"Fan F, Feng Y, Zhao D (2018) Multi-grained attention network for aspect-level sentiment classification. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp 3433\u20133442","DOI":"10.18653\/v1\/D18-1380"},{"key":"7613_CR50","doi-asserted-by":"publisher","unstructured":"Sun C, Huang L, Qiu X (2019) Utilizing BERT for aspect-based sentiment analysis via constructing auxiliary sentence. In: Burstein J, Doran C, Solorio T (eds) Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL-HLT 2019, Minneapolis, MN, USA, June 2-7, 2019, Vol 1 (Long and Short Papers). pp 380\u2013385. https:\/\/doi.org\/10.18653\/V1\/N19-1035","DOI":"10.18653\/V1\/N19-1035"},{"key":"7613_CR51","doi-asserted-by":"publisher","first-page":"90586","DOI":"10.1109\/ACCESS.2024.3404261","volume":"12","author":"H Fan","year":"2024","unstructured":"Fan H, Chen J (2024) Position perceptive multi-hop fusion network for multimodal aspect-based sentiment analysis. IEEE Access 12:90586\u201390595. https:\/\/doi.org\/10.1109\/ACCESS.2024.3404261","journal-title":"IEEE Access"},{"key":"7613_CR52","doi-asserted-by":"crossref","unstructured":"Ling Y, Yu J, Xia R (2022) Vision-language pre-training for multimodal aspect-based sentiment analysis. In: Muresan S, Nakov P, Villavicencio A (eds) Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), ACL 2022, Dublin, Ireland, May 22-27, 2022, pp 2149\u20132159. https:\/\/doi.org\/10\/gtcnxz","DOI":"10.18653\/v1\/2022.acl-long.152"},{"key":"7613_CR53","doi-asserted-by":"crossref","unstructured":"Zhao F, Li C, Wu Z, Ouyang Y, Zhang J, Dai X (2023) M2df: Multi-grained multi-curriculum denoising framework for multimodal aspect-based sentiment analysis. In: Bouamor H, Pino J, Bali K (eds) Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, EMNLP 2023, Singapore, December 6-10, 2023, pp 9057\u20139070","DOI":"10.18653\/v1\/2023.emnlp-main.561"},{"key":"7613_CR54","doi-asserted-by":"crossref","unstructured":"Zhou R, Guo W, Liu X, Yu S, Zhang Y, Yuan X (2023) AOM: Detecting aspect-oriented information for multimodal aspect-based sentiment analysis. In: Rogers A, Boyd-Graber JL, Okazaki N (eds) Findings of the Association for Computational Linguistics: ACL 2023, Toronto, Canada, July 9-14, 2023, pp 8184\u20138196. https:\/\/doi.org\/10\/gs59fg","DOI":"10.18653\/v1\/2023.findings-acl.519"}],"container-title":["The Journal of Supercomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-07613-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11227-025-07613-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11227-025-07613-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,7]],"date-time":"2025-09-07T09:15:08Z","timestamp":1757236508000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11227-025-07613-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,7,14]]},"references-count":54,"journal-issue":{"issue":"10","published-online":{"date-parts":[[2025,7]]}},"alternative-id":["7613"],"URL":"https:\/\/doi.org\/10.1007\/s11227-025-07613-z","relation":{},"ISSN":["1573-0484"],"issn-type":[{"type":"electronic","value":"1573-0484"}],"subject":[],"published":{"date-parts":[[2025,7,14]]},"assertion":[{"value":"25 June 2025","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 July 2025","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"1154"}}