{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T20:23:48Z","timestamp":1773260628985,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":32,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819666010","type":"print"},{"value":"9789819665990","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-981-96-6599-0_16","type":"book-chapter","created":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T22:12:30Z","timestamp":1751407950000},"page":"228-243","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Enhanced Multimodal Aspect-Based Sentiment Analysis by\u00a0LLM-Generated Rationales"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0009-0006-0344-520X","authenticated-orcid":false,"given":"Jun","family":"Cao","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4997-3850","authenticated-orcid":false,"given":"Jiyi","family":"Li","sequence":"additional","affiliation":[]},{"given":"Ziwei","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Renjie","family":"Zhou","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,7,2]]},"reference":[{"key":"16_CR1","unstructured":"Anthropic: model card and evaluations for claude models (2023). https:\/\/www-files.anthropic.com\/production\/images\/Model-Card-Claude-2.pdf, Accessed 13 June 2024"},{"key":"16_CR2","doi-asserted-by":"crossref","unstructured":"Chen, G., Tian, Y., Song, Y.: Joint aspect extraction and sentiment analysis with directional graph convolutional networks. In: COLING, pp. 272\u2013279, December 2020","DOI":"10.18653\/v1\/2020.coling-main.24"},{"key":"16_CR3","doi-asserted-by":"crossref","unstructured":"Chen, H., Zhai, Z., Feng, F., Li, R., Wang, X.: Enhanced multi-channel graph convolutional network for aspect sentiment triplet extraction. In: ACL, pp. 2974\u20132985, May 2022","DOI":"10.18653\/v1\/2022.acl-long.212"},{"key":"16_CR4","unstructured":"Chen, T., Borth, D., Darrell, T., Chang, S.F.: Deepsentibank: visual sentiment concept classification with deep convolutional neural networks. arXiv preprint arXiv:1410.8586 (2014)"},{"key":"16_CR5","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"16_CR6","doi-asserted-by":"crossref","unstructured":"Hu, M., Peng, Y., Huang, Z., Li, D., Lv, Y.: Open-domain targeted sentiment analysis via span-based extraction and classification. In: ACL, pp. 537\u2013546. Florence, Italy, July 2019","DOI":"10.18653\/v1\/P19-1051"},{"key":"16_CR7","doi-asserted-by":"crossref","unstructured":"Ju, X., et al.: Joint multi-modal aspect-sentiment analysis with auxiliary cross-modal relation detection. In: EMNLP, pp. 4395\u20134405, November 2021","DOI":"10.18653\/v1\/2021.emnlp-main.360"},{"key":"16_CR8","doi-asserted-by":"crossref","unstructured":"Khan, Z., Fu, Y.: Exploiting bert for multimodal target sentiment classification through input space translation. In: Proceedings of the 29th ACM International Conference on Multimedia, MM 2021 p. 3034\u20133042 (2021)","DOI":"10.1145\/3474085.3475692"},{"key":"16_CR9","doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: BART: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: ACL, pp. 7871\u20137880, July 2020","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"16_CR10","doi-asserted-by":"publisher","first-page":"107643","DOI":"10.1016\/j.knosys.2021.107643","volume":"235","author":"B Liang","year":"2022","unstructured":"Liang, B., Su, H., Gui, L., Cambria, E., Xu, R.: Aspect-based sentiment analysis via affective knowledge enhanced graph convolutional networks. Knowl.-Based Syst. 235, 107643 (2022)","journal-title":"Knowl.-Based Syst."},{"key":"16_CR11","doi-asserted-by":"crossref","unstructured":"Ling, Y., Yu, J., Xia, R.: Vision-language pre-training for multimodal aspect-based sentiment analysis. In: ACL, pp. 2149\u20132159, May 2022","DOI":"10.18653\/v1\/2022.acl-long.152"},{"key":"16_CR12","unstructured":"Liu, Y., et al.: Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"key":"16_CR13","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1016\/j.neucom.2020.11.049","volume":"428","author":"Y Lv","year":"2021","unstructured":"Lv, Y., Wei, F., Cao, L., Peng, S., Niu, J., Yu, S., Wang, C.: Aspect-level sentiment analysis using context and aspect memory network. Neurocomputing 428, 195\u2013205 (2021)","journal-title":"Neurocomputing"},{"key":"16_CR14","doi-asserted-by":"crossref","unstructured":"Ma, Y., Cao, Y., Hong, Y., Sun, A.: Large language model is not a good few-shot information extractor, but a good reranker for hard samples! arXiv preprint arXiv:2303.08559 (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.710"},{"key":"16_CR15","unstructured":"Mokady, R., Hertz, A., Bermano, A.H.: Clipcap: clip prefix for image captioning. arXiv preprint arXiv:2111.09734 (2021)"},{"key":"16_CR16","unstructured":"OpenAI: Chatgpt: optimizing language models for dialogue (2022). https:\/\/openai.com\/blog\/chatgpt\/, Accessed 13 June 2024"},{"issue":"05","key":"16_CR17","doi-asserted-by":"publisher","first-page":"8600","DOI":"10.1609\/aaai.v34i05.6383","volume":"34","author":"H Peng","year":"2020","unstructured":"Peng, H., Xu, L., Bing, L., Huang, F., Lu, W., Si, L.: Knowing what, how and why: a near complete solution for aspect-based sentiment analysis. AAAI 34(05), 8600\u20138607 (2020)","journal-title":"AAAI"},{"key":"16_CR18","unstructured":"Reid, M., et\u00a0al.: Gemini 1.5: unlocking multimodal understanding across millions of tokens of context. arXiv preprint arXiv:2403.05530 (2024)"},{"key":"16_CR19","unstructured":"Rietzler, A., Stabinger, S., Opitz, P., Engl, S.: Adapt or get left behind: domain adaptation through BERT language model finetuning for aspect-target sentiment classification. In: LREC, pp. 4933\u20134941, May 2020"},{"issue":"15","key":"16_CR20","doi-asserted-by":"publisher","first-page":"13860","DOI":"10.1609\/aaai.v35i15.17633","volume":"35","author":"L Sun","year":"2021","unstructured":"Sun, L., Wang, J., Zhang, K., Su, Y., Weng, F.: Rpbert: a text-image relation propagation-based bert model for multimodal ner. AAAI 35(15), 13860\u201313868 (2021)","journal-title":"AAAI"},{"key":"16_CR21","doi-asserted-by":"crossref","unstructured":"Tian, Y., Chen, G., Song, Y.: Aspect-based sentiment analysis with type-aware graph convolutional networks and layer ensemble. In: NAACL-HLT, pp. 2910\u20132922, June 2021","DOI":"10.18653\/v1\/2021.naacl-main.231"},{"key":"16_CR22","doi-asserted-by":"crossref","unstructured":"Wang, K., Shen, W., Yang, Y., Quan, X., Wang, R.: Relational graph attention network for aspect-based sentiment analysis. In: ACL, pp. 3229\u20133238, July 2020","DOI":"10.18653\/v1\/2020.acl-main.295"},{"key":"16_CR23","unstructured":"Wei, J., et\u00a0al.: Emergent abilities of large language models. arXiv preprint arXiv:2206.07682 (2022)"},{"key":"16_CR24","doi-asserted-by":"crossref","unstructured":"Wu, H., Cheng, S., Wang, J., Li, S., Chi, L.: Multimodal aspect extraction with region-aware alignment network. In: NLPCC, pp. 145\u2013156. Springer (2020)","DOI":"10.1007\/978-3-030-60450-9_12"},{"key":"16_CR25","doi-asserted-by":"crossref","unstructured":"Wu, Z., Zheng, C., Cai, Y., Chen, J., Leung, H.f., Li, Q.: Multimodal representation with embedded visual guiding objects for named entity recognition in social media posts. In: Proceedings of the 28th ACM International Conference on Multimedia, MM 2020, pp. 1038\u20131046 (2020)","DOI":"10.1145\/3394171.3413650"},{"key":"16_CR26","unstructured":"Yan, H., Dai, J., Ji, T., Qiu, X., Zhang, Z.: A unified generative framework for aspect-based sentiment analysis. In: ACL-IJCNLP, pp. 2416\u20132429, August 2021"},{"issue":"5","key":"16_CR27","doi-asserted-by":"publisher","first-page":"103038","DOI":"10.1016\/j.ipm.2022.103038","volume":"59","author":"L Yang","year":"2022","unstructured":"Yang, L., Na, J.C., Yu, J.: Cross-modal multitask transformer for end-to-end multimodal aspect-based sentiment analysis. Inf. Process. Manage. 59(5), 103038 (2022)","journal-title":"Inf. Process. Manage."},{"key":"16_CR28","doi-asserted-by":"crossref","unstructured":"Yang, X., et al.: Few-shot joint multimodal aspect-sentiment analysis based on generative multimodal prompt. In: ACL, pp. 11575\u201311589, July 2023","DOI":"10.18653\/v1\/2023.findings-acl.735"},{"key":"16_CR29","doi-asserted-by":"crossref","unstructured":"Yu, J., Jiang, J.: Adapting bert for target-oriented multimodal sentiment classification. In: IJCAI, pp. 5408\u20135414, July 2019","DOI":"10.24963\/ijcai.2019\/751"},{"key":"16_CR30","doi-asserted-by":"publisher","first-page":"429","DOI":"10.1109\/TASLP.2019.2957872","volume":"28","author":"J Yu","year":"2020","unstructured":"Yu, J., Jiang, J., Xia, R.: Entity-sensitive attention and fusion network for entity-level multimodal sentiment classification. IEEE\/ACM Trans. Audio Speech Lang. Process. 28, 429\u2013439 (2020)","journal-title":"IEEE\/ACM Trans. Audio Speech Lang. Process."},{"key":"16_CR31","doi-asserted-by":"crossref","unstructured":"Yu, J., Jiang, J., Yang, L., Xia, R.: Improving multimodal named entity recognition via entity span detection with unified multimodal transformer. In: ACL, pp. 3342\u20133352, July 2020","DOI":"10.18653\/v1\/2020.acl-main.306"},{"key":"16_CR32","doi-asserted-by":"crossref","unstructured":"Zhou, R., Guo, W., Liu, X., Yu, S., Zhang, Y., Yuan, X.: AoM: detecting aspect-oriented information for multimodal aspect-based sentiment analysis. In: ACL, pp. 8184\u20138196, July 2023","DOI":"10.18653\/v1\/2023.findings-acl.519"}],"container-title":["Lecture Notes in Computer Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-96-6599-0_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,7,1]],"date-time":"2025-07-01T22:12:35Z","timestamp":1751407955000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-96-6599-0_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9789819666010","9789819665990"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-981-96-6599-0_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"2 July 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Auckland","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/iconip2024.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}