{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T15:33:12Z","timestamp":1769182392908,"version":"3.49.0"},"publisher-location":"Singapore","reference-count":34,"publisher":"Springer Nature Singapore","isbn-type":[{"value":"9789819555666","type":"print"},{"value":"9789819555673","type":"electronic"}],"license":[{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,1]],"date-time":"2026-01-01T00:00:00Z","timestamp":1767225600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2026]]},"DOI":"10.1007\/978-981-95-5567-3_8","type":"book-chapter","created":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:08Z","timestamp":1769116388000},"page":"106-120","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["CAHN: Category-Aware Hypergraph Network for\u00a0Multimodal Aspect-Based Sentiment Analysis"],"prefix":"10.1007","author":[{"given":"Jiayi","family":"Wu","sequence":"first","affiliation":[]},{"given":"Siyu","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Muchen","family":"Lan","sequence":"additional","affiliation":[]},{"given":"Yaoru","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,23]]},"reference":[{"key":"8_CR1","doi-asserted-by":"crossref","unstructured":"Ju, X., et al.: Joint multi-modal aspect-sentiment analysis with auxiliary cross-modal relation detection. In: Proceedings of EMNLP, pp. 4395\u20134405 (2021)","DOI":"10.18653\/v1\/2021.emnlp-main.360"},{"key":"8_CR2","doi-asserted-by":"crossref","unstructured":"Ling, Y., Yu, J., Xia, R.: Vision-language pre-training for multimodal aspect-based sentiment analysis. In: Proceedings of ACL, pp. 2149\u20132159 (2022)","DOI":"10.18653\/v1\/2022.acl-long.152"},{"key":"8_CR3","doi-asserted-by":"crossref","unstructured":"Zhao, T., ang Meng, L., Song, D.: Multimodal aspect-based sentiment analysis: a survey of tasks, methods, challenges and future directions. Inf. Fusion 112, 102552 (2024)","DOI":"10.1016\/j.inffus.2024.102552"},{"issue":"5","key":"8_CR4","doi-asserted-by":"publisher","first-page":"103038","DOI":"10.1016\/j.ipm.2022.103038","volume":"59","author":"L Yang","year":"2022","unstructured":"Yang, L., Na, J.C., Yu, J.: Cross-modal multitask transformer for end-to-end multimodal aspect-based sentiment analysis. Inf. Process. Manage. 59(5), 103038 (2022)","journal-title":"Inf. Process. Manage."},{"key":"8_CR5","doi-asserted-by":"crossref","unstructured":"Zhou, R., Guo, W., Liu, X., Yu, S., Zhang, Y., Yuan, X.: Aom: detecting aspect-oriented information for multimodal aspect-based sentiment analysis. In: Findings of ACL, pp. 8184\u20138196 (2023)","DOI":"10.18653\/v1\/2023.findings-acl.519"},{"key":"8_CR6","doi-asserted-by":"crossref","unstructured":"Liu, Y., et al.: Rng: reducing multi-level noise and multi-grained semantic gap for joint multimodal aspect-sentiment analysis. In: Proceedings of ICME, pp.\u00a01\u20136 (2024)","DOI":"10.1109\/ICME57554.2024.10687372"},{"key":"8_CR7","doi-asserted-by":"crossref","unstructured":"Yu, J., Jiang, J., Yang, L., Xia, R.: Improving multimodal named entity recognition via entity span detection with unified multimodal transformer. In: Proceedings of ACL, pp. 3342\u20133352 (2020)","DOI":"10.18653\/v1\/2020.acl-main.306"},{"key":"8_CR8","doi-asserted-by":"crossref","unstructured":"Wu, Z., Zheng, C., Cai, Y., Chen, J., Leung, H.F., Li, Q.: Multimodal representation with embedded visual guiding objects for named entity recognition in social media posts. In: Proceedings of ACM MM, pp. 1038\u20131046 (2020)","DOI":"10.1145\/3394171.3413650"},{"key":"8_CR9","doi-asserted-by":"crossref","unstructured":"Li, J., Li, H., Pan, Z., Sun, D., Wang, J., Zhang, W., Pan, G.: Prompting chatgpt in mner: enhanced multimodal named entity recognition with auxiliary refined knowledge. In: Findings of EMNLP (2023)","DOI":"10.18653\/v1\/2023.findings-emnlp.184"},{"key":"8_CR10","doi-asserted-by":"crossref","unstructured":"Khan, Z., Fu, Y.: Exploiting bert for multimodal target sentiment classification through input space translation. In: Proceedings of ACM MM, pp. 3034\u20133042 (2021)","DOI":"10.1145\/3474085.3475692"},{"key":"8_CR11","doi-asserted-by":"crossref","unstructured":"Yang, H., Zhao, Y., Qin, B.: Face-sensitive image-to-emotional-text cross-modal translation for multimodal aspect-based sentiment analysis. In: Proceedings of EMNLP, pp. 3324\u20133335 (2022)","DOI":"10.18653\/v1\/2022.emnlp-main.219"},{"key":"8_CR12","doi-asserted-by":"crossref","unstructured":"Huang, Y., Chen, Z., Chen, J., Pan, J.Z., Yao, Z., Zhang, W.: Target-oriented sentiment classification with sequential cross-modal semantic graph. In: Proceedings of CANN. pp. 587\u2013599 (2023)","DOI":"10.1007\/978-3-031-44216-2_48"},{"key":"8_CR13","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Proceedings of NeurIPS, vol. 30 (2017)"},{"key":"8_CR14","doi-asserted-by":"crossref","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of NAACL-HLT, pp. 4171\u20134186 (2019)","DOI":"10.18653\/v1\/N19-1423"},{"key":"8_CR15","doi-asserted-by":"crossref","unstructured":"Lewis, M., et al.: Bart: denoising sequence-to-sequence pre-training for natural language generation, translation, and comprehension. In: Proceedings of ACL, pp. 7871\u20137880 (2020)","DOI":"10.18653\/v1\/2020.acl-main.703"},{"key":"8_CR16","doi-asserted-by":"crossref","unstructured":"Zhao, F., Li, C., Wu, Z., Ouyang, Y., Zhang, J., Dai, X.: M2df: multi-grained multi-curriculum denoising framework for multimodal aspect-based sentiment analysis. In: Proceedings of EMNLP, pp. 9057\u20139070 (2023)","DOI":"10.18653\/v1\/2023.emnlp-main.561"},{"key":"8_CR17","doi-asserted-by":"crossref","unstructured":"Zhang, J., Qu, J., Liu, J., Wang, Z.: Mcpl: multi-model co-guided progressive learning for multimodal aspect-based sentiment analysis. Knowl.-Based Syst. 301, 112331 (2024)","DOI":"10.1016\/j.knosys.2024.112331"},{"key":"8_CR18","unstructured":"Feng, Y., You, H., Zhang, Z., Ji, R., Gao, Y.: Hypergraph neural networks. In: Proceedings of AAAI (2018)"},{"key":"8_CR19","unstructured":"Yadati, N., Nimishakavi, M., Yadav, P., Nitin, V., Louis, A., Talukdar, P.: Hypergcn: a new method for training graph convolutional networks on hypergraphs. In: Proceedings of NeurIPS, vol. 32 (2019)"},{"issue":"3","key":"8_CR20","doi-asserted-by":"publisher","first-page":"3181","DOI":"10.1109\/TPAMI.2022.3182052","volume":"45","author":"Y Gao","year":"2022","unstructured":"Gao, Y., Feng, Y., Ji, S., Ji, R.: Hgnn+: general hypergraph neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 3181\u20133199 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"1","key":"8_CR21","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1109\/TNNLS.2020.2978386","volume":"32","author":"Z Wu","year":"2020","unstructured":"Wu, Z., Pan, S., Chen, F., Long, G., Zhang, C., Yu, P.S.: A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 32(1), 4\u201324 (2020)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"4","key":"8_CR22","first-page":"1843","volume":"34","author":"D Yang","year":"2022","unstructured":"Yang, D., Qu, B., Yang, J., Cudr\u00e9-Mauroux, P.: Lbsn2vec++: heterogeneous hypergraph embedding for location-based social networks. IEEE Trans. Knowl. Data Eng. 34(4), 1843\u20131855 (2022)","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"8_CR23","doi-asserted-by":"crossref","unstructured":"Xia, X., Yin, H., Yu, J., Wang, Q., Cui, L., Zhang, X.: Self-supervised hypergraph convolutional networks for session-based recommendation. In: Proceedings of AAAI, vol.\u00a035, pp. 4503\u20134511 (2021)","DOI":"10.1609\/aaai.v35i5.16578"},{"key":"8_CR24","doi-asserted-by":"publisher","first-page":"126992","DOI":"10.1016\/j.neucom.2023.126992","volume":"565","author":"J Huang","year":"2024","unstructured":"Huang, J., et al.: Dynamic hypergraph convolutional network for multimodal sentiment analysis. Neurocomputing 565, 126992 (2024)","journal-title":"Neurocomputing"},{"key":"8_CR25","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"8_CR26","unstructured":"Jocher, G., Chaurasia, A., Qiu, J.: Ultralytics yolo (2023)"},{"key":"8_CR27","doi-asserted-by":"crossref","unstructured":"Cambria, E., Zhang, X., Mao, R., Chen, M., Kwok, K.: Senticnet 8: fusing emotion ai and commonsense ai for interpretable, trustworthy, and explainable affective computing. In: Proceedings of HCII, pp. 197\u2013216 (2024)","DOI":"10.1007\/978-3-031-76827-9_11"},{"key":"8_CR28","unstructured":"Dozat, T., Manning, C.D.: Deep biaffine attention for neural dependency parsing. In: Proceedings of ICLR (2017)"},{"key":"8_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Fu, J., Liu, X., Huang, X.: Adaptive co-attention network for named entity recognition in tweets. In: Proceedings of AAAI, vol.\u00a032 (2018)","DOI":"10.1609\/aaai.v32i1.11962"},{"key":"8_CR30","doi-asserted-by":"crossref","unstructured":"Lu, D., Neves, L., Carvalho, V., Zhang, N., Ji, H.: Visual attention model for name tagging in multimodal social media. In: Proceedings of ACL, pp. 1990\u20131999 (2018)","DOI":"10.18653\/v1\/P18-1185"},{"key":"8_CR31","doi-asserted-by":"crossref","unstructured":"Yu, J., Jiang, J.: Adapting bert for target-oriented multimodal sentiment classification. In: Proceedings of IJCAI, pp. 5408\u20135414 (2019)","DOI":"10.24963\/ijcai.2019\/751"},{"key":"8_CR32","doi-asserted-by":"crossref","unstructured":"Vempala, A., Preo\u0163iuc-Pietro, D.: Categorizing and inferring the relationship between the text and image of twitter posts. In: Proceedings of ACL, pp. 2830\u20132840 (2019)","DOI":"10.18653\/v1\/P19-1272"},{"issue":"15","key":"8_CR33","doi-asserted-by":"publisher","first-page":"13860","DOI":"10.1609\/aaai.v35i15.17633","volume":"35","author":"L Sun","year":"2021","unstructured":"Sun, L., Wang, J., Zhang, K., Su, Y., Weng, F.: Rpbert: a text-image relation propagation-based bert model for multimodal ner. Proc. AAAI 35(15), 13860\u201313868 (2021)","journal-title":"Proc. AAAI"},{"key":"8_CR34","doi-asserted-by":"publisher","first-page":"102304","DOI":"10.1016\/j.inffus.2024.102304","volume":"106","author":"L Xiao","year":"2024","unstructured":"Xiao, L., Wu, X., Xu, J., Li, W., Jin, C., He, L.: Atlantis: aesthetic-oriented multiple granularities fusion network for joint multimodal aspect-based sentiment analysis. Inf. Fusion 106, 102304 (2024)","journal-title":"Inf. Fusion"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-95-5567-3_8","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,22]],"date-time":"2026-01-22T21:13:15Z","timestamp":1769116395000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-95-5567-3_8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026]]},"ISBN":["9789819555666","9789819555673"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-981-95-5567-3_8","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026]]},"assertion":[{"value":"23 January 2026","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2025","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"15 October 2025","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2025","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2025","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/2025.prcv.cn\/index.asp","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}