{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,29]],"date-time":"2025-11-29T08:03:48Z","timestamp":1764403428987,"version":"3.40.3"},"publisher-location":"Cham","reference-count":24,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031781278"},{"type":"electronic","value":"9783031781285"}],"license":[{"start":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T00:00:00Z","timestamp":1732924800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T00:00:00Z","timestamp":1732924800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-78128-5_29","type":"book-chapter","created":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T07:42:30Z","timestamp":1732952550000},"page":"451-466","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Multi-modal Deep Emotion-Cause Pair Extraction for\u00a0Video Corpus"],"prefix":"10.1007","author":[{"given":"Qianli","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Linlin","family":"Zong","sequence":"additional","affiliation":[]},{"given":"Bo","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Xianchao","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xinyue","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,30]]},"reference":[{"key":"29_CR1","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107965","volume":"238","author":"F Chen","year":"2022","unstructured":"Chen, F., Shi, Z., Yang, Z., Huang, Y.: Recurrent synchronization network for emotion-cause pair extraction. Knowl.-Based Syst. 238, 107965 (2022)","journal-title":"Knowl.-Based Syst."},{"doi-asserted-by":"crossref","unstructured":"Chen, Y., Hou, W., Cheng, X., Li, S.: Joint learning for emotion classification and emotion cause detection. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 646\u2013651 (2018)","key":"29_CR2","DOI":"10.18653\/v1\/D18-1066"},{"doi-asserted-by":"crossref","unstructured":"Chen, Y., Hou, W., Li, S., Wu, C., Zhang, X.: End-to-end emotion-cause pair extraction with graph convolutional network. In: Proceedings of the 28th International Conference on Computational Linguistics, pp. 198\u2013207 (2020)","key":"29_CR3","DOI":"10.18653\/v1\/2020.coling-main.17"},{"doi-asserted-by":"crossref","unstructured":"Chudasama, V., Kar, P., Gudmalwar, A., Shah, N., Wasnik, P., Onoe, N.: M2fnet: multi-modal fusion network for emotion recognition in conversation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4652\u20134661 (2022)","key":"29_CR4","DOI":"10.1109\/CVPRW56347.2022.00511"},{"doi-asserted-by":"publisher","unstructured":"Ding, Z., Xia, R., Yu, J.: ECPE-2D: emotion-cause pair extraction based on joint two-dimensional representation, interaction and prediction. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 3161\u20133170. Association for Computational Linguistics (2020). https:\/\/doi.org\/10.18653\/v1\/2020.acl-main.288","key":"29_CR5","DOI":"10.18653\/v1\/2020.acl-main.288"},{"doi-asserted-by":"crossref","unstructured":"Eyben, F., W\u00f6llmer, M., Schuller, B.: Opensmile: the munich versatile and fast open-source audio feature extractor. In: Proceedings of the 18th ACM International Conference on Multimedia, pp. 1459\u20131462 (2010)","key":"29_CR6","DOI":"10.1145\/1873951.1874246"},{"doi-asserted-by":"crossref","unstructured":"Fan, C., Yuan, C., Du, J., Gui, L., Yang, M., Xu, R.: Transition-based directed graph construction for emotion-cause pair extraction. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 3707\u20133717 (2020)","key":"29_CR7","DOI":"10.18653\/v1\/2020.acl-main.342"},{"key":"29_CR8","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-319-18032-8_1","volume-title":"Advances in Knowledge Discovery and Data Mining","author":"K Gao","year":"2015","unstructured":"Gao, K., Xu, H., Wang, J.: Emotion cause detection for Chinese micro-blogs based on ECOCC model. In: Cao, T., Lim, E.-P., Zhou, Z.-H., Ho, T.-B., Cheung, D., Motoda, H. (eds.) PAKDD 2015. LNCS (LNAI), vol. 9078, pp. 3\u201314. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-18032-8_1"},{"doi-asserted-by":"crossref","unstructured":"Gui, L., Xu, R., Wu, D., Lu, Q., Zhou, Y.: Event-driven emotion cause extraction with corpus construction. In: Social Media Content Analysis: Natural Language Processing and Beyond, pp. 145\u2013160. World Scientific (2018)","key":"29_CR9","DOI":"10.1142\/9789813223615_0011"},{"doi-asserted-by":"crossref","unstructured":"Hu, D., Hou, X., Wei, L., Jiang, L., Mo, Y.: Mm-dfn: multimodal dynamic fusion network for emotion recognition in conversations. In: ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 7037\u20137041. IEEE (2022)","key":"29_CR10","DOI":"10.1109\/ICASSP43922.2022.9747397"},{"doi-asserted-by":"crossref","unstructured":"Hu, G., Lin, T.E., Zhao, Y., Lu, G., Wu, Y., Li, Y.: UniMSE: towards unified multimodal sentiment analysis and emotion recognition. In: Goldberg, Y., Kozareva, Z., Zhang, Y. (eds.) Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing. pp. 7837\u20137851. Association for Computational Linguistics, Abu Dhabi, United Arab Emirates (Dec 2022)","key":"29_CR11","DOI":"10.18653\/v1\/2022.emnlp-main.534"},{"issue":"1","key":"29_CR12","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/TPAMI.2012.59","volume":"35","author":"S Ji","year":"2012","unstructured":"Ji, S., Xu, W., Yang, M., Yu, K.: 3d convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 35(1), 221\u2013231 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"unstructured":"Lee, S.Y.M., Chen, Y., Huang, C.R.: A text-driven rule-based system for emotion cause detection. In: Proceedings of the NAACL HLT 2010 Workshop on Computational Approaches to Analysis and Generation of Emotion in Text, pp. 45\u201353 (2010)","key":"29_CR13"},{"doi-asserted-by":"crossref","unstructured":"Li, J., Wang, X., Lv, G., Zeng, Z.: Ga2mif: graph and attention based two-stage multi-source information fusion for conversational emotion detection. IEEE Trans. Affective Comput. (2023)","key":"29_CR14","DOI":"10.1109\/TAFFC.2023.3261279"},{"doi-asserted-by":"crossref","unstructured":"Li, S., Yan, H., Qiu, X.: Contrast and generation make bart a good dialogue emotion recognizer. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a036, pp. 11002\u201311010 (2022)","key":"29_CR15","DOI":"10.1609\/aaai.v36i10.21348"},{"doi-asserted-by":"crossref","unstructured":"Li, X., Song, K., Feng, S., Wang, D., Zhang, Y.: A co-attention neural network model for emotion cause analysis with emotional context awareness. In: Proceedings of the 2018 Conference on Empirical Methods in Natural Language Processing, pp. 4752\u20134757. Association for Computational Linguistics, Brussels (2018)","key":"29_CR16","DOI":"10.18653\/v1\/D18-1506"},{"doi-asserted-by":"crossref","unstructured":"Li, Z., Tang, F., Zhao, M., Zhu, Y.: EmoCaps: emotion capsule based model for conversational emotion recognition. In: Muresan, S., Nakov, P., Villavicencio, A. (eds.) Findings of the Association for Computational Linguistics: ACL 2022, pp. 1610\u20131618. Association for Computational Linguistics, Dublin (2022)","key":"29_CR17","DOI":"10.18653\/v1\/2022.findings-acl.126"},{"doi-asserted-by":"crossref","unstructured":"Poria, S., Hazarika, D., Majumder, N., Naik, G., Cambria, E., Mihalcea, R.: MELD: a multimodal multi-party dataset for emotion recognition in conversations. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, pp. 527\u2013536 (2019)","key":"29_CR18","DOI":"10.18653\/v1\/P19-1050"},{"issue":"3","key":"29_CR19","doi-asserted-by":"publisher","first-page":"1832","DOI":"10.1109\/TAFFC.2022.3226559","volume":"14","author":"F Wang","year":"2023","unstructured":"Wang, F., Ding, Z., Xia, R., Li, Z., Yu, J.: Multimodal emotion-cause pair extraction in conversations. IEEE Trans. Affect. Comput. 14(3), 1832\u20131844 (2023)","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"9","key":"29_CR20","doi-asserted-by":"publisher","first-page":"5643","DOI":"10.1002\/int.22805","volume":"37","author":"L Wang","year":"2022","unstructured":"Wang, L., Li, R., Wu, Y., Jiang, Z.: A multiturn complementary generative framework for conversational emotion recognition. Int. J. Intell. Syst. 37(9), 5643\u20135671 (2022)","journal-title":"Int. J. Intell. Syst."},{"doi-asserted-by":"crossref","unstructured":"Wei, P., Zhao, J., Mao, W.: Effective inter-clause modeling for end-to-end emotion-cause pair extraction. In: Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 3171\u20133181 (2020)","key":"29_CR21","DOI":"10.18653\/v1\/2020.acl-main.289"},{"doi-asserted-by":"crossref","unstructured":"Wu, Z., Dai, X., Xia, R.: Pairwise tagging framework for end-to-end emotion-cause pair extraction. Front. Comput. Sci. 17(2) (2022)","key":"29_CR22","DOI":"10.1007\/s11704-022-1409-x"},{"doi-asserted-by":"crossref","unstructured":"Xia, R., Ding, Z.: Emotion-cause pair extraction: a new task to emotion analysis in texts. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, pp. 1003\u20131012. Association for Computational Linguistics, Florence (2019)","key":"29_CR23","DOI":"10.18653\/v1\/P19-1096"},{"key":"29_CR24","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2022.109978","volume":"258","author":"S Zou","year":"2022","unstructured":"Zou, S., Huang, X., Shen, X., Liu, H.: Improving multimodal fusion with main modal transformer for emotion recognition in conversation. Knowl.-Based Syst. 258, 109978 (2022)","journal-title":"Knowl.-Based Syst."}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-78128-5_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,30]],"date-time":"2024-11-30T08:07:08Z","timestamp":1732954028000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-78128-5_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,30]]},"ISBN":["9783031781278","9783031781285"],"references-count":24,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-78128-5_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,30]]},"assertion":[{"value":"30 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kolkata","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icpr2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icpr2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}