{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T06:46:25Z","timestamp":1762325185020,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":28,"publisher":"ACM","license":[{"start":{"date-parts":[[2021,10,22]],"date-time":"2021-10-22T00:00:00Z","timestamp":1634860800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,10,22]]},"DOI":"10.1145\/3501409.3501645","type":"proceedings-article","created":{"date-parts":[[2022,1,2]],"date-time":"2022-01-02T06:18:08Z","timestamp":1641104288000},"page":"1336-1340","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Dialogue emotion recognition method based on dynamic graph network"],"prefix":"10.1145","author":[{"given":"Jun","family":"Cao","sequence":"first","affiliation":[{"name":"Computer and information engineering college, Tianjin Normal University, Tianjin, China"}]},{"given":"Wei","family":"Xia","sequence":"additional","affiliation":[{"name":"Computer and information engineering college, Tianjin Normal University, Tianjin, China"}]},{"given":"Simin","family":"Wang","sequence":"additional","affiliation":[{"name":"Computer and information engineering college, Tianjin Normal University, Tianjin, China"}]},{"given":"Ziping","family":"Zhao","sequence":"additional","affiliation":[{"name":"Computer and information engineering college, Tianjin Normal University, Tianjin, China"}]}],"member":"320","published-online":{"date-parts":[[2021,12,31]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1081"},{"key":"e_1_3_2_1_2_1","first-page":"397","volume-title":"Lyu","author":"Jiao Wenxiang","year":"2019","unstructured":"Wenxiang Jiao, Haiqin Yang, Irwin King, and Michael R. Lyu. 2019. HiGRU: Hierarchical gated recurrent units for utterance-level emotion recognition. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguis-tics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 397--406, Minneapolis, Minnesota, June. Association for Computational Linguistics."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1016"},{"key":"e_1_3_2_1_4_1","first-page":"8665","volume-title":"Dcr-net: A deep co-interactive relation network for joint dialog act recognition and sentiment classification. In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI","author":"Qin Libo","year":"2020","unstructured":"Libo Qin, Wanxiang Che, Yangming Li, Minheng Ni, and Ting Liu. 2020. Dcr-net: A deep co-interactive relation network for joint dialog act recognition and sentiment classification. In The Thirty-Fourth AAAI Conference on Artificial Intelligence, AAAI 2020, The Thirty-Second Innovative Applications of Artificial Intelligence Conference, IAAI 2020, The Tenth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2020, New York, NY, USA, February 7--12, 2020, pages 8665--8672. AAAI Press."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"crossref","unstructured":"Chul Min Lee Shrikanth S Narayanan et al. 2005. Toward detecting emotions in spoken di-alogs. IEEE transactions on speech and audio processing 13(2):293--303.","DOI":"10.1109\/TSA.2004.838534"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"crossref","unstructured":"Laurence Devillers and Laurence Vidrascu. 2006. Reallife emotions detection with lexical and paralinguistic cues on human-human call center dialogs. In Ninth International Confer-ence on Spoken Language Processing","DOI":"10.21437\/Interspeech.2006-275"},{"key":"e_1_3_2_1_7_1","volume-title":"Iemocap: Interactive emotional dyadic motion capture database. Language resources and evaluation, 42(4):335","author":"Busso Carlos","year":"2008","unstructured":"Carlos Busso, Murtaza Bulut, Chi-Chun Lee, Abe Kazemzadeh, Emily Mower, Samuel Kim, Jeannette N Chang, Sungbok Lee, and Shrikanth S Narayanan. 2008. Iemocap: Interactive emotional dyadic motion capture database. Language resources and evaluation, 42(4):335"},{"key":"e_1_3_2_1_8_1","volume-title":"Meld: A multimodal multi-party dataset for emotion recognition in conversations. arXiv preprint arXiv:1810.02508","author":"Poria Soujanya","year":"2018","unstructured":"Soujanya Poria, Devamanyu Hazarika, Navonil Majumder, Gautam Naik, Erik Cambria, and Rada Mihalcea. 2018. Meld: A multimodal multi-party dataset for emotion recognition in conversations. arXiv preprint arXiv:1810.02508"},{"key":"e_1_3_2_1_9_1","first-page":"6818","volume-title":"Dialoguernn: An attentive RNN for emotion detection in conversations. In The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI","author":"Majumder Navonil","year":"2019","unstructured":"Navonil Majumder, Soujanya Poria, Devamanyu Hazarika, Rada Mihalcea, Alexander F. Gelbukh, and Erik Cambria. 2019. Dialoguernn: An attentive RNN for emotion detection in conversations. In The Thirty-Third AAAI Conference on Artificial Intelligence, AAAI 2019, The Thirty-First Innovative Applications of Artificial Intelligence Conference, IAAI 2019, The Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, EAAI 2019, Honolulu, Hawaii, USA, January 27 - February 1, 2019, pages 6818--6825. AAAI Press."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1015"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.597"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/752"},{"key":"e_1_3_2_1_13_1","first-page":"5998","volume-title":"Proceedings of the 31st International Conference on Neural Information Processing Systems","author":"Vaswani Ashish","year":"2017","unstructured":"Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, pages 5998--6008."},{"key":"e_1_3_2_1_14_1","volume-title":"Dialogxl: All-in-one xlnet for multi-party conversation emotion recognition. arXiv preprint arXiv:2012.08695.","author":"Shen Weizhou","year":"2020","unstructured":"Weizhou Shen, Junqing Chen, Xiaojun Quan, and Zhixian Xie. 2020. Dialogxl: All-in-one xlnet for multi-party conversation emotion recognition. arXiv preprint arXiv:2012.08695."},{"key":"e_1_3_2_1_15_1","first-page":"5753","volume-title":"Advances in Neural Information Processing Systems","author":"Yang Zhilin","year":"2019","unstructured":"Zhilin Yang, Zihang Dai, Yiming Yang, Jaime Carbonell, Russ R Salakhutdinov, and Quoc V Le. 2019. Xlnet: Generalized autoregressive pretraining for language understanding. In Advances in Neural Information Processing Systems, pages 5753--5763."},{"key":"e_1_3_2_1_16_1","volume-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Lin-guistics: Human Language Technologies","volume":"1","author":"Devlin Jacob","unstructured":"Jacob Devlin, Ming-Wei Chang, Kenton Lee, and Kristina N. Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Lin-guistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4171--4186."},{"key":"e_1_3_2_1_17_1","volume-title":"Alec and Karthik Narasimhan","author":"Radford","year":"2018","unstructured":"Radford, Alec and Karthik Narasimhan. 2018. Improving Language Understanding by Gen-erative Pre-Training."},{"key":"e_1_3_2_1_18_1","first-page":"7370","volume-title":"Proc. AAAI Conf. Artif. Intell.","volume":"33","author":"Liang Yao","year":"2019","unstructured":"Yao Liang, Mao Chengsheng, Luo Yuan.: Graph convolutional networks for text classification, in Proc. AAAI Conf. Artif. Intell., vol. 33, 2019, pp. 7370--7377."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.acl-main.553"},{"key":"e_1_3_2_1_20_1","volume-title":"IEMOCAP: Interac-tive emotional dyadic motion capture database. Language resources and evaluation, 42(4):335--359.","author":"Busso Carlos","year":"2008","unstructured":"Carlos Busso, Murtaza Bulut, Chi-Chun Lee, Abe Kazemzadeh, Emily Mower, Samuel Kim, Jeannette N Chang, Sungbok Lee, and Shrikanth S Narayanan. 2008. IEMOCAP: Interac-tive emotional dyadic motion capture database. Language resources and evaluation, 42(4):335--359."},{"key":"e_1_3_2_1_21_1","first-page":"986","volume-title":"Proceedings of the Eighth Interna-tional Joint Conference on Natural Language Processing (Volume 1: Long Papers)","volume":"1","author":"Li Yanran","year":"2017","unstructured":"Yanran Li, Hui Su, Xiaoyu Shen, Wenjie Li, Ziqiang Cao, and Shuzi Niu. 2017. Dailyd-ialog: A manually labelled multi-turn dialogue dataset. In Proceedings of the Eighth Interna-tional Joint Conference on Natural Language Processing (Volume 1: Long Papers), volume 1, pages 986--995."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1050"},{"key":"e_1_3_2_1_23_1","first-page":"44","volume-title":"AAAI Workshops","author":"Sayyed","unstructured":"Sayyed M. Zahiri and Jinho D. Choi. 2017. Emotion detection on tv show transcripts with sequence based convolutional neural networks. In AAAI Workshops, pages 44--52."},{"key":"e_1_3_2_1_24_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.224"},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.123"},{"key":"e_1_3_2_1_26_1","volume-title":"International Con-ference on Learning Representations (ICLR)","author":"Veli\u010dkovi\u0107","year":"2018","unstructured":"Veli\u010dkovi\u0107 P, Cucurull G, Casanova A, et al. Graph Attention Networks. International Con-ference on Learning Representations (ICLR), 2018."},{"key":"e_1_3_2_1_27_1","volume-title":"WWW '20: The Web Conference 2020","author":"Link Continuous-Time","year":"2020","unstructured":"Qu, Liang, et al. \"Continuous-Time Link Prediction via Temporal Dependent Graph Neural Network.\" WWW '20: The Web Conference 2020 2020."},{"key":"e_1_3_2_1_28_1","volume-title":"Semi-Supervised Classification with Graph Convolutional Networks","author":"Kip F, T.","year":"2016","unstructured":"Kip F, T. N., and M. Welling. \"Semi-Supervised Classification with Graph Convolutional Networks.\" (2016)."}],"event":{"name":"EITCE 2021: 2021 5th International Conference on Electronic Information Technology and Computer Engineering","acronym":"EITCE 2021","location":"Xiamen China"},"container-title":["Proceedings of the 2021 5th International Conference on Electronic Information Technology and Computer Engineering"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3501409.3501645","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3501409.3501645","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T20:49:16Z","timestamp":1750193356000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3501409.3501645"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,10,22]]},"references-count":28,"alternative-id":["10.1145\/3501409.3501645","10.1145\/3501409"],"URL":"https:\/\/doi.org\/10.1145\/3501409.3501645","relation":{},"subject":[],"published":{"date-parts":[[2021,10,22]]},"assertion":[{"value":"2021-12-31","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}