{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,8]],"date-time":"2026-03-08T17:41:29Z","timestamp":1772991689024,"version":"3.50.1"},"reference-count":43,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"4","license":[{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2022,10,1]],"date-time":"2022-10-01T00:00:00Z","timestamp":1664582400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"name":"Natural Science Foundation of Beijing","award":["4222036"],"award-info":[{"award-number":["4222036"]}]},{"DOI":"10.13039\/501100003816","name":"Huawei Technologies","doi-asserted-by":"publisher","award":["TC20201228005"],"award-info":[{"award-number":["TC20201228005"]}],"id":[{"id":"10.13039\/501100003816","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Affective Comput."],"published-print":{"date-parts":[[2022,10,1]]},"DOI":"10.1109\/taffc.2022.3212994","type":"journal-article","created":{"date-parts":[[2022,10,10]],"date-time":"2022-10-10T20:18:08Z","timestamp":1665433088000},"page":"1879-1891","source":"Crossref","is-referenced-by-count":13,"title":["Towards Contrastive Context-Aware Conversational Emotion Recognition"],"prefix":"10.1109","volume":"13","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-8715-0532","authenticated-orcid":false,"given":"Hanqing","family":"Zhang","sequence":"first","affiliation":[{"name":"School of Computer Science &#x0026; Technology, Beijing Institute of Technology, Beijing, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8660-3608","authenticated-orcid":false,"given":"Dawei","family":"Song","sequence":"additional","affiliation":[{"name":"School of Computer Science &#x0026; Technology, Beijing Institute of Technology, Beijing, China"}]}],"member":"263","reference":[{"key":"ref39","article-title":"Approximate nearest neighbor negative contrastive learning for dense text retrieval","author":"xiong","year":"2021","journal-title":"Proc 9th Int Conf Learn Representations"},{"key":"ref38","article-title":"Contrastive learning with adversarial perturbations for conditional text generation","author":"lee","year":"2021","journal-title":"Proc 9th Int Conf Learn Representations"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"ref32","first-page":"2238","article-title":"Contrastive learning using spectral methods","author":"zou","year":"2013","journal-title":"Proc Adv Neural Inf Process Syst"},{"key":"ref31","first-page":"307","article-title":"Noise-contrastive estimation of unnormalized statistical models, with applications to natural image statistics","volume":"13","author":"gutmann","year":"2012","journal-title":"J Mach Learn Res"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2006.100"},{"key":"ref37","article-title":"FairFil: Contrastive neural debiasing method for pretrained text encoders","author":"cheng","year":"2021","journal-title":"Proc 9th Int Conf Learn Representations"},{"key":"ref36","article-title":"An efficient framework for learning sentence representations","author":"logeswaran","year":"2018","journal-title":"Proc 6th Int Conf Learn Representations"},{"key":"ref35","article-title":"Representation learning with contrastive predictive coding","author":"van den oord","year":"2018","journal-title":"CoRR"},{"key":"ref34","first-page":"1597","article-title":"A simple framework for contrastive learning of visual representations","author":"chen","year":"2020","journal-title":"Proc 37th Int Conf Mach Learn"},{"key":"ref10","article-title":"A hierarchical transformer with speaker modeling for emotion recognition in conversation","author":"li","year":"2020","journal-title":"CoRR"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.70"},{"key":"ref11","article-title":"DialogXL: All-in-one XLNet for multi-party conversation emotion recognition","volume":"35","year":"0"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/TASLP.2021.3049898"},{"key":"ref13","article-title":"DialogueGCN: A Graph Convolutional Neural Network for Emotion Recognition in Conversation,&#x201D; Hong Kong, China","year":"0"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2019\/752"},{"key":"ref15","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.123"},{"key":"ref16","article-title":"MELD: A Multimodal Multi-Party Dataset for Emotion Recognition in Conversations, Florence, Italy","year":"0"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P19-1004"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1145\/3340531.3411918"},{"key":"ref19","article-title":"Utterance-level dialogue understanding: An empirical study","author":"ghosal","year":"2020","journal-title":"CoRR"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1016\/j.knosys.2021.107751"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2019.2929050"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.547"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2020.3005660"},{"key":"ref6","article-title":"ICON: Interactive Conversational Memory Network for Multimodal Emotion Detection, Brussels, Belgium","year":"0"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.09.057"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/P17-1081"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v33i01.33016818"},{"key":"ref7","article-title":"HiGRU: Hierarchical Gated Recurrent Units for Utterance-Level Emotion Recognition, Minneapolis, Minnesota","year":"0"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TCSS.2021.3095479"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6309"},{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v32i1.11325"},{"key":"ref20","article-title":"IEMOCAP: Interactive emotional dyadic motion capture database","volume":"42","year":"0"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1109\/TAFFC.2020.3005660"},{"key":"ref21","article-title":"Large-scale simple question answering with memory networks","author":"bordes","year":"2015","journal-title":"CoRR"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.findings-emnlp.106"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/D19-1016"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.emnlp-main.66"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2020.findings-emnlp.224"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2021.acl-long.125"},{"key":"ref43","article-title":"RoBERTa: A robustly optimized BERT pretraining approach","author":"liu","year":"2019","journal-title":"CoRR"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v34i05.6454"}],"container-title":["IEEE Transactions on Affective Computing"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx7\/5165369\/9964459\/09914616.pdf?arnumber=9914616","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,12,19]],"date-time":"2022-12-19T19:41:22Z","timestamp":1671478882000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/9914616\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,10,1]]},"references-count":43,"journal-issue":{"issue":"4"},"URL":"https:\/\/doi.org\/10.1109\/taffc.2022.3212994","relation":{},"ISSN":["1949-3045","2371-9850"],"issn-type":[{"value":"1949-3045","type":"electronic"},{"value":"2371-9850","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,10,1]]}}}