{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T13:07:05Z","timestamp":1774530425382,"version":"3.50.1"},"reference-count":48,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["69189338"],"award-info":[{"award-number":["69189338"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neurocomputing"],"published-print":{"date-parts":[[2026,6]]},"DOI":"10.1016\/j.neucom.2026.133306","type":"journal-article","created":{"date-parts":[[2026,3,12]],"date-time":"2026-03-12T07:52:21Z","timestamp":1773301941000},"page":"133306","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Relational graph-driven differential denoising and diffusion attention fusion for multimodal conversational emotion recognition"],"prefix":"10.1016","volume":"681","author":[{"given":"Ying","family":"Liu","sequence":"first","affiliation":[]},{"given":"Yuntao","family":"Shou","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Ai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9787-2002","authenticated-orcid":false,"given":"Tao","family":"Meng","sequence":"additional","affiliation":[]},{"given":"Keqin","family":"Li","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neucom.2026.133306_bib0005","doi-asserted-by":"crossref","unstructured":"A. Ahmed, S. Aleem, Enhancing human-robot collaboration through multimodal emotion and context-aware. Available at SSRN 5237859, 2025.","DOI":"10.2139\/ssrn.5237859"},{"issue":"2","key":"10.1016\/j.neucom.2026.133306_bib0010","doi-asserted-by":"crossref","first-page":"2099","DOI":"10.1007\/s10586-023-04073-z","article-title":"Emotion and sentiment analysis for intelligent customer service conversation using a multi-task ensemble framework","volume":"27","author":"Chen","year":"2024","journal-title":"Cluster Comput."},{"key":"10.1016\/j.neucom.2026.133306_bib0015","first-page":"1","article-title":"Reinforcement learning techniques for autonomous robots in virtual environments with LLM-based multimodal data integration and virtual embodiment","volume":"12","author":"Lee","year":"2025","journal-title":"Clin. Res. Clin. Trials"},{"key":"10.1016\/j.neucom.2026.133306_bib0020","series-title":"2024 IEEE Frontiers in Education Conference (FIE)","first-page":"1","article-title":"A multimodal approach for real-time engagement monitoring in E-learning using machine learning","author":"Shankar","year":"2024"},{"issue":"4","key":"10.1016\/j.neucom.2026.133306_bib0025","doi-asserted-by":"crossref","first-page":"4787","DOI":"10.1007\/s11042-022-12315-2","article-title":"A novel multi-modal depression detection approach based on mobile crowd sensing and task-based mechanisms","volume":"82","author":"Thati","year":"2023","journal-title":"Multimed. Tools Appl."},{"issue":"2","key":"10.1016\/j.neucom.2026.133306_bib0030","doi-asserted-by":"crossref","first-page":"34","DOI":"10.1007\/s10462-023-10645-7","article-title":"An efficient multimodal sentiment analysis in social media using hybrid optimal multi-scale residual attention network","volume":"57","author":"Subbaiah","year":"2024","journal-title":"Artif. Intell. Rev."},{"key":"10.1016\/j.neucom.2026.133306_bib0035","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"4625","article-title":"Joint multimodal transformer for emotion recognition in the wild","author":"Waligora","year":"2024"},{"issue":"3","key":"10.1016\/j.neucom.2026.133306_bib0040","doi-asserted-by":"crossref","first-page":"4908","DOI":"10.1109\/TNNLS.2024.3367940","article-title":"DER-GCN: Dialog and event relation-aware graph convolutional neural network for multimodal dialog emotion recognition","volume":"36","author":"Ai","year":"2024","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"issue":"2","key":"10.1016\/j.neucom.2026.133306_bib0045","doi-asserted-by":"crossref","first-page":"712","DOI":"10.1109\/TAFFC.2024.3461148","article-title":"Dynamic emotion-dependent network with relational subgraph interaction for multimodal emotion recognition","volume":"16","author":"Wang","year":"2024","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.neucom.2026.133306_bib0050","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"1574","article-title":"Big-fusion: brain-inspired global-local context fusion framework for multimodal emotion recognition in conversations","volume":"vol. 39","author":"Wang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133306_bib0055","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"1438","article-title":"Multi-to-single: reducing multimodal dependency in emotion recognition through contrastive learning","volume":"vol. 39","author":"Liu","year":"2025"},{"key":"10.1016\/j.neucom.2026.133306_bib0060","author":"Jin"},{"issue":"2","key":"10.1016\/j.neucom.2026.133306_bib0065","doi-asserted-by":"crossref","first-page":"1177","DOI":"10.1109\/TAFFC.2024.3498443","article-title":"A low-rank matching attention based cross-modal feature fusion method for conversational emotion recognition","volume":"16","author":"Shou","year":"2024","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.neucom.2026.133306_bib0070","author":"Shou"},{"key":"10.1016\/j.neucom.2026.133306_bib0075","series-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition","first-page":"4652","article-title":"M2FNet: multi-modal fusion network for emotion recognition in conversation","author":"Chudasama","year":"2022"},{"issue":"1","key":"10.1016\/j.neucom.2026.133306_bib0080","doi-asserted-by":"crossref","first-page":"2126","DOI":"10.1038\/s41598-025-85859-6","article-title":"Multimodal sentiment analysis based on multi-layer feature fusion and multi-task learning","volume":"15","author":"Cai","year":"2025","journal-title":"Sci. Rep."},{"key":"10.1016\/j.neucom.2026.133306_bib0085","author":"Jin"},{"issue":"4","key":"10.1016\/j.neucom.2026.133306_bib0090","doi-asserted-by":"crossref","first-page":"1856","DOI":"10.1109\/TAFFC.2024.3378570","article-title":"Contrastive learning based modality-invariant feature acquisition for robust multimodal emotion recognition with missing modalities","volume":"15","author":"Liu","year":"2024","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.neucom.2026.133306_bib0095","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2024.112825","article-title":"SDR-GNN: spectral domain reconstruction graph neural network for incomplete multimodal learning in conversational emotion recognition","volume":"309","author":"Fu","year":"2025","journal-title":"Knowl.-Based Syst."},{"key":"10.1016\/j.neucom.2026.133306_bib0100","first-page":"39552","article-title":"Text-difuse: an interactive multi-modal image fusion framework based on text-modulated diffusion model","volume":"37","author":"Zhang","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"2","key":"10.1016\/j.neucom.2026.133306_bib0105","doi-asserted-by":"crossref","first-page":"304","DOI":"10.1080\/10864415.2025.2471673","article-title":"Mitigating the proliferation of fake image-text reviews: A two-tier intra-and inter-modal fusion framework","volume":"29","author":"Du","year":"2025","journal-title":"Int. J. Electron. Commer."},{"key":"10.1016\/j.neucom.2026.133306_bib0110","series-title":"Proceedings of the AAAI conference on artificial intelligence","first-page":"6818","article-title":"DialogueRNN: an attentive rnn for emotion detection in conversations","volume":"vol. 33","author":"Majumder","year":"2019"},{"key":"10.1016\/j.neucom.2026.133306_bib0115","series-title":"Proceedings of the AAAI conference on artificial intelligence","first-page":"10790","article-title":"Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis","volume":"vol. 35","author":"Yu","year":"2021"},{"key":"10.1016\/j.neucom.2026.133306_bib0120","series-title":"Proceedings of the 58th annual meeting of the association for computational linguistics","first-page":"3718","article-title":"CH-SIMS: A chinese multimodal sentiment analysis dataset with fine-grained annotation of modality","author":"Yu","year":"2020"},{"key":"10.1016\/j.neucom.2026.133306_bib0125","series-title":"Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","first-page":"5669","article-title":"CoMPM: context modeling with speaker\u2019s pre-trained memory tracking for emotion recognition in conversation","author":"Lee","year":"2022"},{"key":"10.1016\/j.neucom.2026.133306_bib0130","author":"Li"},{"key":"10.1016\/j.neucom.2026.133306_bib0135","series-title":"Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","first-page":"14752","article-title":"MultiEMO: An attention-based correlation-aware multimodal fusion framework for emotion recognition in conversations","author":"Shi","year":"2023"},{"issue":"30","key":"10.1016\/j.neucom.2026.133306_bib0140","doi-asserted-by":"crossref","first-page":"25445","DOI":"10.1007\/s00521-025-11586-8","article-title":"Financial asset price prediction with graph neural network-based temporal deep learning models","volume":"37","author":"Uygun","year":"2025","journal-title":"Neural Comput. Appl."},{"key":"10.1016\/j.neucom.2026.133306_bib0145","first-page":"1","article-title":"Analyzing transaction graphs via motif-based graph representation learning for cryptocurrency price prediction","author":"Celik","year":"2025","journal-title":"Comput. Econ."},{"key":"10.1016\/j.neucom.2026.133306_bib0150","author":"Ghosal"},{"key":"10.1016\/j.neucom.2026.133306_bib0155","author":"Joshi"},{"key":"10.1016\/j.neucom.2026.133306_bib0160","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102590","article-title":"Adversarial alignment and graph fusion via information bottleneck for multimodal emotion recognition in conversations","volume":"112","author":"Shou","year":"2024","journal-title":"Inf. Fusion."},{"issue":"3","key":"10.1016\/j.neucom.2026.133306_bib0165","doi-asserted-by":"crossref","first-page":"1885","DOI":"10.1109\/TAFFC.2025.3544608","article-title":"Semantic and emotional dual channel for emotion recognition in conversation","volume":"16","author":"Yang","year":"2025","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.neucom.2026.133306_bib0170","series-title":"2024 IEEE International Conference on Evolving and Adaptive Intelligent Systems (EAIS)","first-page":"1","article-title":"Unveiling graph power: segmentanything and GCN synergy for instance segmentation and classification","author":"Scarrica","year":"2024"},{"key":"10.1016\/j.neucom.2026.133306_bib0175","series-title":"Proceedings of the 31st International Conference on Computational Linguistics","first-page":"256","article-title":"Dynamic graph neural ODE network for multi-modal emotion recognition in conversation","author":"Shou","year":"2025"},{"key":"10.1016\/j.neucom.2026.133306_bib0180","series-title":"Proceedings of the conference. Association for computational linguistics. Meeting","first-page":"6558","article-title":"Multimodal transformer for unaligned multimodal language sequences","volume":"vol. 2019","author":"Tsai","year":"2019"},{"key":"10.1016\/j.neucom.2026.133306_bib0185","series-title":"Proceedings of the 28th ACM international conference on multimedia","first-page":"1122","article-title":"MISA: modality-invariant and-specific representations for multimodal sentiment analysis","author":"Hazarika","year":"2020"},{"key":"10.1016\/j.neucom.2026.133306_bib0190","series-title":"Proceedings of the conference. Association for computational linguistics. Meeting","first-page":"2359","article-title":"Integrating multimodal information in large pretrained transformers","volume":"vol. 2020","author":"Rahman","year":"2020"},{"key":"10.1016\/j.neucom.2026.133306_bib0195","series-title":"ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"7037","article-title":"MM-DFN: Multimodal dynamic fusion network for emotion recognition in conversations","author":"Hu","year":"2022"},{"key":"10.1016\/j.neucom.2026.133306_bib0200","author":"Mao"},{"key":"10.1016\/j.neucom.2026.133306_bib0205","series-title":"Proceedings of the 31st ACM International Conference on Multimedia","first-page":"5923","article-title":"Revisiting disentanglement and fusion on modality and context in conversational multimodal emotion recognition","author":"Li","year":"2023"},{"key":"10.1016\/j.neucom.2026.133306_bib0210","doi-asserted-by":"crossref","first-page":"776","DOI":"10.1109\/TMM.2023.3271019","article-title":"A transformer-based model with self-distillation for multimodal emotion recognition in conversations","volume":"26","author":"Ma","year":"2023","journal-title":"IEEE Trans. Multimedia"},{"key":"10.1016\/j.neucom.2026.133306_bib0215","series-title":"ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"1","article-title":"MGAT: multi-granularity attention based transformers for multi-modal emotion recognition","author":"Fan","year":"2023"},{"issue":"4","key":"10.1016\/j.neucom.2026.133306_bib0220","doi-asserted-by":"crossref","first-page":"335","DOI":"10.1007\/s10579-008-9076-6","article-title":"IEMOCAP: Interactive emotional dyadic motion capture database","volume":"42","author":"Busso","year":"2008","journal-title":"Lang. Resour. Eval."},{"key":"10.1016\/j.neucom.2026.133306_bib0225","author":"Poria"},{"key":"10.1016\/j.neucom.2026.133306_bib0230","author":"Hu"},{"key":"10.1016\/j.neucom.2026.133306_bib0235","series-title":"2023 IEEE 29th International Conference on Parallel and Distributed Systems (ICPADS)","first-page":"397","article-title":"A two-stage multimodal emotion recognition model based on graph contrastive learning","author":"Ai","year":"2023"},{"key":"10.1016\/j.neucom.2026.133306_bib0240","doi-asserted-by":"crossref","first-page":"77","DOI":"10.1109\/TMM.2023.3260635","article-title":"GraphCFC: A directed graph based cross-modal feature complementation approach for multimodal conversational emotion recognition","volume":"26","author":"Li","year":"2023","journal-title":"IEEE Trans. Multimedia"}],"container-title":["Neurocomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231226007034?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231226007034?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,3,26]],"date-time":"2026-03-26T12:30:40Z","timestamp":1774528240000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0925231226007034"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,6]]},"references-count":48,"alternative-id":["S0925231226007034"],"URL":"https:\/\/doi.org\/10.1016\/j.neucom.2026.133306","relation":{},"ISSN":["0925-2312"],"issn-type":[{"value":"0925-2312","type":"print"}],"subject":[],"published":{"date-parts":[[2026,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Relational graph-driven differential denoising and diffusion attention fusion for multimodal conversational emotion recognition","name":"articletitle","label":"Article Title"},{"value":"Neurocomputing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neucom.2026.133306","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"133306"}}