{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T21:04:29Z","timestamp":1776891869549,"version":"3.51.2"},"reference-count":42,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,6,1]],"date-time":"2026-06-01T00:00:00Z","timestamp":1780272000000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62171391"],"award-info":[{"award-number":["62171391"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Neurocomputing"],"published-print":{"date-parts":[[2026,6]]},"DOI":"10.1016\/j.neucom.2026.133397","type":"journal-article","created":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T16:18:19Z","timestamp":1774628299000},"page":"133397","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["Adaptive prototype and relation-aware distillation for multimodal sentiment analysis with incomplete data"],"prefix":"10.1016","volume":"682","author":[{"given":"Xiyang","family":"Sun","sequence":"first","affiliation":[]},{"given":"Linkai","family":"Luo","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6314-1159","authenticated-orcid":false,"given":"Hong","family":"Peng","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.neucom.2026.133397_bib0005","series-title":"ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)","first-page":"1","article-title":"Exploiting modality-invariant feature for robust multimodal emotion recognition with missing modalities","author":"Zuo","year":"2023"},{"key":"10.1016\/j.neucom.2026.133397_bib0010","series-title":"Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition","first-page":"12458","article-title":"Correlation-decoupled knowledge distillation for multimodal sentiment analysis with incomplete modalities","author":"Li","year":"2024"},{"key":"10.1016\/j.neucom.2026.133397_bib0015","doi-asserted-by":"crossref","first-page":"209","DOI":"10.1016\/j.inffus.2019.06.019","article-title":"A snapshot research and implementation of multimodal information fusion for data-driven emotion recognition","volume":"53","author":"Jiang","year":"2020","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.neucom.2026.133397_bib0020","doi-asserted-by":"crossref","DOI":"10.1016\/j.neucom.2023.127181","article-title":"Multimodal transformer with adaptive modality weighting for multimodal sentiment analysis","volume":"572","author":"Wang","year":"2024","journal-title":"Neurocomputing"},{"key":"10.1016\/j.neucom.2026.133397_bib0025","series-title":"Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition","first-page":"1405","article-title":"Missing modalities imputation via cascaded residual autoencoder","author":"Tran","year":"2017"},{"key":"10.1016\/j.neucom.2026.133397_bib0030","series-title":"Companion Publication of the 2020 International Conference on Multimodal Interaction","first-page":"400","article-title":"Training strategies to handle missing modalities for audio-visual expression recognition","author":"Parthasarathy","year":"2020"},{"key":"10.1016\/j.neucom.2026.133397_bib0035","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"10074","article-title":"A unified self-distillation framework for multimodal sentiment analysis with uncertain missing modalities","volume":"vol. 38","author":"Li","year":"2024"},{"key":"10.1016\/j.neucom.2026.133397_bib0040","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"2302","article-title":"SMIL: multimodal learning with severely missing modality","volume":"vol. 35","author":"Ma","year":"2021"},{"key":"10.1016\/j.neucom.2026.133397_bib0045","series-title":"IJCAI","first-page":"4026","article-title":"Deep correlated predictive subspace learning for incomplete multi-view semi-supervised classification","author":"Xue","year":"2019"},{"key":"10.1016\/j.neucom.2026.133397_bib0050","first-page":"28515","article-title":"Toward robust incomplete multimodal sentiment analysis via hierarchical representation learning","volume":"37","author":"Li","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"issue":"8","key":"10.1016\/j.neucom.2026.133397_bib0055","doi-asserted-by":"crossref","first-page":"7657","DOI":"10.1109\/TCSVT.2024.3376564","article-title":"Trustworthy multimodal fusion for sentiment analysis in ordinal sentiment space","volume":"34","author":"Xie","year":"2024","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"10.1016\/j.neucom.2026.133397_bib0060","series-title":"Proceedings of the 29th ACM International Conference on Multimedia","first-page":"4400","article-title":"Transformer-based feature reconstruction network for robust multimodal sentiment analysis","author":"Yuan","year":"2021"},{"key":"10.1016\/j.neucom.2026.133397_bib0065","doi-asserted-by":"crossref","first-page":"1686","DOI":"10.1162\/tacl_a_00628","article-title":"MissModal: increasing robustness to missing modality in multimodal sentiment analysis","volume":"11","author":"Lin","year":"2023","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"10.1016\/j.neucom.2026.133397_bib0070","series-title":"Proceedings of the 45th International ACM SIGIR Conference on Research and Development in Information Retrieval","first-page":"1545","article-title":"Tag-assisted multimodal sentiment analysis under uncertain missing modalities","author":"Zeng","year":"2022"},{"key":"10.1016\/j.neucom.2026.133397_bib0075","doi-asserted-by":"crossref","first-page":"4796","DOI":"10.1109\/TIP.2024.3445740","article-title":"Relation knowledge distillation by auxiliary learning for object detection","volume":"33","author":"Wang","year":"2024","journal-title":"IEEE Trans. Image Process."},{"key":"10.1016\/j.neucom.2026.133397_bib0080","series-title":"Proceedings of the 28th ACM International Conference on Multimedia","first-page":"1122","article-title":"MISA: modality-invariant and-specific representations for multimodal sentiment analysis","author":"Hazarika","year":"2020"},{"key":"10.1016\/j.neucom.2026.133397_bib0085","series-title":"Proceedings of the AAAI Conference on Artificial Intelligence","first-page":"10790","article-title":"Learning modality-specific representations with self-supervised multi-task learning for multimodal sentiment analysis","volume":"vol. 35","author":"Yu","year":"2021"},{"key":"10.1016\/j.neucom.2026.133397_bib0090","series-title":"Proceedings of the 30th ACM International Conference on Multimedia","first-page":"1642","article-title":"Disentangled representation learning for multimodal emotion recognition","author":"Yang","year":"2022"},{"key":"10.1016\/j.neucom.2026.133397_bib0095","doi-asserted-by":"crossref","first-page":"652","DOI":"10.1162\/tacl_a_00567","article-title":"How much do language models copy from their training data? Evaluating linguistic novelty in text generation using Raven","volume":"11","author":"McCoy","year":"2023","journal-title":"Trans. Assoc. Comput. Linguist."},{"key":"10.1016\/j.neucom.2026.133397_bib0100","series-title":"Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing","first-page":"9180","article-title":"Improving multimodal fusion with hierarchical mutual information maximization for multimodal sentiment analysis","author":"Han","year":"2021"},{"key":"10.1016\/j.neucom.2026.133397_bib0105","series-title":"Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing","first-page":"1103","article-title":"Tensor fusion network for multimodal sentiment analysis","author":"Zadeh","year":"2017"},{"key":"10.1016\/j.neucom.2026.133397_bib0110","series-title":"Proceedings of the Conference. Association for Computational Linguistics. Meeting","first-page":"6558","article-title":"Multimodal transformer for unaligned multimodal language sequences","volume":"vol. 2019","author":"Tsai","year":"2019"},{"issue":"3","key":"10.1016\/j.neucom.2026.133397_bib0115","doi-asserted-by":"crossref","first-page":"2276","DOI":"10.1109\/TAFFC.2022.3172360","article-title":"Hybrid contrastive learning of tri-modal representation for multimodal sentiment analysis","volume":"14","author":"Mai","year":"2022","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.neucom.2026.133397_bib0120","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2024.102338","article-title":"Online multi-hypergraph fusion learning for cross-subject emotion recognition","volume":"108","author":"Pan","year":"2024","journal-title":"Inf. Fusion."},{"key":"10.1016\/j.neucom.2026.133397_bib0125","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2022.119240","article-title":"Heterogeneous graph convolution based on in-domain self-supervision for multimodal sentiment analysis","volume":"213","author":"Zeng","year":"2023","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.neucom.2026.133397_bib0130","series-title":"Proceedings of the Computer Vision and Pattern Recognition Conference","first-page":"14314","article-title":"EMOE: modality-specific enhanced dynamic emotion experts","author":"Fang","year":"2025"},{"key":"10.1016\/j.neucom.2026.133397_bib0135","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2025.129341","article-title":"A multi-scale representation and multi-level decision learning network for multimodal sentiment analysis","volume":"297","author":"Li","year":"2026","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.neucom.2026.133397_bib0140","series-title":"Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers)","first-page":"2608","article-title":"Missing modality imagination network for emotion recognition with uncertain missing modalities","author":"Zhao","year":"2021"},{"key":"10.1016\/j.neucom.2026.133397_bib0145","author":"Hazarika"},{"key":"10.1016\/j.neucom.2026.133397_bib0150","series-title":"Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing (EMNLP 2023)","first-page":"756","article-title":"Learning language-guided adaptive hyper-modality representation for multimodal sentiment analysis","author":"Zhang","year":"2023"},{"key":"10.1016\/j.neucom.2026.133397_bib0155","series-title":"Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies","first-page":"4171","article-title":"BERT: pre-training of deep bidirectional transformers for language understanding","author":"Devlin","year":"2019"},{"key":"10.1016\/j.neucom.2026.133397_bib0160","doi-asserted-by":"crossref","first-page":"18","DOI":"10.25080\/Majora-7b98e3ed-003","article-title":"librosa: audio and music signal analysis in Python","volume":"2015","author":"McFee","year":"2015","journal-title":"SciPy"},{"key":"10.1016\/j.neucom.2026.133397_bib0165","first-page":"59","article-title":"OpenFace 2.0: facial behavior analysis toolkit","author":"Baltrusaitis","year":"2018","journal-title":"IEEE Comput. Soc."},{"key":"10.1016\/j.neucom.2026.133397_bib0170","doi-asserted-by":"crossref","first-page":"23716","DOI":"10.52202\/068431-1723","article-title":"Flamingo: a visual language model for few-shot learning","volume":"35","author":"Alayrac","year":"2022","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133397_bib0175","author":"Zadeh"},{"key":"10.1016\/j.neucom.2026.133397_bib0180","series-title":"Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)","first-page":"2236","article-title":"Multimodal language analysis in the wild: CMU-MOSEI dataset and interpretable dynamic fusion graph","author":"Zadeh","year":"2018"},{"key":"10.1016\/j.neucom.2026.133397_bib0185","series-title":"Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics","first-page":"3718","article-title":"CH-SIMS: a Chinese multimodal sentiment analysis dataset with fine-grained annotation of modality","author":"Yu","year":"2020"},{"key":"10.1016\/j.neucom.2026.133397_bib0190","first-page":"55943","article-title":"Towards robust multimodal sentiment analysis with incomplete data","volume":"37","author":"Zhang","year":"2024","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.neucom.2026.133397_bib0195","doi-asserted-by":"crossref","first-page":"4909","DOI":"10.1109\/TMM.2022.3183830","article-title":"Cross-modal enhancement network for multimodal sentiment analysis","volume":"25","author":"Wang","year":"2022","journal-title":"IEEE Trans. Multimedia"},{"key":"10.1016\/j.neucom.2026.133397_bib0200","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.109259","article-title":"TETFN: a text enhanced transformer fusion network for multimodal sentiment analysis","volume":"136","author":"Wang","year":"2023","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.neucom.2026.133397_bib0205","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.126274","article-title":"Learning fine-grained representation with token-level alignment for multimodal sentiment analysis","volume":"269","author":"Li","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.neucom.2026.133397_bib0210","series-title":"Findings of the Association for Computational Linguistics: EMNLP 2025","first-page":"11252","article-title":"TF-Mamba: text-enhanced fusion mamba with missing modalities for robust multimodal sentiment analysis","author":"Li","year":"2025"}],"container-title":["Neurocomputing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231226007940?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0925231226007940?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T20:30:58Z","timestamp":1776889858000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0925231226007940"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,6]]},"references-count":42,"alternative-id":["S0925231226007940"],"URL":"https:\/\/doi.org\/10.1016\/j.neucom.2026.133397","relation":{},"ISSN":["0925-2312"],"issn-type":[{"value":"0925-2312","type":"print"}],"subject":[],"published":{"date-parts":[[2026,6]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Adaptive prototype and relation-aware distillation for multimodal sentiment analysis with incomplete data","name":"articletitle","label":"Article Title"},{"value":"Neurocomputing","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.neucom.2026.133397","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"133397"}}