{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,4]],"date-time":"2026-05-04T07:04:14Z","timestamp":1777878254971,"version":"3.51.4"},"reference-count":55,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2026,10,1]],"date-time":"2026-10-01T00:00:00Z","timestamp":1790812800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2026,10,1]],"date-time":"2026-10-01T00:00:00Z","timestamp":1790812800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2026,10,1]],"date-time":"2026-10-01T00:00:00Z","timestamp":1790812800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2026,10,1]],"date-time":"2026-10-01T00:00:00Z","timestamp":1790812800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2026,10,1]],"date-time":"2026-10-01T00:00:00Z","timestamp":1790812800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2026,10,1]],"date-time":"2026-10-01T00:00:00Z","timestamp":1790812800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2026,10,1]],"date-time":"2026-10-01T00:00:00Z","timestamp":1790812800000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["52205275"],"award-info":[{"award-number":["52205275"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100009108","name":"Shandong University","doi-asserted-by":"publisher","award":["SDUQM2537"],"award-info":[{"award-number":["SDUQM2537"]}],"id":[{"id":"10.13039\/100009108","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012166","name":"National Key Research and Development Program of China","doi-asserted-by":"publisher","award":["2023YFC3604500"],"award-info":[{"award-number":["2023YFC3604500"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Information Fusion"],"published-print":{"date-parts":[[2026,10]]},"DOI":"10.1016\/j.inffus.2026.104329","type":"journal-article","created":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T23:31:46Z","timestamp":1775604706000},"page":"104329","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":0,"special_numbering":"C","title":["ASAC-Net: A novel multimodal alignment-complementary fusion framework for EEG-fNIRS emotion recognition"],"prefix":"10.1016","volume":"134","author":[{"ORCID":"https:\/\/orcid.org\/0009-0009-3653-7743","authenticated-orcid":false,"given":"Kaining","family":"Fang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4783-9244","authenticated-orcid":false,"given":"Jing","family":"Qu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-4704-4208","authenticated-orcid":false,"given":"Zixing","family":"Ding","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0597-7964","authenticated-orcid":false,"given":"Junhang","family":"Ding","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2340-6460","authenticated-orcid":false,"given":"Lingguo","family":"Bu","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"key":"10.1016\/j.inffus.2026.104329_bib0001","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2024.111241","article-title":"Driver multi-task emotion recognition network based on multi-modal facial video analysis","volume":"161","author":"Xiang","year":"2025","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.inffus.2026.104329_bib0002","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2023.107450","article-title":"Emotion recognition in EEG signals using deep learning methods: a review","volume":"165","author":"Jafari","year":"2023","journal-title":"Comput. Biol. Med."},{"issue":"7","key":"10.1016\/j.inffus.2026.104329_bib0003","doi-asserted-by":"crossref","first-page":"1324","DOI":"10.1109\/TNSRE.2018.2842464","article-title":"Intersession instability in fNIRS-based emotion recognition","volume":"26","author":"Wang","year":"2018","journal-title":"IEEE Trans. Neural Syst. Rehabil. Eng."},{"key":"10.1016\/j.inffus.2026.104329_bib0004","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2025.103354","article-title":"A hybrid model using multimodal feature perception and multiple cross-attention fusion for depressive episodes detection","volume":"124","author":"Wang","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.inffus.2026.104329_bib0005","doi-asserted-by":"crossref","first-page":"10","DOI":"10.1016\/j.inffus.2018.10.009","article-title":"Human emotion recognition using deep belief network architecture","volume":"51","author":"Hassan","year":"2019","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.inffus.2026.104329_bib0006","doi-asserted-by":"crossref","first-page":"64","DOI":"10.1016\/j.neuroimage.2013.06.062","article-title":"A wearable multi-channel fNIRS system for brain imaging in freely moving subjects","volume":"85","author":"Piper","year":"2014","journal-title":"Neuroimage"},{"key":"10.1016\/j.inffus.2026.104329_bib0007","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2024.106943","article-title":"E-FNet: a EEG-fNIRS dual-stream model for brain-computer interfaces","volume":"100","author":"Yu","year":"2025","journal-title":"Biomed. Signal Process. Control"},{"issue":"39\u201358","key":"10.1016\/j.inffus.2026.104329_bib0008","first-page":"3","article-title":"International affective picture system (IAPS): technical manual and affective ratings","volume":"1","author":"Lang","year":"1997","journal-title":"NIMH Center Study Emotion Attention"},{"issue":"10","key":"10.1016\/j.inffus.2026.104329_bib0009","doi-asserted-by":"crossref","first-page":"1314","DOI":"10.1109\/JPROC.2023.3277471","article-title":"Affective brain-computer interfaces (aBCIs): a tutorial","volume":"111","author":"Wu","year":"2023","journal-title":"Proc. IEEE"},{"key":"10.1016\/j.inffus.2026.104329_bib0010","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.102220","article-title":"Incongruity-aware multimodal physiology signals fusion for emotion recognition","volume":"105","author":"Li","year":"2024","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.inffus.2026.104329_bib0011","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2023.101847","article-title":"Emotion recognition from unimodal to multimodal analysis: a review","volume":"99","author":"Ezzameli","year":"2023","journal-title":"Inf. Fusion"},{"issue":"1\u20133","key":"10.1016\/j.inffus.2026.104329_bib0012","doi-asserted-by":"crossref","first-page":"36","DOI":"10.1049\/ccs2.12111","article-title":"EF-CorrCA:a multi-modal EEG-fNIRS subject independent model to assess speech quality on brain activity using correlated component analysis","volume":"6","author":"Charly","year":"2024","journal-title":"Cogn. Comput. Syst."},{"key":"10.1016\/j.inffus.2026.104329_bib0013","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2025.102971","article-title":"Explainable multi-frequency and multi-region fusion model for affective brain-computer interfaces","volume":"118","author":"Wang","year":"2025","journal-title":"Inf. Fusion"},{"issue":"1","key":"10.1016\/j.inffus.2026.104329_bib0014","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1109\/T-AFFC.2011.15","article-title":"DEAP: a database for emotion analysis using physiological signals","volume":"3","author":"Koelstra","year":"2012","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"3","key":"10.1016\/j.inffus.2026.104329_bib0015","doi-asserted-by":"crossref","first-page":"162","DOI":"10.1109\/TAMD.2015.2431497","article-title":"Investigating critical frequency bands and channels for EEG-based emotion recognition with deep neural networks","volume":"7","author":"Zheng","year":"2015","journal-title":"IEEE Trans. Auton. Ment. Dev."},{"issue":"1","key":"10.1016\/j.inffus.2026.104329_bib0016","doi-asserted-by":"crossref","first-page":"15","DOI":"10.1109\/TAFFC.2024.3407380","article-title":"FEAD: introduction to the fNIRS-EEG affective database-video stimuli","volume":"16","author":"Nia","year":"2024","journal-title":"IEEE Trans. Affect. Comput."},{"issue":"4","key":"10.1016\/j.inffus.2026.104329_bib0017","doi-asserted-by":"crossref","first-page":"2626","DOI":"10.1109\/TAFFC.2022.3181053","article-title":"Virtual reality for emotion elicitation - a review","volume":"14","author":"Somarathna","year":"2023","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.inffus.2026.104329_bib0018","series-title":"2024 IEEE Conference On Virtual Reality And 3D User Interfaces, VR 2024","first-page":"398","article-title":"Design and validation of a library of active affective tasks for emotion elicitation in VR","author":"Woodworth","year":"2024"},{"key":"10.1016\/j.inffus.2026.104329_bib0019","series-title":"2020 Int. Jt. Conf. Neural Netw. (IJCNN)","first-page":"1","article-title":"TSception: a deep learning framework for emotion detection using EEG","author":"Ding","year":"2020"},{"issue":"5","key":"10.1016\/j.inffus.2026.104329_bib0020","doi-asserted-by":"crossref","DOI":"10.1088\/1741-2552\/aace8c","article-title":"EEGNet: a compact convolutional neural network for EEG-based brain-computer interfaces","volume":"15","author":"Lawhern","year":"2018","journal-title":"J. Neural Eng."},{"issue":"3","key":"10.1016\/j.inffus.2026.104329_bib0021","doi-asserted-by":"crossref","first-page":"532","DOI":"10.1109\/TAFFC.2018.2817622","article-title":"EEG emotion recognition using dynamical graph convolutional neural networks","volume":"11","author":"Song","year":"2018","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.inffus.2026.104329_bib0022","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2023.104799","article-title":"GLFANet: a global to local feature aggregation network for EEG emotion recognition","volume":"85","author":"Liu","year":"2023","journal-title":"Biomed. Signal Process. Control"},{"key":"10.1016\/j.inffus.2026.104329_bib0023","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2025.114115","article-title":"Spatiotemporal isomorphic cross-brain region interaction network for cross-subject EEG emotion recognition","volume":"327","author":"An","year":"2025","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.inffus.2026.104329_bib0024","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2025.103135","article-title":"Exploring EEG and eye movement fusion for multi-class target RSVP-BCI","volume":"121","author":"Li","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.inffus.2026.104329_bib0025","series-title":"Proceedings of the 31st ACM International Conference on Multimedia","first-page":"5975","article-title":"Multimodal adaptive emotion transformer with flexible modality inputs on a novel dataset with continuous labels","author":"Jiang","year":"2023"},{"key":"10.1016\/j.inffus.2026.104329_bib0026","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2025.103023","article-title":"STA-Net: spatial-temporal alignment network for hybrid EEG-fNIRS decoding","volume":"119","author":"Liu","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.inffus.2026.104329_bib0027","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2024.126081","article-title":"A bidirectional cross-modal transformer representation learning model for EEG-fNIRS multimodal affective BCI","volume":"266","author":"Si","year":"2025","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.inffus.2026.104329_bib0028","series-title":"2013 6th International IEEE\/EMBS Conference on Neural Engineering (NER)","first-page":"81","article-title":"Differential entropy feature for EEG-based emotion classification","author":"Duan","year":"2013"},{"issue":"1","key":"10.1016\/j.inffus.2026.104329_bib0029","doi-asserted-by":"crossref","first-page":"18","DOI":"10.1109\/T-AFFC.2011.15","article-title":"DEAP: a database for emotion analysis ;using physiological signals","volume":"3","author":"Koelstra","year":"2012","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.inffus.2026.104329_bib0030","doi-asserted-by":"crossref","DOI":"10.1016\/j.jad.2025.119397","article-title":"HBUED: an EEG dataset for emotion recognition","volume":"385","author":"Liu","year":"2025","journal-title":"J. Affect. Disord."},{"issue":"11","key":"10.1016\/j.inffus.2026.104329_bib0031","doi-asserted-by":"crossref","first-page":"7343","DOI":"10.1109\/TVCG.2024.3456202","article-title":"An immersive and interactive VR dataset to elicit emotions","volume":"30","author":"Jiang","year":"2024","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"key":"10.1016\/j.inffus.2026.104329_bib0032","series-title":"2005 IEEE Workshop On Automatic Speech Recognition and Understanding (ASRU)","first-page":"381","article-title":"Evaluation of natural emotions using self assessment manikins","author":"Grimm","year":"2005"},{"key":"10.1016\/j.inffus.2026.104329_bib0033","doi-asserted-by":"crossref","first-page":"24675","DOI":"10.1109\/ACCESS.2023.3254892","article-title":"How much presence is enough? qualitative scales for interpreting the igroup presence questionnaire score","volume":"11","author":"Melo","year":"2023","journal-title":"IEEE Access"},{"key":"10.1016\/j.inffus.2026.104329_bib0034","doi-asserted-by":"crossref","first-page":"16","DOI":"10.3390\/virtualworlds2010002","article-title":"Cybersickness in Virtual Reality Questionnaire (CSQ-VR): A Validation and Comparison against SSQ and VRSQ","volume":"2","author":"Kourtesis","year":"2023","journal-title":"Virtual Worlds"},{"key":"10.1016\/j.inffus.2026.104329_bib0035","article-title":"Physiology, cerebral cortex functions","author":"Jawabri","year":"2023","journal-title":"StatPearls"},{"issue":"7","key":"10.1016\/j.inffus.2026.104329_bib0036","doi-asserted-by":"crossref","DOI":"10.1371\/journal.pbio.0060159","article-title":"Mapping the structural core of human cerebral cortex","volume":"6","author":"Hagmann","year":"2008","journal-title":"PLoS Biol."},{"key":"10.1016\/j.inffus.2026.104329_bib0037","doi-asserted-by":"crossref","first-page":"171","DOI":"10.1016\/j.neuroimage.2018.09.025","article-title":"Temporal derivative distribution repair (TDDR): a motion correction method for fNIRS","volume":"184","author":"Fishburn","year":"2019","journal-title":"Neuroimage"},{"key":"10.1016\/j.inffus.2026.104329_bib0038","doi-asserted-by":"crossref","first-page":"120","DOI":"10.3389\/fnhum.2019.00120","article-title":"fNIRS evidence for recognizably different positive emotions","volume":"13","author":"Hu","year":"2019","journal-title":"Front. Hum. Neurosci."},{"issue":"2","key":"10.1016\/j.inffus.2026.104329_bib0039","doi-asserted-by":"crossref","first-page":"568","DOI":"10.1109\/TAFFC.2019.2922912","article-title":"From regional to global brain: a novel hierarchical spatial-temporal neural network model for EEG emotion recognition","volume":"13","author":"Li","year":"2022","journal-title":"IEEE Trans. Affect. Comput."},{"key":"10.1016\/j.inffus.2026.104329_bib0040","series-title":"Advances in Neural Information Processing Systems 29 (NIPS 2016)","article-title":"Convolutional neural networks on graphs with fast localized spectral filtering","volume":"29","author":"Defferrard","year":"2016"},{"key":"10.1016\/j.inffus.2026.104329_bib0041","series-title":"2022 IEEE International Conference on Multimedia and Expo (ICME)","first-page":"1","article-title":"CAT: cross attention in vision transformer","author":"Lin","year":"2022"},{"issue":"4","key":"10.1016\/j.inffus.2026.104329_bib0042","doi-asserted-by":"crossref","first-page":"1374","DOI":"10.1109\/TCDS.2024.3357618","article-title":"Husformer: a multimodal transformer for multimodal human state recognition","volume":"16","author":"Wang","year":"2024","journal-title":"IEEE Trans. Cogn. Dev. Syst."},{"key":"10.1016\/j.inffus.2026.104329_bib0043","series-title":"Advances in Neural Information Processing Systems 30 (NIPS 2017)","article-title":"Dynamic routing between capsules","volume":"30","author":"Sabour","year":"2017"},{"key":"10.1016\/j.inffus.2026.104329_bib0044","series-title":"Companion of the 2024 on ACM International Joint Conference on Pervasive and Ubiquitous Computing","first-page":"612","article-title":"GNN4EEG: a benchmark and toolkit for electroencephalography classification with graph neural network","author":"Zhang","year":"2024"},{"key":"10.1016\/j.inffus.2026.104329_bib0045","series-title":"Proceedings of the 30th ACM International Conference on Multimedia","first-page":"209","article-title":"VigilanceNet: decouple intra-and inter-modality learning for multimodal vigilance estimation in RSVP-based BCI","author":"Cheng","year":"2022"},{"issue":"6","key":"10.1016\/j.inffus.2026.104329_bib0046","doi-asserted-by":"crossref","first-page":"1983","DOI":"10.1109\/TBME.2021.3132861","article-title":"A bimodal deep learning architecture for EEG-fNIRS decoding of overt and imagined speech","volume":"69","author":"Cooney","year":"2022","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"6","key":"10.1016\/j.inffus.2026.104329_bib0047","doi-asserted-by":"crossref","DOI":"10.3390\/s24061889","article-title":"EF-net: mental state recognition by analyzing multimodal EEG-fNIRS via CNN","volume":"24","author":"Arif","year":"2024","journal-title":"Sensors"},{"key":"10.1016\/j.inffus.2026.104329_bib0048","doi-asserted-by":"crossref","DOI":"10.1016\/j.inffus.2025.103023","article-title":"STA-Net: spatial-temporal alignment network for hybrid EEG-fNIRS decoding","volume":"119","author":"Liu","year":"2025","journal-title":"Inf. Fusion"},{"key":"10.1016\/j.inffus.2026.104329_bib0049","doi-asserted-by":"crossref","DOI":"10.1016\/j.compbiomed.2024.108973","article-title":"Temporal aware mixed attention-based convolution and transformer network for cross-subject EEG emotion recognition","volume":"181","author":"Si","year":"2024","journal-title":"Comput. Biol. Med."},{"key":"10.1016\/j.inffus.2026.104329_bib0050","doi-asserted-by":"crossref","DOI":"10.1016\/j.knosys.2024.112599","article-title":"MAS-DGAT-Net: a dynamic graph attention network with multibranch feature extraction and staged fusion for EEG emotion recognition","volume":"305","author":"Liu","year":"2024","journal-title":"Knowl. Based Syst."},{"key":"10.1016\/j.inffus.2026.104329_bib0051","doi-asserted-by":"crossref","DOI":"10.1016\/j.bspc.2025.108231","article-title":"Cross-subject emotion recognition by EEG driven spatio-temporal hybrid network based on domain adaptation and dynamic graph attention","volume":"110","author":"Liu","year":"2025","journal-title":"Biomed. Signal Process. Control"},{"key":"10.1016\/j.inffus.2026.104329_bib0052","doi-asserted-by":"crossref","first-page":"9070","DOI":"10.1109\/TMM.2024.3385676","article-title":"PGCN: pyramidal graph convolutional network for EEG emotion recognition","volume":"26","author":"Jin","year":"2024","journal-title":"IEEE Trans. Multimedia"},{"key":"10.1016\/j.inffus.2026.104329_bib0053","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2025.129654","article-title":"Dynamic collaborative evolutionary network: a novel spatio-temporal feature extraction framework for EEG emotion recognition","volume":"298","author":"Liu","year":"2026","journal-title":"Expert Syst. Appl."},{"key":"10.1016\/j.inffus.2026.104329_bib0054","doi-asserted-by":"crossref","first-page":"9070","DOI":"10.1109\/TMM.2024.3385676","article-title":"PGCN: pyramidal graph convolutional network for EEG emotion recognition","volume":"26","author":"Jin","year":"2024","journal-title":"IEEE Trans. Multimedia"},{"issue":"6","key":"10.1016\/j.inffus.2026.104329_bib0055","doi-asserted-by":"crossref","DOI":"10.1088\/1741-2552\/ad085a","article-title":"Identifying sex differences in EEG-based emotion recognition using graph convolutional network with attention mechanism","volume":"20","author":"Peng","year":"2023","journal-title":"J. Neural Eng."}],"container-title":["Information Fusion"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1566253526002083?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S1566253526002083?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2026,4,30]],"date-time":"2026-04-30T23:15:32Z","timestamp":1777590932000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S1566253526002083"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,10]]},"references-count":55,"alternative-id":["S1566253526002083"],"URL":"https:\/\/doi.org\/10.1016\/j.inffus.2026.104329","relation":{},"ISSN":["1566-2535"],"issn-type":[{"value":"1566-2535","type":"print"}],"subject":[],"published":{"date-parts":[[2026,10]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"ASAC-Net: A novel multimodal alignment-complementary fusion framework for EEG-fNIRS emotion recognition","name":"articletitle","label":"Article Title"},{"value":"Information Fusion","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.inffus.2026.104329","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2026 Elsevier B.V. All rights are reserved, including those for text and data mining, AI training, and similar technologies.","name":"copyright","label":"Copyright"}],"article-number":"104329"}}