{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,28]],"date-time":"2025-11-28T06:51:11Z","timestamp":1764312671990,"version":"3.46.0"},"reference-count":58,"publisher":"Institute of Electrical and Electronics Engineers (IEEE)","issue":"12","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/ieeexplore.ieee.org\/Xplorehelp\/downloads\/license-information\/IEEE.html"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":["IEEE Trans. Artif. Intell."],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1109\/tai.2025.3560590","type":"journal-article","created":{"date-parts":[[2025,4,14]],"date-time":"2025-04-14T13:42:03Z","timestamp":1744638123000},"page":"3207-3216","source":"Crossref","is-referenced-by-count":0,"title":["Mixture-of-Experts for Open Set Domain Adaptation: A Dual-Space Detection Approach"],"prefix":"10.1109","volume":"6","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-1386-8381","authenticated-orcid":false,"given":"Zhenbang","family":"Du","sequence":"first","affiliation":[{"name":"Key Laboratory of the Ministry of Education for Image Processing and Intelligent Control, School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-1458-3248","authenticated-orcid":false,"given":"Jiayu","family":"An","sequence":"additional","affiliation":[{"name":"Key Laboratory of the Ministry of Education for Image Processing and Intelligent Control, School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0002-2539-036X","authenticated-orcid":false,"given":"Yunlu","family":"Tu","sequence":"additional","affiliation":[{"name":"Key Laboratory of the Ministry of Education for Image Processing and Intelligent Control, School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0000-2389-560X","authenticated-orcid":false,"given":"Jiahao","family":"Hong","sequence":"additional","affiliation":[{"name":"School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7153-9703","authenticated-orcid":false,"given":"Dongrui","family":"Wu","sequence":"additional","affiliation":[{"name":"Key Laboratory of the Ministry of Education for Image Processing and Intelligent Control, School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China"}]}],"member":"263","reference":[{"key":"ref1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref2","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","volume-title":"Int. Conf. Learn. Representations","author":"Dosovitskiy","year":"2021"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/TNNLS.2020.3028503"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/TAI.2024.3350543"},{"key":"ref5","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2020.04.045"},{"key":"ref6","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2020.3042300"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1016\/j.ymssp.2021.108487"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/TIM.2023.3282664"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2024.3362343"},{"key":"ref10","first-page":"1180","article-title":"Unsupervised domain adaptation by backpropagation","volume-title":"Proc. Int\u2019l Conf. Mach. Learn.","author":"Ganin","year":"2015"},{"key":"ref11","first-page":"97","article-title":"Learning transferable features with deep adaptation networks","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Long","year":"2015"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-01228-1_10"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00304"},{"key":"ref14","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58517-4_25"},{"key":"ref15","first-page":"6468","article-title":"Progressive graph learning for open-set domain adaptation","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Luo","year":"2020"},{"key":"ref16","first-page":"16755","article-title":"Unknown-aware domain adversarial learning for open-set domain adaptation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Jang","year":"2022"},{"key":"ref17","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02309"},{"key":"ref18","doi-asserted-by":"publisher","DOI":"10.1016\/j.inffus.2023.101912"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00963"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01917"},{"key":"ref21","article-title":"A baseline for detecting misclassified and out-of-distribution examples in neural networks","volume-title":"Int. Conf. Learn. Represent.","author":"Hendrycks","year":"2017"},{"key":"ref22","first-page":"9175","article-title":"Reducing network agnostophobia","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"31","author":"Dhamija","year":"2018"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01087"},{"key":"ref24","first-page":"8583","article-title":"Scaling vision with sparse mixture of experts","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"34","author":"Riquelme","year":"2021"},{"article-title":"\u201cA review of sparse expert models in deep learning,\u201d","year":"2022","author":"Fedus","key":"ref25"},{"key":"ref26","first-page":"9564","article-title":"Multimodal contrastive learning with limoe: The language-image mixture of experts","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Mustafa","year":"2022"},{"key":"ref27","article-title":"Sparse mixture-of-experts are domain generalizable learners","volume-title":"Int. Conf. Learn. Represent.","author":"Li","year":"2023"},{"key":"ref28","first-page":"8291","article-title":"Vision gnn: An image is worth graph of nodes","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"35","author":"Han","year":"2022"},{"key":"ref29","article-title":"Semi-supervised classification with graph convolutional networks","volume-title":"Int. Conf. Learn. Represent.","author":"Kipf","year":"2017"},{"key":"ref30","article-title":"Graph attention networks","volume-title":"Int. Conf. Learn. Represent.","author":"Veli\u010dkovi\u0107","year":"2018"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-58347-1_8"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.01053"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46493-0_36"},{"key":"ref34","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00334"},{"key":"ref35","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01362"},{"key":"ref36","first-page":"1647","article-title":"Conditional adversarial domain adaptation","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"31","author":"Long","year":"2018"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00712"},{"key":"ref38","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1991.3.1.79"},{"key":"ref39","article-title":"Outrageously large neural networks: The sparsely-gated mixture-of-experts layer","volume-title":"Proc. Int. Conf. Learn. Represent.","author":"Shazeer","year":"2017"},{"key":"ref40","article-title":"Mixture-of-experts meets instruction tuning: A winning combination for large language models","volume-title":"Int\u2019l Conf. Learn. Represent.","author":"Shen","year":"2024"},{"key":"ref41","doi-asserted-by":"publisher","DOI":"10.1109\/TNSRE.2023.3337802"},{"key":"ref42","first-page":"9358","article-title":"Graphmetro: Mitigating complex distribution shifts in gnns via mixture of aligned experts","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"37","author":"Wu","year":"2024"},{"key":"ref43","article-title":"Shiftaddvit: Mixture of multiplication primitives towards efficient vision transformer","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"36","author":"You","year":"2024"},{"key":"ref44","article-title":"Fast graph representation learning with PyTorch geometric,\u201d in Proc.","volume-title":"ICLR Workshop Represent. Learn. Graphs Manifolds","author":"Fey","year":"2019"},{"key":"ref45","doi-asserted-by":"publisher","DOI":"10.5555\/3495724.3496673"},{"key":"ref46","doi-asserted-by":"publisher","DOI":"10.1016\/0377-0427(87)90125-7"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01566"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-642-15561-1_16"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.572"},{"article-title":"\u201cVisda: The visual domain adaptation challenge,\u201d","year":"2017","author":"Peng","key":"ref50"},{"key":"ref51","first-page":"16282","article-title":"Universal domain adaptation through self supervision","volume-title":"Proc. Adv. Neural Inf. Process. Syst.","volume":"33","author":"Saito","year":"2020"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.00887"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"ref54","first-page":"10347","article-title":"Training data-efficient image transformers & distillation through attention","volume-title":"Proc. Int. Conf. Mach. Learn.","author":"Touvron","year":"2021"},{"key":"ref55","first-page":"4696","article-title":"When does label smoothing help?","volume-title":"Proc. Advances in Neural Inf. Process. Syst.","volume":"32","author":"M\u00fcller","year":"2019"},{"issue":"11","key":"ref56","first-page":"2579","article-title":"Visualizing data using t-SNE","volume":"9","author":"Van der Maaten","year":"2008","journal-title":"J. Mach. Learn. Res."},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1145\/3219819.3220007"},{"key":"ref58","article-title":"Long-tailed recognition by routing diverse distribution-aware experts","volume-title":"Int. Conf. Learn. Represent.","author":"Wang","year":"2021"}],"container-title":["IEEE Transactions on Artificial Intelligence"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/9078688\/11269951\/10964735.pdf?arnumber=10964735","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,28]],"date-time":"2025-11-28T06:49:22Z","timestamp":1764312562000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10964735\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12]]},"references-count":58,"journal-issue":{"issue":"12"},"URL":"https:\/\/doi.org\/10.1109\/tai.2025.3560590","relation":{},"ISSN":["2691-4581"],"issn-type":[{"type":"electronic","value":"2691-4581"}],"subject":[],"published":{"date-parts":[[2025,12]]}}}