{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T18:06:28Z","timestamp":1773079588236,"version":"3.50.1"},"reference-count":60,"publisher":"IEEE","license":[{"start":{"date-parts":[[2024,9,15]],"date-time":"2024-09-15T00:00:00Z","timestamp":1726358400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2024,9,15]],"date-time":"2024-09-15T00:00:00Z","timestamp":1726358400000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"}],"funder":[{"DOI":"10.13039\/100006190","name":"Research and Development","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100006190","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100002858","name":"China Postdoctoral Science Foundation","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100002858","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100003009","name":"Science and Technology Development Fund","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100003009","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":[],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024,9,15]]},"DOI":"10.1109\/ijcb62174.2024.10744523","type":"proceedings-article","created":{"date-parts":[[2024,11,11]],"date-time":"2024-11-11T18:38:05Z","timestamp":1731350285000},"page":"1-11","source":"Crossref","is-referenced-by-count":9,"title":["La-SoftMoE CLIP for Unified Physical-Digital Face Attack Detection"],"prefix":"10.1109","author":[{"given":"Hang","family":"Zou","sequence":"first","affiliation":[{"name":"China Telecom Research Institute (CTRI)"}]},{"given":"Chenxi","family":"Du","sequence":"additional","affiliation":[{"name":"Chinese Academy of Sciences,SIAT"}]},{"given":"Hui","family":"Zhang","sequence":"additional","affiliation":[{"name":"Tianjin University of Science &#x0026; Technology"}]},{"given":"Yuan","family":"Zhang","sequence":"additional","affiliation":[{"name":"China Telecom Research Institute (CTRI)"}]},{"given":"Ajian","family":"Liu","sequence":"additional","affiliation":[{"name":"MAIS, CASIA,China"}]},{"given":"Jun","family":"Wan","sequence":"additional","affiliation":[{"name":"MAIS, CASIA,China"}]},{"given":"Zhen","family":"Lei","sequence":"additional","affiliation":[{"name":"MAIS, CASIA,China"}]}],"member":"263","reference":[{"key":"ref1","first-page":"1597","article-title":"A simple framework for contrastive learning of visual representations","volume-title":"International conference on machine learning","author":"Chen"},{"key":"ref2","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.3009287"},{"key":"ref3","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00582"},{"key":"ref4","doi-asserted-by":"publisher","DOI":"10.1109\/FG57933.2023.10042500"},{"key":"ref5","article-title":"The deepfake detection challenge (dfdc) preview dataset","author":"Dolhansky","year":"2019"},{"key":"ref6","article-title":"An image is worth 16x16 words: Transformers for image recognition at scale","author":"Dosovitskiy","year":"2020"},{"key":"ref7","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2014.2322255"},{"key":"ref8","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW59228.2023.00677"},{"key":"ref9","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2023.3337970"},{"key":"ref10","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2024\/83"},{"key":"ref11","article-title":"A review of sparse expert models in deep learning","author":"Fedus","year":"2022"},{"key":"ref12","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"ref13","doi-asserted-by":"publisher","DOI":"10.1162\/neco.1991.3.1.79"},{"key":"ref14","article-title":"Transformers are rnns: Fast autoregressive transformers with linear attention","volume-title":"Proceedings of the International Conference on Machine Learning (ICML)","author":"Katharopoulos"},{"key":"ref15","article-title":"Deepfakes: a new threat to face recognition? assessment and detection","author":"Korshunov","year":"2018"},{"key":"ref16","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"ref17","article-title":"Gshard: Scaling giant models with conditional computation and automatic sharding","author":"Lepikhin","year":"2020"},{"key":"ref18","article-title":"Mimic-it: Multi-modal in-context instruction tuning","author":"Li","year":"2023"},{"key":"ref19","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2018.2801312"},{"key":"ref20","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00327"},{"key":"ref21","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-024-02135-2"},{"key":"ref22","doi-asserted-by":"publisher","DOI":"10.1049\/bme2.12002"},{"key":"ref23","doi-asserted-by":"publisher","DOI":"10.24963\/ijcai.2022\/165"},{"key":"ref24","doi-asserted-by":"publisher","DOI":"10.1109\/WACV48630.2021.00122"},{"key":"ref25","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2021.3065495"},{"key":"ref26","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2023.3296330"},{"key":"ref27","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2019.00202"},{"key":"ref28","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.00029"},{"key":"ref29","doi-asserted-by":"publisher","DOI":"10.1109\/ICCVW54120.2021.00096"},{"key":"ref30","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2022.3188149"},{"key":"ref31","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-46478-7_6"},{"key":"ref32","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00048"},{"key":"ref33","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00048"},{"key":"ref34","article-title":"From sparse to soft mixtures of experts","author":"Puigcerver","year":"2023"},{"key":"ref35","first-page":"8748","article-title":"Learning transferable visual models from natural language supervision","volume-title":"International conference on machine learning","author":"Radford"},{"key":"ref36","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00009"},{"key":"ref37","doi-asserted-by":"publisher","DOI":"10.1109\/ICASSP43922.2022.9746477"},{"key":"ref38","article-title":"Outrageously large neural networks: The sparsely-gated mixture-of-experts layer","author":"Shazeer","year":"2017"},{"key":"ref39","doi-asserted-by":"publisher","DOI":"10.1109\/wacv48630.2021.00357"},{"key":"ref40","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.01803"},{"issue":"11","key":"ref41","article-title":"Visualizing data using t-sne","volume":"9","author":"Van der Maaten","year":"2008","journal-title":"Journal of machine learning research"},{"key":"ref42","doi-asserted-by":"publisher","DOI":"10.48550\/ARXIV.1706.03762"},{"key":"ref43","article-title":"Linformer: Self-attention with linear complexity","author":"Wang","year":"2020"},{"key":"ref44","doi-asserted-by":"publisher","DOI":"10.1109\/TIFS.2015.2400395"},{"key":"ref45","article-title":"A survey on multimodal large language models","author":"Yin","year":"2023"},{"key":"ref46","article-title":"Visual prompt flexible-modal face anti-spoofing","author":"Yu","year":"2023"},{"key":"ref47","doi-asserted-by":"publisher","DOI":"10.1109\/TDSC.2024.3352049"},{"key":"ref48","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-030-58571-6_33"},{"key":"ref49","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00024"},{"key":"ref50","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2022.3215850"},{"key":"ref51","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2020.3036338"},{"key":"ref52","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00534"},{"key":"ref53","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW63382.2024.00097"},{"key":"ref54","doi-asserted-by":"publisher","DOI":"10.1109\/TBIOM.2020.2973001"},{"key":"ref55","doi-asserted-by":"publisher","DOI":"10.18653\/v1\/2023.findings-emnlp.734"},{"key":"ref56","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.00222"},{"key":"ref57","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01631"},{"key":"ref58","doi-asserted-by":"publisher","DOI":"10.1007\/s11263-022-01653-1"},{"key":"ref59","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2017.229"},{"key":"ref60","doi-asserted-by":"publisher","DOI":"10.1109\/ipdpsw55747.2022.00171"}],"event":{"name":"2024 IEEE International Joint Conference on Biometrics (IJCB)","location":"Buffalo, NY, USA","start":{"date-parts":[[2024,9,15]]},"end":{"date-parts":[[2024,9,18]]}},"container-title":["2024 IEEE International Joint Conference on Biometrics (IJCB)"],"original-title":[],"link":[{"URL":"http:\/\/xplorestaging.ieee.org\/ielx8\/10744408\/10744424\/10744523.pdf?arnumber=10744523","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,27]],"date-time":"2024-11-27T10:12:51Z","timestamp":1732702371000},"score":1,"resource":{"primary":{"URL":"https:\/\/ieeexplore.ieee.org\/document\/10744523\/"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,15]]},"references-count":60,"URL":"https:\/\/doi.org\/10.1109\/ijcb62174.2024.10744523","relation":{},"subject":[],"published":{"date-parts":[[2024,9,15]]}}}