{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T04:09:52Z","timestamp":1765339792929,"version":"3.46.0"},"publisher-location":"New York, NY, USA","reference-count":28,"publisher":"ACM","funder":[{"name":"Institute of Information & communications Technology Planning & Evaluation(IITP) grant funded by the Korea government(MSIT)","award":["No. RS-2020-II201336"],"award-info":[{"award-number":["No. RS-2020-II201336"]}]},{"name":"ITRC(Information Technology Research Center) grant funded by the Korea government(Ministry of Science and ICT)","award":["IITP-2025-RS-2024-00436936"],"award-info":[{"award-number":["IITP-2025-RS-2024-00436936"]}]},{"name":"National Research Foundation of Korea(NRF) grant funded by the Korea government(MSIT)","award":["RS-2025-00515481"],"award-info":[{"award-number":["RS-2025-00515481"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,10,27]]},"DOI":"10.1145\/3746027.3761986","type":"proceedings-article","created":{"date-parts":[[2025,10,25]],"date-time":"2025-10-25T07:26:51Z","timestamp":1761377211000},"page":"13730-13736","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["MIG-COW: Transferable Adversarial Attacks on Deepfake Detectors via Gradient Decomposition"],"prefix":"10.1145","author":[{"given":"Wonjune","family":"Seo","sequence":"first","affiliation":[{"name":"Ulsan National Institute of Science and Technology, Ulsan, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-7470-0255","authenticated-orcid":false,"given":"Joonhyuk","family":"Baek","sequence":"additional","affiliation":[{"name":"Ulsan National Institute of Science and Technology, Ulsan, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-3240-3442","authenticated-orcid":false,"given":"Yeseong","family":"Jung","sequence":"additional","affiliation":[{"name":"Ulsan National Institute of Science and Technology, Ulsan, Republic of Korea"}]},{"given":"Saerom","family":"Park","sequence":"additional","affiliation":[{"name":"Ulsan National Institute of Science and Technology, Ulsan, Republic of Korea"}]}],"member":"320","published-online":{"date-parts":[[2025,10,27]]},"reference":[{"key":"e_1_3_2_1_1_1","unstructured":"Stability AI. 2024. sd3.5: Inference?only tiny reference implementation of Stable Diffusion 3.5 and SD3. https:\/\/github.com\/Stability-AI\/sd3.5."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV48922.2021.01007"},{"key":"e_1_3_2_1_3_1","volume-title":"Attend-and-excite: Attention-based semantic guidance for text-to-image diffusion models. ACM transactions on Graphics (TOG)","author":"Chefer Hila","year":"2023","unstructured":"Hila Chefer, Yuval Alaluf, Yael Vinker, Lior Wolf, and Daniel Cohen-Or. 2023. Attend-and-excite: Attention-based semantic guidance for text-to-image diffusion models. ACM transactions on Graphics (TOG), Vol. 42, 4 (2023), 1-10."},{"key":"e_1_3_2_1_4_1","unstructured":"DeepAI. 2022. Text-to-Image API. https:\/\/deepai.org\/machine-learning-model\/text2img."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00925"},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00957"},{"key":"e_1_3_2_1_7_1","volume-title":"International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=YicbFdNTTy","author":"Dosovitskiy Alexey","year":"2021","unstructured":"Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations. https:\/\/openreview.net\/forum?id=YicbFdNTTy"},{"key":"e_1_3_2_1_8_1","unstructured":"Ian J Goodfellow Jonathon Shlens and Christian Szegedy. 2014. Explaining and harnessing adversarial examples."},{"key":"e_1_3_2_1_9_1","volume-title":"TESSER: Transfer-Enhancing Adversarial Attacks from Vision Transformers via Spectral and Semantic Regularization.","author":"Guesmi Amira","year":"2025","unstructured":"Amira Guesmi, Bassem Ouni, and Muhammad Shafique. 2025. TESSER: Transfer-Enhancing Adversarial Attacks from Vision Transformers via Spectral and Semantic Regularization."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_11_1","unstructured":"Hotpot.ai. 2021. Hotpot.ai: AI Image Generator & Creative Tools. https:\/\/hotpot.ai\/."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2017.243"},{"key":"e_1_3_2_1_13_1","unstructured":"Adobe Inc. 2023. Adobe Firefly. https:\/\/firefly.adobe.com\/."},{"key":"e_1_3_2_1_14_1","volume-title":"Alias-free generative adversarial networks. Advances in neural information processing systems","author":"Karras Tero","year":"2021","unstructured":"Tero Karras, Miika Aittala, Samuli Laine, Erik H\u00e4rk\u00f6nen, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. 2021. Alias-free generative adversarial networks. Advances in neural information processing systems, Vol. 34 (2021), 852-863."},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00453"},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"e_1_3_2_1_17_1","unstructured":"Black Forest Labs. 2024. flux: Official inference repo for FLUX.1 models. https:\/\/github.com\/black-forest-labs\/flux."},{"key":"e_1_3_2_1_18_1","volume-title":"Hunyuan-dit: A powerful multi-resolution diffusion transformer with fine-grained chinese understanding. arXiv preprint arXiv:2405.08748","author":"Li Zhimin","year":"2024","unstructured":"Zhimin Li, Jianwei Zhang, Qin Lin, Jiangfeng Xiong, Yanxin Long, Xinchi Deng, Yingfang Zhang, Xingchao Liu, Minbin Huang, Zedong Xiao, et al., 2024. Hunyuan-dit: A powerful multi-resolution diffusion transformer with fine-grained chinese understanding. arXiv preprint arXiv:2405.08748 (2024)."},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00427"},{"key":"e_1_3_2_1_20_1","unstructured":"Aleksander Madry Aleksandar Makelov Ludwig Schmidt Dimitris Tsipras and Adrian Vladu. 2017. Towards deep learning models resistant to adversarial attacks."},{"key":"e_1_3_2_1_21_1","unstructured":"prithivMLmods. [n.d.]. Deep-Fake-Detector-v2-Model."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v39i7.32722"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00009"},{"volume-title":"DFDC 1st Place Solution. https:\/\/github.com\/selimsef\/dfdc_deepfake_challenge","author":"Seferbekov Selim","key":"e_1_3_2_1_24_1","unstructured":"Selim Seferbekov. 2020. DFDC 1st Place Solution. https:\/\/github.com\/selimsef\/dfdc_deepfake_challenge."},{"key":"e_1_3_2_1_25_1","unstructured":"Freepik Company S.L. 2024. Freepik. https:\/\/www.freepik.com."},{"key":"e_1_3_2_1_26_1","first-page":"3319","volume-title":"International conference on machine learning. PMLR","author":"Sundararajan Mukund","year":"2017","unstructured":"Mukund Sundararajan, Ankur Taly, and Qiqi Yan. 2017. Axiomatic attribution for deep networks. In International conference on machine learning. PMLR, Sydney, 3319-3328."},{"key":"e_1_3_2_1_27_1","volume-title":"Sana: Efficient high-resolution image synthesis with linear diffusion transformers. arXiv preprint arXiv:2410.10629","author":"Xie Enze","year":"2024","unstructured":"Enze Xie, Junsong Chen, Junyu Chen, Han Cai, Haotian Tang, Yujun Lin, Zhekai Zhang, Muyang Li, Ligeng Zhu, Yao Lu, et al., 2024. Sana: Efficient high-resolution image synthesis with linear diffusion transformers. arXiv preprint arXiv:2410.10629 (2024)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-031-20065-6_23"}],"event":{"name":"MM '25: The 33rd ACM International Conference on Multimedia","sponsor":["SIGMM ACM Special Interest Group on Multimedia"],"location":"Dublin Ireland","acronym":"MM '25"},"container-title":["Proceedings of the 33rd ACM International Conference on Multimedia"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3746027.3761986","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T04:06:51Z","timestamp":1765339611000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3746027.3761986"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,27]]},"references-count":28,"alternative-id":["10.1145\/3746027.3761986","10.1145\/3746027"],"URL":"https:\/\/doi.org\/10.1145\/3746027.3761986","relation":{},"subject":[],"published":{"date-parts":[[2025,10,27]]},"assertion":[{"value":"2025-10-27","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}