{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,29]],"date-time":"2025-12-29T18:47:01Z","timestamp":1767034021650,"version":"3.44.0"},"publisher-location":"New York, NY, USA","reference-count":18,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,5,30]],"date-time":"2024-05-30T00:00:00Z","timestamp":1717027200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,5,30]]},"DOI":"10.1145\/3652583.3657597","type":"proceedings-article","created":{"date-parts":[[2024,6,7]],"date-time":"2024-06-07T06:30:40Z","timestamp":1717741840000},"page":"1288-1293","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":4,"title":["A Generative Adaptive Context Learning Framework for Large Language Models in Cheapfake Detection"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-6197-3428","authenticated-orcid":false,"given":"Long-Khanh","family":"Pham","sequence":"first","affiliation":[{"name":"University of Science &amp; Vietnam National University, Ho Chi Minh City, Vietnam"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-0187-1777","authenticated-orcid":false,"given":"Hoa-Vien","family":"Vo-Hoang","sequence":"additional","affiliation":[{"name":"University of Science &amp; Vietnam National University, Ho Chi Minh City, Vietnam"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8036-954X","authenticated-orcid":false,"given":"Anh-Duy","family":"Tran","sequence":"additional","affiliation":[{"name":"imec-DistriNet, KU Leuven, Leuven, Belgium"}]}],"member":"320","published-online":{"date-parts":[[2024,6,7]]},"reference":[{"volume-title":"Retrieved","year":"2024","key":"e_1_3_2_1_1_1","unstructured":"2023. Types and origin of cheapfakes. Retrieved Mar 08, 2024 from https:\/\/antispoofing.org\/cheapfakes-definition-overview-and-potential-threats\/"},{"key":"e_1_3_2_1_2_1","volume-title":"Cosmos: Catching out-of-context misinformation with self-supervised learning. arXiv preprint arXiv:2101.06278","author":"Aneja Shivangi","year":"2021","unstructured":"Shivangi Aneja, Chris Bregler, and Matthias Nie\u00dfner. 2021. Cosmos: Catching out-of-context misinformation with self-supervised learning. arXiv preprint arXiv:2101.06278 (2021)."},{"key":"e_1_3_2_1_3_1","volume-title":"Michael Riegler, P\u00e5l Halvorsen, Chris Bregler, and Balu Adsumilli.","author":"Aneja Shivangi","year":"2022","unstructured":"Shivangi Aneja, Cise Midoglu, Duc-Tien Dang-Nguyen, Sohail Ahmed Khan, Michael Riegler, P\u00e5l Halvorsen, Chris Bregler, and Balu Adsumilli. 2022. Acm multimedia grand challenge on detecting cheapfakes. arXiv preprint arXiv:2207.14534 (2022)."},{"key":"e_1_3_2_1_4_1","volume-title":"Cise Midoglu, Michael Riegler, P\u00e5l Halvorsen, and Minh-Son Dao.","author":"Dang-Nguyen Duc-Tien","year":"2023","unstructured":"Duc-Tien Dang-Nguyen, Sohail Ahmed Khan, Cise Midoglu, Michael Riegler, P\u00e5l Halvorsen, and Minh-Son Dao. 2023. Grand Challenge On Detecting Cheapfakes. arXiv preprint arXiv:2304.01328 (2023)."},{"key":"e_1_3_2_1_5_1","doi-asserted-by":"publisher","DOI":"10.1145\/3652583.3657587"},{"key":"e_1_3_2_1_6_1","unstructured":"Peng Gao Jiaming Han Renrui Zhang Ziyi Lin Shijie Geng Aojun Zhou Wei Zhang Pan Lu Conghui He Xiangyu Yue Hongsheng Li and Yu Qiao. 2023. LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model. arXiv:2304.15010 [cs.CV]"},{"key":"e_1_3_2_1_7_1","volume-title":"Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing. arXiv preprint arXiv:2111.09543","author":"He Pengcheng","year":"2021","unstructured":"Pengcheng He, Jianfeng Gao, and Weizhu Chen. 2021. Debertav3: Improving deberta using electra-style pre-training with gradient-disentangled embedding sharing. arXiv preprint arXiv:2111.09543 (2021)."},{"key":"e_1_3_2_1_8_1","volume-title":"Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, et al.","author":"Jiang Albert Q","year":"2023","unstructured":"Albert Q Jiang, Alexandre Sablayrolles, Arthur Mensch, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Florian Bressand, Gianna Lengyel, Guillaume Lample, Lucile Saulnier, et al. 2023. Mistral 7B. arXiv preprint arXiv:2310.06825 (2023)."},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3551595"},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1145\/3512731.3534210"},{"key":"e_1_3_2_1_11_1","unstructured":"Haotian Liu Chunyuan Li QingyangWu and Yong Jae Lee. 2023. Visual Instruction Tuning. arXiv:2304.08485 [cs.CV]"},{"volume-title":"Natural language inference","author":"MacCartney Bill","key":"e_1_3_2_1_12_1","unstructured":"Bill MacCartney. 2009. Natural language inference. Stanford University."},{"key":"e_1_3_2_1_13_1","unstructured":"Britt Paris and Joan Donovan. 2019. Deepfakes and cheap fakes. (2019)."},{"key":"e_1_3_2_1_14_1","volume-title":"Retrieved","author":"Sample Ian","year":"2020","unstructured":"Ian Sample. 2020. What are deepfakes -- and how can you spot them? Retrieved Mar 08, 2024 from https:\/\/www.theguardian.com\/technology\/2020\/jan\/13\/whatare- deepfakes-and-how-can-you-spot-them"},{"key":"e_1_3_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1145\/3503161.3551596"},{"key":"e_1_3_2_1_16_1","volume-title":"Cheap-fake Detection with LLM using Prompt Engineering. arXiv preprint arXiv:2306.02776","author":"Wu Guangyang","year":"2023","unstructured":"Guangyang Wu, Weijie Wu, Xiaohong Liu, Kele Xu, Tianjiao Wan, and Wenyi Wang. 2023. Cheap-fake Detection with LLM using Prompt Engineering. arXiv preprint arXiv:2306.02776 (2023)."},{"key":"e_1_3_2_1_17_1","volume-title":"Retrieved","author":"You Hyosun","year":"2022","unstructured":"Hyosun You. 2022. What Are Cheapfakes. Retrieved Mar 08, 2024 from https:\/\/www.samsungsds.com\/en\/insights\/what-are-cheapfakes.html"},{"key":"e_1_3_2_1_18_1","unstructured":"Deyao Zhu Jun Chen Xiaoqian Shen Xiang Li and Mohamed Elhoseiny. 2023. MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models. arXiv:2304.10592 [cs.CV]"}],"event":{"name":"ICMR '24: International Conference on Multimedia Retrieval","sponsor":["SIGMM ACM Special Interest Group on Multimedia","SIGSOFT ACM Special Interest Group on Software Engineering"],"location":"Phuket Thailand","acronym":"ICMR '24"},"container-title":["Proceedings of the 2024 International Conference on Multimedia Retrieval"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3652583.3657597","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3652583.3657597","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,8,21]],"date-time":"2025-08-21T08:51:52Z","timestamp":1755766312000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3652583.3657597"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,30]]},"references-count":18,"alternative-id":["10.1145\/3652583.3657597","10.1145\/3652583"],"URL":"https:\/\/doi.org\/10.1145\/3652583.3657597","relation":{},"subject":[],"published":{"date-parts":[[2024,5,30]]},"assertion":[{"value":"2024-06-07","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}