{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,5]],"date-time":"2025-10-05T19:50:12Z","timestamp":1759693812623,"version":"3.41.0"},"reference-count":57,"publisher":"Association for Computing Machinery (ACM)","issue":"4","license":[{"start":{"date-parts":[[2023,3,15]],"date-time":"2023-03-15T00:00:00Z","timestamp":1678838400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"name":"National Key Research and Development Program of China","award":["2020AAA0107400"],"award-info":[{"award-number":["2020AAA0107400"]}]},{"name":"Zhejiang Provincial Natural Science Foundation of China","award":["LR19F020004"],"award-info":[{"award-number":["LR19F020004"]}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["U20A20222"],"award-info":[{"award-number":["U20A20222"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100014219","name":"National Science Foundation for Distinguished Young Scholars","doi-asserted-by":"crossref","award":["62225605"],"award-info":[{"award-number":["62225605"]}],"id":[{"id":"10.13039\/501100014219","id-type":"DOI","asserted-by":"crossref"}]},{"name":"Ant Group, and sponsored by CAAI-HUAWEI MindSpore Open Fund"}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Multimedia Comput. Commun. Appl."],"published-print":{"date-parts":[[2023,7,31]]},"abstract":"<jats:p>\n            As an important and challenging problem, image generation with limited data aims at generating realistic images through training a GAN model given few samples. A typical solution is to transfer a well-trained GAN model from a data-rich source domain to the data-deficient target domain. In this paper, we propose a novel self-supervised transfer scheme termed D\n            <jats:sup>3<\/jats:sup>\n            T-GAN, addressing the cross-domain GANs transfer in limited image generation. Specifically, we design two individual strategies to transfer knowledge between generators and discriminators, respectively. To transfer knowledge between generators, we conduct a data-dependent transformation, which projects target samples into the latent space of source generator and reconstructs them back. Then, we perform knowledge transfer from transformed samples to generated samples. To transfer knowledge between discriminators, we design a multi-level discriminant knowledge distillation from the source discriminator to the target discriminator on both the real and fake samples. Extensive experiments show that our method improves the quality of generated images and achieves the state-of-the-art FID scores on commonly used datasets.\n          <\/jats:p>","DOI":"10.1145\/3576858","type":"journal-article","created":{"date-parts":[[2023,2,6]],"date-time":"2023-02-06T09:53:02Z","timestamp":1675677182000},"page":"1-20","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":6,"title":["D\n            <sup>3<\/sup>\n            T-GAN: Data-Dependent Domain Transfer GANs for Image Generation with Limited Data"],"prefix":"10.1145","volume":"19","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2988-5215","authenticated-orcid":false,"given":"Xintian","family":"Wu","sequence":"first","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6941-4445","authenticated-orcid":false,"given":"Huanyu","family":"Wang","sequence":"additional","affiliation":[{"name":"Zhejiang University, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-9866-669X","authenticated-orcid":false,"given":"Yiming","family":"Wu","sequence":"additional","affiliation":[{"name":"Zhejiang University of Technology, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3023-1662","authenticated-orcid":false,"given":"Xi","family":"Li","sequence":"additional","affiliation":[{"name":"Zhejiang University, Shanghai Institute for Advanced Study of Zhejiang University, Zhejiang-Singapore Innovation and AI Joint Research Lab, Shanghai AI Lab College of Computer Science, China"}]}],"member":"320","published-online":{"date-parts":[[2023,3,15]]},"reference":[{"key":"e_1_3_1_2_2","first-page":"4432","volume-title":"Proc. ICCV","author":"Abdal Rameen","year":"2019","unstructured":"Rameen Abdal, Yipeng Qin, and Peter Wonka. 2019. Image2StyleGAN: How to embed images into the StyleGAN latent space?. In Proc. ICCV. 4432\u20134441."},{"key":"e_1_3_1_3_2","volume-title":"Proc. ICLR","author":"Antoniou Antreas","year":"2017","unstructured":"Antreas Antoniou, Amos Storkey, and Harrison Edwards. 2017. Data augmentation generative adversarial networks. In Proc. ICLR."},{"key":"e_1_3_1_4_2","first-page":"214","volume-title":"Proc. ICML","author":"Arjovsky Martin","year":"2017","unstructured":"Martin Arjovsky, Soumith Chintala, and L\u00e9on Bottou. 2017. Wasserstein generative adversarial networks. In Proc. ICML. PMLR, 214\u2013223."},{"key":"e_1_3_1_5_2","first-page":"670","volume-title":"Proc. AISTATS","author":"Bartunov Sergey","year":"2018","unstructured":"Sergey Bartunov and Dmitry Vetrov. 2018. Few-shot generative modelling with generative matching networks. In Proc. AISTATS. PMLR, 670\u2013678."},{"key":"e_1_3_1_6_2","volume-title":"Proc. ICLR","author":"Brock Andrew","year":"2019","unstructured":"Andrew Brock, Jeff Donahue, and Karen Simonyan. 2019. Large scale GAN training for high fidelity natural image synthesis. In Proc. ICLR."},{"key":"e_1_3_1_7_2","volume-title":"arXiv preprint arXiv:1901.02199","author":"Clou\u00e2tre Louis","year":"2019","unstructured":"Louis Clou\u00e2tre and Marc Demers. 2019. FIGR: Few-shot image generation with reptile. In arXiv preprint arXiv:1901.02199."},{"key":"e_1_3_1_8_2","first-page":"658","volume-title":"Proc. NeurIPS","author":"Dosovitskiy Alexey","year":"2016","unstructured":"Alexey Dosovitskiy and Thomas Brox. 2016. Generating images with perceptual similarity metrics based on deep networks. In Proc. NeurIPS, Vol. 29. 658\u2013666."},{"key":"e_1_3_1_9_2","volume-title":"arXiv preprint arXiv:2104.00365","author":"Fan Chenyou","year":"2021","unstructured":"Chenyou Fan and Jianwei Huang. 2021. Federated few-shot learning with adversarial learning. In arXiv preprint arXiv:2104.00365."},{"key":"e_1_3_1_10_2","first-page":"767","volume-title":"IEEE Trans. Image Process.","author":"Gao Zan","year":"2020","unstructured":"Zan Gao, Leming Guo, Weili Guan, An-An Liu, Tongwei Ren, and Shengyong Chen. 2020. A pairwise attentive adversarial spatiotemporal network for cross-domain few-shot action recognition-R2. In IEEE Trans. Image Process., Vol. 30. IEEE, 767\u2013782."},{"key":"e_1_3_1_11_2","volume-title":"Proc. NeurIPS","author":"Goodfellow Ian J.","year":"2014","unstructured":"Ian J. Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. 2014. Generative adversarial networks. In Proc. NeurIPS."},{"key":"e_1_3_1_12_2","first-page":"3012","volume-title":"Proc. CVPR","author":"Gu Jinjin","year":"2020","unstructured":"Jinjin Gu, Yujun Shen, and Bolei Zhou. 2020. Image processing using multi-code GAN prior. In Proc. CVPR. 3012\u20133021."},{"key":"e_1_3_1_13_2","volume-title":"Proc. NeurIPS","author":"Gulrajani Ishaan","year":"2017","unstructured":"Ishaan Gulrajani, Faruk Ahmed, Martin Arjovsky, Vincent Dumoulin, and Aaron Courville. 2017. Improved training of Wasserstein GANs. In Proc. NeurIPS."},{"key":"e_1_3_1_14_2","first-page":"1","volume-title":"Global Oceans","author":"Guo Zonghui","year":"2020","unstructured":"Zonghui Guo, Liqiang Zhang, Yufeng Jiang, Wenjie Niu, Zhaorui Gu, Haiyong Zheng, Guoyu Wang, and Bing Zheng. 2020. Few-shot fish image generation and classification. In Global Oceans. IEEE, 1\u20136."},{"key":"e_1_3_1_15_2","first-page":"1","volume-title":"ACM Trans. Multimedia Comput. Commun. Appl.","author":"Ho Trang-Thi","year":"2020","unstructured":"Trang-Thi Ho, John Jethro Virtusio, Yung-Yao Chen, Chih-Ming Hsu, and Kai-Lung Hua. 2020. Sketch-guided deep portrait generation. In ACM Trans. Multimedia Comput. Commun. Appl., Vol. 16. ACM New York, NY, USA, 1\u201318."},{"key":"e_1_3_1_16_2","first-page":"1","volume-title":"Proc. IEEE ICME","author":"Hong Yan","year":"2020","unstructured":"Yan Hong, Li Niu, Jianfu Zhang, and Liqing Zhang. 2020. MatchingGAN: Matching-based few-shot image generation. In Proc. IEEE ICME. IEEE, 1\u20136."},{"key":"e_1_3_1_17_2","first-page":"2535","volume-title":"Proc. ACM MM","author":"Hong Yan","year":"2020","unstructured":"Yan Hong, Li Niu, Jianfu Zhang, Weijie Zhao, Chen Fu, and Liqing Zhang. 2020. F2GAN: Fusing-and-filling GAN for few-shot image generation. In Proc. ACM MM. 2535\u20132543."},{"key":"e_1_3_1_18_2","first-page":"1501","volume-title":"Proc. ICCV","author":"Huang Xun","year":"2017","unstructured":"Xun Huang and Serge Belongie. 2017. Arbitrary style transfer in real-time with adaptive instance normalization. In Proc. ICCV. 1501\u20131510."},{"key":"e_1_3_1_19_2","first-page":"1","volume-title":"ACM Trans. Multimedia Comput. Commun. Appl.","author":"Jiang Shuqiang","year":"2020","unstructured":"Shuqiang Jiang, Weiqing Min, Yongqiang Lyu, and Linhu Liu. 2020. Few-shot food recognition via multi-view representation learning. In ACM Trans. Multimedia Comput. Commun. Appl., Vol. 16. ACM New York, NY, USA, 1\u201320."},{"key":"e_1_3_1_20_2","volume-title":"Proc. ICLR","author":"Karras Tero","year":"2018","unstructured":"Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen. 2018. Progressive growing of GANs for improved quality, stability, and variation. In Proc. ICLR."},{"key":"e_1_3_1_21_2","volume-title":"Proc. NeurIPS","author":"Karras Tero","year":"2020","unstructured":"Tero Karras, Miika Aittala, Janne Hellsten, Samuli Laine, Jaakko Lehtinen, and Timo Aila. 2020. Training generative adversarial networks with limited data. In Proc. NeurIPS."},{"key":"e_1_3_1_22_2","first-page":"4401","volume-title":"Proc. CVPR","author":"Karras Tero","year":"2019","unstructured":"Tero Karras, Samuli Laine, and Timo Aila. 2019. A style-based generator architecture for generative adversarial networks. In Proc. CVPR. 4401\u20134410."},{"key":"e_1_3_1_23_2","first-page":"8110","volume-title":"Proc. CVPR","author":"Karras Tero","year":"2020","unstructured":"Tero Karras, Samuli Laine, Miika Aittala, Janne Hellsten, Jaakko Lehtinen, and Timo Aila. 2020. Analyzing and improving the image quality of StyleGAN. In Proc. CVPR. 8110\u20138119."},{"key":"e_1_3_1_24_2","volume-title":"Proc. ICLR","author":"Kingma Diederik P.","year":"2015","unstructured":"Diederik P. Kingma and Jimmy Ba. 2015. Adam: A method for stochastic optimization. In Proc. ICLR."},{"key":"e_1_3_1_25_2","first-page":"5881","volume-title":"IEEE Trans. Image Process.","author":"Li Yu","year":"2019","unstructured":"Yu Li, Sheng Tang, Rui Zhang, Yongdong Zhang, Jintao Li, and Shuicheng Yan. 2019. Asymmetric GAN for unpaired image-to-image translation. In IEEE Trans. Image Process., Vol. 28. IEEE, 5881\u20135896."},{"key":"e_1_3_1_26_2","volume-title":"Proc. NeurIPS","author":"Li Yijun","year":"2020","unstructured":"Yijun Li, Richard Zhang, Jingwan Lu, and Eli Shechtman. 2020. Few-shot image generation with elastic weight consolidation. In Proc. NeurIPS."},{"key":"e_1_3_1_27_2","volume-title":"arXiv preprint arXiv:2001.00576","author":"Liang Weixin","year":"2020","unstructured":"Weixin Liang, Zixuan Liu, and Can Liu. 2020. DAWSON: A domain adaptive few shot generation framework. In arXiv preprint arXiv:2001.00576."},{"key":"e_1_3_1_28_2","volume-title":"arXiv preprint arXiv:2012.04256","author":"Mangla Puneet","year":"2020","unstructured":"Puneet Mangla, Nupur Kumari, Mayank Singh, Vineeth N. Balasubramanian, and Balaji Krishnamurthy. 2020. Data instance prior for transfer learning in GANs. In arXiv preprint arXiv:2012.04256."},{"key":"e_1_3_1_29_2","first-page":"3481","volume-title":"Proc. ICML","author":"Mescheder Lars","year":"2018","unstructured":"Lars Mescheder, Andreas Geiger, and Sebastian Nowozin. 2018. Which training methods for GANs do actually converge?. In Proc. ICML. PMLR, 3481\u20133490."},{"key":"e_1_3_1_30_2","volume-title":"Proc. ICLR","author":"Miyato Takeru","year":"2018","unstructured":"Takeru Miyato, Toshiki Kataoka, Masanori Koyama, and Yuichi Yoshida. 2018. Spectral normalization for generative adversarial networks. In Proc. ICLR."},{"key":"e_1_3_1_31_2","volume-title":"Proc. CVPR Workshop","author":"Mo Sangwoo","year":"2020","unstructured":"Sangwoo Mo, Minsu Cho, and Jinwoo Shin. 2020. Freeze discriminator: A simple baseline for fine-tuning GANs. In Proc. CVPR Workshop."},{"key":"e_1_3_1_32_2","volume-title":"arXiv preprint arXiv:1803.02999","author":"Nichol Alex","year":"2018","unstructured":"Alex Nichol, Joshua Achiam, and John Schulman. 2018. On first-order meta-learning algorithms. In arXiv preprint arXiv:1803.02999."},{"key":"e_1_3_1_33_2","first-page":"2750","volume-title":"Proc. ICCV","author":"Noguchi Atsuhiro","year":"2019","unstructured":"Atsuhiro Noguchi and Tatsuya Harada. 2019. Image generation from small datasets via batch statistics adaptation. In Proc. ICCV. 2750\u20132758."},{"key":"e_1_3_1_34_2","volume-title":"Proc. NeurIPS","author":"Nowozin Sebastian","year":"2016","unstructured":"Sebastian Nowozin, Botond Cseke, and Ryota Tomioka. 2016. F-GAN: Training generative neural samplers using variational divergence minimization. In Proc. NeurIPS."},{"key":"e_1_3_1_35_2","volume-title":"IEEE Trans. Multimedia","author":"Phaphuangwittayakul Aniwat","year":"2021","unstructured":"Aniwat Phaphuangwittayakul, Yi Guo, and Fangli Ying. 2021. Fast adaptive meta-learning for few-shot image generation. In IEEE Trans. Multimedia. IEEE."},{"key":"e_1_3_1_36_2","volume-title":"Proc. ICLR","author":"Radford Alec","year":"2016","unstructured":"Alec Radford, Luke Metz, and Soumith Chintala. 2016. Unsupervised representation learning with deep convolutional generative adversarial networks. In Proc. ICLR."},{"key":"e_1_3_1_37_2","volume-title":"arXiv preprint arXiv:2010.11943","author":"Robb Esther","year":"2020","unstructured":"Esther Robb, Wen-Sheng Chu, Abhishek Kumar, and Jia-Bin Huang. 2020. Few-shot adaptation of generative adversarial networks. In arXiv preprint arXiv:2010.11943."},{"key":"e_1_3_1_38_2","first-page":"1354","volume-title":"IEEE Trans. Pattern Anal. Mach. Intell.","author":"Si Zhangzhang","year":"2011","unstructured":"Zhangzhang Si and Song-Chun Zhu. 2011. Learning hybrid image templates (HIT) by information projection. In IEEE Trans. Pattern Anal. Mach. Intell., Vol. 34. IEEE, 1354\u20131367."},{"key":"e_1_3_1_39_2","first-page":"2818","volume-title":"Proc. CVPR","author":"Szegedy Christian","year":"2016","unstructured":"Christian Szegedy, Vincent Vanhoucke, Sergey Ioffe, Jon Shlens, and Zbigniew Wojna. 2016. Rethinking the inception architecture for computer vision. In Proc. CVPR. 2818\u20132826."},{"key":"e_1_3_1_40_2","first-page":"8916","volume-title":"IEEE Trans. Image Process.","author":"Tang Hao","year":"2020","unstructured":"Hao Tang, Hong Liu, and Nicu Sebe. 2020. Unified generative adversarial networks for controllable image-to-image translation. In IEEE Trans. Image Process., Vol. 29. IEEE, 8916\u20138929."},{"key":"e_1_3_1_41_2","first-page":"1882","volume-title":"IEEE Trans. Image Process.","author":"Tran Ngoc-Trung","year":"2021","unstructured":"Ngoc-Trung Tran, Viet-Hung Tran, Ngoc-Bao Nguyen, Trung-Kien Nguyen, and Ngai-Man Cheung. 2021. On data augmentation for GAN training. In IEEE Trans. Image Process.IEEE, 1882\u20131897."},{"key":"e_1_3_1_42_2","volume-title":"Proc. NeurIPS","author":"Wang Ting-Chun","year":"2019","unstructured":"Ting-Chun Wang, Ming-Yu Liu, Andrew Tao, Guilin Liu, Jan Kautz, and Bryan Catanzaro. 2019. Few-shot video-to-video synthesis. In Proc. NeurIPS."},{"key":"e_1_3_1_43_2","first-page":"1","volume-title":"ACM Trans. Multimedia Comput. Commun. Appl.","author":"Wang Xueping","year":"2019","unstructured":"Xueping Wang, Yunhong Wang, and Weixin Li. 2019. U-Net conditional GANs for photo-realistic and identity-preserving facial expression synthesis. In ACM Trans. Multimedia Comput. Commun. Appl., Vol. 15. ACM New York, NY, USA, 1\u201323."},{"key":"e_1_3_1_44_2","first-page":"9332","volume-title":"Proc. CVPR","author":"Wang Yaxing","year":"2020","unstructured":"Yaxing Wang, Abel Gonzalez-Garcia, David Berga, Luis Herranz, Fahad Shahbaz Khan, and Joost van de Weijer. 2020. MineGAN: Effective knowledge transfer from GANs to target domains with few images. In Proc. CVPR. 9332\u20139341."},{"key":"e_1_3_1_45_2","first-page":"218","volume-title":"Proc. ECCV","author":"Wang Yaxing","year":"2018","unstructured":"Yaxing Wang, Chenshen Wu, Luis Herranz, Joost van de Weijer, Abel Gonzalez-Garcia, and Bogdan Raducanu. 2018. Transferring GANs: Generating images from limited data. In Proc. ECCV. 218\u2013234."},{"key":"e_1_3_1_46_2","volume-title":"arXiv preprint arXiv:1904.05046","author":"Wang Yaqing","year":"2019","unstructured":"Yaqing Wang and Quanming Yao. 2019. Few-shot learning: A survey. In arXiv preprint arXiv:1904.05046."},{"key":"e_1_3_1_47_2","volume-title":"arXiv preprint arXiv:2106.04566","author":"Yang Ceyuan","year":"2021","unstructured":"Ceyuan Yang, Yujun Shen, Yinghao Xu, and Bolei Zhou. 2021. Data-efficient instance generation from instance discrimination. In arXiv preprint arXiv:2106.04566."},{"key":"e_1_3_1_48_2","volume-title":"arXiv preprint arXiv:1506.03365","author":"Yu Fisher","year":"2015","unstructured":"Fisher Yu, Yinda Zhang, Shuran Song, Ari Seff, and Jianxiong Xiao. 2015. LSUN: Construction of a large-scale image dataset using deep learning with humans in the loop. In arXiv preprint arXiv:1506.03365."},{"key":"e_1_3_1_49_2","first-page":"7354","volume-title":"Proc. ICML","author":"Zhang Han","year":"2019","unstructured":"Han Zhang, Ian Goodfellow, Dimitris Metaxas, and Augustus Odena. 2019. Self-attention generative adversarial networks. In Proc. ICML. PMLR, 7354\u20137363."},{"key":"e_1_3_1_50_2","volume-title":"Proc. ICLR","author":"Zhang Han","year":"2019","unstructured":"Han Zhang, Zizhao Zhang, Augustus Odena, and Honglak Lee. 2019. Consistency regularization for generative adversarial networks. In Proc. ICLR."},{"key":"e_1_3_1_51_2","volume-title":"Proc. ICLR","author":"Zhao Junbo","year":"2017","unstructured":"Junbo Zhao, Michael Mathieu, and Yann LeCun. 2017. Energy-based generative adversarial network. In Proc. ICLR."},{"key":"e_1_3_1_52_2","first-page":"11340","volume-title":"Proc. ICML","author":"Zhao Miaoyun","year":"2020","unstructured":"Miaoyun Zhao, Yulai Cong, and Lawrence Carin. 2020. On leveraging pretrained GANs for generation with limited data. In Proc. ICML. PMLR, 11340\u201311351."},{"key":"e_1_3_1_53_2","volume-title":"Proc. NeurIPS","author":"Zhao Shengyu","year":"2020","unstructured":"Shengyu Zhao, Zhijian Liu, Ji Lin, Jun-Yan Zhu, and Song Han. 2020. Differentiable augmentation for data-efficient GAN training. In Proc. NeurIPS."},{"key":"e_1_3_1_54_2","volume-title":"Proc. AAAI","author":"Zhao Zhengli","year":"2021","unstructured":"Zhengli Zhao, Sameer Singh, Honglak Lee, Zizhao Zhang, Augustus Odena, and Han Zhang. 2021. Improved consistency regularization for GANs. In Proc. AAAI."},{"key":"e_1_3_1_55_2","volume-title":"arXiv preprint arXiv:2006.02595","author":"Zhao Zhengli","year":"2020","unstructured":"Zhengli Zhao, Zizhao Zhang, Ting Chen, Sameer Singh, and Han Zhang. 2020. Image augmentations for GAN training. In arXiv preprint arXiv:2006.02595."},{"key":"e_1_3_1_56_2","first-page":"592","volume-title":"Proc. ECCV","author":"Zhu Jiapeng","year":"2020","unstructured":"Jiapeng Zhu, Yujun Shen, Deli Zhao, and Bolei Zhou. 2020. In-domain GAN inversion for real image editing. In Proc. ECCV. Springer, 592\u2013608."},{"key":"e_1_3_1_57_2","first-page":"597","volume-title":"Proc. ECCV","author":"Zhu Jun-Yan","year":"2016","unstructured":"Jun-Yan Zhu, Philipp Kr\u00e4henb\u00fchl, Eli Shechtman, and Alexei A. Efros. 2016. Generative visual manipulation on the natural image manifold. In Proc. ECCV. Springer, 597\u2013613."},{"key":"e_1_3_1_58_2","first-page":"4253","volume-title":"IEEE Trans. Image Process.","author":"Zhu Linchao","year":"2021","unstructured":"Linchao Zhu, Hehe Fan, Yawei Luo, Mingliang Xu, and Yi Yang. 2021. Few-shot common-object reasoning using common-centric localization network. In IEEE Trans. Image Process., Vol. 30. IEEE, 4253\u20134262."}],"container-title":["ACM Transactions on Multimedia Computing, Communications, and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3576858","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3576858","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T18:08:59Z","timestamp":1750183739000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3576858"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,15]]},"references-count":57,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2023,7,31]]}},"alternative-id":["10.1145\/3576858"],"URL":"https:\/\/doi.org\/10.1145\/3576858","relation":{},"ISSN":["1551-6857","1551-6865"],"issn-type":[{"type":"print","value":"1551-6857"},{"type":"electronic","value":"1551-6865"}],"subject":[],"published":{"date-parts":[[2023,3,15]]},"assertion":[{"value":"2022-06-25","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2022-11-29","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-03-15","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}