{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T02:40:25Z","timestamp":1760668825026,"version":"build-2065373602"},"reference-count":83,"publisher":"Association for Computing Machinery (ACM)","issue":"7","content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["Proc. ACM Hum.-Comput. Interact."],"published-print":{"date-parts":[[2025,10,18]]},"abstract":"<jats:p>Photorealistic avatars are human avatars that look, move, and talk like real people. The performance of photorealistic avatars has significantly improved recently based on objective metrics such as PSNR, SSIM, LPIPS, FID, and FVD. However, recent photorealistic avatar publications do not provide subjective tests of the avatars to measure human usability factors. We provide an open source test framework to subjectively measure photorealistic avatar performance in ten dimensions: realism, trust, comfortableness using, comfortableness interacting with, appropriateness for work, creepiness, formality, affinity, resemblance to the person, and emotion accuracy. Using telecommunication scenarios, we show that the correlation of nine of these subjective metrics with PSNR, SSIM, LPIPS, FID, and FVD is weak, and moderate for emotion accuracy. The crowdsourced subjective test framework is highly reproducible and accurate when compared to a panel of experts. We analyze a wide range of avatars from photorealistic to cartoon-like and show that some photorealistic avatars are approaching real video performance based on these dimensions. We also find that for avatars above a certain level of realism, eight of these measured dimensions are strongly correlated. This means that avatars that are not as realistic as real video will have lower trust, comfortableness using, comfortableness interacting with, appropriateness for work, formality, and affinity, and higher creepiness compared to real video. In addition, because there is a strong linear relationship between avatar affinity and realism, there is no uncanny valley effect for photorealistic avatars in the telecommunication scenario. We suggest several extensions of this test framework for future work and discuss design implications for telecommunication systems. The test framework is available at https:\/\/github.com\/microsoft\/P.910.<\/jats:p>","DOI":"10.1145\/3757542","type":"journal-article","created":{"date-parts":[[2025,10,16]],"date-time":"2025-10-16T17:06:01Z","timestamp":1760634361000},"page":"1-29","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["A Multidimensional Measurement of Photorealistic Avatars Quality of Experience"],"prefix":"10.1145","volume":"9","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2004-3003","authenticated-orcid":false,"given":"Ross","family":"Cutler","sequence":"first","affiliation":[{"name":"Microsoft, Redmond, WA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0009-0006-4778-5417","authenticated-orcid":false,"given":"Babak","family":"Naderi","sequence":"additional","affiliation":[{"name":"Microsoft, Berlin, Germany"}]},{"ORCID":"https:\/\/orcid.org\/0009-0005-3887-1439","authenticated-orcid":false,"given":"Vishak","family":"Gopal","sequence":"additional","affiliation":[{"name":"Microsoft, Redmond, WA, USA"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2002-8409","authenticated-orcid":false,"given":"Dharmendar","family":"Palle","sequence":"additional","affiliation":[{"name":"Microsoft, Redmond, WA, USA"}]}],"member":"320","published-online":{"date-parts":[[2025,10,16]]},"reference":[{"key":"e_1_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1037\/tmb0000030"},{"key":"e_1_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.1145\/2675133.2675220"},{"key":"e_1_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/VR58804.2024.00063"},{"key":"e_1_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/3528223.3530143"},{"key":"e_1_2_1_5_1","volume-title":"Bovik","author":"Chen Yu-Chih","year":"2024","unstructured":"Yu-Chih Chen, Avinab Saha, Alexandre Chapiro, Christian H\u00e4ne, Jean-Charles Bazin, Bo Qiu, Stefano Zanetti, Ioannis Katsavounidis, and Alan C. Bovik. 2024. Subjective and Objective Quality Assessment of Rendered Human Avatar Videos in Virtual Reality. http:\/\/arxiv.org\/abs\/2408.07041 arXiv:2408.07041 [eess]."},{"key":"e_1_2_1_6_1","volume-title":"The Ishihara test for color blindness. American Journal of Physiological Optics","author":"Clark JH","year":"1924","unstructured":"JH Clark. 1924. The Ishihara test for color blindness. American Journal of Physiological Optics (1924)."},{"key":"e_1_2_1_7_1","volume-title":"Influence functions of the Spearman and Kendall correlation measures. Statistical methods & applications","author":"Croux Christophe","year":"2010","unstructured":"Christophe Croux and Catherine Dehon. 2010. Influence functions of the Spearman and Kendall correlation measures. Statistical methods & applications, Vol. 19 (2010), 497-515."},{"key":"e_1_2_1_8_1","doi-asserted-by":"crossref","unstructured":"Yu Deng Duomin Wang Xiaohang Ren Xingyu Chen and Baoyuan Wang. 2024. Portrait4D: Learning One-Shot 4D Head Avatar Synthesis using Synthetic Data. In CVPR.","DOI":"10.1109\/CVPR52733.2024.00680"},{"key":"e_1_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1145\/3470742"},{"key":"e_1_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.2139\/ssrn.3820035"},{"key":"e_1_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3562939.3565636"},{"key":"e_1_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1145\/3432938"},{"key":"e_1_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1016\/0301-0511(75)90023-X"},{"key":"e_1_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1145\/642611.642703"},{"key":"e_1_2_1_15_1","doi-asserted-by":"publisher","DOI":"10.1109\/AIxVR59861.2024.00022"},{"key":"e_1_2_1_16_1","unstructured":"R. Gonzalez and R. Woods. 2006. Digital image processing (3rd ed.). Prentice Hall."},{"key":"e_1_2_1_17_1","unstructured":"Jianzhu Guo Dingyun Zhang Xiaoqiang Liu Zhizhou Zhong Yuan Zhang Pengfei Wan and Di Zhang. 2024. LivePortrait: Efficient Portrait Animation with Stitching and Retargeting Control. http:\/\/arxiv.org\/abs\/2407.03168 arXiv:2407.03168."},{"key":"e_1_2_1_18_1","volume-title":"Advances in Neural Information Processing Systems","volume":"30","author":"Heusel Martin","year":"2017","unstructured":"Martin Heusel, Hubert Ramsauer, Thomas Unterthiner, Bernhard Nessler, and Sepp Hochreiter. 2017. GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. In Advances in Neural Information Processing Systems, Vol. 30. Curran Associates, Inc. https:\/\/proceedings.neurips.cc\/paper\/2017\/hash\/8a1d694707eb0fefe65871369074926d-Abstract.html"},{"key":"e_1_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1007\/s12369-016-0380-9"},{"key":"e_1_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2013.2291663"},{"key":"e_1_2_1_21_1","doi-asserted-by":"crossref","unstructured":"Ziyao Huang Fan Tang Yong Zhang Xiaodong Cun Juan Cao Jintao Li and Tong-Yee Lee. 2024. Make-Your-Anchor: A Diffusion-based 2D Avatar Generation Framework. In CVPR.","DOI":"10.1109\/CVPR52733.2024.00668"},{"key":"e_1_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1145\/1958824.1958883"},{"key":"e_1_2_1_23_1","first-page":"2017","article-title":"Ophthalmic optics - Visual acuity testing - Standard and clinical optotypes and their presentation","volume":"8596","author":"ISO","year":"2017","unstructured":"ISO Ophthalmic optics and instruments. 2017. Ophthalmic optics - Visual acuity testing - Standard and clinical optotypes and their presentation. Standard ISO 8596:2017. International Organization for Standardization, Geneva, CH. https:\/\/www.iso.org\/standard\/69042.html","journal-title":"Standard ISO"},{"volume-title":"Methodologies for the subjective assessment of the quality of television images","author":"ITU-R","key":"e_1_2_1_24_1","unstructured":"ITU-R Recommendation BT.500-14. 2019. Methodologies for the subjective assessment of the quality of television images. International Telecommunication Union, Geneva."},{"volume-title":"Subjective evaluation of media quality using a crowdsourcing approach","author":"ITU-T PSTR-CROWDS.","key":"e_1_2_1_25_1","unstructured":"ITU-T PSTR-CROWDS. 2018. Subjective evaluation of media quality using a crowdsourcing approach. International Telecommunication Union, Geneva."},{"volume-title":"Quality of experience assessment of extended reality meetings","author":"ITU-T","key":"e_1_2_1_26_1","unstructured":"ITU-T Recommendation P.1320. 2022. Quality of experience assessment of extended reality meetings. International Telecommunication Union, Geneva."},{"volume-title":"Subjective evaluation of speech quality with a crowdsourcing approach","author":"ITU-T","key":"e_1_2_1_27_1","unstructured":"ITU-T Recommendation P.808. 2021. Subjective evaluation of speech quality with a crowdsourcing approach. International Telecommunication Union, Geneva."},{"volume-title":"Subjective video quality assessment methods for multimedia applications","author":"ITU-T","key":"e_1_2_1_28_1","unstructured":"ITU-T Recommendation P.910. 2021. Subjective video quality assessment methods for multimedia applications. International Telecommunication Union, Geneva, Switzerland."},{"volume-title":"Subjective video quality assessment methods for multimedia applications","author":"ITU-T","key":"e_1_2_1_29_1","unstructured":"ITU-T Recommendation P.910. 2021. Subjective video quality assessment methods for multimedia applications. International Telecommunication Union, Geneva."},{"volume-title":"Methods for the subjective assessment of video quality, audio quality and audiovisual quality of Internet video and distribution quality television in any environment","author":"ITU-T","key":"e_1_2_1_30_1","unstructured":"ITU-T Recommendation P.913. 2021. Methods for the subjective assessment of video quality, audio quality and audiovisual quality of Internet video and distribution quality television in any environment. International Telecommunication Union, Geneva."},{"volume-title":"Dimension-based subjective quality evaluation for video content","author":"ITU-T","key":"e_1_2_1_31_1","unstructured":"ITU-T Recommendation P.918. 2021. Dimension-based subjective quality evaluation for video content. International Telecommunication Union, Geneva."},{"key":"e_1_2_1_32_1","unstructured":"J. Jung M. Wien and V. Baroncini. 2021. ISO\/IEC JTC 1\/SC 29\/AG 5: Guidelines for remote experts viewing sessions."},{"key":"e_1_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1145\/2389176.2389181"},{"key":"e_1_2_1_34_1","doi-asserted-by":"publisher","unstructured":"Christian Keimel Julian Habigt Clemens Horch and Klaus Diepold. 2012. QualityCrowd-A Framework for Crowd-based Quality Evaluation. In Picture coding symposium. doi:10.1109\/PCS.2012.6213338","DOI":"10.1109\/PCS.2012.6213338"},{"key":"e_1_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.2307\/2332226"},{"key":"e_1_2_1_36_1","doi-asserted-by":"crossref","unstructured":"Tobias Kirschstein Simon Giebenhain and Matthias Nie\u00dfner. 2024. DiffusionAvatars: Deferred Diffusion for High-fidelity 3D Head Avatars. In CVPR.","DOI":"10.1109\/CVPR52733.2024.00524"},{"key":"e_1_2_1_37_1","doi-asserted-by":"publisher","DOI":"10.1177\/0301006619869134"},{"key":"e_1_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2023.3320234"},{"key":"e_1_2_1_39_1","unstructured":"Jinlin Liu Kai Yu Mengyang Feng Xiefan Guo and Miaomiao Cui. 2024. Disentangling Foreground and Background Motion for Enhanced Realism in Human Video Generation. http:\/\/arxiv.org\/abs\/2405.16393 arXiv:2405.16393 [cs]."},{"key":"e_1_2_1_40_1","volume-title":"Ruben Alvarez Martinez, and Ross Cutler","author":"Majeedi Abrar","year":"2023","unstructured":"Abrar Majeedi, Babak Naderi, Yasaman Hosseinkashi, Juhee Cho, Ruben Alvarez Martinez, and Ross Cutler. 2023. Full Reference Video Quality Assessment for Machine Learning-Based Video Codecs. http:\/\/arxiv.org\/abs\/2309.00769 arXiv:2309.00769 [cs, eess]."},{"key":"e_1_2_1_41_1","first-page":"22","volume-title":"Cognition","volume":"146","author":"Mathur Maya","year":"2016","unstructured":"Maya Mathur and David Reichling. 2016. Navigating a social world with robot partners: A quantitative cartography of the Uncanny Valley. Cognition, Vol. 146 (Jan. 2016), 22-32."},{"key":"e_1_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1145\/2185520.2185587"},{"key":"e_1_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/MRA.2012.2192811"},{"key":"e_1_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.21437\/Interspeech.2020-2665"},{"key":"e_1_2_1_45_1","unstructured":"Babak Naderi and Ross Cutler. 2023. A crowdsourcing approach to video quality assessment. http:\/\/arxiv.org\/abs\/2204.06784 arXiv:2204.06784."},{"key":"e_1_2_1_46_1","doi-asserted-by":"crossref","unstructured":"Babak Naderi and Ross Cutler. 2024. A crowdsourcing approach to video quality assessment. In ICASSP.","DOI":"10.1109\/ICASSP48485.2024.10446509"},{"key":"e_1_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/QoMEX48832.2020.9123093"},{"key":"e_1_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1145\/1240624.1240846"},{"key":"e_1_2_1_49_1","volume-title":"Nguyen and John Canny","author":"David","year":"2009","unstructured":"David T. Nguyen and John Canny. 2009. More than Face-to-Face: Empathy Effects of Video Framing."},{"key":"e_1_2_1_50_1","unstructured":"ITU-T Recommendation P.912. 2016. Subjective video quality assessment methods for recognition tasks."},{"key":"e_1_2_1_51_1","volume-title":"Avatars in Work Meetings: Correlation Between Photorealism and Appeal. http:\/\/arxiv.org\/abs\/2304","author":"Phadnis Vrushank","year":"2023","unstructured":"Vrushank Phadnis, Kristin Moore, and Mar Gonzalez Franco. 2023. Avatars in Work Meetings: Correlation Between Photorealism and Appeal. http:\/\/arxiv.org\/abs\/2304.01405 arXiv:2304.01405 [cs]."},{"key":"e_1_2_1_52_1","volume-title":"A new subjective audiovisual & video quality testing recommendation. An Era of Change","author":"Pinson Margaret H","year":"2014","unstructured":"Margaret H Pinson and Lucjan Janowski. 2014. A new subjective audiovisual & video quality testing recommendation. An Era of Change (2014), 51."},{"key":"e_1_2_1_53_1","doi-asserted-by":"crossref","unstructured":"Shenhan Qian Zhi Tu Yihao Zhi Wen Liu and Shenghua Gao. 2021. Speech Drives Templates: Co-Speech Gesture Synthesis with Learned Templates. http:\/\/arxiv.org\/abs\/2108.08020 arXiv:2108.08020 [cs].","DOI":"10.1109\/ICCV48922.2021.01089"},{"key":"e_1_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/QoMEX.2013.6603196"},{"key":"e_1_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1109\/QoMEX51781.2021.9465425"},{"key":"e_1_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/2531602.2531732"},{"key":"e_1_2_1_57_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW.2013.132"},{"key":"e_1_2_1_58_1","doi-asserted-by":"crossref","unstructured":"Shunsuke Saito Gabriel Schwartz Tomas Simon Junxuan Li and Giljoo Nam. 2024. Relightable Gaussian Codec Avatars. In CVPR.","DOI":"10.1109\/CVPR52733.2024.00021"},{"key":"e_1_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1109\/QoMEX.2017.7965662"},{"volume-title":"CVPR","author":"Shao Zhijing","key":"e_1_2_1_60_1","unstructured":"Zhijing Shao, Zhaolong Wang, Zhuang Li, Duotun Wang, Xiangru Lin, Yu Zhang, Mingming Fan, and Zeyu Wang. 2024. SplattingAvatar: Realistic Real-Time Human Avatars with Mesh-Embedded Gaussian Splatting. In CVPR. http:\/\/arxiv.org\/abs\/2403.05087"},{"key":"e_1_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.3390\/electronics12030583"},{"key":"e_1_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1117\/12.2597813"},{"key":"e_1_2_1_63_1","volume-title":"EMO: Emote Portrait Alive - Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions","author":"Tian Linrui","year":"2024","unstructured":"Linrui Tian, Qi Wang, Bang Zhang, and Liefeng Bo. 2024. EMO: Emote Portrait Alive - Generating Expressive Portrait Videos with Audio2Video Diffusion Model under Weak Conditions. http:\/\/arxiv.org\/abs\/2402.17485 arXiv:2402.17485 [cs]."},{"key":"e_1_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.chb.2013.01.008"},{"key":"e_1_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.1109\/QOMEX.2010.5517948"},{"key":"e_1_2_1_66_1","doi-asserted-by":"publisher","DOI":"10.1109\/VR.2019.8798108"},{"key":"e_1_2_1_67_1","volume-title":"Adaptive psychophysical procedures. Vision research","author":"Treutwein Bernhard","year":"1995","unstructured":"Bernhard Treutwein. 1995. Adaptive psychophysical procedures. Vision research, Vol. 35, 17 (1995), 2503-2522."},{"key":"e_1_2_1_68_1","volume-title":"FVD: A new metric for video generation. ICLR","author":"Unterthiner Thomas","year":"2019","unstructured":"Thomas Unterthiner, Sjoerd van Steenkiste, Karol Kurach, Raphael Marinier, Marcin Michalski, and Sylvain Gelly. 2019. FVD: A new metric for video generation. ICLR (2019)."},{"key":"e_1_2_1_69_1","doi-asserted-by":"crossref","unstructured":"Evgeniy Upenik Michela Testolina Joao Ascenso Fernando Pereira and Touradj Ebrahimi (Eds.). 2021. Large-Scale Crowdsourcing Subjective Quality Evaluation of Learning-Based Image Coding. IEEE Visual Communications and Image Processing (VCIP 2021) (2021).","DOI":"10.1109\/VCIP53242.2021.9675314"},{"key":"e_1_2_1_70_1","unstructured":"Video Quality Experts Group. 2010. Report on the validation of video quality models for high definition video content. (2010). https:\/\/www.its.bldrdoc.gov\/media\/4212\/vqeg_hdtv_final_report_version_2.0.zip"},{"key":"e_1_2_1_71_1","doi-asserted-by":"crossref","unstructured":"Tan Wang Linjie Li Kevin Lin Yuanhao Zhai Chung-Ching Lin Zhengyuan Yang Hanwang Zhang Zicheng Liu and Lijuan Wang. 2024a. DisCo: Disentangled Control for Realistic Human Dance Generation. http:\/\/arxiv.org\/abs\/2307.00040 arXiv:2307.00040 [cs].","DOI":"10.1109\/CVPR52733.2024.00891"},{"key":"e_1_2_1_72_1","doi-asserted-by":"publisher","unstructured":"Xuanyu Wang Weizhan Zhang Christian Sandor and Hongbo Fu. 2024b. Real-and-Present: Investigating the Use of Life-Size 2D Video Avatars in HMD-Based AR Teleconferencing. doi:10.48550\/arXiv.2401.02171 arXiv:2401.02171 [cs].","DOI":"10.48550\/arXiv.2401.02171"},{"key":"e_1_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2003.819861"},{"key":"e_1_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACSSC.2003.1292216"},{"key":"e_1_2_1_75_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2023.3247072"},{"key":"e_1_2_1_76_1","unstructured":"Sicheng Xu Guojun Chen Yu-Xiao Guo Jiaolong Yang Chong Li Zhenyu Zang Yizhong Zhang Xin Tong and Baining Guo. 2024a. VASA-1: Lifelike Audio-Driven Talking Faces Generated in Real Time. http:\/\/arxiv.org\/abs\/2404.10667 arXiv:2404.10667 [cs]."},{"key":"e_1_2_1_77_1","unstructured":"Yuelang Xu Benwang Chen Zhe Li Hongwen Zhang Lizhen Wang Zerong Zheng and Yebin Liu. 2024b. Gaussian Head Avatar: Ultra High-fidelity Head Avatar via Dynamic Gaussians. In CVPR."},{"key":"e_1_2_1_78_1","unstructured":"Shurong Yang Huadong Li Juhao Wu Minhao Jing Linze Li Renhe Ji Jiajun Liang and Haoqiang Fan. 2024. MegActor: Harness the Power of Raw Video for Vivid Portrait Animation. http:\/\/arxiv.org\/abs\/2405.20851 arXiv:2405.20851 [cs]."},{"key":"e_1_2_1_79_1","unstructured":"Gaoxiong Yi Wei Xiao Yiming Xiao Babak Naderi Sebastian M\u00f6ller Wafaa Wardah Gabriel Mittag Ross Cutler Zhuohuang Zhang Donald S. Williamson Fei Chen Fuzheng Yang and Shidong Shang. 2022. ConferencingSpeech 2022 Challenge: Non-intrusive Objective Speech Quality Assessment (NISQA) Challenge for Online Conferencing Applications. In INTERSPEECH."},{"key":"e_1_2_1_80_1","doi-asserted-by":"publisher","DOI":"10.1109\/TVCG.2021.3106480"},{"key":"e_1_2_1_81_1","doi-asserted-by":"publisher","DOI":"10.1145\/2816795.2818126"},{"key":"e_1_2_1_82_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00068"},{"key":"e_1_2_1_83_1","doi-asserted-by":"crossref","unstructured":"Mingyuan Zhou Rakib Hyder Ziwei Xuan and Guojun Qi. 2024. UltrAvatar: A Realistic Animatable 3D Avatar Diffusion Model with Authenticity Guided Textures. In CVPR.","DOI":"10.1109\/CVPR52733.2024.00124"}],"container-title":["Proceedings of the ACM on Human-Computer Interaction"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3757542","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,10,17]],"date-time":"2025-10-17T01:58:12Z","timestamp":1760666292000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3757542"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,10,16]]},"references-count":83,"journal-issue":{"issue":"7","published-print":{"date-parts":[[2025,10,18]]}},"alternative-id":["10.1145\/3757542"],"URL":"https:\/\/doi.org\/10.1145\/3757542","relation":{},"ISSN":["2573-0142"],"issn-type":[{"type":"electronic","value":"2573-0142"}],"subject":[],"published":{"date-parts":[[2025,10,16]]},"assertion":[{"value":"2025-10-16","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}