{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,10]],"date-time":"2025-12-10T09:03:51Z","timestamp":1765357431469,"version":"3.41.0"},"reference-count":54,"publisher":"Association for Computing Machinery (ACM)","issue":"5s","license":[{"start":{"date-parts":[[2023,9,9]],"date-time":"2023-09-09T00:00:00Z","timestamp":1694217600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/501100004663","name":"Ministry of Science and Technology, Taiwan","doi-asserted-by":"crossref","award":["MOST 110-2222-E-001-003-MY3"],"award-info":[{"award-number":["MOST 110-2222-E-001-003-MY3"]}],"id":[{"id":"10.13039\/501100004663","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":["ACM Trans. Embed. Comput. Syst."],"published-print":{"date-parts":[[2023,10,31]]},"abstract":"<jats:p>Intermittent deep neural network (DNN) inference is a promising technique to enable intelligent applications on tiny devices powered by ambient energy sources. Nonetheless, intermittent execution presents inherent challenges, primarily involving accumulating progress across power cycles and having to refetch volatile data lost due to power loss in each power cycle. Existing approaches typically optimize the inference configuration to maximize data reuse. However, we observe that such a fixed configuration may be significantly inefficient due to the fluctuating balance point between data reuse and data refetch caused by the dynamic nature of ambient energy.<\/jats:p>\n          <jats:p>\n            This work proposes\n            <jats:italic>DynBal<\/jats:italic>\n            , an approach to dynamically reconfigure the inference engine at runtime. DynBal is realized as a middleware plugin that improves inference performance by exploring the interplay between data reuse and data refetch to maintain their balance with respect to the changing level of intermittency. An indirect metric is developed to easily evaluate an inference configuration considering the variability in intermittency, and a lightweight reconfiguration algorithm is employed to efficiently optimize the configuration at runtime. We evaluate the improvement brought by integrating DynBal into a recent intermittent inference approach that uses a fixed configuration. Evaluations were conducted on a Texas Instruments device with various network models and under varied intermittent power strengths. Our experimental results demonstrate that DynBal can speed up intermittent inference by 3.26 times, achieving a greater improvement for a large network under high intermittency and a large gap between memory and computation performance.\n          <\/jats:p>","DOI":"10.1145\/3607918","type":"journal-article","created":{"date-parts":[[2023,9,9]],"date-time":"2023-09-09T13:33:18Z","timestamp":1694266398000},"page":"1-25","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":7,"title":["Keep in Balance: Runtime-reconfigurable Intermittent Deep Inference"],"prefix":"10.1145","volume":"22","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4793-1721","authenticated-orcid":false,"given":"Chih-Hsuan","family":"Yen","sequence":"first","affiliation":[{"name":"National Taiwan University and Academia Sinica, Taiwan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3082-7688","authenticated-orcid":false,"given":"Hashan Roshantha","family":"Mendis","sequence":"additional","affiliation":[{"name":"Academia Sinica, Taiwan"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1974-0394","authenticated-orcid":false,"given":"Tei-Wei","family":"Kuo","sequence":"additional","affiliation":[{"name":"National Taiwan University, Taiwan and Mohamed bin Zayed University of Artificial Intelligence, United Arab Emirates"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8035-4033","authenticated-orcid":false,"given":"Pi-Cheng","family":"Hsiu","sequence":"additional","affiliation":[{"name":"Academia Sinica, National Taiwan University and National Chi Nan University, Taiwan"}]}],"member":"320","published-online":{"date-parts":[[2023,9,9]]},"reference":[{"key":"e_1_3_2_2_2","first-page":"324","volume-title":"Proc. of DSD","author":"Aghapour Ehsan","year":"2022","unstructured":"Ehsan Aghapour, Dolly Sapra, Andy Pimentel, and Anuj Pathania. 2022. CPU-GPU layer-switched low latency CNN inference. In Proc. of DSD. 324\u2013331."},{"key":"e_1_3_2_3_2","first-page":"437","volume-title":"Proc. of ESANN","author":"Anguita Davide","year":"2013","unstructured":"Davide Anguita, Alessandro Ghio, Luca Oneto, Xavier Parra Perez, and Jorge Luis Reyes Ortiz. 2013. A public domain dataset for human activity recognition using smartphones. In Proc. of ESANN. 437\u2013442."},{"key":"e_1_3_2_4_2","unstructured":"ARM. 2010. Cortex-M4 instructions. https:\/\/developer.arm.com\/documentation\/ddi0439\/b\/CHDDIGAC"},{"key":"e_1_3_2_5_2","article-title":"Benchmarking TinyML systems: Challenges and direction","author":"Banbury Colby R.","year":"2020","unstructured":"Colby R. Banbury, Vijay Janapa Reddi, Max Lam, William Fu, Amin Fazel, Jeremy Holleman, Xinyuan Huang, Robert Hurtado, David Kanter, Anton Lokhmotov, David Patterson, Danilo Pau, Jae sun Seo, Jeff Sieracki, Urmish Thakker, Marian Verhelst, and Poonam Yadav. 2020. Benchmarking TinyML systems: Challenges and direction. arXiv:2003.04821 (2020).","journal-title":"arXiv:2003.04821"},{"key":"e_1_3_2_6_2","first-page":"40","volume-title":"Proc. of IEEE RTAS","author":"Choi Jongouk","year":"2022","unstructured":"Jongouk Choi, Larry Kittinger, Qingrui Liu, and Changhee Jung. 2022. Compiler-directed high-performance intermittent computation with power failure immunity. In Proc. of IEEE RTAS. 40\u201354."},{"key":"e_1_3_2_7_2","unstructured":"IXYS Corporation. 2010. IXOLAR high efficiency SolarMD. https:\/\/ixapps.ixys.com\/DataSheet\/SM111K04L.pdf"},{"key":"e_1_3_2_8_2","unstructured":"Cypress. 2019. 4-Mbit SPI FRAM. https:\/\/www.cypress.com\/file\/209146\/download"},{"issue":"4","key":"e_1_3_2_9_2","first-page":"46:1\u201328","article-title":"Human activity recognition on microcontrollers with quantized and adaptive deep neural networks","volume":"21","author":"Daghero Francesco","year":"2022","unstructured":"Francesco Daghero, Alessio Burrello, Chen Xie, Marco Castellano, Luca Gandolfi, Andrea Calimera, Enrico Macii, Massimo Poncino, and Daniele Jahier Pagliari. 2022. Human activity recognition on microcontrollers with quantized and adaptive deep neural networks. ACM TECS 21, 4 (2022), 46:1\u201328.","journal-title":"ACM TECS"},{"key":"e_1_3_2_10_2","first-page":"16344","volume-title":"Proc. of NeurIPS","author":"Dao Tri","year":"2022","unstructured":"Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher R\u00e9. 2022. Flashattention: Fast and memory-efficient exact attention with io-awareness. In Proc. of NeurIPS. 16344\u201316359."},{"key":"e_1_3_2_11_2","first-page":"77","volume-title":"Proc. of ENSsys","author":"Daulby Tim","year":"2020","unstructured":"Tim Daulby, Anand Savanth, Alex S Weddell, and Geoff V Merrett. 2020. Comparing NVM technologies through the lens of intermittent computation. In Proc. of ENSsys. 77\u201378."},{"issue":"5","key":"e_1_3_2_12_2","first-page":"70:1\u201327","article-title":"Dmazerunner: Executing perfectly nested loops on dataflow accelerators","volume":"18","author":"Dave Shail","year":"2019","unstructured":"Shail Dave, Youngbin Kim, Sasikanth Avancha, Kyoungwoo Lee, and Aviral Shrivastava. 2019. Dmazerunner: Executing perfectly nested loops on dataflow accelerators. ACM TECS 18, 5s (2019), 70:1\u201327.","journal-title":"ACM TECS"},{"issue":"3","key":"e_1_3_2_13_2","first-page":"111:1\u201334","article-title":"Battery-free game boy","volume":"4","author":"Winkel Jasper De","year":"2020","unstructured":"Jasper De Winkel, Vito Kortbeek, Josiah Hester, and Przemys\u0142aw Pawe\u0142czak. 2020. Battery-free game boy. ACM IMWUT 4, 3 (2020), 111:1\u201334.","journal-title":"ACM IMWUT"},{"key":"e_1_3_2_14_2","first-page":"199","volume-title":"Proc. of ACM ASPLOS","author":"Gobieski Graham","year":"2019","unstructured":"Graham Gobieski, Nathan Beckmann, and Brandon Lucia. 2019. Intelligence beyond the edge: Inference on intermittent embedded systems. In Proc. of ACM ASPLOS. 199\u2013213."},{"key":"e_1_3_2_15_2","unstructured":"Burak Himmetoglu. 2017. CNN Model for Human Activity Recognition. https:\/\/github.com\/healthDataScience\/deep-learning-HAR"},{"key":"e_1_3_2_16_2","article-title":"SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and  \\(\\lt\\) 0.5 MB model size","author":"Iandola Forrest N.","year":"2016","unstructured":"Forrest N. Iandola, Song Han, Matthew W. Moskewicz, Khalid Ashraf, William J. Dally, and Kurt Keutzer. 2016. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and \\(\\lt\\) 0.5 MB model size. arXiv:1602.07360 (2016).","journal-title":"arXiv:1602.07360"},{"key":"e_1_3_2_17_2","first-page":"921","volume-title":"Proc. of IEEE\/ACM DATE","author":"Islam Sahidul","year":"2022","unstructured":"Sahidul Islam, Jieren Deng, Shanglin Zhou, Chen Pan, Caiwen Ding, and Mimi Xie. 2022. Enabling fast deep learning on tiny energy-harvesting IoT devices. In Proc. of IEEE\/ACM DATE. 921\u2013926."},{"key":"e_1_3_2_18_2","first-page":"1","volume-title":"Proc. of IEEE\/ACM ICCAD","author":"Islam Sahidul","year":"2022","unstructured":"Sahidul Islam, Shanglin Zhou, Ran Ran, Yu-Fang Jin, Wujie Wen, Caiwen Ding, and Mimi Xie. 2022. EVE: Environmental adaptive neural network models for low-power energy harvesting system. In Proc. of IEEE\/ACM ICCAD. 1\u20139."},{"key":"e_1_3_2_19_2","first-page":"187","volume-title":"Proc. of ACM SIGCOMM","author":"Jang Junsu","year":"2019","unstructured":"Junsu Jang and Fadel Adib. 2019. Underwater backscatter networking. In Proc. of ACM SIGCOMM. 187\u2013199."},{"issue":"11","key":"e_1_3_2_20_2","first-page":"4154","article-title":"Standing on the shoulders of giants: Hardware and neural architecture co-search with hot start","volume":"39","author":"Jiang Weiwen","year":"2020","unstructured":"Weiwen Jiang, Lei Yang, Sakyasingha Dasgupta, Jingtong Hu, and Yiyu Shi. 2020. Standing on the shoulders of giants: Hardware and neural architecture co-search with hot start. IEEE TCAD 39, 11 (2020), 4154\u20134165.","journal-title":"IEEE TCAD"},{"key":"e_1_3_2_21_2","first-page":"2146","volume-title":"Proc. of IEEE CVPR","author":"Jin Qing","year":"2020","unstructured":"Qing Jin, Linjie Yang, and Zhenyu Liao. 2020. AdaBits: Neural network quantization with adaptive bit-widths. In Proc. of IEEE CVPR. 2146\u20132156."},{"key":"e_1_3_2_22_2","article-title":"More is less: Model augmentation for intermittent deep inference","author":"Kang Chih-Kai","year":"2022","unstructured":"Chih-Kai Kang, Hashan Roshantha Mendis, Chun-Han Lin, Ming-Syan Chen, and Pi-Cheng Hsiu. 2022. More is less: Model augmentation for intermittent deep inference. ACM TECS (2022).","journal-title":"ACM TECS"},{"key":"e_1_3_2_23_2","unstructured":"Kendryte. 2018. K210 AI Chip Datasheet. https:\/\/github.com\/kendryte\/kendryte-doc-datasheet"},{"key":"e_1_3_2_24_2","volume-title":"Learning Multiple Layers of Features from Tiny Images","author":"Krizhevsky Alex","year":"2009","unstructured":"Alex Krizhevsky. 2009. Learning Multiple Layers of Features from Tiny Images. Technical Report. University of Toronto."},{"key":"e_1_3_2_25_2","first-page":"138","volume-title":"Proc. of ACM SenSys","author":"Lee Seulki","year":"2019","unstructured":"Seulki Lee and Shahriar Nirjon. 2019. Neuro.ZERO: A zero-energy neural network accelerator for embedded sensing and inference systems. In Proc. of ACM SenSys. 138\u2013152."},{"key":"e_1_3_2_26_2","first-page":"8607","volume-title":"Proc. of IEEE CVPR","author":"Li Changlin","year":"2021","unstructured":"Changlin Li, Guangrun Wang, Bing Wang, Xiaodan Liang, Zhihui Li, and Xiaojun Chang. 2021. Dynamic slimmable network. In Proc. of IEEE CVPR. 8607\u20138617."},{"issue":"1","key":"e_1_3_2_27_2","first-page":"15","article-title":"Checkpointing-aware loop tiling for energy harvesting powered nonvolatile processors","volume":"38","author":"Li Fuyang","year":"2019","unstructured":"Fuyang Li, Keni Qiu, Mengying Zhao, Jingtong Hu, Yongpan Liu, Yong Guan, and Chun Jason Xue. 2019. Checkpointing-aware loop tiling for energy harvesting powered nonvolatile processors. IEEE TCAD 38, 1 (2019), 15\u201328.","journal-title":"IEEE TCAD"},{"issue":"11","key":"e_1_3_2_28_2","first-page":"3614","article-title":"Fusion-catalyzed pruning for optimizing deep learning on intelligent edge devices","volume":"39","author":"Li Guangli","year":"2020","unstructured":"Guangli Li, Xiu Ma, Xueying Wang, Lei Liu, Jingling Xue, and Xiaobing Feng. 2020. Fusion-catalyzed pruning for optimizing deep learning on intelligent edge devices. IEEE TCAD 39, 11 (2020), 3614\u20133626.","journal-title":"IEEE TCAD"},{"key":"e_1_3_2_29_2","first-page":"343","volume-title":"Proc. of IEEE\/ACM DATE","author":"Li Jiajun","year":"2018","unstructured":"Jiajun Li, Guihai Yan, Wenyan Lu, Shuhao Jiang, Shijun Gong, Jingya Wu, and Xiaowei Li. 2018. SmartShuttle: Optimizing off-chip memory accesses for deep learning accelerators. In Proc. of IEEE\/ACM DATE. 343\u2013348."},{"key":"e_1_3_2_30_2","first-page":"1","volume-title":"Proc. of IEEE\/ACM DAC","author":"Lin Chih-Chia","year":"2023","unstructured":"Chih-Chia Lin, Chia-Yin Liu, Chih-Hsuan Yen, Tei-Wei Kuo, and Pi-Cheng Hsiu. 2023. Intermittent-aware neural network pruning. In Proc. of IEEE\/ACM DAC. 1\u20137."},{"key":"e_1_3_2_31_2","first-page":"1","volume-title":"Proc. of NeurIPS","author":"Lin Ji","year":"2021","unstructured":"Ji Lin, Wei-Ming Chen, Han Cai, Chuang Gan, and Song Han. 2021. Memory-efficient patch-based inference for tiny deep learning. In Proc. of NeurIPS. 1\u201313."},{"key":"e_1_3_2_32_2","first-page":"11711","volume-title":"Proc. of NeurIPS","author":"Lin Ji","year":"2020","unstructured":"Ji Lin, Wei-Ming Chen, John Cohn, Chuang Gan, and Song Han. 2020. MCUNet: Tiny deep learning on IoT devices. In Proc. of NeurIPS. 11711\u201311722."},{"key":"e_1_3_2_33_2","first-page":"1025","volume-title":"Proc. of USENIX ATC","author":"Liu Yizhi","year":"2019","unstructured":"Yizhi Liu, Yao Wang, Ruofei Yu, Mu Li, Vin Sharma, and Yida Wang. 2019. Optimizing CNN model inference on CPUs. In Proc. of USENIX ATC. 1025\u20131040."},{"key":"e_1_3_2_34_2","first-page":"526","volume-title":"Proc. of IEEE HPCA","author":"Ma Kaisheng","year":"2015","unstructured":"Kaisheng Ma, Yang Zheng, Shuangchen Li, Karthik Swaminathan, Xueqing Li, Yongpan Liu, Jack Sampson, Yuan Xie, and Vijaykrishnan Narayanan. 2015. Architecture exploration for ambient energy harvesting nonvolatile processors. In Proc. of IEEE HPCA. 526\u2013537."},{"key":"e_1_3_2_35_2","unstructured":"Maxim Integrated. 2021. MAX78000 Ultra-low-power MCU with Arm Cortex-M4 and a CNN Accelerator. https:\/\/datasheets.maximintegrated.com\/en\/ds\/MAX78000.pdf"},{"issue":"5","key":"e_1_3_2_36_2","first-page":"64:1\u201327","article-title":"Intermittent-aware neural architecture search","volume":"20","author":"Mendis Hashan Roshantha","year":"2021","unstructured":"Hashan Roshantha Mendis, Chih-Kai Kang, and Pi-Cheng Hsiu. 2021. Intermittent-aware neural architecture search. ACM TECS 20, 5s (2021), 64:1\u201327.","journal-title":"ACM TECS"},{"key":"e_1_3_2_37_2","unstructured":"National Renewable Energy Laboratory (NREL). 2023. Measurement and Instrumentation Data Center (MIDC). https:\/\/midcdmz.nrel.gov\/"},{"issue":"2","key":"e_1_3_2_38_2","first-page":"604","article-title":"A survey of the usages of deep learning for natural language processing","volume":"32","author":"Otter Daniel W","year":"2020","unstructured":"Daniel W Otter, Julian R Medina, and Jugal K Kalita. 2020. A survey of the usages of deep learning for natural language processing. IEEE TNNLS 32, 2 (2020), 604\u2013624.","journal-title":"IEEE TNNLS"},{"key":"e_1_3_2_39_2","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2021.3064066"},{"key":"e_1_3_2_40_2","first-page":"1303","volume-title":"Proc. of IEEE\/ACM DAC","author":"Shafique Muhammad","year":"2021","unstructured":"Muhammad Shafique, Theocharis Theocharides, Vijay Janapa Reddy, and Boris Murmann. 2021. TinyML: Current progress, research challenges, and future roadmap. In Proc. of IEEE\/ACM DAC. 1303\u20131306."},{"key":"e_1_3_2_41_2","first-page":"440","volume-title":"Operating Systems: Internals and Design Principles (8 ed.)","author":"Stallings William","year":"2014","unstructured":"William Stallings. 2014. Operating Systems: Internals and Design Principles (8 ed.). Pearson New York, 440\u2013442."},{"issue":"2","key":"e_1_3_2_42_2","first-page":"25:1\u201320","article-title":"Battery-free cellphone","volume":"1","author":"Talla Vamsi","year":"2017","unstructured":"Vamsi Talla, Bryce Kellogg, Shyamnath Gollakota, and Joshua R. Smith. 2017. Battery-free cellphone. ACM IMWUT 1, 2 (2017), 25:1\u201320.","journal-title":"ACM IMWUT"},{"key":"e_1_3_2_43_2","unstructured":"Texas Instruments. 2016. Low-Energy Accelerator. https:\/\/www.ti.com\/lit\/an\/slaa720\/slaa720.pdf"},{"issue":"11","key":"e_1_3_2_44_2","first-page":"3479","article-title":"Everything leaves footprints: Hardware accelerated intermittent deep inference","volume":"39","author":"Kang Chih-Kai","year":"2020","unstructured":"Chih-Kai Kang, Hashan Roshantha Mendis, Chun-Han Lin, Ming-Syan Chen, and Pi-Cheng Hsiu. 2020. Everything leaves footprints: Hardware accelerated intermittent deep inference. IEEE TCAD 39, 11 (2020), 3479\u20133491.","journal-title":"IEEE TCAD"},{"key":"e_1_3_2_45_2","unstructured":"Texas Instruments. 2019. MSP432P401R MCU. https:\/\/www.ti.com\/product\/MSP432P401R"},{"key":"e_1_3_2_46_2","doi-asserted-by":"publisher","DOI":"10.1016\/j.sysarc.2020.101859"},{"key":"e_1_3_2_47_2","doi-asserted-by":"publisher","DOI":"10.1109\/MDAT.2020.2968258"},{"key":"e_1_3_2_48_2","article-title":"Speech commands: A dataset for limited-vocabulary speech recognition","author":"Warden Pete","year":"2018","unstructured":"Pete Warden. 2018. Speech commands: A dataset for limited-vocabulary speech recognition. arXiv:1804.03209 (2018).","journal-title":"arXiv:1804.03209"},{"key":"e_1_3_2_49_2","first-page":"1","volume-title":"Proc. of IEEE\/ACM DAC","author":"Yawen Wu","year":"2020","unstructured":"Wu Yawen, Wang Zhepeng, Jia Zhenge, Shi Yiyu, and Hu Jingtong. 2020. Intermittent inference with nonuniformly compressed multi-exit neural network for energy harvesting powered devices. In Proc. of IEEE\/ACM DAC. 1\u20136."},{"issue":"11","key":"e_1_3_2_50_2","first-page":"4229","article-title":"Stateful neural networks for intermittent systems","volume":"41","author":"Yen Chih-Hsuan","year":"2022","unstructured":"Chih-Hsuan Yen, Hashan Roshantha Mendis, Tei-Wei Kuo, and Pi-Cheng Hsiu. 2022. Stateful neural networks for intermittent systems. IEEE TCAD 41, 11 (2022), 4229\u20134240.","journal-title":"IEEE TCAD"},{"key":"e_1_3_2_51_2","unstructured":"Chih-Hsuan Yen Hashan Roshantha Mendis Tei-Wei Kuo and Pi-Cheng Hsiu. 2023. DynBal open source project. https:\/\/github.com\/EMCLab-Sinica\/DynBal"},{"key":"e_1_3_2_52_2","first-page":"161","volume-title":"Proc. of ACM FPGA","author":"Zhang Chen","year":"2015","unstructured":"Chen Zhang, Peng Li, Guangyu Sun, Yijin Guan, Bingjun Xiao, and Jason Cong. 2015. Optimizing FPGA-based accelerator design for deep convolutional neural networks. In Proc. of ACM FPGA. 161\u2013170."},{"key":"e_1_3_2_53_2","article-title":"Hello edge: Keyword spotting on microcontrollers","author":"Zhang Yundong","year":"2017","unstructured":"Yundong Zhang, Naveen Suda, Liangzhen Lai, and Vikas Chandra. 2017. Hello edge: Keyword spotting on microcontrollers. arXiv:1711.07128 (2017).","journal-title":"arXiv:1711.07128"},{"key":"e_1_3_2_54_2","first-page":"644","volume-title":"Proc. of IEEE ICCD","author":"Zhang Yujie","year":"2022","unstructured":"Yujie Zhang, Dhananjaya Wijerathne, Zhaoying Li, and Tulika Mitra. 2022. Power-performance characterization of TinyML systems. In Proc. of IEEE ICCD. 644\u2013651."},{"issue":"5","key":"e_1_3_2_55_2","first-page":"50:1\u201326","article-title":"TAB: Unified and optimized ternary, binary, and mixed-precision neural network inference on the edge","volume":"21","author":"Zhu Shien","year":"2022","unstructured":"Shien Zhu, Luan HK Duong, and Weichen Liu. 2022. TAB: Unified and optimized ternary, binary, and mixed-precision neural network inference on the edge. ACM TECS 21, 5 (2022), 50:1\u201326.","journal-title":"ACM TECS"}],"container-title":["ACM Transactions on Embedded Computing Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3607918","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3607918","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,17]],"date-time":"2025-06-17T16:38:06Z","timestamp":1750178286000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3607918"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,9]]},"references-count":54,"journal-issue":{"issue":"5s","published-print":{"date-parts":[[2023,10,31]]}},"alternative-id":["10.1145\/3607918"],"URL":"https:\/\/doi.org\/10.1145\/3607918","relation":{},"ISSN":["1539-9087","1558-3465"],"issn-type":[{"type":"print","value":"1539-9087"},{"type":"electronic","value":"1558-3465"}],"subject":[],"published":{"date-parts":[[2023,9,9]]},"assertion":[{"value":"2023-03-23","order":0,"name":"received","label":"Received","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-07-13","order":1,"name":"accepted","label":"Accepted","group":{"name":"publication_history","label":"Publication History"}},{"value":"2023-09-09","order":2,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}