{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,11]],"date-time":"2025-12-11T03:06:23Z","timestamp":1765422383006,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":64,"publisher":"ACM","license":[{"start":{"date-parts":[[2025,3,31]],"date-time":"2025-03-31T00:00:00Z","timestamp":1743379200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/501100003725","name":"National Research Foundation of Korea","doi-asserted-by":"publisher","award":["RS-2023-00229822"],"award-info":[{"award-number":["RS-2023-00229822"]}],"id":[{"id":"10.13039\/501100003725","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2025,3,31]]},"DOI":"10.1145\/3672608.3707816","type":"proceedings-article","created":{"date-parts":[[2025,5,14]],"date-time":"2025-05-14T18:30:17Z","timestamp":1747247417000},"page":"687-694","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Advanced Knowledge Transfer: Refined Feature Distillation for Zero-Shot Quantization in Edge Computing"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-1679-1268","authenticated-orcid":false,"given":"Inpyo","family":"Hong","sequence":"first","affiliation":[{"name":"Department of Computer Science, Yonsei University, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2773-7670","authenticated-orcid":false,"given":"Youngwan","family":"Jo","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Yonsei University, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2046-3091","authenticated-orcid":false,"given":"Hyojeong","family":"Lee","sequence":"additional","affiliation":[{"name":"Department of Artificial Intelligence, Yonsei University, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9881-5731","authenticated-orcid":false,"given":"Sunghyun","family":"Ahn","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Yonsei University, Seoul, Republic of Korea"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5196-6193","authenticated-orcid":false,"given":"Sanghyun","family":"Park","sequence":"additional","affiliation":[{"name":"Department of Computer Science, Yonsei University, Seoul, Republic of Korea"}]}],"member":"320","published-online":{"date-parts":[[2025,5,14]]},"reference":[{"key":"e_1_3_2_1_1_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v38i10.28972"},{"key":"e_1_3_2_1_2_1","volume-title":"Proceedings of the 1988 connectionist models summer school. 29\u201337","author":"Becker Sue","year":"1988","unstructured":"Sue Becker, Yann Le Cun, et al. 1988. Improving the convergence of back-propagation learning with second order methods. In Proceedings of the 1988 connectionist models summer school. 29\u201337."},{"key":"e_1_3_2_1_3_1","series-title":"SIAM review 60, 2","volume-title":"Optimization methods for large-scale machine learning","author":"Bottou L\u00e9on","year":"2018","unstructured":"L\u00e9on Bottou, Frank E Curtis, and Jorge Nocedal. 2018. Optimization methods for large-scale machine learning. SIAM review 60, 2 (2018), 223\u2013311."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01318"},{"key":"e_1_3_2_1_5_1","volume-title":"Efficientqat: Efficient quantization-aware training for large language models. arXiv preprint arXiv:2407.11062","author":"Chen Mengzhao","year":"2024","unstructured":"Mengzhao Chen, Wenqi Shao, Peng Xu, Jiahao Wang, Peng Gao, Kaipeng Zhang, Yu Qiao, and Ping Luo. 2024. Efficientqat: Efficient quantization-aware training for large language models. arXiv preprint arXiv:2407.11062 (2024)."},{"key":"e_1_3_2_1_6_1","volume-title":"TexQ: zero-shot network quantization with texture feature distribution calibration. Advances in Neural Information Processing Systems 36","author":"Chen Xinrui","year":"2024","unstructured":"Xinrui Chen, Yizhi Wang, Renao Yan, Yiqing Liu, Tian Guan, and Yonghong He. 2024. TexQ: zero-shot network quantization with texture feature distribution calibration. Advances in Neural Information Processing Systems 36 (2024)."},{"key":"e_1_3_2_1_7_1","first-page":"14835","article-title":"Qimera: Data-free quantization with synthetic boundary supporting samples","volume":"34","author":"Choi Kanghyun","year":"2021","unstructured":"Kanghyun Choi, Deokki Hong, Noseong Park, Youngsok Kim, and Jinho Lee. 2021. Qimera: Data-free quantization with synthetic boundary supporting samples. Advances in Neural Information Processing Systems 34 (2021), 14835\u201314847.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.00813"},{"key":"e_1_3_2_1_9_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPRW50498.2020.00363"},{"key":"e_1_3_2_1_10_1","volume-title":"Towards the limit of network quantization. arXiv preprint arXiv:1612.01543","author":"Choi Yoojin","year":"2016","unstructured":"Yoojin Choi, Mostafa El-Khamy, and Jungwon Lee. 2016. Towards the limit of network quantization. arXiv preprint arXiv:1612.01543 (2016)."},{"key":"e_1_3_2_1_11_1","volume-title":"Binaryconnect: Training deep neural networks with binary weights during propagations. Advances in neural information processing systems 28","author":"Courbariaux Matthieu","year":"2015","unstructured":"Matthieu Courbariaux, Yoshua Bengio, and Jean-Pierre David. 2015. Binaryconnect: Training deep neural networks with binary weights during propagations. Advances in neural information processing systems 28 (2015)."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52733.2024.01531"},{"key":"e_1_3_2_1_13_1","volume-title":"Training with quantization noise for extreme model compression. arXiv preprint arXiv:2004.07320","author":"Fan Angela","year":"2020","unstructured":"Angela Fan, Pierre Stock, Benjamin Graham, Edouard Grave, R\u00e9mi Gribonval, Herve Jegou, and Armand Joulin. 2020. Training with quantization noise for extreme model compression. arXiv preprint arXiv:2004.07320 (2020)."},{"key":"e_1_3_2_1_14_1","volume-title":"Edge AI: Evaluation of Model Compression Techniques for Convolutional Neural Networks. arXiv preprint arXiv:2409.02134","author":"Francy Samer","year":"2024","unstructured":"Samer Francy and Raghubir Singh. 2024. Edge AI: Evaluation of Model Compression Techniques for Convolutional Neural Networks. arXiv preprint arXiv:2409.02134 (2024)."},{"volume-title":"Low-Power Computer Vision","author":"Gholami Amir","key":"e_1_3_2_1_15_1","unstructured":"Amir Gholami, Sehoon Kim, Zhen Dong, Zhewei Yao, Michael W Mahoney, and Kurt Keutzer. 2022. A survey of quantization methods for efficient neural network inference. In Low-Power Computer Vision. Chapman and Hall\/CRC, 291\u2013326."},{"key":"e_1_3_2_1_16_1","volume-title":"Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752","author":"Gu Albert","year":"2023","unstructured":"Albert Gu and Tri Dao. 2023. Mamba: Linear-time sequence modeling with selective state spaces. arXiv preprint arXiv:2312.00752 (2023)."},{"key":"e_1_3_2_1_17_1","volume-title":"Squant: On-the-fly data-free quantization via diagonal hessian approximation. arXiv preprint arXiv:2202.07471","author":"Guo Cong","year":"2022","unstructured":"Cong Guo, Yuxian Qiu, Jingwen Leng, Xiaotian Gao, Chen Zhang, Yunxin Liu, Fan Yang, Yuhao Zhu, and Minyi Guo. 2022. Squant: On-the-fly data-free quantization via diagonal hessian approximation. arXiv preprint arXiv:2202.07471 (2022)."},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1145\/3007787.3001163"},{"key":"e_1_3_2_1_19_1","volume-title":"Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149","author":"Han Song","year":"2015","unstructured":"Song Han, Huizi Mao, and William J Dally. 2015. Deep compression: Compressing deep neural networks with pruning, trained quantization and huffman coding. arXiv preprint arXiv:1510.00149 (2015)."},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_21_1","volume-title":"Efficientdm: Efficient quantization-aware fine-tuning of low-bit diffusion models. arXiv preprint arXiv:2310.03270","author":"He Yefei","year":"2023","unstructured":"Yefei He, Jing Liu, Weijia Wu, Hong Zhou, and Bohan Zhuang. 2023. Efficientdm: Efficient quantization-aware fine-tuning of low-bit diffusion models. arXiv preprint arXiv:2310.03270 (2023)."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00201"},{"key":"e_1_3_2_1_23_1","volume-title":"Distilling the Knowledge in a Neural Network. arXiv preprint arXiv:1503.02531","author":"Hinton Geoffrey","year":"2015","unstructured":"Geoffrey Hinton. 2015. Distilling the Knowledge in a Neural Network. arXiv preprint arXiv:1503.02531 (2015)."},{"key":"e_1_3_2_1_24_1","volume-title":"Knowledge distillation vulnerability of DeiT through CNN adversarial attack. Neural Computing and Applications","author":"Hong Inpyo","year":"2023","unstructured":"Inpyo Hong and Chang Choi. 2023. Knowledge distillation vulnerability of DeiT through CNN adversarial attack. Neural Computing and Applications (2023), 1\u201311."},{"key":"e_1_3_2_1_25_1","doi-asserted-by":"publisher","DOI":"10.3390\/app142310872"},{"key":"e_1_3_2_1_26_1","volume-title":"Proceedings of the 38th ACM\/SIGAPP Symposium on Applied Computing. 1669\u20131672","author":"Choi Chang","year":"2023","unstructured":"In-pyo Hong, Gyu-ho Choi, Pan-koo Kim, and Chang Choi. 2023. Security verification software platform of data-efficient image transformer based on fast gradient sign method. In Proceedings of the 38th ACM\/SIGAPP Symposium on Applied Computing. 1669\u20131672."},{"key":"e_1_3_2_1_27_1","volume-title":"Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861","author":"Howard Andrew G","year":"2017","unstructured":"Andrew G Howard. 2017. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1145\/3555802"},{"key":"e_1_3_2_1_29_1","first-page":"1","article-title":"Quantized neural networks: Training neural networks with low precision weights and activations","volume":"18","author":"Hubara Itay","year":"2018","unstructured":"Itay Hubara, Matthieu Courbariaux, Daniel Soudry, Ran El-Yaniv, and Yoshua Bengio. 2018. Quantized neural networks: Training neural networks with low precision weights and activations. Journal of Machine Learning Research 18, 187 (2018), 1\u201330.","journal-title":"Journal of Machine Learning Research"},{"key":"e_1_3_2_1_30_1","volume-title":"Quantizing deep convolutional networks for efficient inference: A whitepaper. arXiv preprint arXiv:1806.08342","author":"Krishnamoorthi Raghuraman","year":"2018","unstructured":"Raghuraman Krishnamoorthi. 2018. Quantizing deep convolutional networks for efficient inference: A whitepaper. arXiv preprint arXiv:1806.08342 (2018)."},{"key":"e_1_3_2_1_31_1","unstructured":"Alex Krizhevsky Geoffrey Hinton et al. 2009. Learning multiple layers of features from tiny images. (2009)."},{"volume-title":"Modern Approaches in Machine Learning and Cognitive Science: A Walkthrough: Latest Trends in AI","author":"Kulkarni Uday","key":"e_1_3_2_1_32_1","unstructured":"Uday Kulkarni, SM Meena, Sunil V Gurlahosur, Pratiksha Benagi, Atul Kashyap, Ayub Ansari, and Vinay Karnam. 2021. AI model compression for edge devices using optimization techniques. In Modern Approaches in Machine Learning and Cognitive Science: A Walkthrough: Latest Trends in AI, Volume 2. Springer, 227\u2013240."},{"key":"e_1_3_2_1_33_1","volume-title":"Arash Behboodi, and Tijmen Blankevoort.","author":"Kuzmin Andrey","year":"2024","unstructured":"Andrey Kuzmin, Markus Nagel, Mart Van Baalen, Arash Behboodi, and Tijmen Blankevoort. 2024. Pruning vs quantization: which is better? Advances in neural information processing systems 36 (2024)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.02339"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neucom.2021.07.045"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01554"},{"key":"e_1_3_2_1_37_1","volume-title":"International conference on machine learning. PMLR, 2849\u20132858","author":"Lin Darryl","year":"2016","unstructured":"Darryl Lin, Sachin Talathi, and Sreekanth Annapureddy. 2016. Fixed point quantization of deep convolutional networks. In International conference on machine learning. PMLR, 2849\u20132858."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.01383"},{"key":"e_1_3_2_1_39_1","volume-title":"PTQ-SO: A Scale Optimization-based Approach for Post-training Quantization of Edge Computing. In 2024 27th International Conference on Computer Supported Cooperative Work in Design (CSCWD). IEEE","author":"Liu Kangkang","year":"2024","unstructured":"Kangkang Liu and Ningjiang Chen. 2024. PTQ-SO: A Scale Optimization-based Approach for Post-training Quantization of Edge Computing. In 2024 27th International Conference on Computer Supported Cooperative Work in Design (CSCWD). IEEE, 2078\u20132083."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2022.3176400"},{"key":"e_1_3_2_1_41_1","first-page":"11913","article-title":"High-fidelity generative image compression","volume":"33","author":"Mentzer Fabian","year":"2020","unstructured":"Fabian Mentzer, George D Toderici, Michael Tschannen, and Eirikur Agustsson. 2020. High-fidelity generative image compression. Advances in Neural Information Processing Systems 33 (2020), 11913\u201311924.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_42_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV.2019.00141"},{"key":"e_1_3_2_1_43_1","doi-asserted-by":"publisher","DOI":"10.1109\/ACCESS.2023.3275173"},{"key":"e_1_3_2_1_44_1","volume-title":"Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems 32","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, et al. 2019. Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems 32 (2019)."},{"key":"e_1_3_2_1_45_1","volume-title":"Fast exact multiplication by the Hessian. Neural computation 6, 1","author":"Pearlmutter Barak A","year":"1994","unstructured":"Barak A Pearlmutter. 1994. Fast exact multiplication by the Hessian. Neural computation 6, 1 (1994), 147\u2013160."},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00769"},{"key":"e_1_3_2_1_47_1","volume-title":"Proceedings of the AAAI conference on artificial intelligence","volume":"37","author":"Qian Biao","year":"2023","unstructured":"Biao Qian, Yang Wang, Richang Hong, and Meng Wang. 2023. Rethinking datafree quantization as a zero-sum game. In Proceedings of the AAAI conference on artificial intelligence, Vol. 37. 9489\u20139497."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00474"},{"key":"e_1_3_2_1_49_1","volume-title":"Infrared Domain Adaptation with Zero-Shot Quantization. arXiv preprint arXiv:2408.13925","author":"Sevsay Burak","year":"2024","unstructured":"Burak Sevsay and Erdem Akag\u00fcnd\u00fcz. 2024. Infrared Domain Adaptation with Zero-Shot Quantization. arXiv preprint arXiv:2408.13925 (2024)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52729.2023.00196"},{"key":"e_1_3_2_1_51_1","volume-title":"Rate-distortion optimized post-training quantization for learned image compression","author":"Shi Junqi","year":"2023","unstructured":"Junqi Shi, Ming Lu, and Zhan Ma. 2023. Rate-distortion optimized post-training quantization for learned image compression. IEEE Transactions on Circuits and Systems for Video Technology (2023)."},{"key":"e_1_3_2_1_52_1","first-page":"6906","article-title":"Does knowledge distillation really work","volume":"34","author":"Stanton Samuel","year":"2021","unstructured":"Samuel Stanton, Pavel Izmailov, Polina Kirichenko, Alexander A Alemi, and Andrew G Wilson. 2021. Does knowledge distillation really work? Advances in Neural Information Processing Systems 34 (2021), 6906\u20136919.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_53_1","volume-title":"Yolov10: Real-time end-to-end object detection. arXiv preprint arXiv:2405.14458","author":"Wang Ao","year":"2024","unstructured":"Ao Wang, Hui Chen, Lihao Liu, Kai Chen, Zijia Lin, Jungong Han, and Guiguang Ding. 2024. Yolov10: Real-time end-to-end object detection. arXiv preprint arXiv:2405.14458 (2024)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1145\/3652583.3658018"},{"key":"e_1_3_2_1_55_1","volume-title":"Integer quantization for deep learning inference: Principles and empirical evaluation. arXiv preprint arXiv:2004.09602","author":"Wu Hao","year":"2020","unstructured":"Hao Wu, Patrick Judd, Xiaojie Zhang, Mikhail Isaev, and Paulius Micikevicius. 2020. Integer quantization for deep learning inference: Principles and empirical evaluation. arXiv preprint arXiv:2004.09602 (2020)."},{"key":"e_1_3_2_1_56_1","volume-title":"Proceedings, Part XII 16","author":"Xu Shoukai","year":"2020","unstructured":"Shoukai Xu, Haokun Li, Bohan Zhuang, Jing Liu, Jiezhang Cao, Chuangrun Liang, and Mingkui Tan. 2020. Generative low-bitwidth data free quantization. In Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XII 16. Springer, 1\u201317."},{"key":"e_1_3_2_1_57_1","volume-title":"Proceedings, Part XII 16","author":"Xu Shoukai","year":"2020","unstructured":"Shoukai Xu, Haokun Li, Bohan Zhuang, Jing Liu, Jiezhang Cao, Chuangrun Liang, and Mingkui Tan. 2020. Generative low-bitwidth data free quantization. In Computer Vision-ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XII 16. Springer, 1\u201317."},{"key":"e_1_3_2_1_58_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.00396"},{"key":"e_1_3_2_1_59_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103277"},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR46437.2021.01540"},{"key":"e_1_3_2_1_61_1","volume-title":"Self-Supervised Quantization-Aware Knowledge Distillation. arXiv preprint arXiv:2403.11106","author":"Zhao Kaiqi","year":"2024","unstructured":"Kaiqi Zhao and Ming Zhao. 2024. Self-Supervised Quantization-Aware Knowledge Distillation. arXiv preprint arXiv:2403.11106 (2024)."},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01202"},{"key":"e_1_3_2_1_63_1","volume-title":"Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160","author":"Zhou Shuchang","year":"2016","unstructured":"Shuchang Zhou, Yuxin Wu, Zekun Ni, Xinyu Zhou, He Wen, and Yuheng Zou. 2016. Dorefa-net: Training low bitwidth convolutional neural networks with low bitwidth gradients. arXiv preprint arXiv:1606.06160 (2016)."},{"key":"e_1_3_2_1_64_1","volume-title":"Neural architecture search with reinforcement learning. arXiv preprint arXiv:1611.01578","author":"Zoph B","year":"2016","unstructured":"B Zoph. 2016. Neural architecture search with reinforcement learning. arXiv preprint arXiv:1611.01578 (2016)."}],"event":{"name":"SAC '25: 40th ACM\/SIGAPP Symposium on Applied Computing","sponsor":["SIGAPP ACM Special Interest Group on Applied Computing"],"location":"Catania International Airport Catania Italy","acronym":"SAC '25"},"container-title":["Proceedings of the 40th ACM\/SIGAPP Symposium on Applied Computing"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3672608.3707816","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3672608.3707816","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:57:33Z","timestamp":1750298253000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3672608.3707816"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,3,31]]},"references-count":64,"alternative-id":["10.1145\/3672608.3707816","10.1145\/3672608"],"URL":"https:\/\/doi.org\/10.1145\/3672608.3707816","relation":{},"subject":[],"published":{"date-parts":[[2025,3,31]]},"assertion":[{"value":"2025-05-14","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}