{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,5]],"date-time":"2026-01-05T22:27:23Z","timestamp":1767652043564,"version":"3.41.0"},"publisher-location":"New York, NY, USA","reference-count":68,"publisher":"ACM","license":[{"start":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T00:00:00Z","timestamp":1727654400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0\/"}],"funder":[{"DOI":"10.13039\/100000001","name":"NSF (National Science Foundation)","doi-asserted-by":"publisher","award":["2315612","2315614","2315613","2327480","2348452"],"award-info":[{"award-number":["2315612","2315614","2315613","2327480","2348452"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2024,9,30]]},"DOI":"10.1145\/3678890.3678915","type":"proceedings-article","created":{"date-parts":[[2024,9,29]],"date-time":"2024-09-29T22:23:36Z","timestamp":1727648616000},"page":"496-512","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":1,"title":["Enhancing Model Poisoning Attacks to Byzantine-Robust Federated Learning via Critical Learning Periods"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-7734-1589","authenticated-orcid":false,"given":"Gang","family":"Yan","sequence":"first","affiliation":[{"name":"UC Merced, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1444-2657","authenticated-orcid":false,"given":"Hao","family":"Wang","sequence":"additional","affiliation":[{"name":"Stevens Institute of Technology, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3775-3033","authenticated-orcid":false,"given":"Xu","family":"Yuan","sequence":"additional","affiliation":[{"name":"University of Delaware, United States of America"}]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3642-3569","authenticated-orcid":false,"given":"Jian","family":"Li","sequence":"additional","affiliation":[{"name":"Stony Brook University, USA"}]}],"member":"320","published-online":{"date-parts":[[2024,9,30]]},"reference":[{"key":"e_1_3_2_1_1_1","volume-title":"Proc. of ICLR.","author":"Achille Alessandro","year":"2019","unstructured":"Alessandro Achille, Matteo Rovere, and Stefano Soatto. 2019. Critical Learning Periods in Deep Networks. In Proc. of ICLR."},{"key":"e_1_3_2_1_2_1","volume-title":"Proc. of NeurIPS.","author":"Alistarh Dan","year":"2018","unstructured":"Dan Alistarh, Zeyuan Allen-Zhu, and Jerry Li. 2018. Byzantine stochastic gradient descent. In Proc. of NeurIPS."},{"key":"e_1_3_2_1_3_1","volume-title":"Proc. of ASIA CCS.","author":"Qiang Qiu Saurabh\u00a0Bagchi Joshua Zhao","year":"2023","unstructured":"Joshua Zhao Qiang Qiu Saurabh\u00a0Bagchi Atul\u00a0Sharma, Wei\u00a0Chen and Somali Chaterji. 2023. FLAIR: Defense against Model Poisoning Attack in Federated Learning. In Proc. of ASIA CCS."},{"key":"e_1_3_2_1_4_1","volume-title":"Proc. of USENIX Security Symposium.","author":"Bagdasaryan Eugene","year":"2021","unstructured":"Eugene Bagdasaryan and Vitaly Shmatikov. 2021. Blind backdoors in deep learning models. In Proc. of USENIX Security Symposium."},{"key":"e_1_3_2_1_5_1","volume-title":"Proc. of AISTATS.","author":"Bagdasaryan Eugene","year":"2020","unstructured":"Eugene Bagdasaryan, Andreas Veit, Yiqing Hua, Deborah Estrin, and Vitaly Shmatikov. 2020. How to backdoor federated learning. In Proc. of AISTATS."},{"key":"e_1_3_2_1_6_1","volume-title":"Proc. of NeurIPS.","author":"Baruch Gilad","year":"2019","unstructured":"Gilad Baruch, Moran Baruch, and Yoav Goldberg. 2019. A little is enough: Circumventing defenses for distributed learning. In Proc. of NeurIPS."},{"key":"e_1_3_2_1_7_1","volume-title":"Proc. of ICML.","author":"Bhagoji Arjun\u00a0Nitin","year":"2019","unstructured":"Arjun\u00a0Nitin Bhagoji, Supriyo Chakraborty, Prateek Mittal, and Seraphin Calo. 2019. Analyzing federated learning through an adversarial lens. In Proc. of ICML."},{"key":"e_1_3_2_1_8_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICB.2013.6613006"},{"key":"e_1_3_2_1_9_1","volume-title":"Poisoning attacks against support vector machines. arXiv preprint arXiv:1206.6389","author":"Biggio Battista","year":"2012","unstructured":"Battista Biggio, Blaine Nelson, and Pavel Laskov. 2012. Poisoning attacks against support vector machines. arXiv preprint arXiv:1206.6389 (2012)."},{"key":"e_1_3_2_1_10_1","volume-title":"Proc. of NeurIPS.","author":"Blanchard Peva","year":"2017","unstructured":"Peva Blanchard, El\u00a0Mahdi El\u00a0Mhamdi, Rachid Guerraoui, and Julien Stainer. 2017. Machine learning with adversaries: Byzantine tolerant gradient descent. In Proc. of NeurIPS."},{"key":"e_1_3_2_1_11_1","volume-title":"Proc. of MLSys.","author":"Bonawitz Keith","year":"2019","unstructured":"Keith Bonawitz, Hubert Eichner, Wolfgang Grieskamp, Dzmitry Huba, Alex Ingerman, Vladimir Ivanov, Chloe Kiddon, Jakub Kone\u010dn\u1ef3, Stefano Mazzocchi, Brendan McMahan, 2019. Towards federated learning at scale: System design. In Proc. of MLSys."},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2021.24434"},{"key":"e_1_3_2_1_13_1","volume-title":"MPAF: Model Poisoning Attacks to Federated Learning based on Fake Clients. arXiv preprint arXiv:2203.08669","author":"Cao Xiaoyu","year":"2022","unstructured":"Xiaoyu Cao and Neil\u00a0Zhenqiang Gong. 2022. MPAF: Model Poisoning Attacks to Federated Learning based on Fake Clients. arXiv preprint arXiv:2203.08669 (2022)."},{"key":"e_1_3_2_1_14_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP46215.2023.10179336"},{"key":"e_1_3_2_1_15_1","volume-title":"Towards multi-party targeted model poisoning attacks against federated learning systems. High-Confidence Computing","author":"Chen Zheyi","year":"2021","unstructured":"Zheyi Chen, Pu Tian, Weixian Liao, and Wei Yu. 2021. Towards multi-party targeted model poisoning attacks against federated learning systems. High-Confidence Computing (2021)."},{"key":"e_1_3_2_1_16_1","volume-title":"Proc of ICLR Workshop.","author":"Duygu Nur\u00a0Yaldiz Tuo\u00a0Zhang","year":"2023","unstructured":"Tuo\u00a0Zhang Duygu Nur\u00a0Yaldiz and Salman Avestimehr. 2023. Secure Federated Learning against Model Poisoning Attacks via Client Filtering. In Proc of ICLR Workshop."},{"key":"e_1_3_2_1_17_1","volume-title":"Proc. of ICML.","author":"El\u00a0El\u00a0Mhamdi Mahdi","year":"2018","unstructured":"Mahdi El\u00a0El\u00a0Mhamdi, Rachid Guerraoui, and S\u00e9bastien Rouault. 2018. The hidden vulnerability of distributed learning in byzantium. In Proc. of ICML."},{"key":"e_1_3_2_1_18_1","volume-title":"Proc. of USENIX Security.","author":"Fang Minghong","year":"2020","unstructured":"Minghong Fang, Xiaoyu Cao, Jinyuan Jia, and Neil Gong. 2020. Local Model Poisoning Attacks to Byzantine-Robust Federated Learning. In Proc. of USENIX Security."},{"key":"e_1_3_2_1_19_1","volume-title":"Proc. of ICLR.","author":"Frankle Jonathan","year":"2020","unstructured":"Jonathan Frankle, David\u00a0J Schwab, and Ari\u00a0S Morcos. 2020. The Early Phase of Neural Network Training. In Proc. of ICLR."},{"key":"e_1_3_2_1_20_1","volume-title":"Proc. of NeurIPS","author":"Golatkar Aditya\u00a0Sharad","year":"2019","unstructured":"Aditya\u00a0Sharad Golatkar, Alessandro Achille, and Stefano Soatto. 2019. Time Matters in Regularizing Deep Networks: Weight Decay and Data Augmentation Affect Early Learning Dynamics, Matter Little Near Convergence. Proc. of NeurIPS (2019)."},{"key":"e_1_3_2_1_21_1","doi-asserted-by":"publisher","DOI":"10.1080\/00031305.1978.10479256"},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1214\/aoms\/1177700066"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2016.90"},{"key":"e_1_3_2_1_24_1","unstructured":"Lie He Sai\u00a0Praneeth Karimireddy and Martin Jaggi. 2020. Byzantine-robust learning on heterogeneous datasets via resampling. (2020)."},{"key":"e_1_3_2_1_25_1","volume-title":"Proc. of ICML.","author":"Hsieh Kevin","year":"2020","unstructured":"Kevin Hsieh, Amar Phanishayee, Onur Mutlu, and Phillip Gibbons. 2020. The non-iid data quagmire of decentralized machine learning. In Proc. of ICML."},{"key":"e_1_3_2_1_26_1","doi-asserted-by":"publisher","DOI":"10.1109\/JIOT.2021.3095077"},{"key":"e_1_3_2_1_27_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2018.00057"},{"key":"e_1_3_2_1_28_1","volume-title":"Proc. of ICML.","author":"Jastrzebski Stanislaw","year":"2021","unstructured":"Stanislaw Jastrzebski, Devansh Arpit, Oliver Astrand, Giancarlo\u00a0B Kerg, Huan Wang, Caiming Xiong, Richard Socher, Kyunghyun Cho, and Krzysztof\u00a0J Geras. 2021. Catastrophic Fisher Explosion: Early Phase Fisher Matrix Impacts Generalization. In Proc. of ICML."},{"key":"e_1_3_2_1_29_1","volume-title":"Proc. of ICLR.","author":"Jastrzebski Stanislaw","year":"2019","unstructured":"Stanislaw Jastrzebski, Zachary Kenton, Nicolas Ballas, Asja Fischer, Yoshua Bengio, and Amos\u00a0J Storkey. 2019. On the Relation Between the Sharpest Directions of DNN Loss and the SGD Step Length. In Proc. of ICLR."},{"key":"e_1_3_2_1_30_1","volume-title":"Advances and Open Problems in Federated Learning. arXiv preprint arXiv:1912.04977","author":"Kairouz Peter","year":"2019","unstructured":"Peter Kairouz, H\u00a0Brendan McMahan, Brendan Avent, Aur\u00e9lien Bellet, Mehdi Bennis, Arjun\u00a0Nitin Bhagoji, Kallista Bonawitz, Zachary Charles, Graham Cormode, Rachel Cummings, 2019. Advances and Open Problems in Federated Learning. arXiv preprint arXiv:1912.04977 (2019)."},{"key":"e_1_3_2_1_31_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v30i1.10362"},{"key":"e_1_3_2_1_32_1","unstructured":"Alex Krizhevsky Geoffrey Hinton 2009. Learning Multiple Layers of Features from Tiny Images. (2009)."},{"key":"e_1_3_2_1_33_1","volume-title":"Proc. of NIPS","author":"Krizhevsky Alex","year":"2012","unstructured":"Alex Krizhevsky, Ilya Sutskever, and Geoffrey\u00a0E Hinton. 2012. Imagenet Classification with Deep Convolutional Neural Networks. Proc. of NIPS (2012)."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/5.726791"},{"key":"e_1_3_2_1_35_1","volume-title":"Proc. of NIPS.","author":"Li Bo","year":"2016","unstructured":"Bo Li, Yining Wang, Aarti Singh, and Yevgeniy Vorobeychik. 2016. Data poisoning attacks on factorization-based collaborative filtering. In Proc. of NIPS."},{"key":"e_1_3_2_1_36_1","volume-title":"Proc. of MLSys.","author":"Li Tian","year":"2020","unstructured":"Tian Li, Anit\u00a0Kumar Sahu, Manzil Zaheer, Maziar Sanjabi, Ameet Talwalkar, and Virginia Smith. 2020. Federated Optimization in Heterogeneous Networks. In Proc. of MLSys."},{"key":"e_1_3_2_1_37_1","volume-title":"Proc. of ICLR.","author":"Li Xiang","year":"2020","unstructured":"Xiang Li, Kaixuan Huang, Wenhao Yang, Shusen Wang, and Zhihua Zhang. 2020. On the Convergence of FedAvg on Non-IID Data. In Proc. of ICLR."},{"key":"e_1_3_2_1_38_1","volume-title":"Proc. of ICML.","author":"Mahloujifar Saeed","year":"2019","unstructured":"Saeed Mahloujifar, Mohammad Mahmoody, and Ameer Mohammed. 2019. Universal multi-party poisoning attacks. In Proc. of ICML."},{"key":"e_1_3_2_1_39_1","volume-title":"Proc. of AISTATS.","author":"McMahan Brendan","year":"2017","unstructured":"Brendan McMahan, Eider Moore, Daniel Ramage, Seth Hampson, and Blaise\u00a0Aguera y Arcas. 2017. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proc. of AISTATS."},{"key":"e_1_3_2_1_40_1","doi-asserted-by":"publisher","DOI":"10.1145\/3128572.3140451"},{"key":"e_1_3_2_1_41_1","volume-title":"Byzantine-robust federated machine learning through adaptive model averaging. arXiv preprint arXiv:1909.05125","author":"Mu\u00f1oz-Gonz\u00e1lez Luis","year":"2019","unstructured":"Luis Mu\u00f1oz-Gonz\u00e1lez, Kenneth\u00a0T Co, and Emil\u00a0C Lupu. 2019. Byzantine-robust federated machine learning through adaptive model averaging. arXiv preprint arXiv:1909.05125 (2019)."},{"key":"e_1_3_2_1_42_1","volume-title":"Proc. of USENIX Security Symposium.","author":"Nguyen Thien\u00a0Duc","year":"2022","unstructured":"Thien\u00a0Duc Nguyen, Phillip Rieger, Huili Chen, Hossein Yalame, Helen M\u00f6llering, and Hossein Fereidooni. 2022. FLAME: Taming backdoors in federated learning. In Proc. of USENIX Security Symposium."},{"key":"e_1_3_2_1_43_1","volume-title":"Proc. of AISTATS.","author":"Panda Ashwinee","year":"2022","unstructured":"Ashwinee Panda and Saeed Mahloujifar. 2022. SparseFed: Mitigating Model Poisoning Attacks in Federated Learning with Sparsification. In Proc. of AISTATS."},{"key":"e_1_3_2_1_44_1","unstructured":"Adam Paszke Sam Gross Soumith Chintala Gregory Chanan Edward Yang Zachary DeVito Zeming Lin Alban Desmaison Luca Antiga and Adam Lerer. 2017. Automatic differentiation in pytorch. In NIPS-W."},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1145\/1835449.1835482"},{"key":"e_1_3_2_1_46_1","doi-asserted-by":"publisher","DOI":"10.1145\/1644893.1644895"},{"key":"e_1_3_2_1_47_1","volume-title":"Poisoning Attacks and Defenses in Federated Learning: A Survey. arXiv preprint arXiv:2301.05795","author":"Sagar Subhash","year":"2023","unstructured":"Subhash Sagar, Chang-Tsun Li, Seng\u00a0W. Loke, and Jinho Choi. 2023. Poisoning Attacks and Defenses in Federated Learning: A Survey. arXiv preprint arXiv:2301.05795 (2023)."},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2021.24498"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP46214.2022.9833647"},{"key":"e_1_3_2_1_50_1","volume-title":"MPHM: Model poisoning attacks on federal learning using historical information momentum. Security and Safety","author":"Shi Lei","year":"2023","unstructured":"Lei Shi, Zhen Chen, Yucheng Shi, Lin Wei, Yongcai Tao, Mengyang He, Qingxian Wang, Yuan Zhou, and Yufei Gao. 2023. MPHM: Model poisoning attacks on federal learning using historical information momentum. Security and Safety (2023)."},{"key":"e_1_3_2_1_51_1","volume-title":"Proc. of ICLR.","author":"Simonyan Karen","year":"2015","unstructured":"Karen Simonyan and Andrew Zisserman. 2015. Very Deep Convolutional Networks for Large-scale Image Recognition. In Proc. of ICLR."},{"key":"e_1_3_2_1_52_1","volume-title":"Can you really backdoor federated learning?arXiv preprint arXiv:1911.07963","author":"Sun Ziteng","year":"2019","unstructured":"Ziteng Sun, Peter Kairouz, Ananda\u00a0Theertha Suresh, and H\u00a0Brendan McMahan. 2019. Can you really backdoor federated learning?arXiv preprint arXiv:1911.07963 (2019)."},{"key":"e_1_3_2_1_53_1","volume-title":"Proc. of NeurIPS.","author":"Wang Hongyi","year":"2020","unstructured":"Hongyi Wang, Kartik Sreenivasan, Shashank Rajput, Harit Vishwakarma, Saurabh Agarwal, Jy-yong Sohn, Kangwook Lee, and Dimitris Papailiopoulos. 2020. Attack of the tails: Yes, you really can backdoor federated learning. In Proc. of NeurIPS."},{"key":"e_1_3_2_1_54_1","volume-title":"Proc. of ICLR.","author":"Wang Hongyi","year":"2020","unstructured":"Hongyi Wang, Mikhail Yurochkin, Yuekai Sun, Dimitris Papailiopoulos, and Yasaman Khazaeni. 2020. Federated Learning with Matched Averaging. In Proc. of ICLR."},{"key":"e_1_3_2_1_55_1","volume-title":"Proc. of NeurIPS","author":"Wang Jianyu","year":"2020","unstructured":"Jianyu Wang, Qinghua Liu, Hao Liang, Gauri Joshi, and H\u00a0Vincent Poor. 2020. Tackling the Objective Inconsistency Problem in Heterogeneous Federated Optimization. Proc. of NeurIPS (2020)."},{"key":"e_1_3_2_1_56_1","volume-title":"Proc. of ICML.","author":"Xiao Huang","year":"2015","unstructured":"Huang Xiao, Battista Biggio, Gavin Brown, Giorgio Fumera, Claudia Eckert, and Fabio Roli. 2015. Is feature selection secure against training data poisoning?. In Proc. of ICML."},{"key":"e_1_3_2_1_57_1","volume-title":"Proc. of ICML.","author":"Xie Chulin","year":"2021","unstructured":"Chulin Xie, Minghao Chen, Pin-Yu Chen, and Bo Li. 2021. CRFL:Certifiably Robust Federated Learning against Backdoor Attacks. In Proc. of ICML."},{"key":"e_1_3_2_1_58_1","volume-title":"Proc. of ICLR.","author":"Xie Chulin","year":"2019","unstructured":"Chulin Xie, Keli Huang, Pin-Yu Chen, and Bo Li. 2019. Dba: Distributed backdoor attacks against federated learning. In Proc. of ICLR."},{"key":"e_1_3_2_1_59_1","volume-title":"Generalized byzantine-tolerant sgd. arXiv preprint arXiv:1802.10116","author":"Xie Cong","year":"2018","unstructured":"Cong Xie, Oluwasanmi Koyejo, and Indranil Gupta. 2018. Generalized byzantine-tolerant sgd. arXiv preprint arXiv:1802.10116 (2018)."},{"key":"e_1_3_2_1_60_1","volume-title":"Proc. of UAI.","author":"Xie Cong","year":"2020","unstructured":"Cong Xie, Oluwasanmi Koyejo, and Indranil Gupta. 2020. Fall of empires: Breaking byzantine-tolerant sgd by inner product manipulation. In Proc. of UAI."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1145\/3576915.3623193"},{"key":"e_1_3_2_1_62_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v36i8.20859"},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1145\/3580305.3599293"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1609\/aaai.v37i9.26271"},{"key":"e_1_3_2_1_65_1","doi-asserted-by":"publisher","DOI":"10.14722\/ndss.2017.23020"},{"key":"e_1_3_2_1_66_1","volume-title":"Proc. of ICML.","author":"Yin Dong","year":"2018","unstructured":"Dong Yin, Yudong Chen, Ramchandran Kannan, and Peter Bartlett. 2018. Byzantine-robust distributed learning: Towards optimal statistical rates. In Proc. of ICML."},{"key":"e_1_3_2_1_67_1","volume-title":"Proc. of ICML.","author":"Zhu Chaoyi","year":"2023","unstructured":"Chaoyi Zhu, Stefanie Roos, and Lydia\u00a0Y. Chen. 2023. LeadFL: Client Self-Defense against Model Poisoning in Federated Learning. In Proc. of ICML."},{"key":"e_1_3_2_1_68_1","volume-title":"Backdoor Federated Learning by Poisoning Backdoor-Critical Layers. arXiv preprint arXiv:2308.04466","author":"Zhuang Haomin","year":"2023","unstructured":"Haomin Zhuang, Mingxian Yu, Hao Wang, Yang Hua, Jian Li, and Xu Yuan. 2023. Backdoor Federated Learning by Poisoning Backdoor-Critical Layers. arXiv preprint arXiv:2308.04466 (2023)."}],"event":{"name":"RAID '24: The 27th International Symposium on Research in Attacks, Intrusions and Defenses","acronym":"RAID '24","location":"Padua Italy"},"container-title":["The 27th International Symposium on Research in Attacks, Intrusions and Defenses"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3678890.3678915","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3678890.3678915","content-type":"text\/html","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3678890.3678915","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,19]],"date-time":"2025-06-19T01:18:00Z","timestamp":1750295880000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3678890.3678915"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,9,30]]},"references-count":68,"alternative-id":["10.1145\/3678890.3678915","10.1145\/3678890"],"URL":"https:\/\/doi.org\/10.1145\/3678890.3678915","relation":{},"subject":[],"published":{"date-parts":[[2024,9,30]]},"assertion":[{"value":"2024-09-30","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}