{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,27]],"date-time":"2026-02-27T00:06:23Z","timestamp":1772150783833,"version":"3.50.1"},"publisher-location":"New York, NY, USA","reference-count":74,"publisher":"ACM","license":[{"start":{"date-parts":[[2022,11,13]],"date-time":"2022-11-13T00:00:00Z","timestamp":1668297600000},"content-version":"vor","delay-in-days":366,"URL":"http:\/\/www.acm.org\/publications\/policies\/copyright_policy#Background"}],"funder":[{"DOI":"10.13039\/100000001","name":"National Science Foundation","doi-asserted-by":"publisher","award":["CNS-1805310"],"award-info":[{"award-number":["CNS-1805310"]}],"id":[{"id":"10.13039\/100000001","id-type":"DOI","asserted-by":"publisher"}]},{"name":"U.S. Army Research, Development and Engineering Command Acquisition Center","award":["W911NF-13-2-0045"],"award-info":[{"award-number":["W911NF-13-2-0045"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2021,11,12]]},"DOI":"10.1145\/3460120.3484570","type":"proceedings-article","created":{"date-parts":[[2021,11,13]],"date-time":"2021-11-13T12:05:27Z","timestamp":1636805127000},"page":"495-515","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":25,"title":["On the Robustness of Domain Constraints"],"prefix":"10.1145","author":[{"given":"Ryan","family":"Sheatsley","sequence":"first","affiliation":[{"name":"The Pennsylvania State University, State College, PA, USA"}]},{"given":"Blaine","family":"Hoak","sequence":"additional","affiliation":[{"name":"The Pennsylvania State University, State College, PA, USA"}]},{"given":"Eric","family":"Pauley","sequence":"additional","affiliation":[{"name":"The Pennsylvania State University, State College, PA, USA"}]},{"given":"Yohan","family":"Beugin","sequence":"additional","affiliation":[{"name":"The Pennsylvania State University, State College, PA, USA"}]},{"given":"Michael J.","family":"Weisman","sequence":"additional","affiliation":[{"name":"United States Army Research Laboratory, Adelphi, MD, USA"}]},{"given":"Patrick","family":"McDaniel","sequence":"additional","affiliation":[{"name":"The Pennsylvania State University, State College, PA, USA"}]}],"member":"320","published-online":{"date-parts":[[2021,11,13]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"177","volume-title":"Machine-Learning-Based Feature Selection Techniques for Large-Scale Network Intrusion Detection. In 2014 IEEE 34th International Conference on Distributed Computing Systems Workshops","author":"Al-Jarrah O.Y.","year":"2014","unstructured":"O.Y. Al-Jarrah, A. Siddiqui, M. Elsalamouny, P.D. Yoo, S. Muhaidat, and K. Kim. Machine-Learning-Based Feature Selection Techniques for Large-Scale Network Intrusion Detection. In 2014 IEEE 34th International Conference on Distributed Computing Systems Workshops, pages 177--181, Madrid, Spain, June 2014. IEEE."},{"key":"e_1_3_2_1_2_1","doi-asserted-by":"publisher","DOI":"10.5555\/3016100.3016102"},{"key":"e_1_3_2_1_3_1","unstructured":"Hyrum S Anderson Anant Kharkar and Bobby Filar. Evading Machine Learning Malware Detection. page 6."},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.1145\/304182.304187"},{"key":"e_1_3_2_1_5_1","volume-title":"Konrad Rieck. Drebin: Effective and Explainable Detection of Android Malware in Your Pocket. In Proceedings 2014 Network and Distributed System Security Symposium","author":"Arp Daniel","year":"2014","unstructured":"Daniel Arp, Michael Spreitzenbarth, Malte H\u00fcbner, Hugo Gascon, and Konrad Rieck. Drebin: Effective and Explainable Detection of Android Malware in Your Pocket. In Proceedings 2014 Network and Distributed System Security Symposium, San Diego, CA, 2014. Internet Society."},{"key":"e_1_3_2_1_6_1","doi-asserted-by":"publisher","DOI":"10.1515\/9781400874668"},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2018.07.023"},{"key":"e_1_3_2_1_8_1","volume-title":"Decision-based adversarial attacks: Reliable attacks against black-box machine learning models. arXiv preprint arXiv:1712.04248","author":"Brendel Wieland","year":"2017","unstructured":"Wieland Brendel, Jonas Rauber, and Matthias Bethge. Decision-based adversarial attacks: Reliable attacks against black-box machine learning models. arXiv preprint arXiv:1712.04248, 2017."},{"key":"e_1_3_2_1_9_1","volume-title":"Adversarial patch. arXiv preprint arXiv:1712.09665","author":"Brown Tom B","year":"2017","unstructured":"Tom B Brown, Dandelion Man\u00e9, Aurko Roy, Mart\u00edn Abadi, and Justin Gilmer. Adversarial patch. arXiv preprint arXiv:1712.09665, 2017."},{"key":"e_1_3_2_1_10_1","doi-asserted-by":"publisher","DOI":"10.1109\/COMST.2015.2494502"},{"key":"e_1_3_2_1_11_1","doi-asserted-by":"publisher","DOI":"10.1145\/3128572.3140444"},{"key":"e_1_3_2_1_12_1","doi-asserted-by":"publisher","DOI":"10.1109\/SP.2017.49"},{"key":"e_1_3_2_1_13_1","doi-asserted-by":"publisher","DOI":"10.1109\/SPW.2018.00009"},{"key":"e_1_3_2_1_14_1","first-page":"1905","author":"Chandrasekaran Varun","year":"2019","unstructured":"Varun Chandrasekaran, Brian Tang, Nicolas Papernot, Kassem Fawaz, Somesh Jha, and Xi Wu. Rearchitecting Classification Frameworks For Increased Robustness, 2019. _eprint: 1905.10900.","journal-title":"Rearchitecting Classification Frameworks For Increased Robustness"},{"key":"e_1_3_2_1_15_1","volume-title":"Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning","author":"Chen Xinyun","year":"2017","unstructured":"Xinyun Chen, Chang Liu, Bo Li, Kimberly Lu, and Dawn Song. Targeted Backdoor Attacks on Deep Learning Systems Using Data Poisoning. 2017. _eprint: 1712.05526."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2019.01.064"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/321033.321034"},{"key":"e_1_3_2_1_18_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2019.00444"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1038\/s41591-018-0316-z"},{"key":"e_1_3_2_1_20_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR.2018.00175"},{"key":"e_1_3_2_1_21_1","volume-title":"Detecting adversarial samples from artifacts. arXiv preprint arXiv:1703.00410","author":"Feinman Reuben","year":"2017","unstructured":"Reuben Feinman, Ryan R Curtin, Saurabh Shintre, and Andrew B Gardner. Detecting adversarial samples from artifacts. arXiv preprint arXiv:1703.00410, 2017."},{"key":"e_1_3_2_1_22_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-540-28645-5_29"},{"key":"e_1_3_2_1_23_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-08867-9_5"},{"key":"e_1_3_2_1_24_1","volume-title":"Popular Science","author":"Gershgorn Dave","year":"2019","unstructured":"Dave Gershgorn. Fooling the machine. Popular Science, April 2019. Publication Title: Popular Science."},{"key":"e_1_3_2_1_25_1","first-page":"2672","volume-title":"Advances in neural information processing systems","author":"Goodfellow Ian","year":"2014","unstructured":"Ian Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, Bing Xu, David Warde-Farley, Sherjil Ozair, Aaron Courville, and Yoshua Bengio. Generative adversarial nets. In Advances in neural information processing systems, pages 2672--2680, 2014."},{"key":"e_1_3_2_1_26_1","volume-title":"Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572","author":"Goodfellow Ian J","year":"2014","unstructured":"Ian J Goodfellow, Jonathon Shlens, and Christian Szegedy. Explaining and harnessing adversarial examples. arXiv preprint arXiv:1412.6572, 2014."},{"key":"e_1_3_2_1_27_1","volume-title":"On the (statistical) detection of adversarial examples. arXiv preprint arXiv:1702.06280","author":"Grosse Kathrin","year":"2017","unstructured":"Kathrin Grosse, Praveen Manoharan, Nicolas Papernot, Michael Backes, and Patrick McDaniel. On the (statistical) detection of adversarial examples. arXiv preprint arXiv:1702.06280, 2017."},{"key":"e_1_3_2_1_28_1","doi-asserted-by":"publisher","DOI":"10.1007\/978-3-319-66399-9_4"},{"key":"e_1_3_2_1_29_1","volume-title":"Deep learning in finance. arXiv preprint arXiv:1602.06561","author":"Heaton JB","year":"2016","unstructured":"JB Heaton, Nicholas G Polson, and Jan Hendrik Witte. Deep learning in finance. arXiv preprint arXiv:1602.06561, 2016."},{"key":"e_1_3_2_1_30_1","volume-title":"Dezeen","author":"Hitti Natashah","year":"2019","unstructured":"Natashah Hitti. Science Museum curator picks five designs for a driverless future. Dezeen, July 2019. Publication Title: Dezeen."},{"key":"e_1_3_2_1_31_1","volume-title":"Black-box adversarial attacks with limited queries and information. arXiv preprint arXiv:1804.08598","author":"Ilyas Andrew","year":"2018","unstructured":"Andrew Ilyas, Logan Engstrom, Anish Athalye, and Jessy Lin. Black-box adversarial attacks with limited queries and information. arXiv preprint arXiv:1804.08598, 2018."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1137\/0222052"},{"key":"e_1_3_2_1_33_1","volume-title":"WIRED UK","author":"Kobie Nicole","year":"2018","unstructured":"Nicole Kobie. To cripple AI, hackers are turning data against itself. WIRED UK, September 2018. Publication Title: WIRED."},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.5555\/3001460.3001502"},{"key":"e_1_3_2_1_35_1","doi-asserted-by":"publisher","DOI":"10.23919\/EUSIPCO.2018.8553214"},{"key":"e_1_3_2_1_36_1","volume-title":"Adversarial examples in the physical world. arXiv preprint arXiv:1607.02533","author":"Kurakin Alexey","year":"2016","unstructured":"Alexey Kurakin, Ian Goodfellow, and Samy Bengio. Adversarial examples in the physical world. arXiv preprint arXiv:1607.02533, 2016."},{"key":"e_1_3_2_1_37_1","volume-title":"Adversarial machine learning at scale. arXiv preprint arXiv:1611.01236","author":"Kurakin Alexey","year":"2017","unstructured":"Alexey Kurakin, Ian Goodfellow, and Samy Bengio. Adversarial machine learning at scale. arXiv preprint arXiv:1611.01236, 2017."},{"key":"e_1_3_2_1_38_1","article-title":"What Happens When Television News Gets The Deep Fake Treatment?","author":"Leetaru Kalev","year":"2019","unstructured":"Kalev Leetaru. What Happens When Television News Gets The Deep Fake Treatment? Forbes Magazine, May 2019. Publication Title: Forbes.","journal-title":"Forbes Magazine"},{"key":"e_1_3_2_1_39_1","unstructured":"Junhong Li. DETECTION OF DDOS ATTACKS BASED ON DENSE NEURAL NETWORKS AUTOENCODERS AND PEARSON CORRELATION COEFFICIENT. page 89."},{"key":"e_1_3_2_1_40_1","volume-title":"IDSGAN: Generative Adversarial Networks for Attack Generation against Intrusion Detection","author":"Lin Zilong","year":"2018","unstructured":"Zilong Lin, Yong Shi, and Zhi Xue. IDSGAN: Generative Adversarial Networks for Attack Generation against Intrusion Detection. 2018. _eprint: 1809.02077."},{"key":"e_1_3_2_1_41_1","volume-title":"Delving into transferable adversarial examples and black-box attacks. arXiv preprint arXiv:1611.02770","author":"Liu Yanpei","year":"2016","unstructured":"Yanpei Liu, Xinyun Chen, Chang Liu, and Dawn Song. Delving into transferable adversarial examples and black-box attacks. arXiv preprint arXiv:1611.02770, 2016."},{"key":"e_1_3_2_1_42_1","volume-title":"Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083","author":"Madry Aleksander","year":"2017","unstructured":"Aleksander Madry, Aleksandar Makelov, Ludwig Schmidt, Dimitris Tsipras, and Adrian Vladu. Towards deep learning models resistant to adversarial attacks. arXiv preprint arXiv:1706.06083, 2017."},{"key":"e_1_3_2_1_43_1","volume-title":"May","author":"Mok Kimberley","year":"2018","unstructured":"Kimberley Mok. Google Develops 'Adversarial Example' Images that Fool Both Humans and Computers. The New Stack, May 2018. Publication Title: The New Stack."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1007\/3-540-57370-4_35"},{"key":"e_1_3_2_1_45_1","volume-title":"cleverhans v1.0.0: an adversarial machine learning library. arXiv preprint arXiv:1610.00768","author":"Papernot Nicolas","year":"2016","unstructured":"Nicolas Papernot, Ian Goodfellow, Ryan Sheatsley, Reuben Feinman, and Patrick McDaniel. cleverhans v1.0.0: an adversarial machine learning library. arXiv preprint arXiv:1610.00768, 2016."},{"key":"e_1_3_2_1_46_1","volume-title":"Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial Samples","author":"Papernot Nicolas","year":"2016","unstructured":"Nicolas Papernot, Patrick McDaniel, and Ian Goodfellow. Transferability in Machine Learning: from Phenomena to Black-Box Attacks using Adversarial Samples. 2016. _eprint: 1605.07277."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1145\/3052973.3053009"},{"key":"e_1_3_2_1_48_1","doi-asserted-by":"publisher","DOI":"10.1109\/EuroSP.2016.36"},{"key":"e_1_3_2_1_49_1","doi-asserted-by":"publisher","DOI":"10.1109\/MILCOM.2016.7795300"},{"key":"e_1_3_2_1_50_1","first-page":"8024","volume-title":"Advances in Neural Information Processing Systems 32","author":"Paszke Adam","year":"2019","unstructured":"Adam Paszke, Sam Gross, Francisco Massa, Adam Lerer, James Bradbury, Gregory Chanan, Trevor Killeen, Zeming Lin, Natalia Gimelshein, Luca Antiga, Alban Desmaison, Andreas Kopf, Edward Yang, Zachary DeVito, Martin Raison, Alykhan Tejani, Sasank Chilamkurthy, Benoit Steiner, Lu Fang, Junjie Bai, and Soumith Chintala. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In H. Wallach, H. Larochelle, A. Beygelzimer, F. dtextbackslashtextquotesingle Alch\u00e9-Buc, E. Fox, and R. Garnett, editors, Advances in Neural Information Processing Systems 32, pages 8024--8035. Curran Associates, Inc., 2019."},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1186\/s41039-017-0062-8"},{"key":"e_1_3_2_1_52_1","unstructured":"Luc De Raedt Andrea Passerini and Stefano Teso. Learning Constraints from Examples. page 6."},{"key":"e_1_3_2_1_53_1","volume-title":"Certified defenses against adversarial examples. arXiv preprint arXiv:1801.09344","author":"Raghunathan Aditi","year":"2018","unstructured":"Aditi Raghunathan, Jacob Steinhardt, and Percy Liang. Certified defenses against adversarial examples. arXiv preprint arXiv:1801.09344, 2018."},{"key":"e_1_3_2_1_54_1","volume-title":"Adversarial Deep Learning Against Intrusion Detection Classifiers","author":"Rigaki Maria","year":"2017","unstructured":"Maria Rigaki. Adversarial Deep Learning Against Intrusion Detection Classifiers. 2017."},{"key":"e_1_3_2_1_55_1","first-page":"1","volume-title":"Developing Realistic Distributed Denial of Service (DDoS) Attack Dataset and Taxonomy. In 2019 International Carnahan Conference on Security Technology (ICCST)","author":"Sharafaldin I.","year":"2019","unstructured":"I. Sharafaldin, A. H. Lashkari, S. Hakak, and A. A. Ghorbani. Developing Realistic Distributed Denial of Service (DDoS) Attack Dataset and Taxonomy. In 2019 International Carnahan Conference on Security Technology (ICCST), pages 1--8, 2019."},{"key":"e_1_3_2_1_56_1","doi-asserted-by":"publisher","DOI":"10.1145\/2976749.2978392"},{"key":"e_1_3_2_1_57_1","volume-title":"A general framework for adversarial examples with objectives. ACM Transactions on Privacy and Security (TOPS), 22(3):1--30","author":"Sharif Mahmood","year":"2019","unstructured":"Mahmood Sharif, Sruti Bhagavatula, Lujo Bauer, and Michael K Reiter. A general framework for adversarial examples with objectives. ACM Transactions on Privacy and Security (TOPS), 22(3):1--30, 2019. Publisher: ACM New York, NY, USA."},{"key":"e_1_3_2_1_58_1","volume-title":"November","author":"Sheatsley Ryan","year":"2020","unstructured":"Ryan Sheatsley, Nicolas Papernot, Michael Weisman, Gunjan Verma, and Patrick McDaniel. Adversarial Examples in Constrained Domains. arXiv:2011.01183 [cs], November 2020. arXiv: 2011.01183."},{"key":"e_1_3_2_1_59_1","first-page":"3520","volume-title":"Percy Liang. Certified Defenses for Data Poisoning Attacks. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS'17","author":"Steinhardt Jacob","year":"2017","unstructured":"Jacob Steinhardt, Pang Wei Koh, and Percy Liang. Certified Defenses for Data Poisoning Attacks. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS'17, pages 3520--3532, Red Hook, NY, USA, 2017. Curran Associates Inc. event-place: Long Beach, California, USA."},{"key":"e_1_3_2_1_60_1","doi-asserted-by":"crossref","first-page":"130","DOI":"10.1109\/DISCEX.2000.821515","volume-title":"Proceedings DARPA Information Survivability Conference and Exposition. DISCEX'00","volume":"2","author":"Stolfo S.J.","year":"2000","unstructured":"S.J. Stolfo, Wei Fan, Wenke Lee, A. Prodromidis, and P.K. Chan. Cost-based modeling for fraud and intrusion detection: results from the JAM project. In Proceedings DARPA Information Survivability Conference and Exposition. DISCEX'00, volume 2, pages 130--144 vol.2, 2000."},{"key":"e_1_3_2_1_61_1","doi-asserted-by":"publisher","DOI":"10.1109\/TEVC.2019.2890858"},{"key":"e_1_3_2_1_62_1","volume-title":"Intriguing properties of neural networks","author":"Szegedy Christian","year":"2013","unstructured":"Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. Intriguing properties of neural networks. 2013. _eprint: 1312.6199."},{"key":"e_1_3_2_1_63_1","doi-asserted-by":"publisher","DOI":"10.1109\/CISDA.2009.5356528"},{"key":"e_1_3_2_1_64_1","doi-asserted-by":"publisher","DOI":"10.1109\/65.642356"},{"key":"e_1_3_2_1_65_1","volume-title":"The space of transferable adversarial examples. arXiv preprint arXiv:1704.03453","author":"Tram\u00e8r Florian","year":"2017","unstructured":"Florian Tram\u00e8r, Nicolas Papernot, Ian Goodfellow, Dan Boneh, and Patrick McDaniel. The space of transferable adversarial examples. arXiv preprint arXiv:1704.03453, 2017."},{"key":"e_1_3_2_1_66_1","volume-title":"Tech. rep.","author":"Uther William","year":"1997","unstructured":"William Uther and Manuela Veloso. Adversarial reinforcement learning. Technical report, Tech. rep., Carnegie Mellon University. Unpublished, 1997."},{"key":"e_1_3_2_1_67_1","doi-asserted-by":"publisher","DOI":"10.1145\/1968.1972"},{"key":"e_1_3_2_1_68_1","doi-asserted-by":"publisher","DOI":"10.1007\/s10618-015-0448-4"},{"key":"e_1_3_2_1_69_1","doi-asserted-by":"publisher","DOI":"10.1023\/A:1018046501280"},{"key":"e_1_3_2_1_70_1","doi-asserted-by":"publisher","DOI":"10.1109\/4236.968834"},{"key":"e_1_3_2_1_71_1","first-page":"5286","volume-title":"International Conference on Machine Learning","author":"Wong Eric","year":"2018","unstructured":"Eric Wong and Zico Kolter. Provable defenses against adversarial examples via the convex outer adversarial polytope. In International Conference on Machine Learning, pages 5286--5295, 2018."},{"key":"e_1_3_2_1_72_1","volume-title":"Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv preprint arXiv:1704.01155","author":"Xu Weilin","year":"2017","unstructured":"Weilin Xu, David Evans, and Yanjun Qi. Feature squeezing: Detecting adversarial examples in deep neural networks. arXiv preprint arXiv:1704.01155, 2017."},{"key":"e_1_3_2_1_73_1","doi-asserted-by":"publisher","DOI":"10.1109\/MILCOM.2018.8599759"},{"key":"e_1_3_2_1_74_1","doi-asserted-by":"publisher","DOI":"10.1145\/3296979.3192416"}],"event":{"name":"CCS '21: 2021 ACM SIGSAC Conference on Computer and Communications Security","location":"Virtual Event Republic of Korea","acronym":"CCS '21","sponsor":["SIGSAC ACM Special Interest Group on Security, Audit, and Control"]},"container-title":["Proceedings of the 2021 ACM SIGSAC Conference on Computer and Communications Security"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3460120.3484570","content-type":"unspecified","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3460120.3484570","content-type":"application\/pdf","content-version":"vor","intended-application":"syndication"},{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3460120.3484570","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,11,18]],"date-time":"2025-11-18T20:44:07Z","timestamp":1763498647000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3460120.3484570"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,11,12]]},"references-count":74,"alternative-id":["10.1145\/3460120.3484570","10.1145\/3460120"],"URL":"https:\/\/doi.org\/10.1145\/3460120.3484570","relation":{},"subject":[],"published":{"date-parts":[[2021,11,12]]},"assertion":[{"value":"2021-11-13","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}