{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,27]],"date-time":"2026-04-27T14:41:43Z","timestamp":1777300903751,"version":"3.51.4"},"publisher-location":"New York, NY, USA","reference-count":55,"publisher":"ACM","funder":[{"name":"National Key R&D Program of China","award":["2023YFB3106404"],"award-info":[{"award-number":["2023YFB3106404"]}]},{"name":"National Natural Science Foundations of China","award":["62302442"],"award-info":[{"award-number":["62302442"]}]}],"content-domain":{"domain":["dl.acm.org"],"crossmark-restriction":true},"short-container-title":[],"published-print":{"date-parts":[[2026,4,13]]},"DOI":"10.1145\/3774904.3792211","type":"proceedings-article","created":{"date-parts":[[2026,4,27]],"date-time":"2026-04-27T13:28:36Z","timestamp":1777296516000},"page":"2695-2706","update-policy":"https:\/\/doi.org\/10.1145\/crossmark-policy","source":"Crossref","is-referenced-by-count":0,"title":["Privacy-Friendly Adaptation of Vision Transformers for Communication and Latency-Efficient Private Inference"],"prefix":"10.1145","author":[{"ORCID":"https:\/\/orcid.org\/0009-0004-5233-0113","authenticated-orcid":false,"given":"Zhi","family":"Pang","sequence":"first","affiliation":[{"name":"School of Cyber Science and Engineering, Wuhan University, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9764-7457","authenticated-orcid":false,"given":"Bo","family":"Feng","sequence":"additional","affiliation":[{"name":"The State Key Laboratory of Blockchain and Data Security, Zhejiang University, Hangzhou, China and Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0001-9018-1367","authenticated-orcid":false,"given":"Meng","family":"Luo","sequence":"additional","affiliation":[{"name":"The State Key Laboratory of Blockchain and Data Security, Zhejiang University, Hangzhou, China and Hangzhou High-Tech Zone (Binjiang) Institute of Blockchain and Data Security, Hangzhou, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0004-9790-2250","authenticated-orcid":false,"given":"Chenhao","family":"Liu","sequence":"additional","affiliation":[{"name":"School of Cyber Science and Engineering, Wuhan University, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0009-0007-8460-3747","authenticated-orcid":false,"given":"Shuwang","family":"Xu","sequence":"additional","affiliation":[{"name":"School of Cyber Science and Engineering, Wuhan University, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7410-3112","authenticated-orcid":false,"given":"Kai","family":"Zhao","sequence":"additional","affiliation":[{"name":"School of Cyber Science and Engineering, Wuhan University, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6779-4226","authenticated-orcid":false,"given":"Yadi","family":"Wu","sequence":"additional","affiliation":[{"name":"School of Cyber Science and Engineering, Wuhan University, Wuhan, China"}]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3334-6150","authenticated-orcid":false,"given":"Bo","family":"Zeng","sequence":"additional","affiliation":[{"name":"School of Cyber Science and Engineering, Wuhan University, Wuhan, China"}]}],"member":"320","published-online":{"date-parts":[[2026,4,12]]},"reference":[{"key":"e_1_3_2_1_1_1","first-page":"2173","volume-title":"33rd USENIX Security Symposium (USENIX Security 24)","author":"Ao Wei","year":"2024","unstructured":"Wei Ao and Vishnu Naresh Boddeti. 2024. AutoFHE: Automated Adaption of CNNs for Efficient Evaluation over FHE. In 33rd USENIX Security Symposium (USENIX Security 24). 2173-2190."},{"key":"e_1_3_2_1_2_1","volume-title":"Leveraging redundancy in attention with reuse transformers. arXiv preprint arXiv:2110.06821","author":"Bhojanapalli Srinadh","year":"2021","unstructured":"Srinadh Bhojanapalli, Ayan Chakrabarti, Andreas Veit, Michal Lukasik, Himanshu Jain, Frederick Liu, Yin-Wen Chang, and Sanjiv Kumar. 2021. Leveraging redundancy in attention with reuse transformers. arXiv preprint arXiv:2110.06821 (2021)."},{"key":"e_1_3_2_1_3_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCAD57390.2023.10323702"},{"key":"e_1_3_2_1_4_1","doi-asserted-by":"publisher","DOI":"10.52202\/079017-1459"},{"key":"e_1_3_2_1_5_1","volume-title":"International Conference on Machine Learning. PMLR, 3947-3961","author":"Cho Minsu","year":"2022","unstructured":"Minsu Cho, Ameya Joshi, Brandon Reagen, Siddharth Garg, and Chinmay Hegde. 2022. Selective network linearization for efficient private inference. In International Conference on Machine Learning. PMLR, 3947-3961."},{"key":"e_1_3_2_1_6_1","volume-title":"Patrick Yubeaton, Minsu Cho, Ameya Joshi, Siddharth Garg, Brandon Reagen, and Chinmay Hegde.","author":"Dhyani Naren","year":"2024","unstructured":"Naren Dhyani, Jianqiao Cambridge Mo, Patrick Yubeaton, Minsu Cho, Ameya Joshi, Siddharth Garg, Brandon Reagen, and Chinmay Hegde. 2024. PriViT: Vision Transformers for Private Inference. Transactions on Machine Learning Research (2024)."},{"key":"e_1_3_2_1_7_1","doi-asserted-by":"publisher","DOI":"10.1145\/3543507.3583272"},{"key":"e_1_3_2_1_8_1","volume-title":"International Conference on Learning Representations.","author":"Dosovitskiy Alexey","year":"2021","unstructured":"Alexey Dosovitskiy, Lucas Beyer, Alexander Kolesnikov, Dirk Weissenborn, Xiaohua Zhai, Thomas Unterthiner, Mostafa Dehghani, Matthias Minderer, Georg Heigold, Sylvain Gelly, Jakob Uszkoreit, and Neil Houlsby. 2021. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_9_1","volume-title":"International Conference on Machine Learning. PMLR, 201-210","author":"Dowlin Nathan","year":"2016","unstructured":"Nathan Dowlin, Ran Gilad-Bachrach, Kim Laine, Kristin Lauter, Michael Naehrig, and John Wernsing. 2016. Cryptonets: Applying neural networks to encrypted data with high throughput and accuracy. In International Conference on Machine Learning. PMLR, 201-210."},{"key":"e_1_3_2_1_10_1","first-page":"16961","article-title":"Cryptonas: Private inference on a relu budget","volume":"33","author":"Ghodsi Zahra","year":"2020","unstructured":"Zahra Ghodsi, Akshaj Kumar Veldanda, Brandon Reagen, and Siddharth Garg. 2020. Cryptonas: Private inference on a relu budget. Advances in Neural Information Processing Systems, Vol. 33 (2020), 16961-16971.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_11_1","volume-title":"Iron: Private inference on transformers. Advances in neural information processing systems","author":"Hao Meng","year":"2022","unstructured":"Meng Hao, Hongwei Li, Hanxiao Chen, Pengzhi Xing, Guowen Xu, and Tianwei Zhang. 2022. Iron: Private inference on transformers. Advances in neural information processing systems, Vol. 35 (2022), 15718-15731."},{"key":"e_1_3_2_1_12_1","volume-title":"Escaping the big data paradigm with compact transformers. arXiv preprint arXiv:2104.05704","author":"Hassani Ali","year":"2021","unstructured":"Ali Hassani, Steven Walton, Nikhil Shah, Abulikemu Abuduweili, Jiachen Li, and Humphrey Shi. 2021. Escaping the big data paradigm with compact transformers. arXiv preprint arXiv:2104.05704 (2021)."},{"key":"e_1_3_2_1_13_1","volume-title":"Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531","author":"Hinton Geoffrey","year":"2015","unstructured":"Geoffrey Hinton, Oriol Vinyals, and Jeff Dean. 2015. Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)."},{"key":"e_1_3_2_1_14_1","first-page":"809","volume-title":"31st USENIX Security Symposium (USENIX Security 22)","author":"Huang Zhicong","year":"2022","unstructured":"Zhicong Huang, Wen-jie Lu, Cheng Hong, and Jiansheng Ding. 2022. Cheetah: Lean and fast secure two-party deep neural network inference. In 31st USENIX Security Symposium (USENIX Security 22). 809-826."},{"key":"e_1_3_2_1_15_1","volume-title":"International Conference on Machine Learning. PMLR, 4839-4849","author":"Jha Nandan Kumar","year":"2021","unstructured":"Nandan Kumar Jha, Zahra Ghodsi, Siddharth Garg, and Brandon Reagen. 2021. Deepreduce: Relu reduction for fast private inference. In International Conference on Machine Learning. PMLR, 4839-4849."},{"key":"e_1_3_2_1_16_1","doi-asserted-by":"publisher","DOI":"10.1145\/3243734.3243837"},{"key":"e_1_3_2_1_17_1","doi-asserted-by":"publisher","DOI":"10.1145\/3658644.3690375"},{"key":"e_1_3_2_1_18_1","first-page":"4961","article-title":"Crypten: Secure multi-party computation meets machine learning","volume":"34","author":"Knott Brian","year":"2021","unstructured":"Brian Knott, Shobha Venkataraman, Awni Hannun, Shubho Sengupta, Mark Ibrahim, and Laurens van der Maaten. 2021. Crypten: Secure multi-party computation meets machine learning. Advances in Neural Information Processing Systems, Vol. 34 (2021), 4961-4973.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_19_1","doi-asserted-by":"publisher","DOI":"10.1109\/WACV57701.2024.00259"},{"key":"e_1_3_2_1_20_1","volume-title":"International Conference on Machine Learning. PMLR, 3519-3529","author":"Kornblith Simon","year":"2019","unstructured":"Simon Kornblith, Mohammad Norouzi, Honglak Lee, and Geoffrey Hinton. 2019. Similarity of neural network representations revisited. In International Conference on Machine Learning. PMLR, 3519-3529."},{"key":"e_1_3_2_1_21_1","volume-title":"International Conference on Learning Representations.","author":"Kundu Souvik","year":"2023","unstructured":"Souvik Kundu, Shunlin Lu, Yuke Zhang, Jacqueline Liu, and Peter A Beerel. 2023. Learning to linearize deep neural networks for secure and efficient private inference. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_22_1","volume-title":"International Conference on Learning Representations.","author":"Li Dacheng","year":"2023","unstructured":"Dacheng Li, Rulin Shao, Hongyi Wang, Han Guo, Eric P Xing, and Hao Zhang. 2023. Mpcformer: fast, performant and private transformer inference with mpc. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_23_1","volume-title":"Seesaw: Compensating for Nonlinear Reduction with Linear Computations for Private Inference. In International Conference on Machine Learning. PMLR.","author":"Li Fabing","year":"2024","unstructured":"Fabing Li, Yuanhao Zhai, Shuangyu Cai, and mingyu Gao. 2024b. Seesaw: Compensating for Nonlinear Reduction with Linear Computations for Private Inference. In International Conference on Machine Learning. PMLR."},{"key":"e_1_3_2_1_24_1","first-page":"21572","article-title":"Nimbus: Secure and Efficient Two-Party Inference for Transformers","volume":"37","author":"Li Zhengyi","year":"2024","unstructured":"Zhengyi Li, Kang Yang, Jin Tan, Wen-jie Lu, Haoqi Wu, Xiao Wang, Yu Yu, Derun Zhao, Yancheng Zheng, Minyi Guo, et al., 2024a. Nimbus: Secure and Efficient Two-Party Inference for Transformers. Advances in Neural Information Processing Systems, Vol. 37 (2024), 21572-21600.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_25_1","volume-title":"International Conference on Learning Representations.","author":"Liang Chen","year":"2023","unstructured":"Chen Liang, Haoming Jiang, Zheng Li, Xianfeng Tang, Bin Yin, and Tuo Zhao. 2023. Homodistil: Homotopic task-agnostic distillation of pre-trained transformers. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_26_1","volume-title":"International Conference on Learning Representations.","author":"Liu Yong","year":"2024","unstructured":"Yong Liu, Tengge Hu, Haoran Zhang, Haixu Wu, Shiyu Wang, Lintao Ma, and Mingsheng Long. 2024. itransformer: Inverted transformers are effective for time series forecasting. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_27_1","first-page":"21297","article-title":"Soft: Softmax-free transformer with linear complexity","volume":"34","author":"Lu Jiachen","year":"2021","unstructured":"Jiachen Lu, Jinghan Yao, Junge Zhang, Xiatian Zhu, Hang Xu, Weiguo Gao, Chunjing Xu, Tao Xiang, and Li Zhang. 2021. Soft: Softmax-free transformer with linear complexity. Advances in Neural Information Processing Systems, Vol. 34 (2021), 21297-21309.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_28_1","unstructured":"Wen-jie Lu Zhicong Huang Zhen Gu Jingyu Li Jian Liu Cheng Hong Kui Ren Tao Wei and WenGuang Chen. 2025. BumbleBee: Secure two-party inference framework for large transformers. In Network and Distributed System Security (NDSS)."},{"key":"e_1_3_2_1_29_1","first-page":"17","volume-title":"SecretFlow-SPU: A Performant and User-Friendly Framework for Privacy-Preserving Machine Learning. In 2023 USENIX Annual Technical Conference (USENIX ATC 23)","author":"Ma Junming","year":"2023","unstructured":"Junming Ma, Yancheng Zheng, Jun Feng, Derun Zhao, Haoqi Wu, Wenjing Fang, Jin Tan, Chaofan Yu, Benyu Zhang, and Lei Wang. 2023. SecretFlow-SPU: A Performant and User-Friendly Framework for Privacy-Preserving Machine Learning. In 2023 USENIX Annual Technical Conference (USENIX ATC 23). 17-33."},{"key":"e_1_3_2_1_30_1","first-page":"2505","volume-title":"Delphi: A Cryptographic Inference Service for Neural Networks. In 29th USENIX Security Symposium (USENIX Security 20)","author":"Mishra Pratyush","year":"2020","unstructured":"Pratyush Mishra, Ryan Lehmkuhl, Akshayaram Srinivasan, Wenting Zheng, and Raluca Ada Popa. 2020. Delphi: A Cryptographic Inference Service for Neural Networks. In 29th USENIX Security Symposium (USENIX Security 20). 2505-2522."},{"key":"e_1_3_2_1_31_1","volume-title":"SecureML: A System for Scalable Privacy-Preserving Machine Learning. In 2017 IEEE Symposium on Security and Privacy (SP). IEEE Computer Society, 19-38","author":"Mohassel Payman","year":"2017","unstructured":"Payman Mohassel and Yupeng Zhang. 2017. SecureML: A System for Scalable Privacy-Preserving Machine Learning. In 2017 IEEE Symposium on Security and Privacy (SP). IEEE Computer Society, 19-38."},{"key":"e_1_3_2_1_32_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.ins.2024.120463"},{"key":"e_1_3_2_1_33_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.neunet.2025.107307"},{"key":"e_1_3_2_1_34_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00478"},{"key":"e_1_3_2_1_35_1","first-page":"20679","article-title":"Lingcn: Structural linearized graph convolutional network for homomorphically encrypted inference","volume":"36","author":"Peng Hongwu","year":"2023","unstructured":"Hongwu Peng, Ran Ran, Yukui Luo, Jiahui Zhao, Shaoyi Huang, Kiran Thorat, Tong Geng, Chenghong Wang, Xiaolin Xu, Wujie Wen, et al., 2023b. Lingcn: Structural linearized graph convolutional network for homomorphically encrypted inference. Advances in Neural Information Processing Systems, Vol. 36 (2023), 20679-20694.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_36_1","doi-asserted-by":"publisher","DOI":"10.1109\/DAC56929.2023.10247663"},{"key":"e_1_3_2_1_37_1","volume-title":"International Conference on Learning Representations.","author":"Qin Zhen","year":"2022","unstructured":"Zhen Qin, Weixuan Sun, Hui Deng, Dongxu Li, Yunshen Wei, Baohong Lv, Junjie Yan, Lingpeng Kong, and Yiran Zhong. 2022. cosFormer: Rethinking Softmax In Attention. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_38_1","doi-asserted-by":"publisher","DOI":"10.1145\/3372297.3417274"},{"key":"e_1_3_2_1_39_1","first-page":"20378","article-title":"Movement pruning: Adaptive sparsity by fine-tuning","volume":"33","author":"Sanh Victor","year":"2020","unstructured":"Victor Sanh, Thomas Wolf, and Alexander Rush. 2020. Movement pruning: Adaptive sparsity by fine-tuning. Advances in Neural Information Processing Systems, Vol. 33 (2020), 20378-20389.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_40_1","volume-title":"Ufo-vit: High performance linear vision transformer without softmax. arXiv preprint arXiv:2109.14382","year":"2021","unstructured":"Jeong-geun Song. 2021. Ufo-vit: High performance linear vision transformer without softmax. arXiv preprint arXiv:2109.14382 (2021)."},{"key":"e_1_3_2_1_41_1","volume-title":"How to train your vit? data, augmentation, and regularization in vision transformers. arXiv preprint arXiv:2106.10270","author":"Steiner Andreas","year":"2021","unstructured":"Andreas Steiner, Alexander Kolesnikov, Xiaohua Zhai, Ross Wightman, Jakob Uszkoreit, and Lucas Beyer. 2021. How to train your vit? data, augmentation, and regularization in vision transformers. arXiv preprint arXiv:2106.10270 (2021)."},{"key":"e_1_3_2_1_42_1","first-page":"35612","article-title":"Cemiface: Center-based semi-hard synthetic face generation for face recognition","volume":"37","author":"Sun Zhonglin","year":"2024","unstructured":"Zhonglin Sun, Siyang Song, Ioannis Patras, and Georgios Tzimiropoulos. 2024. Cemiface: Center-based semi-hard synthetic face generation for face recognition. Advances in Neural Information Processing Systems, Vol. 37 (2024), 35612-35638.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"e_1_3_2_1_43_1","volume-title":"International Conference on Learning Representations.","author":"Venkataramanan Shashanka","year":"2024","unstructured":"Shashanka Venkataramanan, Amir Ghodrati, Yuki M Asano, Fatih Porikli, and Amirhossein Habibian. 2024. Skip-attention: Improving vision transformers by paying less attention. In International Conference on Learning Representations."},{"key":"e_1_3_2_1_44_1","doi-asserted-by":"publisher","DOI":"10.1145\/3696410.3714580"},{"key":"e_1_3_2_1_45_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"e_1_3_2_1_46_1","volume-title":"Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768","author":"Wang Sinong","year":"2020","unstructured":"Sinong Wang, Belinda Z Li, Madian Khabsa, Han Fang, and Hao Ma. 2020a. Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768 (2020)."},{"key":"e_1_3_2_1_47_1","doi-asserted-by":"publisher","DOI":"10.1109\/CVPR52688.2022.01716"},{"key":"e_1_3_2_1_48_1","volume-title":"Privcirnet: Efficient private inference via block circulant transformation. Advances in Neural Information Processing Systems","author":"Xu Tianshi","year":"2024","unstructured":"Tianshi Xu, Lemeng Wu, Runsheng Wang, and Meng Li. 2024. Privcirnet: Efficient private inference via block circulant transformation. Advances in Neural Information Processing Systems (2024)."},{"key":"e_1_3_2_1_49_1","volume-title":"Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. arXiv preprint arXiv:1612.03928","author":"Zagoruyko Sergey","year":"2016","unstructured":"Sergey Zagoruyko and Nikos Komodakis. 2016. Paying more attention to attention: Improving the performance of convolutional neural networks via attention transfer. arXiv preprint arXiv:1612.03928 (2016)."},{"key":"e_1_3_2_1_50_1","doi-asserted-by":"publisher","DOI":"10.1016\/j.aei.2025.104081"},{"key":"e_1_3_2_1_51_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00466"},{"key":"e_1_3_2_1_52_1","volume-title":"Advances in Neural Information Processing Systems","volume":"36","author":"Zeng Wenxuan","year":"2023","unstructured":"Wenxuan Zeng, Meng Li, Haichuan Yang, Wen-jie Lu, Runsheng Wang, and Ru Huang. 2023b. CoPriv: Network\/Protocol Co-Optimization for Communication-Efficient Private Inference. Advances in Neural Information Processing Systems, Vol. 36 (2023)."},{"key":"e_1_3_2_1_53_1","volume-title":"Towards Efficient Privacy-Preserving Machine Learning: A Systematic Review from Protocol, Model, and System Perspectives. arXiv preprint arXiv:2507.14519","author":"Zeng Wenxuan","year":"2025","unstructured":"Wenxuan Zeng, Tianshi Xu, Yi Chen, Yifan Zhou, Mingzhe Zhang, Jin Tan, Cheng Hong, and Meng Li. 2025. Towards Efficient Privacy-Preserving Machine Learning: A Systematic Review from Protocol, Model, and System Perspectives. arXiv preprint arXiv:2507.14519 (2025)."},{"key":"e_1_3_2_1_54_1","doi-asserted-by":"publisher","DOI":"10.1109\/ICCV51070.2023.00472"},{"key":"e_1_3_2_1_55_1","doi-asserted-by":"publisher","DOI":"10.1145\/3696410.3714535"}],"event":{"name":"WWW '26: The ACM Web Conference 2026","location":"Dubai United Arab Emirates","sponsor":["SIGWEB ACM Special Interest Group on Hypertext, Hypermedia, and Web"]},"container-title":["Proceedings of the ACM Web Conference 2026"],"original-title":[],"link":[{"URL":"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3774904.3792211","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,27]],"date-time":"2026-04-27T13:48:18Z","timestamp":1777297698000},"score":1,"resource":{"primary":{"URL":"https:\/\/dl.acm.org\/doi\/10.1145\/3774904.3792211"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4,12]]},"references-count":55,"alternative-id":["10.1145\/3774904.3792211","10.1145\/3774904"],"URL":"https:\/\/doi.org\/10.1145\/3774904.3792211","relation":{},"subject":[],"published":{"date-parts":[[2026,4,12]]},"assertion":[{"value":"2026-04-12","order":3,"name":"published","label":"Published","group":{"name":"publication_history","label":"Publication History"}}]}}