{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,10]],"date-time":"2026-04-10T16:04:52Z","timestamp":1775837092626,"version":"3.50.1"},"publisher-location":"Singapore","reference-count":59,"publisher":"Springer Singapore","isbn-type":[{"value":"9789811591280","type":"print"},{"value":"9789811591297","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-981-15-9129-7_31","type":"book-chapter","created":{"date-parts":[[2020,10,21]],"date-time":"2020-10-21T23:07:29Z","timestamp":1603321649000},"page":"443-468","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["Adversarial Examples Attack and Countermeasure for Speech Recognition System: A Survey"],"prefix":"10.1007","author":[{"given":"Donghua","family":"Wang","sequence":"first","affiliation":[]},{"given":"Rangding","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Li","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Diqun","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Xueyuan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yongkang","family":"Gong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,10,22]]},"reference":[{"key":"31_CR1","unstructured":"Vaidya, T., Zhang, Y., Sherr, M., Shields, C.: Cocaine noodles: exploiting the gap between human and machine speech recognition. In: 9th $$\\{$$USENIX$$\\}$$ Workshop on Offensive Technologies ($$\\{$$WOOT$$\\}$$ 2015) (2015)"},{"key":"31_CR2","unstructured":"Carlini, N., et al.: Hidden voice commands. In: 25th $$\\{$$USENIX$$\\}$$ Security Symposium ($$\\{$$USENIX$$\\}$$ Security 2016), pp. 513\u2013530 (2016)"},{"issue":"10","key":"31_CR3","doi-asserted-by":"publisher","first-page":"120","DOI":"10.1109\/MCOM.2019.1900006","volume":"57","author":"S Hu","year":"2019","unstructured":"Hu, S., Shang, X., Qin, Z., Li, M., Wang, Q., Wang, C.: Adversarial examples for automatic speech recognition: attacks and countermeasures. IEEE Commun. Mag. 57(10), 120\u2013126 (2019)","journal-title":"IEEE Commun. Mag."},{"key":"31_CR4","doi-asserted-by":"crossref","unstructured":"Audhkhasi, K., Ramabhadran, B., Saon, G., Picheny, M., Nahamoo, D.: Direct acoustics-to-word models for English conversational speech recognition, arXiv preprint arXiv:1703.07754 (2017)","DOI":"10.21437\/Interspeech.2017-546"},{"key":"31_CR5","doi-asserted-by":"crossref","unstructured":"Xiong, W., Wu, L., Alleva, F., Droppo, J., Huang, X., Stolcke, A.: The Microsoft 2017 conversational speech recognition system. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5934\u20135938. IEEE (2018)","DOI":"10.1109\/ICASSP.2018.8461870"},{"key":"31_CR6","unstructured":"Povey, D., et al.: The kaldispeech recognition toolkit. In: IEEE 2011 Workshop on Automatic Speech Recognition and Understanding, no. CONF. IEEE Signal Processing Society (2011)"},{"key":"31_CR7","unstructured":"Kaldi. https:\/\/github.com\/kaldi-asr\/kaldi"},{"key":"31_CR8","doi-asserted-by":"crossref","unstructured":"Sch\u00f6nherr, L., Kohls, K., Zeiler, S., Holz, T., Kolossa, D.: Adversarial attacks against automatic speech recognition systems via psychoacoustic hiding, arXiv preprint arXiv:1808.05665 (2018)","DOI":"10.14722\/ndss.2019.23288"},{"key":"31_CR9","unstructured":"Hannun, A., et al.: Deep speech: scaling up end-to-end speech recognition, arXiv preprint arXiv:1412.5567 (2014)"},{"key":"31_CR10","unstructured":"DeepSpeech. https:\/\/github.com\/mozilla\/DeepSpeech"},{"key":"31_CR11","doi-asserted-by":"crossref","unstructured":"Du, T., Ji, S., Li, J., Gu, Q., Wang, T., Beyah, R.: Sirenattack: generating adversarial audio for end-to-end acoustic systems, arXiv preprint arXiv:1901.07846 (2019)","DOI":"10.1145\/3320269.3384733"},{"key":"31_CR12","doi-asserted-by":"crossref","unstructured":"Graves, A., Fern\u00e1ndez, S., Gomez, F., Schmidhuber, J.: Connectionist temporal classification: labelling unsegmented sequence data with recurrent neural networks. In: Proceedings of the 23rd International Conference on Machine Learning, pp. 369\u2013376 (2006)","DOI":"10.1145\/1143844.1143891"},{"key":"31_CR13","unstructured":"Szegedy, C., et al.: Intriguing properties of neural networks, arXiv preprint arXiv:1312.6199 (2013)"},{"key":"31_CR14","doi-asserted-by":"crossref","unstructured":"Papernot, N., McDaniel, P., Goodfellow, I., Jha, S., Celik, Z.B., Swami, A.: Practical black-box attacks against machine learning. In: Proceedings of the 2017 ACM on Asia Conference on Computer and Communications Security, pp. 506\u2013519 (2017)","DOI":"10.1145\/3052973.3053009"},{"key":"31_CR15","doi-asserted-by":"crossref","unstructured":"Moosavi-Dezfooli, S.M., Fawzi, A., Fawzi, O., Frossard, P.: Universal adversarial perturbations. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1765\u20131773 (2017)","DOI":"10.1109\/CVPR.2017.17"},{"issue":"5","key":"31_CR16","doi-asserted-by":"publisher","first-page":"828","DOI":"10.1109\/TEVC.2019.2890858","volume":"23","author":"J Su","year":"2019","unstructured":"Su, J., Vargas, D.V., Sakurai, K.: One pixel attack for fooling deep neural networks. IEEE Trans. Evol. Comput. 23(5), 828\u2013841 (2019)","journal-title":"IEEE Trans. Evol. Comput."},{"key":"31_CR17","doi-asserted-by":"crossref","unstructured":"Dong, Y., et al.: Efficient decision-based black-box adversarial attacks on face recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7714\u20137722 (2019)","DOI":"10.1109\/CVPR.2019.00790"},{"key":"31_CR18","unstructured":"Mozilla common voice (2017). https:\/\/voice.mozilla.org\/en"},{"key":"31_CR19","unstructured":"Warden, P.: Speech commands: a public dataset for single-word speech recognition, vol. 1 (2017). Dataset. http:\/\/download.tensorflow.org\/data\/speech_commands_v0.01.tar.gz"},{"key":"31_CR20","doi-asserted-by":"crossref","unstructured":"Panayotov, V., Chen, G., Povey, D., Khudanpur, S.: Librispeech: an ASR corpus based on public domain audio books. In: 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5206\u20135210. IEEE (2015)","DOI":"10.1109\/ICASSP.2015.7178964"},{"issue":"4","key":"31_CR21","doi-asserted-by":"publisher","first-page":"335","DOI":"10.1007\/s10579-008-9076-6","volume":"42","author":"C Busso","year":"2008","unstructured":"Busso, C., et al.: Iemocap: interactive emotional dyadic motion capture database. Lang. Resour. Eval. 42(4), 335 (2008)","journal-title":"Lang. Resour. Eval."},{"key":"31_CR22","doi-asserted-by":"crossref","unstructured":"Zhang, G., Yan, C., Ji, X., Zhang, T., Zhang, T., Xu, W.: Dolphinattack: inaudible voice commands. In: Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, pp. 103\u2013117 (2017)","DOI":"10.1145\/3133956.3134052"},{"key":"31_CR23","unstructured":"Yuan, X., et al.: Commandersong: a systematic approach for practical adversarial voice recognition. In: 27th $$\\{$$USENIX$$\\}$$ Security Symposium ($$\\{$$USENIX$$\\}$$ Security 2018), pp. 49\u201364 (2018)"},{"key":"31_CR24","unstructured":"Cisse, M., Adi, Y., Neverova, N., Keshet, J.: Houdini: fooling deep structured prediction models, arXiv preprint arXiv:1707.05373 (2017)"},{"key":"31_CR25","unstructured":"Amodei, D., et al.: Deep speech 2: end-to-end speech recognition in English and mandarin. In: International Conference on Machine Learning, pp. 173\u2013182 (2016)"},{"key":"31_CR26","unstructured":"Iter, D., Huang, J., Jermann, M.: Generating adversarial examples for speech recognition. Stanford Technical Report (2017)"},{"key":"31_CR27","doi-asserted-by":"crossref","unstructured":"Abdullah, H., Garcia, W., Peeters, C., Traynor, P., Butler, K.R., Wilson, J.: Practical hidden voice attacks against speech and speaker recognition systems, arXiv preprint arXiv:1904.05734 (2019)","DOI":"10.14722\/ndss.2019.23362"},{"key":"31_CR28","doi-asserted-by":"crossref","unstructured":"Carlini, N., Wagner, D.: Audio adversarial examples: targeted attacks on speech-to-text. In: 2018 IEEE Security and Privacy Workshops (SPW), pp. 1\u20137. IEEE (2018)","DOI":"10.1109\/SPW.2018.00009"},{"key":"31_CR29","doi-asserted-by":"crossref","unstructured":"Yakura, H., Sakuma, J.: Robust audio adversarial example for a physical attack, arXiv preprint arXiv:1810.11793 (2018)","DOI":"10.24963\/ijcai.2019\/741"},{"key":"31_CR30","unstructured":"Qin, Y., Carlini, N., Cottrell, G., Goodfellow, I., Raffel, C.: Imperceptible, robust, and targeted adversarial examples for automatic speech recognition, arXiv preprint arXiv:1903.10346 (2019)"},{"key":"31_CR31","unstructured":"Shen, J., et al.: Lingvo: a modular and scalable framework for sequence-to-sequence modeling, arXiv preprint arXiv:1902.08295 (2019)"},{"key":"31_CR32","doi-asserted-by":"crossref","unstructured":"Sch\u00f6nherr, L., Zeiler, S., Holz, T., Kolossa, D.: Robust over-the-air adversarial examples against automatic speech recognition systems, arXiv preprint arXiv:1908.01551 (2019)","DOI":"10.1145\/3427228.3427276"},{"key":"31_CR33","unstructured":"Szurley, J., Kolter, J.Z.: Perceptual based adversarial audio attacks, arXiv preprint arXiv:1906.06355 (2019)"},{"key":"31_CR34","doi-asserted-by":"crossref","unstructured":"Liu, X., Zhang, X., Wan, K., Zhu, Q., Ding, Y.: Towards weighted-sampling audio adversarial example attack. arXiv, Audio and Speech Processing (2019)","DOI":"10.1609\/aaai.v34i04.5928"},{"key":"31_CR35","doi-asserted-by":"publisher","first-page":"526","DOI":"10.1109\/TIFS.2019.2925452","volume":"15","author":"HW Kwon","year":"2020","unstructured":"Kwon, H.W., Kwon, H., Yoon, H., Choi, D.: Selective audio adversarial example in evasion attack on speech recognition system. IEEE Trans. Inf. Forensics Secur. 15, 526\u2013538 (2020)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"31_CR36","unstructured":"Abdoli, S., Hafemann, L.G., Rony, J., Ayed, I.B., Cardinal, P., Koerich, A.L.: Universal adversarial audio perturbations, arXiv, vol. abs\/1908.03173 (2019)"},{"key":"31_CR37","unstructured":"Vadillo, J., Santana, R.: Universal adversarial examples in speech command classification, arXiv, vol. abs\/1911.10182 (2019)"},{"key":"31_CR38","doi-asserted-by":"crossref","unstructured":"Moosavi-Dezfooli, S.M., Fawzi, A., Frossard, P.: Deepfool: a simple and accurate method to fool deep neural networks. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2574\u20132582 (2015)","DOI":"10.1109\/CVPR.2016.282"},{"key":"31_CR39","doi-asserted-by":"crossref","unstructured":"Neekhara, P., Hussain, S., Pandey, P., Dubnov, S., McAuley, J., Koushanfar, F.: Universal adversarial perturbations for speech recognition systems, arXiv, vol. abs\/1905.03828 (2019)","DOI":"10.21437\/Interspeech.2019-1353"},{"key":"31_CR40","unstructured":"Gong, Y., Poellabauer, C.: Crafting adversarial examples for speech paralinguistics applications, arXiv, vol. abs\/1711.03280 (2017)"},{"key":"31_CR41","doi-asserted-by":"crossref","unstructured":"Kreuk, F., Adi, Y., Ciss\u00e9, M., Keshet, J.: Fooling end-to-end speaker verification with adversarial examples. In: 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 1962\u20131966 (2018)","DOI":"10.1109\/ICASSP.2018.8462693"},{"key":"31_CR42","unstructured":"Alzantot, M., Balaji, B., Srivastava, M.: Did you hear that? Adversarial examples against automatic speech recognition, arXiv, vol. abs\/1801.00554 (2018)"},{"key":"31_CR43","doi-asserted-by":"crossref","unstructured":"Taori, R., Kamsetty, A., Chu, B., Vemuri, N.: Targeted adversarial examples for black box audio systems. In: 2019 IEEE Security and Privacy Workshops (SPW), pp. 15\u201320 (2018)","DOI":"10.1109\/SPW.2019.00016"},{"key":"31_CR44","doi-asserted-by":"crossref","unstructured":"Khare, S., Aralikatte, R., Mani, S.: Adversarial black-box attacks on automatic speech recognition systems using multi-objective evolutionary optimization, arXiv preprint arXiv:1811.01312 (2018)","DOI":"10.21437\/Interspeech.2019-2420"},{"key":"31_CR45","doi-asserted-by":"crossref","unstructured":"Zagoruyko, S., Komodakis, N.: Wide residual networks, arXiv preprint arXiv:1605.07146 (2016)","DOI":"10.5244\/C.30.87"},{"key":"31_CR46","doi-asserted-by":"crossref","unstructured":"Sun, S., Yeh, C.-F., Ostendorf, M., Hwang, M.-Y., Xie, L.: Training augmentation with adversarial examples for robust speech recognition, arXiv preprint arXiv:1806.02782 (2018)","DOI":"10.21437\/Interspeech.2018-1247"},{"key":"31_CR47","doi-asserted-by":"crossref","unstructured":"Zeng, Q., et al.: A multiversion programming inspired approach to detecting audio adversarial examples. In: 2019 49th Annual IEEE\/IFIP International Conference on Dependable Systems and Networks (DSN), pp. 39\u201351. IEEE (2019)","DOI":"10.1109\/DSN.2019.00019"},{"key":"31_CR48","unstructured":"Latif, S., Rana, R., Qadir, J.: Adversarial machine learning and speech emotion recognition: utilizing generative adversarial networks for robustness, arXiv preprint arXiv:1811.11402 (2018)"},{"key":"31_CR49","doi-asserted-by":"crossref","unstructured":"Rajaratnam, K., Shah, K., Kalita, J.: Isolated and ensemble audio preprocessing methods for detecting adversarial examples against automatic speech recognition, arXiv preprint arXiv:1809.04397 (2018)","DOI":"10.1109\/ISSPIT.2018.8642623"},{"key":"31_CR50","doi-asserted-by":"crossref","unstructured":"Rajaratnam, K., Kalita, J.: Noise flooding for detecting audio adversarial examples against automatic speech recognition. In: 2018 IEEE International Symposium on Signal Processing and Information Technology (ISSPIT), pp. 197\u2013201. IEEE (2018)","DOI":"10.1109\/ISSPIT.2018.8642623"},{"key":"31_CR51","doi-asserted-by":"crossref","unstructured":"Samizade, S., Tan, Z.-H., Shen, C., Guan, X.: Adversarial example detection by classification for deep speech recognition, arXiv preprint arXiv:1910.10013 (2019)","DOI":"10.1109\/ICASSP40776.2020.9054750"},{"key":"31_CR52","unstructured":"Yang, Z., Li, B., Chen, P.-Y., Song, D.: Characterizing audio adversarial examples using temporal dependency, arXiv preprint arXiv:1809.10875 (2018)"},{"key":"31_CR53","doi-asserted-by":"crossref","unstructured":"Kwon, H., Yoon, H., Park, K.-W.: Poster: detecting audio adversarial example through audio modification. In: Proceedings of the 2019 ACM SIGSAC Conference on Computer and Communications Security, pp. 2521\u20132523 (2019)","DOI":"10.1145\/3319535.3363246"},{"key":"31_CR54","unstructured":"Ma, P., Petridis, S., Pantic, M.: Detecting adversarial attacks on audio-visual speech recognition, arXiv preprint arXiv:1912.08639 (2019)"},{"key":"31_CR55","doi-asserted-by":"publisher","first-page":"2147","DOI":"10.1109\/TIFS.2019.2956591","volume":"15","author":"M Esmaeilpour","year":"2019","unstructured":"Esmaeilpour, M., Cardinal, P., Koerich, A.L.: A robust approach for securing audio classification against adversarial attacks. IEEE Trans. Inf. Forensics Secur. 15, 2147\u20132159 (2019)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"31_CR56","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, pp. 2672\u20132680 (2014)"},{"key":"31_CR57","doi-asserted-by":"crossref","unstructured":"Roy, N., Hassanieh, H., Roy Choudhury, R.: Backdoor: making microphones hear inaudible sounds. In: Proceedings of the 15th Annual International Conference on Mobile Systems, Applications, and Services, pp. 2\u201314 (2017)","DOI":"10.1145\/3081333.3081366"},{"key":"31_CR58","doi-asserted-by":"crossref","unstructured":"Tamura, K., Omagari, A., Hashida, S.: Novel defense method against audio adversarial example for speech-to-text transcription neural networks. In: 2019 IEEE 11th International Workshop on Computational Intelligence and Applications (IWCIA), pp. 115\u2013120. IEEE (2019)","DOI":"10.1109\/IWCIA47330.2019.8955062"},{"key":"31_CR59","doi-asserted-by":"crossref","unstructured":"Yang, C.-H., Qi, J., Chen, P.-Y., Ma, X., Lee, C.-H.: Characterizing speech adversarial examples using self-attention u-net enhancement, arXiv preprint arXiv:2003.13917 (2020)","DOI":"10.1109\/ICASSP40776.2020.9053288"}],"container-title":["Communications in Computer and Information Science","Security and Privacy in Digital Economy"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-15-9129-7_31","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,4,25]],"date-time":"2021-04-25T02:33:58Z","timestamp":1619318038000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-15-9129-7_31"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9789811591280","9789811591297"],"references-count":59,"URL":"https:\/\/doi.org\/10.1007\/978-981-15-9129-7_31","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"22 October 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"SPDE","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Security and Privacy in Digital Economy","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Quzhou","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 October 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 November 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"spde2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/spde2020.csp.escience.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"132","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"48","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"36% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}