{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,19]],"date-time":"2026-02-19T15:17:49Z","timestamp":1771514269022,"version":"3.50.1"},"publisher-location":"Cham","reference-count":71,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197864","type":"print"},{"value":"9783031197871","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19787-1_15","type":"book-chapter","created":{"date-parts":[[2022,10,20]],"date-time":"2022-10-20T22:16:11Z","timestamp":1666304171000},"page":"259-276","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":17,"title":["DeltaGAN: Towards Diverse Few-Shot Image Generation with\u00a0Sample-Specific Delta"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6401-0812","authenticated-orcid":false,"given":"Yan","family":"Hong","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1970-8634","authenticated-orcid":false,"given":"Li","family":"Niu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-2673-5860","authenticated-orcid":false,"given":"Jianfu","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7597-8503","authenticated-orcid":false,"given":"Liqing","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,21]]},"reference":[{"key":"15_CR1","unstructured":"Almahairi, A., Rajeswar, S., Sordoni, A., Bachman, P., Courville, A.C.: Augmented cycleGAN: learning many-to-many mappings from unpaired data. In: ICML (2018)"},{"key":"15_CR2","doi-asserted-by":"crossref","unstructured":"Antoniou, A., Storkey, A., Edwards, H.: Data augmentation generative adversarial networks. arXiv preprint arXiv:1711.04340 (2017)","DOI":"10.1007\/978-3-030-01424-7_58"},{"key":"15_CR3","doi-asserted-by":"crossref","unstructured":"Bao, J., Chen, D., Wen, F., Li, H., Hua, G.: CVAE-GAN: fine-grained image generation through asymmetric training. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.299"},{"key":"15_CR4","unstructured":"Bartunov, S., Vetrov, D.: Few-shot generative modelling with generative matching networks. In: AISTATS (2018)"},{"key":"15_CR5","unstructured":"Binkowski, M., Sutherland, D.J., Arbel, M., Gretton, A.: Demystifying mmd GANs. In: ICLR (2018)"},{"key":"15_CR6","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. In: ICLR (2018)"},{"key":"15_CR7","doi-asserted-by":"publisher","first-page":"1731","DOI":"10.1109\/TNNLS.2018.2872675","volume":"30","author":"B Cao","year":"2019","unstructured":"Cao, B., Wang, N., Li, J., Gao, X.: Data augmentation-based joint learning for heterogeneous face recognition. IEEE Trans. Neural Netw. Learn. Syst. (TNNLS) 30, 1731\u20131743 (2019)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst. (TNNLS)"},{"key":"15_CR8","doi-asserted-by":"crossref","unstructured":"Cao, Q., Shen, L., Xie, W., Parkhi, O.M., Zisserman, A.: VGGFace2: a dataset for recognising faces across pose and age. In: FG (2018)","DOI":"10.1109\/FG.2018.00020"},{"key":"15_CR9","doi-asserted-by":"crossref","unstructured":"Chen, M., et al.: Diversity transfer network for few-shot learning. In: AAAI (2020)","DOI":"10.1609\/aaai.v34i07.6628"},{"key":"15_CR10","doi-asserted-by":"crossref","unstructured":"Chen, T., Zhai, X., Ritter, M., Lucic, M., Houlsby, N.: Self-supervised GANs via auxiliary rotation loss. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01243"},{"key":"15_CR11","doi-asserted-by":"crossref","unstructured":"Chen, Z., Fu, Y., Wang, Y.X., Ma, L., Liu, W., Hebert, M.: Image deformation meta-networks for one-shot learning. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00888"},{"issue":"9","key":"15_CR12","first-page":"4594","volume":"28","author":"Z Chen","year":"2019","unstructured":"Chen, Z., Fu, Y., Zhang, Y., Jiang, Y.G., Xue, X., Sigal, L.: Multi-level semantic feature augmentation for one-shot learning. TIP 28(9), 4594\u20134605 (2019)","journal-title":"TIP"},{"key":"15_CR13","unstructured":"Clou\u00e2tre, L., Demers, M.: FIGR: few-shot image generation with reptile. arXiv preprint arXiv:1901.02199 (2019)"},{"key":"15_CR14","doi-asserted-by":"crossref","unstructured":"Cohen, G., Afshar, S., Tapson, J., van Schaik, A.: EMNIST: an extension of MNIST to handwritten letters. In: IJCNN (2017)","DOI":"10.1109\/IJCNN.2017.7966217"},{"key":"15_CR15","doi-asserted-by":"crossref","unstructured":"Cubuk, E.D., Zoph, B., Mane, D., Vasudevan, V.K., Le, Q.V.: Autoaugment: learning augmentation strategies from data. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00020"},{"key":"15_CR16","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"15_CR17","doi-asserted-by":"crossref","unstructured":"Dixit, M., Kwitt, R., Niethammer, M., Vasconcelos, N.: AGA: attribute-guided augmentation. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.355"},{"key":"15_CR18","unstructured":"Donahue, J., Simonyan, K.: Large scale adversarial representation learning. In: Advances in Neural Information Processing Systems (2019)"},{"key":"15_CR19","doi-asserted-by":"crossref","unstructured":"Feng, R., Gu, J., Qiao, Y., Dong, C.: Suppressing model overfitting for image super-resolution networks. In: CVPR (2019)","DOI":"10.1109\/CVPRW.2019.00248"},{"key":"15_CR20","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: ICML (2017)"},{"key":"15_CR21","unstructured":"Gao, H., Shou, Z., Zareian, A., Zhang, H., Chang, S.: Low-shot learning via covariance-preserving adversarial augmentation networks. In: NeurIPS (2018)"},{"key":"15_CR22","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: NeurIPS (2014)"},{"key":"15_CR23","doi-asserted-by":"crossref","unstructured":"Gu, Z., Li, W., Huo, J., Wang, L., Gao, Y.: LoFGAN: fusing local representations for few-shot image generation. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00835"},{"key":"15_CR24","doi-asserted-by":"crossref","unstructured":"Hariharan, B., Girshick, R.B.: Low-shot visual recognition by shrinking and hallucinating features. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.328"},{"key":"15_CR25","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"15_CR26","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local Nash equilibrium. In: NeurIPS (2017)"},{"key":"15_CR27","unstructured":"Ho, D., Liang, E., Chen, X., Stoica, I., Abbeel, P.: Population based augmentation: efficient learning of augmentation policy schedules. In: ICML (2019)"},{"key":"15_CR28","unstructured":"Hoffman, J., et al.: CyCADA: cycle-consistent adversarial domain adaptation. In: ICML (2018)"},{"key":"15_CR29","doi-asserted-by":"crossref","unstructured":"Hong, Y., Niu, L., Zhang, J., Zhang, L.: MatchingGAN: matching-based few-shot image generation. In: ICME (2020)","DOI":"10.1109\/ICME46284.2020.9102917"},{"key":"15_CR30","doi-asserted-by":"crossref","unstructured":"Hong, Y., Niu, L., Zhang, J., Zhao, W., Fu, C., Zhang, L.: F2GAN: fusing-and-filling GAN for few-shot image generation. In: ACM MM (2020)","DOI":"10.1145\/3394171.3413561"},{"key":"15_CR31","doi-asserted-by":"crossref","unstructured":"Jo, H.J., Min, C.H., Song, J.B.: Bin picking system using object recognition based on automated synthetic dataset generation. In: 2018 15th International Conference on Ubiquitous Robots (UR) (2018)","DOI":"10.1109\/URAI.2018.8441811"},{"key":"15_CR32","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"15_CR33","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of StyleGAN. arXiv preprint arXiv:1912.04958 (2019)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"15_CR34","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-319-16199-0_1","volume-title":"Computer Vision - ECCV 2014 Workshops","author":"Y Kawano","year":"2015","unstructured":"Kawano, Y., Yanai, K.: Automatic expansion of a food image dataset leveraging existing categories with domain adaptation. In: Agapito, L., Bronstein, M.M., Rother, C. (eds.) ECCV 2014. LNCS, vol. 8927, pp. 3\u201317. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-16199-0_1"},{"key":"15_CR35","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational Bayes. In: ICLR (2014)"},{"key":"15_CR36","unstructured":"Lake, B., Salakhutdinov, R., Gross, J., Tenenbaum, J.: One shot learning of simple visual concepts. In: CogSci (2011)"},{"key":"15_CR37","unstructured":"Li, Y., Zhang, R., Lu, J., Shechtman, E.: Few-shot image generation with elastic weight consolidation. In: NeurIPS (2020)"},{"key":"15_CR38","unstructured":"Liang, W., Liu, Z., Liu, C.: DAWSON: a domain adaptive few shot generation framework. arXiv preprint arXiv:2001.00576 (2020)"},{"key":"15_CR39","unstructured":"Lim, S., Kim, I., Kim, T., Kim, C., Kim, S.: Fast autoaugment. In: NeurIPS (2019)"},{"key":"15_CR40","doi-asserted-by":"crossref","unstructured":"Liu, J., Sun, Y., Han, C., Dou, Z., Li, W.: Deep representation learning on long-tailed data: a learnable embedding augmentation perspective. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00304"},{"key":"15_CR41","doi-asserted-by":"crossref","unstructured":"Liu, L., et al.: GenDet: meta learning to generate detectors from few shots. TNNLS (2021)","DOI":"10.1109\/TNNLS.2021.3053005"},{"key":"15_CR42","doi-asserted-by":"crossref","unstructured":"Liu, M., et al.: Few-shot unsupervised image-to-image translation. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.01065"},{"key":"15_CR43","unstructured":"Makhzani, A., Frey, B.J.: PixelGAN autoencoders. In: Advances in Neural Information Processing Systems (2017)"},{"key":"15_CR44","doi-asserted-by":"crossref","unstructured":"Mao, Q., Lee, H.Y., Tseng, H.Y., Ma, S., Yang, M.H.: Mode seeking generative adversarial networks for diverse image synthesis. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00152"},{"key":"15_CR45","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul Smolley, S.: Least squares generative adversarial networks. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"15_CR46","unstructured":"Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral normalization for generative adversarial networks. In: Proceedings of 6th International Conference on Learning Representations (ICLR) (2018)"},{"key":"15_CR47","unstructured":"Miyato, T., Koyama, M.: CGANs with projection discriminator. In: ICLR (2018)"},{"key":"15_CR48","unstructured":"Nichol, A., Achiam, J., Schulman, J.: On first-order meta-learning algorithms. arXiv preprint arXiv:1803.02999 (2018)"},{"key":"15_CR49","doi-asserted-by":"crossref","unstructured":"Nilsback, M.E., Zisserman, A.: Automated flower classification over a large number of classes. In: CVGIP (2008)","DOI":"10.1109\/ICVGIP.2008.47"},{"key":"15_CR50","doi-asserted-by":"crossref","unstructured":"Ojha, U., et al.: Few-shot image generation via cross-domain correspondence. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01060"},{"key":"15_CR51","unstructured":"Perez, L., Wang, J.: The effectiveness of data augmentation in image classification using deep learning. In: CVPR (2017)"},{"key":"15_CR52","unstructured":"Pu, Y., Zhe, G., Henao, R., Xin, Y., Carin, L.: Variational autoencoder for deep learning of images, labels and captions. In: NeurIPS (2016)"},{"key":"15_CR53","unstructured":"Ratner, A.J., Ehrenberg, H., Hussain, Z., Dunnmon, J., R\u00e9, C.: Learning to compose domain-specific transformations for data augmentation. In: NeurIPS (2017)"},{"key":"15_CR54","unstructured":"Reed, S., et al.: Few-shot autoregressive density estimation: towards learning to learn distributions. In: ICLR (2018)"},{"key":"15_CR55","unstructured":"Rezende, D.J., Mohamed, S., Danihelka, I., Gregor, K., Wierstra, D.: One-shot generalization in deep generative models. In: ICML (2016)"},{"key":"15_CR56","unstructured":"Robb, E., Chu, W.S., Kumar, A., Huang, J.B.: Few-shot adaptation of generative adversarial networks. arXiv preprint arXiv:2010.11943 (2020)"},{"key":"15_CR57","unstructured":"Schwartz, E., et al.: Delta-encoder: an effective sample synthesis method for few-shot object recognition. In: NeurIPS (2018)"},{"key":"15_CR58","doi-asserted-by":"crossref","unstructured":"Sun, Q., Liu, Y., Chua, T.S., Schiele, B.: Meta-transfer learning for few-shot learning. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00049"},{"key":"15_CR59","doi-asserted-by":"crossref","unstructured":"Sung, F., Yang, Y., Zhang, L., Xiang, T., Torr, P.H., Hospedales, T.M.: Learning to compare: relation network for few-shot learning. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00131"},{"key":"15_CR60","unstructured":"Tian, K., Lin, C., Sun, M., Zhou, L., Yan, J., Ouyang, W.: Improving auto-augment via augmentation-wise weight sharing. In: NeurIPS (2020)"},{"key":"15_CR61","unstructured":"Tseng, H., Lee, H., Huang, J., Yang, M.: Cross-domain few-shot classification via learned feature-wise transformation. In: Proceedings of 8th International Conference on Learning Representations (ICLR) (2020)"},{"key":"15_CR62","unstructured":"Tsutsui, S., Fu, Y., Crandall, D.: Meta-reinforced synthetic data for one-shot fine-grained visual recognition. In: NeurIPS (2019)"},{"key":"15_CR63","doi-asserted-by":"crossref","unstructured":"Van Horn, G., et al.: Building a bird recognition app and large scale dataset with citizen scientists: the fine print in fine-grained dataset collection. In: CVPR (2015)","DOI":"10.1109\/CVPR.2015.7298658"},{"key":"15_CR64","unstructured":"Vinyals, O., Blundell, C., Lillicrap, T., Wierstra, D., et al.: Matching networks for one shot learning. In: NeurIPS (2016)"},{"key":"15_CR65","doi-asserted-by":"crossref","unstructured":"Wang, Y., Gonzalez-Garcia, A., Berga, D., Herranz, L., Khan, F.S., van de Weijer, J.: MineGAN: effective knowledge transfer from GANs to target domains with few images. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00935"},{"key":"15_CR66","doi-asserted-by":"crossref","unstructured":"Wang, Y., Girshick, R.B., Hebert, M., Hariharan, B.: Low-shot learning from imaginary data. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00760"},{"key":"15_CR67","doi-asserted-by":"crossref","unstructured":"Yang, L., Li, L., Zhang, Z., Zhou, X., Zhou, E., Liu, Y.: DPGN: distribution propagation graph network for few-shot learning. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01340"},{"key":"15_CR68","doi-asserted-by":"crossref","unstructured":"Zhang, C., Cai, Y., Lin, G., Shen, C.: DeepEMD: few-shot image classification with differentiable earth mover\u2019s distance and structured classifiers. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01222"},{"key":"15_CR69","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"15_CR70","doi-asserted-by":"crossref","unstructured":"Zhang, X., Wang, Z., Liu, D., Ling, Q.: DADA: deep adversarial data augmentation for extremely low data regime classification. In: Proceedings of International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2019)","DOI":"10.1109\/ICASSP.2019.8683197"},{"key":"15_CR71","unstructured":"Zhao, M., Cong, Y., Carin, L.: On leveraging pretrained GANs for generation with limited data. In: ICML (2020)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19787-1_15","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T19:18:56Z","timestamp":1710357536000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19787-1_15"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197864","9783031197871"],"references-count":71,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19787-1_15","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"21 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}