{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T18:34:09Z","timestamp":1776882849908,"version":"3.51.2"},"publisher-location":"Cham","reference-count":49,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198205","type":"print"},{"value":"9783031198212","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19821-2_41","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T12:12:59Z","timestamp":1666440779000},"page":"717-733","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":134,"title":["ARF: Artistic Radiance Fields"],"prefix":"10.1007","author":[{"given":"Kai","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Nick","family":"Kolkin","sequence":"additional","affiliation":[]},{"given":"Sai","family":"Bi","sequence":"additional","affiliation":[]},{"given":"Fujun","family":"Luan","sequence":"additional","affiliation":[]},{"given":"Zexiang","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Eli","family":"Shechtman","sequence":"additional","affiliation":[]},{"given":"Noah","family":"Snavely","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"41_CR1","doi-asserted-by":"crossref","unstructured":"Yu, A., Fridovich-Keil, S., Tancik, M., Chen, Q., Recht, B., Kanazawa, A.: Plenoxels: radiance fields without neural networks (2021)","DOI":"10.1109\/CVPR52688.2022.00542"},{"key":"41_CR2","doi-asserted-by":"crossref","unstructured":"An, J., Huang, S., Song, Y., Dou, D., Liu, W., Luo, J.: ArtFlow: unbiased image style transfer via reversible neural flows. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 862\u2013871 (2021)","DOI":"10.1109\/CVPR46437.2021.00092"},{"key":"41_CR3","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Mip-NeRF 360: unbounded anti-aliased neural radiance fields. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"41_CR4","doi-asserted-by":"crossref","unstructured":"Chen, A., Xu, Z., Geiger, A., Yu, J., Su, H.: TensoRF: tensorial radiance fields. arXiv preprint arXiv:2203.09517 (2022)","DOI":"10.1007\/978-3-031-19824-3_20"},{"key":"41_CR5","doi-asserted-by":"crossref","unstructured":"Chen, A., et al.: MVSNeRF: fast generalizable radiance field reconstruction from multi-view stereo. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14124\u201314133 (2021)","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"41_CR6","doi-asserted-by":"crossref","unstructured":"Chen, D., Liao, J., Yuan, L., Yu, N., Hua, G.: Coherent online video style transfer. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1105\u20131114 (2017)","DOI":"10.1109\/ICCV.2017.126"},{"key":"41_CR7","unstructured":"Chen, T.Q., Schmidt, M.: Fast patch-based style transfer of arbitrary style. arXiv preprint arXiv:1612.04337 (2016)"},{"key":"41_CR8","doi-asserted-by":"crossref","unstructured":"Chiang, P.Z., Tsai, M.S., Tseng, H.Y., Lai, W.S., Chiu, W.C.: Stylizing 3D scene via implicit representation and hypernetwork. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1475\u20131484 (2022)","DOI":"10.1109\/WACV51458.2022.00029"},{"key":"41_CR9","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"169","DOI":"10.1007\/978-3-030-58529-7_11","volume-title":"Computer Vision \u2013 ECCV 2020","author":"T-Y Chiu","year":"2020","unstructured":"Chiu, T.-Y., Gurari, D.: Iterative feature transformation for fast and versatile universal style transfer. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 169\u2013184. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58529-7_11"},{"key":"41_CR10","doi-asserted-by":"crossref","unstructured":"Deng, Y., Tang, F., Dong, W., Huang, H., Ma, C., Xu, C.: Arbitrary video style transfer via multi-channel correlation. arXiv preprint arXiv:2009.08003 (2020)","DOI":"10.1145\/3394171.3414015"},{"key":"41_CR11","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2414\u20132423 (2016)","DOI":"10.1109\/CVPR.2016.265"},{"key":"41_CR12","doi-asserted-by":"crossref","unstructured":"Gu, S., Chen, C., Liao, J., Yuan, L.: Arbitrary style transfer with deep feature reshuffle. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8222\u20138231 (2018)","DOI":"10.1109\/CVPR.2018.00858"},{"key":"41_CR13","doi-asserted-by":"crossref","unstructured":"Heitz, E., Vanhoey, K., Chambon, T., Belcour, L.: A sliced wasserstein loss for neural texture synthesis. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9412\u20139420 (2021)","DOI":"10.1109\/CVPR46437.2021.00929"},{"key":"41_CR14","doi-asserted-by":"crossref","unstructured":"H\u00f6llein, L., Johnson, J., Niessner, M.: StyleMesh: style transfer for indoor 3D scene reconstructions. arXiv preprint arXiv:2112.01530 (2021)","DOI":"10.1109\/CVPR52688.2022.00610"},{"key":"41_CR15","doi-asserted-by":"crossref","unstructured":"Huang, H., et al.: Real-time neural style transfer for videos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 783\u2013791 (2017)","DOI":"10.1109\/CVPR.2017.745"},{"key":"41_CR16","doi-asserted-by":"crossref","unstructured":"Huang, H.P., Tseng, H.Y., Saini, S., Singh, M., Yang, M.H.: Learning to stylize novel views. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13869\u201313878 (2021)","DOI":"10.1109\/ICCV48922.2021.01361"},{"key":"41_CR17","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"issue":"11","key":"41_CR18","doi-asserted-by":"publisher","first-page":"3365","DOI":"10.1109\/TVCG.2019.2921336","volume":"26","author":"Y Jing","year":"2019","unstructured":"Jing, Y., Yang, Y., Feng, Z., Ye, J., Yu, Y., Song, M.: Neural style transfer: a review. IEEE Trans. Vis. Comput. Graph. 26(11), 3365\u20133385 (2019)","journal-title":"IEEE Trans. Vis. Comput. Graph."},{"issue":"4","key":"41_CR19","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073599","volume":"36","author":"A Knapitsch","year":"2017","unstructured":"Knapitsch, A., Park, J., Zhou, Q.Y., Koltun, V.: Tanks and temples: benchmarking large-scale scene reconstruction. ACM Trans. Graph. 36(4), 1\u201313 (2017)","journal-title":"ACM Trans. Graph."},{"key":"41_CR20","unstructured":"Kolkin, N., Kucera, M., Paris, S., Sykora, D., Shechtman, E., Shakhnarovich, G.: Neural neighbor style transfer. arXiv e-prints pp. arXiv-2203 (2022)"},{"key":"41_CR21","doi-asserted-by":"crossref","unstructured":"Kolkin, N., Salavon, J., Shakhnarovich, G.: Style transfer by relaxed optimal transport and self-similarity. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10051\u201310060 (2019)","DOI":"10.1109\/CVPR.2019.01029"},{"key":"41_CR22","doi-asserted-by":"crossref","unstructured":"Li, C., Wand, M.: Combining Markov random fields and convolutional neural networks for image synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2479\u20132486 (2016)","DOI":"10.1109\/CVPR.2016.272"},{"key":"41_CR23","doi-asserted-by":"crossref","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Diversified texture synthesis with feed-forward networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3920\u20133928 (2017)","DOI":"10.1109\/CVPR.2017.36"},{"key":"41_CR24","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Universal style transfer via feature transforms. In: Advances in Neural Information Processing Systems 30 (2017)"},{"issue":"4","key":"41_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3072959.3073683","volume":"36","author":"J Liao","year":"2017","unstructured":"Liao, J., Yao, Y., Yuan, L., Hua, G., Kang, S.B.: Visual attribute transfer through deep image analogy. ACM Trans. Graph 36(4), 1\u201315 (2017)","journal-title":"ACM Trans. Graph"},{"key":"41_CR26","unstructured":"Liu, L., Gu, J., Zaw Lin, K., Chua, T.S., Theobalt, C.: Neural sparse voxel fields. In: Advances in Neural Information Processing Systems, vol. 33, pp. 15651\u201315663 (2020)"},{"key":"41_CR27","doi-asserted-by":"crossref","unstructured":"Liu, S., et al.: AdaAttN: revisit attention mechanism in arbitrary neural style transfer. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6649\u20136658 (2021)","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"41_CR28","doi-asserted-by":"crossref","unstructured":"Luan, F., Paris, S., Shechtman, E., Bala, K.: Deep photo style transfer. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4990\u20134998 (2017)","DOI":"10.1109\/CVPR.2017.740"},{"issue":"4","key":"41_CR29","doi-asserted-by":"publisher","first-page":"95","DOI":"10.1111\/cgf.13478","volume":"37","author":"F Luan","year":"2018","unstructured":"Luan, F., Paris, S., Shechtman, E., Bala, K.: Deep painterly harmonization. Comput. Graph. Forum 37(4), 95\u2013106 (2018)","journal-title":"Comput. Graph. Forum"},{"key":"41_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"800","DOI":"10.1007\/978-3-030-01264-9_47","volume-title":"Computer Vision \u2013 ECCV 2018","author":"R Mechrez","year":"2018","unstructured":"Mechrez, R., Talmi, I., Zelnik-Manor, L.: The contextual loss for image transformation with non-aligned data. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. LNCS, vol. 11218, pp. 800\u2013815. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01264-9_47"},{"key":"41_CR31","doi-asserted-by":"crossref","unstructured":"Michel, O., Bar-On, R., Liu, R., Benaim, S., Hanocka, R.: Text2Mesh: text-driven neural stylization for meshes. arXiv preprint arXiv:2112.03221 (2021)","DOI":"10.1109\/CVPR52688.2022.01313"},{"key":"41_CR32","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"405","DOI":"10.1007\/978-3-030-58452-8_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"B Mildenhall","year":"2020","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 405\u2013421. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_24"},{"key":"41_CR33","doi-asserted-by":"crossref","unstructured":"Mu, F., Wang, J., Wu, Y., Li, Y.: 3D photo stylization: learning to generate stylized novel views from a single image. arXiv preprint arXiv:2112.00169 (2021)","DOI":"10.1109\/CVPR52688.2022.01579"},{"key":"41_CR34","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Barron, J.T., Mildenhall, B., Sajjadi, M.S.M., Geiger, A., Radwan, N.: RegNeRF: regularizing neural radiance fields for view synthesis from sparse inputs. In: Proceedings IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2022)","DOI":"10.1109\/CVPR52688.2022.00540"},{"key":"41_CR35","doi-asserted-by":"crossref","unstructured":"Park, D.Y., Lee, K.H.: Arbitrary style transfer with style-attentional networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5880\u20135888 (2019)","DOI":"10.1109\/CVPR.2019.00603"},{"key":"41_CR36","unstructured":"Risser, E., Wilmot, P., Barnes, C.: Stable and controllable neural texture synthesis and style transfer using histogram losses. arXiv preprint arXiv:1701.08893 (2017)"},{"issue":"11","key":"41_CR37","doi-asserted-by":"publisher","first-page":"1199","DOI":"10.1007\/s11263-018-1089-z","volume":"126","author":"M Ruder","year":"2018","unstructured":"Ruder, M., Dosovitskiy, A., Brox, T.: Artistic style transfer for videos and spherical images. Int. J. Comput. Vis. 126(11), 1199\u20131219 (2018)","journal-title":"Int. J. Comput. Vis."},{"key":"41_CR38","doi-asserted-by":"crossref","unstructured":"Sheng, L., Lin, Z., Shao, J., Wang, X.: Avatar-Net: multi-scale zero-shot style transfer by feature decoration. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8242\u20138250 (2018)","DOI":"10.1109\/CVPR.2018.00860"},{"key":"41_CR39","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"41_CR40","doi-asserted-by":"crossref","unstructured":"Tsai, Y.H., Shen, X., Lin, Z., Sunkavalli, K., Lu, X., Yang, M.H.: Deep image harmonization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3789\u20133797 (2017)","DOI":"10.1109\/CVPR.2017.299"},{"key":"41_CR41","doi-asserted-by":"crossref","unstructured":"Wang, W., Xu, J., Zhang, L., Wang, Y., Liu, J.: Consistent video style transfer via compound regularization. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 34, pp. 12233\u201312240 (2020)","DOI":"10.1609\/aaai.v34i07.6905"},{"key":"41_CR42","doi-asserted-by":"crossref","unstructured":"Xia, X., et al.: Real-time localized photorealistic video style transfer. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1089\u20131098 (2021)","DOI":"10.1109\/WACV48630.2021.00113"},{"key":"41_CR43","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"327","DOI":"10.1007\/978-3-030-58598-3_20","volume-title":"Computer Vision \u2013 ECCV 2020","author":"X Xia","year":"2020","unstructured":"Xia, X., et al.: Joint bilateral learning for real-time universal photorealistic style transfer. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12353, pp. 327\u2013342. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58598-3_20"},{"key":"41_CR44","doi-asserted-by":"crossref","unstructured":"Yao, Y., Ren, J., Xie, X., Liu, W., Liu, Y.J., Wang, J.: Attention-aware multi-stroke style transfer. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1467\u20131475 (2019)","DOI":"10.1109\/CVPR.2019.00156"},{"key":"41_CR45","doi-asserted-by":"crossref","unstructured":"Yin, K., Gao, J., Shugrina, M., Khamis, S., Fidler, S.: 3DStyleNet: creating 3D shapes with geometric and texture style variations. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12456\u201312465 (2021)","DOI":"10.1109\/ICCV48922.2021.01223"},{"key":"41_CR46","doi-asserted-by":"crossref","unstructured":"Yu, A., Li, R., Tancik, M., Li, H., Ng, R., Kanazawa, A.: PlenOctrees for real-time rendering of neural radiance fields. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5752\u20135761 (2021)","DOI":"10.1109\/ICCV48922.2021.00570"},{"key":"41_CR47","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelNeRF: neural radiance fields from one or few images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4578\u20134587 (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"41_CR48","unstructured":"Zhang, K., Riegler, G., Snavely, N., Koltun, V.: Nerf++: analyzing and improving neural radiance fields. arXiv preprint arXiv:2010.07492 (2020)"},{"key":"41_CR49","doi-asserted-by":"crossref","unstructured":"Zhang, L., Wen, T., Shi, J.: Deep image blending. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 231\u2013240 (2020)","DOI":"10.1109\/WACV45572.2020.9093632"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19821-2_41","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,9]],"date-time":"2023-03-09T15:41:38Z","timestamp":1678376498000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19821-2_41"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198205","9783031198212"],"references-count":49,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19821-2_41","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}