{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,2]],"date-time":"2025-11-02T08:13:34Z","timestamp":1762071214013,"version":"build-2065373602"},"publisher-location":"Cham","reference-count":74,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031197680"},{"type":"electronic","value":"9783031197697"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19769-7_29","type":"book-chapter","created":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T11:40:06Z","timestamp":1666438806000},"page":"497-514","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["2D GANs Meet Unsupervised Single-View 3D Reconstruction"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2103-4659","authenticated-orcid":false,"given":"Feng","family":"Liu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3215-8753","authenticated-orcid":false,"given":"Xiaoming","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,23]]},"reference":[{"key":"29_CR1","doi-asserted-by":"crossref","unstructured":"Abdal, R., Zhu, P., Mitra, N.J., Wonka, P.: StyleFlow: attribute-conditioned exploration of styleGAN-generated images using conditional continuous normalizing flows. TOG (2021)","DOI":"10.1145\/3447648"},{"key":"29_CR2","unstructured":"Asnani, V., Yin, X., Hassner, T., Liu, X.: Reverse engineering of generative models: Inferring model hyperparameters from generated images. arXiv:2106.07873 (2021)"},{"key":"29_CR3","doi-asserted-by":"crossref","unstructured":"Bai, Z., Cui, Z., Rahim, J.A., Liu, X., Tan, P.: Deep facial non-rigid multi-view stereo. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00589"},{"key":"29_CR4","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale GAN training for high fidelity natural image synthesis. In: ICLR (2019)"},{"key":"29_CR5","doi-asserted-by":"crossref","unstructured":"Chan, E.R., Monteiro, M., Kellnhofer, P., Wu, J., Wetzstein, G.: pi-GAN: periodic implicit generative adversarial networks for 3D-aware image synthesis. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00574"},{"key":"29_CR6","doi-asserted-by":"crossref","unstructured":"Dang, H., Liu, F., Stehouwer, J., Liu, X., Jain, A.K.: On the detection of digital face manipulation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00582"},{"key":"29_CR7","unstructured":"Dolhansky, B., et al.: The deepfake detection challenge (DFDC) dataset. arXiv preprint arXiv:2006.07397 (2020)"},{"key":"29_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"88","DOI":"10.1007\/978-3-030-58555-6_6","volume-title":"Computer Vision \u2013 ECCV 2020","author":"S Goel","year":"2020","unstructured":"Goel, S., Kanazawa, A., Malik, J.: Shape and viewpoint without keypoints. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12360, pp. 88\u2013104. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58555-6_6"},{"key":"29_CR9","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: NeurIPS (2014)"},{"key":"29_CR10","unstructured":"Gropp, A., Yariv, L., Haim, N., Atzmon, M., Lipman, Y.: Implicit geometric regularization for learning shapes. arXiv preprint arXiv:2002.10099 (2020)"},{"key":"29_CR11","unstructured":"Gu, J., Liu, L., Wang, P., Theobalt, C.: StyleNeRF: a style-based 3D-aware generator for high-resolution image synthesis. arXiv preprint arXiv:2110.08985 (2021)"},{"key":"29_CR12","doi-asserted-by":"crossref","unstructured":"Gu, J., Shen, Y., Zhou, B.: Image processing using multi-code GAN prior. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00308"},{"key":"29_CR13","unstructured":"Ha, D., Dai, A., Le, Q.V.: Hypernetworks. arXiv preprint arXiv:1609.09106 (2016)"},{"key":"29_CR14","unstructured":"H\u00e4rk\u00f6nen, E., Hertzmann, A., Lehtinen, J., Paris, S.: GANSpace: discovering interpretable GAN controls. In: NeurIPS (2020)"},{"key":"29_CR15","doi-asserted-by":"crossref","unstructured":"Hart, J.C.: Sphere tracing: a geometric method for the antialiased ray tracing of implicit surfaces. Vis. Comput. 12, 527\u2013545 (1996)","DOI":"10.1007\/s003710050084"},{"key":"29_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"29_CR17","doi-asserted-by":"crossref","unstructured":"Jiang, Y., Ji, D., Han, Z., Zwicker, M.: SDFDiff: differentiable rendering of signed distance fields for 3D shape optimization. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00133"},{"key":"29_CR18","doi-asserted-by":"crossref","unstructured":"Kanazawa, A., Tulsiani, S., Efros, A.A., Malik, J.: Learning category-specific mesh reconstruction from image collections. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01267-0_23"},{"key":"29_CR19","unstructured":"Kar, A., H\u00e4ne, C., Malik, J.: Learning a multi-view stereo machine. In: NeurIPS (2017)"},{"key":"29_CR20","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of GANs for improved quality, stability, and variation. In: ICLR (2018)"},{"key":"29_CR21","unstructured":"Karras, T., Aittala, M., Hellsten, J., Laine, S., Lehtinen, J., Aila, T.: Training generative adversarial networks with limited data. In: NeurIPS (2020)"},{"key":"29_CR22","unstructured":"Karras, T., et al.: Alias-free generative adversarial networks. In: NeurIPS (2021)"},{"key":"29_CR23","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"29_CR24","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of StyleGan. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"29_CR25","unstructured":"Kendall, A., Gal, Y.: What uncertainties do we need in Bayesian deep learning for computer vision? In: NeurIPS (2017)"},{"key":"29_CR26","doi-asserted-by":"crossref","unstructured":"Kulkarni, N., Gupta, A., Fouhey, D.F., Tulsiani, S.: Articulation-aware canonical surface mapping. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00053"},{"key":"29_CR27","doi-asserted-by":"crossref","unstructured":"Kulkarni, N., Gupta, A., Tulsiani, S.: Canonical surface mapping via geometric cycle consistency. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00229"},{"key":"29_CR28","doi-asserted-by":"publisher","unstructured":"Li, X., et al.: Self-supervised Single-View 3D Reconstruction via Semantic Consistency. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12359, pp. 677\u2013693. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58568-6_40","DOI":"10.1007\/978-3-030-58568-6_40"},{"key":"29_CR29","unstructured":"Lin, C.H., Wang, C., Lucey, S.: SDF-SRN: learning signed distance 3D object reconstruction from static images. In: NeurIPS (2020)"},{"key":"29_CR30","doi-asserted-by":"crossref","unstructured":"Liu, F., Kim, M., Jain, A., Liu, X.: Controllable and guided face synthesis for unconstrained face recognition. In: ECCV (2022)","DOI":"10.1007\/978-3-031-19775-8_41"},{"key":"29_CR31","unstructured":"Liu, F., Liu, X.: Voxel-based 3D detection and reconstruction of multiple objects from a single image. NeurIPS (2021)"},{"key":"29_CR32","doi-asserted-by":"crossref","unstructured":"Liu, F., Tran, L., Liu, X.: Fully understanding generic objects: modeling, segmentation, and reconstruction. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00734"},{"key":"29_CR33","unstructured":"Liu, F., Zhao, Q., Liu, X., Zeng, D.: Joint face alignment and 3D face reconstruction with application to face recognition. TPAMI (2018)"},{"key":"29_CR34","doi-asserted-by":"crossref","unstructured":"Liu, S., Zhang, Y., Peng, S., Shi, B., Pollefeys, M., Cui, Z.: DIST: rendering deep implicit signed distance function with differentiable sphere tracing. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00209"},{"key":"29_CR35","doi-asserted-by":"crossref","unstructured":"Liu, S., Chen, W., Li, T., Li, H.: Soft rasterizer: differentiable rendering for unsupervised single-view mesh reconstruction. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00780"},{"key":"29_CR36","unstructured":"Lunz, S., Li, Y., Fitzgibbon, A., Kushman, N.: Inverse graphics GAN: learning to generate 3D shapes from unstructured 2D data. In: NeurIPS (2020)"},{"key":"29_CR37","doi-asserted-by":"crossref","unstructured":"Medin, S.C., et al.: MOST-GAN: 3D morphable StyleGan for disentangled face image manipulation. In: AAAI (2022)","DOI":"10.1609\/aaai.v36i2.20091"},{"key":"29_CR38","doi-asserted-by":"crossref","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: NeRF: representing scenes as neural radiance fields for view synthesis. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58452-8_24"},{"key":"29_CR39","doi-asserted-by":"crossref","unstructured":"Nguyen-Phuoc, T., Li, C., Theis, L., Richardt, C., Yang, Y.L.: HoloGAN: unsupervised learning of 3D representations from natural images. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00768"},{"key":"29_CR40","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Geiger, A.: CAMPARI: camera-aware decomposed generative neural radiance fields. arXiv preprint arXiv:2103.17269 (2021)","DOI":"10.1109\/3DV53792.2021.00103"},{"key":"29_CR41","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Geiger, A.: GIRAFFE: representing scenes as compositional generative neural feature fields. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01129"},{"key":"29_CR42","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Mescheder, L., Oechsle, M., Geiger, A.: Differentiable volumetric rendering: learning implicit 3D representations without 3D supervision. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00356"},{"key":"29_CR43","doi-asserted-by":"crossref","unstructured":"Osher, S., Fedkiw, R., Piechor, K.: Level set methods and dynamic implicit surfaces. Appl. Mech. Rev. 57, B15 (2004)","DOI":"10.1115\/1.1760520"},{"key":"29_CR44","unstructured":"Pan, X., Dai, B., Liu, Z., Loy, C.C., Luo, P.: Do 2D GANs know 3D shape? Unsupervised 3D shape reconstruction from 2D image GANs. In: ICLR (2021)"},{"key":"29_CR45","doi-asserted-by":"crossref","unstructured":"Pan, X., Zhan, X., Dai, B., Lin, D., Loy, C.C., Luo, P.: Exploiting deep generative prior for versatile image restoration and manipulation. TPAMI (2021)","DOI":"10.1007\/978-3-030-58536-5_16"},{"key":"29_CR46","doi-asserted-by":"crossref","unstructured":"Qin, X., Zhang, Z., Huang, C., Dehghan, M., Zaiane, O.R., Jagersand, M.: U2-Net: going deeper with nested u-structure for salient object detection. Patt. Recogn. 106, 107404 (2020)","DOI":"10.1016\/j.patcog.2020.107404"},{"key":"29_CR47","doi-asserted-by":"crossref","unstructured":"Rojtberg, P., P\u00f6llabauer, T., Kuijper, A.: Style-transfer GANs for bridging the domain gap in synthetic pose estimator training. In: AIVR (2020)","DOI":"10.1109\/AIVR50618.2020.00039"},{"key":"29_CR48","doi-asserted-by":"crossref","unstructured":"Rossler, A., Cozzolino, D., Verdoliva, L., Riess, C., Thies, J., Nie\u00dfner, M.: FaceForensics++: learning to detect manipulated facial images. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00009"},{"key":"29_CR49","unstructured":"Schwarz, K., Liao, Y., Niemeyer, M., Geiger, A.: GRAF: generative radiance fields for 3D-aware image synthesis. In: NeurIPS (2020)"},{"key":"29_CR50","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: InterfaceGAN: interpreting the disentangled face representation learned by GANs. TPAMI (2020)"},{"key":"29_CR51","doi-asserted-by":"crossref","unstructured":"Shen, Y., Zhou, B.: Closed-form factorization of latent semantics in GANs. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00158"},{"key":"29_CR52","doi-asserted-by":"crossref","unstructured":"Shi, Y., Aggarwal, D., Jain, A.K.: Lifting 2D StyleGAN for 3D-aware face generation. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00619"},{"key":"29_CR53","unstructured":"Sitzmann, V., Zollh\u00f6fer, M., Wetzstein, G.: Scene representation networks: Continuous 3D-structure-aware neural scene representations. In: NeurIPS (2019)"},{"key":"29_CR54","series-title":"Lecture Notes in Computer Science (Lecture Notes in Artificial Intelligence)","doi-asserted-by":"publisher","first-page":"339","DOI":"10.1007\/978-3-030-65390-3_26","volume-title":"Advanced Data Mining and Applications","author":"K Su","year":"2020","unstructured":"Su, K., Zhou, E., Sun, X., Wang, C., Yu, D., Luo, X.: Pre-trained StyleGAN based data augmentation for small sample brain CT motion artifacts detection. In: Yang, X., Wang, C.-D., Islam, M.S., Zhang, Z. (eds.) ADMA 2020. LNCS (LNAI), vol. 12447, pp. 339\u2013346. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-65390-3_26"},{"key":"29_CR55","unstructured":"Suzuki, R., Koyama, M., Miyato, T., Yonetsuji, T., Zhu, H.: Spatially controllable image synthesis with internal representation collaging. arXiv preprint arXiv:1811.10153 (2018)"},{"key":"29_CR56","unstructured":"Szab\u00f3, A., Meishvili, G., Favaro, P.: Unsupervised generative 3D shape learning from natural images. arXiv preprint arXiv:1910.00287 (2019)"},{"key":"29_CR57","doi-asserted-by":"crossref","unstructured":"Tran, L., Liu, X.: On learning 3D face morphable model from in-the-wild images. TPAMI (2019)","DOI":"10.1109\/TPAMI.2019.2927975"},{"key":"29_CR58","doi-asserted-by":"crossref","unstructured":"Tulsiani, S., Kar, A., Carreira, J., Malik, J.: Learning category-specific deformable 3D models for object reconstruction. TPAMI (2016)","DOI":"10.1109\/TPAMI.2016.2574713"},{"key":"29_CR59","doi-asserted-by":"crossref","unstructured":"Tulsiani, S., Zhou, T., Efros, A.A., Malik, J.: Multi-view supervision for single-view reconstruction via differentiable ray consistency. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.30"},{"key":"29_CR60","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Liu, M.Y., Yang, X., Kautz, J.: MoCoGAN: decomposing motion and content for video generation. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00165"},{"key":"29_CR61","doi-asserted-by":"crossref","unstructured":"Wang, X., Li, Y., Zhang, H., Shan, Y.: Towards real-world blind face restoration with generative facial prior. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00905"},{"key":"29_CR62","unstructured":"Wu, J., Zhang, C., Xue, T., Freeman, W.T., Tenenbaum, J.B.: Learning a probabilistic latent space of object shapes via 3D generative-adversarial modeling. In: NeurIPS (2016)"},{"key":"29_CR63","doi-asserted-by":"crossref","unstructured":"Wu, S., Rupprecht, C., Vedaldi, A.: Unsupervised learning of probably symmetric deformable 3D objects from images in the wild. In: CVPR (2020)","DOI":"10.24963\/ijcai.2021\/665"},{"key":"29_CR64","doi-asserted-by":"crossref","unstructured":"Wu, Y., Sun, Z., Song, Y., Sun, Y., Zhong, Y., Shi, J.: Shape-pose ambiguity in learning 3D reconstruction from images. In: AAAI (2021)","DOI":"10.1609\/aaai.v35i4.16405"},{"key":"29_CR65","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Mottaghi, R., Savarese, S.: Beyond pascal: a benchmark for 3D object detection in the wild. In: WACV (2014)","DOI":"10.1109\/WACV.2014.6836101"},{"key":"29_CR66","doi-asserted-by":"crossref","unstructured":"Xu, H., et al.: Digging into uncertainty in self-supervised multi-view stereo. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00602"},{"key":"29_CR67","doi-asserted-by":"crossref","unstructured":"Yang, T., Ren, P., Xie, X., Zhang, L.: GAN prior embedded network for blind face restoration in the wild. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00073"},{"key":"29_CR68","unstructured":"Yariv, L., et al.: Multiview neural surface reconstruction by disentangling geometry and appearance. In: NeurIPS (2020)"},{"key":"29_CR69","doi-asserted-by":"crossref","unstructured":"Ye, Y., Tulsiani, S., Gupta, A.: Shelf-supervised mesh prediction in the wild. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00873"},{"key":"29_CR70","unstructured":"Yu, F., Seff, A., Zhang, Y., Song, S., Funkhouser, T., Xiao, J.: LSUN: construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365 (2015)"},{"key":"29_CR71","unstructured":"Zhang, Y., et al.: Image GANs meet differentiable rendering for inverse graphics and interpretable 3D neural rendering. In: ICLR (2021)"},{"key":"29_CR72","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: DatasetGAN: efficient labeled data factory with minimal human effort. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01001"},{"key":"29_CR73","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"29_CR74","unstructured":"Zhu, J.Y., et al.: Visual object networks: image generation with disentangled 3D representation. In: NeurIPS (2018)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19769-7_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T14:10:04Z","timestamp":1710339004000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19769-7_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197680","9783031197697"],"references-count":74,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19769-7_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"23 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}