{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,3]],"date-time":"2026-02-03T19:34:48Z","timestamp":1770147288445,"version":"3.49.0"},"publisher-location":"Cham","reference-count":65,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197864","type":"print"},{"value":"9783031197871","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19787-1_36","type":"book-chapter","created":{"date-parts":[[2022,10,20]],"date-time":"2022-10-20T22:16:11Z","timestamp":1666304171000},"page":"632-650","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":16,"title":["CoGS: Controllable Generation and\u00a0Search from\u00a0Sketch and\u00a0Style"],"prefix":"10.1007","author":[{"given":"Cusuh","family":"Ham","sequence":"first","affiliation":[]},{"given":"Gemma Canet","family":"Tarr\u00e9s","sequence":"additional","affiliation":[]},{"given":"Tu","family":"Bui","sequence":"additional","affiliation":[]},{"given":"James","family":"Hays","sequence":"additional","affiliation":[]},{"given":"Zhe","family":"Lin","sequence":"additional","affiliation":[]},{"given":"John","family":"Collomosse","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,21]]},"reference":[{"key":"36_CR1","doi-asserted-by":"crossref","unstructured":"Ashual, O., Wolf, L.: Specifying object attributes and relations in interactive scene generation. In: Proceedings of the CVPR (2019)","DOI":"10.1109\/ICCV.2019.00466"},{"issue":"3","key":"36_CR2","doi-asserted-by":"publisher","first-page":"24","DOI":"10.1145\/1531326.1531330","volume":"28","author":"C Barnes","year":"2009","unstructured":"Barnes, C., Shechtman, E., Finkelstein, A., Goldman, D.B.: PatchMatch: a randomized correspondence algorithm for structural image editing. ACM Trans. Graph. 28(3), 24 (2009)","journal-title":"ACM Trans. Graph."},{"issue":"1","key":"36_CR3","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/s41095-016-0064-2","volume":"3","author":"C Barnes","year":"2016","unstructured":"Barnes, C., Zhang, F.-L.: A survey of the state-of-the-art in patch-based synthesis. Comput. Visual Media 3(1), 3\u201320 (2016). https:\/\/doi.org\/10.1007\/s41095-016-0064-2","journal-title":"Comput. Visual Media"},{"key":"36_CR4","doi-asserted-by":"crossref","unstructured":"Bui, T., Ribeiro, L., Collomosse, J., Ponti, M.: Sketching out the details: Sketch-based image retrieval using convolutional neural networks with multi-stage regression. Comput. Graph. 71, 77\u201387 (2018)","DOI":"10.1016\/j.cag.2017.12.006"},{"key":"36_CR5","doi-asserted-by":"publisher","first-page":"679","DOI":"10.1109\/TPAMI.1986.4767851","volume":"6","author":"J Canny","year":"1986","unstructured":"Canny, J.: A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 6, 679\u2013698 (1986)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"36_CR6","unstructured":"Casanova, A., Careil, M., Verbeek, J., Drozdzal, M., Romero-Soriano, A.: Instance-conditioned gan. arXiv preprint arXiv:2109.05070 (2021)"},{"issue":"5","key":"36_CR7","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1145\/1618452.1618470","volume":"28","author":"T Chen","year":"2009","unstructured":"Chen, T., Cheng, M.M., Tan, P., Shamir, A., Hu, S.M.: Sketch2Photo: Internet image montage. Proc ACM SIGGRAPH 28(5), 124 (2009)","journal-title":"Proc ACM SIGGRAPH"},{"key":"36_CR8","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: III, H.D., Singh, A. (eds.) Proceedings of the 37th International Conference on Machine Learning. Proceedings of Machine Learning Research, vol. 119, pp. 1597\u20131607. PMLR, 13\u201318 July 2020. https:\/\/proceedings.mlr.press\/v119\/chen20j.html"},{"key":"36_CR9","doi-asserted-by":"crossref","unstructured":"Chen, W., Hays, J.: SketchyGAN: towards diverse and realistic sketch to image synthesis. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2018","DOI":"10.1109\/CVPR.2018.00981"},{"key":"36_CR10","unstructured":"Chen, X., Duan, Y., Houthooft, R., Schulman, J., Sutskever, I., Abbeel, P.: InfoGAN: interpretable representation learning by information maximizing generative adversarial nets. In: 30th Conference on Neural Information Processing Systems (NIPS 2016), Barcelona, Spain, June 2016"},{"key":"36_CR11","doi-asserted-by":"crossref","unstructured":"Collomosse, J., Bui, T., Wilber, M., Fang, C., Jin, H.: Sketching with style: Visual search with sketches and aesthetic context. In: Proceedings of the ICCV (2017)","DOI":"10.1109\/ICCV.2017.290"},{"key":"36_CR12","doi-asserted-by":"crossref","unstructured":"Collomosse, J.P., McNeill, G., Watts, L.: Free-hand sketch grouping for video retrieval. In: Proceedings of the ICPR (2008)","DOI":"10.1109\/ICPR.2008.4761466"},{"key":"36_CR13","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"36_CR14","doi-asserted-by":"crossref","unstructured":"Efros, A., Freeman, W.: Image quilting for texture synthesis and transfer. In: Proceedings of the SIGGRAPH (2001)","DOI":"10.1145\/383259.383296"},{"key":"36_CR15","doi-asserted-by":"crossref","unstructured":"Eitz, M., Hays, J., Alexa, M.: How do humans sketch objects? ACM Trans. Graph. (Proc. SIGGRAPH) 31(4), 44:1\u201344:10 (2012)","DOI":"10.1145\/2185520.2335395"},{"key":"36_CR16","doi-asserted-by":"crossref","unstructured":"Esser, P., Rombach, R., Ommer, B.: Taming transformers for high-resolution image synthesis (2020)","DOI":"10.1109\/CVPR46437.2021.01268"},{"key":"36_CR17","doi-asserted-by":"crossref","unstructured":"Gao, C., Liu, Q., Xu, Q., Wang, L., Liu, J., Zou, C.: SketchyCOCO: image generation from freehand scene sketches. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), June 2020","DOI":"10.1109\/CVPR42600.2020.00522"},{"issue":"13","key":"36_CR18","doi-asserted-by":"publisher","first-page":"3006","DOI":"10.1049\/iet-ipr.2018.5767","volume":"14","author":"H Gao","year":"2020","unstructured":"Gao, H., Chen, Z., Huang, B., Chen, J., Li, Z.: Image super-resolution based on conditional generative adversarial network. IET Image Proc. 14(13), 3006\u20133013 (2020)","journal-title":"IET Image Proc."},{"key":"36_CR19","doi-asserted-by":"publisher","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2414\u20132423 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.265","DOI":"10.1109\/CVPR.2016.265"},{"key":"36_CR20","doi-asserted-by":"crossref","unstructured":"Ghosh, A., et al.: Interactive sketch & fill: multiclass sketch-to-image translation. In: Proceedings of the IEEE International Conference on Computer Vision (2019)","DOI":"10.1109\/ICCV.2019.00126"},{"key":"36_CR21","doi-asserted-by":"crossref","unstructured":"Gucluturk, Y., Guclu, U., van Lier, R., van Gerven, M.A.: Convolutional sketch inversion. In: Proceedings of the ECCV Workshop on Vision and Art (VISART) (2016)","DOI":"10.1007\/978-3-319-46604-0_56"},{"key":"36_CR22","doi-asserted-by":"crossref","unstructured":"Guo, X., Yang, H., Huang, D.: Image inpainting via conditional texture and structure dual generation. In: Conference: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.01387"},{"issue":"3","key":"36_CR23","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1145\/1276377.1276382","volume":"26","author":"J Hays","year":"2007","unstructured":"Hays, J., Efros, A.A.: Scene completion using millions of photographs. ACM Trans. Graph. 26(3), 4 (2007)","journal-title":"ACM Trans. Graph."},{"key":"36_CR24","doi-asserted-by":"crossref","unstructured":"Hertzmann, A., Jacobs, C.E., Oliver, N., Curless, B., Salesin, D.H.: Image analogies. In: Proceedings of the ACM SIGGRAPH. pp. 327\u2013340 (2001)","DOI":"10.1145\/383259.383295"},{"key":"36_CR25","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local Nash equilibrium. Adv. Neural Inf. Process. Syst. 30 (2017)"},{"key":"36_CR26","unstructured":"Hjelm, R.D., et al.: Learning deep representations by mutual information estimation and maximization. In: International Conference on Learning Representations (2019). https:\/\/openreview.net\/forum?id=Bklr3j0cKX"},{"key":"36_CR27","unstructured":"Hospedales, T., Song, Y.Z.: Sketch me that shoe. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), January 2016"},{"key":"36_CR28","doi-asserted-by":"crossref","unstructured":"Huang, X., Mallya, A., Wang, T.C., Liu, M.Y.: Multimodal conditional image synthesis with product-of-experts GANs (2021)","DOI":"10.1007\/978-3-031-19787-1_6"},{"key":"36_CR29","doi-asserted-by":"crossref","unstructured":"Hwang, J., Oh, S.W., Lee, J., Han, B.: Exemplar-based open-set panoptic segmentation network. CoRR abs\/2105.08336 (2021). https:\/\/arxiv.org\/abs\/2105.08336","DOI":"10.1109\/CVPR46437.2021.00123"},{"key":"36_CR30","unstructured":"H\u00e9naff, O.J., Razavi, A., Doersch, C., Eslami, S.M.A., Oord, A.v.d.: Data-efficient image recognition with contrastive predictive coding (2019). https:\/\/arxiv.org\/abs\/1905.09272, cite arxiv:1905.09272"},{"key":"36_CR31","doi-asserted-by":"crossref","unstructured":"Iizuka, S., Simo-Serra, E., Ishikawa, H.: Let there be color!: Joint end-to-end learning of global and local image priors for automatic image colorization with simultaneous classification. ACM Trans. Graph. (Proc. of SIGGRAPH 2016) 35(6) (2016)","DOI":"10.1145\/2897824.2925974"},{"key":"36_CR32","doi-asserted-by":"crossref","unstructured":"Inoue, N., Ito, D., Xu, N., Yang, J., Price, B., Yamasaki, T.: Learning to trace: expressive line drawing generation from photographs. Comput. Graph. Forum 38(7), 69\u201380 (2019)","DOI":"10.1111\/cgf.13817"},{"key":"36_CR33","doi-asserted-by":"publisher","unstructured":"Isola, P., Zhu, J., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 5967\u20135976 (2017). https:\/\/doi.org\/10.1109\/CVPR.2017.632","DOI":"10.1109\/CVPR.2017.632"},{"key":"36_CR34","unstructured":"Johnson, J., Gupta, A., Fei-Fei, L.: Image synthesis from reconfigurable layout and style. In: Proceedings of the CVPR (2018)"},{"key":"36_CR35","unstructured":"Jongejan, J., Rowley, H., Kawashima, T., Kim, J., Fox-Gieg, N.: The quick, draw! A.I. experiment (2016). https:\/\/quickdraw.withgoogle.com\/"},{"key":"36_CR36","unstructured":"Kingma, D.P., Welling, M.: Auto-encoding variational bayes. ArXiv e-prints, December 2013"},{"key":"36_CR37","doi-asserted-by":"crossref","unstructured":"Lu, Y., Wu, S., Tai, Y.W., Tang, C.K.: Image generation from sketch constraint using contextual GAN. In: The European Conference on Computer Vision (ECCV), September 2018","DOI":"10.1007\/978-3-030-01270-0_13"},{"key":"36_CR38","unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint arXiv:1411.1784 (2014)"},{"key":"36_CR39","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"36_CR40","unstructured":"Ramesh, A., et al.: Zero-shot text-to-image generation. arXiv preprint arXiv:2102.12092 (2021)"},{"key":"36_CR41","unstructured":"Reed, S., Akata, Z., Mohan, S., Tenka, S., Schiele, B., Lee, H.: Learning what and where to draw. In: Advances in Neural Information Processing Systems (NIPS) (2016)"},{"key":"36_CR42","unstructured":"Reed, S., Akata, Z., Yan, X., Logeswaran, L., Schiele, B., Lee, H.: Generative adversarial text-to-image synthesis. In: Proceedings ICML (2016)"},{"key":"36_CR43","unstructured":"Ribeiro, L., Bui, T., Collomosse, J., Ponti, M.: Scene designer: a unified model for scene search and synthesis from sketch. In: Proceedings of CVPRW on Sketch and Human Expressivity (SHE) (2021)"},{"key":"36_CR44","unstructured":"Ribeiro, L.S.F., Bui, T., Collomosse, J., Ponti, M.: Sketchformer: transformer-based representation for sketched structure. In: Proceedings of CVPR (2020)"},{"key":"36_CR45","doi-asserted-by":"crossref","unstructured":"Ruta, D., et al.: Aladin: all layer adaptive instance normalization for fine-grained style similarity. In: 2021 IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 11906\u201311915 (2021)","DOI":"10.1109\/ICCV48922.2021.01171"},{"issue":"4","key":"36_CR46","doi-asserted-by":"publisher","first-page":"119","DOI":"10.1145\/2897824.2925954","volume":"35","author":"P Sangkloy","year":"2016","unstructured":"Sangkloy, P., Burnell, N., Ham, C., Hays, J.: The sketchy database: learning to retrieve badly drawn bunnies. ACM Trans. Graph. 35(4), 119 (2016)","journal-title":"ACM Trans. Graph."},{"key":"36_CR47","doi-asserted-by":"publisher","unstructured":"Sangkloy, P., Burnell, N., Ham, C., Hays, J.: The sketchy database: Learning to retrieve badly drawn bunnies. ACM Trans. Graph. 35(4) (2016). https:\/\/doi.org\/10.1145\/2897824.2925954, https:\/\/doi.org\/10.1145\/2897824.2925954","DOI":"10.1145\/2897824.2925954"},{"key":"36_CR48","doi-asserted-by":"crossref","unstructured":"Sangkloy, P., Lu, J., Fang, C., Yu, F., Hays, J.: Scribbler: controlling deep image synthesis with sketch and color. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5400\u20135409 (2017)","DOI":"10.1109\/CVPR.2017.723"},{"key":"36_CR49","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"36_CR50","doi-asserted-by":"crossref","unstructured":"Song, J., Song, Y.Z., Xiang, T., Hospedales, T., Ruan, X.: Deep multi-task attribute-driven ranking for fine-grained sketch-based image retrieval. In: British Machine Vision Conference (2016)","DOI":"10.5244\/C.30.132"},{"key":"36_CR51","doi-asserted-by":"crossref","unstructured":"Song, J., Yu, Q., Song, Y.Z., Xiang, T., Hospedales, T.M.: Deep spatial-semantic attention for fine-grained sketch-based image retrieval. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), October 2017","DOI":"10.1109\/ICCV.2017.592"},{"key":"36_CR52","doi-asserted-by":"crossref","unstructured":"Sun, W., Wu, T.: Image synthesis from reconfigurable layout and style. In: Proceedings of CVPR (2019)","DOI":"10.1109\/ICCV.2019.01063"},{"key":"36_CR53","doi-asserted-by":"crossref","unstructured":"Sylvain, T., Zhang, P., Bengio, Y., Hjelm, D., Sharma, S.: Object-centric image generation from layouts. arXiv preprint arXiv:2003.07449 (2020)","DOI":"10.1609\/aaai.v35i3.16368"},{"key":"36_CR54","unstructured":"Tang, H., Liu, H., Xu, D., Torr, P., Sebe, N.: Attentiongan: unpaired image-to-image translation using attention-guided generative adversarial networks. arXiv preprint arXiv:1911.11897 (2019)"},{"key":"36_CR55","unstructured":"Tian, Y., Krishnan, D., Isola, P.: Contrastive multiview coding. CoRR abs\/1906.05849 (2019). https:\/\/arxiv.org\/abs\/1906.05849"},{"key":"36_CR56","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"key":"36_CR57","unstructured":"Wexler, Y., Shechtman, E., Irani, M.: Space-time video completion. In: Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2004. CVPR 2004. vol. 1, pp. I-I. IEEE (2004)"},{"key":"36_CR58","doi-asserted-by":"crossref","unstructured":"Xian, W., et al.: TextureGAN: controlling deep image synthesis with texture patches. arXiv preprint arXiv:1706.02823 (2017)","DOI":"10.1109\/CVPR.2018.00882"},{"issue":"1","key":"36_CR59","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/s41095-021-0234-8","volume":"8","author":"Y Xue","year":"2021","unstructured":"Xue, Y., Guo, Y.-C., Zhang, H., Xu, T., Zhang, S.-H., Huang, X.: Deep image synthesis from intuitive user input: a review and perspectives. Comput. Visual Media 8(1), 3\u201331 (2021). https:\/\/doi.org\/10.1007\/s41095-021-0234-8","journal-title":"Comput. Visual Media"},{"key":"36_CR60","unstructured":"Yang, Y., Hossain, M.Z., Gedeon, T., Rahman, S.: S2FGAN: semantically aware interactive sketch-to-face translation. arXiv preprint arXiv:2011.14785 (2020)"},{"key":"36_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"36_CR62","doi-asserted-by":"crossref","unstructured":"Zhao, B., Meng, L., Yin, W., Sigal, L.: Image generation from layout. In: Proceedings of CVPR (2019)","DOI":"10.1109\/CVPR.2019.00878"},{"key":"36_CR63","unstructured":"Zhou, X., et al.: Full-resolution correspondence learning for image translation. CoRR abs\/2012.02047 (2020). https:\/\/arxiv.org\/abs\/2012.02047"},{"key":"36_CR64","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Krahenbuhl, P., Shechtman, E., Efros, A.A.: Generative visual manipulation on the natural image manifold. In: Proceedings of ECCV (2016)","DOI":"10.1007\/978-3-319-46454-1_36"},{"key":"36_CR65","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. arXiv preprint arXiv:1703.10593 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19787-1_36","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T19:25:42Z","timestamp":1710357942000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19787-1_36"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197864","9783031197871"],"references-count":65,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19787-1_36","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"21 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}