{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T01:39:57Z","timestamp":1773193197569,"version":"3.50.1"},"publisher-location":"Cham","reference-count":66,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031197833","type":"print"},{"value":"9783031197840","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19784-0_21","type":"book-chapter","created":{"date-parts":[[2022,10,30]],"date-time":"2022-10-30T14:02:50Z","timestamp":1667138570000},"page":"357-374","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":21,"title":["Temporally Consistent Semantic Video Editing"],"prefix":"10.1007","author":[{"given":"Yiran","family":"Xu","sequence":"first","affiliation":[]},{"given":"Badour","family":"AlBahar","sequence":"additional","affiliation":[]},{"given":"Jia-Bin","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,31]]},"reference":[{"key":"21_CR1","doi-asserted-by":"crossref","unstructured":"Abdal, R., Qin, Y., Wonka, P.: Image2stylegan: how to embed images into the stylegan latent space? In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00453"},{"key":"21_CR2","doi-asserted-by":"crossref","unstructured":"Abdal, R., Qin, Y., Wonka, P.: Image2stylegan++: how to edit the embedded images? In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00832"},{"issue":"3","key":"21_CR3","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3447648","volume":"40","author":"R Abdal","year":"2021","unstructured":"Abdal, R., Zhu, P., Mitra, N.J., Wonka, P.: Styleflow: attribute-conditioned exploration of stylegan-generated images using conditional continuous normalizing flows. ACM Trans. Graph. (TOG) 40(3), 1\u201321 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"21_CR4","doi-asserted-by":"crossref","unstructured":"Afifi, M., Brubaker, M.A., Brown, M.S.: Histogan: controlling colors of gan-generated and real images via color histograms. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2021)","DOI":"10.1109\/CVPR46437.2021.00785"},{"key":"21_CR5","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., Patashnik, O., Cohen-Or, D.: Only a matter of style: age transformation using a style-based regression model. arXiv preprint arXiv:2102.02754 (2021)","DOI":"10.1145\/3450626.3459805"},{"key":"21_CR6","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., Patashnik, O., Cohen-Or, D.: Restyle: a residual-based stylegan encoder via iterative refinement. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) (2021)","DOI":"10.1109\/ICCV48922.2021.00664"},{"key":"21_CR7","doi-asserted-by":"crossref","unstructured":"Alaluf, Y., et al.: Third time\u2019s the charm? image and video editing with stylegan3. arXiv preprint arXiv:2201.13433 (2022)","DOI":"10.1007\/978-3-031-25063-7_13"},{"key":"21_CR8","doi-asserted-by":"crossref","unstructured":"Bau, D., Strobelt, H., Peebles, W., Zhou, B., Zhu, J.Y., Torralba, A., et al.: Semantic photo manipulation with a generative image prior. arXiv preprint arXiv:2005.07727 (2020)","DOI":"10.1145\/3306346.3323023"},{"issue":"6","key":"21_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2816795.2818107","volume":"34","author":"N Bonneel","year":"2015","unstructured":"Bonneel, N., Tompkin, J., Sunkavalli, K., Sun, D., Paris, S., Pfister, H.: Blind video temporal consistency. ACM TOG 34(6), 1\u20139 (2015)","journal-title":"ACM TOG"},{"key":"21_CR10","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large scale gan training for high fidelity natural image synthesis (2019)"},{"key":"21_CR11","unstructured":"Chai, L., Wulff, J., Isola, P.: Using latent space regression to analyze and leverage compositionality in gans. In: International Conference on Learning Representations (2021)"},{"key":"21_CR12","doi-asserted-by":"crossref","unstructured":"Chen, D., Liao, J., Yuan, L., Yu, N., Hua, G.: Coherent online video style transfer. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.126"},{"key":"21_CR13","doi-asserted-by":"crossref","unstructured":"Collins, E., Bala, R., Price, B., Susstrunk, S.: Editing in style: uncovering the local semantics of gans. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5771\u20135780 (2020)","DOI":"10.1109\/CVPR42600.2020.00581"},{"key":"21_CR14","doi-asserted-by":"crossref","unstructured":"Daras, G., Odena, A., Zhang, H., Dimakis, A.G.: Your local gan: designing two dimensional local attention mechanisms for generative models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14531\u201314539 (2020)","DOI":"10.1109\/CVPR42600.2020.01454"},{"key":"21_CR15","doi-asserted-by":"crossref","unstructured":"Gal, R., Patashnik, O., Maron, H., Chechik, G., Cohen-Or, D.: Stylegan-nada: clip-guided domain adaptation of image generators (2021)","DOI":"10.1145\/3528223.3530164"},{"key":"21_CR16","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems, pp. 2672\u20132680 (2014)"},{"key":"21_CR17","doi-asserted-by":"crossref","unstructured":"Gu, J., Shen, Y., Zhou, B.: Image processing using multi-code gan prior. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3012\u20133021 (2020)","DOI":"10.1109\/CVPR42600.2020.00308"},{"key":"21_CR18","unstructured":"Gulrajani, I., Ahmed, F., Arjovsky, M., Dumoulin, V., Courville, A.C.: Improved training of wasserstein gans (2017)"},{"key":"21_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"152","DOI":"10.1007\/978-3-030-58529-7_10","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Guo","year":"2020","unstructured":"Guo, J., Zhu, X., Yang, Y., Yang, F., Lei, Z., Li, S.Z.: Towards fast, accurate and stable 3D dense face alignment. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12364, pp. 152\u2013168. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58529-7_10"},{"issue":"6","key":"21_CR20","first-page":"1","volume":"35","author":"JB Huang","year":"2016","unstructured":"Huang, J.B., Kang, S.B., Ahuja, N., Kopf, J.: Temporally coherent completion of dynamic video. ACM TOG 35(6), 1\u201311 (2016)","journal-title":"ACM TOG"},{"key":"21_CR21","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1007\/978-3-030-58536-5_2","volume-title":"Computer Vision \u2013 ECCV 2020","author":"M Huh","year":"2020","unstructured":"Huh, M., Zhang, R., Zhu, J.-Y., Paris, S., Hertzmann, A.: Transforming and projecting images into class-conditional generative networks. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 17\u201334. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_2"},{"key":"21_CR22","unstructured":"H\u00e4rk\u00f6nen, E., Hertzmann, A., Lehtinen, J., Paris, S.: Ganspace: discovering interpretable gan controls. In: Proceeding NeurIPS (2020)"},{"issue":"4","key":"21_CR23","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459860","volume":"40","author":"W Jang","year":"2021","unstructured":"Jang, W., Ju, G., Jung, Y., Yang, J., Tong, X., Lee, S.: Stylecarigan: caricature generation via stylegan feature map modulation. ACM Trans. Graph. (TOG) 40(4), 1\u201316 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"21_CR24","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive growing of gans for improved quality, stability, and variation. In: International Conference on Learning Representations (2018)"},{"key":"21_CR25","unstructured":"Karras, T., Aittala, M., Hellsten, J., Laine, S., Lehtinen, J., Aila, T.: Training generative adversarial networks with limited data. arXiv preprint arXiv:2006.06676 (2020)"},{"key":"21_CR26","unstructured":"Karras, T., Aittala, M., Hellsten, J., Laine, S., Lehtinen, J., Aila, T.: Training generative adversarial networks with limited data. In: Proceeding NeurIPS (2020)"},{"key":"21_CR27","unstructured":"Karras, T., et al.: Alias-free generative adversarial networks. In: Proceeding NeurIPS (2021)"},{"key":"21_CR28","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"21_CR29","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and improving the image quality of stylegan. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"21_CR30","doi-asserted-by":"crossref","unstructured":"Kasten, Y., Ofri, D., Wang, O., Dekel, T.: Layered neural atlases for consistent video editing. In: ACM TOG (2021)","DOI":"10.1145\/3478513.3480546"},{"key":"21_CR31","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"21_CR32","first-page":"1097","volume":"25","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. Adv. Neural. Inf. Process. Syst. 25, 1097\u20131105 (2012)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"21_CR33","first-page":"1435","volume":"24","author":"S Kwong","year":"2021","unstructured":"Kwong, S., Huang, J., Liao, J.: Unsupervised image-to-image translation via pre-trained stylegan2 network. IEEE Trans. Multimedia 24, 1435\u20131448 (2021)","journal-title":"IEEE Trans. Multimedia"},{"key":"21_CR34","doi-asserted-by":"crossref","unstructured":"Lai, W.S., Huang, J.B., Wang, O., Shechtman, E., Yumer, E., Yang, M.H.: Learning blind video temporal consistency. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01267-0_11"},{"key":"21_CR35","doi-asserted-by":"crossref","unstructured":"Lai, W.S., Huang, J.B., Wang, O., Shechtman, E., Yumer, E., Yang, M.H.: Learning blind video temporal consistency. In: European Conference on Computer Vision (2018)D","DOI":"10.1007\/978-3-030-01267-0_11"},{"key":"21_CR36","unstructured":"Lei, C., Xing, Y., Chen, Q.: Blind video temporal consistency via deep video prior. In: Advances in Neural Information Processing Systems (2020)"},{"key":"21_CR37","unstructured":"Li, B., et al.: Dystyle: dynamic neural network for multi-attribute-conditioned style editing. arXiv preprint arXiv:2109.10737 (2021)"},{"key":"21_CR38","doi-asserted-by":"crossref","unstructured":"Liu, Y.L., Lai, W.S., Yang, M.H., Chuang, Y.Y., Huang, J.B.: Learning to see through obstructions. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01422"},{"key":"21_CR39","doi-asserted-by":"crossref","unstructured":"Livingstone, S.R., Russo, F.A.: The ryerson audio-visual database of emotional speech and song (ravdess): a dynamic, multimodal set of facial and vocal expressions in North American English. PLoS ONE, 13(5), e0196391 (2018)","DOI":"10.1371\/journal.pone.0196391"},{"key":"21_CR40","doi-asserted-by":"publisher","unstructured":"Luo, J., Xu, Y., Tang, C., Lv, J.: Learning inverse mapping by autoencoder based generative adversarial nets. In: International Conference on Neural Information Processing, pp. 207\u2013216. Springer (2017). https:\/\/doi.org\/10.1007\/978-3-319-70096-0_22","DOI":"10.1007\/978-3-319-70096-0_22"},{"key":"21_CR41","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul Smolley, S.: Least squares generative adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2794\u20132802 (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"21_CR42","unstructured":"Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral normalization for generative adversarial networks (2018)"},{"key":"21_CR43","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3414685.3417826","volume":"39","author":"Y Nitzan","year":"2020","unstructured":"Nitzan, Y., Bermano, A., Li, Y., Cohen-Or, D.: Face identity disentanglement via latent space mapping. ACM Trans. Graph. (TOG) 39, 1\u201314 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"21_CR44","doi-asserted-by":"crossref","unstructured":"Patashnik, O., Wu, Z., Shechtman, E., Cohen-Or, D., Lischinski, D.: Styleclip: text-driven manipulation of stylegan imagery. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV), pp. 2085\u20132094 (2021)","DOI":"10.1109\/ICCV48922.2021.00209"},{"key":"21_CR45","doi-asserted-by":"crossref","unstructured":"Raj, A., Li, Y., Bresler, Y.: Gan-based projector for faster recovery with convergence guarantees in linear inverse problems. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5602\u20135611 (2019)","DOI":"10.1109\/ICCV.2019.00570"},{"key":"21_CR46","doi-asserted-by":"crossref","unstructured":"Rav-Acha, A., Kohli, P., Rother, C., Fitzgibbon, A.: Unwrap mosaics: a new representation for video editing. In: ACM TOG (2008)","DOI":"10.1145\/1399504.1360616"},{"key":"21_CR47","doi-asserted-by":"crossref","unstructured":"Rho, D., Cho, J., Ko, J.H., Park, E.: Neural residual flow fields for efficient video representations. arXiv preprint arXiv:2201.04329 (2022)","DOI":"10.1007\/978-3-031-26284-5_28"},{"key":"21_CR48","doi-asserted-by":"crossref","unstructured":"Richardson, E., et al.: Encoding in style: a stylegan encoder for image-to-image translation. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2021)","DOI":"10.1109\/CVPR46437.2021.00232"},{"key":"21_CR49","doi-asserted-by":"crossref","unstructured":"Roich, D., Mokady, R., Bermano, A.H., Cohen-Or, D.: Pivotal tuning for latent-based editing of real images. arXiv preprint arXiv:2106.05744 (2021)","DOI":"10.1145\/3544777"},{"key":"21_CR50","doi-asserted-by":"crossref","unstructured":"Saha, R., Duke, B., Shkurti, F., Taylor, G.W., Aarabi, P.: Loho: latent optimization of hairstyles via orthogonalization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1984\u20131993 (2021)","DOI":"10.1109\/CVPR46437.2021.00202"},{"key":"21_CR51","unstructured":"Shen, Y., Yang, C., Tang, X., Zhou, B.: Interfacegan: interpreting the disentangled face representation learned by gans. In: TPAMI (2020)"},{"key":"21_CR52","doi-asserted-by":"crossref","unstructured":"Shen, Y., Zhou, B.: Closed-form factorization of latent semantics in gans. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00158"},{"key":"21_CR53","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: Raft: recurrent all-pairs field transforms for optical flow. arXiv preprint arXiv:2003.12039 (2020)","DOI":"10.24963\/ijcai.2021\/662"},{"key":"21_CR54","doi-asserted-by":"crossref","unstructured":"Tewari, A., et al.: Pie: portrait image embedding for semantic control. arXiv preprint arXiv:2009.09485 (2020)","DOI":"10.1145\/3414685.3417803"},{"key":"21_CR55","doi-asserted-by":"crossref","unstructured":"Tewari, A., et al.: Stylerig: rigging stylegan for 3d control over portrait images. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 6142\u20136151 (2020)","DOI":"10.1109\/CVPR42600.2020.00618"},{"issue":"4","key":"21_CR56","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459838","volume":"40","author":"O Tov","year":"2021","unstructured":"Tov, O., Alaluf, Y., Nitzan, Y., Patashnik, O., Cohen-Or, D.: Designing an encoder for stylegan image manipulation. ACM Trans. Graph. (TOG) 40(4), 1\u201314 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"21_CR57","doi-asserted-by":"crossref","unstructured":"Tzaban, R., Mokady, R., Gal, R., Bermano, A.H., Cohen-Or, D.: Stitch it in time: gan-based facial editing of real videos. arXiv preprint arXiv:2201.08361 (2022)","DOI":"10.1145\/3550469.3555382"},{"key":"21_CR58","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"170","DOI":"10.1007\/978-3-030-58542-6_11","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Viazovetskyi","year":"2020","unstructured":"Viazovetskyi, Y., Ivashkin, V., Kashin, E.: StyleGAN2 distillation for feed-forward image manipulation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12367, pp. 170\u2013186. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58542-6_11"},{"issue":"4","key":"21_CR59","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3450626.3459814","volume":"40","author":"Y Wu","year":"2021","unstructured":"Wu, Y., Yang, Y.L., Xiao, Q., Jin, X.: Coarse-to-fine: facial structure editing of portrait images via latent space classifications. ACM Trans. Graph. (TOG) 40(4), 1\u201313 (2021)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"21_CR60","doi-asserted-by":"crossref","unstructured":"Wu, Z., Lischinski, D., Shechtman, E.: Stylespace analysis: disentangled controls for stylegan image generation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12863\u201312872 (2021)","DOI":"10.1109\/CVPR46437.2021.01267"},{"key":"21_CR61","doi-asserted-by":"crossref","unstructured":"Xia, W., Zhang, Y., Yang, Y., Xue, J.H., Zhou, B., Yang, M.H.: Gan inversion: a survey. arXiv preprint arXiv:2101.05278 (2021)","DOI":"10.1109\/TPAMI.2022.3181070"},{"key":"21_CR62","doi-asserted-by":"crossref","unstructured":"Yao, X., Newson, A., Gousseau, Y., Hellier, P.: A latent transformer for disentangled face editing in images and videos. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13789\u201313798 (2021)","DOI":"10.1109\/ICCV48922.2021.01353"},{"key":"21_CR63","doi-asserted-by":"crossref","unstructured":"Y\u00fcksel, O.K., Simsar, E., Er, E.G., Yanardag, P.: Latentclr: a contrastive learning approach for unsupervised discovery of interpretable directions. arXiv preprint arXiv:2104.00820 (2021)","DOI":"10.1109\/ICCV48922.2021.01400"},{"key":"21_CR64","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"21_CR65","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"592","DOI":"10.1007\/978-3-030-58520-4_35","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Zhu","year":"2020","unstructured":"Zhu, J., Shen, Y., Zhao, D., Zhou, B.: In-Domain GAN inversion for real image editing. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12362, pp. 592\u2013608. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58520-4_35"},{"key":"21_CR66","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"597","DOI":"10.1007\/978-3-319-46454-1_36","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J-Y Zhu","year":"2016","unstructured":"Zhu, J.-Y., Kr\u00e4henb\u00fchl, P., Shechtman, E., Efros, A.A.: Generative visual manipulation on the natural image manifold. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9909, pp. 597\u2013613. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46454-1_36"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19784-0_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,6]],"date-time":"2024-10-06T21:41:29Z","timestamp":1728250889000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19784-0_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197833","9783031197840"],"references-count":66,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19784-0_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"31 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}