{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T16:55:05Z","timestamp":1775580905781,"version":"3.50.1"},"publisher-location":"Cham","reference-count":60,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031263125","type":"print"},{"value":"9783031263132","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-26313-2_26","type":"book-chapter","created":{"date-parts":[[2023,3,1]],"date-time":"2023-03-01T08:02:32Z","timestamp":1677657752000},"page":"427-443","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":11,"title":["Fine-Grained Image Style Transfer with\u00a0Visual Transformers"],"prefix":"10.1007","author":[{"given":"Jianbo","family":"Wang","sequence":"first","affiliation":[]},{"given":"Huan","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Jianlong","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Toshihiko","family":"Yamasaki","sequence":"additional","affiliation":[]},{"given":"Baining","family":"Guo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,3,2]]},"reference":[{"key":"26_CR1","unstructured":"Date, P., Ganesan, A., Oates, T.: Fashioning with networks: neural style transfer to design clothes. In: KDD ML4Fashion workshop, vol. 2 (2017)"},{"key":"26_CR2","doi-asserted-by":"crossref","unstructured":"Chen, D., Liao, J., Yuan, L., Yu, N., Hua, G.: Coherent online video style transfer. In: ICCV, pp. 1105\u20131114 (2017)","DOI":"10.1109\/ICCV.2017.126"},{"key":"26_CR3","first-page":"1594","volume":"15","author":"W Zhang","year":"2013","unstructured":"Zhang, W., Cao, C., Chen, S., Liu, J., Tang, X.: Style transfer via image component analysis. TMM 15, 1594\u20131601 (2013)","journal-title":"TMM"},{"key":"26_CR4","first-page":"1724","volume":"20","author":"J Liu","year":"2017","unstructured":"Liu, J., Yang, W., Sun, X., Zeng, W.: Photo stylistic brush: robust style transfer via superpixel-based bipartite graph. TMM 20, 1724\u20131737 (2017)","journal-title":"TMM"},{"key":"26_CR5","first-page":"2245","volume":"23","author":"JJ Virtusio","year":"2021","unstructured":"Virtusio, J.J., Ople, J.J.M., Tan, D.S., Tanveer, M., Kumar, N., Hua, K.L.: Neural style palette: a multimodal and interactive style transfer from a single style image. TMM 23, 2245\u20132258 (2021)","journal-title":"TMM"},{"key":"26_CR6","doi-asserted-by":"crossref","unstructured":"Matsuo, S., Shimoda, W., Yanai, K.: Partial style transfer using weakly supervised semantic segmentation. In: ICME Workshops, pp. 267\u2013272. IEEE (2017)","DOI":"10.1109\/ICMEW.2017.8026228"},{"key":"26_CR7","first-page":"298","volume":"22","author":"BK Kim","year":"2019","unstructured":"Kim, B.K., Kim, G., Lee, S.Y.: Style-controlled synthesis of clothing segments for fashion image manipulation. TMM 22, 298\u2013310 (2019)","journal-title":"TMM"},{"key":"26_CR8","first-page":"2209","volume":"21","author":"Y Liu","year":"2019","unstructured":"Liu, Y., Chen, W., Liu, L., Lew, M.S.: SwapGAN: a multistage generative approach for person-to-person fashion style transfer. TMM 21, 2209\u20132222 (2019)","journal-title":"TMM"},{"key":"26_CR9","doi-asserted-by":"crossref","unstructured":"Castillo, C., De, S., Han, X., Singh, B., Yadav, A.K., Goldstein, T.: Son of Zorn\u2019s lemma: targeted style transfer using instance-aware semantic segmentation. In: ICASSP, pp. 1348\u20131352. IEEE (2017)","DOI":"10.1109\/ICASSP.2017.7952376"},{"key":"26_CR10","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: Image style transfer using convolutional neural networks. In: CVPR, pp. 2414\u20132423 (2016)","DOI":"10.1109\/CVPR.2016.265"},{"key":"26_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"26_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"702","DOI":"10.1007\/978-3-319-46487-9_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"C Li","year":"2016","unstructured":"Li, C., Wand, M.: Precomputed real-time texture synthesis with Markovian generative adversarial networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 702\u2013716. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_43"},{"key":"26_CR13","unstructured":"Ulyanov, D., Lebedev, V., Vedaldi, A., Lempitsky, V.S.: Texture networks: feed-forward synthesis of textures and stylized images. In: ICML, p. 4 (2016)"},{"key":"26_CR14","unstructured":"Dumoulin, V., Shlens, J., Kudlur, M.: A learned representation for artistic style. arXiv preprint arXiv:1610.07629 (2016)"},{"key":"26_CR15","doi-asserted-by":"crossref","unstructured":"Chen, D., Yuan, L., Liao, J., Yu, N., Hua, G.: Stylebank: an explicit representation for neural image style transfer. In: CVPR, pp. 1897\u20131906 (2017)","DOI":"10.1109\/CVPR.2017.296"},{"key":"26_CR16","doi-asserted-by":"crossref","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Diversified texture synthesis with feed-forward networks. In: CVPR, pp. 3920\u20133928 (2017)","DOI":"10.1109\/CVPR.2017.36"},{"key":"26_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"349","DOI":"10.1007\/978-3-030-11018-5_32","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"H Zhang","year":"2019","unstructured":"Zhang, H., Dana, K.: Multi-style generative network for real-time transfer. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11132, pp. 349\u2013365. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11018-5_32"},{"key":"26_CR18","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: ICCV, pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"26_CR19","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Universal style transfer via feature transforms. arXiv preprint arXiv:1705.08086 (2017)"},{"key":"26_CR20","unstructured":"Kitov, V., Kozlovtsev, K., Mishustina, M.: Depth-Aware Arbitrary style transfer using instance normalization. arXiv preprint arXiv:1906.01123 (2019)"},{"key":"26_CR21","doi-asserted-by":"crossref","unstructured":"Hu, Z., Jia, J., Liu, B., Bu, Y., Fu, J.: Aesthetic-aware image style transfer. In: ACM MM, pp. 3320\u20133329 (2020)","DOI":"10.1145\/3394171.3413853"},{"key":"26_CR22","doi-asserted-by":"crossref","unstructured":"Deng, Y., Tang, F., Pan, X., Dong, W., Ma, C., Xu, C.: Stytr$$\\hat{\\,}$$2: unbiased image style transfer with transformers. In: CVPR (2021)","DOI":"10.1109\/CVPR52688.2022.01104"},{"key":"26_CR23","unstructured":"Vaswani, A., et al.: Attention is all you need. arXiv preprint arXiv:1706.03762 (2017)"},{"key":"26_CR24","doi-asserted-by":"crossref","unstructured":"Mahendran, A., Vedaldi, A.: Understanding deep image representations by inverting them. In: CVPR, pp. 5188\u20135196 (2015)","DOI":"10.1109\/CVPR.2015.7299155"},{"key":"26_CR25","doi-asserted-by":"crossref","unstructured":"Gatys, L.A., Ecker, A.S., Bethge, M.: A neural algorithm of artistic style. arXiv preprint arXiv:1508.06576 (2015)","DOI":"10.1167\/16.12.326"},{"key":"26_CR26","unstructured":"Chen, T.Q., Schmidt, M.: Fast patch-based style transfer of arbitrary style. arXiv preprint arXiv:1612.04337 (2016)"},{"key":"26_CR27","doi-asserted-by":"crossref","unstructured":"An, J., Huang, S., Song, Y., Dou, D., Liu, W., Luo, J.: Artflow: unbiased image style transfer via reversible neural flows. In: CVPR, pp. 862\u2013871 (2021)","DOI":"10.1109\/CVPR46437.2021.00092"},{"key":"26_CR28","doi-asserted-by":"crossref","unstructured":"Deng, Y., Tang, F., Dong, W., Huang, H., Ma, C., Xu, C.: Arbitrary video style transfer via multi-channel correlation. In: AAA, vol. 1, pp. 1210\u20131217 (2021)","DOI":"10.1609\/aaai.v35i2.16208"},{"key":"26_CR29","doi-asserted-by":"crossref","unstructured":"Park, D.Y., Lee, K.H.: Arbitrary style transfer with style-attentional networks. In: CVPR, pp. 5880\u20135888 (2019)","DOI":"10.1109\/CVPR.2019.00603"},{"key":"26_CR30","doi-asserted-by":"crossref","unstructured":"Liu, S., et al.: AdaAttN: revisit attention mechanism in arbitrary neural style transfer. In: ICCV, pp. 6649\u20136658 (2021)","DOI":"10.1109\/ICCV48922.2021.00658"},{"key":"26_CR31","doi-asserted-by":"crossref","unstructured":"Hong, K., Jeon, S., Yang, H., Fu, J., Byun, H.: Domain-aware universal style transfer. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01434"},{"key":"26_CR32","unstructured":"Dosovitskiy, A., et al.: An Image is Worth 16x16 Words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"26_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer Vision \u2013 ECCV 2020","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213\u2013229. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13"},{"key":"26_CR34","unstructured":"Beal, J., Kim, E., Tzeng, E., Park, D.H., Zhai, A., Kislyuk, D.: Toward transformer-based object detection. arXiv preprint arXiv:2012.09958 (2020)"},{"key":"26_CR35","doi-asserted-by":"crossref","unstructured":"Pan, X., Xia, Z., Song, S., Li, L.E., Huang, G.: 3D object detection with pointformer. arXiv preprint arXiv:2012.11409 (2020)","DOI":"10.1109\/CVPR46437.2021.00738"},{"key":"26_CR36","unstructured":"Yuan, Z., Song, X., Bai, L., Zhou, W., Wang, Z., Ouyang, W.: Temporal-Channel Transformer for 3D Lidar-Based Video Object Detection in Autonomous Driving. arXiv preprint arXiv:2011.13628 (2020)"},{"key":"26_CR37","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable DETR: Deformable Transformers for End-to-End Object Detection. arXiv preprint arXiv:2010.04159 (2020)"},{"key":"26_CR38","doi-asserted-by":"crossref","unstructured":"Wang, H., Zhu, Y., Adam, H., Yuille, A., Chen, L.C.: MaX-DeepLab: End-to-End Panoptic Segmentation with Mask Transformers. arXiv preprint arXiv:2012.00759 (2020)","DOI":"10.1109\/CVPR46437.2021.00542"},{"key":"26_CR39","doi-asserted-by":"crossref","unstructured":"Wang, Y., et al.: End-to-End Video Instance Segmentation with Transformers. arXiv preprint arXiv:2011.14503 (2020)","DOI":"10.1109\/CVPR46437.2021.00863"},{"key":"26_CR40","doi-asserted-by":"crossref","unstructured":"Zheng, S., et al.: Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. arXiv preprint arXiv:2012.15840 (2020)","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"26_CR41","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1007\/978-3-030-58595-2_2","volume-title":"Computer Vision \u2013 ECCV 2020","author":"L Huang","year":"2020","unstructured":"Huang, L., Tan, J., Liu, J., Yuan, J.: Hand-transformer: non-autoregressive structured modeling for 3D hand pose estimation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12370, pp. 17\u201333. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58595-2_2"},{"key":"26_CR42","doi-asserted-by":"crossref","unstructured":"Huang, L., Tan, J., Meng, J., Liu, J., Yuan, J.: HOT-net: non-autoregressive transformer for 3D hand-object pose estimation. In: ACM MM, pp. 3136\u20133145 (2020)","DOI":"10.1145\/3394171.3413775"},{"key":"26_CR43","doi-asserted-by":"crossref","unstructured":"Lin, K., Wang, L., Liu, Z.: End-to-End Human Pose and Mesh Reconstruction with Transformer. arXiv preprint arXiv:2012.09760 (2020)","DOI":"10.1109\/CVPR46437.2021.00199"},{"key":"26_CR44","unstructured":"Yang, S., Quan, Z., Nie, M., Yang, W.: TransPose: Towards Explainable Human Pose Estimation by Transformer. arXiv preprint arXiv:2012.14214 (2020)"},{"key":"26_CR45","unstructured":"Chen, M., et al.: Generative pretraining from pixels. In: ICML, pp. 1691\u20131703. PMLR (2020)"},{"key":"26_CR46","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., Wu, J., Child, R., Luan, D., Amodei, D., Sutskever, I.: Language models are unsupervised multitask learners. OpenAI Blog 1, 9 (2019)","journal-title":"OpenAI Blog"},{"key":"26_CR47","unstructured":"Zeng, Y., Yang, H., Chao, H., Wang, J., Fu, J.: Improving visual quality of image synthesis by a token-based generator with transformers. In: NeurIPS (2021)"},{"key":"26_CR48","doi-asserted-by":"crossref","unstructured":"Yang, F., Yang, H., Fu, J., Lu, H., Guo, B.: Learning texture transformer network for image super-resolution. In: CVPR, pp. 5791\u20135800 (2020)","DOI":"10.1109\/CVPR42600.2020.00583"},{"key":"26_CR49","doi-asserted-by":"crossref","unstructured":"Liu, C., Yang, H., Fu, J., Qian, X.: Learning trajectory-aware transformer for video super-resolution. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00560"},{"key":"26_CR50","doi-asserted-by":"crossref","unstructured":"Qiu, Z., Yang, H., Fu, J., Fu, D.: Learning spatiotemporal frequency-transformer for compressed video super-resolution. arXiv preprint arXiv:2208.03012 (2022)","DOI":"10.1007\/978-3-031-19797-0_15"},{"key":"26_CR51","doi-asserted-by":"crossref","unstructured":"Liu, C., Yang, H., Fu, J., Qian, X.: TTVFI: Learning Trajectory-Aware Transformer for Video Frame Interpolation. arXiv preprint arXiv:2207.09048 (2022)","DOI":"10.1109\/CVPR52688.2022.00560"},{"key":"26_CR52","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"26_CR53","doi-asserted-by":"crossref","unstructured":"Ulyanov, D., Vedaldi, A., Lempitsky, V.: Improved texture networks: maximizing quality and diversity in feed-forward stylization and texture synthesis. In: CVPR, pp. 6924\u20136932 (2017)","DOI":"10.1109\/CVPR.2017.437"},{"key":"26_CR54","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"26_CR55","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"26_CR56","unstructured":"Nichol, K.: Painter by numbers, wikiart (2016)"},{"key":"26_CR57","unstructured":"Kingma, D.P., Ba, J.: Adam: A Method for Stochastic Optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"26_CR58","doi-asserted-by":"crossref","unstructured":"Zheng, H., Yang, H., Fu, J., Zha, Z.J., Luo, J.: Learning conditional knowledge distillation for degraded-reference image quality assessment. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01008"},{"key":"26_CR59","first-page":"600","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. TIP 13, 600\u2013612 (2004)","journal-title":"TIP"},{"key":"26_CR60","doi-asserted-by":"crossref","unstructured":"Sheng, L., Lin, Z., Shao, J., Wang, X.: Avatar-net: multi-scale zero-shot style transfer by feature decoration. In: CVPR, pp. 8242\u20138250 (2018)","DOI":"10.1109\/CVPR.2018.00860"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ACCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-26313-2_26","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,1]],"date-time":"2023-03-01T08:11:52Z","timestamp":1677658312000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-26313-2_26"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031263125","9783031263132"],"references-count":60,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-26313-2_26","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"2 March 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asian Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Macao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 December 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 December 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"accv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.accv2022.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT Microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"836","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"277","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.6","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"For the ACCV 2022 workshops 25 papers have been accepted from 40 submissions","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}