{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,5,1]],"date-time":"2026-05-01T17:15:48Z","timestamp":1777655748912,"version":"3.51.4"},"publisher-location":"Cham","reference-count":59,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031263156","type":"print"},{"value":"9783031263163","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-26316-3_29","type":"book-chapter","created":{"date-parts":[[2023,3,1]],"date-time":"2023-03-01T08:02:32Z","timestamp":1677657752000},"page":"485-501","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":14,"title":["BaSSL: Boundary-aware Self-Supervised Learning for\u00a0Video Scene Segmentation"],"prefix":"10.1007","author":[{"given":"Jonghwan","family":"Mun","sequence":"first","affiliation":[]},{"given":"Minchul","family":"Shin","sequence":"additional","affiliation":[]},{"given":"Gunsoo","family":"Han","sequence":"additional","affiliation":[]},{"given":"Sangho","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Seongsu","family":"Ha","sequence":"additional","affiliation":[]},{"given":"Joonseok","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Eun-Sol","family":"Kim","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,3,2]]},"reference":[{"key":"29_CR1","unstructured":"Abu-El-Haija, S., et al.: YouTube-8M: a large-scale video classification benchmark. arXiv:1609.08675 (2016)"},{"key":"29_CR2","unstructured":"Ahsan, U., Sun, C., Essa, I.: DiscrimNet: semi-supervised action recognition from videos using generative adversarial networks. arXiv:1801.07230 (2018)"},{"key":"29_CR3","doi-asserted-by":"crossref","unstructured":"Baraldi, L., Grana, C., Cucchiara, R.: A deep Siamese network for scene detection in broadcast videos. In: ACM MM (2015)","DOI":"10.1145\/2733373.2806316"},{"key":"29_CR4","unstructured":"Berndt, D.J., Clifford, J.: Using dynamic time warping to find patterns in time series. In: SIGKDD Workshop (1994)"},{"key":"29_CR5","unstructured":"Caron, M., Misra, I., Mairal, J., Goyal, P., Bojanowski, P., Joulin, A.: Unsupervised learning of visual features by contrasting cluster assignments. arXiv:2006.09882 (2020)"},{"key":"29_CR6","doi-asserted-by":"crossref","unstructured":"Chang, C.Y., Huang, D.A., Sui, Y., Fei-Fei, L., Niebles, J.C.: D3TW: discriminative differentiable dynamic time warping for weakly supervised action alignment and segmentation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00366"},{"issue":"1","key":"29_CR7","doi-asserted-by":"publisher","first-page":"89","DOI":"10.1109\/TMM.2008.2008924","volume":"11","author":"VT Chasanis","year":"2008","unstructured":"Chasanis, V.T., Likas, A.C., Galatsanos, N.P.: Scene detection in videos using shot clustering and sequence alignment. IEEE Trans. Multimedia 11(1), 89\u2013100 (2008)","journal-title":"IEEE Trans. Multimedia"},{"key":"29_CR8","doi-asserted-by":"crossref","unstructured":"Chen, S., Nie, X., Fan, D., Zhang, D., Bhat, V., Hamid, R.: Shot contrastive self-supervised learning for scene boundary detection. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00967"},{"key":"29_CR9","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: a simple framework for contrastive learning of visual representations. In: ICML (2020)"},{"key":"29_CR10","doi-asserted-by":"crossref","unstructured":"Chen, Y.C., Li, L., Yu, L., El Kholy, A., Ahmed, F., Gan, Z., Cheng, Y., Liu, J.: UNITER: UNiversal Image-TExt representation learning. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58577-8_7"},{"key":"29_CR11","doi-asserted-by":"crossref","unstructured":"Dave, I., Gupta, R., Rizve, M.N., Shah, M.: TCLR: temporal contrastive learning for video representation. arXiv:2101.07974 (2021)","DOI":"10.1016\/j.cviu.2022.103406"},{"key":"29_CR12","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., Li, L.J., Li, K., Fei-Fei, L.: ImageNet: a large-scale hierarchical image database. In: CVPR (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"29_CR13","doi-asserted-by":"crossref","unstructured":"Farha, Y.A., Gall, J.: MS-TCN: multi-stage temporal convolutional network for action segmentation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00369"},{"key":"29_CR14","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Xiong, B., Girshick, R., He, K.: A large-scale study on unsupervised spatiotemporal representation learning. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00331"},{"key":"29_CR15","doi-asserted-by":"crossref","unstructured":"Fried, D., Alayrac, J.B., Blunsom, P., Dyer, C., Clark, S., Nematzadeh, A.: Learning to segment actions from observation and narration. arXiv:2005.03684 (2020)","DOI":"10.18653\/v1\/2020.acl-main.231"},{"key":"29_CR16","unstructured":"Han, B., Wu, W.: Video scene segmentation using a novel boundary evaluation criterion and dynamic programming. In: IEEE International Conference on Multimedia and Expo (2011)"},{"key":"29_CR17","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"29_CR18","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"29_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"709","DOI":"10.1007\/978-3-030-58548-8_41","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Q Huang","year":"2020","unstructured":"Huang, Q., Xiong, Yu., Rao, A., Wang, J., Lin, D.: MovieNet: a holistic dataset for movie understanding. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 709\u2013727. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_41"},{"key":"29_CR20","unstructured":"Jing, L., Tian, Y.: Self-supervised spatiotemporal feature learning by video geometric transformations. arXiv preprint arXiv:1811.11387 (2018)"},{"issue":"4","key":"29_CR21","doi-asserted-by":"publisher","first-page":"765","DOI":"10.1109\/TPAMI.2018.2884469","volume":"42","author":"H Kuehne","year":"2018","unstructured":"Kuehne, H., Richard, A., Gall, J.: A hybrid RNN-HMM approach for weakly supervised temporal action segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 42(4), 765\u2013779 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"29_CR22","doi-asserted-by":"crossref","unstructured":"Kukleva, A., Kuehne, H., Sener, F., Gall, J.: Unsupervised learning of action classes with continuous temporal embedding. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01234"},{"key":"29_CR23","doi-asserted-by":"crossref","unstructured":"Kumar, S., Haresh, S., Ahmed, A., Konin, A., Zia, M.Z., Tran, Q.H.: Unsupervised activity segmentation by joint representation learning and online clustering. arXiv:2105.13353 (2021)","DOI":"10.1109\/CVPR52688.2022.01954"},{"key":"29_CR24","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1007\/978-3-319-46487-9_3","volume-title":"Computer Vision \u2013 ECCV 2016","author":"C Lea","year":"2016","unstructured":"Lea, C., Reiter, A., Vidal, R., Hager, G.D.: Segmental spatiotemporal CNNs for fine-grained action segmentation. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9907, pp. 36\u201352. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46487-9_3"},{"key":"29_CR25","doi-asserted-by":"crossref","unstructured":"Lee, H.Y., Huang, J.B., Singh, M., Yang, M.H.: Unsupervised Representation Learning by Sorting Sequences. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.79"},{"key":"29_CR26","doi-asserted-by":"crossref","unstructured":"Li, J., Lei, P., Todorovic, S.: Weakly Supervised Energy-based Learning for Action Segmentation. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00634"},{"key":"29_CR27","doi-asserted-by":"crossref","unstructured":"Li, J., Todorovic, S.: Set-constrained Viterbi for set-supervised action segmentation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01083"},{"key":"29_CR28","doi-asserted-by":"crossref","unstructured":"Li, J., Todorovic, S.: Action shuffle alternating learning for unsupervised action segmentation. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01244"},{"key":"29_CR29","unstructured":"Li, L., et al.: VALUE: a multi-task benchmark for video-and-language understanding evaluation. In: NeurIPS (2021)"},{"key":"29_CR30","doi-asserted-by":"crossref","unstructured":"Liang, C., Zhang, Y., Cheng, J., Xu, C., Lu, H.: A novel role-based movie scene segmentation method. In: Pacific-Rim Conference on Multimedia (2009)","DOI":"10.1007\/978-3-642-10467-1_82"},{"key":"29_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"527","DOI":"10.1007\/978-3-319-46448-0_32","volume-title":"Computer Vision \u2013 ECCV 2016","author":"I Misra","year":"2016","unstructured":"Misra, I., Zitnick, C.L., Hebert, M.: Shuffle and learn: unsupervised learning using temporal order verification. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 527\u2013544. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_32"},{"key":"29_CR32","unstructured":"Oord, A.v.d., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding. arXiv:1807.03748 (2018)"},{"key":"29_CR33","doi-asserted-by":"crossref","unstructured":"Qian, R., et al.: Spatiotemporal contrastive video representation learning. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00689"},{"key":"29_CR34","doi-asserted-by":"crossref","unstructured":"Rao, A., et al.: A local-to-global approach to multi-modal movie scene segmentation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01016"},{"key":"29_CR35","unstructured":"Rasheed, Z., Shah, M.: Scene detection in Hollywood movies and TV shows. In: CVPR (2003)"},{"issue":"6","key":"29_CR36","doi-asserted-by":"publisher","first-page":"1097","DOI":"10.1109\/TMM.2005.858392","volume":"7","author":"Z Rasheed","year":"2005","unstructured":"Rasheed, Z., Shah, M.: Detection and representation of scenes in videos. IEEE Trans. Multimedia 7(6), 1097\u20131105 (2005)","journal-title":"IEEE Trans. Multimedia"},{"key":"29_CR37","doi-asserted-by":"crossref","unstructured":"Roh, B., Shin, W., Kim, I., Kim, S.: Spatially consistent representation learning. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00120"},{"key":"29_CR38","doi-asserted-by":"crossref","unstructured":"Rotman, D., Porat, D., Ashour, G.: Robust and efficient video scene detection using optimal sequential grouping. In: IEEE International Symposium on Multimedia (ISM) (2016)","DOI":"10.1109\/ISM.2016.0061"},{"issue":"02","key":"29_CR39","doi-asserted-by":"publisher","first-page":"193","DOI":"10.1142\/S1793351X17400086","volume":"11","author":"D Rotman","year":"2017","unstructured":"Rotman, D., Porat, D., Ashour, G.: Optimal sequential grouping for robust video scene detection using multiple modalities. Int. J. Seman. Comput. 11(02), 193\u2013208 (2017)","journal-title":"Int. J. Seman. Comput."},{"key":"29_CR40","unstructured":"Rui, Y., Huang, T.S., Mehrotra, S.: Exploring video structure beyond the shots. In: Proceedings of the IEEE International Conference on Multimedia Computing and Systems (1998)"},{"key":"29_CR41","doi-asserted-by":"crossref","unstructured":"Shen, Y., Wang, L., Elhamifar, E.: Learning to segment actions from visual and language instructions via differentiable weak sequence alignment. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01002"},{"key":"29_CR42","doi-asserted-by":"crossref","unstructured":"Shou, M.Z., Lei, S.W., Wang, W., Ghadiyaram, D., Feiszli, M.: Generic event boundary detection: a benchmark for event segmentation. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00797"},{"issue":"8","key":"29_CR43","doi-asserted-by":"publisher","first-page":"1163","DOI":"10.1109\/TCSVT.2011.2138830","volume":"21","author":"P Sidiropoulos","year":"2011","unstructured":"Sidiropoulos, P., Mezaris, V., Kompatsiaris, I., Meinedo, H., Bugalho, M., Trancoso, I.: Temporal Video Segmentation to Scenes using High-level Audiovisual Features. IEEE Trans. Circuits Syst. Video Technol. 21(8), 1163\u20131177 (2011)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"29_CR44","doi-asserted-by":"crossref","first-page":"6196","DOI":"10.1109\/TPAMI.2021.3089127","volume":"44","author":"Y Souri","year":"2021","unstructured":"Souri, Y., Fayyaz, M., Minciullo, L., Francesca, G., Gall, J.: Fast weakly supervised action segmentation using mutual consistency. IEEE Trans. Pattern Anal. Mach. Intell. 44, 6196\u20136208 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"29_CR45","unstructured":"Srivastava, N., Mansimov, E., Salakhudinov, R.: Unsupervised learning of video representations using LSTMs. In: ICML (2015)"},{"key":"29_CR46","unstructured":"Sun, C., Baradel, F., Murphy, K., Schmid, C.: Learning video representations using contrastive bidirectional transformer. arXiv:1906.05743 (2019)"},{"key":"29_CR47","doi-asserted-by":"crossref","unstructured":"Sun, C., Myers, A., Vondrick, C., Murphy, K., Schmid, C.: VideoBERT: a joint model for video and language representation learning. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00756"},{"key":"29_CR48","doi-asserted-by":"crossref","unstructured":"Tapaswi, M., Bauml, M., Stiefelhagen, R.: StoryGraphs: visualizing character interactions as a timeline. In: CVPR (2014)","DOI":"10.1109\/CVPR.2014.111"},{"key":"29_CR49","doi-asserted-by":"crossref","unstructured":"Tversky, B., Zacks, J.M.: Event perception. Oxford Handbook of Cognitive Psychology (2013)","DOI":"10.1093\/oxfordhb\/9780195376746.013.0006"},{"key":"29_CR50","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NIPS (2017)"},{"key":"29_CR51","doi-asserted-by":"crossref","unstructured":"VidalMata, R.G., Scheirer, W.J., Kukleva, A., Cox, D., Kuehne, H.: Joint visual-temporal embedding for unsupervised learning of actions in untrimmed sequences. In: WACV (2021)","DOI":"10.1109\/WACV48630.2021.00128"},{"key":"29_CR52","unstructured":"Vondrick, C., Pirsiavash, H., Torralba, A.: Generating videos with scene dynamics. NIPS (2016)"},{"key":"29_CR53","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1007\/978-3-030-01261-8_24","volume-title":"Computer Vision \u2013 ECCV 2018","author":"C Vondrick","year":"2018","unstructured":"Vondrick, C., Shrivastava, A., Fathi, A., Guadarrama, S., Murphy, K.: Tracking emerges by colorizing videos. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11217, pp. 402\u2013419. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01261-8_24"},{"key":"29_CR54","unstructured":"Wang, Z., et al.: Unsupervised action segmentation with self-supervised feature learning and co-occurrence parsing. arXiv:2105.14158 (2021)"},{"key":"29_CR55","doi-asserted-by":"crossref","unstructured":"Xu, D., Xiao, J., Zhao, Z., Shao, J., Xie, D., Zhuang, Y.: Self-supervised spatiotemporal learning via video clip order prediction. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.01058"},{"key":"29_CR56","doi-asserted-by":"crossref","unstructured":"Xu, M., et al.: Boundary-sensitive pre-training for temporal localization in videos. arXiv:2011.10830 (2020)","DOI":"10.1109\/ICCV48922.2021.00713"},{"key":"29_CR57","unstructured":"Zhang, B., et al.: A hierarchical multi-modal encoder for moment localization in video corpus. arXiv:2011.09046 (2020)"},{"key":"29_CR58","doi-asserted-by":"publisher","first-page":"1452","DOI":"10.1109\/TPAMI.2017.2723009","volume":"40","author":"B Zhou","year":"2017","unstructured":"Zhou, B., Lapedriza, A., Khosla, A., Oliva, A., Torralba, A.: Places: a 10 million image database for scene recognition. IEEE Trans. Pattern Anal. Mach. Intell. 40, 1452\u20131464 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"29_CR59","doi-asserted-by":"crossref","unstructured":"Zhukov, D., Alayrac, J.B., Cinbis, R.G., Fouhey, D., Laptev, I., Sivic, J.: Cross-task weakly supervised learning from instructional videos. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00365"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ACCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-26316-3_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,8]],"date-time":"2023-12-08T01:15:57Z","timestamp":1701998157000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-26316-3_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031263156","9783031263163"],"references-count":59,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-26316-3_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"2 March 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asian Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Macao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 December 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 December 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"accv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.accv2022.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT Microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"836","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"277","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.6","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"For the ACCV 2022 workshops 25 papers have been accepted from 40 submissions","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}