{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,21]],"date-time":"2026-03-21T02:20:28Z","timestamp":1774059628274,"version":"3.50.1"},"publisher-location":"Cham","reference-count":57,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198175","type":"print"},{"value":"9783031198182","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19818-2_25","type":"book-chapter","created":{"date-parts":[[2022,10,21]],"date-time":"2022-10-21T16:21:10Z","timestamp":1666369270000},"page":"430-448","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["One-Trimap Video Matting"],"prefix":"10.1007","author":[{"given":"Hongje","family":"Seong","sequence":"first","affiliation":[]},{"given":"Seoung Wug","family":"Oh","sequence":"additional","affiliation":[]},{"given":"Brian","family":"Price","sequence":"additional","affiliation":[]},{"given":"Euntai","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Joon-Young","family":"Lee","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,22]]},"reference":[{"key":"25_CR1","unstructured":"Apostoloff, N., Fitzgibbon, A.: Bayesian video matting using learnt image priors. In: CVPR (2004)"},{"key":"25_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"63","DOI":"10.1007\/978-3-642-24136-9_6","volume-title":"Computer Vision\/Computer Graphics Collaboration Techniques","author":"X Bai","year":"2011","unstructured":"Bai, X., Wang, J., Simons, D.: Towards temporally-coherent video matting. In: Gagalowicz, A., Philips, W. (eds.) MIRAGE 2011. LNCS, vol. 6930, pp. 63\u201374. Springer, Heidelberg (2011). https:\/\/doi.org\/10.1007\/978-3-642-24136-9_6"},{"issue":"9","key":"25_CR3","doi-asserted-by":"publisher","first-page":"2175","DOI":"10.1109\/TPAMI.2013.18","volume":"35","author":"Q Chen","year":"2013","unstructured":"Chen, Q., Li, D., Tang, C.K.: KNN matting. IEEE Trans. Pattern Anal. Mach. Intell. 35(9), 2175\u20132188 (2013)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR4","doi-asserted-by":"crossref","unstructured":"Chen, X., Zou, D., Zhiying Zhou, S., Zhao, Q., Tan, P.: Image matting with local and nonlocal smooth priors. In: CVPR, pp. 1902\u20131907 (2013)","DOI":"10.1109\/CVPR.2013.248"},{"key":"25_CR5","unstructured":"Cheng, H.K., Tai, Y.W., Tang, C.K.: Rethinking space-time networks with improved memory coverage for efficient video object segmentation. In: NeurIPS (2021)"},{"issue":"3","key":"25_CR6","doi-asserted-by":"publisher","first-page":"569","DOI":"10.1109\/TPAMI.2014.2345401","volume":"37","author":"MM Cheng","year":"2014","unstructured":"Cheng, M.M., Mitra, N.J., Huang, X., Torr, P.H., Hu, S.M.: Global contrast based salient region detection. IEEE Trans. Pattern Anal. Mach. Intell. 37(3), 569\u2013582 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"626","DOI":"10.1007\/978-3-319-46475-6_39","volume-title":"Computer Vision \u2013 ECCV 2016","author":"D Cho","year":"2016","unstructured":"Cho, D., Tai, Y.-W., Kweon, I.: Natural image matting using deep convolutional neural networks. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 626\u2013643. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_39"},{"key":"25_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"540","DOI":"10.1007\/978-3-642-33783-3_39","volume-title":"Computer Vision \u2013 ECCV 2012","author":"I Choi","year":"2012","unstructured":"Choi, I., Lee, M., Tai, Y.-W.: video matting using multi-frame nonlocal matting Laplacian. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7577, pp. 540\u2013553. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33783-3_39"},{"key":"25_CR9","doi-asserted-by":"crossref","unstructured":"Chuang, Y.Y., Agarwala, A., Curless, B., Salesin, D., Szeliski, R.: Video matting of complex scenes. In: SIGGRAPH (2002)","DOI":"10.1145\/566570.566572"},{"key":"25_CR10","unstructured":"Chuang, Y.Y., Curless, B., Salesin, D.H., Szeliski, R.: A Bayesian approach to digital matting. In: CVPR. IEEE (2001)"},{"key":"25_CR11","unstructured":"Eisemann, M., Wolf, J., Magnor, M.A.: Spectral video matting. In: VMV, pp. 121\u2013126. Citeseer (2009)"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Erofeev, M., Gitman, Y., Vatolin, D.S., Fedorov, A., Wang, J.: Perceptually motivated benchmark for video matting. In: BMVC, pp. 99\u20131 (2015)","DOI":"10.5244\/C.29.99"},{"issue":"2","key":"25_CR13","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vision 88(2), 303\u2013338 (2010)","journal-title":"Int. J. Comput. Vision"},{"key":"25_CR14","unstructured":"Forte, M., Piti\u00e9, F.: $$ f $$, $$ b $$, alpha matting. arXiv preprint arXiv:2003.07711 (2020)"},{"issue":"2","key":"25_CR15","doi-asserted-by":"publisher","first-page":"575","DOI":"10.1111\/j.1467-8659.2009.01627.x","volume":"29","author":"ESL Gastal","year":"2010","unstructured":"Gastal, E.S.L., Oliveira, M.M.: Shared sampling for real-time alpha matting. Comput. Graph. Forum 29(2), 575\u2013584 (2010)","journal-title":"Comput. Graph. Forum"},{"key":"25_CR16","unstructured":"Gong, M., Wang, L., Yang, R., Yang, Y.H.: Real-time video matting using multichannel Poisson equations. In: Graphics Interface (2010)"},{"key":"25_CR17","unstructured":"Grady, L., Schiwietz, T., Aharon, S., Westermann, R.: Random walks for interactive alpha-matting. In: Proceedings of VIIP, vol. 2005, pp. 423\u2013429 (2005)"},{"key":"25_CR18","doi-asserted-by":"crossref","unstructured":"Hariharan, B., Arbel\u00e1ez, P., Bourdev, L., Maji, S., Malik, J.: Semantic contours from inverse detectors. In: ICCV, pp. 991\u2013998. IEEE (2011)","DOI":"10.1109\/ICCV.2011.6126343"},{"key":"25_CR19","doi-asserted-by":"crossref","unstructured":"He, K., Rhemann, C., Rother, C., Tang, X., Sun, J.: A global sampling method for alpha matting. In: CVPR, pp. 2049\u20132056. IEEE (2011)","DOI":"10.1109\/CVPR.2011.5995495"},{"key":"25_CR20","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"25_CR21","doi-asserted-by":"crossref","unstructured":"Hou, Q., Liu, F.: Context-aware image matting for simultaneous foreground and alpha estimation. In: ICCV, pp. 4130\u20134139 (2019)","DOI":"10.1109\/ICCV.2019.00423"},{"key":"25_CR22","doi-asserted-by":"publisher","first-page":"315","DOI":"10.1145\/1015706.1015721","volume":"23","author":"S Jian","year":"2004","unstructured":"Jian, S., Jia, J., Tang, C.K., Shum, H.Y.: Poisson matting. ACM Trans. Graph. 23, 315\u2013321 (2004)","journal-title":"ACM Trans. Graph."},{"key":"25_CR23","doi-asserted-by":"crossref","unstructured":"Ke, Z., Sun, J., Li, K., Yan, Q., Lau, R.W.: ModNet: real-time trimap-free portrait matting via objective decomposition. In: AAAI (2022)","DOI":"10.1609\/aaai.v36i1.19999"},{"key":"25_CR24","doi-asserted-by":"crossref","unstructured":"Lee, P., Wu, Y.: Nonlocal matting. In: CVPR, pp. 2193\u20132200. IEEE (2011)","DOI":"10.1109\/CVPR.2011.5995665"},{"issue":"2","key":"25_CR25","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1109\/TPAMI.2007.1177","volume":"30","author":"A Levin","year":"2007","unstructured":"Levin, A., Lischinski, D., Weiss, Y.: A closed-form solution to natural image matting. IEEE Trans. Pattern Anal. Mach. Intell. 30(2), 228\u2013242 (2007)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"10","key":"25_CR26","doi-asserted-by":"publisher","first-page":"1699","DOI":"10.1109\/TPAMI.2008.168","volume":"30","author":"A Levin","year":"2008","unstructured":"Levin, A., Rav-Acha, A., Lischinski, D.: Spectral matting. IEEE Trans. Pattern Anal. Mach. Intell. 30(10), 1699\u20131712 (2008)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR27","doi-asserted-by":"crossref","unstructured":"Li, D., Chen, Q., Tang, C.K.: Motion-aware KNN Laplacian for video matting. In: ICCV (2013)","DOI":"10.1109\/ICCV.2013.447"},{"key":"25_CR28","doi-asserted-by":"crossref","unstructured":"Li, Y., Lu, H.: Natural image matting via guided contextual attention. In: AAAI (2020)","DOI":"10.1609\/aaai.v34i07.6809"},{"key":"25_CR29","doi-asserted-by":"crossref","unstructured":"Lin, S., Ryabtsev, A., Sengupta, S., Curless, B.L., Seitz, S.M., Kemelmacher-Shlizerman, I.: Real-time high-resolution background matting. In: CVPR, pp. 8762\u20138771 (2021)","DOI":"10.1109\/CVPR46437.2021.00865"},{"key":"25_CR30","doi-asserted-by":"crossref","unstructured":"Lin, S., Yang, L., Saleemi, I., Sengupta, S.: Robust high-resolution video matting with temporal guidance. In: WACV (2022)","DOI":"10.1109\/WACV51458.2022.00319"},{"key":"25_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"25_CR32","unstructured":"Liu, L., et al.: On the variance of the adaptive learning rate and beyond. In: ICLR, April 2020"},{"key":"25_CR33","doi-asserted-by":"crossref","unstructured":"Lu, H., Dai, Y., Shen, C., Xu, S.: Indices matter: learning to index for deep image matting. In: ICCV, pp. 3266\u20133275 (2019)","DOI":"10.1109\/ICCV.2019.00336"},{"key":"25_CR34","doi-asserted-by":"crossref","unstructured":"Oh, S.W., Lee, J.Y., Xu, N., Kim, S.J.: Video object segmentation using space-time memory networks. In: ICCV, October 2019","DOI":"10.1109\/ICCV.2019.00932"},{"key":"25_CR35","unstructured":"Pont-Tuset, J., Perazzi, F., Caelles, S., Arbel\u00e1ez, P., Sorkine-Hornung, A., Van Gool, L.: The 2017 Davis challenge on video object segmentation. arXiv preprint arXiv:1704.00675 (2017)"},{"key":"25_CR36","doi-asserted-by":"crossref","unstructured":"Porter, T., Duff, T.: Compositing digital images. In: SIGGRAPH, pp. 253\u2013259 (1984)","DOI":"10.1145\/964965.808606"},{"issue":"3","key":"25_CR37","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., et al.: ImageNet large scale visual recognition challenge. Int. J. Comput. Vision 115(3), 211\u2013252 (2015)","journal-title":"Int. J. Comput. Vision"},{"key":"25_CR38","doi-asserted-by":"crossref","unstructured":"Sengupta, S., Jayaram, V., Curless, B., Seitz, S.M., Kemelmacher-Shlizerman, I.: Background matting: the world is your green screen. In: CVPR, pp. 2291\u20132300 (2020)","DOI":"10.1109\/CVPR42600.2020.00236"},{"key":"25_CR39","doi-asserted-by":"crossref","unstructured":"Seong, H., Oh, S.W., Lee, J.Y., Lee, S., Lee, S., Kim, E.: Hierarchical memory matching network for video object segmentation. In: ICCV, pp. 12889\u201312898 (2021)","DOI":"10.1109\/ICCV48922.2021.01265"},{"key":"25_CR40","doi-asserted-by":"crossref","unstructured":"Shahrian, E., Price, B., Cohen, S., Rajan, D.: Temporally consistent and spatially accurate video matting. In: Eurographics (2014)","DOI":"10.1111\/cgf.12297"},{"key":"25_CR41","doi-asserted-by":"crossref","unstructured":"Shahrian, E., Rajan, D., Price, B., Cohen, S.: Improving image matting using comprehensive sampling sets. In: CVPR, pp. 636\u2013643 (2013)","DOI":"10.1109\/CVPR.2013.88"},{"key":"25_CR42","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"92","DOI":"10.1007\/978-3-319-46448-0_6","volume-title":"Computer Vision \u2013 ECCV 2016","author":"X Shen","year":"2016","unstructured":"Shen, X., Tao, X., Gao, H., Zhou, C., Jia, J.: Deep automatic portrait matting. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 92\u2013107. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_6"},{"issue":"4","key":"25_CR43","doi-asserted-by":"publisher","first-page":"717","DOI":"10.1109\/TPAMI.2015.2465960","volume":"38","author":"J Shi","year":"2015","unstructured":"Shi, J., Yan, Q., Xu, L., Jia, J.: Hierarchical image saliency detection on extended CSSD. IEEE Trans. Pattern Anal. Mach. Intell. 38(4), 717\u2013729 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"25_CR44","unstructured":"Sun, J., Ke, Z., Zhang, L., Lu, H., Lau, R.W.: ModNet-V: improving portrait video matting via background restoration. arXiv preprint arXiv:2109.11818 (2021)"},{"key":"25_CR45","doi-asserted-by":"crossref","unstructured":"Sun, Y., Wang, G., Gu, Q., Tang, C.K., Tai, Y.W.: Deep video matting via spatio-temporal alignment and aggregation. In: CVPR, pp. 6975\u20136984 (2021)","DOI":"10.1109\/CVPR46437.2021.00690"},{"key":"25_CR46","doi-asserted-by":"publisher","first-page":"25","DOI":"10.1016\/j.gmod.2010.03.001","volume":"72","author":"S-Y Lee","year":"2010","unstructured":"Lee, S.-Y., Yoon, J.-C., Lee, I.K.: Temporally coherent video matting. Graph. Models 72, 25\u201333 (2010)","journal-title":"Graph. Models"},{"key":"25_CR47","doi-asserted-by":"crossref","unstructured":"Tang, Z., Miao, Z., Wan, Y.: Temporally consistent video matting based on bilayer segmentation. In: ICME (2010)","DOI":"10.1109\/ICME.2010.5583894"},{"key":"25_CR48","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1007\/s00371-011-0598-3","volume":"28","author":"Z Tang","year":"2012","unstructured":"Tang, Z., Miao, Z., Wan, Y., Zhang, D.: Video matting via opacity propagation. Visual Comput. 28, 47\u201351 (2012)","journal-title":"Visual Comput."},{"key":"25_CR49","doi-asserted-by":"crossref","unstructured":"Wang, J., Cohen, M.F.: Optimized color sampling for robust matting. In: CVPR. IEEE (2007)","DOI":"10.1109\/CVPR.2007.383006"},{"key":"25_CR50","doi-asserted-by":"crossref","unstructured":"Wang, X., Chan, K.C., Yu, K., Dong, C., Change Loy, C.: EDVR: video restoration with enhanced deformable convolutional networks. In: CVPRW (2019)","DOI":"10.1109\/CVPRW.2019.00247"},{"key":"25_CR51","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01234-2_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"S Woo","year":"2018","unstructured":"Woo, S., Park, J., Lee, J.-Y., Kweon, I.S.: CBAM: convolutional block attention module. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11211, pp. 3\u201319. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01234-2_1"},{"key":"25_CR52","doi-asserted-by":"crossref","unstructured":"Xu, N., Price, B., Cohen, S., Huang, T.: Deep image matting. In: CVPR, pp. 2970\u20132979 (2017)","DOI":"10.1109\/CVPR.2017.41"},{"key":"25_CR53","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"603","DOI":"10.1007\/978-3-030-01228-1_36","volume-title":"Computer Vision \u2013 ECCV 2018","author":"N Xu","year":"2018","unstructured":"Xu, N., et al.: YouTube-VOS: sequence-to-sequence video object segmentation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11209, pp. 603\u2013619. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01228-1_36"},{"key":"25_CR54","unstructured":"Xu, N., et al.: Youtube-vos: A large-scale video object segmentation benchmark. arXiv preprint arXiv:1809.03327 (2018)"},{"key":"25_CR55","doi-asserted-by":"crossref","unstructured":"Yu, Q., et al.: Mask guided matting via progressive refinement network. In: CVPR, pp. 1154\u20131163 (2021)","DOI":"10.1109\/CVPR46437.2021.00121"},{"key":"25_CR56","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: Attention-guided temporal coherent video object matting. In: ACM MM (2021)","DOI":"10.1145\/3474085.3475623"},{"key":"25_CR57","doi-asserted-by":"crossref","unstructured":"Zhu, B., Chen, Y., Wang, J., Liu, S., Zhang, B., Tang, M.: Fast deep matting for portrait animation on mobile phone. In: ACM MM, pp. 297\u2013305 (2017)","DOI":"10.1145\/3123266.3123286"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19818-2_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T14:32:15Z","timestamp":1710340335000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19818-2_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198175","9783031198182"],"references-count":57,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19818-2_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"22 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}