{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T19:22:49Z","timestamp":1774380169977,"version":"3.50.1"},"publisher-location":"Cham","reference-count":42,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031262838","type":"print"},{"value":"9783031262845","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-26284-5_37","type":"book-chapter","created":{"date-parts":[[2023,2,22]],"date-time":"2023-02-22T08:02:59Z","timestamp":1677052979000},"page":"613-628","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":14,"title":["Robust Human Matting via\u00a0Semantic Guidance"],"prefix":"10.1007","author":[{"given":"Xiangguang","family":"Chen","sequence":"first","affiliation":[]},{"given":"Ye","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Yu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Bingtao","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Ying","family":"Shan","sequence":"additional","affiliation":[]},{"given":"Shan","family":"Liu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,23]]},"reference":[{"key":"37_CR1","doi-asserted-by":"crossref","unstructured":"Lin, S., Ryabtsev, A., Sengupta, S., Curless, B., Seitz, S., Kemelmacher-Shlizerman, I.: Real-time high-resolution background matting. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00865"},{"key":"37_CR2","doi-asserted-by":"crossref","unstructured":"Sengupta, S., Jayaram, V., Curless, B., Seitz, S., Kemelmacher-Shlizerman, I.: Background matting: the world is your green screen. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00236"},{"key":"37_CR3","doi-asserted-by":"crossref","unstructured":"Ke, Z., Sun, J., Li, K., Yan, Q., Lau, R.W.: Modnet: real-time trimap-free portrait matting via objective decomposition. In: AAAI (2022)","DOI":"10.1609\/aaai.v36i1.19999"},{"key":"37_CR4","doi-asserted-by":"crossref","unstructured":"Yu, Q., et al.: Mask guided matting via progressive refinement network. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00121"},{"issue":"2","key":"37_CR5","doi-asserted-by":"publisher","first-page":"246","DOI":"10.1007\/s11263-021-01541-0","volume":"130","author":"J Li","year":"2022","unstructured":"Li, J., Zhang, J., Maybank, S.J., Tao, D.: Bridging composite and real: towards end-to-end deep image matting. Int. J. Comput. Vision 130(2), 246\u2013266 (2022)","journal-title":"Int. J. Comput. Vision"},{"key":"37_CR6","doi-asserted-by":"crossref","unstructured":"Chen, Q., Ge, T., Xu, Y., Zhang, Z., Yang, X., Gai, K.: Semantic human matting. In: ACM MM (2018)","DOI":"10.1145\/3240508.3240610"},{"key":"37_CR7","doi-asserted-by":"crossref","unstructured":"Shen, X., Tao, X., Gao, H., Zhou, C., Jia, J.: Deep automatic portrait matting. In: ECCV (2016)","DOI":"10.1007\/978-3-319-46448-0_6"},{"key":"37_CR8","doi-asserted-by":"crossref","unstructured":"Lin, S., Yang, L., Saleemi, I., Sengupta, S.: Robust high-resolution video matting with temporal guidance. In: WACV (2022)","DOI":"10.1109\/WACV51458.2022.00319"},{"key":"37_CR9","doi-asserted-by":"crossref","unstructured":"Aksoy, Y., Ozan Aydin, T., Pollefeys, M.: Designing effective inter-pixel information flow for natural image matting. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.32"},{"issue":"9","key":"37_CR10","doi-asserted-by":"publisher","first-page":"2175","DOI":"10.1109\/TPAMI.2013.18","volume":"35","author":"Q Chen","year":"2013","unstructured":"Chen, Q., Li, D., Tang, C.K.: KNN matting. TPAMI 35(9), 2175\u20132188 (2013)","journal-title":"TPAMI"},{"key":"37_CR11","unstructured":"Chuang, Y.Y., Curless, B., Salesin, D.H., Szeliski, R.: A bayesian approach to digital matting. In: CVPR (2001)"},{"key":"37_CR12","doi-asserted-by":"crossref","unstructured":"Gastal, E.S., Oliveira, M.M.: Shared sampling for real-time alpha matting. In: Computer Graphics Forum, vol. 29, pp. 575\u2013584 (2010)","DOI":"10.1111\/j.1467-8659.2009.01627.x"},{"issue":"2","key":"37_CR13","doi-asserted-by":"publisher","first-page":"228","DOI":"10.1109\/TPAMI.2007.1177","volume":"30","author":"A Levin","year":"2007","unstructured":"Levin, A., Lischinski, D., Weiss, Y.: A closed-form solution to natural image matting. TPAMI 30(2), 228\u2013242 (2007)","journal-title":"TPAMI"},{"issue":"10","key":"37_CR14","doi-asserted-by":"publisher","first-page":"1699","DOI":"10.1109\/TPAMI.2008.168","volume":"30","author":"A Levin","year":"2008","unstructured":"Levin, A., Rav-Acha, A., Lischinski, D.: Spectral matting. TPAMI 30(10), 1699\u20131712 (2008)","journal-title":"TPAMI"},{"key":"37_CR15","doi-asserted-by":"crossref","unstructured":"Sun, J., Jia, J., Tang, C.K., Shum, H.Y.: Poisson matting. In: ToG, vol. 23, pp. 315\u2013321 (2004)","DOI":"10.1145\/1015706.1015721"},{"key":"37_CR16","unstructured":"Chen, T., Wang, Y., Schillings, V., Meinel, C.: Grayscale image matting and colorization. In: ACCV (2004)"},{"key":"37_CR17","doi-asserted-by":"crossref","unstructured":"Pham, V.Q., Takahashi, K., Naemura, T.: Real-time video matting based on bilayer segmentation. In: ACCV (2009)","DOI":"10.1007\/978-3-642-12304-7_46"},{"key":"37_CR18","doi-asserted-by":"crossref","unstructured":"Park, Y., Yoo, S.I.: A convex image segmentation: Extending graph cuts and closed-form matting. In: ACCV (2010)","DOI":"10.1007\/978-3-642-19318-7_28"},{"key":"37_CR19","doi-asserted-by":"crossref","unstructured":"Sindeev, M., Konushin, A., Rother, C.: Alpha-flow for video matting. In: ACCV (2012)","DOI":"10.1007\/978-3-642-37431-9_34"},{"key":"37_CR20","doi-asserted-by":"crossref","unstructured":"Xu, N., Price, B., Cohen, S., Huang, T.: Deep image matting. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.41"},{"key":"37_CR21","unstructured":"Forte, M., Piti\u00e9, F.: F, B, alpha matting. CoRR abs\/2003.07711 (2020)"},{"key":"37_CR22","doi-asserted-by":"crossref","unstructured":"Liu, Y., et al.: Tripartite information mining and integration for image matting. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00746"},{"key":"37_CR23","unstructured":"Yang, S., Wang, B., Li, W., Lin, Y., He, C., et al.: Unified interactive image matting. arXiv preprint arXiv:2205.08324 (2022)"},{"key":"37_CR24","doi-asserted-by":"crossref","unstructured":"Dai, Y., Price, B., Zhang, H., Shen, C.: Boosting robustness of image matting with context assembling and strong data augmentation. In: CVPR, pp. 11707\u201311716 (2022)","DOI":"10.1109\/CVPR52688.2022.01141"},{"key":"37_CR25","unstructured":"Chen, G., et al.: PP-matting: high-accuracy natural image matting. arXiv preprint arXiv:2204.09433 (2022)"},{"key":"37_CR26","doi-asserted-by":"crossref","unstructured":"Qiao, Y., et al.: Attention-guided hierarchical structure aggregation for image matting. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01369"},{"key":"37_CR27","doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: A late fusion CNN for digital matting. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00765"},{"key":"37_CR28","doi-asserted-by":"crossref","unstructured":"Zhu, B., Chen, Y., Wang, J., Liu, S., Zhang, B., Tang, M.: Fast deep matting for portrait animation on mobile phone. In: ACM MM (2017)","DOI":"10.1145\/3123266.3123286"},{"key":"37_CR29","doi-asserted-by":"crossref","unstructured":"Sun, Y., Tang, C.K., Tai, Y.W.: Human instance matting via mutual guidance and multi-instance refinement. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00267"},{"key":"37_CR30","doi-asserted-by":"crossref","unstructured":"Xing, Y., Li, Y., Wang, X., Zhu, Y., Chen, Q.: Composite photograph harmonization with complete background cues. In: ACM MM (2022)","DOI":"10.1145\/3503161.3548031"},{"key":"37_CR31","doi-asserted-by":"crossref","unstructured":"Li, J., Ma, S., Zhang, J., Tao, D.: Privacy-preserving portrait matting. arXiv (2021)","DOI":"10.1145\/3474085.3475512"},{"key":"37_CR32","unstructured":"Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation. arXiv (2017)"},{"key":"37_CR33","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"37_CR34","unstructured":"Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral normalization for generative adversarial networks. arXiv (2018)"},{"key":"37_CR35","doi-asserted-by":"crossref","unstructured":"Wang, Q., Wu, B., Zhu, P., Li, P., Zuo, W., Hu, Q.: ECA-Net: efficient channel attention for deep convolutional neural networks. In: CVPR. IEEE (2020)","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"37_CR36","unstructured":"supervise.ly: Supervisely person dataset. supervise.ly (2018)"},{"key":"37_CR37","doi-asserted-by":"crossref","unstructured":"Wu, Z., Huang, Y., Yu, Y., Wang, L., Tan, T.: Early hierarchical contexts learned by convolutional networks for image segmentation. In: ICPR. IEEE (2014)","DOI":"10.1109\/ICPR.2014.273"},{"key":"37_CR38","doi-asserted-by":"crossref","unstructured":"Gong, K., Liang, X., Li, Y., Chen, Y., Yang, M., Lin, L.: Instance-level human parsing via part grouping network. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01225-0_47"},{"key":"37_CR39","doi-asserted-by":"crossref","unstructured":"Hou, Q., Liu, F.: Context-aware image matting for simultaneous foreground and alpha estimation. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00423"},{"key":"37_CR40","unstructured":"Li, J., Zhang, J., Maybank, S.J., Tao, D.: End-to-end animal image matting. arXiv (2020)"},{"key":"37_CR41","doi-asserted-by":"crossref","unstructured":"Liu, J., et al.: Boosting semantic human matting with coarse annotations. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00859"},{"key":"37_CR42","doi-asserted-by":"crossref","unstructured":"Rhemann, C., Rother, C., Wang, J., Gelautz, M., Kohli, P., Rott, P.: A perceptually motivated online benchmark for image matting. In: CVPR. IEEE (2009)","DOI":"10.1109\/CVPR.2009.5206503"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ACCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-26284-5_37","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T21:33:38Z","timestamp":1728941618000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-26284-5_37"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031262838","9783031262845"],"references-count":42,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-26284-5_37","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"23 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asian Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Macao","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 December 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 December 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"accv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.accv2022.org","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT Microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"836","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"277","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"33% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.6","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"For the ACCV 2022 workshops 25 papers have been accepted from 40 submissions","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}