{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T17:43:34Z","timestamp":1772905414841,"version":"3.50.1"},"publisher-location":"Cham","reference-count":78,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031198113","type":"print"},{"value":"9783031198120","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19812-0_42","type":"book-chapter","created":{"date-parts":[[2022,10,29]],"date-time":"2022-10-29T14:03:42Z","timestamp":1667052222000},"page":"729-747","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":37,"title":["Panoptic-PartFormer: Learning a\u00a0Unified Model for\u00a0Panoptic Part Segmentation"],"prefix":"10.1007","author":[{"given":"Xiangtai","family":"Li","sequence":"first","affiliation":[]},{"given":"Shilin","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Yibo","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Guangliang","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Yunhai","family":"Tong","sequence":"additional","affiliation":[]},{"given":"Dacheng","family":"Tao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,30]]},"reference":[{"key":"42_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer Vision","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213\u2013229. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13"},{"key":"42_CR2","unstructured":"Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation. arXiv:1706.05587 (2017)"},{"key":"42_CR3","doi-asserted-by":"crossref","unstructured":"Chen, L.C., Zhu, Y., Papandreou, G., Schroff, F., Adam, H.: Encoder-decoder with atrous separable convolution for semantic image segmentation. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01234-2_49"},{"key":"42_CR4","doi-asserted-by":"crossref","unstructured":"Chen, Y., et al.: Banet: bidirectional aggregation network with occlusion handling for panoptic segmentation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00385"},{"key":"42_CR5","doi-asserted-by":"crossref","unstructured":"Cheng, B., et al.: Panoptic-deeplab: a simple, strong, and fast baseline for bottom-up panoptic segmentation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01249"},{"key":"42_CR6","unstructured":"Cheng, B., Schwing, A.G., Kirillov, A.: Per-pixel classification is not all you need for semantic segmentation. In: NeurIPS (2021)"},{"key":"42_CR7","doi-asserted-by":"crossref","unstructured":"Chollet, F.: Xception: Deep learning with depthwise separable convolutions. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.195"},{"key":"42_CR8","doi-asserted-by":"crossref","unstructured":"Cordts, M., et al.: The cityscapes dataset for semantic urban scene understanding. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.350"},{"key":"42_CR9","unstructured":"Dosovitskiy, A., et al.: An image is worth 16 x 16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"issue":"2","key":"42_CR10","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. IJCV 88(2), 303\u2013338 (2010)","journal-title":"IJCV"},{"key":"42_CR11","doi-asserted-by":"crossref","unstructured":"Fang, H.S., et al.: Weakly and semi supervised human body part parsing via pose-guided knowledge transfer. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00015"},{"key":"42_CR12","unstructured":"Fang, Y., et al.: Instances as queries. arXiv preprint arXiv:2105.01928 (2021)"},{"key":"42_CR13","doi-asserted-by":"crossref","unstructured":"Gao, N., et al.: SSAP: single-shot instance segmentation with affinity pyramid. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00073"},{"issue":"8","key":"42_CR14","first-page":"4291","volume":"44","author":"Q Geng","year":"2021","unstructured":"Geng, Q., et al.: Part-level car parsing and reconstruction in single street view images. IEEE Trans. Pattern Anal. Mach. Intell. 44(8), 4291\u20134305 (2021)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"42_CR15","doi-asserted-by":"crossref","unstructured":"de Geus, D., Meletis, P., Lu, C., Wen, X., Dubbelman, G.: Part-aware panoptic segmentation. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00544"},{"key":"42_CR16","unstructured":"Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 249\u2013256. JMLR Workshop and Conference Proceedings (2010)"},{"key":"42_CR17","doi-asserted-by":"crossref","unstructured":"Gong, K., Liang, X., Li, Y., Chen, Y., Yang, M., Lin, L.: Instance-level human parsing via part grouping network. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01225-0_47"},{"key":"42_CR18","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., Girshick, R.: Mask R-CNN. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.322"},{"key":"42_CR19","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"42_CR20","doi-asserted-by":"crossref","unstructured":"Hou, R., et al.: Real-time panoptic segmentation from dense detections. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00855"},{"key":"42_CR21","doi-asserted-by":"crossref","unstructured":"Ji, R., et al.: Learning semantic neural tree for human parsing. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58601-0_13"},{"key":"42_CR22","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Girshick, R., He, K., Doll\u00e1r, P.: Panoptic feature pyramid networks. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00656"},{"key":"42_CR23","doi-asserted-by":"crossref","unstructured":"Kirillov, A., He, K., Girshick, R., Rother, C., Doll\u00e1r, P.: Panoptic segmentation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00963"},{"key":"42_CR24","unstructured":"Li, J., et al.: Multiple-human parsing in the wild. arXiv preprint arXiv:1705.07206 (2017)"},{"key":"42_CR25","unstructured":"Li, J., Raventos, A., Bhargava, A., Tagawa, T., Gaidon, A.: Learning to fuse things and stuff. arXiv:1812.01192 (2018)"},{"key":"42_CR26","doi-asserted-by":"crossref","unstructured":"Li, Q., Arnab, A., Torr, P.H.: Holistic, instance-level human parsing. arXiv preprint arXiv:1709.03612 (2017)","DOI":"10.5244\/C.31.25"},{"key":"42_CR27","doi-asserted-by":"crossref","unstructured":"Li, Q., Qi, X., Torr, P.H.: Unifying training and inference for panoptic segmentation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.01333"},{"key":"42_CR28","doi-asserted-by":"crossref","unstructured":"Li, X., et al.: Semantic flow for fast and accurate scene parsing. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58452-8_45"},{"key":"42_CR29","doi-asserted-by":"crossref","unstructured":"Li, Y., et al.: Attention-guided unified network for panoptic segmentation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00719"},{"key":"42_CR30","doi-asserted-by":"crossref","unstructured":"Li, Y., et al.: Fully convolutional networks for panoptic segmentation with point-based supervision. arXiv preprint arXiv:2108.07682 (2021)","DOI":"10.1109\/TPAMI.2022.3200416"},{"key":"42_CR31","doi-asserted-by":"crossref","unstructured":"Li, Y., et al.: Fully convolutional networks for panoptic segmentation. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00028"},{"key":"42_CR32","doi-asserted-by":"crossref","unstructured":"Liang, J., Homayounfar, N., Ma, W.C., Xiong, Y., Hu, R., Urtasun, R.: PolyTransform: deep polygon transformer for instance segmentation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00915"},{"key":"42_CR33","doi-asserted-by":"crossref","unstructured":"Liang, X., et al.: Human parsing with contextualized convolutional neural network. In: ICCV (2015)","DOI":"10.1109\/ICCV.2015.163"},{"key":"42_CR34","doi-asserted-by":"crossref","unstructured":"Lin, J., Yang, H., Chen, D., Zeng, M., Wen, F., Yuan, L.: Face Parsing with RoI Tanh-Warping. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00580"},{"key":"42_CR35","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R.B., He, K., Hariharan, B., Belongie, S.J.: Feature pyramid networks for object detection. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"42_CR36","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., Doll\u00e1r, P.: Focal loss for dense object detection. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.324"},{"key":"42_CR37","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., et al.: Microsoft coco: Common objects in context. In: ECCV (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"42_CR38","doi-asserted-by":"crossref","unstructured":"Liu, S., et al.: Cross-domain human parsing via adversarial feature and label adaptation. In: AAAI (2018)","DOI":"10.1609\/aaai.v32i1.12320"},{"key":"42_CR39","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin transformer: Hierarchical vision transformer using shifted windows. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"42_CR40","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization (2017)"},{"key":"42_CR41","doi-asserted-by":"crossref","unstructured":"Meinhardt, T., Kirillov, A., Leal-Taixe, L., Feichtenhofer, C.: TrackFormer: multi-object tracking with transformers. arXiv preprint arXiv:2101.02702 (2021)","DOI":"10.1109\/CVPR52688.2022.00864"},{"key":"42_CR42","doi-asserted-by":"crossref","unstructured":"Michieli, U., Borsato, E., Rossi, L., Zanuttigh, P.: GMNet: graph matching network for large scale part semantic segmentation in the wild. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58598-3_24"},{"key":"42_CR43","doi-asserted-by":"crossref","unstructured":"Milletari, F., Navab, N., Ahmadi, S.: V-Net: fully convolutional neural networks for volumetric medical image segmentation. In: 3DV (2016)","DOI":"10.1109\/3DV.2016.79"},{"key":"42_CR44","doi-asserted-by":"crossref","unstructured":"Milletari, F., Navab, N., Ahmadi, S.A.: V-net: fully convolutional neural networks for volumetric medical image segmentation. In: 3DV (2016)","DOI":"10.1109\/3DV.2016.79"},{"issue":"5","key":"42_CR45","doi-asserted-by":"publisher","first-page":"1551","DOI":"10.1007\/s11263-021-01445-z","volume":"129","author":"R Mohan","year":"2021","unstructured":"Mohan, R., Valada, A.: EfficientPS: efficient panoptic segmentation. Int. J. Comput. Vis. 129(5), 1551\u20131579 (2021)","journal-title":"Int. J. Comput. Vis."},{"key":"42_CR46","doi-asserted-by":"crossref","unstructured":"Neuhold, G., Ollmann, T., Rota Bulo, S., Kontschieder, P.: The mapillary vistas dataset for semantic understanding of street scenes. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.534"},{"key":"42_CR47","doi-asserted-by":"crossref","unstructured":"Porzi, L., Bulo, S.R., Colovic, A., Kontschieder, P.: Seamless scene segmentation. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00847"},{"key":"42_CR48","doi-asserted-by":"crossref","unstructured":"Qi, S., Wang, W., Jia, B., Shen, J., Zhu, S.C.: Learning human-object interactions by graph parsing neural networks. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01240-3_25"},{"key":"42_CR49","doi-asserted-by":"crossref","unstructured":"Qiao, S., Chen, L.C., Yuille, A.: Detectors: Detecting objects with recursive feature pyramid and switchable atrous convolution. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01008"},{"key":"42_CR50","doi-asserted-by":"crossref","unstructured":"Ruan, T., Liu, T., Huang, Z., Wei, Y., Wei, S., Zhao, Y.: Devil in the details: Towards accurate single and multiple human parsing. In: AAAI (2019)","DOI":"10.1609\/aaai.v33i01.33014814"},{"key":"42_CR51","doi-asserted-by":"crossref","unstructured":"Shen, Z., et al.: Human-aware motion deblurring. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00567"},{"key":"42_CR52","doi-asserted-by":"crossref","unstructured":"Sun, P., et al.: Sparse R-CNN: end-to-end object detection with learnable proposals. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01422"},{"key":"42_CR53","unstructured":"Tan, M., Le, Q.: EfficientNet: rethinking model scaling for convolutional neural networks. In: ICML, pp. 6105\u20136114. PMLR (2019)"},{"key":"42_CR54","doi-asserted-by":"crossref","unstructured":"Tian, Z., Shen, C., Chen, H.: Conditional convolutions for instance segmentation. arXiv preprint arXiv:2003.05664 (2020)","DOI":"10.1007\/978-3-030-58452-8_17"},{"key":"42_CR55","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: ICML, PMLR (2021)"},{"key":"42_CR56","unstructured":"Vaswani, A., et al.: Attention is all you need. arXiv preprint arXiv:1706.03762 (2017)"},{"key":"42_CR57","doi-asserted-by":"crossref","unstructured":"Wang, H., Zhu, Y., Adam, H., Yuille, A., Chen, L.C.: MaX-DeepLab: end-to-end panoptic segmentation with mask transformers. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00542"},{"key":"42_CR58","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1007\/978-3-030-58548-8_7","volume-title":"Computer Vision","author":"H Wang","year":"2020","unstructured":"Wang, H., Zhu, Y., Green, B., Adam, H., Yuille, A., Chen, L.-C.: Axial-DeepLab: stand-alone axial-attention for panoptic segmentation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12349, pp. 108\u2013126. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58548-8_7"},{"key":"42_CR59","unstructured":"Wang, J., et al.: Deep high-resolution representation learning for visual recognition. In: PAMI (2020)"},{"key":"42_CR60","doi-asserted-by":"crossref","unstructured":"Wang, W., Zhang, Z., Qi, S., Shen, J., Pang, Y., Shao, L.: Learning compositional neural information fusion for human parsing. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00580"},{"key":"42_CR61","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"649","DOI":"10.1007\/978-3-030-58523-5_38","volume-title":"Computer Vision","author":"X Wang","year":"2020","unstructured":"Wang, X., Kong, T., Shen, C., Jiang, Y., Li, L.: SOLO: segmenting objects by locations. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12363, pp. 649\u2013665. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58523-5_38"},{"key":"42_CR62","unstructured":"Wang, X., Zhang, R., Kong, T., Li, L., Shen, C.: SOLOv2: dynamic and fast instance segmentation. In: NeurIPS (2020)"},{"key":"42_CR63","unstructured":"Wu, Y., Zhang, G., Xu, H., Liang, X., Lin, L.: Auto-panoptic: Cooperative multi-component architecture search for panoptic segmentation. In: NIPS (2020)"},{"key":"42_CR64","doi-asserted-by":"crossref","unstructured":"Xiong, Y., et al.: UPSNet: a unified panoptic segmentation network. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00902"},{"key":"42_CR65","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1007\/978-3-030-58610-2_25","volume-title":"Computer Vision","author":"L Yang","year":"2020","unstructured":"Yang, L., et al.: Renovating parsing R-CNN for accurate multiple human parsing. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12357, pp. 421\u2013437. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58610-2_25"},{"key":"42_CR66","doi-asserted-by":"crossref","unstructured":"Yang, L., Song, Q., Wang, Z., Jiang, M.: Parsing R-CNN for instance-level human analysis. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00045"},{"key":"42_CR67","unstructured":"Yang, T.J., et al.: DeeperLab: single-shot image parser. arXiv:1902.05093 (2019)"},{"key":"42_CR68","doi-asserted-by":"crossref","unstructured":"Yang, Y., Li, H., Li, X., Zhao, Q., Wu, J., Lin, Z.: Sognet: Scene overlap graph network for panoptic segmentation. In: AAAI (2020)","DOI":"10.1609\/aaai.v34i07.6955"},{"key":"42_CR69","doi-asserted-by":"crossref","unstructured":"Yu, F., et al.: Bdd100k: a diverse driving dataset for heterogeneous multitask learning. In: CVPR, pp. 2636\u20132645 (2020)","DOI":"10.1109\/CVPR42600.2020.00271"},{"key":"42_CR70","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"173","DOI":"10.1007\/978-3-030-58539-6_11","volume-title":"Computer Vision","author":"Y Yuan","year":"2020","unstructured":"Yuan, Y., Chen, X., Wang, J.: Object-contextual representations for semantic segmentation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12351, pp. 173\u2013190. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58539-6_11"},{"key":"42_CR71","unstructured":"Zhang, H., et al.: ResNeSt: split-attention networks. arXiv preprint arXiv:2004.08955 (2020)"},{"key":"42_CR72","unstructured":"Zhang, W., Pang, J., Chen, K., Loy, C.C.: K-net: towards unified image segmentation. In: NeurIPS (2021)"},{"key":"42_CR73","doi-asserted-by":"crossref","unstructured":"Zhao, H., Shi, J., Qi, X., Wang, X., Jia, J.: Pyramid scene parsing network. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.660"},{"key":"42_CR74","doi-asserted-by":"crossref","unstructured":"Zhao, J., Li, J., Cheng, Y., Sim, T., Yan, S., Feng, J.: Understanding humans in crowded scenes: deep nested adversarial learning and a new benchmark for multi-human parsing. In: MM (2018)","DOI":"10.1145\/3240508.3240509"},{"key":"42_CR75","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Li, J., Zhang, Y., Tian, Y.: Multi-class part parsing with joint boundary-semantic awareness. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00927"},{"key":"42_CR76","doi-asserted-by":"crossref","unstructured":"Zhou, T., Wang, W., Liu, S., Yang, Y., Van Gool, L.: Differentiable multi-granularity human representation learning for instance-aware human semantic parsing. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00167"},{"key":"42_CR77","doi-asserted-by":"crossref","unstructured":"Zhu, X., Hu, H., Lin, S., Dai, J.: Deformable convnets v2: More deformable, better results. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00953"},{"key":"42_CR78","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable DETR: deformable transformers for end-to-end object detection. In: ICLR (2020)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19812-0_42","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,4,27]],"date-time":"2023-04-27T13:10:46Z","timestamp":1682601046000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19812-0_42"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198113","9783031198120"],"references-count":78,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19812-0_42","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"30 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}