{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,17]],"date-time":"2026-02-17T14:20:38Z","timestamp":1771338038160,"version":"3.50.1"},"publisher-location":"Cham","reference-count":68,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031200618","type":"print"},{"value":"9783031200625","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20062-5_3","type":"book-chapter","created":{"date-parts":[[2022,11,10]],"date-time":"2022-11-10T10:31:55Z","timestamp":1668076315000},"page":"37-54","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":34,"title":["MeshMAE: Masked Autoencoders for\u00a03D Mesh Data Analysis"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-2188-4290","authenticated-orcid":false,"given":"Yaqian","family":"Liang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0682-8645","authenticated-orcid":false,"given":"Shanshan","family":"Zhao","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0761-7893","authenticated-orcid":false,"given":"Baosheng","family":"Yu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6595-7661","authenticated-orcid":false,"given":"Jing","family":"Zhang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7016-3698","authenticated-orcid":false,"given":"Fazhi","family":"He","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,11]]},"reference":[{"key":"3_CR1","unstructured":"Adobe.com: animate 3d characters for games, film, and more. Accessed 24 2021. https:\/\/www.mixamo.com (2021)"},{"key":"3_CR2","doi-asserted-by":"crossref","unstructured":"Agrawal, P., Carreira, J., Malik, J.: Learning to see by moving. In: ICCV, pp. 37\u201345 (2015)","DOI":"10.1109\/ICCV.2015.13"},{"key":"3_CR3","doi-asserted-by":"crossref","unstructured":"Anguelov, D., Srinivasan, P., Koller, D., Thrun, S., Rodgers, J., Davis, J.: Scape: shape completion and animation of people. In: ACM Transactions on Graphics (TOG), pp. 408\u2013416 (2005)","DOI":"10.1145\/1073204.1073207"},{"key":"3_CR4","doi-asserted-by":"crossref","unstructured":"Aumentado-Armstrong, T., Tsogkas, S., Jepson, A., Dickinson, S.: Geometric disentanglement for generative latent shape models. In: CVPR, pp. 8181\u20138190 (2019)","DOI":"10.1109\/ICCV.2019.00827"},{"key":"3_CR5","unstructured":"Bao, H., Dong, L., Wei, F.: Beit: bert pre-training of image transformers. arXiv preprint arXiv:2106.08254 (2021)"},{"key":"3_CR6","doi-asserted-by":"crossref","unstructured":"Bogo, F., Romero, J., Loper, M., Black, M.J.: Faust: dataset and evaluation for 3d mesh registration. In: CVPR, pp. 3794\u20133801 (2014)","DOI":"10.1109\/CVPR.2014.491"},{"key":"3_CR7","first-page":"1877","volume":"33","author":"T Brown","year":"2020","unstructured":"Brown, T., et al.: Language models are few-shot learners. NeurIPS 33, 1877\u20131901 (2020)","journal-title":"NeurIPS"},{"key":"3_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-58452-8_13","volume-title":"Computer Vision","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-end object detection with transformers. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 213\u2013229. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_13"},{"key":"3_CR9","doi-asserted-by":"crossref","unstructured":"Chen, C.F., Fan, Q., Panda, R.: CrossViT: cross-attention multi-scale vision transformer for image classification. arXiv preprint arXiv:2103.14899 (2021)","DOI":"10.1109\/ICCV48922.2021.00041"},{"key":"3_CR10","unstructured":"Chen, M., et al.: Generative pretraining from pixels. In: ICML, pp. 1691\u20131703. PMLR (2020)"},{"key":"3_CR11","unstructured":"Cheng, B., Schwing, A., Kirillov, A.: Per-pixel classification is not all you need for semantic segmentation. In: NeurIPS, pp. 17864\u201317875 (2021)"},{"key":"3_CR12","unstructured":"Conneau, A., Lample, G.: Cross-lingual language model pretraining. In: NeurIPS, pp. 7059\u20137069 (2019)"},{"key":"3_CR13","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/978-3-030-58580-8_2","volume-title":"Computer Vision","author":"L Cosmo","year":"2020","unstructured":"Cosmo, L., Norelli, A., Halimi, O., Kimmel, R., Rodol\u00e0, E.: LIMP: learning latent shape representations with metric preservation\u00a0priors. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12348, pp. 19\u201335. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58580-8_2"},{"key":"3_CR14","doi-asserted-by":"crossref","unstructured":"Dai, Z., Cai, B., Lin, Y., Chen, J.: UP-DETR: unsupervised pre-training for object detection with transformers. In: CVPR, pp. 1601\u20131610 (2021)","DOI":"10.1109\/CVPR46437.2021.00165"},{"issue":"5","key":"3_CR15","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1111\/cgf.12693","volume":"34","author":"B Davide","year":"2015","unstructured":"Davide, B., Jonathan, M., Simone, M., Michael, M.B., Umberto, C., Pierre, V.: Learning class-specific descriptors for deformable shapes using localized spectral convolutional networks. Comput. Graph. Forum 34(5), 13\u201323 (2015)","journal-title":"Comput. Graph. Forum"},{"key":"3_CR16","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"3_CR17","doi-asserted-by":"crossref","unstructured":"Ding, L., et al.: Looking outside the window: wide-context transformer for the semantic segmentation of high-resolution remote sensing images. arXiv preprint arXiv:2106.15754 (2021)","DOI":"10.1109\/TGRS.2022.3168697"},{"key":"3_CR18","unstructured":"Dosovitskiy, A., et al.: An image is worth 16 x 16 words: transformers for image recognition at scale. In: ICLR (2020)"},{"key":"3_CR19","doi-asserted-by":"crossref","unstructured":"Feng, Y., Feng, Y., You, H., Zhao, X., Gao, Y.: MeshNet: mesh neural network for 3d shape representation. In: AAAI, pp. 8279\u20138286 (2019)","DOI":"10.1609\/aaai.v33i01.33018279"},{"key":"3_CR20","unstructured":"Gidaris, S., Singh, P., Komodakis, N.: Unsupervised representation learning by predicting image rotations. In: ICLR (2018)"},{"key":"3_CR21","unstructured":"Giorgi, D., Biasotti, S., Paraboschi, L.: Shape retrieval contest 2007: watertight models track. SHREC Compet. 8(7) (2007)"},{"key":"3_CR22","doi-asserted-by":"crossref","unstructured":"Guan, S., Xu, J., Wang, Y., Ni, B., Yang, X.: Bilevel online adaptation for out-of-domain human mesh reconstruction. In: CVPR, pp. 10472\u201310481 (2021)","DOI":"10.1109\/CVPR46437.2021.01033"},{"issue":"2","key":"3_CR23","doi-asserted-by":"publisher","first-page":"187","DOI":"10.1007\/s41095-021-0229-5","volume":"7","author":"MH Guo","year":"2021","unstructured":"Guo, M.H., Cai, J.X., Liu, Z.N., Mu, T.J., Martin, R.R., Hu, S.M.: PCT: point cloud transformer. Comput. Visual Media 7(2), 187\u2013199 (2021)","journal-title":"Comput. Visual Media"},{"key":"3_CR24","doi-asserted-by":"crossref","unstructured":"Haim, N., Segol, N., Ben-Hamu, H., Maron, H., Lipman, Y.: Surface networks via general covers. In: ICCV, pp. 632\u2013641 (2019)","DOI":"10.1109\/ICCV.2019.00072"},{"issue":"4","key":"3_CR25","doi-asserted-by":"publisher","first-page":"90","DOI":"10.1145\/3306346.3322959","volume":"38","author":"R Hanocka","year":"2019","unstructured":"Hanocka, R., Hertz, A., Fish, N., Giryes, R., Fleishman, S., Cohenor, D.: MeshCNN: a network with an edge. ACM Trans. Graph. (TOG) 38(4), 90 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"3_CR26","doi-asserted-by":"crossref","unstructured":"He, K., Chen, X., Xie, S., Li, Y., Doll\u00e1r, P., Girshick, R.: Masked autoencoders are scalable vision learners. In: CVPR, pp. 16000\u201316009 (2022)","DOI":"10.1109\/CVPR52688.2022.01553"},{"key":"3_CR27","unstructured":"Hjelm, R.D., et al.: Learning deep representations by mutual information estimation and maximization. In: ICLR (2018)"},{"issue":"12","key":"3_CR28","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11432-020-3097-4","volume":"63","author":"S-M Hu","year":"2020","unstructured":"Hu, S.-M., Liang, D., Yang, G.-Y., Yang, G.-W., Zhou, W.-Y.: Jittor: a novel deep learning framework with meta-operators and unified graph execution. Sci. China Inf. Sci. 63(12), 1\u201321 (2020). https:\/\/doi.org\/10.1007\/s11432-020-3097-4","journal-title":"Sci. China Inf. Sci."},{"key":"3_CR29","unstructured":"Hu, S.M., et al.: Subdivision-based mesh convolution networks. ACM Trans. Graph. (TOG) (2021)"},{"key":"3_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"17","DOI":"10.1007\/978-3-030-58595-2_2","volume-title":"Computer Vision","author":"L Huang","year":"2020","unstructured":"Huang, L., Tan, J., Liu, J., Yuan, J.: Hand-transformer: non-autoregressive structured modeling for 3d hand pose estimation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12370, pp. 17\u201333. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58595-2_2"},{"key":"3_CR31","doi-asserted-by":"publisher","first-page":"64","DOI":"10.1162\/tacl_a_00300","volume":"8","author":"M Joshi","year":"2020","unstructured":"Joshi, M., Chen, D., Liu, Y., Weld, D.S., Zettlemoyer, L., Levy, O.: SpanBERT: improving pre-training by representing and predicting spans. Trans. Assoc. Comput. Linguist. 8, 64\u201377 (2020)","journal-title":"Trans. Assoc. Comput. Linguist."},{"issue":"6","key":"3_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3414685.3417806","volume":"39","author":"A Lahav","year":"2020","unstructured":"Lahav, A., Tal, A.: MeshWalker: deep mesh understanding by random walks. ACM Trans. Graph. (TOG) 39(6), 1\u201313 (2020)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"3_CR33","doi-asserted-by":"crossref","unstructured":"Lee, A.W., Sweldens, W., Schr\u00f6der, P., Cowsar, L., Dobkin, D.: Maps: multiresolution adaptive parameterization of surfaces. In: ACM SIGGRAPH, pp. 95\u2013104 (1998)","DOI":"10.1145\/280814.280828"},{"key":"3_CR34","doi-asserted-by":"crossref","unstructured":"Li, W., Liu, H., Tang, H., Wang, P., Van Gool, L.: MHFormer: multi-hypothesis transformer for 3d human pose estimation. arXiv preprint arXiv:2111.12707 (2021)","DOI":"10.1109\/CVPR52688.2022.01280"},{"key":"3_CR35","doi-asserted-by":"crossref","unstructured":"Li, Z., Liu, X., Drenkow, N., Ding, A., Creighton, F.X., Taylor, R.H., Unberath, M.: Revisiting stereo depth estimation from a sequence-to-sequence perspective with transformers. In: CVPR, pp. 6197\u20136206 (2021)","DOI":"10.1109\/ICCV48922.2021.00614"},{"key":"3_CR36","doi-asserted-by":"crossref","unstructured":"Lin, K., Wang, L., Liu, Z.: End-to-end human pose and mesh reconstruction with transformers. In: CVPR, pp. 1954\u20131963 (2021)","DOI":"10.1109\/CVPR46437.2021.00199"},{"key":"3_CR37","doi-asserted-by":"crossref","unstructured":"Lin, K., Wang, L., Liu, Z.: Mesh graphormer. arXiv preprint arXiv:2104.00272 (2021)","DOI":"10.1109\/ICCV48922.2021.01270"},{"key":"3_CR38","doi-asserted-by":"crossref","unstructured":"Liu, H.T.D., Kim, V.G., Chaudhuri, S., Aigerman, N., Jacobson, A.: Neural subdivision. arXiv preprint arXiv:2005.01819 (2020)","DOI":"10.1145\/3386569.3392418"},{"key":"3_CR39","unstructured":"Liu, Y., et al.: Roberta: a robustly optimized bert pretraining approach. arXiv preprint arXiv:1907.11692 (2019)"},{"issue":"4","key":"3_CR40","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1145\/3072959.3073616","volume":"36","author":"H Maron","year":"2017","unstructured":"Maron, H., et al.: Convolutional neural networks on surfaces via seamless toric covers. ACM Trans. Graph. (TOG) 36(4), 71\u20131 (2017)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"3_CR41","first-page":"952","volume":"33","author":"F Milano","year":"2020","unstructured":"Milano, F., Loquercio, A., Rosinol, A., Scaramuzza, D., Carlone, L.: Primal-dual mesh convolutional neural networks. NeurIPS 33, 952\u2013963 (2020)","journal-title":"NeurIPS"},{"key":"3_CR42","doi-asserted-by":"crossref","unstructured":"Misra, I., Girdhar, R., Joulin, A.: An end-to-end transformer model for 3d object detection. In: CVPR, pp. 2906\u20132917 (2021)","DOI":"10.1109\/ICCV48922.2021.00290"},{"key":"3_CR43","unstructured":"Monti, F., Shchur, O., Bojchevski, A., Litany, O., G\u00fcnnemann, S., Bronstein, M.M.: Dual-primal graph convolutional networks. arXiv preprint arXiv:1806.00770 (2018)"},{"key":"3_CR44","unstructured":"Nash, C., Ganin, Y., Eslami, S.A., Battaglia, P.: PolyGen: an autoregressive generative model of 3d meshes. In: PMLR, pp. 7220\u20137229. PMLR (2020)"},{"key":"3_CR45","doi-asserted-by":"crossref","unstructured":"Pathak, D., Girshick, R., Doll\u00e1r, P., Darrell, T., Hariharan, B.: Learning features by watching objects move. In: CVPR, pp. 2701\u20132710 (2017)","DOI":"10.1109\/CVPR.2017.638"},{"key":"3_CR46","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)"},{"issue":"8","key":"3_CR47","first-page":"9","volume":"1","author":"A Radford","year":"2019","unstructured":"Radford, A., et al.: Language models are unsupervised multitask learners. OpenAI blog 1(8), 9 (2019)","journal-title":"OpenAI blog"},{"key":"3_CR48","unstructured":"Rolfe, J.T.: Discrete variational autoencoders. arXiv preprint arXiv:1609.02200 (2016)"},{"key":"3_CR49","doi-asserted-by":"crossref","unstructured":"Saleh, M., Wu, S.C., Cosmo, L., Navab, N., Busam, B., Tombari, F.: Bending graphs: Hierarchical shape matching using gated optimal transport. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01146"},{"key":"3_CR50","doi-asserted-by":"crossref","unstructured":"Sederberg, T.W., Parry, S.R.: Free-form deformation of solid geometric models. In: ACM SIGGRAPH Computer Graphics, pp. 151\u2013160 (1986)","DOI":"10.1145\/15886.15903"},{"key":"3_CR51","doi-asserted-by":"crossref","unstructured":"Tianyu, L., Yali, W., Junhao, Z., Zhe, W., Zhipeng, Z., Yu, Q.: PC-HMR: pose calibration for 3d human mesh recovery from 2d images\/videos. In: AAAI, pp. 2269\u20132276. AAAI Press (2021)","DOI":"10.1609\/aaai.v35i3.16326"},{"key":"3_CR52","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: ICLR, pp. 10347\u201310357 (2021)"},{"key":"3_CR53","unstructured":"Trappolini, G., Cosmo, L., Moschella, L., Marin, R., Melzi, S., Rodol\u00e0, E.: Shape registration in the time of transformers. In: NeurIPS, pp. 5731\u20135744 (2021)"},{"key":"3_CR54","unstructured":"Trinh, T.H., Luong, M.T., Le, Q.V.: Selfie: self-supervised pretraining for image embedding. arXiv preprint arXiv:1906.02940 (2019)"},{"key":"3_CR55","unstructured":"Vaswani, A., et al.: Attention is all you need. In: NeurIPS, pp. 5998\u20136008 (2017)"},{"key":"3_CR56","doi-asserted-by":"crossref","unstructured":"Verma, N., Boyer, E., Verbeek, J.: FeaStNet: feature-steered graph convolutions for 3d shape analysis. In: CVPR, pp. 2598\u20132606 (2018)","DOI":"10.1109\/CVPR.2018.00275"},{"key":"3_CR57","doi-asserted-by":"crossref","unstructured":"Vlasic, D., Baran, I., Matusik, W., Popovi\u0107, J.: Articulated mesh animation from multi-view silhouettes. In: ACM SIGGRAPH, pp. 1\u20139 (2008)","DOI":"10.1145\/1360612.1360696"},{"key":"3_CR58","doi-asserted-by":"crossref","unstructured":"Wang, H., Zhu, Y., Adam, H., Yuille, A., Chen, L.C.: MaX-DeepLab: end-to-end panoptic segmentation with mask transformers. In: CVPR, pp. 5463\u20135474 (2021)","DOI":"10.1109\/CVPR46437.2021.00542"},{"key":"3_CR59","doi-asserted-by":"crossref","unstructured":"Wang, X., Gupta, A.: Unsupervised learning of visual representations using videos. In: ICCV, pp. 2794\u20132802 (2015)","DOI":"10.1109\/ICCV.2015.320"},{"issue":"6","key":"3_CR60","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2366145.2366184","volume":"31","author":"Y Wang","year":"2012","unstructured":"Wang, Y., Asafi, S., Van Kaick, O., Zhang, H., Cohen-Or, D., Chen, B.: Active co-analysis of a set of shapes. ACM Trans. Graph. (TOG) 31(6), 1\u201310 (2012)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"3_CR61","unstructured":"Xu, Y., Zhang, Q., Zhang, J., Tao, D.: Vitae: vision transformer advanced by exploring intrinsic inductive bias. In: NeurIPS, pp. 28522\u201328535 (2021)"},{"key":"3_CR62","doi-asserted-by":"crossref","unstructured":"Yang, G., Tang, H., Ding, M., Sebe, N., Ricci, E.: Transformer-based attention networks for continuous pixel-wise prediction. In: CVPR, pp. 16269\u201316279 (2021)","DOI":"10.1109\/ICCV48922.2021.01596"},{"key":"3_CR63","doi-asserted-by":"crossref","unstructured":"Yu, X., Rao, Y., Wang, Z., Liu, Z., Lu, J., Zhou, J.: PoinTr: diverse point cloud completion with geometry-aware transformers. In: ICCV, pp. 12498\u201312507 (2021)","DOI":"10.1109\/ICCV48922.2021.01227"},{"key":"3_CR64","doi-asserted-by":"crossref","unstructured":"Yu, X., Tang, L., Rao, Y., Huang, T., Zhou, J., Lu, J.: Point-BERT: pre-training 3d point cloud transformers with masked point modeling. arXiv preprint arXiv:2111.14819 (2021)","DOI":"10.1109\/CVPR52688.2022.01871"},{"key":"3_CR65","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Xu, Y., Zhang, J., Tao, D.: Vitaev2: vision transformer advanced by exploring inductive bias for image recognition and beyond. arXiv preprint arXiv:2202.10108 (2022)","DOI":"10.1007\/s11263-022-01739-w"},{"key":"3_CR66","doi-asserted-by":"crossref","unstructured":"Zhao, H., Jiang, L., Jia, J., Torr, P.H., Koltun, V.: Point transformer. In: ICCV, pp. 16259\u201316268 (2021)","DOI":"10.1109\/ICCV48922.2021.01595"},{"key":"3_CR67","doi-asserted-by":"crossref","unstructured":"Zheng, S., et al.: Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: CVPR, pp. 6881\u20136890 (2021)","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"3_CR68","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., Dai, J.: Deformable detr: deformable transformers for end-to-end object detection. arXiv preprint arXiv:2010.04159 (2020)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20062-5_3","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,3,11]],"date-time":"2023-03-11T18:42:35Z","timestamp":1678560155000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20062-5_3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031200618","9783031200625"],"references-count":68,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20062-5_3","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"11 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}