{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,4]],"date-time":"2026-03-04T17:15:21Z","timestamp":1772644521497,"version":"3.50.1"},"publisher-location":"Cham","reference-count":66,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031200649","type":"print"},{"value":"9783031200656","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20065-6_21","type":"book-chapter","created":{"date-parts":[[2022,11,2]],"date-time":"2022-11-02T20:24:03Z","timestamp":1667420643000},"page":"357-374","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":46,"title":["MetaGait: Learning to\u00a0Learn an\u00a0Omni Sample Adaptive Representation for\u00a0Gait Recognition"],"prefix":"10.1007","author":[{"given":"Huanzhang","family":"Dou","sequence":"first","affiliation":[]},{"given":"Pengyi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Su","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0294-2099","authenticated-orcid":false,"given":"Yunlong","family":"Yu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3023-1662","authenticated-orcid":false,"given":"Xi","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,3]]},"reference":[{"key":"21_CR1","unstructured":"Antoniou, A., Edwards, H., Storkey, A.: How to train your MAML. arXiv preprint arXiv:1810.09502 (2018)"},{"key":"21_CR2","doi-asserted-by":"crossref","unstructured":"Ariyanto, G., Nixon, M.S.: Model-based 3D gait biometrics. In: International Joint Conference on Biometrics, pp. 1\u20137 (2011)","DOI":"10.1109\/IJCB.2011.6117582"},{"key":"21_CR3","doi-asserted-by":"publisher","first-page":"129","DOI":"10.1049\/iet-bmt.2015.0072","volume":"6","author":"M Balazia","year":"2017","unstructured":"Balazia, M., Plataniotis, K.N.: Human gait recognition from motion capture data in signature poses. IET Biom. 6, 129\u2013137 (2017)","journal-title":"IET Biom."},{"key":"21_CR4","doi-asserted-by":"crossref","unstructured":"Bashir, K., Xiang, T., Gong, S.: Gait recognition using gait entropy image. In: IET International Conference on Imaging for Crime Detection and Prevention, pp. 1\u20136 (2009)","DOI":"10.1049\/ic.2009.0230"},{"key":"21_CR5","unstructured":"Bengio, E., Bacon, P.L., Pineau, J., Precup, D.: Conditional computation in neural networks for faster models. arXiv preprint arXiv:1511.06297 (2015)"},{"issue":"8","key":"21_CR6","doi-asserted-by":"publisher","first-page":"1194","DOI":"10.1016\/j.imavis.2008.11.008","volume":"27","author":"R Bodor","year":"2009","unstructured":"Bodor, R., Drenner, A., Fehr, D., Masoud, O., Papanikolopoulos, N.: View-independent human motion classification using image-based reconstruction. Image Vis. Comput. 27(8), 1194\u20131206 (2009)","journal-title":"Image Vis. Comput."},{"key":"21_CR7","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-319-68533-5_1","volume-title":"Surveillance in Action","author":"I Bouchrika","year":"2018","unstructured":"Bouchrika, I.: A survey of using biometrics for smart visual surveillance: gait recognition. In: Karampelas, P., Bourlai, T. (eds.) Surveillance in Action, pp. 3\u201323. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-68533-5_1"},{"key":"21_CR8","doi-asserted-by":"crossref","unstructured":"Boulgouris, N.V., Chi, Z.X.: Gait recognition based on human body components. In: IEEE International Conference on Image Processing, pp. 353\u2013356 (2007)","DOI":"10.1109\/ICIP.2007.4378964"},{"key":"21_CR9","doi-asserted-by":"crossref","unstructured":"Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multi-person 2D pose estimation using part affinity fields. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 7291\u20137299 (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"21_CR10","doi-asserted-by":"crossref","unstructured":"Chao, H., He, Y., Zhang, J., Feng, J.: GaitSet: regarding gait as a set for cross-view gait recognition. In: AAAI (2019)","DOI":"10.1609\/aaai.v33i01.33018126"},{"issue":"11","key":"21_CR11","doi-asserted-by":"publisher","first-page":"1843","DOI":"10.1109\/TIFS.2014.2352114","volume":"9","author":"P Chattopadhyay","year":"2014","unstructured":"Chattopadhyay, P., Sural, S., Mukherjee, J.: Frontal gait recognition from incomplete sequences using RGB-D camera. IEEE Trans. Inf. Forensics Secur. 9(11), 1843\u20131856 (2014)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"21_CR12","doi-asserted-by":"crossref","unstructured":"Chen, Y., Dai, X., Liu, M., Chen, D., Yuan, L., Liu, Z.: Dynamic convolution: attention over convolution kernels. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 11030\u201311039 (2020)","DOI":"10.1109\/CVPR42600.2020.01104"},{"key":"21_CR13","unstructured":"Cheng, H.P., et al.: Swiftnet: using graph propagation as meta-knowledge to search highly representative neural architectures. arXiv preprint arXiv:1906.08305 (2019)"},{"key":"21_CR14","doi-asserted-by":"crossref","unstructured":"Dai, J., et al.: Deformable convolutional networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 764\u2013773 (2017)","DOI":"10.1109\/ICCV.2017.89"},{"key":"21_CR15","unstructured":"Denil, M., Shakibi, B., Dinh, L., Ranzato, M., De Freitas, N.: Predicting parameters in deep learning. In: Advances in Neural Information Processing Systems (2013)"},{"key":"21_CR16","unstructured":"Devos, A., Chatel, S., Grossglauser, M.: Reproducing meta-learning with differentiable closed-form solvers. In: International Conference on Learning Representations (2019)"},{"key":"21_CR17","unstructured":"Dou, H., et al.: Versatilegait: a large-scale synthetic gait dataset with fine-grainedattributes and complicated scenarios. arXiv preprint arXiv:2101.01394 (2021)"},{"key":"21_CR18","unstructured":"Du, Y., Wang, W., Wang, L.: Hierarchical recurrent neural network for skeleton based action recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 1110\u20131118 (2015)"},{"key":"21_CR19","unstructured":"Eigen, D., Ranzato, M., Sutskever, I.: Learning factored representations in a deep mixture of experts. arXiv preprint arXiv:1312.4314 (2013)"},{"key":"21_CR20","doi-asserted-by":"crossref","unstructured":"Fan, C., et al.: Gaitpart: temporal part-based model for gait recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.01423"},{"key":"21_CR21","unstructured":"Finn, C., Rajeswaran, A., Kakade, S., Levine, S.: Online meta-learning. In: International Conference on Machine Learning, pp. 1920\u20131930 (2019)"},{"key":"21_CR22","unstructured":"Gao, H., Zhu, X., Lin, S., Dai, J.: Deformable kernels: adapting effective receptive fields for object deformation. In: International Conference on Learning Representations (2019)"},{"issue":"4","key":"21_CR23","doi-asserted-by":"publisher","first-page":"997","DOI":"10.1109\/TSMCB.2009.2031091","volume":"40","author":"M Goffredo","year":"2009","unstructured":"Goffredo, M., Bouchrika, I., Carter, J.N., Nixon, M.S.: Self-calibrating view-invariant gait biometrics. IEEE Trans. Cybern. 40(4), 997\u20131008 (2009)","journal-title":"IEEE Trans. Cybern."},{"key":"21_CR24","unstructured":"Zhao, G., Liu, G., Li, H., Pietikainen, M.: 3D gait recognition using multiple cameras. In: International Conference on Automatic Face Gesture Recognition, pp. 529\u2013534 (2006)"},{"issue":"2","key":"21_CR25","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1109\/TPAMI.2006.38","volume":"28","author":"J Han","year":"2006","unstructured":"Han, J., Bhanu, B.: Individual recognition using gait energy image. IEEE Trans. Pattern Anal. Mach. Intell. 28(2), 316\u2013322 (2006)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"21_CR26","doi-asserted-by":"crossref","unstructured":"Harley, A.W., Derpanis, K.G., Kokkinos, I.: Segmentation-aware convolutional networks using local attention masks. In: International Conference on Computer Vision, pp. 5038\u20135047 (2017)","DOI":"10.1109\/ICCV.2017.539"},{"key":"21_CR27","unstructured":"He, B., Yang, X., Wu, Z., Chen, H., Lim, S.N., Shrivastava, A.: GTA: global temporal attention for video action understanding. arXiv preprint arXiv:2012.08510 (2020)"},{"key":"21_CR28","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"21_CR29","unstructured":"Hermans, A., Beyer, L., Leibe, B.: In defense of the triplet loss for person re-identification. arXiv preprint arXiv:1703.07737 (2017)"},{"key":"21_CR30","doi-asserted-by":"crossref","unstructured":"Hospedales, T., Antoniou, A., Micaelli, P., Storkey, A.: Meta-learning in neural networks: a survey. arXiv preprint arXiv:2004.05439 (2020)","DOI":"10.1109\/TPAMI.2021.3079209"},{"key":"21_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"382","DOI":"10.1007\/978-3-030-58545-7_22","volume-title":"Computer Vision \u2013 ECCV 2020","author":"S Hou","year":"2020","unstructured":"Hou, S., Cao, C., Liu, X., Huang, Y.: Gait lateral network: learning discriminative and compact representations for gait recognition. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12354, pp. 382\u2013398. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58545-7_22"},{"key":"21_CR32","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"21_CR33","doi-asserted-by":"crossref","unstructured":"Huang, X., et al.: Context-sensitive temporal feature learning for gait recognition. In: International Conference on Computer Vision, pp. 12909\u201312918, October 2021","DOI":"10.1109\/ICCV48922.2021.01267"},{"key":"21_CR34","doi-asserted-by":"crossref","unstructured":"Huang, Y., Zhang, J., Zhao, H., Zhang, L.: Attention-based network for cross-view gait recognition. In: Advances in Neural Information Processing Systems, pp. 489\u2013498 (2018)","DOI":"10.1007\/978-3-030-04239-4_44"},{"key":"21_CR35","doi-asserted-by":"crossref","unstructured":"Huang, Y., et al.: Curricularface: adaptive curriculum learning loss for deep face recognition. In: IEEE Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00594"},{"key":"21_CR36","doi-asserted-by":"crossref","unstructured":"Huang, Z., et al.: 3D local convolutional neural networks for gait recognition. In: International Conference on Computer Vision, pp. 14920\u201314929, October 2021","DOI":"10.1109\/ICCV48922.2021.01465"},{"issue":"1","key":"21_CR37","doi-asserted-by":"publisher","first-page":"79","DOI":"10.1162\/neco.1991.3.1.79","volume":"3","author":"RA Jacobs","year":"1991","unstructured":"Jacobs, R.A., Jordan, M.I., Nowlan, S.J., Hinton, G.E.: Adaptive mixtures of local experts. Neural Comput. 3(1), 79\u201387 (1991)","journal-title":"Neural Comput."},{"issue":"6","key":"21_CR38","doi-asserted-by":"publisher","first-page":"063019","DOI":"10.1117\/1.JEI.25.6.063019","volume":"25","author":"D Kastaniotis","year":"2016","unstructured":"Kastaniotis, D., Theodorakopoulos, I., Fotopoulos, S.: Pose-based gait recognition with local gradient descriptors and hierarchically aggregated residuals. J. Electron. Imaging 25(6), 063019 (2016)","journal-title":"J. Electron. Imaging"},{"issue":"2","key":"21_CR39","doi-asserted-by":"publisher","first-page":"696","DOI":"10.1109\/TIP.2013.2294552","volume":"23","author":"W Kusakunniran","year":"2014","unstructured":"Kusakunniran, W., Wu, Q., Zhang, J., Li, H., Wang, L.: Recognizing gaits across views through correlated motion co-clustering. IEEE Trans. Image Process. 23(2), 696\u2013709 (2014)","journal-title":"IEEE Trans. Image Process."},{"issue":"10","key":"21_CR40","doi-asserted-by":"publisher","first-page":"1642","DOI":"10.1109\/TIFS.2013.2252342","volume":"8","author":"W Kusakunniran","year":"2013","unstructured":"Kusakunniran, W., Wu, Q., Zhang, J., Ma, Y., Li, H.: A new view-invariant feature for cross-view gait recognition. IEEE Trans. Inf. Forensics Secur. 8(10), 1642\u20131653 (2013)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"issue":"9","key":"21_CR41","doi-asserted-by":"publisher","first-page":"2361","DOI":"10.1109\/TMM.2019.2900134","volume":"21","author":"S Li","year":"2019","unstructured":"Li, S., Liu, W., Ma, H.: Attentive spatial-temporal summary networks for feature learning in irregular gait recognition. IEEE Trans. Multimedia 21(9), 2361\u20132375 (2019)","journal-title":"IEEE Trans. Multimedia"},{"key":"21_CR42","doi-asserted-by":"publisher","first-page":"107069","DOI":"10.1016\/j.patcog.2019.107069","volume":"98","author":"R Liao","year":"2020","unstructured":"Liao, R., Yu, S., An, W., Huang, Y.: A model-based gait recognition method with body pose and human prior knowledge. Pattern Recog. 98, 107069 (2020)","journal-title":"Pattern Recog."},{"key":"21_CR43","doi-asserted-by":"crossref","unstructured":"Lin, B., Zhang, S., Bao, F.: Gait recognition with multiple-temporal-scale 3D convolutional neural network. In: ACM International Conference on Multimedia, pp. 3054\u20133062 (2020)","DOI":"10.1145\/3394171.3413861"},{"key":"21_CR44","doi-asserted-by":"crossref","unstructured":"Lin, B., Zhang, S., Yu, X.: Gait recognition via effective global-local feature representation and local temporal aggregation. In: International Conference on Computer Vision, pp. 14648\u201314656, October 2021","DOI":"10.1109\/ICCV48922.2021.01438"},{"key":"21_CR45","unstructured":"Lin, J., Rao, Y., Lu, J., Zhou, J.: Runtime neural pruning. In: Advances in Neural Information Processing Systems (2017)"},{"key":"21_CR46","doi-asserted-by":"crossref","unstructured":"Lin, P., Sun, P., Cheng, G., Xie, S., Li, X., Shi, J.: Graph-guided architecture search for real-time semantic segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition, June 2020","DOI":"10.1109\/CVPR42600.2020.00426"},{"key":"21_CR47","doi-asserted-by":"crossref","unstructured":"Ma, J., Zhao, Z., Yi, X., Chen, J., Hong, L., Chi, E.H.: Modeling task relationships in multi-task learning with multi-gate mixture-of-experts. In: SIGKDD, pp. 1930\u20131939 (2018)","DOI":"10.1145\/3219819.3220007"},{"key":"21_CR48","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11) (2008)"},{"issue":"5","key":"21_CR49","doi-asserted-by":"publisher","first-page":"1294","DOI":"10.1111\/1556-4029.14036","volume":"64","author":"I Macoveciuc","year":"2019","unstructured":"Macoveciuc, I., Rando, C.J., Borrion, H.: Forensic gait analysis and recognition: standards of evidence admissibility. J. Forensic Sci. 64(5), 1294\u20131303 (2019)","journal-title":"J. Forensic Sci."},{"key":"21_CR50","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"151","DOI":"10.1007\/11744078_12","volume-title":"Computer Vision \u2013 ECCV 2006","author":"Y Makihara","year":"2006","unstructured":"Makihara, Y., Sagawa, R., Mukaigawa, Y., Echigo, T., Yagi, Y.: Gait recognition using a view transformation model in the frequency domain. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3953, pp. 151\u2013163. Springer, Heidelberg (2006). https:\/\/doi.org\/10.1007\/11744078_12"},{"issue":"7","key":"21_CR51","doi-asserted-by":"publisher","first-page":"1655","DOI":"10.1109\/TPAMI.2018.2846566","volume":"41","author":"F Radenovi\u0107","year":"2018","unstructured":"Radenovi\u0107, F., Tolias, G., Chum, O.: Fine-tuning CNN image retrieval with no human annotation. IEEE Trans. Pattern Anal. Mach. Intell. 41(7), 1655\u20131668 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"21_CR52","doi-asserted-by":"crossref","unstructured":"Samangooei, S., Nixon, M.S.: Performing content-based retrieval of humans using gait biometrics. Multimed. Tools Appl. 49, 195\u2013212 (2010)","DOI":"10.1007\/s11042-009-0391-8"},{"key":"21_CR53","doi-asserted-by":"crossref","unstructured":"Sepas-Moghaddam, A., Etemad, A.: Deep gait recognition: a survey. arXiv preprint arXiv:2102.09546 (2021)","DOI":"10.1109\/TPAMI.2022.3151865"},{"key":"21_CR54","unstructured":"Shan, S., Li, Y., Oliva, J.B.: Meta-neighborhoods. In: Advances in Neural Information Processing Systems, vol. 33, pp. 5047\u20135057 (2020)"},{"key":"21_CR55","doi-asserted-by":"crossref","unstructured":"Su, H., Jampani, V., Sun, D., Gallo, O., Learned-Miller, E., Kautz, J.: Pixel-adaptive convolutional neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 11166\u201311175 (2019)","DOI":"10.1109\/CVPR.2019.01142"},{"key":"21_CR56","doi-asserted-by":"crossref","unstructured":"Sun, K., Xiao, B., Liu, D., Wang, J.: Deep high-resolution representation learning for human pose estimation. In: IEEE Conference on Computer Vision and Pattern Recognition (2019)","DOI":"10.1109\/CVPR.2019.00584"},{"issue":"1","key":"21_CR57","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s41074-018-0039-6","volume":"10","author":"N Takemura","year":"2018","unstructured":"Takemura, N., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition. IPSJ Trans. Comput. Vis. Appl. 10(1), 1\u201314 (2018). https:\/\/doi.org\/10.1186\/s41074-018-0039-6","journal-title":"IPSJ Trans. Comput. Vis. Appl."},{"key":"21_CR58","doi-asserted-by":"crossref","unstructured":"Veit, A., Belongie, S.: Convolutional networks with adaptive inference graphs. In: European Conference on Computer Vision, pp. 3\u201318 (2018)","DOI":"10.1007\/978-3-030-01246-5_1"},{"key":"21_CR59","doi-asserted-by":"crossref","unstructured":"Wang, D., Zhang, S.: Unsupervised person re-identification via multi-label classification. In: IEEE Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.01099"},{"issue":"2","key":"21_CR60","first-page":"149","volume":"14","author":"L Wang","year":"2004","unstructured":"Wang, L., Ning, H., Tan, T., Hu, W.: Fusion of static and dynamic body biometrics for gait recognition. IEEE TCSVT 14(2), 149\u2013158 (2004)","journal-title":"IEEE TCSVT"},{"key":"21_CR61","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, F., Dou, Z.Y., Darrell, T., Gonzalez, J.E.: Skipnet: learning dynamic routing in convolutional networks. In: European Conference on Computer Vision, pp. 409\u2013424 (2018)","DOI":"10.1007\/978-3-030-01261-8_25"},{"key":"21_CR62","doi-asserted-by":"crossref","unstructured":"Woo, S., Park, J., Lee, J.Y., Kweon, I.S.: CBAM: convolutional block attention module. In: European Conference on Computer Vision, pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"21_CR63","unstructured":"Yang, B., Bender, G., Le, Q.V., Ngiam, J.: Condconv: conditionally parameterized convolutions for efficient inference. In: Advances in Neural Information Processing Systems (2019)"},{"key":"21_CR64","unstructured":"Yu, S., Tan, D., Tan, T.: A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: International Conference on Pattern Recognition, pp. 441\u2013444 (2006)"},{"key":"21_CR65","doi-asserted-by":"crossref","unstructured":"Zhang, F., Wah, B.W.: Supplementary meta-learning: towards a dynamic model for deep neural networks. In: International Conference on Computer Vision, pp. 4344\u20134353 (2017)","DOI":"10.1109\/ICCV.2017.466"},{"key":"21_CR66","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/TIP.2019.2926208","volume":"29","author":"Y Zhang","year":"2019","unstructured":"Zhang, Y., Huang, Y., Yu, S., Wang, L.: Cross-view gait recognition by discriminative feature learning. IEEE Trans. Image Process. 29, 1001\u20131015 (2019)","journal-title":"IEEE Trans. Image Process."}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20065-6_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,7]],"date-time":"2022-11-07T00:13:44Z","timestamp":1667780024000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20065-6_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031200649","9783031200656"],"references-count":66,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20065-6_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"3 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}