{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T04:08:27Z","timestamp":1748664507583,"version":"3.41.0"},"publisher-location":"Cham","reference-count":58,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031895593","type":"print"},{"value":"9783031895609","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-89560-9_4","type":"book-chapter","created":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T14:12:35Z","timestamp":1748614355000},"page":"35-59","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Datasets"],"prefix":"10.1007","author":[{"given":"Diogo R. M.","family":"Bastos","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7603-6526","authenticated-orcid":false,"given":"Jo\u00e3o Manuel","family":"R. S. Tavares","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,31]]},"reference":[{"key":"4_CR1","doi-asserted-by":"publisher","DOI":"10.1155\/2021\/3110416","author":"WA Alsaggaf","year":"2021","unstructured":"Alsaggaf, W.A., Mehmood, I., Khairullah, E.F., Alhuraiji, S., Sabir, M., Alghamdi, A.S., Abd El-Latif, A.A.: A smart surveillance system for uncooperative gait recognition using cycle consistent generative adversarial networks (CCGANs). Comput. Intell. Neurosci. (2021). https:\/\/doi.org\/10.1155\/2021\/3110416","journal-title":"Comput. Intell. Neurosci."},{"key":"4_CR2","doi-asserted-by":"publisher","unstructured":"An, W., Yu, S., Makihara, Y., Wu, X., Xu, C., Yu, Y., Liao, R., Yagi, Y.: Performance evaluation of model-based gait on multi-view very large population database with pose sequences. IEEE Trans. Biometrics Behav. Identity Sci. 1\u20131 (2020). https:\/\/doi.org\/10.1109\/TBIOM.2020.3008862","DOI":"10.1109\/TBIOM.2020.3008862"},{"key":"4_CR3","unstructured":"Cao, Z., Hidalgo, G., Simon, T., Wei, S.-E., Sheikh, Y.: Openpose: realtime multi-person 2d pose estimation using part affinity fields (2019)"},{"issue":"17","key":"4_CR4","doi-asserted-by":"publisher","first-page":"14173","DOI":"10.1007\/s00521-020-04811-z","volume":"32","author":"F Castro","year":"2020","unstructured":"Castro, F., Mar\u00edn-Jim\u00e9nez, M., Guil, N., Blanca, N.: Multimodal feature fusion for CNN-based gait recognition: an empirical comparison. Neural Comput. Appl. 32(17), 14173\u201314193 (2020). https:\/\/doi.org\/10.1007\/s00521-020-04811-z","journal-title":"Neural Comput. Appl."},{"issue":"7","key":"4_CR5","doi-asserted-by":"publisher","first-page":"3467","DOI":"10.1109\/TPAMI.2021.3057879","volume":"44","author":"H Chao","year":"2021","unstructured":"Chao, H., Wang, K., He, Y., Zhang, J., Feng, J.: GaitSet: cross-view gait recognition through utilizing gait as a deep set. IEEE Trans. Pattern Anal. Mach. Intell. 44(7), 3467\u20133478 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2021.3057879","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4_CR6","doi-asserted-by":"crossref","unstructured":"Cosma, A., Radoi, E.: Learning gait representations with noisy multi-task learning. Sensors 22(18) (2022). https:\/\/www.mdpi.com\/1424-8220\/22\/18\/6803","DOI":"10.3390\/s22186803"},{"key":"4_CR7","doi-asserted-by":"publisher","unstructured":"Fan, C., Peng, Y., Cao, C., Liu, X., Hou, S., Chi, J., Huang, Y., Li, Q., He, Z.: Gaitpart: temporal part-based model for gait recognition. In: IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), vol. 2020, pp. 14213\u201314221 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.01423","DOI":"10.1109\/CVPR42600.2020.01423"},{"issue":"2","key":"4_CR8","doi-asserted-by":"publisher","first-page":"111","DOI":"10.1049\/cvi2.12070","volume":"16","author":"S Gao","year":"2021","unstructured":"Gao, S., Yun, J., Zhao, Y., Liu, L.: Gait-D: skeleton-based gait feature decomposition for gait recognition. IET Comput. Vis. 16(2), 111\u2013125 (2021). https:\/\/doi.org\/10.1049\/cvi2.12070","journal-title":"IET Comput. Vis."},{"key":"4_CR9","doi-asserted-by":"publisher","unstructured":"Gao, Z., Wu, J., Wu, T., Huang, R., Zhang, A., Zhao, J.: Robust clothing-independent gait recognition using hybrid part-based gait features. PeerJ Comput. Sci. 8 (2022). https:\/\/doi.org\/10.7717\/peerj-cs.996","DOI":"10.7717\/peerj-cs.996"},{"issue":"3","key":"4_CR10","doi-asserted-by":"publisher","first-page":"1290","DOI":"10.1111\/coin.12361","volume":"36","author":"M George","year":"2020","unstructured":"George, M., Govindarajan, T., Rajasekaran, K., Bandi, S.: A robust similarity based deep Siamese convolutional neural network for gait recognition across views. Comput. Intell. 36(3), 1290\u20131319 (2020). https:\/\/doi.org\/10.1111\/coin.12361","journal-title":"Comput. Intell."},{"key":"4_CR11","doi-asserted-by":"publisher","unstructured":"Gul, S., Malik, M.I., Khan, G.M., Shafait, F.: Multi-view gait recognition system using spatio-temporal features and deep learning. Expert Syst. Appl. 179 (2021). https:\/\/doi.org\/10.1016\/j.eswa.2021.115057","DOI":"10.1016\/j.eswa.2021.115057"},{"issue":"28\u201329","key":"4_CR12","doi-asserted-by":"publisher","first-page":"35903","DOI":"10.1007\/s11042-020-10071-9","volume":"80","author":"S Gupta","year":"2020","unstructured":"Gupta, S., Chattopadhyay, P.: Exploiting pose dynamics for human recognition from their gait signatures. Multimedia Tools Appl. 80(28\u201329), 35903\u201335921 (2020). https:\/\/doi.org\/10.1007\/s11042-020-10071-9","journal-title":"Multimedia Tools Appl."},{"key":"4_CR13","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1016\/j.neucom.2021.04.113","volume":"454","author":"S Gupta","year":"2021","unstructured":"Gupta, S., Chattopadhyay, P.: Gait recognition in the presence of co-variate conditions. Neurocomputing 454, 76\u201387 (2021). https:\/\/doi.org\/10.1016\/j.neucom.2021.04.113","journal-title":"Neurocomputing"},{"key":"4_CR14","doi-asserted-by":"publisher","unstructured":"Gupta, S., Chattopadhyay, P.: Pose-based boundary energy image for gait recognition from silhouette contours. Sadhana\u2014Acad. Proc. Eng. Sci. 48(4) (2023). https:\/\/doi.org\/10.1007\/s12046-023-02234-9","DOI":"10.1007\/s12046-023-02234-9"},{"key":"4_CR15","doi-asserted-by":"publisher","unstructured":"Han, F., Li, X., Zhao, J., Shen, F.: A unified perspective of classification-based loss and distance-based loss for cross-view gait recognition. Pattern Recogn. 125 (2022). https:\/\/doi.org\/10.1016\/j.patcog.2021.108519","DOI":"10.1016\/j.patcog.2021.108519"},{"key":"4_CR16","unstructured":"Hang, C., Deva, R.: 3d human pose estimation = 2d pose estimation + matching (2017)"},{"issue":"18","key":"4_CR17","doi-asserted-by":"publisher","first-page":"25873","DOI":"10.1007\/s11042-019-7638-9","volume":"78","author":"A Hawas","year":"2019","unstructured":"Hawas, A., El-Khobby, H., Elnaby, M., El-Samie, F.: Gait identification by convolutional neural networks and optical flow. Multimedia Tools Appl. 78(18), 25873\u201325888 (2019). https:\/\/doi.org\/10.1007\/s11042-019-7638-9","journal-title":"Multimedia Tools Appl."},{"key":"4_CR18","doi-asserted-by":"publisher","unstructured":"Hofmann, M., Geiger, J., Bachmann, S., Schuller, B., Rigoll, G.: The tum gait from audio, image and depth (GAID) database: multimodal recognition of subjects and traits [Visual understanding and applications with RGB-D cameras]. J. Vis. Commun. Image Represent. 25(1), 195\u2013206 (2014). https:\/\/doi.org\/10.1016\/j.jvcir.2013.02.006","DOI":"10.1016\/j.jvcir.2013.02.006"},{"issue":"11","key":"4_CR19","doi-asserted-by":"publisher","first-page":"8978","DOI":"10.1109\/TNNLS.2022.3154723","volume":"34","author":"S Hou","year":"2022","unstructured":"Hou, S., Liu, X., Cao, C., Huang, Y.: Gait quality aware network: toward the interpretability of silhouette-based gait recognition. IEEE Trans. Neural Netw. Learn. Syst. 34(11), 8978\u20138988 (2022). https:\/\/doi.org\/10.1109\/TNNLS.2022.3154723","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"4_CR20","doi-asserted-by":"publisher","unstructured":"Hua, C., Pan, Y., Li, J., Wang, Z.:. Gait recognition by combining the long-short-term attention network and personal physiological features. Sensors 22(22) (2022). https:\/\/doi.org\/10.3390\/s22228779","DOI":"10.3390\/s22228779"},{"issue":"5","key":"4_CR21","doi-asserted-by":"publisher","first-page":"1511","DOI":"10.1109\/TIFS.2012.2204253","volume":"7","author":"H Iwama","year":"2012","unstructured":"Iwama, H., Okumura, M., Makihara, Y., Yagi, Y.: The OU-ISIR gait database comprising the large population dataset and performance evaluation of gait recognition. IEEE Trans. Inform. Forensics Secur. 7(5), 1511\u20131521 (2012). https:\/\/doi.org\/10.1109\/TIFS.2012.2204253","journal-title":"IEEE Trans. Inform. Forensics Secur."},{"key":"4_CR22","unstructured":"Ke, S., Bin, X., Dong, L., Jingdong, W.: Deep high-resolution representation learning for human pose estimation (2019)"},{"key":"4_CR23","doi-asserted-by":"publisher","unstructured":"Khaliluzzaman, M., Uddin, A., Deb, K., Hasan, M.J.: Person recognition based on deep gait: a survey. Sensors 23(10) (2023). https:\/\/doi.org\/10.3390\/s23104875","DOI":"10.3390\/s23104875"},{"key":"4_CR24","doi-asserted-by":"crossref","unstructured":"Li, J., Wang, C., Zhu, H., Mao, Y., Fang, H.-S., Lu, C.: CrowdPose: efficient crowded scenes pose estimation and a new benchmark (2019)","DOI":"10.1109\/CVPR.2019.01112"},{"issue":"12","key":"4_CR25","doi-asserted-by":"publisher","first-page":"3102","DOI":"10.1109\/TIFS.2019.2912577","volume":"14","author":"X Li","year":"2019","unstructured":"Li, X., Makihara, Y., Xu, C., Yagi, Y., Ren, M.: Joint intensity transformer network for gait recognition robust against clothing and carrying status. IEEE Trans. Inform. Forensics Secur. 14(12), 3102\u20133115 (2019). https:\/\/doi.org\/10.1109\/TIFS.2019.2912577","journal-title":"IEEE Trans. Inform. Forensics Secur."},{"key":"4_CR26","unstructured":"Lin, B., Zhang, S., Wang, M., Li, L., Yu, X.: GaitGL: learning discriminative global-local feature representations for gait recognition (2022)"},{"key":"4_CR27","doi-asserted-by":"publisher","unstructured":"Linda, G.M., Themozhi, G., Bandi, S.R.: Color-mapped contour gait image for cross-view gait recognition using deep convolutional neural network. Int. J. Wavelets Multiresolut. Inform. Process. 18(1) (2019). https:\/\/doi.org\/10.1142\/S0219691319410121","DOI":"10.1142\/S0219691319410121"},{"key":"4_CR28","doi-asserted-by":"publisher","unstructured":"Makihara, Y., Mannami, H., Tsuji, A., Hossain, M.A., Sugiura, K., Mori, A., Yagi, Y.: The OU-ISIR gait database comprising the treadmill dataset. IPSJ Trans. Comput. Vis. Appl. 4 (2012). https:\/\/doi.org\/10.2197\/ipsjtcva.4.53","DOI":"10.2197\/ipsjtcva.4.53"},{"key":"4_CR29","doi-asserted-by":"publisher","unstructured":"Mansur, A., Makihara, Y., Aqmar, R., Yagi, Y.: Gait recognition under speed transition. In: IEEE Conference on Computer Vision and Pattern Recognition, vol. 2014, pp. 2521\u20132528 (2014). https:\/\/doi.org\/10.1109\/CVPR.2014.323","DOI":"10.1109\/CVPR.2014.323"},{"key":"4_CR30","doi-asserted-by":"publisher","first-page":"5452","DOI":"10.1109\/TIFS.2021.3132579","volume":"16","author":"MJ Marin-Jimenez","year":"2021","unstructured":"Marin-Jimenez, M.J., Castro, F.M., Delgado-Escano, R., Kalogeiton, V., Guil, N.: UGaitNet: multimodal gait recognition with missing input modalities. IEEE Trans. Inform. Forensics Secur. 16, 5452\u20135462 (2021). https:\/\/doi.org\/10.1109\/TIFS.2021.3132579","journal-title":"IEEE Trans. Inform. Forensics Secur."},{"key":"4_CR31","doi-asserted-by":"publisher","unstructured":"Merlin Linda, G., Lakshmi, Sree Rathna, N.V.S., Murugan, N.S., Mahapatra, R.P., Muthukumaran, V., Sivaram, M.: Intelligent recognition system for viewpoint variations on gait and speech using CNN-CapsNet. Int. J. Intell. Comput. Cybern. 15(3), 363\u2013382 (2021). https:\/\/doi.org\/10.1108\/IJICC-08-2021-0178","DOI":"10.1108\/IJICC-08-2021-0178"},{"key":"4_CR32","doi-asserted-by":"publisher","unstructured":"Mogan, J., Lee, C., Lim, K., Ali, M., Alqahtani, A.: Gait-CNN-ViT: multi-model gait recognition with convolutional neural networks and vision transformer. Sensors 23(8) (2023). https:\/\/doi.org\/10.3390\/s23083809","DOI":"10.3390\/s23083809"},{"issue":"1","key":"4_CR33","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1109\/TBIOM.2020.3031470","volume":"3","author":"A Moghaddam","year":"2020","unstructured":"Moghaddam, A., Etemad, A.: View-invariant gait recognition with attentive recurrent learning of partial representations. IEEE Trans. Biometrics Behav. Identity Sci. 3(1), 124\u2013137 (2020). https:\/\/doi.org\/10.1109\/TBIOM.2020.3031470","journal-title":"IEEE Trans. Biometrics Behav. Identity Sci."},{"key":"4_CR34","doi-asserted-by":"publisher","unstructured":"Pin\u010di\u0107, D., Su\u0161anj, D., Lenac, K.:. Gait recognition with self-supervised learning of gait features based on vision transformers. Sensors 22(19) (2022). https:\/\/doi.org\/10.3390\/s22197140","DOI":"10.3390\/s22197140"},{"key":"4_CR35","doi-asserted-by":"publisher","unstructured":"Sayeed, M.S., Min, P.P., Bari, M.A.: Deep learning based gait recognition using convolutional neural network in the COVID-19 pandemic. Emerg. Sci. J. 6(5), 1086\u20131099 (2022). https:\/\/doi.org\/10.28991\/ESJ-2022-06-05-012","DOI":"10.28991\/ESJ-2022-06-05-012"},{"issue":"3","key":"4_CR36","doi-asserted-by":"publisher","first-page":"2801","DOI":"10.1109\/TPAMI.2022.3183288","volume":"45","author":"C Song","year":"2023","unstructured":"Song, C., Huang, Y., Wang, W., Wang, L.: CASIA-E: a large comprehensive dataset for gait recognition. IEEE Trans. Pattern Anal. Mach. Intell. 45(3), 2801\u20132815 (2023). https:\/\/doi.org\/10.1109\/TPAMI.2022.3183288","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4_CR37","doi-asserted-by":"publisher","unstructured":"Takemura, N., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition. IPSJ Trans. Comput. Vis. Appl. 10 (2018). https:\/\/doi.org\/10.1186\/s41074-018-0039-6","DOI":"10.1186\/s41074-018-0039-6"},{"key":"4_CR38","doi-asserted-by":"publisher","unstructured":"Tan, D., Huang, K., Yu, S., Tan, T.: Efficient night gait recognition based on template matching. In: Proceedings\u2014International Conference on Pattern Recognition, vol. 3, pp. 1000\u20131003 (2006). https:\/\/doi.org\/10.1109\/ICPR.2006.478","DOI":"10.1109\/ICPR.2006.478"},{"key":"4_CR39","doi-asserted-by":"publisher","first-page":"646","DOI":"10.1016\/j.patrec.2019.07.008","volume":"125","author":"D Thapar","year":"2019","unstructured":"Thapar, D., Jaswal, G., Nigam, A., Arora, C.: Gait metric learning Siamese network exploiting dual of spatio-temporal 3D-CNN intra and LSTM based inter gait-cycle-segment features. Pattern Recogn. Lett. 125, 646\u2013653 (2019). https:\/\/doi.org\/10.1016\/j.patrec.2019.07.008","journal-title":"Pattern Recogn. Lett."},{"key":"4_CR40","doi-asserted-by":"publisher","unstructured":"Uddin, M., Thanh Trung, N., Makihara, Y., Takemura, N., Li, X., Muramatsu, D., Yagi, Y.: The OU-ISIR large population gait database with real-life carried object and its performance evaluation. IPSJ Trans. Comput. Vis. Appl. 10 (2018). https:\/\/doi.org\/10.1186\/s41074-018-0041-z","DOI":"10.1186\/s41074-018-0041-z"},{"key":"4_CR41","doi-asserted-by":"publisher","first-page":"1505","DOI":"10.1109\/TPAMI.2003.1251144","volume":"25","author":"L Wang","year":"2004","unstructured":"Wang, L., Tan, T., Ning, H., Hu, W.: Silhouette analysis-based gait recognition for human identification. IEEE Trans. Pattern Anal. Mach. Intell. 25, 1505\u20131518 (2004). https:\/\/doi.org\/10.1109\/TPAMI.2003.1251144","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"issue":"2","key":"4_CR42","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1109\/TBIOM.2023.3244206","volume":"5","author":"R Wang","year":"2023","unstructured":"Wang, R., Shi, Y., Ling, H., Li, Z., Li, P., Liu, B., Zheng, H., Wang, Q.: Gait recognition via gait period set. IEEE Trans. Biometrics Behav. Identity Sci. 5(2), 183\u2013195 (2023). https:\/\/doi.org\/10.1109\/TBIOM.2023.3244206","journal-title":"IEEE Trans. Biometrics Behav. Identity Sci."},{"issue":"31\u201332","key":"4_CR43","doi-asserted-by":"publisher","first-page":"22653","DOI":"10.1007\/s11042-020-09003-4","volume":"79","author":"Y Wang","year":"2020","unstructured":"Wang, Y., Chen, Z., Wu, Q., Rong, X.: Deep mutual learning network for gait recognition. Multimedia Tools Appl. 79(31\u201332), 22653\u201322672 (2020). https:\/\/doi.org\/10.1007\/s11042-020-09003-4","journal-title":"Multimedia Tools Appl."},{"key":"4_CR44","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1016\/j.neucom.2019.02.025","volume":"339","author":"Y Wang","year":"2019","unstructured":"Wang, Y., Song, C., Huang, Y., Wang, Z., Wang, L.: Learning view invariant gait features with two-stream GAN. Neurocomputing 339, 245\u2013254 (2019). https:\/\/doi.org\/10.1016\/j.neucom.2019.02.025","journal-title":"Neurocomputing"},{"issue":"24","key":"4_CR45","doi-asserted-by":"publisher","first-page":"35789","DOI":"10.1007\/s11042-019-08153-4","volume":"78","author":"C Wu","year":"2019","unstructured":"Wu, C., Song, Y., Zhang, Y.: Multi-view gait recognition using NMF and 2DLDA. Multimedia Tools Appl. 78(24), 35789\u201335811 (2019). https:\/\/doi.org\/10.1007\/s11042-019-08153-4","journal-title":"Multimedia Tools Appl."},{"key":"4_CR46","doi-asserted-by":"publisher","first-page":"2734","DOI":"10.1109\/TIP.2020.3039888","volume":"30","author":"H Wu","year":"2020","unstructured":"Wu, H., Tian, J., Fu, Y., Li, B., Li, X.: Condition-aware comparison scheme for gait recognition. IEEE Trans. Image Process. 30, 2734\u20132744 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.3039888","journal-title":"IEEE Trans. Image Process."},{"key":"4_CR47","unstructured":"Xianda, G., Zheng, Z., Tian, Y., Beibei, L., Junjie, H., Jiankang, D., Guan, H., Jie, Z., Jiwen, L.: Gait recognition in the wild: a large-scale benchmark and NAS-based baseline (2022)"},{"key":"4_CR48","doi-asserted-by":"publisher","unstructured":"Xu, C., Makihara, Y., Li, X., Yagi, Y., Lu, J.: Gait recognition from a single image using a phase-aware gait cycle reconstruction network. In: Computer Vision\u2014ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XIX, pp. 386\u2013403 (2020). https:\/\/doi.org\/10.1007\/978-3-030-58529-7_23","DOI":"10.1007\/978-3-030-58529-7_23"},{"key":"4_CR49","doi-asserted-by":"publisher","unstructured":"Xu, C., Makihara, Y., Ogi, G., Li, X., Yagi, Y., Lu, J.: The OU-ISIR gait database comprising the large population dataset with age and performance evaluation of age estimation. IPSJ Trans. Comput. Vis. Appl. 9 (2017). https:\/\/doi.org\/10.1186\/s41074-017-0035-2","DOI":"10.1186\/s41074-017-0035-2"},{"key":"4_CR50","doi-asserted-by":"publisher","first-page":"3265","DOI":"10.1109\/TMM.2021.3095809","volume":"24","author":"K Xu","year":"2021","unstructured":"Xu, K., Jiang, X., Sun, T.: Gait recognition based on local graphical skeleton descriptor with pairwise similarity network. IEEE Trans. Multimedia 24, 3265\u20133275 (2021). https:\/\/doi.org\/10.1109\/TMM.2021.3095809","journal-title":"IEEE Trans. Multimedia"},{"key":"4_CR51","doi-asserted-by":"publisher","first-page":"95","DOI":"10.1016\/j.neucom.2019.11.015","volume":"380","author":"W Xue","year":"2019","unstructured":"Xue, W., Ai, H., Sun, T., Song, C., Huang, Y., Wang, L.: Frame-GAN: increasing the frame rate of gait videos with generative adversarial networks. Neurocomputing 380, 95\u2013104 (2019). https:\/\/doi.org\/10.1016\/j.neucom.2019.11.015","journal-title":"Neurocomputing"},{"key":"4_CR52","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2022.3171961","author":"L Yao","year":"2022","unstructured":"Yao, L., Kusakunniran, W., Zhang, P., Wu, Q., Zhang, J.: Improving disentangled representation learning for gait recognition using group supervision. IEEE Trans. Multimedia (2022). https:\/\/doi.org\/10.1109\/TMM.2022.3171961","journal-title":"IEEE Trans. Multimedia"},{"key":"4_CR53","doi-asserted-by":"publisher","unstructured":"Yeo, S.-S., Rho, S., Kim, H., Safdar, J., Zia, U., Durrani, M.Y.: A triplet-branch convolutional neural network for part-based gait recognition. Comput. Syst. Sci. Eng. 47(2), 2027\u20132047 (2023). https:\/\/doi.org\/10.32604\/csse.2023.040327","DOI":"10.32604\/csse.2023.040327"},{"key":"4_CR54","doi-asserted-by":"publisher","unstructured":"Yu, S., Tan, D., Tan, T.: A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: 18th International Conference on Pattern Recognition (ICPR\u201906), vol. 4, pp. 441\u2013444 (2006). https:\/\/doi.org\/10.1109\/ICPR.2006.67","DOI":"10.1109\/ICPR.2006.67"},{"key":"4_CR55","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/TIP.2019.2926208","volume":"29","author":"Y Zhang","year":"2019","unstructured":"Zhang, Y., Huang, Y., Yu, S., Wang, L.: Cross-view gait recognition by discriminative feature learning. IEEE Trans. Image Process. 29, 1001\u20131015 (2019). https:\/\/doi.org\/10.1109\/TIP.2019.2926208","journal-title":"IEEE Trans. Image Process."},{"key":"4_CR56","doi-asserted-by":"crossref","unstructured":"Zheng, J., Liu, X., Liu, W., He, L., Yan, C., Mei, T.: Gait recognition in the wild with dense 3d representations and a benchmark. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 20228\u201320237 (2022)","DOI":"10.1109\/CVPR52688.2022.01959"},{"issue":"2","key":"4_CR57","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1049\/csy2.12052","volume":"4","author":"L Zheng","year":"2022","unstructured":"Zheng, L., Zha, Y., Kong, D., Yang, H., Zhang, Y.: Multi-branch angle aware spatial temporal graph convolutional neural network for model-based gait recognition. IET Cyber-Syst. Robot. 4(2), 97\u2013106 (2022). https:\/\/doi.org\/10.1049\/csy2.12052","journal-title":"IET Cyber-Syst. Robot."},{"key":"4_CR58","doi-asserted-by":"publisher","unstructured":"Zheng, S., Huang, K., Tan, T., Tao, D.: A cascade fusion scheme for gait and cumulative foot pressure image recognition. Pattern Recogn. 45(10), 3603\u20133610 (2012). https:\/\/doi.org\/10.1016\/j.patcog.2012.03.008","DOI":"10.1016\/j.patcog.2012.03.008"}],"container-title":["Studies in Systems, Decision and Control","Advances in Gait-Based Identification"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-89560-9_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T14:12:39Z","timestamp":1748614359000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-89560-9_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031895593","9783031895609"],"references-count":58,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-89560-9_4","relation":{},"ISSN":["2198-4182","2198-4190"],"issn-type":[{"value":"2198-4182","type":"print"},{"value":"2198-4190","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"31 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}}]}}