{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,31]],"date-time":"2025-05-31T04:08:27Z","timestamp":1748664507972,"version":"3.41.0"},"publisher-location":"Cham","reference-count":64,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031895593","type":"print"},{"value":"9783031895609","type":"electronic"}],"license":[{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,1,1]],"date-time":"2025-01-01T00:00:00Z","timestamp":1735689600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-89560-9_5","type":"book-chapter","created":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T14:12:39Z","timestamp":1748614359000},"page":"61-94","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Comparison of\u00a0the\u00a0Reviewed Methods"],"prefix":"10.1007","author":[{"given":"Diogo R. M.","family":"Bastos","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-7603-6526","authenticated-orcid":false,"given":"Jo\u00e3o Manuel","family":"R. S. Tavares","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,31]]},"reference":[{"key":"5_CR1","doi-asserted-by":"publisher","first-page":"31177","DOI":"10.1109\/ACCESS.2023.3262271","volume":"11","author":"M Aljazaerly","year":"2022","unstructured":"Aljazaerly, M., Makihara, Y., Muramatsu, D., Yagi, Y.: Batch hard contrastive loss and its application to cross-view gait recognition. IEEE Access 11, 31177\u201331187 (2022). https:\/\/doi.org\/10.1109\/ACCESS.2023.3262271","journal-title":"IEEE Access"},{"key":"5_CR2","doi-asserted-by":"publisher","DOI":"10.1155\/2021\/3110416","author":"WA Alsaggaf","year":"2021","unstructured":"Alsaggaf, W.A., Mehmood, I., Khairullah, E.F., Alhuraiji, S., Sabir, M., Alghamdi, A.S., Abd El-Latif, A.A.: A smart surveillance system for uncooperative gait recognition using cycle consistent generative adversarial networks (CCGANs). Comput. Intell. Neurosci. (2021). https:\/\/doi.org\/10.1155\/2021\/3110416","journal-title":"Comput. Intell. Neurosci."},{"key":"5_CR3","doi-asserted-by":"publisher","first-page":"116","DOI":"10.1016\/j.neucom.2019.01.091","volume":"338","author":"M Babaee","year":"2019","unstructured":"Babaee, M., Li, L., Rigoll, G.: Person identification from partial gait cycle using fully convolutional neural networks. Neurocomputing 338, 116\u2013125 (2019). https:\/\/doi.org\/10.1016\/j.neucom.2019.01.091","journal-title":"Neurocomputing"},{"issue":"6","key":"5_CR4","doi-asserted-by":"publisher","first-page":"269","DOI":"10.1049\/iet-bmt.2020.0001","volume":"9","author":"A Bekhouch","year":"2020","unstructured":"Bekhouch, A., Bouchrika, I., Doghmane, N.: Gait biometrics: investigating the use of the lower inner regions for people identification from landmark frames. IET Biometrics 9(6), 269\u2013277 (2020). https:\/\/doi.org\/10.1049\/iet-bmt.2020.0001","journal-title":"IET Biometrics"},{"issue":"17","key":"5_CR5","doi-asserted-by":"publisher","first-page":"14173","DOI":"10.1007\/s00521-020-04811-z","volume":"32","author":"F Castro","year":"2020","unstructured":"Castro, F., Mar\u00edn-Jim\u00e9nez, M., Guil, N., Blanca, N.: Multimodal feature fusion for CNN-based gait recognition: an empirical comparison. Neural Comput. Appl. 32(17), 14173\u201314193 (2020). https:\/\/doi.org\/10.1007\/s00521-020-04811-z","journal-title":"Neural Comput. Appl."},{"issue":"7","key":"5_CR6","doi-asserted-by":"publisher","first-page":"3467","DOI":"10.1109\/TPAMI.2021.3057879","volume":"44","author":"H Chao","year":"2021","unstructured":"Chao, H., Wang, K., He, Y., Zhang, J., Feng, J.: GaitSet: cross-view gait recognition through utilizing gait as a deep set. IEEE Trans. Pattern Anal. Mach. Intell. 44(7), 3467\u20133478 (2021). https:\/\/doi.org\/10.1109\/TPAMI.2021.3057879","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"5_CR7","doi-asserted-by":"publisher","unstructured":"Cosma, A., Catruna, A., Radoi, E.:. Exploring self-supervised vision transformers for gait recognition in the wild. Sensors 23(5) (2023). https:\/\/doi.org\/10.3390\/s23052680","DOI":"10.3390\/s23052680"},{"key":"5_CR8","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., Houlsby, N.: An image is worth 16$$\\,\\times \\,$$16 words: transformers for image recognition at scale (2021)"},{"key":"5_CR9","doi-asserted-by":"publisher","first-page":"14213","DOI":"10.1109\/CVPR42600.2020.01423","volume":"2020","author":"C Fan","year":"2020","unstructured":"Fan, C., Peng, Y., Cao, C., Liu, X., Hou, S., Chi, J., Huang, Y., Li, Q., He, Z.: Gaitpart: temporal part-based model for gait recognition. IEEE\/CVF Conf. Comput. Vis. Pattern Recogn. (CVPR) 2020, 14213\u201314221 (2020). https:\/\/doi.org\/10.1109\/CVPR42600.2020.01423","journal-title":"IEEE\/CVF Conf. Comput. Vis. Pattern Recogn. (CVPR)"},{"key":"5_CR10","doi-asserted-by":"publisher","unstructured":"Gao, Z., Wu, J., Wu, T., Huang, R., Zhang, A., Zhao, J.: Robust clothing-independent gait recognition using hybrid part-based gait features. PeerJ Comput. Sci. 8 (2022). https:\/\/doi.org\/10.7717\/peerj-cs.996","DOI":"10.7717\/peerj-cs.996"},{"issue":"3","key":"5_CR11","doi-asserted-by":"publisher","first-page":"1290","DOI":"10.1111\/coin.12361","volume":"36","author":"M George","year":"2020","unstructured":"George, M., Govindarajan, T., Rajasekaran, K., Bandi, S.: A robust similarity based deep Siamese convolutional neural network for gait recognition across views. Comput. Intell. 36(3), 1290\u20131319 (2020). https:\/\/doi.org\/10.1111\/coin.12361","journal-title":"Comput. Intell."},{"key":"5_CR12","doi-asserted-by":"publisher","unstructured":"Ghosh, R.: A faster R-CNN and recurrent neural network based approach of gait recognition with and without carried objects. Expert Syst. Appl. 205 (2022). https:\/\/doi.org\/10.1016\/j.eswa.2022.117730","DOI":"10.1016\/j.eswa.2022.117730"},{"key":"5_CR13","unstructured":"Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial networks (2014)"},{"key":"5_CR14","doi-asserted-by":"publisher","unstructured":"Gul, S., Malik, M.I., Khan, G.M., Shafait, F.: Multi-view gait recognition system using spatio-temporal features and deep learning. Expert Syst. Appl. 179 (2021). https:\/\/doi.org\/10.1016\/j.eswa.2021.115057","DOI":"10.1016\/j.eswa.2021.115057"},{"key":"5_CR15","doi-asserted-by":"publisher","unstructured":"Guo, H., Li, B., Zhang, Y., Zhang, Y., Li, W., Qiao, F., Rong, X., Zhou, S.: Gait recognition based on the feature extraction of gabor filter and linear discriminant analysis and improved local coupled extreme learning machine. Math. Prob. Eng. 2020 (2020). https:\/\/doi.org\/10.1155\/2020\/5393058","DOI":"10.1155\/2020\/5393058"},{"issue":"28\u201329","key":"5_CR16","doi-asserted-by":"publisher","first-page":"35903","DOI":"10.1007\/s11042-020-10071-9","volume":"80","author":"S Gupta","year":"2020","unstructured":"Gupta, S., Chattopadhyay, P.: Exploiting pose dynamics for human recognition from their gait signatures. Multimedia Tools Appl. 80(28\u201329), 35903\u201335921 (2020). https:\/\/doi.org\/10.1007\/s11042-020-10071-9","journal-title":"Multimedia Tools Appl."},{"key":"5_CR17","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1016\/j.neucom.2021.04.113","volume":"454","author":"S Gupta","year":"2021","unstructured":"Gupta, S., Chattopadhyay, P.: Gait recognition in the presence of co-variate conditions. Neurocomputing 454, 76\u201387 (2021). https:\/\/doi.org\/10.1016\/j.neucom.2021.04.113","journal-title":"Neurocomputing"},{"key":"5_CR18","doi-asserted-by":"publisher","unstructured":"Gupta, S., Chattopadhyay, P.: Pose-based boundary energy image for gait recognition from silhouette contours. Sadhana\u2014Acad. Proc. Eng. Sci. 48(4) (2023). https:\/\/doi.org\/10.1007\/s12046-023-02234-9","DOI":"10.1007\/s12046-023-02234-9"},{"key":"5_CR19","doi-asserted-by":"publisher","unstructured":"Han, F., Li, X., Zhao, J., Shen, F.: A unified perspective of classification-based loss and distance-based loss for cross-view gait recognition. Pattern Recogn. 125 (2022). https:\/\/doi.org\/10.1016\/j.patcog.2021.108519","DOI":"10.1016\/j.patcog.2021.108519"},{"issue":"18","key":"5_CR20","doi-asserted-by":"publisher","first-page":"25873","DOI":"10.1007\/s11042-019-7638-9","volume":"78","author":"A Hawas","year":"2019","unstructured":"Hawas, A., El-Khobby, H., Elnaby, M., El-Samie, F.: Gait identification by convolutional neural networks and optical flow. Multimedia Tools Appl. 78(18), 25873\u201325888 (2019). https:\/\/doi.org\/10.1007\/s11042-019-7638-9","journal-title":"Multimedia Tools Appl."},{"key":"5_CR21","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition (2015)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"11","key":"5_CR22","doi-asserted-by":"publisher","first-page":"8978","DOI":"10.1109\/TNNLS.2022.3154723","volume":"34","author":"S Hou","year":"2022","unstructured":"Hou, S., Liu, X., Cao, C., Huang, Y.: Gait quality aware network: toward the interpretability of silhouette-based gait recognition. IEEE Trans. Neural Netw. Learn. Syst. 34(11), 8978\u20138988 (2022). https:\/\/doi.org\/10.1109\/TNNLS.2022.3154723","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"5_CR23","doi-asserted-by":"publisher","unstructured":"Hua, C., Pan, Y., Li, J., Wang, Z.: Gait recognition by combining the long-short-term attention network and personal physiological features. Sensors 22(22) (2022). https:\/\/doi.org\/10.3390\/s22228779","DOI":"10.3390\/s22228779"},{"key":"5_CR24","doi-asserted-by":"publisher","unstructured":"Iwashita, Y., Sakano, H., Kurazume, R., Stoica, A.: Speed invariant gait recognition\u2014the enhanced mutual subspace method. PLoS ONE 16(8 August) (2021). https:\/\/doi.org\/10.1371\/journal.pone.0255927","DOI":"10.1371\/journal.pone.0255927"},{"key":"5_CR25","doi-asserted-by":"publisher","unstructured":"Jameel, H., Dhannoon, B.: Gait recognition based on deep learning. Iraqi J. Sci. 63(1), 397\u2013408 (2021). https:\/\/doi.org\/10.24996\/ijs.2022.63.1.36","DOI":"10.24996\/ijs.2022.63.1.36"},{"issue":"8","key":"5_CR26","doi-asserted-by":"publisher","first-page":"1239","DOI":"10.1587\/transinf.2020BDP0010","volume":"E104D","author":"P Jia","year":"2021","unstructured":"Jia, P., Zhao, Q., Li, B., Zhang, J.: CJAM: convolutional neural network joint attention mechanism in gait recognition. IEICE Trans. Inform. Syst. E104D(8), 1239\u20131249 (2021). https:\/\/doi.org\/10.1587\/transinf.2020BDP0010","journal-title":"IEICE Trans. Inform. Syst."},{"key":"5_CR27","doi-asserted-by":"publisher","unstructured":"Khaliluzzaman, M., Uddin, A., Deb, K., Hasan, M.J.: Person recognition based on deep gait: a survey. Sensors, 23(10) (2023). https:\/\/doi.org\/10.3390\/s23104875","DOI":"10.3390\/s23104875"},{"issue":"12","key":"5_CR28","doi-asserted-by":"publisher","first-page":"3102","DOI":"10.1109\/TIFS.2019.2912577","volume":"14","author":"X Li","year":"2019","unstructured":"Li, X., Makihara, Y., Xu, C., Yagi, Y., Ren, M.: Joint intensity transformer network for gait recognition robust against clothing and carrying status. IEEE Trans. Inform. Forensics Secur. 14(12), 3102\u20133115 (2019). https:\/\/doi.org\/10.1109\/TIFS.2019.2912577","journal-title":"IEEE Trans. Inform. Forensics Secur."},{"issue":"2","key":"5_CR29","doi-asserted-by":"publisher","first-page":"497","DOI":"10.1007\/s10044-020-00935-z","volume":"24","author":"V Lima","year":"2020","unstructured":"Lima, V., Melo, V., Schwartz, W.: Simple and efficient pose-based gait recognition method for challenging environments. Pattern Anal Appl 24(2), 497\u2013507 (2020). https:\/\/doi.org\/10.1007\/s10044-020-00935-z","journal-title":"Pattern Anal Appl"},{"key":"5_CR30","unstructured":"Lin, B., Zhang, S., Wang, M., Li, L., Yu, X.: GaitGL: learning discriminative global-local feature representations for gait recognition (2022)"},{"key":"5_CR31","doi-asserted-by":"publisher","unstructured":"Linda, G.M., Themozhi, G., Bandi, S.R.: Color-mapped contour gait image for cross-view gait recognition using deep convolutional neural network. Int. J. Wavelets, Multiresolut. Inform. Process. 18(1) (2019). https:\/\/doi.org\/10.1142\/S0219691319410121","DOI":"10.1142\/S0219691319410121"},{"key":"5_CR32","doi-asserted-by":"publisher","first-page":"5452","DOI":"10.1109\/TIFS.2021.3132579","volume":"16","author":"MJ Marin-Jimenez","year":"2021","unstructured":"Marin-Jimenez, M.J., Castro, F.M., Delgado-Escano, R., Kalogeiton, V., Guil, N.: UGaitNet: multimodal gait recognition with missing input modalities. IEEE Trans. Inform. Forensics Secur. 16, 5452\u20135462 (2021). https:\/\/doi.org\/10.1109\/TIFS.2021.3132579","journal-title":"IEEE Trans. Inform. Forensics Secur."},{"key":"5_CR33","doi-asserted-by":"publisher","unstructured":"Merlin Linda, G., Lakshmi, Sree Rathna, N.V.S., Murugan, N.S., Mahapatra, R.P., Muthukumaran, V., Sivaram, M.: Intelligent recognition system for viewpoint variations on gait and speech using CNN-CapsNet. Int. J. Intell. Computi. Cybern. 15(3):363\u2013382 (2021). https:\/\/doi.org\/10.1108\/IJICC-08-2021-0178","DOI":"10.1108\/IJICC-08-2021-0178"},{"key":"5_CR34","doi-asserted-by":"publisher","unstructured":"Mogan, J., Lee, C., Lim, K., Ali, M., Alqahtani, A.: Gait-CNN-ViT: multi-model gait recognition with convolutional neural networks and vision transformer. Sensors 23(8) (2023). https:\/\/doi.org\/10.3390\/s23083809","DOI":"10.3390\/s23083809"},{"key":"5_CR35","doi-asserted-by":"publisher","unstructured":"Mogan, J., Lee, C., Lim, K., Muthu, K.: VGG16-MLP: gait recognition with fine-tuned VGG-16 and multilayer perceptron. Appl. Sci. (Switzerl.) 12(15) (2022). https:\/\/doi.org\/10.3390\/app12157639","DOI":"10.3390\/app12157639"},{"key":"5_CR36","doi-asserted-by":"publisher","unstructured":"Mogan, J.N., Lee, C.P., Lim, K.M., Muthu, K.S.: Gait-ViT: gait recognition with vision transformer. Sensors 22(19) (2022). https:\/\/doi.org\/10.3390\/s22197362","DOI":"10.3390\/s22197362"},{"issue":"1","key":"5_CR37","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1109\/TBIOM.2020.3031470","volume":"3","author":"A Moghaddam","year":"2020","unstructured":"Moghaddam, A., Etemad, A.: View-invariant gait recognition with attentive recurrent learning of partial representations. IEEE Trans. Biometrics Behav. Identity Sci. 3(1), 124\u2013137 (2020). https:\/\/doi.org\/10.1109\/TBIOM.2020.3031470","journal-title":"IEEE Trans. Biometrics Behav. Identity Sci."},{"key":"5_CR38","doi-asserted-by":"publisher","unstructured":"Pin\u010di\u0107, D., Su\u0161anj, D., Lenac, K.: Gait recognition with self-supervised learning of gait features based on vision transformers. Sensors 22(19) (2022). https:\/\/doi.org\/10.3390\/s22197140","DOI":"10.3390\/s22197140"},{"key":"5_CR39","unstructured":"Sabour, S., Frosst, N., Hinton, G.E.: Dynamic routing between capsules. Adv. Neural Inform. Process. Syst. 30 (2017)"},{"key":"5_CR40","doi-asserted-by":"publisher","unstructured":"Sayeed, M.S., Min, P.P., Bari, M.A.: Deep learning based gait recognition using convolutional neural network in the COVID-19 pandemic. Emerg. Sci. J. 6(5), 1086\u20131099 (2022). https:\/\/doi.org\/10.28991\/ESJ-2022-06-05-012","DOI":"10.28991\/ESJ-2022-06-05-012"},{"key":"5_CR41","unstructured":"Sepas-Moghaddam, A., Ghorbani, S., Troje, N.F., Etemad, A.: Gait recognition using multi-scale partial representation transformation with capsules (2020)"},{"key":"5_CR42","doi-asserted-by":"publisher","unstructured":"Sharma, P., Khan, A., Jain, S., Srivastava, A.: Artificial intelligence driven human identification. J. Inform. Technol. Manag. 15(3), 113\u2013133 (2023). https:\/\/doi.org\/10.22059\/JITM.2023.93629","DOI":"10.22059\/JITM.2023.93629"},{"key":"5_CR43","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/ICB.2016.7550060","volume":"2016","author":"K Shiraga","year":"2016","unstructured":"Shiraga, K., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: Geinet: view-invariant gait recognition using a convolutional neural network. Int. Conf. Biometrics (ICB) 2016, 1\u20138 (2016). https:\/\/doi.org\/10.1109\/ICB.2016.7550060","journal-title":"Int. Conf. Biometrics (ICB)"},{"key":"5_CR44","doi-asserted-by":"publisher","first-page":"646","DOI":"10.1016\/j.patrec.2019.07.008","volume":"125","author":"D Thapar","year":"2019","unstructured":"Thapar, D., Jaswal, G., Nigam, A., Arora, C.: Gait metric learning Siamese network exploiting dual of spatio-temporal 3D-CNN intra and LSTM based inter gait-cycle-segment features. Pattern Recogn. Lett. 125, 646\u2013653 (2019). https:\/\/doi.org\/10.1016\/j.patrec.2019.07.008","journal-title":"Pattern Recogn. Lett."},{"key":"5_CR45","doi-asserted-by":"publisher","unstructured":"Upadhyay, J., Gonsalves, T.: An enhanced gait recognition system based on the features fusion methodology with recurrent neural network (RNN). Indian J. Comput. Sci. Eng. 13(5), 1483\u20131496 (2022). https:\/\/doi.org\/10.21817\/indjcse\/2022\/v13i5\/221305042","DOI":"10.21817\/indjcse\/2022\/v13i5\/221305042"},{"issue":"2","key":"5_CR46","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1109\/TBIOM.2023.3244206","volume":"5","author":"R Wang","year":"2023","unstructured":"Wang, R., Shi, Y., Ling, H., Li, Z., Li, P., Liu, B., Zheng, H., Wang, Q.: Gait recognition via gait period set. IEEE Trans. Biometrics Behav. Identity Sci. 5(2), 183\u2013195 (2023). https:\/\/doi.org\/10.1109\/TBIOM.2023.3244206","journal-title":"IEEE Trans. Biometrics Behav. Identity Sci."},{"issue":"20","key":"5_CR47","doi-asserted-by":"publisher","first-page":"29459","DOI":"10.1007\/s11042-022-12831-1","volume":"81","author":"X Wang","year":"2022","unstructured":"Wang, X., Hu, S.: Visual gait recognition based on convolutional block attention network. Multimedia Tools Appl. 81(20), 29459\u201329476 (2022). https:\/\/doi.org\/10.1007\/s11042-022-12831-1","journal-title":"Multimedia Tools Appl."},{"key":"5_CR48","doi-asserted-by":"publisher","unstructured":"Wang, X., Yan, W.Q.: Human gait recognition based on frame-by-frame gait energy images and convolutional long short-term memory. Int. J. Neural Syst. 30(1) (2019). https:\/\/doi.org\/10.1142\/S0129065719500278","DOI":"10.1142\/S0129065719500278"},{"issue":"3\u20134","key":"5_CR49","doi-asserted-by":"publisher","first-page":"2917","DOI":"10.1007\/s11042-019-08509-w","volume":"79","author":"X Wang","year":"2019","unstructured":"Wang, X., Zhang, J.: Gait feature extraction and gait classification using two-branch CNN. Multimedia Tools Appl. 79(3\u20134), 2917\u20132930 (2019). https:\/\/doi.org\/10.1007\/s11042-019-08509-w","journal-title":"Multimedia Tools Appl."},{"issue":"31\u201332","key":"5_CR50","doi-asserted-by":"publisher","first-page":"22653","DOI":"10.1007\/s11042-020-09003-4","volume":"79","author":"Y Wang","year":"2020","unstructured":"Wang, Y., Chen, Z., Wu, Q., Rong, X.: Deep mutual learning network for gait recognition. Multimedia Tools Appl. 79(31\u201332), 22653\u201322672 (2020). https:\/\/doi.org\/10.1007\/s11042-020-09003-4","journal-title":"Multimedia Tools Appl."},{"key":"5_CR51","doi-asserted-by":"publisher","first-page":"245","DOI":"10.1016\/j.neucom.2019.02.025","volume":"339","author":"Y Wang","year":"2019","unstructured":"Wang, Y., Song, C., Huang, Y., Wang, Z., Wang, L.: Learning view invariant gait features with Two-Stream GAN. Neurocomputing 339, 245\u2013254 (2019). https:\/\/doi.org\/10.1016\/j.neucom.2019.02.025","journal-title":"Neurocomputing"},{"issue":"19","key":"5_CR52","doi-asserted-by":"publisher","first-page":"28777","DOI":"10.1007\/s11042-021-11107-4","volume":"80","author":"J Wen","year":"2021","unstructured":"Wen, J., Wang, X.: Cross-view gait recognition based on residual long short-term memory. Multimedia Tools Appl. 80(19), 28777\u201328788 (2021). https:\/\/doi.org\/10.1007\/s11042-021-11107-4","journal-title":"Multimedia Tools Appl."},{"issue":"12","key":"5_CR53","doi-asserted-by":"publisher","first-page":"2761","DOI":"10.1049\/ipr2.12260","volume":"15","author":"J Wen","year":"2021","unstructured":"Wen, J., Wang, X.: Gait recognition based on sparse linear subspace. IET Image Process. 15(12), 2761\u20132769 (2021). https:\/\/doi.org\/10.1049\/ipr2.12260","journal-title":"IET Image Process."},{"issue":"24","key":"5_CR54","doi-asserted-by":"publisher","first-page":"35789","DOI":"10.1007\/s11042-019-08153-4","volume":"78","author":"C Wu","year":"2019","unstructured":"Wu, C., Song, Y., Zhang, Y.: Multi-view gait recognition using NMF and 2DLDA. Multimedia Tools Appl. 78(24), 35789\u201335811 (2019). https:\/\/doi.org\/10.1007\/s11042-019-08153-4","journal-title":"Multimedia Tools Appl."},{"key":"5_CR55","doi-asserted-by":"publisher","first-page":"2734","DOI":"10.1109\/TIP.2020.3039888","volume":"30","author":"H Wu","year":"2020","unstructured":"Wu, H., Tian, J., Fu, Y., Li, B., Li, X.: Condition-aware comparison scheme for gait recognition. IEEE Trans. Image Process. 30, 2734\u20132744 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.3039888","journal-title":"IEEE Trans. Image Process."},{"key":"5_CR56","doi-asserted-by":"publisher","first-page":"3265","DOI":"10.1109\/TMM.2021.3095809","volume":"24","author":"K Xu","year":"2021","unstructured":"Xu, K., Jiang, X., Sun, T.: Gait recognition based on local graphical skeleton descriptor with pairwise similarity network. IEEE Trans. Multimedia 24, 3265\u20133275 (2021). https:\/\/doi.org\/10.1109\/TMM.2021.3095809","journal-title":"IEEE Trans. Multimedia"},{"key":"5_CR57","doi-asserted-by":"publisher","first-page":"159","DOI":"10.1016\/j.jvcir.2019.01.023","volume":"59","author":"Z Xu","year":"2019","unstructured":"Xu, Z., Lu, W., Zhang, Q., Yeung, Y., Chen, X.: Gait recognition based on capsule network. J. Vis. Commun. Image Represent. 59, 159\u2013167 (2019). https:\/\/doi.org\/10.1016\/j.jvcir.2019.01.023","journal-title":"J. Vis. Commun. Image Represent."},{"key":"5_CR58","doi-asserted-by":"publisher","first-page":"95","DOI":"10.1016\/j.neucom.2019.11.015","volume":"380","author":"W Xue","year":"2019","unstructured":"Xue, W., Ai, H., Sun, T., Song, C., Huang, Y., Wang, L.: Frame-GAN: increasing the frame rate of gait videos with generative adversarial networks. Neurocomputing 380, 95\u2013104 (2019). https:\/\/doi.org\/10.1016\/j.neucom.2019.11.015","journal-title":"Neurocomputing"},{"key":"5_CR59","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2022.3171961","author":"L Yao","year":"2022","unstructured":"Yao, L., Kusakunniran, W., Zhang, P., Wu, Q., Zhang, J.: Improving disentangled representation learning for gait recognition using group supervision. IEEE Trans. Multimedia (2022). https:\/\/doi.org\/10.1109\/TMM.2022.3171961","journal-title":"IEEE Trans. Multimedia"},{"key":"5_CR60","doi-asserted-by":"publisher","unstructured":"Yeo, S.-S., Rho, S., Kim, H., Safdar, J., Zia, U., Durrani, M.Y.: A triplet-branch convolutional neural network for part-based gait recognition. Comput. Syst. Sci. Eng. 47(2), 2027\u20132047 (2023). https:\/\/doi.org\/10.32604\/csse.2023.040327","DOI":"10.32604\/csse.2023.040327"},{"key":"5_CR61","doi-asserted-by":"publisher","unstructured":"Yu, Y., Si, X., Hu, C., Zhang, J.: A review of recurrent neural networks: LSTM cells and network architectures. Neural Comput. 31(7), 1235\u20131270 (2019). https:\/\/doi.org\/10.1162\/neco_a_01199","DOI":"10.1162\/neco_a_01199"},{"key":"5_CR62","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/TIP.2019.2926208","volume":"29","author":"Y Zhang","year":"2019","unstructured":"Zhang, Y., Huang, Y., Yu, S., Wang, L.: Cross-view gait recognition by discriminative feature learning. IEEE Trans. Image Process. 29, 1001\u20131015 (2019). https:\/\/doi.org\/10.1109\/TIP.2019.2926208","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"5_CR63","doi-asserted-by":"publisher","first-page":"97","DOI":"10.1049\/csy2.12052","volume":"4","author":"L Zheng","year":"2022","unstructured":"Zheng, L., Zha, Y., Kong, D., Yang, H., Zhang, Y.: Multi-branch angle aware spatial temporal graph convolutional neural network for model-based gait recognition. IET Cyber-Syst. Robot. 4(2), 97\u2013106 (2022). https:\/\/doi.org\/10.1049\/csy2.12052","journal-title":"IET Cyber-Syst. Robot."},{"key":"5_CR64","doi-asserted-by":"publisher","unstructured":"Zhou, J., Cui, G., Hu, S., Zhang, Z., Yang, C., Liu, Z., Wang, L., Li, C., Sun, M.: Graph neural networks: a review of methods and applications. AI Open 1, 57\u201381 (2020). https:\/\/doi.org\/10.1016\/j.aiopen.2021.01.001","DOI":"10.1016\/j.aiopen.2021.01.001"}],"container-title":["Studies in Systems, Decision and Control","Advances in Gait-Based Identification"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-89560-9_5","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,30]],"date-time":"2025-05-30T14:12:42Z","timestamp":1748614362000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-89560-9_5"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025]]},"ISBN":["9783031895593","9783031895609"],"references-count":64,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-89560-9_5","relation":{},"ISSN":["2198-4182","2198-4190"],"issn-type":[{"value":"2198-4182","type":"print"},{"value":"2198-4190","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025]]},"assertion":[{"value":"31 May 2025","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}}]}}