{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,9]],"date-time":"2026-03-09T10:15:22Z","timestamp":1773051322799,"version":"3.50.1"},"reference-count":31,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,12,11]],"date-time":"2023-12-11T00:00:00Z","timestamp":1702252800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2023,12,11]],"date-time":"2023-12-11T00:00:00Z","timestamp":1702252800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["52074064"],"award-info":[{"award-number":["52074064"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62173073"],"award-info":[{"award-number":["62173073"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"the National Key Research and Development Program of China","award":["2020AAA0109200"],"award-info":[{"award-number":["2020AAA0109200"]}]},{"name":"the Natural Science Foundation of Science and Technology Department of Liaoning Province","award":["2021-BS-054"],"award-info":[{"award-number":["2021-BS-054"]}]},{"name":"the Fundamental Research Funds for the Central Universities, China","award":["N2204006"],"award-info":[{"award-number":["N2204006"]}]},{"name":"the Fundamental Research Funds for the Central Universities, China","award":["N2104026"],"award-info":[{"award-number":["N2104026"]}]},{"name":"the Fundamental Research Funds for the Central Universities, China","award":["N2018008"],"award-info":[{"award-number":["N2018008"]}]},{"name":"the Fundamental Research Funds for the Central Universities, China","award":["N2001002"],"award-info":[{"award-number":["N2001002"]}]},{"DOI":"10.13039\/501100018617","name":"Liaoning Revitalization Talents Program","doi-asserted-by":"publisher","award":["XLYC2008020"],"award-info":[{"award-number":["XLYC2008020"]}],"id":[{"id":"10.13039\/501100018617","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Complex Intell. Syst."],"published-print":{"date-parts":[[2024,4]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Despite the ability of 3D convolutional methods to extract spatio-temporal information simultaneously, they also increase parameter redundancy and computational and storage costs. Previous work that has utilized the 2D convolution method has approached the problem in one of two ways: either using the entire body sequence as input to extract global features or dividing the body sequence into several parts to extract local features. However, global information tends to overlook detailed information specific to each body part, while local information fails to capture relationships between local regions. Therefore, this study proposes a new framework for constructing spatio-temporal representations, which involves extracting and fusing features in a novel manner. To achieve this, we introduce the multi-feature extraction-fusion (MFEF) module, which includes two branches: each branch extracts global features or local features individually, after which they are fused using multiple strategies. Additionally, as gait is a periodic action and different body parts contribute unequally to recognition during each cycle, we propose the periodic temporal feature modeling (PTFM) module, which extracts temporal features from adjacent frame parts during the complete gait cycle, based on the fused features. Furthermore, to capture fine-grained information specific to each body part, our framework utilizes multiple parallel PTFMs to correspond with each body part. We conducted a comprehensive experimental study on the widely used public dataset CASIA-B. Results indicate that the proposed approach achieved an average rank-1 accuracy of 97.2% in normal walking conditions, 92.3% while carrying a bag during walking, and 80.5% while wearing a jacket during walking.<\/jats:p>","DOI":"10.1007\/s40747-023-01293-z","type":"journal-article","created":{"date-parts":[[2023,12,11]],"date-time":"2023-12-11T07:02:16Z","timestamp":1702278136000},"page":"2673-2688","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Gait recognition based on multi-feature representation and temporal modeling of periodic parts"],"prefix":"10.1007","volume":"10","author":[{"given":"Zhenni","family":"Li","sequence":"first","affiliation":[]},{"given":"Shiqiang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Dong","family":"Xiao","sequence":"additional","affiliation":[]},{"given":"Zhengmin","family":"Gu","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Yu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,11]]},"reference":[{"key":"1293_CR1","first-page":"1","volume":"27","author":"Y Sun","year":"2014","unstructured":"Sun Y, Chen Y, Wang X, Tang X (2014) Deep learning face representation by joint identification\u2013verification. Adv Neural Inf Process Syst 27:1","journal-title":"Adv Neural Inf Process Syst"},{"issue":"5\/6","key":"1293_CR2","doi-asserted-by":"publisher","first-page":"1063","DOI":"10.1108\/03684920410534128","volume":"33","author":"AM Andrew","year":"2004","unstructured":"Andrew AM (2004) Handbook of fingerprint recognition. Kybernetes 33(5\/6):1063\u20131064","journal-title":"Kybernetes"},{"issue":"9","key":"1293_CR3","doi-asserted-by":"publisher","first-page":"1348","DOI":"10.1109\/5.628669","volume":"85","author":"RP Wildes","year":"1997","unstructured":"Wildes RP (1997) Iris recognition: an emerging biometric technology. Proc IEEE 85(9):1348\u20131363","journal-title":"Proc IEEE"},{"key":"1293_CR4","doi-asserted-by":"crossref","unstructured":"Li X, Makihara Y, Xu C, Yagi Y, Yu S, Ren M (2020) End-to-end model-based gait recognition. In: Proceedings of the Asian conference on computer vision","DOI":"10.1007\/978-3-030-69535-4_1"},{"issue":"2","key":"1293_CR5","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1109\/TPAMI.2006.38","volume":"28","author":"J Han","year":"2005","unstructured":"Han J, Bhanu B (2005) Individual recognition using gait energy image. IEEE Trans Pattern Anal Mach Intell 28(2):316\u2013322","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"1293_CR6","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/TPAMI.2016.2545669","volume":"39","author":"Z Wu","year":"2016","unstructured":"Wu Z, Huang Y, Wang L, Wang X, Tan T (2016) A comprehensive study on cross-view gait based human identification with deep CNNs. IEEE Trans Pattern Anal Mach Intell 39(2):209\u2013226","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1293_CR7","doi-asserted-by":"crossref","unstructured":"Wolf T, Babaee M, Rigoll G (2016) Multi-view gait recognition using 3D convolutional neural networks. In: 2016 IEEE international conference on image processing (ICIP). IEEE, pp 4165\u20134169","DOI":"10.1109\/ICIP.2016.7533144"},{"key":"1293_CR8","doi-asserted-by":"crossref","unstructured":"Chao H, He Y, Zhang J, Feng J (2019) Gaitset: regarding gait as a set for cross-view gait recognition. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 8126\u20138133","DOI":"10.1609\/aaai.v33i01.33018126"},{"key":"1293_CR9","doi-asserted-by":"crossref","unstructured":"Hou S, Cao C, Liu X, Huang Y (2020) Gait lateral network: learning discriminative and compact representations for gait recognition. In: Computer vision\u2014ECCV 2020: 16th European conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part IX. Springer, London, pp 382\u2013398","DOI":"10.1007\/978-3-030-58545-7_22"},{"key":"1293_CR10","doi-asserted-by":"publisher","first-page":"57583","DOI":"10.1109\/ACCESS.2018.2874073","volume":"6","author":"S Tong","year":"2018","unstructured":"Tong S, Fu Y, Yue X, Ling H (2018) Multi-view gait recognition based on a spatial\u2013temporal deep neural network. IEEE Access 6:57583\u201357596","journal-title":"IEEE Access"},{"key":"1293_CR11","doi-asserted-by":"publisher","first-page":"3041","DOI":"10.1109\/TIP.2021.3055936","volume":"30","author":"X Chen","year":"2021","unstructured":"Chen X, Luo X, Weng J, Luo W, Li H, Tian Q (2021) Multi-view gait image generation for cross-view gait recognition. IEEE Trans Image Process 30:3041\u20133055","journal-title":"IEEE Trans Image Process"},{"key":"1293_CR12","doi-asserted-by":"crossref","unstructured":"Fu Y, Wei Y, Zhou Y, Shi H, Huang G, Wang X, Yao Z, Huang T (2019) Horizontal pyramid matching for person re-identification. In: Proceedings of the AAAI conference on artificial intelligence, vol 33, pp 8295\u20138302","DOI":"10.1609\/aaai.v33i01.33018295"},{"key":"1293_CR13","doi-asserted-by":"crossref","unstructured":"Sun Y, Zheng L, Yang Y, Tian Q, Wang S (2018) Beyond part models: person retrieval with refined part pooling (and a strong convolutional baseline). In: Proceedings of the European conference on computer vision (ECCV), pp 480\u2013496","DOI":"10.1007\/978-3-030-01225-0_30"},{"key":"1293_CR14","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/TIP.2019.2926208","volume":"29","author":"Y Zhang","year":"2019","unstructured":"Zhang Y, Huang Y, Yu S, Wang L (2019) Cross-view gait recognition by discriminative feature learning. IEEE Trans Image Process 29:1001\u20131015","journal-title":"IEEE Trans Image Process"},{"key":"1293_CR15","doi-asserted-by":"crossref","unstructured":"Fan C, Peng Y, Cao C, Liu X, Hou S, Chi J, Huang Y, Li Q, He Z (2020) Gaitpart: temporal part-based model for gait recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14225\u201314233","DOI":"10.1109\/CVPR42600.2020.01423"},{"key":"1293_CR16","unstructured":"Yu S, Tan D, Tan T (2006) A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: 18th international conference on pattern recognition (ICPR\u201906), vol 4. IEEE, London, pp 441\u2013444"},{"key":"1293_CR17","doi-asserted-by":"crossref","unstructured":"Liao R, Cao C, Garcia EB, Yu S, Huang Y (2017) Pose-based temporal\u2013spatial network (PTSN) for gait recognition with carrying and clothing variations. In: Biometric recognition: 12th Chinese conference, CCBR 2017, Shenzhen, China, October 28\u201329, 2017, proceedings 12. Springer, London, pp 474\u2013483","DOI":"10.1007\/978-3-319-69923-3_51"},{"key":"1293_CR18","doi-asserted-by":"crossref","unstructured":"Teepe T, Khan A, Gilg J, Herzog F, H\u00f6rmann S, Rigoll G (2021) Gaitgraph: graph convolutional network for skeleton-based gait recognition. In: 2021 IEEE international conference on image processing (ICIP). IEEE, London, pp 2314\u20132318","DOI":"10.1109\/ICIP42928.2021.9506717"},{"key":"1293_CR19","doi-asserted-by":"crossref","unstructured":"Zhang K, Luo W, Ma L, Liu W, Li H (2019) Learning joint gait representation via quintuplet loss minimization. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 4700\u20134709","DOI":"10.1109\/CVPR.2019.00483"},{"key":"1293_CR20","doi-asserted-by":"publisher","first-page":"457","DOI":"10.1007\/s12021-018-9362-4","volume":"16","author":"W Liu","year":"2018","unstructured":"Liu W, Zhang C, Ma H, Li S (2018) Learning efficient spatial\u2013temporal gait features with deep learning for human identification. Neuroinformatics 16:457\u2013471","journal-title":"Neuroinformatics"},{"issue":"7","key":"1293_CR21","first-page":"3467","volume":"44","author":"H Chao","year":"2021","unstructured":"Chao H, Wang K, He Y, Zhang J, Feng J (2021) Gaitset: cross-view gait recognition through utilizing gait as a deep set. IEEE Trans Pattern Anal Mach Intell 44(7):3467\u20133478","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"1293_CR22","doi-asserted-by":"crossref","unstructured":"Shiraga K, Makihara Y, Muramatsu D, Echigo T, Yagi Y (2016) Geinet: view-invariant gait recognition using a convolutional neural network. In: 2016 international conference on biometrics (ICB). IEEE, pp 1\u20138","DOI":"10.1109\/ICB.2016.7550060"},{"key":"1293_CR23","doi-asserted-by":"publisher","first-page":"2734","DOI":"10.1109\/TIP.2020.3039888","volume":"30","author":"H Wu","year":"2020","unstructured":"Wu H, Tian J, Fu Y, Li B, Li X (2020) Condition-aware comparison scheme for gait recognition. IEEE Trans Image Process 30:2734\u20132744","journal-title":"IEEE Trans Image Process"},{"key":"1293_CR24","first-page":"1","volume":"2022","author":"L Yao","year":"2022","unstructured":"Yao L, Kusakunniran W, Zhang P, Wu Q, Zhang J (2022) Improving disentangled representation learning for gait recognition using group supervision. IEEE Trans Multimed 2022:1","journal-title":"IEEE Trans Multimed"},{"key":"1293_CR25","doi-asserted-by":"crossref","unstructured":"Rokanujjaman M, Hossain MA, Islam MR (2012) Effective part selection for part-based gait identification. In: 2012 7th international conference on electrical and computer engineering. IEEE, pp 17\u201319","DOI":"10.1109\/ICECE.2012.6471473"},{"key":"1293_CR26","doi-asserted-by":"publisher","first-page":"75381","DOI":"10.1109\/ACCESS.2020.2986554","volume":"8","author":"G Huang","year":"2020","unstructured":"Huang G, Lu Z, Pun C-M, Cheng L (2020) Flexible gait recognition based on flow regulation of local features between key frames. IEEE Access 8:75381\u201375392","journal-title":"IEEE Access"},{"key":"1293_CR27","doi-asserted-by":"crossref","unstructured":"Kusakunniran W, Wu Q, Zhang J, Li H (2010) Support vector regression for multi-view gait recognition based on local motion feature selection. In: 2010 IEEE computer society conference on computer vision and pattern recognition. IEEE, pp 974\u2013981","DOI":"10.1109\/CVPR.2010.5540113"},{"key":"1293_CR28","doi-asserted-by":"crossref","unstructured":"An W, Liao R, Yu S, Huang Y, Yuen PC (2018) Improving gait recognition with 3D pose estimation. In: Biometric recognition: 13th Chinese conference, CCBR 2018, Urumqi, China, August 11\u201312, 2018, proceedings 13. Springer, London, pp 137\u2013147","DOI":"10.1007\/978-3-319-97909-0_15"},{"key":"1293_CR29","doi-asserted-by":"publisher","first-page":"63164","DOI":"10.1109\/ACCESS.2018.2876890","volume":"6","author":"G Batchuluun","year":"2018","unstructured":"Batchuluun G, Yoon HS, Kang JK, Park KR (2018) Gait-based human identification by combining shallow convolutional neural network-stacked long short-term memory and deep convolutional neural network. IEEE Access 6:63164\u201363186","journal-title":"IEEE Access"},{"key":"1293_CR30","doi-asserted-by":"crossref","unstructured":"Takemura N, Makihara Y, Muramatsu D, Echigo T, Yagi Y (2018) Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition. IPSJ Trans Comput Vis Appl 10:1\u201314","DOI":"10.1186\/s41074-018-0039-6"},{"key":"1293_CR31","unstructured":"Hermans A, Beyer L, Leibe B (2017) In defense of the triplet loss for person re-identification. arXiv Preprint arXiv:1703.07737"}],"container-title":["Complex &amp; Intelligent Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01293-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s40747-023-01293-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s40747-023-01293-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,5]],"date-time":"2024-11-05T18:35:39Z","timestamp":1730831739000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s40747-023-01293-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,12,11]]},"references-count":31,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["1293"],"URL":"https:\/\/doi.org\/10.1007\/s40747-023-01293-z","relation":{},"ISSN":["2199-4536","2198-6053"],"issn-type":[{"value":"2199-4536","type":"print"},{"value":"2198-6053","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,12,11]]},"assertion":[{"value":"5 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 November 2023","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 December 2023","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}