{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T17:46:06Z","timestamp":1772905566865,"version":"3.50.1"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2024,2,1]],"date-time":"2024-02-01T00:00:00Z","timestamp":1706745600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,2,1]],"date-time":"2024-02-01T00:00:00Z","timestamp":1706745600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the STI 2030-Major Projects","award":["2022ZD0208900"],"award-info":[{"award-number":["2022ZD0208900"]}]},{"DOI":"10.13039\/501100001809","name":"the National Natural Science Foundation of China","doi-asserted-by":"crossref","award":["62076103"],"award-info":[{"award-number":["62076103"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"crossref"}]},{"name":"the Special Innovation Projects of Colleges and Universities in Guangdong Province","award":["2022KTSCX035"],"award-info":[{"award-number":["2022KTSCX035"]}]},{"name":"the Special Funds for the Cultivation of Guangdong College Students\u2019 Scientific and Technological Innovation","award":["pdjh2022a0125"],"award-info":[{"award-number":["pdjh2022a0125"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2024,2]]},"DOI":"10.1007\/s10489-023-05241-9","type":"journal-article","created":{"date-parts":[[2024,2,6]],"date-time":"2024-02-06T05:29:41Z","timestamp":1707197381000},"page":"2428-2444","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["GaitCTCG: cross-view gait recognition via cascaded residual temporal shift and comprehensive multi-granularity learning"],"prefix":"10.1007","volume":"54","author":[{"given":"Binyuan","family":"Huang","sequence":"first","affiliation":[]},{"given":"Chengju","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Lewei","family":"He","sequence":"additional","affiliation":[]},{"given":"Chi","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Jiahui","family":"Pan","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,2,6]]},"reference":[{"key":"5241_CR1","doi-asserted-by":"crossref","unstructured":"Premalatha G, Chandramani VP (2020) Improved gait recognition through gait energy image partitioning. Comput Intell 36(3):1261\u20131274","DOI":"10.1111\/coin.12340"},{"issue":"5","key":"5241_CR2","doi-asserted-by":"publisher","first-page":"1294","DOI":"10.1111\/1556-4029.14036","volume":"64","author":"I Macoveciuc","year":"2019","unstructured":"Macoveciuc I, Rando CJ, Borrion H (2019) Forensic gait analysis and recognition: standards of evidence admissibility. J Forensic Sci 64(5):1294\u20131303","journal-title":"J Forensic Sci"},{"key":"5241_CR3","doi-asserted-by":"crossref","unstructured":"Zhang Z, Tran L, Yin X, Atoum Y, Liu X, Wan J, Wang N (2019) Gait recognition via disentangled representation learning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 4710\u20134719","DOI":"10.1109\/CVPR.2019.00484"},{"key":"5241_CR4","doi-asserted-by":"crossref","unstructured":"Fan C, Peng Y, Cao C, Liu X, Hou S, Chi J, Huang Y, Li Q, He Z (2020) Gaitpart: temporal part\u2013based model for gait recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14225\u201314233","DOI":"10.1109\/CVPR42600.2020.01423"},{"key":"5241_CR5","doi-asserted-by":"crossref","unstructured":"Lin B, Zhang S, Bao F (2020) Gait recognition with multipletemporal- scale 3d convolutional neural network. In: Proceedings of the 28th ACM international conference on multimedia, pp 3054\u20133062","DOI":"10.1145\/3394171.3413861"},{"key":"5241_CR6","doi-asserted-by":"crossref","unstructured":"Chao H, He Y, Zhang J, Feng J (2019) Gaitset: regarding gait as a set for cross-view gait recognition. Proceedings of the AAAI Conference on Artificial Intelligence 33:8126\u20138133","DOI":"10.1609\/aaai.v33i01.33018126"},{"key":"5241_CR7","doi-asserted-by":"crossref","unstructured":"Hou S, Cao C, Liu X, Huang Y (2020) In: European conference on computer vision, pp 382\u2013398","DOI":"10.1007\/978-3-030-58545-7_22"},{"issue":"3","key":"5241_CR8","doi-asserted-by":"publisher","first-page":"384","DOI":"10.1109\/TBIOM.2021.3074963","volume":"3","author":"S Hou","year":"2021","unstructured":"Hou S, Liu X, Cao C, Huang Y (2021) Set residual network for silhouette-based gait recognition. IEEE Transactions on Biometrics, Behavior, and Identity Science 3(3):384\u2013393","journal-title":"IEEE Transactions on Biometrics, Behavior, and Identity Science"},{"issue":"1","key":"5241_CR9","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1109\/TBIOM.2020.3031470","volume":"3","author":"A Sepas-Moghaddam","year":"2020","unstructured":"Sepas-Moghaddam A, Etemad A (2020) View-invariant gait recognition with attentive recurrent learning of partial representations. IEEE Transactions on Biometrics, Behavior, and Identity Science 3(1):124\u2013137","journal-title":"IEEE Transactions on Biometrics, Behavior, and Identity Science"},{"key":"5241_CR10","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/TIP.2019.2926208","volume":"29","author":"Y Zhang","year":"2019","unstructured":"Zhang Y, Huang Y, Yu S, Wang L (2019) Cross-view gait recognition by discriminative feature learning. IEEE Trans Image Process 29:1001\u20131015","journal-title":"IEEE Trans Image Process"},{"key":"5241_CR11","doi-asserted-by":"crossref","unstructured":"Lin B, Zhang S, Yu X (2021) Gait recognition via effective global\u2013local feature representation and local temporal aggregation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 14648\u201314656","DOI":"10.1109\/ICCV48922.2021.01438"},{"key":"5241_CR12","doi-asserted-by":"crossref","unstructured":"Li X, Makihara Y, Xu C, Yagi Y, Ren M (2020) Gait recognition via semi\u2013supervised disentangled representation learning to identity and covariate features. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 13309\u201313319","DOI":"10.1109\/CVPR42600.2020.01332"},{"issue":"1","key":"5241_CR13","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1109\/TIFS.2018.2844819","volume":"14","author":"Y He","year":"2018","unstructured":"He Y, Zhang J, Shan H, Wang L (2018) Multi-task gans for viewspecific feature learning in gait recognition. IEEE Trans Inf Forensics Secur 14(1):102\u2013113","journal-title":"IEEE Trans Inf Forensics Secur"},{"issue":"2","key":"5241_CR14","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1109\/TPAMI.2006.38","volume":"28","author":"J Han","year":"2005","unstructured":"Han J, Bhanu B (2005) Individual recognition using gait energy image. IEEE Trans Pattern Anal Mach Intell 28(2):316\u2013322","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"5241_CR15","doi-asserted-by":"crossref","unstructured":"Lin J, Gan C, Han S (2019) Tsm: temporal shift module for efficient video understanding. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 7083\u20137093","DOI":"10.1109\/ICCV.2019.00718"},{"issue":"2","key":"5241_CR16","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/TPAMI.2016.2545669","volume":"39","author":"Z Wu","year":"2016","unstructured":"Wu Z, Huang Y, Wang L, Wang X, Tan T (2016) A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans Pattern Anal Mach Intell 39(2):209\u2013226","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"5241_CR17","doi-asserted-by":"crossref","unstructured":"Shiraga K, Makihara Y, Muramatsu D, Echigo T, Yagi Y (2016) Geinet: View\u2013invariant gait recognition using a convolutional neural network. In: 2016 International conference on biometrics (ICB), pp 1\u20138","DOI":"10.1109\/ICB.2016.7550060"},{"key":"5241_CR18","doi-asserted-by":"crossref","unstructured":"Xiao J, Yang H, Xie K, Zhu J, Zhang J (2021) Learning discriminative representation with global and fine-grained features for cross\u2013view gait recognition. CAAI Transactions on Intelligence Technology","DOI":"10.1049\/cit2.12051"},{"key":"5241_CR19","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.neucom.2017.02.006","volume":"239","author":"S Yu","year":"2017","unstructured":"Yu S, Chen H, Wang Q, Shen L, Huang Y (2017) Invariant feature extraction for gait recognition using only one uniform model. Neurocomputing 239:81\u201393","journal-title":"Neurocomputing"},{"key":"5241_CR20","unstructured":"Hu B, Gao Y, Guan Y, Long Y, Lane N, Ploetz T (2018) Robust cross-view gait identification with evidence: a discriminant gait gan (diggan) approach on 10000 people"},{"issue":"11","key":"5241_CR21","doi-asserted-by":"publisher","first-page":"8149","DOI":"10.1007\/s10489-021-02322-5","volume":"51","author":"W Xu","year":"2021","unstructured":"Xu W (2021) Graph-optimized coupled discriminant projections for cross-view gait recognition. Appl Intell 51(11):8149\u20138161","journal-title":"Appl Intell"},{"issue":"2","key":"5241_CR22","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1109\/TPAMI.2006.38","volume":"28","author":"J Han","year":"2005","unstructured":"Han J, Bhanu B (2005) Individual recognition using gait energy image. IEEE Trans Pattern Anal Mach Intell 28(2):316\u2013322","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"5241_CR23","doi-asserted-by":"publisher","first-page":"2023","DOI":"10.1007\/s10489-021-02484-2","volume":"52","author":"L Zhao","year":"2022","unstructured":"Zhao L, Guo L, Zhang R, Xie X, Ye X (2022) Mmgaitset: multimodal based gait recognition for countering carrying and clothing changes. Appl Intell 52(2):2023\u20132036","journal-title":"Appl Intell"},{"key":"5241_CR24","doi-asserted-by":"crossref","unstructured":"Wolf T, Babaee M, Rigoll G (2016) Multi\u2013view gait recognition using 3d convolutional neural networks. In: 2016 IEEE International conference on image processing (ICIP), pp 4165\u20134169","DOI":"10.1109\/ICIP.2016.7533144"},{"key":"5241_CR25","doi-asserted-by":"crossref","unstructured":"Liu Y, Zeng Y, Pu J, Shan H, He P, Zhang J (2021) Selfgait: A spatiotemporal representation learning method for self-supervised gait recognition. In: ICASSP 2021\u20132021 IEEE international conference on acoustics, speech and signal processing (ICASSP), pp 2570\u20132574","DOI":"10.1109\/ICASSP39728.2021.9413894"},{"key":"5241_CR26","doi-asserted-by":"publisher","first-page":"2734","DOI":"10.1109\/TIP.2020.3039888","volume":"30","author":"H Wu","year":"2020","unstructured":"Wu H, Tian J, Fu Y, Li B, Li X (2020) Condition-aware comparison scheme for gait recognition. IEEE Trans Image Process 30:2734\u20132744","journal-title":"IEEE Trans Image Process"},{"key":"5241_CR27","doi-asserted-by":"crossref","unstructured":"Wang Z, She Q, Smolic A (2021) Action-net: multipath excitation for action recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 13214\u201313223","DOI":"10.1109\/CVPR46437.2021.01301"},{"key":"5241_CR28","doi-asserted-by":"crossref","unstructured":"Sudhakaran S, Escalera S, Lanz O (2020) Gate\u2013shift networks for video action recognition. In: Proceedings of the IEEE\/CVF Conference on computer vision and pattern recognition, pp 1102\u20131111","DOI":"10.1109\/CVPR42600.2020.00118"},{"key":"5241_CR29","doi-asserted-by":"crossref","unstructured":"Gao S-H, Cheng M-M, Zhao K, Zhang X-Y, Yang M-H, Torr P (2019) Res2net: a new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell 43(2):652\u2013662","DOI":"10.1109\/TPAMI.2019.2938758"},{"key":"5241_CR30","doi-asserted-by":"crossref","unstructured":"Zhao T, Wu X (2019) Pyramid feature attention network for saliency detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3085\u20133094","DOI":"10.1109\/CVPR.2019.00320"},{"key":"5241_CR31","unstructured":"Yu S, Tan D, Tan T (2006) A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: 18th International conference on pattern recognition (ICPR\u201906), vol 4, pp 441\u2013444"},{"issue":"1","key":"5241_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s41074-018-0039-6","volume":"10","author":"N Takemura","year":"2018","unstructured":"Takemura N, Makihara Y, Muramatsu D, Echigo T, Yagi Y (2018) Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition. IPSJ Transactions on Computer Vision and Applications 10(1):1\u201314","journal-title":"IPSJ Transactions on Computer Vision and Applications"},{"key":"5241_CR33","doi-asserted-by":"crossref","unstructured":"Ye M, Shen J, Lin G, Xiang T, Shao L, Hoi SCH (2021) Deep learning for person re-identification: a survey and outlook. IEEE transactions on pattern analysis and machine intelligence 44(6):2872\u20132893","DOI":"10.1109\/TPAMI.2021.3054775"},{"key":"5241_CR34","doi-asserted-by":"crossref","unstructured":"Han Y, Huang G, Song S, Yang L, Wang H, Wang Y (2021) Dynamic neural networks: a survey. IEEE Trans Pattern Anal Mach Intell 44(11):7436\u20137456","DOI":"10.1109\/TPAMI.2021.3117837"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-05241-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-023-05241-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-023-05241-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T20:40:30Z","timestamp":1710362430000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-023-05241-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,2]]},"references-count":34,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2024,2]]}},"alternative-id":["5241"],"URL":"https:\/\/doi.org\/10.1007\/s10489-023-05241-9","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,2]]},"assertion":[{"value":"14 December 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 February 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interests"}}]}}