{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,17]],"date-time":"2026-03-17T19:27:48Z","timestamp":1773775668276,"version":"3.50.1"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,4,1]],"date-time":"2024-04-01T00:00:00Z","timestamp":1711929600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Shenzhen Startup Funding","award":["QD2023014C"],"award-info":[{"award-number":["QD2023014C"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1007\/s10489-024-05422-0","type":"journal-article","created":{"date-parts":[[2024,5,8]],"date-time":"2024-05-08T12:01:46Z","timestamp":1715169706000},"page":"6154-6174","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":13,"title":["Spatiotemporal smoothing aggregation enhanced multi-scale residual deep graph convolutional networks for skeleton-based gait recognition"],"prefix":"10.1007","volume":"54","author":[{"given":"Guanghai","family":"Chen","sequence":"first","affiliation":[]},{"given":"Xin","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Chengzhi","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Junshu","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xinchao","family":"Liu","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3553-6764","authenticated-orcid":false,"given":"Yuxing","family":"Han","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,8]]},"reference":[{"issue":"2","key":"5422_CR1","doi-asserted-by":"publisher","first-page":"1517","DOI":"10.1007\/s10489-022-03474-8","volume":"53","author":"N Li","year":"2023","unstructured":"Li N, Zhao X (2023) A multi-modal dataset for gait recognition under occlusion. Appl Intell 53(2):1517\u20131534","journal-title":"Appl Intell"},{"issue":"2","key":"5422_CR2","doi-asserted-by":"publisher","first-page":"1535","DOI":"10.1007\/s10489-022-03543-y","volume":"53","author":"G Li","year":"2023","unstructured":"Li G, Guo L, Zhang R et al (2023) Transgait: Multimodal-based gait recognition with set transformer. Appl Intell 53(2):1535\u20131547","journal-title":"Appl Intell"},{"issue":"3","key":"5422_CR3","doi-asserted-by":"publisher","first-page":"734","DOI":"10.1109\/TCSVT.2019.2893736","volume":"30","author":"X Ben","year":"2019","unstructured":"Ben X, Gong C, Zhang P et al (2019) Coupled bilinear discriminant projection for cross-view gait recognition. IEEE Trans Circuits Syst Video Technol 30(3):734\u2013747","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"5422_CR4","doi-asserted-by":"crossref","unstructured":"Chao H, He Y, Zhang J, et\u00a0al (2019) Gaitset: Regarding gait as a set for cross-view gait recognition. In: Proceedings of the AAAI conference on artificial intelligence, pp 8126\u20138133","DOI":"10.1609\/aaai.v33i01.33018126"},{"key":"5422_CR5","doi-asserted-by":"crossref","unstructured":"Dang L, Nie Y, Long C, et\u00a0al (2021) Msrgcn: Multi-scale residual graph convolution networks for human motion prediction. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 11467\u201311476","DOI":"10.1109\/ICCV48922.2021.01127"},{"key":"5422_CR6","doi-asserted-by":"crossref","unstructured":"Huang X, Zhu D, Wang H, et\u00a0al (2021) Context-sensitive temporal feature learning for gait recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 12909\u201312918","DOI":"10.1109\/ICCV48922.2021.01267"},{"key":"5422_CR7","doi-asserted-by":"crossref","unstructured":"Fan C, Peng Y, Cao C, et\u00a0al (2020) Gaitpart: Temporal part-based model for gait recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14225\u201314233","DOI":"10.1109\/CVPR42600.2020.01423"},{"issue":"9","key":"5422_CR8","doi-asserted-by":"publisher","first-page":"6448","DOI":"10.1109\/TCSVT.2022.3161515","volume":"32","author":"Y Yang","year":"2022","unstructured":"Yang Y, Yang X, Sakamoto T et al (2022) Unsupervised domain adaptation for disguised-gait-based person identification on micro-doppler signatures. IEEE Trans Circuits Syst Video Technol 32(9):6448\u20136460","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"issue":"4","key":"5422_CR9","doi-asserted-by":"publisher","first-page":"4592","DOI":"10.1007\/s10489-022-03589-y","volume":"53","author":"Y Xing","year":"2023","unstructured":"Xing Y, Zhu J, Li Y et al (2023) An improved spatial temporal graph convolutional network for robust skeleton-based action recognition. Appl Intell 53(4):4592\u20134608","journal-title":"Appl Intell"},{"issue":"12","key":"5422_CR10","doi-asserted-by":"publisher","first-page":"14838","DOI":"10.1007\/s10489-022-04179-8","volume":"53","author":"L Yu","year":"2023","unstructured":"Yu L, Tian L, Du Q et al (2023) Multi-stream adaptive 3d attention graph convolution network for skeleton-based action recognition. Appl Intell 53(12):14838\u201314854","journal-title":"Appl Intell"},{"issue":"1","key":"5422_CR11","doi-asserted-by":"publisher","first-page":"574","DOI":"10.1007\/s10489-022-03436-0","volume":"53","author":"W Yang","year":"2023","unstructured":"Yang W, Zhang J, Cai J et al (2023) Hybridnet: Integrating gcn and cnn for skeleton-based action recognition. Appl Intell 53(1):574\u2013585","journal-title":"Appl Intell"},{"key":"5422_CR12","doi-asserted-by":"crossref","unstructured":"Sun K, Xiao B, Liu D, et\u00a0al (2019) Deep high-resolution representation learning for human pose estimation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 5693\u20135703","DOI":"10.1109\/CVPR.2019.00584"},{"key":"5422_CR13","doi-asserted-by":"crossref","unstructured":"Gianaria E, Balossino N, Grangetto M, et\u00a0al (2013) Gait characterization using dynamic skeleton acquisition. In: 2013 IEEE 15th International Workshop on Multimedia Signal Processing (MMSP). IEEE, pp 440\u2013445","DOI":"10.1109\/MMSP.2013.6659329"},{"key":"5422_CR14","doi-asserted-by":"crossref","unstructured":"Cao Z, Simon T, Wei SE, et\u00a0al (2017) Realtime multi-person 2d pose estimation using part affinity fields. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 7291\u20137299","DOI":"10.1109\/CVPR.2017.143"},{"key":"5422_CR15","doi-asserted-by":"crossref","unstructured":"Fang HS, Xie S, Tai YW, et\u00a0al (2017) Rmpe: Regional multi-person pose estimation. In: Proceedings of the IEEE international conference on computer vision. pp 2334\u20132343","DOI":"10.1109\/ICCV.2017.256"},{"key":"5422_CR16","doi-asserted-by":"crossref","unstructured":"Chou CJ, Chien JT, Chen HT (2018) Self adversarial training for human pose estimation. In: 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC). IEEE, pp 17\u201330","DOI":"10.23919\/APSIPA.2018.8659538"},{"key":"5422_CR17","doi-asserted-by":"crossref","unstructured":"Zheng J, Liu X, Liu W, et\u00a0al (2022) Gait recognition in the wild with dense 3d representations and a benchmark. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 20228\u201320237","DOI":"10.1109\/CVPR52688.2022.01959"},{"key":"5422_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.107069","volume":"98","author":"R Liao","year":"2020","unstructured":"Liao R, Yu S, An W et al (2020) A model-based gait recognition method with body pose and human prior knowledge. Pattern Recogn 98:107069","journal-title":"Pattern Recogn"},{"key":"5422_CR19","doi-asserted-by":"crossref","unstructured":"Teepe T, Khan A, Gilg J, et\u00a0al (2021) Gaitgraph: Graph convolutional network for skeleton-based gait recognition. In: 2021 IEEE International Conference on Image Processing (ICIP). IEEE, pp 2314\u20132318","DOI":"10.1109\/ICIP42928.2021.9506717"},{"issue":"24","key":"5422_CR20","doi-asserted-by":"publisher","first-page":"8387","DOI":"10.3390\/s21248387","volume":"21","author":"A Cosma","year":"2021","unstructured":"Cosma A, Radoi IE (2021) Wildgait: Learning gait representations from raw surveillance streams. Sensors 21(24):8387","journal-title":"Sensors"},{"key":"5422_CR21","doi-asserted-by":"crossref","unstructured":"Pinyoanuntapong E, Ali A, Wang P, et\u00a0al (2023) Gaitmixer: skeleton-based gait representation learning via wide-spectrum multi-axial mixer. In: ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP). IEEE, pp 1\u20135","DOI":"10.1109\/ICASSP49357.2023.10096917"},{"key":"5422_CR22","doi-asserted-by":"crossref","unstructured":"Hua G, Long C, Yang M, et\u00a0al (2013) Collaborative active learning of a kernel machine ensemble for recognition. In: Proceedings of the IEEE international conference on computer vision. pp 1209\u20131216","DOI":"10.1109\/ICCV.2013.153"},{"key":"5422_CR23","doi-asserted-by":"publisher","first-page":"3499","DOI":"10.1109\/TIP.2021.3061927","volume":"30","author":"T Hu","year":"2021","unstructured":"Hu T, Long C, Xiao C (2021) A novel visual representation on text using diverse conditional gan for visual recognition. IEEE Trans Image Process 30:3499\u20133512","journal-title":"IEEE Trans Image Process"},{"key":"5422_CR24","doi-asserted-by":"crossref","unstructured":"Long C, Hua G (2015) Multi-class multi-annotator active learning with robust gaussian process for visual recognition. In: Proceedings of the IEEE international conference on computer vision. pp 2839\u20132847","DOI":"10.1109\/ICCV.2015.325"},{"key":"5422_CR25","doi-asserted-by":"crossref","unstructured":"Wang Y, Kitani K, Weng X (2021) Joint object detection and multi-object tracking with graph neural networks. In: 2021 IEEE International Conference on Robotics and Automation (ICRA). IEEE, pp 13708\u201313715","DOI":"10.1109\/ICRA48506.2021.9561110"},{"key":"5422_CR26","doi-asserted-by":"crossref","unstructured":"Zeng R, Huang W, Tan M, et\u00a0al (2019) Graph convolutional networks for temporal action localization. In: Proceedings of the IEEE\/CVF international conference on computer vision. pp 7094\u20137103","DOI":"10.1109\/ICCV.2019.00719"},{"key":"5422_CR27","doi-asserted-by":"crossref","unstructured":"Islam A, Long C, Radke R (2021) A hybrid attention mechanism for weakly-supervised temporal action localization. In: Proceedings of the AAAI conference on artificial intelligence. pp 1637\u20131645","DOI":"10.1609\/aaai.v35i2.16256"},{"key":"5422_CR28","doi-asserted-by":"crossref","unstructured":"Shi L, Wang L, Long C, et\u00a0al (2021) Sgcn: Sparse graph convolution network for pedestrian trajectory prediction. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 8994\u20139003","DOI":"10.1109\/CVPR46437.2021.00888"},{"key":"5422_CR29","doi-asserted-by":"crossref","unstructured":"Yan S, Xiong Y, Lin D (2018) Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Proceedings of the AAAI conference on artificial intelligence","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"5422_CR30","doi-asserted-by":"crossref","unstructured":"Shi L, Zhang Y, Cheng J, et\u00a0al (2019) Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12026\u201312035","DOI":"10.1109\/CVPR.2019.01230"},{"key":"5422_CR31","doi-asserted-by":"crossref","unstructured":"Teepe T, Gilg J, Herzog F, et\u00a0al (2022) Towards a deeper understanding of skeleton-based gait recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 1569\u20131577","DOI":"10.1109\/CVPRW56347.2022.00163"},{"key":"5422_CR32","doi-asserted-by":"crossref","unstructured":"Liao R, Cao C, Garcia EB, et\u00a0al (2017) Pose-based temporal-spatial network (ptsn) for gait recognition with carrying and clothing variations. In: Biometric Recognition: 12th Chinese Conference, CCBR 2017, Shenzhen, China, October 28-29, 2017, Proceedings 12. Springer, pp 474\u2013483","DOI":"10.1007\/978-3-319-69923-3_51"},{"key":"5422_CR33","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention\u2013MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Springer, pp 234\u2013241","DOI":"10.1007\/978-3-319-24574-4_28"},{"issue":"2","key":"5422_CR34","first-page":"134","volume":"8","author":"A Sokolova","year":"2019","unstructured":"Sokolova A, Konushin A (2019) Pose-based deep gait recognition. IET. Biometrics 8(2):134\u2013143","journal-title":"Biometrics"},{"key":"5422_CR35","unstructured":"Yu S, Tan D, Tan T (2006) A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: 18th international conference on pattern recognition (ICPR\u201906). IEEE, pp 441\u2013444"},{"key":"5422_CR36","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2022.108520","volume":"125","author":"X Liu","year":"2022","unstructured":"Liu X, You Z, He Y et al (2022) Symmetry-driven hyper feature gcn for skeleton-based gait recognition. Pattern Recogn 125:108520","journal-title":"Pattern Recogn"},{"key":"5422_CR37","doi-asserted-by":"crossref","unstructured":"Tian H, Ma X, Wu H et al (2022) Skeleton-based abnormal gait recognition with spatio-temporal attention enhanced gait-structural graph convolutional networks. Neurocomputing 473:116\u2013126","DOI":"10.1016\/j.neucom.2021.12.004"},{"key":"5422_CR38","doi-asserted-by":"crossref","unstructured":"Liao R, Li Z, Bhattacharyya SS et al (2022) Posemapgait: A model-based gait recognition method with pose estimation maps and graph convolutional networks. Neurocomputing 501:514\u2013528","DOI":"10.1016\/j.neucom.2022.06.048"},{"key":"5422_CR39","doi-asserted-by":"crossref","unstructured":"Li Q, Han Z, Wu XM (2018) Deeper insights into graph convolutional networks for semi-supervised learning. In: Proceedings of the AAAI conference on artificial intelligence","DOI":"10.1609\/aaai.v32i1.11604"},{"key":"5422_CR40","doi-asserted-by":"crossref","unstructured":"Mao W, Liu M, Salzmann M, et\u00a0al (2019) Learning trajectory dependencies for human motion prediction. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 9489\u20139497","DOI":"10.1109\/ICCV.2019.00958"},{"key":"5422_CR41","doi-asserted-by":"crossref","unstructured":"Song YF, Zhang Z, Shan C, et\u00a0al (2020) Stronger, faster and more explainable: A graph convolutional baseline for skeleton-based action recognition. In: proceedings of the 28th ACM international conference on multimedia. pp 1625\u20131633","DOI":"10.1145\/3394171.3413802"},{"key":"5422_CR42","first-page":"18661","volume":"33","author":"P Khosla","year":"2020","unstructured":"Khosla P, Teterwak P, Wang C et al (2020) Supervised contrastive learning. Adv Neural Inf Process Syst 33:18661\u201318673","journal-title":"Adv Neural Inf Process Syst"},{"key":"5422_CR43","doi-asserted-by":"crossref","unstructured":"Cheng K, Zhang Y, Cao C, et\u00a0al (2020) Decoupling gcn with dropgraph module for skeleton-based action recognition. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXIV 16. Springer, pp 536\u2013553","DOI":"10.1007\/978-3-030-58586-0_32"},{"key":"5422_CR44","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.106988","volume":"96","author":"C Song","year":"2019","unstructured":"Song C, Huang Y, Huang Y et al (2019) Gaitnet: An end-to-end network for gait based human identification. Pattern Recogn 96:106988","journal-title":"Pattern Recogn"},{"issue":"2","key":"5422_CR45","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/TPAMI.2016.2545669","volume":"39","author":"Z Wu","year":"2016","unstructured":"Wu Z, Huang Y, Wang L et al (2016) A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans Pattern Anal Mach Intell 39(2):209\u2013226","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"5422_CR46","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1109\/TPAMI.2006.38","volume":"28","author":"J Han","year":"2005","unstructured":"Han J, Bhanu B (2005) Individual recognition using gait energy image. IEEE Trans Pattern Anal Mach Intell 28(2):316\u2013322","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"5422_CR47","doi-asserted-by":"crossref","unstructured":"Wang C, Zhang J, Pu J, et\u00a0al (2010) Chrono-gait image: A novel temporal template for gait recognition. In: Computer Vision\u2013ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, September 5-11, 2010, Proceedings, Part I 11. Springer, pp 257\u2013270","DOI":"10.1007\/978-3-642-15549-9_19"},{"key":"5422_CR48","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/TIP.2019.2926208","volume":"29","author":"Y Zhang","year":"2019","unstructured":"Zhang Y, Huang Y, Yu S et al (2019) Cross-view gait recognition by discriminative feature learning. IEEE Trans Image Process 29:1001\u20131015","journal-title":"IEEE Trans Image Process"},{"issue":"1","key":"5422_CR49","doi-asserted-by":"publisher","first-page":"260","DOI":"10.1109\/TCSVT.2020.2975671","volume":"31","author":"C Xu","year":"2020","unstructured":"Xu C, Makihara Y, Li X et al (2020) Cross-view gait recognition using pairwise spatial transformer networks. IEEE Trans Circuits Syst Video Technol 31(1):260\u2013274","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"issue":"9","key":"5422_CR50","doi-asserted-by":"publisher","first-page":"2708","DOI":"10.1109\/TCSVT.2017.2760835","volume":"29","author":"N Takemura","year":"2017","unstructured":"Takemura N, Makihara Y, Muramatsu D et al (2017) On input\/output architectures for convolutional neural network-based cross-view gait recognition. IEEE Trans Circuits Syst Video Technol 29(9):2708\u20132719","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"5422_CR51","doi-asserted-by":"crossref","unstructured":"Lin B, Zhang S, Bao F (2020) Gait recognition with multiple-temporal-scale 3d convolutional neural network. In: Proceedings of the 28th ACM international conference on multimedia. pp 3054\u20133062","DOI":"10.1145\/3394171.3413861"},{"key":"5422_CR52","doi-asserted-by":"crossref","unstructured":"Si C, Chen W, Wang W, et\u00a0al (2019) An attention enhanced graph convolutional lstm network for skeleton-based action recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1227\u20131236","DOI":"10.1109\/CVPR.2019.00132"},{"key":"5422_CR53","unstructured":"Li N, Zhao X, Ma C (2020) Jointsgait: A model-based gait recognition method based on gait graph convolutional networks and joints relationship pyramid mapping. arXiv:2005.08625"},{"key":"5422_CR54","doi-asserted-by":"crossref","unstructured":"Smith LN, Topin N (2019) Super-convergence: Very fast training of neural networks using large learning rates. In: Artificial intelligence and machine learning for multi-domain operations applications, SPIE, pp 369\u2013386","DOI":"10.1117\/12.2520589"},{"key":"5422_CR55","unstructured":"Zhu Z, Guo X, Yang T, et\u00a0al (2021) Gait recognition in the wild: A benchmark. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 14789\u201314799"},{"key":"5422_CR56","doi-asserted-by":"crossref","unstructured":"Selvaraju RR, Cogswell M, Das A, et\u00a0al (2017) Grad-cam: Visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE international conference on computer vision, pp 618\u2013626","DOI":"10.1109\/ICCV.2017.74"},{"key":"5422_CR57","doi-asserted-by":"crossref","unstructured":"Yu S, Chen H, Garcia\u00a0Reyes EB, et\u00a0al (2017) Gaitgan: Invariant gait feature extraction using generative adversarial networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops, pp 30\u201337","DOI":"10.1109\/CVPRW.2017.80"},{"key":"5422_CR58","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1016\/j.patcog.2018.10.019","volume":"87","author":"S Yu","year":"2019","unstructured":"Yu S, Liao R, An W et al (2019) Gaitganv 2: Invariant gait feature extraction using generative adversarial networks. Pattern Recogn 87:179\u2013189","journal-title":"Pattern Recogn"},{"issue":"1","key":"5422_CR59","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1109\/TIFS.2018.2844819","volume":"14","author":"Y He","year":"2018","unstructured":"He Y, Zhang J, Shan H et al (2018) Multi-task gans for view-specific feature learning in gait recognition. IEEE Trans Inf Forensics Secur 14(1):102\u2013113","journal-title":"IEEE Trans Inf Forensics Secur"},{"key":"5422_CR60","doi-asserted-by":"crossref","unstructured":"Shiraga K, Makihara Y, Muramatsu D, et\u00a0al (2016) Geinet: View-invariant gait recognition using a convolutional neural network. In: 2016 international conference on biometrics (ICB). IEEE, pp 1\u20138","DOI":"10.1109\/ICB.2016.7550060"},{"issue":"2","key":"5422_CR61","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/TPAMI.2016.2545669","volume":"39","author":"Z Wu","year":"2016","unstructured":"Wu Z, Huang Y, Wang L et al (2016) A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans Pattern Anal Mach Intell 39(2):209\u2013226","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"5422_CR62","doi-asserted-by":"crossref","unstructured":"Hou S, Cao C, Liu X, et\u00a0al (2020) Gait lateral network: Learning discriminative and compact representations for gait recognition. In: European conference on computer vision, Springer, pp 382\u2013398","DOI":"10.1007\/978-3-030-58545-7_22"},{"key":"5422_CR63","doi-asserted-by":"crossref","unstructured":"Lin B, Zhang S, Yu X (2021) Gait recognition via effective global-local feature representation and local temporal aggregation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 14648\u201314656","DOI":"10.1109\/ICCV48922.2021.01438"},{"issue":"11","key":"5422_CR64","doi-asserted-by":"publisher","first-page":"11819","DOI":"10.1109\/TCYB.2021.3082114","volume":"52","author":"Y Wu","year":"2022","unstructured":"Wu Y, Wang Y, Li Y et al (2022) Top-k self-adaptive contrast sequential pattern mining. IEEE transactions on cybernetics 52(11):11819\u201311833","journal-title":"IEEE transactions on cybernetics"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-024-05422-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-024-05422-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-024-05422-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T12:13:21Z","timestamp":1718453601000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-024-05422-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,4]]},"references-count":64,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["5422"],"URL":"https:\/\/doi.org\/10.1007\/s10489-024-05422-0","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,4]]},"assertion":[{"value":"27 March 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 May 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}},{"value":"Not applicable. The work in this paper has no ethical or moral implications such as human or animal experimentation. The work presented in this article is entirely original and has not been published in any other journals. This journal is the premiere and exclusive contributing journal for the paper. There are no violations of academic ethics. The right to use the data used in this study has been approved by the owner.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical and informed consent for data used"}},{"value":"All participants in this study were informed and consented to participate in the study.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"Participants in this study gave their consent for the results to be used for publication, presentation, or sharing.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}]}}