{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,29]],"date-time":"2026-03-29T17:35:48Z","timestamp":1774805748115,"version":"3.50.1"},"reference-count":45,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T00:00:00Z","timestamp":1651190400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T00:00:00Z","timestamp":1651190400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2023,1]]},"DOI":"10.1007\/s10489-022-03474-8","type":"journal-article","created":{"date-parts":[[2022,4,29]],"date-time":"2022-04-29T09:04:00Z","timestamp":1651223040000},"page":"1517-1534","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":22,"title":["A multi-modal dataset for gait recognition under occlusion"],"prefix":"10.1007","volume":"53","author":[{"given":"Na","family":"Li","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-2618-9289","authenticated-orcid":false,"given":"Xinbo","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,29]]},"reference":[{"key":"3474_CR1","doi-asserted-by":"crossref","unstructured":"Sengupta S, Jayaram V, Curless B, Seitz S M, Kemelmacher-Shlizerman I (2020) Background matting: The world is your green screen. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 2291\u20132300","DOI":"10.1109\/CVPR42600.2020.00236"},{"key":"3474_CR2","doi-asserted-by":"crossref","unstructured":"Zhang S-H, Li R, Dong X, Rosin P, Cai Z, Han X, Yang D, Huang H, Hu S-M (2019) Pose2seg: Detection free human instance segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 889\u2013898","DOI":"10.1109\/CVPR.2019.00098"},{"key":"3474_CR3","doi-asserted-by":"crossref","unstructured":"Rogez G, Weinzaepfel P, Schmid C (2017) Lcr-net: Localization-classification-regression for human pose. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 3433\u20133441","DOI":"10.1109\/CVPR.2017.134"},{"key":"3474_CR4","doi-asserted-by":"crossref","unstructured":"Kocabas M, Athanasiou N, Black M J (2020) Vibe: Video inference for human body pose and shape estimation. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR)","DOI":"10.1109\/CVPR42600.2020.00530"},{"key":"3474_CR5","doi-asserted-by":"crossref","unstructured":"Mehta D, Rhodin H, Casas D, Fua P, Sotnychenko O, Xu W, Theobalt C (2017) Monocular 3d human pose estimation in the wild using improved cnn supervision. In: 2017 international conference on 3D vision (3DV). IEEE, pp 506\u2013516","DOI":"10.1109\/3DV.2017.00064"},{"key":"3474_CR6","doi-asserted-by":"crossref","unstructured":"Mehta D, Sotnychenko O, Mueller F, Xu W, Sridhar S, Pons-Moll G, Theobalt C (2018) Single-shot multi-person 3d pose estimation from monocular rgb. In: 2018 International Conference on 3D Vision (3DV). IEEE, pp 120\u2013130","DOI":"10.1109\/3DV.2018.00024"},{"key":"3474_CR7","doi-asserted-by":"crossref","unstructured":"Moon G, Chang J Y, Lee K M (2019) Camera distance-aware top-down approach for 3d multi-person pose estimation from a single rgb image. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp 10133\u201310142","DOI":"10.1109\/ICCV.2019.01023"},{"key":"3474_CR8","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1016\/j.jvcir.2016.05.020","volume":"39","author":"K Yang","year":"2016","unstructured":"Yang K, Dou Y, Lv S, Zhang F, Lv Q (2016) Relative distance features for gait recognition with kinect. J Vis Commun Image R 39:209\u2013217","journal-title":"J Vis Commun Image R"},{"key":"3474_CR9","unstructured":"Yu S, Tan D, Tan T (2006) A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: 18th International Conference on Pattern Recognition (ICPR\u201906), vol 4. IEEE, pp 441\u2013444"},{"issue":"3","key":"3474_CR10","doi-asserted-by":"publisher","first-page":"3260","DOI":"10.1109\/JSEN.2020.3022374","volume":"21","author":"T Wang","year":"2020","unstructured":"Wang T, Li C, Wu C, Zhao C, Sun J, Peng H, Hu X, Hu B (2020) A gait assessment framework for depression detection using kinect sensors. IEEE Sens J 21(3):3260\u20133270","journal-title":"IEEE Sens J"},{"issue":"17","key":"3474_CR11","doi-asserted-by":"publisher","first-page":"7305","DOI":"10.1109\/JSEN.2018.2839732","volume":"18","author":"S Bei","year":"2018","unstructured":"Bei S, Zhen Z, Xing Z, Taocheng L, Qin L (2018) Movement disorder detection via adaptively fused gait analysis based on kinect sensors. IEEE Sens J 18(17):7305\u20137314","journal-title":"IEEE Sens J"},{"key":"3474_CR12","doi-asserted-by":"crossref","unstructured":"Brown Kramer J, Sabalka L, Rush B, Jones K, Nolte T (2020) Automated depth video monitoring for fall reduction: A case study. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp 294\u2013295","DOI":"10.1109\/CVPRW50498.2020.00155"},{"key":"3474_CR13","doi-asserted-by":"crossref","unstructured":"Andersson V O, Araujo R M (2015) Person identification using anthropometric and gait data from kinect sensor. In: Twenty-Ninth AAAI Conference on Artificial Intelligence, pp 425\u201343","DOI":"10.1609\/aaai.v29i1.9212"},{"key":"3474_CR14","doi-asserted-by":"crossref","unstructured":"Baisware A, Sayankar B, Hood S (2019) Review on recent advances in human action recognition in video data. In: 2019 9th International Conference on Emerging Trends in Engineering and Technology-Signal and Information Processing (ICETET-SIP-19). IEEE, pp 1\u20135","DOI":"10.1109\/ICETET-SIP-1946815.2019.9092193"},{"key":"3474_CR15","doi-asserted-by":"crossref","unstructured":"Borr\u00e0s R, Lapedriza A, Igual L (2012) Depth information in human gait analysis: an experimental study on gender recognition. In: International Conference Image Analysis and Recognition. Springer, pp 98\u2013105","DOI":"10.1007\/978-3-642-31298-4_12"},{"key":"3474_CR16","doi-asserted-by":"crossref","unstructured":"Sivapalan S, Chen D, Denman S, Sridharan S, Fookes C (2011) Gait energy volumes and frontal gait recognition using depth images. In: 2011 International Joint Conference on Biometrics (IJCB). IEEE, pp 1\u20136","DOI":"10.1109\/IJCB.2011.6117504"},{"issue":"1","key":"3474_CR17","doi-asserted-by":"publisher","first-page":"195","DOI":"10.1016\/j.jvcir.2013.02.006","volume":"25","author":"M Hofmann","year":"2014","unstructured":"Hofmann M, Geiger J, Bachmann S, Schuller B, Rigoll G (2014) The tum gait from audio, image and depth (gaid) database: Multimodal recognition of subjects and traits. J Vis Commun Image R 25 (1):195\u2013206","journal-title":"J Vis Commun Image R"},{"key":"3474_CR18","doi-asserted-by":"crossref","unstructured":"Wang Y, Sun J, Li J, Zhao D (2016) Gait recognition based on 3d skeleton joints captured by kinect. In: 2016 IEEE International Conference on Image Processing (ICIP). IEEE, pp 3151\u20133155","DOI":"10.1109\/ICIP.2016.7532940"},{"key":"3474_CR19","doi-asserted-by":"publisher","first-page":"53","DOI":"10.2197\/ipsjtcva.4.53","volume":"4","author":"Y Makihara","year":"2012","unstructured":"Makihara Y, Mannami H, Tsuji A, Hossain M A, Sugiura K, Mori A, Yagi Y (2012) The ou-isir gait database comprising the treadmill dataset. IPSJ Trans Comput Vis Appl 4:53\u201362","journal-title":"IPSJ Trans Comput Vis Appl"},{"key":"3474_CR20","unstructured":"Weizhi A, Shiqi Y, Yasushi M, Xinhui W, Chi X, Yang Y, Rijun L, Yasushi Y (2020) Performance evaluation of model-based gait on multi-view very large population database with pose sequences. IEEE Trans. on Biometrics, Behavior, and Identity Science"},{"key":"3474_CR21","unstructured":"Hofmann M, Sural S, Rigoll G (2011) Gait recognition in the presence of occlusion: A new dataset and baseline algorithms"},{"issue":"2","key":"3474_CR22","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/TPAMI.2016.2545669","volume":"39","author":"Z Wu","year":"2016","unstructured":"Wu Z, Huang Y, Wang L, Wang X, Tan T (2016) A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans Pattern Anal Mach Intell 39(2):209\u2013226","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3474_CR23","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1016\/j.neucom.2017.02.006","volume":"239","author":"S Yu","year":"2017","unstructured":"Yu S, Chen H, Wang Q, Shen L, Huang Y (2017) Invariant feature extraction for gait recognition using only one uniform model. Neurocomputing 239:81\u201393","journal-title":"Neurocomputing"},{"key":"3474_CR24","doi-asserted-by":"crossref","unstructured":"Yu S, Chen H, Garcia Reyes E B, Poh N (2017) Gaitgan: Invariant gait feature extraction using generative adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp 30\u201337","DOI":"10.1109\/CVPRW.2017.80"},{"issue":"1","key":"3474_CR25","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1109\/TIFS.2018.2844819","volume":"14","author":"Y He","year":"2018","unstructured":"He Y, Zhang J, Shan H, Wang L (2018) Multi-task gans for view-specific feature learning in gait recognition. IEEE Trans Inf Forensic Secur 14(1):102\u2013113","journal-title":"IEEE Trans Inf Forensic Secur"},{"key":"3474_CR26","doi-asserted-by":"crossref","unstructured":"Yu S, Liao R, An W, Chen H, B E, Huang Y, Poh N (2019) Gaitganv2: invariant gait feature extraction using generative adversarial networks. Pattern Recogn:179\u2013189","DOI":"10.1016\/j.patcog.2018.10.019"},{"key":"3474_CR27","doi-asserted-by":"crossref","unstructured":"Chao H, He Y, Zhang J, Feng J (2019) Gaitset: Regarding gait as a set for cross-view gait recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 33, pp 8126\u20138133","DOI":"10.1609\/aaai.v33i01.33018126"},{"key":"3474_CR28","doi-asserted-by":"crossref","unstructured":"Fan C, Peng Y, Cao C, Liu X, Hou S, Chi J, Huang Y, Li Q, He Z (2020) Gaitpart: Temporal part-based model for gait recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp 14225\u201314233","DOI":"10.1109\/CVPR42600.2020.01423"},{"key":"3474_CR29","doi-asserted-by":"crossref","unstructured":"Xu W (2021) Graph-optimized coupled discriminant projections for cross-view gait recognition. Appl Intell:1\u201313","DOI":"10.1007\/s10489-021-02322-5"},{"key":"3474_CR30","doi-asserted-by":"crossref","unstructured":"Zhao L, Guo L, Zhang R, Xie X, Ye X (2021) mmgaitset: multimodal based gait recognition for countering carrying and clothing changes. Appl Intell:1\u201314","DOI":"10.1007\/s10489-021-02484-2"},{"issue":"5","key":"3474_CR31","first-page":"1200","volume":"48","author":"J Portillo-Portillo","year":"2018","unstructured":"Portillo-Portillo J, Leyva R, Sanchez V, Sanchez-Perez G, Perez-Meana H, Olivares-Mercado J, Toscano-Medina K, Nakano-Miyatake M (2018) A view-invariant gait recognition algorithm based on a joint-direct linear discriminant analysis. Appl Intell 48(5):1200\u20131217","journal-title":"Appl Intell"},{"key":"3474_CR32","doi-asserted-by":"crossref","unstructured":"Liao R, Cao C, Garcia E B, Yu S, Huang Y (2017) Pose-based temporal-spatial network (ptsn) for gait recognition with carrying and clothing variations. In: Chinese Conference on Biometric Recognition. Springer, pp 474\u2013483","DOI":"10.1007\/978-3-319-69923-3_51"},{"key":"3474_CR33","doi-asserted-by":"crossref","unstructured":"An W, Liao R, Yu S, Huang Y, Yuen P C (2018) Improving gait recognition with 3d pose estimation. In: Chinese Conference on Biometric Recognition. Springer, pp 137\u2013147","DOI":"10.1007\/978-3-319-97909-0_15"},{"key":"3474_CR34","first-page":"107069","volume":"98","author":"R Liao","year":"2020","unstructured":"Liao R, Yu S, An W, Huang Y (2020) A model-based gait recognition method with body pose and human prior knowledge. Lect Notes Comput Sc 98:107069","journal-title":"Lect Notes Comput Sc"},{"key":"3474_CR35","unstructured":"Li N, Zhao X, Ma C (2020) Jointsgait: Gait recognition based on graph convolutional networks and joints relationship pyramid mapping. arXiv:2005.08625"},{"key":"3474_CR36","unstructured":"Sendhil Kumara S, Chattopadhyaya P, Wang L (2021) Bgaitr-net: Occluded gait sequence reconstructionwith temporally constrained model for gait recognition. arXiv e-prints, pp arXiv\u20132110"},{"key":"3474_CR37","doi-asserted-by":"crossref","unstructured":"Babaee M, Li L, Rigoll G (2018) Gait recognition from incomplete gait cycle. In: 2018 25th IEEE International Conference on Image Processing (ICIP). IEEE, pp 768\u2013772","DOI":"10.1109\/ICIP.2018.8451785"},{"issue":"11","key":"3474_CR38","doi-asserted-by":"publisher","first-page":"977","DOI":"10.1016\/j.patrec.2009.04.012","volume":"30","author":"C Chen","year":"2009","unstructured":"Chen C, Liang J, Zhao H, Hu H, Tian J (2009) Frame difference energy image for gait recognition with incomplete silhouettes. Pattern Recogn Lett 30(11):977\u2013984","journal-title":"Pattern Recogn Lett"},{"issue":"1","key":"3474_CR39","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s41074-019-0053-3","volume":"11","author":"MZ Uddin","year":"2019","unstructured":"Uddin M Z, Muramatsu D, Takemura N, Ahad M A R, Yagi Y (2019) Spatio-temporal silhouette sequence reconstruction for gait recognition against occlusion. IPSJ Trans Comput Vis Appl 11(1):1\u201318","journal-title":"IPSJ Trans Comput Vis Appl"},{"key":"3474_CR40","doi-asserted-by":"crossref","unstructured":"Yan S, Xiong Y, Lin D (2018) Spatial temporal graph convolutional networks for skeleton-based action recognition. arXiv:1801.07455","DOI":"10.1609\/aaai.v32i1.12328"},{"issue":"7","key":"3474_CR41","doi-asserted-by":"publisher","first-page":"1697","DOI":"10.1109\/TPAMI.2017.2726061","volume":"40","author":"X Chen","year":"2017","unstructured":"Chen X, Weng J, Lu W, Xu J (2017) Multi-gait recognition based on attribute discovery. IEEE Trans Pattern Anal Mach Intell 40(7):1697\u20131710","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3474_CR42","doi-asserted-by":"crossref","unstructured":"Kozlowski L T, Cutting J E (1978) Recognizing the gender of walkers from point-lights mounted on ankles: some second thoughts. Perception & Psychophysics","DOI":"10.3758\/BF03204150"},{"key":"3474_CR43","unstructured":"Hermans A, Beyer L, Leibe B (2017) In defense of the triplet loss for person re-identification. arXiv:1703.07737"},{"key":"3474_CR44","doi-asserted-by":"crossref","unstructured":"Deng J, Guo J, Xue N, Zafeiriou S (2019) Arcface: Additive angular margin loss for deep face recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp 4690\u20134699","DOI":"10.1109\/CVPR.2019.00482"},{"key":"3474_CR45","doi-asserted-by":"crossref","unstructured":"Teepe T, Khan A, Gilg J, Herzog F, H\u00f6rmann S, Rigoll G (2021) Gaitgraph: Graph convolutional network for skeleton-based gait recognition. arXiv:2101.11228","DOI":"10.1109\/ICIP42928.2021.9506717"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03474-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-022-03474-8\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-022-03474-8.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,4]],"date-time":"2023-01-04T04:37:59Z","timestamp":1672807079000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-022-03474-8"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,29]]},"references-count":45,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2023,1]]}},"alternative-id":["3474"],"URL":"https:\/\/doi.org\/10.1007\/s10489-022-03474-8","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,29]]},"assertion":[{"value":"4 March 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 April 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}