{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,24]],"date-time":"2026-03-24T03:10:55Z","timestamp":1774321855501,"version":"3.50.1"},"reference-count":34,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2021,6,2]],"date-time":"2021-06-02T00:00:00Z","timestamp":1622592000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,6,2]],"date-time":"2021-06-02T00:00:00Z","timestamp":1622592000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100010248","name":"Zhejiang Province Public Welfare Technology Application Research Project","doi-asserted-by":"publisher","award":["No.LGF21F020008"],"award-info":[{"award-number":["No.LGF21F020008"]}],"id":[{"id":"10.13039\/501100010248","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["No.61906101"],"award-info":[{"award-number":["No.61906101"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100011442","name":"State Key Laboratory of Computer Aided Design and Computer Graphics","doi-asserted-by":"publisher","award":["No.A2119"],"award-info":[{"award-number":["No.A2119"]}],"id":[{"id":"10.13039\/501100011442","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100005315","name":"Natural Science Foundation of Ningbo Municipality","doi-asserted-by":"publisher","award":["No.2018A610057"],"award-info":[{"award-number":["No.2018A610057"]}],"id":[{"id":"10.13039\/501100005315","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Appl Intell"],"published-print":{"date-parts":[[2022,1]]},"DOI":"10.1007\/s10489-021-02484-2","type":"journal-article","created":{"date-parts":[[2021,6,2]],"date-time":"2021-06-02T21:03:12Z","timestamp":1622667792000},"page":"2023-2036","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":29,"title":["mmGaitSet: multimodal based gait recognition for countering carrying and clothing changes"],"prefix":"10.1007","volume":"52","author":[{"given":"Liming","family":"Zhao","sequence":"first","affiliation":[]},{"given":"Lijun","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Rong","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xijiong","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Xulun","family":"Ye","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,6,2]]},"reference":[{"key":"2484_CR1","doi-asserted-by":"crossref","unstructured":"Fan C, Peng Y, Cao C, Liu X, Hou S, Chi J, Huang Y, Li Q, He Z (2020) GaitPart: temporal part-based model for gait recognition. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14225\u201314233","DOI":"10.1109\/CVPR42600.2020.01423"},{"key":"2484_CR2","doi-asserted-by":"publisher","first-page":"107069","DOI":"10.1016\/j.patcog.2019.107069","volume":"98","author":"R Liao","year":"2020","unstructured":"Liao R, Yu S, An W, Huang Y (2020) A model-based gait recognition method with body pose and human prior knowledge. Pattern Recogn 98:107069","journal-title":"Pattern Recogn"},{"issue":"4","key":"2484_CR3","first-page":"1187","volume":"8","author":"H Iwama","year":"2013","unstructured":"Iwama H, Muramatsu D, Makihara Y, Yagi Y (2013) Gait verification system for criminal investigation. Inf Med Technol 8(4):1187\u20131199","journal-title":"Inf Med Technol"},{"issue":"2","key":"2484_CR4","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1049\/iet-bmt.2013.0090","volume":"3","author":"N Lynnerup","year":"2014","unstructured":"Lynnerup N, Larsen PK (2014) Gait as evidence. IET Biom 3(2):47\u201354","journal-title":"IET Biom"},{"issue":"1","key":"2484_CR5","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1109\/TIFS.2018.2844819","volume":"14","author":"Y He","year":"2019","unstructured":"He Y, Zhang J, Shan H, Wang L (2019) Multi-task GANs for view-specific feature learning in gait recognition. IEEE Trans Inf Forens Secur 14(1):102\u2013113","journal-title":"IEEE Trans Inf Forens Secur"},{"issue":"2","key":"2484_CR6","doi-asserted-by":"publisher","first-page":"209","DOI":"10.1109\/TPAMI.2016.2545669","volume":"39","author":"Z Wu","year":"2016","unstructured":"Wu Z, Huang Y, Wang L, Wang X, Tan T (2016) A comprehensive study on cross-view gait based human identification with deep cnns. IEEE Trans Pattern Anal Mach Intell 39(2):209\u2013226","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"2484_CR7","doi-asserted-by":"crossref","unstructured":"Wolf T, Babaee M, Rigoll G (2016) Multi-view gait recognition using 3D convolutional neural networks. In: 2016 IEEE international conference on image processing (ICIP). IEEE, pp 4165\u20134169","DOI":"10.1109\/ICIP.2016.7533144"},{"key":"2484_CR8","doi-asserted-by":"crossref","unstructured":"Chao H, He Y, Zhang J, Feng J (2019) Gaitset: Regarding gait as a set for cross-view gait recognition. In: Proceedings of the AAAI conference on artificial intelligence, pp 8126\u20138133","DOI":"10.1609\/aaai.v33i01.33018126"},{"key":"2484_CR9","doi-asserted-by":"crossref","unstructured":"Zhao G, Liu G, Li H, Pietikainen M (2006) 3D gait recognition using multiple cameras. In: 7th international conference on automatic face and gesture recognition (FGR06). IEEE, pp 529\u2013534","DOI":"10.1109\/FGR.2006.2"},{"key":"2484_CR10","unstructured":"Yu S, Tan D, Tan T (2006) A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: 18th international conference on pattern recognition (ICPR\u201906). IEEE, pp 441\u2013444"},{"key":"2484_CR11","unstructured":"Hermans A, Beyer L, Leibe B (2017) In defense of the triplet loss for person re-identification. arXiv:170307737"},{"issue":"2","key":"2484_CR12","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1109\/TPAMI.2006.38","volume":"28","author":"J Han","year":"2006","unstructured":"Han J, Bhanu B (2006) Individual recognition using gait energy image. IEEE Trans Pattern Anal Mach Intell 28(2):316\u2013322","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"2484_CR13","doi-asserted-by":"crossref","unstructured":"Bashir K, Xiang T, Gong S (2009) Gait recognition using gait entropy image","DOI":"10.1049\/ic.2009.0230"},{"issue":"13","key":"2484_CR14","doi-asserted-by":"publisher","first-page":"2052","DOI":"10.1016\/j.patrec.2010.05.027","volume":"31","author":"K Bashir","year":"2010","unstructured":"Bashir K, Xiang T, Gong S (2010) Gait recognition without subject cooperation. Pattern Recogn Lett 31(13):2052\u20132060","journal-title":"Pattern Recogn Lett"},{"key":"2484_CR15","doi-asserted-by":"crossref","unstructured":"Makihara Y, Suzuki A, Muramatsu D, Li X, Yagi Y (2017) Joint intensity and spatial metric learning for robust gait recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5705\u20135715","DOI":"10.1109\/CVPR.2017.718"},{"issue":"7","key":"2484_CR16","doi-asserted-by":"publisher","first-page":"1521","DOI":"10.1109\/TPAMI.2014.2366766","volume":"37","author":"Y Guan","year":"2014","unstructured":"Guan Y, Li C-T, Roli F (2014) On reducing the effect of covariate factors in gait recognition: a classifier ensemble method. IEEE Trans Pattern Anal Mach Intell 37(7):1521\u20131528","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"11","key":"2484_CR17","doi-asserted-by":"publisher","first-page":"1960","DOI":"10.1109\/TMM.2015.2477681","volume":"17","author":"Z Wu","year":"2015","unstructured":"Wu Z, Huang Y, Wang L (2015) Learning representative deep features for image set analysis. IEEE Trans Multimed 17(11):1960\u20131968","journal-title":"IEEE Trans Multimed"},{"key":"2484_CR18","doi-asserted-by":"publisher","first-page":"1001","DOI":"10.1109\/TIP.2019.2926208","volume":"29","author":"Y Zhang","year":"2020","unstructured":"Zhang Y, Huang Y, Yu S, Wang L (2020) Cross-View Gait Recognition by Discriminative Feature Learning. IEEE Trans Image Process 29:1001\u20131015","journal-title":"IEEE Trans Image Process"},{"key":"2484_CR19","doi-asserted-by":"crossref","unstructured":"Hou S, Cao C, Liu X, Huang Y (2020) Gait lateral network: learning discriminative and compact representations for gait recognition. In: European conference on computer vision. Springer, pp 382\u2013398","DOI":"10.1007\/978-3-030-58545-7_22"},{"key":"2484_CR20","doi-asserted-by":"crossref","unstructured":"Ariyanto G, Nixon MS (2011) Model-based 3D gait biometrics. In: 2011 international joint conference on biometrics (IJCB). IEEE, pp 1\u20137","DOI":"10.1109\/IJCB.2011.6117582"},{"issue":"8","key":"2484_CR21","doi-asserted-by":"publisher","first-page":"1194","DOI":"10.1016\/j.imavis.2008.11.008","volume":"27","author":"R Bodor","year":"2009","unstructured":"Bodor R, Drenner A, Fehr D, Masoud O, Papanikolopoulos N (2009) View-independent human motion classification using image-based reconstruction. Image Vis Comput 27(8):1194\u20131206","journal-title":"Image Vis Comput"},{"key":"2484_CR22","doi-asserted-by":"crossref","unstructured":"Urtasun R, Fua P (2004) 3D tracking for gait characterization and recognition. In: Sixth IEEE international conference on automatic face and gesture recognition, 2004. Proceedings. IEEE, pp 17\u201322","DOI":"10.1109\/AFGR.2004.1301503"},{"key":"2484_CR23","doi-asserted-by":"crossref","unstructured":"Urtasun R, Fua P (2004) 3D tracking for gait characterization and recognition. In: Sixth IEEE international conference on automatic face and gesture recognition, 2004. Proceedings. IEEE, pp 17\u201322","DOI":"10.1109\/AFGR.2004.1301503"},{"issue":"5","key":"2484_CR24","doi-asserted-by":"publisher","first-page":"1057","DOI":"10.1016\/j.patcog.2003.09.012","volume":"37","author":"C Yam","year":"2004","unstructured":"Yam C, Nixon MS, Carter JN (2004) Automated person recognition by walking and running via model-based approaches. Pattern Recognit 37(5):1057\u20131072","journal-title":"Pattern Recognit"},{"key":"2484_CR25","doi-asserted-by":"crossref","unstructured":"Feng Y, Li Y, Luo J (2016) Learning effective gait features using LSTM. In: 2016 23rd international conference on pattern recognition (ICPR). IEEE, pp 325\u2013330","DOI":"10.1109\/ICPR.2016.7899654"},{"key":"2484_CR26","doi-asserted-by":"crossref","unstructured":"Liao R, Cao C, Garcia EB, Yu S, Huang Y (2017) Pose-based temporal-spatial network (PTSN) for gait recognition with carrying and clothing variations. In: Chinese conference on biometric recognition. Springer, pp 474\u2013483","DOI":"10.1007\/978-3-319-69923-3_51"},{"key":"2484_CR27","doi-asserted-by":"crossref","unstructured":"Li X, Makihara Y, Xu C, Yagi Y, Yu S, Ren M (2020) End-to-end model-based gait recognition. In: Proceedings of the Asian conference on computer vision","DOI":"10.1007\/978-3-030-69535-4_1"},{"key":"2484_CR28","unstructured":"Ngiam J, Khosla A, Kim M, Nam J, Lee H, Ng AY (2011) Multimodal deep learning. In: ICML"},{"issue":"5","key":"2484_CR29","doi-asserted-by":"publisher","first-page":"956","DOI":"10.1109\/TFUZZ.2018.2870590","volume":"27","author":"P Kumar","year":"2018","unstructured":"Kumar P, Mukherjee S, Saini R, Kaushik P, Roy PP, Dogra DP (2018) Multimodal gait recognition with inertial sensor data and video using evolutionary algorithm. IEEE Trans Fuzzy Syst 27(5):956\u2013965","journal-title":"IEEE Trans Fuzzy Syst"},{"issue":"8","key":"2484_CR30","doi-asserted-by":"publisher","first-page":"1213","DOI":"10.1007\/s00138-016-0767-5","volume":"27","author":"FM Castro","year":"2016","unstructured":"Castro FM, Mar\u00edn-Jim\u00e9nez MJ, Guil N (2016) Multimodal features fusion for gait, gender and shoes recognition. Mach Vis Appl 27(8):1213\u20131228","journal-title":"Mach Vis Appl"},{"key":"2484_CR31","unstructured":"Zaheer M, Kottur S, Ravanbakhsh S, Poczos B, Salakhutdinov R, Smola A (2017) Deep sets. arXiv:1703.06114"},{"issue":"1","key":"2484_CR32","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1186\/s41074-018-0039-6","volume":"10","author":"N Takemura","year":"2018","unstructured":"Takemura N, Makihara Y, Muramatsu D, Echigo T, Yagi Y (2018) Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition. IPSJ Trans Comput Vis Appl 10(1):4","journal-title":"IPSJ Trans Comput Vis Appl"},{"key":"2484_CR33","doi-asserted-by":"crossref","unstructured":"Wei S-E, Ramakrishna V, Kanade T, Sheikh Y (2016) Convolutional pose machines. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4724\u20134732","DOI":"10.1109\/CVPR.2016.511"},{"key":"2484_CR34","doi-asserted-by":"crossref","unstructured":"Fu Y, Wei Y, Zhou Y, Shi H, Huang G, Wang X, Yao Z, Huang T (2019) Horizontal pyramid matching for person re-identification. In: Proceedings of the AAAI conference on artificial intelligence, pp 8295\u20138302","DOI":"10.1609\/aaai.v33i01.33018295"}],"container-title":["Applied Intelligence"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02484-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s10489-021-02484-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s10489-021-02484-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,1]],"date-time":"2024-09-01T02:57:06Z","timestamp":1725159426000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s10489-021-02484-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021,6,2]]},"references-count":34,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2022,1]]}},"alternative-id":["2484"],"URL":"https:\/\/doi.org\/10.1007\/s10489-021-02484-2","relation":{},"ISSN":["0924-669X","1573-7497"],"issn-type":[{"value":"0924-669X","type":"print"},{"value":"1573-7497","type":"electronic"}],"subject":[],"published":{"date-parts":[[2021,6,2]]},"assertion":[{"value":"28 April 2021","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 June 2021","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"We have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"<!--Emphasis Type='Bold' removed-->Conflict of Interests"}}]}}