{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T02:19:35Z","timestamp":1743041975073,"version":"3.40.3"},"publisher-location":"Cham","reference-count":34,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031533105"},{"type":"electronic","value":"9783031533112"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-53311-2_28","type":"book-chapter","created":{"date-parts":[[2024,1,27]],"date-time":"2024-01-27T21:37:36Z","timestamp":1706391456000},"page":"381-395","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["DFGait: Decomposition Fusion Representation Learning for\u00a0Multimodal Gait Recognition"],"prefix":"10.1007","author":[{"given":"Jianbo","family":"Xiong","sequence":"first","affiliation":[]},{"given":"Shinan","family":"Zou","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Tang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,1,28]]},"reference":[{"key":"28_CR1","doi-asserted-by":"crossref","unstructured":"Castro, F.M., Marin-Jimenez, M.J., Guil, N., P\u00e9rez de la Blanca, N.: Multimodal feature fusion for CNN-based gait recognition: an empirical comparison. Neural Comput. Appl. 32, 14173\u201314193 (2020)","DOI":"10.1007\/s00521-020-04811-z"},{"key":"28_CR2","doi-asserted-by":"crossref","unstructured":"Chao, H., He, Y., Zhang, J., Feng, J.: GaitSet: regarding gait as a set for cross-view gait recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 8126\u20138133 (2019)","DOI":"10.1609\/aaai.v33i01.33018126"},{"key":"28_CR3","doi-asserted-by":"crossref","unstructured":"Cui, Y., Kang, Y.: Multi-modal gait recognition via effective spatial-temporal feature fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17949\u201317957 (2023)","DOI":"10.1109\/CVPR52729.2023.01721"},{"issue":"1","key":"28_CR4","doi-asserted-by":"publisher","first-page":"625","DOI":"10.1109\/JIOT.2022.3203559","volume":"10","author":"L Deng","year":"2022","unstructured":"Deng, L., Yang, J., Yuan, S., Zou, H., Lu, C.X., Xie, L.: GaitFi: robust device-free human identification via WiFi and vision multimodal learning. IEEE Internet Things J. 10(1), 625\u2013636 (2022)","journal-title":"IEEE Internet Things J."},{"key":"28_CR5","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"28_CR6","doi-asserted-by":"crossref","unstructured":"Fan, C., Liang, J., Shen, C., Hou, S., Huang, Y., Yu, S.: OpenGait: revisiting gait recognition toward better practicality (2022)","DOI":"10.1109\/CVPR52729.2023.00936"},{"key":"28_CR7","doi-asserted-by":"crossref","unstructured":"Fan, C., et al.: GaitPart: temporal part-based model for gait recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14225\u201314233 (2020)","DOI":"10.1109\/CVPR42600.2020.01423"},{"key":"28_CR8","doi-asserted-by":"crossref","unstructured":"Guo, W., Huang, H., Kong, X., He, R.: Learning disentangled representation for cross-modal retrieval with deep mutual information estimation. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 1712\u20131720 (2019)","DOI":"10.1145\/3343031.3351053"},{"issue":"2","key":"28_CR9","doi-asserted-by":"publisher","first-page":"316","DOI":"10.1109\/TPAMI.2006.38","volume":"28","author":"J Han","year":"2005","unstructured":"Han, J., Bhanu, B.: Individual recognition using gait energy image. IEEE Trans. Pattern Anal. Mach. Intell. 28(2), 316\u2013322 (2005)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"28_CR10","unstructured":"Hermans, A., Beyer, L., Leibe, B.: In defense of the triplet loss for person re-identification. arXiv preprint arXiv:1703.07737 (2017)"},{"issue":"4","key":"28_CR11","first-page":"1187","volume":"8","author":"H Iwama","year":"2013","unstructured":"Iwama, H., Muramatsu, D., Makihara, Y., Yagi, Y.: Gait verification system for criminal investigation. Inf. Media Technol. 8(4), 1187\u20131199 (2013)","journal-title":"Inf. Media Technol."},{"key":"28_CR12","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"issue":"5","key":"28_CR13","doi-asserted-by":"publisher","first-page":"956","DOI":"10.1109\/TFUZZ.2018.2870590","volume":"27","author":"P Kumar","year":"2018","unstructured":"Kumar, P., Mukherjee, S., Saini, R., Kaushik, P., Roy, P.P., Dogra, D.P.: Multimodal gait recognition with inertial sensor data and video using evolutionary algorithm. IEEE Trans. Fuzzy Syst. 27(5), 956\u2013965 (2018)","journal-title":"IEEE Trans. Fuzzy Syst."},{"issue":"2","key":"28_CR14","doi-asserted-by":"publisher","first-page":"1535","DOI":"10.1007\/s10489-022-03543-y","volume":"53","author":"G Li","year":"2023","unstructured":"Li, G., Guo, L., Zhang, R., Qian, J., Gao, S.: TransGait: multimodal-based gait recognition with set transformer. Appl. Intell. 53(2), 1535\u20131547 (2023)","journal-title":"Appl. Intell."},{"key":"28_CR15","doi-asserted-by":"crossref","unstructured":"Li, X., Makihara, Y., Xu, C., Yagi, Y., Ren, M.: Gait recognition via semi-supervised disentangled representation learning to identity and covariate features. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13309\u201313319 (2020)","DOI":"10.1109\/CVPR42600.2020.01332"},{"key":"28_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2019.107069","volume":"98","author":"R Liao","year":"2020","unstructured":"Liao, R., Yu, S., An, W., Huang, Y.: A model-based gait recognition method with body pose and human prior knowledge. Pattern Recogn. 98, 107069 (2020)","journal-title":"Pattern Recogn."},{"key":"28_CR17","doi-asserted-by":"crossref","unstructured":"Lin, B., Zhang, S., Yu, X.: Gait recognition via effective global-local feature representation and local temporal aggregation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14648\u201314656 (2021)","DOI":"10.1109\/ICCV48922.2021.01438"},{"key":"28_CR18","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"issue":"2","key":"28_CR19","doi-asserted-by":"publisher","first-page":"47","DOI":"10.1049\/iet-bmt.2013.0090","volume":"3","author":"N Lynnerup","year":"2014","unstructured":"Lynnerup, N., Larsen, P.K.: Gait as evidence. IET Biometrics 3(2), 47\u201354 (2014)","journal-title":"IET Biometrics"},{"key":"28_CR20","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11) (2008)"},{"key":"28_CR21","doi-asserted-by":"publisher","DOI":"10.1016\/j.smhl.2020.100162","volume":"19","author":"I Papavasileiou","year":"2021","unstructured":"Papavasileiou, I., Qiao, Z., Zhang, C., Zhang, W., Bi, J., Han, S.: GaitCode: gait-based continuous authentication using multimodal learning and wearable sensors. Smart Health 19, 100162 (2021)","journal-title":"Smart Health"},{"key":"28_CR22","doi-asserted-by":"crossref","unstructured":"Peng, Y., Ma, K., Zhang, Y., He, Z.: Learning rich features for gait recognition by integrating skeletons and silhouettes. Multimedia Tools Appl., 1\u201322 (2023)","DOI":"10.1007\/s11042-023-15483-x"},{"key":"28_CR23","doi-asserted-by":"crossref","unstructured":"Qiu, Z., Yao, T., Mei, T.: Learning spatio-temporal representation with pseudo-3D residual networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5533\u20135541 (2017)","DOI":"10.1109\/ICCV.2017.590"},{"issue":"1","key":"28_CR24","doi-asserted-by":"publisher","first-page":"264","DOI":"10.1109\/TPAMI.2022.3151865","volume":"45","author":"A Sepas-Moghaddam","year":"2022","unstructured":"Sepas-Moghaddam, A., Etemad, A.: Deep gait recognition: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 45(1), 264\u2013284 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"28_CR25","doi-asserted-by":"crossref","unstructured":"Shiraga, K., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: GeiNet: view-invariant gait recognition using a convolutional neural network. In: 2016 International Conference on Biometrics (ICB), pp. 1\u20138. IEEE (2016)","DOI":"10.1109\/ICB.2016.7550060"},{"key":"28_CR26","doi-asserted-by":"crossref","unstructured":"Teepe, T., Khan, A., Gilg, J., Herzog, F., H\u00f6rmann, S., Rigoll, G.: GaitGraph: graph convolutional network for skeleton-based gait recognition. In: 2021 IEEE International Conference on Image Processing (ICIP), pp. 2314\u20132318. IEEE (2021)","DOI":"10.1109\/ICIP42928.2021.9506717"},{"key":"28_CR27","first-page":"4835","volume":"33","author":"Y Wang","year":"2020","unstructured":"Wang, Y., Huang, W., Sun, F., Xu, T., Rong, Y., Huang, J.: Deep multimodal fusion by channel exchanging. Adv. Neural. Inf. Process. Syst. 33, 4835\u20134845 (2020)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"28_CR28","doi-asserted-by":"crossref","unstructured":"Wu, X., Huang, H., Patel, V.M., He, R., Sun, Z.: Disentangled variational representation for heterogeneous face recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 9005\u20139012 (2019)","DOI":"10.1609\/aaai.v33i01.33019005"},{"key":"28_CR29","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"28_CR30","doi-asserted-by":"crossref","unstructured":"Zhang, Z., et al.: Gait recognition via disentangled representation learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4710\u20134719 (2019)","DOI":"10.1109\/CVPR.2019.00484"},{"issue":"2","key":"28_CR31","doi-asserted-by":"publisher","first-page":"2023","DOI":"10.1007\/s10489-021-02484-2","volume":"52","author":"L Zhao","year":"2022","unstructured":"Zhao, L., Guo, L., Zhang, R., Xie, X., Ye, X.: mmGaitSet: multimodal based gait recognition for countering carrying and clothing changes. Appl. Intell. 52(2), 2023\u20132036 (2022)","journal-title":"Appl. Intell."},{"key":"28_CR32","doi-asserted-by":"crossref","unstructured":"Zheng, J., et al.: Gait recognition in the wild with multi-hop temporal switch. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 6136\u20136145 (2022)","DOI":"10.1145\/3503161.3547897"},{"key":"28_CR33","doi-asserted-by":"crossref","unstructured":"Zheng, J., Liu, X., Liu, W., He, L., Yan, C., Mei, T.: Gait recognition in the wild with dense 3D representations and a benchmark. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20228\u201320237 (2022)","DOI":"10.1109\/CVPR52688.2022.01959"},{"key":"28_CR34","unstructured":"Zhu, Z., et al.: Gait recognition in the wild: a benchmark. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14789\u201314799 (2021)"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-53311-2_28","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,12]],"date-time":"2024-03-12T15:30:01Z","timestamp":1710257401000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-53311-2_28"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031533105","9783031533112"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-53311-2_28","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"28 January 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Amsterdam","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"The Netherlands","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 January 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 February 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ConfTool Pro","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"297","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"112","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"38% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}