{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,25]],"date-time":"2026-04-25T03:37:47Z","timestamp":1777088267013,"version":"3.51.4"},"reference-count":51,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T00:00:00Z","timestamp":1718409600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T00:00:00Z","timestamp":1718409600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,10]]},"DOI":"10.1007\/s00371-024-03426-y","type":"journal-article","created":{"date-parts":[[2024,6,15]],"date-time":"2024-06-15T03:02:42Z","timestamp":1718420562000},"page":"7221-7234","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["MCDGait: multimodal co-learning distillation network with spatial-temporal graph reasoning for gait recognition in the wild"],"prefix":"10.1007","volume":"40","author":[{"given":"Jianbo","family":"Xiong","sequence":"first","affiliation":[]},{"given":"Shinan","family":"Zou","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Tardi","family":"Tjahjadi","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,6,15]]},"reference":[{"key":"3426_CR1","doi-asserted-by":"crossref","unstructured":"Gao, L., Hu, L., Lyu, F., Zhu, L., Wan, L., Pun, C.M., Feng, W.: Difference-guided multi-scale spatial-temporal representation for sign language recognition. Vis. Comput. 39(8), 3417\u20133428 (2023)","DOI":"10.1007\/s00371-023-02979-8"},{"key":"3426_CR2","doi-asserted-by":"publisher","unstructured":"Dong, Y., Yu, C., Ha, R., Shi, Y., Ma, Y., Xu, L., Fu, Y., Wang, J.: HybridGait: A benchmark for spatial-temporal cloth-changing gait recognition with hybrid explorations. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 38(2), pp. 1600\u20131608 (2024). https:\/\/doi.org\/10.1609\/aaai.v38i2.27926","DOI":"10.1609\/aaai.v38i2.27926"},{"key":"3426_CR3","doi-asserted-by":"crossref","unstructured":"Teepe, T., Khan, A., Gilg, J., Herzog, F., H\u00f6rmann, S., Rigoll, G.: GaitGraph: graph convolutional network for skeleton-based gait recognition. In: 2021 IEEE International Conference on Image Processing (ICIP), pp. 2314\u20132318. IEEE (2021)","DOI":"10.1109\/ICIP42928.2021.9506717"},{"key":"3426_CR4","doi-asserted-by":"crossref","unstructured":"Shiraga, K., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: GeiNet: view-invariant gait recognition using a convolutional neural network. In: 2016 International Conference on Biometrics (ICB), pp. 1\u20138. IEEE (2016)","DOI":"10.1109\/ICB.2016.7550060"},{"key":"3426_CR5","doi-asserted-by":"crossref","unstructured":"Chao, H., He, Y., Zhang, J., Feng, J.: GaitSet: regarding gait as a set for cross-view gait recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 33, pp. 8126\u20138133 (2019)","DOI":"10.1609\/aaai.v33i01.33018126"},{"key":"3426_CR6","doi-asserted-by":"crossref","unstructured":"Fan, C., Peng, Y., Cao, C., Liu, X., Hou, S., Chi, J., Huang, Y., Li, Q., He, Z.: GaitPart: temporal part-based model for gait recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 14225\u201314233 (2020)","DOI":"10.1109\/CVPR42600.2020.01423"},{"key":"3426_CR7","doi-asserted-by":"crossref","unstructured":"Hou, S., Cao, C., Liu, X., Huang, Y.: Gait lateral network: learning discriminative and compact representations for gait recognition. In: Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part IX, pp. 382\u2013398. Springer (2020)","DOI":"10.1007\/978-3-030-58545-7_22"},{"key":"3426_CR8","doi-asserted-by":"crossref","unstructured":"Lin, B., Zhang, S., Yu, X.: Gait recognition via effective global-local feature representation and local temporal aggregation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14648\u201314656 (2021)","DOI":"10.1109\/ICCV48922.2021.01438"},{"key":"3426_CR9","doi-asserted-by":"crossref","unstructured":"Huang, X., Zhu, D., Wang, H., Wang, X., Yang, B., He, B., Liu, W., Feng, B.: Context-sensitive temporal feature learning for gait recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 12909\u201312918 (2021)","DOI":"10.1109\/ICCV48922.2021.01267"},{"key":"3426_CR10","doi-asserted-by":"crossref","unstructured":"Huang, Z., Xue, D., Shen, X., Tian, X., Li, H., Huang, J., Hua, X.-S.: 3D local convolutional neural networks for gait recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14920\u201314929 (2021)","DOI":"10.1109\/ICCV48922.2021.01465"},{"key":"3426_CR11","doi-asserted-by":"crossref","unstructured":"Lin, B., Zhang, S., Bao, F.: Gait recognition with multiple-temporal-scale 3D convolutional neural network. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 3054\u20133062 (2020)","DOI":"10.1145\/3394171.3413861"},{"key":"3426_CR12","unstructured":"Yu, S., Tan, D., Tan, T.: A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition. In: 18th International Conference on Pattern Recognition (ICPR\u201906), vol. 4, pp. 441\u2013444. IEEE (2006)"},{"key":"3426_CR13","doi-asserted-by":"crossref","first-page":"1","DOI":"10.1186\/s41074-017-0037-0","volume":"10","author":"N Takemura","year":"2018","unstructured":"Takemura, N., Makihara, Y., Muramatsu, D., Echigo, T., Yagi, Y.: Multi-view large population gait dataset and its performance evaluation for cross-view gait recognition. IPSJ Trans. Comput. Vis. Appl. 10, 1\u201314 (2018)","journal-title":"IPSJ Trans. Comput. Vis. Appl."},{"key":"3426_CR14","doi-asserted-by":"crossref","unstructured":"Zheng, J., Liu, X., Liu, W., He, L., Yan, C., Mei, T.: Gait recognition in the wild with dense 3D representations and a benchmark. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 20228\u201320237 (2022)","DOI":"10.1109\/CVPR52688.2022.01959"},{"key":"3426_CR15","unstructured":"Zhu, Z., Guo, X., Yang, T., Huang, J., Deng, J., Huang, G., Du, D., Lu, J., Zhou, J.: Gait recognition in the wild: a benchmark. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 14789\u201314799 (2021)"},{"key":"3426_CR16","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)"},{"key":"3426_CR17","unstructured":"Tian, Y., Krishnan, D., Isola, P.: Contrastive representation distillation. arXiv preprint arXiv:1910.10699 (2019)"},{"key":"3426_CR18","doi-asserted-by":"publisher","first-page":"4499","DOI":"10.1109\/TNNLS.2021.3116209","volume":"34","author":"Z Xie","year":"2021","unstructured":"Xie, Z., Zhang, W., Sheng, B., Li, P., Chen, C.P.: BaGFN: broad attentive graph fusion network for high-order feature interactions. IEEE Trans. Neural Netw. Learn. Syst. 34, 4499\u20134513 (2021)","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"3426_CR19","doi-asserted-by":"crossref","unstructured":"Wang, W., Lu, X., Shen, J., Crandall, D.J., Shao, L.: Zero-shot video object segmentation via attentive graph neural networks. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9236\u20139245 (2019)","DOI":"10.1109\/ICCV.2019.00933"},{"key":"3426_CR20","doi-asserted-by":"crossref","unstructured":"Fan, L., Wang, W., Huang, S., Tang, X., Zhu, S.-C.: Understanding human gaze communication by spatio-temporal graph reasoning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5724\u20135733 (2019)","DOI":"10.1109\/ICCV.2019.00582"},{"key":"3426_CR21","doi-asserted-by":"crossref","unstructured":"Qi, S., Wang, W., Jia, B., Shen, J., Zhu, S.-C.: Learning human-object interactions by graph parsing neural networks. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 401\u2013417 (2018)","DOI":"10.1007\/978-3-030-01240-3_25"},{"key":"3426_CR22","doi-asserted-by":"crossref","unstructured":"Wang, W., Zhu, H., Dai, J., Pang, Y., Shen, J., Shao, L.: Hierarchical human parsing with typed part-relation reasoning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8929\u20138939 (2020)","DOI":"10.1109\/CVPR42600.2020.00895"},{"key":"3426_CR23","unstructured":"Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)"},{"key":"3426_CR24","doi-asserted-by":"crossref","unstructured":"Qiu, Z., Yao, T., Mei, T.: Learning spatio-temporal representation with pseudo-3D residual networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5533\u20135541 (2017)","DOI":"10.1109\/ICCV.2017.590"},{"key":"3426_CR25","doi-asserted-by":"publisher","first-page":"14173","DOI":"10.1007\/s00521-020-04811-z","volume":"32","author":"FM Castro","year":"2020","unstructured":"Castro, F.M., Marin-Jimenez, M.J., Guil, N., Blanca, N.: Multimodal feature fusion for CNN-based gait recognition: an empirical comparison. Neural Comput. Appl. 32, 14173\u201314193 (2020)","journal-title":"Neural Comput. Appl."},{"issue":"2","key":"3426_CR26","doi-asserted-by":"publisher","first-page":"1535","DOI":"10.1007\/s10489-022-03543-y","volume":"53","author":"G Li","year":"2023","unstructured":"Li, G., Guo, L., Zhang, R., Qian, J., Gao, S.: TransGait: multimodal-based gait recognition with set transformer. Appl. Intell. 53(2), 1535\u20131547 (2023)","journal-title":"Appl. Intell."},{"key":"3426_CR27","doi-asserted-by":"publisher","first-page":"100162","DOI":"10.1016\/j.smhl.2020.100162","volume":"19","author":"I Papavasileiou","year":"2021","unstructured":"Papavasileiou, I., Qiao, Z., Zhang, C., Zhang, W., Bi, J., Han, S.: GaitCode: gait-based continuous authentication using multimodal learning and wearable sensors. Smart Health 19, 100162 (2021)","journal-title":"Smart Health"},{"issue":"5","key":"3426_CR28","doi-asserted-by":"publisher","first-page":"956","DOI":"10.1109\/TFUZZ.2018.2870590","volume":"27","author":"P Kumar","year":"2018","unstructured":"Kumar, P., Mukherjee, S., Saini, R., Kaushik, P., Roy, P.P., Dogra, D.P.: Multimodal gait recognition with inertial sensor data and video using evolutionary algorithm. IEEE Trans. Fuzzy Syst. 27(5), 956\u2013965 (2018)","journal-title":"IEEE Trans. Fuzzy Syst."},{"key":"3426_CR29","doi-asserted-by":"publisher","first-page":"5452","DOI":"10.1109\/TIFS.2021.3132579","volume":"16","author":"MJ Mar\u00edn-Jim\u00e9nez","year":"2021","unstructured":"Mar\u00edn-Jim\u00e9nez, M.J., Castro, F.M., Delgado-Esca\u00f1o, R., Kalogeiton, V., Guil, N.: UGaitNet: multimodal gait recognition with missing input modalities. IEEE Trans. Inf. Forensics Secur. 16, 5452\u20135462 (2021)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"key":"3426_CR30","doi-asserted-by":"publisher","first-page":"203","DOI":"10.1016\/j.inffus.2021.12.003","volume":"81","author":"A Rahate","year":"2022","unstructured":"Rahate, A., Walambe, R., Ramanna, S., Kotecha, K.: Multimodal co-learning: challenges, applications with datasets, recent advances and future directions. Inf Fusion 81, 203\u2013239 (2022)","journal-title":"Inf Fusion"},{"key":"3426_CR31","doi-asserted-by":"publisher","first-page":"140426","DOI":"10.1109\/ACCESS.2020.3006563","volume":"8","author":"S Seo","year":"2020","unstructured":"Seo, S., Na, S., Kim, J.: HMTL: heterogeneous modality transfer learning for audio-visual sentiment analysis. IEEE Access 8, 140426\u2013140437 (2020)","journal-title":"IEEE Access"},{"issue":"2","key":"3426_CR32","doi-asserted-by":"publisher","first-page":"117","DOI":"10.1109\/TETCI.2017.2784878","volume":"2","author":"J-C Hou","year":"2018","unstructured":"Hou, J.-C., Wang, S.-S., Lai, Y.-H., Tsao, Y., Chang, H.-W., Wang, H.-M.: Audio-visual speech enhancement using multimodal deep convolutional neural networks. IEEE Trans. Emerging Top. Comput. Intell. 2(2), 117\u2013128 (2018)","journal-title":"IEEE Trans. Emerging Top. Comput. Intell."},{"issue":"4","key":"3426_CR33","doi-asserted-by":"publisher","first-page":"1083","DOI":"10.1109\/TMI.2022.3223683","volume":"42","author":"R Liu","year":"2022","unstructured":"Liu, R., Wang, T., Li, H., Zhang, P., Li, J., Yang, X., Shen, D., Sheng, B.: TMM-Nets: transferred multi-to mono-modal generation for lupus retinopathy diagnosis. IEEE Trans. Med. Imaging 42(4), 1083\u20131094 (2022)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"3426_CR34","doi-asserted-by":"crossref","unstructured":"Mao, Y., Zhou, W., Lu, Z., Deng, J., Li, H.: CMD: self-supervised 3d action representation learning with cross-modal mutual distillation. In: Computer Vision\u2013ECCV 2022: 17th European Conference, Tel Aviv, Israel, October 23\u201327, 2022, Proceedings, Part III, pp. 734\u2013752. Springer (2022)","DOI":"10.1007\/978-3-031-20062-5_42"},{"key":"3426_CR35","doi-asserted-by":"crossref","unstructured":"Pei, Y., Huang, T., Ipenburg, W., Pechenizkiy, M.: ResGCN: attention-based deep residual modeling for anomaly detection on attributed networks. In: 2021 IEEE 8th International Conference on Data Science and Advanced Analytics (DSAA), pp. 1\u20132 (2021). IEEE","DOI":"10.1109\/DSAA53316.2021.9564233"},{"key":"3426_CR36","unstructured":"Han, K., Wang, Y., Guo, J., Tang, Y., Wu, E.: Vision GNN: an image is worth graph of nodes. arXiv preprint arXiv:2206.00272 (2022)"},{"key":"3426_CR37","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, Y., Lin, D.: Spatial temporal graph convolutional networks for skeleton-based action recognition. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 32 (2018)","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"3426_CR38","doi-asserted-by":"crossref","unstructured":"Li, G., Muller, M., Thabet, A., Ghanem, B.: DeepGCNs: Can GCNs go as deep as CNNs? In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9267\u20139276 (2019)","DOI":"10.1109\/ICCV.2019.00936"},{"key":"3426_CR39","unstructured":"Hermans, A., Beyer, L., Leibe, B.: In defense of the triplet loss for person re-identification. arXiv preprint arXiv:1703.07737 (2017)"},{"key":"3426_CR40","doi-asserted-by":"crossref","unstructured":"Park, W., Kim, D., Lu, Y., Cho, M.: Relational knowledge distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3967\u20133976 (2019)","DOI":"10.1109\/CVPR.2019.00409"},{"key":"3426_CR41","doi-asserted-by":"crossref","unstructured":"Peng, B., Jin, X., Liu, J., Li, D., Wu, Y., Liu, Y., Zhou, S., Zhang, Z.: Correlation congruence for knowledge distillation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5007\u20135016 (2019)","DOI":"10.1109\/ICCV.2019.00511"},{"key":"3426_CR42","doi-asserted-by":"crossref","unstructured":"Zhu, J., Tang, S., Chen, D., Yu, S., Liu, Y., Rong, M., Yang, A., Wang, X.: Complementary relation contrastive distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9260\u20139269 (2021)","DOI":"10.1109\/CVPR46437.2021.00914"},{"key":"3426_CR43","unstructured":"Jang, E., Gu, S., Poole, B.: Categorical reparameterization with Gumbel-Softmax. arXiv preprint arXiv:1611.01144 (2016)"},{"key":"3426_CR44","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft COCO: common objects in context. In: Computer Vision\u2013ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6\u201312, 2014, Proceedings, Part V 13, pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"3426_CR45","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"3426_CR46","doi-asserted-by":"crossref","unstructured":"Fan, C., Liang, J., Shen, C., Hou, S., Huang, Y., Yu, S.: OpenGait: revisiting gait recognition toward better practicality. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (2022)","DOI":"10.1109\/CVPR52729.2023.00936"},{"key":"3426_CR47","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L. et al.: PyTorch: an imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"3426_CR48","doi-asserted-by":"crossref","unstructured":"Zheng, J., Liu, X., Gu, X., Sun, Y., Gan, C., Zhang, J., Liu, W., Yan, C.: Gait recognition in the wild with multi-hop temporal switch. In: Proceedings of the 30th ACM International Conference on Multimedia, pp. 6136\u20136145 (2022)","DOI":"10.1145\/3503161.3547897"},{"key":"3426_CR49","doi-asserted-by":"crossref","unstructured":"Ma, K., Fu, Y., Zheng, D., Cao, C., Hu, X., Huang, Y.: Dynamic aggregated network for gait recognition. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 22076\u201322085 (2023)","DOI":"10.1109\/CVPR52729.2023.02114"},{"key":"3426_CR50","doi-asserted-by":"crossref","unstructured":"Zhu, H., Zheng, W., Zheng, Z., Nevatia, R.: GaitRef: gait recognition with refined sequential skeletons. In: 2023 IEEE International Joint Conference on Biometrics (IJCB), pp. 1\u201310 (2023). IEEE","DOI":"10.1109\/IJCB57857.2023.10448634"},{"issue":"11","key":"3426_CR51","first-page":"2579","volume":"9","author":"L Maaten","year":"2008","unstructured":"Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11), 2579\u20132605 (2008)","journal-title":"J. Mach. Learn. Res."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03426-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-024-03426-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-024-03426-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T06:17:32Z","timestamp":1727677052000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-024-03426-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,6,15]]},"references-count":51,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2024,10]]}},"alternative-id":["3426"],"URL":"https:\/\/doi.org\/10.1007\/s00371-024-03426-y","relation":{"has-preprint":[{"id-type":"doi","id":"10.21203\/rs.3.rs-4037908\/v1","asserted-by":"object"}]},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,6,15]]},"assertion":[{"value":"21 April 2024","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 June 2024","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"We strictly adhere to the application protocols for public datasets (Gait3D and GREW). The data are used for academic research only and are not copied or sold. In addition, we adhere to both public datasets\u2019 ethics and privacy statements.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}}]}}