{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,8]],"date-time":"2026-01-08T22:56:17Z","timestamp":1767912977722,"version":"3.49.0"},"publisher-location":"Cham","reference-count":34,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031463044","type":"print"},{"value":"9783031463051","type":"electronic"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-46305-1_17","type":"book-chapter","created":{"date-parts":[[2023,10,28]],"date-time":"2023-10-28T07:02:41Z","timestamp":1698476561000},"page":"204-215","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["HuMoMM: A Multi-Modal Dataset and\u00a0Benchmark for\u00a0Human Motion Analysis"],"prefix":"10.1007","author":[{"given":"Xiong","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Minghui","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ming","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Wenxiong","family":"Kang","sequence":"additional","affiliation":[]},{"given":"Feiqi","family":"Deng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,10,29]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., Zitnick, C.L.: Microsoft coco: Common objects in context. In: ECCV. pp. 740\u2013755. Springer (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"17_CR2","doi-asserted-by":"crossref","unstructured":"Ionescu, C., Papava, D., Olaru, V., Sminchisescu, C.: Human3. 6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. IEEE TPAMI 36(7), 1325\u20131339 (2013)","DOI":"10.1109\/TPAMI.2013.248"},{"key":"17_CR3","doi-asserted-by":"crossref","unstructured":"Joo, H., Liu, H., Tan, L., Gui, L., Nabbe, B., Matthews, I., Kanade, T., Nobuhara, S., Sheikh, Y.: Panoptic studio: A massively multiview system for social motion capture. In: IEEE ICCV. pp. 3334\u20133342 (2015)","DOI":"10.1109\/ICCV.2015.381"},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"Von Marcard, T., Henschel, R., Black, M.J., Rosenhahn, B., Pons-Moll, G.: Recovering accurate 3d human pose in the wild using imus and a moving camera. In: ECCV. pp. 601\u2013617 (2018)","DOI":"10.1007\/978-3-030-01249-6_37"},{"key":"17_CR5","doi-asserted-by":"crossref","unstructured":"Mahmood, N., Ghorbani, N., Troje, N.F., Pons-Moll, G., Black, M.J.: Amass: Archive of motion capture as surface shapes. In: IEEE ICCV. pp. 5442\u20135451 (2019)","DOI":"10.1109\/ICCV.2019.00554"},{"key":"17_CR6","doi-asserted-by":"crossref","unstructured":"Cai, Z., Ren, D., Zeng, A., Lin, Z., Yu, T., Wang, W., Fan, X., Gao, Y., Yu, Y., Pan, L., et al.: Humman: Multi-modal 4d human dataset for versatile sensing and modeling. In: ECCV. pp. 557\u2013577. Springer (2022)","DOI":"10.1007\/978-3-031-20071-7_33"},{"key":"17_CR7","unstructured":"Soomro, K., Zamir, A.R., Shah, M.: Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402 (2012)"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Gu, C., Sun, C., Ross, D.A., Vondrick, C., Pantofaru, C., Li, Y., Vijayanarasimhan, S., Toderici, G., Ricco, S., Sukthankar, R., et al.: Ava: A video dataset of spatio-temporally localized atomic visual actions. In: IEEE CVPR. pp. 6047\u20136056 (2018)","DOI":"10.1109\/CVPR.2018.00633"},{"key":"17_CR9","doi-asserted-by":"crossref","unstructured":"Shahroudy, A., Liu, J., Ng, T.T., Wang, G.: Ntu rgb+ d: A large scale dataset for 3d human activity analysis. In: IEEE CVPR. pp. 1010\u20131019 (2016)","DOI":"10.1109\/CVPR.2016.115"},{"issue":"10","key":"17_CR10","doi-asserted-by":"publisher","first-page":"2684","DOI":"10.1109\/TPAMI.2019.2916873","volume":"42","author":"J Liu","year":"2019","unstructured":"Liu, J., Shahroudy, A., Perez, M., Wang, G., Duan, L.Y., Kot, A.C.: Ntu rgb+ d 120: A large-scale benchmark for 3d human activity understanding. IEEE TPAMI 42(10), 2684\u20132701 (2019)","journal-title":"IEEE TPAMI"},{"key":"17_CR11","doi-asserted-by":"crossref","unstructured":"Andriluka, M., Pishchulin, L., Gehler, P., Schiele, B.: 2d human pose estimation: New benchmark and state of the art analysis. In: IEEE CVPR. pp. 3686\u20133693 (2014)","DOI":"10.1109\/CVPR.2014.471"},{"key":"17_CR12","doi-asserted-by":"crossref","unstructured":"Andriluka, M., Iqbal, U., Insafutdinov, E., Pishchulin, L., Milan, A., Gall, J., Schiele, B.: Posetrack: A benchmark for human pose estimation and tracking. In: IEEE CVPR. pp. 5167\u20135176 (2018)","DOI":"10.1109\/CVPR.2018.00542"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"Luvizon, D.C., Picard, D., Tabia, H.: 2d\/3d pose estimation and action recognition using multitask deep learning. In: IEEE CVPR. pp. 5137\u20135146 (2018)","DOI":"10.1109\/CVPR.2018.00539"},{"issue":"1\u20132","key":"17_CR14","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1007\/s11263-009-0273-6","volume":"87","author":"L Sigal","year":"2010","unstructured":"Sigal, L., Balan, A.O., Black, M.J.: Humaneva: Synchronized video and motion capture dataset and baseline algorithm for evaluation of articulated human motion. International journal of computer vision 87(1\u20132), 4 (2010)","journal-title":"International journal of computer vision"},{"key":"17_CR15","doi-asserted-by":"crossref","unstructured":"Mehta, D., Rhodin, H., Casas, D., Fua, P., Sotnychenko, O., Xu, W., Theobalt, C.: Monocular 3d human pose estimation in the wild using improved cnn supervision. In: international conference on 3D vision (3DV). pp. 506\u2013516. IEEE (2017)","DOI":"10.1109\/3DV.2017.00064"},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"Li, R., Yang, S., Ross, D.A., Kanazawa, A.: Ai choreographer: Music conditioned 3d dance generation with aist++. In: IEEE ICCV. pp. 13401\u201313412 (2021)","DOI":"10.1109\/ICCV48922.2021.01315"},{"issue":"6","key":"17_CR17","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2816795.2818013","volume":"34","author":"M Loper","year":"2015","unstructured":"Loper, M., Mahmood, N., Romero, J., Pons-Moll, G., Black, M.J.: Smpl: A skinned multi-person linear model. ACM transactions on graphics (TOG) 34(6), 1\u201316 (2015)","journal-title":"ACM transactions on graphics (TOG)"},{"issue":"6","key":"17_CR18","doi-asserted-by":"publisher","first-page":"220","DOI":"10.1145\/2661229.2661273","volume":"33","author":"M Loper","year":"2014","unstructured":"Loper, M., Mahmood, N., Black, M.J.: Mosh: motion and shape capture from sparse markers. ACM Trans. Graph. 33(6), 220\u20131 (2014)","journal-title":"ACM Trans. Graph."},{"issue":"11","key":"17_CR19","doi-asserted-by":"publisher","first-page":"1330","DOI":"10.1109\/34.888718","volume":"22","author":"Z Zhang","year":"2000","unstructured":"Zhang, Z.: A flexible new technique for camera calibration. IEEE Transactions on pattern analysis and machine intelligence 22(11), 1330\u20131334 (2000)","journal-title":"IEEE Transactions on pattern analysis and machine intelligence"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Sun, K., Xiao, B., Liu, D., Wang, J.: Deep high-resolution representation learning for human pose estimation. In: IEEE CVPR. pp. 5693\u20135703 (2019)","DOI":"10.1109\/CVPR.2019.00584"},{"key":"17_CR21","doi-asserted-by":"crossref","unstructured":"Calic, J., Izuierdo, E.: Efficient key-frame extraction and video analysis. In: Proceedings. International Conference on Information Technology: Coding and Computing. pp. 28\u201333. IEEE (2002)","DOI":"10.1109\/ITCC.2002.1000355"},{"key":"17_CR22","first-page":"41","volume":"15","author":"BP Wrobel","year":"2001","unstructured":"Wrobel, B.P.: Multiple view geometry in computer vision. K\u00fcnstliche Intell. 15, 41 (2001)","journal-title":"K\u00fcnstliche Intell."},{"key":"17_CR23","doi-asserted-by":"publisher","unstructured":"F\u00f6rstner, W., Wrobel, B.P.: Bundle Adjustment, pp. 643\u2013725. Springer International Publishing, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-11550-4_15, https:\/\/doi.org\/10.1007\/978-3-319-11550-4_15","DOI":"10.1007\/978-3-319-11550-4_15"},{"key":"17_CR24","unstructured":"Easymocap - make human motion capture easier. Github (2021), https:\/\/github.com\/zju3dv\/EasyMocap"},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Bogo, F., Kanazawa, A., Lassner, C., Gehler, P., Romero, J., Black, M.J.: Keep it smpl: Automatic estimation of 3d human pose and shape from a single image. In: ECCV. pp. 561\u2013578. Springer (2016)","DOI":"10.1007\/978-3-319-46454-1_34"},{"issue":"10","key":"17_CR26","doi-asserted-by":"publisher","first-page":"6981","DOI":"10.1109\/TPAMI.2021.3098052","volume":"44","author":"J Dong","year":"2021","unstructured":"Dong, J., Fang, Q., Jiang, W., Yang, Y., Huang, Q., Bao, H., Zhou, X.: Fast and robust multi-person 3d pose estimation and tracking from multiple views. IEEE TPAMI 44(10), 6981\u20136992 (2021)","journal-title":"IEEE TPAMI"},{"key":"17_CR27","doi-asserted-by":"crossref","unstructured":"Han, H., Zhang, M., Hou, M., Zhang, F., Wang, Z., Chen, E., Wang, H., Ma, J., Liu, Q.: Stgcn: a spatial-temporal aware graph learning method for poi recommendation. In: 2020 IEEE International Conference on Data Mining (ICDM). pp. 1052\u20131057. IEEE (2020)","DOI":"10.1109\/ICDM50108.2020.00124"},{"key":"17_CR28","doi-asserted-by":"crossref","unstructured":"Duan, H., Zhao, Y., Chen, K., Lin, D., Dai, B.: Revisiting skeleton-based action recognition. In: IEEE CVPR. pp. 2969\u20132978 (2022)","DOI":"10.1109\/CVPR52688.2022.00298"},{"key":"17_CR29","doi-asserted-by":"crossref","unstructured":"Osokin, D.: Real-time 2d multi-person pose estimation on cpu: Lightweight openpose. arXiv preprint arXiv:1811.12004 (2018)","DOI":"10.5220\/0007555407440748"},{"key":"17_CR30","doi-asserted-by":"crossref","unstructured":"Wang, Y., Li, M., Cai, H., Chen, W.M., Han, S.: Lite pose: Efficient architecture design for 2d human pose estimation. In: IEEE CVPR. pp. 13126\u201313136 (2022)","DOI":"10.1109\/CVPR52688.2022.01278"},{"key":"17_CR31","doi-asserted-by":"crossref","unstructured":"Osman, A.A., Bolkart, T., Black, M.J.: Star: Sparse trained articulated human body regressor. In: ECCV. pp. 598\u2013613. Springer (2020)","DOI":"10.1007\/978-3-030-58539-6_36"},{"key":"17_CR32","doi-asserted-by":"crossref","unstructured":"Pavllo, D., Feichtenhofer, C., Grangier, D., Auli, M.: 3d human pose estimation in video with temporal convolutions and semi-supervised training. In: IEEE CVPR. pp. 7753\u20137762 (2019)","DOI":"10.1109\/CVPR.2019.00794"},{"key":"17_CR33","doi-asserted-by":"crossref","unstructured":"Liu, J., Rojas, J., Li, Y., Liang, Z., Guan, Y., Xi, N., Zhu, H.: A graph attention spatio-temporal convolutional network for 3d human pose estimation in video. In: 2021 IEEE International Conference on Robotics and Automation (ICRA). pp. 3374\u20133380. IEEE (2021)","DOI":"10.1109\/ICRA48506.2021.9561605"},{"key":"17_CR34","doi-asserted-by":"crossref","unstructured":"Sun, Y., Bao, Q., Liu, W., Fu, Y., Black, M.J., Mei, T.: Monocular, one-stage, regression of multiple 3d people. In: IEEE ICCV. pp. 11179\u201311188 (2021)","DOI":"10.1109\/ICCV48922.2021.01099"}],"container-title":["Lecture Notes in Computer Science","Image and Graphics"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-46305-1_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,1]],"date-time":"2024-11-01T00:00:32Z","timestamp":1730419232000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-46305-1_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031463044","9783031463051"],"references-count":34,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-46305-1_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"29 October 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIG","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Image and Graphics","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Nanjing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 September 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"12","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icig2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/icig2023.csig.org.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Conference Management Toolkit","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"409","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"166","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"41% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}