{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,8]],"date-time":"2026-02-08T12:43:18Z","timestamp":1770554598955,"version":"3.49.0"},"publisher-location":"Cham","reference-count":52,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030585440","type":"print"},{"value":"9783030585457","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58545-7_17","type":"book-chapter","created":{"date-parts":[[2020,11,4]],"date-time":"2020-11-04T10:04:51Z","timestamp":1604484291000},"page":"284-300","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":33,"title":["3D Human Shape and Pose from a Single Low-Resolution Image with Self-Supervised Learning"],"prefix":"10.1007","author":[{"given":"Xiangyu","family":"Xu","sequence":"first","affiliation":[]},{"given":"Hao","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Francesc","family":"Moreno-Noguer","sequence":"additional","affiliation":[]},{"given":"L\u00e1szl\u00f3 A.","family":"Jeni","sequence":"additional","affiliation":[]},{"given":"Fernando","family":"De la Torre","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,11,5]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Alldieck, T., Magnor, M., Bhatnagar, B.L., Theobalt, C., Pons-Moll, G.: Learning to reconstruct people in clothing from a single rgb camera. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00127"},{"key":"17_CR2","doi-asserted-by":"crossref","unstructured":"Alldieck, T., Magnor, M., Xu, W., Theobalt, C., Pons-Moll, G.: Video based reconstruction of 3d people models. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00875"},{"key":"17_CR3","doi-asserted-by":"crossref","unstructured":"Alldieck, T., Pons-Moll, G., Theobalt, C., Magnor, M.: Tex2shape: Detailed full human body geometry from a single image. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00238"},{"key":"17_CR4","doi-asserted-by":"crossref","unstructured":"Andriluka, M., Pishchulin, L., Gehler, P., Schiele, B.: 2d human pose estimation: New benchmark and state of the art analysis. In: CVPR (2014)","DOI":"10.1109\/CVPR.2014.471"},{"key":"17_CR5","unstructured":"Arjovsky, M., Chintala, S., Bottou, L.: Wasserstein generative adversarial networks. In: ICML (2017)"},{"key":"17_CR6","doi-asserted-by":"crossref","unstructured":"Bogo, F., Kanazawa, A., Lassner, C., Gehler, P., Romero, J., Black, M.J.: Keep it smpl: Automatic estimation of 3d human pose and shape from a single image. In: ECCV (2016)","DOI":"10.1007\/978-3-319-46454-1_34"},{"key":"17_CR7","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: ICML (2020)"},{"key":"17_CR8","doi-asserted-by":"crossref","unstructured":"Cheng, Z., Zhu, X., Gong, S.: Low-resolution face recognition. In: ACCV (2018)","DOI":"10.1007\/978-3-030-20893-6_38"},{"key":"17_CR9","unstructured":"Doersch, C., Zisserman, A.: Sim2real transfer learning for 3d human pose estimation: motion to the rescue. In: NeurIPS (2019)"},{"issue":"4","key":"17_CR10","first-page":"2051","volume":"28","author":"S Ge","year":"2018","unstructured":"Ge, S., Zhao, S., Li, C., Li, J.: Low-resolution face recognition in the wild via selective knowledge distillation. TIP 28(4), 2051\u20132062 (2018)","journal-title":"TIP"},{"key":"17_CR11","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: NIPS (2014)"},{"key":"17_CR12","unstructured":"Haris, M., Shakhnarovich, G., Ukita, N.: Task-driven super resolution: Object detection in low-resolution images. arXiv:1803.11316 (2018)"},{"key":"17_CR13","doi-asserted-by":"crossref","unstructured":"Hartley, R., Zisserman, A.: Multiple view geometry in computer vision. Cambridge University Press (2003)","DOI":"10.1017\/CBO9780511811685"},{"key":"17_CR14","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"17_CR15","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Identity mappings in deep residual networks. In: ECCV (2016)","DOI":"10.1007\/978-3-319-46493-0_38"},{"key":"17_CR17","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv:1503.02531 (2015)"},{"issue":"7","key":"17_CR18","doi-asserted-by":"publisher","first-page":"1325","DOI":"10.1109\/TPAMI.2013.248","volume":"36","author":"C Ionescu","year":"2013","unstructured":"Ionescu, C., Papava, D., Olaru, V., Sminchisescu, C.: Human3.6m: Large scale datasets and predictive methods for 3d human sensing in natural environments. TPAMI 36(7), 1325\u20131339 (2013)","journal-title":"TPAMI"},{"key":"17_CR19","doi-asserted-by":"crossref","unstructured":"Johnson, S., Everingham, M.: Clustered pose and nonlinear appearance models for human pose estimation. In: BMVC (2010)","DOI":"10.5244\/C.24.12"},{"key":"17_CR20","doi-asserted-by":"crossref","unstructured":"Johnson, S., Everingham, M.: Learning effective human pose estimation from inaccurate annotation. In: CVPR (2011)","DOI":"10.1109\/CVPR.2011.5995318"},{"key":"17_CR21","doi-asserted-by":"crossref","unstructured":"Kanazawa, A., Black, M.J., Jacobs, D.W., Malik, J.: End-to-end recovery of human shape and pose. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00744"},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Kanazawa, A., Zhang, J.Y., Felsen, P., Malik, J.: Learning 3d human dynamics from video. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00576"},{"key":"17_CR23","unstructured":"Kingma, D., Ba, J.: Adam: A method for stochastic optimization. In: ICLR (2014)"},{"key":"17_CR24","doi-asserted-by":"crossref","unstructured":"Kocabas, M., Athanasiou, N., Black, M.J.: Vibe: Video inference for human body pose and shape estimation. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00530"},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Kolotouros, N., Pavlakos, G., Black, M.J., Daniilidis, K.: Learning to reconstruct 3d human pose and shape via model-fitting in the loop. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00234"},{"key":"17_CR26","unstructured":"Laine, S., Aila, T.: Temporal ensembling for semi-supervised learning. In: ICLR (2017)"},{"key":"17_CR27","doi-asserted-by":"crossref","unstructured":"Li, J., Liang, X., Wei, Y., Xu, T., Feng, J., Yan, S.: Perceptual generative adversarial networks for small object detection. In: CVPR (2017)","DOI":"10.1109\/CVPR.2017.211"},{"key":"17_CR28","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., et al.: Microsoft coco: Common objects in context. In: ECCV (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"issue":"6","key":"17_CR29","doi-asserted-by":"publisher","first-page":"248","DOI":"10.1145\/2816795.2818013","volume":"34","author":"M Loper","year":"2015","unstructured":"Loper, M., Mahmood, N., Romero, J., Pons-Moll, G., Black, M.J.: Smpl: A skinned multi-person linear model. ACM Trans. Graph. 34(6), 248 (2015)","journal-title":"ACM Trans. Graph."},{"key":"17_CR30","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul Smolley, S.: Least squares generative adversarial networks. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"17_CR31","doi-asserted-by":"crossref","unstructured":"von Marcard, T., Henschel, R., Black, M.J., Rosenhahn, B., Pons-Moll, G.: Recovering accurate 3d human pose in the wild using imus and a moving camera. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01249-6_37"},{"key":"17_CR32","doi-asserted-by":"crossref","unstructured":"Mehta, D., et al.: Monocular 3d human pose estimation in the wild using improved cnn supervision. In: 3DV (2017)","DOI":"10.1109\/3DV.2017.00064"},{"key":"17_CR33","unstructured":"Nair, V., Hinton, G.E.: Rectified linear units improve restricted boltzmann machines. In: ICML (2010)"},{"key":"17_CR34","doi-asserted-by":"crossref","unstructured":"Natsume, R., et al.: Siclope: Silhouette-based clothed people. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00461"},{"key":"17_CR35","doi-asserted-by":"crossref","unstructured":"Neumann, L., Vedaldi, A.: Tiny people pose. In: ACCV (2018)","DOI":"10.1007\/978-3-030-20893-6_35"},{"key":"17_CR36","unstructured":"Nishibori, K., Takahashi, T., Deguchi, D., Ide, I., Murase, H.: Exemplar-based human body super-resolution for surveillance camera systems. In: International Conference on Computer Vision Theory and Applications (VISAPP) (2014)"},{"key":"17_CR37","doi-asserted-by":"crossref","unstructured":"Noh, J., Bae, W., Lee, W., Seo, J., Kim, G.: Better to follow, follow to be better: Towards precise supervision of feature super-resolution for small object detection. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00982"},{"key":"17_CR38","doi-asserted-by":"crossref","unstructured":"Oh, S., et al.: A large-scale benchmark dataset for event recognition in surveillance video. In: CVPR (2011)","DOI":"10.1109\/CVPR.2011.5995586"},{"key":"17_CR39","unstructured":"Oord, A.v.d., Li, Y., Vinyals, O.: Representation learning with contrastive predictive coding. arXiv:1807.03748 (2018)"},{"key":"17_CR40","doi-asserted-by":"crossref","unstructured":"Pavlakos, G., Zhu, L., Zhou, X., Daniilidis, K.: Learning to estimate 3d human pose and shape from a single color image. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00055"},{"key":"17_CR41","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Sanchez-Riera, J., Choi, G., Sanfeliu, A., Moreno-Noguer, F.: 3dpeople: Modeling the geometry of dressed humans. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00233"},{"key":"17_CR42","doi-asserted-by":"crossref","unstructured":"Saito, S., Huang, Z., Natsume, R., Morishima, S., Kanazawa, A., Li, H.: Pifu: Pixel-aligned implicit function for high-resolution clothed human digitization. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00239"},{"key":"17_CR43","doi-asserted-by":"crossref","unstructured":"Tan, W., Yan, B., Bare, B.: Feature super-resolution: Make machine see more clearly. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00420"},{"key":"17_CR44","unstructured":"Tarvainen, A., Valpola, H.: Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. In: NIPS (2017)"},{"key":"17_CR45","doi-asserted-by":"crossref","unstructured":"Tian, Y., Krishnan, D., Isola, P.: Contrastive multiview coding. arXiv preprint arXiv:1906.05849 (2019)","DOI":"10.1007\/978-3-030-58621-8_45"},{"key":"17_CR46","doi-asserted-by":"crossref","unstructured":"Wang, Z., Chang, S., Yang, Y., Liu, D., Huang, T.S.: Studying very low resolution recognition using deep networks. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.518"},{"key":"17_CR47","doi-asserted-by":"crossref","unstructured":"Xu, X., Ma, Y., Sun, W.: Towards real scene super-resolution with raw images. In: CVPR (2019)","DOI":"10.1109\/CVPR.2019.00182"},{"key":"17_CR48","doi-asserted-by":"crossref","unstructured":"Xu, X., Sun, D., Pan, J., Zhang, Y., Pfister, H., Yang, M.H.: Learning to super-resolve blurry face and text images. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.36"},{"key":"17_CR49","doi-asserted-by":"crossref","unstructured":"Zanfir, A., Marinoiu, E., Sminchisescu, C.: Monocular 3d pose and shape estimation of multiple people in natural scenes-the importance of multiple scene constraints. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00229"},{"key":"17_CR50","doi-asserted-by":"crossref","unstructured":"Zhang, J.Y., Felsen, P., Kanazawa, A., Malik, J.: Predicting 3d human dynamics from video. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00721"},{"key":"17_CR51","doi-asserted-by":"crossref","unstructured":"Zhang, Y., Tian, Y., Kong, Y., Zhong, B., Fu, Y.: Residual dense network for image super-resolution. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00262"},{"key":"17_CR52","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Yu, T., Wei, Y., Dai, Q., Liu, Y.: Deephuman: 3d human reconstruction from a single image. In: ICCV (2019)","DOI":"10.1109\/ICCV.2019.00783"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58545-7_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,4]],"date-time":"2024-11-04T01:10:44Z","timestamp":1730682644000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58545-7_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030585440","9783030585457"],"references-count":52,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58545-7_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"5 November 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic. From the ECCV Workshops 249 full papers, 18 short papers, and 21 further contributions were published out of a total of 467 submissions.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}