{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T17:46:31Z","timestamp":1772905591738,"version":"3.50.1"},"publisher-location":"Cham","reference-count":26,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030208721","type":"print"},{"value":"9783030208738","type":"electronic"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-20873-8_36","type":"book-chapter","created":{"date-parts":[[2019,5,25]],"date-time":"2019-05-25T16:32:03Z","timestamp":1558801923000},"page":"562-576","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":114,"title":["VIPL-HR: A Multi-modal Database for Pulse Estimation from Less-Constrained Face Video"],"prefix":"10.1007","author":[{"given":"Xuesong","family":"Niu","sequence":"first","affiliation":[]},{"given":"Hu","family":"Han","sequence":"additional","affiliation":[]},{"given":"Shiguang","family":"Shan","sequence":"additional","affiliation":[]},{"given":"Xilin","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,5,26]]},"reference":[{"key":"36_CR1","doi-asserted-by":"crossref","unstructured":"Balakrishnan, G., Durand, F., Guttag, J.: Detecting pulse from head motions in video. In: Proceedings of IEEE CVPR, pp. 3430\u20133437. IEEE (2013)","DOI":"10.1109\/CVPR.2013.440"},{"key":"36_CR2","doi-asserted-by":"crossref","unstructured":"Hsu, G.-S., Ambikapathi, A., Chen, M.S.C.: Deep learning with time-frequency representation for pulse estimation. In: Proceedings of IJCB (2017)","DOI":"10.1109\/BTAS.2017.8272721"},{"issue":"10","key":"36_CR3","doi-asserted-by":"publisher","first-page":"2878","DOI":"10.1109\/TBME.2013.2266196","volume":"60","author":"G Haan de","year":"2013","unstructured":"de Haan, G., Jeanne, V.: Robust pulse rate from chrominance-based rPPG. IEEE Trans. Biomed. Eng. 60(10), 2878\u20132886 (2013)","journal-title":"IEEE Trans. Biomed. Eng."},{"key":"36_CR4","doi-asserted-by":"crossref","unstructured":"Han, H., Jain, A.K.: 3D face texture modeling from uncalibrated frontal and profile images. In: Proceedings of IEEE BTAS, pp. 223\u2013230. IEEE (2012)","DOI":"10.1109\/BTAS.2012.6374581"},{"issue":"11","key":"36_CR5","doi-asserted-by":"publisher","first-page":"2597","DOI":"10.1109\/TPAMI.2017.2738004","volume":"40","author":"H Han","year":"2017","unstructured":"Han, H., Jain, A.K., Shan, S., Chen, X.: Heterogeneous face attribute estimation: a deep multi-task learning approach. IEEE Trans. Pattern Anal. Mach. Intell. 40(11), 2597\u20132609 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"36_CR6","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of IEEE CVPR, pp. 770\u2013778. IEEE (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"36_CR7","unstructured":"Kwon, S., Kim, J., Lee, D., Park, K.: ROI analysis for remote photoplethysmography on facial video. In: Proceedings of EMBS, pp. 851\u2013862. IEEE (2015)"},{"key":"36_CR8","doi-asserted-by":"crossref","unstructured":"Lam, A., Kuno, Y.: Robust heart rate measurement from video using select random patches. In: Proceedings of IEEE ICCV, pp. 3640\u20133648. IEEE (2015)","DOI":"10.1109\/ICCV.2015.415"},{"key":"36_CR9","doi-asserted-by":"crossref","unstructured":"Li, X., et al.: The OBF database: a large face video database for remote physiological signal measurement and atrial fibrillation detection. In: Proceedings of IEEE FG, pp. 242\u2013249. IEEE (2018)","DOI":"10.1109\/FG.2018.00043"},{"key":"36_CR10","doi-asserted-by":"crossref","unstructured":"Li, X., Chen, J., Zhao, G., Pietikainen, M.: Remote heart rate measurement from face videos under realistic situations. In: Proceedings of IEEE CVPR, pp. 4264\u20134271. IEEE (2014)","DOI":"10.1109\/CVPR.2014.543"},{"key":"36_CR11","doi-asserted-by":"crossref","unstructured":"McDuff, D.J., Blackford, E.B., Estepp, J.R.: The impact of video compression on remote cardiac pulse measurement using imaging photoplethysmography. In: Proceedings of IEEE FG, pp. 63\u201370. IEEE (2017)","DOI":"10.1109\/FG.2017.17"},{"key":"36_CR12","doi-asserted-by":"crossref","unstructured":"Niu, X., Han, H., Shan, S., Chen, X.: Continuous heart rate measurement from face: a robust rPPG approach with distribution learning. In: Proceedings of IJCB (2017)","DOI":"10.1109\/BTAS.2017.8272752"},{"key":"36_CR13","doi-asserted-by":"crossref","unstructured":"Niu, X., Han, H., Shan, S., Chen, X.: Synrhythm: learning a deep heart rate estimator from general to specific. In: Proceedings of IEEE ICPR. IEEE (2018)","DOI":"10.1109\/ICPR.2018.8546321"},{"issue":"10","key":"36_CR14","doi-asserted-by":"publisher","first-page":"2268","DOI":"10.1109\/TIFS.2016.2578288","volume":"11","author":"K Patel","year":"2016","unstructured":"Patel, K., Han, H., Jain, A.K.: Secure face unlock: spoof detection on smartphones. IEEE Trans. Inf. Forensics Secur. 11(10), 2268\u20132283 (2016)","journal-title":"IEEE Trans. Inf. Forensics Secur."},{"issue":"10","key":"36_CR15","doi-asserted-by":"publisher","first-page":"10762","DOI":"10.1364\/OE.18.010762","volume":"18","author":"MZ Poh","year":"2010","unstructured":"Poh, M.Z., McDuff, D.J., Picard, R.W.: Non-contact, automated cardiac pulse measurements using video imaging and blind source separation. Opt. Exp. 18(10), 10762\u201310774 (2010)","journal-title":"Opt. Exp."},{"issue":"1","key":"36_CR16","doi-asserted-by":"publisher","first-page":"7","DOI":"10.1109\/TBME.2010.2086456","volume":"58","author":"MZ Poh","year":"2011","unstructured":"Poh, M.Z., McDuff, D.J., Picard, R.W.: Advancements in noncontact, multiparameter physiological measurements using a webcam. IEEE Trans. Biomed. Eng. 58(1), 7\u201311 (2011)","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"3","key":"36_CR17","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., et al.: Imagenet large scale visual recognition challenge. IJCV 115(3), 211\u2013252 (2015)","journal-title":"IJCV"},{"issue":"1","key":"36_CR18","doi-asserted-by":"publisher","first-page":"42","DOI":"10.1109\/T-AFFC.2011.25","volume":"3","author":"M Soleymani","year":"2012","unstructured":"Soleymani, M., Lichtenauer, J., Pun, T., Pantic, M.: A multimodal database for affect recognition and implicit tagging. IEEE Trans. Affect. Comput. 3(1), 42\u201355 (2012)","journal-title":"IEEE Trans. Affect. Comput."},{"key":"36_CR19","doi-asserted-by":"crossref","unstructured":"Stricker, R., M\u00fcller, S., Gross, H.M.: Non-contact video-based pulse rate measurement on a mobile service robot. In: Proceedings of IEEE RO-MAN, pp. 1056\u20131062. IEEE (2014)","DOI":"10.1109\/ROMAN.2014.6926392"},{"issue":"4","key":"36_CR20","doi-asserted-by":"publisher","first-page":"048002","DOI":"10.1117\/1.JBO.20.4.048002","volume":"20","author":"GR Tsouri","year":"2015","unstructured":"Tsouri, G.R., Li, Z.: On the benefits of alternative color spaces for noncontact heart rate measurements using standard red-green-blue cameras. J. Biomed. Opt. 20(4), 048002 (2015)","journal-title":"J. Biomed. Opt."},{"key":"36_CR21","doi-asserted-by":"crossref","unstructured":"Tulyakov, S., Alameda-Pineda, X., Ricci, E., Yin, L., Cohn, J.F., Sebe, N.: Self-adaptive matrix completion for heart rate estimation from face videos under realistic conditions. In: Proceedings of IEEE CVPR. IEEE (2016)","DOI":"10.1109\/CVPR.2016.263"},{"issue":"26","key":"36_CR22","doi-asserted-by":"publisher","first-page":"21434","DOI":"10.1364\/OE.16.021434","volume":"16","author":"W Verkruysse","year":"2008","unstructured":"Verkruysse, W., Svaasand, L.O., Nelson, J.S.: Remote plethysmographic imaging using ambient light. Opt. Exp. 16(26), 21434\u201321445 (2008)","journal-title":"Opt. Exp."},{"issue":"7","key":"36_CR23","doi-asserted-by":"publisher","first-page":"1479","DOI":"10.1109\/TBME.2016.2609282","volume":"64","author":"W Wang","year":"2017","unstructured":"Wang, W., den Brinker, A.C., Stuijk, S., de Haan, G.: Algorithmic principles of remote PPG. IEEE Trans. Biomed. Eng. 64(7), 1479\u20131491 (2017)","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"2","key":"36_CR24","doi-asserted-by":"publisher","first-page":"415","DOI":"10.1109\/TBME.2014.2356291","volume":"62","author":"W Wang","year":"2015","unstructured":"Wang, W., Stuijk, S., De Haan, G.: Exploiting spatial redundancy of image sensor for motion robust rPPG. IEEE Trans. Biomed. Eng. 62(2), 415\u2013425 (2015)","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"4","key":"36_CR25","doi-asserted-by":"publisher","first-page":"65","DOI":"10.1145\/2185520.2185561","volume":"31","author":"HY Wu","year":"2012","unstructured":"Wu, H.Y., Rubinstein, M., Shih, E., Guttag, J., Durand, F., Freeman, W.: Eulerian video magnification for revealing subtle changes in the world. ACM Trans. Graph. 31(4), 65 (2012)","journal-title":"ACM Trans. Graph."},{"key":"36_CR26","doi-asserted-by":"crossref","unstructured":"Zhang, Z., et al.: Multimodal spontaneous emotion corpus for human behavior analysis. In: Proceedings of IEEE CVPR, pp. 3438\u20133446. IEEE (2016)","DOI":"10.1109\/CVPR.2016.374"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ACCV 2018"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-20873-8_36","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,5,26]],"date-time":"2019-05-26T02:51:14Z","timestamp":1558839074000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-20873-8_36"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030208721","9783030208738"],"references-count":26,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-20873-8_36","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"26 May 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ACCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Asian Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Perth, WA","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Australia","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2018","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 December 2018","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 December 2018","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"14","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"accv2018","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/accv2018.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"Microsoft CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"979","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"274","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"2.7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information"}}]}}