{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T14:42:11Z","timestamp":1742913731539,"version":"3.40.3"},"publisher-location":"Cham","reference-count":28,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030678340"},{"type":"electronic","value":"9783030678357"}],"license":[{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2021,1,1]],"date-time":"2021-01-01T00:00:00Z","timestamp":1609459200000},"content-version":"vor","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2021]]},"DOI":"10.1007\/978-3-030-67835-7_11","type":"book-chapter","created":{"date-parts":[[2021,4,1]],"date-time":"2021-04-01T07:44:49Z","timestamp":1617263089000},"page":"121-133","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Unsupervised Gaze: Exploration of Geometric Constraints for 3D Gaze Estimation"],"prefix":"10.1007","author":[{"given":"Yawen","family":"Lu","sequence":"first","affiliation":[]},{"given":"Yuxing","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yuan","family":"Xin","sequence":"additional","affiliation":[]},{"given":"Di","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Guoyu","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2021,1,21]]},"reference":[{"key":"11_CR1","doi-asserted-by":"publisher","first-page":"241","DOI":"10.1177\/1553350612449075","volume":"20","author":"MS Atkins","year":"2013","unstructured":"Atkins, M.S., Tien, G., Khan, R.S., Meneghetti, A., Zheng, B.: What do surgeons see: capturing and synchronizing eye gaze for surgery applications. Surg. Innov. 20, 241\u2013248 (2013)","journal-title":"Surg. Innov."},{"key":"11_CR2","doi-asserted-by":"crossref","unstructured":"Bas, A., Huber, P., Smith, W.A., Awais, M., Kittler, J.: 3D morphable models as spatial transformer networks. In: IEEE ICCVW (2017)","DOI":"10.1109\/ICCVW.2017.110"},{"key":"11_CR3","doi-asserted-by":"crossref","unstructured":"Drewes, H., De Luca, A., Schmidt, A.: Eye-gaze interaction for mobile phones. In: Proceedings of the Conference on Mobile Technology, Applications, and Systems (2007)","DOI":"10.1145\/1378063.1378122"},{"key":"11_CR4","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1109\/MIS.2016.47","volume":"31","author":"L Fridman","year":"2016","unstructured":"Fridman, L., Langhans, P., Lee, J., Reimer, B.: Driver gaze region estimation without use of eye movement. IEEE Intell. Syst. 31, 49\u201356 (2016)","journal-title":"IEEE Intell. Syst."},{"key":"11_CR5","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac Aodha, O., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: IEEE ICCV (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"11_CR6","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1109\/TPAMI.2009.30","volume":"32","author":"DW Hansen","year":"2009","unstructured":"Hansen, D.W., Ji, Q.: In the eye of the beholder: a survey of models for eyes and gaze. IEEE TPAMI 32, 478\u2013500 (2009)","journal-title":"IEEE TPAMI"},{"key":"11_CR7","doi-asserted-by":"publisher","first-page":"1527","DOI":"10.1109\/21.44068","volume":"19","author":"TE Hutchinson","year":"1989","unstructured":"Hutchinson, T.E., White, K.P., Martin, W.N., Reichert, K.C., Frey, L.A.: Human-computer interaction using eye-gaze input. IEEE Trans. Syst. Man Cybern. 19, 1527\u20131534 (1989)","journal-title":"IEEE Trans. Syst. Man Cybern."},{"key":"11_CR8","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"11_CR9","doi-asserted-by":"crossref","unstructured":"Liang, K., Chahir, Y., Molina, M., Tijus, C., Jouen, F.: Appearance-based gaze tracking with spectral clustering and semi-supervised Gaussian process regression. In: Proceedings of Conference on Eye Tracking South Africa (2013)","DOI":"10.1145\/2509315.2509318"},{"key":"11_CR10","doi-asserted-by":"publisher","first-page":"2033","DOI":"10.1109\/TPAMI.2014.2313123","volume":"36","author":"F Lu","year":"2014","unstructured":"Lu, F., Sugano, Y., Okabe, T., Sato, Y.: Adaptive linear regression for appearance-based gaze estimation. IEEE TPAMI 36, 2033\u20132046 (2014)","journal-title":"IEEE TPAMI"},{"key":"11_CR11","doi-asserted-by":"crossref","unstructured":"Morimoto, C.H., Amir, A., Flickner, M.: Detecting eye position and gaze from a single camera and 2 light sources. In: Object Recognition Supported by User Interaction for Service Robots, vol. 4, pp. 314\u2013317. IEEE (2002)","DOI":"10.1109\/ICPR.2002.1047459"},{"key":"11_CR12","doi-asserted-by":"crossref","unstructured":"Park, S., Mello, S.D., Molchanov, P., Iqbal, U., Hilliges, O., Kautz, J.: Few-shot adaptive gaze estimation. In: IEEE ICCV (2019)","DOI":"10.1109\/ICCV.2019.00946"},{"key":"11_CR13","doi-asserted-by":"crossref","unstructured":"Park, S., Spurr, A., Hilliges, O.: Deep pictorial gaze estimation. In: ECCV, pp. 721\u2013738 (2018)","DOI":"10.1007\/978-3-030-01261-8_44"},{"key":"11_CR14","doi-asserted-by":"crossref","unstructured":"Sela, M., Richardson, E., Kimmel, R.: Unrestricted facial geometry reconstruction using image-to-image translation. In: IEEE ICCV (2017)","DOI":"10.1109\/ICCV.2017.175"},{"key":"11_CR15","doi-asserted-by":"crossref","unstructured":"Smith, B.A., Yin, Q., Feiner, S.K., Nayar, S.K.: Gaze locking: passive eye contact detection for human-object interaction. In: ACM Symposium on User Interface Software and Technology, pp. 271\u2013280 (2013)","DOI":"10.1145\/2501988.2501994"},{"key":"11_CR16","doi-asserted-by":"crossref","unstructured":"Tsukada, A., Shino, M., Devyver, M., Kanade, T.: Illumination-free gaze estimation method for first-person vision wearable device. In: IEEE ICCVW (2011)","DOI":"10.1109\/ICCVW.2011.6130505"},{"key":"11_CR17","first-page":"802","volume":"21","author":"R Valenti","year":"2011","unstructured":"Valenti, R., Sebe, N., Gevers, T.: Combining head pose and eye location information for gaze estimation. IEEE TIP 21, 802\u2013815 (2011)","journal-title":"IEEE TIP"},{"issue":"4","key":"11_CR18","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"key":"11_CR19","doi-asserted-by":"crossref","unstructured":"Whitmire, E., et al.: EyeContact: scleral coil eye tracking for virtual reality. In: Proceedings of ACM International Symposium on Wearable Computers, pp. 184\u2013191 (2016)","DOI":"10.1145\/2971763.2971771"},{"key":"11_CR20","doi-asserted-by":"crossref","unstructured":"Wood, E., Baltrusaitis, T., Zhang, X., Sugano, Y., Robinson, P., Bulling, A.: Rendering of eyes for eye-shape registration and gaze estimation. In: IEEE ICCV (2015)","DOI":"10.1109\/ICCV.2015.428"},{"key":"11_CR21","doi-asserted-by":"crossref","unstructured":"Yin, Z., Shi, J.: GeoNet: unsupervised learning of dense depth, optical flow and camera pose. In: IEEE CVPR (2018)","DOI":"10.1109\/CVPR.2018.00212"},{"key":"11_CR22","first-page":"25","volume":"98","author":"DH Yoo","year":"2005","unstructured":"Yoo, D.H., Chung, M.J.: A novel non-intrusive eye gaze estimation using cross-ratio under large head motion. CVIU 98, 25\u201351 (2005)","journal-title":"CVIU"},{"key":"11_CR23","doi-asserted-by":"crossref","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: Appearance-based gaze estimation in the wild. In: IEEE CVPR, pp. 4511\u20134520 (2015)","DOI":"10.1109\/CVPR.2015.7299081"},{"key":"11_CR24","doi-asserted-by":"crossref","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: It\u2019s written all over your face: full-face appearance-based gaze estimation. In: IEEE CVPRW (2017)","DOI":"10.1109\/CVPRW.2017.284"},{"key":"11_CR25","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/TPAMI.2017.2778103","volume":"41","author":"X Zhang","year":"2017","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: MPIIGaze: real-world dataset and deep appearance-based gaze estimation. IEEE TPAMI 41, 162\u2013175 (2017)","journal-title":"IEEE TPAMI"},{"key":"11_CR26","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.G.: Unsupervised learning of depth and ego-motion from video. In: IEEE CVPR (2017)","DOI":"10.1109\/CVPR.2017.700"},{"key":"11_CR27","unstructured":"Zhu, Z., Ji, Q.: Eye gaze tracking under natural head movements. In: IEEE CVPR (2005)"},{"key":"11_CR28","unstructured":"Zhu, Z., Ji, Q., Bennett, K.P.: Nonlinear eye gaze mapping function estimation via support vector regression. In: ICPR (2006)"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-67835-7_11","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,8,27]],"date-time":"2024-08-27T08:35:16Z","timestamp":1724747716000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-67835-7_11"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2021]]},"ISBN":["9783030678340","9783030678357"],"references-count":28,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-67835-7_11","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2021]]},"assertion":[{"value":"21 January 2021","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Prague","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Czech Republic","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 June 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 June 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/mmm2021.cz\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"easychair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"211","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"73","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"35% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,63","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}