{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,11]],"date-time":"2024-09-11T15:14:54Z","timestamp":1726067694607},"publisher-location":"Singapore","reference-count":29,"publisher":"Springer Singapore","isbn-type":[{"type":"print","value":"9789811533402"},{"type":"electronic","value":"9789811533419"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-981-15-3341-9_11","type":"book-chapter","created":{"date-parts":[[2020,2,15]],"date-time":"2020-02-15T07:02:24Z","timestamp":1581750144000},"page":"126-139","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Adaptive Person-Specific Appearance-Based Gaze Estimation"],"prefix":"10.1007","author":[{"given":"Chuanyang","family":"Zheng","sequence":"first","affiliation":[]},{"given":"Jun","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Jun","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Lihua","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,2,16]]},"reference":[{"key":"11_CR1","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1007\/978-3-030-01264-9_7","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Yihua Cheng","year":"2018","unstructured":"Cheng, Y., Lu, F., Zhang, X.: Appearance-based gaze estimation via evaluation-guided asymmetric regression. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 100\u2013115 (2018)"},{"key":"11_CR2","unstructured":"Finn, C., Abbeel, P., Levine, S.: Model-agnostic meta-learning for fast adaptation of deep networks. In: Proceedings of the 34th International Conference on Machine Learning, vol. 70, pp. 1126\u20131135. JMLR. org (2017)"},{"key":"11_CR3","doi-asserted-by":"publisher","first-page":"339","DOI":"10.1007\/978-3-030-01249-6_21","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Tobias Fischer","year":"2018","unstructured":"Fischer, T., Jin Chang, H., Demiris, Y.: Rt-gene: real-time eye gaze estimation in natural environments. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 334\u2013352 (2018)"},{"key":"11_CR4","doi-asserted-by":"publisher","unstructured":"Funes Mora, K.A., Monay, F., Odobez, J.M.: EYEDIAP: a database for the development and evaluation of gaze estimation algorithms from RGB and RGB-D cameras. In: Proceedings of the ACM Symposium on Eye Tracking Research and Applications. ACM, March 2014. https:\/\/doi.org\/10.1145\/2578153.2578190","DOI":"10.1145\/2578153.2578190"},{"key":"11_CR5","unstructured":"Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, pp. 249\u2013256 (2010)"},{"key":"11_CR6","doi-asserted-by":"crossref","unstructured":"Guasconi, S., Porta, M., Resta, C., Rottenbacher, C.: A low-cost implementation of an eye tracking system for driver\u2019s gaze analysis. In: 2017 10th International Conference on Human System Interactions (HSI), pp. 264\u2013269. IEEE (2017)","DOI":"10.1109\/HSI.2017.8005043"},{"issue":"6","key":"11_CR7","doi-asserted-by":"publisher","first-page":"1124","DOI":"10.1109\/TBME.2005.863952","volume":"53","author":"ED Guestrin","year":"2006","unstructured":"Guestrin, E.D., Eizenman, M.: General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans. Biomed. Eng. 53(6), 1124\u20131133 (2006)","journal-title":"IEEE Trans. Biomed. Eng."},{"issue":"3","key":"11_CR8","doi-asserted-by":"publisher","first-page":"478","DOI":"10.1109\/TPAMI.2009.30","volume":"32","author":"DW Hansen","year":"2009","unstructured":"Hansen, D.W., Ji, Q.: In the eye of the beholder: a survey of models for eyes and gaze. IEEE Trans. Pattern Anal. Mach. Intell. 32(3), 478\u2013500 (2009)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"11_CR9","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"11_CR10","doi-asserted-by":"crossref","unstructured":"Krafka, K., et al.: Eye tracking for everyone. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2176\u20132184 (2016)","DOI":"10.1109\/CVPR.2016.239"},{"key":"11_CR11","doi-asserted-by":"crossref","unstructured":"Lind\u00e9n, E., Sj\u00f6strand, J., Proutiere, A.: Learning to personalize in appearance-based gaze tracking. arXiv e-prints arXiv:1807.00664 , July 2018","DOI":"10.1109\/ICCVW.2019.00145"},{"key":"11_CR12","doi-asserted-by":"crossref","unstructured":"Liu, G., Yu, Y., Funes Mora, K.A., Odobez, J.M.: A differential approach for gaze estimation. arXiv e-prints arXiv:1904.09459 , April 2019","DOI":"10.1109\/TPAMI.2019.2957373"},{"key":"11_CR13","doi-asserted-by":"crossref","unstructured":"Liu, G., Yu, Y., Funes Mora, K.A., Odobez, J.M.: A differential approach for gaze estimation with calibration. In: 29th British Machine Vision Conference (2018)","DOI":"10.1109\/TPAMI.2019.2957373"},{"issue":"10","key":"11_CR14","doi-asserted-by":"publisher","first-page":"2033","DOI":"10.1109\/TPAMI.2014.2313123","volume":"36","author":"F Lu","year":"2014","unstructured":"Lu, F., Sugano, Y., Okabe, T., Sato, Y.: Adaptive linear regression for appearance-based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 36(10), 2033\u20132046 (2014)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"11_CR15","doi-asserted-by":"crossref","unstructured":"Mavely, A.G., Judith, J., Sahal, P., Kuruvilla, S.A.: Eye gaze tracking based driver monitoring system. In: 2017 IEEE International Conference on Circuits and Systems (ICCS), pp. 364\u2013367. IEEE (2017)","DOI":"10.1109\/ICCS1.2017.8326022"},{"key":"11_CR16","doi-asserted-by":"crossref","unstructured":"Mutlu, B., Shiwa, T., Kanda, T., Ishiguro, H., Hagita, N.: Footing in human-robot conversations: how robots might shape participant roles using gaze cues. In: Proceedings of the 4th ACM\/IEEE International Conference on Human Robot Interaction, pp. 61\u201368. ACM (2009)","DOI":"10.1145\/1514095.1514109"},{"issue":"1696","key":"11_CR17","doi-asserted-by":"publisher","first-page":"2935","DOI":"10.1098\/rspb.2010.0587","volume":"277","author":"T Nakano","year":"2010","unstructured":"Nakano, T., et al.: Atypical gaze patterns in children and adults with autism spectrum disorders dissociated from developmental changes in gaze behaviour. Proc. R. Soc. B: Biol. Sci. 277(1696), 2935\u20132943 (2010)","journal-title":"Proc. R. Soc. B: Biol. Sci."},{"issue":"9","key":"11_CR18","doi-asserted-by":"publisher","first-page":"2183","DOI":"10.1073\/pnas.1617251114","volume":"114","author":"N Padmanaban","year":"2017","unstructured":"Padmanaban, N., Konrad, R., Stramer, T., Cooper, E.A., Wetzstein, G.: Optimizing virtual reality for all users through gaze-contingent and adaptive focus displays. Proc. Natl. Acad. Sci. 114(9), 2183\u20132188 (2017)","journal-title":"Proc. Natl. Acad. Sci."},{"issue":"6","key":"11_CR19","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1145\/2980179.2980246","volume":"35","author":"A Patney","year":"2016","unstructured":"Patney, A., et al.: Towards foveated rendering for gaze-tracked virtual reality. ACM Trans. Graph. (TOG) 35(6), 179 (2016)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"11_CR20","doi-asserted-by":"crossref","unstructured":"Ranjan, R., De Mello, S., Kautz, J.: Light-weight head pose invariant gaze tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 2156\u20132164 (2018)","DOI":"10.1109\/CVPRW.2018.00290"},{"key":"11_CR21","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:1409.1556 (2014)"},{"key":"11_CR22","unstructured":"Sugano, Y., Fritz, M., Andreas Bulling, X., et al.: It\u2019s written all over your face: full-face appearance-based gaze estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 51\u201360 (2017)"},{"issue":"2","key":"11_CR23","doi-asserted-by":"publisher","first-page":"329","DOI":"10.1109\/TPAMI.2012.101","volume":"35","author":"Y Sugano","year":"2012","unstructured":"Sugano, Y., Matsushita, Y., Sato, Y.: Appearance-based gaze estimation using visual saliency. IEEE Trans. Pattern Anal. Mach. Intell. 35(2), 329\u2013341 (2012)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"11_CR24","doi-asserted-by":"crossref","unstructured":"Sugano, Y., Matsushita, Y., Sato, Y.: Learning-by-synthesis for appearance-based 3D gaze estimation. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2014","DOI":"10.1109\/CVPR.2014.235"},{"key":"11_CR25","unstructured":"Wang, Y., Yao, Q.: Few-shot learning: a survey. CoRR abs\/1904.05046, http:\/\/arxiv.org\/abs\/1904.05046 (2019)"},{"key":"11_CR26","doi-asserted-by":"crossref","unstructured":"Yu, Y., Liu, G., Odobez, J.M.: Improving few-shot user-specific gaze adaptation via gaze redirection synthesis. In: The IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019","DOI":"10.1109\/CVPR.2019.01221"},{"key":"11_CR27","doi-asserted-by":"crossref","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: Appearance-based gaze estimation in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4511\u20134520 (2015)","DOI":"10.1109\/CVPR.2015.7299081"},{"issue":"1","key":"11_CR28","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/TPAMI.2017.2778103","volume":"41","author":"X Zhang","year":"2017","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: MPIIGaze: real-world dataset and deep appearance-based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 41(1), 162\u2013175 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"11_CR29","doi-asserted-by":"crossref","unstructured":"Zhu, W., Deng, H.: Monocular free-head 3D gaze tracking with deep learning and geometry constraints. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3143\u20133152 (2017)","DOI":"10.1109\/ICCV.2017.341"}],"container-title":["Communications in Computer and Information Science","Digital TV and Wireless Multimedia Communication"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-15-3341-9_11","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2021,2,27]],"date-time":"2021-02-27T12:23:01Z","timestamp":1614428581000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-981-15-3341-9_11"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9789811533402","9789811533419"],"references-count":29,"URL":"https:\/\/doi.org\/10.1007\/978-981-15-3341-9_11","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"16 February 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"IFTC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Forum on Digital TV and Wireless Multimedia Communications","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Shanghai","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19 September 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 September 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iftc2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.siga.com.cn\/iftc2019.html","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"120","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"34","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"-","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}