{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,27]],"date-time":"2025-07-27T07:33:06Z","timestamp":1753601586901,"version":"3.40.3"},"publisher-location":"Cham","reference-count":41,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031730092"},{"type":"electronic","value":"9783031730108"}],"license":[{"start":{"date-parts":[[2024,11,10]],"date-time":"2024-11-10T00:00:00Z","timestamp":1731196800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,10]],"date-time":"2024-11-10T00:00:00Z","timestamp":1731196800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73010-8_1","type":"book-chapter","created":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T13:10:41Z","timestamp":1731157841000},"page":"1-17","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["LG-Gaze: Learning Geometry-Aware Continuous Prompts for\u00a0Language-Guided Gaze Estimation"],"prefix":"10.1007","author":[{"given":"Pengwei","family":"Yin","sequence":"first","affiliation":[]},{"given":"Jingjing","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Guanzhong","family":"Zeng","sequence":"additional","affiliation":[]},{"given":"Di","family":"Xie","sequence":"additional","affiliation":[]},{"given":"Jiang","family":"Zhu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,10]]},"reference":[{"key":"1_CR1","doi-asserted-by":"crossref","unstructured":"Andrist, S., Tan, X.Z., Gleicher, M., Mutlu, B.: Conversational gaze aversion for humanlike robots. In: 2014 9th ACM\/IEEE International Conference on Human-Robot Interaction (HRI), pp. 25\u201332 (2014)","DOI":"10.1145\/2559636.2559666"},{"key":"1_CR2","doi-asserted-by":"crossref","unstructured":"Bao, Y., Liu, Y., Wang, H., Lu, F.: Generalizing gaze estimation with rotation consistency. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4207\u20134216 (2022)","DOI":"10.1109\/CVPR52688.2022.00417"},{"key":"1_CR3","doi-asserted-by":"crossref","unstructured":"Cai, X., Zeng, J., Shan, S., Chen, X.: Source-free adaptive gaze estimation by uncertainty reduction. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 22035\u201322045 (2023)","DOI":"10.1109\/CVPR52729.2023.02110"},{"key":"1_CR4","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: International Conference on Machine Learning, pp. 1597\u20131607. PMLR (2020)"},{"key":"1_CR5","unstructured":"Chen, X., Fan, H., Girshick, R., He, K.: Improved baselines with momentum contrastive learning. arXiv preprint arXiv:2003.04297 (2020)"},{"key":"1_CR6","doi-asserted-by":"crossref","unstructured":"Cheng, Y., Bao, Y.: Puregaze: Purifying gaze feature for generalizable gaze estimation. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a036, pp. 436\u2013443 (2022)","DOI":"10.1609\/aaai.v36i1.19921"},{"key":"1_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"105","DOI":"10.1007\/978-3-030-01264-9_7","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Cheng","year":"2018","unstructured":"Cheng, Y., Lu, F., Zhang, X.: Appearance-based gaze estimation via evaluation-guided asymmetric regression. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) Computer Vision \u2013 ECCV 2018. LNCS, vol. 11218, pp. 105\u2013121. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01264-9_7"},{"key":"1_CR8","doi-asserted-by":"publisher","first-page":"5259","DOI":"10.1109\/TIP.2020.2982828","volume":"29","author":"Y Cheng","year":"2020","unstructured":"Cheng, Y., Zhang, X., Lu, F., Sato, Y.: Gaze estimation by exploring two-eye asymmetry. IEEE Trans. Image Process. 29, 5259\u20135272 (2020)","journal-title":"IEEE Trans. Image Process."},{"key":"1_CR9","doi-asserted-by":"crossref","unstructured":"Funes\u00a0Mora, K.A., Monay, F., Odobez, J.M.: EYEDIAP: a database for the development and evaluation of gaze estimation algorithms from RGB and RGB-D cameras. In: Proceedings of the Symposium on Eye Tracking Research and Applications, pp. 255\u2013258 (2014)","DOI":"10.1145\/2578153.2578190"},{"key":"1_CR10","doi-asserted-by":"publisher","first-page":"49","DOI":"10.1007\/s11004-009-9257-x","volume":"42","author":"\u00c1 Gonz\u00e1lez","year":"2010","unstructured":"Gonz\u00e1lez, \u00c1.: Measurement of areas on a sphere using Fibonacci and latitude-longitude lattices. Math. Geosci. 42, 49\u201364 (2010)","journal-title":"Math. Geosci."},{"key":"1_CR11","doi-asserted-by":"crossref","unstructured":"He, K., Fan, H., Wu, Y., Xie, S., Girshick, R.: Momentum contrast for unsupervised visual representation learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9729\u20139738 (2020)","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"1_CR12","doi-asserted-by":"crossref","unstructured":"Kellnhofer, P., Recasens, A., Stent, S., Matusik, W., Torralba, A.: Gaze360: physically unconstrained gaze estimation in the wild. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 6912\u20136921 (2019)","DOI":"10.1109\/ICCV.2019.00701"},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Krafka, K., et al.: Eye tracking for everyone. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2176\u20132184 (2016)","DOI":"10.1109\/CVPR.2016.239"},{"key":"1_CR14","doi-asserted-by":"crossref","unstructured":"Lee, I., Yun, J.S., Kim, H.H., Na, Y., Yoo, S.B.: LatentGaze: cross-domain gaze estimation through gaze-aware analytic latent code manipulation. In: Proceedings of the Asian Conference on Computer Vision, pp. 3379\u20133395 (2022)","DOI":"10.1007\/978-3-031-26348-4_10"},{"key":"1_CR15","first-page":"35313","volume":"35","author":"W Li","year":"2022","unstructured":"Li, W., et al.: OrdinalCLIP: learning rank prompts for language-guided ordinal regression. Adv. Neural. Inf. Process. Syst. 35, 35313\u201335325 (2022)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"1_CR16","unstructured":"Liu, R., Bao, Y., Xu, M., Wang, H., Liu, Y., Lu, F.: Jitter does matter: adapting gaze estimation to new domains. arXiv preprint arXiv:2210.02082 (2022)"},{"key":"1_CR17","doi-asserted-by":"crossref","unstructured":"Liu, Y., Liu, R., Wang, H., Lu, F.: Generalizing gaze estimation with outlier-guided collaborative adaptation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision (2021)","DOI":"10.1109\/ICCV48922.2021.00381"},{"key":"1_CR18","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts. arXiv preprint arXiv:1608.03983 (2016)"},{"key":"1_CR19","unstructured":"Van\u00a0der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11) (2008)"},{"key":"1_CR20","doi-asserted-by":"publisher","unstructured":"Mavely, A.G., Judith, J.E., Sahal, P.A., Kuruvilla, S.A.: Eye gaze tracking based driver monitoring system. In: 2017 IEEE International Conference on Circuits and Systems (ICCS), pp. 364\u2013367 (2017). https:\/\/doi.org\/10.1109\/ICCS1.2017.8326022","DOI":"10.1109\/ICCS1.2017.8326022"},{"key":"1_CR21","doi-asserted-by":"publisher","unstructured":"Padmanaban, N., Konrad, R., Cooper, E.A., Wetzstein, G.: Optimizing VR for all users through adaptive focus displays. In: ACM SIGGRAPH 2017 Talks. SIGGRAPH 2017. Association for Computing Machinery, New York (2017). https:\/\/doi.org\/10.1145\/3084363.3085029","DOI":"10.1145\/3084363.3085029"},{"key":"1_CR22","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"1_CR23","unstructured":"Radford, A., et\u00a0al.: Language models are unsupervised multitask learners. OpenAI blog 9 (2019)"},{"key":"1_CR24","doi-asserted-by":"crossref","unstructured":"Rao, Y., et al.: DenseCLIP: language-guided dense prediction with context-aware prompting. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18082\u201318091 (2022)","DOI":"10.1109\/CVPR52688.2022.01755"},{"key":"1_CR25","doi-asserted-by":"crossref","unstructured":"Shen, S., et al.: CLIP-cluster: CLIP-guided attribute hallucination for face clustering. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 20786\u201320795 (2023)","DOI":"10.1109\/ICCV51070.2023.01900"},{"key":"1_CR26","doi-asserted-by":"crossref","unstructured":"Shoemake, K.: Animating rotation with quaternion curves. In: Proceedings of the 12th Annual Conference on Computer Graphics and Interactive Techniques, pp. 245\u2013254 (1985)","DOI":"10.1145\/325334.325242"},{"key":"1_CR27","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in neural information processing systems (2017)"},{"key":"1_CR28","doi-asserted-by":"crossref","unstructured":"Vidit, V., Engilberge, M., Salzmann, M.: CLIP the gap: a single domain generalization approach for object detection. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3219\u20133229 (2023)","DOI":"10.1109\/CVPR52729.2023.00314"},{"key":"1_CR29","unstructured":"Wang, R., Li, P., Huang, H., Cao, C., He, R., He, Z.: Learning-to-rank meets language: boosting language-driven ordering alignment for ordinal classification. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"},{"key":"1_CR30","doi-asserted-by":"publisher","unstructured":"Wang, Y., et al.: Contrastive regression for domain adaptation on gaze estimation. In: 2022 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 19354\u201319363 (2022). https:\/\/doi.org\/10.1109\/CVPR52688.2022.01877","DOI":"10.1109\/CVPR52688.2022.01877"},{"key":"1_CR31","doi-asserted-by":"crossref","unstructured":"Xu, M., Wang, H., Lu, F.: Learning a generalized gaze estimator from gaze-consistent feature. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 3, pp. 3027\u20133035 (2023)","DOI":"10.1609\/aaai.v37i3.25406"},{"key":"1_CR32","unstructured":"Yao, L., et al.: DetCLIP: dictionary-enriched visual-concept paralleled pre-training for open-world detection. In: Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A. (eds.) Advances in Neural Information Processing Systems, vol.\u00a035, pp. 9125\u20139138. Curran Associates, Inc. (2022). https:\/\/proceedings.neurips.cc\/paper_files\/paper\/2022\/file\/3ba960559212691be13fa81d9e5e0047-Paper-Conference.pdf"},{"key":"1_CR33","doi-asserted-by":"crossref","unstructured":"Yin, P., Wang, J., Dai, J., Wu, X.: NeRF-gaze: a head-eye redirection parametric model for gaze estimation. In: IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (2024)","DOI":"10.1109\/ICASSP48485.2024.10446677"},{"key":"1_CR34","doi-asserted-by":"crossref","unstructured":"Yin, P., Zeng, G., Wang, J., Xie, D.: CLIP-gaze: towards general gaze estimation via visual-linguistic model. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol.\u00a038, pp. 6729\u20136737 (2024)","DOI":"10.1609\/aaai.v38i7.28496"},{"key":"1_CR35","unstructured":"Zha, K., Cao, P., Son, J., Yang, Y., Katabi, D.: Rank-N-contrast: learning continuous representations for regression. In: Thirty-seventh Conference on Neural Information Processing Systems (2023)"},{"key":"1_CR36","unstructured":"Zha, K., Cao, P., Son, J., Yang, Y., Katabi, D.: Rank-N-contrast: learning continuous representations for regression. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"},{"key":"1_CR37","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"365","DOI":"10.1007\/978-3-030-58558-7_22","volume-title":"Computer Vision \u2013 ECCV 2020","author":"X Zhang","year":"2020","unstructured":"Zhang, X., Park, S., Beeler, T., Bradley, D., Tang, S., Hilliges, O.: ETH-XGaze: a large scale dataset for gaze estimation under extreme head pose and gaze variation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12350, pp. 365\u2013381. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58558-7_22"},{"key":"1_CR38","doi-asserted-by":"crossref","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: Appearance-based gaze estimation in the wild. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4511\u20134520 (2015)","DOI":"10.1109\/CVPR.2015.7299081"},{"issue":"1","key":"1_CR39","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/TPAMI.2017.2778103","volume":"41","author":"X Zhang","year":"2017","unstructured":"Zhang, X., Sugano, Y., Fritz, M., Bulling, A.: MPIIGaze: real-world dataset and deep appearance-based gaze estimation. IEEE Trans. Pattern Anal. Mach. Intell. 41(1), 162\u2013175 (2017)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"1_CR40","doi-asserted-by":"crossref","unstructured":"Zhou, K., Yang, J., Loy, C.C., Liu, Z.: Conditional prompt learning for vision-language models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 16816\u201316825 (2022)","DOI":"10.1109\/CVPR52688.2022.01631"},{"issue":"9","key":"1_CR41","doi-asserted-by":"publisher","first-page":"2337","DOI":"10.1007\/s11263-022-01653-1","volume":"130","author":"K Zhou","year":"2022","unstructured":"Zhou, K., Yang, J., Loy, C.C., Liu, Z.: Learning to prompt for vision-language models. Int. J. Comput. Vision 130(9), 2337\u20132348 (2022)","journal-title":"Int. J. Comput. Vision"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73010-8_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,9]],"date-time":"2024-11-09T14:02:06Z","timestamp":1731160926000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73010-8_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,10]]},"ISBN":["9783031730092","9783031730108"],"references-count":41,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73010-8_1","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,10]]},"assertion":[{"value":"10 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}