{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T17:17:28Z","timestamp":1770830248881,"version":"3.50.1"},"reference-count":64,"publisher":"Elsevier BV","license":[{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/tdm\/userlicense\/1.0\/"},{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.elsevier.com\/legal\/tdmrep-license"},{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-017"},{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-037"},{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-012"},{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-029"},{"start":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T00:00:00Z","timestamp":1746057600000},"content-version":"stm-asf","delay-in-days":0,"URL":"https:\/\/doi.org\/10.15223\/policy-004"}],"funder":[{"DOI":"10.13039\/501100002920","name":"Research Grants Council, University Grants Committee","doi-asserted-by":"publisher","id":[{"id":"10.13039\/501100002920","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["elsevier.com","sciencedirect.com"],"crossmark-restriction":true},"short-container-title":["Pattern Recognition"],"published-print":{"date-parts":[[2025,5]]},"DOI":"10.1016\/j.patcog.2024.111244","type":"journal-article","created":{"date-parts":[[2024,12,16]],"date-time":"2024-12-16T05:36:51Z","timestamp":1734327411000},"page":"111244","update-policy":"https:\/\/doi.org\/10.1016\/elsevier_cm_policy","source":"Crossref","is-referenced-by-count":3,"special_numbering":"C","title":["Collaborative contrastive learning for cross-domain gaze estimation"],"prefix":"10.1016","volume":"161","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-9572-3299","authenticated-orcid":false,"given":"Lifan","family":"Xia","sequence":"first","affiliation":[]},{"given":"Yong","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0005-9121-6900","authenticated-orcid":false,"given":"Xin","family":"Cai","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-8292-6389","authenticated-orcid":false,"given":"Zhen","family":"Cui","sequence":"additional","affiliation":[]},{"given":"Chunyan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Antoni B.","family":"Chan","sequence":"additional","affiliation":[]}],"member":"78","reference":[{"issue":"1","key":"10.1016\/j.patcog.2024.111244_b1","doi-asserted-by":"crossref","first-page":"78","DOI":"10.1037\/0033-2909.100.1.78","article-title":"Gaze and eye contact: a research review","volume":"100","author":"Kleinke","year":"1986","journal-title":"Psychol. Bull."},{"key":"10.1016\/j.patcog.2024.111244_b2","doi-asserted-by":"crossref","first-page":"105","DOI":"10.3389\/fnhum.2018.00105","article-title":"Eye movements during everyday behavior predict personality traits","author":"Hoppe","year":"2018","journal-title":"Front. Hum. Neurosci."},{"key":"10.1016\/j.patcog.2024.111244_b3","doi-asserted-by":"crossref","unstructured":"X. Zhang, Y. Sugano, M. Fritz, A. Bulling, Appearance-based gaze estimation in the wild, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2015, pp. 4511\u20134520.","DOI":"10.1109\/CVPR.2015.7299081"},{"key":"10.1016\/j.patcog.2024.111244_b4","series-title":"Asian Conference on Computer Vision","first-page":"309","article-title":"Appearance-based gaze estimation using dilated-convolutions","author":"Chen","year":"2018"},{"key":"10.1016\/j.patcog.2024.111244_b5","doi-asserted-by":"crossref","unstructured":"Y. Cheng, F. Lu, X. Zhang, Appearance-based gaze estimation via evaluation-guided asymmetric regression, in: Proceedings of the European Conference on Computer Vision, ECCV, 2018, pp. 100\u2013115.","DOI":"10.1007\/978-3-030-01264-9_7"},{"key":"10.1016\/j.patcog.2024.111244_b6","doi-asserted-by":"crossref","unstructured":"Y. Yu, G. Liu, J.-M. Odobez, Improving few-shot user-specific gaze adaptation via gaze redirection synthesis, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 11937\u201311946.","DOI":"10.1109\/CVPR.2019.01221"},{"key":"10.1016\/j.patcog.2024.111244_b7","doi-asserted-by":"crossref","unstructured":"Y. Liu, R. Liu, H. Wang, F. Lu, Generalizing gaze estimation with outlier-guided collaborative adaptation, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 3835\u20133844.","DOI":"10.1109\/ICCV48922.2021.00381"},{"key":"10.1016\/j.patcog.2024.111244_b8","doi-asserted-by":"crossref","unstructured":"Y. Bao, Y. Liu, H. Wang, F. Lu, Generalizing gaze estimation with rotation consistency, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 4207\u20134216.","DOI":"10.1109\/CVPR52688.2022.00417"},{"key":"10.1016\/j.patcog.2024.111244_b9","doi-asserted-by":"crossref","unstructured":"Y. Wang, Y. Jiang, J. Li, B. Ni, W. Dai, C. Li, H. Xiong, T. Li, Contrastive regression for domain adaptation on gaze estimation, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 19376\u201319385.","DOI":"10.1109\/CVPR52688.2022.01877"},{"key":"10.1016\/j.patcog.2024.111244_b10","doi-asserted-by":"crossref","unstructured":"X. Cai, J. Zeng, S. Shan, X. Chen, Source-Free Adaptive Gaze Estimation by Uncertainty Reduction, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 22035\u201322045.","DOI":"10.1109\/CVPR52729.2023.02110"},{"key":"10.1016\/j.patcog.2024.111244_b11","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2023.109980","article-title":"Gaze estimation with semi-supervised eye landmark detection as an auxiliary task","volume":"146","author":"Sun","year":"2024","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.111244_b12","doi-asserted-by":"crossref","first-page":"36","DOI":"10.1016\/j.patcog.2017.04.026","article-title":"Gazing point dependent eye gaze estimation","volume":"71","author":"Cheng","year":"2017","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.111244_b13","doi-asserted-by":"crossref","DOI":"10.1016\/j.patcog.2022.108944","article-title":"In the eye of the beholder: A survey of gaze tracking techniques","volume":"132","author":"Liu","year":"2022","journal-title":"Pattern Recognit."},{"key":"10.1016\/j.patcog.2024.111244_b14","series-title":"Advances in Neural Information Processing Systems","article-title":"Where are they looking?","volume":"Vol. 28","author":"Recasens","year":"2015"},{"key":"10.1016\/j.patcog.2024.111244_b15","doi-asserted-by":"crossref","first-page":"445","DOI":"10.1007\/s00138-017-0852-4","article-title":"Tabletgaze: dataset and analysis for unconstrained appearance-based gaze estimation in mobile tablets","volume":"28","author":"Huang","year":"2017","journal-title":"Mach. Vis. Appl."},{"issue":"2","key":"10.1016\/j.patcog.2024.111244_b16","doi-asserted-by":"crossref","first-page":"329","DOI":"10.1109\/TPAMI.2012.101","article-title":"Appearance-based gaze estimation using visual saliency","volume":"35","author":"Sugano","year":"2013","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.patcog.2024.111244_b17","doi-asserted-by":"crossref","DOI":"10.1016\/j.eswa.2022.116894","article-title":"Eye gaze estimation: A survey on deep learning-based approaches","volume":"199","author":"Pathirana","year":"2022","journal-title":"Expert Syst. Appl.","ISSN":"https:\/\/id.crossref.org\/issn\/0957-4174","issn-type":"print"},{"key":"10.1016\/j.patcog.2024.111244_b18","doi-asserted-by":"crossref","unstructured":"K. Alberto Funes Mora, J.-M. Odobez, Geometric generative gaze estimation (g3e) for remote rgb-d cameras, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 1773\u20131780.","DOI":"10.1109\/CVPR.2014.229"},{"key":"10.1016\/j.patcog.2024.111244_b19","doi-asserted-by":"crossref","unstructured":"N.H. Cuong, H.T. Hoang, Eye-gaze detection with a single WebCAM based on geometry features extraction, in: 2010 11th International Conference on Control Automation Robotics & Vision, 2010, pp. 2507\u20132512.","DOI":"10.1109\/ICARCV.2010.5707319"},{"key":"10.1016\/j.patcog.2024.111244_b20","series-title":"2020 IEEE International Symposium on Mixed and Augmented Reality","first-page":"320","article-title":"Improved vergence and accommodation via purkinje image tracking with multiple cameras for ar glasses","author":"Lu","year":"2020"},{"key":"10.1016\/j.patcog.2024.111244_b21","doi-asserted-by":"crossref","first-page":"346","DOI":"10.1016\/j.ins.2015.02.004","article-title":"Real time gaze estimation with a consumer depth camera","volume":"320","author":"Sun","year":"2015","journal-title":"Inform. Sci."},{"key":"10.1016\/j.patcog.2024.111244_b22","doi-asserted-by":"crossref","unstructured":"K. Wang, Q. Ji, Real time eye gaze tracking with 3d deformable eye-face model, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 1003\u20131011.","DOI":"10.1109\/ICCV.2017.114"},{"key":"10.1016\/j.patcog.2024.111244_b23","doi-asserted-by":"crossref","unstructured":"D. Lian, Z. Zhang, W. Luo, L. Hu, M. Wu, Z. Li, J. Yu, S. Gao, RGBD based gaze estimation via multi-task CNN, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 33, 2019, pp. 2488\u20132495.","DOI":"10.1609\/aaai.v33i01.33012488"},{"key":"10.1016\/j.patcog.2024.111244_b24","doi-asserted-by":"crossref","unstructured":"Y. Yu, J.-M. Odobez, Unsupervised representation learning for gaze estimation, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 7314\u20137324.","DOI":"10.1109\/CVPR42600.2020.00734"},{"key":"10.1016\/j.patcog.2024.111244_b25","doi-asserted-by":"crossref","unstructured":"S. Park, S.D. Mello, P. Molchanov, U. Iqbal, O. Hilliges, J. Kautz, Few-shot adaptive gaze estimation, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2019, pp. 9368\u20139377.","DOI":"10.1109\/ICCV.2019.00946"},{"key":"10.1016\/j.patcog.2024.111244_b26","doi-asserted-by":"crossref","unstructured":"X. Zhang, Y. Sugano, M. Fritz, A. Bulling, It\u2019s written all over your face: Full-face appearance-based gaze estimation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 2017, pp. 51\u201360.","DOI":"10.1109\/CVPRW.2017.284"},{"key":"10.1016\/j.patcog.2024.111244_b27","article-title":"MPIIGaze: Real-world dataset and deep appearance-based gaze estimation","author":"Zhang","year":"2017","journal-title":"Cornell Univ. - arXiv IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.patcog.2024.111244_b28","doi-asserted-by":"crossref","unstructured":"Y. Cheng, S. Huang, F. Wang, C. Qian, F. Lu, A coarse-to-fine adaptive network for appearance-based gaze estimation, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 34, 2020, pp. 10623\u201310630.","DOI":"10.1609\/aaai.v34i07.6636"},{"key":"10.1016\/j.patcog.2024.111244_b29","doi-asserted-by":"crossref","unstructured":"K. Wang, R. Zhao, H. Su, Q. Ji, Generalizing eye tracking with bayesian adversarial learning, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 11907\u201311916.","DOI":"10.1109\/CVPR.2019.01218"},{"key":"10.1016\/j.patcog.2024.111244_b30","doi-asserted-by":"crossref","unstructured":"Y. Sun, J. Zeng, S. Shan, X. Chen, Cross-encoder for unsupervised gaze representation learning, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2021, pp. 3702\u20133711.","DOI":"10.1109\/ICCV48922.2021.00368"},{"issue":"10","key":"10.1016\/j.patcog.2024.111244_b31","doi-asserted-by":"crossref","first-page":"3010","DOI":"10.1109\/TNNLS.2018.2865525","article-title":"Multiview multitask gaze estimation with deep convolutional neural networks","volume":"30","author":"Lian","year":"2019","journal-title":"IEEE Trans. Neural Netw. Learn. Syst."},{"key":"10.1016\/j.patcog.2024.111244_b32","doi-asserted-by":"crossref","unstructured":"Y. Sugano, Y. Matsushita, Y. Sato, Learning-by-Synthesis for Appearance-Based 3D Gaze Estimation, in: 2014 IEEE Conference on Computer Vision and Pattern Recognition, 2014, pp. 1821\u20131828, http:\/\/dx.doi.org\/10.1109\/CVPR.2014.235.","DOI":"10.1109\/CVPR.2014.235"},{"key":"10.1016\/j.patcog.2024.111244_b33","doi-asserted-by":"crossref","first-page":"64904","DOI":"10.1109\/ACCESS.2022.3183357","article-title":"Customer gaze estimation in retail using deep learning","volume":"10","author":"Senarath","year":"2022","journal-title":"IEEE Access"},{"key":"10.1016\/j.patcog.2024.111244_b34","doi-asserted-by":"crossref","unstructured":"P. Pathirana, S. Senarath, D. Meedeniya, S. Jayarathna, Single-User 2D Gaze Estimation in Retail Environment Using Deep Learning, in: 2022 2nd International Conference on Advanced Research in Computing, ICARC, 2022, pp. 206\u2013211, http:\/\/dx.doi.org\/10.1109\/ICARC54489.2022.9754167.","DOI":"10.1109\/ICARC54489.2022.9754167"},{"key":"10.1016\/j.patcog.2024.111244_b35","doi-asserted-by":"crossref","first-page":"135","DOI":"10.1016\/j.neucom.2018.05.083","article-title":"Deep visual domain adaptation: A survey","volume":"312","author":"Wang","year":"2018","journal-title":"Neurocomputing"},{"key":"10.1016\/j.patcog.2024.111244_b36","doi-asserted-by":"crossref","DOI":"10.1109\/TKDE.2022.3178128","article-title":"Generalizing to unseen domains: A survey on domain generalization","author":"Wang","year":"2022","journal-title":"IEEE Trans. Knowl. Data Eng."},{"key":"10.1016\/j.patcog.2024.111244_b37","doi-asserted-by":"crossref","unstructured":"X. Yue, Y. Zhang, S. Zhao, A. Sangiovanni-Vincentelli, K. Keutzer, B. Gong, Domain randomization and pyramid consistency: Simulation-to-real generalization without accessing target domain data, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, 2019, pp. 2100\u20132110.","DOI":"10.1109\/ICCV.2019.00219"},{"key":"10.1016\/j.patcog.2024.111244_b38","doi-asserted-by":"crossref","unstructured":"D. Li, J. Yang, K. Kreis, A. Torralba, S. Fidler, Semantic segmentation with generative models: Semi-supervised learning and strong out-of-domain generalization, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 8300\u20138311.","DOI":"10.1109\/CVPR46437.2021.00820"},{"key":"10.1016\/j.patcog.2024.111244_b39","series-title":"Feature alignment and restoration for domain generalization and adaptation","author":"Jin","year":"2020"},{"key":"10.1016\/j.patcog.2024.111244_b40","article-title":"Metareg: Towards domain generalization using meta-regularization","volume":"31","author":"Balaji","year":"2018","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.patcog.2024.111244_b41","series-title":"Domain generalization with mixstyle","author":"Zhou","year":"2021"},{"key":"10.1016\/j.patcog.2024.111244_b42","series-title":"Generalizing across domains via cross-gradient training","author":"Shankar","year":"2018"},{"key":"10.1016\/j.patcog.2024.111244_b43","doi-asserted-by":"crossref","unstructured":"Y. Shu, Z. Cao, C. Wang, J. Wang, M. Long, Open domain generalization with domain-augmented meta-learning, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 9624\u20139633.","DOI":"10.1109\/CVPR46437.2021.00950"},{"key":"10.1016\/j.patcog.2024.111244_b44","series-title":"Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part VI 16","first-page":"756","article-title":"Domain2vec: Domain embedding for unsupervised domain adaptation","author":"Peng","year":"2020"},{"key":"10.1016\/j.patcog.2024.111244_b45","doi-asserted-by":"crossref","unstructured":"D. Li, J. Yang, K. Kreis, A. Torralba, S. Fidler, Semantic segmentation with generative models: Semi-supervised learning and strong out-of-domain generalization, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 8300\u20138311.","DOI":"10.1109\/CVPR46437.2021.00820"},{"key":"10.1016\/j.patcog.2024.111244_b46","series-title":"Pattern Recognition and Computer Vision: First Chinese Conference, PRCV 2018, Guangzhou, China, November 23-26, 2018, Proceedings, Part IV 1","first-page":"27","article-title":"Domain attention model for domain generalization in object detection","author":"He","year":"2018"},{"key":"10.1016\/j.patcog.2024.111244_b47","article-title":"Domain generalization via model-agnostic learning of semantic features","volume":"32","author":"Dou","year":"2019","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.patcog.2024.111244_b48","doi-asserted-by":"crossref","unstructured":"P. Kellnhofer, A. Recasens, S. Stent, W. Matusik, A. Torralba, Gaze360: Physically Unconstrained Gaze Estimation in the Wild, in: Proceedings of the IEEE\/CVF International Conference on Computer Vision, ICCV, 2019.","DOI":"10.1109\/ICCV.2019.00701"},{"key":"10.1016\/j.patcog.2024.111244_b49","doi-asserted-by":"crossref","unstructured":"Z. Guo, Z. Yuan, C. Zhang, W. Chi, Y. Ling, S. Zhang, Domain adaptation gaze estimation by embedding with prediction consistency, in: Proceedings of the Asian Conference on Computer Vision, 2020.","DOI":"10.1007\/978-3-030-69541-5_18"},{"key":"10.1016\/j.patcog.2024.111244_b50","doi-asserted-by":"crossref","unstructured":"Y. Wang, Y. Jiang, J. Li, B. Ni, W. Dai, C. Li, H. Xiong, T. Li, Contrastive regression for domain adaptation on gaze estimation, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 19376\u201319385.","DOI":"10.1109\/CVPR52688.2022.01877"},{"key":"10.1016\/j.patcog.2024.111244_b51","doi-asserted-by":"crossref","unstructured":"Y. Cheng, Y. Bao, F. Lu, PureGaze: Purifying Gaze Feature for Generalizable Gaze Estimation., in: Proceedings of the AAAI Conference on Artificial Intelligence, 2022, pp. 436\u2013443, http:\/\/dx.doi.org\/10.1609\/aaai.v36i1.19921.","DOI":"10.1609\/aaai.v36i1.19921"},{"key":"10.1016\/j.patcog.2024.111244_b52","doi-asserted-by":"crossref","unstructured":"I. Lee, J.-S. Yun, H.H. Kim, Y. Na, S.B. Yoo, LatentGaze: Cross-Domain Gaze Estimation through Gaze-Aware Analytic Latent Code Manipulation, in: Proceedings of the Asian Conference on Computer Vision, 2022, pp. 3379\u20133395.","DOI":"10.1007\/978-3-031-26348-4_10"},{"key":"10.1016\/j.patcog.2024.111244_b53","doi-asserted-by":"crossref","unstructured":"M. Xu, H. Wang, F. Lu, Learning a generalized gaze estimator from gaze-consistent feature, in: Proceedings of the AAAI Conference on Artificial Intelligence, Vol. 37, 2023, pp. 3027\u20133035.","DOI":"10.1609\/aaai.v37i3.25406"},{"key":"10.1016\/j.patcog.2024.111244_b54","article-title":"Domain separation networks","volume":"29","author":"Bousmalis","year":"2016","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"10.1016\/j.patcog.2024.111244_b55","series-title":"Adversarial multi-task learning for text classification","author":"Liu","year":"2017"},{"key":"10.1016\/j.patcog.2024.111244_b56","doi-asserted-by":"crossref","unstructured":"J.-Y. Zhu, T. Park, P. Isola, A.A. Efros, Unpaired image-to-image translation using cycle-consistent adversarial networks, in: Proceedings of the IEEE International Conference on Computer Vision, 2017, pp. 2223\u20132232.","DOI":"10.1109\/ICCV.2017.244"},{"key":"10.1016\/j.patcog.2024.111244_b57","series-title":"Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part V 16","first-page":"365","article-title":"Eth-xgaze: A large scale dataset for gaze estimation under extreme head pose and gaze variation","author":"Zhang","year":"2020"},{"key":"10.1016\/j.patcog.2024.111244_b58","doi-asserted-by":"crossref","unstructured":"K.A. Funes Mora, F. Monay, J.-M. Odobez, EYEDIAP: a database for the development and evaluation of gaze estimation algorithms from RGB and RGB-D cameras, in: Proceedings of the Symposium on Eye Tracking Research and Applications, 2014, http:\/\/dx.doi.org\/10.1145\/2578153.2578190.","DOI":"10.1145\/2578153.2578190"},{"key":"10.1016\/j.patcog.2024.111244_b59","doi-asserted-by":"crossref","DOI":"10.1109\/TPAMI.2024.3393571","article-title":"Appearance-based gaze estimation with deep learning: A review and benchmark","author":"Cheng","year":"2024","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"10.1016\/j.patcog.2024.111244_b60","series-title":"Adam: A method for stochastic optimization","author":"Kingma","year":"2014"},{"key":"10.1016\/j.patcog.2024.111244_b61","doi-asserted-by":"crossref","unstructured":"T. Fischer, H.J. Chang, Y. Demiris, Rt-gene: Real-time eye gaze estimation in natural environments, in: Proceedings of the European Conference on Computer Vision, ECCV, 2018, pp. 334\u2013352.","DOI":"10.1007\/978-3-030-01249-6_21"},{"key":"10.1016\/j.patcog.2024.111244_b62","doi-asserted-by":"crossref","unstructured":"E. Tzeng, J. Hoffman, K. Saenko, T. Darrell, Adversarial discriminative domain adaptation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 7167\u20137176.","DOI":"10.1109\/CVPR.2017.316"},{"key":"10.1016\/j.patcog.2024.111244_b63","doi-asserted-by":"crossref","unstructured":"S. Cui, S. Wang, J. Zhuo, C. Su, Q. Huang, Q. Tian, Gradually vanishing bridge for adversarial domain adaptation, in: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 12455\u201312464.","DOI":"10.1109\/CVPR42600.2020.01247"},{"key":"10.1016\/j.patcog.2024.111244_b64","doi-asserted-by":"crossref","unstructured":"M. Cai, F. Lu, Y. Sato, Generalizing hand segmentation in egocentric videos with uncertainty-guided model adaptation, in: Proceedings of the Ieee\/Cvf Conference on Computer Vision and Pattern Recognition, 2020, pp. 14392\u201314401.","DOI":"10.1109\/CVPR42600.2020.01440"}],"container-title":["Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320324009956?httpAccept=text\/xml","content-type":"text\/xml","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/api.elsevier.com\/content\/article\/PII:S0031320324009956?httpAccept=text\/plain","content-type":"text\/plain","content-version":"vor","intended-application":"text-mining"}],"deposited":{"date-parts":[[2025,2,20]],"date-time":"2025-02-20T12:05:45Z","timestamp":1740053145000},"score":1,"resource":{"primary":{"URL":"https:\/\/linkinghub.elsevier.com\/retrieve\/pii\/S0031320324009956"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5]]},"references-count":64,"alternative-id":["S0031320324009956"],"URL":"https:\/\/doi.org\/10.1016\/j.patcog.2024.111244","relation":{},"ISSN":["0031-3203"],"issn-type":[{"value":"0031-3203","type":"print"}],"subject":[],"published":{"date-parts":[[2025,5]]},"assertion":[{"value":"Elsevier","name":"publisher","label":"This article is maintained by"},{"value":"Collaborative contrastive learning for cross-domain gaze estimation","name":"articletitle","label":"Article Title"},{"value":"Pattern Recognition","name":"journaltitle","label":"Journal Title"},{"value":"https:\/\/doi.org\/10.1016\/j.patcog.2024.111244","name":"articlelink","label":"CrossRef DOI link to publisher maintained version"},{"value":"article","name":"content_type","label":"Content Type"},{"value":"\u00a9 2024 Published by Elsevier Ltd.","name":"copyright","label":"Copyright"}],"article-number":"111244"}}