{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,11]],"date-time":"2025-12-11T21:04:14Z","timestamp":1765487054187,"version":"3.37.3"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"36","license":[{"start":{"date-parts":[[2024,3,26]],"date-time":"2024-03-26T00:00:00Z","timestamp":1711411200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,3,26]],"date-time":"2024-03-26T00:00:00Z","timestamp":1711411200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100004608","name":"Natural Science Foundation of Jiangsu Province","doi-asserted-by":"publisher","award":["BK20180594","BK20231036"],"award-info":[{"award-number":["BK20180594","BK20231036"]}],"id":[{"id":"10.13039\/501100004608","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-18941-2","type":"journal-article","created":{"date-parts":[[2024,3,26]],"date-time":"2024-03-26T08:04:21Z","timestamp":1711440261000},"page":"83867-83887","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["Highly efficient gaze estimation method using online convolutional re-parameterization"],"prefix":"10.1007","volume":"83","author":[{"given":"De","family":"Gu","sequence":"first","affiliation":[]},{"given":"Minghao","family":"Lv","sequence":"additional","affiliation":[]},{"given":"Jianchu","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Mari Anne","family":"Snow","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,26]]},"reference":[{"key":"18941_CR1","doi-asserted-by":"crossref","unstructured":"Bao Y, Cheng Y, Liu Y et\u00a0al (2021) Adaptive feature fusion network for gaze tracking in mobile tablets. In: 2020 25th international conference on pattern recognition (ICPR). IEEE, pp 9936\u20139943","DOI":"10.1109\/ICPR48806.2021.9412205"},{"key":"18941_CR2","doi-asserted-by":"crossref","unstructured":"Chen Z, Shi BE (2018) Appearance-based gaze estimation using dilated-convolutions. In: Asian conference on computer vision. Springer, pp 309\u2013324","DOI":"10.1007\/978-3-030-20876-9_20"},{"key":"18941_CR3","doi-asserted-by":"crossref","unstructured":"Cheng Y, Lu F (2022) Gaze estimation using transformer. In: 2022 26th international conference on pattern recognition (ICPR). IEEE, pp 3341\u20133347","DOI":"10.1109\/ICPR56361.2022.9956687"},{"key":"18941_CR4","doi-asserted-by":"crossref","unstructured":"Cheng Y, Lu F, Zhang X (2018) Appearance-based gaze estimation via evaluation-guided asymmetric regression. In: Proceedings of the European conference on computer vision (ECCV). pp 100\u2013115","DOI":"10.1007\/978-3-030-01264-9_7"},{"key":"18941_CR5","doi-asserted-by":"crossref","unstructured":"Cheng Y, Huang S, Wang F et\u00a0al (2020) A coarse-to-fine adaptive network for appearance-based gaze estimation. In: Proceedings of the AAAI conference on artificial intelligence. pp 10623\u201310630","DOI":"10.1609\/aaai.v34i07.6636"},{"key":"18941_CR6","doi-asserted-by":"publisher","first-page":"5259","DOI":"10.1109\/TIP.2020.2982828","volume":"29","author":"Y Cheng","year":"2020","unstructured":"Cheng Y, Zhang X, Lu F et al (2020) Gaze estimation by exploring two-eye asymmetry. IEEE Trans Image Process 29:5259\u20135272","journal-title":"IEEE Trans Image Process"},{"key":"18941_CR7","unstructured":"Cheng Y, Wang H, Bao Y et\u00a0al (2021) Appearance-based gaze estimation with deep learning: a review and benchmark. arXiv:2104.12668"},{"key":"18941_CR8","doi-asserted-by":"crossref","unstructured":"Ding X, Guo Y, Ding G et\u00a0al (2019) ACNet: strengthening the kernel skeletons for powerful CNN via asymmetric convolution blocks. In: Proceedings of the IEEE\/CVF international conference on computer vision (ICCV)","DOI":"10.1109\/ICCV.2019.00200"},{"key":"18941_CR9","doi-asserted-by":"crossref","unstructured":"Ding X, Zhang X, Han J et\u00a0al (2021) Diverse branch block: building a convolution as an inception-like unit. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 10886\u201310895","DOI":"10.1109\/CVPR46437.2021.01074"},{"key":"18941_CR10","doi-asserted-by":"crossref","unstructured":"Ding X, Zhang X, Ma N et\u00a0al (2021) RepVGG: making VGG-style convnets great again. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 13733\u201313742","DOI":"10.1109\/CVPR46437.2021.01352"},{"key":"18941_CR11","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A et\u00a0al (2020) An image is worth 16x16 words: transformers for image recognition at scale. arXiv:2010.11929"},{"key":"18941_CR12","doi-asserted-by":"crossref","unstructured":"Fischer T, Chang HJ, Demiris Y (2018) RT-GENE: real-time eye gaze estimation in natural environments. In: Proceedings of the European conference on computer vision (ECCV). pp 334\u2013352","DOI":"10.1007\/978-3-030-01249-6_21"},{"key":"18941_CR13","doi-asserted-by":"crossref","unstructured":"Funes\u00a0Mora KA, Monay F, Odobez JM (2014) Eyediap: a database for the development and evaluation of gaze estimation algorithms from RGB and RGB-D cameras. In: Proceedings of the symposium on eye tracking research and applications. pp 255\u2013258","DOI":"10.1145\/2578153.2578190"},{"issue":"6","key":"18941_CR14","doi-asserted-by":"publisher","first-page":"1124","DOI":"10.1109\/TBME.2005.863952","volume":"53","author":"ED Guestrin","year":"2006","unstructured":"Guestrin ED, Eizenman M (2006) General theory of remote gaze estimation using the pupil center and corneal reflections. IEEE Trans Biomed Eng 53(6):1124\u20131133","journal-title":"IEEE Trans Biomed Eng"},{"key":"18941_CR15","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S et\u00a0al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"18941_CR16","doi-asserted-by":"crossref","unstructured":"Hu M, Feng J, Hua J et\u00a0al (2022) Online convolutional re-parameterization. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 568\u2013577","DOI":"10.1109\/CVPR52688.2022.00065"},{"key":"18941_CR17","doi-asserted-by":"crossref","unstructured":"Huang T, You S, Zhang B et\u00a0al (2022) DyRep: bootstrapping training with dynamic re-parameterization. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 588\u2013597","DOI":"10.1109\/CVPR52688.2022.00067"},{"key":"18941_CR18","doi-asserted-by":"crossref","unstructured":"Kellnhofer P, Recasens A, Stent S et\u00a0al (2019) Gaze360: physically unconstrained gaze estimation in the wild. In: Proceedings of the IEEE\/CVF international conference on computer vision. pp 6912\u20136921","DOI":"10.1109\/ICCV.2019.00701"},{"key":"18941_CR19","doi-asserted-by":"crossref","unstructured":"Krafka K, Khosla A, Kellnhofer P et\u00a0al (2016) Eye tracking for everyone. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 2176\u20132184","DOI":"10.1109\/CVPR.2016.239"},{"key":"18941_CR20","unstructured":"Krizhevsky A, Sutskever I, Hinton GE (2012) ImageNet classification with deep convolutional neural networks. Adv Neural Inf Process Syst 25"},{"key":"18941_CR21","unstructured":"Li Y, Zhang K, Cao J et\u00a0al (2021) LocalViT: bringing locality to vision transformers. arXiv:2104.05707"},{"key":"18941_CR22","doi-asserted-by":"crossref","unstructured":"Ma N, Zhang X, Zheng HT et\u00a0al (2018) ShuffleNet V2: practical guidelines for efficient CNN architecture design. In: Proceedings of the European conference on computer vision (ECCV). pp 116\u2013131","DOI":"10.1007\/978-3-030-01264-9_8"},{"issue":"2","key":"18941_CR23","doi-asserted-by":"publisher","first-page":"141","DOI":"10.1109\/TIV.2018.2804160","volume":"3","author":"S Martin","year":"2018","unstructured":"Martin S, Vora S, Yuen K et al (2018) Dynamics of driver\u2019s gaze: explorations in behavior modeling and maneuver prediction. IEEE Trans Intell Veh 3(2):141\u2013150","journal-title":"IEEE Trans Intell Veh"},{"issue":"11","key":"18941_CR24","doi-asserted-by":"publisher","first-page":"2711","DOI":"10.1109\/TPAMI.2017.2782819","volume":"40","author":"B Mass\u00e9","year":"2017","unstructured":"Mass\u00e9 B, Ba S, Horaud R (2017) Tracking gaze and visual focus of attention of people involved in social interaction. IEEE Trans Pattern Anal Mach Intell 40(11):2711\u20132724","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"2","key":"18941_CR25","doi-asserted-by":"publisher","first-page":"590","DOI":"10.1177\/1094428117744882","volume":"22","author":"M Mei\u00dfner","year":"2019","unstructured":"Mei\u00dfner M, Oll J (2019) The promise of eye-tracking methodology in organizational research: a taxonomy, review, and future avenues. Organ Res Methods 22(2):590\u2013617","journal-title":"Organ Res Methods"},{"key":"18941_CR26","doi-asserted-by":"crossref","unstructured":"Murthy L, Biswas P (2021) Appearance-based gaze estimation using attention and difference mechanism. In: 2021 IEEE\/CVF conference on computer vision and pattern recognition workshops (CVPRW). IEEE, pp 3137\u20133146","DOI":"10.1109\/CVPRW53098.2021.00351"},{"key":"18941_CR27","doi-asserted-by":"crossref","unstructured":"O\u00a0Oh J, Chang HJ, Choi SI (2022) Self-attention with convolution and deconvolution for efficient eye gaze estimation from a full face image. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 4992\u20135000","DOI":"10.1109\/CVPRW56347.2022.00547"},{"key":"18941_CR28","doi-asserted-by":"publisher","first-page":"103369","DOI":"10.1016\/j.jvcir.2021.103369","volume":"81","author":"D Ren","year":"2021","unstructured":"Ren D, Chen J, Zhong J et al (2021) Gaze estimation via bilinear pooling-based attention networks. Journal of Visual CommunImage Represent 81:103369. https:\/\/doi.org\/10.1016\/j.jvcir.2021.103369","journal-title":"Journal of Visual CommunImage Represent"},{"issue":"2","key":"18941_CR29","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1002\/npr2.12046","volume":"39","author":"E Shishido","year":"2019","unstructured":"Shishido E, Ogawa S, Miyata S et al (2019) Application of eye trackers for understanding mental disorders: cases for schizophrenia and autism spectrum disorder. Neuropsychopharmacology Rep 39(2):72\u201377. https:\/\/doi.org\/10.1002\/npr2.12046","journal-title":"Neuropsychopharmacology Rep"},{"key":"18941_CR30","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556"},{"key":"18941_CR31","unstructured":"Vasu PKA, Gabriel J, Zhu J et\u00a0al (2023) FastViT: a fast hybrid vision transformer using structural reparameterization. arXiv:2303.14189"},{"key":"18941_CR32","doi-asserted-by":"crossref","unstructured":"Vasu PKA, Gabriel J, Zhu J et\u00a0al (2023) Mobileone: an improved one millisecond mobile backbone. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition. pp 7907\u20137917","DOI":"10.1109\/CVPR52729.2023.00764"},{"key":"18941_CR33","unstructured":"Vaswani A, Shazeer N, Parmar N et\u00a0al (2017) Attention is all you need. Adv Neural Inf Process Syst 30"},{"key":"18941_CR34","unstructured":"Wadekar SN, Chaurasia A (2022) MobileViTv3: mobile-friendly vision transformer with simple and effective fusion of local, global and input features. arXiv:2209.15159"},{"issue":"3","key":"18941_CR35","doi-asserted-by":"publisher","first-page":"415","DOI":"10.1007\/s41095-022-0274-8","volume":"8","author":"W Wang","year":"2022","unstructured":"Wang W, Xie E, Li X et al (2022) PVT v2: improved baselines with pyramid vision transformer. Comput Vis Media 8(3):415\u2013424","journal-title":"Comput Vis Media"},{"key":"18941_CR36","doi-asserted-by":"publisher","unstructured":"Wang X, Zhou J, Wang L et\u00a0al (2023) BoT2L-Net: appearance-based gaze estimation using bottleneck transformer block and two identical losses in unconstrained environments. Electron 12(7). https:\/\/doi.org\/10.3390\/electronics12071704","DOI":"10.3390\/electronics12071704"},{"key":"18941_CR37","unstructured":"Xu T, Wu B, Fan R et\u00a0al (2023) FR-Net: a light-weight FFT residual net for gaze estimation. arXiv:2305.11875"},{"key":"18941_CR38","doi-asserted-by":"crossref","unstructured":"Xu Y, Dong Y, Wu J et\u00a0al (2018) Gaze prediction in dynamic 360 immersive videos. In: Proceedings of the IEEE conference on computer vision and pattern recognition. pp 5333\u20135342","DOI":"10.1109\/CVPR.2018.00559"},{"key":"18941_CR39","doi-asserted-by":"crossref","unstructured":"Zhang X, Sugano Y, Fritz M et\u00a0al (2015) Appearance-based gaze estimation in the wild. In: The IEEE conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR.2015.7299081"},{"key":"18941_CR40","doi-asserted-by":"crossref","unstructured":"Zhang X, Sugano Y, Fritz M et\u00a0al (2017) It\u2019s written all over your face: full-face appearance-based gaze estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition workshops. pp 51\u201360","DOI":"10.1109\/CVPRW.2017.284"},{"issue":"1","key":"18941_CR41","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1109\/TPAMI.2017.2778103","volume":"41","author":"X Zhang","year":"2019","unstructured":"Zhang X, Sugano Y, Fritz M et al (2019) MPIIGaze: real-world dataset and deep appearance-based gaze estimation. IEEE Trans Pattern Anal Mach Intell 41(1):162\u2013175. https:\/\/doi.org\/10.1109\/TPAMI.2017.2778103","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"18941_CR42","doi-asserted-by":"crossref","unstructured":"Zhang X, Park S, Beeler T et\u00a0al (2020) ETH-XGaze: a large scale dataset for gaze estimation under extreme head pose and gaze variation. In: Computer vision\u2013ECCV 2020: 16th European conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part V 16. Springer, pp 365\u2013381","DOI":"10.1007\/978-3-030-58558-7_22"},{"issue":"12","key":"18941_CR43","doi-asserted-by":"publisher","first-page":"2246","DOI":"10.1109\/TBME.2007.895750","volume":"54","author":"Z Zhu","year":"2007","unstructured":"Zhu Z, Ji Q (2007) Novel eye gaze tracking techniques under natural head movement. IEEE Trans Biomed Eng 54(12):2246\u20132260","journal-title":"IEEE Trans Biomed Eng"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-18941-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-18941-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-18941-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,14]],"date-time":"2024-11-14T13:16:41Z","timestamp":1731590201000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-18941-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,26]]},"references-count":43,"journal-issue":{"issue":"36","published-online":{"date-parts":[[2024,11]]}},"alternative-id":["18941"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-18941-2","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,3,26]]},"assertion":[{"value":"23 November 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"28 February 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 March 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 March 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that there is no conflict of interest related to this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}]}}