{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T21:17:58Z","timestamp":1769894278797,"version":"3.49.0"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2022,4,20]],"date-time":"2022-04-20T00:00:00Z","timestamp":1650412800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2022,4,20]],"date-time":"2022-04-20T00:00:00Z","timestamp":1650412800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Sci. China Inf. Sci."],"published-print":{"date-parts":[[2022,6]]},"abstract":"<jats:title>Abstract<\/jats:title><jats:p>Onfocus detection aims at identifying whether the focus of the individual captured by a camera is on the camera or not. Based on the behavioral research, the focus of an individual during face-to-camera communication leads to a special type of eye contact, i.e., the individual-camera eye contact, which is a powerful signal in social communication and plays a crucial role in recognizing irregular individual status (e.g., lying or suffering mental disease) and special purposes (e.g., seeking help or attracting fans). Thus, developing effective onfocus detection algorithms is of significance for assisting the criminal investigation, disease discovery, and social behavior analysis. However, the review of the literature shows that very few efforts have been made toward the development of onfocus detector owing to the lack of large-scale public available datasets as well as the challenging nature of this task. To this end, this paper engages in the onfocus detection research by addressing the above two issues. Firstly, we build a large-scale onfocus detection dataset, named as the onfocus detection in the wild (OFDIW). It consists of 20623 images in unconstrained capture conditions (thus called \u201cin the wild\u201d) and contains individuals with diverse emotions, ages, facial characteristics, and rich interactions with surrounding objects and background scenes. On top of that, we propose a novel end-to-end deep model, i.e., the eye-context interaction inferring network (ECIIN), for onfocus detection, which explores eye-context interaction via dynamic capsule routing. Finally, comprehensive experiments are conducted on the proposed OFDIW dataset to benchmark the existing learning models and demonstrate the effectiveness of the proposed ECIIN.<\/jats:p>","DOI":"10.1007\/s11432-020-3181-9","type":"journal-article","created":{"date-parts":[[2022,5,12]],"date-time":"2022-05-12T15:07:30Z","timestamp":1652368050000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":21,"title":["Onfocus detection: identifying individual-camera eye contact from unconstrained images"],"prefix":"10.1007","volume":"65","author":[{"given":"Dingwen","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Bo","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Gerong","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Qiang","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jiajia","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Jungong","family":"Han","sequence":"additional","affiliation":[]},{"given":"Zheng","family":"You","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,20]]},"reference":[{"key":"3181_CR1","doi-asserted-by":"publisher","first-page":"887","DOI":"10.1177\/0093854812437014","volume":"39","author":"J J Walczyk","year":"2012","unstructured":"Walczyk J J, Griffith D A, Yates R, et al. LIE detection by inducing cognitive load. Criminal Justice Behav, 2012, 39: 887\u2013909","journal-title":"Criminal Justice Behav"},{"key":"3181_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3131902","volume":"1","author":"E Chong","year":"2017","unstructured":"Chong E, Chanda K, Ye Z, et al. Detecting gaze towards eyes in natural social interactions and its use in child assessment. Proc ACM Interact Mob Wearable Ubiquitous Technol, 2017, 1: 1\u201320","journal-title":"Proc ACM Interact Mob Wearable Ubiquitous Technol"},{"key":"3181_CR3","doi-asserted-by":"publisher","first-page":"107","DOI":"10.1177\/1745691616654457","volume":"12","author":"T Grossmann","year":"2017","unstructured":"Grossmann T. The eyes as windows into other minds. Perspect Psychol Sci, 2017, 12: 107\u2013121","journal-title":"Perspect Psychol Sci"},{"key":"3181_CR4","doi-asserted-by":"publisher","first-page":"73","DOI":"10.1016\/j.jcms.2018.05.034","volume":"47","author":"L Prantl","year":"2019","unstructured":"Prantl L, Heidekrueger P I, Broer P N, et al. Female eye attractiveness\u2014Where beauty meets science. J Cranio-Maxillofacial Surgery, 2019, 47: 73\u201379","journal-title":"J Cranio-Maxillofacial Surgery"},{"key":"3181_CR5","doi-asserted-by":"publisher","first-page":"1884","DOI":"10.1109\/TPAMI.2019.2906588","volume":"42","author":"W Zhao","year":"2020","unstructured":"Zhao W, Zhao F, Wang D, et al. Defocus blur detection via multi-stream bottom-top-bottom network. IEEE Trans Pattern Anal Mach Intell, 2020, 42: 1884\u20131897","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3181_CR6","unstructured":"Huang G B, Learned-Miller E. Labeled Faces in the Wild: Updates and New Reporting Procedures. Technical Report, 14-003. 2014"},{"key":"3181_CR7","doi-asserted-by":"crossref","unstructured":"Parkhi O M, Vedaldi A, Zisserman A, et al. Cats and dogs. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Providence, 2012. 3498\u20133505","DOI":"10.1109\/CVPR.2012.6248092"},{"key":"3181_CR8","unstructured":"Hinton G E, Sabour S, Frosst N. Matrix capsules with EM routing. In: Proceedings of International Conference on Learning Representations, Vancouver, 2018"},{"key":"3181_CR9","doi-asserted-by":"crossref","unstructured":"Qodseya M, Panta F, Sedes F. Visual-based eye contact detection in multi-person interactions. In: Proceedings of 2019 International Conference on Content-Based Multimedia Indexing, Dublin, 2019. 1\u20136","DOI":"10.1109\/CBMI.2019.8877471"},{"key":"3181_CR10","doi-asserted-by":"crossref","unstructured":"Li B, Zhang H, He L, et al. Eye contact detection via CNN-based gaze direction regression. In: Proceedings of 2019 WRC Symposium on Advanced Robotics and Automation, Beijing, 2019. 425\u2013431","DOI":"10.1109\/WRC-SARA.2019.8931950"},{"key":"3181_CR11","doi-asserted-by":"crossref","unstructured":"Mitsuzumi Y, Nakazawa A, Nishida T. DEEP eye contact detector: robust eye contact bid detection using convolutional neural network. In: Proceedings of British Machine Vision Conference, London, 2017","DOI":"10.5244\/C.31.59"},{"key":"3181_CR12","doi-asserted-by":"crossref","unstructured":"Ye Z, Li Y, Fathi A, et al. Detecting eye contact using wearable eye-tracking glasses. In: Proceedings of the 2012 ACM Conference on Ubiquitous Computing, Pittsburgh, 2012. 699\u2013704","DOI":"10.1145\/2370216.2370368"},{"key":"3181_CR13","doi-asserted-by":"crossref","unstructured":"Zhang X, Sugano Y, Bulling A. Everyday eye contact detection using unsupervised gaze target discovery. In: Proceedings of the 30th Annual ACM Symposium on User Interface Software and Technology, Qu\u00e9bec City, 2017. 193\u2013203","DOI":"10.1145\/3126594.3126614"},{"key":"3181_CR14","doi-asserted-by":"crossref","unstructured":"Smith B A, Yin Q, Feiner S K, et al. Gaze locking: passive eye contact detection for human-object interaction. In: Proceedings of the 26th Annual ACM Symposium on User Interface Software and Technology, St. Andrews Scotland, 2013. 271\u2013280","DOI":"10.1145\/2501988.2501994"},{"key":"3181_CR15","doi-asserted-by":"crossref","unstructured":"Zhang X, Sugano Y, Fritz M, et al. Appearance-based gaze estimation in the wild. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, 2015. 4511\u20134520","DOI":"10.1109\/CVPR.2015.7299081"},{"key":"3181_CR16","doi-asserted-by":"publisher","first-page":"445","DOI":"10.1007\/s00138-017-0852-4","volume":"28","author":"Q Huang","year":"2017","unstructured":"Huang Q, Veeraraghavan A, Sabharwal A. TabletGaze: dataset and analysis for unconstrained appearance-based gaze estimation in mobile tablets. Machine Vision Appl, 2017, 28: 445\u2013461","journal-title":"Machine Vision Appl"},{"key":"3181_CR17","doi-asserted-by":"crossref","unstructured":"Park S, Spurr A, Hilliges O. Deep pictorial gaze estimation. In: Proceedings of the European Conference on Computer Vision, Munich, 2018. 721\u2013738","DOI":"10.1007\/978-3-030-01261-8_44"},{"key":"3181_CR18","doi-asserted-by":"publisher","first-page":"3010","DOI":"10.1109\/TNNLS.2018.2865525","volume":"30","author":"D Lian","year":"2019","unstructured":"Lian D, Hu L, Luo W, et al. Multiview multitask gaze estimation with deep convolutional neural networks. IEEE Trans Neural Netw Learn Syst, 2019, 30: 3010\u20133023","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"key":"3181_CR19","doi-asserted-by":"crossref","unstructured":"Kellnhofer P, Recasens A, Stent S, et al. Gaze360: physically unconstrained gaze estimation in the wild. In: Proceedings of the IEEE International Conference on Computer Vision, Long Beach, 2019. 6912\u20136921","DOI":"10.1109\/ICCV.2019.00701"},{"key":"3181_CR20","doi-asserted-by":"crossref","unstructured":"M\u00fcller P, Huang M X, Zhang X, et al. Robust eye contact detection in natural multi-person interactions using gaze and speaking behaviour. In: Proceedings of the 2018 ACM Symposium on Eye Tracking Research & Applications, Warsaw Poland, 2018. 1\u201310","DOI":"10.1145\/3204493.3204549"},{"key":"3181_CR21","unstructured":"Mora K A F, Monay F, Odobez J M. Eyediap: a database for the development and evaluation of gaze estimation algorithms from RGB and RGB-D cameras. In: Proceedings of the Symposium on Eye Tracking Research and Applications, Safety Harbor, 2014. 255\u2013258"},{"key":"3181_CR22","doi-asserted-by":"crossref","unstructured":"Rehg J, Abowd G, Rozga A, et al. Decoding children\u2019s social behavior. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland Oregon, 2013. 3414\u20133421","DOI":"10.1109\/CVPR.2013.438"},{"key":"3181_CR23","doi-asserted-by":"crossref","unstructured":"Zhang X, Sugano Y, Fritz M, et al. It\u2019s written all over your face: full-face appearance-based gaze estimation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu Hawaii, 2017. 51\u201360","DOI":"10.1109\/CVPRW.2017.284"},{"key":"3181_CR24","doi-asserted-by":"crossref","unstructured":"Krafka K, Khosla A, Kellnhofer P, et al. Eye tracking for everyone. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas Nevada, 2016. 2176\u20132184","DOI":"10.1109\/CVPR.2016.239"},{"key":"3181_CR25","doi-asserted-by":"crossref","unstructured":"Parekh V, Subramanian R, Jawahar C V. Eye contact detection via deep neural networks. In: Proceedings of International Conference on Human-Computer Interaction, Vancouver, 2017. 366\u2013374","DOI":"10.1007\/978-3-319-58750-9_51"},{"key":"3181_CR26","doi-asserted-by":"crossref","unstructured":"Ye Z, Li Y, Liu Y, et al. Detecting bids for eye contact using a wearable camera. In: Proceedings of the 11th IEEE International Conference and Workshops on Automatic Face and Gesture Recognition, Ljubljana, 2015. 1\u20138","DOI":"10.1109\/FG.2015.7163095"},{"key":"3181_CR27","unstructured":"Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. 2014. ArXiv:1409.1556"},{"key":"3181_CR28","doi-asserted-by":"crossref","unstructured":"Zhou B, Khosla A, Lapedriza A, et al. Learning deep features for discriminative localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas Nevada, 2016. 2921\u20132929","DOI":"10.1109\/CVPR.2016.319"},{"key":"3181_CR29","doi-asserted-by":"crossref","unstructured":"Cheng Y, Huang S, Wang F, et al. A coarse-to-fine adaptive network for appearance-based gaze estimation. In: Proceedings of the AAAI Conference on Artificial Intelligence, New York, 2020. 10623\u201310630","DOI":"10.1609\/aaai.v34i07.6636"},{"key":"3181_CR30","doi-asserted-by":"crossref","unstructured":"Yang Z, Luo T, Wang D, et al. Learning to navigate for fine-grained classification. In: Proceedings of the European Conference on Computer Vision, Munich, 2018. 420\u2013435","DOI":"10.1007\/978-3-030-01264-9_26"},{"key":"3181_CR31","doi-asserted-by":"crossref","unstructured":"Li P, Xie J, Wang Q, et al. Towards faster training of global covariance pooling networks by iterative matrix square root normalization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, 2018. 947\u2013955","DOI":"10.1109\/CVPR.2018.00105"},{"key":"3181_CR32","doi-asserted-by":"crossref","unstructured":"Wang Y, Morariu V I, Davis L S. Learning a discriminative filter bank within a CNN for fine-grained recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, 2018. 4148\u20134157","DOI":"10.1109\/CVPR.2018.00436"},{"key":"3181_CR33","doi-asserted-by":"publisher","first-page":"1309","DOI":"10.1109\/TPAMI.2017.2723400","volume":"40","author":"T Y Lin","year":"2018","unstructured":"Lin T Y, RoyChowdhury A, Maji S. Bilinear convolutional neural networks for fine-grained visual recognition. IEEE Trans Pattern Anal Mach Intell, 2018, 40: 1309\u20131322","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3181_CR34","doi-asserted-by":"crossref","unstructured":"Chen Y, Bai Y, Zhang W, et al. Destruction and construction learning for fine-grained image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, 2019. 5157\u20135166","DOI":"10.1109\/CVPR.2019.00530"},{"key":"3181_CR35","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, et al. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, 2016. 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"3181_CR36","doi-asserted-by":"publisher","first-page":"652","DOI":"10.1109\/TPAMI.2019.2938758","volume":"43","author":"S H Gao","year":"2021","unstructured":"Gao S H, Cheng M M, Zhao K, et al. Res2Net: a new multi-scale backbone architecture. IEEE Trans Pattern Anal Mach Intell, 2021, 43: 652\u2013662","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3181_CR37","doi-asserted-by":"publisher","unstructured":"Huang G, Liu Z, Pleiss G, et al. Convolutional networks with dense connectivity. IEEE Trans Pattern Anal Mach Intell, 2019. doi: https:\/\/doi.org\/10.1109\/TPAMI.2019.2918284","DOI":"10.1109\/TPAMI.2019.2918284"},{"key":"3181_CR38","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, 2018. 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"3181_CR39","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky O, Deng J, Su H, et al. ImageNet large scale visual recognition challenge. Int J Comput Vis, 2015, 115: 211\u2013252","journal-title":"Int J Comput Vis"},{"key":"3181_CR40","doi-asserted-by":"crossref","unstructured":"Wang G, Wang K, Lin L. Adaptively connected neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Long Beach, 2019. 1781\u20131790","DOI":"10.1109\/CVPR.2019.00188"},{"key":"3181_CR41","doi-asserted-by":"crossref","unstructured":"Tokozume Y, Ushiku Y, Harada T. Between-class learning for image classification. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, 2018. 5486\u20135494","DOI":"10.1109\/CVPR.2018.00575"}],"container-title":["Science China Information Sciences"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11432-020-3181-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11432-020-3181-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11432-020-3181-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,9,24]],"date-time":"2024-09-24T15:28:18Z","timestamp":1727191698000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11432-020-3181-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,20]]},"references-count":41,"journal-issue":{"issue":"6","published-print":{"date-parts":[[2022,6]]}},"alternative-id":["3181"],"URL":"https:\/\/doi.org\/10.1007\/s11432-020-3181-9","relation":{},"ISSN":["1674-733X","1869-1919"],"issn-type":[{"value":"1674-733X","type":"print"},{"value":"1869-1919","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,4,20]]},"assertion":[{"value":"16 September 2020","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 December 2020","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 January 2021","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 April 2022","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"160101"}}