{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,28]],"date-time":"2025-03-28T01:48:27Z","timestamp":1743126507165,"version":"3.40.3"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031202322"},{"type":"electronic","value":"9783031202339"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-20233-9_16","type":"book-chapter","created":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T00:02:48Z","timestamp":1667433768000},"page":"158-168","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Estimation of Gaze-Following Based on Transformer and the Guiding Offset"],"prefix":"10.1007","author":[{"given":"Sheng","family":"Gao","sequence":"first","affiliation":[]},{"given":"Xiao","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Jia","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,3]]},"reference":[{"key":"16_CR1","unstructured":"Recasens, A., Khosla, A., Vondrick, C., Torralba, A.: Where are they looking? Advances in neural information processing systems, 28 (2015)"},{"key":"16_CR2","doi-asserted-by":"crossref","unstructured":"Lian, D., Yu, Z., Gao, S.: Believe it or not, we know what you are looking at! In Asian Conference on Computer Vision, pp. 35\u201350. Springer (2018)","DOI":"10.1007\/978-3-030-20893-6_3"},{"key":"16_CR3","doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R., Doll\u00b4ar, P., Tu, Z., He, K.: Aggregated residual transformations for deep neural networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1492\u20131500 (2017)","DOI":"10.1109\/CVPR.2017.634"},{"key":"16_CR4","doi-asserted-by":"crossref","unstructured":"Jeong, J.E., Choi, Y.S.: Depth-enhanced gaze following method. In: Proceedings of the 36th Annual ACM Symposium on Applied Computing, pp. 1090\u20131093 (2021)","DOI":"10.1145\/3412841.3442107"},{"key":"16_CR5","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., Sun, G.: Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7132\u20137141 (2018)","DOI":"10.1109\/CVPR.2018.00745"},{"key":"16_CR6","doi-asserted-by":"crossref","unstructured":"Krafka, K., et al.: Eye tracking for everyone. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2176\u20132184 (2016)","DOI":"10.1109\/CVPR.2016.239"},{"key":"16_CR7","doi-asserted-by":"crossref","unstructured":"Judd, T., Ehinger, K., Durand, F., Torralba, A.: Learning to predict where humans look. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 2106\u20132113. IEEE (2009)","DOI":"10.1109\/ICCV.2009.5459462"},{"key":"16_CR8","doi-asserted-by":"crossref","unstructured":"Marin-Jimenez, M.J., Zisserman, A., Eichner, M., Ferrari, V.: Detecting people looking at each other in videos. Int. J. Comput. Vis. 106(3), 282\u2013296 (2014)","DOI":"10.1007\/s11263-013-0655-7"},{"key":"16_CR9","doi-asserted-by":"crossref","unstructured":"Jin, T., Lin, Z., Zhu, S., Wang, W., Hu, S.: Multiperson gaze-following with numerical coordinate regression. In: 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021), pp. 01\u201308. IEEE (2021)","DOI":"10.1109\/FG52635.2021.9666980"},{"issue":"3","key":"16_CR10","doi-asserted-by":"publisher","first-page":"1390","DOI":"10.1109\/TCSVT.2021.3071621","volume":"32","author":"W Chen","year":"2021","unstructured":"Chen, W., et al.: Gaze estimation via the joint modeling of multiple cues. IEEE Trans. Circuits Syst. Video Technol. 32(3), 1390\u20131402 (2021)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"16_CR11","doi-asserted-by":"crossref","unstructured":"Park, H.S., Shi, J.: Social saliency prediction. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4777\u20134785 (2015)","DOI":"10.1109\/CVPR.2015.7299110"},{"key":"16_CR12","doi-asserted-by":"crossref","unstructured":"Leifman, G., Rudoy, D., Swedish, T., Bayro-Corrochano, E., Raskar, R.: Learning gaze transitions from depth to improve video saliency estimation. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1698\u20131707 (2017)","DOI":"10.1109\/ICCV.2017.188"},{"key":"16_CR13","doi-asserted-by":"crossref","unstructured":"Pfister, T., Charles, J., Zisserman, A.: Flowing convnets for human pose estimation in videos. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1913\u20131921 (2015)","DOI":"10.1109\/ICCV.2015.222"},{"key":"16_CR14","doi-asserted-by":"crossref","unstructured":"Pathirana, P., Senarath, S., Meedeniya, D., Jayarathna, S.: Single-user 2d gaze estimation in retail environment using deep learning. In: 2022 2nd International Conference on Advanced Research in Computing (ICARC), pp. 206\u2013211. IEEE (2022)","DOI":"10.1109\/ICARC54489.2022.9754167"},{"key":"16_CR15","doi-asserted-by":"crossref","unstructured":"Peng, Z., et al.: Conformer: local features coupling global representations for visual recognition. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 367\u2013376 (2021)","DOI":"10.1109\/ICCV48922.2021.00042"},{"key":"16_CR16","doi-asserted-by":"crossref","unstructured":"Lin, T.-Y., Doll\u00b4ar, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117\u20132125 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"16_CR17","doi-asserted-by":"crossref","unstructured":"MohebAli, R., Toroghi, R.M., Zareian, H.: Human action recognition using attention mechanism and gaze information. In: Mediterranean Conference on Pattern Recognition and Artificial Intelligence, pp. 3\u201317. Springer (2022)","DOI":"10.1007\/978-3-031-04112-9_1"},{"key":"16_CR18","unstructured":"K\u00fcmmerer, M., Theis, L., Bethge, M.: Deep gaze i: boosting saliency prediction with feature maps trained on imagenet. arXiv preprint arXiv:1411.1045 (2014)"},{"key":"16_CR19","doi-asserted-by":"crossref","unstructured":"Recasens, A., Vondrick, C., Khosla, A., Torralba, A.: Following gaze in video. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 1435\u20131443 (2017)","DOI":"10.1109\/ICCV.2017.160"},{"issue":"2","key":"16_CR20","doi-asserted-by":"publisher","first-page":"302","DOI":"10.1109\/THMS.2022.3145097","volume":"52","author":"L Dai","year":"2022","unstructured":"Dai, L., Liu, J., Zhaojie, J.: Binocular feature fusion and spatial attention mechanism based gaze tracking. IEEE Trans. Hum.-Mach. Syst. 52(2), 302\u2013311 (2022)","journal-title":"IEEE Trans. Hum.-Mach. Syst."},{"key":"16_CR21","doi-asserted-by":"crossref","unstructured":"Tu, D., Min, X., Duan, H., Guo, G., Zhai, G., Shen, W.: End-to-end human-gaze-target detection with transformers. arXiv preprint arXiv:2203.10433 (2022)","DOI":"10.1109\/CVPR52688.2022.00224"}],"container-title":["Lecture Notes in Computer Science","Biometric Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-20233-9_16","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,11,3]],"date-time":"2022-11-03T00:26:41Z","timestamp":1667435201000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-20233-9_16"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031202322","9783031202339"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-20233-9_16","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"3 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"CCBR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Biometric Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Beijing","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccbr2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ccbr99.cn\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"115","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"70","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"61% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}