{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T14:39:31Z","timestamp":1743086371135,"version":"3.40.3"},"publisher-location":"Cham","reference-count":21,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030317225"},{"type":"electronic","value":"9783030317232"}],"license":[{"start":{"date-parts":[[2019,1,1]],"date-time":"2019-01-01T00:00:00Z","timestamp":1546300800000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2019]]},"DOI":"10.1007\/978-3-030-31723-2_55","type":"book-chapter","created":{"date-parts":[[2019,10,31]],"date-time":"2019-10-31T00:05:31Z","timestamp":1572480331000},"page":"644-656","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["Stage-by-Stage Based Design Paradigm of Two-Pathway Model for Gaze Following"],"prefix":"10.1007","author":[{"given":"Zhongping","family":"Cao","sequence":"first","affiliation":[]},{"given":"Guoli","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xuemei","family":"Guo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,10,31]]},"reference":[{"key":"55_CR1","unstructured":"Recasens, A., Khosla, A., Vondrick, C., Torralba, A.: Where are they looking? In: Advances in Neural Information Processing Systems (NIPS) (2015)"},{"key":"55_CR2","doi-asserted-by":"crossref","unstructured":"Jiang, M., Huang, S., Duan, J., et al.: Salicon: saliency in context. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1072\u20131080 (2015)","DOI":"10.1109\/CVPR.2015.7298710"},{"key":"55_CR3","doi-asserted-by":"crossref","unstructured":"Krafka, K., Khosla, A., Kellnhofer, P., et al.: Eye tracking for everyone. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2176\u20132184 (2016)","DOI":"10.1109\/CVPR.2016.239"},{"key":"55_CR4","doi-asserted-by":"crossref","unstructured":"Recasens, A., Vondrick, C., Khosla, A., Torralba, A.: Following gaze in video. In: IEEE International Conference on Computer Vision (2017)","DOI":"10.1109\/ICCV.2017.160"},{"key":"55_CR5","doi-asserted-by":"crossref","unstructured":"Chong, E., Ruiz, N., et al.: Connecting gaze, scene, and attention: generalized attention estimation via joint modeling of gaze and scene saliency. In: The European Conference on Computer Vision (2018)","DOI":"10.1007\/978-3-030-01228-1_24"},{"key":"55_CR6","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, et al.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"issue":"11","key":"55_CR7","doi-asserted-by":"publisher","first-page":"1254","DOI":"10.1109\/34.730558","volume":"20","author":"L Itti","year":"1998","unstructured":"Itti, L., Koch, C., Niebur, E.: A model of saliency-based visual attention for rapid scene analysis. IEEE Trans. Pattern Anal. Mach. Intell. 20(11), 1254\u20131259 (1998)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"55_CR8","doi-asserted-by":"crossref","unstructured":"Deng, J., Dong, W., Socher, R., et al.: ImageNet: a large-scale hierarchical image database. In: IEEE Conference on Computer Vision and Pattern Recognition, pp. 248\u2013255 (2009)","DOI":"10.1109\/CVPR.2009.5206848"},{"key":"55_CR9","unstructured":"Matthias, K., Lucas, T., Matthias, B.: Deep gaze I: boosting saliency prediction with feature maps trained on imagenet. CoRR, vol.abs\/1411.1045 (2014)"},{"issue":"9","key":"55_CR10","doi-asserted-by":"publisher","first-page":"4446","DOI":"10.1109\/TIP.2017.2710620","volume":"26","author":"SS Kruthiventi","year":"2017","unstructured":"Kruthiventi, S.S., Ayush, K., et al.: DeepFix: a fully convolutional neural network for predicting human eye fixations. IEEE Trans. Image Process. 26(9), 4446\u20134455 (2017)","journal-title":"IEEE Trans. Image Process."},{"issue":"2","key":"55_CR11","first-page":"1097","volume":"25","author":"A Krizhevsky","year":"2012","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.: ImageNet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 25(2), 1097\u20131105 (2012)","journal-title":"Adv. Neural Inf. Process. Syst."},{"key":"55_CR12","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. CoRR, vol.abs\/1409.1556 (2014)"},{"key":"55_CR13","unstructured":"Sun, X., Xiao, B., Liang, S., et al.: Integral human pose regression. CoRR, vol. abs\/1711.08229 (2017)"},{"key":"55_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"TY Lin","year":"2014","unstructured":"Lin, T.Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"issue":"2","key":"55_CR15","doi-asserted-by":"publisher","first-page":"303","DOI":"10.1007\/s11263-009-0275-4","volume":"88","author":"M Everingham","year":"2010","unstructured":"Everingham, M., Van Gool, L., Williams, C.K., Winn, J., Zisserman, A.: The pascal visual object classes (VOC) challenge. Int. J. Comput. Vis. 88(2), 303\u2013338 (2010)","journal-title":"Int. J. Comput. Vis."},{"key":"55_CR16","unstructured":"Pytorch: tensors and dynamic neural networks in python with strong GPU acceleration. https:\/\/github.com\/pytorch\/pytorch. Accessed 03 Nov 2017"},{"key":"55_CR17","doi-asserted-by":"crossref","unstructured":"Judd, T., Ehinger, K., Durand, F., et al.: Learning to predict where humans look. In: Proceedings of the 2009 IEEE International Conference on Computer Vision (2009)","DOI":"10.1109\/ICCV.2009.5459462"},{"issue":"3","key":"55_CR18","doi-asserted-by":"publisher","first-page":"254","DOI":"10.1109\/TIV.2018.2843120","volume":"3","author":"V Sourabh","year":"2018","unstructured":"Sourabh, V., Akshay, R., Trivedi, M.M.: Gaze zone estimation using convolutional neural networks: a general framework and ablative analysis. IEEE Trans. Intell. Veh. 3(3), 254\u2013265 (2018)","journal-title":"IEEE Trans. Intell. Veh."},{"issue":"3","key":"55_CR19","doi-asserted-by":"publisher","first-page":"569","DOI":"10.1109\/TPAMI.2014.2345401","volume":"37","author":"MM Cheng","year":"2018","unstructured":"Cheng, M.M., Mitra, N.J., Huang, X., et al.: Global contrast based salient region detection. IEEE Trans. Pattern Anal. Mach. Intell. 37(3), 569\u2013582 (2018)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"55_CR20","doi-asserted-by":"crossref","unstructured":"Saran, A., Majumdar, S., Shor, E.S., et al.: Human gaze following for human-robot interaction. In: IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 8615\u20138621 (2018)","DOI":"10.1109\/IROS.2018.8593580"},{"key":"55_CR21","doi-asserted-by":"crossref","unstructured":"Zhao, J.X., Cao, Y., Cheng, M.M., et al.: Contrast prior and fluid pyramid integration for RGBD salient object detection. In: Proceedings of the IEEE International Conference on Computer Vision (2019)","DOI":"10.1109\/CVPR.2019.00405"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition and Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-31723-2_55","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,31]],"date-time":"2024-10-31T00:28:11Z","timestamp":1730334491000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-31723-2_55"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019]]},"ISBN":["9783030317225","9783030317232"],"references-count":21,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-31723-2_55","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2019]]},"assertion":[{"value":"31 October 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"PRCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Chinese Conference on Pattern Recognition and Computer Vision  (PRCV)","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Xi'an","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2019","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 November 2019","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"11 November 2019","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ccprcv2019","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.prcv2019.com\/en\/index.html","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"412","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"165","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"40% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}