{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,13]],"date-time":"2026-04-13T18:43:15Z","timestamp":1776105795066,"version":"3.50.1"},"publisher-location":"Cham","reference-count":73,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783030584511","type":"print"},{"value":"9783030584528","type":"electronic"}],"license":[{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2020,1,1]],"date-time":"2020-01-01T00:00:00Z","timestamp":1577836800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-58452-8_25","type":"book-chapter","created":{"date-parts":[[2020,11,3]],"date-time":"2020-11-03T00:34:03Z","timestamp":1604363643000},"page":"422-440","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":186,"title":["ReferIt3D: Neural Listeners for Fine-Grained 3D Object Identification in Real-World Scenes"],"prefix":"10.1007","author":[{"given":"Panos","family":"Achlioptas","sequence":"first","affiliation":[]},{"given":"Ahmed","family":"Abdelreheem","sequence":"additional","affiliation":[]},{"given":"Fei","family":"Xia","sequence":"additional","affiliation":[]},{"given":"Mohamed","family":"Elhoseiny","sequence":"additional","affiliation":[]},{"given":"Leonidas","family":"Guibas","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2020,11,3]]},"reference":[{"key":"25_CR1","unstructured":"Abdelkarim, S., Achlioptas, P., Huang, J., Li, B., Church, K., Elhoseiny, M.: Long-tail visual relationship recognition with a visiolinguistic hubless loss. CoRR abs\/2004.00436 (2020)"},{"key":"25_CR2","doi-asserted-by":"crossref","unstructured":"Achlioptas, P., Abdelreheem, A., Xia, F., Elhoseiny, M., Guibas, L.: Supplementary material for: ReferIt3D: neural listeners for fine-grained 3D object identification in real world 3D scenes (2020)","DOI":"10.1007\/978-3-030-58452-8_25"},{"key":"25_CR3","unstructured":"Achlioptas, P., Diamanti, O., Mitliagkas, I., Guibas, L.: Learning representations and generative models for 3D point clouds. In: International Conference on Machine Learning (ICML) (2018)"},{"key":"25_CR4","doi-asserted-by":"crossref","unstructured":"Achlioptas, P., Fan, J., Hawkins, R.X., Goodman, N.D., Guibas, L.J.: ShapeGlot: learning language for shape differentiation. In: International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00903"},{"key":"25_CR5","doi-asserted-by":"crossref","unstructured":"Agrawal, A., Batra, D., Parikh, D.: Analyzing the behavior of visual question answering models. In: Empirical Methods in Natural Language Processing (EMNLP) (2016)","DOI":"10.18653\/v1\/D16-1203"},{"key":"25_CR6","doi-asserted-by":"crossref","unstructured":"Agrawal, H., et al.: nocaps: novel object captioning at scale. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/ICCV.2019.00904"},{"key":"25_CR7","doi-asserted-by":"crossref","unstructured":"Akata, Z., Reed, S., Walter, D., Lee, H., Schiele, B.: Evaluation of output embeddings for fine-grained image classification. In: CVPR (2015)","DOI":"10.1109\/CVPR.2015.7298911"},{"key":"25_CR8","unstructured":"Anayurt, H., Ozyegin, S.A., Cetin, U., Aktas, U., Kalkan, S.: Searching for ambiguous objects in videos using relational referring expressions. CoRR abs\/1908.01189 (2019)"},{"key":"25_CR9","doi-asserted-by":"crossref","unstructured":"Anderson, P., et al.: Vision-and-language navigation: interpreting visually-grounded navigation instructions in real environments. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00387"},{"key":"25_CR10","doi-asserted-by":"crossref","unstructured":"Andreas, J., Klein, D.: Reasoning about pragmatics with neural listeners and speakers. CoRR abs\/1604.00562 (2016)","DOI":"10.18653\/v1\/D16-1125"},{"key":"25_CR11","doi-asserted-by":"crossref","unstructured":"Andreas, J., Rohrbach, M., Darrell, T., Klein, D.: Neural module networks. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.12"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Antol, S., et al.: VQA: visual question answering. In: International Conference on Computer Vision (ICCV) (2015)","DOI":"10.1109\/ICCV.2015.279"},{"key":"25_CR13","doi-asserted-by":"crossref","unstructured":"Armeni, I., et al.: 3D semantic parsing of large-scale indoor spaces. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2016)","DOI":"10.1109\/CVPR.2016.170"},{"key":"25_CR14","doi-asserted-by":"crossref","unstructured":"Avetisyan, A., Dahnert, M., Dai, A., Savva, M., Chang, A.X., Nie\u00dfner, M.: Scan2CAD: learning cad model alignment in RGB-D scans. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00272"},{"key":"25_CR15","doi-asserted-by":"crossref","unstructured":"Kong, C., Lin, D., Bansal, M., Urtasun, R., Fidler, S.: What are you talking about? Text-to-image coreference. In: CVPR (2014)","DOI":"10.1109\/CVPR.2014.455"},{"key":"25_CR16","doi-asserted-by":"crossref","unstructured":"Chen, K., Choy, C.B., Savva, M., Chang, A.X., Funkhouser, T., Savarese, S.: Text2Shape: generating shapes from natural language by learning joint embeddings. CoRR abs\/1803.08495 (2018)","DOI":"10.1007\/978-3-030-20893-6_7"},{"key":"25_CR17","unstructured":"Chen, Z.D., Chang, A.X., Nie\u00dfner, M.: https:\/\/github.com\/daveredrum\/ScanRefer. Accessed 17 July 2020"},{"key":"25_CR18","doi-asserted-by":"crossref","unstructured":"Chen, Z.D., Chang, A.X., Nie\u00dfner, M.: ScanRefer: 3D object localization in RGB-D scans using natural language. CoRR abs\/1912.08830 (2019)","DOI":"10.1007\/978-3-030-58565-5_13"},{"key":"25_CR19","doi-asserted-by":"crossref","unstructured":"Cohn-Gordon, R., Goodman, N., Potts, C.: Pragmatically informative image captioning with character-level inference. CoRR abs\/1804.05417 (2018)","DOI":"10.18653\/v1\/N18-2070"},{"key":"25_CR20","doi-asserted-by":"crossref","unstructured":"Dai, A., Chang, A.X., Savva, M., Halber, M., Funkhouser, T., Nie\u00dfner, M.: ScanNet: richly-annotated 3D reconstructions of indoor scenes. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.261"},{"key":"25_CR21","doi-asserted-by":"crossref","unstructured":"Das, A., Datta, S., Gkioxari, G., Lee, S., Parikh, D., Batra, D.: Embodied question answering. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00008"},{"key":"25_CR22","unstructured":"Devlin, J., Gupta, S., Girshick, R., Mitchell, M., Zitnick, C.L.: Exploring nearest neighbor approaches for image captioning. arXiv preprint arXiv:1505.04467 (2015)"},{"key":"25_CR23","doi-asserted-by":"crossref","unstructured":"Do\u011fan, F.I., Kalkan, S., Leite, I.: Learning to generate unambiguous spatial referring expressions for real-world environments. CoRR (2019)","DOI":"10.1109\/IROS40897.2019.8968510"},{"key":"25_CR24","doi-asserted-by":"crossref","unstructured":"Elhoseiny, M., Elfeki, M.: Creativity inspired zero-shot learning. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 5784\u20135793 (2019)","DOI":"10.1109\/ICCV.2019.00588"},{"key":"25_CR25","doi-asserted-by":"crossref","unstructured":"Elhoseiny, M., Saleh, B., Elgammal, A.: Write a classifier: zero-shot learning using purely textual descriptions. In: ICCV (2013)","DOI":"10.1109\/ICCV.2013.321"},{"key":"25_CR26","doi-asserted-by":"crossref","unstructured":"Elhoseiny, M., Zhu, Y., Zhang, H., Elgammal, A.: Link the head to the \u201cbeak\u201d: zero shot learning from noisy text description at part precision. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.666"},{"key":"25_CR27","doi-asserted-by":"crossref","unstructured":"Goyal, Y., Khot, T., Summers-Stay, D., Batra, D., Parikh, D.: Making the V in VQA matter: elevating the role of image understanding in visual question answering. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2017)","DOI":"10.1109\/CVPR.2017.670"},{"key":"25_CR28","unstructured":"Karpathy, A., Joulin, A., Li, F.F.F.: Deep fragment embeddings for bidirectional image sentence mapping. In: Advances in Neural Information Processing Systems (NeurIPS) (2014)"},{"key":"25_CR29","doi-asserted-by":"crossref","unstructured":"Kazemzadeh, S., Ordonez, V., Matten, M., Berg, T.: ReferitGame: referring to objects in photographs of natural scenes. In: Empirical Methods in Natural Language Processing (EMNLP) (2014)","DOI":"10.3115\/v1\/D14-1086"},{"key":"25_CR30","unstructured":"Kim, J.H., Jun, J., Zhang, B.T.: Bilinear attention networks. In: Advances in Neural Information Processing Systems (NeurIPS) (2018)"},{"key":"25_CR31","unstructured":"Kiros, R., Salakhutdinov, R., Zemel, R.S., et al.: Unifying visual-semantic embeddings with multimodal neural language models. Trans. Assoc. Comput. Linguist. (TACL) (2015)"},{"issue":"1","key":"25_CR32","doi-asserted-by":"publisher","first-page":"32","DOI":"10.1007\/S11263-016-0981-7","volume":"123","author":"R Krishna","year":"2017","unstructured":"Krishna, R., et al.: Visual genome: connecting language and vision using crowdsourced dense image annotations. Int. J. Comput. Vis. 123(1), 32\u201373 (2017). https:\/\/doi.org\/10.1007\/S11263-016-0981-7","journal-title":"Int. J. Comput. Vis."},{"key":"25_CR33","doi-asserted-by":"crossref","unstructured":"Kulkarni, N., Misra, I., Tulsiani, S., Gupta, A.: 3D-RelNet: joint object and relational network for 3D prediction. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/ICCV.2019.00230"},{"key":"25_CR34","unstructured":"Lazaridou, A., Hermann, K.M., Tuyls, K., Clark, S.: Emergence of linguistic communication from referential games with symbolic and pixel input. arXiv preprint arXiv:1804.03984 (2018)"},{"key":"25_CR35","doi-asserted-by":"crossref","unstructured":"Ba, J.L., Swersky, K., Fidler, S., et al.: Predicting deep zero-shot convolutional neural networks using textual descriptions. In: ICCV (2015)","DOI":"10.1109\/ICCV.2015.483"},{"key":"25_CR36","volume-title":"Convention: A Philosophical Study","author":"D Lewis","year":"2008","unstructured":"Lewis, D.: Convention: A Philosophical Study. Wiley, Hoboken (2008)"},{"key":"25_CR37","unstructured":"Li, C., Xia, F., Mart\u00edn-Mart\u00edn, R., Savarese, S.: HRL4IN: hierarchical reinforcement learning for interactive navigation with mobile manipulators. In: Conference on Robot Learning (2020)"},{"key":"25_CR38","doi-asserted-by":"crossref","unstructured":"Long, Y., Shao, L.: Describing unseen classes by exemplars: zero-shot learning using grouped simile ensemble. In: Winter Conference on Applications of Computer Vision (WACV) (2017)","DOI":"10.1109\/WACV.2017.106"},{"key":"25_CR39","doi-asserted-by":"crossref","unstructured":"Long, Y., Shao, L.: Learning to recognise unseen classes by a few similes. In: Proceedings of the 25th ACM International Conference on Multimedia, pp. 636\u2013644. ACM (2017)","DOI":"10.1145\/3123266.3123323"},{"key":"25_CR40","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"852","DOI":"10.1007\/978-3-319-46448-0_51","volume-title":"Computer Vision \u2013 ECCV 2016","author":"C Lu","year":"2016","unstructured":"Lu, C., Krishna, R., Bernstein, M., Fei-Fei, L.: Visual relationship detection with language priors. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 852\u2013869. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_51"},{"key":"25_CR41","unstructured":"Mao, J., Xu, W., Yang, Y., Wang, J., Yuille, A.: Deep captioning with multimodal recurrent neural networks (m-RNN). In: International Conference on Learning Representations (ICLR) (2015)"},{"key":"25_CR42","doi-asserted-by":"crossref","unstructured":"Mauceri, C., Palmer, M., Heckman, C.: SUN-Spot: an RGB-D dataset with spatial referring expressions. In: International Conference on Computer Vision Workshop on Closing the Loop Between Vision and Language (2019)","DOI":"10.1109\/ICCVW.2019.00236"},{"key":"25_CR43","unstructured":"Mitchell, M., van Deemter, K., Reiter, E.: Generating expressions that refer to visible objects. In: North American Chapter of the Association for Computational Linguistics (NAACL) (2013)"},{"key":"25_CR44","doi-asserted-by":"crossref","unstructured":"Mo, K., et al.: A large-scale benchmark for fine-grained and hierarchical part-level 3D object understanding. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00100"},{"key":"25_CR45","doi-asserted-by":"crossref","unstructured":"Monroe, W., Hawkins, R.X., Goodman, N.D., Potts, C.: Colors in context: a pragmatic neural model for grounded language understanding. Trans. Assoc. Comput. Linguist. (TACL) (2017)","DOI":"10.1162\/tacl_a_00064"},{"key":"25_CR46","unstructured":"Paetzel, M., Racca, D.N., DeVault, D.: A multimodal corpus of rapid dialogue games. In: LREC, pp. 4189\u20134195 (2014)"},{"key":"25_CR47","unstructured":"Plummer, B.A., et al.: Revisiting image-language networks for open-ended phrase detection. CoRR abs\/1811.07212 (2018)"},{"key":"25_CR48","doi-asserted-by":"crossref","unstructured":"Plummer, B.A., Wang, L., Cervantes, C.M., Caicedo, J.C., Hockenmaier, J., Lazebnik, S.: Flickr30k entities: collecting region-to-phrase correspondences for richer image-to-sentence models. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2015)","DOI":"10.1109\/ICCV.2015.303"},{"key":"25_CR49","doi-asserted-by":"crossref","unstructured":"Prabhudesai, M., Tung, H.Y.F., Javed, S.A., Sieb, M., Harley, A.W., Fragkiadaki, K.: Embodied language grounding with implicit 3D visual feature representations. CoRR abs\/1910.01210 (2019)","DOI":"10.1109\/CVPR42600.2020.00229"},{"key":"25_CR50","doi-asserted-by":"crossref","unstructured":"Qi, C.R., Litany, O., He, K., Guibas, L.: Deep hough voting for 3D object detection in point clouds. In: International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/CVPR42600.2020.00446"},{"key":"25_CR51","unstructured":"Qi, C.R., Yi, L., Su, H., Guibas, L.J.: PointNet++: deep hierarchical feature learning on point sets in a metric space. In: Advances in Neural Information Processing Systems (NeurIPS) (2017)"},{"key":"25_CR52","doi-asserted-by":"crossref","unstructured":"Ramanathan, V., Joulin, A., Liang, P., Fei-Fei, L.: Linking people with \u201ctheir\u201d names using coreference resolution. In: European Conference on Computer Vision (ECCV) (2014)","DOI":"10.1007\/978-3-319-10590-1_7"},{"key":"25_CR53","unstructured":"Ren, M., Kiros, R., Zemel, R.: Exploring models and data for image question answering. In: Advances in Neural Information Processing Systems (NeurIPS) (2015)"},{"key":"25_CR54","unstructured":"Romera-Paredes, B., Torr, P.: An embarrassingly simple approach to zero-shot learning. In: ICML, pp. 2152\u20132161 (2015)"},{"issue":"11","key":"25_CR55","doi-asserted-by":"publisher","first-page":"1328","DOI":"10.1177\/0278364911408155","volume":"30","author":"B Rosman","year":"2011","unstructured":"Rosman, B., Ramamoorthy, S.: Learning spatial relationships between objects. Int. J. Rob. Res. 30(11), 1328\u20131342 (2011)","journal-title":"Int. J. Rob. Res."},{"key":"25_CR56","doi-asserted-by":"crossref","unstructured":"Savva, M., et al.: Habitat: a platform for embodied AI research. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/ICCV.2019.00943"},{"key":"25_CR57","doi-asserted-by":"crossref","unstructured":"Shridhar, M., et al.: ALFRED: a benchmark for interpreting grounded instructions for everyday tasks. CoRR abs\/1912.01734 (2019)","DOI":"10.1109\/CVPR42600.2020.01075"},{"key":"25_CR58","unstructured":"Socher, R., Ganjoo, M., Manning, C.D., Ng, A.: Zero-shot learning through cross-modal transfer. In: NeurIPS, pp. 935\u2013943 (2013)"},{"key":"25_CR59","doi-asserted-by":"crossref","unstructured":"Su, J.C., Wu, C., Jiang, H., Maji, S.: Reasoning about fine-grained attribute phrases using reference games. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 418\u2013427 (2017)","DOI":"10.1109\/ICCV.2017.53"},{"key":"25_CR60","doi-asserted-by":"crossref","unstructured":"Tsai, Y.H.H., Huang, L.K., Salakhutdinov, R.: Learning robust visual-semantic embeddings. In: ICCV (2017)","DOI":"10.1109\/ICCV.2017.386"},{"key":"25_CR61","doi-asserted-by":"crossref","unstructured":"Vedantam, R., Bengio, S., Murphy, K., Parikh, D., Chechik, G.: Context-aware captions from context-agnostic supervision. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 251\u2013260 (2017)","DOI":"10.1109\/CVPR.2017.120"},{"key":"25_CR62","doi-asserted-by":"crossref","unstructured":"Vinyals, O., Toshev, A., Bengio, S., Erhan, D.: Show and tell: a neural image caption generator. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2015)","DOI":"10.1109\/CVPR.2015.7298935"},{"issue":"5","key":"25_CR63","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3326362","volume":"38","author":"Y Wang","year":"2019","unstructured":"Wang, Y., Sun, Y., Liu, Z., Sarma, S.E., Bronstein, M.M., Solomon, J.M.: Dynamic graph CNN for learning on point clouds. ACM Trans. Graph. (TOG) 38(5), 1\u201312 (2019)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"25_CR64","unstructured":"Wittgenstein., L.: Philosophical Investigations: the English text of the third edition. Wiley, Hoboken (1953)"},{"key":"25_CR65","doi-asserted-by":"crossref","unstructured":"Xia, F., Zamir, A.R., He, Z., Sax, A., Malik, J., Savarese, S.: Gibson Env: real-world perception for embodied agents. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2018)","DOI":"10.1109\/CVPR.2018.00945"},{"key":"25_CR66","doi-asserted-by":"crossref","unstructured":"Xiang, F., et al.: SAPIEN: a simulated part-based interactive environment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11097\u201311107 (2020)","DOI":"10.1109\/CVPR42600.2020.01111"},{"key":"25_CR67","unstructured":"Xu, K., et al.: Show, attend and tell: neural image caption generation with visual attention. In: International Conference on Machine Learning (ICML) (2015)"},{"key":"25_CR68","unstructured":"Yang, J., et al.: Embodied visual recognition. CoRR abs\/1904.04404 (2019)"},{"key":"25_CR69","unstructured":"Yang, Y., Hospedales, T.M.: A unified perspective on multi-domain and multi-task learning. In: ICLR (2015)"},{"key":"25_CR70","doi-asserted-by":"crossref","unstructured":"Yu, Z., Yu, J., Cui, Y., Tao, D., Tian, Q.: Deep modular co-attention networks for visual question answering. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00644"},{"key":"25_CR71","doi-asserted-by":"crossref","unstructured":"Zhang, J., Kalantidis, Y., Rohrbach, M., Paluri, M., Elgammal, A., Elhoseiny, M.: Large-scale visual relationship understanding. In: AAAI Conference on Artificial Intelligence (2019)","DOI":"10.1609\/aaai.v33i01.33019185"},{"key":"25_CR72","doi-asserted-by":"crossref","unstructured":"Zhou, L., Kalantidis, Y., Chen, X., Corso, J.J., Rohrbach, M.: Grounded video description. In: Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","DOI":"10.1109\/CVPR.2019.00674"},{"key":"25_CR73","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Elhoseiny, M., Liu, B., Peng, X., Elgammal, A.: A generative adversarial approach for zero-shot learning from noisy texts. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00111"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2020"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-58452-8_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,3]],"date-time":"2024-11-03T00:13:10Z","timestamp":1730592790000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-58452-8_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2020]]},"ISBN":["9783030584511","9783030584528"],"references-count":73,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-58452-8_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2020]]},"assertion":[{"value":"3 November 2020","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Glasgow","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 August 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2020.eu\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"OpenReview","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5025","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1360","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"27% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"7","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"The conference was held virtually due to the COVID-19 pandemic.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}