{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,1]],"date-time":"2026-04-01T11:18:16Z","timestamp":1775042296537,"version":"3.50.1"},"publisher-location":"Cham","reference-count":60,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031167874","type":"print"},{"value":"9783031167881","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-16788-1_25","type":"book-chapter","created":{"date-parts":[[2022,9,22]],"date-time":"2022-09-22T20:35:56Z","timestamp":1663878956000},"page":"409-425","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":12,"title":["Diverse Video Captioning by\u00a0Adaptive Spatio-temporal Attention"],"prefix":"10.1007","author":[{"given":"Zohreh","family":"Ghaderi","sequence":"first","affiliation":[]},{"given":"Leonard","family":"Salewski","sequence":"additional","affiliation":[]},{"given":"Hendrik P. A.","family":"Lensch","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,20]]},"reference":[{"key":"25_CR1","doi-asserted-by":"crossref","unstructured":"Aafaq, N., Akhtar, N., Liu, W., Gilani, S.Z., Mian, A.: Spatio-temporal dynamics and semantic attribute enriched visual encoding for video captioning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12487\u201312496 (2019)","DOI":"10.1109\/CVPR.2019.01277"},{"key":"25_CR2","doi-asserted-by":"publisher","unstructured":"Aafaq, N., Mian, A., Liu, W., Gilani, S.Z., Shah, M.: Video description: a survey of methods, datasets, and evaluation metrics. ACM Comput. Surv. 52(6) (2019). https:\/\/doi.org\/10.1145\/3355390","DOI":"10.1145\/3355390"},{"key":"25_CR3","doi-asserted-by":"crossref","unstructured":"Arnab, A., Dehghani, M., Heigold, G., Sun, C., Lu\u010di\u0107, M., Schmid, C.: Vivit: A video vision transformer. arXiv preprint arXiv:2103.15691 (2021)","DOI":"10.1109\/ICCV48922.2021.00676"},{"key":"25_CR4","unstructured":"Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473 (2014)"},{"key":"25_CR5","unstructured":"Banerjee, S., Lavie, A.: METEOR: an automatic metric for MT evaluation with improved correlation with human judgments. In: Proceedings of the ACL Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and\/or Summarization, pp. 65\u201372 (2005)"},{"key":"25_CR6","unstructured":"Black, S., Gao, L., Wang, P., Leahy, C., Biderman, S.: GPT-Neo: large scale autoregressive language modeling with mesh-tensorflow. If you use this Software, Please Cite it using these Metadata 58 (2021)"},{"key":"25_CR7","unstructured":"Chen, D., Dolan, W.B.: Collecting highly parallel data for paraphrase evaluation. In: Proceedings of the 49th Annual Meeting of the Association for Computational Linguistics: Human Language Technologies, pp. 190\u2013200 (2011)"},{"key":"25_CR8","doi-asserted-by":"publisher","DOI":"10.3389\/frobt.2020.475767","volume":"7","author":"H Chen","year":"2020","unstructured":"Chen, H., Lin, K., Maye, A., Li, J., Hu, X.: A semantics-assisted video captioning model trained with scheduled sampling. Front. Robot. AI 7, 475767 (2020)","journal-title":"Front. Robot. AI"},{"key":"25_CR9","unstructured":"Chen, M., Li, Y., Zhang, Z., Huang, S.: TVT: two-view transformer network for video captioning. In: Asian Conference on Machine Learning, pp. 847\u2013862. PMLR (2018)"},{"key":"25_CR10","doi-asserted-by":"crossref","unstructured":"Chen, S., Jiang, Y.G.: Motion guided region message passing for video captioning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1543\u20131552 (2021)","DOI":"10.1109\/ICCV48922.2021.00157"},{"key":"25_CR11","unstructured":"Chen, X., et al.: Microsoft coco captions: data collection and evaluation server. arXiv preprint arXiv:1504.00325 (2015)"},{"key":"25_CR12","doi-asserted-by":"crossref","unstructured":"Chen, Y., Wang, S., Zhang, W., Huang, Q.: Less is more: picking informative frames for video captioning. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 358\u2013373 (2018)","DOI":"10.1007\/978-3-030-01261-8_22"},{"key":"25_CR13","doi-asserted-by":"crossref","unstructured":"Cherian, A., Wang, J., Hori, C., Marks, T.: Spatio-temporal ranked-attention networks for video captioning. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1617\u20131626 (2020)","DOI":"10.1109\/WACV45572.2020.9093291"},{"key":"25_CR14","first-page":"658","volume":"31","author":"B Dai","year":"2018","unstructured":"Dai, B., Fidler, S., Lin, D.: A neural compositional paradigm for image captioning. NIPS 31, 658\u2013668 (2018)","journal-title":"NIPS"},{"key":"25_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"428","DOI":"10.1007\/11744047_33","volume-title":"Computer Vision \u2013 ECCV 2006","author":"N Dalal","year":"2006","unstructured":"Dalal, N., Triggs, B., Schmid, C.: Human detection using oriented histograms of flow and appearance. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006. LNCS, vol. 3952, pp. 428\u2013441. Springer, Heidelberg (2006). https:\/\/doi.org\/10.1007\/11744047_33"},{"key":"25_CR16","unstructured":"Devlin, J., Chang, M.W., Lee, K., Toutanova, K.: Bert: pre-training of deep bidirectional transformers for language understanding. arXiv preprint arXiv:1810.04805 (2018)"},{"key":"25_CR17","doi-asserted-by":"crossref","unstructured":"Donahue, J., et al.: Long-term recurrent convolutional networks for visual recognition and description. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2625\u20132634 (2015)","DOI":"10.1109\/CVPR.2015.7298878"},{"key":"25_CR18","doi-asserted-by":"crossref","unstructured":"Gan, Z., et al.: Semantic compositional networks for visual captioning. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5630\u20135639 (2017)","DOI":"10.1109\/CVPR.2017.127"},{"key":"25_CR19","doi-asserted-by":"crossref","unstructured":"Gowda, S.N., Rohrbach, M., Sevilla-Lara, L.: Smart frame selection for action recognition. arXiv preprint arXiv:2012.10671 (2020)","DOI":"10.1609\/aaai.v35i2.16235"},{"key":"25_CR20","doi-asserted-by":"crossref","unstructured":"Guadarrama, S., et al.: YouTube2Text: recognizing and describing arbitrary activities using semantic hierarchies and zero-shot recognition. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2712\u20132719 (2013)","DOI":"10.1109\/ICCV.2013.337"},{"key":"25_CR21","doi-asserted-by":"crossref","unstructured":"Jin, T., Huang, S., Chen, M., Li, Y., Zhang, Z.: SBAT: video captioning with sparse boundary-aware transformer. arXiv preprint arXiv:2007.11888 (2020)","DOI":"10.24963\/ijcai.2020\/88"},{"key":"25_CR22","unstructured":"Kay, W., et al.: The kinetics human action video dataset. arXiv preprint arXiv:1705.06950 (2017)"},{"key":"25_CR23","doi-asserted-by":"crossref","unstructured":"Kayser, M., et al.: e-Vil: a dataset and benchmark for natural language explanations in vision-language tasks. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1244\u20131254 (2021)","DOI":"10.1109\/ICCV48922.2021.00128"},{"key":"25_CR24","doi-asserted-by":"crossref","unstructured":"Khan, M.U.G., Zhang, L., Gotoh, Y.: Human focused video description. In: 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops), pp. 1480\u20131487. IEEE (2011)","DOI":"10.1109\/ICCVW.2011.6130425"},{"issue":"2","key":"25_CR25","doi-asserted-by":"publisher","first-page":"171","DOI":"10.1023\/A:1020346032608","volume":"50","author":"A Kojima","year":"2002","unstructured":"Kojima, A., Tamura, T., Fukunaga, K.: Natural language description of human activities from video images based on concept hierarchy of actions. Int. J. Comput. Vis. 50(2), 171\u2013184 (2002). https:\/\/doi.org\/10.1023\/A:1020346032608","journal-title":"Int. J. Comput. Vis."},{"key":"25_CR26","doi-asserted-by":"crossref","unstructured":"Krishnamoorthy, N., Malkarnenkar, G., Mooney, R., Saenko, K., Guadarrama, S.: Generating natural-language video descriptions using text-mined knowledge. In: Twenty-Seventh AAAI Conference on Artificial Intelligence (2013)","DOI":"10.1609\/aaai.v27i1.8679"},{"key":"25_CR27","doi-asserted-by":"crossref","unstructured":"Lee, M.W., Hakeem, A., Haering, N., Zhu, S.C.: Save: a framework for semantic annotation of visual events. In: 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, pp. 1\u20138. IEEE (2008)","DOI":"10.1109\/CVPRW.2008.4562954"},{"key":"25_CR28","unstructured":"Lin, C.Y.: Rouge: a package for automatic evaluation of summaries. In: Text Summarization Branches Out, pp. 74\u201381 (2004)"},{"key":"25_CR29","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Swin Transformer: hierarchical vision transformer using shifted windows. arXiv preprint arXiv:2103.14030 (2021)","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"25_CR30","unstructured":"Liu, Z., et al.: Video Swin Transformer. arXiv preprint arXiv:2106.13230 (2021)"},{"key":"25_CR31","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"25_CR32","doi-asserted-by":"crossref","unstructured":"Pan, B., et al.: Spatio-temporal graph for video captioning with knowledge distillation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10870\u201310879 (2020)","DOI":"10.1109\/CVPR42600.2020.01088"},{"key":"25_CR33","doi-asserted-by":"crossref","unstructured":"Papineni, K., Roukos, S., Ward, T., Zhu, W.J.: Bleu: a method for automatic evaluation of machine translation. In: Proceedings of the 40th annual meeting of the Association for Computational Linguistics, pp. 311\u2013318 (2002)","DOI":"10.3115\/1073083.1073135"},{"key":"25_CR34","doi-asserted-by":"crossref","unstructured":"Perez-Martin, J., Bustos, B., P\u00e9rez, J.: Improving video captioning with temporal composition of a visual-syntactic embedding. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 3039\u20133049 (2021)","DOI":"10.1109\/WACV48630.2021.00308"},{"key":"25_CR35","unstructured":"Radford, A., Narasimhan, K., Salimans, T., Sutskever, I.: Improving language understanding by generative pre-training (2018)"},{"key":"25_CR36","volume-title":"Speech Understanding Systems: a Summary Of Results of the Five-year Research Effort","author":"DR Reddy","year":"1977","unstructured":"Reddy, D.R., et al.: Speech Understanding Systems: a Summary Of Results of the Five-year Research Effort. Department of Computer Science, CMU, Pittsburgh, PA (1977)"},{"key":"25_CR37","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 28 (2015)"},{"key":"25_CR38","unstructured":"Shekhar, C.C., et al.: Domain-specific semantics guided approach to video captioning. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1587\u20131596 (2020)"},{"key":"25_CR39","unstructured":"Singh, A., Singh, T.D., Bandyopadhyay, S.: NITS-VC system for VATEX video captioning challenge 2020. arXiv preprint arXiv:2006.04058 (2020)"},{"key":"25_CR40","unstructured":"Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. In: Advances in Neural Information Processing Systems, pp. 3104\u20133112 (2014)"},{"key":"25_CR41","unstructured":"Thomason, J., Venugopalan, S., Guadarrama, S., Saenko, K., Mooney, R.: Integrating Language and Vision to Generate Natural Language Descriptions of Videos in the Wild. In: Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, pp. 1218\u20131227 (2014)"},{"key":"25_CR42","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3D convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4489\u20134497 (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"25_CR43","doi-asserted-by":"crossref","unstructured":"Vaidya, J., Subramaniam, A., Mittal, A.: Co-Segmentation aided two-stream architecture for video captioning. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 2774\u20132784 (2022)","DOI":"10.1109\/WACV51458.2022.00250"},{"key":"25_CR44","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"25_CR45","doi-asserted-by":"crossref","unstructured":"Vedantam, R., Lawrence Zitnick, C., Parikh, D.: Cider: consensus-based image description evaluation. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4566\u20134575 (2015)","DOI":"10.1109\/CVPR.2015.7299087"},{"key":"25_CR46","doi-asserted-by":"crossref","unstructured":"Venugopalan, S., Rohrbach, M., Donahue, J., Mooney, R., Darrell, T., Saenko, K.: Sequence to sequence-video to text. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4534\u20134542 (2015)","DOI":"10.1109\/ICCV.2015.515"},{"key":"25_CR47","doi-asserted-by":"crossref","unstructured":"Venugopalan, S., Xu, H., Donahue, J., Rohrbach, M., Mooney, R., Saenko, K.: Translating videos to natural language using deep recurrent neural networks. arXiv preprint arXiv:1412.4729 (2014)","DOI":"10.3115\/v1\/N15-1173"},{"key":"25_CR48","doi-asserted-by":"crossref","unstructured":"Wang, H., Ullah, M.M., Klaser, A., Laptev, I., Schmid, C.: Evaluation of local spatio-temporal features for action recognition. In: BMVC 2009-British Machine Vision Conference, pp. 124\u20131. BMVA Press (2009)","DOI":"10.5244\/C.23.124"},{"key":"25_CR49","doi-asserted-by":"crossref","unstructured":"Wang, X., Han, T.X., Yan, S.: An HOG-LBP human detector with partial occlusion handling. In: 2009 IEEE 12th International Conference on Computer Vision, pp. 32\u201339. IEEE (2009)","DOI":"10.1109\/ICCV.2009.5459207"},{"key":"25_CR50","doi-asserted-by":"crossref","unstructured":"Wang, X., Wu, J., Chen, J., Li, L., Wang, Y.F., Wang, W.Y.: VATEX: a large-scale, high-quality multilingual dataset for video-and-language research. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 4581\u20134591 (2019)","DOI":"10.1109\/ICCV.2019.00468"},{"key":"25_CR51","doi-asserted-by":"crossref","unstructured":"Xu, J., Mei, T., Yao, T., Rui, Y.: MSR-VTT: a large video description dataset for bridging video and language. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5288\u20135296 (2016)","DOI":"10.1109\/CVPR.2016.571"},{"issue":"1","key":"25_CR52","doi-asserted-by":"publisher","first-page":"229","DOI":"10.1109\/TMM.2019.2924576","volume":"22","author":"C Yan","year":"2019","unstructured":"Yan, C., et al.: STAT: spatial-temporal attention mechanism for video captioning. IEEE Trans. Multimedia 22(1), 229\u2013241 (2019)","journal-title":"IEEE Trans. Multimedia"},{"key":"25_CR53","doi-asserted-by":"crossref","unstructured":"Yao, L., Torabi, A., Cho, K., Ballas, N., Pal, C., Larochelle, H., Courville, A.: Describing videos by exploiting temporal structure. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 4507\u20134515 (2015)","DOI":"10.1109\/ICCV.2015.512"},{"key":"25_CR54","doi-asserted-by":"crossref","unstructured":"Zhang, J., Peng, Y.: Object-aware aggregation with bidirectional temporal graph for video captioning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8327\u20138336 (2019)","DOI":"10.1109\/CVPR.2019.00852"},{"key":"25_CR55","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"25_CR56","unstructured":"Zhang, T., Kishore, V., Wu, F., Weinberger, K.Q., Artzi, Y.: Bertscore: evaluating text generation with Bert. In: International Conference on Learning Representations (2020). https:\/\/openreview.net\/forum?id=SkeHuCVFDr"},{"key":"25_CR57","doi-asserted-by":"crossref","unstructured":"Zhang, Z., et al.: Object relational graph with teacher-recommended learning for video captioning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13278\u201313288 (2020)","DOI":"10.1109\/CVPR42600.2020.01329"},{"key":"25_CR58","doi-asserted-by":"crossref","unstructured":"Zheng, Q., Wang, C., Tao, D.: Syntax-aware action targeting for video captioning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13096\u201313105 (2020)","DOI":"10.1109\/CVPR42600.2020.01311"},{"key":"25_CR59","doi-asserted-by":"crossref","unstructured":"Zhou, L., Zhou, Y., Corso, J.J., Socher, R., Xiong, C.: End-to-end dense video captioning with masked transformer. In: CVPR, pp. 8739\u20138748 (2018)","DOI":"10.1109\/CVPR.2018.00911"},{"key":"25_CR60","doi-asserted-by":"crossref","unstructured":"Zhu, Y., et al.: Texygen: a benchmarking platform for text generation models. In: ACM SIGIR, pp. 1097\u20131100 (2018)","DOI":"10.1145\/3209978.3210080"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-16788-1_25","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,4]],"date-time":"2024-10-04T12:36:53Z","timestamp":1728045413000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-16788-1_25"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031167874","9783031167881"],"references-count":60,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-16788-1_25","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"20 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"DAGM GCPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"DAGM German Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Konstanz","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Germany","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"44","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"dagm2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/gcpr-vmv-2022.uni-konstanz.de\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"78","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"37","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"47% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.6","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}