{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T17:24:28Z","timestamp":1743009868518,"version":"3.40.3"},"publisher-location":"Cham","reference-count":61,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031198328"},{"type":"electronic","value":"9783031198335"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19833-5_12","type":"book-chapter","created":{"date-parts":[[2022,11,4]],"date-time":"2022-11-04T00:40:30Z","timestamp":1667522430000},"page":"195-212","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["GraphVid: It only\u00a0Takes a\u00a0Few Nodes to\u00a0Understand a\u00a0Video"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5538-0616","authenticated-orcid":false,"given":"Eitan","family":"Kosman","sequence":"first","affiliation":[]},{"given":"Dotan","family":"Di Castro","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,4]]},"reference":[{"issue":"9","key":"12_CR1","first-page":"1","volume":"54","author":"S Abadal","year":"2021","unstructured":"Abadal, S., Jain, A., Guirado, R., L\u00f3pez-Alonso, J., Alarc\u00f3n, E.: Computing graph neural networks: a survey from algorithms to accelerators. ACM Computi. Surv. (CSUR) 54(9), 1\u201338 (2021)","journal-title":"ACM Computi. Surv. (CSUR)"},{"key":"12_CR2","unstructured":"Achanta, R., Shaji, A., Smith, K., Lucchi, A., Fua, P., S\u00fcsstrunk, S.: Slic superpixels. Technical report (2010)"},{"key":"12_CR3","unstructured":"Akbari, H., et al.: Vatt: transformers for multimodal self-supervised learning from raw video, audio and text. arXiv preprint arXiv:2104.11178 (2021)"},{"key":"12_CR4","doi-asserted-by":"crossref","unstructured":"Arnab, A., Dehghani, M., Heigold, G., Sun, C., Lu\u010di\u0107, M., Schmid, C.: ViViT: a video vision transformer. arXiv preprint arXiv:2103.15691 (2021)","DOI":"10.1109\/ICCV48922.2021.00676"},{"key":"12_CR5","doi-asserted-by":"crossref","unstructured":"Avelar, P.H., Tavares, A.R., da Silveira, T.L., Jung, C.R., Lamb, L.C.: Superpixel image classification with graph attention networks. In: SIBGRAPI, pp. 203\u2013209. IEEE (2020)","DOI":"10.1109\/SIBGRAPI51738.2020.00035"},{"key":"12_CR6","unstructured":"Bertasius, G., Wang, H., Torresani, L.: Is space-time attention all you need for video understanding? arXiv preprint arXiv:2102.05095 (2021)"},{"key":"12_CR7","doi-asserted-by":"crossref","unstructured":"Cao, Y., Xu, J., Lin, S., Wei, F., Hu, H.: GCNet: non-local networks meet squeeze-excitation networks and beyond. In: ICCV Workshops, pp. 0\u20130 (2019)","DOI":"10.1109\/ICCVW.2019.00246"},{"key":"12_CR8","doi-asserted-by":"crossref","unstructured":"Chang, J., Wei, D., Fisher, J.W.: A video representation using temporal superpixels. In: CVPR, pp. 2051\u20132058 (2013)","DOI":"10.1109\/CVPR.2013.267"},{"key":"12_CR9","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"364","DOI":"10.1007\/978-3-030-01246-5_22","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Chen","year":"2018","unstructured":"Chen, Y., Kalantidis, Y., Li, J., Yan, S., Feng, J.: Multi-fiber networks for video recognition. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11205, pp. 364\u2013380. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01246-5_22"},{"key":"12_CR10","unstructured":"Corso, G., Cavalleri, L., Beaini, D., Li\u00f2, P., Veli\u010dkovi\u0107, P.: Principal neighbourhood aggregation for graph nets. arXiv preprint arXiv:2004.05718 (2020)"},{"key":"12_CR11","doi-asserted-by":"crossref","unstructured":"Dadsetan, S., Pichler, D., Wilson, D., Hovakimyan, N., Hobbs, J.: Superpixels and graph convolutional neural networks for efficient detection of nutrient deficiency stress from aerial imagery. In: CVPR, pp. 2950\u20132959 (2021)","DOI":"10.1109\/CVPRW53098.2021.00330"},{"key":"12_CR12","unstructured":"Dokania, S., Singh, V.: Graph representation learning for audio & music genre classification. arXiv preprint arXiv:1910.11117 (2019)"},{"key":"12_CR13","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"12_CR14","doi-asserted-by":"crossref","unstructured":"Du Tran, H.W., Torresani, L., Ray, J., Lecun, Y., Paluri, M.: A closer look at spatiotemporal convolutions for action recognition (2017)","DOI":"10.1109\/CVPR.2018.00675"},{"key":"12_CR15","doi-asserted-by":"crossref","unstructured":"Fan, H., et al.: Multiscale vision transformers. arXiv preprint arXiv:2104.11227 (2021)","DOI":"10.1109\/ICCV48922.2021.00675"},{"key":"12_CR16","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C.: X3D: expanding architectures for efficient video recognition. In: CVPR, pp. 203\u2013213 (2020)","DOI":"10.1109\/CVPR42600.2020.00028"},{"key":"12_CR17","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Fan, H., Malik, J., He, K.: Slowfast networks for video recognition. In: ICCV, pp. 6202\u20136211 (2019)","DOI":"10.1109\/ICCV.2019.00630"},{"key":"12_CR18","doi-asserted-by":"crossref","unstructured":"Fern\u00e1ndez, D., et al.: ViTs: video tagging system from massive web multimedia collections. In: ICCV Workshops, pp. 337\u2013346 (2017)","DOI":"10.1109\/ICCVW.2017.48"},{"key":"12_CR19","unstructured":"Fey, M., Lenssen, J.E.: Fast graph representation learning with PyTorch geometric. arXiv preprint arXiv:1903.02428 (2019)"},{"key":"12_CR20","doi-asserted-by":"crossref","unstructured":"Girdhar, R., Carreira, J., Doersch, C., Zisserman, A.: Video action transformer network. In: CVPR, pp. 244\u2013253 (2019)","DOI":"10.1109\/CVPR.2019.00033"},{"key":"12_CR21","doi-asserted-by":"crossref","unstructured":"Gonzalo-Martin, C., Garcia-Pedrero, A., Lillo-Saavedra, M., Menasalvas, E.: Deep learning for superpixel-based classification of remote sensing images, September 2016. https:\/\/proceedings.utwente.nl\/401\/","DOI":"10.3990\/2.401"},{"key":"12_CR22","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"12_CR23","unstructured":"Kay, W., et al.: The kinetics human action video dataset. arXiv preprint arXiv:1705.06950 (2017)"},{"key":"12_CR24","unstructured":"Kim, A.: fast-slic (2019). https:\/\/github.com\/Algy\/fast-slic"},{"key":"12_CR25","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"12_CR26","unstructured":"Kipf, T.N., Welling, M.: Semi-supervised classification with graph convolutional networks. arXiv preprint arXiv:1609.02907 (2016)"},{"key":"12_CR27","doi-asserted-by":"crossref","unstructured":"Kondratyuk, D., et al.: MoviNets: mobile video networks for efficient video recognition. In: CVPR, pp. 16020\u201316030 (2021)","DOI":"10.1109\/CVPR46437.2021.01576"},{"key":"12_CR28","doi-asserted-by":"crossref","unstructured":"Kumar, A., Singh, S.S., Singh, K., Biswas, B.: Link prediction techniques, applications, and performance: a survey. Physica A: Stat. Mech. Appl. 553, 124289 (2020)","DOI":"10.1016\/j.physa.2020.124289"},{"issue":"7553","key":"12_CR29","doi-asserted-by":"publisher","first-page":"436","DOI":"10.1038\/nature14539","volume":"521","author":"Y LeCun","year":"2015","unstructured":"LeCun, Y., Bengio, Y., Hinton, G.: Deep learning. Nature 521(7553), 436\u2013444 (2015)","journal-title":"Nature"},{"key":"12_CR30","unstructured":"Li, X., et al.: ViDTR: video transformer without convolutions. arXiv preprint arXiv:2104.11746 (2021)"},{"key":"12_CR31","unstructured":"Liu, Z., et al.: Video swin transformer. arXiv preprint arXiv:2106.13230 (2021)"},{"key":"12_CR32","doi-asserted-by":"crossref","unstructured":"Mittal, S., et al.: A survey of accelerator architectures for 3D convolution neural networks. J. Syst. Archit. 102041 (2021)","DOI":"10.1016\/j.sysarc.2021.102041"},{"key":"12_CR33","doi-asserted-by":"crossref","unstructured":"Monti, F., Boscaini, D., Masci, J., Rodola, E., Svoboda, J., Bronstein, M.M.: Geometric deep learning on graphs and manifolds using mixture model CNNs. In: CVPR, pp. 5115\u20135124 (2017)","DOI":"10.1109\/CVPR.2017.576"},{"key":"12_CR34","doi-asserted-by":"crossref","unstructured":"Neimark, D., Bar, O., Zohar, M., Asselmann, D.: Video transformer network. arXiv preprint arXiv:2102.00719 (2021)","DOI":"10.1109\/ICCVW54120.2021.00355"},{"key":"12_CR35","unstructured":"Oprea, S., et al.: A review on deep learning techniques for video prediction. IEEE Trans. Pattern Anal. Mach. Intell. (2020)"},{"key":"12_CR36","doi-asserted-by":"crossref","unstructured":"Papon, J., Abramov, A., Schoeler, M., Worgotter, F.: Voxel cloud connectivity segmentation-supervoxels for point clouds. In: CVPR, pp. 2027\u20132034 (2013)","DOI":"10.1109\/CVPR.2013.264"},{"issue":"3","key":"12_CR37","doi-asserted-by":"publisher","first-page":"2259","DOI":"10.1007\/s10462-020-09904-8","volume":"54","author":"P Pareek","year":"2021","unstructured":"Pareek, P., Thakkar, A.: A survey on video-based human action recognition: recent updates, datasets, challenges, and applications. Artif. Intell. Rev. 54(3), 2259\u20132322 (2021)","journal-title":"Artif. Intell. Rev."},{"key":"12_CR38","unstructured":"Piergiovanni, A., Angelova, A., Ryoo, M.S.: Tiny video networks. Appl. AI Lett. e38 (2019)"},{"key":"12_CR39","unstructured":"Rong, Y., Huang, W., Xu, T., Huang, J.: DropEdge: towards deep graph convolutional networks on node classification. arXiv preprint arXiv:1907.10903 (2019)"},{"key":"12_CR40","unstructured":"Ryoo, M.S., Piergiovanni, A., Tan, M., Angelova, A.: AssembleNet: searching for multi-stream neural connectivity in video architectures. arXiv preprint arXiv:1905.13209 (2019)"},{"key":"12_CR41","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"593","DOI":"10.1007\/978-3-319-93417-4_38","volume-title":"The Semantic Web","author":"M Schlichtkrull","year":"2018","unstructured":"Schlichtkrull, M., Kipf, T.N., Bloem, P., van\u00a0den Berg, R., Titov, I., Welling, M.: Modeling relational data with graph convolutional networks. In: Gangemi, A., et al. (eds.) ESWC 2018. LNCS, vol. 10843, pp. 593\u2013607. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-319-93417-4_38"},{"key":"12_CR42","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"510","DOI":"10.1007\/978-3-319-46448-0_31","volume-title":"Computer Vision \u2013 ECCV 2016","author":"GA Sigurdsson","year":"2016","unstructured":"Sigurdsson, G.A., Varol, G., Wang, X., Farhadi, A., Laptev, I., Gupta, A.: Hollywood in homes: crowdsourcing data collection for activity understanding. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 510\u2013526. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_31"},{"key":"12_CR43","unstructured":"Simonyan, K., Zisserman, A.: Two-stream convolutional networks for action recognition in videos. arXiv preprint arXiv:1406.2199 (2014)"},{"key":"12_CR44","unstructured":"Srivastava, N., Mansimov, E., Salakhudinov, R.: Unsupervised learning of video representations using LSTMs. In: ICML, pp. 843\u2013852. PMLR (2015)"},{"key":"12_CR45","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.cviu.2017.03.007","volume":"166","author":"D Stutz","year":"2018","unstructured":"Stutz, D., Hermans, A., Leibe, B.: Superpixels: an evaluation of the state-of-the-art. Comput. Vis. Image Underst. 166, 1\u201327 (2018)","journal-title":"Comput. Vis. Image Underst."},{"key":"12_CR46","unstructured":"Suarez, J.J.P., Naval Jr., P.C.: A survey on deep learning techniques for video anomaly detection. arXiv preprint arXiv:2009.14146 (2020)"},{"key":"12_CR47","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., J\u00e9gou, H.: Training data-efficient image transformers & distillation through attention. In: ICML, pp. 10347\u201310357. PMLR (2021)"},{"key":"12_CR48","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., Paluri, M.: Learning spatiotemporal features with 3D convolutional networks. In: ICCV, pp. 4489\u20134497 (2015)","DOI":"10.1109\/ICCV.2015.510"},{"key":"12_CR49","doi-asserted-by":"crossref","unstructured":"Tran, D., Wang, H., Torresani, L., Ray, J., LeCun, Y., Paluri, M.: A closer look at spatiotemporal convolutions for action recognition. In: CVPR, pp. 6450\u20136459 (2018)","DOI":"10.1109\/CVPR.2018.00675"},{"key":"12_CR50","doi-asserted-by":"publisher","first-page":"1155","DOI":"10.1109\/ACCESS.2017.2778011","volume":"6","author":"A Ullah","year":"2017","unstructured":"Ullah, A., Ahmad, J., Muhammad, K., Sajjad, M., Baik, S.W.: Action recognition in video sequences using deep bi-directional LSTM with CNN features. IEEE Access 6, 1155\u20131166 (2017)","journal-title":"IEEE Access"},{"key":"12_CR51","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, pp. 5998\u20136008 (2017)"},{"key":"12_CR52","unstructured":"Veli\u010dkovi\u0107, P., Cucurull, G., Casanova, A., Romero, A., Lio, P., Bengio, Y.: Graph attention networks. arXiv preprint arXiv:1710.10903 (2017)"},{"key":"12_CR53","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., He, K.: Non-local neural networks. In: CVPR, pp. 7794\u20137803 (2018)","DOI":"10.1109\/CVPR.2018.00813"},{"key":"12_CR54","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"413","DOI":"10.1007\/978-3-030-01228-1_25","volume-title":"Computer Vision \u2013 ECCV 2018","author":"X Wang","year":"2018","unstructured":"Wang, X., Gupta, A.: Videos as space-time region graphs. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11209, pp. 413\u2013431. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01228-1_25"},{"key":"12_CR55","doi-asserted-by":"crossref","unstructured":"Xie, R., Liu, Z., Jia, J., Luan, H., Sun, M.: Representation learning of knowledge graphs with entity descriptions. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol. 30 (2016)","DOI":"10.1609\/aaai.v30i1.10329"},{"key":"12_CR56","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"318","DOI":"10.1007\/978-3-030-01267-0_19","volume-title":"Computer Vision \u2013 ECCV 2018","author":"S Xie","year":"2018","unstructured":"Xie, S., Sun, C., Huang, J., Tu, Z., Murphy, K.: Rethinking spatiotemporal feature learning: speed-accuracy trade-offs in video classification. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11219, pp. 318\u2013335. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01267-0_19"},{"key":"12_CR57","doi-asserted-by":"publisher","unstructured":"Xu, C., Corso, J.J.: Evaluation of super-voxel methods for early video processing. In: 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1202\u20131209 (2012). https:\/\/doi.org\/10.1109\/CVPR.2012.6247802","DOI":"10.1109\/CVPR.2012.6247802"},{"key":"12_CR58","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"626","DOI":"10.1007\/978-3-642-33783-3_45","volume-title":"Computer Vision \u2013 ECCV 2012","author":"C Xu","year":"2012","unstructured":"Xu, C., Xiong, C., Corso, J.J.: Streaming hierarchical video segmentation. In: Fitzgibbon, A., Lazebnik, S., Perona, P., Sato, Y., Schmid, C. (eds.) ECCV 2012. LNCS, vol. 7577, pp. 626\u2013639. Springer, Heidelberg (2012). https:\/\/doi.org\/10.1007\/978-3-642-33783-3_45"},{"key":"12_CR59","doi-asserted-by":"crossref","unstructured":"Yue-Hei Ng, J., Hausknecht, M., Vijayanarasimhan, S., Vinyals, O., Monga, R., Toderici, G.: Beyond short snippets: deep networks for video classification. In: CVPR, pp. 4694\u20134702 (2015)","DOI":"10.1109\/CVPR.2015.7299101"},{"issue":"1","key":"12_CR60","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1109\/TBDATA.2018.2850013","volume":"6","author":"D Zhang","year":"2018","unstructured":"Zhang, D., Yin, J., Zhu, X., Zhang, C.: Network representation learning: a survey. IEEE Trans. Big Data 6(1), 3\u201328 (2018)","journal-title":"IEEE Trans. Big Data"},{"key":"12_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, S., Qin, Y., Sun, K., Lin, Y.: Few-shot audio classification with attentional graph neural networks. In: INTERSPEECH, pp. 3649\u20133653 (2019)","DOI":"10.21437\/Interspeech.2019-1532"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19833-5_12","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,1,9]],"date-time":"2023-01-09T15:33:43Z","timestamp":1673278423000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19833-5_12"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031198328","9783031198335"],"references-count":61,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19833-5_12","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"4 November 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}