{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,4,4]],"date-time":"2025-04-04T07:48:42Z","timestamp":1743752922891,"version":"3.40.3"},"publisher-location":"Cham","reference-count":46,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031422850"},{"type":"electronic","value":"9783031422867"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-42286-7_32","type":"book-chapter","created":{"date-parts":[[2023,8,24]],"date-time":"2023-08-24T17:02:36Z","timestamp":1692896556000},"page":"567-587","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Sample-Based Human Movement Detection for Interactive Videos Applied to Performing Arts"],"prefix":"10.1007","author":[{"given":"Rui","family":"Rodrigues","sequence":"first","affiliation":[]},{"given":"Jo\u00e3o","family":"Diogo","sequence":"additional","affiliation":[]},{"given":"Stephan","family":"Jurgens","sequence":"additional","affiliation":[]},{"given":"Carla","family":"Fernandes","sequence":"additional","affiliation":[]},{"given":"Nuno","family":"Correia","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,8,25]]},"reference":[{"key":"32_CR1","doi-asserted-by":"crossref","unstructured":"Torre, I., Galluccio, I., Coccoli, M.: Video augmentation to support video-based learning. In: ACM International Conference Proceeding Series (2022)","DOI":"10.1145\/3531073.3531179"},{"key":"32_CR2","unstructured":"Dias Pereira Dos, A., Loke, L., Martinez-Maldonado, R.: Exploring video annotation as a tool to support dance teaching. In: ACM International Conference Proceeding Series, pp. 448\u2013452 (2018)"},{"key":"32_CR3","doi-asserted-by":"publisher","unstructured":"Paul, S., Saha, R., Padhi, S., Majumdar, S., Das, P.P., Rao, S.: NrityaManch: an annotation and retrieval system for bharatanatyam dance ACM reference format. Proceedings of the 14th Annual Meeting of the Forum for Information Retrieval Evaluation, p. 22 (2022). https:\/\/doi.org\/10.1145\/3574318","DOI":"10.1145\/3574318"},{"key":"32_CR4","doi-asserted-by":"publisher","unstructured":"Singh, V., Latulipe, C., Carroll, E., Lottridge, D.: The choreographer\u2019s notebook-a video annotation system for dancers and choreographers. C and C 2011 - Proceedings of the 8th ACM Conference on Creativity and Cognition, pp. 197\u2013206 (2011). https:\/\/doi.org\/10.1145\/2069618.2069653","DOI":"10.1145\/2069618.2069653"},{"key":"32_CR5","doi-asserted-by":"publisher","unstructured":"Lagrue, S., et al.: An ontology web application-based annotation tool for intangible culture heritage dance videos. SUMAC 2019 - Proceedings of the 1st Workshop on Structuring and Understanding of Multimedia heritAge Contents, co-located with MM 2019, pp. 75\u201381 (2019). https:\/\/doi.org\/10.1145\/3347317.3357245","DOI":"10.1145\/3347317.3357245"},{"key":"32_CR6","doi-asserted-by":"publisher","unstructured":"Chen, K., Zhang, D., Yao, L., Guo, B., Yu, Z., Liu, Y.: Deep learning for sensor-based human activity recognition. ACM Computing Surveys (CSUR), p. 54 (2021). https:\/\/doi.org\/10.1145\/3447744","DOI":"10.1145\/3447744"},{"key":"32_CR7","doi-asserted-by":"publisher","unstructured":"Kuffner dos Anjos, R., Ribeiro, C., Fernandes, C., Sartor, J.B.: Three-Dimensional Visualization of Movement Qualities in Contemporary Dance (2018). https:\/\/doi.org\/10.1145\/3212721.3212812","DOI":"10.1145\/3212721.3212812"},{"key":"32_CR8","doi-asserted-by":"publisher","unstructured":"Ho, C., Tsai, W.T., Lin, K.S., Chen, H.H.: Extraction and alignment evaluation of motion beats for street dance. ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing \u2013 Proceedings, pp. 2429\u20132433 (2013). https:\/\/doi.org\/10.1109\/ICASSP.2013.6638091","DOI":"10.1109\/ICASSP.2013.6638091"},{"key":"32_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/J.CAEAI.2023.100125","volume":"4","author":"TS Ashwin","year":"2023","unstructured":"Ashwin, T.S., Prakash, V., Rajendran, R.: A systematic review of intelligent tutoring sys-tems based on Gross body movement detected using computer vision. Computers and Education: Artificial Intell. 4, 100125 (2023). https:\/\/doi.org\/10.1016\/J.CAEAI.2023.100125","journal-title":"Computers and Education: Artificial Intell."},{"key":"32_CR10","doi-asserted-by":"publisher","first-page":"1366","DOI":"10.1007\/S11263-022-01594-9","volume":"130","author":"Y Kong","year":"2022","unstructured":"Kong, Y., Fu, Y.: Human action recognition and prediction: a survey. Int J Comput Vis. 130, 1366\u20131401 (2022). https:\/\/doi.org\/10.1007\/S11263-022-01594-9","journal-title":"Int J Comput Vis."},{"key":"32_CR11","doi-asserted-by":"publisher","first-page":"133330","DOI":"10.1109\/ACCESS.2020.3010248","volume":"8","author":"TL Munea","year":"2020","unstructured":"Munea, T.L., Jembre, Y.Z., Weldegebriel, H.T., Chen, L., Huang, C., Yang, C.: The progress of human pose estimation: a survey and taxonomy of models applied in 2D human pose estimation. IEEE Access. 8, 133330\u2013133348 (2020). https:\/\/doi.org\/10.1109\/ACCESS.2020.3010248","journal-title":"IEEE Access."},{"key":"32_CR12","doi-asserted-by":"publisher","unstructured":"Difini, G.M., Martins, M.G., Barbosa, J.L.V.: Human pose estimation for training assistance: a systematic literature review. ACM International Conference Proceeding Series, pp.189\u2013196 (2021). https:\/\/doi.org\/10.1145\/3470482.3479633","DOI":"10.1145\/3470482.3479633"},{"key":"32_CR13","doi-asserted-by":"crossref","unstructured":"Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multi-person 2D pose estimation using part affinity fields. In: Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017 (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"32_CR14","doi-asserted-by":"publisher","unstructured":"Li, Y.C., Chang, C.T., Cheng, C.C., Huang, Y.L.: Baseball swing pose estimation using openpose. 2021 IEEE International Conference on Robotics, Automation and Artificial Intelligence, RAAI 2021, pp. 6\u20139 (2021). https:\/\/doi.org\/10.1109\/RAAI52226.2021.9507807","DOI":"10.1109\/RAAI52226.2021.9507807"},{"key":"32_CR15","doi-asserted-by":"publisher","unstructured":"Potempski, F., Sabo, A., Patterson, K.K.: Technical note: quantifying music-dance synchrony during salsa dancing with a deep learning-based 2D pose estimator. J Biomech 141, 111178 (2022). https:\/\/doi.org\/10.1016\/J.JBIOMECH.2022.111178","DOI":"10.1016\/J.JBIOMECH.2022.111178"},{"key":"32_CR16","doi-asserted-by":"publisher","unstructured":"Muhammad, K., et al.: Human action recognition using attention based LSTM network with dilated CNN features. Future Generation Computer Syst. 125, 820\u2013830 (2021). https:\/\/doi.org\/10.1016\/J.FUTURE.2021.06.045","DOI":"10.1016\/J.FUTURE.2021.06.045"},{"key":"32_CR17","doi-asserted-by":"publisher","unstructured":"Yen, C.C., Pin, T., Xu, H.: Bilateral pose transformer for human pose estimation. ACM International Conference Proceeding Series. Par F180473, pp. 21\u201329 (2022). https:\/\/doi.org\/10.1145\/3532342.3532346","DOI":"10.1145\/3532342.3532346"},{"key":"32_CR18","doi-asserted-by":"publisher","unstructured":"Xia, H., Zhang, Q.: VitPose: multi-view 3D human pose estimation with vision transformer. 2022 IEEE 8th International Conference on Computer and Communications, ICCC 2022, pp. 1922\u20131927 (2022). https:\/\/doi.org\/10.1109\/ICCC56324.2022.10065997","DOI":"10.1109\/ICCC56324.2022.10065997"},{"key":"32_CR19","doi-asserted-by":"publisher","unstructured":"Janardhanan, J., Umamaheswari, S.: A comprehensive study on human pose estimation. 8th International Conference on Advanced Computing and Communication Systems, ICACCS 2022, pp. 535\u2013541 (2022). https:\/\/doi.org\/10.1109\/ICACCS54159.2022.9784965","DOI":"10.1109\/ICACCS54159.2022.9784965"},{"key":"32_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3323335","volume":"52","author":"KE Raheb","year":"2020","unstructured":"Raheb, K.E., Stergiou, M., Katifori, A., Ioannidis, Y.: Dance interactive learning systems. ACM Comput Surv. 52, 1\u201337 (2020). https:\/\/doi.org\/10.1145\/3323335","journal-title":"ACM Comput Surv."},{"key":"32_CR21","doi-asserted-by":"crossref","unstructured":"Towey, D., et al.: Students as partners in a multi-media note-taking app development: best practices. In: Proceedings - 2017 IEEE\/ACM 39th International Conference on Software Engineering Companion, ICSE-C, pp. 334\u2013335. Institute of Electrical and Electronics Engineers Inc. (2017)","DOI":"10.1109\/ICSE-C.2017.58"},{"key":"32_CR22","doi-asserted-by":"crossref","unstructured":"Rodrigues, R., Neves Madeira, R.: Studying natural user interfaces for smart video annotation towards ubiquitous environments. In: 20th International Conference on Mobile and Ubiquitous Multimedia (MUM 2021), pp. 1\u201318. ACM, New York, NY, USA, Leuven, Belgium (2021)","DOI":"10.1145\/3490632.3490672"},{"key":"32_CR23","doi-asserted-by":"publisher","unstructured":"de Sousa, L., Richter, B., Nel, C.: The effect of multimedia use on the teaching and learning of Social Sciences at tertiary level: a case study. Yesterday and Today, pp. 1\u201322 (2017). https:\/\/doi.org\/10.17159\/2223-0386\/2017\/n17a1","DOI":"10.17159\/2223-0386\/2017\/n17a1"},{"key":"32_CR24","doi-asserted-by":"publisher","first-page":"1752","DOI":"10.1111\/JCAL.12714","volume":"38","author":"Y Sidi","year":"2022","unstructured":"Sidi, Y., Blau, I., Shamir-Inbal, T.: Mapping active and collaborative learning in higher education through annotations in hyper-video by learning analytics. J. Comput. Assist Learn. 38, 1752\u20131764 (2022). https:\/\/doi.org\/10.1111\/JCAL.12714","journal-title":"J. Comput. Assist Learn."},{"key":"32_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/S41239-021-00242-5\/FIGURES\/5","volume":"18","author":"SS Tseng","year":"2021","unstructured":"Tseng, S.S.: The influence of teacher annotations on student learning engagement and video watching behaviors. Int. J. Educ. Technol. High. Educ. 18, 1\u201317 (2021). https:\/\/doi.org\/10.1186\/S41239-021-00242-5\/FIGURES\/5","journal-title":"Int. J. Educ. Technol. High. Educ."},{"key":"32_CR26","doi-asserted-by":"publisher","first-page":"2571","DOI":"10.1007\/S11423-021-10041-4\/TABLES\/3","volume":"69","author":"N Mirriahi","year":"2021","unstructured":"Mirriahi, N., Jovanovi\u0107, J., Lim, L.A., Lodge, J.M.: Two sides of the same coin: video an-notations and in-video questions for active learning. Education Tech. Research Dev. 69, 2571\u20132588 (2021). https:\/\/doi.org\/10.1007\/S11423-021-10041-4\/TABLES\/3","journal-title":"Education Tech. Research Dev."},{"key":"32_CR27","doi-asserted-by":"crossref","unstructured":"Rijmer, S.: Negotiating deliberate choice-making: Insights from an interdisciplinary and mul-timodal encounter during the making of a new contemporary dance. In: Dance Data, Cognition, and Multimodal Communication, pp. 15\u201337. Routledge, London (2022)","DOI":"10.4324\/9781003106401-3"},{"key":"32_CR28","unstructured":"Jurgens, S.: Three methods of designing a workflow with multimodal video annotation in interdisciplinary choreographic processes. Multimodality and Performance, Fernandes, C. (Ed.). Cambridge Scholars Publishing, Newcastle upon Tyne, pp. 159\u2013178 (2016)"},{"key":"32_CR29","unstructured":"Butterworth, J., Wildschut, L.: Contemporary Choreography: A Critical Reader. Routledge (2009)"},{"key":"32_CR30","unstructured":"Harvie, J., Lavender, A.: Making Contemporary Theatre: International Rehearsal Processes, p. 252 (2010)"},{"key":"32_CR31","doi-asserted-by":"crossref","unstructured":"J\u00fcrgens, S., Fernandes, C.: Choreographic practice-as-research. In: Performance as Research, pp. 249\u2013274. Routledge, First edition. Routledge, New York (2017)","DOI":"10.4324\/9781315157672-14"},{"key":"32_CR32","doi-asserted-by":"publisher","unstructured":"Bargeron, D., Gupta, A., Grudin, J., Sanocki, E.: Annotations for streaming video on the web. Conference on Human Factors in Computing Systems \u2013 Proceedings, pp. 278\u2013279 (1999). https:\/\/doi.org\/10.1145\/632716.632887","DOI":"10.1145\/632716.632887"},{"key":"32_CR33","unstructured":"Wittenburg, P., Brugman, H., Russel, A., Klassmann, A., Sloetjes, H.: ELAN: a professional framework for multimodality research. In: Proceedings of the 5th International Conference on Language Resources and Evaluation, LREC 2006, pp. 1556\u20131559 (2006)"},{"key":"32_CR34","unstructured":"Buono, P., Desolda, G., Lanzilotti, R.: Scenes extraction from videos of telementored surgeries. Proceedings: DMS 2013 - 19th International Conference on Distributed Multimedia Systems, pp. 106\u2013111 (2013)"},{"key":"32_CR35","doi-asserted-by":"publisher","unstructured":"Pless, A., Hari, R., Brem, B., Woermamm, U., Schnabel, K.P.: Using self and peer video annotations of simulated patient encounters in communication training to facilitate the reflection of communication skills: an implementation study. GMS J Med Educ. 38 (2021). https:\/\/doi.org\/10.3205\/ZMA001451","DOI":"10.3205\/ZMA001451"},{"key":"32_CR36","doi-asserted-by":"publisher","unstructured":"Deng, D., et al.: EventAnchor: reducing Human Interactions in Event Annotation of Racket Sports Videos KEYWORDS. https:\/\/doi.org\/10.1145\/3411764.3445431","DOI":"10.1145\/3411764.3445431"},{"key":"32_CR37","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1109\/TLT.2012.15","volume":"6","author":"EF Risko","year":"2013","unstructured":"Risko, E.F., Foulsham, T., Dawson, S., Kingstone, A.: The collaborative lecture annotation system (CLAS): a new TOOL for distributed learning. IEEE Trans. Learn. Technol. 6, 4\u201313 (2013). https:\/\/doi.org\/10.1109\/TLT.2012.15","journal-title":"IEEE Trans. Learn. Technol."},{"key":"32_CR38","doi-asserted-by":"publisher","unstructured":"Cabral, D., Valente, J., Silva, J., Arag\u00e3o, U., Fernandes, C., Correia, N.: A creation-tool for contemporary dance using multimodal video annotation. MM\u201911 - Proceedings of the 2011 ACM Multimedia Conference and Co-Located Workshops, pp. 905\u2013908 (2011). https:\/\/doi.org\/10.1145\/2072298.2071899","DOI":"10.1145\/2072298.2071899"},{"key":"32_CR39","doi-asserted-by":"crossref","unstructured":"Ribeiro, C., Kuffner, R., Fernandes, C., Pereira, J.: 3D annotation in contemporary dance. In: Proceedings of the 3rd International Symposium on Movement and Computing, pp. 1\u20134. ACM, New York, NY, USA (2016)","DOI":"10.1145\/2948910.2948961"},{"key":"32_CR40","doi-asserted-by":"crossref","unstructured":"de Lahunta, S., Jenett, F.: Making digital choreographic objects interrelate. In: Performing the Digital, pp. 63\u201380. transcript Verlag (2016)","DOI":"10.14361\/9783839433553-003"},{"key":"32_CR41","unstructured":"Rizzo, A., et al.: WhoLoDancE: whole-body interaction learning for dance education. In: CEUR Workshop Proceedings, pp. 41\u201350 (2018)"},{"key":"32_CR42","doi-asserted-by":"publisher","unstructured":"Lagrue, S., et al.: An Ontology Web Application-based Annotation Tool for Intangible Culture Heritage Dance Videos (2019). https:\/\/doi.org\/10.1145\/3347317.3357245","DOI":"10.1145\/3347317.3357245"},{"key":"32_CR43","doi-asserted-by":"publisher","unstructured":"Rivi\u00e8re, J.P., Alaoui, S.F., Caramiaux, B., Mackay, W.E.: Capturing movement decomposition to support learning and teaching in contemporary dance. Proc ACM Hum Comput Interact. 3 (2019). https:\/\/doi.org\/10.1145\/3359188","DOI":"10.1145\/3359188"},{"key":"32_CR44","doi-asserted-by":"publisher","unstructured":"Rodrigues, R., Madeira, R.N., Correia, N.: Exploring the user interaction with a multimodal web-based video annotator. Lecture Notes of the Institute for Computer Sciences, Social-Informatics and Telecommunications Engineering, LNICST. 429 LNICST, pp. 13\u201322 (2022). https:\/\/doi.org\/10.1007\/978-3-030-99188-3_2\/COVER","DOI":"10.1007\/978-3-030-99188-3_2\/COVER"},{"key":"32_CR45","doi-asserted-by":"publisher","unstructured":"Rodrigues, R., Madeira, R.N., Correia, N., Fernandes, C., Ribeiro, S.: Multimodal web based video annotator with real-time human pose estimation. Lecture Notes in Comput-er Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics). 11872 LNCS, pp. 23\u201330 (2019). https:\/\/doi.org\/10.1007\/978-3-030-33617-2_3\/COVER","DOI":"10.1007\/978-3-030-33617-2_3\/COVER"},{"key":"32_CR46","doi-asserted-by":"publisher","unstructured":"Diogo, J., Rodrigues, R., Madeira, R., Correia, N.: Video annotation tool using human pose estimation for sports training. ACM International Conference Proceeding Series, pp. 262\u2013265 (2022). https:\/\/doi.org\/10.1145\/3568444.3570592","DOI":"10.1145\/3568444.3570592"}],"container-title":["Lecture Notes in Computer Science","Human-Computer Interaction \u2013 INTERACT 2023"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-42286-7_32","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,25]],"date-time":"2024-04-25T18:08:22Z","timestamp":1714068502000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-42286-7_32"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031422850","9783031422867"],"references-count":46,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-42286-7_32","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"25 August 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"INTERACT","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"IFIP Conference on Human-Computer Interaction","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"York","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28 August 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 September 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"19","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"interact2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"PCS","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"406","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"71","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"17% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4,01","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2,75","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Courses: 6 Industrial Experiences: 2 Interactive demos: 10 Panels: 2 Keynotes: 2 Posters: 44 Workshop summaries: 16- Submissions -As for full paper: 220 As for short papers: 186","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}