{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T02:18:24Z","timestamp":1742955504436,"version":"3.40.3"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030991876"},{"type":"electronic","value":"9783030991883"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-030-99188-3_2","type":"book-chapter","created":{"date-parts":[[2022,3,25]],"date-time":"2022-03-25T10:51:26Z","timestamp":1648205486000},"page":"13-22","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Exploring the User Interaction with a Multimodal Web-Based Video Annotator"],"prefix":"10.1007","author":[{"given":"Rui","family":"Rodrigues","sequence":"first","affiliation":[]},{"given":"Rui Neves","family":"Madeira","sequence":"additional","affiliation":[]},{"given":"Nuno","family":"Correia","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,3,25]]},"reference":[{"key":"2_CR1","doi-asserted-by":"publisher","first-page":"189","DOI":"10.1016\/j.patrec.2013.07.003","volume":"36","author":"M Turk","year":"2014","unstructured":"Turk, M.: Multimodal interaction: a review. Pattern Recogn. Lett. 36, 189\u2013195 (2014)","journal-title":"Pattern Recogn. Lett."},{"key":"2_CR2","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-642-00437-7_1","volume-title":"Human Machine Interaction","author":"B Dumas","year":"2009","unstructured":"Dumas, B., Lalanne, D., Oviatt, S.: Multimodal interfaces: a survey of principles, models and frameworks. In: Lalanne, D., Kohlas, J. (eds.) Human Machine Interaction. LNCS, vol. 5440, pp. 3\u201326. Springer, Heidelberg (2009). https:\/\/doi.org\/10.1007\/978-3-642-00437-7_1"},{"key":"2_CR3","first-page":"86","volume":"9","author":"\u00c1 Abuczki","year":"2013","unstructured":"Abuczki, \u00c1., Esfandiari Baiat, G.: An overview of multimodal corpora, annotation tools and schemes. Argumentu 9, 86\u201398 (2013)","journal-title":"Argumentu"},{"key":"2_CR4","unstructured":"CultureMoves: Culture Moves. https:\/\/culturemoves.eu\/. Accessed 17 Jun 2021"},{"key":"2_CR5","unstructured":"Europeana: Europeana. www.europeana.eu. Accessed 16 May 2021"},{"key":"2_CR6","doi-asserted-by":"crossref","unstructured":"Cao, Z., Simon, T., Wei, S.E., Sheikh, Y.: Realtime multi-person 2D pose estimation using part affinity fields. In: Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017 (2017)","DOI":"10.1109\/CVPR.2017.143"},{"key":"2_CR7","doi-asserted-by":"crossref","unstructured":"Cabral, D., Valente, J., Silva, J., Arag\u00e3o, U., Fernandes, C., Correia, N.: A creation-tool for contemporary dance using multimodal video annotation. In: Proceedings of the 2011 ACM Multimedia Conference and Workshops, MM 2011 (2011)","DOI":"10.1145\/2072298.2071899"},{"key":"2_CR8","unstructured":"Wittenburg, P., Brugman, H., Russel, A., Klassmann, A., Sloetjes, H.: ELAN: a professional framework for multimodality research. In: Proceedings of the 5th International Conference on Language Resources and Evaluation, LREC 2006 (2006)"},{"key":"2_CR9","doi-asserted-by":"crossref","unstructured":"Goldman, D.B., Gonterman, C., Curless, B., Salesin, D., Seitz, S.M.: Video object annotation, navigation, and composition. In: Proceedings of the 21st Annual ACM Symposium on User Interface Software and Technology, UIST 2008 (2008)","DOI":"10.1145\/1449715.1449719"},{"key":"2_CR10","doi-asserted-by":"crossref","unstructured":"Singh, V., Latulipe, C., Carroll, E., Lottridge, D.: The choreographer\u2019s notebook-a video annotation system for dancers and choreographers. In: Proceedings of the 8th ACM Conference on Creativity and Cognition, C and C 2011 (2011)","DOI":"10.1145\/2069618.2069653"},{"key":"2_CR11","doi-asserted-by":"crossref","unstructured":"El Raheb, K., Kasomoulis, A., Katifori, A., Rezkalla, M., Ioannidis, Y.: A web-based system for annotation of dance multimodal recordings by dance practitioners and experts. In: ACM International Conference Proceeding Series (2018)","DOI":"10.1145\/3212721.3212722"},{"key":"2_CR12","doi-asserted-by":"crossref","unstructured":"Cabral, D., Valente, J.G., Arag\u00e3o, U., Fernandes, C., Correia, N.: Evaluation of a multimodal video annotator for contemporary dance. In: Proceedings of the Workshop on Advanced Visual Interfaces AVI (2012)","DOI":"10.1145\/2254556.2254663"},{"issue":"5","key":"2_CR13","doi-asserted-by":"publisher","first-page":"6889","DOI":"10.1007\/s11042-016-3329-y","volume":"76","author":"D Cabral","year":"2016","unstructured":"Cabral, D., Correia, N.: Video editing with pen-based technology. Multimedia Tools Appl. 76(5), 6889\u20136914 (2016)","journal-title":"Multimedia Tools Appl."},{"key":"2_CR14","doi-asserted-by":"crossref","unstructured":"Silva, J., Fernandes, C., Cabral, D., Correia, N.: Real-time annotation of video objects on tablet computers. In: Proceedings of the 11th International Conference on Mobile and Ubiquitous Multimedia, MUM 2012 (2012)","DOI":"10.1145\/2406367.2406391"},{"key":"2_CR15","doi-asserted-by":"crossref","unstructured":"Ribeiro, C., Kuffner, R., Fernandes, C., Pereira, J.: 3D annotation in contemporary dance: Enhancing the creation-tool video annotator. In: ACM International Conference Proceeding Series (2016)","DOI":"10.1145\/2948910.2948961"},{"key":"2_CR16","unstructured":"Wipster | Review Software. https:\/\/wipster.io\/. Accessed 15 Jun 2021"},{"key":"2_CR17","unstructured":"Camtasia. https:\/\/www.techsmith.com\/video-editor.html. Accessed 2 Jun 2021"},{"key":"2_CR18","unstructured":"Frame.io. https:\/\/www.frame.io\/. Accessed 25 May 2021"},{"key":"2_CR19","unstructured":"Vimeo. https:\/\/vimeo.com\/features\/video-collaboration. Accessed 5 Jun 2021"},{"key":"2_CR20","doi-asserted-by":"publisher","unstructured":"Felzenszwalb, P., McAllester, D., Ramanan, D.: A discriminatively trained, multiscale, deformable part model. In: 26th IEEE Conference on Computer Vision and Pattern Recognition, CVPR (2008). https:\/\/doi.org\/10.1109\/CVPR.2008.4587597","DOI":"10.1109\/CVPR.2008.4587597"},{"key":"2_CR21","doi-asserted-by":"publisher","unstructured":"Andriluka, M., Roth, S., Schiele, B.: Pictorial structures revisited: people detection and articulated pose estimation, pp. 1014\u20131021 (2010). https:\/\/doi.org\/10.1109\/CVPR.2009.5206754","DOI":"10.1109\/CVPR.2009.5206754"},{"key":"2_CR22","doi-asserted-by":"publisher","unstructured":"Yang, Y., Ramanan, D.: Articulated pose estimation with flexible mixtures-of-parts. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 1385\u20131392 (2011). https:\/\/doi.org\/10.1109\/CVPR.2011.5995741","DOI":"10.1109\/CVPR.2011.5995741"},{"key":"2_CR23","unstructured":"Markoff, J.: Scientists See Promise in Deep-Learning Program. Nyt. (2012)"},{"key":"2_CR24","doi-asserted-by":"crossref","unstructured":"Toshev, A., Szegedy, C.: DeepPose: human pose estimation via deep neural networks (2014)","DOI":"10.1109\/CVPR.2014.214"},{"key":"2_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"34","DOI":"10.1007\/978-3-319-46466-4_3","volume-title":"Computer Vision \u2013 ECCV 2016","author":"E Insafutdinov","year":"2016","unstructured":"Insafutdinov, E., Pishchulin, L., Andres, B., Andriluka, M., Schiele, B.: DeeperCut: a deeper, stronger, and faster multi-person pose estimation model. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9910, pp. 34\u201350. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46466-4_3"},{"key":"2_CR26","doi-asserted-by":"crossref","unstructured":"Sun, K., Xiao, B., Liu, D., Wang, J.: Deep high-resolution representation learning for human pose estimation. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 5686\u20135696 (2019)","DOI":"10.1109\/CVPR.2019.00584"},{"key":"2_CR27","doi-asserted-by":"crossref","unstructured":"Wei, S.E., Ramakrishna, V., Kanade, T., Sheikh, Y.: Convolutional pose machines. In: Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 4724\u20134732 (2016)","DOI":"10.1109\/CVPR.2016.511"},{"key":"2_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"483","DOI":"10.1007\/978-3-319-46484-8_29","volume-title":"Computer Vision \u2013 ECCV 2016","author":"A Newell","year":"2016","unstructured":"Newell, A., Yang, K., Deng, J.: Stacked hourglass networks for human pose estimation. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 483\u2013499. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46484-8_29"},{"key":"2_CR29","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"23","DOI":"10.1007\/978-3-030-33617-2_3","volume-title":"Intelligent Data Engineering and Automated Learning \u2013 IDEAL 2019","author":"R Rodrigues","year":"2019","unstructured":"Rodrigues, R., Madeira, R.N., Correia, N., Fernandes, C., Ribeiro, S.: Multimodal web based video annotator with real-time human pose estimation. In: Yin, H., Camacho, D., Tino, P., Tall\u00f3n-Ballesteros, A.J., Menezes, R., Allmendinger, R. (eds.) IDEAL 2019. LNCS, vol. 11872, pp. 23\u201330. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-33617-2_3"},{"key":"2_CR30","unstructured":"PoseNet. https:\/\/learn.ml5js.org\/#\/reference\/posenet?id=posenet. Accessed 15 Nov 2021"}],"container-title":["Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering","Intelligent Technologies for Interactive Entertainment"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-99188-3_2","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,3,25]],"date-time":"2022-03-25T10:52:25Z","timestamp":1648205545000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-030-99188-3_2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783030991876","9783030991883"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-99188-3_2","relation":{},"ISSN":["1867-8211","1867-822X"],"issn-type":[{"type":"print","value":"1867-8211"},{"type":"electronic","value":"1867-822X"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"25 March 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"INTETAIN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Intelligent Technologies for Interactive Entertainment","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2021","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"3 December 2021","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 December 2021","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"13","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"intetain2021","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/intetain.eai-conferences.org\/2021\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Confy+","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"50","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"25","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"50% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}