{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T15:36:30Z","timestamp":1742916990779,"version":"3.40.3"},"publisher-location":"Cham","reference-count":39,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031783401"},{"type":"electronic","value":"9783031783418"}],"license":[{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,12,2]],"date-time":"2024-12-02T00:00:00Z","timestamp":1733097600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-78341-8_13","type":"book-chapter","created":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T15:14:48Z","timestamp":1733066088000},"page":"194-209","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["AnnChor: A Video Dataset for\u00a0Temporal Action Localization in\u00a0Classical Ballet Choreography"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-2171-3275","authenticated-orcid":false,"given":"Margaux","family":"Bowditch","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-5632-1220","authenticated-orcid":false,"given":"Dustin","family":"van der Haar","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,12,2]]},"reference":[{"key":"13_CR1","doi-asserted-by":"crossref","unstructured":"Barioni, R.R., Costa, W., Aleluia, A., Teichrieb, V.: BalletVR: a virtual reality system for ballet arm positions training. In: 2019 21st Symposium on Virtual and Augmented Reality (SVR), pp. 10\u201316. IEEE (2019)","DOI":"10.1109\/SVR.2019.00018"},{"key":"13_CR2","doi-asserted-by":"publisher","unstructured":"Bisig, D.: Generative dance - a taxonomy and survey. In: Proceedings of the 8th International Conference on Movement and Computing. MOCO 2022, Association for Computing Machinery, New York, NY, USA (2022). https:\/\/doi.org\/10.1145\/3537972.3537978","DOI":"10.1145\/3537972.3537978"},{"key":"13_CR3","doi-asserted-by":"crossref","unstructured":"Caba\u00a0Heilbron, F., Escorcia, V., Ghanem, B., Carlos\u00a0Niebles, J.: Activitynet: A large-scale video benchmark for human activity understanding. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 961\u2013970 (2015)","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"13_CR4","unstructured":"Castro, D., et al.: Let\u2019s dance: learning from online dance videos. arXiv preprint arXiv:1801.07388 (2018)"},{"issue":"2","key":"13_CR5","doi-asserted-by":"publisher","first-page":"142","DOI":"10.1080\/14647893.2014.950643","volume":"16","author":"E Choi","year":"2015","unstructured":"Choi, E., Kim, N.Y.: Whole ballet education: exploring direct and indirect teaching methods. Res. Dance Educ. 16(2), 142\u2013160 (2015)","journal-title":"Res. Dance Educ."},{"key":"13_CR6","unstructured":"of\u00a0Dance, R.A.: Benesh international: Benesh movement notation: Royal academy of dance (2022). https:\/\/www.royalacademyofdance.org\/benesh-international-benesh-movement-notation\/"},{"key":"13_CR7","unstructured":"Dutta, A., Gupta, A., Zissermann, A.: VGG image annotator (VIA) (2016). http:\/\/www.robots.ox.ac.uk\/~vgg\/software\/via\/"},{"key":"13_CR8","doi-asserted-by":"crossref","unstructured":"Dutta, A., Zisserman, A.: The via annotation software for images, audio and video. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 2276\u20132279 (2019)","DOI":"10.1145\/3343031.3350535"},{"key":"13_CR9","doi-asserted-by":"crossref","unstructured":"Fathi, A., Mori, G.: Action recognition by learning mid-level motion features. In: 2008 IEEE Conference on Computer Vision and Pattern Recognition, pp.\u00a01\u20138 (2008)","DOI":"10.1109\/CVPR.2008.4587735"},{"key":"13_CR10","doi-asserted-by":"crossref","unstructured":"Gleason, J., Castillo, C.D., Chellappa, R.: Real-time detection of activities in untrimmed videos. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision Workshops, pp. 117\u2013125 (2020)","DOI":"10.1109\/WACVW50321.2020.9096937"},{"key":"13_CR11","doi-asserted-by":"crossref","unstructured":"Gleason, J., Schwarcz, S., Ranjan, R., Castillo, C.D., Chen, J.C., Chellappa, R.: Activity detection in untrimmed videos using chunk-based classifiers. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision Workshops, pp. 107\u2013116 (2020)","DOI":"10.1109\/WACVW50321.2020.9096912"},{"key":"13_CR12","doi-asserted-by":"publisher","first-page":"134611","DOI":"10.1109\/ACCESS.2021.3115476","volume":"9","author":"MS Hutchinson","year":"2021","unstructured":"Hutchinson, M.S., Gadepally, V.N.: Video action understanding. IEEE Access 9, 134611\u2013134637 (2021)","journal-title":"IEEE Access"},{"key":"13_CR13","doi-asserted-by":"crossref","unstructured":"Kahatapitiya, K., Ryoo, M.S.: Coarse-fine networks for temporal activity detection in videos. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8385\u20138394 (2021)","DOI":"10.1109\/CVPR46437.2021.00828"},{"key":"13_CR14","doi-asserted-by":"crossref","unstructured":"Kaspersen, E.T., G\u00f3rny, D., Erkut, C., Palamas, G.: Generative choreographies: the performance dramaturgy of the machine. In: VISIGRAPP (1: GRAPP), pp. 319\u2013326 (2020)","DOI":"10.5220\/0008990403190326"},{"key":"13_CR15","unstructured":"Kay, W., et\u00a0al.: The kinetics human action video dataset. arXiv preprint arXiv:1705.06950 (2017)"},{"issue":"2","key":"13_CR16","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/2735951","volume":"6","author":"M Kyan","year":"2015","unstructured":"Kyan, M., et al.: An approach to ballet dance training through MS Kinect and visualization in a cave virtual reality environment. ACM Trans. Intell. Syst. Technol. (TIST) 6(2), 1\u201337 (2015)","journal-title":"ACM Trans. Intell. Syst. Technol. (TIST)"},{"issue":"3","key":"13_CR17","doi-asserted-by":"publisher","first-page":"87","DOI":"10.1109\/MRA.2011.942118","volume":"18","author":"A LaViers","year":"2011","unstructured":"LaViers, A., Chen, Y., Belta, C., Egerstedt, M.: Automatic sequencing of ballet poses. IEEE Robot. Autom. Mag. 18(3), 87\u201395 (2011). https:\/\/doi.org\/10.1109\/MRA.2011.942118","journal-title":"IEEE Robot. Autom. Mag."},{"key":"13_CR18","doi-asserted-by":"crossref","unstructured":"Li, Y., Li, Y., Vasconcelos, N.: Resound: towards action recognition without representation bias. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 513\u2013528 (2018)","DOI":"10.1007\/978-3-030-01231-1_32"},{"key":"13_CR19","doi-asserted-by":"crossref","unstructured":"Li, Y., Chen, L., He, R., Wang, Z., Wu, G., Wang, L.: Multisports: a multi-person video dataset of spatio-temporally localized sports actions. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 13536\u201313545 (2021)","DOI":"10.1109\/ICCV48922.2021.01328"},{"key":"13_CR20","doi-asserted-by":"publisher","first-page":"6937","DOI":"10.1109\/TIP.2022.3217368","volume":"31","author":"Y Liu","year":"2022","unstructured":"Liu, Y., Wang, L., Wang, Y., Ma, X., Qiao, Y.: FineAction: a fine-grained video dataset for temporal action localization. IEEE Trans. Image Process. 31, 6937\u20136950 (2022)","journal-title":"IEEE Trans. Image Process."},{"issue":"1","key":"13_CR21","doi-asserted-by":"publisher","first-page":"39","DOI":"10.1080\/14647890801924550","volume":"9","author":"G Morris","year":"2008","unstructured":"Morris, G.: Artistry or mere technique? The value of the ballet competition. Res. Dance Educ. 9(1), 39\u201354 (2008)","journal-title":"Res. Dance Educ."},{"key":"13_CR22","unstructured":"Neagle, R., Ng, K., Ruddle, R.: Developing a virtual ballet dancer to visualise choreography. In: Proceedings of the Symposium on Language, Speech and Gesture for Expressive Characters (2004)"},{"issue":"12","key":"13_CR23","doi-asserted-by":"publisher","first-page":"155014771668179","DOI":"10.1177\/1550147716681794","volume":"12","author":"SH Park","year":"2016","unstructured":"Park, S.H., et al.: A feasibility study of ballet education using measurement and analysis on partial features of still scenes. Int. J. Distrib. Sens. Netw. 12(12), 1550147716681794 (2016)","journal-title":"Int. J. Distrib. Sens. Netw."},{"key":"13_CR24","unstructured":"Paskevska, A.: Ballet: From the First Pli\u00e9 to Mastery, an Eight-Year Course (2002)"},{"key":"13_CR25","unstructured":"Ryman, R., of\u00a0Dance\u00a0Enterprises, R.A., of\u00a0Dance Enterprises\u00a0Staff, R.A., Chivers, F.: The Foundations of Classical Ballet Technique. Royal Academy of Dancing (2008). https:\/\/books.google.co.za\/books?id=yGSKoAEACAAJ"},{"key":"13_CR26","doi-asserted-by":"crossref","unstructured":"Shao, D., Zhao, Y., Dai, B., Lin, D.: FineGym: a hierarchical video dataset for fine-grained action understanding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2616\u20132625 (2020)","DOI":"10.1109\/CVPR42600.2020.00269"},{"key":"13_CR27","doi-asserted-by":"crossref","unstructured":"Shi, D., Zhong, Y., Cao, Q., Ma, L., Li, J., Tao, D.: TriDet: temporal action detection with relative boundary modeling. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 18857\u201318866 (2023)","DOI":"10.1109\/CVPR52729.2023.01808"},{"key":"13_CR28","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"510","DOI":"10.1007\/978-3-319-46448-0_31","volume-title":"Computer Vision \u2013 ECCV 2016","author":"GA Sigurdsson","year":"2016","unstructured":"Sigurdsson, G.A., Varol, G., Wang, X., Farhadi, A., Laptev, I., Gupta, A.: Hollywood in homes: crowdsourcing data collection for activity understanding. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9905, pp. 510\u2013526. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46448-0_31"},{"key":"13_CR29","series-title":"Advances in Computer Vision and Pattern Recognition","doi-asserted-by":"publisher","first-page":"181","DOI":"10.1007\/978-3-319-09396-3_9","volume-title":"Computer Vision in Sports","author":"K Soomro","year":"2014","unstructured":"Soomro, K., Zamir, A.R.: Action recognition in realistic sports videos. In: Moeslund, T.B., Thomas, G., Hilton, A. (eds.) Computer Vision in Sports. ACVPR, pp. 181\u2013208. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-09396-3_9"},{"key":"13_CR30","doi-asserted-by":"crossref","unstructured":"Sun, J.: A digital feature recognition technology used in ballet training action correction. Comput. Intell. Neurosci. 2022 (2022)","DOI":"10.1155\/2022\/7953172"},{"key":"13_CR31","doi-asserted-by":"crossref","unstructured":"Trajkova, M., Cafaro, F.: E-ballet: designing for remote ballet learning. In: Proceedings of the 2016 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct, pp. 213\u2013216 (2016)","DOI":"10.1145\/2968219.2971442"},{"key":"13_CR32","doi-asserted-by":"publisher","unstructured":"Trajkova, M., Cafaro, F.: Takes tutu to ballet: designing visual and verbal feedback for augmented mirrors. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2(1) (2018). https:\/\/doi.org\/10.1145\/3191770","DOI":"10.1145\/3191770"},{"issue":"20","key":"13_CR33","doi-asserted-by":"publisher","first-page":"6957","DOI":"10.1016\/j.eswa.2015.04.039","volume":"42","author":"DK Vishwakarma","year":"2015","unstructured":"Vishwakarma, D.K., Kapoor, R.: Hybrid classifier based human activity recognition using the silhouette and cells. Expert Syst. Appl. 42(20), 6957\u20136965 (2015)","journal-title":"Expert Syst. Appl."},{"key":"13_CR34","doi-asserted-by":"publisher","first-page":"157","DOI":"10.1016\/j.aeue.2019.05.023","volume":"107","author":"DK Vishwakarma","year":"2019","unstructured":"Vishwakarma, D.K., Singh, T.: A visual cognizance based multi-resolution descriptor for human action recognition using key pose. AEU-Int. J. Electron. Commun. 107, 157\u2013169 (2019)","journal-title":"AEU-Int. J. Electron. Commun."},{"key":"13_CR35","doi-asserted-by":"publisher","first-page":"70477","DOI":"10.1109\/ACCESS.2020.2986861","volume":"8","author":"H Xia","year":"2020","unstructured":"Xia, H., Zhan, Y.: A survey on temporal action localization. IEEE Access 8, 70477\u201370487 (2020)","journal-title":"IEEE Access"},{"key":"13_CR36","doi-asserted-by":"crossref","unstructured":"Xu, J., Rao, Y., Yu, X., Chen, G., Zhou, J., Lu, J.: FineDiving: a fine-grained dataset for procedure-aware action quality assessment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2949\u20132958 (2022)","DOI":"10.1109\/CVPR52688.2022.00296"},{"issue":"2","key":"13_CR37","doi-asserted-by":"publisher","first-page":"375","DOI":"10.1007\/s11263-017-1013-y","volume":"126","author":"S Yeung","year":"2018","unstructured":"Yeung, S., Russakovsky, O., Jin, N., Andriluka, M., Mori, G., Fei-Fei, L.: Every moment counts: dense detailed labeling of actions in complex videos. Int. J. Comput. Visi. 126(2), 375\u2013389 (2018)","journal-title":"Int. J. Comput. Visi."},{"key":"13_CR38","doi-asserted-by":"crossref","unstructured":"Yu, T., Johnson, P.: Tour jet\u00e9, pirouette: dance choreographing by computers. Lecture notes in computer science, pp. 156\u2013157 (2003)","DOI":"10.1007\/3-540-45105-6_21"},{"key":"13_CR39","doi-asserted-by":"crossref","unstructured":"Zhao, H., Torralba, A., Torresani, L., Yan, Z.: HACS: human action clips and segments dataset for recognition and temporal localization. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 8668\u20138678 (2019)","DOI":"10.1109\/ICCV.2019.00876"}],"container-title":["Lecture Notes in Computer Science","Pattern Recognition"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-78341-8_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,12,1]],"date-time":"2024-12-01T16:06:08Z","timestamp":1733069168000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-78341-8_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,12,2]]},"ISBN":["9783031783401","9783031783418"],"references-count":39,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-78341-8_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,12,2]]},"assertion":[{"value":"2 December 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICPR","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Pattern Recognition","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Kolkata","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"India","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"1 December 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 December 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icpr2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/icpr2024.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}