{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,31]],"date-time":"2026-01-31T04:52:19Z","timestamp":1769835139005,"version":"3.49.0"},"reference-count":70,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2024,7,14]],"date-time":"2024-07-14T00:00:00Z","timestamp":1720915200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,7,14]],"date-time":"2024-07-14T00:00:00Z","timestamp":1720915200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2024,12]]},"DOI":"10.1007\/s11263-024-02146-z","type":"journal-article","created":{"date-parts":[[2024,7,14]],"date-time":"2024-07-14T02:02:21Z","timestamp":1720922541000},"page":"6069-6090","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Procedure-Aware Action Quality Assessment: Datasets and Performance Evaluation"],"prefix":"10.1007","volume":"132","author":[{"given":"Jinglin","family":"Xu","sequence":"first","affiliation":[]},{"given":"Yongming","family":"Rao","sequence":"additional","affiliation":[]},{"given":"Jie","family":"Zhou","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6121-5529","authenticated-orcid":false,"given":"Jiwen","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,7,14]]},"reference":[{"key":"2146_CR1","doi-asserted-by":"crossref","unstructured":"Bai, Y., Zhou, D., Zhang, S., Wang, J., Ding, E., Guan, Y., Long, Y., & Wang, J. (2022). Action quality assessment with temporal parsing transformer. In ECCV, (pp. 422\u2013438).","DOI":"10.1007\/978-3-031-19772-7_25"},{"key":"2146_CR2","doi-asserted-by":"crossref","unstructured":"Bertasius, G., Soo\u00a0Park, H., Yu, S.X., & Shi, J. (2017). Am i a baller? basketball performance assessment from first-person videos. In ICCV, (pp. 2177\u20132185).","DOI":"10.1109\/ICCV.2017.239"},{"key":"2146_CR3","doi-asserted-by":"crossref","unstructured":"Caba\u00a0Heilbron, F.,\u00a0Escorcia, V., Ghanem, B., & Niebles, J.C. (2015). Activitynet: A large-scale video benchmark for human activity understanding. In CVPR, (pp. 961\u2013970).","DOI":"10.1109\/CVPR.2015.7298698"},{"key":"2146_CR4","doi-asserted-by":"crossref","unstructured":"Carreira, J., & Zisserman, A. (2017). Quo vadis, action recognition? a new model and the kinetics dataset. In CVPR, (pp. 6299\u20136308).","DOI":"10.1109\/CVPR.2017.502"},{"key":"2146_CR5","doi-asserted-by":"publisher","first-page":"2846","DOI":"10.1007\/s11263-021-01486-4","volume":"129","author":"X Chen","year":"2021","unstructured":"Chen, X., Pang, A., Yang, W., Ma, Y., Xu, L., & Yu, J. (2021). Sportscap: Monocular 3d human motion capture and fine-grained understanding in challenging sports videos. IJCV, 129, 2846\u20132864.","journal-title":"IJCV"},{"key":"2146_CR6","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., et\u00a0al. (2020). An image is worth 16x16 words: Transformers for image recognition at scale. arXiv:2010.11929."},{"key":"2146_CR7","doi-asserted-by":"crossref","unstructured":"Doughty, H., Damen, D., & Mayol-Cuevas, W. (2018). Who\u2019s better? who\u2019s best? pairwise deep ranking for skill determination. In CVPR, (pp. 6057\u20136066).","DOI":"10.1109\/CVPR.2018.00634"},{"key":"2146_CR8","doi-asserted-by":"crossref","unstructured":"Doughty, H., Mayol-Cuevas, W., & Damen, D. (2019). The pros and cons: Rank-aware temporal attention for skill determination in long videos. In CVPR, (pp. 7862\u20137871).","DOI":"10.1109\/CVPR.2019.00805"},{"key":"2146_CR9","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Pinz, A., & Zisserman, A. (2016). Convolutional two-stream network fusion for video action recognition. In CVPR, (pp. 1933\u20131941).","DOI":"10.1109\/CVPR.2016.213"},{"key":"2146_CR10","doi-asserted-by":"crossref","unstructured":"Gao, J., Zheng, W.S., Pan, J.H., Gao, C., Wang, Y., Zeng, W., & Lai, J. (2020). An asymmetric modeling for action assessment. In ECCV, (pp. 222\u2013238).","DOI":"10.1007\/978-3-030-58577-8_14"},{"issue":"3","key":"2146_CR11","doi-asserted-by":"publisher","first-page":"659","DOI":"10.1007\/s11263-022-01695-5","volume":"131","author":"J Gao","year":"2023","unstructured":"Gao, J., Pan, J. H., Zhang, S. J., & Zheng, W. S. (2023). Automatic modelling for interactive action assessment. IJCV, 131(3), 659\u2013679.","journal-title":"IJCV"},{"key":"2146_CR12","doi-asserted-by":"crossref","unstructured":"Gattupalli, S., Ebert, D., Papakostas, M., Makedon, F., & Athitsos, V. (2017). Cognilearn: A deep learning-based interface for cognitive behavior assessment. In IUI, (pp. 577\u2013587).","DOI":"10.1145\/3025171.3025213"},{"key":"2146_CR13","unstructured":"Gorban, A., Idrees, H., Jiang, Y.G., Roshan\u00a0Zamir, A., Laptev, I., Shah, M., & Sukthankar, R. (2015). THUMOS challenge: Action recognition with a large number of classes. http:\/\/www.thumos.info\/."},{"key":"2146_CR14","doi-asserted-by":"crossref","unstructured":"Gu, C., Sun, C., Ross, D.A., Vondrick, C., Pantofaru, C., Li, Y., Vijayanarasimhan, S., Toderici, G., Ricco, S., Sukthankar, R., et\u00a0al. (2018). Ava: A video dataset of spatio-temporally localized atomic visual actions. In CVPR, (pp. 6047\u20136056).","DOI":"10.1109\/CVPR.2018.00633"},{"issue":"6","key":"2146_CR15","first-page":"2260","volume":"31","author":"H Jain","year":"2020","unstructured":"Jain, H., Harit, G., & Sharma, A. (2020). Action quality assessment using siamese network-based deep metric learning. TCSVT, 31(6), 2260\u20132273.","journal-title":"TCSVT"},{"issue":"1","key":"2146_CR16","doi-asserted-by":"publisher","first-page":"221","DOI":"10.1109\/TPAMI.2012.59","volume":"35","author":"S Ji","year":"2012","unstructured":"Ji, S., Xu, W., Yang, M., & Yu, K. (2012). 3d convolutional neural networks for human action recognition. TPAMI, 35(1), 221\u2013231.","journal-title":"TPAMI"},{"key":"2146_CR17","doi-asserted-by":"crossref","unstructured":"Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R., & Fei-Fei, L. (2014). Large-scale video classification with convolutional neural networks. In CVPR, (pp. 1725\u20131732).","DOI":"10.1109\/CVPR.2014.223"},{"key":"2146_CR18","unstructured":"Kingma, D.P., & Ba, J. (2014). Adam: A method for stochastic optimization. arXiv:1412.6980."},{"key":"2146_CR19","doi-asserted-by":"crossref","unstructured":"Kuehne, H., Jhuang, H., Garrote, E., Poggio, T., & Serre, T. (2011). Hmdb: a large video database for human motion recognition. In ICCV, (pp. 2556\u20132563).","DOI":"10.1109\/ICCV.2011.6126543"},{"key":"2146_CR20","doi-asserted-by":"crossref","unstructured":"Li, Y., Chai, X., & Chen, X. (2018). End-to-end learning for action quality assessment. In PRCM, (pp. 125\u2013134).","DOI":"10.1007\/978-3-030-00767-6_12"},{"key":"2146_CR21","doi-asserted-by":"crossref","unstructured":"Li, Y., Chen, L., He, R., Wang, Z., Wu, G., & Wang, L. (2021). Multisports: A multi-person video dataset of spatio-temporally localized sports actions. In ICCV, (pp. 13536\u201313545).","DOI":"10.1109\/ICCV48922.2021.01328"},{"key":"2146_CR22","doi-asserted-by":"crossref","unstructured":"Li, H., Chen, J., Hu, R., Yu, M., Chen, H., & Xu, Z. (2019). Action recognition using visual attention with reinforcement learning. In ICMM, (pp. 365\u2013376).","DOI":"10.1007\/978-3-030-05716-9_30"},{"key":"2146_CR23","doi-asserted-by":"crossref","unstructured":"Li, Y., Li, Y., & Vasconcelos, N. (2018). Resound: Towards action recognition without representation bias. In ECCV, (pp. 513\u2013528).","DOI":"10.1007\/978-3-030-01231-1_32"},{"key":"2146_CR24","doi-asserted-by":"crossref","unstructured":"Lin, T., Liu, X., Li, X., Ding, E., & Wen, S. (2019). Bmn: Boundary-matching network for temporal action proposal generation. In ICCV, (pp. 3889\u20133898).","DOI":"10.1109\/ICCV.2019.00399"},{"key":"2146_CR25","unstructured":"Liu, S., Liu, X., Huang, G., Feng, L., Hu, L., Jiang, D., Zhang, A., Liu, Y., & Qiao, H. (2020). Fsd-10: a dataset for competitive sports content analysis. arXiv:2002.03312."},{"key":"2146_CR26","doi-asserted-by":"crossref","unstructured":"Liu, S., Zhang, A., Li, Y., Zhou, J., Xu, L., Dong, Z., & Zhang, R. (2021). Temporal segmentation of fine-gained semantic action: A motion-centered figure skating dataset. In AAAI, (pp. 2163\u20132171).","DOI":"10.1609\/aaai.v35i3.16314"},{"key":"2146_CR27","unstructured":"Meyer, M., Baldwin, D.A., & Sage, K. (2011). Assessing young children\u2019s hierarchical action segmentation. In CogSci, (pp. 3156\u20133161)."},{"key":"2146_CR28","doi-asserted-by":"publisher","unstructured":"Monfort, M., Andonian, A., Zhou, B., Ramakrishnan, K., Bargal, S.A., Yan, T., Brown, L., Fan, Q., Gutfruend, D., Vondrick, C., et\u00a0al. (2019). Moments in time dataset: one million videos for event understanding. In TPAMI, (pp. 1\u20138). https:\/\/doi.org\/10.1109\/TPAMI.2019.2901464.","DOI":"10.1109\/TPAMI.2019.2901464"},{"key":"2146_CR29","unstructured":"Montes, A., Salvador, A., Pascual, S., & Giro-i Nieto, X. (2016). Temporal activity detection in untrimmed videos with recurrent neural networks. arXiv:1608.08128."},{"key":"2146_CR30","doi-asserted-by":"crossref","unstructured":"Niebles, J.C., Chen, C.W., & Fei-Fei, L. (2010). Modeling temporal structure of decomposable motion segments for activity classification. In ECCV, (pp. 392\u2013405).","DOI":"10.1007\/978-3-642-15552-9_29"},{"key":"2146_CR31","doi-asserted-by":"crossref","unstructured":"Oneata, D., Verbeek, J., & Schmid, C. (2013). Action and event recognition with fisher vectors on a compact feature set. In ICCV, (pp. 1817\u20131824).","DOI":"10.1109\/ICCV.2013.228"},{"key":"2146_CR32","doi-asserted-by":"crossref","unstructured":"Pan, J.H., Gao, J., & Zheng, W.S. (2019). Action assessment by joint relation graphs. In ICCV, (pp. 6331\u20136340).","DOI":"10.1109\/ICCV.2019.00643"},{"key":"2146_CR33","doi-asserted-by":"crossref","unstructured":"Parisi, G.I., Magg, S., & Wermter, S. (2016). Human motion assessment in real time using recurrent self-organization. In RO-MAN, (pp. 71\u201376).","DOI":"10.1109\/ROMAN.2016.7745093"},{"key":"2146_CR34","doi-asserted-by":"crossref","unstructured":"Parmar, P., & Morris, B. (2019). Action quality assessment across multiple actions. In WACV, (pp. 1468\u20131476).","DOI":"10.1109\/WACV.2019.00161"},{"key":"2146_CR35","doi-asserted-by":"crossref","unstructured":"Parmar, P., & Tran\u00a0Morris, B. (2017). Learning to score olympic events. In CVPRW, (pp. 20\u201328).","DOI":"10.1109\/CVPRW.2017.16"},{"key":"2146_CR36","doi-asserted-by":"crossref","unstructured":"Parmar, P., & Tran\u00a0Morris, B. (2019). What and how well you performed? a multitask learning approach to action quality assessment. In CVPR, (pp. 304\u2013313).","DOI":"10.1109\/CVPR.2019.00039"},{"key":"2146_CR37","doi-asserted-by":"crossref","unstructured":"Pirsiavash, H., Vondrick, C., & Torralba, A. (2014). Assessing the quality of actions. In ECCV, (pp. 556\u2013571).","DOI":"10.1007\/978-3-319-10599-4_36"},{"key":"2146_CR38","doi-asserted-by":"crossref","unstructured":"Safdarnejad, S.M., Liu, X., Udpa, L., Andrus, B., Wood, J., & Craven, D. (2015). Sports videos in the wild (svw): A video dataset for sports analysis. In FG, (pp. 1\u20137).","DOI":"10.1109\/FG.2015.7163105"},{"key":"2146_CR39","unstructured":"Schmidt, C.F. (1976). Understanding human action: Recognizing the plans and motives of other persons. In ACL Anthology, (pp. 196\u2013200)."},{"key":"2146_CR40","doi-asserted-by":"crossref","unstructured":"Schuldt, C., Laptev, I., & Caputo, B. (2004). Recognizing human actions: A local svm approach. In ICPR, (pp. 32\u201336).","DOI":"10.1109\/ICPR.2004.1334462"},{"key":"2146_CR41","doi-asserted-by":"crossref","unstructured":"Shao, D., Zhao, Y., Dai, B., & Lin, D. (2020). Finegym: A hierarchical video dataset for fine-grained action understanding. In CVPR, (pp. 2616\u20132625).","DOI":"10.1109\/CVPR42600.2020.00269"},{"key":"2146_CR42","doi-asserted-by":"crossref","unstructured":"Shao, D., Zhao, Y., Dai, B., & Lin, D. (2020). Intra-and inter-action understanding via temporal action parsing. In CVPR, (pp. 730\u2013739).","DOI":"10.1109\/CVPR42600.2020.00081"},{"key":"2146_CR43","unstructured":"Simonyan, K., & Zisserman, A. (2014). Two-stream convolutional networks for action recognition in videos. arXiv:1406.2199."},{"key":"2146_CR44","unstructured":"Soomro, K., Zamir, A.R., & Shah, M. (2012). Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv:1212.0402."},{"key":"2146_CR45","doi-asserted-by":"crossref","unstructured":"Tang, Y., Ding, D., Rao, Y., Zheng, Y., Zhang, D., Zhao, L., Lu, J., & Zhou, J. (2019). Coin: A large-scale dataset for comprehensive instructional video analysis. In CVPR, (pp. 1207\u20131216).","DOI":"10.1109\/CVPR.2019.00130"},{"key":"2146_CR46","doi-asserted-by":"crossref","unstructured":"Tang, Y., Ni, Z., Zhou, J., Zhang, D., Lu, J., Wu, Y., & Zhou, J. (2020). Uncertainty-aware score distribution learning for action quality assessment. In CVPR, (pp. 9839\u20139848).","DOI":"10.1109\/CVPR42600.2020.00986"},{"key":"2146_CR47","unstructured":"Tong, Z., Song, Y., Wang, J., & Wang, L. (2022). Videomae: Masked autoencoders are data-efficient learners for self-supervised video pre-training. arXiv:2203.12602."},{"key":"2146_CR48","doi-asserted-by":"crossref","unstructured":"Tran, D., Bourdev, L., Fergus, R., Torresani, L., & Paluri, M. (2015). Learning spatiotemporal features with 3d convolutional networks. In ICCV, (pp. 4489\u20134497).","DOI":"10.1109\/ICCV.2015.510"},{"issue":"6","key":"2146_CR49","doi-asserted-by":"publisher","first-page":"1510","DOI":"10.1109\/TPAMI.2017.2712608","volume":"40","author":"G Varol","year":"2017","unstructured":"Varol, G., Laptev, I., & Schmid, C. (2017). Long-term temporal convolutions for action recognition. TPAMI, 40(6), 1510\u20131517.","journal-title":"TPAMI"},{"key":"2146_CR50","doi-asserted-by":"crossref","unstructured":"Wang, H., & Schmid, C. (2013). Action recognition with improved trajectories. In ICCV, (pp. 3551\u20133558).","DOI":"10.1109\/ICCV.2013.441"},{"key":"2146_CR51","doi-asserted-by":"crossref","unstructured":"Wang, X., Girshick, R., Gupta, A., & He, K. (2018). Non-local neural networks. In CVPR, (pp. 7794\u20137803).","DOI":"10.1109\/CVPR.2018.00813"},{"key":"2146_CR52","doi-asserted-by":"crossref","unstructured":"Wang, L., Xiong, Y., Wang, Z., Qiao, Y., Lin, D., Tang, X., & Van\u00a0Gool, L. (2016). Temporal segment networks: Towards good practices for deep action recognition. In ECCV, (pp. 20\u201336).","DOI":"10.1007\/978-3-319-46484-8_2"},{"key":"2146_CR53","doi-asserted-by":"crossref","unstructured":"Wang, S., Yang, D., Zhai, P., Chen, C., & Zhang, L. (2021). Tsa-net: Tube self-attention network for action quality assessment. In ACM-MM, (pp. 4902\u20134910).","DOI":"10.1145\/3474085.3475438"},{"key":"2146_CR54","doi-asserted-by":"crossref","unstructured":"Xiang, X., Tian, Y., Reiter, A., Hager, G.D., & Tran, T.D. (2018). S3d: Stacking segmental p3d for action quality assessment. In ICIP, (pp. 928\u2013932).","DOI":"10.1109\/ICIP.2018.8451364"},{"key":"2146_CR55","doi-asserted-by":"crossref","unstructured":"Xu, J., Rao, Y., Yu, X., Chen, G., Zhou, J., & Lu, J. (2022). Finediving: A fine-grained dataset for procedure-aware action quality assessment. In CVPR, (pp. 2949\u20132958).","DOI":"10.1109\/CVPR52688.2022.00296"},{"key":"2146_CR56","doi-asserted-by":"crossref","unstructured":"Xu, A., Zeng, L.A., & Zheng, W.S. (2022). Likert scoring with grade decoupling for long-term action assessment. In CVPR, (pp. 3232\u20133241).","DOI":"10.1109\/CVPR52688.2022.00323"},{"issue":"12","key":"2146_CR57","first-page":"4578","volume":"30","author":"C Xu","year":"2019","unstructured":"Xu, C., Fu, Y., Zhang, B., Chen, Z., Jiang, Y. G., & Xue, X. (2019). Learning to score figure skating sport videos. TCSVT, 30(12), 4578\u20134590.","journal-title":"TCSVT"},{"key":"2146_CR58","doi-asserted-by":"crossref","unstructured":"Yan, S., Xiong, Y., & Lin, D. (2018). Spatial temporal graph convolutional networks for skeleton-based action recognition. In AAAI, 32.","DOI":"10.1609\/aaai.v32i1.12328"},{"key":"2146_CR59","doi-asserted-by":"crossref","unstructured":"Yang, C., Xu, Y., Shi, J., Dai, B., & Zhou, B. (2020). Temporal pyramid network for action recognition. In CVPR, (pp. 591\u2013600).","DOI":"10.1109\/CVPR42600.2020.00067"},{"key":"2146_CR60","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3132058","author":"L Yang","year":"2021","unstructured":"Yang, L., Han, J., Zhao, T., Lin, T., Zhang, D., & Chen, J. (2021). Background-click supervision for temporal action localization. TPAMI. https:\/\/doi.org\/10.1109\/TPAMI.2021.3132058","journal-title":"TPAMI"},{"key":"2146_CR61","doi-asserted-by":"crossref","unstructured":"Yeung, S., Russakovsky, O., Mori, G., & Fei-Fei, L. (2016). End-to-end learning of action detection from frame glimpses in videos. In CVPR, (pp. 2678\u20132687).","DOI":"10.1109\/CVPR.2016.293"},{"key":"2146_CR62","doi-asserted-by":"crossref","unstructured":"Yu, X., Rao, Y., Zhao, W., Lu, J., & Zhou, J. (2021). Group-aware contrastive regression for action quality assessment. In ICCV, (pp. 7919\u20137928).","DOI":"10.1109\/ICCV48922.2021.00782"},{"key":"2146_CR63","doi-asserted-by":"crossref","unstructured":"Zeng, L.A., Hong, F.T., Zheng, W.S., Yu, Q.Z., Zeng, W., Wang, Y.W., & Lai, J.H. (2020). Hybrid dynamic-static context-aware attention network for action assessment in long videos. In ACM MM, (pp. 2526\u20132534).","DOI":"10.1145\/3394171.3413560"},{"key":"2146_CR64","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3090167","author":"R Zeng","year":"2021","unstructured":"Zeng, R., Huang, W., Tan, M., Rong, Y., Zhao, P., Huang, J., & Gan, C. (2021). Graph convolutional module for temporal action localization in videos. TPAMI. https:\/\/doi.org\/10.1109\/TPAMI.2021.3090167","journal-title":"TPAMI"},{"key":"2146_CR65","unstructured":"Zhang, B., Chen, J., Xu, Y., Zhang, H., Yang, X., & Geng, X. (2021). Auto-encoding score distribution regression for action quality assessment. arXiv:2111.11029."},{"key":"2146_CR66","doi-asserted-by":"crossref","unstructured":"Zhang, S., Dai, W., Wang, S., Shen, X., Lu, J., Zhou, J., & Tang, Y. (2023). Logo: A long-form video dataset for group action quality assessment. In CVPR, (pp. 2405\u20132414).","DOI":"10.1109\/CVPR52729.2023.00238"},{"key":"2146_CR67","doi-asserted-by":"crossref","unstructured":"Zhang, C., Gupta, A., & Zisserman, A. (2021). Temporal query networks for fine-grained video understanding. In CVPR, (pp. 4486\u20134496).","DOI":"10.1109\/CVPR46437.2021.00446"},{"issue":"6","key":"2146_CR68","doi-asserted-by":"publisher","first-page":"1206","DOI":"10.1109\/TPAMI.2014.2361121","volume":"37","author":"Q Zhang","year":"2014","unstructured":"Zhang, Q., & Li, B. (2014). Relative hidden markov models for video-based evaluation of motion skills in surgical training. TPAMI, 37(6), 1206\u20131218.","journal-title":"TPAMI"},{"key":"2146_CR69","doi-asserted-by":"crossref","unstructured":"Zhao, H., Torralba, A., Torresani, L., & Yan, Z. (2019). Hacs: Human action clips and segments dataset for recognition and temporal localization. In ICCV, (pp. 8668\u20138678).","DOI":"10.1109\/ICCV.2019.00876"},{"key":"2146_CR70","doi-asserted-by":"crossref","unstructured":"Zhao, Y., Xiong, Y., Wang, L., Wu, Z., Tang, X., & Lin, D. (2017). Temporal action detection with structured segment networks. In ICCV, (pp. 2914\u20132923).","DOI":"10.1109\/ICCV.2017.317"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02146-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-024-02146-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-024-02146-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,15]],"date-time":"2024-11-15T10:27:58Z","timestamp":1731666478000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-024-02146-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,7,14]]},"references-count":70,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2024,12]]}},"alternative-id":["2146"],"URL":"https:\/\/doi.org\/10.1007\/s11263-024-02146-z","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,7,14]]},"assertion":[{"value":"14 June 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 June 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 July 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}