{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T09:09:55Z","timestamp":1774602595497,"version":"3.50.1"},"reference-count":60,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T00:00:00Z","timestamp":1770336000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T00:00:00Z","timestamp":1770336000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100021856","name":"Ministero dell\u2019Universit\u00e0 e della Ricerca","doi-asserted-by":"publisher","award":["CUP E63C22001940006"],"award-info":[{"award-number":["CUP E63C22001940006"]}],"id":[{"id":"10.13039\/501100021856","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1007\/s11263-025-02675-1","type":"journal-article","created":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T07:11:41Z","timestamp":1770361901000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Exocentric-to-Egocentric Adaptation for Temporal Action Segmentation with Unlabeled Synchronized Video Pairs"],"prefix":"10.1007","volume":"134","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4999-8698","authenticated-orcid":false,"given":"Camillo","family":"Quattrocchi","sequence":"first","affiliation":[]},{"given":"Antonino","family":"Furnari","sequence":"additional","affiliation":[]},{"given":"Daniele Di","family":"Mauro","sequence":"additional","affiliation":[]},{"given":"Mario Valerio","family":"Giuffrida","sequence":"additional","affiliation":[]},{"given":"Giovanni Maria","family":"Farinella","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,2,6]]},"reference":[{"key":"2675_CR1","doi-asserted-by":"crossref","unstructured":"Camporese, G., Coscia, P., Furnari, A., Farinella, G.M., & Ballan, L. (2021).Knowledge distillation for action anticipation via label smoothing. In 2020 25th International Conference on Pattern Recognition (ICPR) (pp. 3312\u20133319). IEEE.","DOI":"10.1109\/ICPR48806.2021.9412660"},{"key":"2675_CR2","doi-asserted-by":"crossref","unstructured":"Chen, M.H., Kira, Z., AlRegib, G., Yoo, J., Chen, R., & Zheng, J. (2019) Temporal attentive alignment for large-scale video domain adaptation. In Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV)","DOI":"10.1109\/ICCV.2019.00642"},{"key":"2675_CR3","doi-asserted-by":"crossref","unstructured":"Chen, M.H., Li, B., Bao, Y., & AlRegib, G. (2020). Action segmentation with mixed temporal domain adaptation. In Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision pp. 605\u2013614.","DOI":"10.1109\/WACV45572.2020.9093535"},{"key":"2675_CR4","doi-asserted-by":"crossref","unstructured":"Chen, X., Qin, S., Bai, Z., Guo, Y., Yang, Z., Jiang, R., Wu, C., & Su, S. (2022). Long video-based action segmentation for earthmoving excavators using improved Temporal Convolutional Network models. In IOP Conference Series: Earth and Environmental Science (Vol. 1101, pp. 092003). IOP Publishing.","DOI":"10.1088\/1755-1315\/1101\/9\/092003"},{"key":"2675_CR5","doi-asserted-by":"crossref","unstructured":"Choi, J., Sharma, G., Chandraker, M., & Huang, J.B. (2020). Unsupervised and semi-supervised domain adaptation for action recognition from drones. In Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision pp. 1717\u20131726.","DOI":"10.1109\/WACV45572.2020.9093511"},{"key":"2675_CR6","doi-asserted-by":"crossref","unstructured":"Choi, J., Sharma, G., Schulter, S., Huang, J.B. (2020). Shuffle and attend: Video domain adaptation. In Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XII 16 (pp. 678\u2013695). Springer","DOI":"10.1007\/978-3-030-58610-2_40"},{"key":"2675_CR7","doi-asserted-by":"crossref","unstructured":"Crasto, N., Weinzaepfel, P., Alahari, K., & Schmid, C. (2019). Mars: Motion-augmented rgb stream for action recognition. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition pp. 7882\u20137891.","DOI":"10.1109\/CVPR.2019.00807"},{"key":"2675_CR8","doi-asserted-by":"crossref","unstructured":"Csurka, G. (2017). Domain adaptation for visual applications: A comprehensive survey. CoRR abs\/1702.05374 arxiv:1702.05374","DOI":"10.1007\/978-3-319-58347-1_1"},{"key":"2675_CR9","unstructured":"Damen, D., Doughty, H., Farinella, G.M., Furnari, A., Ma, J., Kazakos, E., Moltisanti, D., Munro, J., Perrett, T., Price, W., & Wray, M. (2020). Rescaling egocentric vision. CoRR abs\/2006.13256."},{"key":"2675_CR10","unstructured":"Ding, G., Sener, F., Yao, A. (2022). Temporal action segmentation: An analysis of modern technique. arXiv preprint arXiv:2210.10352"},{"key":"2675_CR11","doi-asserted-by":"crossref","unstructured":"Dwibedi, D., Aytar, Y., Tompson, J., Sermanet, P., & Zisserman, A. (2019). Temporal cycle-consistency learning. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition pp. 1801\u20131810.","DOI":"10.1109\/CVPR.2019.00190"},{"key":"2675_CR12","doi-asserted-by":"crossref","unstructured":"Fernando, B., & Herath, S. (2021). Anticipating human actions by correlating past with the future with jaccard similarity measures. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition pp. 13224\u201313233.","DOI":"10.1109\/CVPR46437.2021.01302"},{"key":"2675_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.cviu.2023.103763","volume":"234","author":"A Furnari","year":"2023","unstructured":"Furnari, A., & Farinella, G. M. (2023). Streaming egocentric action anticipation: An evaluation scheme and approach. Computer Vision and Image Understanding, 234, Article 103763.","journal-title":"Computer Vision and Image Understanding"},{"key":"2675_CR14","unstructured":"Ganin, Y., & Lempitsky, V. (2015). Unsupervised domain adaptation by backpropagation. In International conference on machine learning (pp. 1180\u20131189). PMLR."},{"key":"2675_CR15","doi-asserted-by":"crossref","unstructured":"Girdhar, R., Singh, M., Ravi, N., van\u00a0der Maaten, L., Joulin, A., & Misra, I. (2022). Omnivore: A single model for many visual modalities. In CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01563"},{"key":"2675_CR16","doi-asserted-by":"crossref","unstructured":"Grauman, K., Westbury, A., Torresani, L., Kitani, K., Malik, J., Afouras, T., Ashutosh, K., Baiyya, V., Bansal, S., & Boote, B. et\u00a0al. (2024). Ego-exo4d: Understanding skilled human activity from first-and third-person perspectives. In CVPR","DOI":"10.1109\/CVPR52733.2024.01834"},{"issue":"1","key":"2675_CR17","first-page":"723","volume":"13","author":"A Gretton","year":"2012","unstructured":"Gretton, A., Borgwardt, K. M., Rasch, M. J., Sch\u00f6lkopf, B., & Smola, A. (2012). A kernel two-sample test. The Journal of Machine Learning Research, 13(1), 723\u2013773.","journal-title":"The Journal of Machine Learning Research"},{"key":"2675_CR18","doi-asserted-by":"crossref","unstructured":"Hadji, I., Derpanis, K.G., & Jepson, A.D. (2021). Representation learning via global temporal alignment and cycle-consistency. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 11068\u201311077.","DOI":"10.1109\/CVPR46437.2021.01092"},{"key":"2675_CR19","unstructured":"Hinton, G., Vinyals, O., & Dean, J. (2015). Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531"},{"key":"2675_CR20","unstructured":"Hodgins, F., & Macey, J. (2009). Guide to the carnegie mellon university multimodal activity (cmu-mmac) database. CMU-RI-TR-08-22"},{"key":"2675_CR21","unstructured":"Huang, Z. & Wang, N. (2017). Like what you like: Knowledge distill via neuron selectivity transfer."},{"key":"2675_CR22","doi-asserted-by":"crossref","unstructured":"Kim, D., Tsai, Y.H., Zhuang, B., Yu, X., Sclaroff, S., Saenko, K., & Chandraker, M. (2021). Learning cross-modal contrastive features for video domain adaptation. In Proceedings of the IEEE\/CVF International Conference on Computer Vision (ICCV) pp. 13618\u201313627.","DOI":"10.1109\/ICCV48922.2021.01336"},{"key":"2675_CR23","doi-asserted-by":"crossref","unstructured":"Kim, D., Tsai, Y.H., Zhuang, B., Yu, X., Sclaroff, S., Saenko, K., & Chandraker, M. (2021). Learning cross-modal contrastive features for video domain adaptation. In Proceedings of the IEEE\/CVF International Conference on Computer Vision pp. 13618\u201313627.","DOI":"10.1109\/ICCV48922.2021.01336"},{"key":"2675_CR24","unstructured":"Kingma, D.P. (2014). Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980"},{"issue":"4","key":"2675_CR25","doi-asserted-by":"publisher","first-page":"765","DOI":"10.1109\/TPAMI.2018.2884469","volume":"42","author":"H Kuehne","year":"2018","unstructured":"Kuehne, H., Richard, A., & Gall, J. (2018). A hybrid rnn-hmm approach for weakly supervised temporal action segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(4), 765\u2013779.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"issue":"4","key":"2675_CR26","doi-asserted-by":"publisher","first-page":"765","DOI":"10.1109\/TPAMI.2018.2884469","volume":"42","author":"H Kuehne","year":"2020","unstructured":"Kuehne, H., Richard, A., & Gall, J. (2020). A hybrid rnn-hmm approach for weakly supervised temporal action segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 42(4), 765\u2013779. https:\/\/doi.org\/10.1109\/TPAMI.2018.2884469","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2675_CR27","doi-asserted-by":"crossref","unstructured":"Lea, C., Flynn, MD., Vidal, R., Reiter, A., & Hager, G.D. (2017). Temporal convolutional networks for action segmentation and detection. In proceedings of the IEEE Conference on Computer Vision and Pattern Recognition pp. 156\u2013165.","DOI":"10.1109\/CVPR.2017.113"},{"key":"2675_CR28","doi-asserted-by":"crossref","unstructured":"Li, Z., Abu\u00a0Farha, Y., & Gall, J. (2021). Temporal action segmentation from timestamp supervision. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 8365\u20138374.","DOI":"10.1109\/CVPR46437.2021.00826"},{"key":"2675_CR29","unstructured":"Li, S.J., AbuFarha, Y., Liu, Y., Cheng, M.M., & Gall, J. (2020). Ms-tcn++: Multi-stage temporal convolutional network for action segmentation. IEEE transactions on pattern analysis and machine intelligence"},{"key":"2675_CR30","doi-asserted-by":"crossref","unstructured":"Li, G., Jampani, V., Sun, D., Sevilla-Lara, L. (2023). LOCATE: Localize and transfer object parts for weakly supervised affordance grounding. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 10922\u201310931.","DOI":"10.1109\/CVPR52729.2023.01051"},{"key":"2675_CR31","doi-asserted-by":"crossref","unstructured":"Li, Y., Liu, M., Rehg, J.M. (2018). In the eye of beholder: Joint learning of gaze and actions in first person video. In Proceedings of the European conference on computer vision (ECCV). pp. 619\u2013635.","DOI":"10.1007\/978-3-030-01228-1_38"},{"key":"2675_CR32","doi-asserted-by":"crossref","unstructured":"Li, Y., Nagarajan, T., Xiong, B., & Grauman, K. (2021). Ego-Exo: Transferring visual representations from third-person to first-person videos. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) pp. 6943\u20136953.","DOI":"10.1109\/CVPR46437.2021.00687"},{"key":"2675_CR33","doi-asserted-by":"crossref","unstructured":"Lin, J., Gan, C., & Han, S. (2019). TSM: Temporal shift module for efficient video understanding. In Proceedings of the IEEE International Conference on Computer Vision","DOI":"10.1109\/ICCV.2019.00718"},{"key":"2675_CR34","doi-asserted-by":"crossref","unstructured":"Liu, M., Chen, X., Zhang,Y., Li, Y., & Rehg, J.M. (2019). Attention distillation for learning video representations. arXiv preprint arXiv:1904.03249","DOI":"10.5244\/C.34.1"},{"key":"2675_CR35","doi-asserted-by":"crossref","unstructured":"Liu, D., Li, Q., Dinh, A., Jiang, T., Shah, M., & Xu, C. (2023). Diffusion action segmentation. arXiv preprint arXiv:2303.17959.","DOI":"10.1109\/ICCV51070.2023.00930"},{"key":"2675_CR36","doi-asserted-by":"publisher","unstructured":"Liu, X., Zhou, S., Lei, T., Jiang, P., Chen, Z., & Lu, H. (2023). First-person video domain adaptation with multi-scene cross-site datasets and attention-based methods. IEEE Transactions on Circuits and Systems for Video Technology pp. 1\u20131. https:\/\/doi.org\/10.1109\/TCSVT.2023.3281671","DOI":"10.1109\/TCSVT.2023.3281671"},{"key":"2675_CR37","doi-asserted-by":"crossref","unstructured":"Munro, J., Damen, D. (2020). Multi-modal domain adaptation for fine-grained action recognition. In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition pp. 122\u2013132.","DOI":"10.1109\/CVPR42600.2020.00020"},{"key":"2675_CR38","unstructured":"Oquab, M., Darcet, T., Moutakanni, T., Vo, H., Szafraniec, M., Khalidov, V., Fernandez, P., Haziza, D., Massa, F., & El-Nouby, A., et\u00a0al. (2023). Dinov2: Learning robust visual features without supervision. arXiv preprint arXiv:2304.07193"},{"key":"2675_CR39","doi-asserted-by":"crossref","unstructured":"Passalis, N., Tzelepi, M., & Tefas, A. (2020). Heterogeneous knowledge distillation using information flow modeling. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 2339\u20132348.","DOI":"10.1109\/CVPR42600.2020.00241"},{"key":"2675_CR40","doi-asserted-by":"crossref","unstructured":"Quattrocchi, C., Furnari, A., Di\u00a0Mauro, D., Giuffrida, M.V., & Farinella, G.M. (2023). Synchronization is all you need: Exocentric-to-egocentric transfer for temporal action segmentation with unlabeled synchronized video pairs. arXiv preprint arXiv:2312.02638.","DOI":"10.1007\/978-3-031-73220-1_15"},{"key":"2675_CR41","unstructured":"Reza, S., Sundareshan, B., Moghaddam, M., & Camps, O. (2023). Enhancing transformer backbone for egocentric video action segmentation. arXiv preprint arXiv:2305.11365"},{"key":"2675_CR42","unstructured":"Romero, A., Ballas, N., Kahou, S.E., Chassang, A., Gatta, C., & Bengio, Y. (2014). Fitnets: Hints for thin deep nets. arXiv preprint arXiv:1412.6550"},{"key":"2675_CR43","doi-asserted-by":"crossref","unstructured":"Sayed, S., Ghoddoosian, R., Trivedi, B., & Athitsos, V. (2023). A new dataset and approach for timestamp supervised action segmentation using human object interaction. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 3132\u20133141.","DOI":"10.1109\/CVPRW59228.2023.00315"},{"key":"2675_CR44","doi-asserted-by":"crossref","unstructured":"Sener, F., Chatterjee, D., Shelepov, D., He, K., Singhania, D., Wang, R., & Yao, A. (2022). Assembly101: A large-scale multi-view video dataset for understanding procedural activities. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 21096\u201321106.","DOI":"10.1109\/CVPR52688.2022.02042"},{"key":"2675_CR45","doi-asserted-by":"crossref","unstructured":"Sener, F., Singhania, D., Yao, A. (2020). Temporal aggregate representations for long-range video understanding. In Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XVI 16 (pp. 154\u2013171) Springer.","DOI":"10.1007\/978-3-030-58517-4_10"},{"key":"2675_CR46","doi-asserted-by":"crossref","unstructured":"Sigurdsson, G.A., Gupta, A., Schmid, C., Farhadi, A., Alahari, K. (2018). Actor and observer: Joint modeling of first and third-person videos. In proceedings of the IEEE conference on computer vision and pattern recognition pp. 7396\u20137404.","DOI":"10.1109\/CVPR.2018.00772"},{"key":"2675_CR47","doi-asserted-by":"crossref","unstructured":"Singhania, D., Rahaman, R., & Yao, A. (2022). Iterative contrast-classify for semi-supervised temporal action segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 36, pp. 2262\u20132270).","DOI":"10.1609\/aaai.v36i2.20124"},{"key":"2675_CR48","unstructured":"Singhania, D., Rahaman, R., Yao, A. (2021). Coarse to fine multi-resolution temporal convolutional network. arXiv preprint arXiv:2105.10859"},{"key":"2675_CR49","unstructured":"Somasundaram, K., Dong, J., Tang, H., Straub, J., Yan, M., Goesele, M., Engel, J.J., De\u00a0Nardi, R., & Newcombe, R. (2023). Project aria: A new tool for egocentric multi-modal ai research. arXiv preprint arXiv:2308.13561"},{"key":"2675_CR50","doi-asserted-by":"publisher","unstructured":"Spriggs, E.H., De\u00a0La\u00a0Torre, F., & Hebert, M. (2009). Temporal segmentation and activity classification from first-person sensing. In 2009 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops pp. 17\u201324https:\/\/doi.org\/10.1109\/CVPRW.2009.5204354","DOI":"10.1109\/CVPRW.2009.5204354"},{"key":"2675_CR51","doi-asserted-by":"crossref","unstructured":"Stroud, J., Ross, D., Sun, C., Deng, J. & Sukthankar, R. (2020). D3d: Distilled 3d networks for video action recognition. In Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision pp. 625\u2013634.","DOI":"10.1109\/WACV45572.2020.9093274"},{"key":"2675_CR52","doi-asserted-by":"crossref","unstructured":"Sun, B., & Saenko, K. (2016). Deep coral: Correlation alignment for deep domain adaptation. In Computer Vision\u2013ECCV 2016 Workshops: Amsterdam, The Netherlands, October 8-10 and 15-16, 2016, Proceedings, Part III 14 (pp. 443\u2013450). Springer.","DOI":"10.1007\/978-3-319-49409-8_35"},{"key":"2675_CR53","doi-asserted-by":"crossref","unstructured":"Tran, V., Wang, Y., Zhang, Z., & Hoai, M. (2021). Knowledge distillation for human action anticipation. In 2021 IEEE International Conference on Image Processing (ICIP) (pp. 2518\u20132522). IEEE.","DOI":"10.1109\/ICIP42928.2021.9506693"},{"key":"2675_CR54","doi-asserted-by":"crossref","unstructured":"Wang, Z., Gao, Z., Wang, L., Li, Z., Wu, G. (2020). Boundary-aware cascade networks for temporal action segmentation. In Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part XXV 16 (pp. 34\u201351) Springer.","DOI":"10.1007\/978-3-030-58595-2_3"},{"key":"2675_CR55","doi-asserted-by":"crossref","unstructured":"Wang, X., Hu, J.F., Lai, J.H., Zhang, J., & Zheng, W.S. (2019). Progressive teacher-student learning for early action prediction. In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition pp. 3556\u20133565.","DOI":"10.1109\/CVPR.2019.00367"},{"issue":"5","key":"2675_CR56","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3400066","volume":"11","author":"G Wilson","year":"2020","unstructured":"Wilson, G., & Cook, D. J. (2020). A survey of unsupervised deep domain adaptation. ACM Transactions on Intelligent Systems and Technology (TIST), 11(5), 1\u201346.","journal-title":"ACM Transactions on Intelligent Systems and Technology (TIST)"},{"key":"2675_CR57","doi-asserted-by":"crossref","unstructured":"Xue, Z., & Grauman, K. (2023). Learning fine-grained view-invariant representations from unpaired ego-exo videos via temporal alignment. arXiv preprint arXiv:2306.05526","DOI":"10.52202\/075280-2336"},{"key":"2675_CR58","first-page":"14890","volume":"35","author":"Z Xu","year":"2022","unstructured":"Xu, Z., Rawat, Y., Wong, Y., Kankanhalli, M. S., & Shah, M. (2022). Don\u2019t pour cereal into coffee: Differentiable temporal logic for temporal action segmentation. Advances in Neural Information Processing Systems, 35, 14890\u201314903.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2675_CR59","doi-asserted-by":"crossref","unstructured":"Yim, J., Joo, D., Bae, J., & Kim, J. (2017). A gift from knowledge distillation: Fast optimization, network minimization and transfer learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition pp. 4133\u20134141","DOI":"10.1109\/CVPR.2017.754"},{"key":"2675_CR60","doi-asserted-by":"crossref","unstructured":"Yu, H., Cai, M., Liu, Y., Lu, F. (2019). What i see is what you see: Joint attention learning for first and third person video co-analysis. In Proceedings of the 27th ACM International Conference on Multimedia pp. 1358\u20131366.","DOI":"10.1145\/3343031.3350896"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02675-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02675-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02675-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T08:32:13Z","timestamp":1774600333000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02675-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,2,6]]},"references-count":60,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["2675"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02675-1","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,2,6]]},"assertion":[{"value":"23 January 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"23 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 February 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"112"}}