{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T04:06:31Z","timestamp":1775534791588,"version":"3.50.1"},"reference-count":64,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T00:00:00Z","timestamp":1775520000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T00:00:00Z","timestamp":1775520000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2026,5]]},"DOI":"10.1007\/s11263-026-02790-7","type":"journal-article","created":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T03:22:04Z","timestamp":1775532124000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Distractor-Aware Memory-Based Visual Object Tracking"],"prefix":"10.1007","volume":"134","author":[{"ORCID":"https:\/\/orcid.org\/0009-0005-8614-3305","authenticated-orcid":false,"given":"Jovana","family":"Videnovi\u0107","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4252-4342","authenticated-orcid":false,"given":"Matej","family":"Kristan","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-6316-2707","authenticated-orcid":false,"given":"Alan","family":"Luke\u017ei\u010d","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,4,7]]},"reference":[{"key":"2790_CR1","unstructured":"Adamyan, A., \u010c\u00ed\u017eek, T., Straka, M., Janouskova, K., & Schmid, M. (2025). Sam2rl: Towards reinforcement learning memory control in segment anything model 2. arXiv preprint arXiv:2507.08548."},{"key":"2790_CR2","doi-asserted-by":"crossref","unstructured":"Alawode, B., Guo, Y., Ummar, M., Werghi, N., Dias, J., Mian, A., & Javed, S. (2022). Utb180: A high-quality benchmark for underwater tracking. In Proceedings asian conference computer vision, pp 3326\u20133342.","DOI":"10.1007\/978-3-031-26348-4_26"},{"key":"2790_CR3","doi-asserted-by":"crossref","unstructured":"Bertinetto, L., Valmadre, J., Henriques, J. F., Vedaldi, A., & Torr, P. H. (2016). Fully-convolutional siamese networks for object tracking. In Proceedings european conference computer vision workshops, pp 850\u2013865.","DOI":"10.1007\/978-3-319-48881-3_56"},{"key":"2790_CR4","unstructured":"Carion, N., Gustafson, L., Hu, Y. T., Debnath, S., Hu, R., Suris, D., Ryali, C., Alwala, K. V., Khedr, H., Huang, A., et al. (2025). SAM 3: Segment anything with concepts. In arXiv preprint arXiv:2511.16719."},{"key":"2790_CR5","doi-asserted-by":"crossref","unstructured":"Chen, B., Li, P., Bai, L., Qiao, L., Shen, Q., Li, B., Gan, W., Wu, W., & Ouyang, W. (2022). Backbone is all your need: A simplified architecture for visual object tracking. In Proceedings european conference computer vision, pp 375\u2013392.","DOI":"10.1007\/978-3-031-20047-2_22"},{"key":"2790_CR6","doi-asserted-by":"crossref","unstructured":"Chen, X., Yan, B., Zhu, J., Wang, D., Yang, X., & Lu, H. (2021). Transformer tracking. In Computer vision pattern recognition, pp 8126\u20138135.","DOI":"10.1109\/CVPR46437.2021.00803"},{"key":"2790_CR7","doi-asserted-by":"crossref","unstructured":"Chen, X., Peng, H., Wang, D., Lu, H., & Hu, H. (2023). Seqtrack: Sequence to sequence learning for visual object tracking. In Computer vision pattern recognition, pp 14572\u201314581.","DOI":"10.1109\/CVPR52729.2023.01400"},{"key":"2790_CR8","doi-asserted-by":"crossref","unstructured":"Cheng, H. K., & Schwing, A. G. (2022). Xmem: Long-term video object segmentation with an atkinson-shiffrin memory model. In Proceedings european conference computer, pp 640\u2013658.","DOI":"10.1007\/978-3-031-19815-1_37"},{"key":"2790_CR9","doi-asserted-by":"crossref","unstructured":"Cheng, H. K., Oh, S. W., Price, B., Lee, J. Y., & Schwing, A. (2024). Putting the object back into video object segmentation. In Computer vision pattern recognition, pp 3151\u20133161.","DOI":"10.1109\/CVPR52733.2024.00304"},{"key":"2790_CR10","doi-asserted-by":"crossref","unstructured":"Cui, Y., Jiang, C., Wang, L., & Wu, G. (2022). Mixformer: End-to-end tracking with iterative mixed attention. In Computer vision pattern recognition, pp 13608\u201313618.","DOI":"10.1109\/CVPR52688.2022.01324"},{"issue":"6","key":"2790_CR11","doi-asserted-by":"publisher","first-page":"4129","DOI":"10.1109\/TPAMI.2024.3349519","volume":"46","author":"Y Cui","year":"2024","unstructured":"Cui, Y., Jiang, C., Wu, G., et al. (2024). Mixformer: End-to-end tracking with iterative mixed attention. IEEE Transactions on Pattern Analysis and Machine Intelligence, 46(6), 4129\u20134146.","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"2790_CR12","doi-asserted-by":"crossref","unstructured":"Cuttano, C., Trivigno, G., Rosi, G., Masone, C., & Averta, G. (2025). Samwise: Infusing wisdom in sam2 for text-driven video segmentation. In Computer vision pattern recognition.","DOI":"10.1109\/CVPR52734.2025.00322"},{"key":"2790_CR13","doi-asserted-by":"crossref","unstructured":"Danelljan, M., Bhat, G., Khan, F. S., & Felsberg, M. (2019). ATOM: Accurate tracking by overlap maximization. In Computer vision pattern recognition.","DOI":"10.1109\/CVPR.2019.00479"},{"key":"2790_CR14","doi-asserted-by":"crossref","unstructured":"Ding, H., Liu, C., He, S., Jiang, X., Torr, P. H., & Bai, S. (2023). Mose: A new dataset for video object segmentation in complex scenes. In Computer vision pattern recognition, pp 20224\u201320234.","DOI":"10.1109\/ICCV51070.2023.01850"},{"key":"2790_CR15","unstructured":"Ding, S., Qian, R., Dong, X., Zhang, P., Zang, Y., Cao, Y., Guo, Y., Lin, D., & Wang, J. (2024). Sam2long: Enhancing sam 2 for long video segmentation with a training-free memory tree. arXiv preprint arXiv:2410.16268."},{"key":"2790_CR16","doi-asserted-by":"crossref","unstructured":"Fan, H., Lin, L., Yang, F., Chu, P., Deng, G., Yu, S., Bai, H., Xu, Y., Liao, C., & Ling, H. (2019). Lasot: A high-quality benchmark for large-scale single object tracking. In Computer vision pattern recognition.","DOI":"10.1109\/CVPR.2019.00552"},{"key":"2790_CR17","doi-asserted-by":"crossref","unstructured":"Fan, H., Bai, H., Lin, L., Yang, F., Chu, P., Deng, G., Yu, S., Harshit, Huang, M., Liu, J., et al. (2021). Lasot: A high-quality large-scale single object tracking benchmark. In International journal of computer vision, pp 439\u2013461.","DOI":"10.1007\/s11263-020-01387-y"},{"key":"2790_CR18","doi-asserted-by":"crossref","unstructured":"Gao, Y., Xu, H., Li, J., Wang, N., & Gao, X. (2024). Multi-scene generalized trajectory global graph solver with composite nodes for multiple object tracking. In Proceedings of the AAAI conference on artificial intelligence.","DOI":"10.1609\/aaai.v38i3.27953"},{"key":"2790_CR19","doi-asserted-by":"crossref","unstructured":"Guo, D., Wang, J., Cui, Y., Wang, Z., & Chen, S. (2020). Siamcar: Siamese fully convolutional classification and regression for visual tracking. In Computer vision pattern recognition.","DOI":"10.1109\/CVPR42600.2020.00630"},{"key":"2790_CR20","doi-asserted-by":"crossref","unstructured":"Henriques, J. F., Caseiro, R., Martins, P., & Batista, J. (2014). High-speed tracking with kernelized correlation filters. In The IEEE transactions on pattern analysis and machine intelligence, pp 583\u2013596.","DOI":"10.1109\/TPAMI.2014.2345390"},{"key":"2790_CR21","doi-asserted-by":"crossref","unstructured":"Hong, L., Chen, W., Liu, Z., Zhang, W., Guo, P., Chen, Z., & Zhang, W. (2023). Lvos: A benchmark for long-term video object segmentation. In International conference computer vision, pp 13480\u201313492.","DOI":"10.1109\/ICCV51070.2023.01240"},{"key":"2790_CR22","unstructured":"Hong, L., Liu, Z., Chen, W., Tan, C., Feng, Y., Zhou, X., Guo, P., Li, J., Chen, Z., Gao, S., et al. (2024). Lvos: A benchmark for large-scale long-term video object segmentation. arXiv preprint arXiv:2404.19326."},{"key":"2790_CR23","doi-asserted-by":"crossref","unstructured":"Huang, L., Zhao, X., & Huang, K. (2018). Got-10k: A large high-diversity benchmark for generic object tracking in the wild. In The IEEE transactions on pattern analysis and machine intelligence, pp 1562\u20131577.","DOI":"10.1109\/TPAMI.2019.2957464"},{"key":"2790_CR24","doi-asserted-by":"crossref","unstructured":"Kiani Galoogahi, H., Fagg, A., Huang, C., Ramanan, D., & Lucey, S. (2017). Need for speed: A benchmark for higher frame rate object tracking. In International conference computer vision, pp 1125\u20131134.","DOI":"10.1109\/ICCV.2017.128"},{"key":"2790_CR25","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A., Lo, W.-Y., et al. (2023). Segment anything. In International conference computer vision, pp 4015\u20134026.","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"2790_CR26","doi-asserted-by":"crossref","unstructured":"Kristan, M., Matas, J., Leonardis, A., Voj\u00ed\u0159, T., Pflugfelder, R., Fernandez, G., Nebehay, G., Porikli, F., & \u010cehovin, L. (2016). A novel performance evaluation methodology for single-target trackers. In The IEEE transactions on pattern analysis and machine intelligence, pp 213\u2013215.","DOI":"10.1109\/TPAMI.2016.2516982"},{"key":"2790_CR27","doi-asserted-by":"crossref","unstructured":"Kristan, M., Matas, J., Leonardis, A., Felsberg, M., Pflugfelder, R., Kamarainen, J.-K., \u010cehovin Zajc, L., Drbohlav, O., Lukezic, A., Berg, A., et al. (2019). The seventh visual object tracking vot2019 challenge results. In Proceedings of the IEEE\/CVF international conference on computer vision (ICCV) workshops.","DOI":"10.1109\/ICCVW.2019.00276"},{"key":"2790_CR28","doi-asserted-by":"crossref","unstructured":"Kristan, M., Leonardis, A., Matas, J., Felsberg, M., Pflugfelder, R., K\u00e4m\u00e4r\u00e4inen, J.-K., Danelljan, M., Zajc, L. \u010c., Luke\u017ei\u010d, A., Drbohlav, O., et al. (2020). The eighth visual object tracking VOT2020 challenge results. In Proceedings european conference computer workshops, pp 547\u2013601.","DOI":"10.1007\/978-3-030-68238-5_39"},{"key":"2790_CR29","unstructured":"Kristan, M., Leonardis, A., Matas, J., Felsberg, M., Pflugfelder, R., K\u00e4m\u00e4r\u00e4inen, J.-K., Chang, H. J., Danelljan, M., Zajc, L. \u010c., Luke\u017ei\u010d, A., et al. (2022). The tenth visual object tracking vot2022 challenge results. In Proceedings european conference computer workshops, pp 431\u2013460."},{"key":"2790_CR30","doi-asserted-by":"crossref","unstructured":"Kristan, M., Matas, J., Danelljan, M., Felsberg, M., Chang, H. J., Zajc, L. \u010c., Luke\u017ei\u010d, A., Drbohlav, O., Zhang, Z., Tran, K.-T., et al. (2023). The first visual object tracking segmentation vots2023 challenge results. In International conference computer vision workshops, pp 1796\u20131818.","DOI":"10.1109\/ICCVW60793.2023.00195"},{"key":"2790_CR31","unstructured":"Kristan, M., Matas, J., Tokmakov, P., Felsberg, M., Zajc, L. \u010c., Luke\u017ei\u010d, A., Tran, K.-T., Vu, X.-S., Bj\u00f6rklund, J., Chang, H. J., et al. (2024). The second visual object tracking segmentation vots2024 challenge results."},{"key":"2790_CR32","doi-asserted-by":"crossref","unstructured":"Li, B., Yan, J., Wu, W., Zhu, Z., & Hu, X. (2018). High performance visual tracking with siamese region proposal network. In Computer vision pattern recognition.","DOI":"10.1109\/CVPR.2018.00935"},{"key":"2790_CR33","doi-asserted-by":"crossref","unstructured":"Li, B., Wu, W., Wang, Q., Zhang, F., Xing, J., & Yan, J. (2019). SiamRPN++: Evolution of siamese visual tracking with very deep networks. In Computer vision pattern recognition.","DOI":"10.1109\/CVPR.2019.00441"},{"key":"2790_CR34","doi-asserted-by":"crossref","unstructured":"Li, Y., & Zhu, J. (2014). A scale adaptive kernel correlation filter tracker with feature integration. In Proceedings european conference computer workshops, pp 254\u2013265.","DOI":"10.1007\/978-3-319-16181-5_18"},{"key":"2790_CR35","doi-asserted-by":"crossref","unstructured":"Lin, L., Fan, H., Zhang, Z., Wang, Y., Xu, Y., & Ling, H. (2025). Tracking meets lora: Faster training, larger model, stronger performance. In Proceedings european conference computer, pp 300\u2013318.","DOI":"10.1007\/978-3-031-73232-4_17"},{"key":"2790_CR36","doi-asserted-by":"crossref","unstructured":"Luke\u017ei\u010d, A., Matas, J., & Kristan, M. (2021). A discriminative single-shot segmentation network for visual object tracking. In IEEE transactions on pattern analysis and machine intelligence, pp 9742\u20139755.","DOI":"10.1109\/TPAMI.2021.3137933"},{"key":"2790_CR37","doi-asserted-by":"crossref","unstructured":"Luke\u017ei\u010d, A., Trojer, \u017d., Matas, J., & Kristan, M. (2024). A new dataset and a distractor-aware architecture for transparent object tracking. International Journal of Computer Vision. pp 1\u201314.","DOI":"10.1007\/s11263-024-02010-0"},{"key":"2790_CR38","doi-asserted-by":"crossref","unstructured":"Luo, R., Song, Z., Ma, L., Wei, J., Yang, W., & Yang, M. (2024). Diffusiontrack: Diffusion model for multi-object tracking. In Proceedings of the AAAI conference on artificial intelligence, pp 3991\u20133999.","DOI":"10.1609\/aaai.v38i5.28192"},{"key":"2790_CR39","doi-asserted-by":"crossref","unstructured":"Lv, W., Huang, Y., Zhang, N., Lin, R. S., Han, M., & Zeng, D. (2024). Diffmot: A real-time diffusion-based multiple object tracker with non-linear prediction. In Computer vision pattern recognition, pp 19321\u201319330.","DOI":"10.1109\/CVPR52733.2024.01828"},{"key":"2790_CR40","doi-asserted-by":"crossref","unstructured":"Ma, Z., Wang, L., Zhang, H., Lu, W., & Yin, J. (2020). Rpt: Learning point set representation for siamese visual tracking. In Proceedings european conference computer workshops, pp 653\u2013665.","DOI":"10.1007\/978-3-030-68238-5_43"},{"key":"2790_CR41","doi-asserted-by":"crossref","unstructured":"Mayer, C., Danelljan, M., Paudel, D. P., & Van Gool, L. (2021). Learning target candidate association to keep track of what not to track. In Computer vision pattern recognition, pp 13444\u201313454.","DOI":"10.1109\/ICCV48922.2021.01319"},{"key":"2790_CR42","unstructured":"Missaoui, B., Cetintas, O., Bras\u00f3, G., Meinhardt, T., & Leal-Taix\u00e1, L. (2025). Noougat: Towards unified online and offline multi-object tracking. arXiv preprint arXiv:2509.02111."},{"key":"2790_CR43","doi-asserted-by":"crossref","unstructured":"Mueller, M., Smith, N., & Ghanem, B. (2016). A benchmark and simulator for uav tracking. In Proceedings european conference computer, pp 445\u2013461.","DOI":"10.1007\/978-3-319-46448-0_27"},{"key":"2790_CR44","doi-asserted-by":"crossref","unstructured":"Muller, M., Bibi, A., Giancola, S., Alsubaihi, S., & Ghanem, B. (2018). TrackingNet: A large-scale dataset and benchmark for object tracking in the wild. In Proceedings european conference computer vision.","DOI":"10.1007\/978-3-030-01246-5_19"},{"key":"2790_CR45","doi-asserted-by":"crossref","unstructured":"Nam, H., & Han, B. (2016). Learning multi-domain convolutional neural networks for visual tracking. In Computer vision pattern recognition, pp 4293\u20134302.","DOI":"10.1109\/CVPR.2016.465"},{"key":"2790_CR46","unstructured":"Oquab, M., Darcet, T., Moutakanni, T., Vo, H., Szafraniec, M., Khalidov, V., Fernandez, P., Haziza, D., Massa, F., El-Nouby, A., et al. (2023). Dinov2: Learning robust visual features without supervision. arXiv:2304.07193."},{"key":"2790_CR47","unstructured":"Pont-Tuset, J., Perazzi, F., Caelles, S., Arbel\u00e1ez, P., Sorkine-Hornung, A., & Van Gool, L. (2017). The 2017 davis challenge on video object segmentation. arXiv:1704.00675."},{"key":"2790_CR48","unstructured":"Ravi, N., Gabeur, V., Hu, Y. T., Hu, R., Ryali, C., Ma, T., Khedr, H., R\u00e4dle, R., Rolland, C., Gustafson, L., et al. (2025). Sam 2: Segment anything in images and videos. In Proceedings international conference learning representations."},{"key":"2790_CR49","unstructured":"Ryali, C., Hu, Y. T., Bolya, D., Wei, C., Fan, H., Huang, P.-Y., Aggarwal, V., Chowdhury, A., Poursaeed, O., Hoffman, J., et al. (2023). Hiera: A hierarchical vision transformer without the bells-and-whistles. In Proceedings international conference machine learning, pp 29441\u201329454."},{"key":"2790_CR50","doi-asserted-by":"crossref","unstructured":"Tokmakov, P., Li, J., & Gaidon, A. (2023). Breaking the \"object\" in video object segmentation. In Computer vision pattern recognition, pp 22836\u201322845.","DOI":"10.1109\/CVPR52729.2023.02187"},{"key":"2790_CR51","doi-asserted-by":"crossref","unstructured":"Videnovic, J., Lukezic, A., & Kristan, M. (2025). A distractor-aware memory for visual object tracking with SAM2. In Computer vision pattern recognition.","DOI":"10.1109\/CVPR52734.2025.02259"},{"key":"2790_CR52","unstructured":"VOTS (2025). VOT challenge. https:\/\/www.votchallenge.net, accessed: 2025\u201308-11."},{"key":"2790_CR53","doi-asserted-by":"crossref","unstructured":"Wu, Y., Lim, J., & Yang, M. H. (2013). Online object tracking: A benchmark. In Computer Vision Pattern Recognition, pp 2411\u20132418.","DOI":"10.1109\/CVPR.2013.312"},{"key":"2790_CR54","doi-asserted-by":"crossref","unstructured":"Wu, Y., Lim, J., & Yang, M. H. (2015). Object tracking benchmark. In IEEE transactions on pattern analysis and machine intelligence, pp 1834\u20131848.","DOI":"10.1109\/TPAMI.2014.2388226"},{"key":"2790_CR55","doi-asserted-by":"crossref","unstructured":"Xie, J., Zhong, B., Mo, Z., Zhang, S., Shi, L., Song, S., & Ji, R. (2024). Autoregressive queries for adaptive tracking with spatio-temporal transformers. In Computer vision pattern recognition, pp 19300\u201319309.","DOI":"10.1109\/CVPR52733.2024.01826"},{"key":"2790_CR56","unstructured":"Xiong, Y., Zhou, C., Xiang, X., Wu, L., Zhu, C., Liu, Z., Suri, S., Varadarajan, B., Akula, R., Iandola, F., et al. (2024). Efficient track anything. arXiv preprint arXiv:2411.18933."},{"key":"2790_CR57","doi-asserted-by":"crossref","unstructured":"Xu, N., Yang, L., Fan, Y., Yue, D., Liang, Y., Yang, J., & Huang, T. (2018). Youtube-vos: A large-scale video object segmentation benchmark. arXiv preprint arXiv:1809.03327.","DOI":"10.1007\/978-3-030-01228-1_36"},{"key":"2790_CR58","doi-asserted-by":"crossref","unstructured":"Yan, B., Peng, H., Fu, J., Wang, D., & Lu, H. (2021). Learning spatio-temporal transformer for visual tracking. In International conference computer vision, pp 10448\u201310457.","DOI":"10.1109\/ICCV48922.2021.01028"},{"key":"2790_CR59","unstructured":"Yang, C. Y., Huang, H. W., Chai, W., Jiang, Z., & Hwang, J. N. (2024). Samurai: Adapting segment anything model for zero-shot visual tracking with motion-aware memory. arXiv:2411.11922."},{"key":"2790_CR60","first-page":"2491","volume":"34","author":"Z Yang","year":"2021","unstructured":"Yang, Z., Wei, Y., & Yang, Y. (2021). Associating objects with transformers for video object segmentation. Neural Information Processing Systems, 34, 2491\u20132502.","journal-title":"Neural Information Processing Systems"},{"key":"2790_CR61","doi-asserted-by":"crossref","unstructured":"Ye, B., Chang, H., Ma, B., Shan, S., & Chen, X. (2022). Joint feature learning and relation modeling for tracking: A one-stream framework. In Avidan, S., Brostow, G., Ciss\u00e9, M., et al. (eds). Proceedings european conference computer, pp 341\u2013357.","DOI":"10.1007\/978-3-031-20047-2_20"},{"key":"2790_CR62","doi-asserted-by":"crossref","unstructured":"Zheng, Y., Zhong, B., Liang, Q., Mo, Z., Zhang, S., & Li, X. (2024). Odtrack: Online dense temporal token learning for visual tracking. In Proceedings of the AAAI conference on artificial intelligence, pp 7588\u20137596.","DOI":"10.1609\/aaai.v38i7.28591"},{"key":"2790_CR63","doi-asserted-by":"crossref","unstructured":"Zhou, C., Zhu, C., Xiong, Y., Suri, S., Xiao, F., Wu, L., Krishnamoorthi, R., Dai, B., Loy, C. C., Chandra, V., et al. (2025). Edgetam: On-device track anything model. In Computer vision pattern recognition, pp 13832\u201313842.","DOI":"10.1109\/CVPR52734.2025.01291"},{"key":"2790_CR64","doi-asserted-by":"crossref","unstructured":"Zhou, J., Pang, Z., & Wang, Y. X. (2024). Rmem: Restricted memory banks improve video object segmentation. In computer vision pattern recognition, pp 18602\u201318611.","DOI":"10.1109\/CVPR52733.2024.01760"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-026-02790-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-026-02790-7","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-026-02790-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,4,7]],"date-time":"2026-04-07T03:22:22Z","timestamp":1775532142000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-026-02790-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,4,7]]},"references-count":64,"journal-issue":{"issue":"5","published-print":{"date-parts":[[2026,5]]}},"alternative-id":["2790"],"URL":"https:\/\/doi.org\/10.1007\/s11263-026-02790-7","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,4,7]]},"assertion":[{"value":"12 June 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 February 2026","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 April 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"211"}}