{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T09:12:19Z","timestamp":1774602739410,"version":"3.50.1"},"reference-count":76,"publisher":"Springer Science and Business Media LLC","issue":"3","license":[{"start":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T00:00:00Z","timestamp":1769731200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T00:00:00Z","timestamp":1769731200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62372251"],"award-info":[{"award-number":["62372251"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2026,3]]},"DOI":"10.1007\/s11263-025-02700-3","type":"journal-article","created":{"date-parts":[[2026,1,30]],"date-time":"2026-01-30T05:07:41Z","timestamp":1769749661000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Practical Video Object Detection via Feature Selection and Aggregation"],"prefix":"10.1007","volume":"134","author":[{"given":"Yuheng","family":"Shi","sequence":"first","affiliation":[]},{"given":"Tong","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Xiaojie","family":"Guo","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,1,30]]},"reference":[{"key":"2700_CR1","doi-asserted-by":"crossref","unstructured":"Belhassen, H., Zhang, H., Fresse, V., & Bourennane, E.B. (2019). Improving video object detection by seq-bbox matching. In: VISIGRAPP (5: VISAPP)","DOI":"10.5220\/0007260000002108"},{"key":"2700_CR2","unstructured":"Bochkovskiy, A., Wang, C.Y., & Liao, H.Y.M. (2020). Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934"},{"key":"2700_CR3","doi-asserted-by":"crossref","unstructured":"Bolya, D., Foley, S., Hays, J., & Hoffman, J. (2020). Tide: A general toolbox for identifying object detection errors. In: ECCV","DOI":"10.1007\/978-3-030-58580-8_33"},{"key":"2700_CR4","doi-asserted-by":"crossref","unstructured":"Cai, Z., & Vasconcelos, N. (2018). Cascade r-cnn: Delving into high quality object detection. In: CVPR","DOI":"10.1109\/CVPR.2018.00644"},{"key":"2700_CR5","doi-asserted-by":"crossref","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., & Zagoruyko, S. (2020). End-to-end object detection with transformers. In: ECCV","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"2700_CR6","doi-asserted-by":"crossref","unstructured":"Chang, M.F., Lambert, J.W., Sangkloy, P., Singh, J., Bak, S., Hartnett, A., Wang, D., Carr, P., Lucey, S., Ramanan, D., & Hays, J. (2019). Argoverse: 3d tracking and forecasting with rich maps. In: CVPR","DOI":"10.1109\/CVPR.2019.00895"},{"key":"2700_CR7","doi-asserted-by":"crossref","unstructured":"Chen, Y., Cao, Y., Hu, H., & Wang, L. (2020). Memory enhanced global-local aggregation for video object detection. In: CVPR","DOI":"10.1109\/CVPR42600.2020.01035"},{"key":"2700_CR8","unstructured":"Dai, J., Li, Y., He, K., & Sun, J. (2016). R-fcn: Object detection via region-based fully convolutional networks. In: NeurIPS"},{"key":"2700_CR9","unstructured":"Dalal, N., & Triggs, B. (2005). Histograms of oriented gradients for human detection. In: CVPR"},{"key":"2700_CR10","doi-asserted-by":"crossref","unstructured":"Deng, C., Chen, D., & Wu, Q. (2023). Identity-consistent aggregation for video object detection. In: ICCV","DOI":"10.1109\/ICCV51070.2023.01236"},{"key":"2700_CR11","doi-asserted-by":"crossref","unstructured":"Deng, J., Pan, Y., Yao, T., Zhou, W., Li, H., & Mei, T. (2019). Relation distillation networks for video object detection. In: ICCV","DOI":"10.1109\/ICCV.2019.00712"},{"key":"2700_CR12","unstructured":"Dosovitskiy, A., Beyer, L., Kolesnikov, A., Weissenborn, D., Zhai, X., Unterthiner, T., Dehghani, M., Minderer, M., Heigold, G., Gelly, S., Uszkoreit, J., & Houlsby, N. (2021). An image is worth 16x16 words: Transformers for image recognition at scale. In: ICLR"},{"key":"2700_CR13","doi-asserted-by":"crossref","unstructured":"Feichtenhofer, C., Pinz, A., & Zisserman, A. (2017). Detect to track and track to detect. In: ICCV","DOI":"10.1109\/ICCV.2017.330"},{"key":"2700_CR14","doi-asserted-by":"crossref","unstructured":"Felzenszwalb, P., McAllester, D., & Ramanan, D. (2008). A discriminatively trained, multiscale, deformable part model. In: CVPR","DOI":"10.1109\/CVPR.2008.4587597"},{"key":"2700_CR15","doi-asserted-by":"publisher","first-page":"65886","DOI":"10.1109\/ACCESS.2022.3184031","volume":"10","author":"M Fujitake","year":"2022","unstructured":"Fujitake, M., & Sugimoto, A. (2022). Video sparse transformer with attention-guided memory for video object detection. IEEE Access, 10, 65886\u201365900.","journal-title":"IEEE Access"},{"key":"2700_CR16","doi-asserted-by":"crossref","unstructured":"Ge, Z., Liu, S., Li, Z., Yoshie, O., & Sun, J. (2021). Ota: Optimal transport assignment for object detection. In: CVPR","DOI":"10.1109\/CVPR46437.2021.00037"},{"key":"2700_CR17","unstructured":"Ge, Z., Liu, S., Wang, F., Li, Z., & Sun, J. (2021). Yolox: Exceeding yolo series in 2021. arXiv preprint arXiv:2107.08430"},{"key":"2700_CR18","doi-asserted-by":"crossref","unstructured":"Girshick, R. (2015). Fast r-cnn. In: ICCV","DOI":"10.1109\/ICCV.2015.169"},{"key":"2700_CR19","doi-asserted-by":"crossref","unstructured":"Girshick, R., Donahue, J., Darrell, T., & Malik, J. (2014). Rich feature hierarchies for accurate object detection and semantic segmentation. In: CVPR","DOI":"10.1109\/CVPR.2014.81"},{"key":"2700_CR20","doi-asserted-by":"crossref","unstructured":"Gong, T., Chen, K., Wang, X., Chu, Q., Zhu, F., Lin, D., Yu, N., & Feng, H. (2021). Temporal roi align for video object recognition. In: AAAI","DOI":"10.1609\/aaai.v35i2.16234"},{"key":"2700_CR21","unstructured":"Han, W., Jun, T., Xiaodong, L., Shanyan, G., Rong, X., & Li, S. (2022). Ptseformer: Progressive temporal-spatial enhanced transformer towards video object detection. In: ECCV"},{"key":"2700_CR22","unstructured":"Han, W., Khorrami, P., Paine, T.L., Ramachandran, P., Babaeizadeh, M., Shi, H., Li, J., Yan, S., & Huang, T.S. (2016). Seq-nms for video object detection. arXiv preprint arXiv:1602.08465"},{"key":"2700_CR23","doi-asserted-by":"crossref","unstructured":"Han, M., Wang, Y., Chang, X., & Qiao, Y. (2020). Mining inter-video proposal relations for video object detection. In: ECCV","DOI":"10.1007\/978-3-030-58589-1_26"},{"key":"2700_CR24","doi-asserted-by":"crossref","unstructured":"He, F., Gao, N., Jia, J., Zhao, X., & Huang, K. (2022). Queryprop: Object query propagation for high-performance video object detection. In: AAAI","DOI":"10.1609\/aaai.v36i1.19965"},{"key":"2700_CR25","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Doll\u00e1r, P., & Girshick, R. (2017). Mask r-cnn. In: ICCV","DOI":"10.1109\/ICCV.2017.322"},{"key":"2700_CR26","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In: CVPR","DOI":"10.1109\/CVPR.2016.90"},{"key":"2700_CR27","doi-asserted-by":"crossref","unstructured":"He, L., Zhou, Q., Li, X., Niu, L., Cheng, G., Li, X., Liu, W., Tong, Y., Ma, L., & Zhang, L. (2021). End-to-end video object detection with spatial-temporal transformers. In: ACM MM","DOI":"10.1145\/3474085.3475285"},{"key":"2700_CR28","unstructured":"Howard, A.G., Zhu, M., Chen, B., Kalenichenko, D., Wang, W., Weyand, T., Andreetto, M., & Adam, H. (2017). Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861"},{"key":"2700_CR29","doi-asserted-by":"crossref","unstructured":"Hu, H., Gu, J., Zhang, Z., Dai, J., & Wei, Y. (2018). Relation networks for object detection. In: CVPR","DOI":"10.1109\/CVPR.2018.00378"},{"key":"2700_CR30","doi-asserted-by":"crossref","unstructured":"Hu, J., Shen, L., & Sun, G. (2018). Squeeze-and-excitation networks. In: CVPR","DOI":"10.1109\/CVPR.2018.00745"},{"key":"2700_CR31","doi-asserted-by":"crossref","unstructured":"Kim, K., & Lee, H.S. (2020). Probabilistic anchor assignment with iou prediction for object detection. In: ECCV. Springer","DOI":"10.1007\/978-3-030-58595-2_22"},{"key":"2700_CR32","unstructured":"Krizhevsky, A., Sutskever, I., & Hinton, G.E. (2012). Imagenet classification with deep convolutional neural networks. In: NeurIPS"},{"key":"2700_CR33","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., & Belongie, S. (2017). Feature pyramid networks for object detection. In: CVPR","DOI":"10.1109\/CVPR.2017.106"},{"key":"2700_CR34","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Goyal, P., Girshick, R., He, K., & Doll\u00e1r, P. (2017). Focal loss for dense object detection. In: ICCV","DOI":"10.1109\/ICCV.2017.324"},{"key":"2700_CR35","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Maire, M., Belongie, S., Hays, J., Perona, P., Ramanan, D., Doll\u00e1r, P., & Zitnick, C.L. (2014). Microsoft coco: Common objects in context. In: ECCV","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"2700_CR36","unstructured":"Lin, J., Mao, X., Chen, Y., Xu, L., He, Y., & Xue, H. (2022). D2etr: Decoder-only detr with computationally efficient cross-scale attention. arXiv preprint arXiv:2203.00860"},{"key":"2700_CR37","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y.,& Berg, A.C. (2016). Ssd: Single shot multibox detector. In: ECCV","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"2700_CR38","unstructured":"Liu, S., Li, F., Zhang, H., Yang, X., Qi, X., Su, H., Zhu, J., & Zhang, L. (2022). DAB-DETR: Dynamic anchor boxes are better queries for DETR. In: ICLR"},{"key":"2700_CR39","doi-asserted-by":"crossref","unstructured":"Liu, Z., Lin, Y., Cao, Y., Hu, H., Wei, Y., Zhang, Z., Lin, S., & Guo, B. (2021). Swin transformer: Hierarchical vision transformer using shifted windows. In: ICCV","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"2700_CR40","doi-asserted-by":"crossref","unstructured":"Liu, Z., Mao, H., Wu, C.Y., Feichtenhofer, C., Darrell, T., & Xie, S. (2022). A convnet for the 2020s. In: CVPR","DOI":"10.1109\/CVPR52688.2022.01167"},{"key":"2700_CR41","doi-asserted-by":"crossref","unstructured":"Liu, X., Nejadasl, F.K., van Gemert, J.C., Booij, O., & Pintea, S.L. (2023). Objects do not disappear: Video object detection by single-frame object location anticipation. In: ICCV","DOI":"10.1109\/ICCV51070.2023.00640"},{"key":"2700_CR42","doi-asserted-by":"crossref","unstructured":"Liu, S., Qi, L., Qin, H., Shi, J., & Jia, J. (2018). Path aggregation network for instance segmentation. In: CVPR","DOI":"10.1109\/CVPR.2018.00913"},{"key":"2700_CR43","doi-asserted-by":"publisher","first-page":"2022","DOI":"10.1007\/s11263-022-01629-1","volume":"130","author":"J Qi","year":"2022","unstructured":"Qi, J., Gao, Y., Hu, Y., Wang, X., Liu, X., Bai, X., Belongie, S., Yuille, A., Torr, P., & Bai, S. (2022). Occluded video instance segmentation: A benchmark. IJCV, 130, 2022\u20132039.","journal-title":"IJCV"},{"key":"2700_CR44","unstructured":"Radford, A., Kim, J. W., Hallacy, C., Ramesh, A., Goh, G., Agarwal, S., Sastry, G., Askell, A., Mishkin, P., Clark, J., et\u00a0al. (2021). Learning transferable visual models from natural language supervision."},{"key":"2700_CR45","unstructured":"Ravi, N., Gabeur, V., Hu, Y.T., Hu, R., Ryali, C., Ma, T., Khedr, H., R\u00e4dle, R., Rolland, C., Gustafson, L., Mintun, E., Pan, J., Alwala, K.V., Carion, N., Wu, C.Y., Girshick, R., Doll\u00e1r, P., & Feichtenhofer, C. (2024). Sam 2: Segment anything in images and videos. https:\/\/arxiv.org\/abs\/2408.00714"},{"key":"2700_CR46","doi-asserted-by":"crossref","unstructured":"Redmon, J., & Farhadi, A. (2017). Yolo9000: better, faster, stronger. In: CVPR","DOI":"10.1109\/CVPR.2017.690"},{"key":"2700_CR47","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look once: Unified, real-time object detection. In: CVPR","DOI":"10.1109\/CVPR.2016.91"},{"key":"2700_CR48","doi-asserted-by":"crossref","unstructured":"Ren, S., He, K., Girshick, R., & Sun, J. (2017). Faster r-cnn: Towards real-time object detection with region proposal networks. In: NeurIPS (2015)","DOI":"10.1109\/TPAMI.2016.2577031"},{"key":"2700_CR49","doi-asserted-by":"publisher","first-page":"211","DOI":"10.1007\/s11263-015-0816-y","volume":"115","author":"O Russakovsky","year":"2015","unstructured":"Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy, A., Khosla, A., Bernstein, M., et al. (2015). Imagenet large scale visual recognition challenge. IJCV, 115, 211\u2013252.","journal-title":"IJCV"},{"key":"2700_CR50","doi-asserted-by":"crossref","unstructured":"Sabater, A., Montesano, L., & Murillo, A.C. (2020). Robust and efficient post-processing for video object detection. In: IROS","DOI":"10.1109\/IROS45743.2020.9341600"},{"key":"2700_CR51","doi-asserted-by":"crossref","unstructured":"Shao, S., Li, Z., Zhang, T., Peng, C., Yu, G., Zhang, X., Li, J., & Sun, J. (2019). Objects365: A large-scale, high-quality dataset for object detection. In: ICCV","DOI":"10.1109\/ICCV.2019.00852"},{"key":"2700_CR52","doi-asserted-by":"crossref","unstructured":"Shi, Y., Wang, N., & Guo, X. (2023). Yolov: Making still image object detectors great at video object detection. In: AAAI","DOI":"10.1609\/aaai.v37i2.25320"},{"key":"2700_CR53","doi-asserted-by":"crossref","unstructured":"Shvets, M., Liu, W., & Berg, A.C. (2019). Leveraging long-range temporal relationships between proposals for video object detection. In: ICCV","DOI":"10.1109\/ICCV.2019.00985"},{"key":"2700_CR54","unstructured":"Simonyan, K., & Zisserman, A. (2015). Very deep convolutional networks for large-scale image recognition. In: ICLR"},{"key":"2700_CR55","doi-asserted-by":"crossref","unstructured":"Sun, G., Hua, Y., Hu, G., & Robertson, N. (2021). Mamba: Multi-level aggregation via memory bank for video object detection. In: AAAI","DOI":"10.1609\/aaai.v35i3.16365"},{"key":"2700_CR56","doi-asserted-by":"crossref","unstructured":"Sun, G., Hua, Y., Hu, G., & Robertson, N. (2022). Efficient one-stage video object detection by exploiting temporal consistency. In: ECCV","DOI":"10.1007\/978-3-031-19833-5_1"},{"key":"2700_CR57","unstructured":"Tan, M., & Le, Q. (2019). Efficientnet: Rethinking model scaling for convolutional neural networks. In: ICML"},{"key":"2700_CR58","doi-asserted-by":"crossref","unstructured":"Tian, Z., Shen, C., Chen, H., & He, T. (2019). Fcos: Fully convolutional one-stage object detection. In: ICCV","DOI":"10.1109\/ICCV.2019.00972"},{"key":"2700_CR59","unstructured":"Touvron, H., Cord, M., Douze, M., Massa, F., Sablayrolles, A., & J\u00e9gou, H. (2021). Training data-efficient image transformers & distillation through attention. In: ICML"},{"key":"2700_CR60","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.N., Kaiser, \u0141., & Polosukhin, I. (2017). Attention is all you need. In: NeurIPS"},{"key":"2700_CR61","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., & Liao, H.Y.M. (2021). Scaled-yolov4: Scaling cross stage partial network. In: CVPR","DOI":"10.1109\/CVPR46437.2021.01283"},{"key":"2700_CR62","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Bochkovskiy, A., & Liao, H.Y.M. (2023). Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: CVPR","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"2700_CR63","doi-asserted-by":"crossref","unstructured":"Wang, C.Y., Liao, H.Y.M., Wu, Y.H., Chen, P.Y., Hsieh, J.W., & Yeh, I.H.(2020). Cspnet: A new backbone that can enhance learning capability of cnn. In: CVPR","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"2700_CR64","doi-asserted-by":"crossref","unstructured":"Wang, S., Zhou, Y., Yan, J., & Deng, Z. (2018). Fully motion-aware network for video object detection. In: ECCV","DOI":"10.1007\/978-3-030-01261-8_33"},{"key":"2700_CR65","doi-asserted-by":"crossref","unstructured":"Wu, H., Chen, Y., Wang, N., & Zhang, Z. (2019). Sequence level semantics aggregation for video object detection. In: ICCV","DOI":"10.1109\/ICCV.2019.00931"},{"key":"2700_CR66","doi-asserted-by":"crossref","unstructured":"Xie, S., Girshick, R., Doll\u00e1r, P., Tu, Z., & He, K. (2017). Aggregated residual transformations for deep neural networks. In: CVPR","DOI":"10.1109\/CVPR.2017.634"},{"key":"2700_CR67","unstructured":"Xu, S., Wang, X., Lv, W., Chang, Q., Cui, C., Deng, K., Wang, G., Dang, Q., Wei, S., & Du, Y. et\u00a0al. (2022). Pp-yoloe: An evolved version of yolo. arXiv preprint arXiv:2203.16250"},{"key":"2700_CR68","unstructured":"Yang, J., Li, C., Dai, X., & Gao, J. (2022). Focal modulation networks. NeurIPS"},{"key":"2700_CR69","unstructured":"Zhang, Z., Cheng, D., Zhu, X., Lin, S., & Dai, J. (2018). Integrated object detection and tracking with tracklet-conditioned detection. arXiv preprint arXiv:1811.11167"},{"key":"2700_CR70","doi-asserted-by":"crossref","unstructured":"Zhang, S., Chi, C., Yao, Y., Lei, Z., & Li, S.Z. (2020). Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection. In: CVPR","DOI":"10.1109\/CVPR42600.2020.00978"},{"key":"2700_CR71","unstructured":"Zhang, H., Li, F., Liu, S., Zhang, L., Su, H., Zhu, J., Ni, L.M., & Shum, H.Y. (2022). Dino: Detr with improved denoising anchor boxes for end-to-end object detection. arXiv preprint arXiv:2203.03605"},{"issue":"6","key":"2700_CR72","doi-asserted-by":"publisher","first-page":"7853","DOI":"10.1109\/TPAMI.2022.3223955","volume":"45","author":"Q Zhou","year":"2023","unstructured":"Zhou, Q., Li, X., He, L., Yang, Y., Cheng, G., Tong, Y., Ma, L., & Tao, D. (2023). Transvod: End-to-end video object detection with spatial-temporal transformers. TPAMI, 45(6), 7853\u20137869.","journal-title":"TPAMI"},{"key":"2700_CR73","doi-asserted-by":"crossref","unstructured":"Zhu, X., Dai, J., Yuan, L., & Wei, Y. (2018). Towards high performance video object detection. In: CVPR","DOI":"10.1109\/CVPR.2018.00753"},{"key":"2700_CR74","unstructured":"Zhu, X., Su, W., Lu, L., Li, B., Wang, X., & Dai, J. (2021). Deformable detr: Deformable transformers for end-to-end object detection. arXiv preprint arXiv:2010.04159"},{"key":"2700_CR75","doi-asserted-by":"crossref","unstructured":"Zhu, X., Wang, Y., Dai, J., Yuan, L., & Wei, Y. (2017). Flow-guided feature aggregation for video object detection. In: ICCV","DOI":"10.1109\/ICCV.2017.52"},{"key":"2700_CR76","doi-asserted-by":"crossref","unstructured":"Zhu, X., Xiong, Y., Dai, J., Yuan, L., & Wei, Y. (2017). Deep feature flow for video recognition. In: CVPR","DOI":"10.1109\/CVPR.2017.441"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02700-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02700-3","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02700-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,27]],"date-time":"2026-03-27T08:37:15Z","timestamp":1774600635000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02700-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,1,30]]},"references-count":76,"journal-issue":{"issue":"3","published-print":{"date-parts":[[2026,3]]}},"alternative-id":["2700"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02700-3","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,1,30]]},"assertion":[{"value":"9 May 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"2 September 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 January 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"95"}}