{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,2]],"date-time":"2025-05-02T04:03:20Z","timestamp":1746158600398,"version":"3.40.4"},"reference-count":33,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2024,5,14]],"date-time":"2024-05-14T00:00:00Z","timestamp":1715644800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,14]],"date-time":"2024-05-14T00:00:00Z","timestamp":1715644800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-19362-x","type":"journal-article","created":{"date-parts":[[2024,5,14]],"date-time":"2024-05-14T04:17:00Z","timestamp":1715660220000},"page":"10221-10239","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Dynamic distillation based multi-scale lightweight target detection"],"prefix":"10.1007","volume":"84","author":[{"given":"Kai","family":"Sun","sequence":"first","affiliation":[]},{"given":"Danjing","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,14]]},"reference":[{"key":"19362_CR1","doi-asserted-by":"publisher","unstructured":"Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 580\u2013587. https:\/\/doi.org\/10.18127\/j00338486-202109-11","DOI":"10.18127\/j00338486-202109-11"},{"key":"19362_CR2","doi-asserted-by":"publisher","unstructured":"Girshick R (2015) Fast r-cnn. In: Proceedings of the IEEE international conference on computer vision, pp 1440\u20131448. https:\/\/doi.org\/10.1109\/iccv.2015.169","DOI":"10.1109\/iccv.2015.169"},{"key":"19362_CR3","doi-asserted-by":"publisher","unstructured":"Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: Towards real-time object detection with region proposal networks. Adv Neural Inf Process Syst 28. https:\/\/doi.org\/10.48550\/arXiv.1506.01497","DOI":"10.48550\/arXiv.1506.01497"},{"key":"19362_CR4","doi-asserted-by":"publisher","unstructured":"Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv:1804.02767. https:\/\/doi.org\/10.48550\/arXiv.1804.02767","DOI":"10.48550\/arXiv.1804.02767"},{"key":"19362_CR5","doi-asserted-by":"publisher","unstructured":"Bochkovskiy A, Wang C-Y, Liao H-YM (2020) Yolov4: Optimal speed and accuracy of object detection. arXiv:2004.10934. https:\/\/doi.org\/10.48550\/arXiv.2004.10934","DOI":"10.48550\/arXiv.2004.10934"},{"key":"19362_CR6","doi-asserted-by":"publisher","unstructured":"Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) Ssd: Single shot multibox detector. In: Computer vision\u2013ECCV 2016: 14th european conference, Amsterdam, The Netherlands, October 11\u201314, 2016, Proceedings, Part I 14, pp 21\u201337. https:\/\/doi.org\/10.48550\/arXiv.1512.02325","DOI":"10.48550\/arXiv.1512.02325"},{"key":"19362_CR7","doi-asserted-by":"publisher","unstructured":"Lin T-Y, Goyal P, Girshick R, He K, Doll\u00e1r P (2017) Focal loss for dense object detection. In: Proceedings of the IEEE international conference on computer vision, pp 2980\u20132988. https:\/\/doi.org\/10.48550\/arXiv.1708.02002","DOI":"10.48550\/arXiv.1708.02002"},{"key":"19362_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2022.104914","volume":"113","author":"X Dong","year":"2022","unstructured":"Dong X, Yan S, Duan C (2022) A lightweight vehicles detection network model based on yolov5. Eng Appl Artif Intell 113:104914. https:\/\/doi.org\/10.1016\/j.engappai.2022.104914","journal-title":"Eng Appl Artif Intell"},{"key":"19362_CR9","doi-asserted-by":"publisher","DOI":"10.1016\/j.patcog.2023.109377","volume":"138","author":"C Wang","year":"2023","unstructured":"Wang C, Wang H (2023) Cascaded feature fusion with multi-level self-attention mechanism for object detection. Pattern Recogn 138:109377. https:\/\/doi.org\/10.1016\/j.patcog.2023.109377","journal-title":"Pattern Recogn"},{"issue":"4","key":"19362_CR10","doi-asserted-by":"publisher","first-page":"2402","DOI":"10.3390\/app13042402","volume":"13","author":"Y Yang","year":"2023","unstructured":"Yang Y, Zhou Y, Din NU, Li J, He Y, Zhang L (2023) An improved yolov5 model for detecting laser welding defects of lithium battery pole. Appl Sci 13(4):2402. https:\/\/doi.org\/10.3390\/app13042402","journal-title":"Appl Sci"},{"issue":"2","key":"19362_CR11","doi-asserted-by":"publisher","first-page":"486","DOI":"10.1007\/s12559-022-10052-0","volume":"15","author":"Y Lu","year":"2023","unstructured":"Lu Y, Su M, Wang Y, Liu Z, Peng T (2023) Learning discriminated features based on feature pyramid networks and attention for multi-scale object detection. Cogn Comput 15(2):486\u2013495. https:\/\/doi.org\/10.1007\/s12559-022-10052-0","journal-title":"Cogn Comput"},{"key":"19362_CR12","doi-asserted-by":"publisher","unstructured":"Wang C-Y, Bochkovskiy A, Liao H-YM (2023) Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 7464\u20137475. https:\/\/doi.org\/10.48550\/arXiv.2207.02696","DOI":"10.48550\/arXiv.2207.02696"},{"key":"19362_CR13","doi-asserted-by":"publisher","unstructured":"Yu F, Koltun V (2015) Multi-scale context aggregation by dilated convolutions. arXiv:1511.07122. https:\/\/doi.org\/10.48550\/arXiv.1511.07122","DOI":"10.48550\/arXiv.1511.07122"},{"key":"19362_CR14","doi-asserted-by":"publisher","unstructured":"Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Eca-net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 11534\u201311542. https:\/\/doi.org\/10.48550\/arXiv.1910.03151","DOI":"10.48550\/arXiv.1910.03151"},{"key":"19362_CR15","doi-asserted-by":"publisher","first-page":"6101","DOI":"10.1609\/aaai.v35i7.16760","volume":"35","author":"D Liu","year":"2021","unstructured":"Liu D, Cui Y, Yan L, Mousas C, Yang B, Chen Y (2021) Densernet: Weakly supervised visual localization using multi-scale feature aggregation. Proceedings of the AAAI conference on artificial intelligence 35:6101\u20136109. https:\/\/doi.org\/10.1609\/aaai.v35i7.16760","journal-title":"Proceedings of the AAAI conference on artificial intelligence"},{"key":"19362_CR16","doi-asserted-by":"publisher","unstructured":"Zhu X, Hu H, Lin S, Dai J (2019) Deformable convnets v2: More deformable, better results. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 9308\u20139316. https:\/\/doi.org\/10.48550\/arXiv.1811.11168","DOI":"10.48550\/arXiv.1811.11168"},{"key":"19362_CR17","doi-asserted-by":"publisher","DOI":"10.1109\/TIP.2023.3272826","author":"D Liu","year":"2023","unstructured":"Liu D, Liang J, Geng T, Loui A, Zhou T (2023) Tripartite feature enhanced pyramid network for dense prediction. IEEE Trans Image Process. https:\/\/doi.org\/10.1109\/TIP.2023.3272826","journal-title":"IEEE Trans Image Process"},{"key":"19362_CR18","doi-asserted-by":"publisher","unstructured":"Lee J, Park S, Mo S, Ahn S, Shin J (2020) Layer-adaptive sparsity for the magnitude-based pruning. arXiv:2010.07611. https:\/\/doi.org\/10.48550\/arXiv.2010.07611","DOI":"10.48550\/arXiv.2010.07611"},{"key":"19362_CR19","doi-asserted-by":"publisher","unstructured":"Chen H, Wang Y, Xu C, Yang Z, Liu C, Shi B, Xu C, Xu C, Tian Q (2019) Data-free learning of student networks. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 3514\u20133522. https:\/\/doi.org\/10.48550\/arXiv.1904.01186","DOI":"10.48550\/arXiv.1904.01186"},{"key":"19362_CR20","doi-asserted-by":"publisher","unstructured":"Hinton G, Vinyals O, Dean J (2015) Distilling the knowledge in a neural network. arXiv:1503.02531. https:\/\/doi.org\/10.48550\/arXiv.1503.02531","DOI":"10.48550\/arXiv.1503.02531"},{"key":"19362_CR21","doi-asserted-by":"publisher","unstructured":"Tang R, Lu Y, Liu L, Mou L, Vechtomova O, Lin J (2019) Distilling task-specific knowledge from bert into simple neural networks. arXiv:1903.12136. https:\/\/doi.org\/10.48550\/arXiv.1903.12136","DOI":"10.48550\/arXiv.1903.12136"},{"key":"19362_CR22","doi-asserted-by":"publisher","first-page":"98","DOI":"10.1007\/s11263-014-0733-5","volume":"111","author":"M Everingham","year":"2015","unstructured":"Everingham M, Eslami SA, Van Gool L, Williams CK, Winn J, Zisserman A (2015) The pascal visual object classes challenge: A retrospective. Int J Comput Vision 111:98\u2013136. https:\/\/doi.org\/10.1007\/s11263-014-0733-5","journal-title":"Int J Comput Vision"},{"issue":"11","key":"19362_CR23","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger A, Lenz P, Stiller C, Urtasun R (2013) Vision meets robotics: The kitti dataset. Int J Robot Res 32(11):1231\u20131237. https:\/\/doi.org\/10.1177\/0278364913491297","journal-title":"Int J Robot Res"},{"key":"19362_CR24","doi-asserted-by":"publisher","unstructured":"Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, Franke U, Roth S, Schiele B (2016) The cityscapes dataset for semantic urban scene understanding. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3213\u20133223. https:\/\/doi.org\/10.48550\/arXiv.1604.01685","DOI":"10.48550\/arXiv.1604.01685"},{"key":"19362_CR25","doi-asserted-by":"publisher","unstructured":"Lin T-Y, Maire M, Belongie S, Hays J, Perona P, Ramanan D, Doll\u00e1r P, Zitnick CL (2014) Microsoft coco: Common objects in context. In: Computer vision\u2013ECCV 2014: 13th European conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pp 740\u2013755. Springer. https:\/\/doi.org\/10.48550\/arXiv.1405.0312","DOI":"10.48550\/arXiv.1405.0312"},{"key":"19362_CR26","doi-asserted-by":"publisher","unstructured":"Jocher G (2020) Ultralytics YOLOv5. https:\/\/doi.org\/10.5281\/zenodo.3908559. https:\/\/github.com\/ultralytics\/yolov5","DOI":"10.5281\/zenodo.3908559"},{"key":"19362_CR27","doi-asserted-by":"publisher","unstructured":"Ge Z, Liu S, Wang F, Li Z, Sun J (2021) Yolox: Exceeding yolo series in 2021. arXiv:2107.08430. https:\/\/doi.org\/10.48550\/arXiv.2107.08430","DOI":"10.48550\/arXiv.2107.08430"},{"key":"19362_CR28","unstructured":"Jocher G, Chaurasia A, Qiu J (2023) Ultralytics YOLOv8. https:\/\/github.com\/ultralytics\/ultralytics"},{"key":"19362_CR29","doi-asserted-by":"publisher","unstructured":"Wang C-Y, Bochkovskiy A, Liao H-YM (2021) Scaled-yolov4: Scaling cross stage partial network. In: Proceedings of the IEEE\/cvf conference on computer vision and pattern recognition, pp 13029\u201313038. https:\/\/doi.org\/10.48550\/arXiv.2011.08036","DOI":"10.48550\/arXiv.2011.08036"},{"key":"19362_CR30","doi-asserted-by":"publisher","unstructured":"Xu S, Wang X, Lv W, Chang Q, Cui C, Deng K, Wang G, Dang Q, Wei S, Du Y et al (2022) Pp-yoloe: An evolved version of yolo. arXiv:2203.16250. https:\/\/doi.org\/10.48550\/arXiv.2203.16250","DOI":"10.48550\/arXiv.2203.16250"},{"key":"19362_CR31","doi-asserted-by":"publisher","unstructured":"Zhang H, Li F, Liu S, Zhang L, Su H, Zhu J, Ni LM, Shum H-Y (2022) Dino: Detr with improved denoising anchor boxes for end-to-end object detection. arXiv:2203.03605. https:\/\/doi.org\/10.48550\/arXiv.2203.03605","DOI":"10.48550\/arXiv.2203.03605"},{"key":"19362_CR32","doi-asserted-by":"publisher","unstructured":"Lv W, Xu S, Zhao Y, Wang G, Wei J, Cui C, Du Y, Dang Q, Liu Y (2023) Detrs beat yolos on real-time object detection. arXiv:2304.08069. https:\/\/doi.org\/10.48550\/arXiv.2304.08069","DOI":"10.48550\/arXiv.2304.08069"},{"key":"19362_CR33","doi-asserted-by":"publisher","unstructured":"Wang C-Y, Yeh I-H, Liao H-YM (2024) Yolov9: Learning what you want to learn using programmable gradient information. arXiv:2402.13616. https:\/\/doi.org\/10.48550\/arXiv.2402.13616","DOI":"10.48550\/arXiv.2402.13616"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19362-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-19362-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19362-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T05:04:28Z","timestamp":1746075868000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-19362-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,14]]},"references-count":33,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2025,4]]}},"alternative-id":["19362"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-19362-x","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,5,14]]},"assertion":[{"value":"14 January 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"6 March 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"1 May 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 May 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The data in this paper did not have an unethical mandate and the informed consent of all authors for publication of this article has been obtained","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical and informed consent for data used"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper. The authors have no relevant financial or non-financial interests to disclose.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing Interests"}}]}}