{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,19]],"date-time":"2025-03-19T12:06:15Z","timestamp":1742385975269,"version":"3.37.3"},"reference-count":67,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2023,9,4]],"date-time":"2023-09-04T00:00:00Z","timestamp":1693785600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,9,4]],"date-time":"2023-09-04T00:00:00Z","timestamp":1693785600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100012166","name":"National key research and development program","doi-asserted-by":"crossref","award":["2020YFB160070301"],"award-info":[{"award-number":["2020YFB160070301"]}],"id":[{"id":"10.13039\/501100012166","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100015335","name":"Modern Agricultural Technology Industry System of Shandong province","doi-asserted-by":"publisher","award":["CX(20)2013"],"award-info":[{"award-number":["CX(20)2013"]}],"id":[{"id":"10.13039\/501100015335","id-type":"DOI","asserted-by":"publisher"}]},{"name":"the Key R &D Program of Jiangsu Province","award":["BE2019311"],"award-info":[{"award-number":["BE2019311"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-023-16568-3","type":"journal-article","created":{"date-parts":[[2023,9,4]],"date-time":"2023-09-04T08:07:33Z","timestamp":1693814853000},"page":"28197-28213","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["AG-YOLO: Attention-guided network for real-time object detection"],"prefix":"10.1007","volume":"83","author":[{"given":"Hangyu","family":"Zhu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7838-9410","authenticated-orcid":false,"given":"Libo","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Wenhu","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Feng","family":"Tian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,9,4]]},"reference":[{"key":"16568_CR1","doi-asserted-by":"crossref","unstructured":"Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 580\u2013587","DOI":"10.1109\/CVPR.2014.81"},{"key":"16568_CR2","doi-asserted-by":"crossref","unstructured":"Girshick R (2015) Fast r-cnn. In: Proceedings of the IEEE international conference on computer vision, pp 1440\u20131448","DOI":"10.1109\/ICCV.2015.169"},{"issue":"9","key":"16568_CR3","doi-asserted-by":"publisher","first-page":"1904","DOI":"10.1109\/TPAMI.2015.2389824","volume":"37","author":"K He","year":"2015","unstructured":"He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Transactions Pattern Analysis and Machine Intelligence 37(9):1904\u20131916","journal-title":"IEEE Transactions Pattern Analysis and Machine Intelligence"},{"key":"16568_CR4","unstructured":"Ren S, He K, Girshick R, Sun J (2015) Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in Neural Information Processing Systems 28"},{"key":"16568_CR5","doi-asserted-by":"crossref","unstructured":"Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) Ssd: Single shot multibox detector. In: European conference on computer vision, pp 21\u201337. Springer","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"16568_CR6","doi-asserted-by":"crossref","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 779\u2013788","DOI":"10.1109\/CVPR.2016.91"},{"key":"16568_CR7","doi-asserted-by":"crossref","unstructured":"Redmon J, Farhadi A (2017) Yolo9000: better, faster, stronger. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7263\u20137271","DOI":"10.1109\/CVPR.2017.690"},{"key":"16568_CR8","unstructured":"Redmon J, Farhadi A (2018) Yolov3: An incremental improvement. arXiv preprint arXiv:1804.02767"},{"key":"16568_CR9","unstructured":"Bochkovskiy A, Wang C-Y, Liao H-YM (2020) Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934"},{"key":"16568_CR10","unstructured":"Li C, Li L, Jiang H, Weng K, Geng Y, Li L, Ke Z, Li Q, Cheng M, Nie W et al (2022) Yolov6: A single-stage object detection framework for industrial applications. arXiv preprint arXiv:2209.02976"},{"key":"16568_CR11","doi-asserted-by":"crossref","unstructured":"Wang C-Y, Bochkovskiy A, Liao H-YM (2023) Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 7464\u20137475","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"16568_CR12","doi-asserted-by":"crossref","unstructured":"Zhang S, Wen L, Bian X, Lei Z, Li SZ (2018) Single-shot refinement neural network for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4203\u20134212","DOI":"10.1109\/CVPR.2018.00442"},{"key":"16568_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2020.3044719","volume":"70","author":"B Wang","year":"2020","unstructured":"Wang B, Ma F, Ge L, Ma H, Wang H, Mohamed MA (2020) Icing-edgenet: A pruning lightweight edge intelligent method of discriminative driving channel for ice thickness of transmission lines. IEEE Trans Instrum Meas 70:1\u201312","journal-title":"IEEE Trans Instrum Meas"},{"issue":"4","key":"16568_CR14","doi-asserted-by":"publisher","first-page":"683","DOI":"10.3390\/rs13040683","volume":"13","author":"L Huyan","year":"2021","unstructured":"Huyan L, Bai Y, Li Y, Jiang D, Zhang Y, Zhou Q, Wei J, Liu J, Zhang Y, Cui T (2021) A lightweight object detection framework for remote sensing images. Remote Sens 13(4):683","journal-title":"Remote Sens"},{"issue":"22","key":"16568_CR15","doi-asserted-by":"publisher","first-page":"4706","DOI":"10.3390\/rs13224706","volume":"13","author":"M Zhang","year":"2021","unstructured":"Zhang M, Xu S, Song W, He Q, Wei Q (2021) Lightweight underwater object detection based on yolo v4 and multi-scale attentional feature fusion. Remote Sens 13(22):4706","journal-title":"Remote Sens"},{"key":"16568_CR16","doi-asserted-by":"crossref","unstructured":"Yang Y, Sun X, Diao W, Li H, Wu Y, Li X, Fu K (2022) Adaptive knowledge distillation for lightweight remote sensing object detectors optimizing. IEEE Trans Geosci Remote Sens","DOI":"10.1109\/TGRS.2022.3175213"},{"key":"16568_CR17","doi-asserted-by":"crossref","unstructured":"Liu Y, Zhang C, Wu W, Zhang B, Zhou F (2022) Miniyolo: A lightweight object detection algorithm that realizes the trade-off between model size and detection accuracy. Int J Intell Syst","DOI":"10.1002\/int.23079"},{"key":"16568_CR18","doi-asserted-by":"crossref","unstructured":"Zhao H, Shi J, Qi X, Wang X, Jia J (2017) Pyramid scene parsing network. In: Proceedings of the IEEE conference on computer vision and pattern recognition pp 2881\u20142890","DOI":"10.1109\/CVPR.2017.660"},{"issue":"4","key":"16568_CR19","doi-asserted-by":"publisher","first-page":"834","DOI":"10.1109\/TPAMI.2017.2699184","volume":"40","author":"L-C Chen","year":"2017","unstructured":"Chen L-C, Papandreou G, Kokkinos I, Murphy K, Yuille AL (2017) Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Transactions on Pattern Analysis and Machine Intelligence 40(4):834\u2013848","journal-title":"IEEE Transactions on Pattern Analysis and Machine Intelligence"},{"key":"16568_CR20","doi-asserted-by":"crossref","unstructured":"Liu S, Huang D et al (2018) Receptive field block net for accurate and fast object detection. In: Proceedings of the european conference on computer vision (ECCV), pp 385\u2013400","DOI":"10.1007\/978-3-030-01252-6_24"},{"key":"16568_CR21","doi-asserted-by":"crossref","unstructured":"Lin T-Y, Doll\u00e1r P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2117\u20132125","DOI":"10.1109\/CVPR.2017.106"},{"key":"16568_CR22","doi-asserted-by":"crossref","unstructured":"Liu S, Qi L, Qin H, Shi J, Jia J (2018) Path aggregation network for instance segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8759\u20138768","DOI":"10.1109\/CVPR.2018.00913"},{"key":"16568_CR23","doi-asserted-by":"crossref","unstructured":"Ghiasi G, Lin T-Y, Le QV (2019) Nas-fpn: Learning scalable feature pyramid architecture for object detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 7036\u20137045","DOI":"10.1109\/CVPR.2019.00720"},{"key":"16568_CR24","first-page":"1","volume":"71","author":"N Zeng","year":"2022","unstructured":"Zeng N, Wu P, Wang Z, Li H, Liu W, Liu X (2022) A small-sized object detection oriented multi-scale feature fusion approach with application to defect detection. IEEE Trans Instrum Meas 71:1\u201314","journal-title":"IEEE Trans Instrum Meas"},{"issue":"4","key":"16568_CR25","doi-asserted-by":"publisher","first-page":"681","DOI":"10.1109\/LGRS.2019.2930462","volume":"17","author":"J Chen","year":"2019","unstructured":"Chen J, Wan L, Zhu J, Xu G, Deng M (2019) Multi-scale spatial and channel-wise attention for improving object detection in remote sensing imagery. IEEE Geosci Remote Sens Lett 17(4):681\u2013685","journal-title":"IEEE Geosci Remote Sens Lett"},{"key":"16568_CR26","unstructured":"Wang Z, Wang J, Li Y, Wang S (2020) Traffic sign recognition with lightweight two-stage model in complex scenes. IEEE Trans Intell Transp Syst"},{"issue":"8","key":"16568_CR27","doi-asserted-by":"publisher","first-page":"11539","DOI":"10.1007\/s11042-020-10191-2","volume":"80","author":"S Zhou","year":"2021","unstructured":"Zhou S, Qiu J (2021) Enhanced ssd with interactive multi-scale attention features for object detection. Multimedia Tools and Applications 80(8):11539\u201311556","journal-title":"Multimedia Tools and Applications"},{"key":"16568_CR28","doi-asserted-by":"publisher","first-page":"3804","DOI":"10.1109\/TIP.2021.3065239","volume":"30","author":"Y Liu","year":"2021","unstructured":"Liu Y, Zhang X-Y, Bian J-W, Zhang L, Cheng M-M (2021) Samnet: Stereoscopically attentive multi-scale network for lightweight salient object detection. IEEE Trans Image Process 30:3804\u20133814","journal-title":"IEEE Trans Image Process"},{"key":"16568_CR29","doi-asserted-by":"crossref","unstructured":"Li Y, Zhou S, Chen H (2022) Attention-based fusion factor in fpn for object detection. Appl Intell 1\u201310","DOI":"10.1007\/s10489-022-03220-0"},{"issue":"4","key":"16568_CR30","doi-asserted-by":"publisher","first-page":"469","DOI":"10.3390\/jmse10040469","volume":"10","author":"A Li","year":"2022","unstructured":"Li A, Yu L, Tian S (2022) Underwater biological detection based on yolov4 combined with channel attention. Journal of Marine Science and Engineering 10(4):469","journal-title":"Journal of Marine Science and Engineering"},{"issue":"3","key":"16568_CR31","doi-asserted-by":"publisher","first-page":"755","DOI":"10.3390\/rs14030755","volume":"14","author":"K Zhou","year":"2022","unstructured":"Zhou K, Zhang M, Wang H, Tan J (2022) Ship detection in sar images based on multi-scale feature extraction and adaptive feature fusion. Remote Sens 14(3):755","journal-title":"Remote Sens"},{"key":"16568_CR32","doi-asserted-by":"crossref","unstructured":"Dai Y, Gieseke F, Oehmcke S, Wu Y, Barnard K (2021) Attentional feature fusion. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 3560\u20133569","DOI":"10.1109\/WACV48630.2021.00360"},{"key":"16568_CR33","unstructured":"Everingham M (2007) The pascal visual object classes challenge,(voc2007) results. http:\/\/pascallin.ecs.soton.ac.uk\/challenges\/VOC\/voc2007\/index.html"},{"key":"16568_CR34","first-page":"1","volume":"2007","author":"M Everingham","year":"2012","unstructured":"Everingham M, Winn J (2012) The pascal visual object classes challenge 2012 (voc2012) development kit. Pattern Anal Stat Model Comput Learn Tech Rep 2007:1\u201345","journal-title":"Pattern Anal Stat Model Comput Learn Tech Rep"},{"key":"16568_CR35","doi-asserted-by":"crossref","unstructured":"Cai Z, Vasconcelos N (2018) Cascade r-cnn: Delving into high quality object detection. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6154\u20136162","DOI":"10.1109\/CVPR.2018.00644"},{"issue":"5","key":"16568_CR36","doi-asserted-by":"publisher","first-page":"4226","DOI":"10.1109\/JIOT.2022.3215469","volume":"10","author":"Z Ou","year":"2022","unstructured":"Ou Z, Wang Z, Xiao F, Xiong B, Zhang H, Song M, Zheng Y, Hui P (2022) Ad-rcnn: Adaptive dynamic neural network for small object detection. IEEE Internet of Things Journal 10(5):4226\u20134238","journal-title":"IEEE Internet of Things Journal"},{"key":"16568_CR37","doi-asserted-by":"crossref","unstructured":"Wang C-Y, Liao H-YM, Wu Y-H, Chen P-Y, Hsieh J-W, Yeh I-H (2020) Cspnet: A new backbone that can enhance learning capability of cnn. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition workshops, pp 390\u2013391","DOI":"10.1109\/CVPRW50498.2020.00203"},{"key":"16568_CR38","doi-asserted-by":"publisher","DOI":"10.1016\/j.eswa.2022.117682","volume":"206","author":"K Wang","year":"2022","unstructured":"Wang K, Wang Y, Zhang S, Tian Y, Li D (2022) Slms-ssd: Improving the balance of semantic and spatial information in object detection. Expert Syst Appl 206:117682","journal-title":"Expert Syst Appl"},{"issue":"2","key":"16568_CR39","doi-asserted-by":"publisher","first-page":"1586","DOI":"10.1007\/s10489-022-03549-6","volume":"53","author":"G Wen","year":"2023","unstructured":"Wen G, Cao P, Wang H, Chen H, Liu X, Xu J, Zaiane O (2023) Ms-ssd: Multi-scale single shot detector for ship detection in remote sensing images. Appl Intell 53(2):1586\u20131604","journal-title":"Appl Intell"},{"key":"16568_CR40","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"16568_CR41","doi-asserted-by":"crossref","unstructured":"Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, Erhan D, Vanhoucke V, Rabinovich A (2015) Going deeper with convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1\u20139","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"16568_CR42","unstructured":"Ioffe S, Szegedy C (2015) Batch normalization: Accelerating deep network training by reducing internal covariate shift. In: International conference on machine learning, pp 448\u2013456"},{"key":"16568_CR43","doi-asserted-by":"crossref","unstructured":"Szegedy C, Vanhoucke V, Ioffe S, Shlens J, Wojna Z (2016) Rethinking the inception architecture for computer vision. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2818\u20132826","DOI":"10.1109\/CVPR.2016.308"},{"key":"16568_CR44","doi-asserted-by":"crossref","unstructured":"Szegedy C, Ioffe S, Vanhoucke V, Alemi AA (2017) Inception-v4, inception-resnet and the impact of residual connections on learning. In: Thirty-first AAAI conference on artificial intelligence","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"16568_CR45","unstructured":"Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, Andreetto M, Adam H (2017) Mobilenets: Efficient convolutional neural networks for mobile vision applications.arXiv preprint arXiv:1704.04861"},{"key":"16568_CR46","doi-asserted-by":"crossref","unstructured":"Sandler M, Howard A, Zhu M, Zhmoginov A, Chen L-C (2018) Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4510\u20134520","DOI":"10.1109\/CVPR.2018.00474"},{"key":"16568_CR47","doi-asserted-by":"crossref","unstructured":"Howard A, Sandler M, Chu G, Chen L-C, Chen B, Tan M, Wang W, Zhu Y, Pang R, Vasudevan V et al (2019) Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 1314\u20131324","DOI":"10.1109\/ICCV.2019.00140"},{"key":"16568_CR48","doi-asserted-by":"crossref","unstructured":"Zhang X, Zhou X, Lin M, Sun J (2018) Shufflenet: An extremely efficient convolutional neural network for mobile devices. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6848\u20136856","DOI":"10.1109\/CVPR.2018.00716"},{"key":"16568_CR49","doi-asserted-by":"crossref","unstructured":"Ma N, Zhang X, Zheng H-T, Sun J (2018) Shufflenet v2: Practical guidelines for efficient cnn architecture design. In: Proceedings of the european conference on computer vision (ECCV), pp 116\u2013131","DOI":"10.1007\/978-3-030-01264-9_8"},{"key":"16568_CR50","doi-asserted-by":"crossref","unstructured":"Han K, Wang Y, Tian Q, Guo J, Xu C, Xu C (2020) Ghostnet: More features from cheap operations. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 1580\u20131589","DOI":"10.1109\/CVPR42600.2020.00165"},{"key":"16568_CR51","unstructured":"Hinton G, Vinyals O, Dean J et al (2015) Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531"},{"key":"16568_CR52","doi-asserted-by":"crossref","unstructured":"Fan D-P, Wang W, Cheng M-M, Shen J (2019) Shifting more attention to video salient object detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 8554\u20138564","DOI":"10.1109\/CVPR.2019.00875"},{"key":"16568_CR53","doi-asserted-by":"crossref","unstructured":"Wang X, Girshick R, Gupta A, He K (2018) Non-local neural networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7794\u20137803","DOI":"10.1109\/CVPR.2018.00813"},{"key":"16568_CR54","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"16568_CR55","doi-asserted-by":"crossref","unstructured":"Cao Y, Xu J, Lin S, Wei F, Hu H (2019) Gcnet: Non-local networks meet squeeze-excitation networks and beyond. In: Proceedings of the IEEE\/CVF international conference on computer vision workshops, pp 0\u20130","DOI":"10.1109\/ICCVW.2019.00246"},{"key":"16568_CR56","unstructured":"Park J, Woo S, Lee J-Y, Kweon IS (2018) Bam: Bottleneck attention module. arXiv preprint arXiv:1807.06514"},{"key":"16568_CR57","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: Convolutional block attention module. In: Proceedings of the european conference on computer vision (ECCV), pp 3\u201319","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"16568_CR58","doi-asserted-by":"crossref","unstructured":"Fu J, Liu J, Tian H, Li Y, Bao Y, Fang Z, Lu H (2019) Dual attention network for scene segmentation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 3146\u20133154","DOI":"10.1109\/CVPR.2019.00326"},{"key":"16568_CR59","doi-asserted-by":"crossref","unstructured":"Zhang Q-L, Yang Y-B (2021) Sa-net: Shuffle attention for deep convolutional neural networks. In: ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp 2235\u20132239","DOI":"10.1109\/ICASSP39728.2021.9414568"},{"key":"16568_CR60","doi-asserted-by":"crossref","unstructured":"Zhu X, Cheng D, Zhang Z, Lin S, Dai J (2019) An empirical study of spatial attention mechanisms in deep networks. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 6688\u20136697","DOI":"10.1109\/ICCV.2019.00679"},{"key":"16568_CR61","doi-asserted-by":"crossref","unstructured":"Qin Z, Zhang P, Wu F, Li X (2021) Fcanet: Frequency channel attention networks. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 783\u2013792","DOI":"10.1109\/ICCV48922.2021.00082"},{"key":"16568_CR62","doi-asserted-by":"crossref","unstructured":"Yu C, Wang J, Peng C, Gao C, Yu G, Sang N (2018) Bisenet: Bilateral segmentation network for real-time semantic segmentation. In: Proceedings of the european conference on computer vision (ECCV), pp 325\u2013341","DOI":"10.1007\/978-3-030-01261-8_20"},{"issue":"5","key":"16568_CR63","doi-asserted-by":"publisher","first-page":"3545","DOI":"10.1007\/s11063-021-10560-4","volume":"53","author":"L Mao","year":"2021","unstructured":"Mao L, Li X, Yang D, Zhang R (2021) Convolutional feature frequency adaptive fusion object detection network. Neural Process Lett 53(5):3545\u20133560","journal-title":"Neural Process Lett"},{"key":"16568_CR64","unstructured":"Tan M, Le QV (2019) Mixconv: Mixed depthwise convolutional kernels. arXiv preprint arXiv:1907.09595"},{"key":"16568_CR65","doi-asserted-by":"crossref","unstructured":"Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Supplementary material for \u2019eca-net: Efficient channel attention for deep convolutional neural networks. In: Proceedings of the 2020 IEEE\/CVF conference on computer vision and pattern recognition, IEEE, Seattle, WA, USA, pp 13\u201319","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"16568_CR66","unstructured":"Lin M, Chen Q, Yan S (2013) Network in network. arXiv preprint arXiv:1312.4400"},{"key":"16568_CR67","doi-asserted-by":"crossref","unstructured":"Chen L-C, Zhu Y, Papandreou G, Schroff F, Adam H (2018) Encoder-decoder with atrous separable convolution for semantic image segmentation. In: Proceedings of the european conference on computer vision (ECCV), pp 801\u2013818","DOI":"10.1007\/978-3-030-01234-2_49"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16568-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-023-16568-3\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-023-16568-3.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,2,29]],"date-time":"2024-02-29T10:54:21Z","timestamp":1709204061000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-023-16568-3"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,9,4]]},"references-count":67,"journal-issue":{"issue":"9","published-online":{"date-parts":[[2024,3]]}},"alternative-id":["16568"],"URL":"https:\/\/doi.org\/10.1007\/s11042-023-16568-3","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2023,9,4]]},"assertion":[{"value":"12 January 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 July 2023","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"18 August 2023","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 September 2023","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}]}}