{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,10]],"date-time":"2026-02-10T09:00:49Z","timestamp":1770714049228,"version":"3.49.0"},"reference-count":58,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2023,3,9]],"date-time":"2023-03-09T00:00:00Z","timestamp":1678320000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,3,9]],"date-time":"2023-03-09T00:00:00Z","timestamp":1678320000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"Key-Area Research and Development Program of Guangdong Province","award":["2020B1111010002"],"award-info":[{"award-number":["2020B1111010002"]}]},{"name":"Key-Area Research and Development Program of Guangdong Province","award":["2018B010109001"],"award-info":[{"award-number":["2018B010109001"]}]},{"name":"2021 Guangdong Provincial Science and Technology Special Fund","award":["210719145863737"],"award-info":[{"award-number":["210719145863737"]}]},{"DOI":"10.13039\/501100018568","name":"Special Fund Project for Science and Technology Innovation Strategy of Guangdong Province","doi-asserted-by":"publisher","award":["[2020]018"],"award-info":[{"award-number":["[2020]018"]}],"id":[{"id":"10.13039\/501100018568","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Laboratory of Autonomous Systems and Network Control of Ministry of Education"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Vis Comput"],"published-print":{"date-parts":[[2024,2]]},"DOI":"10.1007\/s00371-023-02809-x","type":"journal-article","created":{"date-parts":[[2023,3,26]],"date-time":"2023-03-26T21:05:47Z","timestamp":1679864747000},"page":"681-698","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["An adaptive loss weighting multi-task network with attention-guide proposal generation for small size defect inspection"],"prefix":"10.1007","volume":"40","author":[{"given":"Huangyuan","family":"Wu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-6767-5350","authenticated-orcid":false,"given":"Bin","family":"Li","sequence":"additional","affiliation":[]},{"given":"Lianfang","family":"Tian","sequence":"additional","affiliation":[]},{"given":"Junjian","family":"Feng","sequence":"additional","affiliation":[]},{"given":"Chao","family":"Dong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,3,9]]},"reference":[{"key":"2809_CR1","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-021-02210-6","author":"W Hu","year":"2021","unstructured":"Hu, W., Wang, T., Wang, Y., Chen, Z., Huang, G.: LE\u2013MSFE\u2013DDNet: a defect detection network based on low-light enhancement and multi-scale feature extraction. Vis. Comput. (2021). https:\/\/doi.org\/10.1007\/s00371-021-02210-6","journal-title":"Vis. Comput."},{"key":"2809_CR2","doi-asserted-by":"publisher","first-page":"639","DOI":"10.1007\/s00371-020-02040-y","volume":"38","author":"G Liu","year":"2022","unstructured":"Liu, G., Li, F.: Fabric defect detection based on low-rank decomposition with structural constraints. Vis. Comput. 38, 639\u2013653 (2022). https:\/\/doi.org\/10.1007\/s00371-020-02040-y","journal-title":"Vis. Comput."},{"key":"2809_CR3","doi-asserted-by":"publisher","first-page":"7293","DOI":"10.1109\/tie.2017.2682037","volume":"64","author":"L Wang","year":"2017","unstructured":"Wang, L., Zhang, Z.: Automatic detection of wind turbine blade surface cracks based on UAV-taken images. IEEE Trans. Ind. Electron. 64, 7293\u20137303 (2017). https:\/\/doi.org\/10.1109\/tie.2017.2682037","journal-title":"IEEE Trans. Ind. Electron."},{"key":"2809_CR4","doi-asserted-by":"publisher","first-page":"2824","DOI":"10.1109\/tsg.2016.2621135","volume":"9","author":"L Wang","year":"2018","unstructured":"Wang, L., Zhang, Z., Xu, J., Liu, R.: Wind turbine blade breakage monitoring with deep autoencoders. IEEE Trans. Smart Grid 9, 2824\u20132833 (2018). https:\/\/doi.org\/10.1109\/tsg.2016.2621135","journal-title":"IEEE Trans. Smart Grid"},{"key":"2809_CR5","doi-asserted-by":"publisher","first-page":"1271","DOI":"10.1109\/TMECH.2019.2908233","volume":"24","author":"L Wang","year":"2019","unstructured":"Wang, L., Zhang, Z., Luo, X.: A two-stage data-driven approach for image-based wind turbine blade crack inspections. IEEE\/ASME Trans. Mechatron. 24, 1271\u20131281 (2019). https:\/\/doi.org\/10.1109\/TMECH.2019.2908233","journal-title":"IEEE\/ASME Trans. Mechatron."},{"key":"2809_CR6","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1016\/j.neucom.2019.09.071","volume":"376","author":"YJ Yu","year":"2020","unstructured":"Yu, Y.J., Cao, H., Yan, X.Y., Wang, T., Ge, S.S.: Defect identification of wind turbine blades based on defect semantic features with transfer feature extractor. Neurocomputing 376, 1\u20139 (2020). https:\/\/doi.org\/10.1016\/j.neucom.2019.09.071","journal-title":"Neurocomputing"},{"key":"2809_CR7","doi-asserted-by":"publisher","first-page":"1493","DOI":"10.1109\/TIM.2019.2915404","volume":"69","author":"Y He","year":"2020","unstructured":"He, Y., Song, K., Meng, Q., Yan, Y.: An end-to-end steel surface defect detection approach via fusing multiple hierarchical features. IEEE Trans. Instrum. Meas. 69, 1493\u20131504 (2020). https:\/\/doi.org\/10.1109\/TIM.2019.2915404","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2809_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3176239","volume":"71","author":"CC Yeung","year":"2022","unstructured":"Yeung, C.C., Lam, K.M.: Efficient fused-attention model for steel surface defect detection. IEEE Trans. Instrum. Meas. 71, 1\u201311 (2022). https:\/\/doi.org\/10.1109\/TIM.2022.3176239","journal-title":"IEEE Trans. Instrum. Meas."},{"key":"2809_CR9","doi-asserted-by":"publisher","first-page":"1343","DOI":"10.1109\/TII.2019.2945403","volume":"16","author":"J Lian","year":"2020","unstructured":"Lian, J., Jia, W., Zareapoor, M., Zheng, Y., Luo, R., Jain, D.K., Kumar, N.: Deep-learning-based small surface defect detection via an exaggerated local variation-based generative adversarial network. IEEE Trans. Ind. Inf. 16, 1343\u20131351 (2020). https:\/\/doi.org\/10.1109\/TII.2019.2945403","journal-title":"IEEE Trans. Ind. Inf."},{"key":"2809_CR10","doi-asserted-by":"publisher","first-page":"936","DOI":"10.1109\/TSMC.2020.3005231","volume":"52","author":"G Chen","year":"2022","unstructured":"Chen, G., Wang, H., Chen, K., Li, Z., Song, Z., Liu, Y., Chen, W., Knoll, A.: A survey of the four pillars for small object detection: multiscale representation, contextual information, super-resolution, and region proposal. IEEE Trans. Syst. Man Cybern. Syst. 52, 936\u2013953 (2022). https:\/\/doi.org\/10.1109\/TSMC.2020.3005231","journal-title":"IEEE Trans. Syst. Man Cybern. Syst."},{"key":"2809_CR11","doi-asserted-by":"publisher","first-page":"541","DOI":"10.1109\/TIP.2021.3132828","volume":"31","author":"Y Liu","year":"2022","unstructured":"Liu, Y., Xia, C., Zhu, X., Xu, S.: Two-stage copy-move forgery detection with self deep matching and proposal SuperGlue. IEEE Trans. Image Process. 31, 541\u2013555 (2022). https:\/\/doi.org\/10.1109\/TIP.2021.3132828","journal-title":"IEEE Trans. Image Process."},{"key":"2809_CR12","doi-asserted-by":"publisher","first-page":"1331","DOI":"10.1109\/TGRS.2020.3005151","volume":"59","author":"J Fu","year":"2021","unstructured":"Fu, J., Sun, X., Wang, Z., Fu, K.: An anchor-free method based on feature balancing and refinement network for multiscale ship detection in SAR images. IEEE Trans. Geosci. Remote Sens. 59, 1331\u20131344 (2021). https:\/\/doi.org\/10.1109\/TGRS.2020.3005151","journal-title":"IEEE Trans. Geosci. Remote Sens."},{"key":"2809_CR13","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2017","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 39, 1137\u20131149 (2017). https:\/\/doi.org\/10.1109\/TPAMI.2016.2577031","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2809_CR14","doi-asserted-by":"publisher","first-page":"520","DOI":"10.1016\/j.tics.2007.09.009","volume":"11","author":"A Oliva","year":"2007","unstructured":"Oliva, A., Torralba, A.: The role of context in object recognition. Trends Cogn. Sci. 11, 520\u2013527 (2007). https:\/\/doi.org\/10.1016\/j.tics.2007.09.009","journal-title":"Trends Cogn. Sci."},{"key":"2809_CR15","doi-asserted-by":"publisher","unstructured":"Jha, A., Kumar, A., Banerjee, B., Chaudhuri, S.: AdaMT-Net: an adaptive weight learning based multi-task learning model for scene understanding. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops. https:\/\/doi.org\/10.1109\/cvprw50498.2020.00361 (2020)","DOI":"10.1109\/cvprw50498.2020.00361"},{"key":"2809_CR16","doi-asserted-by":"crossref","unstructured":"Jha, A., Kumar, A., Banerjee, B., Chaudhuri, S.: AdaMT-Net: an adaptive weight learning based multi-task learning model for scene understanding. In: Proceedings of 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 14\u201319 June 2020, pp. 3027\u20133035 (2020)","DOI":"10.1109\/CVPRW50498.2020.00361"},{"key":"2809_CR17","unstructured":"Chen, Z., Badrinarayanan, V., Lee, C.-Y., Rabinovich, A.: Gradnorm: gradient normalization for adaptive loss balancing in deep multitask networks. In: Proceedings of International Conference on Machine Learning. PMLR, pp. 794\u2013803 (2018)"},{"key":"2809_CR18","unstructured":"Lin, X., Baweja, H.S., Kantor, G., Held, D.: Adaptive auxiliary task weighting for reinforcement learning. In Proceedings of 33rd Conference on Neural Information Processing Systems (NeurIPS), Vancouver, CANADA, Dec 08\u201314. Advances in Neural Information Processing Systems (2019)"},{"key":"2809_CR19","doi-asserted-by":"publisher","first-page":"121","DOI":"10.1109\/TPAMI.2017.2781233","volume":"41","author":"R Ranjan","year":"2019","unstructured":"Ranjan, R., Patel, V.M., Chellappa, R.: HyperFace: a deep multi-task learning framework for face detection, landmark localization, pose estimation, and gender recognition. IEEE Trans. Pattern Anal. Mach. Intell. 41, 121\u2013135 (2019). https:\/\/doi.org\/10.1109\/TPAMI.2017.2781233","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2809_CR20","doi-asserted-by":"publisher","DOI":"10.1007\/s00371-022-02472-8","author":"H Yang","year":"2022","unstructured":"Yang, H., Fan, Y., Lv, G., Liu, S., Guo, Z.: Exploiting emotional concepts for image emotion recognition. Vis. Comput. (2022). https:\/\/doi.org\/10.1007\/s00371-022-02472-8","journal-title":"Vis. Comput."},{"key":"2809_CR21","doi-asserted-by":"publisher","first-page":"2473","DOI":"10.1007\/s00371-021-02124-3","volume":"38","author":"M Jiang","year":"2022","unstructured":"Jiang, M., Zhai, F., Kong, J.: Sparse attention module for optimizing semantic segmentation performance combined with a multi-task feature extraction network. Vis. Comput. 38, 2473\u20132488 (2022). https:\/\/doi.org\/10.1007\/s00371-021-02124-3","journal-title":"Vis. Comput."},{"key":"2809_CR22","doi-asserted-by":"crossref","unstructured":"Cipolla, R., Gal, Y., Kendall, A.: Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition, 18\u201323 June 2018, pp. 7482\u20137491 (2018)","DOI":"10.1109\/CVPR.2018.00781"},{"key":"2809_CR23","doi-asserted-by":"crossref","unstructured":"Gao, Y., Ma, J., Zhao, M., Liu, W., Yuille, A.L.: NDDR-CNN: layerwise feature fusing in multi-task CNNs by neural discriminative dimensionality reduction. In Proceedings of 2019 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 15\u201320 June 2019, pp. 3200\u20133209 (2019)","DOI":"10.1109\/CVPR.2019.00332"},{"key":"2809_CR24","doi-asserted-by":"crossref","unstructured":"Misra, I., Shrivastava, A., Gupta, A., Hebert, M.: Cross-stitch networks for multi-task learning. In: Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 27\u201330 June 2016, pp. 3994\u20134003 (2016)","DOI":"10.1109\/CVPR.2016.433"},{"key":"2809_CR25","doi-asserted-by":"publisher","DOI":"10.1109\/TPAMI.2021.3054719","author":"S Vandenhende","year":"2021","unstructured":"Vandenhende, S., Georgoulis, S., Gansbeke, W.V., Proesmans, M., Dai, D., Gool, L.V.: Multi-task learning for dense prediction tasks: a survey. IEEE Trans. Pattern Anal. Mach. Intell. (2021). https:\/\/doi.org\/10.1109\/TPAMI.2021.3054719","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"2809_CR26","doi-asserted-by":"crossref","unstructured":"Wilms, C., Frintrop, S.: AttentionMask: attentive, efficient object proposal generation focusing on small objects. In: Proceedings of 14th Asian Conference on Computer Vision (ACCV), Perth, Australia, Dec 02\u201306. Lecture Notes in Computer Science, pp. 678\u2013694 (2018)","DOI":"10.1007\/978-3-030-20890-5_43"},{"key":"2809_CR27","doi-asserted-by":"publisher","DOI":"10.3390\/rs13173362","author":"RC Dong","year":"2021","unstructured":"Dong, R.C., Jiao, L.C., Zhang, Y., Zhao, J., Shen, W.Y.: A multi-scale spatial attention region proposal network for high-resolution optical remote sensing imagery. Remote Sens. (2021). https:\/\/doi.org\/10.3390\/rs13173362","journal-title":"Remote Sens."},{"key":"2809_CR28","doi-asserted-by":"crossref","unstructured":"Quan, Y., Li, Z.X., Zhang, C.L., Ma, H.F., IEEE Computer Society: Object detection model based on scene-level region proposal self-attention. In: Proceedings of 25th International Conference on Pattern Recognition (ICPR), Electrical Network, Jan 10\u201315. International Conference on Pattern Recognition, pp. 954\u2013961 (2021)","DOI":"10.1109\/ICPR48806.2021.9412726"},{"key":"2809_CR29","doi-asserted-by":"crossref","unstructured":"Guo, M., Haque, A., Huang, D.-A., Yeung, S., Fei-Fei, L.: Dynamic task prioritization for multitask learning. In: Proceedings of Proceedings of the European Conference on Computer Vision (ECCV), pp. 270\u2013287 (2018)","DOI":"10.1007\/978-3-030-01270-0_17"},{"key":"2809_CR30","doi-asserted-by":"crossref","unstructured":"Lin, T., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 21\u201326 July 2017, pp. 936\u2013944 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"2809_CR31","doi-asserted-by":"publisher","first-page":"16153","DOI":"10.1007\/s11042-019-08578-x","volume":"80","author":"JH Zhang","year":"2021","unstructured":"Zhang, J.H., Min, X.K., Jia, J., Zhu, Z.H., Wang, J., Zhai, G.T.: Fine localization and distortion resistant detection of multi-class barcode in complex environments. Multimed. Tools Appl. 80, 16153\u201316172 (2021). https:\/\/doi.org\/10.1007\/s11042-019-08578-x","journal-title":"Multimed. Tools Appl."},{"key":"2809_CR32","doi-asserted-by":"crossref","unstructured":"Bell, S., Zitnick, C.L., Bala, K., Girshick, R.: Inside-outside net: detecting objects in context with skip pooling and recurrent neural networks. In: Proceedings of 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 27\u201330 June 2016, pp. 2874\u20132883 (2016)","DOI":"10.1109\/CVPR.2016.314"},{"key":"2809_CR33","doi-asserted-by":"publisher","first-page":"9919","DOI":"10.1109\/JIOT.2019.2933254","volume":"6","author":"J Jia","year":"2019","unstructured":"Jia, J., Zhai, G., Zhang, J., Gao, Z., Zhu, Z., Min, X., Yang, X., Guo, G.: EMBDN: an efficient multiclass barcode detection network for complicated environments. IEEE Internet Things J. 6, 9919\u20139933 (2019). https:\/\/doi.org\/10.1109\/JIOT.2019.2933254","journal-title":"IEEE Internet Things J."},{"key":"2809_CR34","doi-asserted-by":"publisher","first-page":"688","DOI":"10.1109\/JSTSP.2020.2976566","volume":"14","author":"J Jia","year":"2020","unstructured":"Jia, J., Zhai, G., Ren, P., Zhang, J., Gao, Z., Min, X., Yang, X.: Tiny-BDN: an efficient and compact barcode detection network. IEEE J. Sel. Top. Signal Process. 14, 688\u2013699 (2020). https:\/\/doi.org\/10.1109\/JSTSP.2020.2976566","journal-title":"IEEE J. Sel. Top. Signal Process."},{"key":"2809_CR35","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/LGRS.2021.3103069","volume":"19","author":"M Hong","year":"2022","unstructured":"Hong, M., Li, S., Yang, Y., Zhu, F., Zhao, Q., Lu, L.: SSPNet: scale selection pyramid network for tiny person detection from UAV images. IEEE Geosci. Remote Sens. Lett. 19, 1\u20135 (2022). https:\/\/doi.org\/10.1109\/LGRS.2021.3103069","journal-title":"IEEE Geosci. Remote Sens. Lett."},{"key":"2809_CR36","unstructured":"Bochkovskiy, A., Wang, C.-Y., Liao, H.-Y.M.Jae-p: YOLOv4: optimal speed and accuracy of object detection. arXiv:2004.10934 (2020)"},{"key":"2809_CR37","unstructured":"Redmon, J., Farhadi, A.Jae-p.: YOLOv3: an incremental improvement. arXiv:1804.02767 (2018)"},{"key":"2809_CR38","unstructured":"Zheng, Z., Wang, P., Liu, W., Li, J., Ren, D.: Distance-IoU loss: faster and better learning for bounding box regression. In: Proceedings of AAAI Conference on Artificial Intelligence"},{"key":"2809_CR39","doi-asserted-by":"publisher","first-page":"4188","DOI":"10.1109\/TCSVT.2021.3126590","volume":"32","author":"Y Zhu","year":"2022","unstructured":"Zhu, Y., Zhai, G., Yang, Y., Duan, H., Min, X., Yang, X.: Viewing behavior supported visual saliency predictor for 360 degree videos. IEEE Trans. Circuits Syst. Video Technol. 32, 4188\u20134201 (2022). https:\/\/doi.org\/10.1109\/TCSVT.2021.3126590","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2809_CR40","doi-asserted-by":"publisher","first-page":"2331","DOI":"10.1109\/TMM.2019.2957986","volume":"22","author":"Y Zhu","year":"2020","unstructured":"Zhu, Y., Zhai, G., Min, X., Zhou, J.: The prediction of saliency map for head and eye movements in 360 degree images. iEEE Trans. Multimed. 22, 2331\u20132344 (2020). https:\/\/doi.org\/10.1109\/TMM.2019.2957986","journal-title":"iEEE Trans. Multimed."},{"key":"2809_CR41","doi-asserted-by":"publisher","first-page":"3805","DOI":"10.1109\/TIP.2020.2966082","volume":"29","author":"X Min","year":"2020","unstructured":"Min, X., Zhai, G., Zhou, J., Zhang, X.P., Yang, X., Guan, X.: A multimodal saliency model for videos with high audio-visual correspondence. IEEE Trans. Image Process. 29, 3805\u20133819 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.2966082","journal-title":"IEEE Trans. Image Process."},{"key":"2809_CR42","doi-asserted-by":"crossref","unstructured":"Milletari, F., Navab, N., Ahmadi, S.: V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In: Proceedings of 2016 Fourth International Conference on 3D Vision (3DV), 25\u201328 Oct. 2016, pp. 565\u2013571 (2016)","DOI":"10.1109\/3DV.2016.79"},{"key":"2809_CR43","doi-asserted-by":"crossref","unstructured":"Zeiler, M.D., Fergus, R.: Visualizing and understanding convolutional networks. In: Proceedings of Computer Vision\u2014ECCV 2014. Springer, Cham, pp. 818\u2013833 (2014)","DOI":"10.1007\/978-3-319-10590-1_53"},{"key":"2809_CR44","doi-asserted-by":"publisher","first-page":"15","DOI":"10.3390\/en12040676","volume":"12","author":"ASM Shihavuddin","year":"2019","unstructured":"Shihavuddin, A.S.M., Chen, X., Fedorov, V., Christensen, A.N., Riis, N.A.B., Branner, K., Dahl, A.B., Paulsen, R.R.: Wind turbine surface damage detection by deep learning aided drone inspection analysis. Energies 12, 15 (2019). https:\/\/doi.org\/10.3390\/en12040676","journal-title":"Energies"},{"key":"2809_CR45","doi-asserted-by":"crossref","unstructured":"Krishna, H., Jawahar, C.V.: Improving small object detection. In: Proceedings of 2017 4th IAPR Asian Conference on Pattern Recognition (ACPR), 26\u201329 Nov. 2017, pp. 340\u2013345 (2017)","DOI":"10.1109\/ACPR.2017.149"},{"key":"2809_CR46","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A.C.: SSD: single shot multibox detector. In: Proceedings of Computer Vision\u2014ECCV 2016. Springer, pp. 21\u201337 (2016)","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"2809_CR47","unstructured":"Redmon, J., Farhadi, A.Jae-p: YOLOv3: an incremental improvement (2018)"},{"key":"2809_CR48","doi-asserted-by":"crossref","unstructured":"Tan, M., Pang, R., Le, Q.V.: EfficientDet: scalable and efficient object detection. In: Proceedings of 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 13\u201319 June 2020, pp. 10778\u201310787 (2020)","DOI":"10.1109\/CVPR42600.2020.01079"},{"key":"2809_CR49","doi-asserted-by":"crossref","unstructured":"Duan, K., Bai, S., Xie, L., Qi, H., Huang, Q., Tian, Q.: CenterNet: keypoint triplets for object detection. In: Proceedings of 2019 IEEE\/CVF International Conference on Computer Vision (ICCV), 27 Oct.\u20132 Nov. 2019, pp. 6568\u20136577 (2019)","DOI":"10.1109\/ICCV.2019.00667"},{"key":"2809_CR50","unstructured":"Ge, Z., Liu, S., Wang, F., Li, Z., Sun, J.: YOLOX: exceeding YOLO series in 2021 (2021). arxiv:2107.08430"},{"key":"2809_CR51","doi-asserted-by":"publisher","first-page":"1639","DOI":"10.1109\/TCSVT.2019.2906246","volume":"30","author":"K Duan","year":"2020","unstructured":"Duan, K., Du, D., Qi, H., Huang, Q.: Detecting small objects using a channel-aware deconvolutional network. IEEE Trans. Circuits Syst. Video Technol. 30, 1639\u20131652 (2020). https:\/\/doi.org\/10.1109\/TCSVT.2019.2906246","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2809_CR52","doi-asserted-by":"publisher","first-page":"1758","DOI":"10.1109\/TCSVT.2019.2905881","volume":"30","author":"X Liang","year":"2020","unstructured":"Liang, X., Zhang, J., Zhuo, L., Li, Y., Tian, Q.: Small object detection in unmanned aerial vehicle images using feature fusion and scaling-based single shot detector with spatial context analysis. IEEE Trans. Circuits Syst. Video Technol. 30, 1758\u20131770 (2020). https:\/\/doi.org\/10.1109\/TCSVT.2019.2905881","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"2809_CR53","doi-asserted-by":"publisher","first-page":"66","DOI":"10.1109\/LSP.2020.3045638","volume":"28","author":"D Parashar","year":"2021","unstructured":"Parashar, D., Agrawal, D.K.: Automatic classification of glaucoma stages using two-dimensional tensor empirical wavelet transform. IEEE Signal Process. Lett. 28, 66\u201370 (2021). https:\/\/doi.org\/10.1109\/LSP.2020.3045638","journal-title":"IEEE Signal Process. Lett."},{"key":"2809_CR54","doi-asserted-by":"publisher","DOI":"10.1007\/s10278-022-00648-1","author":"D Parashar","year":"2022","unstructured":"Parashar, D., Agrawal, D.K.: Classification of glaucoma stages using image empirical mode decomposition from fundus images. J. Digit. Imaging (2022). https:\/\/doi.org\/10.1007\/s10278-022-00648-1","journal-title":"J. Digit. Imaging"},{"key":"2809_CR55","doi-asserted-by":"publisher","first-page":"5462","DOI":"10.1109\/TIP.2017.2735192","volume":"26","author":"X Min","year":"2017","unstructured":"Min, X., Ma, K., Gu, K., Zhai, G., Wang, Z., Lin, W.: Unified blind quality assessment of compressed natural, graphic, and screen content images. IEEE Trans. Image Process. 26, 5462\u20135474 (2017). https:\/\/doi.org\/10.1109\/TIP.2017.2735192","journal-title":"IEEE Trans. Image Process."},{"key":"2809_CR56","doi-asserted-by":"publisher","DOI":"10.1145\/3470970","author":"XK Min","year":"2022","unstructured":"Min, X.K., Gu, K., Zhai, G.T., Yang, X.K., Zhang, W.J., Le Callet, P., Chen, C.W.: Screen content quality assessment: overview, benchmark, and beyond. ACM Comput Surv (2022). https:\/\/doi.org\/10.1145\/3470970","journal-title":"ACM Comput Surv"},{"key":"2809_CR57","doi-asserted-by":"publisher","first-page":"6054","DOI":"10.1109\/TIP.2020.2988148","volume":"29","author":"X Min","year":"2020","unstructured":"Min, X., Zhai, G., Zhou, J., Farias, M.C.Q., Bovik, A.C.: Study of subjective and objective quality assessment of audio-visual signals. IEEE Trans. Image Process. 29, 6054\u20136068 (2020). https:\/\/doi.org\/10.1109\/TIP.2020.2988148","journal-title":"IEEE Trans. Image Process."},{"key":"2809_CR58","doi-asserted-by":"publisher","first-page":"10","DOI":"10.1007\/s11432-019-2757-1","volume":"63","author":"GT Zhai","year":"2020","unstructured":"Zhai, G.T., Min, X.K.: Perceptual image quality assessment: a survey. Sci. China Inf. Sci. 63, 10 (2020). https:\/\/doi.org\/10.1007\/s11432-019-2757-1","journal-title":"Sci. China Inf. Sci."}],"container-title":["The Visual Computer"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02809-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00371-023-02809-x\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00371-023-02809-x.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,1,23]],"date-time":"2024-01-23T19:04:43Z","timestamp":1706036683000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00371-023-02809-x"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,3,9]]},"references-count":58,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,2]]}},"alternative-id":["2809"],"URL":"https:\/\/doi.org\/10.1007\/s00371-023-02809-x","relation":{},"ISSN":["0178-2789","1432-2315"],"issn-type":[{"value":"0178-2789","type":"print"},{"value":"1432-2315","type":"electronic"}],"subject":[],"published":{"date-parts":[[2023,3,9]]},"assertion":[{"value":"9 February 2023","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 March 2023","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}