{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,2]],"date-time":"2026-04-02T08:53:26Z","timestamp":1775120006056,"version":"3.50.1"},"reference-count":27,"publisher":"Springer Science and Business Media LLC","issue":"1","license":[{"start":{"date-parts":[[2025,11,11]],"date-time":"2025-11-11T00:00:00Z","timestamp":1762819200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,11,11]],"date-time":"2025-11-11T00:00:00Z","timestamp":1762819200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["J Real-Time Image Proc"],"published-print":{"date-parts":[[2026,2]]},"DOI":"10.1007\/s11554-025-01797-6","type":"journal-article","created":{"date-parts":[[2025,11,11]],"date-time":"2025-11-11T16:47:35Z","timestamp":1762879655000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["CMCF-DETR: a real-time lightweight DETR model for foreign object detection on coal mine conveyor belts"],"prefix":"10.1007","volume":"23","author":[{"given":"Zhi","family":"Wang","sequence":"first","affiliation":[]},{"given":"Hanghang","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Haojie","family":"Ye","sequence":"additional","affiliation":[]},{"given":"Zhuang","family":"Chu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,11,11]]},"reference":[{"key":"1797_CR1","doi-asserted-by":"publisher","first-page":"112735","DOI":"10.1016\/j.measurement.2023.112735","volume":"213","author":"M Zhang","year":"2023","unstructured":"Zhang, M., Jiang, K., Cao, Y., Li, M., Wang, Q., Li, D., Zhang, Y.: A new paradigm for intelligent status detection of belt conveyors based on deep learning. Measurement 213, 112735 (2023)","journal-title":"Measurement"},{"key":"1797_CR2","first-page":"21","volume-title":"European conference on computer vision","author":"W Liu","year":"2016","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.Y., Berg, A.C.: Ssd: Single shot multibox detector. In: European conference on computer vision, pp. 21\u201337. Springer, Cham (2016)"},{"issue":"6","key":"1797_CR3","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2016","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intel 39(6), 1137\u20131149 (2016)","journal-title":"IEEE Trans Pattern Anal Mach Intel"},{"key":"1797_CR4","first-page":"1","volume-title":"European conference on computer vision","author":"CY Wang","year":"2024","unstructured":"Wang, C.Y., Yeh, I.H., Mark Liao, H.Y.: YOLOv9: learning what you want to\u00a0learn using programmable gradient information. In: European conference on computer vision, pp. 1\u201321. Springer, Cham (2024)"},{"key":"1797_CR5","unstructured":"Khanam, R., Hussain, M.: Yolov11: an overview of the key architectural enhancements. arXiv:2410.17725 (2024)"},{"key":"1797_CR6","first-page":"107984","volume":"37","author":"A Wang","year":"2024","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J.: Yolov10: real-time end-to-end object detection. Adv Neural Inf Proc Syst 37, 107984\u2013108011 (2024)","journal-title":"Adv Neural Inf Proc Syst"},{"key":"1797_CR7","first-page":"213","volume-title":"European conference on computer vision","author":"N Carion","year":"2020","unstructured":"Carion, N., Massa, F., Synnaeve, G., Usunier, N., Kirillov, A., Zagoruyko, S.: End-to-End Object Detection with Transformers. In: European conference on computer vision, pp. 213\u2013229. Springer, Cham (2020)"},{"key":"1797_CR8","first-page":"16965","volume-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition","author":"Y Zhao","year":"2024","unstructured":"Zhao, Y., Lv, W., Xu, S., Wei, J., Wang, G., Dang, Q., Liu, Y., Chen, J.: DETRs beat YOLOs on real-time object detection. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 16965\u201316974. Springer, Cham (2024)"},{"issue":"6","key":"1797_CR9","first-page":"1","volume":"19","author":"L Yi","year":"2025","unstructured":"Yi, L., Wu, J., Wu, R., Li, H.: CCO-DETR: a lightweight multi-scale object detection model for coal coking operations. Signal Image Video Proc 19(6), 1\u20132 (2025)","journal-title":"Signal Image Video Proc"},{"key":"1797_CR10","doi-asserted-by":"publisher","first-page":"6984","DOI":"10.1109\/TASE.2024.3457829","volume":"22","author":"X Sun","year":"2024","unstructured":"Sun, X., Song, K., Wen, X., Wang, Y., Yan, Y.: SDD-DETR: surface defect detection for no-service aero-engine blades with detection transformer. IEEE Trans Automat Sci Eng 22, 6984\u20136997 (2024)","journal-title":"IEEE Trans Automat Sci Eng"},{"issue":"12","key":"1797_CR11","doi-asserted-by":"publisher","first-page":"6793","DOI":"10.1007\/s00521-024-09430-6","volume":"36","author":"M Wang","year":"2024","unstructured":"Wang, M., Jiao, Z., Huang, Z., Yu, S.: FCDS-DETR: detection transformer based on feature correction and double sampling. Neural Comput. Appl. 36(12), 6793\u20136808 (2024)","journal-title":"Neural Comput. Appl."},{"key":"1797_CR12","first-page":"243","volume-title":"Computer graphics forum","author":"Q Zhang","year":"2019","unstructured":"Zhang, Q., Nie, Y., Zheng, W.S.: Dual illumination estimation for robust exposure correction. In: Computer graphics forum, vol. 38, pp. 243\u2013252. Wiley, New Jersey (2019)"},{"issue":"2","key":"1797_CR13","doi-asserted-by":"publisher","first-page":"982","DOI":"10.1109\/TIP.2016.2639450","volume":"26","author":"X Guo","year":"2017","unstructured":"Guo, X., Li, Yu., Ling, H.: Lime: low-light image enhancement via illumination map estimation. IEEE Trans. Image Process. 26(2), 982\u2013993 (2017)","journal-title":"IEEE Trans. Image Process."},{"key":"1797_CR14","first-page":"5694","volume-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition","author":"X Ma","year":"2024","unstructured":"Ma, X., Dai, X., Bai, Y., Wang, Y., Fu, Y.: Rewrite the stars. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 5694\u20135703. IEEE (2024)"},{"key":"1797_CR15","first-page":"770","volume-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","author":"K He","year":"2016","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 770\u2013778. IEEE (2016)"},{"key":"1797_CR16","first-page":"1251","volume-title":"Proceedings of the IEEE conference on computer vision and pattern recognition","author":"F Chollet","year":"2017","unstructured":"Chollet, F.: Xception: deep learning with depthwise separable convolutions. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 1251\u20131258. IEEE (2017)"},{"key":"1797_CR17","first-page":"805","volume-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition","author":"X Pan","year":"2022","unstructured":"Pan, X., Ge, C., Lu, R., Song, S., Chen, G., Huang, Z., Huang, G.: On the integration of self-attention and convolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 805\u2013815. IEEE (2022)"},{"key":"1797_CR18","doi-asserted-by":"publisher","first-page":"179","DOI":"10.1109\/CW58918.2023.00034","volume-title":"2023 International conference on Cyberworlds (CW)","author":"W Messaoud","year":"2023","unstructured":"Messaoud, W., Trabelsi, R., Cabani, A., Abdelkefi, F.: Multi-head self attention for enhanced object detection in the maritime domain. In: 2023 International conference on Cyberworlds (CW), pp. 179\u2013184. IEEE (2023)"},{"key":"1797_CR19","unstructured":"Lin, Z., Feng, M., Santos, CN., Yu, M., Xiang, B., Zhou, B., Bengio, Y.: A structured self-attentive sentence embedding. arXiv:1703.03130. (2017)"},{"key":"1797_CR20","first-page":"116","volume-title":"Proceedings of the European conference on computer vision (ECCV)","author":"N Ma","year":"2018","unstructured":"Ma, N., Zhang, X., Zheng, H.T., Sun, J.: ShuffleNet V2: practical guidelines for efficient CNN architecture design. In: Proceedings of the European conference on computer vision (ECCV), pp. 116\u2013131. Springer, Cham (2018)"},{"issue":"3","key":"1797_CR21","first-page":"1361","volume":"47","author":"DQ Cheng","year":"2022","unstructured":"Cheng, D.Q., Xu, J.Y., Kou, Q.Q., Zhang, H.X., Han, C.G., Yu, B., Qian, J.S.: Lightweight network based on residual information for foreign body classification on coal conveyor belt. J China Coal Soci 47(3), 1361\u20131369 (2022)","journal-title":"J China Coal Soci"},{"key":"1797_CR22","doi-asserted-by":"publisher","first-page":"109256","DOI":"10.1016\/j.patcog.2022.109256","volume":"137","author":"C Dong","year":"2023","unstructured":"Dong, C., Duoqian, M.: Control distance IoU and control distance IoU loss for better bounding box regression. Patt Recog 137, 109256 (2023)","journal-title":"Patt Recog"},{"key":"1797_CR23","unstructured":"Ma, S., Xu, Y.: Mpdiou: a loss for efficient and accurate bounding box regression. arXiv:2307.07662. (2023)"},{"key":"1797_CR24","doi-asserted-by":"publisher","first-page":"146","DOI":"10.1016\/j.neucom.2022.07.042","volume":"506","author":"YF Zhang","year":"2022","unstructured":"Zhang, Y.F., Ren, W., Zhang, Z., Jia, Z., Wang, L., Tan, T.: Focal and efficient IOU loss for accurate bounding box regression. Neurocomputing 506, 146\u2013157 (2022)","journal-title":"Neurocomputing"},{"key":"1797_CR25","unstructured":"Gevorgyan, Z.: SIoU loss: more powerful learning for bounding box regression. arXiv:2205.12740. (2022)"},{"key":"1797_CR26","first-page":"5901","volume-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition","author":"W Wu","year":"2022","unstructured":"Wu, W., Weng, J., Zhang, P., Wang, X., Yang, W., Jiang, J.: URetinex-Net: Retinex-based deep unfolding network for low-light image enhancement. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 5901\u20135910. IEEE (2022)"},{"key":"1797_CR27","first-page":"1780","volume-title":"Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition","author":"C Guo","year":"2020","unstructured":"Guo, C., Li, C., Guo, J., Loy, C.C., Hou, J., Kwong, S., Cong, R.: Zero-reference deep curve estimation for low-light image enhancement. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 1780\u20131789. IEEE (2020)"}],"container-title":["Journal of Real-Time Image Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01797-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11554-025-01797-6","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11554-025-01797-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T16:52:27Z","timestamp":1770396747000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11554-025-01797-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,11,11]]},"references-count":27,"journal-issue":{"issue":"1","published-print":{"date-parts":[[2026,2]]}},"alternative-id":["1797"],"URL":"https:\/\/doi.org\/10.1007\/s11554-025-01797-6","relation":{},"ISSN":["1861-8200","1861-8219"],"issn-type":[{"value":"1861-8200","type":"print"},{"value":"1861-8219","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,11,11]]},"assertion":[{"value":"23 September 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"27 October 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 November 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no competing interests.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This study did not involve human participants or animal experiments, nor did it include any patient-related data; therefore, no informed consent statement is required.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval and informed consent"}}],"article-number":"5"}}