{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,11]],"date-time":"2026-03-11T18:45:09Z","timestamp":1773254709161,"version":"3.50.1"},"reference-count":41,"publisher":"Springer Science and Business Media LLC","issue":"9","license":[{"start":{"date-parts":[[2022,11,29]],"date-time":"2022-11-29T00:00:00Z","timestamp":1669680000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,11,29]],"date-time":"2022-11-29T00:00:00Z","timestamp":1669680000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Key Research and Development Program of China","award":["2018YFB1003401"],"award-info":[{"award-number":["2018YFB1003401"]}]},{"DOI":"10.13039\/501100010909","name":"Young Scientists Fund","doi-asserted-by":"publisher","award":["61702178"],"award-info":[{"award-number":["61702178"]}],"id":[{"id":"10.13039\/501100010909","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100004735","name":"Natural Science Foundation of Hunan Province","doi-asserted-by":"publisher","award":["2019JJ50123"],"award-info":[{"award-number":["2019JJ50123"]}],"id":[{"id":"10.13039\/501100004735","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Research Foundation of Education Bureau of Hunan Province","award":["20C0625"],"award-info":[{"award-number":["20C0625"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2023,3]]},"DOI":"10.1007\/s00521-022-08035-1","type":"journal-article","created":{"date-parts":[[2022,11,29]],"date-time":"2022-11-29T11:02:37Z","timestamp":1669719757000},"page":"6767-6785","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":10,"title":["MSSIF-Net: an efficient CNN automatic detection method for freight train images"],"prefix":"10.1007","volume":"35","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-4413-9974","authenticated-orcid":false,"given":"Longxin","family":"Zhang","sequence":"first","affiliation":[]},{"given":"Yang","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Jingsheng","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Chuang","family":"Li","sequence":"additional","affiliation":[]},{"given":"Keqin","family":"Li","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,11,29]]},"reference":[{"key":"8035_CR1","doi-asserted-by":"publisher","first-page":"127","DOI":"10.1016\/j.neucom.2016.10.018","volume":"222","author":"J Sun","year":"2017","unstructured":"Sun J, Xiao Z, Xie Y (2017) Automatic multi-fault recognition in tfds based on convolutional neural network. Neurocomputing 222:127\u2013136","journal-title":"Neurocomputing"},{"key":"8035_CR2","doi-asserted-by":"publisher","first-page":"212","DOI":"10.1016\/j.neucom.2019.11.002","volume":"380","author":"X Fu","year":"2020","unstructured":"Fu X, Li K, Liu J, Li K, Zeng Z, Chen C (2020) A two-stage attention aware method for train bearing shed oil inspection based on convolutional neural networks. Neurocomputing 380:212\u2013224","journal-title":"Neurocomputing"},{"issue":"3","key":"8035_CR3","doi-asserted-by":"publisher","first-page":"523","DOI":"10.1109\/TNNLS.2016.2551940","volume":"28","author":"T De Bruin","year":"2016","unstructured":"De Bruin T, Verbert K, Babu\u0161ka R (2016) Railway track circuit fault diagnosis using recurrent neural networks. IEEE Trans Neural Netw Learn Syst 28(3):523\u2013533","journal-title":"IEEE Trans Neural Netw Learn Syst"},{"issue":"11","key":"8035_CR4","doi-asserted-by":"publisher","first-page":"7423","DOI":"10.1109\/TII.2021.3056554","volume":"17","author":"Y Zhang","year":"2021","unstructured":"Zhang Y, Liu M, Yang Y, Guo Y, Zhang H (2021) A unified light framework for real-time fault detection of freight train images. IEEE Trans Industr Inf 17(11):7423\u20137432","journal-title":"IEEE Trans Industr Inf"},{"issue":"8","key":"8035_CR5","doi-asserted-by":"publisher","first-page":"3583","DOI":"10.1007\/s00521-020-05202-0","volume":"33","author":"J Leng","year":"2021","unstructured":"Leng J, Liu Y (2021) Single-shot augmentation detector for object detection. Neural Comput Appl 33(8):3583\u20133596","journal-title":"Neural Comput Appl"},{"key":"8035_CR6","doi-asserted-by":"crossref","unstructured":"Gao F, Ji S, Guo J, Li Q, Ji Y, Liu Y, Feng S, Wei H, Wang N, Yang B (2021) Id-net: an improved mask r-cnn model for intrusion detection under power grid surveillance. Neural Comput Appl 1\u201317 (2021)","DOI":"10.1007\/s00521-021-05688-2"},{"key":"8035_CR7","doi-asserted-by":"crossref","unstructured":"Girshick R, Donahue J, Darrell T, Malik J (2014) Rich feature hierarchies for accurate object detection and semantic segmentation. In: Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition, pp. 580\u2013587","DOI":"10.1109\/CVPR.2014.81"},{"key":"8035_CR8","doi-asserted-by":"crossref","unstructured":"Girshick R (2015) Fast r-cnn. In: Proceedings of The IEEE International Conference on Computer Vision, pp. 1440\u20131448","DOI":"10.1109\/ICCV.2015.169"},{"issue":"6","key":"8035_CR9","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2016","unstructured":"Ren S, He K, Girshick R, Sun J (2016) Faster r-cnn: towards real-time object detection with region proposal networks. IEEE Trans Pattern Anal Mach Intell 39(6):1137\u20131149","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8035_CR10","doi-asserted-by":"crossref","unstructured":"He K, Gkioxari G, Doll\u00e1r P, Girshick R (2017) Mask r-cnn. In: Proceedings of The IEEE International Conference on Computer Vision, pp. 2961\u20132969","DOI":"10.1109\/ICCV.2017.322"},{"issue":"2","key":"8035_CR11","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1007\/s11263-013-0620-5","volume":"104","author":"JR Uijlings","year":"2013","unstructured":"Uijlings JR, Van De Sande KE, Gevers T, Smeulders AW (2013) Selective search for object recognition. Int J Comput Vision 104(2):154\u2013171","journal-title":"Int J Comput Vision"},{"key":"8035_CR12","doi-asserted-by":"crossref","unstructured":"Redmon J, Divvala S, Girshick R, Farhadi A (2016) You only look once: Unified, real-time object detection. In: Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition, pp. 779\u2013788","DOI":"10.1109\/CVPR.2016.91"},{"key":"8035_CR13","doi-asserted-by":"crossref","unstructured":"Redmon J, Farhadi A (2017) Yolo9000: better, faster, stronger. In: Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition, pp. 7263\u20137271","DOI":"10.1109\/CVPR.2017.690"},{"key":"8035_CR14","unstructured":"Farhadi A, Redmon J (2018) Yolov3: An incremental improvement. In: Computer Vision and Pattern Recognition, vol. 1804 (2018). Springer Berlin\/Heidelberg, Germany"},{"key":"8035_CR15","unstructured":"Bochkovskiy A, Wang C-Y, Liao H-YM (2020) Yolov4: Optimal speed and accuracy of object detection. arXiv preprint arXiv:2004.10934"},{"key":"8035_CR16","doi-asserted-by":"crossref","unstructured":"Liu W, Anguelov D, Erhan D, Szegedy C, Reed S, Fu C-Y, Berg AC (2016) Ssd: Single shot multibox detector. In: European Conference on Computer Vision, pp. 21\u201337 (2016). Springer","DOI":"10.1007\/978-3-319-46448-0_2"},{"issue":"2","key":"8035_CR17","doi-asserted-by":"publisher","first-page":"318","DOI":"10.1109\/TPAMI.2018.2858826","volume":"42","author":"T-Y Lin","year":"2020","unstructured":"Lin T-Y, Goyal P, Girshick R, He K, Doll\u00e1r P (2020) Focal loss for dense object detection. IEEE Trans Pattern Anal Mach Intell 42(2):318\u2013327","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8035_CR18","doi-asserted-by":"crossref","unstructured":"Lin T-Y, Doll\u00e1r P, Girshick R, He K, Hariharan B, Belongie S (2017) Feature pyramid networks for object detection. In: Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117\u20132125","DOI":"10.1109\/CVPR.2017.106"},{"key":"8035_CR19","doi-asserted-by":"crossref","unstructured":"Liu S, Qi L, Qin H, Shi J, Jia J (2018) Path aggregation network for instance segmentation. In: Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition, pp. 8759\u20138768","DOI":"10.1109\/CVPR.2018.00913"},{"issue":"9","key":"8035_CR20","doi-asserted-by":"publisher","first-page":"1904","DOI":"10.1109\/TPAMI.2015.2389824","volume":"37","author":"K He","year":"2015","unstructured":"He K, Zhang X, Ren S, Sun J (2015) Spatial pyramid pooling in deep convolutional networks for visual recognition. IEEE Trans Pattern Anal Mach Intell 37(9):1904\u20131916","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"8035_CR21","doi-asserted-by":"crossref","unstructured":"Law H, Deng J (2018) Cornernet: Detecting objects as paired keypoints. In: Proceedings of The European Conference on Computer Vision (ECCV), pp. 734\u2013750","DOI":"10.1007\/978-3-030-01264-9_45"},{"key":"8035_CR22","unstructured":"Zhou X, Wang D, Kr\u00e4henb\u00fchl P (2019) Objects as points. arXiv preprint arXiv:1904.07850"},{"key":"8035_CR23","unstructured":"Ge Z, Liu S, Wang F, Li Z, Sun J (2021) Yolox: Exceeding yolo series in 2021. arXiv preprint arXiv:2107.08430"},{"key":"8035_CR24","doi-asserted-by":"crossref","unstructured":"Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: European Conference on Computer Vision, pp. 213\u2013229. Springer","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"8035_CR25","unstructured":"Zhu X, Su W, Lu L, Li B, Wang X, Dai J (2020) Deformable detr: Deformable transformers for end-to-end object detection. arXiv preprint arXiv:2010.04159"},{"key":"8035_CR26","unstructured":"Fang Y, Liao B, Wang X, Fang J, Qi J, Wu R, Niu J, Liu W (2021) You only look at one sequence: Rethinking transformer in vision through object detection. Adv Neural Inf Proc Syst 34"},{"key":"8035_CR27","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S, et al (2020) An image is worth 16x16 words: Transformers for image recognition at scale. arXiv preprint arXiv:2010.11929"},{"key":"8035_CR28","first-page":"2017","volume":"28","author":"M Jaderberg","year":"2015","unstructured":"Jaderberg M, Simonyan K, Zisserman A (2015) Spatial transformer networks. Adv Neural Inf Process Syst 28:2017\u20132025","journal-title":"Adv Neural Inf Process Syst"},{"key":"8035_CR29","doi-asserted-by":"crossref","unstructured":"Zhu X, Cheng D, Zhang Z, Lin S, Dai J (2019) An empirical study of spatial attention mechanisms in deep networks. In: Proceedings of The IEEE\/CVF International Conference on Computer Vision, pp. 6688\u20136697","DOI":"10.1109\/ICCV.2019.00679"},{"key":"8035_CR30","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition, pp. 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"key":"8035_CR31","doi-asserted-by":"crossref","unstructured":"Li X, Wang W, Hu X, Yang J (2019) Selective kernel networks. In: Proceedings of The IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 510\u2013519","DOI":"10.1109\/CVPR.2019.00060"},{"key":"8035_CR32","doi-asserted-by":"crossref","unstructured":"Wang Q, Wu B, Zhu P, Li P, Zuo W, Hu Q (2020) Eca-net: efficient channel attention for deep convolutional neural networks, 2020 ieee. In: CVF Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, pp. 11531\u201311539","DOI":"10.1109\/CVPR42600.2020.01155"},{"key":"8035_CR33","doi-asserted-by":"crossref","unstructured":"Qin Z, Zhang P, Wu F, Li X (2021) Fcanet: Frequency channel attention networks. In: Proceedings of The IEEE\/CVF International Conference on Computer Vision, pp. 783\u2013792","DOI":"10.1109\/ICCV48922.2021.00082"},{"key":"8035_CR34","doi-asserted-by":"crossref","unstructured":"Woo S, Park J, Lee J-Y, Kweon IS (2018) Cbam: Convolutional block attention module. In: Proceedings of The European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)","DOI":"10.1007\/978-3-030-01234-2_1"},{"key":"8035_CR35","unstructured":"Zhang H, Zu K, Lu J, Zou Y, Meng D (2021) Epsanet: An efficient pyramid split attention block on convolutional neural network. arXiv preprint arXiv:2105.14447"},{"key":"8035_CR36","unstructured":"Liu S, Huang D, Wang Y (2019) Learning spatial fusion for single-shot object detection. arXiv preprint arXiv:1911.09516"},{"key":"8035_CR37","doi-asserted-by":"crossref","unstructured":"Neubeck A, Van\u00a0Gool L (2006) Efficient non-maximum suppression. In: 18th International Conference on Pattern Recognition (ICPR\u201906), vol. 3, pp. 850\u2013855","DOI":"10.1109\/ICPR.2006.479"},{"key":"8035_CR38","doi-asserted-by":"publisher","first-page":"858","DOI":"10.1016\/j.apsusc.2013.09.002","volume":"285","author":"K Song","year":"2013","unstructured":"Song K, Yan Y (2013) A noise robust method based on completed local binary patterns for hot-rolled steel strip surface defects. Appl Surf Sci 285:858\u2013864","journal-title":"Appl Surf Sci"},{"issue":"4","key":"8035_CR39","doi-asserted-by":"publisher","first-page":"1493","DOI":"10.1109\/TIM.2019.2915404","volume":"69","author":"Y He","year":"2019","unstructured":"He Y, Song K, Meng Q, Yan Y (2019) An end-to-end steel surface defect detection approach via fusing multiple hierarchical features. IEEE Trans Instrum Meas 69(4):1493\u20131504","journal-title":"IEEE Trans Instrum Meas"},{"key":"8035_CR40","first-page":"1","volume":"70","author":"Y Bao","year":"2021","unstructured":"Bao Y, Song K, Liu J, Wang Y, Yan Y, Yu H, Li X (2021) Triplet-graph reasoning network for few-shot metal generic surface defect segmentation. IEEE Trans Instrum Meas 70:1\u201311","journal-title":"IEEE Trans Instrum Meas"},{"key":"8035_CR41","doi-asserted-by":"crossref","unstructured":"Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A (2016) Learning deep features for discriminative localization. In: Proceedings of The IEEE Conference on Computer Vision and Pattern Recognition, pp. 2921\u20132929","DOI":"10.1109\/CVPR.2016.319"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-08035-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-022-08035-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-022-08035-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,28]],"date-time":"2023-02-28T20:33:49Z","timestamp":1677616429000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-022-08035-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,11,29]]},"references-count":41,"journal-issue":{"issue":"9","published-print":{"date-parts":[[2023,3]]}},"alternative-id":["8035"],"URL":"https:\/\/doi.org\/10.1007\/s00521-022-08035-1","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"value":"0941-0643","type":"print"},{"value":"1433-3058","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022,11,29]]},"assertion":[{"value":"9 February 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 November 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"29 November 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Compliance with ethical standards"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}]}}