{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:19:47Z","timestamp":1757618387843,"version":"3.44.0"},"reference-count":42,"publisher":"Springer Science and Business Media LLC","issue":"21","license":[{"start":{"date-parts":[[2025,6,9]],"date-time":"2025-06-09T00:00:00Z","timestamp":1749427200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,6,9]],"date-time":"2025-06-09T00:00:00Z","timestamp":1749427200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100012542","name":"Sichuan Province Science and Technology Support Program","doi-asserted-by":"publisher","award":["2023YFG0181"],"award-info":[{"award-number":["2023YFG0181"]}],"id":[{"id":"10.13039\/100012542","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Comput &amp; Applic"],"published-print":{"date-parts":[[2025,7]]},"DOI":"10.1007\/s00521-025-11377-1","type":"journal-article","created":{"date-parts":[[2025,6,8]],"date-time":"2025-06-08T23:45:08Z","timestamp":1749426308000},"page":"17083-17109","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Lightweight network research for robotic visual grasp for deep space exploration"],"prefix":"10.1007","volume":"37","author":[{"given":"Zhichao","family":"Xu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-3398-8708","authenticated-orcid":false,"given":"Junpeng","family":"Xue","sequence":"additional","affiliation":[]},{"given":"Zeyu","family":"Song","sequence":"additional","affiliation":[]},{"given":"Ran","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Wenbo","family":"Lu","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,9]]},"reference":[{"key":"11377_CR1","doi-asserted-by":"publisher","unstructured":"M. Bigdeli, R. Srivastava, M. Scaraggi, \u201cDynamics of space debris removal: A review,\" Instrumentation and Methods for Astrophysics. Apr. 2023. https:\/\/doi.org\/10.48550\/arXiv.2304.05709.","DOI":"10.48550\/arXiv.2304.05709"},{"key":"11377_CR2","doi-asserted-by":"publisher","first-page":"17963","DOI":"10.1007\/s00521-022-07446-4","volume":"34","author":"C Yin","year":"2022","unstructured":"Yin C, Zhang Q (2022) Object affordance detection with boundary-preserving network for robotic manipulation tasks. Neural Comput & Applic 34:17963\u201317980. https:\/\/doi.org\/10.1007\/s00521-022-07446-4","journal-title":"Neural Comput & Applic"},{"key":"11377_CR3","doi-asserted-by":"publisher","first-page":"12283","DOI":"10.1007\/s00521-020-05644-6","volume":"33","author":"SH Zabihifar","year":"2021","unstructured":"Zabihifar SH, Semochkin AN, Seliverstova EV et al (2021) Unreal mask: one-shot multi-object class-based pose estimation for robotic manipulation using keypoints with a synthetic dataset. Neural Comput Applic 33:12283\u201312300. https:\/\/doi.org\/10.1007\/s00521-020-05644-6","journal-title":"Neural Comput Applic"},{"key":"11377_CR4","doi-asserted-by":"publisher","first-page":"21309","DOI":"10.1007\/s00521-023-08886-2","volume":"35","author":"H Cao","year":"2023","unstructured":"Cao H, Zhang Y, Shan D et al (2023) TRF-Net: a transformer-based RGB-D fusion network for desktop object instance segmentation. Neural Comput Applic 35:21309\u201321330. https:\/\/doi.org\/10.1007\/s00521-023-08886-2","journal-title":"Neural Comput Applic"},{"key":"11377_CR5","doi-asserted-by":"publisher","first-page":"18","DOI":"10.1016\/j.paerosci.2015.11.001","volume":"80","author":"M Shan","year":"2016","unstructured":"Shan M, Guo J, Gill E (2016) Review and comparison of active space debris grasping and removal methods. Prog Aerosp Sci 80:18\u201332","journal-title":"Prog Aerosp Sci"},{"key":"11377_CR6","doi-asserted-by":"publisher","unstructured":"Q. Gu, J. Su and X. Bi, \u201cAttention Grasping Network: A Real-time Approach to Generating Grasp Synthesis,\" IEEE international conference on robotics and biomimetics (ROBIO), Dali, China, Jan. 2019, pp.3036\u20133041. https:\/\/doi.org\/10.1109\/ROBIO49542.2019.8961828.","DOI":"10.1109\/ROBIO49542.2019.8961828"},{"key":"11377_CR7","doi-asserted-by":"publisher","unstructured":"Y. Jiang, S. Moseson, A. Saxena, \u201cEfficient grasping from RGBD images: Learning using a new rectangle representation,\" IEEE international conference on robotics and automation, Shanghai, China, May.2011, pp: 3304\u20133311. https:\/\/doi.org\/10.1109\/ICRA.2011.5980145.","DOI":"10.1109\/ICRA.2011.5980145"},{"key":"11377_CR8","doi-asserted-by":"publisher","DOI":"10.1177\/0278364914549607","author":"I Lenz","year":"2015","unstructured":"Lenz I, Lee H, Saxena A (2015) Deep learning for detecting robotic grasps. Int J Robot Res. https:\/\/doi.org\/10.1177\/0278364914549607","journal-title":"Int J Robot Res"},{"key":"11377_CR9","doi-asserted-by":"publisher","first-page":"3355","DOI":"10.48550\/arXiv.1301.3592","volume":"3","author":"F-J Chu","year":"2018","unstructured":"Chu F-J, Xu R, Vela PA (2018) Real-world multi-object, multi-grasp detection. IEEE Robot Autom Lett 3:3355\u20133362. https:\/\/doi.org\/10.48550\/arXiv.1301.3592","journal-title":"IEEE Robot Autom Lett"},{"key":"11377_CR10","doi-asserted-by":"publisher","unstructured":"D. Morrison, P. Corke, J. Leitner, \u201cClosing the Loop for Robotic Grasping: A Real-time, Generative Grasp Synthesis Approach,\" Robotics: Science and Systems, May. 2018. https:\/\/doi.org\/10.48550\/arXiv.1804.05172.","DOI":"10.48550\/arXiv.1804.05172"},{"key":"11377_CR11","doi-asserted-by":"publisher","unstructured":"O. Ronneberger,P. Fischer,T. Brox, 2015 \u201cU-net: Convolutional networks for biomedical image segmentation,\" International conference on medical image computing and computer-assisted intervention, Freiburg, Baden-W\u00fcrttemberg, Germany, pp. 234\u2013241. https:\/\/doi.org\/10.48550\/arXiv.1505.04597.","DOI":"10.48550\/arXiv.1505.04597"},{"key":"11377_CR12","doi-asserted-by":"publisher","unstructured":"S. Kumra, S. Joshi, F. Sahin, \u201cAntipodal Robotic Grasping using Generative Residual Convolutional Neural Network,\" IEEE\/RSJ international conference on intelligent robots and systems (IROS), Las Vegas, NV, USA, Oct. 2020, pp.9626\u20139633. https:\/\/doi.org\/10.1109\/IROS45743.2020.9340777.","DOI":"10.1109\/IROS45743.2020.9340777"},{"key":"11377_CR13","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2021.3129875","volume":"71","author":"D Liu","year":"2022","unstructured":"Liu D, Tao X, Yuan L, Du Y, Cong M (2022) Robotic objects detection and grasping in clutter based on cascaded deep convolutional neural network. IEEE Trans Instrum Meas 71:1\u201310. https:\/\/doi.org\/10.1109\/TIM.2021.3129875","journal-title":"IEEE Trans Instrum Meas"},{"key":"11377_CR14","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3165825","volume":"71","author":"H Cheng","year":"2022","unstructured":"Cheng H, Wang Y, Meng MQ-H (2022) A robot grasping system with single-stage anchor-free deep grasp detector. IEEE Trans Instrum Meas 71:1\u201312. https:\/\/doi.org\/10.1109\/TIM.2022.3165825","journal-title":"IEEE Trans Instrum Meas"},{"key":"11377_CR15","first-page":"1","volume":"71","author":"L Yuanhao","year":"2022","unstructured":"Yuanhao L, Liu Y, Ma Z, Huang P (2022) A novel generative convolutional neural network for robot grasp detection on gaussian guidance. IEEE Trans Instrum Measurement 71:1\u201310","journal-title":"IEEE Trans Instrum Measurement"},{"key":"11377_CR16","doi-asserted-by":"publisher","unstructured":"A Gholami et al., 2021 \u201cA Survey of Quantization Methods for Efficient Neural Network Inference.\" Computer Vision and Pattern Recognition, https:\/\/doi.org\/10.48550\/arXiv.2103.13630.","DOI":"10.48550\/arXiv.2103.13630"},{"key":"11377_CR17","doi-asserted-by":"publisher","unstructured":"Iandola F et al., \u201cSqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5MB model size,\" Computer Vision and Pattern Recognition, Nov. 2016, arXiv:1602.07360. https:\/\/doi.org\/10.48550\/arXiv.1602.07360.","DOI":"10.48550\/arXiv.1602.07360"},{"key":"11377_CR18","doi-asserted-by":"publisher","unstructured":"A. Gholami et al., \u201cSqueezeNext: Hardware-Aware Neural Network Design,\" IEEE\/CVF conference on computer vision and pattern recognition workshops (CVPRW), Salt Lake City, UT, USA, Dec. 2018, pp. 1719\u2013171909. https:\/\/doi.org\/10.48550\/arXiv.1803.10615.","DOI":"10.48550\/arXiv.1803.10615"},{"key":"11377_CR19","doi-asserted-by":"publisher","unstructured":"A. Howard et al, 2017 \u201cMobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications,\" Computer Vision and Pattern Recognition, arXiv:1704.04861. https:\/\/doi.org\/10.48550\/arXiv.1704.04861.","DOI":"10.48550\/arXiv.1704.04861"},{"key":"11377_CR20","doi-asserted-by":"publisher","unstructured":"M. Sandler, A. Howard, M. Zhu, A. Zhmoginov and L. -C. Chen, \u201cMobileNetV2: Inverted Residuals and Linear Bottlenecks,\" IEEE\/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, Jun. 2018, pp. 4510\u20134520. https:\/\/doi.org\/10.48550\/arXiv.1801.04381.","DOI":"10.48550\/arXiv.1801.04381"},{"key":"11377_CR21","doi-asserted-by":"publisher","unstructured":"A. Howard et al., 2019 \u201cSearching for MobileNetV3,\" IEEE\/CVF international conference on computer vision (ICCV), Seoul, Korea (South), pp. 1314\u20131324. https:\/\/doi.org\/10.48550\/arXiv.1905.02244.","DOI":"10.48550\/arXiv.1905.02244"},{"key":"11377_CR22","doi-asserted-by":"publisher","unstructured":"J. Hu, L. Shen and G. Sun, \u201cSqueeze-and-Excitation Networks,\" 2018 IEEE\/CVF conference on computer vision and pattern recognition, Salt Lake City, UT, USA, May 2018, pp. 7132\u20137141. https:\/\/doi.org\/10.48550\/arXiv.1709.01507.","DOI":"10.48550\/arXiv.1709.01507"},{"key":"11377_CR23","doi-asserted-by":"publisher","unstructured":"X. Zhang, X. Zhou, M. Lin and J. Sun, 2018 \u201cShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices,\" IEEE\/CVF conference on computer vision and pattern recognition, Salt Lake City, UT, USA, pp. 6848\u20136856. https:\/\/doi.org\/10.48550\/arXiv.1707.01083.","DOI":"10.48550\/arXiv.1707.01083"},{"key":"11377_CR24","doi-asserted-by":"crossref","unstructured":"Ulyanov D, Vedaldi A, Lempitsky V. Improved texture networks: maximizing quality and diversity in feed-forward stylization and texture synthesis. Computer vision and pattern recognition, 2017.","DOI":"10.1109\/CVPR.2017.437"},{"key":"11377_CR25","doi-asserted-by":"publisher","unstructured":"N. Ma, X. Zhang, H.-T. Zheng, J Sun, 2018 \u201cShuffleNet V2: practical guidelines for efficient cnn architecture design,\" Computer Vision and Pattern Recognition https:\/\/doi.org\/10.48550\/arXiv.1807.11164.","DOI":"10.48550\/arXiv.1807.11164"},{"key":"11377_CR26","doi-asserted-by":"publisher","unstructured":"T. Lin T, P. Doll\u00e1r, R. Girshick et al., 2017 \u201cFeature Pyramid Networks for Object Detection,\" IEEE conference on computer vision and pattern recognition (CVPR), July 21\u201326, Honolulu, HI, USA, 2017: 936\u2013944. https:\/\/doi.org\/10.48550\/arXiv.1612.03144.","DOI":"10.48550\/arXiv.1612.03144"},{"key":"11377_CR27","doi-asserted-by":"publisher","unstructured":"D. Kingma, J. Ba, 2017 \u201cAdam: A Method for Stochastic Optimization,\" Machine Learning, https:\/\/doi.org\/10.48550\/arXiv.1412.6980.","DOI":"10.48550\/arXiv.1412.6980"},{"key":"11377_CR28","doi-asserted-by":"publisher","unstructured":"G. Meyer, 2021 \u201cAn Alternative Probabilistic Interpretation of the Huber Loss,\" IEEE\/CVF conference on computer vision and pattern recognition (CVPR), Nashville, TN, USA, pp. 5257\u20135265. https:\/\/doi.org\/10.48550\/arXiv.1911.02088.","DOI":"10.48550\/arXiv.1911.02088"},{"key":"11377_CR29","doi-asserted-by":"publisher","first-page":"101052","DOI":"10.1016\/j.aei.2020.101052","volume":"44","author":"L Bergamini","year":"2020","unstructured":"Bergamini L, Sposato M, Pellicciari M, Peruzzini M, Calderara S, Schmidt J (2020) Deep learning-based method for vision-guided robotic grasping of unknown objects. Adv Eng Inform 44:101052. https:\/\/doi.org\/10.1016\/j.aei.2020.101052","journal-title":"Adv Eng Inform"},{"issue":"10","key":"11377_CR30","doi-asserted-by":"publisher","first-page":"9610","DOI":"10.1109\/JSEN.2022.3163730","volume":"22","author":"H Cheng","year":"2022","unstructured":"Cheng H, Wang Y, Meng MQH (2022) A vision-based robot grasping system. IEEE Sensors J 22(10):9610\u20139620. https:\/\/doi.org\/10.1109\/JSEN.2022.3163730","journal-title":"IEEE Sensors J"},{"issue":"3","key":"11377_CR31","doi-asserted-by":"publisher","first-page":"8170","DOI":"10.1109\/LRA.2022.3187261","volume":"7","author":"S Wang","year":"2022","unstructured":"Wang S, Zhou Z, Kan Z (2022) When transformer meets robotic grasping: exploits context for efficient grasp detection. IEEE Robot Autom Lett 7(3):8170\u20138177. https:\/\/doi.org\/10.1109\/LRA.2022.3187261","journal-title":"IEEE Robot Autom Lett"},{"key":"11377_CR32","doi-asserted-by":"publisher","unstructured":"J. Xia, J. Chi, C. Wu and F. Zhao, \"Robot Grasping Detection in Object Overlapping Scenes Based on Multi-Stage ROI Extraction,\" 2022 34th Chinese Control and Decision Conference (CCDC), Hefei, China, 2022, pp. 5066\u20135071, https:\/\/doi.org\/10.1109\/CCDC55256.2022.10034365.","DOI":"10.1109\/CCDC55256.2022.10034365"},{"issue":"4","key":"11377_CR33","doi-asserted-by":"publisher","first-page":"199","DOI":"10.1109\/LES.2022.3181892","volume":"14","author":"M-T Le","year":"2022","unstructured":"Le M-T, Lien J-JJ (2022) Lightweight robotic grasping model based on template matching and depth image. IEEE Embed Syst Lett 14(4):199\u2013202. https:\/\/doi.org\/10.1109\/LES.2022.3181892","journal-title":"IEEE Embed Syst Lett"},{"key":"11377_CR34","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1109\/TIM.2022.3196130","volume":"71","author":"H Tian","year":"2022","unstructured":"Tian H, Song K, Li S, Ma S, Yan Y (2022) Lightweight pixel-wise generative robot grasping detection based on RGB-D dense fusion. IEEE Trans Instrum Measurement 71:1\u201312. https:\/\/doi.org\/10.1109\/TIM.2022.3196130","journal-title":"IEEE Trans Instrum Measurement"},{"issue":"3","key":"11377_CR35","doi-asserted-by":"publisher","first-page":"1384","DOI":"10.1109\/TMECH.2022.3224314","volume":"28","author":"H Cao","year":"2023","unstructured":"Cao H, Chen G, Li Z, Feng Q, Lin J, Knoll A (2023) Efficient grasp detection network with gaussian-based grasp representation for robotic manipulation. IEEE\/ASME Trans Mechatron 28(3):1384\u20131394. https:\/\/doi.org\/10.1109\/TMECH.2022.3224314","journal-title":"IEEE\/ASME Trans Mechatron"},{"key":"11377_CR36","first-page":"445","volume":"05","author":"X Yue","year":"2015","unstructured":"Yue X, Hu X, Tang L (2015) The influence of the amount of parameters in different layers on the performance of deep learning models. Computer Sci Appl 05:445\u2013453","journal-title":"Computer Sci Appl"},{"key":"11377_CR37","unstructured":"A. Canziani., P. Adam., C. Eugenio. \u201cAn Analysis of Deep Neural Network Models for Practical Applications.\u201dArXivabs\/1605.07678, 2016: n"},{"key":"11377_CR38","doi-asserted-by":"publisher","first-page":"106059","DOI":"10.1016\/j.engappai.2023.106059","volume":"121","author":"S Duan","year":"2023","unstructured":"Duan S, Tian G, Wang Z et al (2023) A semantic robotic grasping framework based on multi-task learning in stacking scenes. Eng Appl Artif Intell 121:106059","journal-title":"Eng Appl Artif Intell"},{"issue":"2\u20133","key":"11377_CR39","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1177\/0278364919859066","volume":"39","author":"D Morrison","year":"2019","unstructured":"Morrison D, Corke P, Leitner J (2019) Learning robust, real-time, reactive robotic grasping. Int J Robot Res 39(2\u20133):183\u2013201. https:\/\/doi.org\/10.1177\/0278364919859066","journal-title":"Int J Robot Res"},{"key":"11377_CR40","doi-asserted-by":"publisher","unstructured":"J. Mahler, M. Matl, X. Liu, A. Li, D. Gealy and K. Goldberg, \"Dex-Net 3.0: Computing Robust Vacuum Suction Grasp Targets in Point Clouds Using a New Analytic Model and Deep Learning,\" 2018 IEEE international conference on robotics and automation (ICRA), Brisbane, QLD, Australia, 2018, pp. 5620\u20135627, https:\/\/doi.org\/10.1109\/ICRA.2018.8460887.","DOI":"10.1109\/ICRA.2018.8460887"},{"key":"11377_CR41","doi-asserted-by":"publisher","unstructured":"S Ruder 2016 An overview of gradient descent optimization algorithms https:\/\/doi.org\/10.48550\/arXiv.1609.04747","DOI":"10.48550\/arXiv.1609.04747"},{"issue":"9","key":"11377_CR42","doi-asserted-by":"publisher","first-page":"6486","DOI":"10.1109\/TPAMI.2024.3382294","volume":"46","author":"P Zhou","year":"2024","unstructured":"Zhou P, Xie X, Lin Z, Yan S (2024) Towards understanding convergence and generalization of AdamW. IEEE Trans Pattern Anal Mach Intell 46(9):6486\u20136493. https:\/\/doi.org\/10.1109\/TPAMI.2024.3382294","journal-title":"IEEE Trans Pattern Anal Mach Intell"}],"container-title":["Neural Computing and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-025-11377-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00521-025-11377-1\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00521-025-11377-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T18:49:29Z","timestamp":1757184569000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00521-025-11377-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,9]]},"references-count":42,"journal-issue":{"issue":"21","published-print":{"date-parts":[[2025,7]]}},"alternative-id":["11377"],"URL":"https:\/\/doi.org\/10.1007\/s00521-025-11377-1","relation":{},"ISSN":["0941-0643","1433-3058"],"issn-type":[{"type":"print","value":"0941-0643"},{"type":"electronic","value":"1433-3058"}],"subject":[],"published":{"date-parts":[[2025,6,9]]},"assertion":[{"value":"12 January 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 December 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 June 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflicts of interest to this work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of interest"}},{"value":"Not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"Not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent to participate"}},{"value":"All authors have read this manuscript and would like to have it considered exclusively for publication in neural computing and applications.","order":5,"name":"Ethics","group":{"name":"EthicsHeading","label":"Consent for publication"}}]}}