{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,23]],"date-time":"2026-01-23T07:18:47Z","timestamp":1769152727143,"version":"3.49.0"},"reference-count":27,"publisher":"Springer Science and Business Media LLC","issue":"18","license":[{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,12,1]],"date-time":"2025-12-01T00:00:00Z","timestamp":1764547200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100004335","name":"Southwest University of Science and Technology","doi-asserted-by":"crossref","award":["25ycx1044"],"award-info":[{"award-number":["25ycx1044"]}],"id":[{"id":"10.13039\/501100004335","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s11760-025-05035-1","type":"journal-article","created":{"date-parts":[[2025,12,19]],"date-time":"2025-12-19T09:18:26Z","timestamp":1766135906000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["High-Fidelity Object Detection and 6D Pose Estimation for Vision-Guided 6-DoF Grasping of Chemical Vials"],"prefix":"10.1007","volume":"19","author":[{"given":"Yunxiao","family":"Li","sequence":"first","affiliation":[]},{"given":"Yueming","family":"Fang","sequence":"additional","affiliation":[]},{"given":"Hu","family":"Deng","sequence":"additional","affiliation":[]},{"given":"Yuting","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Jiyu","family":"Yang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,12,19]]},"reference":[{"issue":"95","key":"5035_CR1","doi-asserted-by":"publisher","first-page":"6991","DOI":"10.1126\/scirobotics.adm6991","volume":"9","author":"A Angelopoulos","year":"2024","unstructured":"Angelopoulos, A., Cahoon, J.F., Alterovitz, R.: Transforming science labs into automated factories of discovery. Sci. Robot. 9(95), 6991 (2024)","journal-title":"Sci. Robot."},{"key":"5035_CR2","doi-asserted-by":"crossref","unstructured":"Ochiai, K., Tahara-Arai, Y., Kato, A., Kaizu, K., Kariyazaki, H., Umeno, M., Takahashi, K., Kanda, G.N., Ozaki, H.: Automating care by self-maintainability for full laboratory automation. arXiv preprint arXiv:2501.05789 (2025)","DOI":"10.1039\/D5DD00151J"},{"key":"5035_CR3","doi-asserted-by":"crossref","unstructured":"Li, T., Song, W., Chen, N., Wang, Q., Gao, F., Xing, Y., Wu, S., Song, C., Li, J., Liu, Y., et al.: The artificial intelligence-driven intelligent laboratory for organic chemistry synthesis. Applied Sciences (2076-3417) 15(13) (2025)","DOI":"10.3390\/app15137387"},{"key":"5035_CR4","doi-asserted-by":"crossref","unstructured":"Wang, A.-L., Chen, N., Lin, K.-Y., Li, Y.-M., Zheng, W.-S.: Task-oriented 6-dof grasp pose detection in clutters. In: 2025 IEEE International Conference on Robotics and Automation (ICRA), pp. 5692\u20135698 (2025). IEEE","DOI":"10.1109\/ICRA55743.2025.11128749"},{"key":"5035_CR5","doi-asserted-by":"crossref","unstructured":"Mousavian, A., Eppner, C., Fox, D.: 6-dof graspnet: Variational grasp generation for object manipulation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2901\u20132910 (2019)","DOI":"10.1109\/ICCV.2019.00299"},{"key":"5035_CR6","doi-asserted-by":"crossref","unstructured":"Wang, S., Nikoli\u0107, M.N., Lam, T.L., Gao, Q., Ding, R., Zhang, T.: Robot manipulation based on embodied visual perception: A survey. CAAI Transactions on Intelligence Technology (2025)","DOI":"10.1049\/cit2.70022"},{"key":"5035_CR7","doi-asserted-by":"crossref","unstructured":"Chen, Y., Lin, Y., Xu, R., Vela, P.A.: Keypoint-graspnet: Keypoint-based 6-dof grasp generation from the monocular rgb-d input. In: 2023 IEEE International Conference on Robotics and Automation (ICRA), pp. 7988\u20137995 (2023). IEEE","DOI":"10.1109\/ICRA48891.2023.10161284"},{"issue":"4","key":"5035_CR8","doi-asserted-by":"publisher","first-page":"1076","DOI":"10.3390\/s24041076","volume":"24","author":"J Guan","year":"2024","unstructured":"Guan, J., Hao, Y., Wu, Q., Li, S., Fang, Y.: A survey of 6dof object pose estimation methods for different application scenarios. Sensors 24(4), 1076 (2024)","journal-title":"Sensors"},{"key":"5035_CR9","doi-asserted-by":"crossref","unstructured":"Varghese, R., Sambath, M.: Yolov8: A novel object detection algorithm with enhanced performance and robustness. In: 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), pp. 1\u20136 (2024). IEEE","DOI":"10.1109\/ADICS58448.2024.10533619"},{"key":"5035_CR10","doi-asserted-by":"crossref","unstructured":"Liu, Y., Wen, Y., Peng, S., Lin, C., Long, X., Komura, T., Wang, W.: Gen6d: Generalizable model-free 6-dof object pose estimation from rgb images. In: European Conference on Computer Vision, pp. 298\u2013315 (2022). Springer","DOI":"10.1007\/978-3-031-19824-3_18"},{"key":"5035_CR11","unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster r-cnn: Towards real-time object detection with region proposal networks. Advances in neural information processing systems 28 (2015)"},{"key":"5035_CR12","doi-asserted-by":"crossref","unstructured":"Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A.C.: Ssd: Single shot multibox detector. In: European Conference on Computer Vision, pp. 21\u201337 (2016). Springer","DOI":"10.1007\/978-3-319-46448-0_2"},{"key":"5035_CR13","doi-asserted-by":"crossref","unstructured":"Redmon, J., Divvala, S., Girshick, R., Farhadi, A.: You only look once: Unified, real-time object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 779\u2013788 (2016)","DOI":"10.1109\/CVPR.2016.91"},{"key":"5035_CR14","doi-asserted-by":"crossref","unstructured":"Redmon, J., Farhadi, A.: Yolo9000: better, faster, stronger. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 7263\u20137271 (2017)","DOI":"10.1109\/CVPR.2017.690"},{"key":"5035_CR15","unstructured":"Jocher, G., Stoken, A., Borovec, J., Changyu, L., Hogan, A., Diaconu, L., Poznanski, J., Yu, L., Rai, P., Ferriday, R., et al.: ultralytics\/yolov5: v3. 0. Zenodo (2020)"},{"key":"5035_CR16","doi-asserted-by":"crossref","unstructured":"Wang, C.-Y., Bochkovskiy, A., Liao, H.-Y.M.: Yolov7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7464\u20137475 (2023)","DOI":"10.1109\/CVPR52729.2023.00721"},{"key":"5035_CR17","first-page":"107984","volume":"37","author":"A Wang","year":"2024","unstructured":"Wang, A., Chen, H., Liu, L., Chen, K., Lin, Z., Han, J.: Yolov10: Real-time end-to-end object detection. Adv. Neural. Inf. Process. Syst. 37, 107984\u2013108011 (2024)","journal-title":"Adv. Neural. Inf. Process. Syst."},{"key":"5035_CR18","unstructured":"Khanam, R., Hussain, M.: Yolov11: An overview of the key architectural enhancements. arXiv preprint arXiv:2410.17725 (2024)"},{"key":"5035_CR19","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Schmidt, T., Narayanan, V., Fox, D.: Posecnn: A convolutional neural network for 6d object pose estimation in cluttered scenes. CoRR (2017)","DOI":"10.15607\/RSS.2018.XIV.019"},{"key":"5035_CR20","doi-asserted-by":"crossref","unstructured":"Weinzaepfel, P., Br\u00e9gier, R., Combaluzier, H., Leroy, V., Rogez, G.: Dope: Distillation of part experts for whole-body 3d pose estimation in the wild. In: European Conference on Computer Vision, pp. 380\u2013397 (2020). Springer","DOI":"10.1007\/978-3-030-58574-7_23"},{"key":"5035_CR21","doi-asserted-by":"crossref","unstructured":"Wang, C., Xu, D., Zhu, Y., Mart\u00edn-Mart\u00edn, R., Lu, C., Fei-Fei, L., Savarese, S.: Densefusion: 6d object pose estimation by iterative dense fusion. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3343\u20133352 (2019)","DOI":"10.1109\/CVPR.2019.00346"},{"key":"5035_CR22","doi-asserted-by":"crossref","unstructured":"Peng, S., Liu, Y., Huang, Q., Zhou, X., Bao, H.: Pvnet: Pixel-wise voting network for 6dof pose estimation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4561\u20134570 (2019)","DOI":"10.1109\/CVPR.2019.00469"},{"key":"5035_CR23","doi-asserted-by":"crossref","unstructured":"Wen, B., Yang, W., Kautz, J., Birchfield, S.: Foundationpose: Unified 6d pose estimation and tracking of novel objects. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 17868\u201317879 (2024)","DOI":"10.1109\/CVPR52733.2024.01692"},{"key":"5035_CR24","doi-asserted-by":"crossref","unstructured":"Wang, Z., Li, Y., Liu, Y., et al.: Improved object detection via large kernel attention[J]. Expert Systems with Applications 240, 122507 (2024)","DOI":"10.1016\/j.eswa.2023.122507"},{"key":"5035_CR25","doi-asserted-by":"crossref","unstructured":"Zhu, L., Wang, X., Ke, Z., Zhang, W., Lau, R.W.: Biformer: Vision transformer with bi-level routing attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10323\u201310333 (2023)","DOI":"10.1109\/CVPR52729.2023.00995"},{"key":"5035_CR26","doi-asserted-by":"crossref","unstructured":"Zheng, Z., Wang, P., Ren, D., et al.: Enhancing geometric factors in model learning and inference for object detection and instance segmentation[J]. IEEE transactions on cybernetics, 52(8): 8574\u20138586 (2021)","DOI":"10.1109\/TCYB.2021.3095305"},{"key":"5035_CR27","unstructured":"Tong, Z., Chen, Y., Xu, Z., Yu, R.: Wise-iou: bounding box regression loss with dynamic focusing mechanism. arXiv preprint arXiv:2301.10051 (2023)"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-05035-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-05035-1","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-05035-1.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,1,3]],"date-time":"2026-01-03T06:24:04Z","timestamp":1767421444000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-05035-1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,12]]},"references-count":27,"journal-issue":{"issue":"18","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["5035"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-05035-1","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"value":"1863-1703","type":"print"},{"value":"1863-1711","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,12]]},"assertion":[{"value":"17 September 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"4 November 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"5 December 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"19 December 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflicts of Interest"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"1447"}}