{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,3,8]],"date-time":"2026-03-08T05:32:35Z","timestamp":1772947955433,"version":"3.50.1"},"reference-count":72,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T00:00:00Z","timestamp":1772841600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T00:00:00Z","timestamp":1772841600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100003995","name":"Anhui Provincial Natural Science Foundation","doi-asserted-by":"crossref","award":["2508085MF142"],"award-info":[{"award-number":["2508085MF142"]}],"id":[{"id":"10.13039\/501100003995","id-type":"DOI","asserted-by":"crossref"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J Comput Vis"],"published-print":{"date-parts":[[2026,4]]},"DOI":"10.1007\/s11263-025-02716-9","type":"journal-article","created":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T15:56:53Z","timestamp":1772899013000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["TARGO and TARGO-Net: Benchmarking Target-Driven Object Grasping Under Occlusions"],"prefix":"10.1007","volume":"134","author":[{"ORCID":"https:\/\/orcid.org\/0000-0001-6684-9814","authenticated-orcid":false,"given":"Yan","family":"Xia","sequence":"first","affiliation":[]},{"given":"Ran","family":"Ding","sequence":"additional","affiliation":[]},{"given":"Ziyuan","family":"Qin","sequence":"additional","affiliation":[]},{"given":"Guanqi","family":"Zhan","sequence":"additional","affiliation":[]},{"given":"Kaichen","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Long","family":"Yang","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Daniel","family":"Cremers","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2026,3,7]]},"reference":[{"key":"2716_CR1","doi-asserted-by":"crossref","unstructured":"Back, S., Lee, J., Kim, K., Rho, H., Lee, G., Kang, R., Lee, S., Noh, S., Lee, Y., Lee, T.et\u00a0al. (2025). Graspclutter6d: A large-scale real-world dataset for robust perception and grasping in cluttered scenes, arXiv preprint arXiv:2504.06866.","DOI":"10.1109\/LRA.2025.3601045"},{"key":"2716_CR2","unstructured":"Breyer, M., Chung, J.\u00a0J., Ott, L., Siegwart, R., Nieto, J. (2021). Volumetric grasping network: Real-time 6 dof grasp detection in clutter, In Conference on Robot Learning, pp.\u00a01602\u20131611, PML"},{"key":"2716_CR3","doi-asserted-by":"crossref","unstructured":"Breyer, M., Ott, L., Siegwart, R. & Chung, J.\u00a0J. (2022). Closed-loop next-best-view planning for target-driven grasping, In 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp.\u00a01411\u20131416, IEEE.","DOI":"10.1109\/IROS47612.2022.9981472"},{"key":"2716_CR4","doi-asserted-by":"crossref","unstructured":"Calli, B., Singh, A., Walsman, A., Srinivasa, S., Abbeel, P. & Dollar, A.\u00a0M. (2015). The ycb object and model set: Towards common benchmarks for manipulation research, In 2015 international conference on advanced robotics (ICAR), pp.\u00a0510\u2013517, IEEE.","DOI":"10.1109\/ICAR.2015.7251504"},{"key":"2716_CR5","doi-asserted-by":"crossref","unstructured":"Chao, Y.-W., Yang, W., Xiang, Y., Molchanov, P., Handa, A., Tremblay, J., Narang, Y.\u00a0S., Van\u00a0Wyk, K., Iqbal, U., Birchfield, S., et\u00a0al. (2021). Dexycb: A benchmark for capturing hand grasping of objects, In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a09044\u20139053.","DOI":"10.1109\/CVPR46437.2021.00893"},{"key":"2716_CR6","unstructured":"Chen, Y., Zeng, E.\u00a0Z., Gilles, M. & Wong, A. (2021). Metagraspnet_v0: A large-scale benchmark dataset for vision-driven robotic grasping via physics-based metaverse synthesis, arXiv preprint arXiv:2112.14663."},{"key":"2716_CR7","doi-asserted-by":"crossref","unstructured":"Cheong, S.\u00a0H., Cho, B.\u00a0Y., Lee, J., Kim, C. & Nam, C. (2020). Where to relocate?: Object rearrangement inside cluttered and confined environments for robotic manipulation, in 2020 IEEE International Conference on Robotics and Automation (ICRA), pp.\u00a07791\u20137797, IEEE.","DOI":"10.1109\/ICRA40945.2020.9197485"},{"key":"2716_CR8","doi-asserted-by":"crossref","unstructured":"Choy, C., Gwak, J. & Savarese, S. (2019). 4d spatio-temporal convnets: Minkowski convolutional neural networks, In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp.\u00a03075\u20133084.","DOI":"10.1109\/CVPR.2019.00319"},{"issue":"10","key":"2716_CR9","doi-asserted-by":"publisher","first-page":"3941","DOI":"10.1109\/TIE.2009.2025293","volume":"56","author":"W Chung","year":"2009","unstructured":"Chung, W., Kim, S., Choi, M., Choi, J., Kim, H., Moon, C.-B., & Song, J.-B. (2009). Safe navigation of a mobile robot considering visibility of environment. IEEE Transactions on Industrial Electronics, 56(10), 3941\u20133950.","journal-title":"IEEE Transactions on Industrial Electronics"},{"key":"2716_CR10","unstructured":"Coleman, D., Sucan, I., Chitta, S. & Correll, N. (2014). Reducing the barrier to entry of complex robotic software: a moveit! case study, arXiv preprint arXiv:1404.3785."},{"key":"2716_CR11","unstructured":"Coumans, E. & Bai, Y. (2016). Pybullet, a python module for physics simulation for games, robotics and machine learning. http:\/\/pybullet.org, 2020."},{"key":"2716_CR12","doi-asserted-by":"crossref","unstructured":"Eppner, C., Mousavian, A. & Fox, D. (2021). Acronym: A large-scale grasp dataset based on simulation, in 2021 IEEE International Conference on Robotics and Automation (ICRA), pp.\u00a06222\u20136227, IEEE.","DOI":"10.1109\/ICRA48506.2021.9560844"},{"key":"2716_CR13","doi-asserted-by":"crossref","unstructured":"Fang, H.-S., Wang, C., Gou, M. & Lu, C. (2020). Graspnet-1billion: A large-scale benchmark for general object grasping, In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp.\u00a011444\u201311453.","DOI":"10.1109\/CVPR42600.2020.01146"},{"key":"2716_CR14","doi-asserted-by":"crossref","unstructured":"Gilles, M., Chen, Y., Zeng, E.\u00a0Z., Wu, Y., Furmans, K., Wong, A. & Rayyes, R. (2023). Metagraspnetv2: All-in-one dataset enabling fast and reliable robotic bin picking via object relationship reasoning and dexterous grasping, IEEE Transactions on Automation Science and Engineering.","DOI":"10.1109\/TASE.2023.3328964"},{"key":"2716_CR15","first-page":"65642","volume":"36","author":"S Grover","year":"2024","unstructured":"Grover, S., Vineet, V., & Rawat, Y. (2024). Revealing the unseen: Benchmarking video action recognition under occlusion. Advances in Neural Information Processing Systems, 36, 65642.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2716_CR16","doi-asserted-by":"crossref","unstructured":"Guo, A., Wen, B., Yuan, J., Tremblay, J., Tyree, S., Smith, J. & Birchfield, S. (2023). Handal: A dataset of real-world manipulable object categories with pose annotations, affordances, and reconstructions, In 2023 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp.\u00a011428\u201311435, IEEE.","DOI":"10.1109\/IROS55552.2023.10341672"},{"key":"2716_CR17","doi-asserted-by":"crossref","unstructured":"Guo, A., Wen, B., Yuan, J., Tremblay, J., Tyree, S., Smith, J. & Birchfield, S. (2023). HANDAL: A dataset of real-world manipulable object categories with pose annotations, affordances, and reconstructions, In IROS.","DOI":"10.1109\/IROS55552.2023.10341672"},{"key":"2716_CR18","unstructured":"Hoorick, B.\u00a0V., Tokmakov, P., Stent, S., Li, J. & Vondrick, C. (2023). Tracking through containers and occluders in the wild."},{"key":"2716_CR19","unstructured":"Hsieh, C.-Y., Khurana, Dave, T.A. & Ramanan, D. (2023). Tracking any object amodally, arXiv preprint arXiv:2312.12433."},{"key":"2716_CR20","doi-asserted-by":"crossref","unstructured":"Huang, H., Wang, D., Zhu, X., Walters, R. & Platt, R. (2023). Edge grasp network: A graph-based se (3)-invariant approach to grasp detection, in 2023 IEEE International Conference on Robotics and Automation (ICRA), pp.\u00a03882\u20133888, IEEE.","DOI":"10.1109\/ICRA48891.2023.10160728"},{"key":"2716_CR21","doi-asserted-by":"crossref","unstructured":"Jefferies, M.\u00a0E., Baker, J. & Weng, W. (2008). Robot cognitive mapping\u2013a role for a global metric map in a cognitive mapping process, In Robotics and cognitive approaches to spatial mapping, pp.\u00a0265\u2013279, Springer.","DOI":"10.1007\/978-3-540-75388-9_16"},{"key":"2716_CR22","doi-asserted-by":"crossref","unstructured":"Jiang, Y., Moseson, S. & Saxena, A. (2011). Efficient grasping from rgbd images: Learning using a new rectangle representation, In 2011 IEEE International conference on robotics and automation, pp.\u00a03304\u20133311, IEEE.","DOI":"10.1109\/ICRA.2011.5980145"},{"key":"2716_CR23","doi-asserted-by":"crossref","unstructured":"Jiang, Z., Zhu, Y., Svetlik, M., Fang, K. & Zhu, Y. (2021). Synergies between affordance and geometry: 6-dof grasp detection via implicit representations, Robotics: Science and Systems.","DOI":"10.15607\/RSS.2021.XVII.024"},{"issue":"5","key":"2716_CR24","doi-asserted-by":"publisher","first-page":"433","DOI":"10.1109\/34.765655","volume":"21","author":"AE Johnson","year":"1999","unstructured":"Johnson, A. E., & Hebert, M. (1999). Using spin images for efficient object recognition in cluttered 3d scenes. IEEE Transactions on pattern analysis and machine intelligence, 21(5), 433\u2013449.","journal-title":"IEEE Transactions on pattern analysis and machine intelligence"},{"key":"2716_CR25","doi-asserted-by":"crossref","unstructured":"Jung, H., Wu, S.-C., Ruhkamp, P., Zhai, G., Schieber, H., Rizzoli, G., Wang, P., Zhao, H., Garattoni, L., Meier, S. et\u00a0al. (2024). Housecat6d-a large-scale multi-modal category level 6d object perception dataset with household objects in realistic scenarios, In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a022498\u201322508.","DOI":"10.1109\/CVPR52733.2024.02123"},{"key":"2716_CR26","doi-asserted-by":"crossref","unstructured":"Ke, L., Tai, Y.-W. & Tang, C.-K. (2021). Deep occlusion-aware instance segmentation with overlapping bilayers, In Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition (CVPR), pp.\u00a04019\u2013402.","DOI":"10.1109\/CVPR46437.2021.00401"},{"issue":"4","key":"2716_CR27","doi-asserted-by":"publisher","first-page":"8783","DOI":"10.1109\/LRA.2022.3188437","volume":"7","author":"M Kiatos","year":"2022","unstructured":"Kiatos, M., Sarantopoulos, I., Koutras, L., Malassiotis, S., & Doulgeri, Z. (2022). Learning push-grasping in dense clutter. IEEE Robotics and Automation Letters, 7(4), 8783\u20138790.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2716_CR28","unstructured":"Kingma, D.\u00a0P. & Ba, J. Adam: A method for stochastic optimization, The 3rd International Conference for Learning Representations, (2015)."},{"key":"2716_CR29","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.\u00a0C., Lo, W.-Y. et\u00a0al. (2023). Segment anything, arXiv preprint arXiv:2304.02643.","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"2716_CR30","doi-asserted-by":"crossref","unstructured":"Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A.\u00a0C., Lo, W.-Y. et\u00a0al. (2023). Segment anything, in Proceedings of the IEEE\/CVF international conference on computer vision, pp.\u00a04015\u20134026.","DOI":"10.1109\/ICCV51070.2023.00371"},{"key":"2716_CR31","doi-asserted-by":"crossref","unstructured":"Kortylewski, A., He, J., Liu, Q. & Yuille, A.\u00a0L. (2020). Compositional convolutional neural networks: A deep architecture with innate robustness to partial occlusion, In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a08940\u2013894.","DOI":"10.1109\/CVPR42600.2020.00896"},{"key":"2716_CR32","doi-asserted-by":"crossref","unstructured":"Kuffner, J.\u00a0J. (2004). Effective sampling and distance metrics for 3d rigid body path planning, In IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA\u201904. 2004, vol.\u00a04, pp.\u00a03993\u20133998, IEEE.","DOI":"10.1109\/ROBOT.2004.1308895"},{"key":"2716_CR33","doi-asserted-by":"crossref","unstructured":"Kurenkov, A., Taglic, J., Kulkarni, R., Dominguez-Kuhne, M., Garg, A., Mart\u00edn-Mart\u00edn, R. & Savarese, S. (2020). Visuomotor mechanical search: Learning to retrieve target objects in clutter, In 2020 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp.\u00a08408\u20138414, IEEE.","DOI":"10.1109\/IROS45743.2020.9341545"},{"key":"2716_CR34","doi-asserted-by":"crossref","unstructured":"Lee, J., Cho, Y., Nam, C., Park, J. & Kim, C. (2019). Efficient obstacle rearrangement for object manipulation tasks in cluttered environments, in 2019 International Conference on Robotics and Automation (ICRA), pp.\u00a0183\u2013189, IEEE.","DOI":"10.1109\/ICRA.2019.8793616"},{"issue":"4\u20135","key":"2716_CR35","doi-asserted-by":"publisher","first-page":"421","DOI":"10.1177\/0278364917710318","volume":"37","author":"S Levine","year":"2018","unstructured":"Levine, S., Pastor, P., Krizhevsky, A., Ibarz, J., & Quillen, D. (2018). Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection. The International Journal of Robotics Research, 37(4\u20135), 421\u2013436.","journal-title":"The International Journal of Robotics Research"},{"key":"2716_CR36","unstructured":"Li, Y., Cheng, K., Wu, R., Shen, Y., Zhou, K. & Dong, H. (2024). Mobileafford: Mobile robotic manipulation through differentiable affordance learning, in 2nd Workshop on Mobile Manipulation and Embodied Intelligence at ICRA 2024."},{"key":"2716_CR37","doi-asserted-by":"crossref","unstructured":"Li, Y., Wu, R., Lu, H., Ning, C., Shen, Y., Zhan, G. & Dong, H. (2024). Broadcasting support relations recursively from local dynamics for object retrieval in clutters, In Robotics: Science and Systems (RSS).","DOI":"10.15607\/RSS.2024.XX.070"},{"key":"2716_CR38","doi-asserted-by":"crossref","unstructured":"Liu, Z., Wang, Z., Huang, S., Zhou, J. & Lu, J. (2022). Ge-grasp: Efficient target-oriented grasping in dense clutter, In 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp.\u00a01388\u20131395, IEEE.","DOI":"10.1109\/IROS47612.2022.9981499"},{"key":"2716_CR39","unstructured":"Lundberg, S.\u00a0M. & Lee, S.-I. (2017). A unified approach to interpreting model predictions, Advances in neural information processing systems, vol.\u00a030."},{"key":"2716_CR40","first-page":"65642","volume":"36","author":"R Modi","year":"2023","unstructured":"Modi, R., Vineet, V., & Rawat, Y. (2023). On occlusions in video action detection: Benchmark datasets and training recipes. Advances in Neural Information Processing Systems, 36, 65642.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2716_CR41","doi-asserted-by":"publisher","first-page":"45129","DOI":"10.1109\/ACCESS.2023.3273289","volume":"11","author":"T Motoda","year":"2023","unstructured":"Motoda, T., Petit, D., Nishi, T., Nagata, K., Wan, W., & Harada, K. (2023). Multi-step object extraction planning from clutter based on support relations. IEEE Access, 11, 45129\u201345139.","journal-title":"IEEE Access"},{"key":"2716_CR42","doi-asserted-by":"crossref","unstructured":"Murali, A., Mousavian, A., Eppner, C., Paxton, C. & Fox, D. (2020). 6-dof grasping for target-driven object manipulation in clutter, In 2020 IEEE International Conference on Robotics and Automation (ICRA), pp.\u00a06232\u20136238, IEEE.","DOI":"10.1109\/ICRA40945.2020.9197318"},{"key":"2716_CR43","doi-asserted-by":"crossref","unstructured":"Ozguroglu, E., Liu, R., Sur\u00eds, D., Chen, D., Dave, A., Tokmakov, P. & Vondrick, C. (2024). pix2gestalt: Amodal segmentation by synthesizing wholes, arXiv preprint arXiv:2401.14398.","DOI":"10.1109\/CVPR52733.2024.00377"},{"key":"2716_CR44","doi-asserted-by":"crossref","unstructured":"Peng, S., Niemeyer, M., Mescheder, L., Pollefeys, M. & Geiger, A. (2020). Convolutional occupancy networks, in Computer Vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part III 16, pp.\u00a0523\u2013540, Springer.","DOI":"10.1007\/978-3-030-58580-8_31"},{"issue":"2","key":"2716_CR45","doi-asserted-by":"publisher","first-page":"1697","DOI":"10.1109\/LRA.2022.3141662","volume":"7","author":"M St\u00f6lzle","year":"2022","unstructured":"St\u00f6lzle, M., Miki, T., Gerdes, L., Azkarate, M., & Hutter, M. (2022). Reconstructing occluded elevation information in terrain maps with self-supervised learning. IEEE Robotics and Automation Letters, 7(2), 1697\u20131704.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2716_CR46","doi-asserted-by":"crossref","unstructured":"Sundermeyer, M., Mousavian, A., Triebel, R. & Fox, D. (2021). Contact-graspnet: Efficient 6-dof grasp generation in cluttered scenes, In 2021 IEEE International Conference on Robotics and Automation (ICRA), pp.\u00a013438\u201313444, IEEE.","DOI":"10.1109\/ICRA48506.2021.9561877"},{"key":"2716_CR47","unstructured":"Vaswani, A. (2017). Attention is all you need, Advances in Neural Information Processing Systems."},{"key":"2716_CR48","doi-asserted-by":"crossref","unstructured":"Wang, K.-J., Liu, Y.-H., Su, H.-T., Wang, J.-W., Wang, Y.-S., Hsu, W.\u00a0H. & Chen, W.-C. (2021). Ocid-ref: A 3d robotic dataset with embodied language for clutter scene grounding, arXiv preprint arXiv:2103.07679.","DOI":"10.18653\/v1\/2021.naacl-main.419"},{"key":"2716_CR49","doi-asserted-by":"crossref","unstructured":"Wang, A., Sun, Y., Kortylewski, A. & Yuille, A.\u00a0L. (2020). Robust object detection under occlusion with context-aware compositionalnets, In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a012645\u201312654.","DOI":"10.1109\/CVPR42600.2020.01266"},{"key":"2716_CR50","doi-asserted-by":"crossref","unstructured":"Wang, L., Ye, H., Wang, Q., Gao, Y., Xu, C. & Gao, F. (2021). Learning-based 3d occupancy prediction for autonomous navigation in occluded environments, In 2021 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp.\u00a04509\u20134516, IEEE.","DOI":"10.1109\/IROS51168.2021.9636333"},{"key":"2716_CR51","doi-asserted-by":"crossref","unstructured":"Wang, R., Zhang, J., Chen, J., Xu, Y., Li, P., Liu, T. & Wang, H. (2023). Dexgraspnet: A large-scale robotic dexterous grasp dataset for general objects based on simulation, In 2023 IEEE International Conference on Robotics and Automation (ICRA), pp.\u00a011359\u201311366, IEEE.","DOI":"10.1109\/ICRA48891.2023.10160982"},{"key":"2716_CR52","doi-asserted-by":"crossref","unstructured":"Wen, B., Mitash, C., Soorian, S., Kimmel, A., Sintov, A. & Bekris, K.\u00a0E. (2020). Robust, occlusion-aware pose estimation for objects grasped by adaptive hands, in 2020 IEEE International Conference on Robotics and Automation (ICRA), pp.\u00a06210\u20136217, IEEE.","DOI":"10.1109\/ICRA40945.2020.9197350"},{"key":"2716_CR53","first-page":"60966","volume":"36","author":"R Wu","year":"2023","unstructured":"Wu, R., Cheng, K., Zhao, Y., Ning, C., Zhan, G., & Dong, H. (2023). Learning environment-aware affordance for 3d articulated object manipulation under occlusions. Advances in Neural Information Processing Systems, 36, 60966.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2716_CR54","doi-asserted-by":"crossref","unstructured":"Xia, Y., Gladkova, M., Wang, R., Li, Q., Stilla, U., Henriques, J.\u00a0F. & Cremers, D. (2023). Casspr: Cross attention single scan place recognition, In Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp.\u00a08461\u20138472.","DOI":"10.1109\/ICCV51070.2023.00777"},{"key":"2716_CR55","doi-asserted-by":"crossref","unstructured":"Xia, Y., Xia, Y., Li, W., Song, R., Cao, K. & Stilla, U. (2021). Asfm-net: Asymmetrical siamese feature matching network for point completion, In Proceedings of the 29th ACM international conference on multimedia, pp.\u00a01938\u20131947.","DOI":"10.1145\/3474085.3475348"},{"key":"2716_CR56","doi-asserted-by":"publisher","first-page":"166","DOI":"10.1016\/j.isprsjprs.2021.01.027","volume":"174","author":"Y Xia","year":"2021","unstructured":"Xia, Y., Xu, Y., Wang, C., & Stilla, U. (2021). Vpc-net: Completion of 3d vehicles from mls point clouds. ISPRS Journal of Photogrammetry and Remote Sensing, 174, 166\u2013181.","journal-title":"ISPRS Journal of Photogrammetry and Remote Sensing"},{"key":"2716_CR57","doi-asserted-by":"crossref","unstructured":"Xu, K., Zhang, L. & Shi, J. (2023). Amodal completion via progressive mixed context diffusion, arXiv preprint arXiv:2312.15540.","DOI":"10.1109\/CVPR52733.2024.00869"},{"issue":"2","key":"2716_CR58","doi-asserted-by":"publisher","first-page":"2232","DOI":"10.1109\/LRA.2020.2970622","volume":"5","author":"Y Yang","year":"2020","unstructured":"Yang, Y., Liang, H., & Choi, C. (2020). A deep learning approach to grasping the invisible. IEEE Robotics and Automation Letters, 5(2), 2232\u20132239.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2716_CR59","doi-asserted-by":"crossref","unstructured":"Yu, X., Rao, Y., Wang, Z., Lu, J. & Zhou, J. (2023). Adapointr: Diverse point cloud completion with adaptive geometry-aware transformers, IEEE Transactions on Pattern Analysis and Machine Intelligence.","DOI":"10.1109\/TPAMI.2023.3309253"},{"key":"2716_CR60","doi-asserted-by":"crossref","unstructured":"Yuan, X., Kortylewski, A., Sun, Y. & Yuille, A. (2021). Robust instance segmentation through reasoning about multi-object occlusion, In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp.\u00a011141\u201311150.","DOI":"10.1109\/CVPR46437.2021.01099"},{"issue":"18","key":"2716_CR61","doi-asserted-by":"publisher","first-page":"10996","DOI":"10.1109\/JSEN.2020.2995395","volume":"20","author":"Y Yu","year":"2020","unstructured":"Yu, Y., Cao, Z., Liang, S., Geng, W., & Yu, J. (2020). A novel vision-based grasping method under occlusion for manipulating robotic system. IEEE Sensors Journal, 20(18), 10996\u201311006.","journal-title":"IEEE Sensors Journal"},{"issue":"2","key":"2716_CR62","doi-asserted-by":"publisher","first-page":"884","DOI":"10.1109\/TMECH.2022.3209488","volume":"28","author":"S Yu","year":"2022","unstructured":"Yu, S., Zhai, D.-H., & Xia, Y. (2022). Cgnet: Robotic grasp detection in heavily cluttered scenes. IEEE\/ASME Transactions on Mechatronics, 28(2), 884\u2013894.","journal-title":"IEEE\/ASME Transactions on Mechatronics"},{"issue":"7","key":"2716_CR63","doi-asserted-by":"publisher","first-page":"690","DOI":"10.1177\/0278364919868017","volume":"41","author":"A Zeng","year":"2022","unstructured":"Zeng, A., Song, S., Yu, K.-T., Donlon, E., Hogan, F. R., Bauza, M., Ma, D., Taylor, O., Liu, M., Romo, E., et al. (2022). Robotic pick-and-place of novel objects in clutter with multi-affordance grasping and cross-domain image matching. The International Journal of Robotics Research, 41(7), 690\u2013705.","journal-title":"The International Journal of Robotics Research"},{"key":"2716_CR64","doi-asserted-by":"crossref","unstructured":"Zhan, X., Pan, X., Dai, B., Liu, Z., Lin, D. & Loy, C.\u00a0C. (2020). Self-supervised scene de-occlusion, In Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a03784\u20133792.","DOI":"10.1109\/CVPR42600.2020.00384"},{"key":"2716_CR65","unstructured":"Zhan, G., Xie, W. & Zisserman, A. (2022). A tri-layer plugin to improve occluded detection, British Machine Vision Conference (BMVC)."},{"key":"2716_CR66","unstructured":"Zhang, X., Wang, D., Han, S., Li, W., Zhao, B., Wang, Z., Duan, X., Fang, C., Li, X. & He, J. (2023). Affordance-driven next-best-view planning for robotic grasping, arXiv preprint arXiv:2309.09556."},{"key":"2716_CR67","unstructured":"Zhang, J., Wu, M. & Dong, H. (2023). Genpose: Generative category-level object pose estimation via diffusion models, arXiv preprint arXiv:2306.10531."},{"issue":"2","key":"2716_CR68","doi-asserted-by":"publisher","first-page":"2929","DOI":"10.1109\/LRA.2022.3142401","volume":"7","author":"H Zhang","year":"2022","unstructured":"Zhang, H., Yang, D., Wang, H., Zhao, B., Lan, X., Ding, J., & Zheng, N. (2022). Regrad: A large-scale relational grasp dataset for safe and object-specific robotic grasping in clutter. IEEE Robotics and Automation Letters, 7(2), 2929\u20132936.","journal-title":"IEEE Robotics and Automation Letters"},{"key":"2716_CR69","doi-asserted-by":"crossref","unstructured":"Zhou, K., Hong, L., Chen, C., Xu, H., Ye, C., Hu, Q. & Li, Z. (2022). Devnet: Self-supervised monocular depth learning via density volume construction, In European Conference on Computer Vision, pp.\u00a0125\u2013142, Springer.","DOI":"10.1007\/978-3-031-19842-7_8"},{"key":"2716_CR70","first-page":"69532","volume":"36","author":"K Zhou","year":"2024","unstructured":"Zhou, K., Zhong, J.-X., Shin, S., Lu, K., Yang, Y., Markham, A., & Trigoni, N. (2024). Dynpoint: Dynamic neural point for view synthesis. Advances in Neural Information Processing Systems, 36, 69532.","journal-title":"Advances in Neural Information Processing Systems"},{"key":"2716_CR71","doi-asserted-by":"crossref","unstructured":"Zhu, Y., Tian, Y., Metaxas, D. & Doll\u00e1r, P. (2017). Semantic amodal segmentation, In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp.\u00a01464\u2013147.","DOI":"10.1109\/CVPR.2017.320"},{"key":"2716_CR72","doi-asserted-by":"crossref","unstructured":"Zurbr\u00fcgg, R., Liu, Y., Engelmann, F., Kumar, S., Hutter, M., Patil, V. & Yu, F. (2024). Icgnet: A unified approach for instance-centric grasping, arXiv preprint arXiv:2401.09939.","DOI":"10.1109\/ICRA57147.2024.10611725"}],"container-title":["International Journal of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02716-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11263-025-02716-9","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11263-025-02716-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2026,3,7]],"date-time":"2026-03-07T15:57:00Z","timestamp":1772899020000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11263-025-02716-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2026,3,7]]},"references-count":72,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2026,4]]}},"alternative-id":["2716"],"URL":"https:\/\/doi.org\/10.1007\/s11263-025-02716-9","relation":{},"ISSN":["0920-5691","1573-1405"],"issn-type":[{"value":"0920-5691","type":"print"},{"value":"1573-1405","type":"electronic"}],"subject":[],"published":{"date-parts":[[2026,3,7]]},"assertion":[{"value":"30 May 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"15 December 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"7 March 2026","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}],"article-number":"170"}}