{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,6]],"date-time":"2026-02-06T05:13:16Z","timestamp":1770354796341,"version":"3.49.0"},"publisher-location":"Cham","reference-count":52,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031729195","type":"print"},{"value":"9783031729201","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,1]],"date-time":"2024-10-01T00:00:00Z","timestamp":1727740800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72920-1_9","type":"book-chapter","created":{"date-parts":[[2024,9,30]],"date-time":"2024-09-30T08:02:57Z","timestamp":1727683377000},"page":"149-166","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["$$3\\times 2$$: 3D Object Part Segmentation by\u00a02D Semantic Correspondences"],"prefix":"10.1007","author":[{"given":"Anh","family":"Thai","sequence":"first","affiliation":[]},{"given":"Weiyao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Hao","family":"Tang","sequence":"additional","affiliation":[]},{"given":"Stefan","family":"Stojanov","sequence":"additional","affiliation":[]},{"given":"James M.","family":"Rehg","sequence":"additional","affiliation":[]},{"given":"Matt","family":"Feiszli","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,1]]},"reference":[{"key":"9_CR1","doi-asserted-by":"crossref","unstructured":"Abdelreheem, A., Skorokhodov, I., Ovsjanikov, M., Wonka, P.: SATR: zero-shot semantic segmentation of 3D shapes. arXiv preprint arXiv:2304.04909 (2023)","DOI":"10.1109\/ICCV51070.2023.01392"},{"key":"9_CR2","unstructured":"Amir, S., Gandelsman, Y., Bagon, S., Dekel, T.: Deep ViT features as dense visual descriptors. arXiv preprint arXiv:2112.058142(3), 4 (2021)"},{"key":"9_CR3","doi-asserted-by":"crossref","unstructured":"Caron, M., et al.: Emerging properties in self-supervised vision transformers. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9650\u20139660 (2021)","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"9_CR4","unstructured":"Cen, J., et\u00a0al.: Segment anything in 3D with NeRFs. In: Advances in Neural Information Processing Systems, vol. 36 (2024)"},{"key":"9_CR5","unstructured":"Chang, A.X., et\u00a0al.: ShapeNet: an information-rich 3D model repository. arXiv preprint arXiv:1512.03012 (2015)"},{"key":"9_CR6","doi-asserted-by":"crossref","unstructured":"Chen, N., Liu, L., Cui, Z., Chen, R., Ceylan, D., Tu, C., Wang, W.: Unsupervised learning of intrinsic structural representation points. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9121\u20139130 (2020)","DOI":"10.1109\/CVPR42600.2020.00914"},{"key":"9_CR7","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"458","DOI":"10.1007\/978-3-030-01249-6_28","volume-title":"Computer Vision \u2013 ECCV 2018","author":"A Dai","year":"2018","unstructured":"Dai, A., Nie\u00dfner, M.: 3DMV: joint 3D-multi-view prediction for 3D semantic scene segmentation. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11214, pp. 458\u2013474. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01249-6_28"},{"key":"9_CR8","doi-asserted-by":"crossref","unstructured":"Deng, S., Xu, X., Wu, C., Chen, K., Jia, K.: 3D affordancenet: a benchmark for visual object affordance understanding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1778\u20131787 (2021)","DOI":"10.1109\/CVPR46437.2021.00182"},{"key":"9_CR9","doi-asserted-by":"crossref","unstructured":"Gupta, A., Dollar, P., Girshick, R.: LVIS: a dataset for large vocabulary instance segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5356\u20135364 (2019)","DOI":"10.1109\/CVPR.2019.00550"},{"key":"9_CR10","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"128","DOI":"10.1007\/978-3-031-20074-8_8","volume-title":"ECCV 2022","author":"J He","year":"2022","unstructured":"He, J., et al.: PartImageNet: a large, high-quality dataset of parts. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13668, pp. 128\u2013145. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20074-8_8"},{"key":"9_CR11","unstructured":"Hedlin, E., et al.: Unsupervised semantic correspondence using stable diffusion. arXiv preprint arXiv:2305.15581 (2023)"},{"key":"9_CR12","doi-asserted-by":"crossref","unstructured":"Huang, R., et al.: Segment3D: learning fine-grained class-agnostic 3D segmentation without manual labels. arXiv preprint arXiv:2312.17232 (2023)","DOI":"10.1007\/978-3-031-72754-2_16"},{"key":"9_CR13","doi-asserted-by":"crossref","unstructured":"Jaritz, M., Gu, J., Su, H.: Multi-view pointnet for 3D scene understanding. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3995\u20134003 (2019). https:\/\/api.semanticscholar.org\/CorpusID:203593088","DOI":"10.1109\/ICCVW.2019.00494"},{"key":"9_CR14","doi-asserted-by":"crossref","unstructured":"Kalogerakis, E., Hertzmann, A., Singh, K.: Learning 3D mesh segmentation and labeling. ACM Trans. Graph. 29(3) (2010)","DOI":"10.1145\/1833351.1778839"},{"key":"9_CR15","doi-asserted-by":"crossref","unstructured":"Kim, H., Sung, M.: PartSTAD: 2D-to-3D part segmentation task adaptation (2024)","DOI":"10.1007\/978-3-031-72652-1_25"},{"key":"9_CR16","unstructured":"Kirillov, A., et\u00a0al.: Segment anything. arXiv preprint arXiv:2304.02643 (2023)"},{"key":"9_CR17","doi-asserted-by":"crossref","unstructured":"Li, L.H., et\u00a0al.: Grounded language-image pre-training. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10965\u201310975 (2022)","DOI":"10.1109\/CVPR52688.2022.01069"},{"key":"9_CR18","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"110","DOI":"10.1007\/978-3-031-20074-8_7","volume-title":"ECCV 2022","author":"Y Li","year":"2022","unstructured":"Li, Y., et al.: 3D CoMPaT: composition of materials on parts of 3D things. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13668, pp. 110\u2013127. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20074-8_7"},{"key":"9_CR19","doi-asserted-by":"crossref","unstructured":"Liu, M., et al.: Partslip: low-shot part segmentation for 3D point clouds via pretrained image-language models. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 21736\u201321746 (2023)","DOI":"10.1109\/CVPR52729.2023.02082"},{"key":"9_CR20","unstructured":"Liu, W., Mao, J., Hsu, J., Hermans, T., Garg, A., Wu, J.: Composable part-based manipulation. In: 7th Annual Conference on Robot Learning (2023). https:\/\/openreview.net\/forum?id=o-K3HVUeEw"},{"key":"9_CR21","doi-asserted-by":"crossref","unstructured":"Liu, X., Xu, X., Rao, A., Gan, C., Yi, L.: AutoGPart: intermediate supervision search for generalizable 3D part segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11624\u201311634 (2022)","DOI":"10.1109\/CVPR52688.2022.01133"},{"key":"9_CR22","doi-asserted-by":"publisher","first-page":"91","DOI":"10.1023\/B:VISI.0000029664.99615.94","volume":"60","author":"DG Lowe","year":"2004","unstructured":"Lowe, D.G.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60, 91\u2013110 (2004)","journal-title":"Int. J. Comput. Vis."},{"key":"9_CR23","unstructured":"Min, J., Lee, J., Ponce, J., Cho, M.: SPair-71k: a large-scale benchmark for semantic correspondence. arXiv preprint arXiv:1908.10543 (2019)"},{"key":"9_CR24","doi-asserted-by":"crossref","unstructured":"Mo, K., et al.: PartNet: a large-scale benchmark for fine-grained and hierarchical part-level 3D object understanding. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 909\u2013918 (2019)","DOI":"10.1109\/CVPR.2019.00100"},{"key":"9_CR25","doi-asserted-by":"crossref","unstructured":"Nadeau, P., Giamou, M., Kelly, J.: The sum of its parts: visual part segmentation for inertial parameter identification of manipulated objects. arXiv preprint arXiv:2302.06685 (2023)","DOI":"10.1109\/ICRA48891.2023.10160394"},{"key":"9_CR26","doi-asserted-by":"crossref","unstructured":"Nguyen, P.D.A., et al.: Open3DIS: open-vocabulary 3D instance segmentation with 2D mask guidance (2023)","DOI":"10.1109\/CVPR52733.2024.00385"},{"key":"9_CR27","doi-asserted-by":"crossref","unstructured":"Peng, S., et\u00a0al.: OpenScene: 3D scene understanding with open vocabularies. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 815\u2013824 (2023)","DOI":"10.1109\/CVPR52729.2023.00085"},{"key":"9_CR28","unstructured":"Qi, C.R., Yi, L., Su, H., Guibas, L.J.: PointNet++: deep hierarchical feature learning on point sets in a metric space. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"9_CR29","unstructured":"Qian, G., et al.: PointNext: revisiting pointnet++ with improved training and scaling strategies. In: Advances in Neural Information Processing Systems, vol. 35, pp. 23192\u201323204 (2022)"},{"key":"9_CR30","unstructured":"Radford, A., et\u00a0al.: Learning transferable visual models from natural language supervision. In: International Conference on Machine Learning, pp. 8748\u20138763. PMLR (2021)"},{"key":"9_CR31","doi-asserted-by":"crossref","unstructured":"Ramanathan, V., et\u00a0al.: Paco: parts and attributes of common objects. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7141\u20137151 (2023)","DOI":"10.1109\/CVPR52729.2023.00690"},{"key":"9_CR32","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"550","DOI":"10.1007\/978-3-031-20086-1_32","volume-title":"ECCV 2022","author":"G Sharma","year":"2022","unstructured":"Sharma, G., Yin, K., Maji, S., Kalogerakis, E., Litany, O., Fidler, S.: MvDeCor: multi-view dense correspondence learning for fine-grained 3D segmentation. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13662, pp. 550\u2013567. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20086-1_32"},{"key":"9_CR33","unstructured":"Singh, C., Murdoch, W.J., Yu, B.: Hierarchical interpretations for neural network predictions. arXiv preprint arXiv:1806.05337 (2018)"},{"key":"9_CR34","doi-asserted-by":"crossref","unstructured":"Sun, P., et al.: Going denser with open-vocabulary part segmentation. arXiv preprint arXiv:2305.11173 (2023)","DOI":"10.1109\/ICCV51070.2023.01417"},{"key":"9_CR35","unstructured":"Takmaz, A., Fedele, E., Sumner, R.W., Pollefeys, M., Tombari, F., Engelmann, F.: Openmask3D: open-vocabulary 3D instance segmentation. arXiv preprint arXiv:2306.13631 (2023)"},{"key":"9_CR36","unstructured":"Tang, L., Jia, M., Wang, Q., Phoo, C.P., Hariharan, B.: Emergent correspondence from image diffusion. arXiv preprint arXiv:2306.03881 (2023)"},{"key":"9_CR37","doi-asserted-by":"crossref","unstructured":"Varadarajan, K.M., Vincze, M.: Object part segmentation and classification in range images for grasping. In: 2011 15th International Conference on Advanced Robotics (ICAR), pp. 21\u201327. IEEE (2011)","DOI":"10.1109\/ICAR.2011.6088647"},{"key":"9_CR38","doi-asserted-by":"crossref","unstructured":"Vu, T., Kim, K., Luu, T.M., Nguyen, T., Yoo, C.D.: Softgroup for 3D instance segmentation on point clouds. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2708\u20132717 (2022)","DOI":"10.1109\/CVPR52688.2022.00273"},{"key":"9_CR39","doi-asserted-by":"crossref","unstructured":"Wang, L., Li, X., Fang, Y.: Few-shot learning of part-specific probability space for 3D shape segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (2020)","DOI":"10.1109\/CVPR42600.2020.00456"},{"key":"9_CR40","unstructured":"Wang, R., Zhang, Y., Mao, J., Zhang, R., Cheng, C.Y., Wu, J.: Ikea-manual: seeing shape assembly step by step. In: Advances in Neural Information Processing Systems, vol. 35, pp. 28428\u201328440 (2022)"},{"key":"9_CR41","doi-asserted-by":"crossref","unstructured":"Xiang, F., et\u00a0al.: SAPIEN: a simulated part-based interactive environment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 11097\u201311107 (2020)","DOI":"10.1109\/CVPR42600.2020.01111"},{"key":"9_CR42","unstructured":"Xu, M., Yin, X., Qiu, L., Liu, Y., Tong, X., Han, X.: SAMPro3D: locating SAM prompts in 3D for zero-shot scene segmentation. arXiv preprint arXiv:2311.17707 (2023)"},{"key":"9_CR43","unstructured":"Xue, Y., Chen, N., Liu, J., Sun, W.: Zerops: high-quality cross-modal knowledge transfer for zero-shot 3D part segmentation (2023)"},{"key":"9_CR44","unstructured":"Yang, Y., Wu, X., He, T., Zhao, H., Liu, X.: SAM3D: segment anything in 3d scenes. arXiv preprint arXiv:2306.03908 (2023)"},{"key":"9_CR45","doi-asserted-by":"crossref","unstructured":"Yi, L., et al.: A scalable active framework for region annotation in 3D shape collections. In: SIGGRAPH Asia (2016)","DOI":"10.1145\/2980179.2980238"},{"key":"9_CR46","doi-asserted-by":"crossref","unstructured":"Yu, Q., Du, H., Liu, C., Yu, X.: When 3D bounding-box meets SAM: point cloud instance segmentation with weak-and-noisy supervision. arXiv abs\/2309.00828 (2023). https:\/\/api.semanticscholar.org\/CorpusID:261530997","DOI":"10.1109\/WACV57701.2024.00368"},{"key":"9_CR47","unstructured":"Zhang, J., et al.: A tale of two features: stable diffusion complements DINO for zero-shot semantic correspondence. arXiv preprint arXiv:2305.15347 (2023)"},{"key":"9_CR48","doi-asserted-by":"crossref","unstructured":"Zhao, L., Lu, J., Zhou, J.: Similarity-aware fusion network for 3D semantic segmentation. In: 2021 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 1585\u20131592 (2021). https:\/\/api.semanticscholar.org\/CorpusID:235732071","DOI":"10.1109\/IROS51168.2021.9636494"},{"key":"9_CR49","doi-asserted-by":"crossref","unstructured":"Zhao, N., Chua, T.S., Lee, G.H.: Few-shot 3D point cloud semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8873\u20138882 (2021)","DOI":"10.1109\/CVPR46437.2021.00876"},{"key":"9_CR50","unstructured":"Zhou, Y., Gu, J., Li, X., Liu, M., Fang, Y., Su, H.: PartSLIP++: enhancing low-shot 3d part segmentation via multi-view instance segmentation and maximum likelihood estimation. arXiv preprint arXiv:2312.03015 (2023)"},{"key":"9_CR51","doi-asserted-by":"publisher","unstructured":"Zhu, J., et al.: Label transfer between images and 3D shapes via local correspondence encoding. Comput. Aided Geom. Des. 71(C), 255\u2013266 (2019). https:\/\/doi.org\/10.1016\/j.cagd.2019.04.009","DOI":"10.1016\/j.cagd.2019.04.009"},{"key":"9_CR52","doi-asserted-by":"crossref","unstructured":"Zhu, X., et al.: PointCLIP V2: prompting clip and GPT for powerful 3D open-world learning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 2639\u20132650 (2023)","DOI":"10.1109\/ICCV51070.2023.00249"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72920-1_9","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,28]],"date-time":"2024-11-28T21:53:11Z","timestamp":1732830791000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72920-1_9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,1]]},"ISBN":["9783031729195","9783031729201"],"references-count":52,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72920-1_9","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,1]]},"assertion":[{"value":"1 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}