{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,8,29]],"date-time":"2025-08-29T10:25:01Z","timestamp":1756463101193,"version":"3.40.3"},"publisher-location":"Cham","reference-count":61,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031730207"},{"type":"electronic","value":"9783031730214"}],"license":[{"start":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T00:00:00Z","timestamp":1732147200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,11,21]],"date-time":"2024-11-21T00:00:00Z","timestamp":1732147200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-73021-4_21","type":"book-chapter","created":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T09:20:00Z","timestamp":1732094400000},"page":"352-369","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["FAFA: Frequency-Aware Flow-Aided Self-supervision for\u00a0Underwater Object Pose Estimation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6681-2103","authenticated-orcid":false,"given":"Jingyi","family":"Tang","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-0759-0782","authenticated-orcid":false,"given":"Gu","family":"Wang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0009-0004-5418-8827","authenticated-orcid":false,"given":"Zeyu","family":"Chen","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-4504-3659","authenticated-orcid":false,"given":"Shengquan","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-0403-1923","authenticated-orcid":false,"given":"Xiu","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-7333-9975","authenticated-orcid":false,"given":"Xiangyang","family":"Ji","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,11,21]]},"reference":[{"unstructured":"Borkman, S., et al.: Unity perception: generate synthetic data for computer vision. arXiv preprint arXiv:2107.04259 (2021)","key":"21_CR1"},{"unstructured":"Bukschat, Y., Vetter, M.: Efficientpose: an efficient, accurate and scalable end-to-end 6D multi object pose estimation approach. arXiv preprint arXiv:2011.04307 (2020)","key":"21_CR2"},{"issue":"4","key":"21_CR3","doi-asserted-by":"publisher","first-page":"98","DOI":"10.4031\/MTSJ.50.4.7","volume":"50","author":"G Casalino","year":"2016","unstructured":"Casalino, G., et al.: Underwater intervention robotics: an outline of the Italian national project Maris. Mar. Technol. Soc. J. 50(4), 98\u2013107 (2016)","journal-title":"Mar. Technol. Soc. J."},{"doi-asserted-by":"crossref","unstructured":"Chang, W.G., You, T., Seo, S., Kwak, S., Han, B.: Domain-specific batch normalization for unsupervised domain adaptation. In: CVPR (2019)","key":"21_CR4","DOI":"10.1109\/CVPR.2019.00753"},{"key":"21_CR5","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"533","DOI":"10.1007\/978-3-031-19842-7_31","volume-title":"ECCV 2022","author":"K Chen","year":"2022","unstructured":"Chen, K., et al.: Sim-to-real 6D object pose estimation via iterative self-training for robotic bin picking. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13699, pp. 533\u2013550. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-19842-7_31"},{"doi-asserted-by":"crossref","unstructured":"Deng, X., Xiang, Y., Mousavian, A., Eppner, C., Bretl, T., Fox, D.: Self-supervised 6D object pose estimation for robot manipulation. In: ICRA, pp. 3665\u20133671. IEEE (2020)","key":"21_CR6","DOI":"10.1109\/ICRA40945.2020.9196714"},{"unstructured":"Denninger, M., et al.: Blenderproc: reducing the reality gap with photorealistic rendering. In: RSS (2020)","key":"21_CR7"},{"issue":"4","key":"21_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3524496","volume":"55","author":"Z Fan","year":"2022","unstructured":"Fan, Z., Zhu, Y., He, Y., Sun, Q., Liu, H., He, J.: Deep learning on monocular object pose detection and tracking: a comprehensive overview. ACM Comput. Surv. 55(4), 1\u201340 (2022)","journal-title":"ACM Comput. Surv."},{"key":"21_CR9","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3578516","volume":"55","author":"SP Gonz\u00e1lez-Sabbagh","year":"2023","unstructured":"Gonz\u00e1lez-Sabbagh, S.P., Robles-Kelly, A.: A survey on underwater computer vision. ACM Comput. Surv. 55, 1\u201339 (2023)","journal-title":"ACM Comput. Surv."},{"doi-asserted-by":"crossref","unstructured":"Hai, Y., Song, R., Li, J., Ferstl, D., Hu, Y.: Pseudo flow consistency for self-supervised 6D object pose estimation. In: ICCV, pp. 14075\u201314085 (2023)","key":"21_CR10","DOI":"10.1109\/ICCV51070.2023.01294"},{"doi-asserted-by":"crossref","unstructured":"Hai, Y., Song, R., Li, J., Hu, Y.: Shape-constraint recurrent flow for 6D object pose estimation. In: CVPR, pp. 4831\u20134840 (2023)","key":"21_CR11","DOI":"10.1109\/CVPR52729.2023.00468"},{"issue":"7","key":"21_CR12","doi-asserted-by":"publisher","first-page":"1873","DOI":"10.1364\/JOSAA.24.001873","volume":"24","author":"BC Hansen","year":"2007","unstructured":"Hansen, B.C., Hess, R.F.: Structural sparseness and spatial phase alignment in natural scenes. JOSA A 24(7), 1873\u20131885 (2007)","journal-title":"JOSA A"},{"doi-asserted-by":"crossref","unstructured":"He, C., et al.: Camouflaged object detection with feature decomposition and edge reconstruction. In: CVPR, pp. 22046\u201322055 (2023)","key":"21_CR13","DOI":"10.1109\/CVPR52729.2023.02111"},{"key":"21_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"548","DOI":"10.1007\/978-3-642-37331-2_42","volume-title":"Computer Vision \u2013 ACCV 2012","author":"S Hinterstoisser","year":"2013","unstructured":"Hinterstoisser, S., et al.: Model based training, detection and pose estimation of texture-less 3D objects in heavily cluttered scenes. In: Lee, K.M., Matsushita, Y., Rehg, J.M., Hu, Z. (eds.) ACCV 2012. LNCS, vol. 7724, pp. 548\u2013562. Springer, Heidelberg (2013). https:\/\/doi.org\/10.1007\/978-3-642-37331-2_42"},{"key":"21_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"606","DOI":"10.1007\/978-3-319-49409-8_52","volume-title":"Computer Vision \u2013 ECCV 2016 Workshops","author":"T Hoda\u0148","year":"2016","unstructured":"Hoda\u0148, T., Matas, J., Obdr\u017e\u00e1lek, \u0160: On evaluation of 6D object pose estimation. In: Hua, G., J\u00e9gou, H. (eds.) ECCV 2016. LNCS, vol. 9915, pp. 606\u2013619. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-49409-8_52"},{"doi-asserted-by":"crossref","unstructured":"Hodan, T., et al.: BOP challenge 2023 on detection segmentation and pose estimation of seen and unseen rigid objects. In: CVPRW, pp. 5610\u20135619 (2024)","key":"21_CR16","DOI":"10.1109\/CVPRW63382.2024.00570"},{"key":"21_CR17","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"89","DOI":"10.1007\/978-3-031-20086-1_6","volume-title":"ECCV 2022","author":"Y Hu","year":"2022","unstructured":"Hu, Y., Fua, P., Salzmann, M.: Perspective flow aggregation for data-limited 6d object pose estimation. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13662, pp. 89\u2013106. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20086-1_6"},{"doi-asserted-by":"crossref","unstructured":"Hu, Y., Hugonot, J., Fua, P., Salzmann, M.: Segmentation-driven 6D object pose estimation. In: CVPR, pp. 3385\u20133394 (2019)","key":"21_CR18","DOI":"10.1109\/CVPR.2019.00350"},{"key":"21_CR19","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"557","DOI":"10.1007\/978-3-030-58536-5_33","volume-title":"Computer Vision \u2013 ECCV 2020","author":"R Jonschkowski","year":"2020","unstructured":"Jonschkowski, R., Stone, A., Barron, J.T., Gordon, A., Konolige, K., Angelova, A.: What matters in unsupervised optical flow. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 557\u2013572. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_33"},{"doi-asserted-by":"crossref","unstructured":"Joshi, B., et al.: Deepurl: deep pose estimation framework for underwater relative localization. In: IROS, pp. 1777\u20131784. IEEE (2020)","key":"21_CR20","DOI":"10.1109\/IROS45743.2020.9341201"},{"issue":"27","key":"21_CR21","doi-asserted-by":"publisher","first-page":"226","DOI":"10.3182\/20120919-3-IT-2046.00039","volume":"45","author":"J Kalwa","year":"2012","unstructured":"Kalwa, J., et al.: The European R &D-project morph: marine robotic systems of self-organizing, logically linked physical nodes. IFAC Proc. Vol. 45(27), 226\u2013231 (2012)","journal-title":"IFAC Proc. Vol."},{"doi-asserted-by":"crossref","unstructured":"Kehl, W., Manhardt, F., Tombari, F., Ilic, S., Navab, N.: SSD-6D: making RGB-based 3D detection and 6D pose estimation great again. In: ICCV, pp. 1521\u20131529 (2017)","key":"21_CR22","DOI":"10.1109\/ICCV.2017.169"},{"key":"21_CR23","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"574","DOI":"10.1007\/978-3-030-58520-4_34","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Y Labb\u00e9","year":"2020","unstructured":"Labb\u00e9, Y., Carpentier, J., Aubry, M., Sivic, J.: CosyPose: consistent multi-view multi-object 6D pose estimation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12362, pp. 574\u2013591. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58520-4_34"},{"doi-asserted-by":"crossref","unstructured":"Lee, S., Kim, J.H.: Semi-supervised scene change detection by distillation from feature-metric alignment. In: WACV, pp. 1226\u20131235 (2024)","key":"21_CR24","DOI":"10.1109\/WACV57701.2024.00126"},{"key":"21_CR25","doi-asserted-by":"publisher","first-page":"657","DOI":"10.1007\/s11263-019-01250-9","volume":"128","author":"Y Li","year":"2020","unstructured":"Li, Y., Wang, G., Ji, X., Xiang, Y., Fox, D.: Deepim: deep iterative matching for 6D pose estimation. IJCV 128, 657\u2013678 (2020)","journal-title":"IJCV"},{"doi-asserted-by":"crossref","unstructured":"Li, Z., Wang, G., Ji, X.: CDPN: coordinates-based disentangled pose network for real-time RGB-based 6-DoF object pose estimation. In: ICCV, pp. 7678\u20137687 (2019)","key":"21_CR26","DOI":"10.1109\/ICCV.2019.00777"},{"key":"21_CR27","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"19","DOI":"10.1007\/978-3-031-20077-9_2","volume-title":"ECCV 2022","author":"J Lin","year":"2022","unstructured":"Lin, J., Wei, Z., Ding, C., Jia, K.: Category-level 6D object pose and size estimation using self-supervised deep prior deformation networks. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13669, pp. 19\u201334. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20077-9_2"},{"doi-asserted-by":"crossref","unstructured":"Lipson, L., Teed, Z., Goyal, A., Deng, J.: Coupled iterative refinement for 6D multi-object pose estimation. In: CVPR, pp. 6728\u20136737 (2022)","key":"21_CR28","DOI":"10.1109\/CVPR52688.2022.00661"},{"unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)","key":"21_CR29"},{"key":"21_CR30","doi-asserted-by":"publisher","first-page":"1204","DOI":"10.1007\/s11036-017-0863-4","volume":"22","author":"H Lu","year":"2017","unstructured":"Lu, H., Li, Y., Zhang, Y., Chen, M., Serikawa, S., Kim, H.: Underwater optical image processing: a comprehensive review. Mob. Netw. Appl. 22, 1204\u20131211 (2017)","journal-title":"Mob. Netw. Appl."},{"unstructured":"Manderson, T., Karp, I., Dudek, G.: Aqua underwater simulator. In: IROS (2018)","key":"21_CR31"},{"doi-asserted-by":"crossref","unstructured":"Meister, S., Hur, J., Roth, S.: Unflow: unsupervised learning of optical flow with a bidirectional census loss. In: AAAI, vol.\u00a032 (2018)","key":"21_CR32","DOI":"10.1609\/aaai.v32i1.12276"},{"key":"21_CR33","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1007\/978-3-642-81897-4_4","volume-title":"Fast Fourier Transform and Convolution Algorithms","author":"HJ Nussbaumer","year":"1982","unstructured":"Nussbaumer, H.J.: The fast Fourier transform. In: Nussbaumer, H.J. (ed.) Fast Fourier Transform and Convolution Algorithms, pp. 80\u2013111. Springer, Heidelberg (1982). https:\/\/doi.org\/10.1007\/978-3-642-81897-4_4"},{"doi-asserted-by":"crossref","unstructured":"Oppenheim, A., Lim, J., Kopec, G., Pohlig, S.: Phase in speech and pictures. In: ICASSP, vol.\u00a04, pp. 632\u2013637. IEEE (1979)","key":"21_CR34","DOI":"10.1109\/ICASSP.1979.1170798"},{"unstructured":"Paszke, A., et al.: Pytorch: an imperative style, high-performance deep learning library. In: NeurIPS, vol. 32 (2019)","key":"21_CR35"},{"key":"21_CR36","first-page":"1","volume":"60","author":"D Peng","year":"2021","unstructured":"Peng, D., Guan, H., Zang, Y., Bruzzone, L.: Full-level domain adaptation for building extraction in very-high-resolution optical remote-sensing images. IEEE TGRS 60, 1\u201317 (2021)","journal-title":"IEEE TGRS"},{"doi-asserted-by":"crossref","unstructured":"Peng, S., Liu, Y., Huang, Q., Zhou, X., Bao, H.: Pvnet: pixel-wise voting network for 6dof pose estimation. In: CVPR, pp. 4561\u20134570 (2019)","key":"21_CR37","DOI":"10.1109\/CVPR.2019.00469"},{"issue":"3","key":"21_CR38","doi-asserted-by":"publisher","first-page":"337","DOI":"10.1068\/p110337","volume":"11","author":"LN Piotrowski","year":"1982","unstructured":"Piotrowski, L.N., Campbell, F.W.: A demonstration of the visual importance and flexibility of spatial-frequency amplitude and phase. Perception 11(3), 337\u2013346 (1982)","journal-title":"Perception"},{"doi-asserted-by":"crossref","unstructured":"Rad, M., Lepetit, V.: BB8: a scalable, accurate, robust to partial occlusion method for predicting the 3D poses of challenging objects without using depth. In: ICCV, pp. 3828\u20133836 (2017)","key":"21_CR39","DOI":"10.1109\/ICCV.2017.413"},{"doi-asserted-by":"crossref","unstructured":"Sapienza, D., et al.: Model-based underwater 6D pose estimation from RGB. IEEE RA-L (2023)","key":"21_CR40","DOI":"10.1109\/LRA.2023.3320028"},{"doi-asserted-by":"publisher","unstructured":"Shotton, J., Glocker, B., Zach, C., Izadi, S., Criminisi, A., Fitzgibbon, A.: Scene coordinate regression forests for camera relocalization in RGB-D images. In: CVPR, pp. 2930\u20132937 (2013). https:\/\/doi.org\/10.1109\/CVPR.2013.377","key":"21_CR41","DOI":"10.1109\/CVPR.2013.377"},{"doi-asserted-by":"crossref","unstructured":"Smith, L.N., Topin, N.: Super-convergence: very fast training of neural networks using large learning rates. In: Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications, vol. 11006, pp. 369\u2013386. SPIE (2019)","key":"21_CR42","DOI":"10.1117\/12.2520589"},{"doi-asserted-by":"crossref","unstructured":"Sundermeyer, M., et al.: BOP challenge 2022 on detection, segmentation and pose estimation of specific rigid objects. In: CVPRW, pp. 2785\u20132794 (2023)","key":"21_CR43","DOI":"10.1109\/CVPRW59228.2023.00279"},{"issue":"1","key":"21_CR44","first-page":"65","volume":"9","author":"J Tang","year":"2023","unstructured":"Tang, J., et al.: ROV6D: 6D pose estimation benchmark dataset for underwater remotely operated vehicles. IEEE RA-L 9(1), 65\u201372 (2023)","journal-title":"IEEE RA-L"},{"unstructured":"Tarvainen, A., Valpola, H.: Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results. In: NeurIPS, vol. 30 (2017)","key":"21_CR45"},{"key":"21_CR46","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"402","DOI":"10.1007\/978-3-030-58536-5_24","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Teed","year":"2020","unstructured":"Teed, Z., Deng, J.: RAFT: recurrent all-pairs field transforms for optical flow. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12347, pp. 402\u2013419. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58536-5_24"},{"doi-asserted-by":"crossref","unstructured":"Tekin, B., Sinha, S.N., Fua, P.: Real-time seamless single shot 6D object pose prediction. In: CVPR, pp. 292\u2013301 (2018)","key":"21_CR47","DOI":"10.1109\/CVPR.2018.00038"},{"issue":"3","key":"21_CR48","doi-asserted-by":"publisher","first-page":"1788","DOI":"10.1109\/TPAMI.2021.3136301","volume":"46","author":"G Wang","year":"2021","unstructured":"Wang, G., Manhardt, F., Liu, X., Ji, X., Tombari, F.: Occlusion-aware self-supervised monocular 6D object pose estimation. IEEE TPAMI 46(3), 1788\u20131803 (2021)","journal-title":"IEEE TPAMI"},{"key":"21_CR49","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1007\/978-3-030-58452-8_7","volume-title":"Computer Vision \u2013 ECCV 2020","author":"G Wang","year":"2020","unstructured":"Wang, G., Manhardt, F., Shao, J., Ji, X., Navab, N., Tombari, F.: Self6D: self-supervised monocular 6D object pose estimation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12346, pp. 108\u2013125. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58452-8_7"},{"doi-asserted-by":"crossref","unstructured":"Wang, G., Manhardt, F., Tombari, F., Ji, X.: GDR-Net: geometry-guided direct regression network for monocular 6D object pose estimation. In: CVPR, pp. 16611\u201316621 (2021)","key":"21_CR50","DOI":"10.1109\/CVPR46437.2021.01634"},{"issue":"31","key":"21_CR51","doi-asserted-by":"publisher","first-page":"395","DOI":"10.1016\/j.ifacol.2022.10.460","volume":"55","author":"M Xanthidis","year":"2022","unstructured":"Xanthidis, M., Joshi, B., O\u2019Kane, J.M., Rekleitis, I.: Multi-robot exploration of underwater structures. IFAC-PapersOnLine 55(31), 395\u2013400 (2022)","journal-title":"IFAC-PapersOnLine"},{"doi-asserted-by":"crossref","unstructured":"Xiang, Y., Schmidt, T., Narayanan, V., Fox, D.: Posecnn: a convolutional neural network for 6D object pose estimation in cluttered scenes. RSS (2018)","key":"21_CR52","DOI":"10.15607\/RSS.2018.XIV.019"},{"key":"21_CR53","series-title":"LNCS","doi-asserted-by":"publisher","first-page":"398","DOI":"10.1007\/978-3-031-20068-7_23","volume-title":"ECCV 2022","author":"L Xu","year":"2022","unstructured":"Xu, L., et al.: Pose for everything: towards category-agnostic pose estimation. In: Avidan, S., Brostow, G., Ciss\u00e9, M., Farinella, G.M., Hassner, T. (eds.) ECCV 2022. LNCS, vol. 13666, pp. 398\u2013416. Springer, Cham (2022). https:\/\/doi.org\/10.1007\/978-3-031-20068-7_23"},{"doi-asserted-by":"crossref","unstructured":"Yang, Y., Lao, D., Sundaramoorthi, G., Soatto, S.: Phase consistent ecological domain adaptation. In: CVPR (2020)","key":"21_CR54","DOI":"10.1109\/CVPR42600.2020.00903"},{"doi-asserted-by":"crossref","unstructured":"Yang, Y., Soatto, S.: FDA: fourier domain adaptation for semantic segmentation. In: CVPR, pp. 4085\u20134095 (2020)","key":"21_CR55","DOI":"10.1109\/CVPR42600.2020.00414"},{"doi-asserted-by":"crossref","unstructured":"Yi, L., Gong, B., Funkhouser, T.: Complete & label: a domain adaptation approach to semantic segmentation of lidar point clouds. In: CVPR, pp. 15363\u201315373 (2021)","key":"21_CR56","DOI":"10.1109\/CVPR46437.2021.01511"},{"doi-asserted-by":"crossref","unstructured":"Zhang, Y., et al.: Self-supervised exclusive learning for 3D segmentation with cross-modal unsupervised domain adaptation. In: ACM MM, pp. 3338\u20133346 (2022)","key":"21_CR57","DOI":"10.1145\/3503161.3547987"},{"doi-asserted-by":"crossref","unstructured":"Zhang, Y., Qiu, Z., Yao, T., Liu, D., Mei, T.: Fully convolutional adaptation networks for semantic segmentation. In: CVPR, pp. 6810\u20136818 (2018)","key":"21_CR58","DOI":"10.1109\/CVPR.2018.00712"},{"doi-asserted-by":"crossref","unstructured":"Zheng, L., Ma, W., Cai, Y., Lu, T., Wang, S.: Gpdan: grasp pose domain adaptation network for sim-to-real 6-DoF object grasping. IEEE RA-L (2023)","key":"21_CR59","DOI":"10.1109\/LRA.2023.3286816"},{"doi-asserted-by":"crossref","unstructured":"Zhou, Y., Barnes, C., Lu, J., Yang, J., Li, H.: On the continuity of rotation representations in neural networks. In: CVPR, pp. 5745\u20135753 (2019)","key":"21_CR60","DOI":"10.1109\/CVPR.2019.00589"},{"doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: ICCV, pp. 2223\u20132232 (2017)","key":"21_CR61","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-73021-4_21","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,11,20]],"date-time":"2024-11-20T09:47:30Z","timestamp":1732096050000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-73021-4_21"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,11,21]]},"ISBN":["9783031730207","9783031730214"],"references-count":61,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-73021-4_21","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2024,11,21]]},"assertion":[{"value":"21 November 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}