{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,11]],"date-time":"2026-02-11T18:15:19Z","timestamp":1770833719872,"version":"3.50.1"},"publisher-location":"Cham","reference-count":25,"publisher":"Springer International Publishing","isbn-type":[{"value":"9783031063800","type":"print"},{"value":"9783031063817","type":"electronic"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-06381-7_1","type":"book-chapter","created":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T08:03:16Z","timestamp":1652688196000},"page":"3-17","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["6D Pose Estimation of\u00a0Transparent Objects Using Synthetic Data"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-3560-9583","authenticated-orcid":false,"given":"Munkhtulga","family":"Byambaa","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-3414-1085","authenticated-orcid":false,"given":"Gou","family":"Koutaki","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0002-1773-1059","authenticated-orcid":false,"given":"Lodoiravsal","family":"Choimaa","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,5,17]]},"reference":[{"key":"1_CR1","doi-asserted-by":"crossref","unstructured":"Chen, Y., Sun, G., Lin, H., Chen, S.: Random bin picking with multi-view image acquisition and CAD-based pose estimation. In: 2018 IEEE International Conference on Systems, pp. 2218\u20132223 (2018)","DOI":"10.1109\/SMC.2018.00381"},{"key":"1_CR2","doi-asserted-by":"crossref","unstructured":"He, R., Rojas, J., Guan, Y.: A 3D object detection and pose estimation pipeline using RGB-D images. In: 2017 IEEE International Conference on Robotics and Biomimetics (ROBIO), pp. 1527\u20131532 (2017)","DOI":"10.1109\/ROBIO.2017.8324634"},{"key":"1_CR3","doi-asserted-by":"crossref","unstructured":"Wu, C., Jiang, S., Song, K.: CAD-based pose estimation for random bin-picking of multiple objects using a RGB-D camera. In: 2015 15th International Conference on Control Automation and Systems (ICCAS), pp. 1645\u20131649 (2015)","DOI":"10.1109\/ICCAS.2015.7364621"},{"key":"1_CR4","doi-asserted-by":"crossref","unstructured":"Xiang, Y., Schmidt, T., Narayanan, V., Fox, D.: PoseCNN: A Convolutional Neural Network for 6D Object Pose Estimation in Cluttered Scenes. c (RSS) (2018)","DOI":"10.15607\/RSS.2018.XIV.019"},{"key":"1_CR5","doi-asserted-by":"crossref","unstructured":"Wang, C., et al.: DenseFusion: 6D object pose estimation by iterative dense fusion. In: Computer Vision and Pattern Recognition (CVPR), pp. 3343\u20133352 (2019)","DOI":"10.1109\/CVPR.2019.00346"},{"key":"1_CR6","unstructured":"Tremblay, J., To, T., Sundaralingam, B., Xiang, Y., Fox, D., Birchfield, S.: Deep object pose estimation for semantic robotic grasping of household objects. In: Conference on Robot Learning (CoRL), pp. 306\u2013316 (2018)"},{"key":"1_CR7","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. In: International Conference on Learning Representations (2015)"},{"key":"1_CR8","doi-asserted-by":"crossref","unstructured":"Guo-Hua, C., Jun-Yi, W., Ai-Jun, Z.: Transparent object detection and location based on RGB-D camera. J. Phys. Conf. Ser. (JPCS) 1183, 012011 (2019)","DOI":"10.1088\/1742-6596\/1183\/1\/012011"},{"key":"1_CR9","doi-asserted-by":"crossref","unstructured":"Dosovitskiy, A., et al.: FlowNet: learning optical flow with convolutional networks. In: 2015 IEEE International Conference on Computer Vision (ICCV), pp. 2758\u20132766 (2015)","DOI":"10.1109\/ICCV.2015.316"},{"key":"1_CR10","doi-asserted-by":"crossref","unstructured":"Tremblay, J., To, T., Birchfield, S.: Falling things: a synthetic dataset for 3D object detection and pose estimation. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 2119\u201321193 (2018)","DOI":"10.1109\/CVPRW.2018.00275"},{"key":"1_CR11","doi-asserted-by":"publisher","first-page":"902","DOI":"10.1007\/s11263-018-1073-7","volume":"126","author":"M M\u00fcller","year":"2018","unstructured":"M\u00fcller, M., Casser, V., Lahoud, J., Smith, N., Ghanem, B.: Sim4CV: a photo-realistic simulator for computer vision applications. Int. J. Comput. Vis. 126, 902\u2013919 (2018)","journal-title":"Int. J. Comput. Vis."},{"key":"1_CR12","unstructured":"McCormac, J., Handa, A., Leutenegger, S., Davison, A.J.: SceneNet RGB-D: 5M photorealistic images of synthetic indoor trajectories with ground truth. In: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 2678\u20132687 (2016)"},{"key":"1_CR13","doi-asserted-by":"crossref","unstructured":"Ros, G., Sellart, L., Materzynska, J., V\u00e1zquez, D., L\u00f3pez, A.: The SYNTHIA dataset: a large collection of synthetic images for semantic segmentation of urban scenes. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 3234\u20133243 (2016)","DOI":"10.1109\/CVPR.2016.352"},{"key":"1_CR14","doi-asserted-by":"crossref","unstructured":"Tobin, J., Fong, R., Ray, A., Schneider, J., Zaremba, W., Abbeel, P.: Domain randomization for transferring deep neural networks from simulation to the real world. In: 2017 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 23\u201330 (2017)","DOI":"10.1109\/IROS.2017.8202133"},{"key":"1_CR15","doi-asserted-by":"crossref","unstructured":"Sadeghi, F., Levine, S.: CAD2RL: Real Single-Image Flight without a Single Real Image. CoRR (2016)","DOI":"10.15607\/RSS.2017.XIII.034"},{"key":"1_CR16","doi-asserted-by":"crossref","unstructured":"Tremblay, J., et al.: Training deep networks with synthetic data: bridging the reality gap by domain randomization. In: 2018 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1082\u201310828 (2018)","DOI":"10.1109\/CVPRW.2018.00143"},{"key":"1_CR17","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-10602-1_48","volume-title":"Computer Vision \u2013 ECCV 2014","author":"T-Y Lin","year":"2014","unstructured":"Lin, T.-Y., et al.: Microsoft COCO: common objects in context. In: Fleet, D., Pajdla, T., Schiele, B., Tuytelaars, T. (eds.) ECCV 2014. LNCS, vol. 8693, pp. 740\u2013755. Springer, Cham (2014). https:\/\/doi.org\/10.1007\/978-3-319-10602-1_48"},{"key":"1_CR18","doi-asserted-by":"crossref","unstructured":"Johnson-Roberson, M., Barto, C., Mehta, R., Sridhar, S., Rosaen, K., Vasudevan, R.: Driving in the matrix: can virtual worlds replace human-generated annotations for real world tasks? In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 746\u2013753 (2017)","DOI":"10.1109\/ICRA.2017.7989092"},{"key":"1_CR19","doi-asserted-by":"crossref","unstructured":"Mayer, N., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4040\u20134048 (2016)","DOI":"10.1109\/CVPR.2016.438"},{"key":"1_CR20","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"682","DOI":"10.1007\/978-3-030-11009-3_42","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"S Hinterstoisser","year":"2019","unstructured":"Hinterstoisser, S., Lepetit, V., Wohlhart, P., Konolige, K.: On pre-trained image features and synthetic images for deep learning. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11129, pp. 682\u2013697. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11009-3_42"},{"key":"1_CR21","doi-asserted-by":"crossref","unstructured":"Lysenkov, I., Rabaud, V.: Pose estimation of rigid transparent objects in transparent clutter. In: 2013 IEEE International Conference on Robotics and Automation, pp. 162\u2013169 (2013)","DOI":"10.1109\/ICRA.2013.6630571"},{"key":"1_CR22","doi-asserted-by":"crossref","unstructured":"Moreno-Noguer, F., Lepetit, V., Fua, P.: Accurate non-iterative O(n) solution to the PnP problem. In: 2007 IEEE 11th International Conference on Computer Vision, pp. 1\u20138 (2007)","DOI":"10.1109\/ICCV.2007.4409116"},{"key":"1_CR23","doi-asserted-by":"crossref","unstructured":"Sajjan, S., et al.: Clear grasp: 3D shape estimation of transparent objects for manipulation. In: 2020 IEEE International Conference on Robotics and Automation (ICRA), pp. 3634\u20133642 (2020)","DOI":"10.1109\/ICRA40945.2020.9197518"},{"key":"1_CR24","doi-asserted-by":"crossref","unstructured":"Liu, X., Jonschkowski, R., Angelova, A., Konolige, K.: KeyPose: multi-view 3D labeling and keypoint estimation for transparent objects. In: 2020 IEEE\/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 11599\u201311607 (2020)","DOI":"10.1109\/CVPR42600.2020.01162"},{"key":"1_CR25","unstructured":"Byambaa, M., Choimaa, L., Koutaki, G.: 6D pose estimation of transparent object from single RGB image. In: The 25th Conference of FRUCT Association, pp. 444\u2013447 (2019)"}],"container-title":["Communications in Computer and Information Science","Frontiers of Computer Vision"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-06381-7_1","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,5,16]],"date-time":"2022-05-16T08:14:25Z","timestamp":1652688865000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-06381-7_1"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031063800","9783031063817"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-06381-7_1","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"value":"1865-0929","type":"print"},{"value":"1865-0937","type":"electronic"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"17 May 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"IW-FCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Workshop on Frontiers of Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21 February 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"22 February 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"28","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iwfcv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/sites.google.com\/view\/iwfcv2022","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"63","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"24","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"38% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}