{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,10,1]],"date-time":"2025-10-01T16:15:13Z","timestamp":1759335313969,"version":"3.40.3"},"publisher-location":"Cham","reference-count":17,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031479656"},{"type":"electronic","value":"9783031479663"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-47966-3_30","type":"book-chapter","created":{"date-parts":[[2023,12,2]],"date-time":"2023-12-02T09:02:11Z","timestamp":1701507731000},"page":"385-396","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Generation Method of\u00a0Robot Assembly Motion Considering Physicality Gap Between Humans and\u00a0Robots"],"prefix":"10.1007","author":[{"given":"Takahiro","family":"Suzuki","sequence":"first","affiliation":[]},{"given":"Manabu","family":"Hashimoto","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,12,3]]},"reference":[{"key":"30_CR1","doi-asserted-by":"crossref","unstructured":"Domae, Y., Okuda, H., Taguchi, Y., Sumi, K., Hirai, T.: Fast graspability evaluation on single depth maps for bin picking with general gripper. In: Proceedings of the IEEE Conference on ICRA, HongKong, pp. 1997\u20132004 (2014)","DOI":"10.1109\/ICRA.2014.6907124"},{"issue":"8","key":"30_CR2","doi-asserted-by":"publisher","first-page":"559","DOI":"10.7210\/jrsj.36.559","volume":"36","author":"R Araki","year":"2018","unstructured":"Araki, R., et al.: Grasping detection using deep convolutional neural network with graspability. J. Rob. Soc. Japan 36(8), 559\u2013566 (2018)","journal-title":"J. Rob. Soc. Japan"},{"key":"30_CR3","doi-asserted-by":"crossref","unstructured":"Zhang, X., Koyama, K., Domae, Y., Wan, W., Harada, K.: A topological solution of entanglement for complex-shaped parts in robotic bin-picking. In: Proceedings of the IEEE Conference on CASE, pp. 461\u2013467 (2021)","DOI":"10.1109\/CASE49439.2021.9551426"},{"issue":"3","key":"30_CR4","doi-asserted-by":"publisher","first-page":"4978","DOI":"10.1109\/LRA.2020.3004787","volume":"5","author":"S Song","year":"2020","unstructured":"Song, S., Zeng, A., Lee, J., Funkhouser, T.: Grasping in the wild: learning 6dof closed-loop grasping from low-cost demonstrations. IEEE Rob. Autom. Lett. 5(3), 4978\u20134985 (2020)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"30_CR5","doi-asserted-by":"crossref","unstructured":"Chen, X., Ghadirzadeh, A., Bhorkman, M., Jensfelt, P.: Adversarial feature training for generalizable robotic visuomotor control. In: Proceedings of the IEEE Conference on ICRA, pp. 1142\u20131148 (2020)","DOI":"10.1109\/ICRA40945.2020.9197505"},{"key":"30_CR6","unstructured":"Kevin, Z., Andy, Z., Johnny, L., Shuran, S.: Form2Fit: learning shape priors for generalizable assembly from disassembly. In: Proceedings of the IEEE Conference on ICRA, pp. 9404\u20139410 (2020)"},{"key":"30_CR7","doi-asserted-by":"crossref","unstructured":"Turpin, D., Wang, L., Tshogkas, S., Dickinson, S., Garg, A.: GIFT: generalizable interaction-aware functional tool affordances without labels. In Robotics: Science and Systems (2021). https:\/\/arxiv.org\/abs\/2106.14973","DOI":"10.15607\/RSS.2021.XVII.060"},{"key":"30_CR8","doi-asserted-by":"crossref","unstructured":"Qin, Z., Fang, K., Zhu, Y., Fei-Fei, L., Savarese, S.: KETO:learning keypoint representations for tool manipulation. In: Proceedings of the IEEE Conference on ICRA, pp. 7278\u20137285 (2020)","DOI":"10.1109\/ICRA40945.2020.9196971"},{"key":"30_CR9","doi-asserted-by":"crossref","unstructured":"Ardon, P., Pairet, E., Petillot, Y., Petrick, P.A.R., Ramamoorthy, S., Lohan, S.K.: Self-assessment of grasp affordance transfer. In: Proceedings of the IEEE\/RSJ Conference on IROS, pp. 9385\u20139392 (2020)","DOI":"10.1109\/IROS45743.2020.9340841"},{"key":"30_CR10","doi-asserted-by":"crossref","unstructured":"Liang, J., Boularias, A.: Learning category-level manipulation tasks from point clouds with dynamic graph CNNs. In: Proceedings of the IEEE Conference on ICRA, UK, pp. 1807\u20131813 (2023)","DOI":"10.1109\/ICRA48891.2023.10160820"},{"key":"30_CR11","doi-asserted-by":"publisher","first-page":"154","DOI":"10.1007\/978-3-030-90436-4_12","volume":"13018","author":"T Suzuki","year":"2021","unstructured":"Suzuki, T., Hashimoto, M.: A method for transferring robot motion parameters using functional attributes of parts. Lect. Notes Comput. Sci. 13018, 154\u2013165 (2021)","journal-title":"Lect. Notes Comput. Sci."},{"issue":"2","key":"30_CR12","doi-asserted-by":"publisher","first-page":"1140","DOI":"10.1109\/LRA.2019.2894439","volume":"4","author":"F-J Chu","year":"2019","unstructured":"Chu, F.-J., Xu, R., Vela, P.A.: Learning affordance segmentation for real-world robotic manipulation via synthetic images. IEEE Rob. Autom. Lett. 4(2), 1140\u20131447 (2019)","journal-title":"IEEE Rob. Autom. Lett."},{"issue":"4","key":"30_CR13","doi-asserted-by":"publisher","first-page":"4070","DOI":"10.1109\/LRA.2019.2930364","volume":"4","author":"F-J Chu","year":"2019","unstructured":"Chu, F.-J., Xu, R., Vela, P.A.: Toward affordance detection and ranking on novel objects for real-world robotic manipulation. IEEE Rob. Autom. Lett. 4(4), 4070\u20134077 (2019)","journal-title":"IEEE Rob. Autom. Lett."},{"key":"30_CR14","unstructured":"Minh, C., Gilani, S., Islam, S., Suter, D.: Learning affordance segmentation: an investigative study. In: Proceedings of the DICTA, pp. 2870\u20132877 (2020)"},{"key":"30_CR15","doi-asserted-by":"crossref","unstructured":"Luo, H., Zhai, W., Zhang, J., Cao, Y., Tao, D.: Learning visual affordance grounding from demonstration videos (2021). https:\/\/arxiv.org\/pdf\/2108.05675v1.pdf","DOI":"10.1109\/TNNLS.2023.3298638"},{"key":"30_CR16","doi-asserted-by":"crossref","unstructured":"Suzuki, T., Hashimoto, M.: Estimation of robot motion parameters based on functional consistency for randomly stacked parts. In: Proceedings of the VISAPP, Portugal, pp. 519\u2013528 (2023)","DOI":"10.5220\/0011683500003417"},{"key":"30_CR17","doi-asserted-by":"crossref","unstructured":"He, K., Gkioxari, G., Dollar, P., Girshick, R.: Mask R-RCNN. In: Proceedings of the IEEE Conference on ICCV, Italy, pp. 2961\u20132969 (2017)","DOI":"10.1109\/ICCV.2017.322"}],"container-title":["Lecture Notes in Computer Science","Advances in Visual Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-47966-3_30","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,12,2]],"date-time":"2023-12-02T09:06:11Z","timestamp":1701507971000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-47966-3_30"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031479656","9783031479663"],"references-count":17,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-47966-3_30","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"3 December 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ISVC","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Symposium on Visual Computing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Lake Tahoe, NV","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"USA","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"16 October 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18 October 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"isvc2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.isvc.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"25","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"58","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"232% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.3","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"43 (oral), 15 (poster),  25 (special tracks) out of 34 submissions","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}