{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T20:23:51Z","timestamp":1743107031567,"version":"3.40.3"},"publisher-location":"Singapore","reference-count":23,"publisher":"Springer Nature Singapore","isbn-type":[{"type":"print","value":"9789819981403"},{"type":"electronic","value":"9789819981410"}],"license":[{"start":{"date-parts":[[2023,11,26]],"date-time":"2023-11-26T00:00:00Z","timestamp":1700956800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,11,26]],"date-time":"2023-11-26T00:00:00Z","timestamp":1700956800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-981-99-8141-0_43","type":"book-chapter","created":{"date-parts":[[2023,11,25]],"date-time":"2023-11-25T09:02:16Z","timestamp":1700902936000},"page":"579-593","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Real-Time Instance Segmentation and\u00a0Tip Detection for\u00a0Neuroendoscopic Surgical Instruments"],"prefix":"10.1007","author":[{"given":"Rihui","family":"Song","sequence":"first","affiliation":[]},{"given":"Silu","family":"Guo","sequence":"additional","affiliation":[]},{"given":"Ni","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yehua","family":"Ling","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Gong","sequence":"additional","affiliation":[]},{"given":"Kai","family":"Huang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,11,26]]},"reference":[{"key":"43_CR1","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"90","DOI":"10.1007\/978-3-319-29965-5_9","volume-title":"Computer-Assisted and Robotic Endoscopy","author":"A Agustinos","year":"2016","unstructured":"Agustinos, A., Voros, S.: 2D\/3D real-time tracking of surgical instruments based on endoscopic image processing. In: Luo, X., Reichl, T., Reiter, A., Mariottini, G.-L. (eds.) CARE 2015. LNCS, vol. 9515, pp. 90\u2013100. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-29965-5_9"},{"key":"43_CR2","doi-asserted-by":"publisher","unstructured":"Bolya, D., et al.: YOLACT: real-time instance segmentation. In: ICCV, pp. 9157\u20139166 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00925","DOI":"10.1109\/ICCV.2019.00925"},{"issue":"12","key":"43_CR3","first-page":"2603","volume":"34","author":"D Bouget","year":"2015","unstructured":"Bouget, D., Benenson, R., Omran, M., et al.: Detecting surgical tools by modelling local appearance and global shape. IEEE TMI 34(12), 2603\u20132617 (2015)","journal-title":"IEEE TMI"},{"issue":"1","key":"43_CR4","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1023\/A:1007379606734","volume":"28","author":"R Caruana","year":"1997","unstructured":"Caruana, R.: Multitask learning. Mach. Learn. 28(1), 41\u201375 (1997)","journal-title":"Mach. Learn."},{"key":"43_CR5","unstructured":"Chen, L.C., et al.: Rethinking Atrous convolution for semantic image segmentation. arXiv preprint: arXiv:1706.05587 (2017)"},{"key":"43_CR6","doi-asserted-by":"publisher","unstructured":"Dai, J., et al.: Deformable convolutional networks. In: ICCV, pp. 764\u2013773 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.89","DOI":"10.1109\/ICCV.2017.89"},{"key":"43_CR7","doi-asserted-by":"publisher","first-page":"183","DOI":"10.1016\/j.disopt.2015.03.002","volume":"22","author":"M Fischetti","year":"2016","unstructured":"Fischetti, M.: Fast training of support vector machines with gaussian kernel. Discret. Optim. 22, 183\u2013194 (2016)","journal-title":"Discret. Optim."},{"key":"43_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"84","DOI":"10.1007\/978-3-319-54057-3_8","volume-title":"Computer-Assisted and Robotic Endoscopy","author":"LC Garc\u00eda-Peraza-Herrera","year":"2017","unstructured":"Garc\u00eda-Peraza-Herrera, L.C., et al.: Real-time segmentation of non-rigid surgical tools based on deep learning and tracking. In: Peters, T., et al. (eds.) CARE 2016. LNCS, vol. 10170, pp. 84\u201395. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-54057-3_8"},{"key":"43_CR9","doi-asserted-by":"publisher","first-page":"832208","DOI":"10.3389\/frobt.2022.832208","volume":"9","author":"C Gruijthuijsen","year":"2022","unstructured":"Gruijthuijsen, C., Garcia-Peraza-Herrera, L.C., Borghesan, G., et al.: Robotic endoscope control via autonomous instrument tracking. Front. Robot. AI 9, 832208 (2022). https:\/\/doi.org\/10.3389\/frobt.2022.832208","journal-title":"Front. Robot. AI"},{"key":"43_CR10","doi-asserted-by":"publisher","unstructured":"He, K., et al.: Mask R-CNN. In: ICCV, pp. 2961\u20132969 (2017). https:\/\/doi.org\/10.1109\/ICCV.2017.322","DOI":"10.1109\/ICCV.2017.322"},{"key":"43_CR11","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"505","DOI":"10.1007\/978-3-319-66185-8_57","volume-title":"Medical Image Computing and Computer-Assisted Intervention-MICCAI 2017","author":"T Kurmann","year":"2017","unstructured":"Kurmann, T., et al.: Simultaneous recognition and pose estimation of instruments in minimally invasive surgery. In: Descoteaux, M., Maier-Hein, L., Franz, A., Jannin, P., Collins, D.L., Duchesne, S. (eds.) MICCAI 2017. LNCS, vol. 10434, pp. 505\u2013513. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-66185-8_57"},{"key":"43_CR12","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"664","DOI":"10.1007\/978-3-319-66185-8_75","volume-title":"Medical Image Computing and Computer-Assisted Intervention-MICCAI 2017","author":"I Laina","year":"2017","unstructured":"Laina, I., et al.: Concurrent segmentation and localization for tracking of surgical instruments. In: Descoteaux, M., Maier-Hein, L., Franz, A., Jannin, P., Collins, D.L., Duchesne, S. (eds.) MICCAI 2017. LNCS, vol. 10434, pp. 664\u2013672. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-66185-8_75"},{"key":"43_CR13","doi-asserted-by":"publisher","unstructured":"Lin, T.Y., et al.: Focal loss for dense object detection. In: TPAMI, pp. 2980\u20132988 (2017). https:\/\/doi.org\/10.1109\/TPAMI.2018.2858826","DOI":"10.1109\/TPAMI.2018.2858826"},{"key":"43_CR14","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"667","DOI":"10.1007\/978-3-030-00937-3_76","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2018","author":"F Milletari","year":"2018","unstructured":"Milletari, F., Rieke, N., Baust, M., Esposito, M., Navab, N.: CFCM: segmentation via coarse to fine context memory. In: Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-L\u00f3pez, C., Fichtinger, G. (eds.) MICCAI 2018. LNCS, vol. 11073, pp. 667\u2013674. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-00937-3_76"},{"issue":"1","key":"43_CR15","doi-asserted-by":"publisher","first-page":"4","DOI":"10.1002\/rcs.1638","volume":"12","author":"M Niccolini","year":"2016","unstructured":"Niccolini, M., Castelli, V., Diversi, C., et al.: Development and preliminary assessment of a robotic platform for neuroendoscopy based on a lightweight robot. IJMRCAS 12(1), 4\u201317 (2016). https:\/\/doi.org\/10.1002\/rcs.1638","journal-title":"IJMRCAS"},{"key":"43_CR16","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"566","DOI":"10.1007\/978-3-030-32692-0_65","volume-title":"Machine Learning in Medical Imaging","author":"D Pakhomov","year":"2019","unstructured":"Pakhomov, D., Premachandran, V., Allan, M., Azizian, M., Navab, N.: Deep residual learning for instrument segmentation in robotic surgery. In: Suk, H.-I., Liu, M., Yan, P., Lian, C. (eds.) MLMI 2019. LNCS, vol. 11861, pp. 566\u2013573. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-32692-0_65"},{"key":"43_CR17","doi-asserted-by":"crossref","unstructured":"Perez, S.A.H., et al.: Segmentation of endonasal robotic instruments in a head phantom using deep learning and virtual-reality simulation. In: ROBOMEC 2020, pp. 2P2-F01. The Japan Society of Mechanical Engineers (2020)","DOI":"10.1299\/jsmermd.2020.2P2-F01"},{"key":"43_CR18","unstructured":"Redmon, J., Farhadi, A.: Yolov3: an incremental improvement. arXiv preprint: arXiv:1804.02767 (2018)"},{"issue":"6","key":"43_CR19","doi-asserted-by":"publisher","first-page":"1137","DOI":"10.1109\/TPAMI.2016.2577031","volume":"39","author":"S Ren","year":"2016","unstructured":"Ren, S., He, K., Girshick, R., et al.: Faster R-CNN: towards real-time object detection with region proposal networks. TPAMI 39(6), 1137\u20131149 (2016). https:\/\/doi.org\/10.1109\/TPAMI.2016.2577031","journal-title":"TPAMI"},{"issue":"4","key":"43_CR20","doi-asserted-by":"publisher","first-page":"640","DOI":"10.1109\/TPAMI.2016.2572683","volume":"39","author":"E Shelhamer","year":"2017","unstructured":"Shelhamer, E., Long, J., Darrell, T.: Fully convolutional networks for semantic segmentation. TPAMI 39(4), 640\u2013651 (2017). https:\/\/doi.org\/10.1109\/TPAMI.2016.2572683","journal-title":"TPAMI"},{"issue":"1","key":"43_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1186\/s40537-019-0197-0","volume":"6","author":"C Shorten","year":"2019","unstructured":"Shorten, C., Khoshgoftaar, T.M.: A survey on image data augmentation for deep learning. J. Big Data 6(1), 1\u201348 (2019)","journal-title":"J. Big Data"},{"issue":"7","key":"43_CR22","first-page":"732","volume":"17","author":"L Zappella","year":"2013","unstructured":"Zappella, L., B\u00e9jar, B., Hager, G., et al.: Surgical gesture classification from video and kinematic data. MIA 17(7), 732\u2013745 (2013)","journal-title":"MIA"},{"issue":"6","key":"43_CR23","doi-asserted-by":"publisher","first-page":"2665","DOI":"10.3233\/BME-141083","volume":"24","author":"Z Zhao","year":"2014","unstructured":"Zhao, Z.: Real-time 3D visual tracking of laparoscopic instruments for robotized endoscope holder. Bio-Med. Mater. Eng. 24(6), 2665\u20132672 (2014)","journal-title":"Bio-Med. Mater. Eng."}],"container-title":["Communications in Computer and Information Science","Neural Information Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-981-99-8141-0_43","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T15:44:27Z","timestamp":1710344667000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-981-99-8141-0_43"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023,11,26]]},"ISBN":["9789819981403","9789819981410"],"references-count":23,"URL":"https:\/\/doi.org\/10.1007\/978-981-99-8141-0_43","relation":{},"ISSN":["1865-0929","1865-0937"],"issn-type":[{"type":"print","value":"1865-0929"},{"type":"electronic","value":"1865-0937"}],"subject":[],"published":{"date-parts":[[2023,11,26]]},"assertion":[{"value":"26 November 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICONIP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Neural Information Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Changsha","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"China","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2023","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"20 November 2023","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2023","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iconip2023","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/iconip2023.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1274","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"650","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"51% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4.14","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.46","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}