{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,1,13]],"date-time":"2026-01-13T01:48:22Z","timestamp":1768268902757,"version":"3.49.0"},"publisher-location":"Cham","reference-count":30,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031564345","type":"print"},{"value":"9783031564352","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,1,1]],"date-time":"2024-01-01T00:00:00Z","timestamp":1704067200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2024]]},"DOI":"10.1007\/978-3-031-56435-2_7","type":"book-chapter","created":{"date-parts":[[2024,3,19]],"date-time":"2024-03-19T06:51:26Z","timestamp":1710831086000},"page":"82-95","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["Event Recognition in\u00a0Laparoscopic Gynecology Videos with\u00a0Hybrid Transformers"],"prefix":"10.1007","author":[{"given":"Sahar","family":"Nasirihaghighi","sequence":"first","affiliation":[]},{"given":"Negin","family":"Ghamsarian","sequence":"additional","affiliation":[]},{"given":"Heinrich","family":"Husslein","sequence":"additional","affiliation":[]},{"given":"Klaus","family":"Schoeffmann","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,20]]},"reference":[{"key":"7_CR1","doi-asserted-by":"publisher","first-page":"115006","DOI":"10.1109\/ACCESS.2021.3105454","volume":"9","author":"N Aldahoul","year":"2021","unstructured":"Aldahoul, N., Karim, H.A., Tan, M.J.T., Fermin, J.L.: Transfer learning and decision fusion for real time distortion classification in laparoscopic videos. IEEE Access 9, 115006\u2013115018 (2021)","journal-title":"IEEE Access"},{"key":"7_CR2","doi-asserted-by":"crossref","unstructured":"Bello, I., Zoph, B., Vaswani, A., Shlens, J., Le, Q.V.: Attention augmented convolutional networks. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 3286\u20133295 (2019)","DOI":"10.1109\/ICCV.2019.00338"},{"key":"7_CR3","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"343","DOI":"10.1007\/978-3-030-59716-0_33","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2020","author":"T Czempiel","year":"2020","unstructured":"Czempiel, T., et al.: TeCNO: surgical phase recognition with multi-stage temporal convolutional networks. In: Martel, A.L., et al. (eds.) MICCAI 2020. LNCS, vol. 12263, pp. 343\u2013352. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-59716-0_33"},{"key":"7_CR4","unstructured":"Dosovitskiy, A., et al.: An image is worth 16$$\\times $$16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"7_CR5","doi-asserted-by":"publisher","first-page":"1217","DOI":"10.1007\/s11548-019-01995-1","volume":"14","author":"I Funke","year":"2019","unstructured":"Funke, I., Mees, S.T., Weitz, J., Speidel, S.: Video-based surgical skill assessment using 3D convolutional neural networks. Int. J. Comput. Assist. Radiol. Surg. 14, 1217\u20131225 (2019)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"7_CR6","doi-asserted-by":"crossref","unstructured":"Ghamsarian, N.: Enabling relevance-based exploration of cataract videos. In: Proceedings of the 2020 International Conference on Multimedia Retrieval, pp. 378\u2013382 (2020)","DOI":"10.1145\/3372278.3391937"},{"key":"7_CR7","doi-asserted-by":"crossref","unstructured":"Ghamsarian, N., Amirpourazarian, H., Timmerer, C., Taschwer, M., Sch\u00f6ffmann, K.: Relevance-based compression of cataract surgery videos using convolutional neural networks. In: Proceedings of the 28th ACM International Conference on Multimedia, pp. 3577\u20133585 (2020)","DOI":"10.1145\/3394171.3413658"},{"key":"7_CR8","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1007\/978-3-030-87237-3_8","volume-title":"Medical Image Computing and Computer Assisted Intervention \u2013 MICCAI 2021","author":"N Ghamsarian","year":"2021","unstructured":"Ghamsarian, N., Taschwer, M., Putzgruber-Adamitsch, D., Sarny, S., El-Shabrawi, Y., Schoeffmann, K.: LensID: A CNN-RNN-based framework towards lens irregularity detection in cataract surgery videos. In: de Bruijne, M., et al. (eds.) MICCAI 2021. LNCS, vol. 12908, pp. 76\u201386. Springer, Cham (2021). https:\/\/doi.org\/10.1007\/978-3-030-87237-3_8"},{"key":"7_CR9","doi-asserted-by":"crossref","unstructured":"Ghamsarian, N., Taschwer, M., Putzgruber-Adamitsch, D., Sarny, S., Schoeffmann, K.: Relevance detection in cataract surgery videos by spatio-temporal action localization. In: 2020 25th International Conference on Pattern Recognition (ICPR), pp. 10720\u201310727. IEEE (2021)","DOI":"10.1109\/ICPR48806.2021.9412525"},{"issue":"12","key":"7_CR10","doi-asserted-by":"publisher","first-page":"9215","DOI":"10.1007\/s00464-022-09405-5","volume":"36","author":"T Golany","year":"2022","unstructured":"Golany, T., et al.: Artificial intelligence for phase recognition in complex laparoscopic cholecystectomy. Surg. Endosc. 36(12), 9215\u20139223 (2022)","journal-title":"Surg. Endosc."},{"key":"7_CR11","unstructured":"He, Z., Mottaghi, A., Sharghi, A., Jamal, M.A., Mohareri, O.: An empirical study on activity recognition in long surgical videos. In: Machine Learning for Health, pp. 356\u2013372. PMLR (2022)"},{"key":"7_CR12","doi-asserted-by":"crossref","unstructured":"Huang, G.: Surgical action recognition and prediction with transformers. In: 2022 IEEE 2nd International Conference on Software Engineering and Artificial Intelligence (SEAI), pp. 36\u201340. IEEE (2022)","DOI":"10.1109\/SEAI55746.2022.9832094"},{"key":"7_CR13","doi-asserted-by":"publisher","first-page":"101572","DOI":"10.1016\/j.media.2019.101572","volume":"59","author":"Y Jin","year":"2020","unstructured":"Jin, Y., et al.: Multi-task recurrent convolutional network with correlation loss for surgical video analysis. Med. Image Anal. 59, 101572 (2020)","journal-title":"Med. Image Anal."},{"key":"7_CR14","doi-asserted-by":"publisher","first-page":"780","DOI":"10.1038\/s41551-023-01010-8","volume":"7","author":"D Kiyasseh","year":"2023","unstructured":"Kiyasseh, D., et al.: A vision transformer for decoding surgeon activity from surgical videos. Nat. Biomed. Eng. 7, 780\u2013796 (2023)","journal-title":"Nat. Biomed. Eng."},{"key":"7_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"70","DOI":"10.1007\/978-3-319-67543-5_7","volume-title":"Computer Assisted and Robotic Endoscopy and Clinical Image-Based Procedures","author":"A Leibetseder","year":"2017","unstructured":"Leibetseder, A., Primus, M.J., Petscharnig, S., Schoeffmann, K.: Image-based smoke detection in laparoscopic videos. In: Cardoso, M.J., et al. (eds.) CARE\/CLIP -2017. LNCS, vol. 10550, pp. 70\u201387. Springer, Cham (2017). https:\/\/doi.org\/10.1007\/978-3-319-67543-5_7"},{"issue":"3","key":"7_CR16","doi-asserted-by":"publisher","first-page":"e2017.00021","DOI":"10.4293\/JSLS.2017.00021","volume":"21","author":"S Lim","year":"2017","unstructured":"Lim, S., Ghosh, S., Niklewski, P., Roy, S.: Laparoscopic suturing as a barrier to broader adoption of laparoscopic surgery. J. Soc. Laparoendosc. Surg. 21(3), e2017.00021 (2017)","journal-title":"J. Soc. Laparoendosc. Surg."},{"key":"7_CR17","doi-asserted-by":"publisher","first-page":"553","DOI":"10.1007\/s00464-017-5878-1","volume":"32","author":"C Loukas","year":"2018","unstructured":"Loukas, C.: Video content analysis of surgical procedures. Surg. Endosc. 32, 553\u2013568 (2018)","journal-title":"Surg. Endosc."},{"issue":"1","key":"7_CR18","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1002\/rcs.1578","volume":"11","author":"C Loukas","year":"2015","unstructured":"Loukas, C., Georgiou, E.: Smoke detection in endoscopic surgery videos: a first step towards retrieval of semantic events. Int. J. Med. Robot. Comput. Assist. Surg. 11(1), 80\u201394 (2015)","journal-title":"Int. J. Med. Robot. Comput. Assist. Surg."},{"key":"7_CR19","doi-asserted-by":"publisher","first-page":"13","DOI":"10.1016\/j.cmpb.2018.07.004","volume":"165","author":"C Loukas","year":"2018","unstructured":"Loukas, C., Varytimidis, C., Rapantzikos, K., Kanakis, M.A.: Keyframe extraction from laparoscopic videos based on visual saliency detection. Comput. Methods Programs Biomed. 165, 13\u201323 (2018)","journal-title":"Comput. Methods Programs Biomed."},{"key":"7_CR20","doi-asserted-by":"publisher","first-page":"521","DOI":"10.1007\/s11042-009-0353-1","volume":"46","author":"M Lux","year":"2010","unstructured":"Lux, M., Marques, O., Sch\u00f6ffmann, K., B\u00f6sz\u00f6rmenyi, L., Lajtai, G.: A novel tool for summarization of arthroscopic videos. Multimedia Tools Appl. 46, 521\u2013544 (2010)","journal-title":"Multimedia Tools Appl."},{"key":"7_CR21","unstructured":"Namazi, B., Sankaranarayanan, G., Devarajan, V.: Automatic detection of surgical phases in laparoscopic videos. In: Proceedings on the International Conference in Artificial Intelligence (ICAI), pp. 124\u2013130 (2018)"},{"key":"7_CR22","doi-asserted-by":"publisher","first-page":"679","DOI":"10.1007\/s00464-021-08336-x","volume":"36","author":"B Namazi","year":"2022","unstructured":"Namazi, B., Sankaranarayanan, G., Devarajan, V.: A contextual detector of surgical tools in laparoscopic videos using deep learning. Surg. Endosc. 36, 679\u2013688 (2022)","journal-title":"Surg. Endosc."},{"key":"7_CR23","doi-asserted-by":"crossref","unstructured":"Nasirihaghighi, S., Ghamsarian, N., Stefanics, D., Schoeffmann, K., Husslein, H.: Action recognition in video recordings from gynecologic laparoscopy. In: 2023 IEEE 36th International Symposium on Computer-Based Medical Systems (CBMS), pp. 29\u201334 (2023)","DOI":"10.1109\/CBMS58004.2023.00187"},{"issue":"4","key":"7_CR24","doi-asserted-by":"publisher","first-page":"236","DOI":"10.5144\/0256-4947.2023.236","volume":"43","author":"M Polat","year":"2023","unstructured":"Polat, M., Incebiyik, A., Tammo, O.: Abdominal access in laparoscopic surgery of obese patients: a novel abdominal access technique. Ann. Saudi Med. 43(4), 236\u2013242 (2023)","journal-title":"Ann. Saudi Med."},{"key":"7_CR25","doi-asserted-by":"publisher","first-page":"11187","DOI":"10.1007\/s11042-014-2224-7","volume":"74","author":"K Schoeffmann","year":"2015","unstructured":"Schoeffmann, K., Del Fabro, M., Szkaliczki, T., B\u00f6sz\u00f6rmenyi, L., Keckstein, J.: Keyframe extraction in endoscopic video. Multimedia Tools Appl. 74, 11187\u201311206 (2015)","journal-title":"Multimedia Tools Appl."},{"key":"7_CR26","doi-asserted-by":"crossref","unstructured":"Shi, C., Zheng, Y., Fey, A.M.: Recognition and prediction of surgical gestures and trajectories using transformer models in robot-assisted surgery. In: 2022 IEEE\/RSJ International Conference on Intelligent Robots and Systems (IROS), pp. 8017\u20138024. IEEE (2022)","DOI":"10.1109\/IROS47612.2022.9981611"},{"issue":"2","key":"7_CR27","first-page":"406","volume":"9","author":"P Shi","year":"2022","unstructured":"Shi, P., Zhao, Z., Liu, K., Li, F.: Attention-based spatial-temporal neural network for accurate phase recognition in minimally invasive surgery: feasibility and efficiency verification. J. Comput. Des. Eng. 9(2), 406\u2013416 (2022)","journal-title":"J. Comput. Des. Eng."},{"issue":"1","key":"7_CR28","doi-asserted-by":"publisher","first-page":"86","DOI":"10.1109\/TMI.2016.2593957","volume":"36","author":"AP Twinanda","year":"2016","unstructured":"Twinanda, A.P., Shehata, S., Mutter, D., Marescaux, J., De Mathelin, M., Padoy, N.: EndoNet: a deep architecture for recognition tasks on laparoscopic videos. IEEE Trans. Med. Imaging 36(1), 86\u201397 (2016)","journal-title":"IEEE Trans. Med. Imaging"},{"key":"7_CR29","unstructured":"Vaswani, A., et al.: Attention is all you need. In: Advances in Neural Information Processing Systems, vol. 30 (2017)"},{"key":"7_CR30","unstructured":"Wang, C., Cheikh, F.A., Kaaniche, M., Elle, O.J.: A smoke removal method for laparoscopic images. arXiv preprint arXiv:1803.08410 (2018)"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-56435-2_7","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,19]],"date-time":"2024-03-19T07:03:59Z","timestamp":1710831839000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-56435-2_7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024]]},"ISBN":["9783031564345","9783031564352"],"references-count":30,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-56435-2_7","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024]]},"assertion":[{"value":"20 March 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Amsterdam","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"The Netherlands","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 January 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2 February 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"30","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"ConfTool Pro","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"297","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"120","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"40% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.2","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}