{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,2,21]],"date-time":"2025-02-21T13:25:18Z","timestamp":1740144318744,"version":"3.37.3"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2022,4,26]],"date-time":"2022-04-26T00:00:00Z","timestamp":1650931200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,4,26]],"date-time":"2022-04-26T00:00:00Z","timestamp":1650931200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/100010477","name":"Intuitive Surgical","doi-asserted-by":"publisher","id":[{"id":"10.13039\/100010477","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001665","name":"Agence Nationale de la Recherche","doi-asserted-by":"publisher","award":["ANR-10-IAHU-02"],"award-info":[{"award-number":["ANR-10-IAHU-02"]}],"id":[{"id":"10.13039\/501100001665","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"published-print":{"date-parts":[[2022,8]]},"DOI":"10.1007\/s11548-022-02629-9","type":"journal-article","created":{"date-parts":[[2022,4,26]],"date-time":"2022-04-26T09:02:46Z","timestamp":1650963766000},"page":"1469-1476","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["Self-supervised learning via cluster distance prediction for operating room context awareness"],"prefix":"10.1007","volume":"17","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-6791-9703","authenticated-orcid":false,"given":"Idris","family":"Hamoud","sequence":"first","affiliation":[]},{"given":"Alexandros","family":"Karargyris","sequence":"additional","affiliation":[]},{"given":"Aidean","family":"Sharghi","sequence":"additional","affiliation":[]},{"given":"Omid","family":"Mohareri","sequence":"additional","affiliation":[]},{"given":"Nicolas","family":"Padoy","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,4,26]]},"reference":[{"issue":"11","key":"2629_CR1","doi-asserted-by":"publisher","first-page":"2274","DOI":"10.1109\/TPAMI.2012.120","volume":"34","author":"R Achanta","year":"2012","unstructured":"Achanta R, Shaji A, Smith K, Lucchi A, Fua P, S\u00fcsstrunk S (2012) Slic superpixels compared to state-of-the-art superpixel methods. IEEE TPAMI 34(11):2274","journal-title":"IEEE TPAMI"},{"key":"2629_CR2","unstructured":"Adam S, Aidean S, Helene H, Daniel O, Omid M (2021) Multi-view surgical video action detection via mixed global view attention. In: MICCAI"},{"key":"2629_CR3","unstructured":"Aidean S, Helene H, Daniel O, Omid M (2020) Automatic operating room surgical activity recognition for robot-assisted surgery. In: MICCAI"},{"key":"2629_CR4","unstructured":"Asano YM, Rupprecht C, Vedaldi A (2020) A critical analysis of self-supervision, or what we can learn from a single image. In: CVPR"},{"key":"2629_CR5","doi-asserted-by":"crossref","unstructured":"Azizi S, Mustafa B, Ryan F, Beaver Z, Freyberg J, Deaton J, Loh A, Karthikesalingam A, Kornblith S, Chen T, Natarajan V, Norouzi M (2021) Big self-supervised models advance medical image classification (ICCV)","DOI":"10.1109\/ICCV48922.2021.00346"},{"key":"2629_CR6","unstructured":"Caron M, Misra I, Mairal J, Goyal P, Bojanowski P, Joulin A (2021) Unsupervised learning of visual features by contrasting cluster assignments"},{"key":"2629_CR7","doi-asserted-by":"crossref","unstructured":"Caron M, Touvron H, Misra I, J\u00e9gou H, Mairal J, Bojanowski P, Joulin A (2021) Emerging properties in self-supervised vision transformers. In: Proceedings of ICCV","DOI":"10.1109\/ICCV48922.2021.00951"},{"key":"2629_CR8","doi-asserted-by":"publisher","first-page":"3749","DOI":"10.1007\/s00464-015-4671-2","volume":"30","author":"K Catchpole","year":"2015","unstructured":"Catchpole K, Perkins CE, Bresee C, Solnik MJ, Sherman B, Fritch JL, Gross B, Jagannathan S, Hakami-Majd N, Avenido RM, Anger JT (2015) Safety, efficiency and learning curves in robotic surgery: a human factors analysis. Surg Endosc 30:3749\u20133761","journal-title":"Surg Endosc"},{"key":"2629_CR9","doi-asserted-by":"crossref","unstructured":"Chakraborty I, Elgammal A, Burd RS (2013) Video based activity recognition in trauma resuscitation. In: 2013 10th IEEE international conference and workshops on automatic face and gesture recognition (FG), pp 1\u20138","DOI":"10.1109\/FG.2013.6553758"},{"key":"2629_CR10","unstructured":"Chen T, Kornblith S, Norouzi M, Hinton GE (2020) A simple framework for contrastive learning of visual representations. arXiv:2002.05709"},{"key":"2629_CR11","doi-asserted-by":"crossref","unstructured":"Dias RD, Yule SJ, Zenati MA (2020) Augmented cognition in the operating room","DOI":"10.1007\/978-3-030-49100-0_19"},{"key":"2629_CR12","doi-asserted-by":"crossref","unstructured":"Doersch C, Gupta AK, Efros AA (2015) Unsupervised visual representation learning by context prediction. In: 2015 IEEE ICCV","DOI":"10.1109\/ICCV.2015.167"},{"key":"2629_CR13","unstructured":"Gidaris S, Singh P, Komodakis N (2018) Unsupervised representation learning by predicting image rotations. In: ICLR"},{"key":"2629_CR14","unstructured":"Grill JB, Strub F, Altch\u2019e F, Tallec C, Richemond PH, Buchatskaya E, Doersch C, Pires B\u00c1, Guo ZD, Azar MG, Piot B, Kavukcuoglu K, Munos R, Valko M (2020) Bootstrap your own latent: a new approach to self-supervised learning. In: NeurIPS"},{"key":"2629_CR15","first-page":"203","volume":"47","author":"HA Hajj","year":"2018","unstructured":"Hajj HA, Lamard M, Conze PH, Cochener B, Quellec G (2018) Monitoring tool usage in surgery videos using boosted convolutional and recurrent neural networks. MedIA 47:203\u2013218","journal-title":"MedIA"},{"key":"2629_CR16","doi-asserted-by":"crossref","unstructured":"He K, Fan H, Wu Y, Xie S, Girshick RB (2020) Momentum contrast for unsupervised visual representation learning. In: 2020 IEEE\/CVF CVPR, pp 9726\u20139735","DOI":"10.1109\/CVPR42600.2020.00975"},{"key":"2629_CR17","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Identity mappings in deep residual networks. In: Computer vision\u2014ECCV 2016, Springer International Publishing, pp 630\u2013645","DOI":"10.1007\/978-3-319-46493-0_38"},{"key":"2629_CR18","doi-asserted-by":"crossref","unstructured":"Issenhuth T, Srivastav VK, Gangi A, Padoy N (2019) Face detection in the operating room: comparison of state-of-the-art methods and a self-supervised approach. In: IJCARS","DOI":"10.1007\/s11548-019-01944-y"},{"key":"2629_CR19","doi-asserted-by":"crossref","unstructured":"Kadkhodamohammadi A, Gangi A, de\u00a0Mathelin M, Padoy N A multi-view rgb-d approach for human pose estimation in operating rooms. In: 2017 IEEE WACV","DOI":"10.1109\/WACV.2017.47"},{"key":"2629_CR20","unstructured":"Li Z, Shaban A, Simard JG, Rabindran D, DiMaio SP, Mohareri O (2020) A robotic 3d perception system for operating room environment awareness. In: IPCAI"},{"key":"2629_CR21","doi-asserted-by":"crossref","unstructured":"Liu MY, Tuzel O, Ramalingam S, Chellappa R (2011) Entropy rate superpixel segmentation. In: CVPR 2011, pp 2097\u20132104","DOI":"10.1109\/CVPR.2011.5995323"},{"key":"2629_CR22","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: 2015 IEEE CVPR, pp 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"2629_CR23","unstructured":"Luo Z, Hsieh JT, Balachandar N, Yeung S, Pusiol G, Luxenberg JS, Li G, Li LJ, Milstein A, Fei-Fei L (2018) Vision-based descriptive analytics of seniors\u2014daily activities for long-term health monitoring"},{"key":"2629_CR24","doi-asserted-by":"crossref","unstructured":"Newell A, Deng J (2020) How useful is self-supervised pretraining for visual tasks? In: 2020 IEEE\/CVF conference on computer vision and pattern recognition (CVPR)","DOI":"10.1109\/CVPR42600.2020.00737"},{"key":"2629_CR25","doi-asserted-by":"crossref","unstructured":"Noroozi M, Favaro P (2016) Unsupervised learning of visual representations by solving jigsaw puzzles. In: ECCV","DOI":"10.1007\/978-3-319-46466-4_5"},{"key":"2629_CR26","doi-asserted-by":"crossref","unstructured":"Ouyang C, Biffi C, Chen C, Kart T, Qiu H, Rueckert D (2020) Self-supervision with superpixels: training few-shot medical image segmentation without annotation. In: ECCV","DOI":"10.1007\/978-3-030-58526-6_45"},{"key":"2629_CR27","doi-asserted-by":"crossref","unstructured":"Ro\u00df T, Zimmerer D, Vemuri AS, Isensee F, Bodenstedt S, Both F, Kessler P, Wagner M, M\u00fcller-Stich BP, Kenngott H, Speidel S, Maier-Hein K, Maier-Hein L (2018) Exploiting the potential of unlabeled endoscopic video data with self-supervised learning. In: IJCARS","DOI":"10.1007\/s11548-018-1772-0"},{"key":"2629_CR28","doi-asserted-by":"publisher","first-page":"e1918911","DOI":"10.1001\/jamanetworkopen.2019.18911","volume":"3","author":"KH Sheetz","year":"2020","unstructured":"Sheetz KH, Claflin J (2020) Trends in the adoption of robotic surgery for common surgical procedures. JAMA Netw Open 3:e1918911","journal-title":"JAMA Netw Open"},{"key":"2629_CR29","doi-asserted-by":"crossref","unstructured":"Srivastav VK, Gangi A, Padoy N (2019) Human pose estimation on privacy-preserving low-resolution depth images. In: MICCAI. arXiv:2007.08340","DOI":"10.1007\/978-3-030-32254-0_65"},{"key":"2629_CR30","unstructured":"Srivastav VK, Issenhuth T, Kadkhodamohammadi A, de\u00a0Mathelin M, Gangi A, Padoy N (2018) Mvor: a multi-view RGB-d operating room dataset for 2d and 3d human pose estimation. arXiv:1808.08180"},{"key":"2629_CR31","unstructured":"Taleb A, Loetzsch W, Danz N, Severin J thomas. gaertner, Bergner B, Lippert C (2020) 3d self-supervised methods for medical imaging. In: NeurIPS. arXiv:2006.03829"},{"issue":"1","key":"2629_CR32","first-page":"86","volume":"36","author":"AP Twinanda","year":"2017","unstructured":"Twinanda AP, Shehata S, Mutter D, Marescaux J, de Mathelin M, Padoy N (2017) EndoNet: a deep architecture for recognition tasks on laparoscopic videos. IEEE TMI 36(1):86\u201397","journal-title":"IEEE TMI"},{"key":"2629_CR33","unstructured":"Twinanda AP, Winata P, Gangi A, De M, Mathelin PN (2017) Multi-stream deep architecture for surgical phase recognition on multi-view RGBD videos"},{"issue":"86","key":"2629_CR34","first-page":"2579","volume":"9","author":"L van der Maaten","year":"2008","unstructured":"van der Maaten L, Hinton G (2008) Visualizing data using t-sne. J Mach Learn Res 9(86):2579\u20132605","journal-title":"J Mach Learn Res"},{"key":"2629_CR35","unstructured":"van\u00a0den Oord A, Li Y, Vinyals O (2018) Representation learning with contrastive predictive coding. arXiv:1807.03748"},{"key":"2629_CR36","doi-asserted-by":"crossref","unstructured":"Wang X, Zhang R, Shen C, Kong T, Li L (2021) Dense contrastive learning for self-supervised visual pre-training. In: 2021 IEEE\/CVF CVPR, pp 3023\u20133032","DOI":"10.1109\/CVPR46437.2021.00304"},{"key":"2629_CR37","unstructured":"Yu T, Mutter D, Marescaux J, Padoy N (2019) Learning from a tiny dataset of manual annotations: a teacher\/student approach for surgical phase recognition. In: IPCAI"},{"key":"2629_CR38","unstructured":"Zbontar J, Jing L, Misra I, LeCun Y, Deny S (2021) Barlow twins: self-supervised learning via redundancy reduction. In: ICML"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-022-02629-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-022-02629-9\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-022-02629-9.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,7,23]],"date-time":"2022-07-23T03:41:18Z","timestamp":1658547678000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-022-02629-9"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,4,26]]},"references-count":38,"journal-issue":{"issue":"8","published-print":{"date-parts":[[2022,8]]}},"alternative-id":["2629"],"URL":"https:\/\/doi.org\/10.1007\/s11548-022-02629-9","relation":{},"ISSN":["1861-6429"],"issn-type":[{"type":"electronic","value":"1861-6429"}],"subject":[],"published":{"date-parts":[[2022,4,26]]},"assertion":[{"value":"4 March 2022","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 March 2022","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 April 2022","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Idris Hamoud is funded by a research scholarship from Intuitive Surgical. Nicolas Padoy is a scientific advisor to Caresyntax on topics unrelated to this study. The other authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and\/or national research committee and with the 1964 Declaration of Helsinki and its later amendments or comparable ethical standards.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"Data have been collected within an Institutional Review Board (IRB)-approved study, and all participants\u2019 informed consent has been obtained.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}