{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,7,25]],"date-time":"2025-07-25T10:01:14Z","timestamp":1753437674959,"version":"3.41.0"},"reference-count":28,"publisher":"Springer Science and Business Media LLC","issue":"6","license":[{"start":{"date-parts":[[2025,5,14]],"date-time":"2025-05-14T00:00:00Z","timestamp":1747180800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"},{"start":{"date-parts":[[2025,5,14]],"date-time":"2025-05-14T00:00:00Z","timestamp":1747180800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/creativecommons.org\/licenses\/by\/4.0"}],"funder":[{"DOI":"10.13039\/501100013348","name":"Innosuisse - Schweizerische Agentur f\u00fcr Innovationsf\u00f6rderung","doi-asserted-by":"publisher","award":["PFFS-21-19"],"award-info":[{"award-number":["PFFS-21-19"]}],"id":[{"id":"10.13039\/501100013348","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"abstract":"<jats:title>Abstract<\/jats:title>\n          <jats:sec>\n            <jats:title>\n              <jats:bold>Purpose<\/jats:bold>\n            <\/jats:title>\n            <jats:p>Advances in computer vision, particularly in optical image-based 3D reconstruction and feature matching, enable applications like marker-less surgical navigation and digitization of surgery. However, their development is hindered by a lack of suitable datasets with 3D ground truth. This work explores an approach to generating realistic and accurate ex vivo datasets tailored for 3D reconstruction and feature matching in open orthopedic surgery.<\/jats:p>\n          <\/jats:sec>\n          <jats:sec>\n            <jats:title>\n              <jats:bold>Methods<\/jats:bold>\n            <\/jats:title>\n            <jats:p>A set of posed images and an accurately registered ground truth surface mesh of the scene are required to develop vision-based 3D reconstruction and matching methods suitable for surgery. We propose a framework consisting of three core steps and compare different methods for each step: 3D scanning, calibration of viewpoints for a set of high-resolution RGB images, and an optical method for scene registration.<\/jats:p>\n          <\/jats:sec>\n          <jats:sec>\n            <jats:title>\n              <jats:bold>Results<\/jats:bold>\n            <\/jats:title>\n            <jats:p>We evaluate each step of this framework on an ex vivo scoliosis surgery using a pig spine, conducted under real operating room conditions. A mean 3D Euclidean error of 0.35 mm is achieved with respect to the 3D ground truth.<\/jats:p>\n          <\/jats:sec>\n          <jats:sec>\n            <jats:title>\n              <jats:bold>Conclusion<\/jats:bold>\n            <\/jats:title>\n            <jats:p>The proposed method results in submillimeter-accurate 3D ground truths and surgical images with a spatial resolution of 0.1 mm. This opens the door to acquiring future surgical datasets for high-precision applications.<\/jats:p>\n          <\/jats:sec>","DOI":"10.1007\/s11548-025-03385-2","type":"journal-article","created":{"date-parts":[[2025,5,14]],"date-time":"2025-05-14T11:56:03Z","timestamp":1747223763000},"page":"1293-1300","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Acquiring submillimeter-accurate multi-task vision datasets for computer-assisted orthopedic surgery"],"prefix":"10.1007","volume":"20","author":[{"ORCID":"https:\/\/orcid.org\/0009-0007-4064-6841","authenticated-orcid":false,"given":"Emma","family":"Most","sequence":"first","affiliation":[]},{"given":"Jonas","family":"Hein","sequence":"additional","affiliation":[]},{"given":"Fr\u00e9d\u00e9ric","family":"Giraud","sequence":"additional","affiliation":[]},{"given":"Nicola A.","family":"Cavalcanti","sequence":"additional","affiliation":[]},{"given":"Lukas","family":"Zingg","sequence":"additional","affiliation":[]},{"given":"Baptiste","family":"Brument","sequence":"additional","affiliation":[]},{"given":"Nino","family":"Louman","sequence":"additional","affiliation":[]},{"given":"Fabio","family":"Carrillo","sequence":"additional","affiliation":[]},{"given":"Philipp","family":"F\u00fcrnstahl","sequence":"additional","affiliation":[]},{"given":"Lilian","family":"Calvet","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,5,14]]},"reference":[{"key":"3385_CR1","doi-asserted-by":"publisher","DOI":"10.3389\/fsurg.2021.640554","volume":"8","author":"I Kalfas","year":"2021","unstructured":"Kalfas I, Nguyen A (2021) Machine vision navigation in spine surgery. Front Surg 8:640554. https:\/\/doi.org\/10.3389\/fsurg.2021.640554","journal-title":"Front Surg"},{"key":"3385_CR2","doi-asserted-by":"publisher","unstructured":"Kiyasseh D, Ma R, Haque TF, Nguyen J, Wagner C, Anandkumar A, Hung AJ (2022) Quantification of robotic surgeries with vision-based deep learning. CoRR abs\/2205.03028. https:\/\/doi.org\/10.48550\/arXiv.2205.03028","DOI":"10.48550\/arXiv.2205.03028"},{"key":"3385_CR3","doi-asserted-by":"crossref","unstructured":"Hein J, Giraud F, Calvet L, Schwarz A, Cavalcanti NA, Prokudin S, Farshad M, Tang S, Pollefeys M, Carrillo F, F\u00fcrnstahl P (2024) Creating a digital twin of spinal surgery: a proof of concept. In: CVPR workshops, pp 2355\u20132364","DOI":"10.1109\/CVPRW63382.2024.00241"},{"key":"3385_CR4","doi-asserted-by":"publisher","first-page":"164","DOI":"10.3390\/jimaging7090164","volume":"7","author":"F Liebmann","year":"2021","unstructured":"Liebmann F, St\u00fctz D, Suter D, Jecklin S, Snedeker J, Farshad M, F\u00fcrnstahl P, Esfandiari H (2021) Spinedepth: a multi-modal data collection approach for automatic labelling and intraoperative spinal shape reconstruction based on rgb-d data. J Imaging 7:164. https:\/\/doi.org\/10.3390\/jimaging7090164","journal-title":"J Imaging"},{"issue":"1","key":"3385_CR5","doi-asserted-by":"publisher","first-page":"162","DOI":"10.1016\/j.wneu.2012.03.011","volume":"79","author":"R H\u00e4rtl","year":"2013","unstructured":"H\u00e4rtl R, Lam KS, Wang J, Korge A, Kandziora F, Audig\u00e9 L (2013) Worldwide survey on the use of navigation in spine surgery. World Neurosurg 79(1):162\u2013172","journal-title":"World Neurosurg"},{"key":"3385_CR6","doi-asserted-by":"publisher","unstructured":"Barnoy Y, O\u2019Brien M, Wang W, Hager G (2021) Obotic surgery with lean reinforcement learning. arXiv preprint arXiv:2105.01006, https:\/\/doi.org\/10.48550\/arXiv.2105.01006","DOI":"10.48550\/arXiv.2105.01006"},{"issue":"18","key":"3385_CR7","doi-asserted-by":"publisher","first-page":"5861","DOI":"10.3390\/s24185861","volume":"24","author":"Y Lee","year":"2024","unstructured":"Lee Y (2024) Three-dimensional dense reconstruction: a review of algorithms and datasets. Sensors 24(18):5861. https:\/\/doi.org\/10.3390\/s24185861","journal-title":"Sensors"},{"key":"3385_CR8","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2024.103131","volume":"94","author":"A Schmidt","year":"2024","unstructured":"Schmidt A, Mohareri O, DiMaio S, Yip MC, Salcudean SE (2024) Tracking and mapping in medical computer vision: a review. Med Image Anal 94:103131. https:\/\/doi.org\/10.1016\/j.media.2024.103131","journal-title":"Med Image Anal"},{"key":"3385_CR9","doi-asserted-by":"publisher","first-page":"72","DOI":"10.1007\/978-3-319-54057-3_7","volume-title":"Computer-assisted and robotic endoscopy","author":"N Mahmoud","year":"2017","unstructured":"Mahmoud N, Cirauqui I, Hostettler A, Doignon C, Soler L, Marescaux J, Montiel JMM (2017) Orbslam-based endoscope tracking and 3d reconstruction. Computer-assisted and robotic endoscopy. Springer, Cham, pp 72\u201383"},{"key":"3385_CR10","doi-asserted-by":"publisher","DOI":"10.1016\/j.compbiomed.2024.108546","volume":"175","author":"Z Yang","year":"2024","unstructured":"Yang Z, Dai J, Pan J (2024) 3d reconstruction from endoscopy images: a survey. Comput Biol Med 175:108546","journal-title":"Comput Biol Med"},{"key":"3385_CR11","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1117\/12.380037","volume-title":"Three-dimensional image capture and applications III","author":"K Keller","year":"2000","unstructured":"Keller K, Ackerman JD (2000) Real-time structured light depth extraction. Three-dimensional image capture and applications III, vol 3958. SPIE, Bellingham, pp 11\u201318. https:\/\/doi.org\/10.1117\/12.380037"},{"key":"3385_CR12","doi-asserted-by":"crossref","unstructured":"Richa R, Poignet P, Liu C (2008) Efficient 3d tracking for motion compensation in beating heart surgery. In: MICCAI, pp 684\u2013691","DOI":"10.1007\/978-3-540-85990-1_82"},{"key":"3385_CR13","doi-asserted-by":"publisher","unstructured":"Yip MC, Lowe DG, Salcudean SE, Rohling RN, Nguan CY (2012) Tissue tracking and registration for image-guided surgery. IEEE Trans Med Imaging 31(11):2169\u20132182. https:\/\/doi.org\/10.1109\/TMI.2012.2212718","DOI":"10.1109\/TMI.2012.2212718"},{"issue":"3","key":"3385_CR14","doi-asserted-by":"crossref","first-page":"63","DOI":"10.3390\/jimaging8030063","volume":"8","author":"F Liebmann","year":"2022","unstructured":"Liebmann F, St\u00fctz D, Suter D, Jecklin S, Snedeker JG, Farshad M, F\u00fcrnstahl P, Esfandiari H (2022) Spinedepth: a multi-modal data collection approach for automatic labelling and intraoperative spinal shape reconstruction based on rgb-d data. J. Imaging 8(3):63","journal-title":"J. Imaging"},{"key":"3385_CR15","doi-asserted-by":"crossref","unstructured":"Eftekhar A, Sax A, Malik J, Zamir A (2021) Omnidata: a scalable pipeline for making multi-task mid-level vision datasets from 3d scans. In: ICCV, pp 10786\u201310796","DOI":"10.1109\/ICCV48922.2021.01061"},{"key":"3385_CR16","doi-asserted-by":"crossref","unstructured":"Li Z, Snavely N (2018) Megadepth: learning single-view depth prediction from internet photos. In: CVPR, pp 2041\u20132050","DOI":"10.1109\/CVPR.2018.00218"},{"key":"3385_CR17","unstructured":"Yu Z, Peng S, Niemeyer M, Sattler T, Geiger A (2022) Monosdf: exploring monocular geometric cues for neural implicit surface reconstruction. NeurIPS 35:25018\u201325032"},{"key":"3385_CR18","doi-asserted-by":"crossref","unstructured":"Wang S, Leroy V, Cabon Y, Chidlovskii B, Revaud J (2024) Dust3r: geometric 3d vision made easy. In: CVPR","DOI":"10.1109\/CVPR52733.2024.01956"},{"key":"3385_CR19","doi-asserted-by":"crossref","unstructured":"Lindenberger P, Sarlin P-E, Pollefeys M (2023) LightGlue: local feature matching at light speed. In: ICCV","DOI":"10.1109\/ICCV51070.2023.01616"},{"key":"3385_CR20","doi-asserted-by":"crossref","unstructured":"Sarlin P-E, DeTone D, Malisiewicz T, Rabinovich A (2020) Superglue: learning feature matching with graph neural networks. In: CVPR","DOI":"10.1109\/CVPR42600.2020.00499"},{"issue":"1","key":"3385_CR21","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1038\/s41597-023-02564-7","volume":"10","author":"P Azagra","year":"2023","unstructured":"Azagra P, Sostres C, Ferr\u00e1ndez Riazuelo L, Tomasini C, Barbed OL, Morlana J, Recasens D, Batlle VM, G\u00f3mez-Rodr\u00edguez JJ, Elvira R, L\u00f3pez J, Oriol C, Civera J, Tard\u00f3s JD, Murillo AC, Lanas A, Montiel JMM (2023) Endomapper dataset of complete calibrated endoscopy procedures. Sci Data 10(1):1\u201316. https:\/\/doi.org\/10.1038\/s41597-023-02564-7","journal-title":"Sci Data"},{"key":"3385_CR22","doi-asserted-by":"crossref","unstructured":"Sch\u00f6nberger JL, Zheng E, Pollefeys M, Frahm J-M (2016) Pixelwise view selection for unstructured multi-view stereo. In: ECCV","DOI":"10.1007\/978-3-319-46487-9_31"},{"key":"3385_CR23","doi-asserted-by":"crossref","unstructured":"Pan L, Barath D, Pollefeys M, Sch\u00f6nberger JL (2024) Global structure-from-motion revisited. In: ECCV","DOI":"10.1007\/978-3-031-73661-2_4"},{"issue":"3","key":"3385_CR24","doi-asserted-by":"publisher","first-page":"499","DOI":"10.1109\/TPAMI.2007.45","volume":"29","author":"H Zhang","year":"2007","unstructured":"Zhang H, Wong K-yK, Zhang G (2007) Camera calibration from images of spheres. IEEE Trans Pattern Anal Mach Intell 29(3):499\u2013502. https:\/\/doi.org\/10.1109\/TPAMI.2007.45","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"3385_CR25","doi-asserted-by":"publisher","unstructured":"Voynov et al (2023) Multi-sensor large-scale dataset for multi-view 3d reconstruction. In: CVPR, pp 21392\u201321403. https:\/\/doi.org\/10.1109\/CVPR52729.2023.02049","DOI":"10.1109\/CVPR52729.2023.02049"},{"key":"3385_CR26","doi-asserted-by":"crossref","unstructured":"Wang Y, Han Q, Habermann M, Daniilidis K, Theobalt C, Liu L (2023) Neus2: fast learning of neural implicit surfaces for multi-view reconstruction. In: ICCV","DOI":"10.1109\/ICCV51070.2023.00305"},{"issue":"4","key":"3385_CR27","doi-asserted-by":"publisher","first-page":"102","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller T, Evans A, Schied C, Keller A (2022) Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans Graph 41(4):102\u2013110215. https:\/\/doi.org\/10.1145\/3528223.3530127","journal-title":"ACM Trans Graph"},{"key":"3385_CR28","doi-asserted-by":"crossref","unstructured":"Gu\u00e9don A, Lepetit V (2024) Sugar: surface-aligned gaussian splatting for efficient 3d mesh reconstruction and high-quality mesh rendering. In: CVPR","DOI":"10.1109\/CVPR52733.2024.00512"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-025-03385-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-025-03385-2\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-025-03385-2.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,6,14]],"date-time":"2025-06-14T11:34:31Z","timestamp":1749900871000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-025-03385-2"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,5,14]]},"references-count":28,"journal-issue":{"issue":"6","published-online":{"date-parts":[[2025,6]]}},"alternative-id":["3385"],"URL":"https:\/\/doi.org\/10.1007\/s11548-025-03385-2","relation":{},"ISSN":["1861-6429"],"issn-type":[{"type":"electronic","value":"1861-6429"}],"subject":[],"published":{"date-parts":[[2025,5,14]]},"assertion":[{"value":"3 March 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 March 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"14 May 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no relevant financial or non-financial interests to disclose.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"Since this study did not involve human or animal subjects, ethics approval is not applicable.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethical approval"}},{"value":"As this research did not involve human participants, informed consent is not applicable.","order":4,"name":"Ethics","group":{"name":"EthicsHeading","label":"Informed consent"}}]}}