{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,4,22]],"date-time":"2026-04-22T20:10:02Z","timestamp":1776888602999,"version":"3.51.2"},"reference-count":30,"publisher":"Springer Science and Business Media LLC","issue":"5","license":[{"start":{"date-parts":[[2025,2,24]],"date-time":"2025-02-24T00:00:00Z","timestamp":1740355200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,2,24]],"date-time":"2025-02-24T00:00:00Z","timestamp":1740355200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100016073","name":"Key Technologies Research and Development Program of Anhui Province","doi-asserted-by":"publisher","award":["2023YFF0723300"],"award-info":[{"award-number":["2023YFF0723300"]}],"id":[{"id":"10.13039\/100016073","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100012226","name":"Fundamental Research Funds for the Central Universities","doi-asserted-by":"publisher","award":["2042024kf0003, 2042024kf1010, 2042023kf0105"],"award-info":[{"award-number":["2042024kf0003, 2042024kf1010, 2042023kf0105"]}],"id":[{"id":"10.13039\/501100012226","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100019035","name":"Science and Technology Program of Hubei Province","doi-asserted-by":"publisher","award":["202319"],"award-info":[{"award-number":["202319"]}],"id":[{"id":"10.13039\/501100019035","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Int J CARS"],"DOI":"10.1007\/s11548-025-03333-0","type":"journal-article","created":{"date-parts":[[2025,2,24]],"date-time":"2025-02-24T16:24:07Z","timestamp":1740414247000},"page":"971-979","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":3,"title":["SfMDiffusion: self-supervised monocular depth estimation in endoscopy based on diffusion models"],"prefix":"10.1007","volume":"20","author":[{"given":"Yu","family":"Li","sequence":"first","affiliation":[]},{"given":"Da","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Die","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Jin","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Lan","family":"Dong","sequence":"additional","affiliation":[]},{"given":"Du","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Liye","family":"Mei","sequence":"additional","affiliation":[]},{"given":"Cheng","family":"Lei","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,2,24]]},"reference":[{"key":"3333_CR1","doi-asserted-by":"publisher","first-page":"3449","DOI":"10.1007\/s00464-019-07121-1","volume":"34","author":"P Zhang","year":"2020","unstructured":"Zhang P, Luo H, Zhu W, Yang J, Zeng N, Fan Y, Wen S, Xiang N, Jia F, Fang C (2020) Real-time navigation for laparoscopic hepatectomy using image fusion of preoperative 3d surgical plan and intraoperative indocyanine green fluorescence imaging. Surg Endosc 34:3449\u20133459","journal-title":"Surg Endosc"},{"key":"3333_CR2","doi-asserted-by":"publisher","first-page":"101988","DOI":"10.1016\/j.bspc.2020.101988","volume":"60","author":"L Mei","year":"2020","unstructured":"Mei L, Guo X, Huang X, Weng Y, Liu S, Lei C (2020) Dense contour-imbalance aware framework for colon gland instance segmentation. Biomed Signal Process Control 60:101988","journal-title":"Biomed Signal Process Control"},{"key":"3333_CR3","doi-asserted-by":"publisher","first-page":"105099","DOI":"10.1016\/j.cmpb.2019.105099","volume":"187","author":"H Luo","year":"2020","unstructured":"Luo H, Yin D, Zhang S, Xiao D, He B, Meng F, Zhang Y, Cai W, He S, Zhang W et al (2020) Augmented reality navigation for liver resection with a stereoscopic laparoscope. Comput Methods Progr Biomed 187:105099","journal-title":"Comput Methods Progr Biomed"},{"issue":"4","key":"3333_CR4","doi-asserted-by":"publisher","first-page":"71","DOI":"10.1145\/3386569.3392377","volume":"39","author":"X Luo","year":"2020","unstructured":"Luo X, Huang J-B, Szeliski R, Matzen K, Kopf J (2020) Consistent video depth estimation. ACM Trans Graph (ToG) 39(4):71\u20131","journal-title":"ACM Trans Graph (ToG)"},{"key":"3333_CR5","doi-asserted-by":"publisher","first-page":"14","DOI":"10.1016\/j.neucom.2020.12.089","volume":"438","author":"Y Ming","year":"2021","unstructured":"Ming Y, Meng X, Fan C, Yu H (2021) Deep learning for monocular depth estimation: a review. Neurocomputing 438:14\u201333","journal-title":"Neurocomputing"},{"issue":"14","key":"3333_CR6","doi-asserted-by":"publisher","first-page":"5353","DOI":"10.3390\/s22145353","volume":"22","author":"A Masoumian","year":"2022","unstructured":"Masoumian A, Rashwan HA, Cristiano J, Asif MS, Puig D (2022) Monocular depth estimation using deep learning: a review. Sensors 22(14):5353","journal-title":"Sensors"},{"key":"3333_CR7","doi-asserted-by":"crossref","unstructured":"Kuznietsov Y, Stuckler J, Leibe B (2017) Semi-supervised deep learning for monocular depth map prediction. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp 6647\u20136655","DOI":"10.1109\/CVPR.2017.238"},{"key":"3333_CR8","doi-asserted-by":"crossref","unstructured":"Zama Ramirez P, Poggi M, Tosi F, Mattoccia S, Di Stefano L (2019) Geometry meets semantics for semi-supervised monocular depth estimation. In Computer vision\u2013ACCV 2018: 14th Asian conference on computer vision, Perth, Australia, December 2\u20136, 2018, revised selected papers, Part III 14, pp 298\u2013313. Springer","DOI":"10.1007\/978-3-030-20893-6_19"},{"key":"3333_CR9","doi-asserted-by":"crossref","unstructured":"Zhou T, Brown M, Snavely N, Lowe DG (2017) Unsupervised learning of depth and ego-motion from video. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1851\u20131858","DOI":"10.1109\/CVPR.2017.700"},{"key":"3333_CR10","doi-asserted-by":"crossref","unstructured":"Godard C, Mac Aodha O, Firman M, Brostow GJ (2019) Digging into self-supervised monocular depth estimation. In Proceedings of the IEEE\/CVF international conference on computer vision, pp 3828\u20133838","DOI":"10.1109\/ICCV.2019.00393"},{"key":"3333_CR11","unstructured":"Bian J, Li Z, Wang N, Zhan H, Shen C, Cheng M-M, Reid I (2019) Unsupervised scale-consistent depth and ego-motion learning from monocular video. Adv Neural Inf Process Syst 32"},{"key":"3333_CR12","doi-asserted-by":"publisher","first-page":"102058","DOI":"10.1016\/j.media.2021.102058","volume":"71","author":"KB Ozyoruk","year":"2021","unstructured":"Ozyoruk KB, Gokceler GI, Bobrow TL, Coskun G, Incetan K, Almalioglu Y, Mahmood F, Curto E, Perdigoto L, Oliveira M et al (2021) Endoslam dataset and an unsupervised monocular visual odometry and depth estimation approach for endoscopic videos. Med Image Anal 71:102058","journal-title":"Med Image Anal"},{"key":"3333_CR13","doi-asserted-by":"crossref","unstructured":"Croitoru F-A, Hondru V, Ionescu RT, Shah M (2023) Diffusion models in vision: a survey. IEEE transactions on pattern analysis and machine intelligence","DOI":"10.1109\/TPAMI.2023.3261988"},{"key":"3333_CR14","unstructured":"Shao S, Pei Z, Chen W, Sun D, Chen PC, Li Z (2023) Monodiffusion: self-supervised monocular depth estimation using diffusion model. arXiv preprint arXiv:2311.07198"},{"issue":"8","key":"3333_CR15","doi-asserted-by":"publisher","first-page":"1982","DOI":"10.1109\/TMM.2019.2895292","volume":"21","author":"X Guo","year":"2019","unstructured":"Guo X, Nie R, Cao J, Zhou D, Mei L, He K (2019) Fusegan: learning to fuse multi-focus image via conditional generative adversarial network. IEEE Trans Multimed 21(8):1982\u20131996","journal-title":"IEEE Trans Multimed"},{"key":"3333_CR16","doi-asserted-by":"publisher","first-page":"102338","DOI":"10.1016\/j.media.2021.102338","volume":"77","author":"S Shao","year":"2022","unstructured":"Shao S, Pei Z, Chen W, Zhu W, Wu X, Sun D, Zhang B (2022) Self-supervised monocular depth and ego-motion estimation in endoscopy: appearance flow to the rescue. Med Image Anal 77:102338","journal-title":"Med Image Anal"},{"key":"3333_CR17","doi-asserted-by":"crossref","unstructured":"Meng C, Rombach R, Gao R, Kingma D, Ermon S, Ho J, Salimans T (2023) On distillation of guided diffusion models. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14297\u201314306","DOI":"10.1109\/CVPR52729.2023.01374"},{"key":"3333_CR18","unstructured":"Yang R, Yang Y, Zhou F, Sun Q (2024) Directional diffusion models for graph representation learning. Adv Neural Inf Process Syst 36"},{"key":"3333_CR19","doi-asserted-by":"crossref","unstructured":"Zhao W, Liu S, Shu Y, Liu Y-J (2020) Towards better generalization: joint depth-pose learning without posenet. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 9151\u20139161","DOI":"10.1109\/CVPR42600.2020.00917"},{"key":"3333_CR20","doi-asserted-by":"crossref","unstructured":"Liu Z, Li R, Shao S, Wu X, Chen W (2023) Self-supervised monocular depth estimation with self-reference distillation and disparity offset refinement. IEEE transactions on circuits and systems for video technology","DOI":"10.1109\/TCSVT.2023.3275584"},{"key":"3333_CR21","doi-asserted-by":"crossref","unstructured":"Cao H, Tan C, Gao Z, Xu Y, Chen G, Heng P-A, Li SZ (2024) A survey on generative diffusion models. IEEE transactions on knowledge and data engineering","DOI":"10.1109\/TKDE.2024.3361474"},{"key":"3333_CR22","doi-asserted-by":"crossref","unstructured":"Zhang N, Nex F, Vosselman G, Kerle N (2023) Lite-mono: a lightweight cnn and transformer architecture for self-supervised monocular depth estimation. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 18537\u201318546","DOI":"10.1109\/CVPR52729.2023.01778"},{"key":"3333_CR23","doi-asserted-by":"crossref","unstructured":"Bergmann P, Fauser M, Sattlegger D, Steger C (2020) Uninformed students: student-teacher anomaly detection with discriminative latent embeddings. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 4183\u20134192","DOI":"10.1109\/CVPR42600.2020.00424"},{"key":"3333_CR24","unstructured":"Allan M, Mcleod J, Wang C, Rosenthal JC, Hu Z, Gard N, Eisert P, Fu KX, Zeffiro T, Xia W, et al. (2021) Stereo correspondence and reconstruction of endoscopic data challenge. arXiv preprint arXiv:2101.01133"},{"key":"3333_CR25","doi-asserted-by":"crossref","unstructured":"Behley J, Milioto A, Stachniss C (2021) A benchmark for lidar-based panoptic segmentation based on kitti. In: 2021 IEEE international conference on robotics and automation (ICRA), IEEE, pp 13596\u201313603","DOI":"10.1109\/ICRA48506.2021.9561476"},{"issue":"3","key":"3333_CR26","doi-asserted-by":"publisher","first-page":"505","DOI":"10.1016\/j.jksuci.2020.03.007","volume":"34","author":"A Shah","year":"2022","unstructured":"Shah A, Bangash JI, Khan AW, Ahmed I, Khan A, Khan A, Khan A (2022) Comparative analysis of median filter and its variants for removal of impulse noise from gray scale images. J King Saud Univ-Comput Inf Sci 34(3):505\u2013519","journal-title":"J King Saud Univ-Comput Inf Sci"},{"key":"3333_CR27","doi-asserted-by":"crossref","unstructured":"Fang Z, Chen X, Chen Y, Gool LV (2020) Towards good practice for cnn-based monocular depth estimation. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 1091\u20131100","DOI":"10.1109\/WACV45572.2020.9093334"},{"key":"3333_CR28","doi-asserted-by":"crossref","unstructured":"Spencer J, Bowden R, Hadfield S (2020) Defeat-net: general monocular depth via simultaneous unsupervised representation learning. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 14402\u201314413","DOI":"10.1109\/CVPR42600.2020.01441"},{"key":"3333_CR29","unstructured":"Oquab M, Darcet T, Moutakanni T, Vo H, Szafraniec M, Khalidov V, Fernandez P, Haziza D, Massa F, El-Nouby A, et al. (2023) Dinov2: learning robust visual features without supervision. arXiv preprint arXiv:2304.07193"},{"key":"3333_CR30","doi-asserted-by":"crossref","unstructured":"Cui B, Islam M, Bai L, Ren H (2024) Surgical-dino: adapter learning of foundation models for depth estimation in endoscopic surgery. Int J Comput Assist Radiol Surg 1\u20138","DOI":"10.1007\/s11548-024-03083-5"}],"container-title":["International Journal of Computer Assisted Radiology and Surgery"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-025-03333-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11548-025-03333-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11548-025-03333-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,6]],"date-time":"2025-05-06T12:51:38Z","timestamp":1746535898000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11548-025-03333-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,2,24]]},"references-count":30,"journal-issue":{"issue":"5","published-online":{"date-parts":[[2025,5]]}},"alternative-id":["3333"],"URL":"https:\/\/doi.org\/10.1007\/s11548-025-03333-0","relation":{},"ISSN":["1861-6429"],"issn-type":[{"value":"1861-6429","type":"electronic"}],"subject":[],"published":{"date-parts":[[2025,2,24]]},"assertion":[{"value":"16 June 2024","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 February 2025","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"24 February 2025","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}},{"value":"This article does not include any studies involving human participants or animals conducted by the authors. Additionally, it does not contain any patient data.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval and consent to participate"}}]}}