{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,11]],"date-time":"2025-09-11T19:20:52Z","timestamp":1757618452871,"version":"3.44.0"},"reference-count":38,"publisher":"Springer Science and Business Media LLC","issue":"10","license":[{"start":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T00:00:00Z","timestamp":1750896000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T00:00:00Z","timestamp":1750896000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100013061","name":"Jilin Provincial Scientific and Technological Development Program","doi-asserted-by":"publisher","award":["20210201027GX","20220101104JC"],"award-info":[{"award-number":["20210201027GX","20220101104JC"]}],"id":[{"id":"10.13039\/501100013061","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["62271226"],"award-info":[{"award-number":["62271226"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,10]]},"DOI":"10.1007\/s11760-025-04389-w","type":"journal-article","created":{"date-parts":[[2025,6,26]],"date-time":"2025-06-26T10:37:10Z","timestamp":1750934230000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["HADepth: Highlight-aware monocular depth estimation for endoscopy"],"prefix":"10.1007","volume":"19","author":[{"given":"Xiaoyuan","family":"Peng","sequence":"first","affiliation":[]},{"given":"Shigang","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Jian","family":"Wei","sequence":"additional","affiliation":[]},{"given":"Yan","family":"Zhao","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,6,26]]},"reference":[{"issue":"1","key":"4389_CR1","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11760-024-03720-1","volume":"19","author":"I Ardiyanto","year":"2025","unstructured":"Ardiyanto, I., Al-Fahsi, R.D.H.: Lightweight monocular depth estimation network for robotics using intercept block ghostnet. Signal, Image and Video Processing 19(1), 1\u201314 (2025)","journal-title":"Signal, Image and Video Processing"},{"issue":"6","key":"4389_CR2","doi-asserted-by":"publisher","first-page":"1013","DOI":"10.1007\/s11548-024-03083-5","volume":"19","author":"B Cui","year":"2024","unstructured":"Cui, B., Islam, M., Bai, L., Ren, H.: Surgical-dino: adapter learning of foundation models for depth estimation in endoscopic surgery. Int. J. Comput. Assist. Radiol. Surg. 19(6), 1013\u20131020 (2024)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"4389_CR3","doi-asserted-by":"crossref","unstructured":"Cui, B., Islam, M., Bai, L., Wang, A., Ren, H.: Endodac: Efficient adapting foundation model for self-supervised depth estimation from any endoscopic camera, in: Proceedings of International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), pp. 208\u2013218 (2024)","DOI":"10.1007\/978-3-031-72089-5_20"},{"key":"4389_CR4","doi-asserted-by":"crossref","unstructured":"Wang, Y., Li, Y., Wang, G., Liu, X.: Multi-scale attention network for single image super-resolution, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 5950\u20135960 (2024)","DOI":"10.1109\/CVPRW63382.2024.00602"},{"key":"4389_CR5","unstructured":"Hu, E.\u00a0J., Shen, Y., Wallis, P., Allen-Zhu, Z., Li, Y., Wang, S., Wang, L., Chen, W. et\u00a0al.: Lora: Low-rank adaptation of large language models., in: Proceedings of International Conference on Learning Representations (ICLR), pp. 1\u201320 (2022)"},{"key":"4389_CR6","unstructured":"Si, C., Wang, X., Yang, X., Xu, Z., Li, Q., Dai, J., Qiao, Y., Yang, X., Shen, W.: Flora: Low-rank core space for n-dimension, arXiv preprint arXiv:2405.14739 (2024)"},{"key":"4389_CR7","unstructured":"Liu, S.-Y., Wang, C.-Y., Yin, H., Molchanov, P., Wang, Y.-C.\u00a0F., Cheng, K.-T., Chen, M.-H.: Dora: Weight-decomposed low-rank adaptation, in: Proceedings of International Conference on Machine Learning (ICML), pp. 1\u201322 (2024)"},{"key":"4389_CR8","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network, in: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), pp. 2366\u20132374 (2014)"},{"key":"4389_CR9","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac\u00a0Aodha, O., Firman, M., Brostow, G.\u00a0J.: Digging into self-supervised monocular depth estimation, in: Proceedings of the IEEE International Conference on Computer Vision (ICCV), pp. 3828\u20133838 (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"4389_CR10","doi-asserted-by":"crossref","unstructured":"Watson, J., Mac\u00a0Aodha, O., Prisacariu, V., Brostow, G., Firman, M.: The temporal opportunist: Self-supervised multi-frame monocular depth, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1164\u20131174 (2021)","DOI":"10.1109\/CVPR46437.2021.00122"},{"issue":"4","key":"4389_CR11","doi-asserted-by":"publisher","first-page":"2564","DOI":"10.1109\/TCSVT.2023.3305776","volume":"34","author":"X Miao","year":"2023","unstructured":"Miao, X., Bai, Y., Duan, H., Huang, Y., Wan, F., Xu, X., Long, Y., Zheng, Y.: Ds-depth: Dynamic and static depth estimation via a fusion cost volume. IEEE Trans. Circuits Syst. Video Technol. 34(4), 2564\u20132576 (2023)","journal-title":"IEEE Trans. Circuits Syst. Video Technol."},{"key":"4389_CR12","doi-asserted-by":"crossref","unstructured":"Zhang, N., Nex, F., Vosselman, G., Kerle, N.: Lite-mono: A lightweight cnn and transformer architecture for self-supervised monocular depth estimation, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 18537\u201318546 (2023)","DOI":"10.1109\/CVPR52729.2023.01778"},{"key":"4389_CR13","doi-asserted-by":"publisher","DOI":"10.1016\/j.measurement.2022.110742","volume":"190","author":"Q Sheng","year":"2022","unstructured":"Sheng, Q., Zheng, J., Shi, W., Zhao, R., Liu, J., Li, H.: Measurement and modeling of reflection characteristics of hole inner surface based on endoscopic image. Measurement 190, 110742 (2022)","journal-title":"Measurement"},{"key":"4389_CR14","doi-asserted-by":"crossref","unstructured":"Turan, M., Ornek, E.\u00a0P., Ibrahimli, N., Giracoglu, C., Almalioglu, Y., Yanik, M.\u00a0F., Sitti, M.: Unsupervised odometry and depth learning for endoscopic capsule robots, in: Proceedings of IEEE International Conference on Intelligent Robots and Systems (IROS), pp. 1801\u20131807 (2018)","DOI":"10.1109\/IROS.2018.8593623"},{"issue":"4","key":"4389_CR15","doi-asserted-by":"publisher","first-page":"7225","DOI":"10.1109\/LRA.2021.3095528","volume":"6","author":"D Recasens","year":"2021","unstructured":"Recasens, D., Lamarca, J., F\u00e1cil, J.M., Montiel, J., Civera, J.: Endo-depth-and-motion: Reconstruction and tracking in endoscopic videos using depth networks and photometric constraints. IEEE Robot. Autom. Lett. 6(4), 7225\u20137232 (2021)","journal-title":"IEEE Robot. Autom. Lett."},{"key":"4389_CR16","doi-asserted-by":"publisher","DOI":"10.1016\/j.engappai.2023.105989","volume":"122","author":"Y Yang","year":"2023","unstructured":"Yang, Y., Shao, S., Yang, T., Wang, P., Yang, Z., Wu, C., Liu, H.: A geometry-aware deep network for depth estimation in monocular endoscopy. Eng. Appl. Artif. Intell. 122, 105989 (2023)","journal-title":"Eng. Appl. Artif. Intell."},{"key":"4389_CR17","doi-asserted-by":"crossref","unstructured":"Li, Y.: Endodepthl: Lightweight endoscopic monocular depth estimation with cnn-transformer, in: Proceedings of IEEE International Conference on Bioinformatics and Biomedicine (BIBM), pp. 4344\u20134351 (2023)","DOI":"10.1109\/BIBM58861.2023.10386008"},{"key":"4389_CR18","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102338","volume":"77","author":"S Shao","year":"2022","unstructured":"Shao, S., Pei, Z., Chen, W., Zhu, W., Wu, X., Sun, D., Zhang, B.: Self-supervised monocular depth and ego-motion estimation in endoscopy: Appearance flow to the rescue. Med. Image Anal. 77, 102338 (2022)","journal-title":"Med. Image Anal."},{"key":"4389_CR19","doi-asserted-by":"crossref","unstructured":"Yang, L., Kang, B., Huang, Z., Xu, X., Feng, J., Zhao, H.: Depth anything: Unleashing the power of large-scale unlabeled data, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 10371\u201310381 (2024)","DOI":"10.1109\/CVPR52733.2024.00987"},{"key":"4389_CR20","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1155\/2010\/814319","volume":"2010","author":"M Arnold","year":"2010","unstructured":"Arnold, M., Ghosh, A., Ameling, S., Lacey, G.: Automatic segmentation and inpainting of specular highlights for endoscopic imaging. EURASIP Journal on Image and Video Processing 2010, 1\u201312 (2010)","journal-title":"EURASIP Journal on Image and Video Processing"},{"issue":"6","key":"4389_CR21","doi-asserted-by":"publisher","first-page":"2079","DOI":"10.3390\/s21062079","volume":"21","author":"S Kim","year":"2021","unstructured":"Kim, S., Ra, M., Kim, W.-Y.: Specular detection on glossy surface using geometric characteristics of specularity in top-view images. Sensors 21(6), 2079 (2021)","journal-title":"Sensors"},{"issue":"4","key":"4389_CR22","doi-asserted-by":"publisher","first-page":"1954","DOI":"10.1109\/TIP.2018.2880088","volume":"28","author":"Y Chang","year":"2018","unstructured":"Chang, Y., Jung, C.: Single image reflection removal using convolutional neural networks. IEEE Trans. Image Process. 28(4), 1954\u20131966 (2018)","journal-title":"IEEE Trans. Image Process."},{"issue":"6","key":"4389_CR23","doi-asserted-by":"publisher","first-page":"2869","DOI":"10.1364\/BOE.10.002869","volume":"10","author":"TL Bobrow","year":"2019","unstructured":"Bobrow, T.L., Mahmood, F., Inserni, M., Durr, N.J.: Deeplsr: a deep learning approach for laser speckle reduction. Biomed. Opt. Express 10(6), 2869\u20132882 (2019)","journal-title":"Biomed. Opt. Express"},{"key":"4389_CR24","doi-asserted-by":"crossref","unstructured":"Funke, I., Bodenstedt, S., Riediger, C., Weitz, J., Speidel, S.: Generative adversarial networks for specular highlight removal in endoscopic images, in: Proceedings of Medical Imaging: Image-Guided Procedures, Robotic Interventions, and Modeling, Vol. 10576, pp. 8\u201316 (2018)","DOI":"10.1117\/12.2293755"},{"issue":"3","key":"4389_CR25","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/1970392.1970395","volume":"58","author":"EJ Cand\u00e8s","year":"2011","unstructured":"Cand\u00e8s, E.J., Li, X., Ma, Y., Wright, J.: Robust principal component analysis? Journal of the ACM 58(3), 1\u201337 (2011)","journal-title":"Journal of the ACM"},{"key":"4389_CR26","doi-asserted-by":"publisher","first-page":"360","DOI":"10.1109\/JTEHM.2023.3283444","volume":"11","author":"J Joseph","year":"2023","unstructured":"Joseph, J., George, S.N., Raja, K.: Parameter-free matrix decomposition for specular reflections removal in endoscopic images. IEEE J. Transl. Eng. Health Med. 11, 360\u2013374 (2023)","journal-title":"IEEE J. Transl. Eng. Health Med."},{"key":"4389_CR27","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"4389_CR28","unstructured":"Vaswani, A., Shazeer, N., Parmar, N., Uszkoreit, J., Jones, L., Gomez, A.\u00a0N., Kaiser, \u0141., Polosukhin, I.: Attention is all you need, in: Proceedings of Advances in Neural Information Processing Systems (NeurIPS), pp. 6000\u20136010 (2017)"},{"issue":"4","key":"4389_CR29","doi-asserted-by":"publisher","first-page":"1956","DOI":"10.1137\/080738970","volume":"20","author":"J Cai","year":"2010","unstructured":"Cai, J., Candes, E.J., Shen, Z.: A singular value thresholding algorithm for matrix completion. Siam Journal on Optimization 20(4), 1956\u20131982 (2010)","journal-title":"Siam Journal on Optimization"},{"key":"4389_CR30","volume-title":"Image intrinsic-based unsupervised monocular depth estimation in endoscopy","author":"B Li","year":"2024","unstructured":"Li, B., Liu, B., Zhu, M., Luo, X., Zhou, F.: Image intrinsic-based unsupervised monocular depth estimation in endoscopy. IEEE J. Biomed, Health Inform (2024)"},{"key":"4389_CR31","doi-asserted-by":"publisher","DOI":"10.1016\/j.media.2021.102058","volume":"71","author":"KB Ozyoruk","year":"2021","unstructured":"Ozyoruk, K.B., Gokceler, G.I., Bobrow, T.L., Coskun, G., Incetan, K., et al.: Endoslam dataset and an unsupervised monocular visual odometry and depth estimation approach for endoscopic videos. Med. Image Anal. 71, 102058 (2021)","journal-title":"Med. Image Anal."},{"key":"4389_CR32","unstructured":"Allan, M., Mcleod, J., Wang, C., Rosenthal, J.\u00a0C., Hu, Z., Gard, N., Eisert, P., Fu, K.\u00a0X., Zeffiro, T., Xia, W., et\u00a0al., Stereo correspondence and reconstruction of endoscopic data challenge, arXiv preprintarXiv:2101.01133 (2021)"},{"key":"4389_CR33","doi-asserted-by":"publisher","first-page":"1167","DOI":"10.1007\/s11548-019-01962-w","volume":"14","author":"A Rau","year":"2019","unstructured":"Rau, A., Edwards, P.E., Ahmad, O.F., Riordan, P., Janatka, M., Lovat, L.B., Stoyanov, D.: Implicit domain adaptation with conditional generative adversarial networks for depth prediction in endoscopy. Int. J. Comput. Assist. Radiol. Surg. 14, 1167\u20131176 (2019)","journal-title":"Int. J. Comput. Assist. Radiol. Surg."},{"key":"4389_CR34","doi-asserted-by":"crossref","unstructured":"Zhou, T., Brown, M., Snavely, N., Lowe, D.\u00a0G.: Unsupervised learning of depth and ego-motion from video, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1851\u20131858 (2017)","DOI":"10.1109\/CVPR.2017.700"},{"issue":"1","key":"4389_CR35","doi-asserted-by":"publisher","first-page":"11","DOI":"10.1038\/s41598-024-84695-4","volume":"15","author":"J Cai","year":"2025","unstructured":"Cai, J., Chen, T., Qi, Y., Liu, S., Chen, R.: Fibrosis and inflammatory activity diagnosis of chronic hepatitis c based on extreme learning machine. Sci. Rep. 15(1), 11 (2025)","journal-title":"Sci. Rep."},{"key":"4389_CR36","doi-asserted-by":"crossref","unstructured":"Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735\u20131780 (1997)","DOI":"10.1162\/neco.1997.9.8.1735"},{"key":"4389_CR37","doi-asserted-by":"crossref","unstructured":"Nguyen, A., Do, T.-T., Caldwell, D.\u00a0G., Tsagarakis, N.\u00a0G.: Real-time 6dof pose relocalization for event cameras with stacked spatial lstm networks, in: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1638\u20131645 (2019)","DOI":"10.1109\/CVPRW.2019.00207"},{"key":"4389_CR38","doi-asserted-by":"publisher","DOI":"10.1109\/JBHI.2024.3384333","volume-title":"Developing deep lstms with later temporal attention for predicting covid-19 severity, clinical outcome, and antibody level by screening serological indicators over time","author":"J Cai","year":"2024","unstructured":"Cai, J., Li, Y., Liu, B., Wu, Z., Zhu, S., Chen, Q., Lei, Q., Hou, H., Guo, Z., Jiang, H., et al.: Developing deep lstms with later temporal attention for predicting covid-19 severity, clinical outcome, and antibody level by screening serological indicators over time. IEEE J. Biomed, Health Inform (2024)"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04389-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04389-w\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04389-w.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,6]],"date-time":"2025-09-06T22:52:10Z","timestamp":1757199130000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04389-w"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,6,26]]},"references-count":38,"journal-issue":{"issue":"10","published-print":{"date-parts":[[2025,10]]}},"alternative-id":["4389"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04389-w","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2025,6,26]]},"assertion":[{"value":"20 February 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 June 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"12 June 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 June 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare no conflict of interest.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"810"}}