{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,13]],"date-time":"2025-12-13T07:20:42Z","timestamp":1765610442437,"version":"3.40.2"},"reference-count":46,"publisher":"Springer Science and Business Media LLC","issue":"8","license":[{"start":{"date-parts":[[2024,3,25]],"date-time":"2024-03-25T00:00:00Z","timestamp":1711324800000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,3,25]],"date-time":"2024-03-25T00:00:00Z","timestamp":1711324800000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"the National Natural Science Foundation of China under Grant","award":["No. 52204177","No.52304182"],"award-info":[{"award-number":["No. 52204177","No.52304182"]}]},{"name":"the Assistance Program for Future Outstanding Talents of China University of Mining and Technology","award":["No. 2022WLJCRCZL105"],"award-info":[{"award-number":["No. 2022WLJCRCZL105"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-18877-7","type":"journal-article","created":{"date-parts":[[2024,3,25]],"date-time":"2024-03-25T06:02:11Z","timestamp":1711346531000},"page":"4773-4787","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":4,"title":["DCL-depth: monocular depth estimation network based on iam and depth consistency loss"],"prefix":"10.1007","volume":"84","author":[{"given":"Chenggong","family":"Han","sequence":"first","affiliation":[]},{"given":"Chen","family":"Lv","sequence":"additional","affiliation":[]},{"given":"Qiqi","family":"Kou","sequence":"additional","affiliation":[]},{"given":"He","family":"Jiang","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0001-8831-1994","authenticated-orcid":false,"given":"Deqiang","family":"Cheng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,25]]},"reference":[{"key":"18877_CR1","doi-asserted-by":"publisher","first-page":"104261","DOI":"10.1016\/j.imavis.2021.104261","volume":"113","author":"U Ali","year":"2021","unstructured":"Ali U, Bayramli B, Alsarhan T et al (2021) A lightweight network for monocular depth estimation with decoupled body and edge supervision[J]. Image Vis Comput 113:104261","journal-title":"Image Vis Comput"},{"key":"18877_CR2","doi-asserted-by":"publisher","first-page":"103934","DOI":"10.1016\/j.imavis.2020.103934","volume":"100","author":"A Mathew","year":"2020","unstructured":"Mathew A, Mathew J (2020) Monocular depth estimation with SPN loss[J]. Image Vis Comput 100:103934","journal-title":"Image Vis Comput"},{"key":"18877_CR3","doi-asserted-by":"crossref","unstructured":"Bay H, Tuytelaars T, Gool LV (2006) Surf: speeded up robust features[C]. In: Proceedings of the European conference on computer vision, pp 404\u2013417","DOI":"10.1007\/11744023_32"},{"key":"18877_CR4","doi-asserted-by":"crossref","unstructured":"Yu Z, Jin L, Gao S (2020) P2net: Patch-match and plane-regularization for unsupervised indoor depth estimation[C]. In: European conference on computer vision. Springer, pp 206\u2013222","DOI":"10.1007\/978-3-030-58586-0_13"},{"key":"18877_CR5","doi-asserted-by":"crossref","unstructured":"Li B, Huang Y, Liu Z, Zou D, Yu W (2021) StructDepth: leveraging the structural regularities for self-supervised indoor depth estimation[C]. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 12663\u201312673","DOI":"10.1109\/ICCV48922.2021.01243"},{"key":"18877_CR6","doi-asserted-by":"crossref","unstructured":"Patil V, Sakaridis C, Liniger A, Van Gool L (2022) P3Depth: monocular depth estimation with a piecewise planarity prior[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1610\u20131621","DOI":"10.1109\/CVPR52688.2022.00166"},{"issue":"4","key":"18877_CR7","doi-asserted-by":"publisher","first-page":"1889","DOI":"10.1109\/TCSVT.2021.3086598","volume":"32","author":"L Chen","year":"2021","unstructured":"Chen L, Guo L, Cheng D et al (2021) Structure-preserving and color-restoring up-sampling for single low-light image[J]. IEEE Trans Circuits Syst Video Technol 32(4):1889\u20131902","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"18877_CR8","unstructured":"Yang H, Lin W, Deng C (2012) Image activity measure (IAM) for screen image segmentation[C]. In: Proc IEEE Int Conf Image Process, pp 1569\u20131572"},{"key":"18877_CR9","doi-asserted-by":"crossref","unstructured":"Silberman N, Hoiem D, Kohli P, Fergus R (2012) Indoor segmentation and support inference from rgbd images[C]. In: Proceedings of the European conference on computer vision. Springer, pp 746\u2013760","DOI":"10.1007\/978-3-642-33715-4_54"},{"key":"18877_CR10","doi-asserted-by":"crossref","unstructured":"Dai A, Chang AX, Savva M, Halber M, Funkhouser T, Nie\u00dfner M (2017) Scannet: Richly-annotated 3d reconstructions of indoor scenes. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5828\u20135839","DOI":"10.1109\/CVPR.2017.261"},{"key":"18877_CR11","doi-asserted-by":"publisher","first-page":"379","DOI":"10.1109\/LSP.2021.3055116","volume":"28","author":"J Li","year":"2021","unstructured":"Li J, Cheng D, Liu R et al (2021) Unsupervised person re-identification based on measurement Axis[J]. IEEE Signal Process Lett 28:379\u2013383","journal-title":"IEEE Signal Process Lett"},{"key":"18877_CR12","doi-asserted-by":"crossref","unstructured":"Cao Z, Kar A, H\u00e4ne C, Malik J (2019) Learning independent object motion from unlabelled stereoscopic videos[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5594\u20135603","DOI":"10.1109\/CVPR.2019.00574"},{"key":"18877_CR13","doi-asserted-by":"crossref","unstructured":"Zhou T, Brown M, Snavely N, Lowe DG (2017) Unsupervised learning of depth and ego-motion from video[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1851\u20131858","DOI":"10.1109\/CVPR.2017.700"},{"key":"18877_CR14","doi-asserted-by":"crossref","unstructured":"Janai J, G\u00fcney F, Ranjan A, Black M, Geiger A (2018) Unsupervised learning of multi-frame optical flow with occlusions[C]. In: Proceedings of the European conference on computer vision, pp 690\u2013706","DOI":"10.1007\/978-3-030-01270-0_42"},{"key":"18877_CR15","doi-asserted-by":"crossref","unstructured":"Klingner M, Termhlen JA, Mikolajczyk J, Fingscheidt T (2020) Self-supervised monocular depth estimation: solving the dynamic object problem by semantic guidance[J]. In: European conference on computer vision. Springer, pp 582\u2013600","DOI":"10.1007\/978-3-030-58565-5_35"},{"key":"18877_CR16","doi-asserted-by":"crossref","unstructured":"Godard C, Aodha OM, Firman M, Brostow GJ (2019) Digging into self-supervised monocular depth estimation[C]. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 3828\u20133838","DOI":"10.1109\/ICCV.2019.00393"},{"key":"18877_CR17","doi-asserted-by":"crossref","unstructured":"Geiger A, Lenz P, Urtasun R (2012) Are we ready for autonomous driving? the kitti vision benchmark suite[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3354\u20133361","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"18877_CR18","doi-asserted-by":"crossref","unstructured":"Cordts M, Omran M, Ramos S, Rehfeld T, Enzweiler M, Benenson R, Schiele B (2016) The cityscapes dataset for semantic urban scene understanding[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3213\u20133223","DOI":"10.1109\/CVPR.2016.350"},{"key":"18877_CR19","doi-asserted-by":"crossref","unstructured":"Wu C-Y, Wang J, Hall M, Neumann U, Su S (2022) Toward practical monocular indoor depth estimation[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3814\u20133824","DOI":"10.1109\/CVPR52688.2022.00379"},{"issue":"24","key":"18877_CR20","doi-asserted-by":"publisher","first-page":"38035","DOI":"10.1007\/s11042-022-14012-6","volume":"82","author":"C Han","year":"2023","unstructured":"Han C, Cheng D, Kou Q et al (2023) Self-supervised monocular Depth estimation with multi-scale structure similarity loss[J]. Multimed Tools Appl 82(24):38035\u201338050","journal-title":"Multimed Tools Appl"},{"issue":"12","key":"18877_CR21","doi-asserted-by":"publisher","first-page":"8436","DOI":"10.1109\/TCSVT.2022.3194169","volume":"32","author":"D Cheng","year":"2022","unstructured":"Cheng D, Chen L, Lv C et al (2022) Light-guided and cross-fusion U-Net for anti-illumination image super-resolution[J]. IEEE Trans Circuits Syst Video Technol 32(12):8436\u20138449","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"key":"18877_CR22","doi-asserted-by":"crossref","unstructured":"Zhou J, Wang Y, Qin K, Zeng W (2019) Moving indoor: unsupervised video depth learning in challenging environments[C]. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 8618\u20138627","DOI":"10.1109\/ICCV.2019.00871"},{"key":"18877_CR23","doi-asserted-by":"crossref","unstructured":"Zhao W, Liu S, Shu Y, Liu YJ (2020) Towards better generalization: joint depth-pose learning without posenet[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 9151\u20139161","DOI":"10.1109\/CVPR42600.2020.00917"},{"key":"18877_CR24","unstructured":"Bian J-W, Zhan H, Wang N, Chin TJ, Shen C, Reid I (2020) Unsupervised depth learning in challenging indoor video: Weak rectification to rescue[J]. arXiv preprint arXiv:2006.02708 2(5):7"},{"key":"18877_CR25","doi-asserted-by":"crossref","unstructured":"Ji P, Li R, Bhanu B, Xu Y (2021) Monoindoor: towards good practice of self-supervised monocular depth estimation for indoor environments. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 12787\u201312796","DOI":"10.1109\/ICCV48922.2021.01255"},{"issue":"5","key":"18877_CR26","doi-asserted-by":"publisher","first-page":"104","DOI":"10.1109\/97.841153","volume":"7","author":"S Saha","year":"2000","unstructured":"Saha S, Vemuri R (2000) An analysis on the effect of image features on lossy coding performance[J]. IEEE Signal Process Lett 7(5):104\u2013107","journal-title":"IEEE Signal Process Lett"},{"key":"18877_CR27","doi-asserted-by":"publisher","first-page":"104267","DOI":"10.1016\/j.imavis.2021.104267","volume":"114","author":"A Dc","year":"2021","unstructured":"Dc A, Rl A, Jl A et al (2021) Activity guided multi-scales collaboration based on scaled-CNN for saliency prediction[J]. Image Vis Comput 114:104267","journal-title":"Image Vis Comput"},{"key":"18877_CR28","doi-asserted-by":"crossref","unstructured":"Ghiasi G, Lee H, Kudlur M, Dumoulin V, Shlens J (2017) Exploring the structure of a real-time, arbitrary neural artistic stylization network[J]. arXiv preprint arXiv:1705.06830","DOI":"10.5244\/C.31.114"},{"key":"18877_CR29","doi-asserted-by":"crossref","unstructured":"Liu L, Song X, Wang M, Liu Y, Zhang L (2021) Self-supervised monocular depth estimation for all day images using domain separation[C]. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 12737\u201312746","DOI":"10.1109\/ICCV48922.2021.01250"},{"key":"18877_CR30","first-page":"1398","volume":"2","author":"Z Wang","year":"2003","unstructured":"Wang Z, Simoncelli EP, Bovik AC (2003) Multiscale structural similarity for image quality assessment[C]. Proc IEEE Asilomar Conference on Signals 2:1398\u20131402","journal-title":"Proc IEEE Asilomar Conference on Signals"},{"key":"18877_CR31","doi-asserted-by":"crossref","unstructured":"Godard C, Aodha OMGJ (2017) Unsupervised monocular depth estimation with left-right consistency[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 270\u2013279","DOI":"10.1109\/CVPR.2017.699"},{"issue":"5","key":"18877_CR32","doi-asserted-by":"publisher","first-page":"824","DOI":"10.1109\/TPAMI.2008.132","volume":"31","author":"A Saxena","year":"2008","unstructured":"Saxena A, Sun M, Ng AY (2008) Make3d: Learning 3d scene structure from a single still image[J]. IEEE Trans Pattern Anal Mach Intell 31(5):824\u2013840","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"issue":"11","key":"18877_CR33","doi-asserted-by":"publisher","first-page":"2144","DOI":"10.1109\/TPAMI.2014.2316835","volume":"36","author":"K Karsch","year":"2014","unstructured":"Karsch K, Liu C, Kang SB (2014) Depth transfer: Depth extraction from video using non-parametric sampling[J]. IEEE Trans Pattern Anal Mach Intell 36(11):2144\u20132158","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"18877_CR34","doi-asserted-by":"crossref","unstructured":"Ladicky L, Shi J, Pollefeys M (2014) Pulling things out of perspective[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 89\u201396\u00a0","DOI":"10.1109\/CVPR.2014.19"},{"key":"18877_CR35","doi-asserted-by":"crossref","unstructured":"Li B, Shen C, Dai Y, Van Den, Hengel A, He M (2015) Depth and surface normal estimation from monocular images using regression on deep features and hierarchical crfs[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1119\u20131127","DOI":"10.1109\/CVPR.2015.7298715"},{"key":"18877_CR36","doi-asserted-by":"crossref","unstructured":"Wang P, Shen X, Lin Z, Cohen S, Price B, Yuille AL (2015) Towards unified depth and semantic prediction from a single image[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2800\u20132809","DOI":"10.1109\/CVPR.2015.7298897"},{"issue":"10","key":"18877_CR37","doi-asserted-by":"publisher","first-page":"2024","DOI":"10.1109\/TPAMI.2015.2505283","volume":"38","author":"F Liu","year":"2015","unstructured":"Liu F, Shen C, Lin G et al (2015) Learning depth from single monocular images using deep convolutional neural fields[J]. IEEE Trans Pattern Anal Mach Intell 38(10):2024\u20132039","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"18877_CR38","doi-asserted-by":"crossref","unstructured":"Li J, Klein R, Yao A (2017) A two-streamed network for estimating fine-scaled depth maps from single rgb images[C]. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 3372\u20133380","DOI":"10.1109\/ICCV.2017.365"},{"key":"18877_CR39","doi-asserted-by":"crossref","unstructured":"Xu D, Ricci E, Ouyang W, Wang X, Sebe N (2017) Multi-scale continuous crfs as sequential deep networks for monocular depth estimation[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 5354\u20135362","DOI":"10.1109\/CVPR.2017.25"},{"key":"18877_CR40","doi-asserted-by":"crossref","unstructured":"Fu H, Gong M, Wang C, Batmanghelich K, Tao D (2018) Deep ordinal regression network for monocular depth estimation[C]. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 2002\u20132011","DOI":"10.1109\/CVPR.2018.00214"},{"key":"18877_CR41","doi-asserted-by":"crossref","unstructured":"Hu J, Ozay M, Zhang Y, Okatani T (2019) Revisiting single image depth estimation: toward higher resolution maps with accurate object boundaries[C]. In: IEEE winter conference on applications of computer vision, pp 1043\u20131051","DOI":"10.1109\/WACV.2019.00116"},{"key":"18877_CR42","doi-asserted-by":"crossref","unstructured":"Yin W, Liu Y, Shen C, Yan Y (2019) Enforcing geometric constraints of virtual normal for depth prediction[C]. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 5684\u20135693","DOI":"10.1109\/ICCV.2019.00578"},{"issue":"6","key":"18877_CR43","doi-asserted-by":"publisher","first-page":"3839","DOI":"10.1109\/TCSVT.2021.3118681","volume":"32","author":"Y Wei","year":"2021","unstructured":"Wei Y, Guo H, Lu J et al (2021) Iterative feature matching for self-supervised indoor depth estimation[J]. IEEE Trans Circuits Syst Video Technol 32(6):3839\u20133852","journal-title":"IEEE Trans Circuits Syst Video Technol"},{"issue":"12","key":"18877_CR44","doi-asserted-by":"publisher","first-page":"9802","DOI":"10.1109\/TPAMI.2021.3136220","volume":"44","author":"JW Bian","year":"2021","unstructured":"Bian JW, Zhan H, Wang N et al (2021) Auto-rectify network for unsupervised indoor depth estimation[J]. IEEE Trans Pattern Anal Mach Intell 44(12):9802\u20139813","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"18877_CR45","doi-asserted-by":"crossref","unstructured":"Ma J, Lei X, Liu N, Zhao X, Pu S (2022) Towards comprehensive representation enhancement in semantics-guided self-supervised monocular depth estimation[C]. In: European conference on computer vision. Springer, pp 304\u2013321","DOI":"10.1007\/978-3-031-19769-7_18"},{"key":"18877_CR46","doi-asserted-by":"crossref","unstructured":"Qiao S, Xiong NN, Gao Y, Fang Z, Yu W, Zhang J, Jiang X (2023) Self-supervised learning of depth and ego-motion for 3D perception in human computer interaction[J]. ACM Trans Multimed Comput Commun Appl 1\u201321","DOI":"10.1145\/3588571"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-18877-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-18877-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-18877-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,3,23]],"date-time":"2025-03-23T00:16:38Z","timestamp":1742688998000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-18877-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,25]]},"references-count":46,"journal-issue":{"issue":"8","published-online":{"date-parts":[[2025,3]]}},"alternative-id":["18877"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-18877-7","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,3,25]]},"assertion":[{"value":"21 July 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 January 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"11 March 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"25 March 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interest"}}]}}