{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,23]],"date-time":"2025-09-23T03:10:44Z","timestamp":1758597044376,"version":"3.44.0"},"reference-count":48,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T00:00:00Z","timestamp":1757376000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T00:00:00Z","timestamp":1757376000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100015956","name":"Special Project for Research and Development in Key areas of Guangdong Province","doi-asserted-by":"publisher","award":["2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001"],"award-info":[{"award-number":["2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001","2022B0701180001"]}],"id":[{"id":"10.13039\/501100015956","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["SIViP"],"published-print":{"date-parts":[[2025,12]]},"DOI":"10.1007\/s11760-025-04733-0","type":"journal-article","created":{"date-parts":[[2025,9,9]],"date-time":"2025-09-09T00:45:35Z","timestamp":1757378735000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["FASNet: a high-precision, two-stage stereo matching network for real-time depth perception"],"prefix":"10.1007","volume":"19","author":[{"given":"Bifa","family":"Liang","sequence":"first","affiliation":[]},{"given":"Yichao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Ziyang","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Zhicong","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Haifeng","family":"Hu","sequence":"additional","affiliation":[]},{"given":"Jianming","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Dihu","family":"Chen","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2025,9,9]]},"reference":[{"key":"4733_CR1","doi-asserted-by":"crossref","unstructured":"Perri, S., et al.: Stereo vision architecture for heterogeneous systems-on-chip. J. Real-Time Image Proc. 393\u2013415 (2018)","DOI":"10.1007\/s11554-018-0782-z"},{"key":"4733_CR2","unstructured":"Lu, K., et al.: Binocular stereo vision based on OpenCV. IET International Conference on Smart and Sustainable City. IET, (2011)"},{"key":"4733_CR3","doi-asserted-by":"crossref","unstructured":"Brown, M.Z., Burschka, D., Hager, G.D.: Advances in computational stereo. IEEE Trans. Pattern Anal. Mach. Intell. 993\u20131008 (2003)","DOI":"10.1109\/TPAMI.2003.1217603"},{"key":"4733_CR4","unstructured":"Liang, Y., et al.: Research and implementation of adaptive stereo matching algorithm based on ZYNQ. J. Real-Time Image Proc. (2018)"},{"key":"4733_CR5","doi-asserted-by":"crossref","unstructured":"Kriegman, D.J., et al.: Stereo vision and navigation in buildings for mobile robots. IEEE Trans. Robot Autom. 792\u2013803 (1989)","DOI":"10.1109\/70.88100"},{"key":"4733_CR6","doi-asserted-by":"crossref","unstructured":"Dong, P., et al.: A 4.29 nJ\/pixel stereo depth coprocessor with pixel level pipeline and region optimized semi-global matching for IoT application. IEEE Trans. Circuits Syst. I Regul. Pap 334\u2013346 (2022)","DOI":"10.1109\/TCSI.2021.3100071"},{"key":"4733_CR7","unstructured":"Bertozzi, M., et al.: Stereo vision-based vehicle detection. In: Proceedings of the Proceedings of the IEEE Intelligent Vehicles Symposium (2000)"},{"key":"4733_CR8","doi-asserted-by":"crossref","unstructured":"Mayer, N., et al.: A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition. 4040\u20134048 (2016)","DOI":"10.1109\/CVPR.2016.438"},{"key":"4733_CR9","doi-asserted-by":"crossref","unstructured":"Zhou, T., Yang, M., Jiang, K., Wong, H.: and D. Yang.: MMW radar-based technologies in autonomous driving: A review. Sensors, 7283 (2020)","DOI":"10.3390\/s20247283"},{"key":"4733_CR10","doi-asserted-by":"publisher","first-page":"6775","DOI":"10.1109\/ACCESS.2022.3141587","volume":"10","author":"J Fuchs","year":"2022","unstructured":"Fuchs, J., et al.: A machine learning perspective on automotive radar direction of arrival Estimation. IEEE Access. 10, 6775\u20136797 (2022)","journal-title":"IEEE Access."},{"key":"4733_CR11","doi-asserted-by":"crossref","unstructured":"Tian, F., et al.: Depth Estimation using a self-supervised network based on cross-layer feature fusion and the quadtree constraint. IEEE Trans. Circuits Syst. Video Technol. 1751\u20131766 (2022)","DOI":"10.1109\/TCSVT.2021.3080928"},{"key":"4733_CR12","doi-asserted-by":"crossref","unstructured":"Poggi, M., et al.: On the synergies between machine learning and binocular stereo for depth Estimation from images: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 5314\u20135334 (2022)","DOI":"10.1109\/TPAMI.2021.3070917"},{"key":"4733_CR13","doi-asserted-by":"crossref","unstructured":"Pang, J., et al.: Cascade residual learning: A two-stage convolutional neural network for stereo matching. Proceedings of the IEEE international conference on computer vision workshops. 887\u2013895 (2017)","DOI":"10.1109\/ICCVW.2017.108"},{"key":"4733_CR14","doi-asserted-by":"crossref","unstructured":"Kendall, A., et al.: End-to-end learning of geometry and context for deep stereo regression. Proceedings of the IEEE international conference on computer vision. 66\u201375 (2017)","DOI":"10.1109\/ICCV.2017.17"},{"key":"4733_CR15","doi-asserted-by":"crossref","unstructured":"Chang, J.-R., et al.: Pyramid stereo matching network. Proceedings of the IEEE conference on computer vision and pattern recognition (2018)","DOI":"10.1109\/CVPR.2018.00567"},{"key":"4733_CR16","doi-asserted-by":"crossref","unstructured":"Zhang, F., et al.: Ga-net: Guided aggregation net for end-to-end stereo matching. Proceedings of the IEEE conference on computer vision and pattern recognition. 185\u2013194 (2019)","DOI":"10.1109\/CVPR.2019.00027"},{"key":"4733_CR17","doi-asserted-by":"crossref","unstructured":"Guo, X., et al.: Group-wise correlation stereo network. Proceedings of the IEEE conference on computer vision and pattern recognition. (2019)","DOI":"10.1109\/CVPR.2019.00339"},{"key":"4733_CR18","doi-asserted-by":"crossref","unstructured":"Shen, Z., et al.: Cfnet: Cascade and fused cost volume for robust stereo matching. Proceedings of the IEEE conference on computer vision and pattern recognition. 13906\u201313915 (2021)","DOI":"10.1109\/CVPR46437.2021.01369"},{"key":"4733_CR19","doi-asserted-by":"crossref","unstructured":"Xu, G., et al.: Attention concatenation volume for accurate and efficient stereo matching. Proceedings of the IEEE conference on computer vision and pattern recognition. 12981\u201312990 (2022)","DOI":"10.1109\/CVPR52688.2022.01264"},{"key":"4733_CR20","doi-asserted-by":"crossref","unstructured":"Chen, Z., et al.: Mocha-stereo: Motif channel attention network for stereo matching. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 27768\u201327777 (2024)","DOI":"10.1109\/CVPR52733.2024.02623"},{"key":"4733_CR21","unstructured":"Chen, Z., et al.: Motif Channel Opened in a White-Box: Stereo Matching via Motif Correlation Graph. arXiv preprint arXiv:2411.12426 (2024)"},{"key":"4733_CR22","first-page":"590","volume":"573","author":"S Khamis","year":"2018","unstructured":"Khamis, S., et al.: Stereonet: Guided hierarchical refinement for real-time edge-aware depth prediction. Proc. Eur. Conf. Comput. Vis. 573, 590 (2018)","journal-title":"Proc. Eur. Conf. Comput. Vis."},{"key":"4733_CR23","doi-asserted-by":"crossref","unstructured":"Lee, H., Shin, Y.: Real-time stereo matching network with high accuracy. IEEE International Conference on Image Processing, 4280\u20134284 (2019)","DOI":"10.1109\/ICIP.2019.8803514"},{"key":"4733_CR24","doi-asserted-by":"crossref","unstructured":"Duggal, S., et al.: Deeppruner: Learning efficient stereo matching via differentiable patchmatch. Proceedings of the IEEE international conference on computer vision. 4384\u20134393 (2019)","DOI":"10.1109\/ICCV.2019.00448"},{"key":"4733_CR25","first-page":"5900","volume":"5893","author":"Y Wang","year":"2019","unstructured":"Wang, Y., et al.: Anytime stereo image depth Estimation on mobile devices. Int. Conf. Rob. Autom. 5893, 5900 (2019)","journal-title":"Int. Conf. Rob. Autom."},{"key":"4733_CR26","doi-asserted-by":"crossref","unstructured":"Tonioni, A., et al.: Real-time self-adaptive deep stereo. Proceedings of the IEEE conference on computer vision and pattern recognition. 195\u2013204 (2019)","DOI":"10.1109\/CVPR.2019.00028"},{"key":"4733_CR27","doi-asserted-by":"crossref","unstructured":"Wang, Q., et al.: Fadnet: A fast and accurate network for disparity estimation. IEEE international conference on robotics and automation. 101\u2013107 (2020)","DOI":"10.1109\/ICRA40945.2020.9197031"},{"key":"4733_CR28","doi-asserted-by":"crossref","unstructured":"Dovesi, P.L., et al.: Real-time semantic stereo matching. IEEE international conference on robotics and automation. 10780\u201310787 (2020)","DOI":"10.1109\/ICRA40945.2020.9196784"},{"key":"4733_CR29","doi-asserted-by":"crossref","unstructured":"Dai, H., et al.: Adaptive disparity candidates prediction network for efficient real-time stereo matching. IEEE Trans. Circuits Syst. Video Technol. 3099\u20133110 (2021)","DOI":"10.1109\/TCSVT.2021.3102109"},{"key":"4733_CR30","doi-asserted-by":"crossref","unstructured":"Wu, Z., et al.: Toward accurate and real-time binocular vision: A lightweight stereo matching network for RGB stereo images. IEEE Sens. J., (2023)","DOI":"10.1109\/JSEN.2023.3344947"},{"key":"4733_CR31","unstructured":"Baiyu, Pan, et al.: Distill-then-prune: An efficient compression framework for real-time stereo matching network on edge devices. In Proceedings of the IEEE International Conference on Robotics and Automation. (2024)"},{"key":"4733_CR32","doi-asserted-by":"crossref","unstructured":"Ronneberger, O., et al.: U-net: Convolutional networks for biomedical image segmentation. Medical image computing and computer-assisted intervention, 234\u2013241 (2015)","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"4733_CR33","doi-asserted-by":"crossref","unstructured":"Hirschmuller, H.: Stereo processing by semiglobal matching and mutual information. IEEE Trans. Pattern Anal. Mach. Intell. 328\u2013341 (2007)","DOI":"10.1109\/TPAMI.2007.1166"},{"key":"4733_CR34","doi-asserted-by":"crossref","unstructured":"Barnes, C., et al.: PatchMatch: A randomized correspondence algorithm for structural image editing. ACM Trans. Graph, 24 (2009)","DOI":"10.1145\/1576246.1531330"},{"key":"4733_CR35","doi-asserted-by":"crossref","unstructured":"Geiger, A., et al.: Are we ready for autonomous driving? The KITTI vision benchmark suite. Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. 3354\u20133361 (2012)","DOI":"10.1109\/CVPR.2012.6248074"},{"key":"4733_CR36","doi-asserted-by":"crossref","unstructured":"Geiger, A., Lenz, P., et al.: Vision Meets robotics: The KITTI dataset. Int. J. Robot. Res. 1231\u20131237 (2013)","DOI":"10.1177\/0278364913491297"},{"key":"4733_CR37","doi-asserted-by":"crossref","unstructured":"Liu, C.-W., et al.: Playing to Vision Foundation Model\u2019s Strengths in Stereo Matching. IEEE Trans. Intell. Veh. (2024)","DOI":"10.1109\/TIV.2024.3467287"},{"key":"4733_CR38","doi-asserted-by":"crossref","unstructured":"Alper Emlek, et al.: P3SNet: Parallel pyramid pooling stereo network. IEEE Trans. Intell. Transp. Syst., 10433\u201310444 (2023)","DOI":"10.1109\/TITS.2023.3276328"},{"key":"4733_CR39","doi-asserted-by":"crossref","unstructured":"Shamsafar, F., et al.: Mobilestereonet: Towards lightweight deep networks for stereo matching. Proceedings of the ieee winter conference on applications of computer vision (2022)","DOI":"10.1109\/WACV51458.2022.00075"},{"key":"4733_CR40","doi-asserted-by":"crossref","unstructured":"Xu, G., Wang, X., et al.: Igev++: Iterative multi-range geometry encoding volumes for stereo matching. IEEE Trans. Pattern Anal. Mach. Intell. (2025)","DOI":"10.1109\/TPAMI.2025.3569218"},{"key":"4733_CR41","doi-asserted-by":"crossref","unstructured":"Cheng, J., et al.: MonSter: Marry Monodepth to Stereo Unleashes Power. arXiv preprint arXiv:2501.08643 (2025)","DOI":"10.1109\/CVPR52734.2025.00588"},{"key":"4733_CR42","doi-asserted-by":"crossref","unstructured":"Jiang, H., et al.: DEFOM-Stereo: Depth Foundation Model Based Stereo Matching. arXiv preprint arXiv:2501.09466 (2025)","DOI":"10.1109\/CVPR52734.2025.02036"},{"key":"4733_CR43","doi-asserted-by":"crossref","unstructured":"Hashemi, H., et al.: Real-time stereo matching with enhanced geometric comprehension through cross-attention integration. Neurocomputing, 636 (2025)","DOI":"10.1016\/j.neucom.2025.130069"},{"key":"4733_CR44","doi-asserted-by":"crossref","unstructured":"Tahmasebi, M., et al.: DCVSMNet: Double cost volume stereo matching network. Neurocomputing, 618 (2025)","DOI":"10.1016\/j.neucom.2024.129002"},{"key":"4733_CR45","doi-asserted-by":"crossref","unstructured":"Wang, Y., et al.: ADStereo: Efficient stereo matching with adaptive downsampling and disparity alignment. IEEE Trans. Image Process. (2025)","DOI":"10.1109\/TIP.2025.3540282"},{"key":"4733_CR46","doi-asserted-by":"crossref","unstructured":"Wei, Q., et al.: CAFNet: Efficient cost aggregation via Conv-Atten fusion for Real-Time stereo matching toward autonomous driving. IEEE Trans. Intell. Transp. Syst. (2025)","DOI":"10.1109\/TITS.2025.3561309"},{"key":"4733_CR47","doi-asserted-by":"crossref","unstructured":"Fang, I., et al.: ES3Net: accurate and efficient edge-based self-supervised stereo matching network. Proceedings of the IEEE conference on computer vision and pattern recognition. (2023)","DOI":"10.1109\/CVPRW59228.2023.00470"},{"key":"4733_CR48","doi-asserted-by":"crossref","unstructured":"Chang, Q., et al.: TinyStereo: A tiny coarse-to-fine framework for vision-based depth Estimation on embedded GPUs. IEEE Trans. Syst. Man. Cybernetics: Syst. 5196\u20135208 (2024)","DOI":"10.1109\/TSMC.2024.3395464"}],"container-title":["Signal, Image and Video Processing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04733-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11760-025-04733-0\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11760-025-04733-0.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,9,22]],"date-time":"2025-09-22T13:16:36Z","timestamp":1758546996000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11760-025-04733-0"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2025,9,9]]},"references-count":48,"journal-issue":{"issue":"12","published-print":{"date-parts":[[2025,12]]}},"alternative-id":["4733"],"URL":"https:\/\/doi.org\/10.1007\/s11760-025-04733-0","relation":{},"ISSN":["1863-1703","1863-1711"],"issn-type":[{"type":"print","value":"1863-1703"},{"type":"electronic","value":"1863-1711"}],"subject":[],"published":{"date-parts":[[2025,9,9]]},"assertion":[{"value":"18 April 2025","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"13 July 2025","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"31 August 2025","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"9 September 2025","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"Not applicable.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Ethics approval"}},{"value":"The authors declare no competing interests.","order":3,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}],"article-number":"1036"}}