{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,12,6]],"date-time":"2025-12-06T16:47:44Z","timestamp":1765039664408,"version":"3.37.3"},"reference-count":33,"publisher":"Springer Science and Business Media LLC","issue":"4","license":[{"start":{"date-parts":[[2022,9,26]],"date-time":"2022-09-26T00:00:00Z","timestamp":1664150400000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2022,9,26]],"date-time":"2022-09-26T00:00:00Z","timestamp":1664150400000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/100002089","name":"Fight for Sight","doi-asserted-by":"publisher","award":["Summer student fellowship"],"award-info":[{"award-number":["Summer student fellowship"]}],"id":[{"id":"10.13039\/100002089","id-type":"DOI","asserted-by":"publisher"}]},{"DOI":"10.13039\/100000053","name":"National Eye Institute","doi-asserted-by":"publisher","award":["EY020518"],"award-info":[{"award-number":["EY020518"]}],"id":[{"id":"10.13039\/100000053","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Neural Process Lett"],"published-print":{"date-parts":[[2023,8]]},"DOI":"10.1007\/s11063-022-11039-6","type":"journal-article","created":{"date-parts":[[2022,9,26]],"date-time":"2022-09-26T07:03:27Z","timestamp":1664175807000},"page":"4231-4255","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["DDCNet-Multires: Effective Receptive Field Guided Multiresolution CNN for Dense Prediction"],"prefix":"10.1007","volume":"55","author":[{"given":"Ali","family":"Salehi","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4676-106X","authenticated-orcid":false,"given":"Madhusudhanan","family":"Balasubramanian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,26]]},"reference":[{"issue":"3","key":"11039_CR1","doi-asserted-by":"publisher","first-page":"283","DOI":"10.1007\/BF00158167","volume":"2","author":"P Anandan","year":"1989","unstructured":"Anandan P (1989) A computational framework and an algorithm for the measurement of visual motion. Int J Comput Vision 2(3):283\u2013310","journal-title":"Int J Comput Vision"},{"issue":"1","key":"11039_CR2","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1007\/s11263-010-0390-2","volume":"92","author":"S Baker","year":"2011","unstructured":"Baker S, Daniel Scharstein JP, Lewis SR, Black MJ, Szeliski R (2011) A database and evaluation methodology for optical flow. Int J Comput Vision 92(1):1\u201331","journal-title":"Int J Comput Vision"},{"key":"11039_CR3","unstructured":"Balasubramanian M (2006) A Computational framework for the structural change analysis of 3D volumes of microscopic specimens. PhD thesis, Louisiana State University"},{"key":"11039_CR4","doi-asserted-by":"crossref","unstructured":"Bergen JR, Anandan P, Hanna KJ, Hingorani R (1992)Hierarchical model-based motion estimation. In: European conference on computer vision. Springer, pp 237\u2013252","DOI":"10.1007\/3-540-55426-2_27"},{"key":"11039_CR5","doi-asserted-by":"crossref","unstructured":"Butler DJ, Wulff J, Stanley GB, Black MJ (2012) A naturalistic open source movie for optical flow evaluation. In: Fitzgibbon A et al. (ed) European conf. on computer vision (ECCV), Part IV, LNCS 7577. Springer-Verlag, pp 611\u2013625 October","DOI":"10.1007\/978-3-642-33783-3_44"},{"key":"11039_CR6","doi-asserted-by":"crossref","unstructured":"Dosovitskiy A, Fischer P, Ilg E, Hausser P, Hazirbas C, Golkov V, Van\u00a0der\u00a0Smagt P, Cremers D, Brox T (2015) Flownet: Learning optical flow with convolutional networks. In: Proceedings of the IEEE international conference on computer vision, pp 2758\u20132766","DOI":"10.1109\/ICCV.2015.316"},{"key":"11039_CR7","first-page":"2366","volume":"27","author":"E David","year":"2014","unstructured":"David E, Christian P, Rob F (2014) Depth map prediction from a single image using a multi-scale deep network. Adv Neural Inf Process Syst 27:2366\u20132374","journal-title":"Adv Neural Inf Process Syst"},{"issue":"11","key":"11039_CR8","doi-asserted-by":"publisher","first-page":"1231","DOI":"10.1177\/0278364913491297","volume":"32","author":"A Geiger","year":"2013","unstructured":"Geiger A, Lenz P, Stiller C, Urtasun R (2013) Vision meets robotics: the kitti dataset. Int J Robot Res 32(11):1231\u20131237","journal-title":"Int J Robot Res"},{"key":"11039_CR9","doi-asserted-by":"crossref","unstructured":"Gur S, Wolf L (2019) Single image depth estimation trained via depth from defocus cues. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7683\u20137692","DOI":"10.1109\/CVPR.2019.00787"},{"key":"11039_CR10","doi-asserted-by":"crossref","unstructured":"Hui T-W, Loy CC (2020) Liteflownet3: Resolving correspondence ambiguity for more accurate optical flow estimation. In: European conference on computer vision. Springer, pp 169\u2013184","DOI":"10.1007\/978-3-030-58565-5_11"},{"key":"11039_CR11","doi-asserted-by":"crossref","unstructured":"Hui T-W, Tang X, Loy CC (2018) Liteflownet: A lightweight convolutional neural network for optical flow estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8981\u20138989","DOI":"10.1109\/CVPR.2018.00936"},{"key":"11039_CR12","unstructured":"Hui T-W, Tang X, Loy CC (2019) A lightweight optical flow CNN-revisiting data fidelity and regularization. arXiv:1903.07414"},{"key":"11039_CR13","doi-asserted-by":"crossref","unstructured":"Ilg E, Mayer N, Saikia T, Keuper M, Dosovitskiy A, Brox T (2017) Flownet 2.0: Evolution of optical flow estimation with deep networks. In: IEEE conference on computer vision and pattern recognition (CVPR), vol 2, p 6","DOI":"10.1109\/CVPR.2017.179"},{"key":"11039_CR14","doi-asserted-by":"crossref","unstructured":"Ilg E, Saikia T, Keuper M, Brox T (2018) Occlusions, motion and depth boundaries with a generic network for disparity, optical flow or scene flow estimation. In Proceedings of the European conference on computer vision (ECCV), pp 614\u2013630","DOI":"10.1007\/978-3-030-01258-8_38"},{"key":"11039_CR15","doi-asserted-by":"crossref","unstructured":"Janai J, Guney F, Ranjan A, Black M, Geiger A (2018) Unsupervised learning of multi-frame optical flow with occlusions. In: Proceedings of the European conference on computer vision (ECCV), pp 690\u2013706","DOI":"10.1007\/978-3-030-01270-0_42"},{"key":"11039_CR16","doi-asserted-by":"crossref","unstructured":"Jason JY, Harley AW, Derpanis KG (2016) Back to basics: Unsupervised learning of optical flow via brightness constancy and motion smoothness. In European conference on computer vision. Springer, pp 3\u201310","DOI":"10.1007\/978-3-319-49409-8_1"},{"key":"11039_CR17","doi-asserted-by":"publisher","unstructured":"Jiang S, Campbell D, Lu Y, Li H, Hartley R (2021) Learning to estimate hidden motions with global motion aggregation. In: Proceedings of the IEEE international conference on computer vision, pp 9752\u20139761. Institute of Electrical and Electronics Engineers Inc., ISBN 9781665428125. https:\/\/doi.org\/10.1109\/ICCV48922.2021.00963","DOI":"10.1109\/ICCV48922.2021.00963"},{"key":"11039_CR18","doi-asserted-by":"crossref","unstructured":"Liu C, Chen L-C, Schroff F, Adam H, Hua W, Yuille AL, Fei-Fei L (2019) Auto-deeplab: Hierarchical neural architecture search for semantic image segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 82\u201392","DOI":"10.1109\/CVPR.2019.00017"},{"key":"11039_CR19","doi-asserted-by":"crossref","unstructured":"Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431\u20133440","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"11039_CR20","unstructured":"Lotter W, Kreiman G, Cox D (2016) Deep predictive coding networks for video prediction and unsupervised learning. arXiv:1605.08104"},{"key":"11039_CR21","doi-asserted-by":"crossref","unstructured":"Mayer N, Ilg E, Hausser P, Fischer P, Cremers D, Dosovitskiy A, Brox T (2016) A large dataset to train convolutional networks for disparity, optical flow, and scene flow estimation. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4040\u20134048, This is DispNet. They have introduced several methods here","DOI":"10.1109\/CVPR.2016.438"},{"key":"11039_CR22","doi-asserted-by":"crossref","unstructured":"Menze M, Geiger A (2015) Object scene flow for autonomous vehicles. In Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3061\u20133070","DOI":"10.1109\/CVPR.2015.7298925"},{"key":"11039_CR23","doi-asserted-by":"crossref","unstructured":"Ranjan A, Black MJ (2017) Optical flow estimation using a spatial pyramid network. In IEEE conference on computer vision and pattern recognition (CVPR), vol 2, p 2. IEEE","DOI":"10.1109\/CVPR.2017.291"},{"key":"11039_CR24","doi-asserted-by":"crossref","unstructured":"Ren Z, Yan J, Ni B, Liu B, Yang X, Zha H (2017) Unsupervised deep learning for optical flow estimation. In: AAAI, pp 1495\u20131501","DOI":"10.1609\/aaai.v31i1.10723"},{"key":"11039_CR25","unstructured":"Salehi A, Balasubramanian M (2021) DDCNet: Deep dilated convolutional neural network for dense prediction. arXiv preprint arXiv:2107.04715"},{"key":"11039_CR26","doi-asserted-by":"crossref","unstructured":"Sun D, Yang X, Liu M-Y, Kautz J (2018) Pwc-net: Cnns for optical flow using pyramid, warping, and cost volume. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8934\u20138943","DOI":"10.1109\/CVPR.2018.00931"},{"issue":"6","key":"11039_CR27","doi-asserted-by":"publisher","first-page":"1408","DOI":"10.1109\/TPAMI.2019.2894353","volume":"42","author":"D Sun","year":"2019","unstructured":"Sun D, Yang X, Liu M-Y, Kautz J (2019) Models matter, so does training: an empirical study of CNNS for optical flow estimation. IEEE Trans Pattern Anal Mach Intell 42(6):1408\u20131423","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"11039_CR28","doi-asserted-by":"publisher","unstructured":"Teed Z, Deng J (2021) RAFT: Recurrent all-pairs field transforms for optical flow (Extended Abstract). In: IJCAI international joint conference on artificial intelligence, pp 4839\u20134843, mar ISBN 9780999241196. https:\/\/doi.org\/10.24963\/ijcai.2021\/662. arXiv:2003.12039","DOI":"10.24963\/ijcai.2021\/662"},{"key":"11039_CR29","doi-asserted-by":"crossref","unstructured":"Wang Y, Yang Y, Yang Z, Zhao L, Wang P, Xu W (2018) Occlusion aware unsupervised learning of optical flow. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4884\u20134893, URL http:\/\/openaccess.thecvf.com\/content_cvpr_2018\/papers\/Wang_Occlusion_Aware_Unsupervised_CVPR_2018_paper.pdf","DOI":"10.1109\/CVPR.2018.00513"},{"key":"11039_CR30","doi-asserted-by":"publisher","first-page":"124","DOI":"10.1016\/j.neucom.2019.08.040","volume":"368","author":"M Zhai","year":"2019","unstructured":"Zhai M, Xiang X, Zhang R, Lv N, El Saddik A (2019) Optical flow estimation using channel attention mechanism and dilated convolutional neural networks. Neurocomputing 368:124\u2013132. https:\/\/doi.org\/10.1016\/j.neucom.2019.08.040 (ISSN 18728286)","journal-title":"Neurocomputing"},{"key":"11039_CR31","doi-asserted-by":"publisher","DOI":"10.1109\/TMM.2021.3096083","author":"C Zhang","year":"2021","unstructured":"Zhang C, Zhou Z, Chen Z, Weiming H, Li M, Jiang S (2021) Self-attention-based multiscale feature learning optical flow with occlusion feature map prediction. IEEE Trans Multimed. https:\/\/doi.org\/10.1109\/TMM.2021.3096083 (ISSN 19410077)","journal-title":"IEEE Trans Multimed"},{"key":"11039_CR32","doi-asserted-by":"publisher","first-page":"116560","DOI":"10.1016\/j.image.2021.116560","volume":"101","author":"C Zhang","year":"2022","unstructured":"Zhang C, Feng C, Chen Z, Weiming H, Li M (2022) Parallel multiscale context-based edge-preserving optical flow estimation with occlusion detection. Signal Process: Image Commun 101:116560. https:\/\/doi.org\/10.1016\/j.image.2021.116560 (ISSN 09235965)","journal-title":"Signal Process: Image Commun"},{"issue":"3","key":"11039_CR33","doi-asserted-by":"publisher","first-page":"468","DOI":"10.1109\/TKDE.2019.2891537","volume":"32","author":"J Zhang","year":"2019","unstructured":"Zhang J, Zheng Y, Sun J, Qi D (2019) Flow prediction in spatio-temporal networks based on multitask deep learning. IEEE Trans Knowl Data Eng 32(3):468\u2013478","journal-title":"IEEE Trans Knowl Data Eng"}],"container-title":["Neural Processing Letters"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-11039-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11063-022-11039-6\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11063-022-11039-6.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,7,31]],"date-time":"2023-07-31T16:40:13Z","timestamp":1690821613000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11063-022-11039-6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022,9,26]]},"references-count":33,"journal-issue":{"issue":"4","published-print":{"date-parts":[[2023,8]]}},"alternative-id":["11039"],"URL":"https:\/\/doi.org\/10.1007\/s11063-022-11039-6","relation":{},"ISSN":["1370-4621","1573-773X"],"issn-type":[{"type":"print","value":"1370-4621"},{"type":"electronic","value":"1573-773X"}],"subject":[],"published":{"date-parts":[[2022,9,26]]},"assertion":[{"value":"13 September 2022","order":1,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"26 September 2022","order":2,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no conflicts of interests that are relevant to the contents of this research work.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of Interest"}}]}}