{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T05:40:04Z","timestamp":1746078004509,"version":"3.40.4"},"reference-count":66,"publisher":"Springer Science and Business Media LLC","issue":"12","license":[{"start":{"date-parts":[[2024,5,17]],"date-time":"2024-05-17T00:00:00Z","timestamp":1715904000000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,5,17]],"date-time":"2024-05-17T00:00:00Z","timestamp":1715904000000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"name":"National Major Scientific Instruments and Equipments Development Project of National Natural Science Foundation of China","award":["62327808"],"award-info":[{"award-number":["62327808"]}]},{"name":"NSFC Fund","award":["62171288"],"award-info":[{"award-number":["62171288"]}]},{"DOI":"10.13039\/501100012271","name":"Shenzhen Fundamental Research Fund","doi-asserted-by":"crossref","award":["20200810150441003"],"award-info":[{"award-number":["20200810150441003"]}],"id":[{"id":"10.13039\/501100012271","id-type":"DOI","asserted-by":"crossref"}]},{"DOI":"10.13039\/501100021171","name":"Basic and Applied Basic Research Foundation of Guangdong Province","doi-asserted-by":"publisher","award":["2021A1515012287"],"award-info":[{"award-number":["2021A1515012287"]}],"id":[{"id":"10.13039\/501100021171","id-type":"DOI","asserted-by":"publisher"}]},{"name":"Natural Science Foundation of Top Talent of SZTU","award":["20211061010009"],"award-info":[{"award-number":["20211061010009"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"DOI":"10.1007\/s11042-024-19271-z","type":"journal-article","created":{"date-parts":[[2024,5,17]],"date-time":"2024-05-17T03:33:56Z","timestamp":1715916836000},"page":"10609-10630","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["EAT: epipolar-aware Transformer for low-light light field enhancement"],"prefix":"10.1007","volume":"84","author":[{"ORCID":"https:\/\/orcid.org\/0000-0002-5433-3631","authenticated-orcid":false,"given":"Xingzheng","family":"Wang","sequence":"first","affiliation":[]},{"given":"Wenhao","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Kaiqiang","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Zixuan","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yuanlong","family":"Deng","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,5,17]]},"reference":[{"key":"19271_CR1","unstructured":"Ng R, Levoy M, Br\u00e9dif M, Duval G, Horowitz M, Hanrahan P (2005) Light field photography with a hand-held plenoptic camera. PhD thesis, Stanford University"},{"key":"19271_CR2","doi-asserted-by":"crossref","unstructured":"Lumsdaine A, Georgiev T (2009) The focused plenoptic camera. In: 2009 IEEE International Conference on Computational Photography (ICCP), pp 1\u20138. IEEE","DOI":"10.1109\/ICCPHOT.2009.5559008"},{"issue":"7","key":"19271_CR3","doi-asserted-by":"publisher","first-page":"926","DOI":"10.1109\/JSTSP.2017.2747126","volume":"11","author":"G Wu","year":"2017","unstructured":"Wu G, Masia B, Jarabo A, Zhang Y, Wang L, Dai Q, Chai T, Liu Y (2017) Light field image processing: An overview. IEEE J Sel Top Signal Process 11(7):926\u2013954","journal-title":"IEEE J Sel Top Signal Process"},{"issue":"5","key":"19271_CR4","doi-asserted-by":"publisher","first-page":"551","DOI":"10.1038\/s41592-021-01058-x","volume":"18","author":"Z Wang","year":"2021","unstructured":"Wang Z, Zhu L, Zhang H, Li G, Yi C, Li Y, Yang Y, Ding Y, Zhen M, Gao S et al (2021) Real-time volumetric reconstruction of biological dynamics with light-field microscopy and deep learning. Nat Methods 18(5):551\u2013556","journal-title":"Nat Methods"},{"key":"19271_CR5","doi-asserted-by":"crossref","unstructured":"Ding Y, Chen Z, Ji Y, Yu J, Ye J (2023) Light field-based underwater 3d reconstruction via angular resampling. IEEE Trans Comput Imaging","DOI":"10.1109\/TCI.2023.3319983"},{"issue":"12","key":"19271_CR6","doi-asserted-by":"publisher","first-page":"3016","DOI":"10.1364\/AO.484909","volume":"62","author":"W Feng","year":"2023","unstructured":"Feng W, Gao J, Sun J, Wang H (2023) 3d reconstruction of light-field images based on spatiotemporal correlation super-resolution. Appl Opt 62(12):3016\u20133027","journal-title":"Appl Opt"},{"key":"19271_CR7","doi-asserted-by":"publisher","first-page":"146868","DOI":"10.1109\/ACCESS.2021.3123529","volume":"9","author":"D Bonatto","year":"2021","unstructured":"Bonatto D, Fachada S, Rogge S, Munteanu A, Lafruit G (2021) Real-time depth video-based rendering for 6-dof hmd navigation and light field displays. IEEE Access 9:146868\u2013146887","journal-title":"IEEE Access"},{"issue":"6","key":"19271_CR8","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3478513.3480481","volume":"40","author":"N Matsuda","year":"2021","unstructured":"Matsuda N, Wheelwright B, Hegland J, Lanman D (2021) Vr social copresence with light field displays. ACM Trans Graph (TOG) 40(6):1\u201313","journal-title":"ACM Trans Graph (TOG)"},{"key":"19271_CR9","doi-asserted-by":"crossref","unstructured":"Kara PA, Tamboli RR, Adhikarla VK, Balogh T, Guindy M, Simon A (2023) Connected without disconnection: overview of light field metaverse applications and their quality of experience. Displays 102430","DOI":"10.1016\/j.displa.2023.102430"},{"issue":"20","key":"19271_CR10","doi-asserted-by":"publisher","first-page":"29788","DOI":"10.1364\/OE.404318","volume":"28","author":"J-H Lee","year":"2020","unstructured":"Lee J-H, Yanusik I, Choi Y, Kang B, Hwang C, Park J, Nam D, Hong S (2020) Automotive augmented reality 3d head-up display based on light-field rendering with eye-tracking. Opt Express 28(20):29788\u201329804","journal-title":"Opt Express"},{"key":"19271_CR11","doi-asserted-by":"crossref","unstructured":"Shi J, Hua J, Zhou F, Yang M, Qiao W (2021) Augmented reality vector light field display with large viewing distance based on pixelated multilevel blazed gratings. In: Photonics, vol 8, pp 337. MDPI","DOI":"10.3390\/photonics8080337"},{"issue":"3","key":"19271_CR12","doi-asserted-by":"publisher","first-page":"034001","DOI":"10.1117\/1.AP.5.3.034001","volume":"5","author":"Z Liu","year":"2023","unstructured":"Liu Z, Wang D, Gao H, Li M, Zhou H, Zhang C (2023) Metasurface-enabled augmented reality display: a review. Adv Photonics 5(3):034001\u2013034001","journal-title":"Adv Photonics"},{"key":"19271_CR13","doi-asserted-by":"crossref","unstructured":"Wang T, Piao Y, Li X, Zhang L, Lu H (2019) Deep learning for light field saliency detection. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 8838\u20138848","DOI":"10.1109\/ICCV.2019.00893"},{"key":"19271_CR14","doi-asserted-by":"crossref","unstructured":"Wang X, Chen S, Wei G, Liu J (2022) Tenet: Accurate light-field salient object detection with a transformer embedding network. Image Vis Comput 104595","DOI":"10.1016\/j.imavis.2022.104595"},{"key":"19271_CR15","doi-asserted-by":"crossref","unstructured":"Liu N, Zhao W, Zhang D, Han J, Shao L (2021) Light field saliency detection with dual local graph learning and reciprocative guidance. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 4712\u20134721","DOI":"10.1109\/ICCV48922.2021.00467"},{"key":"19271_CR16","doi-asserted-by":"publisher","first-page":"4421","DOI":"10.1109\/TIP.2020.2970529","volume":"29","author":"J Zhang","year":"2020","unstructured":"Zhang J, Liu Y, Zhang S, Poppe R, Wang M (2020) Light field saliency detection with deep convolutional networks. IEEE Trans Image Process 29:4421\u20134434","journal-title":"IEEE Trans Image Process"},{"key":"19271_CR17","first-page":"116853","volume":"109","author":"S Zhang","year":"2022","unstructured":"Zhang S, Chen Y, An P, Huang X, Yang C (2022) Light field occlusion removal network via foreground location and background recovery. Signal Process: Image Commun 109:116853","journal-title":"Signal Process: Image Commun"},{"key":"19271_CR18","doi-asserted-by":"crossref","unstructured":"Wang X, Liu J, Chen S, Wei G (2022) Effective light field de-occlusion network based on swin transformer. IEEE Trans Circuits Syst Video Technol","DOI":"10.1109\/TCSVT.2022.3226227"},{"key":"19271_CR19","doi-asserted-by":"crossref","unstructured":"Wang Y, Wu T, Yang J, Wang L, An W, Guo Y (2020) Deoccnet: Learning to see through foreground occlusions in light fields. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 118\u2013127","DOI":"10.1109\/WACV45572.2020.9093448"},{"key":"19271_CR20","doi-asserted-by":"crossref","unstructured":"Chen C, Chen Q, Xu J, Koltun V (2018) Learning to see in the dark. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3291\u20133300","DOI":"10.1109\/CVPR.2018.00347"},{"key":"19271_CR21","doi-asserted-by":"crossref","unstructured":"Xu K, Yang X, Yin B, Lau RW (2020) Learning to restore low-light images via decomposition-and-enhancement. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2281\u20132290","DOI":"10.1109\/CVPR42600.2020.00235"},{"key":"19271_CR22","doi-asserted-by":"crossref","unstructured":"Shin C, Jeon H-G, Yoon Y, Kweon IS, Kim SJ (2018) Epinet: A fully-convolutional neural network using epipolar geometry for depth from light field images. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4748\u20134757","DOI":"10.1109\/CVPR.2018.00499"},{"key":"19271_CR23","doi-asserted-by":"crossref","unstructured":"Chen J, Zhang S, Lin Y (2021) Attention-based multi-level fusion network for light field depth estimation. In: Proceedings of the AAAI conference on artificial intelligence, vol 35, pp 1009\u20131017","DOI":"10.1609\/aaai.v35i2.16185"},{"issue":"11","key":"19271_CR24","doi-asserted-by":"publisher","first-page":"16329","DOI":"10.1007\/s11042-020-08890-x","volume":"80","author":"X Wang","year":"2021","unstructured":"Wang X, Dong Y, Zhang Q, Wang Q (2021) Region-based depth feature descriptor for saliency detection on light field. Multimed Tools Appl 80(11):16329\u201316346","journal-title":"Multimed Tools Appl"},{"key":"19271_CR25","doi-asserted-by":"publisher","first-page":"30","DOI":"10.1016\/j.cviu.2018.10.010","volume":"178","author":"YP Loh","year":"2019","unstructured":"Loh YP, Chan CS (2019) Getting to know low-light images with the exclusively dark dataset. Comput Vis Image Underst 178:30\u201342","journal-title":"Comput Vis Image Underst"},{"key":"19271_CR26","doi-asserted-by":"crossref","unstructured":"Vogt C, Lyu G, Subr K (2020) Lightless fields: Enhancement and denoising of light-deficient light fields. In: International symposium on visual computing, pp 383\u2013396. Springer","DOI":"10.1007\/978-3-030-64556-4_30"},{"key":"19271_CR27","doi-asserted-by":"publisher","first-page":"76","DOI":"10.1016\/j.neucom.2021.05.074","volume":"456","author":"S Zhang","year":"2021","unstructured":"Zhang S, Lam EY (2021) Learning to restore light fields under low-light imaging. Neurocomputing 456:76\u201387","journal-title":"Neurocomputing"},{"key":"19271_CR28","doi-asserted-by":"crossref","unstructured":"Ge Z, Song L, Lam EY (2020) Light field image restoration in low-light environment. In: SPIE future sensing technologies, vol 11525, pp 300\u2013305. SPIE","DOI":"10.1117\/12.2580033"},{"key":"19271_CR29","doi-asserted-by":"publisher","first-page":"1501","DOI":"10.1109\/TIP.2020.3045617","volume":"30","author":"M Lamba","year":"2020","unstructured":"Lamba M, Rachavarapu KK, Mitra K (2020) Harnessing multi-view perspective of light fields for low-light imaging. IEEE Trans Image Process 30:1501\u20131513","journal-title":"IEEE Trans Image Process"},{"key":"19271_CR30","doi-asserted-by":"crossref","unstructured":"Lamba M, Mitra K (2022) Fast and efficient restoration of extremely dark light fields. In: Proceedings of the IEEE\/CVF winter conference on applications of computer vision, pp 1361\u20131370","DOI":"10.1109\/WACV51458.2022.00321"},{"key":"19271_CR31","doi-asserted-by":"publisher","first-page":"108279","DOI":"10.1016\/j.sigpro.2021.108279","volume":"189","author":"S Zhang","year":"2021","unstructured":"Zhang S, Lam EY (2021) An effective decomposition-enhancement method to restore light field images captured in the dark. Signal Process 189:108279","journal-title":"Signal Process"},{"issue":"6","key":"19271_CR32","doi-asserted-by":"publisher","first-page":"108","DOI":"10.1038\/scientificamerican1277-108","volume":"237","author":"EH Land","year":"1977","unstructured":"Land EH (1977) The retinex theory of color vision. Sci Am 237(6):108\u2013129","journal-title":"Sci Am"},{"key":"19271_CR33","doi-asserted-by":"crossref","unstructured":"Guo M, Hou J, Jin J, Chen J, Chau L-P (2020) Deep spatial-angular regularization for compressive light field reconstruction over coded apertures. In: Computer vision\u2013ECCV 2020: 16th European Conference, Glasgow, UK, August 23\u201328, 2020, Proceedings, Part II 16, pp 278\u2013294. Springer","DOI":"10.1007\/978-3-030-58536-5_17"},{"issue":"10","key":"19271_CR34","doi-asserted-by":"publisher","first-page":"6094","DOI":"10.1109\/TPAMI.2021.3087485","volume":"44","author":"M Guo","year":"2022","unstructured":"Guo M, Hou J, Jin J, Chen J, Chau L-P (2022) Deep spatial-angular regularization for light field imaging, denoising, and super-resolution. IEEE Trans Pattern Anal Mach Intell 44(10):6094\u20136110","journal-title":"IEEE Trans Pattern Anal Mach Intell"},{"key":"19271_CR35","doi-asserted-by":"crossref","unstructured":"Zhang S, Lin Y, Sheng H (2019) Residual networks for light field image super-resolution. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp. 11046\u201311055","DOI":"10.1109\/CVPR.2019.01130"},{"key":"19271_CR36","doi-asserted-by":"crossref","unstructured":"Khan N, Zhang Q, Kasser L, Stone H, Kim MH, Tompkin J (2019) View-consistent 4d light field superpixel segmentation. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp. 7811\u20137819","DOI":"10.1109\/ICCV.2019.00790"},{"key":"19271_CR37","doi-asserted-by":"publisher","first-page":"1507","DOI":"10.1109\/TCI.2020.3037413","volume":"6","author":"D Liu","year":"2020","unstructured":"Liu D, Huang Y, Wu Q, Ma R, An P (2020) Multi-angular epipolar geometry based light field angular reconstruction network. IEEE Trans Comput Imaging 6:1507\u20131522","journal-title":"IEEE Trans Comput Imaging"},{"key":"19271_CR38","doi-asserted-by":"publisher","first-page":"5956","DOI":"10.1109\/TIP.2021.3079805","volume":"30","author":"S Zhang","year":"2021","unstructured":"Zhang S, Chang S, Lin Y (2021) End-to-end light field spatial super-resolution network using multiple epipolar geometry. IEEE Trans Image Process 30:5956\u20135968","journal-title":"IEEE Trans Image Process"},{"key":"19271_CR39","doi-asserted-by":"publisher","first-page":"188","DOI":"10.1016\/j.inffus.2021.10.005","volume":"79","author":"X Wang","year":"2022","unstructured":"Wang X, Ma J, Yi P, Tian X, Jiang J, Zhang X-P (2022) Learning an epipolar shift compensation for light field image super-resolution. Inf Fusion 79:188\u2013199","journal-title":"Inf Fusion"},{"key":"19271_CR40","doi-asserted-by":"crossref","unstructured":"Ronneberger O, Fischer P, Brox T (2015) U-net: Convolutional networks for biomedical image segmentation. In: International conference on medical image computing and computer-assisted intervention, pp 234\u2013241. Springer","DOI":"10.1007\/978-3-319-24574-4_28"},{"key":"19271_CR41","doi-asserted-by":"crossref","unstructured":"He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770\u2013778","DOI":"10.1109\/CVPR.2016.90"},{"key":"19271_CR42","doi-asserted-by":"crossref","unstructured":"Zamir SW, Arora A, Khan S, Hayat M, Khan FS, Yang M-H (2022) Restormer: Efficient transformer for high-resolution image restoration. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5728\u20135739","DOI":"10.1109\/CVPR52688.2022.00564"},{"key":"19271_CR43","unstructured":"Vaswani A, Shazeer N, Parmar N, Uszkoreit J, Jones L, Gomez AN, Kaiser \u0141, Polosukhin I (2017) Attention is all you need. Advances in neural information processing systems 30"},{"key":"19271_CR44","unstructured":"Chen M, Radford A, Child R, Wu J, Jun H, Luan D, Sutskever I (2020) Generative pretraining from pixels. In: International conference on machine learning, pp 1691\u20131703. PMLR"},{"key":"19271_CR45","unstructured":"Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, Dehghani M, Minderer M, Heigold G, Gelly S et al (2020) An image is worth 16x16 words: Transformers for image recognition at scale. Preprint arXiv:2010.11929"},{"key":"19271_CR46","doi-asserted-by":"crossref","unstructured":"Carion N, Massa F, Synnaeve G, Usunier N, Kirillov A, Zagoruyko S (2020) End-to-end object detection with transformers. In: European conference on computer vision, pp 213\u2013229. Springer","DOI":"10.1007\/978-3-030-58452-8_13"},{"key":"19271_CR47","unstructured":"Zhu X, Su W, Lu L, Li B, Wang X, Dai J (2020) Deformable detr: Deformable transformers for end-to-end object detection. Preprint arXiv:2010.04159"},{"key":"19271_CR48","doi-asserted-by":"crossref","unstructured":"Zheng S, Lu J, Zhao H, Zhu X, Luo Z, Wang Y, Fu Y, Feng J, Xiang T, Torr PH et al (2021) Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 6881\u20136890","DOI":"10.1109\/CVPR46437.2021.00681"},{"key":"19271_CR49","doi-asserted-by":"crossref","unstructured":"Chen H, Wang Y, Guo T, Xu C, Deng Y, Liu Z, Ma S, Xu C, Xu C, Gao W (2021) Pre-trained image processing transformer. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 12299\u201312310","DOI":"10.1109\/CVPR46437.2021.01212"},{"key":"19271_CR50","doi-asserted-by":"crossref","unstructured":"Zhou L, Zhou Y, Corso JJ, Socher R, Xiong C (2018) End-to-end dense video captioning with masked transformer. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 8739\u20138748","DOI":"10.1109\/CVPR.2018.00911"},{"key":"19271_CR51","doi-asserted-by":"crossref","unstructured":"Wang H, Zhu Y, Adam H, Yuille A, Chen L-C (2021) Max-deeplab: End-to-end panoptic segmentation with mask transformers. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 5463\u20135474","DOI":"10.1109\/CVPR46437.2021.00542"},{"key":"19271_CR52","doi-asserted-by":"crossref","unstructured":"Lee J, Son H, Rim J, Cho S, Lee S (2021) Iterative filter adaptive network for single image defocus deblurring. In: Proceedings of the IEEE\/CVF conference on computer vision and pattern recognition, pp 2034\u20132042","DOI":"10.1109\/CVPR46437.2021.00207"},{"key":"19271_CR53","doi-asserted-by":"crossref","unstructured":"Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, Lin S, Guo B (2021) Swin transformer: Hierarchical vision transformer using shifted windows. In: Proceedings of the IEEE\/CVF international conference on computer vision, pp 10012\u201310022","DOI":"10.1109\/ICCV48922.2021.00986"},{"key":"19271_CR54","doi-asserted-by":"crossref","unstructured":"Wang S, Zhou T, Lu Y, Di H (2022) Detail-preserving transformer for light field image super-resolution. In: Proceedings of the AAAI conference on artificial intelligence, vol 36, pp 2522\u20132530","DOI":"10.1609\/aaai.v36i3.20153"},{"key":"19271_CR55","doi-asserted-by":"publisher","first-page":"563","DOI":"10.1109\/LSP.2022.3146798","volume":"29","author":"Z Liang","year":"2022","unstructured":"Liang Z, Wang Y, Wang L, Yang J, Zhou S (2022) Light field image super-resolution with transformers. IEEE Signal Process Lett 29:563\u2013567","journal-title":"IEEE Signal Process Lett"},{"key":"19271_CR56","doi-asserted-by":"crossref","unstructured":"Zhang S, Meng N, Lam EY (2023) Lrt: an efficient low-light restoration transformer for dark light field images. IEEE Trans Image Process","DOI":"10.1109\/TIP.2023.3297412"},{"key":"19271_CR57","doi-asserted-by":"crossref","unstructured":"Levoy M, Hanrahan P (1996) Light field rendering. In: Proceedings of the 23rd annual conference on computer graphics and interactive techniques, pp 31\u201342","DOI":"10.1145\/237170.237199"},{"key":"19271_CR58","doi-asserted-by":"crossref","unstructured":"Ledig C, Theis L, Husz\u00e1r F, Caballero J, Cunningham A, Acosta A, Aitken A, Tejani A, Totz J, Wang Z et al (2017) Photo-realistic single image super-resolution using a generative adversarial network. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 4681\u20134690","DOI":"10.1109\/CVPR.2017.19"},{"key":"19271_CR59","unstructured":"Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image recognition. Preprint arXiv:1409.1556"},{"key":"19271_CR60","doi-asserted-by":"crossref","unstructured":"Dansereau DG, Pizarro O, Williams SB (2013) Decoding, calibration and rectification for lenselet-based plenoptic cameras. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 1027\u20131034","DOI":"10.1109\/CVPR.2013.137"},{"issue":"9","key":"19271_CR61","doi-asserted-by":"publisher","first-page":"4274","DOI":"10.1109\/TIP.2018.2834819","volume":"27","author":"Y Wang","year":"2018","unstructured":"Wang Y, Liu F, Zhang K, Hou G, Sun Z, Tan T (2018) Lfnet: A novel bidirectional recurrent convolutional neural network for light-field image super-resolution. IEEE Trans Image Process 27(9):4274\u20134286","journal-title":"IEEE Trans Image Process"},{"issue":"4","key":"19271_CR62","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang Z, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600\u2013612","journal-title":"IEEE Trans Image Process"},{"key":"19271_CR63","unstructured":"Kingma DP, Ba J (2014) Adam: A method for stochastic optimization. Preprint arXiv:1412.6980"},{"key":"19271_CR64","unstructured":"Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, Killeen T, Lin Z, Gimelshein N, Antiga L et al (2019) Pytorch: An imperative style, high-performance deep learning library. Advances in neural information processing systems 32"},{"key":"19271_CR65","doi-asserted-by":"crossref","unstructured":"Hu J, Shen L, Sun G (2018) Squeeze-and-excitation networks. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 7132\u20137141","DOI":"10.1109\/CVPR.2018.00745"},{"issue":"1","key":"19271_CR66","doi-asserted-by":"publisher","first-page":"425","DOI":"10.1109\/TPAMI.2022.3152488","volume":"45","author":"Y Wang","year":"2023","unstructured":"Wang Y, Wang L, Wu G, Yang J, An W, Yu J, Guo Y (2023) Disentangling light fields for super-resolution and disparity estimation. IEEE Trans Pattern Anal Mach Intell 45(1):425\u2013443","journal-title":"IEEE Trans Pattern Anal Mach Intell"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19271-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s11042-024-19271-z\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-024-19271-z.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2025,5,1]],"date-time":"2025-05-01T05:01:30Z","timestamp":1746075690000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s11042-024-19271-z"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,5,17]]},"references-count":66,"journal-issue":{"issue":"12","published-online":{"date-parts":[[2025,4]]}},"alternative-id":["19271"],"URL":"https:\/\/doi.org\/10.1007\/s11042-024-19271-z","relation":{},"ISSN":["1573-7721"],"issn-type":[{"type":"electronic","value":"1573-7721"}],"subject":[],"published":{"date-parts":[[2024,5,17]]},"assertion":[{"value":"21 December 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"20 March 2024","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 April 2024","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"17 May 2024","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors have no competing interests to declare that are relevant to the content of this article.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Competing interests"}}]}}