{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,8]],"date-time":"2026-02-08T06:13:05Z","timestamp":1770531185523,"version":"3.49.0"},"publisher-location":"Cham","reference-count":65,"publisher":"Springer Nature Switzerland","isbn-type":[{"value":"9783031726453","type":"print"},{"value":"9783031726460","type":"electronic"}],"license":[{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T00:00:00Z","timestamp":1730073600000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2025]]},"DOI":"10.1007\/978-3-031-72646-0_13","type":"book-chapter","created":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T08:45:29Z","timestamp":1730105129000},"page":"218-235","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["High-Resolution and\u00a0Few-Shot View Synthesis from\u00a0Asymmetric Dual-Lens Inputs"],"prefix":"10.1007","author":[{"given":"Ruikang","family":"Xu","sequence":"first","affiliation":[]},{"given":"Mingde","family":"Yao","sequence":"additional","affiliation":[]},{"given":"Yue","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yueyi","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Zhiwei","family":"Xiong","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,10,28]]},"reference":[{"key":"13_CR1","doi-asserted-by":"crossref","unstructured":"Alzayer, H., et al.: DC2: dual-camera defocus control by learning to refocus. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.02058"},{"key":"13_CR2","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: MIP-nerf 360: unbounded anti-aliased neural radiance fields. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00539"},{"key":"13_CR3","doi-asserted-by":"crossref","unstructured":"Barron, J.T., Mildenhall, B., Verbin, D., Srinivasan, P.P., Hedman, P.: Zip-nerf: anti-aliased grid-based neural radiance fields. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.01804"},{"key":"13_CR4","unstructured":"Bhat, S.F., Birkl, R., Wofk, D., Wonka, P., M\u00fcller, M.: Zoedepth: zero-shot transfer by combining relative and metric depth. arXiv preprint arXiv:2302.12288 (2023)"},{"key":"13_CR5","unstructured":"Bhoi, A.: Monocular depth estimation: a survey. arXiv preprint arXiv:1901.09402 (2019)"},{"key":"13_CR6","doi-asserted-by":"crossref","unstructured":"Cao, J., et al.: Real-time neural light field on mobile devices. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00805"},{"key":"13_CR7","doi-asserted-by":"crossref","unstructured":"Chan, K.C., Wang, X., Yu, K., Dong, C., Loy, C.C.: Basicvsr: the search for essential components in video super-resolution and beyond. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00491"},{"key":"13_CR8","doi-asserted-by":"crossref","unstructured":"Chen, A., et al.: Mvsnerf: fast generalizable radiance field reconstruction from multi-view stereo. In: CVPR (2021)","DOI":"10.1109\/ICCV48922.2021.01386"},{"key":"13_CR9","doi-asserted-by":"crossref","unstructured":"Chen, X., Wang, X., Zhou, J., Qiao, Y., Dong, C.: Activating more pixels in image super-resolution transformer. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.02142"},{"key":"13_CR10","doi-asserted-by":"crossref","unstructured":"Chen, X., Xiong, Z., Cheng, Z., Peng, J., Zhang, Y., Zha, Z.J.: Degradation-agnostic correspondence from resolution-asymmetric stereo. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01262"},{"key":"13_CR11","doi-asserted-by":"crossref","unstructured":"Chen, Z., Funkhouser, T., Hedman, P., Tagliasacchi, A.: Mobilenerf: exploiting the polygon rasterization pipeline for efficient neural field rendering on mobile architectures. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01590"},{"key":"13_CR12","doi-asserted-by":"crossref","unstructured":"Chung, J., Oh, J., Lee, K.M.: Depth-regularized optimization for 3D gaussian splatting in few-shot images. arXiv preprint arXiv:2311.13398 (2023)","DOI":"10.1109\/CVPRW63382.2024.00086"},{"key":"13_CR13","doi-asserted-by":"crossref","unstructured":"Deng, K., Liu, A., Zhu, J.Y., Ramanan, D.: Depth-supervised nerf: fewer views and faster training for free. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01254"},{"issue":"11","key":"13_CR14","doi-asserted-by":"publisher","first-page":"3854","DOI":"10.1109\/TVCG.2022.3203102","volume":"28","author":"N Deng","year":"2022","unstructured":"Deng, N., et al.: FoV-NeRF: foveated neural radiance fields for virtual reality. IEEE Trans. Visual Comput. Graphics 28(11), 3854\u20133864 (2022)","journal-title":"IEEE Trans. Visual Comput. Graphics"},{"key":"13_CR15","doi-asserted-by":"crossref","unstructured":"Dong, J., Fang, Q., Yang, T., Shuai, Q., Qiao, C., Peng, S.: iVS-Net: learning human view synthesis from internet videos. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.02097"},{"key":"13_CR16","unstructured":"Hattori, H., Maki, A.: Stereo without depth search and metric calibration. In: CVPR (2000)"},{"key":"13_CR17","doi-asserted-by":"crossref","unstructured":"Hu, T., Liu, S., Chen, Y., Shen, T., Jia, J.: Efficientnerf efficient neural radiance fields. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01256"},{"key":"13_CR18","doi-asserted-by":"crossref","unstructured":"Huang, X., Li, W., Hu, J., Chen, H., Wang, Y.: RefSR-NeRF: towards high fidelity and super resolution view synthesis. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00797"},{"key":"13_CR19","doi-asserted-by":"crossref","unstructured":"Kerbl, B., Kopanas, G., Leimk\u00fchler, T., Drettakis, G.: 3D gaussian splatting for real-time radiance field rendering. ACM Trans. Graph. 42(4) (2023)","DOI":"10.1145\/3592433"},{"key":"13_CR20","doi-asserted-by":"crossref","unstructured":"Kim, M., Seo, S., Han, B.: Infonerf: ray entropy minimization for few-shot neural volume rendering. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01257"},{"key":"13_CR21","doi-asserted-by":"crossref","unstructured":"Lai, W.S., Huang, J.B., Wang, O., Shechtman, E., Yumer, E., Yang, M.H.: Learning blind video temporal consistency. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01267-0_11"},{"key":"13_CR22","doi-asserted-by":"crossref","unstructured":"Larsson, V., Zobernig, N., Taskin, K., Pollefeys, M.: Calibration-free structure-from-motion with calibrated radial trifocal tensors. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58558-7_23"},{"key":"13_CR23","doi-asserted-by":"crossref","unstructured":"Lee, J., Lee, M., Cho, S., Lee, S.: Reference-based video super-resolution using multi-camera video triplets. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01730"},{"key":"13_CR24","doi-asserted-by":"crossref","unstructured":"Li, Q., Li, F., Guo, J., Guo, Y.: Uhdnerf: ultra-high-definition neural radiance fields. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.02111"},{"key":"13_CR25","doi-asserted-by":"crossref","unstructured":"Liang, J., Cao, J., Sun, G., Zhang, K., Van\u00a0Gool, L., Timofte, R.: Swinir: image restoration using swin transformer. In: CVPRW (2021)","DOI":"10.1109\/ICCVW54120.2021.00210"},{"key":"13_CR26","doi-asserted-by":"crossref","unstructured":"Lin, C.Y., Fu, Q., Merth, T., Yang, K., Ranjan, A.: Fastsr-nerf: improving nerf efficiency on consumer devices with a simple super-resolution pipeline. In: WACV (2024)","DOI":"10.1109\/WACV57701.2024.00593"},{"key":"13_CR27","doi-asserted-by":"crossref","unstructured":"Manne, S.K.R., Prasad, B., Rosh, K.: Asymmetric wide tele camera fusion for high fidelity digital zoom. In: ICCVIP (2019)","DOI":"10.1007\/978-981-15-4015-8_4"},{"key":"13_CR28","doi-asserted-by":"crossref","unstructured":"Mechrez, R., Talmi, I., Shama, F., Zelnik-Manor, L.: Maintaining natural image statistics with the contextual loss. In: ACCV (2019)","DOI":"10.1007\/978-3-030-20893-6_27"},{"key":"13_CR29","doi-asserted-by":"crossref","unstructured":"Mechrez, R., Talmi, I., Zelnik-Manor, L.: The contextual loss for image transformation with non-aligned data. In: ECCV (2018)","DOI":"10.1007\/978-3-030-01264-9_47"},{"issue":"1","key":"13_CR30","doi-asserted-by":"publisher","first-page":"99","DOI":"10.1145\/3503250","volume":"65","author":"B Mildenhall","year":"2021","unstructured":"Mildenhall, B., Srinivasan, P.P., Tancik, M., Barron, J.T., Ramamoorthi, R., Ng, R.: Nerf: representing scenes as neural radiance fields for view synthesis. Commun. ACM 65(1), 99\u2013106 (2021)","journal-title":"Commun. ACM"},{"key":"13_CR31","doi-asserted-by":"publisher","first-page":"4479","DOI":"10.1109\/TIP.2021.3072856","volume":"30","author":"MM Mohan","year":"2021","unstructured":"Mohan, M.M., Nithin, G., Rajagopalan, A.: Deep dynamic scene deblurring for unconstrained dual-lens cameras. IEEE Trans. Image Process. 30, 4479\u20134491 (2021)","journal-title":"IEEE Trans. Image Process."},{"issue":"4","key":"13_CR32","doi-asserted-by":"publisher","first-page":"1","DOI":"10.1145\/3528223.3530127","volume":"41","author":"T M\u00fcller","year":"2022","unstructured":"M\u00fcller, T., Evans, A., Schied, C., Keller, A.: Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. (ToG) 41(4), 1\u201315 (2022)","journal-title":"ACM Trans. Graph. (ToG)"},{"key":"13_CR33","doi-asserted-by":"crossref","unstructured":"Niemeyer, M., Barron, J.T., Mildenhall, B., Sajjadi, M.S., Geiger, A., Radwan, N.: Regnerf: regularizing neural radiance fields for view synthesis from sparse inputs. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00540"},{"key":"13_CR34","doi-asserted-by":"crossref","unstructured":"Ranftl, R., Bochkovskiy, A., Koltun, V.: Vision transformers for dense prediction. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.01196"},{"key":"13_CR35","doi-asserted-by":"crossref","unstructured":"Saito, S., Simon, T., Saragih, J., Joo, H.: Pifuhd: multi-level pixel-aligned implicit function for high-resolution 3D human digitization. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00016"},{"key":"13_CR36","unstructured":"Santesteban, I., Otaduy, M., Thuerey, N., Casas, D.: Ulnef: untangled layered neural fields for mix-and-match virtual try-on. In: NIPS (2022)"},{"key":"13_CR37","doi-asserted-by":"crossref","unstructured":"Schonberger, J.L., Frahm, J.M.: Structure-from-motion revisited. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.445"},{"key":"13_CR38","doi-asserted-by":"crossref","unstructured":"Sedgwick, P.: Pearson\u2019s correlation coefficient. BMJ 345 (2012)","DOI":"10.1136\/bmj.e4483"},{"key":"13_CR39","doi-asserted-by":"crossref","unstructured":"Seo, S., Chang, Y., Kwak, N.: Flipnerf: flipped reflection rays for few-shot novel view synthesis. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.02092"},{"key":"13_CR40","doi-asserted-by":"crossref","unstructured":"Shao, R., et al.: Doublefield: bridging the neural surface and radiance fields for high-fidelity human reconstruction and rendering. In: CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.01541"},{"key":"13_CR41","doi-asserted-by":"crossref","unstructured":"Somraj, N., Soundararajan, R.: VIP-nerf: visibility prior for sparse input neural radiance fields. In: ACM SIGGRAPH (2023)","DOI":"10.1145\/3588432.3591539"},{"key":"13_CR42","unstructured":"Song, J., et al.: D\u00e4rf: boosting radiance fields from sparse input views with monocular depth adaptation. In: NIPS (2023)"},{"key":"13_CR43","doi-asserted-by":"crossref","unstructured":"Song, T., Kim, S., Sohn, K.: Unsupervised deep asymmetric stereo matching with spatially-adaptive self-similarity. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01314"},{"key":"13_CR44","unstructured":"Tang, J., Ren, J., Zhou, H., Liu, Z., Zeng, G.: Dreamgaussian: generative gaussian splatting for efficient 3D content creation. arXiv preprint arXiv:2309.16653 (2023)"},{"key":"13_CR45","doi-asserted-by":"crossref","unstructured":"Teed, Z., Deng, J.: Raft: recurrent all-pairs field transforms for optical flow. In: ECCV (2020)","DOI":"10.1007\/978-3-030-58536-5_24"},{"key":"13_CR46","doi-asserted-by":"crossref","unstructured":"Tian, Y., Zhang, Y., Fu, Y., Xu, C.: TDAN: temporally-deformable alignment network for video super-resolution. In: CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00342"},{"key":"13_CR47","doi-asserted-by":"crossref","unstructured":"Tosi, F., Tonioni, A., De\u00a0Gregorio, D., Poggi, M.: Nerf-supervised deep stereo. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00089"},{"key":"13_CR48","doi-asserted-by":"crossref","unstructured":"Uy, M.A., Martin-Brualla, R., Guibas, L., Li, K.: Scade: nerfs from space carving with ambiguity-aware depth estimates. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01585"},{"key":"13_CR49","doi-asserted-by":"crossref","unstructured":"Wang, C., Wu, X., Guo, Y.C., Zhang, S.H., Tai, Y.W., Hu, S.M.: Nerf-SR: high quality neural radiance fields using supersampling. In: ACM MM (2022)","DOI":"10.1145\/3503161.3547808"},{"key":"13_CR50","doi-asserted-by":"crossref","unstructured":"Wang, G., Chen, Z., Loy, C.C., Liu, Z.: Sparsenerf: distilling depth ranking for few-shot novel view synthesis. In: ICCV (2023)","DOI":"10.1109\/ICCV51070.2023.00832"},{"key":"13_CR51","doi-asserted-by":"crossref","unstructured":"Wang, T., Xie, J., Sun, W., Yan, Q., Chen, Q.: Dual-camera super-resolution with aligned attention modules. In: ICCV (2021)","DOI":"10.1109\/ICCV48922.2021.00201"},{"key":"13_CR52","doi-asserted-by":"crossref","unstructured":"Wang, X., Chan, K.C., Yu, K., Dong, C., Change\u00a0Loy, C.: EDVR: video restoration with enhanced deformable convolutional networks. In: CVPRW (2019)","DOI":"10.1109\/CVPRW.2019.00247"},{"issue":"1","key":"13_CR53","doi-asserted-by":"publisher","first-page":"425","DOI":"10.1109\/TPAMI.2022.3152488","volume":"45","author":"Y Wang","year":"2022","unstructured":"Wang, Y., et al.: Disentangling light fields for super-resolution and disparity estimation. IEEE Trans. Pattern Anal. Mach. Intell. 45(1), 425\u2013443 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"13_CR54","unstructured":"Wang, Z., Simoncelli, E.P., Bovik, A.C.: Multiscale structural similarity for image quality assessment. In: The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, 2003, vol.\u00a02, pp. 1398\u20131402. IEEE (2003)"},{"key":"13_CR55","unstructured":"Xiong, H., Muttukuru, S., Upadhyay, R., Chari, P., Kadambi, A.: Sparsegs: real-time 360 sparse view synthesis using gaussian splatting. arXiv preprint arXiv:2312.00206 (2023)"},{"key":"13_CR56","doi-asserted-by":"crossref","unstructured":"Xu, R., Yao, M., Xiong, Z.: Zero-shot dual-lens super-resolution. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00881"},{"key":"13_CR57","doi-asserted-by":"crossref","unstructured":"Yang, J., Pavone, M., Wang, Y.: Freenerf: improving few-shot neural rendering with free frequency regularization. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.00798"},{"key":"13_CR58","doi-asserted-by":"crossref","unstructured":"Yoon, Y., Yoon, K.J.: Cross-guided optimization of radiance fields with multi-view image super-resolution for high-resolution novel view synthesis. In: CVPR (2023)","DOI":"10.1109\/CVPR52729.2023.01196"},{"key":"13_CR59","doi-asserted-by":"crossref","unstructured":"Yu, A., Ye, V., Tancik, M., Kanazawa, A.: pixelNeRF: neural radiance fields from one or few images. In: CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.00455"},{"key":"13_CR60","doi-asserted-by":"crossref","unstructured":"Yue, H., Cui, Z., Li, K., Yang, J.: Kedusr: real-world dual-lens super-resolution via kernel-free matching. arXiv preprint arXiv:2312.17050 (2023)","DOI":"10.1609\/aaai.v38i7.28513"},{"key":"13_CR61","doi-asserted-by":"crossref","unstructured":"Zhang, J., et al.: Mobidepth: real-time depth estimation using on-device dual cameras. In: Proceedings of the 28th Annual International Conference on Mobile Computing and Networking, pp. 528\u2013541 (2022)","DOI":"10.1145\/3495243.3560517"},{"key":"13_CR62","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"key":"13_CR63","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Wang, R., Zhang, H., Chen, Y., Zuo, W.: Self-supervised learning for real-world super-resolution from dual zoomed observations. In: ECCV (2022)","DOI":"10.1007\/978-3-031-19797-0_35"},{"issue":"4","key":"13_CR64","first-page":"4396","volume":"45","author":"K Zhou","year":"2022","unstructured":"Zhou, K., Liu, Z., Qiao, Y., Xiang, T., Loy, C.C.: Domain generalization: a survey. IEEE Trans. Pattern Anal. Mach. Intell. 45(4), 4396\u20134415 (2022)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"13_CR65","doi-asserted-by":"crossref","unstructured":"Zhu, Z., Fan, Z., Jiang, Y., Wang, Z.: FSGS: real-time few-shot view synthesis using gaussian splatting. arXiv preprint arXiv:2312.00451 (2023)","DOI":"10.1007\/978-3-031-72933-1_9"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2024"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-72646-0_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,28]],"date-time":"2024-10-28T08:51:57Z","timestamp":1730105517000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-72646-0_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,10,28]]},"ISBN":["9783031726453","9783031726460"],"references-count":65,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-72646-0_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"value":"0302-9743","type":"print"},{"value":"1611-3349","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,10,28]]},"assertion":[{"value":"28 October 2024","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Milan","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2024","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"29 September 2024","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"4 October 2024","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"18","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2024","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2024.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}}]}}