{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,17]],"date-time":"2025-09-17T15:25:55Z","timestamp":1758122755430,"version":"3.40.3"},"publisher-location":"Cham","reference-count":74,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031250651"},{"type":"electronic","value":"9783031250668"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25066-8_6","type":"book-chapter","created":{"date-parts":[[2023,2,17]],"date-time":"2023-02-17T08:18:05Z","timestamp":1676621885000},"page":"130-152","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":8,"title":["Power Efficient Video Super-Resolution on\u00a0Mobile NPUs with\u00a0Deep Learning, Mobile AI &amp; AIM 2022 Challenge: Report"],"prefix":"10.1007","author":[{"given":"Andrey","family":"Ignatov","sequence":"first","affiliation":[]},{"given":"Radu","family":"Timofte","sequence":"additional","affiliation":[]},{"given":"Cheng-Ming","family":"Chiang","sequence":"additional","affiliation":[]},{"given":"Hsien-Kai","family":"Kuo","sequence":"additional","affiliation":[]},{"given":"Yu-Syuan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Man-Yu","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Allen","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Chia-Ming","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Chih-Cheng","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Jia-Ying","family":"Yong","sequence":"additional","affiliation":[]},{"given":"Hong-Han","family":"Shuai","sequence":"additional","affiliation":[]},{"given":"Wen-Huang","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Zhuang","family":"Jia","sequence":"additional","affiliation":[]},{"given":"Tianyu","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Yijian","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Long","family":"Bao","sequence":"additional","affiliation":[]},{"given":"Heng","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Diankai","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Si","family":"Gao","sequence":"additional","affiliation":[]},{"given":"Shaoli","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Biao","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Xiaofeng","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Chengjian","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Kaidi","family":"Lu","sequence":"additional","affiliation":[]},{"given":"Ning","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Xiao","family":"Sun","sequence":"additional","affiliation":[]},{"given":"HaoDong","family":"Wu","sequence":"additional","affiliation":[]},{"given":"Xuncheng","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Weizhan","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Caixia","family":"Yan","sequence":"additional","affiliation":[]},{"given":"Haipeng","family":"Du","sequence":"additional","affiliation":[]},{"given":"Qinghua","family":"Zheng","sequence":"additional","affiliation":[]},{"given":"Qi","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Wangdu","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Ran","family":"Duan","sequence":"additional","affiliation":[]},{"given":"Mengdi","family":"Sun","sequence":"additional","affiliation":[]},{"given":"Dan","family":"Zhu","sequence":"additional","affiliation":[]},{"given":"Guannan","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Hojin","family":"Cho","sequence":"additional","affiliation":[]},{"given":"Steve","family":"Kim","sequence":"additional","affiliation":[]},{"given":"Shijie","family":"Yue","sequence":"additional","affiliation":[]},{"given":"Chenghua","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zhengyang","family":"Zhuge","sequence":"additional","affiliation":[]},{"given":"Wei","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Wenxu","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Yufeng","family":"Zhou","sequence":"additional","affiliation":[]},{"given":"Xiaochen","family":"Cai","sequence":"additional","affiliation":[]},{"given":"Hengxing","family":"Cai","sequence":"additional","affiliation":[]},{"given":"Kele","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Li","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Zehua","family":"Cheng","sequence":"additional","affiliation":[]},{"given":"Wenyi","family":"Lian","sequence":"additional","affiliation":[]},{"given":"Wenjing","family":"Lian","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"key":"6_CR1","unstructured":"Abdelhamed, A., Afifi, M., Timofte, R., Brown, M.S.: NTIRE 2020 challenge on real image denoising: dataset, methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 496\u2013497 (2020)"},{"key":"6_CR2","unstructured":"Abdelhamed, A., Timofte, R., Brown, M.S.: NTIRE 2019 challenge on real image denoising: methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)"},{"key":"6_CR3","unstructured":"Cai, J., Gu, S., Timofte, R., Zhang, L.: NTIRE 2019 challenge on real image super-resolution: methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)"},{"key":"6_CR4","doi-asserted-by":"crossref","unstructured":"Cai, Y., Yao, Z., Dong, Z., Gholami, A., Mahoney, M.W., Keutzer, K.: ZeroQ: a novel zero shot quantization framework. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13169\u201313178 (2020)","DOI":"10.1109\/CVPR42600.2020.01318"},{"key":"6_CR5","doi-asserted-by":"crossref","unstructured":"Chiang, C.M., et al.: Deploying image deblurring across mobile devices: a perspective of quality and latency. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 502\u2013503 (2020)","DOI":"10.1109\/CVPRW50498.2020.00259"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Conde, M.V., Timofte, R., et al.: Reversed image signal processing and RAW reconstruction. AIM 2022 challenge report. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 3\u201326. Springer, Cham (2023)","DOI":"10.1007\/978-3-031-25066-8_1"},{"key":"6_CR7","doi-asserted-by":"crossref","unstructured":"Du, Z., Liu, J., Tang, J., Wu, G.: Anchor-based plain net for mobile image super-resolution. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00283"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Fuoli, D., Gu, S., Timofte, R.: Efficient video super-resolution through recurrent latent space propagation. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3476\u20133485. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00431"},{"key":"6_CR9","series-title":"LNCS","first-page":"540","volume-title":"ECCV 2022","author":"S Gao","year":"2023","unstructured":"Gao, S., et al.: RCBSR: re-parameterization convolution block for super-resolution. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13802, pp. 540\u2013548. Springer, Cham (2023)"},{"key":"6_CR10","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)"},{"key":"6_CR11","doi-asserted-by":"crossref","unstructured":"Howard, A., et al.: Searching for MobileNetV3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"6_CR12","doi-asserted-by":"crossref","unstructured":"Hui, Z., Gao, X., Yang, Y., Wang, X.: Lightweight image super-resolution with information multi-distillation network. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 2024\u20132032 (2019)","DOI":"10.1145\/3343031.3351084"},{"key":"6_CR13","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Byeoung-su, K., Timofte, R.: Fast camera image denoising on mobile GPUs with deep learning, mobile AI 2021 challenge: report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00285"},{"key":"6_CR14","unstructured":"Ignatov, A., Chiang, J., Kuo, H.K., Sycheva, A., Timofte, R.: Learned smartphone ISP on mobile NPUs with deep learning, mobile AI 2021 challenge: report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Kobyshev, N., Timofte, R., Vanhoey, K., Van Gool, L.: DSLR-quality photos on mobile devices with deep convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3277\u20133285 (2017)","DOI":"10.1109\/ICCV.2017.355"},{"key":"6_CR16","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Kobyshev, N., Timofte, R., Vanhoey, K., Van Gool, L.: WESPE: weakly supervised photo enhancer for digital cameras. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 691\u2013700 (2018)","DOI":"10.1109\/CVPRW.2018.00112"},{"key":"6_CR17","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Malivenko, G., Plowman, D., Shukla, S., Timofte, R.: Fast and accurate single-image depth estimation on mobile devices, mobile AI 2021 challenge: report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00288"},{"key":"6_CR18","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Malivenko, G., Timofte, R.: Fast and accurate quantized camera scene detection on smartphones, mobile AI 2021 challenge: report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00289"},{"key":"6_CR19","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: PyNet-V2 mobile: efficient on-device photo processing with neural networks. In: 2021 26th International Conference on Pattern Recognition (ICPR). IEEE (2022)","DOI":"10.1109\/ICPR56361.2022.9956598"},{"key":"6_CR20","series-title":"LNCS","first-page":"71","volume-title":"ECCV 2022","author":"A Ignatov","year":"2023","unstructured":"Ignatov, A., Malivenko, G., Timofte, R., et al.: Efficient single-image depth estimation on mobile devices, mobile AI & AIM 2022 challenge: report. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 71\u201391. Springer, Cham (2023)"},{"key":"6_CR21","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Patel, J., Timofte, R.: Rendering natural camera bokeh effect with deep learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 418\u2013419 (2020)","DOI":"10.1109\/CVPRW50498.2020.00217"},{"key":"6_CR22","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: AIM 2019 challenge on bokeh effect synthesis: methods and results. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3591\u20133598. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00444"},{"key":"6_CR23","series-title":"LNCS","first-page":"729","volume-title":"ECCV 2022","author":"A Ignatov","year":"2023","unstructured":"Ignatov, A., et al.: MicroISP: processing 32MP photos on mobile devices with deep learning. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13802, pp. 729\u2013746. Springer, Cham (2023)"},{"key":"6_CR24","unstructured":"Ignatov, A., Timofte, R.: NTIRE 2019 challenge on image enhancement: methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)"},{"key":"6_CR25","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"288","DOI":"10.1007\/978-3-030-11021-5_19","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"A Ignatov","year":"2019","unstructured":"Ignatov, A., et al.: AI benchmark: running deep neural networks on Android smartphones. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 288\u2013314. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_19"},{"key":"6_CR26","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Timofte, R., Denna, M., Younes, A.: Real-time quantized image super-resolution on mobile NPUs, mobile AI 2021 challenge: report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00286"},{"key":"6_CR27","series-title":"LNCS","first-page":"92","volume-title":"ECCV 2022","author":"A Ignatov","year":"2023","unstructured":"Ignatov, A., Timofte, R., Denna, M., Younes, A., et al.: Efficient and accurate quantized image super-resolution on mobile NPUs, mobile AI & AIM 2022 challenge: report. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 92\u2013129. Springer, Cham (2023)"},{"key":"6_CR28","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: AIM 2019 challenge on raw to RGB mapping: methods and results. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3584\u20133590. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00443"},{"key":"6_CR29","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: AI benchmark: all about deep learning on smartphones in 2019. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3617\u20133635. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00447"},{"key":"6_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-67070-2_13","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"A Ignatov","year":"2020","unstructured":"Ignatov, A., et al.: AIM 2020 challenge on rendering realistic bokeh. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12537, pp. 213\u2013228. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-67070-2_13"},{"key":"6_CR31","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"315","DOI":"10.1007\/978-3-030-11021-5_20","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"A Ignatov","year":"2019","unstructured":"Ignatov, A., et al.: PIRM challenge on perceptual image enhancement on smartphones: report. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 315\u2013333. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_20"},{"key":"6_CR32","unstructured":"Ignatov, A., et al.: AIM 2020 challenge on learned image signal processing pipeline. arXiv preprint arXiv:2011.04994 (2020)"},{"key":"6_CR33","series-title":"LNCS","first-page":"44","volume-title":"ECCV 2022","author":"A Ignatov","year":"2023","unstructured":"Ignatov, A., Timofte, R., et al.: Learned smartphone ISP on mobile GPUs with deep learning, mobile AI & AIM 2022 challenge: report. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 44\u201370. Springer, Cham (2023)"},{"key":"6_CR34","series-title":"LNCS","first-page":"153","volume-title":"ECCV 2022","author":"A Ignatov","year":"2023","unstructured":"Ignatov, A., Timofte, R., et al.: Realistic bokeh effect rendering on mobile GPUs, mobile AI & AIM 2022 challenge: report. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 153\u2013173. Springer, Cham (2023)"},{"key":"6_CR35","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Van Gool, L., Timofte, R.: Replacing mobile camera ISP with a single deep learning model. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 536\u2013537 (2020)","DOI":"10.1109\/CVPRW50498.2020.00276"},{"key":"6_CR36","doi-asserted-by":"publisher","first-page":"276","DOI":"10.1016\/j.patrec.2020.07.033","volume":"138","author":"D Ignatov","year":"2020","unstructured":"Ignatov, D., Ignatov, A.: Controlling information capacity of binary neural network. Pattern Recogn. Lett. 138, 276\u2013281 (2020)","journal-title":"Pattern Recogn. Lett."},{"key":"6_CR37","unstructured":"Isobe, T., Zhu, F., Jia, X., Wang, S.: Revisiting temporal modeling for video super-resolution. arXiv preprint arXiv:2008.05765 (2020)"},{"key":"6_CR38","doi-asserted-by":"crossref","unstructured":"Jacob, B., et al.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2704\u20132713 (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"6_CR39","unstructured":"Jain, S.R., Gural, A., Wu, M., Dick, C.H.: Trained quantization thresholds for accurate and efficient fixed-point inference of deep neural networks. arXiv preprint arXiv:1903.08066 (2019)"},{"issue":"2","key":"6_CR40","doi-asserted-by":"publisher","first-page":"109","DOI":"10.1109\/TCI.2016.2532323","volume":"2","author":"A Kappeler","year":"2016","unstructured":"Kappeler, A., Yoo, S., Dai, Q., Katsaggelos, A.K.: Video super-resolution with convolutional neural networks. IEEE Trans. Comput. Imaging 2(2), 109\u2013122 (2016)","journal-title":"IEEE Trans. Comput. Imaging"},{"key":"6_CR41","doi-asserted-by":"crossref","unstructured":"Kim, J., Lee, J.K., Lee, K.M.: Accurate image super-resolution using very deep convolutional networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1646\u20131654 (2016)","DOI":"10.1109\/CVPR.2016.182"},{"key":"6_CR42","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"6_CR43","series-title":"LNCS","first-page":"27","volume-title":"ECCV 2022","author":"FO K\u0131nl\u0131","year":"2023","unstructured":"K\u0131nl\u0131, F.O., Mente\u015f, S., \u00d6zcan, B., Kirac, F., Timofte, R., et al.: AIM 2022 challenge on Instagram filter removal: methods and results. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 27\u201343. Springer, Cham (2023)"},{"key":"6_CR44","doi-asserted-by":"crossref","unstructured":"Lee, Y.L., Tsung, P.K., Wu, M.: Technology trend of edge AI. In: 2018 International Symposium on VLSI Design, Automation and Test (VLSI-DAT), pp. 1\u20132. IEEE (2018)","DOI":"10.1109\/VLSI-DAT.2018.8373244"},{"key":"6_CR45","doi-asserted-by":"crossref","unstructured":"Li, Y., Gu, S., Gool, L.V., Timofte, R.: Learning filter basis for convolutional neural network compression. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5623\u20135632 (2019)","DOI":"10.1109\/ICCV.2019.00572"},{"key":"6_CR46","unstructured":"Li, Y., et al.: NTIRE 2022 challenge on efficient super-resolution: methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 1062\u20131102 (2022)"},{"key":"6_CR47","series-title":"LNCS","first-page":"591","volume-title":"ECCV 2022","author":"W Lian","year":"2023","unstructured":"Lian, W., Lian, W.: Sliding window recurrent network for efficient video super-resolution. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13802, pp. 591\u2013601. Springer, Cham (2023)"},{"key":"6_CR48","unstructured":"Lian, W., Peng, S.: Kernel-aware raw burst blind super-resolution. arXiv preprint arXiv:2112.07315 (2021)"},{"key":"6_CR49","unstructured":"Liang, J., et al.: VRT: a video restoration transformer. arXiv preprint arXiv:2201.12288 (2022)"},{"key":"6_CR50","doi-asserted-by":"publisher","first-page":"5981","DOI":"10.1007\/s10462-022-10147-y","volume":"55","author":"H Liu","year":"2022","unstructured":"Liu, H., et al.: Video super-resolution based on deep learning: a comprehensive survey. Artif. Intell. Rev. 55, 5981\u20136035 (2022). https:\/\/doi.org\/10.1007\/s10462-022-10147-y","journal-title":"Artif. Intell. Rev."},{"key":"6_CR51","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1007\/978-3-030-67070-2_2","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"J Liu","year":"2020","unstructured":"Liu, J., Tang, J., Wu, G.: Residual feature distillation network for lightweight image super-resolution. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12537, pp. 41\u201355. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-67070-2_2"},{"key":"6_CR52","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: MetaPruning: meta learning for automatic neural network channel pruning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3296\u20133305 (2019)","DOI":"10.1109\/ICCV.2019.00339"},{"key":"6_CR53","doi-asserted-by":"crossref","unstructured":"Liu, Z., Wu, B., Luo, W., Yang, X., Liu, W., Cheng, K.T.: Bi-real net: enhancing the performance of 1-bit CNNs with improved representational capability and advanced training algorithm. In: Proceedings of the European conference on computer vision (ECCV), pp. 722\u2013737 (2018)","DOI":"10.1007\/978-3-030-01267-0_44"},{"key":"6_CR54","unstructured":"Lugmayr, A., Danelljan, M., Timofte, R.: NTIRE 2020 challenge on real-world image super-resolution: methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 494\u2013495 (2020)"},{"key":"6_CR55","doi-asserted-by":"crossref","unstructured":"Luo, Z., et al.: BSRT: improving burst super-resolution with swin transformer and flow-guided deformable alignment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 998\u20131008 (2022)","DOI":"10.1109\/CVPRW56347.2022.00113"},{"key":"6_CR56","doi-asserted-by":"crossref","unstructured":"Luo, Z., et al.: EBSR: feature enhanced burst super-resolution with deformable alignment. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 471\u2013478 (2021)","DOI":"10.1109\/CVPRW53098.2021.00058"},{"key":"6_CR57","doi-asserted-by":"crossref","unstructured":"Nah, S., et al.: NTIRE 2019 challenge on video deblurring and super-resolution: dataset and study. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00251"},{"key":"6_CR58","unstructured":"Nah, S., Son, S., Timofte, R., Lee, K.M.: NTIRE 2020 challenge on image and video deblurring. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 416\u2013417 (2020)"},{"key":"6_CR59","doi-asserted-by":"crossref","unstructured":"Nah, S., et al.: NTIRE 2019 challenge on video super-resolution: methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00251"},{"key":"6_CR60","unstructured":"Obukhov, A., Rakhuba, M., Georgoulis, S., Kanakis, M., Dai, D., Van Gool, L.: T-basis: a compact representation for neural networks. In: International Conference on Machine Learning, pp. 7392\u20137404. PMLR (2020)"},{"key":"6_CR61","doi-asserted-by":"crossref","unstructured":"Romero, A., Ignatov, A., Kim, H., Timofte, R.: Real-time video super-resolution on smartphones with deep learning, mobile AI 2021 challenge: report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2021)","DOI":"10.1109\/CVPRW53098.2021.00287"},{"key":"6_CR62","doi-asserted-by":"crossref","unstructured":"Sajjadi, M.S., Vemulapalli, R., Brown, M.: Frame-recurrent video super-resolution. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6626\u20136634 (2018)","DOI":"10.1109\/CVPR.2018.00693"},{"key":"6_CR63","doi-asserted-by":"crossref","unstructured":"Shi, W., et al.: Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1874\u20131883 (2016)","DOI":"10.1109\/CVPR.2016.207"},{"key":"6_CR64","doi-asserted-by":"crossref","unstructured":"Tan, M., et al.: MnasNet: platform-aware neural architecture search for mobile. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2820\u20132828 (2019)","DOI":"10.1109\/CVPR.2019.00293"},{"key":"6_CR65","unstructured":"TensorFlow-Lite. https:\/\/www.tensorflow.org\/lite"},{"key":"6_CR66","unstructured":"Timofte, R., Gu, S., Wu, J., Van Gool, L.: NTIRE 2018 challenge on single image super-resolution: methods and results. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 852\u2013863 (2018)"},{"key":"6_CR67","unstructured":"Uhlich, S., et al.: Mixed precision DNNs: all you need is a good parametrization. arXiv preprint arXiv:1905.11452 (2019)"},{"key":"6_CR68","doi-asserted-by":"crossref","unstructured":"Wan, A., et al.: FBNetV2: differentiable neural architecture search for spatial and channel dimensions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12965\u201312974 (2020)","DOI":"10.1109\/CVPR42600.2020.01298"},{"key":"6_CR69","doi-asserted-by":"crossref","unstructured":"Wang, X., Chan, K.C., Yu, K., Dong, C., Change Loy, C.: EDVR: video restoration with enhanced deformable convolutional networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (2019)","DOI":"10.1109\/CVPRW.2019.00247"},{"key":"6_CR70","doi-asserted-by":"crossref","unstructured":"Wu, B., et al.: FBNet: hardware-aware efficient convnet design via differentiable neural architecture search. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10734\u201310742 (2019)","DOI":"10.1109\/CVPR.2019.01099"},{"key":"6_CR71","doi-asserted-by":"crossref","unstructured":"Yang, J., et al.: Quantization networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7308\u20137316 (2019)","DOI":"10.1109\/CVPR.2019.00748"},{"key":"6_CR72","series-title":"LNCS","first-page":"174","volume-title":"ECCV 2022","author":"R Yang","year":"2023","unstructured":"Yang, R., Timofte, R., et al.: AIM 2022 challenge on super-resolution of compressed image and video: dataset, methods and results. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13803, pp. 174\u2013202. Springer, Cham (2023)"},{"key":"6_CR73","doi-asserted-by":"crossref","unstructured":"Yue, S., Li, C., Zhuge, Z., Song, R.: EESRNet: a network for energy efficient super-resolution. In: Karlinsky, L., et al. (eds.) ECCV 2022. LNCS, vol. 13802, pp. xx\u2013yy. Springer, Cham (2023)","DOI":"10.1007\/978-3-031-25063-7_38"},{"key":"6_CR74","doi-asserted-by":"crossref","unstructured":"Zhang, X., Zeng, H., Zhang, L.: Edge-oriented convolution block for real-time super resolution on mobile devices. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 4034\u20134043 (2021)","DOI":"10.1145\/3474085.3475291"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25066-8_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T13:12:10Z","timestamp":1728911530000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25066-8_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250651","9783031250668"],"references-count":74,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25066-8_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}