{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,9,17]],"date-time":"2025-09-17T15:21:42Z","timestamp":1758122502616,"version":"3.40.3"},"publisher-location":"Cham","reference-count":81,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031250651"},{"type":"electronic","value":"9783031250668"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25066-8_4","type":"book-chapter","created":{"date-parts":[[2023,2,17]],"date-time":"2023-02-17T08:18:05Z","timestamp":1676621885000},"page":"71-91","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":9,"title":["Efficient Single-Image Depth Estimation on\u00a0Mobile Devices, Mobile AI &amp; AIM 2022 Challenge: Report"],"prefix":"10.1007","author":[{"given":"Andrey","family":"Ignatov","sequence":"first","affiliation":[]},{"given":"Grigory","family":"Malivenko","sequence":"additional","affiliation":[]},{"given":"Radu","family":"Timofte","sequence":"additional","affiliation":[]},{"given":"Lukasz","family":"Treszczotko","sequence":"additional","affiliation":[]},{"given":"Xin","family":"Chang","sequence":"additional","affiliation":[]},{"given":"Piotr","family":"Ksiazek","sequence":"additional","affiliation":[]},{"given":"Michal","family":"Lopuszynski","sequence":"additional","affiliation":[]},{"given":"Maciej","family":"Pioro","sequence":"additional","affiliation":[]},{"given":"Rafal","family":"Rudnicki","sequence":"additional","affiliation":[]},{"given":"Maciej","family":"Smyl","sequence":"additional","affiliation":[]},{"given":"Yujie","family":"Ma","sequence":"additional","affiliation":[]},{"given":"Zhenyu","family":"Li","sequence":"additional","affiliation":[]},{"given":"Zehui","family":"Chen","sequence":"additional","affiliation":[]},{"given":"Jialei","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Xianming","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Junjun","family":"Jiang","sequence":"additional","affiliation":[]},{"given":"XueChao","family":"Shi","sequence":"additional","affiliation":[]},{"given":"Difan","family":"Xu","sequence":"additional","affiliation":[]},{"given":"Yanan","family":"Li","sequence":"additional","affiliation":[]},{"given":"Xiaotao","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Lei","family":"Lei","sequence":"additional","affiliation":[]},{"given":"Ziyu","family":"Zhang","sequence":"additional","affiliation":[]},{"given":"Yicheng","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zilong","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Guozhong","family":"Luo","sequence":"additional","affiliation":[]},{"given":"Gang","family":"Yu","sequence":"additional","affiliation":[]},{"given":"Bin","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Jiaqi","family":"Li","sequence":"additional","affiliation":[]},{"given":"Yiran","family":"Wang","sequence":"additional","affiliation":[]},{"given":"Zihao","family":"Huang","sequence":"additional","affiliation":[]},{"given":"Zhiguo","family":"Cao","sequence":"additional","affiliation":[]},{"given":"Marcos V.","family":"Conde","sequence":"additional","affiliation":[]},{"given":"Denis","family":"Sapozhnikov","sequence":"additional","affiliation":[]},{"given":"Byeong Hyun","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Dongwon","family":"Park","sequence":"additional","affiliation":[]},{"given":"Seongmin","family":"Hong","sequence":"additional","affiliation":[]},{"given":"Joonhee","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Seunggyu","family":"Lee","sequence":"additional","affiliation":[]},{"given":"Se Young","family":"Chun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,18]]},"reference":[{"key":"4_CR1","unstructured":"Abdelhamed, A., Afifi, M., Timofte, R., Brown, M.S.: Ntire 2020 challenge on real image denoising: Dataset, methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 496\u2013497 (2020)"},{"key":"4_CR2","unstructured":"Abdelhamed, A., Timofte, R., Brown, M.S.: Ntire 2019 challenge on real image denoising: Methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2019)"},{"key":"4_CR3","unstructured":"Archive, A.B.: http:\/\/web.archive.org\/web\/20210425131428\/https:\/\/ai-benchmark.com\/ranking_processors.html"},{"key":"4_CR4","doi-asserted-by":"crossref","unstructured":"Barron, J.T.: A general and adaptive robust loss function. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 4331\u20134339 (2019)","DOI":"10.1109\/CVPR.2019.00446"},{"key":"4_CR5","unstructured":"Bhardwaj, K., et al.: Collapsible linear blocks for super-efficient super resolution. In: Marculescu, D., Chi, Y., Wu, C. (eds.) Proceedings of Machine Learning and Systems. vol. 4, pp. 529\u2013547 (2022). https:\/\/proceedings.mlsys.org\/paper\/2022\/file\/ac627ab1ccbdb62ec96e702f07f6425b-Paper.pdf"},{"key":"4_CR6","unstructured":"Cai, J., Gu, S., Timofte, R., Zhang, L.: Ntire 2019 challenge on real image super-resolution: Methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2019)"},{"key":"4_CR7","doi-asserted-by":"crossref","unstructured":"Cai, Y., Yao, Z., Dong, Z., Gholami, A., Mahoney, M.W., Keutzer, K.: Zeroq: a novel zero shot quantization framework. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 13169\u201313178 (2020)","DOI":"10.1109\/CVPR42600.2020.01318"},{"key":"4_CR8","doi-asserted-by":"publisher","unstructured":"Chen, L.C., Papandreou, G., Schroff, F., Adam, H.: Rethinking atrous convolution for semantic image segmentation (2017). https:\/\/doi.org\/10.48550\/ARXIV.1706.05587, https:\/\/arxiv.org\/abs\/1706.05587","DOI":"10.48550\/ARXIV.1706.05587"},{"key":"4_CR9","unstructured":"Chen, W., Fu, Z., Yang, D., Deng, J.: Single-image depth perception in the wild. arXiv preprint arXiv:1604.03901 (2016)"},{"key":"4_CR10","doi-asserted-by":"crossref","unstructured":"Chiang, C.M., et al.: Deploying image deblurring across mobile devices: a perspective of quality and latency. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 502\u2013503 (2020)","DOI":"10.1109\/CVPRW50498.2020.00259"},{"key":"4_CR11","doi-asserted-by":"publisher","unstructured":"Conde, M.V., Timofte, R., et al.: Reversed image signal processing and RAW reconstruction. AIM 2022 challenge report. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022). https:\/\/doi.org\/10.1007\/978-3-030-66415-2","DOI":"10.1007\/978-3-030-66415-2"},{"key":"4_CR12","unstructured":"Dosovitskiy, A., et al.: An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:2010.11929 (2020)"},{"key":"4_CR13","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. arXiv preprint arXiv:1406.2283 (2014)"},{"key":"4_CR14","unstructured":"Eigen, D., Puhrsch, C., Fergus, R.: Depth map prediction from a single image using a multi-scale deep network. In: 27th Proceedings on Advances in Neural Information Processing Systems (2014)"},{"key":"4_CR15","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1007\/978-3-319-46484-8_45","volume-title":"Computer Vision \u2013 ECCV 2016","author":"R Garg","year":"2016","unstructured":"Garg, R., B.G., V.K., Carneiro, G., Reid, I.: Unsupervised CNN for single view depth estimation: geometry to the rescue. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9912, pp. 740\u2013756. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46484-8_45"},{"key":"4_CR16","doi-asserted-by":"crossref","unstructured":"Godard, C., Mac Aodha, O., Firman, M., Brostow, G.J.: Digging into self-supervised monocular depth estimation. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3828\u20133838 (2019)","DOI":"10.1109\/ICCV.2019.00393"},{"key":"4_CR17","doi-asserted-by":"publisher","unstructured":"Han, K., Wang, Y., Tian, Q., Guo, J., Xu, C., Xu, C.: Ghostnet: more features from cheap operations (2019). https:\/\/doi.org\/10.48550\/ARXIV.1911.11907, https:\/\/arxiv.org\/abs\/1911.11907","DOI":"10.48550\/ARXIV.1911.11907"},{"key":"4_CR18","unstructured":"Hinton, G., Vinyals, O., Dean, J.: Distilling the knowledge in a neural network. arXiv preprint arXiv:1503.02531 (2015)"},{"key":"4_CR19","doi-asserted-by":"crossref","unstructured":"Howard, A., et al.: Searching for mobilenetv3. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 1314\u20131324 (2019)","DOI":"10.1109\/ICCV.2019.00140"},{"key":"4_CR20","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Byeoung-su, K., Timofte, R.: Fast camera image denoising on mobile GPUs with deep learning, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2021)","DOI":"10.1109\/CVPRW53098.2021.00285"},{"key":"4_CR21","unstructured":"Ignatov, A., Chiang, J., Kuo, H.K., Sycheva, A., Timofte, R.: Learned smartphone ISP on mobile NPUs with deep learning, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2021)"},{"key":"4_CR22","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Kobyshev, N., Timofte, R., Vanhoey, K., Van Gool, L.: Dslr-quality photos on mobile devices with deep convolutional networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3277\u20133285 (2017)","DOI":"10.1109\/ICCV.2017.355"},{"key":"4_CR23","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Kobyshev, N., Timofte, R., Vanhoey, K., Van Gool, L.: Wespe: weakly supervised photo enhancer for digital cameras. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 691\u2013700 (2018)","DOI":"10.1109\/CVPRW.2018.00112"},{"key":"4_CR24","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Malivenko, G., Plowman, D., Shukla, S., Timofte, R.: Fast and accurate single-image depth estimation on mobile devices, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2021)","DOI":"10.1109\/CVPRW53098.2021.00288"},{"key":"4_CR25","unstructured":"Ignatov, A., Malivenko, G., Timofte, R.: Fast and accurate quantized camera scene detection on smartphones, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshop,. pp. 0\u20130 (2021)"},{"key":"4_CR26","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Pynet-v2 mobile: Efficient on-device photo processing with neural networks. In: 2021 26th International Conference on Pattern Recognition (ICPR), IEEE (2022)","DOI":"10.1109\/ICPR56361.2022.9956598"},{"key":"4_CR27","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Patel, J., Timofte, R.: Rendering natural camera bokeh effect with deep learning. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 418\u2013419 (2020)","DOI":"10.1109\/CVPRW50498.2020.00217"},{"key":"4_CR28","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Aim 2019 challenge on bokeh effect synthesis: methods and results. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3591\u20133598. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00444"},{"key":"4_CR29","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Romero, A., Kim, H., Timofte, R.: Real-time video super-resolution on smartphones with deep learning, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2021)","DOI":"10.1109\/CVPRW53098.2021.00287"},{"key":"4_CR30","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: MicroISP: processing 32mp photos on mobile devices with deep learning. In: European Conference on Computer Vision (2022)","DOI":"10.1007\/978-3-031-25063-7_46"},{"key":"4_CR31","unstructured":"Ignatov, A., Timofte, R.: Ntire 2019 challenge on image enhancement: Methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 0\u20130 (2019)"},{"key":"4_CR32","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Power efficient super-resolution on mobile NPUs with deep learning, mobile AI & aim 2022 challenge: Report. In: European Conference on Computer Vision (2022)","DOI":"10.1109\/CVPRW53098.2021.00286"},{"key":"4_CR33","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"288","DOI":"10.1007\/978-3-030-11021-5_19","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"A Ignatov","year":"2019","unstructured":"Ignatov, A., et al.: AI benchmark: running deep neural networks on android smartphones. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 288\u2013314. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_19"},{"key":"4_CR34","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Timofte, R., Denna, M., Younes, A.: Real-time quantized image super-resolution on mobile NPUs, mobile AI 2021 challenge: Report. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshop,. pp. 0\u20130 (2021)","DOI":"10.1109\/CVPRW53098.2021.00286"},{"key":"4_CR35","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Timofte, R., Denna, M., Younes, A., et al.: Efficient and accurate quantized image super-resolution on mobile NPUs, mobile AI & aim 2022 challenge: Report. In: European Conference on Computer Vision (2022)","DOI":"10.1109\/CVPRW53098.2021.00286"},{"key":"4_CR36","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: Aim 2019 challenge on raw to RGB mapping: methods and results. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW)., pp. 3584\u20133590. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00443"},{"key":"4_CR37","doi-asserted-by":"crossref","unstructured":"Ignatov, A., et al.: AI benchmark: All about deep learning on smartphones in 2019. In: 2019 IEEE\/CVF International Conference on Computer Vision Workshop (ICCVW), pp. 3617\u20133635. IEEE (2019)","DOI":"10.1109\/ICCVW.2019.00447"},{"key":"4_CR38","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/978-3-030-67070-2_13","volume-title":"Computer Vision \u2013 ECCV 2020 Workshops","author":"A Ignatov","year":"2020","unstructured":"Ignatov, A., et al.: AIM 2020 challenge on rendering realistic bokeh. In: Bartoli, A., Fusiello, A. (eds.) ECCV 2020. LNCS, vol. 12537, pp. 213\u2013228. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-67070-2_13"},{"key":"4_CR39","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"315","DOI":"10.1007\/978-3-030-11021-5_20","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"A Ignatov","year":"2019","unstructured":"Ignatov, A., et al.: PIRM challenge on perceptual image enhancement on smartphones: report. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11133, pp. 315\u2013333. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11021-5_20"},{"key":"4_CR40","unstructured":"Ignatov, A., et al.: Aim 2020 challenge on learned image signal processing pipeline. arXiv preprint arXiv:2011.04994 (2020)"},{"key":"4_CR41","unstructured":"Ignatov, A., Timofte, R., et al.: Learned smartphone ISP on mobile GPUs with deep learning, mobile AI & aim 2022 challenge: Report. In: European Conference on Computer Vision (2022)"},{"key":"4_CR42","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Timofte, R., et al.: Realistic bokeh effect rendering on mobile GPUs, mobile AI & aim 2022 challenge: Report (2022)","DOI":"10.1007\/978-3-031-25066-8_7"},{"key":"4_CR43","doi-asserted-by":"crossref","unstructured":"Ignatov, A., Van Gool, L., Timofte, R.: Replacing mobile camera ISP with a single deep learning model. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 536\u2013537 (2020)","DOI":"10.1109\/CVPRW50498.2020.00276"},{"key":"4_CR44","doi-asserted-by":"publisher","first-page":"276","DOI":"10.1016\/j.patrec.2020.07.033","volume":"138","author":"D Ignatov","year":"2020","unstructured":"Ignatov, D., Ignatov, A.: Controlling information capacity of binary neural network. Pattern Recogn. Lett. 138, 276\u2013281 (2020)","journal-title":"Pattern Recogn. Lett."},{"key":"4_CR45","doi-asserted-by":"crossref","unstructured":"Jacob, B., et al.: Quantization and training of neural networks for efficient integer-arithmetic-only inference. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2704\u20132713 (2018)","DOI":"10.1109\/CVPR.2018.00286"},{"key":"4_CR46","unstructured":"Jain, S.R., Gural, A., Wu, M., Dick, C.H.: Trained quantization thresholds for accurate and efficient fixed-point inference of deep neural networks. arXiv preprint arXiv:1903.08066 (2019)"},{"key":"4_CR47","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"4_CR48","doi-asserted-by":"crossref","unstructured":"K\u0131nl\u0131, F.O., Mente\u015f, S., \u00d6zcan, B., Kirac, F., Timofte, R., et al.: Aim 2022 challenge on Instagram filter removal: Methods and results. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022)","DOI":"10.1007\/978-3-031-25066-8_2"},{"key":"4_CR49","doi-asserted-by":"crossref","unstructured":"Laina, I., Rupprecht, C., Belagiannis, V., Tombari, F., Navab, N.: Deeper depth prediction with fully convolutional residual networks. In: 2016 Fourth International Conference on 3D Vision (3DV). pp. 239\u2013248. IEEE (2016)","DOI":"10.1109\/3DV.2016.32"},{"key":"4_CR50","doi-asserted-by":"publisher","unstructured":"Lee, J.H., Han, M.K., Ko, D.W., Suh, I.H.: From big to small: Multi-scale local planar guidance for monocular depth estimation (2019). https:\/\/doi.org\/10.48550\/ARXIV.1907.10326, https:\/\/arxiv.org\/abs\/1907.10326","DOI":"10.48550\/ARXIV.1907.10326"},{"key":"4_CR51","unstructured":"Lee, J.H., Han, M.K., Ko, D.W., Suh, I.H.: From big to small: Multi-scale local planar guidance for monocular depth estimation. arXiv preprint arXiv:1907.10326 (2019)"},{"key":"4_CR52","doi-asserted-by":"crossref","unstructured":"Li, Y., Gu, S., Gool, L.V., Timofte, R.: Learning filter basis for convolutional neural network compression. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 5623\u20135632 (2019)","DOI":"10.1109\/ICCV.2019.00572"},{"key":"4_CR53","doi-asserted-by":"crossref","unstructured":"Li, Z., Snavely, N.: Megadepth: learning single-view depth prediction from internet photos. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2041\u20132050 (2018)","DOI":"10.1109\/CVPR.2018.00218"},{"key":"4_CR54","doi-asserted-by":"crossref","unstructured":"Li, Z., Chen, Z., Xu, J., Liu, X., Jiang, J.: Litedepth: digging into fast and accurate depth estimation on mobile devices. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022)","DOI":"10.1007\/978-3-031-25063-7_31"},{"key":"4_CR55","doi-asserted-by":"crossref","unstructured":"Liu, F., Shen, C., Lin, G.: Deep convolutional neural fields for depth estimation from a single image. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5162\u20135170 (2015)","DOI":"10.1109\/CVPR.2015.7299152"},{"issue":"10","key":"4_CR56","doi-asserted-by":"publisher","first-page":"2024","DOI":"10.1109\/TPAMI.2015.2505283","volume":"38","author":"F Liu","year":"2015","unstructured":"Liu, F., Shen, C., Lin, G., Reid, I.: Learning depth from single monocular images using deep convolutional neural fields. IEEE Trans. Pattern Anal. Mach. Intell. 38(10), 2024\u20132039 (2015)","journal-title":"IEEE Trans. Pattern Anal. Mach. Intell."},{"key":"4_CR57","doi-asserted-by":"crossref","unstructured":"Liu, Y., Chen, K., Liu, C., Qin, Z., Luo, Z., Wang, J.: Structured knowledge distillation for semantic segmentation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2604\u20132613 (2019)","DOI":"10.1109\/CVPR.2019.00271"},{"key":"4_CR58","unstructured":"Liu, Y., Shu, C., Wang, J., Shen, C.: Structured knowledge distillation for dense prediction. IEEE Trans. Pattern Anal. Mach. Intell. (99), 1-1 (2020)"},{"key":"4_CR59","doi-asserted-by":"crossref","unstructured":"Liu, Z., et al.: Metapruning: meta learning for automatic neural network channel pruning. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 3296\u20133305 (2019)","DOI":"10.1109\/ICCV.2019.00339"},{"key":"4_CR60","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"747","DOI":"10.1007\/978-3-030-01267-0_44","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Z Liu","year":"2018","unstructured":"Liu, Z., Wu, B., Luo, W., Yang, X., Liu, W., Cheng, K.-T.: Bi-Real Net: enhancing the performance of 1-Bit CNNs with improved representational capability and advanced training algorithm. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11219, pp. 747\u2013763. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01267-0_44"},{"key":"4_CR61","unstructured":"Loshchilov, I., Hutter, F.: Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101 (2017)"},{"key":"4_CR62","unstructured":"Lugmayr, A., Danelljan, M., Timofte, R.: Ntire 2020 challenge on real-world image super-resolution: Methods and results. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops, pp. 494\u2013495 (2020)"},{"key":"4_CR63","unstructured":"Obukhov, A., Rakhuba, M., Georgoulis, S., Kanakis, M., Dai, D., Van Gool, L.: T-basis: a compact representation for neural networks. In: International Conference on Machine Learning, pp. 7392\u20137404. PMLR (2020)"},{"key":"4_CR64","doi-asserted-by":"crossref","unstructured":"Ortiz, L.E., Cabrera, E.V., Gon\u00e7alves, L.M.: Depth data error modeling of the zed 3d vision sensor from stereolabs. ELCVIA: Electr. Lett. Compu. Visi. Image Anal. 17(1), 0001\u201315 (2018)","DOI":"10.5565\/rev\/elcvia.1084"},{"key":"4_CR65","unstructured":"Ranftl, R., Lasinger, K., Hafner, D., Schindler, K., Koltun, V.: Towards robust monocular depth estimation: Mixing datasets for zero-shot cross-dataset transfer. IEEE Trans. Pattern Anal. Mach. Intell. (2020)"},{"key":"4_CR66","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"4_CR67","doi-asserted-by":"crossref","unstructured":"Tan, M., Chen, B., Pang, R., Vasudevan, V., Sandler, M., Howard, A., Le, Q.V.: Mnasnet: Platform-aware neural architecture search for mobile. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2820\u20132828 (2019)","DOI":"10.1109\/CVPR.2019.00293"},{"key":"4_CR68","unstructured":"TensorFlow-Lite: https:\/\/www.tensorflow.org\/lite"},{"key":"4_CR69","unstructured":"TensorFlow-Lite: https:\/\/www.tensorflow.org\/lite\/guide\/python"},{"key":"4_CR70","unstructured":"Timofte, R., Gu, S., Wu, J., Van Gool, L.: Ntire 2018 challenge on single image super-resolution: Methods and results. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops,. pp. 852\u2013863 (2018)"},{"key":"4_CR71","unstructured":"Uhlich, S., et al.: Mixed precision DNNs: All you need is a good parametrization. arXiv preprint arXiv:1905.11452 (2019)"},{"key":"4_CR72","doi-asserted-by":"crossref","unstructured":"Uhrig, J., Schneider, N., Schneider, L., Franke, U., Brox, T., Geiger, A.: Sparsity invariant CNNS. In: International Conference on 3D Vision (3DV) (2017)","DOI":"10.1109\/3DV.2017.00012"},{"key":"4_CR73","doi-asserted-by":"crossref","unstructured":"Wan, A., et al.: Fbnetv2: differentiable neural architecture search for spatial and channel dimensions. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 12965\u201312974 (2020)","DOI":"10.1109\/CVPR42600.2020.01298"},{"key":"4_CR74","doi-asserted-by":"publisher","unstructured":"Wang, Y., Li, X., Shi, M., Xian, K., Cao, Z.: Knowledge distillation for fast and accurate monocular depth estimation on mobile devices. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 2457\u20132465 (2021). https:\/\/doi.org\/10.1109\/CVPRW53098.2021.00278","DOI":"10.1109\/CVPRW53098.2021.00278"},{"key":"4_CR75","doi-asserted-by":"publisher","unstructured":"Wofk, D., Ma, F., Yang, T.J., Karaman, S., Sze, V.: Fastdepth: Fast monocular depth estimation on embedded systems (2019). https:\/\/doi.org\/10.48550\/ARXIV.1903.03273, https:\/\/arxiv.org\/abs\/1903.03273","DOI":"10.48550\/ARXIV.1903.03273"},{"key":"4_CR76","doi-asserted-by":"crossref","unstructured":"Wu, B., et al.: Fbnet: hardware-aware efficient convnet design via differentiable neural architecture search. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 10734\u201310742 (2019)","DOI":"10.1109\/CVPR.2019.01099"},{"key":"4_CR77","doi-asserted-by":"crossref","unstructured":"Yang, J., et al.: Quantization networks. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 7308\u20137316 (2019)","DOI":"10.1109\/CVPR.2019.00748"},{"key":"4_CR78","unstructured":"Yang, R., Timofte, R., et al.: Aim 2022 challenge on super-resolution of compressed image and video: Dataset, methods and results. In: Proceedings of the European Conference on Computer Vision (ECCV) Workshops (2022)"},{"key":"4_CR79","doi-asserted-by":"crossref","unstructured":"Yin, W., Liu, Y., Shen, C., Yan, Y.: Enforcing geometric constraints of virtual normal for depth prediction. In: The IEEE International Conference on Computer Vision (ICCV) (2019)","DOI":"10.1109\/ICCV.2019.00578"},{"key":"4_CR80","doi-asserted-by":"crossref","unstructured":"Zhang, Z., Wang, Y., Huang, Z., Luo, G., Yu, G., Fu, B.: A simple baseline for fast and accurate depth estimation on mobile devices. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops. pp. 0\u20130 (2021)","DOI":"10.1109\/CVPRW53098.2021.00279"},{"key":"4_CR81","doi-asserted-by":"publisher","unstructured":"Zhang, Z., Wang, Y., Huang, Z., Luo, G., Yu, G., Fu, B.: A simple baseline for fast and accurate depth estimation on mobile devices. In: 2021 IEEE\/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 2466\u20132471 (2021). https:\/\/doi.org\/10.1109\/CVPRW53098.2021.00279","DOI":"10.1109\/CVPRW53098.2021.00279"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25066-8_4","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,10,14]],"date-time":"2024-10-14T13:12:32Z","timestamp":1728911552000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25066-8_4"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031250651","9783031250668"],"references-count":81,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25066-8_4","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"18 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"From the workshops, 367 reviewed full papers have been selected for publication","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}