{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,26]],"date-time":"2025-03-26T22:54:34Z","timestamp":1743029674507,"version":"3.40.3"},"publisher-location":"Cham","reference-count":82,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031197833"},{"type":"electronic","value":"9783031197840"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-19784-0_29","type":"book-chapter","created":{"date-parts":[[2022,10,30]],"date-time":"2022-10-30T14:02:50Z","timestamp":1667138570000},"page":"493-510","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Contrastive Monotonic Pixel-Level Modulation"],"prefix":"10.1007","author":[{"ORCID":"https:\/\/orcid.org\/0000-0003-4698-9769","authenticated-orcid":false,"given":"Kun","family":"Lu","sequence":"first","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-4297-5060","authenticated-orcid":false,"given":"Rongpeng","family":"Li","sequence":"additional","affiliation":[]},{"ORCID":"https:\/\/orcid.org\/0000-0003-1492-1364","authenticated-orcid":false,"given":"Honggang","family":"Zhang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,10,31]]},"reference":[{"key":"29_CR1","doi-asserted-by":"crossref","unstructured":"Abdelhamed, A., Lin, S., Brown, M.S.: A high-quality denoising dataset for smartphone cameras. In: CVPR, pp. 1692\u20131700 (2018)","DOI":"10.1109\/CVPR.2018.00182"},{"key":"29_CR2","unstructured":"Ali, J., Lucy, C., Phillip, I.: On the \u201csteerability\u201d of generative adversarial networks. In: ICLR (2020)"},{"key":"29_CR3","unstructured":"Almahairi, A., Rajeshwar, S., Sordoni, A., Bachman, P., Courville, A.: Augmented CycleGAN: learning many-to-many mappings from unpaired data. In: ICML, pp. 195\u2013204 (2018)"},{"key":"29_CR4","doi-asserted-by":"crossref","unstructured":"Brooks, T., Mildenhall, B., Xue, T., Chen, J., Sharlet, D., Barron, J.T.: Unprocessing images for learned raw denoising. In: CVPR, pp. 11036\u201311045 (2019)","DOI":"10.1109\/CVPR.2019.01129"},{"key":"29_CR5","doi-asserted-by":"crossref","unstructured":"Cai, H., He, J., Qiao, Y., Dong, C.: Toward interactive modulation for photo-realistic image restoration. In: CVPR, pp. 294\u2013303 (2021)","DOI":"10.1109\/CVPRW53098.2021.00039"},{"key":"29_CR6","doi-asserted-by":"crossref","unstructured":"Chen, C., Chen, Q., Xu, J., Koltun, V.: Learning to see in the dark. In: CVPR, pp. 3291\u20133300 (2018)","DOI":"10.1109\/CVPR.2018.00347"},{"key":"29_CR7","first-page":"8043","volume":"29","author":"D Chen","year":"2020","unstructured":"Chen, D., Fan, Q., Liao, J., Aviles-Rivero, A., Yuan, L., Yu, N., Hua, G.: Controllable image processing via adaptive filterbank pyramid. TIP 29, 8043\u20138054 (2020)","journal-title":"TIP"},{"key":"29_CR8","doi-asserted-by":"crossref","unstructured":"Chen, D., Yuan, L., Liao, J., Yu, N., Hua, G.: StyleBank: an explicit representation for neural image style transfer. In: CVPR, pp. 1897\u20131906 (2017)","DOI":"10.1109\/CVPR.2017.296"},{"key":"29_CR9","doi-asserted-by":"crossref","unstructured":"Chen, J., Chen, J., Chao, H., Yang, M.: Image blind denoising with generative adversarial network based noise modeling. In: CVPR, pp. 3155\u20133164 (2018)","DOI":"10.1109\/CVPR.2018.00333"},{"key":"29_CR10","unstructured":"Chen, T., Kornblith, S., Norouzi, M., Hinton, G.: A simple framework for contrastive learning of visual representations. In: ICML, pp. 1597\u20131607 (2020)"},{"key":"29_CR11","doi-asserted-by":"crossref","unstructured":"Chen, Y.C., Xu, X., Tian, Z., Jia, J.: Homomorphic latent space interpolation for unpaired image-to-image translation. In: CVPR, pp. 2408\u20132416 (2019)","DOI":"10.1109\/CVPR.2019.00251"},{"key":"29_CR12","doi-asserted-by":"crossref","unstructured":"Choi, Y., Choi, M., Kim, M., Ha, J.W., Kim, S., Choo, J.: StarGAN: unified generative adversarial networks for multi-domain image-to-image translation. In: CVPR, pp. 8789\u20138797 (2018)","DOI":"10.1109\/CVPR.2018.00916"},{"key":"29_CR13","doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: StarGAN v2: diverse image synthesis for multiple domains. In: CVPR, pp. 8188\u20138197 (2020)","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"29_CR14","doi-asserted-by":"crossref","unstructured":"Collins, E., Bala, R., Price, B., Susstrunk, S.: Editing in style: uncovering the local semantics of GANs. In: CVPR, pp. 5771\u20135780 (2020)","DOI":"10.1109\/CVPR42600.2020.00581"},{"key":"29_CR15","unstructured":"Dumoulin, V., Shlens, J., Kudlur, M.: A learned representation for artistic style. In: ICLR (2017)"},{"key":"29_CR16","doi-asserted-by":"crossref","unstructured":"Gatys, L., Ecker, A.S., Bethge, M.: Texture synthesis using convolutional neural networks. In: NeurIPS, vol. 28 (2015)","DOI":"10.1109\/CVPR.2016.265"},{"key":"29_CR17","doi-asserted-by":"crossref","unstructured":"Girshick, R.: Fast R-CNN. In: ICCV (2015)","DOI":"10.1109\/ICCV.2015.169"},{"key":"29_CR18","doi-asserted-by":"crossref","unstructured":"Gong, R., Li, W., Chen, Y., Gool, L.V.: DLOW: domain flow for adaptation and generalization. In: CVPR, pp. 2477\u20132486 (2019)","DOI":"10.1109\/CVPR.2019.00258"},{"key":"29_CR19","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: NeurIPS, vol. 27 (2014)"},{"key":"29_CR20","doi-asserted-by":"crossref","unstructured":"Guo, S., Yan, Z., Zhang, K., Zuo, W., Zhang, L.: Toward convolutional blind denoising of real photographs. In: CVPR, pp. 1712\u20131722 (2019)","DOI":"10.1109\/CVPR.2019.00181"},{"key":"29_CR21","doi-asserted-by":"crossref","unstructured":"He, J., Dong, C., Qiao, Y.: Modulating image restoration with continual levels via adaptive feature modification layers. In: CVPR, pp. 11056\u201311064 (2019)","DOI":"10.1109\/CVPR.2019.01131"},{"key":"29_CR22","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"53","DOI":"10.1007\/978-3-030-58565-5_4","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J He","year":"2020","unstructured":"He, J., Dong, C., Qiao, Yu.: Interactive multi-dimension modulation with dynamic controllable residual learning for image restoration. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12365, pp. 53\u201368. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58565-5_4"},{"key":"29_CR23","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"},{"key":"29_CR24","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local nash equilibrium. In: NeurIPS, vol. 30 (2017)"},{"key":"29_CR25","doi-asserted-by":"crossref","unstructured":"Huang, X., Belongie, S.: Arbitrary style transfer in real-time with adaptive instance normalization. In: ICCV, pp. 1501\u20131510 (2017)","DOI":"10.1109\/ICCV.2017.167"},{"key":"29_CR26","doi-asserted-by":"crossref","unstructured":"Huang, X., Liu, M.Y., Belongie, S., Kautz, J.: Multimodal unsupervised image-to-image translation. In: ECCV, pp. 172\u2013189 (2018)","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"29_CR27","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: CVPR, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"29_CR28","first-page":"2340","volume":"30","author":"Y Jiang","year":"2021","unstructured":"Jiang, Y., et al.: EnlightenGAN: deep light enhancement without paired supervision. TIP 30, 2340\u20132349 (2021)","journal-title":"TIP"},{"key":"29_CR29","doi-asserted-by":"crossref","unstructured":"Jo, S.Y., Lee, S., Ahn, N., Kang, S.J.: Deep arbitrary HDRI: inverse tone mapping with controllable exposure changes. TMM 24, 2713\u20132726 (2021)","DOI":"10.1109\/TMM.2021.3087034"},{"key":"29_CR30","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"694","DOI":"10.1007\/978-3-319-46475-6_43","volume-title":"Computer Vision \u2013 ECCV 2016","author":"J Johnson","year":"2016","unstructured":"Johnson, J., Alahi, A., Fei-Fei, L.: Perceptual losses for real-time style transfer and super-resolution. In: Leibe, B., Matas, J., Sebe, N., Welling, M. (eds.) ECCV 2016. LNCS, vol. 9906, pp. 694\u2013711. Springer, Cham (2016). https:\/\/doi.org\/10.1007\/978-3-319-46475-6_43"},{"key":"29_CR31","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A style-based generator architecture for generative adversarial networks. In: CVPR, pp. 4401\u20134410 (2019)","DOI":"10.1109\/CVPR.2019.00453"},{"key":"29_CR32","doi-asserted-by":"crossref","unstructured":"Kim, D.W., Ryun Chung, J., Jung, S.W.: GRDN: grouped residual dense network for real image denoising and GAN-based real-world noise modeling. In: CVPRW (2019)","DOI":"10.1109\/CVPRW.2019.00261"},{"key":"29_CR33","doi-asserted-by":"crossref","unstructured":"Kim, H., Choi, Y., Kim, J., Yoo, S., Uh, Y.: Exploiting spatial dimensions of latent in GAN for real-time image editing. In: CVPR, pp. 852\u2013861 (2021)","DOI":"10.1109\/CVPR46437.2021.00091"},{"key":"29_CR34","unstructured":"Klyuchka, M., Neiterman, E.H., Ben-Artzi, G.: CEL-net: continuous exposure for extreme low-light imaging. arXiv:2012.04112 (2020)"},{"key":"29_CR35","doi-asserted-by":"crossref","unstructured":"Lee, C.H., Liu, Z., Wu, L., Luo, P.: MaskGAN: towards diverse and interactive facial image manipulation. In: CVPR, pp. 5549\u20135558 (2020)","DOI":"10.1109\/CVPR42600.2020.00559"},{"key":"29_CR36","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"36","DOI":"10.1007\/978-3-030-01246-5_3","volume-title":"Computer Vision \u2013 ECCV 2018","author":"H-Y Lee","year":"2018","unstructured":"Lee, H.-Y., Tseng, H.-Y., Huang, J.-B., Singh, M., Yang, M.-H.: Diverse image-to-image translation via disentangled representations. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11205, pp. 36\u201352. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01246-5_3"},{"key":"29_CR37","unstructured":"Lee, H., Kim, T., Son, H., Baek, S., Cheon, M., Lee, S.: Smoother network tuning and interpolation for continuous-level image processing. arXiv:2010.02270 (2020)"},{"key":"29_CR38","unstructured":"Li, C., Guo, C., Han, L.H., Jiang, J., Cheng, M.M., Gu, J., Loy, C.C.: Low-light image and video enhancement using deep learning: a survey. TPAMI (2021)"},{"key":"29_CR39","doi-asserted-by":"crossref","unstructured":"Li, C., Wand, M.: Combining Markov random fields and convolutional neural networks for image synthesis. In: CVPR, pp. 2479\u20132486 (2016)","DOI":"10.1109\/CVPR.2016.272"},{"key":"29_CR40","doi-asserted-by":"crossref","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Diversified texture synthesis with feed-forward networks. In: CVPR, pp. 3920\u20133928 (2017)","DOI":"10.1109\/CVPR.2017.36"},{"key":"29_CR41","unstructured":"Li, Y., Fang, C., Yang, J., Wang, Z., Lu, X., Yang, M.H.: Universal style transfer via feature transforms. In: NeurIPS, vol. 30 (2017)"},{"key":"29_CR42","unstructured":"Ling, H., Kreis, K., Li, D., Kim, S.W., Torralba, A., Fidler, S.: EditGAN: high-precision semantic image editing. In: NeurIPS, vol. 34 (2021)"},{"key":"29_CR43","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"363","DOI":"10.1007\/978-3-030-58574-7_22","volume-title":"Computer Vision \u2013 ECCV 2020","author":"W Lira","year":"2020","unstructured":"Lira, W., Merz, J., Ritchie, D., Cohen-Or, D., Zhang, H.: GANHopper: multi-hop GAN for unsupervised image-to-image translation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12371, pp. 363\u2013379. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58574-7_22"},{"key":"29_CR44","doi-asserted-by":"crossref","unstructured":"Liu, Y., et al.: Smoothing the disentangled latent style space for unsupervised image-to-image translation. In: CVPR, pp. 10785\u201310794 (2021)","DOI":"10.1109\/CVPR46437.2021.01064"},{"key":"29_CR45","doi-asserted-by":"crossref","unstructured":"Long, J., Shelhamer, E., Darrell, T.: Fully convolutional networks for semantic segmentation. In: CVPR, pp. 3431\u20133440 (2015)","DOI":"10.1109\/CVPR.2015.7298965"},{"key":"29_CR46","doi-asserted-by":"crossref","unstructured":"Lore, K.G., Akintayo, A., Sarkar, S.: LLNet: a deep autoencoder approach to natural low-light image enhancement. PR 61, 650\u2013662 (2017)","DOI":"10.1016\/j.patcog.2016.06.008"},{"key":"29_CR47","first-page":"4093","volume":"23","author":"K Lu","year":"2020","unstructured":"Lu, K., Zhang, L.: TBEFN: a two-branch exposure-fusion network for low-light image enhancement. TMM 23, 4093\u20134105 (2020)","journal-title":"TMM"},{"key":"29_CR48","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"293","DOI":"10.1007\/978-3-030-01258-8_18","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Y Lu","year":"2018","unstructured":"Lu, Y., Tai, Y.-W., Tang, C.-K.: Attribute-guided face generation using conditional CycleGAN. In: Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y. (eds.) ECCV 2018. LNCS, vol. 11216, pp. 293\u2013308. Springer, Cham (2018). https:\/\/doi.org\/10.1007\/978-3-030-01258-8_18"},{"key":"29_CR49","doi-asserted-by":"crossref","unstructured":"Lv, Z., et al.: Learning semantic person image generation by region-adaptive normalization. In: CVPR, pp. 10806\u201310815 (2021)","DOI":"10.1109\/CVPR46437.2021.01066"},{"key":"29_CR50","unstructured":"Maas, A.L., Hannun, A.Y., Ng, A.Y., et al.: Rectifier nonlinearities improve neural network acoustic models. In: ICML, vol. 30, p. 3 (2013)"},{"key":"29_CR51","doi-asserted-by":"publisher","first-page":"517","DOI":"10.1007\/s11263-021-01557-6","volume":"130","author":"Q Mao","year":"2022","unstructured":"Mao, Q., Tseng, H.Y., Lee, H.Y., Huang, J.B., Ma, S., Yang, M.H.: Continuous and diverse image-to-image translation via signed attribute vectors. IJCV 130, 517\u2013549 (2022)","journal-title":"IJCV"},{"key":"29_CR52","doi-asserted-by":"crossref","unstructured":"Mao, X., Li, Q., Xie, H., Lau, R.Y., Wang, Z., Paul Smolley, S.: Least squares generative adversarial networks. In: ICCV, pp. 2794\u20132802 (2017)","DOI":"10.1109\/ICCV.2017.304"},{"key":"29_CR53","doi-asserted-by":"crossref","unstructured":"Ni, S., Cao, X., Yue, T., Hu, X.: Controlling the rain: from removal to rendering. In: CVPR, pp. 6328\u20136337 (2021)","DOI":"10.1109\/CVPR46437.2021.00626"},{"key":"29_CR54","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"319","DOI":"10.1007\/978-3-030-58545-7_19","volume-title":"Computer Vision \u2013 ECCV 2020","author":"T Park","year":"2020","unstructured":"Park, T., Efros, A.A., Zhang, R., Zhu, J.-Y.: Contrastive learning for unpaired image-to-image translation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12354, pp. 319\u2013345. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58545-7_19"},{"key":"29_CR55","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic image synthesis with spatially-adaptive normalization. In: CVPR, pp. 2337\u20132346 (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"29_CR56","doi-asserted-by":"crossref","unstructured":"Pizzati, F., Cerri, P., de Charette, R.: CoMoGAN: continuous model-guided image-to-image translation. In: CVPR, pp. 14288\u201314298 (2021)","DOI":"10.1109\/CVPR46437.2021.01406"},{"key":"29_CR57","doi-asserted-by":"crossref","unstructured":"Pumarola, A., Agudo, A., Martinez, A.M., Sanfeliu, A., Moreno-Noguer, F.: GANimation: anatomically-aware facial animation from a single image. In: ECCV, pp. 818\u2013833 (2018)","DOI":"10.1007\/978-3-030-01249-6_50"},{"key":"29_CR58","doi-asserted-by":"crossref","unstructured":"Ren, D., Zuo, W., Hu, Q., Zhu, P., Meng, D.: Progressive image deraining networks: a better and simpler baseline. In: CVPR, pp. 3937\u20133946 (2019)","DOI":"10.1109\/CVPR.2019.00406"},{"key":"29_CR59","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"234","DOI":"10.1007\/978-3-319-24574-4_28","volume-title":"Medical Image Computing and Computer-Assisted Intervention \u2013 MICCAI 2015","author":"O Ronneberger","year":"2015","unstructured":"Ronneberger, O., Fischer, P., Brox, T.: U-Net: convolutional networks for biomedical image segmentation. In: Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F. (eds.) MICCAI 2015. LNCS, vol. 9351, pp. 234\u2013241. Springer, Cham (2015). https:\/\/doi.org\/10.1007\/978-3-319-24574-4_28"},{"key":"29_CR60","doi-asserted-by":"crossref","unstructured":"Shen, Y., Zhou, B.: Closed-form factorization of latent semantics in GANs. In: CVPR, pp. 1532\u20131540 (2021)","DOI":"10.1109\/CVPR46437.2021.00158"},{"key":"29_CR61","doi-asserted-by":"crossref","unstructured":"Shi, J., Xu, N., Xu, Y., Bui, T., Dernoncourt, F., Xu, C.: Learning by planning: language-guided global image editing. In: CVPR, pp. 13590\u201313599 (2021)","DOI":"10.1109\/CVPR46437.2021.01338"},{"key":"29_CR62","doi-asserted-by":"crossref","unstructured":"Shoshan, A., Mechrez, R., Zelnik-Manor, L.: Dynamic-net: tuning the objective without re-training for synthesis tasks. In: ICCV, pp. 3215\u20133223 (2019)","DOI":"10.1109\/ICCV.2019.00331"},{"key":"29_CR63","unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv:1409.1556 (2014)"},{"key":"29_CR64","doi-asserted-by":"crossref","unstructured":"Sun, X., Li, M., He, T., Fan, L.: Enhance images as you like with unpaired learning. arXiv:2110.01161 (2021)","DOI":"10.24963\/ijcai.2021\/140"},{"key":"29_CR65","unstructured":"Ulyanov, D., Vedaldi, A., Lempitsky, V.: Instance normalization: the missing ingredient for fast stylization. arXiv:1607.08022 (2016)"},{"key":"29_CR66","doi-asserted-by":"crossref","unstructured":"Upchurch, P., et al.: Deep feature interpolation for image content changes. In: CVPR, pp. 7064\u20137073 (2017)","DOI":"10.1109\/CVPR.2017.645"},{"key":"29_CR67","doi-asserted-by":"crossref","unstructured":"Wang, W., Guo, R., Tian, Y., Yang, W.: CFSNet: toward a controllable feature space for image restoration. In: ICCV, pp. 4140\u20134149 (2019)","DOI":"10.1109\/ICCV.2019.00424"},{"key":"29_CR68","doi-asserted-by":"crossref","unstructured":"Wang, X., Yu, K., Dong, C., Tang, X., Loy, C.C.: Deep network interpolation for continuous imagery effect transition. In: CVPR, pp. 1692\u20131701 (2019)","DOI":"10.1109\/CVPR.2019.00179"},{"issue":"4","key":"29_CR69","first-page":"600","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P.: Image quality assessment: from error visibility to structural similarity. TIP 13(4), 600\u2013612 (2004)","journal-title":"TIP"},{"key":"29_CR70","unstructured":"Wei, C., Wang, W., Yang, W., Liu, J.: Deep Retinex decomposition for low-light enhancement. In: BMVC (2018)"},{"key":"29_CR71","unstructured":"Wu, P.W., Lin, Y.J., Chang, C.H., Chang, E.Y., Liao, S.W.: RelGAN: multi-domain image-to-image translation via relative attributes. In: ICCV, pp. 5914\u20135922 (2019)"},{"key":"29_CR72","doi-asserted-by":"crossref","unstructured":"Ye, M., Zhang, X., Yuen, P.C., Chang, S.F.: Unsupervised embedding learning via invariant and spreading instance feature. In: CVPR, pp. 6210\u20136219 (2019)","DOI":"10.1109\/CVPR.2019.00637"},{"key":"29_CR73","unstructured":"Yu, X., Chen, Y., Liu, S., Li, T., Li, G.: Multi-mapping image-to-image translation via learning disentanglement. In: NeurIPS, vol. 32 (2019)"},{"key":"29_CR74","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"41","DOI":"10.1007\/978-3-030-58607-2_3","volume-title":"Computer Vision \u2013 ECCV 2020","author":"Z Yue","year":"2020","unstructured":"Yue, Z., Zhao, Q., Zhang, L., Meng, D.: Dual adversarial network: toward real-world noise removal and noise generation. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12355, pp. 41\u201358. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58607-2_3"},{"key":"29_CR75","doi-asserted-by":"crossref","unstructured":"Zamir, S.W., et al.: Multi-stage progressive image restoration. In: CVPR, pp. 14821\u201314831 (2021)","DOI":"10.1109\/CVPR46437.2021.01458"},{"key":"29_CR76","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"349","DOI":"10.1007\/978-3-030-11018-5_32","volume-title":"Computer Vision \u2013 ECCV 2018 Workshops","author":"H Zhang","year":"2019","unstructured":"Zhang, H., Dana, K.: Multi-style generative network for real-time transfer. In: Leal-Taix\u00e9, L., Roth, S. (eds.) ECCV 2018. LNCS, vol. 11132, pp. 349\u2013365. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-11018-5_32"},{"key":"29_CR77","doi-asserted-by":"crossref","unstructured":"Zhang, R., Isola, P., Efros, A.A., Shechtman, E., Wang, O.: The unreasonable effectiveness of deep features as a perceptual metric. In: CVPR, pp. 586\u2013595 (2018)","DOI":"10.1109\/CVPR.2018.00068"},{"issue":"4","key":"29_CR78","doi-asserted-by":"publisher","first-page":"1013","DOI":"10.1007\/s11263-020-01407-x","volume":"129","author":"Y Zhang","year":"2021","unstructured":"Zhang, Y., Guo, X., Ma, J., Liu, W., Zhang, J.: Beyond brightening low-light images. IJCV 129(4), 1013\u20131037 (2021)","journal-title":"IJCV"},{"key":"29_CR79","series-title":"Lecture Notes in Computer Science","doi-asserted-by":"publisher","first-page":"592","DOI":"10.1007\/978-3-030-58520-4_35","volume-title":"Computer Vision \u2013 ECCV 2020","author":"J Zhu","year":"2020","unstructured":"Zhu, J., Shen, Y., Zhao, D., Zhou, B.: In-domain GAN inversion for real image editing. In: Vedaldi, A., Bischof, H., Brox, T., Frahm, J.-M. (eds.) ECCV 2020. LNCS, vol. 12362, pp. 592\u2013608. Springer, Cham (2020). https:\/\/doi.org\/10.1007\/978-3-030-58520-4_35"},{"key":"29_CR80","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: ICCV, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"29_CR81","doi-asserted-by":"crossref","unstructured":"Zhu, P., Abdal, R., Qin, Y., Wonka, P.: SEAN: image synthesis with semantic region-adaptive normalization. In: CVPR, pp. 5104\u20135113 (2020)","DOI":"10.1109\/CVPR42600.2020.00515"},{"key":"29_CR82","unstructured":"Zhuang, P., Koyejo, O., Schwing, A.G.: Enjoy your editing: controllable GANs for image editing via latent space navigation. In: ICLR (2021)"}],"container-title":["Lecture Notes in Computer Science","Computer Vision \u2013 ECCV 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-19784-0_29","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,3,13]],"date-time":"2024-03-13T19:37:42Z","timestamp":1710358662000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-19784-0_29"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031197833","9783031197840"],"references-count":82,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-19784-0_29","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"31 October 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ECCV","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"European Conference on Computer Vision","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Tel Aviv","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Israel","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 October 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 October 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"17","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"eccv2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/eccv2022.ecva.net\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"CMT","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5804","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"1645","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"28% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.21","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.91","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}