{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,13]],"date-time":"2026-02-13T04:20:13Z","timestamp":1770956413355,"version":"3.50.1"},"reference-count":37,"publisher":"Springer Science and Business Media LLC","issue":"17","license":[{"start":{"date-parts":[[2018,4,30]],"date-time":"2018-04-30T00:00:00Z","timestamp":1525046400000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"funder":[{"DOI":"10.13039\/501100001809","name":"National Natural Science Foundation of China","doi-asserted-by":"publisher","award":["61303137"],"award-info":[{"award-number":["61303137"]}],"id":[{"id":"10.13039\/501100001809","id-type":"DOI","asserted-by":"publisher"}]},{"name":"National Science and Technology Support Program","award":["2015BAH21F01"],"award-info":[{"award-number":["2015BAH21F01"]}]},{"name":"Art Project for National Social-Science Foundation","award":["15BG084"],"award-info":[{"award-number":["15BG084"]}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimed Tools Appl"],"published-print":{"date-parts":[[2018,9]]},"DOI":"10.1007\/s11042-018-5968-7","type":"journal-article","created":{"date-parts":[[2018,4,30]],"date-time":"2018-04-30T11:01:35Z","timestamp":1525086095000},"page":"22339-22366","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":19,"title":["A one-to-many conditional generative adversarial network framework for multiple image-to-image translations"],"prefix":"10.1007","volume":"77","author":[{"given":"Chunlei","family":"Chai","sequence":"first","affiliation":[]},{"given":"Jing","family":"Liao","sequence":"additional","affiliation":[]},{"given":"Ning","family":"Zou","sequence":"additional","affiliation":[]},{"given":"Lingyun","family":"Sun","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2018,4,30]]},"reference":[{"issue":"11","key":"5968_CR1","doi-asserted-by":"publisher","first-page":"5187","DOI":"10.1109\/TIP.2016.2598681","volume":"25","author":"B Cai","year":"2016","unstructured":"Cai B, Xu X, Jia K, Qing C, Tao D (2016) DehazeNet: an end-to-end system for single image haze removal. IEEE Trans Image Process 25(11):5187\u20135198","journal-title":"IEEE Trans Image Process"},{"issue":"10","key":"5968_CR2","doi-asserted-by":"publisher","first-page":"12433","DOI":"10.1007\/s11042-016-3659-9","volume":"76","author":"F \u00c7al\u0131\u015f\u0131r","year":"2017","unstructured":"\u00c7al\u0131\u015f\u0131r F, Ba\u015ftan M, Ulusoy \u00d6, G\u00fcd\u00fckbay U (2017) Mobile multi-view object image search. Multimedia Tools & Applications 76(10):12433\u201312456","journal-title":"Multimedia Tools & Applications"},{"key":"5968_CR3","unstructured":"Chen M, Denoyer L (2016) Multi-view Generative Adversarial Networks arXiv eprint arXiv:1611.02019"},{"key":"5968_CR4","unstructured":"Elgammal A, Liu B, Elhoseiny M, Mazzone M (2017) CAN: Creative Adversarial Networks, Generating \"Art\" by Learning About Styles and Deviating from Style Norms. arXiv eprint arXiv:1706.07068"},{"key":"5968_CR5","doi-asserted-by":"publisher","first-page":"83","DOI":"10.1016\/j.sigpro.2014.08.034","volume":"112","author":"Z Gao","year":"2015","unstructured":"Gao Z, Zhang H, Xu GP, Xue YB, Hauptmannc AG (2015) Multi-view discriminative and structured dictionary learning with group sparsity for human action recognition. Signal Process 112:83\u201397","journal-title":"Signal Process"},{"key":"5968_CR6","doi-asserted-by":"crossref","unstructured":"Gatys LA, Ecker AS, Bethge M (2016) Image style transfer using convolutional neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition, pp 2414\u20132423","DOI":"10.1109\/CVPR.2016.265"},{"key":"5968_CR7","unstructured":"Ghosh A, Kulharia V, Namboodiri V, Torr PHS, Dokania PK (2017). Multi-Agent Diverse Generative Adversarial Networks. arXiv eprint arXiv:1606.07536"},{"key":"5968_CR8","unstructured":"Goodfellow IJ, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, Courville A, Bengio Y (2014) Generative adversarial nets. In: International Conference on Neural Information Processing Systems, pp 2672\u20132680"},{"key":"5968_CR9","unstructured":"Isola P, Zhu JY, Zhou TH, Efros, AA (2016) Image-to-Image Translation with Conditional Adversarial Networks arXiv eprint arXiv:1611.07004"},{"key":"5968_CR10","doi-asserted-by":"crossref","unstructured":"Jacob VG, Gupta S (2009) Colorization of grayscale images and videos using a semiautomatic approach. In: 2009 16th IEEE International Conference on Image Processing, pp 1653\u20131656. doi:10.1109\/ICIP.2009.5413392","DOI":"10.1109\/ICIP.2009.5413392"},{"key":"5968_CR11","unstructured":"Kim T, Cha M, Kim H, Lee JK, Kim J (2017) Learning to Discover Cross-Domain Relations with Generative Adversarial Networks. arXiv eprint arXiv:1703.05192"},{"key":"5968_CR12","unstructured":"Kwak H, Zhang BT (2016) Ways of Conditioning Generative Adversarial Networks. arXiv eprint arXiv:1611.01455"},{"key":"5968_CR13","unstructured":"Liu MY, Tuzel O (2016) Coupled generative adversarial networks. arXiv preprint arXiv:"},{"issue":"6","key":"5968_CR14","doi-asserted-by":"publisher","first-page":"1194","DOI":"10.1109\/TCYB.2014.2347057","volume":"45","author":"A-A Liu","year":"2015","unstructured":"Liu A-A, Su Y-T, Jia P-P, Gao Z, Hao T, Yang Z-X (Jun. 2015) (2015) Multipe\/single-view human action recognition via part-induced multitask structural learning. IEEE Transactions on Cybernetics 45(6):1194\u20131208","journal-title":"IEEE Transactions on Cybernetics"},{"key":"5968_CR15","unstructured":"Liu Y, Qin Z, Luo Z, Wang H (2017) Auto-painter: Cartoon Image Generation from Sketch by Using Conditional Generative Adversarial Networks. arXiv eprint arXiv:1705.01908"},{"key":"5968_CR16","doi-asserted-by":"crossref","unstructured":"Liu Z et al. (2017) Multiview and multimodal pervasive indoor localization. ACM on Multimedia Conference ACM: 109\u2013117","DOI":"10.1145\/3123266.3123436"},{"key":"5968_CR17","doi-asserted-by":"crossref","unstructured":"Luan F, Paris S, Bala K (2017) Deep Photo Style Transfer. arXiv eprint arXiv:1703.07511","DOI":"10.1109\/CVPR.2017.740"},{"key":"5968_CR18","unstructured":"Mirza M, Osindero S (2014) Conditional generative adversarial nets. Computer Science 2672\u20132680"},{"key":"5968_CR19","doi-asserted-by":"crossref","unstructured":"Nie L, Wang M, Zha Z, et al (2011) Multimedia answering: enriching text QA with media information: 695\u2013704","DOI":"10.1145\/2009916.2010010"},{"key":"5968_CR20","unstructured":"Perarnau G, Weijer JVD, Raducanu B, \u00c1lvarez JM (2016) Invertible Conditional GANs for image editing. In Conference and Workshop on Neural Information Processing Systems 2016. arXiv eprint arXiv:1611.06355"},{"key":"5968_CR21","unstructured":"Salimans T, Goodfellow I, Zaremba W, Cheung V, Radford A, Chen X (2016) Improved Techniques for Training GANs. arXiv eprint arXiv:1606.03498"},{"issue":"2","key":"5968_CR22","doi-asserted-by":"publisher","first-page":"430","DOI":"10.1109\/TIP.2005.859378","volume":"15","author":"HR Sheikh","year":"2006","unstructured":"Sheikh HR, Bovik AC (2006) Image information and visual quality. IEEE Trans Image Process 15(2):430\u2013444. \n                    https:\/\/doi.org\/10.1109\/TIP.2005.859378","journal-title":"IEEE Trans Image Process"},{"key":"5968_CR23","unstructured":"Vedran V, Raymond C, Gravier G (2017) Generative adversarial networks for multimodal representation learning in video hyperlinking. In: ACM on International Conference on Multimedia Retrieval, pp 416\u2013419"},{"key":"5968_CR24","doi-asserted-by":"crossref","unstructured":"Wang X, Gupta A (2016) Generative Image Modeling Using Style and Structure Adversarial Networks. arXiv eprint arXiv:1603.05631","DOI":"10.1007\/978-3-319-46493-0_20"},{"key":"5968_CR25","unstructured":"Wang Y, Zhang L, Weijer JVD (2016) Ensembles of Generative Adversarial Networks. arXiv eprint arXiv:1612.00991"},{"key":"5968_CR26","unstructured":"Wang C, Xu C, Tao D (2017) Perceptual Adversarial Networks for Image-to-Image Transformation. arXiv eprint arXiv:1706.09138"},{"key":"5968_CR27","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/s11263-017-1004-z","volume":"125","author":"S Xie","year":"2017","unstructured":"Xie S, Tu Z (2017) Holistically-nested edge detection. Int J Comput Vis 125:3\u201318","journal-title":"Int J Comput Vis"},{"issue":"3","key":"5968_CR28","doi-asserted-by":"publisher","first-page":"661","DOI":"10.1109\/TMM.2012.2237023","volume":"15","author":"Y Yang","year":"2013","unstructured":"Yang Y, Ma Z, Hauptmann AG, Sebe N (2013) Feature selection for multimedia analysis by sharing information among multiple tasks. IEEE Transactions on Multimedia 15(3):661\u2013669","journal-title":"IEEE Transactions on Multimedia"},{"key":"5968_CR29","doi-asserted-by":"crossref","unstructured":"Yi Z, Zhang H, Tan P, Gong M (2017) DualGAN: Unsupervised Dual Learning for Image-to-Image Translation. arXiv eprint arXiv:1704.02510","DOI":"10.1109\/ICCV.2017.310"},{"key":"5968_CR30","doi-asserted-by":"crossref","unstructured":"Yu A, Grauman K (2014) Fine-grained visual comparisons with local learning. In: Computer Vision and Pattern Recognition, pp 192\u2013199","DOI":"10.1109\/CVPR.2014.32"},{"key":"5968_CR31","doi-asserted-by":"crossref","unstructured":"Zhang L, Zhang L, Mou X, Zhang D (2012) A comprehensive evaluation of full reference image quality assessment algorithms. In: 2012 19th IEEE International Conference on Image Processing, pp 1477\u20131480. doi:10.1109\/ICIP.2012.6467150","DOI":"10.1109\/ICIP.2012.6467150"},{"key":"5968_CR32","doi-asserted-by":"crossref","unstructured":"Zhang R, Isola P, Efros AA (2016). Colorful Image Colorization. arXiv eprint arXiv:1603.08511","DOI":"10.1007\/978-3-319-46487-9_40"},{"key":"5968_CR33","doi-asserted-by":"crossref","unstructured":"Zhang H et al (2016) Online collaborative learning for open-vocabulary visual classifiers. IEEE Computer Vision and Pattern Recognition: 2809\u20132817","DOI":"10.1109\/CVPR.2016.307"},{"key":"5968_CR34","unstructured":"Zhang H, Sindagi V, Patel VM (2017) Image De-raining Using a Conditional Generative Adversarial Network. arXiv eprint arXiv:1701.05957"},{"issue":"3","key":"5968_CR35","doi-asserted-by":"publisher","first-page":"81","DOI":"10.1109\/97.995823","volume":"9","author":"W Zhou","year":"2002","unstructured":"Zhou W, Bovik AC (2002) A universal image quality index. IEEE Signal Processing Letters 9(3):81\u201384. \n                    https:\/\/doi.org\/10.1109\/97.995823","journal-title":"IEEE Signal Processing Letters"},{"issue":"4","key":"5968_CR36","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"W Zhou","year":"2004","unstructured":"Zhou W, Bovik AC, Sheikh HR, Simoncelli EP (2004) Image quality assessment: from error visibility to structural similarity. IEEE Trans Image Process 13(4):600\u2013612. \n                    https:\/\/doi.org\/10.1109\/TIP.2003.819861","journal-title":"IEEE Trans Image Process"},{"key":"5968_CR37","doi-asserted-by":"crossref","unstructured":"Zhu JY, Park T, Isola P, Efros AA (2017) Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. arXiv eprint arXiv:1703.10593","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Multimedia Tools and Applications"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/article\/10.1007\/s11042-018-5968-7\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-018-5968-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/s11042-018-5968-7.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2019,4,30]],"date-time":"2019-04-30T08:34:47Z","timestamp":1556613287000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/s11042-018-5968-7"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2018,4,30]]},"references-count":37,"journal-issue":{"issue":"17","published-print":{"date-parts":[[2018,9]]}},"alternative-id":["5968"],"URL":"https:\/\/doi.org\/10.1007\/s11042-018-5968-7","relation":{},"ISSN":["1380-7501","1573-7721"],"issn-type":[{"value":"1380-7501","type":"print"},{"value":"1573-7721","type":"electronic"}],"subject":[],"published":{"date-parts":[[2018,4,30]]},"assertion":[{"value":"27 September 2017","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"16 March 2018","order":2,"name":"revised","label":"Revised","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"3 April 2018","order":3,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"30 April 2018","order":4,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}}]}}