{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,25]],"date-time":"2025-03-25T21:25:01Z","timestamp":1742937901106,"version":"3.40.3"},"publisher-location":"Cham","reference-count":31,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031159336"},{"type":"electronic","value":"9783031159343"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-15934-3_6","type":"book-chapter","created":{"date-parts":[[2022,9,6]],"date-time":"2022-09-06T00:02:53Z","timestamp":1662422573000},"page":"62-73","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Cross-Domain Learning for\u00a0Reference-Based Sketch Colorization with\u00a0Structural and\u00a0Colorific Strategy"],"prefix":"10.1007","author":[{"given":"Haowei","family":"Zhong","sequence":"first","affiliation":[]},{"given":"Xianzhi","family":"Tu","sequence":"additional","affiliation":[]},{"given":"Hailong","family":"Liu","sequence":"additional","affiliation":[]},{"given":"Yulu","family":"Fu","sequence":"additional","affiliation":[]},{"given":"Jinrong","family":"Cui","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,9,15]]},"reference":[{"issue":"1","key":"6_CR1","doi-asserted-by":"publisher","first-page":"298","DOI":"10.1109\/TIP.2013.2288929","volume":"23","author":"A Bugeau","year":"2013","unstructured":"Bugeau, A., Ta, V.T., Papadakis, N.: Variational exemplar-based image colorization. IEEE Trans. Image Process. 23(1), 298\u2013307 (2013)","journal-title":"IEEE Trans. Image Process."},{"key":"6_CR2","doi-asserted-by":"crossref","unstructured":"Chen, W., Hays, J.: SketchyGAN: towards diverse and realistic sketch to image synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 9416\u20139425 (2018)","DOI":"10.1109\/CVPR.2018.00981"},{"key":"6_CR3","doi-asserted-by":"crossref","unstructured":"Choi, Y., Uh, Y., Yoo, J., Ha, J.W.: StarGAN v2: diverse image synthesis for multiple domains. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2020)","DOI":"10.1109\/CVPR42600.2020.00821"},{"key":"6_CR4","doi-asserted-by":"crossref","unstructured":"Ci, Y., Ma, X., Wang, Z., Li, H., Luo, Z.: User-guided deep anime line art colorization with conditional adversarial networks. In: Proceedings of the 26th ACM International Conference on Multimedia, pp. 1536\u20131544 (2018)","DOI":"10.1145\/3240508.3240661"},{"key":"6_CR5","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Advances in Neural Information Processing Systems 27 (2014)"},{"key":"6_CR6","doi-asserted-by":"crossref","unstructured":"Hati, Y., Jouet, G., Rousseaux, F., Duhart, C.: PaintsTorch: a user-guided anime line art colorization tool with double generator conditional adversarial network. In: European Conference on Visual Media Production, pp. 1\u201310 (2019)","DOI":"10.1145\/3359998.3369401"},{"key":"6_CR7","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a local Nash equilibrium. In: Advances in Neural Information Processing Systems 30 (2017)"},{"key":"6_CR8","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"6_CR9","doi-asserted-by":"crossref","unstructured":"Kim, H., Jhoo, H.Y., Park, E., Yoo, S.: Tag2Pix: line art colorization using text tag with SECat and changing loss. In: Proceedings of the IEEE\/CVF International Conference on Computer Vision, pp. 9056\u20139065 (2019)","DOI":"10.1109\/ICCV.2019.00915"},{"key":"6_CR10","unstructured":"Kingma, D.P., Ba, J.: Adam: a method for stochastic optimization. arXiv preprint arXiv:1412.6980 (2014)"},{"key":"6_CR11","doi-asserted-by":"crossref","unstructured":"Lee, J., Kim, E., Lee, Y., Kim, D., Chang, J., Choo, J.: Reference-based sketch image colorization using augmented-self reference and dense semantic correspondence. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5801\u20135810 (2020)","DOI":"10.1109\/CVPR42600.2020.00584"},{"key":"6_CR12","doi-asserted-by":"crossref","unstructured":"Lin, T.Y., Doll\u00e1r, P., Girshick, R., He, K., Hariharan, B., Belongie, S.: Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2117\u20132125 (2017)","DOI":"10.1109\/CVPR.2017.106"},{"key":"6_CR13","unstructured":"Liu, M.Y., Breuel, T., Kautz, J.: Unsupervised image-to-image translation networks. In: Advances in Neural Information Processing Systems, pp. 700\u2013708 (2017)"},{"key":"6_CR14","doi-asserted-by":"crossref","unstructured":"Liu, Z., Luo, P., Wang, X., Tang, X.: Deep learning face attributes in the wild. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 3730\u20133738 (2015)","DOI":"10.1109\/ICCV.2015.425"},{"key":"6_CR15","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic image synthesis with spatially-adaptive normalization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2337\u20132346 (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"6_CR16","doi-asserted-by":"crossref","unstructured":"Richardson, E., et al.: Encoding in style: a styleGAN encoder for image-to-image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 2287\u20132296 (2021)","DOI":"10.1109\/CVPR46437.2021.00232"},{"issue":"2","key":"6_CR17","doi-asserted-by":"publisher","first-page":"213","DOI":"10.1007\/s00530-014-0446-1","volume":"22","author":"RC Streijl","year":"2016","unstructured":"Streijl, R.C., Winkler, S., Hands, D.S.: Mean opinion score (MOS) revisited: methods and applications, limitations and alternatives. Multimedia Syst. 22(2), 213\u2013227 (2016)","journal-title":"Multimedia Syst."},{"key":"6_CR18","doi-asserted-by":"crossref","unstructured":"Sun, T.H., Lai, C.H., Wong, S.K., Wang, Y.S.: Adversarial colorization of icons based on contour and color conditions. In: Proceedings of the 27th ACM International Conference on Multimedia, pp. 683\u2013691 (2019)","DOI":"10.1145\/3343031.3351041"},{"key":"6_CR19","doi-asserted-by":"crossref","unstructured":"Wang, T.C., Liu, M.Y., Zhu, J.Y., Tao, A., Kautz, J., Catanzaro, B.: High-resolution image synthesis and semantic manipulation with conditional GANs. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 8798\u20138807 (2018)","DOI":"10.1109\/CVPR.2018.00917"},{"issue":"6","key":"6_CR20","doi-asserted-by":"publisher","first-page":"740","DOI":"10.1016\/j.cag.2012.03.004","volume":"36","author":"H Winnem\u00f6ller","year":"2012","unstructured":"Winnem\u00f6ller, H., Kyprianidis, J.E., Olsen, S.C.: XDoG: an extended difference-of-gaussians compendium including advanced image stylization. Comput. Graph. 36(6), 740\u2013753 (2012)","journal-title":"Comput. Graph."},{"issue":"3","key":"6_CR21","doi-asserted-by":"publisher","first-page":"1221","DOI":"10.1145\/1141911.1142018","volume":"25","author":"H Winnem\u00f6ller","year":"2006","unstructured":"Winnem\u00f6ller, H., Olsen, S.C., Gooch, B.: Real-time video abstraction. ACM Trans. Graph. (TOG) 25(3), 1221\u20131226 (2006)","journal-title":"ACM Trans. Graph. (TOG)"},{"key":"6_CR22","doi-asserted-by":"crossref","unstructured":"Xu, Z., Wang, T., Fang, F., Sheng, Y., Zhang, G.: Stylization-based architecture for fast deep exemplar colorization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9363\u20139372 (2020)","DOI":"10.1109\/CVPR42600.2020.00938"},{"key":"6_CR23","doi-asserted-by":"crossref","unstructured":"Yin, W., Lu, P., Zhao, Z., Peng, X.: Yes, \u201cattention is all you need\u201d, for exemplar based colorization. In: Proceedings of the 29th ACM International Conference on Multimedia, pp. 2243\u20132251 (2021)","DOI":"10.1145\/3474085.3475385"},{"key":"6_CR24","doi-asserted-by":"crossref","unstructured":"Yuan, M., Simo-Serra, E.: Line art colorization with concatenated spatial attention. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 3946\u20133950 (2021)","DOI":"10.1109\/CVPRW53098.2021.00442"},{"key":"6_CR25","doi-asserted-by":"crossref","unstructured":"Zhang, B., et al.: Deep exemplar-based video colorization. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 8052\u20138061 (2019)","DOI":"10.1109\/CVPR.2019.00824"},{"key":"6_CR26","doi-asserted-by":"crossref","unstructured":"Zhang, L., Li, C., Simo-Serra, E., Ji, Y., Wong, T.T., Liu, C.: User-guided line art flat filling with split filling mechanism. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 9889\u20139898 (2021)","DOI":"10.1109\/CVPR46437.2021.00976"},{"key":"6_CR27","doi-asserted-by":"crossref","unstructured":"Zhang, P., Zhang, B., Chen, D., Yuan, L., Wen, F.: Cross-domain correspondence learning for exemplar-based image translation. In: Proceedings of the IEEE\/CVF Conference on Computer Vision and Pattern Recognition, pp. 5143\u20135153 (2020)","DOI":"10.1109\/CVPR42600.2020.00519"},{"key":"6_CR28","doi-asserted-by":"crossref","unstructured":"Zhang, Q., Wang, B., Wen, W., Li, H., Liu, J.: Line art correlation matching feature transfer network for automatic animation colorization. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 3872\u20133881 (2021)","DOI":"10.1109\/WACV48630.2021.00392"},{"key":"6_CR29","doi-asserted-by":"crossref","unstructured":"Zhang, R., et al.: Real-time user-guided image colorization with learned deep priors. arXiv preprint arXiv:1705.02999 (2017)","DOI":"10.1145\/3072959.3073703"},{"key":"6_CR30","doi-asserted-by":"crossref","unstructured":"Zhao, M., Zhu, S.C.: Portrait painting using active templates. In: Proceedings of the ACM SIGGRAPH\/Eurographics Symposium on Non-photorealistic Animation and Rendering, pp. 117\u2013124 (2011)","DOI":"10.1145\/2024676.2024696"},{"key":"6_CR31","doi-asserted-by":"crossref","unstructured":"Zhu, J.Y., Park, T., Isola, P., Efros, A.A.: Unpaired image-to-image translation using cycle-consistent adversarial networks. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 2223\u20132232 (2017)","DOI":"10.1109\/ICCV.2017.244"}],"container-title":["Lecture Notes in Computer Science","Artificial Neural Networks and Machine Learning \u2013 ICANN 2022"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-15934-3_6","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,9,14]],"date-time":"2022-09-14T06:05:49Z","timestamp":1663135549000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-15934-3_6"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031159336","9783031159343"],"references-count":31,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-15934-3_6","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"15 September 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICANN","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Artificial Neural Networks","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Bristol","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"United Kingdom","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"6 September 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"9 September 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"31","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"icann2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/e-nns.org\/icann2022\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Single-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"561","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"255","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"45% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}