{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,3,27]],"date-time":"2025-03-27T03:33:52Z","timestamp":1743046432631,"version":"3.40.3"},"publisher-location":"Cham","reference-count":32,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783030377304"},{"type":"electronic","value":"9783030377311"}],"license":[{"start":{"date-parts":[[2019,12,24]],"date-time":"2019-12-24T00:00:00Z","timestamp":1577145600000},"content-version":"tdm","delay-in-days":0,"URL":"http:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2020]]},"DOI":"10.1007\/978-3-030-37731-1_51","type":"book-chapter","created":{"date-parts":[[2019,12,27]],"date-time":"2019-12-27T06:02:51Z","timestamp":1577426571000},"page":"628-641","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":1,"title":["OmniEyes: Analysis and Synthesis of Artistically Painted Eyes"],"prefix":"10.1007","author":[{"given":"Gjorgji","family":"Strezoski","sequence":"first","affiliation":[]},{"given":"Rogier","family":"Knoester","sequence":"additional","affiliation":[]},{"given":"Nanne","family":"van Noord","sequence":"additional","affiliation":[]},{"given":"Marcel","family":"Worring","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2019,12,24]]},"reference":[{"doi-asserted-by":"crossref","unstructured":"Baltru\u0161aitis, T., Robinson, P., Morency, L.P.: Openface: an open source facial behavior analysis toolkit. In: 2016 IEEE Winter Conference on Applications of Computer Vision (WACV), pp. 1\u201310. IEEE (2016)","key":"51_CR1","DOI":"10.1109\/WACV.2016.7477553"},{"key":"51_CR2","doi-asserted-by":"publisher","first-page":"73694","DOI":"10.1109\/ACCESS.2019.2921101","volume":"7","author":"E Cetinic","year":"2019","unstructured":"Cetinic, E., Lipic, T., Grgic, S.: A deep learning perspective on beauty, sentiment, and remembrance of art. IEEE Access 7, 73694\u201373710 (2019)","journal-title":"IEEE Access"},{"doi-asserted-by":"publisher","unstructured":"Ci, Y., Ma, X., Wang, Z., Li, H., Luo, Z.: User-guided deep anime line art colorization with conditional adversarial networks. In: 2018 ACM Multimedia Conference on Multimedia Conference, MM 2018, Seoul, Republic of Korea, October 22\u201326, 2018, pp. 1536\u20131544 (2018). https:\/\/doi.org\/10.1145\/3240508.3240661","key":"51_CR3","DOI":"10.1145\/3240508.3240661"},{"unstructured":"Elgammal, A., Liu, B., Elhoseiny, M., Mazzone, M.: Can: creative adversarial networks, generating \u201cart\u201d by learning about styles and deviating from style norms. arXiv preprint (2017). arXiv:1706.07068","key":"51_CR4"},{"doi-asserted-by":"crossref","unstructured":"Elgammal, A., Liu, B., Kim, D., Elhoseiny, M., Mazzone, M.: The shape of art history in the eyes of the machine. In: Thirty-Second AAAI Conference on Artificial Intelligence (2018)","key":"51_CR5","DOI":"10.1609\/aaai.v32i1.11894"},{"doi-asserted-by":"crossref","unstructured":"Garcia, N., Renoust, B., Nakashima, Y.: Context-aware embeddings for automatic art analysis. In: Proceedings of the 2019 on International Conference on Multimedia Retrieval, pp. 25\u201333. ACM (2019)","key":"51_CR6","DOI":"10.1145\/3323873.3325028"},{"doi-asserted-by":"crossref","unstructured":"Goldfarb, D., Merkl, D.: Visualizing art historical developments using the getty ulan, wikipedia and wikidata. In: 2018 22nd International Conference Information Visualisation (IV), pp. 459\u2013466. IEEE (2018)","key":"51_CR7","DOI":"10.1109\/iV.2018.00086"},{"doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X., Ren, S., Sun, J.: Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 770\u2013778 (2016)","key":"51_CR8","DOI":"10.1109\/CVPR.2016.90"},{"key":"51_CR9","doi-asserted-by":"publisher","first-page":"107:1","DOI":"10.1145\/3072959.3073659","volume":"36","author":"S Iizuka","year":"2017","unstructured":"Iizuka, S., Simo-Serra, E., Ishikawa, H.: Globally and locally consistent image completion. ACM Trans. Graph. 36, 107:1\u2013107:14 (2017)","journal-title":"ACM Trans. Graph."},{"doi-asserted-by":"crossref","unstructured":"Kazemi, V., Sullivan, J.: One millisecond face alignment with an ensemble of regression trees. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1867\u20131874 (2014)","key":"51_CR10","DOI":"10.1109\/CVPR.2014.241"},{"issue":"2","key":"51_CR11","doi-asserted-by":"publisher","first-page":"80","DOI":"10.1037\/1931-3896.1.2.80","volume":"1","author":"A Kozbelt","year":"2007","unstructured":"Kozbelt, A., Seeley, W.P.: Integrating art historical, psychological, and neuroscientific explanations of artists\u2019 advantages in drawing and perception. Psychology of Aesthetics, Creativity, and the Arts 1(2), 80 (2007)","journal-title":"Psychology of Aesthetics, Creativity, and the Arts"},{"unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097\u20131105 (2012)","key":"51_CR12"},{"doi-asserted-by":"crossref","unstructured":"Li, Y., Liu, S., Yang, J., Yang, M.H.: Generative face completion. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 3911\u20133919 (2017)","key":"51_CR13","DOI":"10.1109\/CVPR.2017.624"},{"key":"51_CR14","first-page":"89","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Guilin Liu","year":"2018","unstructured":"Liu, G., Reda, F.A., Shih, K.J., Wang, T.C., Tao, A., Catanzaro, B.: Image inpainting for irregular holes using partial convolutions. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 85\u2013100 (2018)"},{"doi-asserted-by":"publisher","unstructured":"Mao, H., Cheung, M., She, J.: Deepart: learning joint representations of visual arts. In: Proceedings of the 2017 ACM on Multimedia Conference, MM 2017, Mountain View, CA, USA, October 23\u201327, 2017, pp. 1183\u20131191 (2017). https:\/\/doi.org\/10.1145\/3123266.3123405","key":"51_CR15","DOI":"10.1145\/3123266.3123405"},{"unstructured":"Mirza, M., Osindero, S.: Conditional generative adversarial nets. arXiv preprint (2014). arXiv:1411.1784","key":"51_CR16"},{"unstructured":"Oh, C.: Automatically classifying art images using computer vision (2018)","key":"51_CR17"},{"doi-asserted-by":"crossref","unstructured":"Pathak, D., Krahenbuhl, P., Donahue, J., Darrell, T., Efros, A.A.: Context encoders: feature learning by inpainting. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2536\u20132544 (2016)","key":"51_CR18","DOI":"10.1109\/CVPR.2016.278"},{"unstructured":"Ren, S., He, K., Girshick, R., Sun, J.: Faster R-CNN: towards real-time object detection with region proposal networks. In: Advances in Neural Information Processing Systems, pp. 91\u201399 (2015)","key":"51_CR19"},{"doi-asserted-by":"crossref","unstructured":"Rodriguez, C.S., Lech, M., Pirogova, E.: Classification of style in fine-art paintings using transfer learning and weighted image patches. In: 2018 12th International Conference on Signal Processing and Communication Systems (ICSPCS), pp. 1\u20137. IEEE (2018)","key":"51_CR20","DOI":"10.1109\/ICSPCS.2018.8631731"},{"doi-asserted-by":"crossref","unstructured":"Sbai, O., Elhoseiny, M., Bordes, A., LeCun, Y., Couprie, C.: Design: design inspiration from generative networks. In: Proceedings of the European Conference on Computer Vision (ECCV) (2018)","key":"51_CR21","DOI":"10.1007\/978-3-030-11015-4_5"},{"doi-asserted-by":"crossref","unstructured":"Shen, X., Efros, A.A., Aubry, M.: Discovering visual patterns in art collections with spatially-consistent feature learning. In: Proceedings IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2019)","key":"51_CR22","DOI":"10.1109\/CVPR.2019.00950"},{"unstructured":"Simonyan, K., Zisserman, A.: Very deep convolutional networks for large-scale image recognition. arXiv preprint (2014). arXiv:1409.1556","key":"51_CR23"},{"key":"51_CR24","doi-asserted-by":"publisher","first-page":"3","DOI":"10.1007\/978-3-030-01216-8_1","volume-title":"Computer Vision \u2013 ECCV 2018","author":"Yuhang Song","year":"2018","unstructured":"Song, Y., et al.: Contextual-based image inpainting: infer, match, and translate. In: Proceedings of the European Conference on Computer Vision (ECCV), pp. 3\u201319 (2018)"},{"doi-asserted-by":"publisher","unstructured":"Strezoski, G., Groenen, I., Besenbruch, J., Worring, M.: Artsight: an artistic data exploration engine. In: 2018 ACM Multimedia Conference on Multimedia Conference, MM 2018, Seoul, Republic of Korea, October 22\u201326, 2018, pp. 1240\u20131241 (2018). https:\/\/doi.org\/10.1145\/3240508.3241389","key":"51_CR25","DOI":"10.1145\/3240508.3241389"},{"issue":"4","key":"51_CR26","first-page":"88","volume":"14","author":"G Strezoski","year":"2018","unstructured":"Strezoski, G., Worring, M.: Omniart: a large-scale artistic benchmark. ACM Trans. Multimedia Comput. Commun. Appl. (TOMM) 14(4), 88 (2018)","journal-title":"ACM Trans. Multimedia Comput. Commun. Appl. (TOMM)"},{"unstructured":"Van Noord, N., Postma, E.: Light-weight pixel context encoders for image inpainting. arXiv preprint (2018). arXiv:1801.05585","key":"51_CR27"},{"issue":"4","key":"51_CR28","doi-asserted-by":"publisher","first-page":"600","DOI":"10.1109\/TIP.2003.819861","volume":"13","author":"Z Wang","year":"2004","unstructured":"Wang, Z., Bovik, A.C., Sheikh, H.R., Simoncelli, E.P., et al.: Image quality assessment: from error visibility to structural similarity. IEEE Trans. Image Process. 13(4), 600\u2013612 (2004)","journal-title":"IEEE Trans. Image Process."},{"doi-asserted-by":"crossref","unstructured":"Yang, C., Lu, X., Lin, Z., Shechtman, E., Wang, O., Li, H.: High-resolution image inpainting using multi-scale neural patch synthesis. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6721\u20136729 (2017)","key":"51_CR29","DOI":"10.1109\/CVPR.2017.434"},{"key":"51_CR30","series-title":"Lecture Notes in Geoinformation and Cartography","doi-asserted-by":"publisher","first-page":"249","DOI":"10.1007\/978-3-030-19424-6_14","volume-title":"Computational Urban Planning and Management for Smart Cities","author":"Y Yoshimura","year":"2019","unstructured":"Yoshimura, Y., Cai, B., Wang, Z., Ratti, C.: Deep learning architect: classification for architectural design through the eye of artificial intelligence. In: Geertman, S., Zhan, Q., Allan, A., Pettit, C. (eds.) CUPUM 2019. LNGC, pp. 249\u2013265. Springer, Cham (2019). https:\/\/doi.org\/10.1007\/978-3-030-19424-6_14"},{"doi-asserted-by":"crossref","unstructured":"Yu, J., Lin, Z., Yang, J., Shen, X., Lu, X., Huang, T.S.: Generative image inpainting with contextual attention. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 5505\u20135514 (2018)","key":"51_CR31","DOI":"10.1109\/CVPR.2018.00577"},{"doi-asserted-by":"crossref","unstructured":"Zaidel, D.W.: Neuropsychology of art: Neurological, Cognitive, and Evolutionary Perspectives. Psychology Press (2015)","key":"51_CR32","DOI":"10.4324\/9781315719931"}],"container-title":["Lecture Notes in Computer Science","MultiMedia Modeling"],"original-title":[],"language":"en","link":[{"URL":"http:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-030-37731-1_51","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,9]],"date-time":"2022-10-09T11:38:01Z","timestamp":1665315481000},"score":1,"resource":{"primary":{"URL":"http:\/\/link.springer.com\/10.1007\/978-3-030-37731-1_51"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2019,12,24]]},"ISBN":["9783030377304","9783030377311"],"references-count":32,"URL":"https:\/\/doi.org\/10.1007\/978-3-030-37731-1_51","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2019,12,24]]},"assertion":[{"value":"24 December 2019","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"MMM","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Multimedia Modeling","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Daejeon","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Korea (Republic of)","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2020","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"5 January 2020","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"8 January 2020","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"26","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"mmm2020","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"http:\/\/www.mmm2020.kr\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"171","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"40","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"23% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"5","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Of the 171 submissions, 46 were accepted as poster papers; of the 49 special session paper submissions, 28 were accepted for oral presentation and 8 for poster presentation; 9 demo papers and 10 VBS papers were also accepted.","order":10,"name":"additional_info_on_review_process","label":"Additional Info on Review Process","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}