{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2025,11,5]],"date-time":"2025-11-05T21:13:23Z","timestamp":1762377203529,"version":"3.40.3"},"publisher-location":"Cham","reference-count":25,"publisher":"Springer International Publishing","isbn-type":[{"type":"print","value":"9783031133237"},{"type":"electronic","value":"9783031133244"}],"license":[{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"},{"start":{"date-parts":[[2022,1,1]],"date-time":"2022-01-01T00:00:00Z","timestamp":1640995200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springer.com\/tdm"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2022]]},"DOI":"10.1007\/978-3-031-13324-4_17","type":"book-chapter","created":{"date-parts":[[2022,8,3]],"date-time":"2022-08-03T20:21:50Z","timestamp":1659558110000},"page":"191-202","update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":5,"title":["StyleTrendGAN: A Deep Learning Generative Framework for\u00a0Fashion Bag Generation"],"prefix":"10.1007","author":[{"given":"Laura","family":"Della Sciucca","sequence":"first","affiliation":[]},{"given":"Emanuele","family":"Balloni","sequence":"additional","affiliation":[]},{"given":"Marco","family":"Mameli","sequence":"additional","affiliation":[]},{"given":"Emanuele","family":"Frontoni","sequence":"additional","affiliation":[]},{"given":"Primo","family":"Zingaretti","sequence":"additional","affiliation":[]},{"given":"Marina","family":"Paolanti","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2022,8,4]]},"reference":[{"key":"17_CR1","doi-asserted-by":"crossref","unstructured":"Al-Halah, Z., Stiefelhagen, R., Grauman, K.: Fashion forward: forecasting visual style in fashion. arXiv:1705.06394 (2017)","DOI":"10.1109\/ICCV.2017.50"},{"key":"17_CR2","unstructured":"Bi\u0144kowski, M., Sutherland, D.J., Arbel, M., Gretton, A.: Demystifying MMD GANs. arXiv:1801.01401 (2018)"},{"key":"17_CR3","doi-asserted-by":"crossref","unstructured":"Dong, H., Liang, X., Wang, B., Lai, H., Zhu, J., Yin, J.: Towards multi-pose guided virtual try-on network. arXiv:1902.11026 (2019)","DOI":"10.1109\/ICCV.2019.00912"},{"issue":"2","key":"17_CR4","first-page":"3","volume":"14","author":"C Gekombe","year":"2019","unstructured":"Gekombe, C., Tumsifu, E., Jani, D.: Social media use among small and medium enterprises: a case of fashion industry growth. Univ. Dar es Salaam Libr. J. 14(2), 3\u201318 (2019)","journal-title":"Univ. Dar es Salaam Libr. J."},{"key":"17_CR5","unstructured":"Goodfellow, I., et al.: Generative adversarial nets. In: Ghahramani, Z., Welling, M., Cortes, C., Lawrence, N., Weinberger, K.Q. (eds.) Advances in Neural Information Processing Systems. vol. 27. Curran Associates, Inc. (2014). https:\/\/proceedings.neurips.cc\/paper\/2014\/file\/5ca3e9b122f61f8f06494c97b1afccf3-Paper.pdf"},{"key":"17_CR6","doi-asserted-by":"crossref","unstructured":"Han, X., Wu, Z., Wu, Z., Yu, R., Davis, L.S.: VITON: an image-based virtual try-on network. CoRR arXiv:1711.08447 (2017)","DOI":"10.1109\/CVPR.2018.00787"},{"key":"17_CR7","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Klambauer, G., Hochreiter, S.: GANs trained by a two time-scale update rule converge to a Nash equilibrium. CoRR arXiv:1706.08500 (2017)"},{"key":"17_CR8","doi-asserted-by":"publisher","unstructured":"Hsiao, W.L., Katsman, I., Wu, C.Y., Parikh, D., Grauman, K.: Fashion++: minimal edits for outfit improvement, pp. 5046\u20135055 (2019). https:\/\/doi.org\/10.1109\/ICCV.2019.00515","DOI":"10.1109\/ICCV.2019.00515"},{"key":"17_CR9","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.Y., Zhou, T., Efros, A.A.: Image-to-image translation with conditional adversarial networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1125\u20131134 (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"17_CR10","doi-asserted-by":"crossref","unstructured":"Jetchev, N., Bergmann, U.: The conditional analogy GAN: swapping fashion articles on people images. arXiv:1709.04695 (2017)","DOI":"10.1109\/ICCVW.2017.269"},{"key":"17_CR11","doi-asserted-by":"publisher","unstructured":"Jiang, S., Li, J., Fu, Y.: Deep learning for fashion style generation. IEEE Trans. Neural Netw. Learn. Syst., 1\u201313 (2021). https:\/\/doi.org\/10.1109\/TNNLS.2021.3057892","DOI":"10.1109\/TNNLS.2021.3057892"},{"key":"17_CR12","doi-asserted-by":"publisher","first-page":"156","DOI":"10.1016\/j.neucom.2019.03.011","volume":"341","author":"L Liu","year":"2019","unstructured":"Liu, L., Zhang, H., Ji, Y., Wu, Q.M.J.: Toward AI fashion design: an attribute-GAN model for clothing match. Neurocomputing 341, 156\u2013167 (2019). https:\/\/doi.org\/10.1016\/j.neucom.2019.03.011","journal-title":"Neurocomputing"},{"key":"17_CR13","unstructured":"Loshchilov, I., Hutter, F.: SGDR: stochastic gradient descent with warm restarts arXiv:1608.03983 (2016)"},{"key":"17_CR14","doi-asserted-by":"publisher","first-page":"1545","DOI":"10.1109\/ACCESS.2021.3137893","volume":"10","author":"M Mameli","year":"2021","unstructured":"Mameli, M., Paolanti, M., Pietrini, R., Pazzaglia, G., Frontoni, E., Zingaretti, P.: Deep learning approaches for fashion knowledge extraction from social media: a review. IEEE Access 10, 1545\u20131576 (2021). https:\/\/doi.org\/10.1109\/ACCESS.2021.3137893","journal-title":"IEEE Access"},{"key":"17_CR15","doi-asserted-by":"publisher","unstructured":"Namboodiri, R., Singla, K., Kulkarni, V.: GAN based try-on system: improving CAGAN towards commercial viability. In: 2021 12th International Conference on Computing Communication and Networking Technologies (ICCCNT), pp. 1\u20136 (2021). https:\/\/doi.org\/10.1109\/ICCCNT51525.2021.9579703","DOI":"10.1109\/ICCCNT51525.2021.9579703"},{"key":"17_CR16","doi-asserted-by":"crossref","unstructured":"Pandey, N., Savakis, A.E.: Poly-GAN: multi-conditioned GAN for fashion synthesis. Neurocomputing 414, 356\u2013364 (2020). CoRR arXiv:1909.02165","DOI":"10.1016\/j.neucom.2020.07.092"},{"key":"17_CR17","unstructured":"Poma, X.S., Riba, E., Sappa, A.: Dense extreme inception network: towards a robust CNN model for edge detection. In: Proceedings of the IEEE\/CVF Winter Conference on Applications of Computer Vision, pp. 1923\u20131932 (2020)"},{"key":"17_CR18","unstructured":"Raffiee, A.H., Sollami, M.: GarmentGAN: photo-realistic adversarial fashion transfer. arXiv:2003.01894 (2020)"},{"key":"17_CR19","unstructured":"Salimans, T., et al.: Improved techniques for training GANs. In: Lee, D., Sugiyama, M., Luxburg, U., Guyon, I., Garnett, R. (eds.) Advances in Neural Information Processing Systems, vol. 29. Curran Associates, Inc. (2016). https:\/\/proceedings.neurips.cc\/paper\/2016\/file\/8a3363abe792db2d8761d6403605aeb7-Paper.pdf"},{"key":"17_CR20","doi-asserted-by":"publisher","unstructured":"Shastri, H., Lodhavia, D., Purohit, P., Kaoshik, R., Batra, N.: Vastr-GAN: versatile apparel synthesised from text using a robust generative adversarial network. In: 5th Joint International Conference on Data Science and Management of Data (9th ACM IKDD CODS and 27th COMAD), pp. 222\u2013226. Association for Computing Machinery, New York (2022). https:\/\/doi.org\/10.1145\/3493700.3493721","DOI":"10.1145\/3493700.3493721"},{"key":"17_CR21","doi-asserted-by":"crossref","unstructured":"Xian, W., Sangkloy, P., Lu, J., Fang, C., Yu, F., Hays, J.: TextureGAN: controlling deep image synthesis with texture patches. CoRR arXiv:1706.02823 (2017)","DOI":"10.1109\/CVPR.2018.00882"},{"key":"17_CR22","doi-asserted-by":"crossref","unstructured":"Xie, S., Tu, Z.: Holistically-nested edge detection. arXiv:1504.06375 (2015)","DOI":"10.1109\/ICCV.2015.164"},{"key":"17_CR23","unstructured":"Yildirim, G., Seward, C., Bergmann, U.: Disentangling multiple conditional inputs in GANs. arXiv:1806.07819 (2018)"},{"key":"17_CR24","doi-asserted-by":"publisher","unstructured":"Yirui, C., Liu, Q., Gao, C., Su, Z.: FashionGAN: display your fashion design using conditional generative adversarial nets. In: Computer Graphics Forum, vol. 37, pp. 109\u2013119 (2018). https:\/\/doi.org\/10.1111\/cgf.13552","DOI":"10.1111\/cgf.13552"},{"key":"17_CR25","doi-asserted-by":"crossref","unstructured":"Zhu, S., Fidler, S., Urtasun, R., Lin, D., Loy, C.C.: Be your own Prada: fashion synthesis with structural coherence. CoRR arXiv:1710.07346 (2017)","DOI":"10.1109\/ICCV.2017.186"}],"container-title":["Lecture Notes in Computer Science","Image Analysis and Processing. ICIAP 2022 Workshops"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-13324-4_17","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2022,10,22]],"date-time":"2022-10-22T19:06:16Z","timestamp":1666465576000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-13324-4_17"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2022]]},"ISBN":["9783031133237","9783031133244"],"references-count":25,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-13324-4_17","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2022]]},"assertion":[{"value":"4 August 2022","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"ICIAP","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Image Analysis and Processing","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Lecce","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Italy","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 May 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"27 May 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"21","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"iciap2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/www.iciap2021.org\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Microsoft","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"307","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"168","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"0","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"55% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"4","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"Yes","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}