{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2026,2,21]],"date-time":"2026-02-21T18:54:58Z","timestamp":1771700098704,"version":"3.50.1"},"reference-count":43,"publisher":"Springer Science and Business Media LLC","issue":"2","license":[{"start":{"date-parts":[[2024,3,21]],"date-time":"2024-03-21T00:00:00Z","timestamp":1710979200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2024,3,21]],"date-time":"2024-03-21T00:00:00Z","timestamp":1710979200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"funder":[{"DOI":"10.13039\/501100011347","name":"State Key Laboratory of Software Development Environment","doi-asserted-by":"publisher","award":["SKLSDE-2020ZX-02"],"award-info":[{"award-number":["SKLSDE-2020ZX-02"]}],"id":[{"id":"10.13039\/501100011347","id-type":"DOI","asserted-by":"publisher"}]}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":["Multimedia Systems"],"published-print":{"date-parts":[[2024,4]]},"DOI":"10.1007\/s00530-024-01287-y","type":"journal-article","created":{"date-parts":[[2024,3,21]],"date-time":"2024-03-21T11:02:22Z","timestamp":1711018942000},"update-policy":"https:\/\/doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":2,"title":["Boundary-aware GAN for multiple overlapping objects in layout-to-image generation"],"prefix":"10.1007","volume":"30","author":[{"given":"Fengnan","family":"Quan","sequence":"first","affiliation":[]},{"given":"Bo","family":"Lang","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2024,3,21]]},"reference":[{"key":"1287_CR1","unstructured":"Radford, A., Metz, L., Chintala, S.: Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. ICLR (2016)"},{"key":"1287_CR2","unstructured":"Han, Z., Goodfellow, I., Metaxas, D., Odena, A.: Self-Attention Generative Adversarial Networks (2018)"},{"key":"1287_CR3","doi-asserted-by":"publisher","unstructured":"Miyato, T., Kataoka, T., Koyama, M., Yoshida, Y.: Spectral Normalization for Generative Adversarial Networks. https:\/\/doi.org\/10.48550\/arXiv.1802.05957 (2018)","DOI":"10.48550\/arXiv.1802.05957"},{"key":"1287_CR4","unstructured":"Brock, A., Donahue, J., Simonyan, K.: Large Scale GAN Training for High Fidelity Natural Image Synthesis. ICLR (2019)"},{"key":"1287_CR5","unstructured":"Miyato, T., Koyama, M.: CGANs with Projection Discriminator. ICLR (2018)"},{"key":"1287_CR6","unstructured":"Karras, T., Aila, T., Laine, S., Lehtinen, J.: Progressive Growing of GANs for Improved Quality, Stability, and Variation. ICLR (2018)"},{"key":"1287_CR7","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aila, T.: A Style-Based Generator Architecture for Generative Adversarial Networks. TPAMI (2021)","DOI":"10.1109\/TPAMI.2020.2970919"},{"key":"1287_CR8","unstructured":"Goodfellow, I.J., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative Adversarial Networks. NeurIPS (2014)"},{"key":"1287_CR9","doi-asserted-by":"crossref","unstructured":"Karras, T., Laine, S., Aittala, M., Hellsten, J., Lehtinen, J., Aila, T.: Analyzing and Improving the Image Quality of Stylegan. CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00813"},{"key":"1287_CR10","doi-asserted-by":"crossref","unstructured":"Johnson, J., Gupta, A., Fei-Fei, L.: Image Generation from Scene Graphs. CVPR (2018)","DOI":"10.1109\/CVPR.2018.00133"},{"key":"1287_CR11","doi-asserted-by":"crossref","unstructured":"Ashual, O., Wolf, L.: Specifying Object Attributes and Relations in Interactive Scene Generation. ICCV (2019)","DOI":"10.1109\/ICCV.2019.00466"},{"key":"1287_CR12","unstructured":"S. Reed, Z. Akata, X. Yan, L. Logeswaran, B. Schiele, H. Lee, Generative adversarial text to image synthesis, ICML, (2016)."},{"key":"1287_CR13","doi-asserted-by":"crossref","unstructured":"Park, T., Liu, M.Y., Wang, T.C., Zhu, J.Y.: Semantic Image Synthesis with Spatially-Adaptive Normalization, CVPR (2019)","DOI":"10.1109\/CVPR.2019.00244"},{"key":"1287_CR14","doi-asserted-by":"crossref","unstructured":"Zhao, B., Meng, L., Yin, W., Sigal, L.: Image Generation from Layout. CVPR, (2019)","DOI":"10.1109\/CVPR.2019.00878"},{"key":"1287_CR15","doi-asserted-by":"crossref","unstructured":"Sun, W., Wu, T.: Image Synthesis from Reconfigurable Layout and Style. ICCV, (2019)","DOI":"10.1109\/ICCV.2019.01063"},{"key":"1287_CR16","doi-asserted-by":"crossref","unstructured":"Sun, W., Wu, T.: Learning Layout and Style Reconfigurable GANs for Controllable Image Synthesis. TPAMI (2021)","DOI":"10.1109\/TPAMI.2021.3078577"},{"key":"1287_CR17","doi-asserted-by":"crossref","unstructured":"Sylvain, T., Zhang, P., Bengio, Y., Hjelm, R.D., Sharma, S.: Object-Centric Image Generation from Layouts. AAAI (2021)","DOI":"10.1609\/aaai.v35i3.16368"},{"key":"1287_CR18","doi-asserted-by":"crossref","unstructured":"He, S., Liao, W., Yang, M.Y., Yang, Y., Song, Y.Z., Rosenhahn, B., Xiang, T.: Context-Aware Layout to Image Generation with Enhanced Object Appearance. CVPR (2021)","DOI":"10.1109\/CVPR46437.2021.01480"},{"key":"1287_CR19","doi-asserted-by":"crossref","unstructured":"Wang, B., Wu, T., Zhu, M., Du, P.: Interactive Image Synthesis with Panoptic Layout Generation. CVPR (2022)","DOI":"10.1109\/CVPR52688.2022.00763"},{"key":"1287_CR20","unstructured":"Heusel, M., Ramsauer, H., Unterthiner, T., Nessler, B., Hochreiter, S.: GANs Trained by a Two Time-Scale Update Rule Converge to a Local Nash Equilibrium. NeurIPS (2017)"},{"key":"1287_CR21","unstructured":"Odena, A., Olah, C., Shlens, J.: Conditional Image Synthesis with Auxiliary Classifier Gans. ICML (2017)"},{"key":"1287_CR22","unstructured":"Kim, T., Cha, M., Kim, H., Lee, J.K., Kim, J.: Learning to Discover Cross-Domain Relations with Generative Adversarial Networks. ICML (2017)"},{"key":"1287_CR23","doi-asserted-by":"crossref","unstructured":"Zhu, J.-Y., Park, T., Isola, P., Efros, A.A.: Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. CVPR (2017)","DOI":"10.1109\/ICCV.2017.244"},{"key":"1287_CR24","doi-asserted-by":"crossref","unstructured":"Huang, X., Liu, M.Y., Belongie, S., Kautz, J.: Multimodal Unsupervised Image-to-Image Translation. ECCV (2018)","DOI":"10.1007\/978-3-030-01219-9_11"},{"key":"1287_CR25","doi-asserted-by":"crossref","unstructured":"Zhang, H., Xu, T., Li, H., Zhang, S., Wang, X., Huang, X., Metaxas, D.: StackGAN: Text to Photo-realistic Image Synthesis with Stacked Generative Adversarial Networks. ICCV (2017)","DOI":"10.1109\/ICCV.2017.629"},{"key":"1287_CR26","doi-asserted-by":"publisher","DOI":"10.1016\/j.image.2022.116728","volume":"106","author":"F Quan","year":"2022","unstructured":"Quan, F., Lang, B., Liu, Y.: ARRPNGAN: text-to-image GAN with attention regularization and region proposal networks. Signal Process. Image Commun. 106, 116728 (2022)","journal-title":"Signal Process. Image Commun."},{"key":"1287_CR27","doi-asserted-by":"crossref","unstructured":"Isola, P., Zhu, J.-Y., Zhou, T., Efros, A.A.: Image-to-Image Translation with Conditional Adversarial Networks. CVPR (2017)","DOI":"10.1109\/CVPR.2017.632"},{"key":"1287_CR28","unstructured":"Liu, X., Shao, J., Yin, G., Wang, X., Li, H.: Learning to Predict Layout-to-Image Conditional Convolutions for Semantic Image Synthesis. NeurIPS (2019)"},{"key":"1287_CR29","doi-asserted-by":"crossref","unstructured":"Li, Y., Cheng, Y., Gan, Z., Yu, L., Wang, L., Liu, J.: BachGAN: High-Resolution Image Synthesis from Salient Object Layout. CVPR (2020)","DOI":"10.1109\/CVPR42600.2020.00839"},{"key":"1287_CR30","doi-asserted-by":"crossref","unstructured":"Ma, K., Zhao, B., Sigal, L.: Attribute-Guided Image Generation from Layout. BMVC (2020)","DOI":"10.1109\/CVPR.2019.00878"},{"key":"1287_CR31","doi-asserted-by":"crossref","unstructured":"Hong, S., Yang, D., Choi, J., Lee, H.: Inferring Semantic Layout for Hierarchical Text-to-Image Synthesis. CVPR (2018)","DOI":"10.1109\/CVPR.2018.00833"},{"key":"1287_CR32","unstructured":"Ioffe S., Szegedy, C.: Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. ICML (2015)"},{"key":"1287_CR33","unstructured":"Dumoulin, V., Shlens, J., Kudlur, M.: A Learned Representation for Artistic Style. ICLR (2017)"},{"key":"1287_CR34","unstructured":"Kingma, D.P., Welling, M.: Auto-Encoding Variational Bayes. ICLR, pp. 1\u201314 (2014)"},{"key":"1287_CR35","doi-asserted-by":"crossref","unstructured":"Caesar, H., Uijlings, J., Ferrari, V.: COCO-Stuff: Thing and Stuff Classes in Context. CVPR, (2018)","DOI":"10.1109\/CVPR.2018.00132"},{"key":"1287_CR36","doi-asserted-by":"crossref","unstructured":"Krishna, R., Zhu, Y., Groth, O., Johnson, J., Hata, K., Kravitz, J., Chen, S., Kalantidis, Y., Li, L.J., Shamma, D.A., Bernstein, M.S., Fei-Fei, L.: Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations. IJCV, pp. 123 (2017)","DOI":"10.1007\/s11263-016-0981-7"},{"key":"1287_CR37","doi-asserted-by":"crossref","unstructured":"Lin, T., Zitnick, C.L., Doll, P.: Microsoft COCO: Common Objects in Context. ECCV, pp. 1\u201315 (2014)","DOI":"10.1007\/978-3-319-10602-1_48"},{"key":"1287_CR38","unstructured":"Salimans, T., Goodfellow, I., Zaremba, W., Cheung, V., Radford, A., Chen, X.: Improved Techniques for Training GANs. NIPS (2016)"},{"key":"1287_CR39","doi-asserted-by":"publisher","unstructured":"Liu, S., Wei, Y., Lu, J., Zhou, J.: An Improved Evaluation Framework for Generative Adversarial Networks (2018). https:\/\/doi.org\/10.48550\/arXiv.1803.07474","DOI":"10.48550\/arXiv.1803.07474"},{"key":"1287_CR40","unstructured":"Paszke, A., Gross, S., Massa, F., Lerer, A., Bradbury, J., Chanan, G., Killeen, T., Lin, Z., Gimelshein, N., Antiga, L., Desmaison, A., K\u00f6pf, A., Yang, E., DeVito, Z., Raison, M., Tejani, A., Chilamkurthy, S., Steiner, B., Fang, L., Bai, J., Chintala, S.: PyTorch: an imperative style, high-performance deep learning library. NeurIPS (2019)"},{"key":"1287_CR41","doi-asserted-by":"publisher","first-page":"273","DOI":"10.1007\/BF00994018","volume":"20","author":"C Cortes","year":"1995","unstructured":"Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20, 273\u2013297 (1995)","journal-title":"Mach. Learn."},{"key":"1287_CR42","unstructured":"Kingma, D.P., Ba, J.: Adam: A Method for Stochastic Optimization. ICLR, pp. 1\u201315 (2014)"},{"key":"1287_CR43","doi-asserted-by":"crossref","unstructured":"He, K., Zhang, X, Ren, S., Sun, J.: Deep Residual Learning for Image Recognition. CVPR (2016)","DOI":"10.1109\/CVPR.2016.90"}],"container-title":["Multimedia Systems"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01287-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/article\/10.1007\/s00530-024-01287-y\/fulltext.html","content-type":"text\/html","content-version":"vor","intended-application":"text-mining"},{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/s00530-024-01287-y.pdf","content-type":"application\/pdf","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2024,4,12]],"date-time":"2024-04-12T13:10:56Z","timestamp":1712927456000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/s00530-024-01287-y"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2024,3,21]]},"references-count":43,"journal-issue":{"issue":"2","published-print":{"date-parts":[[2024,4]]}},"alternative-id":["1287"],"URL":"https:\/\/doi.org\/10.1007\/s00530-024-01287-y","relation":{},"ISSN":["0942-4962","1432-1882"],"issn-type":[{"value":"0942-4962","type":"print"},{"value":"1432-1882","type":"electronic"}],"subject":[],"published":{"date-parts":[[2024,3,21]]},"assertion":[{"value":"11 July 2023","order":1,"name":"received","label":"Received","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"8 February 2024","order":2,"name":"accepted","label":"Accepted","group":{"name":"ArticleHistory","label":"Article History"}},{"value":"21 March 2024","order":3,"name":"first_online","label":"First Online","group":{"name":"ArticleHistory","label":"Article History"}},{"order":1,"name":"Ethics","group":{"name":"EthicsHeading","label":"Declarations"}},{"value":"The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.","order":2,"name":"Ethics","group":{"name":"EthicsHeading","label":"Conflict of interest"}}],"article-number":"88"}}